Abstract
A CNN+LSTM (Convolutional Neural Network + Long Short-Term Memory) based deep hybrid neural network was established for the citywide daily PM2.5 prediction in South Korea. The structural hyperparameters of the CNN+LSTM model were determined through comprehensive sensitivity tests. The input features were obtained from the ground observations and GFS forecast. The performance of CNN+LSTM was evaluated by comparison with PM2.5 observations and with the 3-D CTM (three-dimensional chemistry transport model)-predicted PM2.5. The newly developed hybrid model estimated more accurate ambient levels of PM2.5 compared to the 3-D CTM. For example, the error and bias of the CNN+LSTM prediction were 1.51 and 6.46 times smaller than those by 3D-CTM simulation. In addition, based on IOA (Index of Agreement), the accuracy of CNN+LSTM prediction was 1.10–1.18 times higher than the 3-D CTM-based prediction. The importance of input features was indirectly investigated by sequential perturbing input variables. The most important meteorological and atmospheric environmental features were geopotential height and previous day PM2.5. The obstacles of the current CNN+LSTM-based PM2.5 prediction were also discussed. The promising result of this study indicates that DNN-based models can be utilized as an effective tool for air quality prediction.
1. Introduction
PM2.5 (particulate matter with an aerodynamic diameter of ≤2.5 μm) is a very harmful air pollutant. Several epidemiological studies have evaluated the harmfulness of this pollutant [1,2]. In South Korea, PM2.5 is one of the most concerning air pollutants. During the cold seasons, high PM2.5 events frequently occur. To prevent public damage caused by high PM2.5, the National Institute of Environmental Research (NIER) of South Korea has conducted air quality forecasts using a three-dimensional chemistry transport model (3-D CTM) based ensemble system since 2014.
However, the accuracy of current CTM-based air quality prediction is known to be relatively low for PM2.5. The CTM estimations have several sources of uncertainty: emission inventories, meteorological fields, initial and boundary conditions, and physico-chemical mechanisms. Great efforts have been made to enhance the predictive performance of 3-D CTMs [3,4,5,6]. Nevertheless, it is not yet clear when the air quality predictions will be sufficiently accurate. Therefore, in order to estimate more accurate ambient levels of PM2.5, it is necessary to develop a new model that makes predictions with a different operating principle from traditional CTMs.
Recently, artificial intelligence (AI) algorithms such as decision tree (DT)-based ensembles and deep neural networks have been utilized in air quality prediction [7,8,9,10,11,12,13,14,15,16,17,18,19]. For the optimal algorithm for PM2.5 prediction, several previous studies have performed comprehensive performance comparisons between AI algorithms [7,8,9,10]. As shown in Table 1, in general, it is well known that artificial neural network (ANN) algorithms tend to predict PM2.5 more accurately than the DT-based ensemble algorithms. Because of the advantages of predictive performance, ANN algorithms have shown a wide range of applications in the field of air quality prediction. For example, general deep neural network (DNNs)-based air quality models have been established to estimate the ground levels of O3 [11,12]. In addition, recurrent neural networks (RNNs), an ANN specialized in time-series prediction, have been used in PM2.5 predictions [13,14,15,16]. Moreover, convolutional neural network (CNN)-based DNNs have been developed to estimate ambient levels of air pollutants by considering the spatial distribution of meteorological and atmospheric environmental variables together [17,18,19]. Unlike the traditional CTMs, these ANNs make predictions in a data-driven manner (i.e., without consideration of sophisticated atmospheric processes). Because of their operation principles, ANN-based prediction is known to be more cost-effective than the CTM-based estimation. In this study, for more accurate daily PM2.5 prediction, a CNN+LSTM hybrid model was established.
Table 1.
Performance comparison between AI-based PM2.5 prediction models *.
ANNs have their own specialized functions depending on the type or structure. For example, RNNs have a unique structure that allows past experiences to be stored in their internal pinholes (or memory cells). Because of this capability, RNNs have shown superior performance in time-series prediction [20,21]. In particular, among the RNNs, long short-term memory (LSTM) cells have shown a wide range of applications because of their low probability of gradient exploding and vanishing [22,23,24]. In addition, CNN is the most specialized ANN for finding the latent information in multi-dimensional data (e.g., language, voice, and image). To analyze multi-dimensional data, general DNNs require the construction of hidden nodes with extremely complex and deep structures. In contrast, CNNs can significantly reduce computation costs by sharing the convolutional kernels (or filters), and this type of ANN has also shown superior analytical capabilities for high-dimensional datasets compared to other ANNs [25,26,27]. For accuracy improvement of air quality prediction, recent studies have proposed the development of hybrid DNNs utilizing both special functions of CNN and LSTM [9,10,19].
It is well known that the concentration of PM2.5 in South Korea is under the great influence of domestic and international emission sources [28]. The East Asian atmospheric pressure pattern determines the degree of the said influences. For example, in stagnant conditions, PM2.5 emitted from domestic sources accumulates, and the concentration of PM2.5 increases. In addition, if the intensity of the westerly wind increases, air pollutants originating from China move to South Korea and the ambient levels follow suit—the levels increase. Therefore, for more accurate DNN-based PM2.5 prediction, it is necessary to utilize ANNs that can effectively analyze atmospheric pressure patterns over East Asia. Furthermore, the possibility of gradient exploding and vanishing caused by using a long-term training dataset should be minimized. For these reasons, we propose a CNN+LSTM hybrid system for daily PM2.5 prediction.
In this study, the CNN+LSTM model was optimized through comprehensive sensitivity tests. The details of the model development are described in Section 2. This section also provides detailed information on the configuration of the input features and the structure of the CNN+LSTM. Evaluation of the newly developed hybrid system is conducted through the comparisons between the CNN+LSTM-based predictions, the observations, and 3-D CTM estimations in Section 3. The discussions of the advantages and limitations of the current CNN+LSTM model are in Section 4.
2. Model Development
2.1. Dataset
We constructed 3.1-year big data for model development (May 2017 to December 2019) and prediction (November 2016 to April 2017). In this study, the data produced after 2019 were not included in the construction of the big data because of COVID-19 (coronavirus disease 2019)-induced anomalies in air quality patterns. As input variables, NCEP (National Center for Environmental Prediction) GFS (Global Forecast System) forecast, KMA (Korea Meteorological Administration), ASOS (Automated Surface Observing System) observations, and NIER AIR KOREA observations were obtained from their official websites (https://rda.ucar.edu/datasets/ds084.1/, last accessed: 11 November 2022; https://rda.ucar.edu/datasets/ds084.1/, last accessed: 11 November 2022; https://www.airkorea.or.kr/web/, last accessed: 11 November 2022). In order to represent the pressure pattern over East Asia, the latitudinal and longitudinal boundaries of GFS were set from 20° N to 52° N and 96° E to 144° E. The forecast product of GFS has a resolution of 0.25° × 0.25°, respectively (i.e., the dimension of the GFS forecast was 128 × 192). Since the time interval of the GFS forecast is three hours, we performed linear interpolation to estimate the hourly forecast dataset.
Figure 1 presents locational information of ASOS and AIR KOREA ground monitoring stations, which consist of 103 and 494 monitoring sites throughout South Korea, respectively. ASOS network provides hourly observed meteorological variables, and the AIR KOREA network provides hourly concentrations of air pollutants (refer to Table 2). Among the monitoring sites, the urban observation sites located in seven major cities (with a population of more than one million) in South Korea were selected for daily PM2.5 predictions (refer to Figure 1a–g). For the prediction of citywide PM2.5, the observations of the selected sites located in each city were averaged. Table 2 is the detailed information on input variables.
Figure 1.
Locations of KMA ASOS and NIER AIR KOREA observation sites in seven major cities: (a) Seoul, (b) Incheon, (c) Daejeon, (d) Gwangju, (e) Daegu, (f) Ulsan, and (g) Busan. The blue circle represents the ASOS site, and the red triangle represents the AIR KOREA site.
Table 2.
Input variables of CNN+LSTM hybrid PM2.5 prediction model.
Similar to this study, several previous studies have proposed the use of the integration of input features of both observation-based meteorological and air quality information for more accurate DNN-based PM2.5 predictions [13,29]. The selected input variables are closely related to the concentrations of PM2.5. For example, wind speed and amount of precipitation can directly represent the intensity of atmospheric turbulent dispersions and wet scavenging processes, and the concentrations of primary pollutants such as CO and SO2 can denote the degree of domestic anthropogenic emissions. In addition, combining input features allows DNN-based models to consider more diverse atmospheric environmental properties. Among the meteorological variables, wind speed, wind direction, pressure, and geopotential height can represent the possibility of long-range transport of air pollutants. Moreover, by co-considering meteorological and air quality variables such as O3, NO2, temperature, and relative humidity, the intensity of atmospheric oxidation processes can be reflected in the DNN-based prediction.
The input variables were normalized before feeding into the CNN+LSTM hybrid model to prevent overestimating the effect of the variables with large-scale values.
here is the normalized value of variable i; is the value of variable i; and are maximum and minimum values of variable i, respectively. The previous day’s observations and the next day’s geopotential height forecasts were normalized and merged with the next day’s PM2.5-observations.
2.2. Model Construction
The procedure of the CNN+LSTM-based daily PM2.5 prediction is summarized in Figure 2. The proposed model consists of three parts: i) feature representation, ii) data vector fusion, and iii) PM2.5 prediction. As shown in the figure, we extracted the spatial feature of the GFS forecast and temporal features of ground observations in the feature representation step. In order to represent the spatial latent features, two CNN layers were embedded. Comprehensive sensitivity tests have been made to determine the number and structure of deep hidden layers. For example, CNN-based autoencoders have been constructed en masse. We elected the type of CNN layers and their structural hyperparameters (the size and number of convolutional filters) based on the reproduction accuracy of the constructed autoencoders. Based on the result, we decided to embed two three-dimensional CNN (3-D CNN) layers to represent the spatial features. The kernel size of CNN layers was set to 3 × 3 × 3. The number of kernels for the first and second CNN layers were 32 and 24. The input dimension of the first 3-D CNN layer was set to 24 × 128 × 192 × 1 for daily PM2.5 prediction. We embedded two 3-D max pooling layers for downscaling the vector matrices extracted from the CNN layers. The pooling size was set to 2 × 2 × 2.
Figure 2.
Flowchart of the CNN+LSTM hybrid ANN-based PM2.5 prediction.
Latent features extracted through convolutional operations were subjected to the next flatten and LSTM layers for data fusion with temporal features. We injected observation-based input features (i.e., ASOS and AIR KOREA observations) into two LSTM-based hidden layers to extract the characteristic of temporal input features. The input dimension of the first LSTM layer was 24 × 14. Additionally, we determined the number of hidden nodes of the entire LSTM layers through sensitive tests. The number of the LSTM hidden nodes was 512. In the second step, the vectors of temporal and spatial latent information extracted through the first step were fused as they passed the concatenate layer. Since the number of LSTM hidden nodes was set to 512, the concatenate layer generates 1024 output vectors. The concatenate layer was connected to the final dense layer. This dense layer identifies the relationship between the outputs of concatenate layer and the true values (observed PM2.5). Table 3 shows the summary of detailed configuration information of the CNN+LSTM hybrid model.
Table 3.
Configuration of the CNN+LSTM hybrid neural network model.
To identify this complicated and non-linear relationship, we need to utilize an activation function for the final dense layer. There are several functions for the activation of DNNs. Among them, rectified linear unit (ReLU) is the most versatile because of its low computation cost [30]. As the output of ReLU ranges from 0 to ∞, its derivative is 0 or 1. Therefore, this function is free from the gradient vanishing and exploding during the back propagations. However, when the hidden node outputs a negative value, the gradient of ReLU is zero, and thus the learnable parameters (weights and biases) associated with the node cannot be adjusted or updated during the model training. In addition, if a huge number of hidden nodes output negative values, training loss can no longer be reduced (i.e., dying ReLU). To prevent the said problem, we used the leaky rectified linear unit (Leaky-ReLU) as the activation function [31]. The Leaky-ReLU expressed by
as shown in the equation, the derivative of Leaky-ReLU is or 1. In this study, we set the value of as 0.3.
2.3. CNN+LSTM Optimization
Model optimization is a process of finding the optimal combination of weight and bias matrices (learnable parameters). As mentioned previously, the structural hyperparameters of the CNN+LSTM model were determined from the massive amounts of sensitivity tests. There are two main components in DNN optimization: (i) cost function and (ii) optimizer. Their respective roles are to evaluate the accuracy of the updated learnable parameters and to discover efficient and/or stable reduction paths for the cost (or loss). Several cost functions have been utilized in the optimization of DNNs. For better predictive performance, it is very important to select an appropriate objective function corresponding to the purpose of the DNN model. Since the proposed hybrid model was the regression model, the mean squared error (MSE) was utilized as a cost function in this study. The MSE expressed as
here is the true value for th training; is input for th training; is the predicted value by the given model . In this study, adaptive moment estimation (ADAM) was used as an optimizer [32]. The ADAM is an extended stochastic gradient descent algorithm. This algorithm determines the individual adaptive learning rates for different learnable parameters by computing the first and second moments of their gradients.
To train the CNN+LSTM hybrid model, the 2.7-year data (May 2017 to December 2019) were used, as mentioned previously. We divided this dataset into two groups with ratios of 80% for model training and 20% for validation. To confirm the suitability of model optimization, the variations of MSE for model training and validation were monitored during the model training. In the early stage of model optimization, the MSEs of the training and validation decrease together by updating weights and biases. Because training data are only considered in the adjustment of weight and bias matrices, the training cost continuously decreases with the iterative update of learnable parameters. However, the validation MSE decreases until the specific update, and then it stagnates or starts to increase. The optimization of DNNs is to update the learnable parameters until this stagnant or inflection point [33]. If the DNN is properly optimized, the validation MSE at this point should be slightly higher than the training MSE. When the validation MSE is much higher than the training MSE, this represents that the learnable parameters are overturned (i.e., overfitting). On the contrary, if the training cost is bigger than the validation cost, it is called underfitting. From the perspective of improving the predictive performance and generalization of DNNs, it is crucial to minimize the possibility of overfitting and underfitting. To achieve this goal, iterative model training was performed, and then the models with the smallest gap between the training and validation cost were selected for daily PM2.5 prediction. In addition, we also compared the training and validation MSEs and root mean squared errors (RMSEs) to re-evaluate the suitability of model training. The results of this evaluation are summarized in Table 4. Based on this, we confirmed that the CNN+LSTM model was suitably trained.
Table 4.
Evaluation of the suitability of model optimization *.
2.4. 3-D CTM-Based PM2.5 Prediction
We made the comparison between the CNN+LSTM-based PM2.5 prediction and the CTM-predicted PM2.5 to evaluate the predictive performance of the CNN+LSTM hybrid neural network model. In this study, we utilized the community multiscale air quality (CMAQ) model v5.2.1 for the 3-D CTM prediction. The meteorological fields for CMAQ prediction were estimated from weather research and forecasting (WRF) v3.8.1 model simulation. Figure 3 presents the boundary of the CMAQ prediction. As shown in the figure, the CMAQ boundary covers northeast Asia. The horizontal resolution of the CMAQ simulation was set to 15 × 15 km2. The KORUS v5.0 emission inventory was used for anthropogenic emissions [34]. The biogenic emissions were acquired from the MEGAN (Model of Emissions of Gases and Aerosols from Nature) v2.1 simulation [35]. The emissions for biomass burning were acquired from FINN (Fire Inventory from National Center for Atmospheric Research, https://www.acom.ucar.edu/Data/fire/, last accesses: 11 November 2022) v1.5 [36]. The lateral boundary conditions were obtained from the official data archive of MOZART-4 (Model for Ozone and Related Chemical Tracers version 4, https://www2.acom.ucar.edu/gcm/mozart-4, last accessed: 11 November 2022), respectively [37].
Figure 3.
Boundary of the CMAQ prediction (red line) and GFS forecast (blue line). The grey triangles and dots represent the locational information of ground monitoring stations in South Korea.
2.5. Evaluation Metric
In this study, to evaluate the prediction performance of the CNN+LSTM and 3-D CTM, we introduced the following six scientific performance metrics, such as IOA (Index or Agreement), R (Pearson correlation coefficient), RMSE, MB (Mean Bias), MNGE (Mean Normalized Gross Error), and MNB (Mean Normalized Bias). These evaluation metrics were estimated as below:
here and represent the modeled and observed PM2.5, and are the averaged and .
3. Results
3.1. Model Evaluation
We performed the daily PM2.5 prediction from November 2016 to April 2017. In order to evaluate the accuracy of the newly developed hybrid model, the CNN+LSTM-based PM2.5 was compared to the observed PM2.5. The CNN+LSTM-predicted PM2.5 was also compared with that by CMAQ prediction. The results of this comparison are summarized in Table 5 and also shown in Figure 4. In the figure, the black-dashed line with an open circle represents the observed daily PM2.5; blue-dashed and red lines represent the CMAQ- and CNN+LSTM-predicted daily PM2.5; the grey shade represents the period with relatively high concentration among nationwide high PM2.5 episodes, respectively. As shown in the figure, the CNN+LSTM predictions showed better agreements with the observations compared to the CMAQ-based predictions.
Table 5.
Scientific evaluations of the CMAQ- and CNN+LSTM-based PM2.5 predictions *.
Figure 4.
Comparisons between the observed, CMAQ-predicted, and the CNN+LSTM-predicted PM2.5 at seven major cities in South Korea. Black-dashed line with an open circle represents the observed PM2.5. A blue-dashed line represents the CMAQ-predicted PM2.5. A red line represents the CNN+LSTM-predicted PM2.5. Grey shade represents the period with relatively high concentration among nationwide high PM2.5 episodes.
As shown in Figure 4 and Table 5, the CNN+LSTM predicted PM2.5 more accurately than the CMAQ. The IOA for the CNN+LSTM-based PM2.5 prediction ranged from 0.72 to 0.82. On the other hand, the CMAQ-based IOA ranged from 0.65 to 0.70. These differences in IOA demonstrate that the CNN+LSTM can produce 1.10–1.18 times better air quality information than the CMAQ. Based on the statistic metric, among seven cities, the CNN+LSTM predictions (0.81 ≤ IOA ≤ 0.82) at Seoul and Incheon showed the best agreements with the observations. The CMAQ-based predictions in these cities also showed better correlations with the observed PM2.5 (IOA = 0.70). However, the accuracies of the CNN+LSTM at Daejeon, Gwangju, and Busan (0.72 ≤ IOA ≤ 0.73) were relatively low. The CMAQ predictions showed similar trends to the CNN+LSTM-based predictions (0.65 ≤ IOA ≤ 0.66).
In addition, the CNN+LSTM-based predictions showed fewer errors and lower biases. The RMSE for CNN+LSTM predictions ranged from 10.96 μg/m3 to 14.66 μg/m3, and those by CMAQ ranged from 17.01 μg/m3 to 19.70 μg/m3. The CNN+LSTM generated 1.51 times smaller deviation compared to the CMAQ. The prediction errors made by the CNN+LSTM showed significant improvement in Seoul and Busan; the improved rates in these cities were 38.17% and 39.21%. The CNN+LSTM-based predictions also showed lower biases than the CMAQ-based estimations. The MBs for the CMAQ predictions ranged from −9.70 μg/m3 to −1.30 μg/m3. These negative MBs indicate that the CMAQ model significantly underestimates ambient levels of PM2.5. Although the CNN+LSTM-based predictions also showed negative biases (−1.64 μg/m3 ≤ MB ≤ 0.11 μg/m3), their values were 6.46 times smaller than those predicted by CMAQ.
As mentioned above, the PM2.5 prediction was conducted from November 2016 to April 2017. During this cold season, high PM2.5 events (PM2.5 ≥ 40 μg/m3) occur frequently in South Korea. There were seventeen nationwide high PM2.5 episodes. These high episodes are mainly attributable to the long-range transport of air pollutants by highly enhanced westerly wind and increased consumption of fossil fuels for domestic heating [22]. In particular, the daily PM2.5 in Seoul increased up to 85 μg/m3 on January 1 and March 20, respectively. As shown in Figure 4, the CNN+LSTM model generated relatively high errors and biases during the nationwide high-PM events. Since DNNs optimize their weight and bias matrices in a data-driven manner, the corresponding errors and biases arise from the quality of the optimization dataset. Because DNNs are black box models, accurate investigation of data quality is practically impossible.
In this study, to indirectly present the data quality, we evaluated the data imbalance. The datasets used for the optimization of the CNN+LSTM model were highly imbalanced (i.e., the majority class of the training data consisted of the data samples with a relatively low concentration of PM2.5). Since DNNs make predictions based on statistically generalized non-linear relationships between input features, it is difficult for the influence of minority-class data samples to be reflected in their estimations. The data samples with high PM2.5 contributed just 11.30% on average. Among the seven training datasets, the training data for Seoul and Gwangju included the highest number of high-PM2.5 incident samples; their contribution was 12.32% and 12.10%, respectively. In addition, Daejeon showed the lowest frequency of high-PM2.5 incident samples (9.14%). However, as mentioned above, data imbalance is only an indirect indicator of data quality. Therefore, it is logically flawed to conclude that the cause of the performance degradation of the CNN+LSTM is entirely due to the said indicator.
3.2. Importance of Input Features
It is impossible to directly comprehend the relationship between input features and predictions of the DNNs due to the complex non-linearity between the outputs of the deeply structured hidden layers. In this study, we indirectly evaluated the influence of the input variables by sequential perturbing each feature. The main idea of this method is that: if the perturbed feature is “important” (i.e., an input feature has a larger influence on performance), the accuracy of the model predictions will be greatly reduced by the perturbation. In contrast, if the perturbed input feature is unnecessary, the predictive performance of the model will be rather improved. In this estimation, IOA was used as a basis for feature importance investigation because it can represent the overall accuracy of model prediction, as the statistical metric includes an error, bias, and correlation. The feature importance was estimated as below:
here is the importance of feature ; and are accuracies of the CNN+LSTM model with and without perturbation of feature , respectively. The results of this estimation are summarized in Table 6.
Table 6.
Importance of the input features in daily PM2.5 prediction *.
As shown in Table 6, the importance of all input variables was positive. These positive values indicate that all input variables are necessary for daily PM2.5 prediction (i.e., the appropriateness of variable selection). Among the atmospheric environmental variables, the previous day’s PM2.5 and O3 showed a great influence on the next day’s PM2.5. Their importance ranged from 31.67% to 41.82% and from 3.33% to 15.73%, respectively. In particular, the previous-day-PM2.5 was the most important input variable. Among the meteorological variables, PM2.5 prediction was greatly influenced by geopotential height (from 7.92% to 25.29%) and temperature (from 3.10% to 12.87%). In particular, the importance of atmospheric pressure-related features was relatively higher compared to other meteorological variables. This result represents the importance of pressure patterns over East Asia and inside cities in the daily PM2.5 prediction in South Korea. Therefore, the spatial and temporal pressure features should be included as the input variables in the construction of the DNN-based PM2.5 model, and it is essential to embed CNN and LSTM together to extract their inherent information.
4. Conclusions
In this study, we developed a new CNN+LSTM hybrid neural network for daily PM2.5 prediction. The performance of the CNN+LSTM model was precisely evaluated by comparing the CNN+LSTM-based PM2.5 with the ground observations. In addition, the accuracy of the CNN+LSTM was compared with that of CMAQ to prove its usefulness as an air quality prediction tool. The CNN+LSTM hybrid model generated 1.51 and 6.46 times less error and lower bias on average than the CMAQ. In particular, based on IOA, the CNN+LSTM predictions were 1.10–1.18 times more accurate than the CMAQ estimations. This promising result clearly demonstrates that the DNN-based prediction model can be utilized as an effective tool for air quality prediction. In addition, a similar CNN+LSTM model can also be used in the prediction of other harmful air pollutants (e.g., NO2, SO2, and O3).
In particular, the newly developed CNN+LSTM model is more cost-effective than the 3-D CTMs. Several pre-processing steps are required to generate a 3-D CTM-based air quality forecast: simulation of the numerical weather prediction model, preparation of biogenic and anthropogenic emission inventories, acquisition of initial and boundary conditions, and chemical speciation of air pollutants. Despite the efforts of these sophisticated pre-processings, the current 3-D CTM showed lower predictive power than the CNN+LSTM (refer to Section 3.1). On the other hand, the pre-processing of the CNN+LSTM-based prediction is very simple. For PM2.5 predictions, it is only necessary to acquire the observational and forecast-based datasets from their official archives and configure input features. Nevertheless, the current CNN+LSTM model is superior to the 3-D CTM. Because of these advantages, for developing countries where it is impossible to establish and operate their own 3-D CTM-based forecast system due to financial limitations, the CNN+LSTM model developed through this study can be used as a more economical and efficient air quality forecast tool.
Although the current hybrid model can accurately predict PM2.5, its performance can be deteriorated by two obstacles: i) the amount of available data and ii) an imbalance in training data. As shown in Section 3.1, the prediction accuracy of the CNN+LSTM model was very high for medium and relatively low concentrations (PM2.5 < 40 μg/m3). In contrast, the CNN+LSTM tends to generate relatively high errors and biases for high PM2.5 episodes. This inaccuracy originates from the amount and quality of the training data. The NIER of South Korea has officially conducted ground-based PM2.5 monitoring since 2016. Although the PM2.5 observations have been accumulated for seven years up to date, the amount of useful data for the CNN+LSTM development is less than four years, excluding the COVID-19 period. In addition, the data samples with high PM2.5 only contributed 11.30% of the total available observations. Since DNNs statistically determine their predictions, it is very difficult to reflect the influence of these minority classes in the model estimation. In other words, from the perspective of the DNN generalization, the high PM2.5 samples are likely to be recognized as noise signals during the model training. We believe that the current CNN+LSTM hybrid model can be more accurate by acquiring more data samples of minority classes through more observations.
Author Contributions
Conceptualization, H.S.K.; methodology, H.S.K., K.M.H. and J.Y.; software, H.S.K. and J.K.; validation, K.K. and H.K.; writing—original draft preparation, H.S.K. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2022R1I1A1A01066083).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Acknowledgments
We obtained KMA ASOS, NIER AIR KOREA, and NCEP GFS datasets from respective official data archives. This work appreciates the said institutions for the provision of useful data.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Dorkery, D.W.; Schwartz, J.; Spengler, J.D. Air pollution and daily mortality: Associations with particles and acid aerosols. Environ. Res. 1992, 59, 362–373. [Google Scholar] [CrossRef] [PubMed]
- Pope III, C.A.; Dorkey, D.W. Health effects of fine particulate air pollution: Lines that connect. J. Air Waste Manag. Assoc. 2006, 56, 709–742. [Google Scholar] [CrossRef] [PubMed]
- Berge, E.; Huang, H.-C.; Chang, J.; Liu, T.-H. A study of the importance of initial conditions for photochemical oxidant modeling. J. Geophys. Res.-Atmos. 2001, 106, 1347–1363. [Google Scholar] [CrossRef]
- Liu, T.-H.; Jeng, F.-T.; Huang, H.-C.; Berger, E.; Chang, J.S. Influences of initial conditions and boundary conditions on regional and urban scale Eulerian air quality transport model simulations. Chem.-Glob. Change Sci. 2001, 3, 175–183. [Google Scholar] [CrossRef]
- Holloway, T.; Spak, S.N.; Barker, D.; Bretl, M.; Moberg, C.; Hayhoe, K.; van Dorn, J.; Wuebbles, D. Change in ozone air pollution over Chicago associated with global climate change. J. Geophys. Res.-Atmos. 2008, 113, D22306. [Google Scholar] [CrossRef]
- Han, K.M.; Lee, C.K.; Lee, J.; Kim, J.; Song, C.H. A comparison study between model-predicted and OMI-retrieved tropospheric NO2 columns over the Korean peninsula. Atmos. Environ. 2011, 45, 2962–2971. [Google Scholar] [CrossRef]
- Joharestani, Z.M.; Cao, C.; Ni, X.; Bashir, B.; Talebiesfandarani, S. PM2.5 prediction based on random forest, XGBoost, and deep learning using multisource remote sensing data. Atmosphere 2019, 10, 373. [Google Scholar] [CrossRef]
- Karimian, H.; Li, Q.; Wu, C.; Qi, Y.; Mo, Y.; Chen, G.; Zhang, X.; Sachdeva, S. Evaluation of different machine learning approaches to forecasting PM2.5 mass concentrations. Aerosol Air Qual. Res. 2019, 19, 1400–1410. [Google Scholar] [CrossRef]
- Li, T.; Hua, M.; Wu, X. A hybrid CNN-LSTM model for forecasting particulate matter (PM2.5). IEEE Access 2020, 8, 26933–26940. [Google Scholar] [CrossRef]
- Park, U.; Ma, J.; Ryu, U.; Ryom, K.; Juhyok, U.; Park, K.; Park, C. Deep learning-based PM2.5 prediction considering the spatiotemporal correlations: A case study of Beijing, China. Sci. Total Environ. 2020, 699, 133561. [Google Scholar] [CrossRef]
- Al-Alawi, S.M.; Abdul-Wahab, S.A.; Bakheit, C.S. Combining principal component regression and artificial neural networks for more accurate predictions of ground-level ozone. Environ. Model. Softw. 2008, 23, 396–403. [Google Scholar] [CrossRef]
- Feng, y.; Zhang, W.; Sun, D.; Zhang, L. Ozone concentration forecast method based on genetic algorithm optimized back propagation neural networks and support vector machine data classification. Atmos. Environ. 2011, 45, 1979–1985. [Google Scholar] [CrossRef]
- Kim, H.S.; Park, I.; Song, C.H.; Lee, K.; Yun, J.W.; Kim, H.K.; Jeon, M.; Lee, J.; Han, K.M. Development of a daily PM10 and PM2.5 prediction system using a deep long short-term memory neural network model. Atmos. Chem. Phys. 2019, 19, 12935–12951. [Google Scholar] [CrossRef]
- Zhao, j.; Deng, F.; Cai, Y.; Chen, J. Long short-term memory—Fully connected (LSTM-FC) neural network for PM2.5 concentration prediction. Chemosphere 2019, 220, 486–492. [Google Scholar] [CrossRef] [PubMed]
- Chang-Hoi, H.; Park, I.; Oh, H.-R.; Gim, H.J.; Hur, S.K.; Kim, J.; Choi, D.-R. Development of a PM2.5 prediction model using a recurrent neural network algorithm for the Seoul metropolitan area, Republic of Korea. Atmos. Environ. 2021, 245, 118021. [Google Scholar] [CrossRef]
- Muruganandam, N.S.; Arumugam, U. Seminal stacked long short-term memory (SS-LSTM) model for forecasting particulate matter (PM2.5 and PM10). Atmosphere 2022, 13, 1726. [Google Scholar] [CrossRef]
- Eslami, E.; Choi, Y.; Lops, Y.; Sayeed, A. A real-time hourly ozone prediction system using deep convolutional neural network. Neural Comput. Appl. 2020, 32, 8783–8797. [Google Scholar] [CrossRef]
- Park, Y.; Kwon, B.; Heo, J.; Hu, X.; Liu, Y.; Moon, T. Estimating PM2.5 concentration of the conterminous United States via interpretable convolutional neural networks. Environ. Pollut. 2020, 256, 113395. [Google Scholar] [CrossRef]
- Dai, H.; Huang, G.; Wang, J.; Zeng, H.; Zhou, F. Prediction of air pollutant concentration based on one-dimensional multi-scale CNN-LSTM considering spatial-temporal characteristics: A case study of Xi’an, China. Atmosphere 2021, 12, 1626. [Google Scholar] [CrossRef]
- Connor, J.T.; Martin, R.D.; Atlas, L.E. Recurrent neural networks and robust time series prediction. IEEE Trans. Neural Netw. 1994, 5, 240–254. [Google Scholar] [CrossRef]
- Sezer, O.B.; Gudelek, M.U.; Ozbayoglu, A.M. Financial time series forecasting with deep learning: A systematic literature review: 2005–2019. Appl. Soft. Comput. 2020, 90, 106181. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1977, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Graves, A. Long short-term memory. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin, Germany, 2012; Volume 385, pp. 37–45. [Google Scholar]
- Sak, H.; Senior, A.; Beaufays, F. Long short-term memory based recurrent neural network architectures for large vocabulary speech recognition. arXiv 2014, arXiv:1402.1128. [Google Scholar]
- Albawi, S.; Mohammed, T.A.; Al-Zawi, S. Understanding of a convolutional neural network. In Proceedings of the 2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey, 21–23 August 2017. [Google Scholar]
- Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
- Rawat, W.; Wang, Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef]
- Kim, H.C.; Kim, E.; Bae, C.; Cho, J.H.; Kim, B.-U.; Kim, S. Regional contributions to particulate matter concentration in the Seoul metropolitan area, South Korea: Seasonal variation and sensitivity to meteorology and emissions inventory. Atmos. Chem. Phys. 2017, 17, 10315–10332. [Google Scholar] [CrossRef]
- Sayeed, A.; Lops, Y.; Choi, Y.; Jung, J.; Salman, A.K. Bias correcting and extending the PM forecast by CMAQ up to 7 days using deep convolutional neural networks. Atmos. Environ. 2021, 253, 118376. [Google Scholar] [CrossRef]
- Nair, V.; Hinton, G.E. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010. [Google Scholar]
- Mass, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
- Kingma, D.; Ba, J. A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 3–8 May 2015. [Google Scholar]
- Mahsereci, M.; Ballers, L.; Lassner, C.; Henning, P. Early stopping without a validation set. arXiv 2017, arXiv:1703.09580. [Google Scholar]
- Woo, J.-H.; Kim, Y.; Kim, H.-K.; Choi, K.-C.; Eum, J.-H.; Lee, J.-B.; Lim, J.-H.; Kim, J.; Seong, M. Development of the CREATE Inventory in Support of Integrated Climate and Air Quality Modeling for Asia. Sustainability 2020, 12, 7930. [Google Scholar] [CrossRef]
- Guenther, A.; Karl, T.; Harley, P.; Wiedinmyer, C.; Palmer, P.I.; Geron, C. Estimates of global terrestrial isoprene emissions using MEGAN (Model of Emissions of Gases and Aerosols from Nature). Atmos. Chem. Phys. 2006, 6, 3181–3210. [Google Scholar] [CrossRef]
- Wiedinmyer, C.; Akagi, S.K.; Yokelson, R.J.; Emmons, L.K.; Al-Saadi, J.A.; Orlando, J.J.; Soja, A.J. The Fire INventory from NCAR (FINN): A high resolution global model to estimate the emissions from open burning. Geosci. Model Dev. 2011, 4, 625–641. [Google Scholar] [CrossRef]
- Emmons, L.K.; Walters, S.; Hess, P.G.; Lamarque, J.-F.; Pfister, G.G.; Fillmore, D.; Granier, C.; Guenther, A.; Kinnison, D.; Laepple, T.; et al. Description and evaluation of the Model for Ozone and Related chemical Tracers, version 4 (MOZART-4). Geosci. Model Dev. 2010, 3, 43–67. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).