Next Article in Journal
Research on Rapid Identification Technology of Sand and Dust Characteristic Monitoring Data Based on Optimized K-Means Clustering
Previous Article in Journal
The Effect of Sea Surface Temperature on Relative Humidity and Atmospheric Visibility of a Winter Sea Fog Event over the Yellow-Bohai Sea
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Multi-Site PM2.5 Concentrations in Beijing Using CNN-Bi LSTM with CBAM

1
Faculty of Geomatics, Lanzhou Jiaotong University, Lanzhou 730070, China
2
Chinese Academy of Surveying and Mapping, Beijing 100830, China
3
National-Local Joint Engineering Research Center of Technologies and Applications for National Geographic State Monitoring, Lanzhou 730070, China
4
Gansu Provincial Engineering Laboratory for National Geographic State Monitoring, Lanzhou 730070, China
*
Author to whom correspondence should be addressed.
Atmosphere 2022, 13(10), 1719; https://doi.org/10.3390/atmos13101719
Submission received: 8 September 2022 / Revised: 8 October 2022 / Accepted: 14 October 2022 / Published: 19 October 2022
(This article belongs to the Section Air Quality)

Abstract

:
Air pollution is a growing problem and poses a challenge to people’s healthy lives. Accurate prediction of air pollutant concentrations is considered the key to air pollution warning and management. In this paper, a novel PM2.5 concentration prediction model, CBAM-CNN-Bi LSTM, is constructed by deep learning techniques based on the principles related to spatial big data. This model consists of the convolutional block attention module (CBAM), the convolutional neural network (CNN), and the bi-directional long short-term memory neural network (Bi LSTM). CBAM is applied to the extraction of feature relationships between pollutant data and meteorological data and assists in deeply obtaining the spatial distribution characteristics of PM2.5 concentrations. As the output layer, Bi LSTM obtains the variation pattern of PM2.5 concentrations from spatial data, overcomes the problem of long-term dependence on PM2.5 concentrations, and achieves the task of accurately forecasting PM2.5 concentrations at multiple sites. Based on real datasets, we perform an experimental evaluation and the results show that, in comparison to other models, CBAM-CNN-Bi LSTM improves the accuracy of PM2.5 concentration prediction. For the prediction tasks from 1 to 12 h, our proposed prediction model performs well. For the 13 to 48 h prediction task, the CBAM-CNN-Bi LSTM also achieves satisfactory results.

1. Introduction

Over the past few years, significant social concern has been generated by the increasingly serious issue of air pollution. [1]. Prediction of pollutant concentrations in the air plays an essential role in environmental management and air pollution prevention [2]. PM2.5 (particulate matter less than 2.5   μ m in diameter) is an important component of air pollutants [3]. Therefore, the prediction of PM2.5 concentration trends is considered a critical issue in the prediction of air pollutant concentrations.
Deterministic and statistical approaches can be used to forecast PM2.5 concentrations based on the features of the study methods. [4]. Deterministic methods simulate the emission, dispersion, transformation, and removal of PM2.5 through meteorological principles and statistical methods [5], thus enabling the prediction of PM2.5 concentrations. There are several representative models for pollutant concentration prediction based on deterministic methods. There are a few representative models for forecasting PM2.5 concentrations based on deterministic methods: a Community Multiscale Air Quality Modeling System (CMAQ) [6], a nested air quality prediction modeling system (NAQPMS) [7], and a Weather Research and Forecasting Model with Chemistry (WRF-Chem) [8].
Different from the deterministic methods, the statistical methods do not have complex theoretical models, and they give better predictions through the learning and analysis of historical data on pollutants. Statistical methods are mainly classified into two approaches: machine learning approaches and deep learning approaches [9]. The classical machine learning models used for the prediction of PM2.5 concentrations include Random Forest (RF) [10] models, Autoregressive Sliding Average (ARMA) Models [11], Autoregressive Integrated Moving Average (ARIMA) Models [12], Support Vector Regression (SVR) [13], and Linear Regression (LR) models [14].
Compared with machine learning methods, which suffer from slow convergence and inadequate generalization [5], deep learning is widely used in PM2.5 concentration prediction due to its ability to fit data more robustly and non-linearly [15]. The following deep learning models have been used to forecast pollution concentrations: Convolutional neural networks (CNN) [16], Back Propagation Neural Networks (BPNN) [17], Recurrent Neural Networks (RNN) [18], Gate Recurrent Units (GRU) [19], Long Short-Term Memory networks (LSTM) [20], and Bi-directional Long Short-Term Memory Neural Networks (Bi LSTM) [21], attention-based ConvLSTM (Att-ConvLSTM) [22], etc. Although the above models are widely used in PM2.5 concentration prediction due to their superiority in handling time series data, the current pollutant concentration prediction models described above have the following problem: owing to the single network model, it is limited by the feature dimension of the input data, in other words, the dimension of the hidden state is influenced by the dimension of the input data.
In order to solve the problem that the predictive power of a single deep learning network is limited, in recent years, hybrid deep learning models have been widely used in the research of pollutant concentration prediction. Hybrid deep learning models have several different network structures that allow better quantification of complex data [23], which have been used for pollutant concentration prediction, including: LSTM-FC [24], AC-LSTM [25], EEMD-GRNN [26], etc. Meanwhile, air pollution is a problem of regional dispersion with spatial dimensions [27], and there is a spatial interrelationship of air pollution between adjacent sites. However, all of the above hybrid deep learning models have focused on the prediction of pollutant concentrations at individual stations and have not considered the spatial correlation of adjacent observation sites. CNN-LSTM [15] is a recently proposed hybrid deep learning model, which can handle time series problems and has successfully interpreted the spatial distribution characteristics of air pollutant concentrations through the image analysis capability of CNN [15]. However, there are also three crucial problems with CNN-LSTM [28]. Firstly, the simple structure of CNN leads to the loss of feature information and the inability to extract deep spatial features of contaminant data [29]. Secondly, CNN-LSTM has difficulty capturing the long time-series variation between pollutant concentrations. Finally, much of the work at the present stage confirms the complex interactions between pollutant data and meteorological data [5]. CNN-LSTM has trouble obtaining complicated correlation characteristics between meteorological input and air pollution input. In view of the above, we introduced the convolutional block attention module (CBAM) to build a CNN-based prediction model: the CBAM-CNN-Bi LSTM model. The reasons are as follows.
(1)
The convolutional block attention module includes the channel attention module (CAM) and the spatial attention module (SAM) [30]. It is a simple and effective attention module that can be arbitrarily embedded into any 2D CNN model and does not consume too much of the computer’s running memory.
(2)
As the network depth is increased, convolutional neural networks degrade and lose feature information. Therefore, we introduce a spatial attention module to efficiently extract spatially relevant features of contaminant data between multiple stations [29].
(3)
The correlation characteristics between pollution data and meteorological data are not taken into account by the aforementioned prediction models. In order to optimize the prediction outcomes based on the intricate correlation characteristics of the model input data, we thoroughly analyze the prediction problems of pollution data and meteorological data at each station and present the channel attention module.
In this study, our proposed prediction model is fully taken into account to produce more accurate predictions of PM2.5 concentrations in the target city in the future, which should achieve the following goals: (1) efficiently utilizing historical pollutant concentration and meteorological big data from multiple stations; (2) accurately achieving long-term predictions of pollutant concentrations in the target city; and (3) in-depth exploration of the spatial and temporal correlation characteristics from multiple stations.
The remainder of the essay is structured as follows. The study area, the experimental data and the procedures for processing the experimental data, as well as the overall framework of the pollutant concentration prediction model, and a detailed explanation of each model component are described in Section 2. The key findings and expectations are covered in Section 3. The work is concluded in Section 4, which also suggests areas for future investigation.

2. Materials and Methods

2.1. Materials

2.1.1. Data

The experimental information in this study consists of pollutant data and meteorological data. The data was monitored by twelve air quality monitoring stations in the Beijing area from 1 March 2013 to 28 February 2017. The distribution of the twelve stations is shown in Figure 1a. The meteorological data and pollutant data selected are from the UCI website (https://archive.ics.uci.edu/ml/datasets/Beijing+Multi-Site+Air-Quality+Data, accessed on 1 April 2022). Figure 1b shows a time series plot of PM2.5 concentration data from twelve stations for the period 1–31 December 2016. The meteorological data include hourly temperature, pressure, dew point, precipitation, wind direction, and wind speed; and the pollutant data include hourly PM2.5, PM10, SO2, NO2, CO, and O3. Table 1 shows the range of values for meteorological data and pollutant data.

2.1.2. Data Preprocessing

Wind direction, as a non-numerical type of data, needs to be converted to a numerical type of data by categorical coding. The average of the data prior to and following the time of the missing value is used to fill in missing values for meteorological and pollution data. Then, in order to eliminate the effect of numerical differences on prediction accuracy, meteorological and pollutant data were converted to the range [0, 1] by the Min-Max function as below.
y = y min ( y ) max ( y ) min ( y )

2.2. Methods

2.2.1. CBAM-CNN-Bi LSTM

Deep learning is a type of machine learning that can train samples of data using unsupervised approaches in order to create deep network structures [31]. We propose a CBAM-CNN-Bi LSTM model, whose intricate structure is depicted in Figure 2, to reliably anticipate PM2.5 concentrations. First, we took advantage of CNN properties to identify key characteristics in the input PM2.5 data and to obtain the spatial dependence of all PM2.5 sites. Then, we capture the time dependence of the PM2.5 series data using the special architecture of Bi LSTM for time-series problems. In addition, we added the convolutional block attention module to the CNN to enhance training accuracy by focusing on channel and spatial information. The convolutional block attention module is classified into the channel attention module and the spatial attention module. The channel attention module enables the network to disregard the remainder and concentrate on the useful feature channels. The spatial attention module enables the network to concentrate on the nearby areas on the feature map [32]. In other words, the channel attention module enables the network to focus on the classes of factors which have a greater impact on PM2.5 concentrations, and the spatial attention module enables the network to focus on areas where there is a stronger spatial relationship between PM2.5 sites.
The CNN serves as the basis layer for the CBAM-CNN-Bi LSTM model, and its convolutional layer is used to extract features. The convolutional block attention module is embedded in the CNN, and the channel and spatial feature information in the convolutional block attention module capture is used as an input for the CNN. For time series prediction, the higher layer Bi LSTM uses the output of the CNN layer as its input. In Figure 2, the prediction model is displayed.
Meteorological data and pollutant concentrations are converted into two-dimensional matrices with time series as inputs to the prediction model. These matrices are then fed into the convolutional block attention module and the CNN network in order to obtain characteristics. Its output serves as the Bi LSTM’s input. The fully connected layer decodes the Bi LSTM’s output to get the final prediction result.

2.2.2. Convolutional Block Attention Module

Figure 3a shows the architecture of the convolutional block attention module [33], which can perform attention in both the channel dimension and the spatial dimension by concentrating on important elements while disregarding unimportant ones. There are two independent sub-modules in the convolutional block attention module, the channel attention module and the spatial attention module, and the structure of each of these two sub-modules is shown in Figure 3b,c. The characteristic map is created by the convolutional layer, and the weighting of the characteristic map is calculated along the order of the channel attention module first and then the spatial attention module. Then, the weight map and the input feature map are multiplied to carry out adaptive feature optimization learning [30]. The convolutional block attention module is designed as a simple attention block which is casually embedded into any 2D CNN model for end-to-end training and improved model representation at a lower cost [33].
As shown in Figure 3a, the convolutional block attention module combines the spatial attention and channel attention modules to infer attention weight maps and produces detailed characteristic maps. These two sub-modules together are referred to as:
F = M c ( F ) F
F = M s ( F ) F
Here, denotes element-wise multiplication, F represents the input feature map, F represents the channel-refined characteristic map, F is the refined characteristic map.
The channel attention module identifies the more important channels based on their relationship to each other. The channel attention module has two pools, M a x p o o l and A v g p o o l ( · ) . Firstly, the input features are extracted by M a x p o o l and A v g p o o l ( · ) for different high-level features. By using shared MLP (multi-layer perceptron), the two categories of high-level characteristics are then combined. Lastly, a sigmoid function activates the fused features to show the channel priority of the input characteristics. The channel attention module’s computation is done in the manner listed below.
M c ( F ) = f s i g m o i d ( M L P ( M a x P o o l ( F ) + M L P ( A v g P o o l ( F ) ) )
The spatial attention module acts as a complement to the channel attention module, and it focuses on which position of information is more important. At first, M a x p o o l and A v g p o o l ( · ) are performed on the input features. Then, the outputs of two different features are connected to generate a novel characteristic descriptor. Lastly, the new feature descriptors are transformed into refined features through the convolution and sigmoid function operations. The following describes how the SAM is calculated.
M s ( F ) = f s i g m o i d ( C o n v [ M a x P o o l ( F ) ; A v g P o o l ( F ) ] ) )
where M a x p o o l ( · ) represents maxpooling, A v g p o o l ( · ) represents averagepooling, M L P ( · ) represents the multi-layer perceptron, and C o n v ( · ) represents a CNN layer.

2.2.3. Convolutional Neural Network

The widely-used CNN in image analysis offers strong grid data processing capabilities [35]. As shown in Figure 4, the input, convolutional, pooling, fully connected, and output layers make up the fundamental architecture of the CNN. The convolution and pooling layers transform and extract features from the information in the input layer. The fully connected layer then performs the mapping between the resulting characteristic maps and output values [16].
The convolutional layer, the most important layer in the CNN, extracts the features of the input image by means of the convolutional kernel. The size of the input matrix is larger than the convolution kernel. The convolution layer uses convolution operations to output the characteristic map. Each component of the feature map has the following calculation formula.
x i , j o u t = f C o n v ( m = 0 k n = 0 k w m , n x i + m , j + n i n + b )
where x i , j o u t is the characteristic map’s output value for row i and column j ; x i + m , j + m i n is the value in the input matrix’s row i and column j ; f C o n v ( · ) is activation function; w m , n is the convolution kernel’s weight in row m and column n ; and b is the convolution kernel’s bias [23].

2.2.4. Bi-Directional Long Short-Term Memory

In order to overcome the difficulty of the long-term dependence of time series data, the LSTM introduces a special cell storage structure. As shown in Figure 5b, the architecture of each LSTM cell has an input gate i t , a forgetting gate f t , and an output gate O t . The specific derivation of the LSTM is as follows.
f t = σ ( W f [ h t 1 ,   x t ] + b f )
i t = σ ( W i [ h t 1 ,   x t ] + b i )
C ˜ t = tan h ( W C [ h t 1 ,   x t ] + b C )
C t = f t C t 1 + i t C ˜ t
O t = σ ( W o [ h t 1 , x t   ] + b o )
h t = O t tanh ( C t )
where W f , W i , W C , and W o are the input weights, b f , b i , b c , and b o are the deviation weights, t 1 is the previous time state, t represents the current time state, x t is the input vector, and h t represents the output vector. Here, f t acts as the forget gate and decides what past PM2.5 information and other factors should be forgotten from the cell state. i t expresses the input gate, which makes the decision on what new data to store in the cell state. Lastly, C ˜ t is a self-recurrent cell in a neuron, and C t is the LSTM block’s internal memory cell.
The Bi LSTM model, as shown in Figure 5, in contrast to the LSTM model, is made up of a forward LSTM layer h and a backward LSTM layer h . Due to the separate hidden layers in both directions, the Bi LSTM can analyze sequence data in both forward and backward directions. Each hidden layer enables the recording of information from the past (forward) and the future (backward) [36]. As a result, a more thorough collection of PM2.5 characteristics may be retrieved to raise the network’s prediction accuracy.

2.3. Experimental Setup

In this experiment, we use Keras based on Tensorflow to construct comparative models of deep learning (CNN, LSTM, Bi LSTM, and CNN-LSTM) and the proposed CBAM-CNN-Bi LSTM model. Table 2 displays the parameters utilized to train the prediction model. Then, we use Adam (adaptive gradient algorithm) as the optimization algorithm and MSE as the loss function for the prediction model. In addition, to enhance the model’s capacity for generalization, we packaged the dataset and broke it up, splitting the dataset so that the training set contains 80% of the dataset and the test set contains 20%.

3. Results and Discussion

3.1. Performance Evaluation Indices

This paper uses RMSE, MAE, R2, and IA to analyze how well the prediction models performed. The RMSE reflects the sensitivity of the model to error, and the MAE reflects the stability of the model; the closer the value of both to 0, the better the prediction result. R2 represents the ability to forecast the actual data, and IA represents the similarity of the distribution between actual and predicted values, both variables’ values span from [0, 1], the closer to 1, the more consistent the predicted results are with the distribution of the true data. The calculation formula is shown as shown.
RMSE = 1 n i = 1 n ( y i y ^ i ) 2
MAE = 1 n i = 1 n | y i y ^ i |
R 2 = i = 1 n ( y ^ i y ¯ i ) 2 i = 1 n ( y i y ¯ i ) 2
IA = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( | y i y ¯ | + | y ^ i y ¯ | ) 2
where n denotes the sample size in the dataset, y ^ i denotes predicted value corresponding to it, y i denotes actual concentration of PM2.5, and y ¯ represents the mean of all measurements of PM2.5 concentrations.

3.2. Correlation Analysis of Variables

In this subsection, we correlated PM2.5 concentrations to achieve two objectives. First, we investigate the relationships between PM2.5 concentrations, pollutant concentrations, and meteorological data. Furthermore, to ensure the convergence of the model, we placed the highly correlated factors of influence in the same channel. The correlations between the 12 variables in all sites are shown in Figure 6. For absolute values of correlation, among the pollutants, the highest correlation between PM2.5 and itself was found, with PM10 (0.89), CO (0.80), NO2 (0.67), SO2 (0.39), and O3 following (−0.17). Among the meteorological factors, wind speed (−0.29) had the strongest association with PM2.5, followed by temperature (−0.15) and dew point (0.11). The absolute values of the correlation coefficients for all other influences were below 0.1. In this paper, the six influential factors with the highest absolute values of correlation coefficients with PM2.5 are put into the same channel; they are PM2.5, PM10, CO, NO2, SO2, and wind speed. We put O3 (−0.17), temperature, pressure, dew point, precipitation, and wind direction into another channel of the model input, forming a 12*6*2 “image”.
In terms of pollutants, PM2.5 concentrations have a highly substantial association with PM10, CO, NO2, and SO2, and a bad relationship with O3 concentrations. This is due to the fact that PM2.5, PM10, CO, NO2, and SO2 are mostly derived from human activities such as coal combustion, vehicle exhausts, and industrial manufacturing. In addition, emissions of PM10, CO, NO2, and SO2 contribute to a certain extent to the increase in PM2.5 concentrations. Unlike other pollutants, O3 comes mainly from nature. As light and temperature increase, the concentration of O3 increases. Due to the high concentration of O3, there is a chance that some photochemical reactions will consume some PM2.5 and lower its concentration.
Compared to pollutants, meteorological factors have a relatively small but integral impact on PM2.5 concentrations. The relationship between wind speed and PM2.5 concentrations is negative. High wind speeds are conducive to PM2.5 dispersion and therefore have a substantial impact on the concentration of PM2.5. The increase in temperature causes instability in the atmosphere, which has a positive effect on the dispersion of PM2.5. The dew point, a measure of air humidity, is positively correlated with PM2.5 concentrations. An environment with high air humidity contributes to the formation of fine particulate matter, making PM2.5 less likely to disperse. Wind direction, barometric pressure, and rainfall are all weakly correlated with PM2.5 and their changes have little effect on PM2.5 concentrations.

3.3. Short-Term Prediction

Pollutant prediction models can be divided into short-term prediction models and long-term prediction models [37]. Short-term prediction focuses on the accuracy of the forecast and ensures the safety of human activities in the short term by keeping the forecast time within 12 h [38]. We will compare and analyze the short-term prediction performance of each model for pollutant concentrations in this section.

3.3.1. Effect of Convolution Kernel Size on Experimental Results

The convolutional kernel is a key part of the convolutional neural network model, which directly affects how well the features are extracted and how quickly the network converges. The size of the convolution kernel should be appropriate to the size of the input “image”. If the convolution kernel is too large, the local features cannot be extracted effectively; if the convolution kernel is too small, the overall features cannot be extracted successfully. Therefore, a convolution kernel of the right size should be selected to fit the input “image”. As the size of the input “image” of the CBAM-CNN-Bi LSTM model is 12*6*2, we set the convolution kernel to 2*2, 3*3, 4*4, 5*5 to forecast the PM2.5 concentrations in the next 6 h.
Table 3 gives the average of the performance evaluation indicators of the CBAM-CNN-Bi LSTM for PM2.5 concentration prediction for the next 6 h at different convolutional kernel sizes. As shown in Table 3, the test errors of the models do not differ significantly when the size of the convolution kernel varies. However, when the convolutional kernel size was 3*3, RMSE and MAE reached a minimum value of 29.65 and 18.58, and R2 and IA reached a maximum value of 0.8192 and 96.01%.
This demonstrates that the highest prediction performance for our proposed model occurs when the convolution kernel is 3*3. When the convolution kernel is smaller than 3*3, the model has an underfitting problem for the overall spatial features of PM2.5 concentrations; when the size of the kernel is larger than 3*3, our model cannot effectively extract the local spatial features of PM2.5 concentrations. When the convolution kernel is 3*3, the model is capable of extracting both global and local spatial information related to PM2.5 concentrations. In the following experiments, the convolution kernel size is set to 3*3.

3.3.2. Effect of Different Models on Experimental Results

The quantitative results for single-step PM2.5 concentrations prediction are given in Table 4, which gives a comparison of the RMSE, MAE, R2 and IA for CNN, LSTM, Bi LSTM, CNN-LSTM and CBAM-CNN-Bi LSTM. As shown in Table 4, our proposed model performed better than other deep learning models in single-step PM2.5 concentration prediction. In contrast to other models, our proposed model has the minimum prediction error and the greatest prediction accuracy and reduces RMSE to 18.90, MAE to 11.20, improves R2 to 0.9397, and IA to 98.54%. However, the prediction results of our proposed model are not much ahead of Bi LSTM because the single-step prediction is relatively simple and does not reflect the advantages of our designed architecture. In addition, CNN has the worst prediction performance, and LSTM and Bi LSTM perform predictions more accurately than CNN-LSTM. This means that the LSTM and Bi LSTM, with their excellent time series data processing capability, are more suitable than CNN-LSTM for the PM2.5 concentration single-step prediction task.
It is well known that with the increase in forecast time steps, forecasting becomes more difficult. To further evaluate the PM2.5 concentrations short-term prediction capability of CBAM-CNN-Bi LSTM and other deep learning models, we predicted PM2.5 concentrations for the next 2–12 h and presented the predicted quantitative results through the change curves of RMSE, MAE, R2 and IA in Figure 7. As shown in Figure 7, the predictive ability of all prediction models declines as the prediction time step rises. We observed that four performance evaluation indicators of CNN were always worse than other deep learning models, and the LSTM and Bi LSTM prediction performance was generally consistent. Interestingly, as shown in Figure 7a,b, in cases where the predicted time step is under five, compared to the LSTM and Bi LSTM, the CNN-LSTM has a greater prediction error. Does this mean that CNN-LSTM has poorer short-term prediction performance than LSTM and Bi LSTM? In fact, we will find that four metrics for evaluating the performance of the CNN-LSTM start to outperform the LSTM and Bi LSTM when the prediction step size is greater than five. This suggests that CNN-LSTM, with its hybrid model structure, can better quantify complex data when prediction problems become difficult. It is worth noting that when prediction time steps are greater than 4, our proposed model consistently maintains optimal prediction performance with the lowest RMSE and MAE, and the highest R2 and IA.
In summary, for PM2.5 concentrations short-term prediction, when the convolution kernel is 3*3, CBAM-CNN-Bi LSTM obtains the best prediction performance and maintains the best results among all deep learning models. This means that CBAM plays a key role in the prediction of deep learning models. CBAM obtains the feature relationship between pollutant data and meteorological data, optimizes the CNN spatial feature extraction, and improves the model prediction accuracy.

3.4. Long-Term Prediction

The research on pollutant concentration prediction has mainly focused on pollutant concentration short-term prediction, but this is not sufficient to meet the actual demand. The purpose of long-term forecasting is to forecast pollutant concentrations for a longer period of time in the future, and its predictions can serve as a useful reference for managers. It can be seen that long-term predictions of pollutant concentrations are very meaningful. We will analyze the pollutant concentration long-term prediction performance of each model in this section.

3.4.1. PM2.5 Concentration Prediction

The quantitative results of the long-term PM2.5 concentration prediction (h13~h18) are given in Table 5, which gives the comparison of RMSE, MAE, R2, and IA for CNN, LSTM, Bi LSTM, CNN-LSTM, and our proposed model. As shown in Table 5, our proposed model outperforms other deep learning models in PM2.5 concentration prediction (h13~h18). In comparison to alternative models, CBAM-CNN-Bi LSTM has the lowest prediction error and the highest prediction accuracy and reduces RMSE to 37.33, MAE to 26.54, R2 to 0.6981, and IA to 93.50%. As shown in Table 5, the R2 of CNN has decreased to −0.2546, which indicates that CNN is inappropriate for long-term PM2.5 concentration prediction. It is worth noting that the prediction performance of CNN-LSTM is superior to Bi LSTM and LSTM, this shows that the hybrid model of CNN-LSTM is more suitable for long-term PM2.5 concentration prediction.
Next, we analyze the effect of the size of the prediction time steps on CNN, LSTM, Bi LSTM, CNN-LSTM, and our proposed model. As shown in Table 6, Table 7 and Table 8, when the prediction time steps are the same as h13~h18, the larger the window size, the more accurate the model’s prediction performance. This means that deep learning models can optimize the prediction results by learning more historical data. Table 5 shows that the prediction performance of all deep learning methods steadily declines as prediction step size increases. It is evident that, in contrast to other deep learning methods, our proposed model has the minimum prediction error (RMSE and MAE) and the highest prediction accuracy (R2 and IA) for different prediction time steps.
To verify the effectiveness of our proposed model, we analyzed the variations of RMSE, MAE, R2, and IA for each model within the prediction step of 48 h. As shown in Table 7 and Table 8, the four performance evaluation indicator metrics of CNN, Bi LSTM, LSTM, and CNN-LSTM fluctuated widely in the long-term prediction. Furthermore, what is interesting about the data in Table 7 and Table 8 is that our proposed model can continue to outperform the CNN-LSTM significantly as the prediction time step grows. In addition, the four evaluation indicators of our proposed model fluctuated less (with little change in values) as the prediction time steps increased, which shows that CBAM-CNN-Bi LSTM is most suitable for long-term PM2.5 concentration prediction. In conclusion, our proposed model can be used to maintain the best and most stable prediction performance in long-term PM2.5 concentration prediction.
To further validate the prediction performance of our proposed model, we analyzed the fitting ability of our proposed model and four other deep learning models for PM2.5 concentrations at a prediction time step of 48 h. As shown in Figure 8a1–e1, we found that the CNN has the worst prediction performance and cannot describe the trend of PM2.5 concentrations. Compared to LSTM and Bi LSTM, CNN-LSTM has a higher long-term prediction ability for PM2.5 concentrations, but the accuracy of the prediction of sudden change points of PM2.5 concentrations is not enough. Our proposed model shown in Figure 8 outperforms other comparative models in the prediction of sudden change points in PM2.5 concentrations. We observed that when the PM2.5 concentrations were larger than 200 μg/m3, the comparison model’s predicted outcomes were unable to capture the true trend. This also reflects that when the PM2.5 concentrations are too high, it makes precise prediction using the model challenging. Moreover, the predictions of our proposed model and the observed outcomes are virtually identical (as shown in the red wireframe part in Figure 8). This means that our proposed model has a good fit for the prediction of high PM2.5 concentration values.
Combined with the capability of fitting the model in Figure 8, we can draw the following conclusions: (1) CBAM-CNN-Bi LSTM has better prediction performance in all time periods and can effectively forecast PM2.5 concentrations in different environments; (2) in the case of high PM2.5 concentration values, the CBAM-CNN-Bi LSTM has a good fitting effect, making the predicted and observed values basically consistent; (3) we can see that the number of sudden change point samples is relatively small. This phenomenon causes the deep learning model to insufficiently learn the change pattern of PM2.5 concentrations in the case of sudden changes. This is the reason why the deep learning model is difficult to fit to the sudden change points.

3.4.2. Other Pollutant Concentration Prediction

To confirm the applicability of our suggested model. In this section, PM10 and SO2 are used as examples, and Table 9 and Table 10 display the evaluation indices of model prediction performance. As shown in Table 9, our proposed model still has the best prediction ability. The prediction performance of our proposed model is significantly better than that of the single framework model. Compared with the CNN-LSTM model, our proposed model reduces RMSE by 3.57, MAE by 2.65, R2 by 0.0115, and IA by 0.59%. This indicates that the convolutional block attention module optimizes the model and improves the prediction accuracy. The predicted results for SO2 are similar to those for PM10. As shown in Table 10, our model reduced the RMSE to 9.87, MAE to 6.14, R2 to 0.7175, and IA to 93.76%. In the prediction of SO2 and PM10, our model also maintains the optimal prediction results. This indicates that our proposed prediction approach is applicable to the prediction of other pollutants and is as successful.

4. Conclusions and Future

In my research, a unique PM2.5 concentration prediction model (CBAM-CNN-Bi LSTM) is proposed, which gives a reasonable prediction by learning from a large amount of pollutant data and meteorological data. CBAM-CNN-Bi LSTM is a hybrid deep learning model which consists of CBAM, CNN, and Bi LSTM. The advantages of CBAM-CNN-Bi LSTM are concluded as below:
(1)
By utilizing the convolutional block attention module, the CNN network degradation issue may be solved. The spatial attention module assists CNN in efficiently acquiring spatial correlation features between multiple sites to extract pollutants and meteorological data. The channel attention module is used to capture the complex relationship features between the influencing factors of model inputs. Convolutional block attention modules optimize convolutional neural networks to provide more reliable data for more precise result prediction;
(2)
By using Bi LSTM as the output prediction layer, the model not only obtains the performance advantage of long-time series prediction through Bi LSTM, avoiding the problem of underutilization of contextual information, but also extracts the effective association features of the output of the convolutional neural network layer to achieve the goal of mining data spatiotemporal association;
(3)
Our proposed model can be simultaneously applied to meteorological and pollution data from multiple stations for environmental monitoring of big data while considering the changes in the spatial and temporal distribution of the data to achieve the prediction of air pollutant concentrations in the target city. Experiments conducted on the dataset show that our framework obtains better results than other methods.
Based on the aforementioned experimental findings, the effectiveness of our proposed model is demonstrated. In comparison to other models, our proposed model gives accurate PM2.5 concentration predictions by fully extracting the temporal and spatial characteristics of PM2.5 and the correlation between pollutant data and meteorological data and overcoming the problem of long-time dependence of PM2.5 concentrations. Therefore, our proposed model overcomes the weaknesses of CNN-LSTM and has more practical value. However, traffic, vegetation cover, and pedestrian flow are not considered in this paper, which will be addressed in future studies.

Author Contributions

Conceptualization, D.L.; methodology, D.L.; software, D.L.; validation, D.L.; formal analysis, Y.Z.; investigation, J.L.; resources, J.L.; data curation, D.L.; writing—original draft preparation, D.L.; writing—review and editing, D.L. and Y.Z.; visualization, Y.Z.; supervision, J.L.; project administration, Y.Z.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Lanzhou Jiaotong University (grant no. EP 201806).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data was obtained from [Songxi Chen] and are available [https://archive.ics.uci.edu/ml/datasets/Beijing+Multi-Site+Air-Quality+Data] with the permission of [Songxi Chen].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fong, I.H.; Li, T.; Fong, S.; Wong, R.K.; Tallon-Ballesteros, A.J. Predicting concentration levels of air pollutants by transfer learning and recurrent neural network. Knowl.-Based Syst. 2020, 192, 105622. [Google Scholar] [CrossRef]
  2. Maleki, H.; Sorooshian, A.; Goudarzi, G.; Baboli, Z.; Birgani, Y.T.; Rahmati, M.J. Air pollution prediction by using an artificial neural network model. Clean Technol. Environ. Policy 2019, 21, 1341–1352. [Google Scholar] [CrossRef] [PubMed]
  3. Chen, F.; Chen, Z. Cost of economic growth: Air pollution and health expenditure. Sci. Total Environ. 2021, 755, 142543. [Google Scholar] [CrossRef]
  4. Park, S.; Kim, M.; Kim, M.; Namgung, H.G.; Kim, K.T.; Cho, K.H.; Kwon, S.B. Predicting PM10 concentration in Seoul metropolitan subway stations using artificial neural network (ANN). J. Hazard. Mater. 2017, 341, 75–82. [Google Scholar] [CrossRef] [PubMed]
  5. Zhang, B.; Zou, G.; Qin, D.; Ni, Q.; Mao, H.; Li, M. RCL-Learning: ResNet and convolutional long short-term memory-based spatiotemporal air pollutant concentration prediction model. Expert Syst. Appl. 2022, 207, 118017. [Google Scholar] [CrossRef]
  6. Djalalova, I.; Delle Monache, L.; Wilczak, J. PM2.5 analog forecast and Kalman filter post-processing for the Community Multiscale Air Quality (CMAQ) model. Atmos. Environ. 2015, 108, 76–87. [Google Scholar] [CrossRef]
  7. Zhu, B.; Akimoto, H.; Wang, Z. The Preliminary Application of a Nested Air Quality Prediction Modeling System in Kanto Area, Japan. AGU Fall Meet. Abstr. 2005, 2005, A33F-08. [Google Scholar]
  8. Saide, P.E.; Carmichael, G.R.; Spak, S.N.; Gallardo, L.; Osses, A.E.; Mena-Carrasco, M.A.; Pagowski, M. Forecasting urban PM10 and PM2.5 pollution episodes in very stable nocturnal conditions and complex terrain using WRF–Chem CO tracer model—ScienceDirect. Atmos. Environ. 2011, 45, 2769–2780. [Google Scholar] [CrossRef]
  9. Zou, G.; Zhang, B.; Yong, R.; Qin, D.; Zhao, Q. FDN-learning: Urban PM 2.5-concentration Spatial Correlation Prediction Model Based on Fusion Deep Neural Network. Big Data Res. 2021, 26, 100269. [Google Scholar] [CrossRef]
  10. Rubal; Kumar, D. Evolving Differential evolution method with random forest for prediction of Air Pollution. Procedia Comput. Sci. 2018, 132, 824–833. [Google Scholar] [CrossRef]
  11. Zhang, H.; Zhang, S.; Wang, P.; Qin, Y.; Wang, H. Forecasting of particulate matter time series using wavelet analysis and wavelet-ARMA/ARIMA model in Taiyuan, China. J. Air Waste Manag. Assoc. 2017, 67, 776–788. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, L.; Lin, J.; Qiu, R.; Hu, X.; Zhang, H.; Chen, Q.; Tan, H.; Lin, D.; Wang, J. Trend analysis and forecast of PM2.5 in Fuzhou, China using the ARIMA model. Ecol. Indic. 2018, 95, 702–710. [Google Scholar] [CrossRef]
  13. Leong, W.; Kelani, R.; Ahmad, Z. Prediction of air pollution index (API) using support vector machine (SVM). J. Environ. Chem. Eng. 2019, 8, 103208. [Google Scholar] [CrossRef]
  14. Yu, Z.; Yi, X.; Ming, L.; Li, R.; Shan, Z. Forecasting Fine-Grained Air Quality Based on Big Data. In Proceedings of the the 21th ACM SIGKDD International Conference, Sydney, Australia, 10–12 August 2015. [Google Scholar]
  15. Huang, C.-J.; Kuo, P.-H. A Deep CNN-LSTM Model for Particulate Matter (PM2.5) Forecasting in Smart Cities. Sensors 2018, 18, 2220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Yang, J.; Yan, R.; Nong, M.; Liao, J.; Li, F.; Sun, W. PM 2.5 concentrations forecasting in Beijing through deep learning with different inputs, model structures and forecast time. Atmos. Pollut. Res. 2021, 12, 101168. [Google Scholar] [CrossRef]
  17. Xin, R.B.; Jiang, Z.F.; Li, N.; Hou, L.J. An Air Quality Predictive Model of Licang of Qingdao City Based on BP Neural Network; Trans Tech Publications Ltd.: Stafa-Zurich, Switzerland, 2012. [Google Scholar]
  18. Fan, J.; Li, Q.; Hou, J.; Feng, X.; Karimian, H.; Lin, S. A Spatiotemporal Prediction Framework for Air Pollution Based on Deep RNN. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 4, 15. [Google Scholar] [CrossRef] [Green Version]
  19. Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  20. Li, X.; Peng, L.; Yao, X.; Cui, S.; Hu, Y.; You, C.; Chi, T. Long short-term memory neural network for air pollutant concentration predictions: Method development and evaluation. Environ. Pollut. 2017, 231, 997–1004. [Google Scholar] [CrossRef]
  21. Prihatno, A.T.; Nurcahyanto, H.; Ahmed, M.F.; Rahman, M.H.; Alam, M.M.; Jang, Y.M. Forecasting PM2.5 Concentration Using a Single-Dense Layer BiLSTM Method. Electronics 2021, 10, 1808. [Google Scholar] [CrossRef]
  22. Hu, X.; Liu, T.; Hao, X.; Lin, C. Attention-based Conv-LSTM and Bi-LSTM networks for large-scale traffic speed prediction. J. Supercomput. 2022, 78, 12686–12709. [Google Scholar] [CrossRef]
  23. Yan, R.; Liao, J.; Yang, J.; Sun, W.; Nong, M.; Li, F. Multi-hour and multi-site air quality index forecasting in Beijing using CNN, LSTM, CNN-LSTM, and spatiotemporal clustering. Expert Syst. Appl. 2020, 169, 114513. [Google Scholar] [CrossRef]
  24. Zhao, J.; Deng, F.; Cai, Y.; Chen, J. Long short-term memory—Fully connected (LSTM-FC) neural network for PM 2.5 concentration prediction. Chemosphere 2019, 220, 486–492. [Google Scholar] [CrossRef] [PubMed]
  25. Li, S.; Xie, G.; Ren, J.; Guo, L.; Yang, Y.; Xu, X. Urban PM2.5 Concentration Prediction via Attention-Based CNN–LSTM. Appl. Sci. 2020, 10, 1953. [Google Scholar] [CrossRef] [Green Version]
  26. Zhou, Q.; Jiang, H.; Wang, J.; Zhou, J. A hybrid model for PM2.5 forecasting based on ensemble empirical mode decomposition and a general regression neural network. Sci. Total Environ. 2014, 496, 264–274. [Google Scholar] [CrossRef]
  27. Qi, Y.; Li, Q.; Karimian, H.; Liu, D. A hybrid model for spatiotemporal forecasting of PM2.5 based on graph convolutional neural network and long short-term memory. Sci. Total Environ. 2019, 664, 1–10. [Google Scholar] [CrossRef]
  28. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. Adv. Neural Inf. Process. Syst. 2015, 28, 802–810. [Google Scholar]
  29. Liang, Z.; Zhu, G.; Shen, P.; Song, J. Learning Spatiotemporal Features Using 3DCNN and Convolutional LSTM for Gesture Recognition. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017. [Google Scholar]
  30. Liang, Y.; Lin, Y.; Lu, Q. Forecasting gold price using a novel hybrid model with ICEEMDAN and LSTM-CNN-CBAM. Expert Syst. Appl. 2022, 206, 117847. [Google Scholar] [CrossRef]
  31. Qin, D.; Yu, J.; Zou, G.; Yong, R.; Zhao, Q.; Zhang, B. A Novel Combined Prediction Scheme Based on CNN and LSTM for Urban PM2.5 Concentration. IEEE Access 2019, 7, 20050–20059. [Google Scholar] [CrossRef]
  32. Ma, K.; Zhan, C.a.A.; Yang, F. Multi-classification of arrhythmias using ResNet with CBAM on CWGAN-GP augmented ECG Gramian Angular Summation Field. Biomed. Signal Process. Control 2022, 77, 103684. [Google Scholar] [CrossRef]
  33. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar]
  34. Wang, Y.; Zhang, Z.; Feng, L.; Ma, Y.; Du, Q. A new attention-based CNN approach for crop mapping using time series Sentinel-2 images—ScienceDirect. Comput. Electron. Agric. 2021, 184, 106090. [Google Scholar] [CrossRef]
  35. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  36. Du, S.; Li, T.; Yang, Y.; Horng, S.J. Deep Air Quality Forecasting Using Hybrid Deep Learning Framework. IEEE Trans. Knowl. Data Eng. 2018, 33, 2412–2424. [Google Scholar] [CrossRef] [Green Version]
  37. Wang, Z.B.; Fang, C.L. Spatial-temporal characteristics and determinants of PM2.5 in the Bohai Rim Urban Agglomeration. Chemosphere 2016, 148, 148–162. [Google Scholar] [CrossRef] [PubMed]
  38. Zhou, J.; Li, W.; Yu, X.; Xu, X.; Yuan, X.; Wang, J. Elman-Based Forecaster Integrated by AdaboostAlgorithm in 15 min and 24 h ahead Power OutputPrediction Using PM 2.5 Values, PV ModuleTemperature, Hours of Sunshine, and Meteorological Data. Pol. J. Environ. Stud. 2019, 28, 1999. [Google Scholar] [CrossRef]
Figure 1. (a). Distribution of air quality monitoring stations in Beijing. (b). Time series plot of PM2.5 concentration data.
Figure 1. (a). Distribution of air quality monitoring stations in Beijing. (b). Time series plot of PM2.5 concentration data.
Atmosphere 13 01719 g001
Figure 2. The architecture of CBAM-CNN-Bi LSTM.
Figure 2. The architecture of CBAM-CNN-Bi LSTM.
Atmosphere 13 01719 g002
Figure 3. The architecture of convolutional block attention module. (a) The convolutional block attention module. (b) The architecture of channel attention. (c) The architecture of spatial attention. Reproduced with permission from Ref. [34]. Copyright 2021 ELSEVIER.
Figure 3. The architecture of convolutional block attention module. (a) The convolutional block attention module. (b) The architecture of channel attention. (c) The architecture of spatial attention. Reproduced with permission from Ref. [34]. Copyright 2021 ELSEVIER.
Atmosphere 13 01719 g003
Figure 4. The fundamental architecture of CNN.
Figure 4. The fundamental architecture of CNN.
Atmosphere 13 01719 g004
Figure 5. The basic structure of the Bi LSTM. (a) The Bi LSTM. (b) The basic structure of LSTM.
Figure 5. The basic structure of the Bi LSTM. (a) The Bi LSTM. (b) The basic structure of LSTM.
Atmosphere 13 01719 g005
Figure 6. Correlation result of multiple variables.
Figure 6. Correlation result of multiple variables.
Atmosphere 13 01719 g006
Figure 7. RMSE, MAE, R2, and IA of the CBAM-CNN-Bi LSTM model at different prediction time steps and comparisons with other deep learning models. (a) RMSE of models at different prediction steps. (b) MAE of models at different prediction steps. (c) R2 of models at different prediction steps. (d) IA of models at different prediction steps.
Figure 7. RMSE, MAE, R2, and IA of the CBAM-CNN-Bi LSTM model at different prediction time steps and comparisons with other deep learning models. (a) RMSE of models at different prediction steps. (b) MAE of models at different prediction steps. (c) R2 of models at different prediction steps. (d) IA of models at different prediction steps.
Atmosphere 13 01719 g007
Figure 8. Comparison of PM2.5 concentration prediction models for the next 48h at station 1001A. (a1) Prediction line graph of CNN model; (a2) Prediction scatter plot of CNN model; (b1) Prediction line graph of LSTM model; (b2) Prediction scatter plot of LSTM model; (c1) Prediction line graph of Bi LSTM model; (c2) Prediction scatter plot of Bi LSTM model; (d1) Prediction line graph of CNN-LSTM model; (d2) Prediction scatter plot of CNN-LSTM model; (e1) Prediction line graph of CBAM-CNN-Bi LSTM model; (e2) Prediction scatter plot of CBAM-CNN-Bi LSTM model.
Figure 8. Comparison of PM2.5 concentration prediction models for the next 48h at station 1001A. (a1) Prediction line graph of CNN model; (a2) Prediction scatter plot of CNN model; (b1) Prediction line graph of LSTM model; (b2) Prediction scatter plot of LSTM model; (c1) Prediction line graph of Bi LSTM model; (c2) Prediction scatter plot of Bi LSTM model; (d1) Prediction line graph of CNN-LSTM model; (d2) Prediction scatter plot of CNN-LSTM model; (e1) Prediction line graph of CBAM-CNN-Bi LSTM model; (e2) Prediction scatter plot of CBAM-CNN-Bi LSTM model.
Atmosphere 13 01719 g008aAtmosphere 13 01719 g008b
Table 1. Experimental data of the PM2.5 concentration prediction model.
Table 1. Experimental data of the PM2.5 concentration prediction model.
TypeVariableUnit
Meteorological DataTemperature [ 19.9 ,   41.6 ]
Pressure [ 982.4 ,   1042.8 ]
Dew Point [ 43.4 ,   29.1 ]
Precipitation [ 0 ,   72.5 ]
Wind Direction [ N ,   E S E ]
Wind Speed [ 0 ,   13.2 ]
Pollutant DataPM2.5 [ 2 ,   999 ]
PM10 [ 2 ,   999 ]
SO2 [ 0.2856 ,   500 ]
NO2 [ 1.0265 ,   290 ]
CO [ 100 ,   10000 ]
O3 [ 0.2142 ,   1071 ]
Table 2. CBAM-CNN-Bi LSTM model parameters setting.
Table 2. CBAM-CNN-Bi LSTM model parameters setting.
ParametersValue
Kernel size of CNN3 × 3
Convolution channels32
Convolution layer1
Bi LSTM nodes128
Bi LSTM layer1
Fully connected layer nodes12
Fully connected layer1
Learning rate0.0001
Batch size128
Epochs50
Table 3. Effect of convolution kernel size on experimental results in CBAM-CNN-Bi LSTM. (RMSE is the root mean squared error, MAE is the mean squared error, R2 is the r- squared, IA is the index of agreement).
Table 3. Effect of convolution kernel size on experimental results in CBAM-CNN-Bi LSTM. (RMSE is the root mean squared error, MAE is the mean squared error, R2 is the r- squared, IA is the index of agreement).
Convolution Kernel SizeRMSEMAER2IA
2*229.7918.840.815395.92%
3*329.6518.580.819296.01%
4*430.2619.120.814195.82%
5*530.1418.830.816295.85%
Note: window size = 6 (the model’s input window size represents historical observations), and model performance evaluation indicators (RMSE, MAE, R2 and IA) are the average of prediction of the next 1–6 h.
Table 4. Performance evaluation indicators for model single-step prediction.
Table 4. Performance evaluation indicators for model single-step prediction.
ModelsRMSEMAER2IA
CNN23.1514.150.913697.83%
LSTM19.5311.490.931498.39%
Bi LSTM18.9411.350.937098.30%
CNN-LSTM20.5111.990.931398.31%
CBAM-CNN-Bi LSTM18.9011.200.939798.54%
Note: window size = 3, and model performance evaluation indicators (RMSE, MAE, R2 and IA) are the prediction of the next 1 h.
Table 5. Performance evaluation indicators for model prediction (h13–h18).
Table 5. Performance evaluation indicators for model prediction (h13–h18).
ModelsRMSEMAER2IA
CNN58.7941.53−0.254679.05%
LSTM49.7134.870.356187.03%
Bi LSTM48.8434.890.413287.73%
CNN-LSTM42.8430.410.587691.12%
CBAM-CNN-Bi LSTM37.3326.540.698193.50%
Note: window size = 18, and model performance evaluation indicators (RMSE, MAE, R2, and IA) are the average of prediction of the next 13–18 h.
Table 6. Performance evaluation indicators for model prediction (h13–h24).
Table 6. Performance evaluation indicators for model prediction (h13–h24).
ModelsRMSEMAER2IA
13–18 h19–24 h13–18 h19–24 h13–18 h19–24 h13–18 h19–24 h
CNN58.5362.5942.2845.33−0.3338−1.007178.25%72.26%
LSTM47.9952.9434.5437.730.42160.274688.06%85.25%
Bi LSTM45.4249.4532.7235.570.51820.397289.67%87.44%
CNN-LSTM40.3043.4428.5431.450.63860.552592.20%90.57%
CBAM-CNN-Bi LSTM34.7537.0024.7026.530.75080.694294.49%93.49%
Note: window size = 24, and model performance evaluation indicators (RMSE, MAE, R2 and IA) are the average of prediction of the next t ~ t + n hours.
Table 7. Testing error for model prediction (h13–h48).
Table 7. Testing error for model prediction (h13–h48).
ModelsRMSEMAE
13–18 h19–24 h25–48 h13–18 h19–24 h25–48 h
CNN57.1961.8265.6940.4144.6348.19
LSTM46.0348.6049.3432.7635.4935.90
Bi LSTM43.0744.9445.6831.0532.0333.43
CNN-LSTM37.2038.5540.2426.5127.4529.16
CBAM-CNN-Bi LSTM31.4731.8432.3421.8621.7822.30
Note: window size = 48, and prediction error is averaged out by model testing error (RMSE and MAE). of the future t ~ t + n hours.
Table 8. Testing accuracy for model prediction (h13–h48).
Table 8. Testing accuracy for model prediction (h13–h48).
ModelsR2IA
13–18 h19–24 h25–48 h13–18 h25–36 h25–48 h
CNN−0.1841−0.9108−1.362280.01%73.68%68.79%
LSTM0.50190.38450.346789.44%87.53%87.02%
Bi LSTM0.55380.52940.493090.76%90.03%89.50%
CNN-LSTM0.71900.70770.663793.74%93.31%92.52%
CBAM-CNN-Bi LSTM0.80680.80420.798095.52%95.50%95.39%
Note: window size = 48, and model testing accuracy (R2 and IA) are the average of prediction accuracy of the next t ~ t + n hours.
Table 9. Performance evaluation indicators for model prediction of PM10 (h13–h48).
Table 9. Performance evaluation indicators for model prediction of PM10 (h13–h48).
ModelsRMSEMAER2IA
CNN73.0553.15−1.163970.69%
LSTM56.1241.120.381687.19%
Bi LSTM54.9740.770.421288.09%
CNN-LSTM45.9232.720.681892.74%
CBAM-CNN-Bi LSTM42.3530.070.693393.33%
Note: window size = 48, and model performance evaluation indicators (RMSE, MAE, R2 and IA) are the average of prediction of the next 13–48 h.
Table 10. Performance evaluation indicators for the model prediction of SO2 (h13–h48).
Table 10. Performance evaluation indicators for the model prediction of SO2 (h13–h48).
ModelsRMSEMAER2IA
CNN15.199.59−0.079380.79%
LSTM13.548.380.317286.35%
Bi LSTM13.158.220.344587.11%
CNN-LSTM11.237.050.656892.53%
CBAM-CNN-Bi LSTM9.876.140.717593.76%
Note: window size = 48, and model performance evaluation indicators (RMSE, MAE, R2, and IA) are the average of prediction of the next 13–48 h.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, D.; Liu, J.; Zhao, Y. Prediction of Multi-Site PM2.5 Concentrations in Beijing Using CNN-Bi LSTM with CBAM. Atmosphere 2022, 13, 1719. https://doi.org/10.3390/atmos13101719

AMA Style

Li D, Liu J, Zhao Y. Prediction of Multi-Site PM2.5 Concentrations in Beijing Using CNN-Bi LSTM with CBAM. Atmosphere. 2022; 13(10):1719. https://doi.org/10.3390/atmos13101719

Chicago/Turabian Style

Li, Dong, Jiping Liu, and Yangyang Zhao. 2022. "Prediction of Multi-Site PM2.5 Concentrations in Beijing Using CNN-Bi LSTM with CBAM" Atmosphere 13, no. 10: 1719. https://doi.org/10.3390/atmos13101719

APA Style

Li, D., Liu, J., & Zhao, Y. (2022). Prediction of Multi-Site PM2.5 Concentrations in Beijing Using CNN-Bi LSTM with CBAM. Atmosphere, 13(10), 1719. https://doi.org/10.3390/atmos13101719

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop