Next Article in Journal
Analysis of Factors Related to Working Status of Dental Hygienists in Japan
Previous Article in Journal
Identification Framework of Contaminant Spill in Rivers Using Machine Learning with Breakthrough Curve Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on a Novel Hybrid Decomposition–Ensemble Learning Paradigm Based on VMD and IWOA for PM2.5 Forecasting

1
School of Geoscience and Technology, Zhengzhou University, Zhengzhou 450001, China
2
College of Atmospheric Sciences, Lanzhou University, Lanzhou 730000, China
3
School of Management, Lanzhou University, Lanzhou 730000, China
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(3), 1024; https://doi.org/10.3390/ijerph18031024
Submission received: 4 January 2021 / Accepted: 20 January 2021 / Published: 24 January 2021

Abstract

:
The non-stationarity, nonlinearity and complexity of the PM2.5 series have caused difficulties in PM2.5 prediction. To improve prediction accuracy, many forecasting methods have been developed. However, these methods usually do not consider the importance of data preprocessing and have limitations only using a single forecasting model. Therefore, this paper proposed a new hybrid decomposition–ensemble learning paradigm based on variation mode decomposition (VMD) and improved whale-optimization algorithm (IWOA) to address complex nonlinear environmental data. First, the VMD is employed to decompose the PM2.5 sequences into a set of variational modes (VMs) with different frequencies. Then, an ensemble method based on four individual forecasting approaches is applied to forecast all the VMs. With regard to ensemble weight coefficients, the IWOA is applied to optimize the weight coefficients, and the final forecasting results were obtained by reconstructing the refined sequences. To verify and validate the proposed learning paradigm, four daily PM2.5 datasets collected from the Jing-Jin-Ji area of China are chosen as the test cases to conduct the empirical research. The experimental results indicated that the proposed learning paradigm has the best results in all cases and metrics.

1. Introduction

Pollution of the environment is one of the most serious issues facing humankind today, and badly polluted air can cause great damage in economics and people’s lives. According to the World Health Organization (WHO), it is known that almost 3 million children die every year from a range of problems caused by air pollution [1]. With the process of industrialization and urbanization, the air pollution is becoming increasingly serious and the hazy weather has grown rapidly, especially in developing countries. In recent years, the foggy weather in many areas of China have become increasingly serious. Since the beginning of 2013, sustained haze weather has turned Beijing-Tianjin-Hebei (Jing-Jin-Ji region) into heavy pollution region. Fine particulate matter is one of the key contributors that leading to air pollution and hazy weather. It carries many adverse health effects, such as respiratory diseases and premature death [2].
Recently, increasingly countries have set up environmental monitoring systems, which can provide a large amount of PM monitoring data. However, PM data are affected by many factors and fluctuates greatly over time, making it very challenging to predict. Therefore, many models and tools have been developed to predict PM2.5 and other air pollutant concentrations to improve the accuracy of the predictions. These models can be generally categorized into physical, statistical and hybrid models. For example, physical methods can be used to simulate the processes of emissions, diffusion and transfer of pollutants through meteorological, emission, and chemical models [3,4,5]. Statistical methods which mainly include autoregressive integrated moving average model (ARIMA), artificial neural networks (ANN) and multiple linear regression (MLR) [2,6,7,8,9], have been broadly applied to the pollutant concentration prediction. For instance, Ref. [10] proposed a forecasting model based on MLR and bivariate correlation analysis to predict the annual and seasonal concentrations of PM10 and PM2.5. Ref. [11] studied the effects of meteorological factors on ultrafine particulate matter (UFP) and PM10 concentrations under traffic congestion conditions using the ARIMA model. However, in practice, most pollutant sequences are non-linear and irregular, which may involve the problem of non-linear dynamical systems, so these linear algorithms are still problematic in predicting PM concentration. On the contrary, using artificial neural network models to predict pollutant concentration can overcome the limitations of traditional linear models and handle nonlinear problems well. [12] developed extended model based on long-term and short-term memory neural network. The model takes into account the spatiotemporal correlation to predict the pollutant concentration and shows excellent performance. [13] applied cuckoo search (CS) to optimize BPNN to predict PM concentrations in four major cities in China.
Recently, to predict air quality more accurately, many hybrid models have been proposed based on ensemble learning paradigms, data preprocessing techniques and heuristic algorithms. For example, Ref. [14] developed a new prediction model based on the multidimensional k-nearest neighbor model and the ensemble empirical mode decomposition (EEMD) method. Ref. [15] developed a novel hybrid model based on wavelet transform (WT) and stacked autoencoder (SAE) and long short-term memory (LSTM) to simulate PM2.5 at six sites in China. Ref. [16] developed a model based on a combination of WT and neural network algorithm to decompose the PM2.5 data and then perform sub-series prediction analysis and finally data reconstruction. Ref. [17] proposed a novel PM2.5 hybrid prediction model, which includes a new pre-processing method (wavelet transform and variational mode decomposition), using differential evolution (DE) algorithm optimized BPNN to predict each decomposition sequence. The drawback of the decomposition-based prediction model is that using a single method to predict all signal sequences. Since different decomposition sequences have different characteristics, a single model does not fit all the characteristics of the decomposition sequences [18]. Thus, ensemble prediction model integrated multiple single models will help avoid the shortcomings of a single model and further improve the prediction accuracy. Furthermore, many heuristic algorithms are used to help optimize the weight coefficients of the ensemble model. [19] developed an ensemble model based on differential evolution (DE) to determine the optimum weights for electricity demand forecasting. Ref. [20] employed the cuckoo search algorithm (CSO) to optimize the weight coefficients of ensemble model. Whale optimization algorithm (WOA), proposed by Ref. [21], is a novel heuristic algorithm by imitating whale behavior in nature. However, the WOA will encounter problems such as being stuck in a local optimal solution and slow in convergence, when solving more complex problems. Thus, a new improved whale optimization algorithm (IWOA) is proposed in this study to strengthen the local seeking capability of the WOA.
Through the above analysis, considering the criticality of data pre-processing and the limitations of one single prediction model, a new hybrid decomposition–ensemble learning paradigm based on variation mode decomposition (VMD) and modified whale-optimization algorithm (IWOA) is introduced. First, the original PM sequence is decomposed into different VM sequences using VMD. Then, the weight-determined ensemble model, which optimized by IWOA, is employed to forecast each decomposition component. Finally, several prediction subsets are assembled into the final prediction result.
The paper is structured as follows: in Section 2, several single forecasting models, the ensemble prediction theory and VMD, are introduced. In Section 3, the proposed decomposition–ensemble model is presented. In Section 4, the study areas and the evaluation criteria are described. In Section 5, the comparative results of the proposed model and other models is in conducted. Finally, in Section 6, the conclusions the important results of this paper are explicitly introduced.

2. Related Methodology

Four individual forecasting models, VMD, IWOA, which employed in the suggested ensemble model, are described as follows.

2.1. Four Individual Prediction Methods

In latest years, many prediction models have been developed and applied to PM2.5 concentration prediction. This paper uses four popular methods, BPNN, ANFIS, ANFIS-FCM and GMDH, which show good performance in PM2.5 prediction, to construct the ensemble models.

2.1.1. The Back Propagation Neural Network (BPNN)

The BPNN is a multi-layer feed-forward neural network, which is widely used in many fields. The BPNN algorithm needs to find the parameter with the minimum error, i.e., the minimum value of the error between the output value and the actual value according to the negative gradient direction. The process of the BPNN is mainly divided into update and learning stages:
k i j ( t ) = w i j ( t 1 ) Δ w i j ( t )
Δ w i j ( t ) = η E / w i j ( t 1 ) + α Δ w i j ( t 1 )
where η denotes the learning speed, wij represents the weights between nodes i and j, α denotes the impulse parameter, E denotes the error super curve face and t denotes the current iterative steps.

2.1.2. The Adaptive Network Based Fuzzy Inference System (ANFIS)

Ref. [22] proposed the ANFIS that combines the blur systems and neural networks. It plays the advantages of both and makes up for the shortcomings of each. ANFIS can form an adaptive neuro-fuzzy controller by using a neural network learning mechanism to automatically retrieve rules from the input and output sample data. Through the offline training and the online learning algorithms, it can create fuzzy inferences and control the self-adjustment of rules, thereby making the system itself develop towards adaptive, self-organizing, and self-learning. ANFIS includes five-layer network, and each layer contains several node functions.
Layer 1: This process is the fuzzy layer. Each node in layer 1 is adaptive, all have node function, and will generate membership of a fuzzy set.
O i 1 = μ A i ( x )
where x denotes the input to node I, and Ai denotes the language label associated with the function of this node. The “μ” denotes the membership functions for Ai, which described by generalized Gaussian functions.
Layer 2: In this process every node is a circular node labeled out, i.e., -norm operation:
O i 2 = μ A i ( x ) + μ B i ( y ) , i = 1 , 2
Layer 3: At each node in this layer, the ratio of the firing weights under the ith rule to the sum of the firing weights under all rules is calculated:
O i 3 = w ¯ i = w i w 1 w 2 , i = 1 , 2
At this level, all outputs are collectively referred to as normalized emission intensity.
Layer 4: In this process, the contribution of the ith rule to the overall output is calculated:
O i 4 = w ¯ i f i = w ¯ i ( a i x + b i y + c i ) , i = 1 , 2
where w ¯ i represents the out of layer 3, and ( a i x + b i y + c i ) is the parameter set.
Layer 5: In this layer, the signal node calculate the final output as the sum of all incoming signals:
O i 5 = i w ¯ i f i = i w i f i i w i
The final output of the adaptive neural fuzzy inference system is:
f o u t = w ¯ 1 f 1 + w ¯ 2 f 2 = w 1 w 1 + w 2 f 1 + w 2 w 1 + w 2 f 2 = ( w ¯ 1 x ) p 1 + ( w ¯ 1 x ) q 1 + ( w ¯ 1 x ) r 1 + ( w ¯ 2 x ) p 2 + ( w ¯ 2 x ) q 2 + ( w ¯ 2 x ) r 2

2.1.3. The Fuzzy C-Means Clustering (FCM)

The FCM is a type of data aggregation method. In the FCM method, each data point needs to be classified as a level assigned at the member level in the cluster. FCM divides a selection of n vector xi, (i = 1, 2, …, n) into fuzzy groups, and then finds a clustering center in each fuzzy group in a way that minimizes the cost function of the similarity measure. The above i = 1, 2, …, c represents random sampling from n points. Here is a brief introduction to the stage of the FCM algorithm. First, choosing the centers of cluster ci, (i = 1, 2, …, c) from the n points (x1, x2, x3, …, xn) randomly. Second, the membership matrix U, is calculated by using the subsequent equation as follows:
μ i j = 1 k = 1 c ( d i j d k j ) 2 m 1
where d i j = c i x j is the Euclidean distance which involves the i-th cluster center and the j-th data point, and m is the fuzziness index. Third, compute the cost function using the following formula.
J ( U , c 1 , , c 2 ) = i = 1 c j i = i = 1 c j = 1 n μ i j m d i j 2
Stop the process when it falls below a certain threshold. Additionally, finally, the new c fuzzy clustering center ci, i = 1, 2, …, c is calculated using the following equation:
C i = j = 1 n μ i j m x j j = 1 n μ i j m

2.1.4. The Group Method of Data Handling (GMDH)

The GMDH is a series of computer-based inductive algorithms for the mathematical modeling of multi-parameter data sets. It is characterized by a fully automatic structure and parameter optimization of the model. The GMDH can be used for data extraction, knowledge detection, prediction, modeling of complex systems, optimization, and pattern recognition [23]. The GMDH algorithm features an induction procedure to classify increasingly complex multinomial models and select the best solution by the externality criterion.
The GMDH pattern generally have multiple sets of inputs and one set of outputs, and is a subset of the components of the basic function:
Y ( x 1 , , x n ) = a 0 + i = 1 m a i f i
where a denotes coefficients and f denotes the fundamental function that depends on different inputs, m represents the number of fundamental function components.
The basic function (12) is called the partial model and the GMDH considers various subsets of this function and thus finds the optimal solution. The coefficients of the model are first estimated using the least squares method. Then, the number of local components of the model is gradually increased. Finally, the GMDH algorithm finds the best complexity model structure by minimizing the external criterion. This process is called self-organizing of the model.
The main idea of the GMDH neural network learning algorithm is as follows: a series of source neurons are generated by performing cross-combining on each entry unit of the system, and the mean square error of the output error corresponding to each neuron is calculated; then, several outputs are chosen from the created neurons with smaller mean square error than a predetermined threshold, the selected neurons are used as the input unit of the new generation; the process of survival of the fittest and gradual evolution is repeated until the new generation of neurons is no better than the previous generation.

2.2. Variation Mode Decomposition (VMD)

The VMD is a non-recursive signal treatment algorithm that decomposes the original signal into a family of patterns with a specific frequency spectrum domain bandwidth [24]. During the decomposition process, each pattern can be compressively pulsed around a certain center. If the bandwidth of each pattern is required, three steps should be completed. After that, a constraint variational problem can be given. The details of VMD are described in [25].

2.3. Optimization Algorithm-IWOA

An improved heuristic algorithm IWOA is developed to enhance the performance of the ensemble model. The IWOA determines the optimal weight coefficient of the ensemble model. The basic whale optimization algorithm (WOA), chaotic local search (CLS), and the WOA modified by CLS will be described below.

2.3.1. Overview of the Whale Optimization Algorithm

The WOA, put forward by S. Mirjalili in 2016, is a simulation of the hunting mechanism of humpback whales, called bubble-net feeding method. Humpback whales form distinctive bubbles by circling around their prey in a circular or “9 shaped” path during foraging. With special exercises, humpback whales first form a spiral bubble 10-15 m below their prey while swimming upstream to the surface. Then, it surrounds the prey with its flashing fins to prevent it from escaping and catch it [21]. The mathematical principles of the above humpback whale behavior are described as follows:
(a) Encircling prey:
The humpback whale orbits its prey, and updates its position to the best search agent as the number of iterations increases. It can be depicted mathematically as:
D = | C X * ( t ) X ( t ) |
X ( t + 1 ) = X * ( t ) A D
where X* denotes the position vector of the best solution obtained thus far, X is the position vector, A and C denote the coefficient vectors, and t denotes the current iteration.
(b) Bubble-net attacking method:
The mathematical modeling of humpback whale’s vesicular behavior is designed for the following two methods. 1. Shrinking encircling mechanism: This is a bracket predation mechanism that requires finding a new agent location, which can be anywhere between the agent’s original location and the current optimal agent location. The values of A in this process is in the interval [−1, 1].
2. Spiral update position: A spiral equation needs to be created between the positions of both the prey and the whale to mimic the spiral motion of the humpback whale, as shown below.
X ( t + 1 ) = D e b l cos ( 2 π l ) + X * ( t )
The probability p, a random number in [0,1], is assumed to select between the shrinking encircling and the spiral-shaped path during the optimization process.
(c) Search for prey:
During the search phase, the variation of vector A can also be used to find prey at random. Therefore, to move away from a reference whale, A can be utilized with the random values greater than 1 or less than −1. The mathematical model at this stage is as follows:
X ( t + 1 ) = X r a n d A D D = | C X r a n d X |
where X r a n d is a random position vector (random whale) selected from the current population.

2.3.2. IWOA

As mentioned earlier, WOA was recently proposed and widely used in many fields. However, it also has shortcomings, like slow conversion in the late stage and easy to fall into a local optimum. Additionally, the chaotic local search (CLS), based on chaotic search, can effectively avoid the local optimization and converge to the global optimization. The blending of WOA and CLS can help improve global conversion and prevent falling into local solutions. To accelerate the local convergence of WOA, the chaotic local search algorithm is also applied. When WOA finishes iterating to find the best solution, the acceptance of these new solutions as determined by CLS will perform to local search a better solution close to the best solution. A logistic equation applied in CLS is defined as follows:
c x i i t e r + 1 = μ c x i i t e r ( 1 c x i i t e r )
where c x i denotes the ith chaotic variable, iter is the iteration number. When μ = 4 , the above equation exhibits chaotic dynamics, c x i denotes range in (0,1) and c x 0 { 0.25 ,   0.5 ,   0.75 } . For more details about CLS, please refer to [26].
The pseudo-code of the IWOA algorithm is outlined as follows:
Algorithm: Improved whale-optimization algorithm (IWOA)
Objective:
Minimize and maximize the objective function f ( x ) , x i = ( x i 1 ,   x i 2 ,   ,   x i d )
Parameters:
iter-iteration number.
Maxiter-the maximum number of iteration.
I-a population pop.
p-the switch probability
1. /*Initialize a population x i = ( x i 1 ,   x i 2 ,   ,   x i d )
2. WHILEiter < Maxiter
3.FORi = 1 to I
Update A , C , l and p
4.IFp > 0.5
5.IF ( | A | < + 1 )
6. Update the position of the current solution by Equation (14)
7.ELSE IF ( | A | > + 1 )
8. Randomly choose a search agent
9. Update the position of the current search agent by Equation (16)
10.END IF
11.ELSE IFp > 0.5
12. Update the position of the current search by Equation (15)
13.END IF
14.END FOR
15. /*Jump out of local optimum by using chaotic local search. */
16. Calculate c x i i t e r = x i i t e r x i min x i max x i min
17. Calculate the next iteration chaotic variable by Equation (16)
18. Transform c x i i t e r + 1 for the next iteration x i i t e r + 1 = x i min + c x i i t e r + 1 ( x i max x i min )
19. /*Evaluate x i i t e r replace x i i t e r by x i i t e r + 1 if the newly generation is better. */
20. /*Find the current best solution gbest*/
21. i t e r   =   i t e r   +   1
22.END WHILE

3. Decomposition–Ensemble Learning Paradigm

In this part, we suggest a new hybrid decomposition–ensemble learning paradigm that integrates VMD method, several prediction models and IWOA optimization. The main process of the developed decomposition–ensemble paradigm is shown in Figure 1. The three main steps of the ensemble model are as follows:
-
Step 1: Decomposition process:
First, the features and noise of the original pollution data needed to be cleaned and processed so that an effective prediction model could be built. In this study, VMD technology was used to disaggregate the original pollution datasets into a set of VMs and the residue component with corresponding frequencies.
-
Step 2: Ensemble forecasting and IWOA optimization:
The decomposition sequences with different characteristics were obtained via the VMD process. However, different sequences had different properties, which meant that a single prediction method could no longer effectively adapt to all the characteristics of the VMs. Thus, the ensemble strategy is adopted to solve this problem, and can be described as that if there are M types of prediction methods with the correct selection of weight coefficients to solve a problem. The results of multiple models were added together. Assume that E m o d e l (Model = “BPNN”, “ANFIS”, “ANFIS-FCM”, “GMDH”) is the ensemble prediction result of each VM by using the above methods. Then, using IWOA to optimize the output of the E m o d e l , it can be expressed as
O u t p u t S C W O A N N C T V M s = w 1 × P m o d e l 1 + w 2 × P m o d e l 2 + w 3 × P m o d e l 3
where w i (i = 1, 2, …, N) is the weight coefficient of the model N. w i [ 2 ,   2 ] is the range of weight coefficients by NNCT [27].
To improve the optimal weight coefficients w i (i = 1, 2, …, N), IWOA was employed to find the optimal solution for the ensemble weight coefficients. Before optimization, the objective equation needed to be confirmed first. The objective function of this paper is set by Equation (18). When the predefined minimum value of the objective function or the maximum iterations was reached, the optimization process was terminated. Nevertheless, the search boundary of the WOA is set to [−2, 2], the nesting dimension is 5 and the maximum number of iterations is 500.
-
Step 3: Assemble forecasting results:
Through the above steps, the overall prediction results of the VMs were obtained. Then, the prediction results were combined to obtain the final result.

4. Study Areas and the Evaluation Criteria

4.1. Data Description

In this paper, the PM2.5 concentration data from the Environmental Protection of the People’s Republic of China (http://www.mep.gov.cn/) were collected to verify the performance of the proposed model. The selected daily PM2.5 concentration sample data are for Beijing, Tianjin, Baoding and Shijiazhuang from 1 August 2015 to 31 August 2017. The total data number of daily PM2.5 concentration for each city were 763. In each experiment, the first 572 data (approximately 75% of the total data) of each VM were used for training subsection, and the rest were the test subsection. When all predicted VMs were integrated into the overall result, the 189 pieces of data (about 25%) in the test results were used for optimizing the weights of the ensemble model and the rest were used for model testing.

4.2. Model Assessment Standards

To effectively assess the prediction performance of the developed model, four popular error criteria, shown in Table 1, were employed to assessment the prediction capacity of the developed model. Smaller values denote better prediction performance.
Here, y n and y ^ n present the actual and predicted values at time n, respectively. N denotes the sample size.

5. Results and Analysis

5.1. Data Decomposition by VMD

In the proposed VMD-IWOA ensemble model, the original PM2.5 concentration sequence is first decomposed into several independent VMs by using VMD. However, too many VMs introduce new problems. During the integrated prediction process, each VM generates estimation errors, and too many VMs cause an accumulation of errors. It also increases the time consumed in a single prediction step. To prevent the above problems, the entire VMs were restructured into three VMs and a residual.

5.2. The Process of Ensemble Forecast on VMs

The BPNN, ANFIS, ANFIS-FCM and GMDH prediction models were applied to forecast each VM, which reconstructed in Section 5.1. Additionally, then, the ensemble model integrates the results of the four prediction models on each VM, and optimizes the weights of the four prediction results based on IWOA. Before the simulation, the parameters of the four neural network model need to be initialized. The input nodes of the neural network are set to four, the hidden nodes to nine and the output nodes to 1. Besides, the rolling single-step forecasting operation method based on PM2.5 concentration data of four cities is used to test the predictive performance. The detailed experimental parameters of the four neural networks are shown in Table 2.
Table 3 shows the prediction results of the single models and the proposed ensemble model for each VM. To evaluate model performance, the RMSE was utilized as a model evaluation index. As can be seen from Table 3, each model performed optimally predictive behavior at a particular VM. For instance, the experimental results in Beijing were shown as follows: the BPNN provides the lowest RMSE values among all single models at VM2 and VM3, while at VM1 and residual, GMDH has the lowest RMSE values. The prediction results in Tianjin show that the ANFIS presents the best results at VM1. The FCM performs best at VM3. At VM2 and Residual, the GMDH provides the best results. The experimental results in Baoding show that among all of the single models, the RMSE value was lower than those of the other methods at VM1 and Residual, when the ANFIS was applied. At VM2 and VM3, the GMDH presents the optimal results. The forecasting results in Shijiazhuang reveal that the GMDH performs better than the others at VM2, VM3 and residual while ANFIS performs the best at VM1.
Based on the above analysis, it can be revealed that each model has its advantages on the particular VMs. A single prediction model cannot be used to predict all decomposition signals uniformly. Thus, the most suitable model is selected according to the different conditions, which reveals that an ensemble model can incorporate the virtues of multiple individual models to overcome the limitations of individual models. Therefore, this study proposed an ensemble model based on the IWOA to seek the best weight coefficients of the ensemble model. The searching boundary is set in [−2, 2] based on the NNCT, and the RMSE criteria is used as fitness function of IWOA. Table 3 presents the best weights and final results of the ensemble model. By comparing with each single model, it indicates that the developed ensemble model can give the desired prediction results.
Comparing the ensemble model with BPNN, ANFIS, FCM and GMDH, the average RMSE of four cities at VM1 was reduced by 26.10%, 2.62%, 7.80% and 3.97%, respectively; At VM2, the average RMSE of four cities was reduced by 5.81%, 11.15%, 11.51% and 3.59%; At VM3, the average RMSE of four sites was reduced by 7.19%, 58.21%, 13.92% and 6.22%, respectively. For Residual, the average RMSE of four sites was reduced by 17.79%, 33.09%, 22.27% and 7.51%, respectively. Consequently, it can be seen that compared with the single models BPNN, ANFIS, FCM and GMDH, the forecasting result of the ensemble model is significantly improved on each VM component.

5.3. Model Performance Evaluation and Comparison

To evaluate the proposed ensemble model, three types of model comparison experiments were designed to compare the proposed ensemble model with other individual models, VMD-based models, and existing benchmark models.

5.3.1. Experiment 1: The Comparison between the Ensemble Model and VMD-Based Models

The experiment compares four VMD-based prediction models with the developed ensemble model. The four VMD-based models are VMD-BPNN, VMD-ANFIS, VMD-FCM and VMD-GMDH, which were constructed to emphasize important usages of the data decomposition technology. The corresponding improvement of the developed ensemble model and the VMD-based models are shown in Table 4 and Figure 2. By comparing the ensemble model with the VMD-BPNN, VMD-ANFIS, VMD-FCM and VMD-GMDH, we can conclude that the ensemble model significantly outperforms the other VMD-based models according to four evaluation criteria. For example, in Beijing, the ensemble model leads to 2.3843, 10.6660, 3.6867 and 2.1953 reductions in MAE, 5.4454, 21.3895, 11.6926 and 9.7510 reductions in RMSE, 0.3159, 11.9748, 12.3061 and 5.3553 reductions in MAPE, 5.2795, 21.3508, 11.6465 and 9.6318 reductions in TIC to compare with VMD-BPNN, VMD-ANFIS, VMD-FCM and VMD-GMDH, respectively. In addition, Figure 2 illustrates the comparison of actual values and the forecast values. The predicted results from the developed ensemble model are better than other VMD-based models.

5.3.2. Experiment 2: The Comparison between the Ensemble Model and Individual Models

This experiment used four individual models to make comparison with the developed ensemble model. The four individual models are BPNN, ANFIS, FCM and GMDH. Table 5 indicates the comparison forecasting results between ensemble model and other single models. From Table 5, by comparing the ensemble model with the BPNN, ANFIS, FCM and GMDH, there are significant improvements in the predictions of the proposed model. For example, in Beijing, the ensemble model leads to 66.4829, 71.3965, 67.8848 and 82.5946 reductions in MAE, 65.7865, 73.3943, 7.7401and 81.1458 reductions in RMSE, 67.7547, 71.7270, 67.5358 and 83.5715 reductions in MAPE, 65.6355, 73.1804, 67.5553 and 80.7598 reductions in TIC to compare with BPNN, ANFIS, FCM and GMDH, respectively. Besides, Figure 3 presents the comparison between the actual values and the forecast values. The forecast results from the developed ensemble model are better than other single models.

5.3.3. Experiment 3: The Comparison between the Proposed Model and the Existing Models

This part was conducted to further verify that the suggested hybrid decomposition–ensemble method can effectively improve performance prediction. Several existing models widely used in environmental prediction were applied to conduct comparative studies to access the suggested models. The existing models include two simple algorithms (i.e., ARIMA and RBFNN) and three hybrid algorithms (i.e., SSA-ENN, EEMD-GRNN and EEND-WOA-BPNN). The results of the comparative study are given in Table 6 and Figure 4. It can be seen from Table 6 and Figure 4 that the values of MAE, RMSE, MAPE and TIC of the developed model are all lower than the other existing models, which further shows the prediction performance of the developed ensemble model has obvious advantages. For example, comparing the proposed model with ARIMA, RBFNN, SSA-ENN, EEMD-GRNN and EEND-WOA-BPNN, the MAPE of Beijing was reduced by 94.01%, 91.14%, 90.22%, 86.73% and 69.20%, respectively. For Tianjing, the average MAPE of was reduced by 92.13%, 89.47%, 89.39%, 82.92% and 57.96%, respectively. For Baoding, the average MAPE was reduced by 94.01%, 91.14%, 90.22%, 86.73% and 69.20%, respectively. For Shijiazhuang, the average MAPE was reduced by 94.01%, 91.14%,90.22%, 86.73% and 69.20%, respectively.
In addition, the error mean and error STD are also used to evaluate the models’ accuracy and stability, and the results shows that the developed model has higher accuracy and stability than other existing models. Therefore, it can be concluded that the proposed ensemble model can be successfully and effectively employed for PM2.5 concentration prediction compared with existing models. Furthermore, the proposed ensemble model has the following highlights compared to previous works [15,16]: 1. the data decomposition; 2. multi-model integration prediction; and 3. the optimized ensemble pattern weighting coefficients.

6. Conclusions

Reliable and precise PM2.5 concentration forecasting is important for air quality early warning and pollution control. Owing to uncertainties and unstable of the PM2.5 datasets, the original PM2.5 series are very difficult to forecast accurately. Thus, it is still a challenging task to predict and simulate the PM2.5 reasonably. In this study, a new hybrid decomposition–ensemble learning paradigm, which based on variation mode decomposition (VMD) and modified whale-optimization algorithm (IWOA), is proposed to predict the PM2.5 concentration. In this developed paradigm, the VMD method was employed to decompose the original PM2.5 sequence into several VM series for forecasting. The prediction results show that the single prediction model used for pollution concentration prediction has limited capability and is not appropriate for all VMs. To this end, an ensemble model, based on four individual forecasting approaches, BPNN, ANFIS, FCM and GMDH, is proposed for predict all the VM components. Furthermore, in order to ascertain the best ensemble weight coefficients, an improved Whale Optimization Algorithm, named IWOA, is proposed and the final forecasting results were achieved by reconstructing the precise sequence. The main contributions of this paper are summarized as follows: (1) A new decomposition–ensemble learning paradigm is developed for PM2.5 concentration forecasting. (2) The VMD technique is adopted to decompose the primary PM2.5 series. (3) ANFIS, ANFIS-FCM and GMDH are utilized for PM2.5 forecasting. (4) An improved heuristic algorithm, IWOA, is developed to improve the weight coefficients of the ensemble model.
To evaluate the developed model, daily PM2.5 sequence from four cities located in Jing-Jin-Ji area of China were collected as the test cases for the comparison study. The comparison results indicated that the developed ensemble model is superior to comparison models, include four VMD-based models, four individual models, two benchmark models and three existing models. Thus, the developed ensemble model provides an effective forecasting ability, especially for the highly volatile and irregular data (e.g., PM2.5 concentration) and can be a powerful tool for decision makers in air quality monitoring and early warning system.

Author Contributions

Conceptualization, H.G. and X.H.; Methodology, H.G. and X.H.; Investigation, Y.G., W.Z. and Z.Q.; Data curation, Y.G., W.Z. and Z.Q.; Writing-original draft preparation, H.G. and X.H.; Writing-review and editing, Y.G., W.Z. and Z.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Project of China Ministry of Education for Philosophy and Social Science: Big Data Driven Risk Research on City’s Public Safety (Grant No. 16JZD023) and the National Natural Science Foundation project (41225018).

Data Availability Statement

In this paper, the PM2.5 concentration data from the Environmental Protection of the People’s Republic of China (http://www.mep.gov.cn/) were collected to verify the performance of the proposed model.

Acknowledgments

This work was supported by the Key Project of China Ministry of Education for Philosophy and Social Science: Big Data Driven Risk Research on City’s Public Safety (Grant No. 16JZD023) and the National Natural Science Foundation project (41225018).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Abdoos, A.A. A new intelligent method based on combination of VMD and ELM for short term wind power forecasting. Neurocomputing 2016, 203, 111–120. [Google Scholar] [CrossRef]
  2. Asadollahfardi, G.; Zangooei, H.; Aria, S.H. Predicting PM2.5 Concentrations Using Artificial Neural Networks and Markov Chain, a Case Study Karaj City. Asian J. Atmos. Environ. 2016, 10, 67–79. [Google Scholar] [CrossRef] [Green Version]
  3. Elbayoumi, M.; Ramli, N.A.; Yusof, N.F.F.M. Development and comparison of regression models and feedforward backpropagation neural network models to predict seasonal indoor PM2.5–10 and PM2.5 concentrations in naturally ventilated schools. Atmos. Pollut. Res. 2015, 6, 1013–1023. [Google Scholar] [CrossRef]
  4. Feng, X.; Li, Q.; Zhu, Y.; Hou, J.; Jin, L.; Wang, J. Artificial neural networks forecasting of PM2.5 pollution using air mass trajectory based geographic model and wavelet transformation. Atmos. Environ. 2015, 107, 118–128. [Google Scholar] [CrossRef]
  5. Fernando, H.; Mammarella, M.; Grandoni, G.; Fedele, P.; Di Marco, R.; Dimitrova, R.; Hyde, P. Forecasting PM10 in metropolitan areas: Efficacy of neural networks. Environ. Pollut. 2012, 163, 62–67. [Google Scholar] [CrossRef]
  6. Dag, O.; Yozgatligil, C. GMDH: An R Package for Short Term Forecasting via GMDH-Type Neural Network Algorithms. R J. 2016, 8, 379–386. [Google Scholar] [CrossRef] [Green Version]
  7. Dragomiretskiy, K.; Zosso, D. Variational Mode Decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  8. Chen, Y.-C.; Lei, T.-C.; Yao, S.; Wang, H.-P. PM2.5 Prediction Model Based on Combinational Hammerstein Recurrent Neural Networks. Mathematics 2020, 8, 2178. [Google Scholar] [CrossRef]
  9. Wang, X.; Yuan, J.; Wang, B. Prediction and analysis of PM2.5 in Fuling District of Chongqing by artificial neural network. Neural Comput. Appl. 2020, 1–8. [Google Scholar] [CrossRef]
  10. Jian, L.; Zhao, Y.; Zhu, Y.-P.; Zhang, M.-B.; Bertolatti, D. An application of ARIMA model to predict submicron particle concentrations from meteorological factors at a busy roadside in Hangzhou, China. Sci. Total Environ. 2012, 426, 336–345. [Google Scholar] [CrossRef] [PubMed]
  11. Konovalov, I.; Beekmann, M.; Meleux, F.; Dutot, A.L.; Foret, G. Combining deterministic and statistical approaches for PM10 forecasting in Europe. Atmos. Environ. 2009, 43, 6425–6434. [Google Scholar] [CrossRef]
  12. Li, X.; Peng, L.; Yao, X.; Cui, S.; Hu, Y.; You, C.; Chi, T. Long short-term memory neural network for air pollutant concentration predictions: Method development and evaluation. Environ. Pollut. 2017, 231, 997–1004. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, B.; Wang, L.; Jin, Y.-H.; Tang, F.; Huang, D.-X. Improved particle swarm optimization combined with chaos. Chaos Solitons Fractals 2005, 25, 1261–1271. [Google Scholar] [CrossRef]
  14. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  15. Qiao, W.; Tian, W.; Tian, Y.; Yang, Q.; Zhang, J. The forecasting of PM2.5 using a hybrid model based on wave-let transform and an improved deep learning algorithm. IEEE Access 2019, 7, 142814–142825. [Google Scholar] [CrossRef]
  16. Liu, H.; Chen, C. Prediction of outdoor PM2.5 concentrations based on a three-stage hybrid neural network model. Atmos. Pollut. Res. 2020, 11, 469–481. [Google Scholar] [CrossRef]
  17. Wang, D.; Liu, Y.; Luo, H.; Yue, C.; Cheng, S. Day-Ahead PM2.5 Concentration Forecasting Using WT-VMD Based Decomposition Method and Back Propagation Neural Network Improved by Differential Evolution. Int. J. Environ. Res. Public Health 2017, 14, 764. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Niu, M.; Wang, Y.; Sun, S.; Li, Y. A novel hybrid decomposition-and-ensemble model based on CEEMD and GWO for short-term PM2.5 concentration forecasting. Atmos. Environ. 2016, 134, 168–180. [Google Scholar] [CrossRef]
  19. Perez, P.; Reyes, J. An integrated neural network model for PM10 forecasting. Atmos. Environ. 2006, 40, 2845–2851. [Google Scholar] [CrossRef]
  20. Qin, S.; Liu, F.; Wang, J.; Sun, B. Analysis and forecasting of the particulate matter (PM) concentration levels over four major cities of China using hybrid models. Atmos. Environ. 2014, 98, 665–675. [Google Scholar] [CrossRef]
  21. Qu, Z.; Zhang, W.; Mao, W.; Wang, J.; Liu, C.; Zhang, W. Research and application of ensemble forecasting based on a novel multi-objective optimization algorithm for wind-speed forecasting. Energy Convers. Manag. 2017, 154, 440–454. [Google Scholar] [CrossRef]
  22. World Health Organization. World Health Statistics; World Health Organization: Geneva, Switzerland, 2011.
  23. Xiao, L.; Shao, W.; Liang, T.; Wang, C. A combined model based on multiple seasonal patterns and modified firefly algorithm for electrical load forecasting. Appl. Energy 2016, 167, 135–153. [Google Scholar] [CrossRef]
  24. Xiao, L.; Wang, J.; Dong, Y.; Wu, J. Combined forecasting models for wind energy forecasting: A case study in China. Renew. Sustain. Energy Rev. 2015, 44, 271–288. [Google Scholar] [CrossRef]
  25. Yang, Y.; Chen, Y.; Wang, Y.; Li, C.; Li, L. Modelling a combined method based on ANFIS and neural network improved by DE algorithm: A case study for short-term electricity demand forecasting. Appl. Soft Comput. 2016, 49, 663–675. [Google Scholar] [CrossRef]
  26. Zhang, N.; Lin, A.; Shang, P. Multidimensionalk-nearest neighbor model based on EEMD for financial time series forecasting. Phys. A Stat. Mech. Appl. 2017, 477, 161–173. [Google Scholar] [CrossRef]
  27. Zhou, Q.; Jiang, H.; Wang, J.; Zhou, J. A hybrid model for PM 2.5 forecasting based on ensemble empirical mode decomposition and a general regression neural network. Sci. Total Environ. 2014, 496, 264–274. [Google Scholar] [CrossRef]
Figure 1. The main procedures of the combinatorial model proposed in the paper. Step 1: Variation mode decomposition (VMD) process; Step 2: Ensemble forecasting on variational modes (VMs); Step 3: Assemble forecasting results.
Figure 1. The main procedures of the combinatorial model proposed in the paper. Step 1: Variation mode decomposition (VMD) process; Step 2: Ensemble forecasting on variational modes (VMs); Step 3: Assemble forecasting results.
Ijerph 18 01024 g001
Figure 2. Forecasting results of the ensemble model and VMD-based models.
Figure 2. Forecasting results of the ensemble model and VMD-based models.
Ijerph 18 01024 g002
Figure 3. Forecasting results of the ensemble model and individual models.
Figure 3. Forecasting results of the ensemble model and individual models.
Ijerph 18 01024 g003
Figure 4. Forecasting results of the ensemble model and existing models.
Figure 4. Forecasting results of the ensemble model and existing models.
Ijerph 18 01024 g004
Table 1. Four evaluation rules.
Table 1. Four evaluation rules.
MetricEquationDefinition
MAE MAE = 1 N n = 1 N | y n y n | The average absolute forecast error of n times forecast results
RMSE RMSE = ( 1 N n = 1 N ( y n y n ) 2 ) 1 / 2 The root mean-square forecast error
MAPE MAPE = 1 N n = 1 N | y n y n y n | × 100 % The average of absolute error
TIC TIC = 1 N n = 1 N ( y n y ^ n ) 2 1 N n = 1 N y n 2 + 1 N n = 1 N y ^ n 2 Theil’s inequality coefficient
Table 2. Experiment parameters of artificial neural networks (ANNs).
Table 2. Experiment parameters of artificial neural networks (ANNs).
ModelExperimental ParametersDefault Value
BPNNThe learning velocity0.01
The maximum number of trainings 1000
Training requirements precision0.00004
ANFISSpread of radial basis functions0.5
Training requirements precision0.00004
FCMThe maximum number of trainings1000
Spread of radial basis functions0.15
GMDHLearning rate0.1
Training requirements precision0.00004
Table 3. VMs forecasting results of the individual model and ensemble models in four cities.
Table 3. VMs forecasting results of the individual model and ensemble models in four cities.
ModelsVM1VM2VM3Residual
WeightsRMSEWeightsRMSEWeightsRMSEWeightsRMSE
BeijingBPNN0.068800.326950.408500.670431.240760.811300.517701.16300
ANFIS0.303310.302560.037010.75516−0.100141.383200.381671.17560
ANFIS-FCM0.006170.306690.171780.732790.188660.87710−0.487221.43640
GMDH0.633060.296720.388930.68048−0.343650.863980.561771.16110
Ensemble model-0.29096-0.65445-0.79141-1.05820
TianjingBPNN0.037030.310100.116450.619970.168650.79890−0.576731.34090
ANFIS0.344670.255750.351670.64571−0.170930.90740−0.003751.73350
ANFIS-FCM0.268160.25998−0.072720.662140.743910.75810−0.149571.23630
GMDH0.349970.256850.617070.595480.240710.780051.714330.99831
Ensemble model-0.24593-0.57482-0.73581-0.93073
BaodingBPNN−0.042980.36294−0.353750.655370.118251.193900.173850.82444
ANFIS0.724310.267150.338400.672080.126881.496900.407900.80775
ANFIS-FCM0.102330.286780.339000.65961−0.030611.35590−0.002820.85015
GMDH0.216070.276190.678520.652210.819671.137500.385440.86024
Ensemble model-0.26360-0.63175-1.08830-0.77429
ShijiazhuangBPNN−0.070310.29393−1.282080.604970.403461.09080−0.018580.78471
ANFIS1.015760.231830.500000.61216−0.040742.99250−0.078900.92644
ANFIS-FCM−0.086460.25504−0.201450.634280.153971.184300.323210.78460
GMDH0.141440.239882.000000.573800.504121.062800.759310.72264
Ensemble model-0.22918-0.55405-1.01030-0.70881
Table 4. The results of the ensemble model and other VMD-based models at four cities.
Table 4. The results of the ensemble model and other VMD-based models at four cities.
DatasetIndicatorEnsemble Model vs. VMD-BPNNEnsemble Model vs. VMD-ANFISEnsemble Model vs. VMD-ANFIS-FCMEnsemble Model vs. VMD-GMDH
BeijingIMAE (%)2.384310.66603.68672.1953
IRMSE (%)5.445421.389511.69269.7510
IMAPE (%)0.315911.974812.30615.3553
ITIC (%)5.279521.350811.64659.6318
TianjingIMAE (%)14.027014.54737.34313.5937
IRMSE (%)18.048421.993911.84527.1982
IMAPE (%)14.559212.925910.84813.3801
ITIC (%)18.027922.009711.78547.1898
BaodingIMAE (%)2.82957.30446.98630.5752
IRMSE (%)3.90047.47847.65382.3325
IMAPE (%)2.15287.05656.01850.0488
ITIC (%)3.81677.50207.68812.3468
ShijiazhuangIMAE (%)1.425118.70684.72264.8292
IRMSE (%)4.867849.98649.27885.0535
IMAPE (%)1.271615.50305.31384.0961
ITIC (%)4.821850.07159.39765.0563
Table 5. The results of the ensemble model and other single models at four cities.
Table 5. The results of the ensemble model and other single models at four cities.
DatasetIndicatorEnsemble Model vs. BPNNEnsemble Model vs. ANFISEnsemble Model vs. ANFIS-FCMEnsemble Model vs. GMDH
BeijingIMAE (%)68.653771.924869.189580.9890
IRMSE (%)66.639576.964468.159678.5772
IMAPE (%)69.679268.009858.932079.7105
ITIC (%)66.382976.707967.949677.7231
TianjingIMAE (%)66.482971.396567.884882.5946
IRMSE (%)65.786573.394367.740181.1458
IMAPE (%)67.754771.727067.535883.5715
ITIC (%)65.635573.180467.555380.7598
BaodingIMAE (%)87.947390.145989.088888.2149
IRMSE (%)88.037190.955888.038287.3972
IMAPE (%)88.024089.655589.264788.3508
ITIC (%)87.741690.639487.790887.0859
ShijiazhuangIMAE (%)88.338488.739689.161688.0327
IRMSE (%)88.818188.932089.578888.2980
IMAPE (%)87.757488.445088.812387.8461
ITIC (%)88.801888.892189.488088.1406
Table 6. Comparison of prediction performances with existing models.
Table 6. Comparison of prediction performances with existing models.
DatasetIndicatorMAERMSEMAPETICError MeanError STD
BeijingARIMA26.486536.237492.82960.3176−0.131736.3334
RBFNN17.081321.765362.72640.1751−11.970318.2263
SSA-ENN14.823718.718256.85340.1593−6.426617.6271
EEMD-GRNN11.356913.857541.91010.1176−6.021212.5142
EEMD-WOA-BPNN6.37487.783218.05440.0647−5.24815.7629
Pro. Ensemble1.68432.33675.5600 0.02020.05772.3422
TianjingARIMA17.461324.695735.40340.2059−1.627324.7075
RBFNN12.506216.278726.46800.1376−2.293316.1591
SSA-ENN12.463616.441326.27100.1335−5.659315.4776
EEMD-GRNN7.68089.788116.31860.0815−2.85099.3886
EEMD-WOA-BPNN3.20473.93046.62890.0328−0.80123.8581
Pro. Ensemble1.47452.09182.78690.01760.00612.0973
BaodingARIMA20.063426.721831.99560.1888−3.988026.4927
RBFNN15.037019.287725.58080.1376−5.402318.5648
SSA-ENN13.411417.431624.11570.1275−4.201016.9627
EEMD-GRNN9.581812.433117.28510.0902−4.282111.7035
EEMD-WOA-BPNN5.09886.76049.22800.0495−2.21836.4030
Pro. Ensemble1.49262.10292.44270.0156−0.04982.1079
ShijiazhuangARIMA15.599119.936325.43030.15411.955319.8929
RBFNN13.498017.789923.62750.1310−5.971716.8022
SSA-ENN11.131915.362418.54690.12040.926615.3752
EEMD-GRNN15.558218.548429.29070.1326−11.155214.8584
EEMD-WOA-BPNN11.996415.274022.62240.1093−9.833311.7187
Pro. Ensemble1.40041.87172.39300.0143−0.04491.8762
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guo, H.; Guo, Y.; Zhang, W.; He, X.; Qu, Z. Research on a Novel Hybrid Decomposition–Ensemble Learning Paradigm Based on VMD and IWOA for PM2.5 Forecasting. Int. J. Environ. Res. Public Health 2021, 18, 1024. https://doi.org/10.3390/ijerph18031024

AMA Style

Guo H, Guo Y, Zhang W, He X, Qu Z. Research on a Novel Hybrid Decomposition–Ensemble Learning Paradigm Based on VMD and IWOA for PM2.5 Forecasting. International Journal of Environmental Research and Public Health. 2021; 18(3):1024. https://doi.org/10.3390/ijerph18031024

Chicago/Turabian Style

Guo, Hengliang, Yanling Guo, Wenyu Zhang, Xiaohui He, and Zongxi Qu. 2021. "Research on a Novel Hybrid Decomposition–Ensemble Learning Paradigm Based on VMD and IWOA for PM2.5 Forecasting" International Journal of Environmental Research and Public Health 18, no. 3: 1024. https://doi.org/10.3390/ijerph18031024

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop