Next Article in Journal
Magnetic Field Characteristics and Stator Core Losses of High-Speed Permanent Magnet Synchronous Motors
Previous Article in Journal
Experimental Investigation of the Spatial and Temporal Evolution of the Tangential and Normal E-Field Components along the Stress Grading System of a Real Stator Bar
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Objective Particle Swarm Optimization Algorithm for Multi-Step Electric Load Forecasting

1
School of Information Science and Engineering, Lanzhou University, Lanzhou 730000, China
2
Department of Mathematics and Computer Science, Free University of Berlin, 14195 Berlin, Germany
3
School of Information Engineering, Zhengzhou University, Zhengzhou 450000, China
*
Author to whom correspondence should be addressed.
Energies 2020, 13(3), 532; https://doi.org/10.3390/en13030532
Submission received: 9 December 2019 / Revised: 9 January 2020 / Accepted: 13 January 2020 / Published: 21 January 2020

Abstract

:
As energy saving becomes more and more popular, electric load forecasting has played a more and more crucial role in power management systems in the last few years. Because of the real-time characteristic of electricity and the uncertainty change of an electric load, realizing the accuracy and stability of electric load forecasting is a challenging task. Many predecessors have obtained the expected forecasting results by various methods. Considering the stability of time series prediction, a novel combined electric load forecasting, which based on extreme learning machine (ELM), recurrent neural network (RNN), and support vector machines (SVMs), was proposed. The combined model first uses three neural networks to forecast the electric load data separately considering that the single model has inevitable disadvantages, the combined model applies the multi-objective particle swarm optimization algorithm (MOPSO) to optimize the parameters. In order to verify the capacity of the proposed combined model, 1-step, 2-step, and 3-step are used to forecast the electric load data of three Australian states, including New South Wales, Queensland, and Victoria. The experimental results intuitively indicate that for these three datasets, the combined model outperforms all three individual models used for comparison, which demonstrates its superior capability in terms of accuracy and stability.

1. Introduction

At present, electricity can be stored on a large scale by hydropower. It is completed at the same time as the production, transmission and distribution. Maintaining the balance between power generation, supply, and use is essential to ensure the quality of power, which is an important part of the stable operation of a power system. In view of the real-time characteristics of electricity and the uncertainty change of the electric load, electric load forecasting has gradually begun to attract people’s attention. Improving the accuracy of short-term electric load forecasting is important for suppliers and consumers. The accuracy results can provide scientific guidance to large-scale factories, thus avoiding the waste of resources.
The electric load forecasting method can be divided into traditional statistical learning-based methods, machine learning methods, and a combined model as well as hybrid forecasting methods [1]. Traditional statistical learning-based methods are accurate and easy to implement, but when the fluctuation of the electric load is severe, they usually perform poorly. These methods include regression analysis [2], trend extrapolation [3], time series method [4,5], autoregressive integrated moving average (ARIMA) [6,7,8], and so on. Liu et al. [9] built a large vector autoregression (VAR) model to forecast three important weather variables for 61 cities around the United States. In addition, traditional statistical learning-based methods generally use statistical models to establish mathematical prediction formulas. However, because the electric load is affected by many uncertain factors, such as time, region, climate, and economic environment, it is difficult to establish an accurate mathematical model for prediction.
The machine learning methods have a strong fitting ability, can fully reflect the nonlinear characteristics of the electric load [10], and can get better forecasting results in short-term electric load forecasting. These methods include the expert system method [11], fuzzy clustering method [12], wavelet analysis method [13,14], artificial neural network [15,16,17,18,19], and so on. For example, Yang et al. [20] first used the autocorrelation function to select informative input variables, and then applied the least-squares support-vector machine (LSSVM) to predict the electric load. The experimental results show that the proposed method can improve the forecasting accuracy. Using the cluster analysis method to divide 12-month data into seasonal-based data, and considering multi-input and multi-output, Bedi and Toshniwal [21] used long short-term memory (LSTM) to forecast the long-term electric load. Eseye et al. [22] first used the binary genetic algorithm (BGA) to select features, then applied gaussian process regression (GPR) to evaluate the fitness function of the feature, and finally employed feedforward artificial neural network (FFANN) to predict the electric load. The annual MAPE value of the proposed method can reach 1.96%. Liu et al. [23] applied the stacked denoising auto-encoder (DAE) method to forecast electricity demand in a city in southern China. Moreover, distance-based predictive clustering [24,25] is used in time series forecasting.
In short-term electric load forecasting, each method has defects, and it is rare for an individual forecasting model to perform well in all cases due to its superiorities and weaknesses. For example, the linear regression model cannot capture the nonlinear and seasonal features of an electric load, and the artificial neural network may fall into the local optimum [26]. Therefore, when there are multiple models that can be used, without considering the best model, we can use a hybrid model and combined method. For example, Foley et al. [27] applied the optimization algorithm to optimize the model parameters, and introduced denoising methods into electric load preprocessing. Li et al. [28] proposed a hybrid model based on wavelet transform, ELM, and improved artificial bee colony compared with the traditional neural network, the number of iterations of the hybrid model is greatly reduced, and the forecasting accuracy of the hybrid model is higher. A novel hybrid model based on the support vector machine (SVM) and cuckoo search optimization (CSO) is presented in [29]. The data is first denoised by singular spectrum analysis (SSA), and then the parameters of SVM are optimized by CSO to overcome the defects caused by manual selection. Xiao et al. [30] used CSO to determine the weight coefficient of the combined model for electric load forecasting. The parameter setting of CSO is simpler, and the result is better than particle swarm optimization (PSO). Wang et al. [31] proposed a combined electric load forecasting model based on ARIMA, seasonal exponential smoothing, and weighted support vector machine. The adaptive PSO was used to determine the weight coefficient of the combined model, and the accuracy was higher than the single model. Liu et al. [32] proposed a hybrid prediction method based on SVM, ensemble empirical mode decomposition (EEMD), and the whale optimization algorithm (WOA), where EEMD is used for denoising and WOA is used to optimize the parameters of SVM. Similar to the literature [32], forecasting the Australian electric load, Zhang et al. [33] proposed a new combined forecasting method based on multi-objective optimization.
Although the above-mentioned method can obtain the expected forecasting results with given the electric load data, these methods usually do not consider the stability of time series prediction. This paper proposes a new combined forecasting model, which not only considers the accuracy of electric load forecasting but also considers the stability of electric load forecasting. The proposed model first uses three neural networks to forecast the electric load data separately, and then the combined theory is applied to combine the three intermediate forecasting results [10]. These three neural networks include extreme learning machine (ELM), recurrent neural network (RNN), and support vector machines (SVM). In addition, the multi-objective particle swarm optimization algorithm (MOPSO) is used to optimize the parameters of the combined model. As a special time series forecasting, multi-step electric load forecasting is usually much more difficult, because as the steps increase, the errors accumulate and the prediction accuracy decreases. In order to verify the proposed model, 1-step, 2-step, and 3-step were used to forecast the electric load data of three Australian states, including New South Wales, Queensland, and Victoria. Finally, we compared the forecasting results of the proposed model with three individual models and evaluated the four models using different evaluation metrics.
The main contribution of this paper is to propose a novel combined electric load forecasting model based on ELM, RNN, SVM, and MOPSO, which can improve the accuracy and stability of electric load forecasting. In addition, to evaluate the model presented in this paper more comprehensively, we used multi-step electric load forecasting and evaluated the forecasting results by three states of Australia. A scientific and reasonable model evaluation system examines whether the proposed prediction model achieves an excellent prediction effect. In this evaluation system, three performance metrics, i.e., the mean absolute percent error (MAPE), mean absolute error (MAE), and root mean square error (RMSE), were used to verify the predictive accuracy of the combined model.
The rest of this paper is organized as follows. Section 2 introduces the related methodology, which includes ELM, RNN, SVM, and MOPSO. The proposed combined model is presented in Section 3. Section 4 presents the experimental results. Section 5 concludes the whole paper.

2. Related Methodology

In this section, four components of the proposed combined forecasting model are presented, namely, the extreme learning machine, recurrent neural network, support vector machine, and multi-objective particle swarm optimization algorithm.

2.1. Extreme Learning Machine

Suppose the extreme learning machine model (ELM) has N samples ( x j , t j ) , where x j = [ x j 1 , x j 2 , , x j n ] represents the j -th input sample of the ELM network, and t j = [ t j 1 , t j 2 , , t j n ] is the expected output of the j -th input sample. If the number of hidden layer nodes in ELM is L , then the mathematical model of ELM can be described as:
i = 1 L β i g i ( x j ) = i = 1 L β i g ( w i x j + b i ) = y j ,       j = 1 , 2 , , N ,
where y j = ( y j 1 , y j 2 , , y j n ) T represents the actual output value of the ELM, w i = ( ω i 1 , ω i 2 , , ω i n ) T is the input weight vector between the input layer node and the i-th hidden layer node, b i is the i -th hidden layer node threshold, β i = ( β i 1 , β i 2 , , β i m ) T is the output weight vector between the i -th hidden layer node and the output layer node, and g ( · ) is the activation function, where the Sigmoid function is often used. At this point, assuming that the actual output of the ELM network is equal to the expected output, then:
i = 1 L β i g ( w i x j + b i ) = t j , j = 1 , 2 , , N .
The N equations in Equation (2) can be abbreviated as:
H β = T ,
where H represents the output matrix of the hidden layer, β is the output weight matrix, and T is the expected output matrix. The expanded formulas are as follows:
H = H ( w 1 , w 2 , , w L , b 1 , b 2 , , b L , x 1 , x 2 , , x N ) = [ g ( w 1 · x 1 + b 1 ) g ( w L · x 1 + b 1 ) g ( w 1 · x N + b 1 ) g ( w L · x N + b L ) ] N × L ,
β = [ β 1 T β L T ] L × m ,    T = [ t 1 T t L T ] N × m .
In the ELM algorithm, the input weight vector and the bias of the hidden layer can be given randomly. From Equation (4), it can be seen that the hidden layer output matrix, H , becomes a fixed matrix. At this time, the training of ELM can be translated into seeking the least squares solution of the output weight matrix, β . The output weight matrix, β , is calculated by the following formula:
β = H + T ,
where H + represents the Moore–Penrose generalized inverse of the hidden layer output matrix, H .

2.2. Recurrent Neural Network

Unlike feedforward neural networks, the recurrent neural network (RNN) can use historical information to help the current decision. RNN mainly uses the neurons with self-feedback to memorize the previous information and applies it to the calculation of the current output. The input of the hidden layer includes not only the input of the input layer at the current time but also the output of the hidden layer at the previous moment. That is, the output of the current time of a sequence is related to the output of the previous time, as shown in Figure 1.
From Figure 1, we can see that a simple RNN model consists of three layers: The input layer, hidden layer, and output layer, and the connection between the layers is fully connected. Where W I H represents the weight matrix between the input layer and the hidden layer, W H H represents the weight matrix between the previous hidden layer to the current hidden layer, and W H O represents the weight matrix between the hidden layer and the output layer. The weight matrix at each moment is shared, and the expanded RNN can be regarded as a multi-layer feedforward neural network with parameter sharing. Suppose the input sequence of RNN with length T is X = { X 1 , , X t , X t + 1 , X t + 2 , , X T } , the dimension of the input data, X t , at each moment is N , namely X t = ( x 1 , x 2 , , x N ) ; the number of nodes in the hidden layer is M , namely H t = ( h 1 , h 2 , , h M ) ; and the number of nodes in the output layer is P, namely Y t = ( y 1 , y 2 , , y P ) . Then, the hidden layer, H t , at time t can be calculated by Equations (7) and (8):
O t = W I H X t + W H H H t 1 + b h ,
H t = f H ( O t ) ,
where f H ( · ) is the activation function of the hidden layer and b h is the bias vector of the hidden layer. It is worth noting that when t = 0 , that is, when the model starts the first calculation, in order to improve the stability and the overall performance, we usually initialize H 0 to a number that is relatively small but not zero [34]. The calculation of the output layer, Y t , at time t can be obtained by the following equation:
Y t = f O ( W H O H t + b o ) ,
where f O ( · ) is the activation function of the output layer, b o is the bias vector of the output layer, and the feedforward process of the model starts from t = 1 .
As a model of supervised learning, the training goal of RNN is to minimize the objective function. In order to achieve the training objectives, RNN uses the backpropagation through time (BPTT) [35] to perform gradient calculation on the model to optimize the parameters of RNN.

2.3. Support Vector Machine

Given a set of points, M = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) } , x n R k . x n is the input feature variable, which belongs to either of the two categories according to its label y n { 1 , + 1 } . For the nonlinearly separable problem, a high-dimensional maximum margin hyperplane can be represented as follows:
y = b +   ω i y i K ( x ( i ) , x ) ,
where x is the test example, x ( i ) is the support vector, and K ( · ) is the kernel function. The kernel function can map low-dimensional feature variables to high-dimensional feature variables. Common choices of the kernel function include the polynomial kernel function and Gaussian radial basis function (RBF).
To obtain an optimal hyperplane, the above problem can be converted into the following quadratic programming (QP) problem:
Minimize   1 2 ω 2 + C i = 1 n ξ i Subject   to   { y i ( ω · ϕ ( x i ) + b ) + ξ i 1 ,   i = 1 , , n ξ i 0 ,   i = 1 , , n ,
where C is the penalty factor, ξ i is the slack variable, · is the dot product, ϕ ( · ) is the nonlinear map, ω is the normal vector of the hyperplane, and b is the bias.

2.4. Multi-Objective Particle Swarm Optimization (MOPSO)

2.4.1. Multi-Objective Optimization

For optimization problems that achieve more than one objective function, it is often referred to as multi-objective optimization. The multi-objective optimization problem can be described by the following mathematical form:
min   f ( x ) = [ f 1 ( x ) , f 2 ( x ) , , f n ( x ) ] ,
subject   to   { g i ( x ) 0 ,       i = 1 , 2 , , m h j ( x ) = 0 ,       j = 1 , 2 , , p ,
where x = [ x 1 , x 2 , , x n ] T represents the decision vector, f i ( x )   i = 1 , 2 , k represents the i -th objective function, and g i ( x ) , h j ( x ) : i = 1 , 2 , m ,   j = 1 , 2 , p represents the constraints of the target problem, respectively, and finally, the optimal solution of the total target, f ( x ) , should be found. In order to better understand the concept of optimal solutions, this paper introduces the following definition [36]:
Definition 1.
Pareto dominance.
Given two vectors x = ( x 1 , x 2 , , x t ) and y = y 1 , y 2 , , y t , x dominates y ( x y ) if:
i [ 1 , t ] ,   [ f ( x i ) f ( y i ) ] [ i [ 1 , t ] : f ( x i ) ] .
Definition 2.
Pareto optimality.
A pareto-optimal x belongs to X if:
y X   subject   to   F ( x ) F ( y ) .
Definition 3.
Pareto optimal set.
P s   : = { x , y X | F ( x ) F ( y ) } .
Definition 4.
Pareto optimal front.
A set includes the value of objective functions for pareto solutions set:
P f   : = { F ( x ) | x P s } .

2.4.2. MOPSO

For the problem of multi-objective function optimization, Moore J et al. [37] first tried to use particle swarm optimization (PSO) to explore multi-objective problems, and then other researchers [38,39] improved the PSO algorithm to solve the multi-objective optimization problem. In 2004, Coello et al. [40] proposed the multi-objective particle swarm optimization (MOPSO) algorithm, which has been widely used in many fields by researchers.
When the PSO algorithm performs single-objective optimization, once the leader in the particle population (the individual optimal solution and the global optimal solution after the last iteration) is determined in each iteration, the search direction is also determined. However, in the optimization process of multi-objective problems, each particle may have multiple different leaders, and one of these leaders needs to be selected to determine the search direction of the particles. In multi-objective optimization, such a group of leaders needs to be stored in a repository, and the repository contains the non-dominated solution set searched so far (generally, its storage location is different from the particle population, called the external archive). In the iterative update of the particle swarm, the algorithm will select the non-dominated solution from the repository as the leader. After the algorithm completes the iterative search, the solution set in the external archive is the output of the algorithm.

3. Proposed Model

In recent years, many combined methods have been proposed to improve the accuracy and performance of electric load forecasting [41]. The main issue in these combinations is how to connect and combine the models with each other so that they have the desired accuracy of prediction and simultaneously exert the benefits of single-mode models. In this paper, the components of the three artificial neural networks (ANNs) are combined, based on their weight, as follows:
Y ^ C o m , t = W 1 Y ^ S V M , t + W 2 Y ^ E L M , t + W 3 Y ^ R N N , t   ( t = 1 , 2 , , m ) ,
where the forecasted value of the combined model at time t is illustrated by f ^ C o m , t   , ( t = 1 , 2 , , m ) ; Y ^ S V M , t , Y ^ E L M , t , and Y ^ R N N , t are the forecasting results at time t by SVM, ELM, and RNN; and W i ,   ( i = 1 , 2 , 3 ) is the weight coefficient allocated to the three ANNs. The weights of each component of the combined model are calculated in such a way that the error values are minimized. Several different heuristic and meta-heuristic optimization techniques, such as genetic algorithms (GAs), differential evolution (DE), particle swarm optimization (PSO), etc., have been proposed in recent years in order to improve the forecasting performance. However, these methods only consider the accuracy of the combined model. In this paper, MOPSO is used to optimize the weight coefficient, which can not only improve the accuracy but also consider the stability of the combined model. The general flowchart of the proposed model is illustrated in Figure 2. From Figure 2, we can see that the proposed combined model includes the following three steps.
Step 1. Three ANNs, which include ELM, RNN, and SVM, are applied separately to forecast the electric load. After that, combined theory [10] is used to combine the three intermediate forecasting results.
Step 2. MOPSO is applied to optimize the weight coefficients of the combined model. When using MOPSO to get the optimal weight coefficient, it will get a lot of weights, and each weight is actually a particle. In this paper, X-Means [42] is used to cluster the particles, and then a cluster that contains the most particles is selected. The selected cluster has a center, and the particle closest to the center is the final weight.
Step 3. If the weight coefficient obtained in Step 2 does not reach the expected forecasting result, we continue to use MOPSO to optimize the weight coefficient; otherwise, the final forecasting result can be obtained.

4. Experimental Results and Discussion

In order to verify whether the combined model proposed in this paper can achieve accurate prediction effect when applied to electric load forecasting, the following two experiments were carried out. All the data collected for electric load forecasting in this experiment came from Australia. For the first experiment, three models, including ELM, RNN, and SVM, were used to forecast the electric load independently. In the second experiment, the combined model, which is aimed at improving the accuracy and stability, was applied for multi-step prediction.

4.1. Dataset

The original data used in this article includes three data sets from Australia, which are new data from New South Wales (NSW)) from May 2 2011 to July 3 2011, data from Queensland (QLD)) from April 20 2015 to June 21 2015, and data from Victoria (VIC) from April 30 2007 to July 1 2007. These data sets were obtained based on observation records with a length of two months. In addition, it is composed of data collected by every half hour from 00:00 to 23:30, so 48 observations could be obtained in one day. To test the stability of this model, datasets were collected from Monday to Sunday, so 48 values were obtained every day. Because eight weeks of data were collected, these data were used to predict values for the same day in the next week. Therefore, the dataset was divided into two parts, namely the training set and testing set. Because the data was recorded every half hour, the length of the training data sequence was 2688 (48*8*7), and the length of the testing data sequence was 336 (48*1*7).
For multi-step forecasting, the rolling prediction mechanism was applied in this paper to deal with the input vector every time the input data slides back once. The steps of forecasting minus one was the delayed operator. More specifically, for one-step-ahead forecasting, the delayed operator was 0; for two-steps-ahead forecasting, the delayed operator was 1; and for three-steps-ahead forecasting, the delayed operator was 2. In addition, every time, the output vector was 1. For the SVM and RNN used in this paper, the following parameters were used: The number of hidden layers in RNN was 4, and the regularization parameter and the kernel parameter of SVM were 100 and 0.1, respectively.

4.2. Forecasting Performance Metrics

There is no doubt that there is no perfect model for obtaining the best electric load forecasting performance. In the process of continuously evaluating the prediction results of various models, many evaluation indexes have appeared, but there is no unified standard. Therefore, in order to fully understand the advantages and disadvantages of a model, this paper used the most commonly used three indicators for the evaluation. These are the mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE), respectively, and the formulas are as follows:
MAE = 1 N n = 1 n | y n y ˜ n | ,
MAPE = 1 N n = 1 n | y n y ˜ n y n | × 100 % , ,
RMSE = 1 N n = 1 n ( y n y ˜ n ) 2 ,
where y n is the original value and y ˜ n is the predicted value, and the smaller the evaluation result value, the higher the accuracy of the model.

4.3. Experimental Results and Analysis

In this section, we first use ELM, RNN, and SVM to predict the electric load, respectively, and then adjust the prediction results by using these models as the three sub-models to construct the combined model. This paper is conducted in two parts: One is the single-model electric load forecasting, which consists of three components; and the other includes a comparison of the forecasting results of the combined model in three datasets against the single models. For the reason that the advantages of the combined model could be shown perfectly, the four forecasting results from New South Wales, Victoria, and Queensland were selected. To put the results in this paper in detail, in this section, the four forecasting models will be described in the following section.

4.3.1. The Forecasting Results of ELM

In order to evaluate ELM from various angles, MAE, MAPE, and RMSE were adopted to reveal the effect of ELM. Figure 3 shows the prediction results of a week. According to the analysis, it can be seen that ELM can achieve a good prediction effect during the week except on Monday, and the best performance is on Sunday. The MAE, MAPE, and RMSE are 66.2811, 0.7571, and 84.2174, respectively. Due to the six days’ good performance, we know that ELM performs well when using this dataset. When looking at the value of Sunday, the forecasting results of the multi-step are shown in Figure 3. We can see that compared to the original values and predicted values, the coincidence of the two curves is very high, thus it can be seen that ELM can obtain a good prediction effect.
However, after observing the more detailed data in Table 1, it was found that only three days out of seven days in one-week ELM achieved a better prediction value while the prediction effect of the remaining four days was not ideal, especially as the difference between the predicted value on Monday and the original value was relatively large. Therefore, there is no doubt that we can draw the following conclusions:
(1) ELM can achieve a perfect prediction effect on most days of the week, but it does not have a stable prediction ability.
(2) The prediction effect of ELM on each day is relatively average, being neither good nor bad.

4.3.2. The Forecasting Results of RNN

Figure 4 shows the prediction results of the RNN model. Similarly, the performance was also evaluated by MAE, MAPE, and RMSE. As can be seen from part A of Figure 4, on the whole, the trend of the predicted results of RNN is similar to ELM, and what impresses us most is that RNN had better predictive values on Thursday than on the other six days. What is more, the RNN model not only performs poorly like ELM on Monday but also has a higher prediction value on Tuesday. In order to estimate the stability of this model intuitively, we used Thursday’s forecasting result as an example. The experimental results on Thursday of RNN are shown in part B of Figure 4. Although Thursday represents the best prediction effect of the RNN model, we cannot ignore the fact that the three curves of step 1, step 2, and step 3 on Thursday are not completely consistent with the actual value from the figure.
In Table 2, we can clearly see each value of the forecasting results of RNN. If the value of MAPE is less than 1.00, the prediction effect of this day is better, and the conditions are met on Thursday and Friday. In addition, the maximum, minimum, and fluctuation of MAE, MAPE, and RMSE can also be observed. Therefore, for the RNN model, the following conclusion can be drawn:
(1) The prediction effect of the RNN model is relatively average, and unacceptable extreme values may occur.
(2) The three evaluation metrics’ values of the RNN model are generally high.

4.3.3. The Forecasting Results of SVM

In part A of Figure 5, the prediction results obtained using SVM in NSW are displayed. According to the bar chart in Figure 5, the performance of Tuesday, Friday, Saturday, and Sunday is better, in which the best one of the four days can be obtained in a week. Besides, no one can deny the fact that the forecasting result of Wednesday was very unsatisfactory. In order to highlight the gap between the original value and the predicted value, it can be clearly seen from part B of Figure 5 that the predicted result of SVM on Friday almost coincides with the curve of the original data. On the one hand, we can see when the accuracy at all the moments is the highest. On the other hand, we can see that the curve of SVM was less consistent with the actual data at a specific time point
From Table 3, we can see that the SVM model does not behave well on one day, as the lowest value of MAPE is 1.2377 on Sunday and the highest value of MAPE is 5.7218 on Thursday. Meanwhile, the value of MAE and RMSE can be easily observed.

4.3.4. The Forecasting Results of Combined

To further show that the combined model could capture the characteristics of each model effectively, the performance of the proposed model, which consists of ELM, RNN, and SVM, is displayed in this part. Firstly, we cannot ignore the fact that the three forecasting metric values (MAE, MAPE, and RMSE) of the three ANNs on Friday are very high, which also means that the result of the combined model on Friday is not accurate. After carefully observing part A of Figure 6, we can see that the three performance metrics are very average expect on Friday, and Figure 6 also shows that there is no negative value. From part B of Figure 6, we can directly know that there is almost no gap between the original value and the predicted value on Monday. Unlike NSW, SVM can perform well in local optimization, and the combined model performs good in all 48 moments. The performance of the combined model for each day is presented in Table 4. All values of MAPE are much smaller than the lower limits at 0.30, and the maximum value is also below 1.00.
When evaluating these four models (ELM, RNN, SVM, and the combined model) systematically, it is very obvious that the prediction effect of the combined model is superior to the three classic models by using the same dataset of NSW. Compared with the other three models, the combined model can obtain the minimum MAPE, with a value of 0.20.

4.3.5. The Forecasting Results of the Four Models and Comparisons

In this subsection, a holistic comparison with the other three methods (ELM, RNN, and SVM) is conducted to estimate the effectiveness of the combined model by putting it under broader time and space conditions. Generally speaking, the combined model has a stronger capability to forecast the electric load than the other three models. Figure 7 shows the prediction values and original values for the seven days of all four models in Queensland. As is vividly shown in the graph, the five curves are representative of the different values. These curves with different colors stand for electric load values by ELM, RNN, SVM, and the combined model, respectively. From Figure 7, there is no doubt that every model can get the best result in some special time points. Among the four models, it is noteworthy that the combined model curve is less consistent with the actual values while the ELM and RNN curves are more consistent with the actual values, with between 15 and 38 time points in one day expect on Monday. From the curves, it can be recognized that the combined model curve and the actual curve almost overlap. However, when we look at the curves on Saturday and Sunday, we could draw a conclusion that the four models only achieve better prediction in the first 10 observations. So, in order to more thoroughly estimate which model can achieve the best forecasting results synthetically, the results taking the three forecasting metrics (MAE, MAPE, and RMSE) into account are recorded in Table 5.
Table 5 reveals the calculated values of three evaluation metrics and the average values of four models for a week in Queensland. In order to compare the four models in detail, we considered MAE, RMSE, and MAPE to estimate which model had the best performance. When we contrasted the combined model with the other three models, the results were as follows.
The combined model vs, ELM, RNN, and SVM: Compared with the other three models, the presented combined model, which contains the lowest MAE, MAPE, and RMSE values in the days of the week, can still get the lowest average values, which means the combined model is suitable for forecasting the electric load in Queensland. In addition, the MAPE of the combined model is not more than 0.15 and is also less than 3.00. When compared with ELM, it could be observed that the great improvements on the forecasting effectiveness of the proposed model are fairly obvious. To be frank, the combined model achieved the maximum improvement of MAE, RMSE, and MAPE, which are 69.19%, 71.00%, and 68.37%, respectively. Similarly, when compared with RNN, the combined model achieved the minimum improvement of MAE, RMSE, and MAPE, which are 37.13%, 38.21%, and 37.28%. At last, a comparison with SVM was made, revealing that combined model reduced MAE by 62.52%, RMSE by 64.99%, and MAPE by 62.5%. In the following, the merits and demerits of the three models are recommended.
RNN vs. ELM and SVM: The three metrics of MAE, RMSE and MAPE of RNN have lower values than ELM and SVM expect on Wednesday. In general, in terms of the average value, RNN is better than ELM and SVM. Even so, compared with ELM, RNN led to a 50.99%, 53.06%, and 49.70% reduction in terms of MAE, RMSE, and MAPE. Compared with SVM, RNN reduced MAE by 40.39%, RMSE by 43.34%, and MAPE by 40.36%.
SVM vs. ELM: It is evident that the forecasting results based on the three error metrics (MAE, RMSE, and MAPE) of SVM are more precise than ELM except on Monday, Friday, and Saturday. Though there are three days of MAE, RMSE and MAPE of ELM are lower than SVM. Compared with the average values, SVM reduced MAE by 17.79%, RMSE by 17.16%, and MAPE by 15.66%.
To further illustrate the quality of the entire models, Figure 8 shows the errors, which are the differences between the actual values and the forecasting values of the four models in New South Wales, Queensland, and Victoria on Monday and Saturday. Though the error rates of the forecasting results are shown in three different sites on Monday, it can obviously be observed that the combined model curve is almost zero and as smooth as a line. On the one hand, that is to say, the combined model obtained the smallest errors with the original values. On the other hand, we can draw the conclusion that the performance of the combined model is remarkably steady overall. The error of the other three models has a great fluctuation range between the positive value of 600 and negative value of 500, among which errors of RNN are almost a line, but all error nodes are under zero. The curves of ELM and SVM fluctuate up and down irregularly. As mentioned above, the smaller the errors are, the better the model is. To sum up, the combined model has the best performance among the four approaches. This demonstrates that the combined model has high accuracy and stability.
At last, based on the analysis of the previous results, the final prediction effect of the four models is visible in Figure 9. The histogram, in terms of the total forecasting metrics of these four models for a week, is presented in Figure 9 part A in detail. The shorter the histogram is, the better the model is. It is intuitive that the red histogram of the combined model is the shortest, which means the combined model performs well. Based on the above results, we can obtain the final evaluation results (MAE, RMSE, and MAPE) from Figure 9 part B. In Figure 9 part B, 1 represents the result of ELM, 2 shows the result of RNN, 3 is the symbol of SVM, and 4 displays the combined model. From Figure 9 part B, firstly, we can see that the MAE, MAPE, and RMSE of the combined model are the shortest. Then, in the last histogram, the difference among the four models can be seen intuitively. The two parts of Figure 9 together reveal that it is important to generate forecasts in terms of forecasting accuracy and stability for a model in a week. It is worth noting that the mean value of the combined model is lower than the three individual models by the forecasting metrics. We can see the differences vividly in Figure 9.
In a word, based on prior results, the proposed combined model can effectively forecast the electric load with a high prediction performance and strong stability.

5. Conclusions

Accurate and stable forecasting results can not only help consumers to save the cast using in electricity but also provide assistance to suppliers to estimate electricity consumption in advance, thus achieving huge economic and social benefits for consumers and suppliers. With respect to the forecasting methods, there is no denying that some approaches cannot be adopted in practice. Furthermore, every single model has inevitable drawbacks, and the combined model takes full advantage of each single model and reduces the negative effects of the individual model.
In this paper, to overcome these drawbacks, this research proposed a novel combined model, which is based on a combination of ELM, RNN, and SVM and employed multi-objective particle swarm optimization (MOPSO). The combined model first used three ANNs, which include ELM, RNN, and SVM, separately to forecast the electric load. Then, the three intermediate forecasting results were obtained. After that, the MOPSO was applied to optimize the weight coefficients of the integrated model for multi-step electric load forecasting. If the weight coefficient could not achieve better results, we continued to use MOPSO to select the weight coefficient, so it is not surprising that the combined model obtained a good prediction result. Considering the time and environmental factors, by simulating the electric load data of three sites from Australia and comparing the combined model with the other three single methods, we found that the behavior of the proposed combined model was superior to the individual model. All in all, the combined model showed great progress in accuracy and stability.

Author Contributions

Conceptualization, Z.S. and Y.C. (Yao Chen); Methodology, Z.S.; Software, Y.Y.; Validation, Y.C. (Yao Chen), Y.C. (Yanhua Chen); Formal Analysis, Z.S. and Y.C. (Yanhua Chen); Investigation, Y.C. (Yanhua Chen); Resources, Y.Y.; Data Curation, Y.C. (Yao Chen); Writing-Original Draft Preparation, Z.S.; Writing-Review & Editing, Z.S. and Y.Y.; Visualization, Z.S. and Y.C. (Yao Chen); Supervision, Y.C. (Yanhua Chen); Project Administration, Y.Y.; Funding Acquisition, Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by National Key R&D Program of China (Grant number 2018YFB1003205, 2017YFE0111900) and 2019 Gansu key talent project.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this paper.

References

  1. Zhang, J.; Wei, Y.M.; Li, D.; Tan, Z.; Zhou, J. Short term electricity load forecasting using a hybrid model. Energy 2018, 158, 774–781. [Google Scholar] [CrossRef]
  2. Velsink, H. Time Series Analysis of 3D Coordinates Using Nonstochastic Observations. J. Appl. Geod. 2016, 10, 5–16. [Google Scholar] [CrossRef] [Green Version]
  3. Meng, M.; Wang, L.; Shang, W. Decomposition and forecasting analysis of China’s household electricity consumption using three-dimensional decomposition and hybrid trend extrapolation models. Energy 2018, 165, 143–152. [Google Scholar] [CrossRef]
  4. Yunishafira, A. Determining the Appropriate Demand Forecasting Using Time Series Method: Study Case at Garment Industry in Indonesia. KnE Soc. Sci. 2018, 3, 553–564. [Google Scholar] [CrossRef] [Green Version]
  5. Verma, A.; Karan, A.; Mathur, A.; Chethan, S. Analysis of time-series method for demand forecasting. J. Eng. Appl. Sci. 2017, 12, 3102–3107. [Google Scholar]
  6. An, Y.; Zhou, Y.; Li, R. Forecasting India’s Electricity Demand Using a Range of Probabilistic Methods. Energies 2019, 12, 2574. [Google Scholar] [CrossRef] [Green Version]
  7. Shah, I.; Iftikhar, H.; Ali, S.; Wang, D. Short-Term Electricity Demand Forecasting Using Components Estimation Technique. Energies 2019, 12, 2532. [Google Scholar] [CrossRef] [Green Version]
  8. Al-Musaylh, M.S.; Deo, R.C.; Adamowski, J.F.; Li, Y. Short-term electricity demand forecasting with MARS, SVR and ARIMA models using aggregated demand data in Queensland, Australia. Adv. Eng. Inf. 2018, 35, 1–16. [Google Scholar] [CrossRef]
  9. Yixian, L.I.U.; Roberts, M.C.; Sioshansi, R. A vector autoregression weather model for electricity supply and demand modeling. J. Mod. Power Syst. Clean Energy 2018, 6, 763–776. [Google Scholar]
  10. Yang, Y.; Chen, Y.; Wang, Y. Modelling a combined method based on ANFIS and neural network improved by DE algorithm: A case study for short-term electricity demand forecasting. Appl. Soft Comput. 2016, 49, 663–675. [Google Scholar] [CrossRef]
  11. Al-Musaylh, M.S.; Deo, R.C.; Li, Y. Two-phase particle swarm optimized-support vector regression hybrid model integrated with improved empirical mode decomposition with adaptive noise for multiple-horizon electricity demand forecasting. Appl. Energy 2018, 217, 422–439. [Google Scholar] [CrossRef]
  12. Hatori, T.; Sato-Ilic, M. A Fuzzy Clustering Method Using the Relative Structure of the Belongingness of Objects to Clusters. Procedia Comput. Sci. 2014, 35, 994–1002. [Google Scholar] [CrossRef] [Green Version]
  13. Majkowski, A.; Kołodziej, M.; Rak, R.J. Joint Time-Frequency and Wavelet Analysis—An Introduction. Metrol. Meas. Syst. 2014, 21, 741–758. [Google Scholar] [CrossRef] [Green Version]
  14. Bento, P.M.R.; Pombo, J.A.N.; Calado, M.R.A. A bat optimized neural network and wavelet transform approach for short-term price forecasting. Appl. Energy 2018, 210, 88–97. [Google Scholar] [CrossRef]
  15. Koroglu, S.; Sergeant, P.; Umurkan, N. Comparison of Analytical, Finite Element and Neural Network Methods to Study Magnetic Shielding. Simul. Model. Pract. Theory 2010, 18, 206–216. [Google Scholar] [CrossRef]
  16. Johannesen, N.J.; Kolhe, M.; Goodwin, M. Relative evaluation of regression tools for urban area electrical energy demand forecasting. J. Clean. Prod. 2019, 218, 555–564. [Google Scholar] [CrossRef]
  17. Trull, Ó.; García-Díaz, J.C.; Troncoso, A. Application of Discrete-Interval Moving Seasonalities to Spanish Electricity Demand Forecasting during Easter. Energies 2019, 12, 1083. [Google Scholar] [CrossRef] [Green Version]
  18. Ceci, M.; Corizzo, R.; Malerba, D. Spatial autocorrelation and entropy for renewable energy forecasting. Data Min. Knowl. Discov. 2019, 33, 698–729. [Google Scholar] [CrossRef]
  19. La Rosa, M.; Rabinovich, M.I.; Huerta, R. Slow regularization through chaotic oscillation transfer in an unidirectional chain of Hindmarsh–Rose models. Phys. Lett. A 2000, 266, 88–93. [Google Scholar] [CrossRef]
  20. Yang, A.; Li, W.; Yang, X. Short-term electricity load forecasting based on feature selection and Least Squares Support Vector Machines. Knowl. Based Syst. 2019, 163, 159–173. [Google Scholar] [CrossRef]
  21. Bedi, J.; Toshniwal, D. Deep learning framework to forecast electricity demand. Appl. Energy 2019, 238, 1312–1326. [Google Scholar] [CrossRef]
  22. Eseye, A.T.; Lehtonen, M.; Tukia, T. Machine learning based integrated feature selection approach for improved electricity demand forecasting in decentralized energy systems. IEEE Access 2019, 7, 91463–91475. [Google Scholar] [CrossRef]
  23. Liu, P.; Zheng, P.; Chen, Z. Deep Learning with Stacked Denoising Auto-Encoder for Short-Term Electric Load Forecasting. Energies 2019, 12, 2445. [Google Scholar] [CrossRef] [Green Version]
  24. Singh, S.; Yassine, A. Big data mining of energy time series for behavioral analytics and energy consumption forecasting. Energies 2018, 11, 452. [Google Scholar] [CrossRef] [Green Version]
  25. Corizzo, R.; Pio, G.; Ceci, M. DENCAST: Distributed density-based clustering for multi-target regression. J. Big Data 2019, 6, 43. [Google Scholar] [CrossRef]
  26. Shi, J.; Guo, J.; Zheng, S. Evaluation of hybrid forecasting approaches for wind speed and power generation time series. Renew. Sustain. Energy Rev. 2012, 16, 3471–3480. [Google Scholar] [CrossRef]
  27. Foley, A.M.; Leahy, P.G.; Marvuglia, A. Current methods and advances in forecasting of wind power generation. Renew. Energy 2012, 37, 1–8. [Google Scholar] [CrossRef] [Green Version]
  28. Li, S.; Wang, P.; Goel, L. Short-term load forecasting by wavelet transform and evolutionary extreme learning machine. Electr. Power Syst. Res. 2015, 122, 96–103. [Google Scholar] [CrossRef]
  29. Zhang, X.; Wang, J.; Zhang, K. Short-term electric load forecasting based on singular spectrum analysis and support vector machine optimized by Cuckoo search algorithm. Electr. Power Syst. Res. 2017, 146, 270–285. [Google Scholar] [CrossRef]
  30. Xiao, L.; Wang, J.; Hou, R. A combined model based on data pre-analysis and weight coefficients optimization for electrical load forecasting. Energy 2015, 82, 524–549. [Google Scholar] [CrossRef]
  31. Wang, J.; Zhu, S.; Zhang, W. Combined modeling for electric load forecasting with adaptive particle swarm optimization. Energy 2010, 35, 1671–1678. [Google Scholar] [CrossRef]
  32. Liu, T.; Jin, Y.; Gao, Y. A New Hybrid Approach for Short-Term Electric Load Forecasting Applying Support Vector Machine with Ensemble Empirical Mode Decomposition and Whale Optimization. Energies 2019, 12, 1520. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, Y.; Wang, J.; Lu, H. Research and Application of a Novel Combined Model Based on Multiobjective Optimization for Multistep-Ahead Electric Load Forecasting. Energies 2019, 12, 1931. [Google Scholar] [CrossRef] [Green Version]
  34. Sutskever, I.; Martens, J.; Dahl, G. On the importance of initialization and momentum in deep learning. Proc. Mach. Learn. Res. 2013, 28, 1139–1147. [Google Scholar]
  35. Werbos, P.J. Backpropagation through time: What it does and how to do it. Proc. IEEE 1990, 78, 1550–1560. [Google Scholar] [CrossRef] [Green Version]
  36. Jin, Y.; Sendhoff, B. Pareto-based multiobjective machine learning: An overview and case studies. IEEE Trans. Syst. Man Cybern. Part C 2008, 38, 397–415. [Google Scholar]
  37. Moore, J. Application of Particle Swarm to Multiobjective Optimization; Technical Report; Department of Computer Science and Software Engineering, Auburn University: Auburn, AL, USA, 1999. [Google Scholar]
  38. Bartz-Beielstein, T.; Limbourg, P.; Mehnen, J. Particle Swarm Optimizers for Pareto Optimization with Enhanced Archiving Techniques. In Proceedings of the 2003 Congress on Evolutionary Computation (CEC’03 IEEE), Canberra, Australia, 8–12 December 2003; Volume 3, pp. 1780–1787. [Google Scholar]
  39. Pulido, G.T.; Coello, C.A.C. Using Clustering Techniques to Improve the Performance of a Multi-Objective Particle Swarm Optimizer. In Genetic and Evolutionary Computation Conference; Springer: Berlin/Heidelberg, Germany, 2004; pp. 225–237. [Google Scholar]
  40. Coello, C.A.C.; Pulido, G.T.; Lechuga, M.S. Handling multiple objectives with particle swarm optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  41. Tian, C.; Hao, Y. A novel nonlinear combined forecasting system for short-term load forecasting. Energies 2018, 11, 712. [Google Scholar] [CrossRef] [Green Version]
  42. Pelleg, D.; Moore, A. X-Means: Extending K-Means with Efficient Estimation of the Number of Clusters; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2000. [Google Scholar]
Figure 1. The structure of RNN.
Figure 1. The structure of RNN.
Energies 13 00532 g001
Figure 2. The flowchart of the proposed model.
Figure 2. The flowchart of the proposed model.
Energies 13 00532 g002
Figure 3. Comparison of the 1-step, 2-step, and 3-step forecasting results of ELM on Monday. (A): three evaluation metrics results of ELM. (B): ELM forecasting results on Monday.
Figure 3. Comparison of the 1-step, 2-step, and 3-step forecasting results of ELM on Monday. (A): three evaluation metrics results of ELM. (B): ELM forecasting results on Monday.
Energies 13 00532 g003
Figure 4. Comparison of the 1-step, 2-step, and 3-step forecasting results of RNN on Thursday. (A): three evaluation metrics results of RNN (B): RNN forecasting results on Thursday.
Figure 4. Comparison of the 1-step, 2-step, and 3-step forecasting results of RNN on Thursday. (A): three evaluation metrics results of RNN (B): RNN forecasting results on Thursday.
Energies 13 00532 g004
Figure 5. Comparison of the 1-step, 2-step, and 3-step forecasting results of SVM on Friday. (A): three evaluation metrics results of SVM. (B): SVM forecasting results on Friday.
Figure 5. Comparison of the 1-step, 2-step, and 3-step forecasting results of SVM on Friday. (A): three evaluation metrics results of SVM. (B): SVM forecasting results on Friday.
Energies 13 00532 g005
Figure 6. Comparison of the 1-step, 2-step, and 3-step forecasting results of the combined model on Monday. (A): three evaluation metrics results of Combined model. (B): Combined model forecasting results on Monday.
Figure 6. Comparison of the 1-step, 2-step, and 3-step forecasting results of the combined model on Monday. (A): three evaluation metrics results of Combined model. (B): Combined model forecasting results on Monday.
Energies 13 00532 g006
Figure 7. Comparison of forecasting performances of four models for Queensland in a week.
Figure 7. Comparison of forecasting performances of four models for Queensland in a week.
Energies 13 00532 g007
Figure 8. Comparison of the final forecasting errors of the four models on Monday and Saturday.
Figure 8. Comparison of the final forecasting errors of the four models on Monday and Saturday.
Energies 13 00532 g008
Figure 9. Comparison of the final forecasting value results of the four models for Queensland in a week.
Figure 9. Comparison of the final forecasting value results of the four models for Queensland in a week.
Energies 13 00532 g009
Table 1. One-step, 2-step, and 3-step of the typical results of ELM for a week in NSW.
Table 1. One-step, 2-step, and 3-step of the typical results of ELM for a week in NSW.
ELMMAEMAPE (%)RMSE
TIME(D)STEP1STEP2STEP3STEP1STEP2STEP3STEP1STEP2STEP3
DAY1180.2620223.5158269.04771.88292.28232.7711199.3975257.5034301.6650
DAY2150.5427119.555895.78081.55681.14240.9093169.7331152.0707129.8965
DAY3115.4631100.0074127.02221.17501.02091.2884134.7088116.1235154.6992
DAY473.833270.035484.20510.76590.73160.879898.669994.7967116.4148
DAY5110.4526129.0427116.53591.12881.29061.2246140.7077156.1950154.2217
DAY668.8109105.4868129.21830.77531.16521.478891.7127134.2359177.7145
DAY7138.0636160.841866.28111.64481.84330.7571207.5540180.595884.2173
Table 2. One-step, 2-step, and 3-step of the typical results of RNN for a week in NSW.
Table 2. One-step, 2-step, and 3-step of the typical results of RNN for a week in NSW.
RNNMAEMAPE (%)RMSE
TIME(D)STEP1STEP2STEP3STEP1STEP2STEP3STEP1STEP2STEP3
DAY1204.1581100.5311101.66792.05801.00041.0509213.0676161.3429110.6075
DAY298.1426167.0452122.84770.92961.66831.1735134.0583182.9610160.6650
DAY3121.852372.247997.62251.24740.76551.0284139.623885.7970118.0915
DAY467.440967.524982.06950.69290.70090.854793.116893.0551113.8564
DAY5114.374395.813492.71501.15880.97400.9485136.1701113.4121114.0553
DAY676.4809109.1112109.33410.86191.23231.2266103.1139136.8135129.6750
DAY7110.821682.7990121.96771.29291.00411.3783130.8743114.4017153.7139
Table 3. One-step, 2-step, and 3-step of the typical results of SVM for a week in NSW.
Table 3. One-step, 2-step, and 3-step of the typical results of SVM for a week in NSW.
SVMMAEMAPE (%)RMSE
TIME(D)STEP1STEP2STEP3STEP1STEP2STEP3STEP1STEP2STEP3
DAY1180.0960172.5924303.21881.80991.72673.0852210.2919198.6691346.2488
DAY272.6219109.3225140.45590.77141.12311.346086.1632122.1922187.4571
DAY3209.4914144.2640439.06662.07441.47694.3661238.7288174.8522506.0270
DAY482.994886.9796554.56020.87160.91585.7218104.0347109.1755109.1755
DAY596.6003191.0145181.12850.97760.97761.7755112.0442240.2994225.3321
DAY678.3938192.2528118.73550.88832.18991.3550110.0911229.1995161.0205
DAY755.2772274.7910110.17460.64063.10981.237768.5575345.1129133.3524
Table 4. One-step, 2-step, and 3-step of the typical results of the combined model for a week in NSW.
Table 4. One-step, 2-step, and 3-step of the typical results of the combined model for a week in NSW.
MIXMAEMAPERMSE
TIME (D)STEP1STEP2STEP3STEP1STEP2STEP3STEP1STEP2STEP3
DAY137.658919.508018.26690.39180.20340.200049.764223.976022.5024
DAY237.025323.749323.74930.37990.24100.248948.667229.370434.0105
DAY312.150472.144624.52230.12370.76440.259015.124485.688629.7709
DAY467.595466.681478.92360.69360.69520.821693.344590.1175110.0905
DAY5115.334095.773192.09741.16870.97520.9449136.8067114.1894113.5286
DAY679.213933.543118.18080.89160.37380.2037105.865944.079621.6340
DAY725.216218.437818.67160.30140.21200.217731.976122.286622.6853
Table 5. Results of the three evaluation metrics of the four models for Queensland.
Table 5. Results of the three evaluation metrics of the four models for Queensland.
DateMondayTuesdayWednesdayThursdayFridaySaturdaySundayAverage
MAE
ELM106.7967430.4047124.4383248.7430137.0113119.1218232.6718199.8839
RNN49.406151.964496.5393105.4656135.3078114.5491132.477497.9585
SVM169.2212130.853774.2149223.3669210.5455139.9597202.0815164.3205
Combined10.346845.783851.526079.4087103.839817.2876122.888861.5831
RMSE
ELM146.4229622.5072195.6884329.6401206.9160177.1841330.4177286.9681
RNN61.117878.9962129.8775141.4667188.2211153.6274189.4702134.6824
SVM196.0913169.0187106.3683340.2963328.5419215.5059308.1820237.7149
Combined13.520260.042579.889998.0779133.278825.3522172.364483.2180
MAPE (%)
ELM1.69796.69001.94414.10412.27232.16414.34663.3170
RNN0.79520.83011.51191.77112.24332.08342.48811.6747
SVM2.72282.06971.18723.73873.50432.55913.84902.8044
Combined0.16810.73460.81771.32941.70870.30372.29581.0511

Share and Cite

MDPI and ACS Style

Yang, Y.; Shang, Z.; Chen, Y.; Chen, Y. Multi-Objective Particle Swarm Optimization Algorithm for Multi-Step Electric Load Forecasting. Energies 2020, 13, 532. https://doi.org/10.3390/en13030532

AMA Style

Yang Y, Shang Z, Chen Y, Chen Y. Multi-Objective Particle Swarm Optimization Algorithm for Multi-Step Electric Load Forecasting. Energies. 2020; 13(3):532. https://doi.org/10.3390/en13030532

Chicago/Turabian Style

Yang, Yi, Zhihao Shang, Yao Chen, and Yanhua Chen. 2020. "Multi-Objective Particle Swarm Optimization Algorithm for Multi-Step Electric Load Forecasting" Energies 13, no. 3: 532. https://doi.org/10.3390/en13030532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop