Next Article in Journal
Mechanical Performance of Methane Hydrate–Coal Mixture
Next Article in Special Issue
Short-Term Forecasting for Energy Consumption through Stacking Heterogeneous Ensemble Learning Model
Previous Article in Journal
Peer to Peer Distributed Energy Trading in Smart Grids: A Survey
Previous Article in Special Issue
Short-Term Load Forecasting for Electric Bus Charging Stations Based on Fuzzy Clustering and Least Squares Support Vector Machine Optimized by Wolf Pack Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Hybrid Interval Prediction Approach Based on Modified Lower Upper Bound Estimation in Combination with Multi-Objective Salp Swarm Algorithm for Short-Term Load Forecasting

1
Faculty of Information Technology, Macau University of Science and Technology, Macau 999078, China
2
School of Statistics, Dongbei University of Finance and Economics, Dalian 116025, China
3
Gansu Meteorological Service Centre, Lanzhou 730020, China
*
Author to whom correspondence should be addressed.
Energies 2018, 11(6), 1561; https://doi.org/10.3390/en11061561
Submission received: 25 May 2018 / Revised: 8 June 2018 / Accepted: 10 June 2018 / Published: 14 June 2018
(This article belongs to the Special Issue Short-Term Load Forecasting by Artificial Intelligent Technologies)

Abstract

:
Effective and reliable load forecasting is an important basis for power system planning and operation decisions. Its forecasting accuracy directly affects the safety and economy of the operation of the power system. However, attaining the desired point forecasting accuracy has been regarded as a challenge because of the intrinsic complexity and instability of the power load. Considering the difficulties of accurate point forecasting, interval prediction is able to tolerate increased uncertainty and provide more information for practical operation decisions. In this study, a novel hybrid system for short-term load forecasting (STLF) is proposed by integrating a data preprocessing module, a multi-objective optimization module, and an interval prediction module. In this system, the training process is performed by maximizing the coverage probability and by minimizing the forecasting interval width at the same time. To verify the performance of the proposed hybrid system, half-hourly load data are set as illustrative cases and two experiments are carried out in four states with four quarters in Australia. The simulation results verified the superiority of the proposed technique and the effects of the submodules were analyzed by comparing the outcomes with those of benchmark models. Furthermore, it is proved that the proposed hybrid system is valuable in improving power grid management.

1. Introduction

Load forecasting is of upmost significance and affects the construction and operation of power systems. In the preparation of the power system planning stage, if the load forecasting result is lower than the real demand, the installed and distribution capacities of the planned power system will be insufficient. The power generated will not be able to meet electricity demand of the community it serves, and the entire system will not be able to operate in a stable manner. Conversely, if the load forecast is too high, it will result in power generation, transmission, and distribution, at a larger scale, that cannot be fully used in the real power system. The investment efficiency and the efficiency of the resource utilization will be reduced in this situation. Therefore, effective and reliable power load forecasting can promote a balanced development of the power system while improving the utilization of energy. There are various power load forecasting methods and, commonly, load forecasting is classified into short-term, medium-term, and long-term, based on the application field and forecasting time. Among these categories, short-term load forecasting (STLF) is an essential tool for the planning and operation [1,2] of energy systems and it has thus been a major area of research during the past few decades.
According to existing research, concern mostly focuses on the point forecasting of STLF. Additionally, the relative algorithms can be mainly classified into three major categories: traditional statistical techniques, computational intelligent methods, and multimodule hybrid models [3].
In the early stages of research, traditional statistical techniques were extensively employed for point forecasting of STLF, such as linear regression methods [4,5], exponential smoothing [6], Kalman filters [7], and other linear time-series models. In general, most of the traditional statistical approaches have been involved in linear analysis and have mainly considered linear factors in time series. However, the short-term load series are a mixture of multiple components which include linear and non-linear factors. Therefore, the traditional statistical approaches encounter difficulties when dealing with the STLF, and the forecasting accuracy is often unsatisfactory. With the development of machine learning and artificial intelligence, an increased number of non-linear computational intelligent methods have been applied to STLF, such as neural network models (NN) [8,9], expert systems [10] and support vector machines (SVM) [11,12]. These approaches have been proved to have advantages in dealing with the non-linear problems of STLF compared to traditional statistical methods, thereby eliciting improved performances in most cases. Most importantly, a key point that influences the performance of computational intelligent methods is the setting of related parameters in algorithms. At this time, efficient hybrid models appeared. In hybrid models, different modules were introduced to improve the performance and accuracy of STLF [13,14,15,16,17,18,19]. Among existing reviews in the literature, two popular and efficient modules include the data preprocessing and optimization modules. In the case of the data preprocessing modules, a multiwavelet transform was used in combination with a three-layer feed-forward neural network to extract the training data and predict the load in [13]. Fan et al. [14] used empirical mode decomposition (EMD) to decompose electric load data, generating high-frequency series and residuals for the forecasting of support vector regression (SVR) and autoregression (AR), respectively. The results showed that the hybrid methods can perform well by eliciting good forecasting accuracy and interpretability. In the case of the optimization modules, AlRashidi et al. [15] employed the particle swarm optimizer (PSO) to fine-tune the model parameters, and the forecasting problem was presented in a state space form. Wang et al. [16] proposed a hybrid forecasting model combining differential evolution (DE) and support vector regression (SVR) for load forecasting, where the DE algorithm was used to choose the appropriate parameters for SVR.
However, as mentioned above, the current research on STLF mainly concentrates on point forecasting in which the accuracy is usually measured by the errors between the predicted and the target values. With power system growth and the increase in its complexity, point forecasting might not be able to provide adequate information support for power system decision making. An increasing number of factors, such as load management, energy conversion, spot pricing, independent power producers and non-conventional energy, make point forecasting undependable in practice. In addition to the fact that most of these point forecasting models do not elicit the required precision, they are also not adequately robust. They fail to yield accurate forecasts when quick exogenous changes occur. Other shortcomings are related to noise immunity, portability, and maintenance [20].
In general, point forecasting cannot properly handle uncertainties associated with load datasets in most cases. To avoid such imperfection, interval prediction (IP) of STLF is an efficient way to deal with the forecast uncertainty in electrical power systems. Prediction intervals (PIs) not only provide a range in which targets are highly likely to be covered, but they also provide an indication of their accuracy, known as the coverage probability. Furthermore, the PIs can take into account more uncertain information and the result of (PIs) commonly form a double output (upper bounds and lower bounds) which can reflect more uncertain information and provide a more adequate basis for power system planning.
With the development of artificial intelligence technology, the interval prediction methods based on NN have been proved to be efficient techniques. According to existing research, the popular techniques for constructing PIs are Bayesian [21], delta [22], bootstrap [23], and mean–variance estimation [24]. In the literature, the Bayesian technique [25] is used for the construction of NN-based PIs. Error bars are assigned to the predicted NN values using the Bayesian technique. Even if the theories are effective in the construction of PIs, the calculation of the Hessian matrix will result in the increase of model complexity and computation cost. In [26], the delta technique was applied to construct PIs for STLF, and a simulated annealing (SA) algorithm was introduced to improve the performance of PIs through the minimization of a loss function. In [27], according to bootstrap, error output, resampling, and multilinear regression, were used with STLF for the construction of confidence intervals with NN models. In [24], a mean–variance estimation-based method used NN to estimate the characteristics of the conditional target distribution. Additive Gaussian noise with non-constant variance was the key assumption of the method for PI construction.
Considering most of the existing research studies of PIs by NN mentioned above, the PIs were usually calculated depending on the point forecasting. The NNs were first trained by minimizing an error-based cost function, and the PIs were then constructed depending on the outcomes of trained and tuned NNs. It may be questionable to construct PIs in this way. Furthermore, it is a more reasonable way to output the upper and lower bounds directly [28]. Compared with the Bayesian, delta, and bootstrap techniques, this approach can output the PIs without being dependent on point prediction. However, in traditional research approaches, the cost function mainly aims at guaranteeing coverage probability (CP). However, a satisfactory coverage probability can be achieved easily by assigning sufficiently large and small values to the upper and lower bounds of the PIs. Thus, the prediction interval width (PIW) is another key characteristic which needs to be considered fully. These two goals, that is, achieving a higher CP and a lower PIW, should be considered in a comprehensive manner when the NN parameters are determined.
Therefore, in this study, a hybrid, lower upper bound estimation (LUBE) based on multi-objective optimization is proposed. The requirements for higher CP and lower PIW constitute a typical case of the Pareto optimization problem. In the present study, a significant and valid approach was used to solve the Pareto optimization problem is the multi-objective optimization [29]. There are many algorithms in the literature for solving multi-objective optimizations. For the GA, the most well-regarded multi-objective algorithm is the non-dominated sorting genetic algorithm (NSGA) [30]. Other popular algorithms include the multi-objective particle swarm optimization (MOPSO) [31,32], multi-objective ant colony optimization (MOACO) [33], multi-objective differential evolution (MODE) [34], multi-objective grasshopper optimization (MOGO) [35], multi-objective evolution strategy (MOES) [36], multi-objective sine cosine (MOSC) [37], and multi-objective ant lion [38]. All these algorithms are proved to be effective in identifying non-dominated solutions for multi-objective problems. According to the “no free lunch theorem” for optimization [39,40], there is no algorithm capable of solving optimization algorithms for all types of problems. This theorem logically proves this and proposes new algorithms, or improves the current ones.
In this study, to achieve a better performance in STLF, one of the novel recurrent neural networks, the Elman neural network (ENN) [41], is applied to construct the structure of a modified LUBE. The Elman neural network has already been extensively used in time-series forecasting [42,43,44]. As a type of recurrent neural network, ENN exhibits superiority on the time delay information because of the existence of the undertaking layer which can connect hidden NN layers and store the historical information in the training process. This structure design of NN commonly leads to a better performance in time-series forecasting.
In traditional STLF, most of the methods construct the training set of the model directly using the original data. However, data in the natural world often receives a lot of noise interference, which will cause more difficulties for desired STLF. Furthermore, improving the signal-to-noise ratio of the training dataset will help the effective training of the model. Amongst the existing denoising methods, empirical mode decomposition (EMD) [45] is extensively used, which is an adaptive method introduced to analyze non-linear and non-stationary signals. In order to alleviate some reconstruction problems, such as “mode mixing” of EMD, some other versions [46,47,48] are proposed. Particularly, the problem of different number of modes for different realizations of signal and noise need to be considered.
Summing up the above, in this study, a hybrid interval prediction system is proposed to solve the STLF problem based on the modified Lower and Upper bound estimate (LUBE) technique, by incorporating the use of a data preprocessing module, an optimization module, and a prediction module. In order to verify the performance of the proposed model, we choose as the experimental case the power loads of four states in Australia. The elicited results are compared with those from basic benchmark models. In summary, the primary contributions of this study are described below:
(1)
A modified LUBE technique is proposed based on a recurrent neural network, which is able to consider previous information of former observations in STLF. The contest layer of ENN can store the outputs of a former hidden layer, and then connect the input layer in the current period. Comparison of the traditional interval predictive model with the basic neural network, this mechanism can improve the performance of time series forecasting methods, such as STLF.
(2)
A more convincing optimization technique based on multi-objective optimization is proposed for LUBE. In LUBE, besides CP, PIW should also be considered in the construction of the cost function. In this study, the novel multi-objective optimization method MOSSA is employed in the optimization module to balance the conflict between higher CP and lower PIW, and to train the parameters in ENN. With this method, the structure of neural networks can provide a better performance in interval prediction.
(3)
A novel and efficient data preprocessing method is introduced to extract the valuable information from raw data. In order to improve the signal noise ratio (SNR) of the input data, an efficient method is used to decompose the raw data into several empirical modal functions (IMFs). According to the entropy theory, the IMFs with little valuable information are ignored. The performance of the proposed model trained with processed data improves significantly.
(4)
The proposed hybrid system for STLF can provide powerful theoretical and practical support for decision making and management in power grids. This hybrid system is simulated and tested depending on the abundant samples involving different regions and different times, which indicate its practicability and applicability in the practical operations of power grids compared to some basic models.
The rest of this study is organized as follows: The relevant methodology, including data preprocessing, Elman neural network, LUBE, and multi-objective algorithms, are introduced in Section 2. Section 3 discusses our proposed model in detail. The specific simulation, comparisons and analyses of the model performances are shown in Section 4. In order to further understand the features of the proposed model, several points are discussed in Section 5. According to the results of our research, conclusions are outlined in Section 6.

2. Methodology

In this section, the theory of the hybrid interval prediction model is elaborated, and the methodology of the components in hybrid models, including complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), Elman neural networks, LUBE, and MOSSA, are explained in detail.

2.1. Data Preprocessing

The EMD technique [45] usually decomposes a signal into several numbers of IMFs. For each IMF, the series have to fulfill two conditions: (i) the number of extrema (maxima and minima) and the number of zero-crossings must be equal or differ at most by one; and (ii) the local mean, defined as the mean of the upper and lower envelopes, must be zero. In order to alleviate mode mixing, the EEMD [46], defines the “true” modes as the average of the corresponding IMFs obtained from an ensemble of the original signal plus different realizations of finite variance white noise. But incompletion of decomposition still exists, and the number of modes will be different due to the noise added. Taking these short comes into account, CEEMDAN is proposed. The details are described as follows: let E k ( ) be the operator which produces the kth mode obtained by EMD and w ( i ) be a realization of white noise with N (0, 1). And then the process of CEEMDAN can be expressed as several stages:
1st step.
For every i = 1 , , I decompose each x ( i ) = x + β 0 w ( i ) by EMD, until the first mode is extracted and compute d ˜ 1 by:
d ˜ 1 = 1 I i = 1 I d 1 i = d ¯ 1
2nd step.
At the first stage (k = 1) calculate the first residue as r 1 = x d ˜ 1 .
3rd step.
Obtain the first mode of r 1 + β 1 E 1 ( w i ) , i = 1 , I , by EMD and define the second CEEMDAN mode as:
d ˜ 2 = 1 I i = 1 I E 1 ( r 1 + β 1 E 1 ( w ( i ) ) )
4th step.
For k = 2 , K calculate the kth residue:
r k = r ( k 1 ) d ˜ k
5th step.
Obtain the first mode of r k + β k E k ( w ( i ) ) , i = 1 , , I , by EMD until define the (k + 1)th CEEMDAN mode as:
d ˜ ( k + 1 ) = 1 I i = 1 I E 1 ( r k + β k E k ( w ( i ) ) )
6th step.
Go to 4th step for the next k.
Iterate the steps 4 to 6 until the obtained residue cannot be further decomposed by EMD, either because it satisfies IMF conditions or because it has less than three local extremums. Observe that, by construction of CEEMDAN, the final residue satisfies:
r K = x k = 1 K d ˜ k
with K being the total number of modes. Therefore, the signal of interest x can be expressed as:
x = k = 1 K d ˜ k + r k
which ensures the completeness property of the proposed decomposition and thus providing an exact reconstruction of the original data. The final number of modes is determined only by the data and the stopping criterion. The coefficients β k = ε k s t d ( r k ) allow the selection of the SNR at each stage.
The CEEMDAN method can add a limited number of self-use white noises at each stage, which can achieve almost zero reconstruction error with fewer average times. Therefore, CEEMDAN can overcome the “mode-mixing” phenomenon existing in EMD, and can also solve the incompleteness of EEMD decomposition and reduce the computational efficiency by reducing the reconstruction error by increasing the number of integrations.

2.2. Elman Neural Network (ENN)

As an important branch of deep learning, recurrent neural networks have been widely used in academic and industrial fields. The common neural network mainly consists of three layers: input layer, hidden layer and output layer. For the hidden layer, the input information only comes from the input layer. For a recurrent neural network, the input information of the hidden layer will not only come from the input layer, but also from the hidden layer itself and the output layer.
In various structures of the recurrent neural network, Elman neural network (ENN) [49] is typical structure in which the lags of hidden layer are delivered into the current hidden layer by a new layer called the context layer. This structure takes the former information of the hidden layer into account and commonly has a better performance in the time-series forecasting such as STLF, wind speed forecasting, financial time-series forecasting. The structure is showed in Figure 1.
The context layer can feed back the hidden layer outputs in the previous time steps and neurons contained in each layer are used to transmit information from one layer to another. The dynamics of the change in hidden state neuron activations in the context layer is as follows:
S i ( t ) = g ( k = 1 K V i k S k ( t 1 ) + j = 1 j W i j I j ( t 1 ) )
where S k ( t ) and I j ( t ) denote the output of the context state and input neurons, respectively; V i k and W i j denote their corresponding weights; and g ( ) is a sigmoid transfer function. The other related theories such as feed-forward and back propagation are similar with the common back propagation neural network.

2.3. Lower Bound and Upper Bound Estimation (LUBE)

In the literature, the traditional interval prediction commonly attempts to construct the PI based on the point prediction. The upper bound and the lower bound are calculated according to the forecasting value and the confidence level. The accuracy of the point forecasting has played a key role in the accuracy of the PI. In this paper, we introduce a novel method of interval prediction called lower bound and upper bound estimation (LUBE). This method directly outputs the lower bound and the upper bound of PI depending on the multi-output neural network. The structure we employed in this paper is shown in Figure 1.
The output of the normal LUBE structure [50] just consist of two neurons which denote the upper bound and the lower bound, while the outputs in our structure of LUBE consist of three neurons. The first output corresponds to the upper bound of the PI, the second output denotes the predicted value, and third output approximates the lower bound of the PI. In the literature, the PI construction techniques attempt to estimate the mean and variance of the targets for construction of PIs. In contrast to existing techniques, the proposed method tries to directly approximate upper and lower bounds of PIs based on the set of inputs. Therefore, in the training process, loss function of this LUBE method based on neural network should be set according to the key criterion of PIs (CP and PIW).

2.4. Multi-Objective Optimization Algorithm

The multi-objective optimization algorithm has been widely used to solve multi-objective optimization problem. In this paper, a novel multi-objective optimization algorithm named Multi-Objective Salp Swarm Algorithm (MOSSA) is introduced.

2.4.1. Multi-Objective Optimization Problem

In multi-objective optimization, all of the objectives are optimized simultaneously. The main concern is formulated as follows:
M i n i m i z e : F ( X ) = { f 1 ( X ) , f 2 ( X ) , , f o ( X ) }
S u b j e c t t o : g i ( X ) 0 , i = 1 , 2 , , m
h i ( X ) = 0 , i = 1 , 2 , , p
l b i x i u b i , i = 1 , 2 , , n
where o is the number of objectives, m is the number of inequality constraints, p is the number of equality constraints, and lbi is the lower bound of the ith variable, and ubi is the upper bound of the ith variable. With one objective we can confidently estimate that a solution is better than another depending on comparing the single criterion, while in a multi-objective problem, there is more than one criterion to compare solutions. The main theory to compare two solutions considering multiple objectives is called Pareto optimal dominance as explained in [51].
There are two main approaches for solving multi-objective problems: a priori and a posteriori [52]. In the priori method, the multi-objective problem is transformed to a single-objective problem by aggregating the objectives with a set of weights determined by experts. The main defect of this method is that the Pareto optimal set and the front need to be constructed by re-running the algorithm and changing the weights [53]. However, the a posteriori method keeps the multi-objective formulation in the solving process, and the Pareto optimal set can be determined in a single run. Without any weight to be defined by experts, this approach can approximate any type of Pareto optimal front. Because of the advantages of a posteriori optimization over the a priori approach, the focus of our research is aimed at a posteriori multi-objective optimization.

2.4.2. Multi-Objective Salp Swarm Algorithm (MOSSA)

As an a posteriori multi-objective optimization, MOSSA [54] is similar to some swarm multi-objective optimization algorithm such as MOPSO [31], MOACO [33] and MOGO [35]. By simulating the biological behavior of ecological communities, the optimal solution is achieved.
Salps belong to the family of Salpidae and have transparent barrel-shaped body. Their tissues are highly similar to jellyfishes. They also move very similar to jellyfish, in which the water is pumped through body as propulsion to move forward. In deep oceans, salps often form a swarm called a salp chain. The main concern about salps in MOSSA is their swarming behavior.
To mathematically model the salp chains, the population is first divided to two groups: leader and followers. The leader is the salp at the front of the chain, whereas the rest of salps are considered as followers. As the name of these salps implies, the leader guides swarm and the followers follow each other.
Similar to other swarm-based techniques, the position of salps is defined in an n-dimensional search space where n is the number of variables of a given problem. Therefore, the positions of all salps are stored in a two-dimensional matrix called x. It is also assumed that there is a food source called F in the search space as the swarm’s target.
Definition 1.
To update the position of the leader the following equation is proposed:
x j 1 = { F j + c 1 ( ( u b j l b j ) c 2 + l b j ) c 3 0 F j c 1 ( ( u b j l b j ) c 2 + l b j ) c 3 < 0
where x j 1 shows the position of the first salp (leader) in the jth dimension, F j is the position of the food source in the jth dimension, u b j indicates the upper bound of jth dimension, l b j indicates the lower bound of jth dimension, c1, c2, and c3 are random numbers. Equation (12) shows that the leader only updates its position with respect to the food source.
Definition 2.
The coefficient c1 is the most important parameter in the Salp swarm algorithm (SSA) because it balances exploration and exploitation is defined as follows:
c 1 = 2 e ( 4 l L ) 2
where l is the current iteration and L is the maximum number of iterations.
The parameter c2 and c3 are random numbers uniformly generated in the interval of [0, 1]. In fact, they dictate if the next position in jth dimension should be towards positive infinity or negative infinity as well as the step size.
Definition 3.
To update the position of the followers, the following equations is utilized depending on Newton’s law of motion:
x j i = 1 2 a i j t 2 + v 0 t
where i 2 , x j i shows the position of ith follower salp in jth dimension, t is time, v0 is the initial speed, and a i j = v i j v 0 t where v i j = x i j x 0 t , i 2 , j 1 .
Because the time in optimization is iteration, the discrepancy between iterations is equal to 1, and considering v0 = 0, this equation can be expressed as follows:
x j ( t ) i = 1 2 ( x j ( t 1 ) i + x j ( t 1 ) i 1 )
where i 2 and x j ( t ) i show the position of ith follower salp in jth dimension at t-th iteration.
According to the mathematical emulation explained above, the swarm behavior of salp chains can be simulated vividly.
When dealing with multi-objective problems, there are two issues that need to be adjusted for SSA. First, MOSSA need to store multiple solutions as the best solutions for a multi-objective problem. Second, in each iteration, SSA updates the food source with the best solution, but in the multi-objective problem, single best solutions does not exist.
In MOSSA, the first issue is settled by equipping the SSA algorithm with a repository of food sources. The repository can store a limited number of non-dominated solutions. In the process of optimization, each salp is compared with all the residents in repository using the Pareto dominance operators. If a salp dominates only one solution in the repository, it will be swapped. If a salp dominates a set of solutions in the repository, they all should be removed from the repository and the salp should be added in the repository. If at least one of the repository residents dominates a salp in the new population, it should be discarded straight away. If a salp is non-dominated in comparison with all repository residents, it has to be added to the archive. If the repository becomes full, we need to remove one of the similar non-dominated solutions in the repository. For the second issue, an appropriate way is to select it from a set of non-dominated solutions with the least crowded neighborhood. This can be done using the same ranking process and roulette wheel selection employed in the repository maintenance operator. The pseudo code of MOSSA is showed in Algorithm 1:
Algorithm 1. Pseudo-code of MOSSA.
Energies 11 01561 i001

3. Proposed Interval Prediction Model for Short-Term Load Forecasting (STLF)

In this paper, we proposed a hybrid model for interval prediction based on the data preprocessing, multi-objective optimization algorithm and LUBE to solve the problem of STLF. This hybrid model consist of two stages: data de-noising and model prediction.
In the first stage, the main task is to refine the original data. The raw power load data is affected by many internal and external factors in the collection process. Therefore, a lot of unrelated information is integrated in the data. Several pieces of information will further affect the quality of the power load data, and increase the difficulty of accurate forecasting of the power load. In the neural network model, the performance of the model is directly affected by the quality of the data. As a type of machine learning algorithm, the neural network uses its multilayered structure to learn the relevant interdependencies of the data and determine the structural parameters of the prediction model, so as to achieve fitting and forecasting. However, if the input set of the model contains too much noise and “false information”, the model will be seriously affected in the training process, and some problems will emerge, such as the overfitting problem. Therefore, we introduced CEEMDAN to eliminate useless information in the raw data. As mentioned above, CEEMDAN can decompose the data series into several IMFs with different frequencies, as shown in Figure 2. Because the IMFs are extracted with envelope curves depending on the extremum, some of the IMFs have higher frequencies, just as the first few IMFs that are shown in Figure 2. In addition, the other IMFs also have lower frequencies and represent the trend factors, thereby formulating the vital basis for time-series prediction. In the actual operations, we can remove the IMFs with higher frequencies, which effectively represent noise to refine the original data. In order to determine which IMFs ought to be abandoned, we calculated the entropy of each IMF and removed the IMFs with lower entropy. After the denoising process, the refined data are transferred to next stage as the input data for training in the predictive model.
In the second stage, the main interval prediction model was proposed. In our hybrid interval prediction model, the PI is output dependent on LUBE, which is based on the multi-output of the Elman neural network (E–LUBE). In the training process, the input set of E–LUBE is constructed as indicated in Formula (16), while the output set is constructed as indicated in Formula (17), where m and s respectively denote the number of features and the numbers of samples, and α denotes the interval width coefficient. In the case of the STLF problem, m indicates the number of previous time-points that we use to forecast the predictive value.
Input   set :   [ x 1 x 2 x m x 2 x 3 x m + 1 x s x s + 1 x s + m ]
Output   set :   [ x m + 1 × ( 1 α ) x m + 2 × ( 1 α ) x m + s + 1 × ( 1 α ) x m + 1 x m + 2 x m + s + 1 x m + 1 × ( 1 + α ) x m + 2 × ( 1 + α ) x m + s + 1 × ( 1 + α ) ]
According to a trained model, when a new series Xi, i = 1, …, m, is input into the model, Xm+1 with an upper bound X m + 1 U and a lower bound X m + 1 L will be output. This is the basic mechanism of interval prediction for STLF in this study. However, in traditional multi-output neural networks, the loss function is always the mean-square-error (MSE), which is a key criterion for point forecasting. In this study, we introduced two new criteria (PIW and CP) to construct the loss function, considering the main purpose of our interval prediction. The traditional neural network parameters were determined by using a gradient descent algorithm, but for two of the set criteria, the calculation of the gradient was difficult. Therefore, we employed MOSSA to realize the multi-objective parameter optimization. Furthermore, the optimization problem can be expressed as,
a r g m i n { PIW ( θ ) 1 CP ( θ )
where θ is a set of parameters in E–LUBE, including the weight and bias.
When the parameters are determined in the training process, the entire model can be applied to the test set to verify the performance of interval prediction.

4. Simulations and Analyses

In order to validate the performance of the proposed hybrid model in STLF, four electrical load datasets collected from four states in Australia are used in our research. The four states include New South Wales (NSW), Tasmania (TAX), Queensland (QLD) and Victoria (VIC), and the specific location is showed in Figure 3. The experiments in this study consist of two parts: experiment I and experiment II. For experiment I, the load data of four states are modeled with interval width coefficient α = 0.05 , and for the experiment II, the interval width coefficient α is set as 0.025 for further analysis. In order to verify the superiority of the proposed hybrid model, several benchmark models which include basic LUBE (LUBE), LUBE with Elman neural network (E–LUBE), E–LUBE with point optimization (PO–E–LUBE), E–LUBE with interval optimization (IO–E–LUBE), and models integrated with CEEMDAN, are exhibited. For persuasive comparability and fairness, the hyper-parameters in each model are consistent, as shown in Table 1. All experiments have been carried out in MATLAB 2016a on a PC with the configuration of Windows 7 64-bit, Inter Core i5-4590 CPU @ 3.30GHz, 8GB RAM.

4.1. Data Descriptions

For each state, we considered the data using half an hour interval in four quarters. The data used in this paper can be obtained on the website of Australian energy market operator (http://www.aemo.com.au/). We chose data from the whole of 2017 from 1 January 2017 0:30 am to 31 December 2017 0:00 am to construct dataset. In each state, the total sample number is 17,520. For each quarter, the number of samples were 4320, 4358, 4416, 4416 respectively. In order to control the comparability, we selected 1200 samples to test the trained model, and used the rest in each quarter to train the models. The proportion of train sets versus the test sets was approximately equal to 3:1. The description of the data characteristics are shown in Figure 4. Considering the structure of the neural network in this study, we set six input neurons, 13 hidden neurons, and three output neurons. Specifically, the output set was formulated in accordance with Formula (17).
During data preprocessing, the input data were divided into several IMFs depending on CEEMDAN, as displayed in Figure 2. According to the energy entropy of each IMF shown in Figure 3, we ignored the IMFs which contained high frequencies, and summed the rest of the IMFs to reconstruct the input set, as shown in Figure 1.

4.2. Performance Metrics

In order to comprehensively assess the performance of the models, some metrics were employed. These metrics primarily focused on the coverage of the real value in the prediction interval and the width of the interval.

4.2.1. Coverage Probability

Coverage probability [50] is usually considered as a basic feature of PIs and CP is calculated according to the ratio of the number of target values covered by PIs:
CP = 1 m i = 1 m θ i
where m denotes the number of samples, and θ i is a binary index which measures whether the target value is covered by PIs:
θ i = { 1 , y i t [ L ^ i , U ^ i ] 0 , y i t [ L ^ i , U ^ i ]
where y i t denote the ith target value and L ^ i , U ^ i represent the ith lower bound and the upper bound, respectively.
A larger CP means more targets are covered by the constructed PIs and a too small CP indicates the unsatisfied coverage behaviors. To have valid PIs, CP should be larger or at least equal to the nominal confidence level of PIs. Furthermore, in this paper, CP is also an important factor in the process of parameter optimization by the multi-objective optimization algorithm.

4.2.2. Prediction Interval (PI) Normalized Average width and PI Normalized Root-Mean-Square Width

In research studies on interval prediction, more attention is usually paid to CP. However, if the lower and upper bounds of the PIs are expanded from either side, any requirement for a larger CP can be satisfied, even for 100%. However, in some cases, a narrower interval width is necessary for a more precise support for electric power supply. Therefore, the width between the lower and upper bounds should be controlled so that the PIs are more convincing. In this study, the prediction interval width (PIW) is another factor in the process of parameter optimization. With CP and PIW, two objects compose the solution space within which the Pareto solution set is estimated.
In order to eliminate the impact of dimension, some relative indexes should be introduced to improve the comparability of width indicators. Inspired by the mean absolute percentage error (MAPE) in point forecasting, we employed PI normalized average width (PINAW) and PI normalized root-mean-square width (PINRW) [50]:
PINAW = 1 m R i = 1 m ( U i L i )
PINRW = 1 R 1 m i = 1 m ( U i L i ) 2
where R equals to the maximum minus minimum of the target values. Normalization by the range R is able to improve comparability of PIs constructed using different methods and for different types of datasets.

4.2.3. Accumulated Width Deviation (AWD)

Accumulated width deviation (AWD) is a criterion that measure the relative deviation degree, and it can be obtained by the cumulative sum of AWDi [55]. The calculation formula of AWD is expressed as Equations (23) and (24), where α denotes the interval width coefficient and Ii represents the i-th prediction interval.
A W D i = { L i ( α ) z i U i ( α ) L i ( α ) , z i < L i ( α ) 0 , z i I i ( α ) z i U i ( α ) U i ( α ) L i ( α ) , z i > U i ( α )
A W D ( α ) = 1 n i = 1 n A W D i α

4.3. Experiment I: Cases with Larger Width Coefficients

In this experiment, we set the interval width coefficient α = 0.05 , which is equivalent to setting the output to [ 0.95 × X , X , 1.05 × X ] for a single sample in the training process of the neural network. Based on this structure, the PIs can be output given an input test set. In order to guarantee the diversity of the samples, we studied four different quarterly data for four different states.
The models involved in our research can be divided into three groups for better explanations for the impact of different components. The first group included LUBE and E–LUBE, and the difference between them were the structures of the neural network. The structure of LUBE consisted of three layers which were similar to the traditional BP neural network. Moreover, in the E–LUBE, an extra context layer was added to the structure so that we could validate the impact of the context layer in prediction by comparing the performance of these two models. The second group included the PO–E–LUBE and IO–E–LUBE, and the difference between them included the optimization algorithm in the training process. PO–E–LUBE used the error and variance of point prediction to construct the cost function in MOSSA, whereby the target of minimizing the cost function effectively denotes a requirement for better prediction accuracy. In addition, IO–E–LUBE employed the CP and PIW of the interval prediction to construct the cost function in multi-objective optimization, while the target of minimizing such a cost function denoted the requirements for a better performance in interval coverage, which is more rational for our goal of interval prediction. The comparison between such models can reflect the influence of different cost functions in the parameter optimization process. Furthermore, in the first group, the parameters of the neural network are determined by a conventional gradient descent algorithm, and in the second group, the parameters are determined by a heuristic optimization algorithm. Therefore, the impact of different optimization algorithms can be shown by comparing the models in different groups. In addition, in the third group, the data preprocessing is introduced. Based on the models in the first two groups, CEEMDAN was used to refine the input dataset. The results of the models in this group will display the effect of data preprocessing in the hybrid model.
The simulation results are shown in Table 2 and Table 3. Also shown in Figure 5 are the principal indices of interval prediction, namely, CP and PINAW. Based on the conducted comparisons referred to earlier, several conclusions can be inferred:
(1)
By comparing the models in the first group, we can conclude that the E–LUBE is superior to LUBE in most cases, such as the fourth quarter in NSW and the first quarter in TAX, as shown in Table 2 and Figure 5. The CP of E–LUBE reached 87.17%, while the CP of LUBE was 72.36% for the fourth quarter in NSW. The rate of improvement was more than 15% with the maintenance of PINAW and PINRW. However, in some cases, the improvement is not remarkable, such as the fourth quarter in QLD, as shown in Table 3 and Figure 5. The performances of these two models are almost the same. In general, the performance of E–LUBE is better than LUBE, which means that E–LUBE with an extra context layer can improve the performance. In theory, the context layers are able to provide more information compared to previous outputs of hidden layers. This superiority has been proved in our experiments. However, owing to the instability of the parameters in the neural network, the improvement is not adequately remarkable in a few cases.
(2)
In terms of the optimization methods, and according to the results shown in Figure 5, and Table 2 and Table 3, the CPs of the second group (PO–E–LUBE and IO–E–LUBE) perform better than E–LUBE in most cases. E–LUBE uses the gradient descent algorithm, which is sensitive to the initialization, in order to obtain the parameters in NN. Furthermore, the models in the second group use the heuristic swarm optimization algorithm which can synthesize the initialization results using an adequate population size. Thus, the models in the second groups should have elicited better performances in theory unless the random initializations of E–LUBE are perfect. Moreover, within the second group, IO–E–LUBE has a larger CP value than PO–E–LUBE, with low levels of PINAW and PINRW. It is just the influence of the cost function that makes such a difference. The main object of the interval prediction is a larger CP value along with a narrow width. Therefore, the IO should have an advantage
(3)
Incorporation of CEEMDAN in the hybrid models is improved the performances significantly because of the denoising preprocessing. In most cases, the CPs are larger than 80% and 90%, which means more than 80% target load values are covered by the predicted intervals. Furthermore, in some cases, the CPs can reach 100%, such as the second and third quarters in NSW, and the second quarter in QLD. Such accuracy can ensure that the power supply meets the demand. Compared with the original LUBE and E–LUBE, the hybrid model we proposed (CEEMDAN–IO–E–LUBE) elicited a significant improvement in the elicited results of interval prediction.
(4)
With a larger width coefficient, the CPs of our models were almost satisfactory. The smallest CP was more than 70%, and the largest CP was able to reach 100%, which is perfect for interval prediction in STLF. However, the PINAW and PINRW were almost all larger than 10, and even reached the value of 20 in second quarter in QLD. But the proposed model still outperforms other models.
(5)
Considering the accumulated width deviation (AWD), for a larger width coefficient, the proposed model (CEEMDAN-IO-E-LUBE) has a smaller AWD compared with other benchmark models in most cases. According to the definition of AWD, a smaller AWD means more target values fall into the predicted intervals. For the results in which the target values are over the bounds, the deviations are relatively small. In this experiment, the AWDs of the proposed model are satisfactory in most case. For some cases, the AWDs is even closed to 0, which means almost all target load values fall into the predicted intervals. According to these predicted intervals, load dispatch will be more rational.

4.4. Experiment II: Cases with Smaller Width Coefficients

In this experiment, we set the interval width coefficient α = 0.025 , which means we set the output to be [ 0.925 × X , X , 1.025 × X ] for a single sample in the training process of the neural network. With a narrow width coefficient, the lower and upper bounds were closer to the target value in the training process, which can provide more valuable information in practice. However, a narrow bound might lead to the increase of CP. Thus, a smaller width coefficient requires the models to have better predictive properties. The results of this simulation are shown in Table 4 and Table 5, and in Figure 6. Correspondingly, the following conclusions can be drawn:
(1)
As Table 4 and Figure 6 show, the distinction of the models is similar to experiment I. The CPs of the original LUBE and E–LUBE are the smallest among the models in our simulation, and our proposed model CEEMDAN–IO–E–LUBE elicits the best performance
(2)
For some benchmark models in this experiment, with a narrow bound in the training process, the performance was not adequately satisfactory. As the cases of the third quarter in NSW denote and the second quarter in TAX show the CPs of LUBE and E–LUBE are close to 50%, which is not conclusive in practice. However, based on the hybrid mechanism we proposed, the performances were improved significantly. The minimum CP values of CEEMDAN–IO–E–LUBE can reach 70%, and the maximum is close to 100%, such as in the third quarter in QLD. Such results show that the predicted intervals can better cover actual electricity demand data and economize spinning reserve in power grid.
(3)
With a smaller width coefficient, the CPs decreased while the PINAW and PINRW are reduced. For the benchmark models, the results mostly display smaller CPs and larger PINAW or PINRW. However, the proposed model is able to demonstrate larger CPs with smaller PINAW and PINRW values, which is equivalent to a good performance in interval prediction. In some cases, the CP values were larger than 95% with PINAW and PINRW values less than 10. In such cases, the CPs are satisfactory and the widths of the PIs are most appropriate.
(4)
In terms of AWD in this experiment, the proposed model still showed a relatively small AWD compared with other benchmark models, which means the proposed model has a better performance at predicted accuracy. Compared with experiment I, the AWDs in this experiment are bigger. For a smaller width coefficient, the predicted interval will be narrower, which means there will be more target points falling outside the intervals. In some situations, a narrower predicted interval is necessary. The proposed model is able to provide a better performance on the condition of the requirement of a narrower predicted interval of electric load.

4.5. Comparisons and Analyses

According to the comparison of the above two experimental results, the width coefficient has a significant influence on performance, as shown in Figure 7. From one perspective, for most models, a coefficient with a larger width may lead to a larger and more satisfactory CP value, but the index about the width of PI may not be desired. From another perspective, for most models, a narrower width coefficient may elicit the desired PINAW and PINRW values, but the CP is not good enough. Considering such a situation, the proposed models alleviate the contradiction. Even though the CP value of the proposed model will decline when the width coefficient decreases, comprehensive performance is satisfactory. In some exceptional cases, owing to the complexity and instability of the datasets, the performance of the proposed models is not adequate, as the description in Figure 3 shows.

5. Discussion

In this section, we discuss some factors which may have an effect on the performances of the proposed models in order to improve the practicability of our hybrid model. The factors involved mainly include the features of the datasets and the setting of the hyperparameters in the algorithm.

5.1. Dataset Features

The feature and quality of the datasets have a significant effect on the performance of the prediction models. In STLF, the data shows periodicity and volatility. The periodicity is attributed to the regularity in the actual use of electricity, and the volatility is attributed to the randomness and occasional use of electricity. Therefore, the linear component and the non-linear components operate simultaneously during the forecasting of the model. Specifically, some outliers may have a negative effect in the process of prediction.
As Figure 4 shows, the dataset features of the different samples are various. According to the boxplot theory, the data points that are larger than Q3 + 1.5IQR or smaller than Q1 − 1.5IQR are regarded as outliers. For the first and fourth quarters in NSW, and the first and fourth quarters in VIC, the distributions of the datasets displayed a number of outliers. Additionally, the results of the models shown in Table 2, Table 3, Table 4 and Table 5 demonstrate that the model performance of the sample whose distribution is not desired may be unremarkable. These outliers are important factors that lead to such results, even if the CEEMDAN model has been applied in data preprocessing.
Another set of data features that may cause an unsatisfactory result are the non-linear characteristics of the dataset. It is well known that in traditional research, the prediction of regular and linear time series are easy to reach the desired accuracy. However, unstable and non-linear time series are more difficult to forecast in spite of the applications of novel models, such as the case of machine learning algorithms. A method used to measure the instability of data series is the recurrence plot (RP) [56]. A recurrence plot is an advanced technique of non-linear data analyses. It is the visualization (or a graph) of a square matrix in which the matrix elements correspond to those times at which the state of the dynamical system recurs. Stationary systems will deliver homogeneous recurrence plots, and unstable systems cause changes in the distribution of recurrence points in the plot, which is visible and identifiable by the brightened areas. In this study, we selected VIC as an example to verify the influence of instability. Before drawing the recurrence plot, the time delay and the dimension of the embedded matrix were determined by the C–C method. Depending on the “CRP Toolbox” released by Norbert Marwan [57], the recurrence plot of the four datasets of the different quarters in VIC is shown in Figure 8.
As the figure shows, the second and third quarters in VIC display relatively homogeneous distributions, while other quarters display isolated brightened areas. According to the theory of the recurrence plot, the instabilities of the former two samples are weaker, and the other two samples reveal stronger instabilities. Furthermore, we can conclude that the performances of the forecasting models shown in Table 5 are remarkable when the dataset is relatively stable, while the unstable dataset results in an unsatisfactory performance, which cannot be avoided.

5.2. Sensitivity Analysis

The hybrid model proposed in this study is based on the structure of the neural network shown in Figure 1. In the hybrid model, the hyperparameter is a key factor that influences the model’s performance. In most studies on machine learning, the setting of the hyperparameters always depends on trials or empirical knowledge. This is the reason why many experimental results cannot be reproduced and why a considerable amount of time and energy is spent on tuning parameters in industrial applications. At present, there is no absolute method to determine the values of all types of hyperparameters. In this study, we also mainly relied on experiences and trials to set the hyperparameters, as shown in Table 1. Among the hyperparameters, several parameters need to be highlighted.
The first one is the number of salp populations in MOSSA. In the swarm heuristic optimization algorithm, the number of swarms is usually a vital factor that needs to be considered. A larger population might provide a larger probability to reach the best individual, but exceeding the desired population may cause an increase in the complexity of the algorithm, which is related to the number of algorithmic iterations. Considering the number of parameters in our proposed model, the population numbers that ranged from 10 to 100 with a step of 10 were evaluated. As a result, we selected the number 50 as the population number (as shown in Table 6) after comprehensively considering the time complexity and model performance.
The second type of hyperparameters that need to be emphasized are the upper and lower bounds of individual parameters in MOSSA. In our simulation, the datasets were normalized within the range of −1 to 1 in order to avoid the influence of dimension and improve the training speed. Therefore, the absolute value of weights and thresholds of neural networks in the training process will not be too large. As Table 6 shows, we set the initial upper and lower bounds to 2 and −2 according to the experiment trials. Excessive range limits may increase the difficulty of searching for the best parameters with a limited number of iterations. Furthermore, the algorithm that operates based on a small range may not elicit the optimal solution.

5.3. Consistency Analysis

In this section, in order to verify the consistency of our proposed model, new datasets involving latest dates are introduced. In addition, several basic compared models including long short-term memory (LSTM) networks, function fitting neural networks (FITNET), and least squares support vector machine (LSSVM) which have been proved to provide good results for STLF are employed to verify the advantages of the proposed model.
We chose NSW and VIC randomly as examples. The new datasets are collected from 1 January 2018 0:30 am to 30 May 2018 0:00 am and the total number of samples is 7152. The samples in the second quarter in NSW and the fourth quarter in VIC are chosen as compared datasets. According to the results shown in Table 7, the proposed model also has a good performance on the new datasets. The CP is almost 90%, which means the predicted interval can cover 90% target load value. The consistency of the proposed can be guaranteed, and the change of the dates of dataset will not risk altering the final conclusion.
Considering different basic models for STLF, we chose three widely used artificial intelligence models (LSTM, FITNET, and LSSVM) as comparators to verify the superiority. As shown in Table 7, the proposed models provide a larger CP and smaller PINAW compared with the other three models. In particular, LSTM reveals desired narrower PINAW and PINRW, but the CPs are not satisfactory. Moreover, the proposed model outperformed than other basic models in AWD. Therefore, the proposed approach have a distinct advantage in the performance of short-term power load interval forecasting. It is able to provide a satisfactory CP and restrict the interval width at the same time, which is the most important aspect of superiority of the proposed model.
On the other hand, in order to obtain a better performance and accuracy, the proposed approach is more complex. The algorithm with higher complexity often takes longer in practice. As Table 7 shows, compared with LSSVM and FITNET, the execution times of the proposed model are longer, which is the major disadvantage. However, with the development of hardware, the operational capability of computer can be improved, and the execution time can be reduced. Furthermore, as a kind of artificial intelligence technique, the fine-tuning of hyper-parameters in the proposed model will take time, which is a common situation in academic and industrial fields.

5.4. Further Research Prospect

This paper proposes a hybrid interval prediction model to predict the power load intervals. Compared with other basic models, this model has achieved good results in terms of coverage, interval width, and deviation error of the prediction interval. The model can obtain relatively high coverage under the condition of relatively narrow interval width, and the interval obtained can accurately reflect the changes of future short-term power load and provide more accurate and reliable support for power dispatch. On the other hand, for datasets with more complex changes and non-linear features, although the performance of proposed model is improved compared with the traditional models, it is still not ideal in some cases. For the unfavorable results caused by the characteristics of datasets, we may explore the following two aspects in future:
(a)
Finding and improving prediction methods that can better solve the non-linear characteristics of electrical loads, and improving the performance of predictive models in complex situations;
(b)
Fully analyzing the relevant characteristics in the power load data, selecting different models for different characteristics, and using ensemble learning to integrate and enhance the prediction results.

6. Conclusions

STLF is the basic work of power system planning and operation. However, the power load has regularity and certain randomness at the same time, which increases the difficulty of desired and reliable STLF. Moreover, compared with the prediction of specific points, interval prediction may provide more information for decision making in STLF. In this study, based on LUBE, we developed a novel hybrid model including data preprocessing, a multi-objective salp algorithm, and E–LUBE. In theory, such a hybrid model can reduce the influence of noise in a dataset and the parameter optimization process is more effective and efficient in E–LUBE.
In our proposed approach, we used a multi-objective optimization algorithm to search for the parameters of the neural network and reconstructed the cost function with double interval criterions instead of point criterions (such as MSE) in the traditional method. As Table 2, Table 3, Table 4 and Table 5 show, by comparing it with traditional methods, the proposed approach provides a higher CP and a lower interval width at the same time, which makes up for the lower CP and higher interval width of traditional methods.
In order to verify the performance of the proposed model and validate the impact of the constituent components in a hybrid model, we collected 16 samples involving four states using four quarters in Australia, and set several model comparisons in experiments
Furthermore, according to the comparison and analyses results, the conclusions are summarized as follows: (a) an efficient data preprocessing method was applied herein. Depending on the decomposition and reconstruction, this method can significantly improve the model performance in STLF. (b) Compared to the traditional prediction models based on neural networks, the newly developed E–LUBE method has an advantage in terms of comprehensive performance in interval prediction. It can be validated that the context layer with the information of the former hidden layer can improve model performance. (c) The introduction of the novel multi-objective algorithm MOSSA optimized the process of parameter search. The new cost function was based on a double-objective interval index that outperformed the traditional single-objective point error index (such as MSE) in interval prediction. (d) For STLF based on the E–LUBE mechanism, the width coefficient is an important factor. A larger width coefficient may lead to satisfactory CPs, and a smaller width coefficient may result in a satisfactory interval width. Therefore, in practice, the decision maker needs to adjust the width coefficient for specific demands. For example, we chose the width coefficient with a minimum interval width at the same time that the minimum demand of CP was guaranteed. (e) No matter how complex is the dataset, the proposed model always provides the best performance compared to benchmark models. However, because of the complexity of the data itself, some of the performance is not remarkable. In general, the proposed model provided a desired result in most cases.
Furthermore, in a power grid operator the proposed method has a strong practical application significance. A highly accurate forecasting method is one of the most important approaches used in improving power system management, especially in the power market [58]. In actual operation, for secure power grid dispatching, a control center has to make a prediction for the subsequent load. According to historical data, the dataset for the predictive model involved can be constructed. The results of the predictive model are able to provide the upper bound and lower bound of the load at some point in the future. Depending on the upper bound and lower bound, the control center can adjust the quantity of electricity on each charging line. Therefore, such a hybrid approach which can provide more accurate results can ensure the safe operation of the power grid and improve the economic efficiency of power grid operation.

Author Contributions

J.W. carried on the validation and visualization of experiment results; Y.G. carried on programming and writing of the whole manuscript; X.C. provided the overall guide of conceptualization and methodology.

Funding

This research was funded by National Natural Science Foundation of China (Grant number: 71671029) and Gansu science and technology program “Study on the forecasting methods of very short-term wind speeds” (Grant number: 1506RJZA187).

Acknowledgments

This work was supported by the National Natural Science Foundation of China (No. 71671029) and the Gansu science and technology program “Study on the forecasting methods of very short-term wind speeds” (No. 1506RJZA187).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviation

STLFShort-term load forecasting
PIPrediction intervals
PIWPrediction intervals width
PINAWPI normalized average width
ENNElman neural network
SNRSignal to noise ratio
IMFIntrinsic mode function
NstdNoise standard deviation
Pop_numTotal population number
MaxiterThe maximum number of iterations
CEEMDANThe complete ensemble empirical mode decomposition with adaptive noise
NNNeural networks
CPCoverage probability
LUBELower upper bound estimation
PINRWPI normalized root-mean-square width
DimIndividual parameter dimension
EMDEmpirical mode decomposition
MSEMean square error
NRNumber of realizations
RPRecurrence plot
MOSSAMulti-objective salp swarm algorithm
E-LUBELower upper bound estimation with ENN

References

  1. Fan, S.; Hyndman, R.J. Short-term load forecasting based on a semi-parametric additive model. IEEE Trans. Power Syst. 2012, 27, 134–141. [Google Scholar] [CrossRef]
  2. Du, P.; Wang, J.; Yang, W.; Niu, T. Multi-step ahead forecasting in electrical power system using a hybrid forecasting system. Renew. Energy 2018, 122, 533–550. [Google Scholar] [CrossRef]
  3. Shrivastava, N.A.; Khosravi, A.; Panigrahi, B.K. Prediction Interval Estimation of Electricity Prices using PSO tuned Support Vector Machines. IEEE Trans. Ind. Inform. 2015, 11. [Google Scholar] [CrossRef]
  4. Hagan, M.T.; Behr, S.M. The Time Series Approach to Short Term Load Forecasting. IEEE Trans. Power Syst. 1987, 2, 785–791. [Google Scholar] [CrossRef]
  5. Papalexopoulos, A.D.; Hesterberg, T.C. A regression-based approach to short-term system load forecasting. IEEE Trans. Power Syst. 1990, 5, 1535–1547. [Google Scholar] [CrossRef]
  6. Christiaanse, W.R. Short-Term Load Forecasting Using General Exponential Smoothing. IEEE Trans. Power Appar. Syst. 1971, PAS-90, 900–911. [Google Scholar] [CrossRef]
  7. Al-Hamadi, H.M.; Soliman, S.A. Short-term electric load forecasting based on Kalman filtering algorithm with moving window weather and load model. Electr. Power Syst. Res. 2004, 68, 47–59. [Google Scholar] [CrossRef]
  8. Metaxiotis, K.; Kagiannas, A.; Askounis, D.; Psarras, J. Artificial intelligence in short term electric load forecasting: A state-of-the-art survey for the researcher. Energy Convers. Manag. 2003, 44, 1525–1534. [Google Scholar] [CrossRef]
  9. Yoo, H.; Pimmel, R.L. Short term load forecasting using a self-supervised adaptive neural network. IEEE Trans. Power Syst. 1999, 14, 779–784. [Google Scholar] [CrossRef]
  10. Ho, K.L.; Hsu, Y.Y.; Chen, C.F.; Lee, T.E.; Liang, C.C.; Lai, T.S.; Chen, K.K. Short term load forecasting of taiwan power system using a knowledge-based expert system. IEEE Trans. Power Syst. 1990, 5, 1214–1221. [Google Scholar] [CrossRef]
  11. Mohandes, M. Support vector machines for short-term electrical load forecasting. Int. J. Energy Res. 2002, 26, 335–345. [Google Scholar] [CrossRef]
  12. Hong, W.C. Electric load forecasting by seasonal recurrent SVR (support vector regression) with chaotic artificial bee colony algorithm. Energy 2011, 36, 5568–5578. [Google Scholar] [CrossRef]
  13. Liu, Z.; Li, W.; Sun, W. A novel method of short-term load forecasting based on multiwavelet transform and multiple neural networks. Neural Comput. Appl. 2013, 22, 271–277. [Google Scholar] [CrossRef]
  14. Fan, G.F.; Peng, L.L.; Hong, W.C.; Sun, F. Electric load forecasting by the SVR model with differential empirical mode decomposition and auto regression. Neurocomputing 2016, 173, 958–970. [Google Scholar] [CrossRef]
  15. AlRashidi, M.R.; EL-Naggar, K.M. Long term electric load forecasting based on particle swarm optimization. Appl. Energy 2010, 87, 320–326. [Google Scholar] [CrossRef]
  16. Wang, J.; Li, L.; Niu, D.; Tan, Z. An annual load forecasting model based on support vector regression with differential evolution algorithm. Appl. Energy 2012, 94, 65–70. [Google Scholar] [CrossRef]
  17. Ghayekhloo, M.; Menhaj, M.B.; Ghofrani, M. A hybrid short-term load forecasting with a new data preprocessing framework. Electr. Power Syst. Res. 2015, 119, 138–148. [Google Scholar] [CrossRef]
  18. Zhang, X.; Wang, J. A novel decomposition-ensemble model for forecasting short-term load-time series with multiple seasonal patterns. Appl. Soft Comput. J. 2018, 65, 478–494. [Google Scholar] [CrossRef]
  19. Tian, C.; Hao, Y. A Novel Nonlinear Combined Forecasting System for Short-Term Load Forecasting. Energies 2018, 11, 714. [Google Scholar]
  20. Khotanzad, A.; Hwang, R.C.; Abaye, A.; Maratukulam, D. An Adaptive Modular Artificial Neural Network Hourly Load Forecaster and its Implementation at Electric Utilities. IEEE Trans. Power Syst. 1995, 10, 1716–1722. [Google Scholar] [CrossRef]
  21. Bishop, C.M. Neural networks for pattern recognition. J. Am. Stat. Assoc. 1995, 92, 482. [Google Scholar] [CrossRef]
  22. Hwang, J.T.G.; Ding, A.A. Prediction Intervals for Artificial Neural Networks. J. Am. Stat. Assoc. 1997, 92, 748–757. [Google Scholar] [CrossRef]
  23. Heskes, T. Practical confidence and prediction intervals. Adv. Neural Inf. Process. Syst. 1997, 9, 176–182. [Google Scholar]
  24. Nix, D.A.; Weigend, A.S. Estimating the mean and variance of the target probability distribution. In Proceedings of the 1994 IEEE International Conference on Neural Networks (ICNN’94), Orlando, FL, USA, 28 June–2 July 1994; Volume 1, pp. 55–60. [Google Scholar]
  25. Van Hinsbergen, C.P.I.; van Lint, J.W.C.; van Zuylen, H.J. Bayesian committee of neural networks to predict travel times with confidence intervals. Transp. Res. Part C Emerg. Technol. 2009, 17, 498–509. [Google Scholar] [CrossRef]
  26. Khosravi, A.; Nahavandi, S.; Creighton, D. Construction of optimal prediction intervals for load forecasting problems. IEEE Trans. Power Syst. 2010, 25, 1496–1503. [Google Scholar] [CrossRef]
  27. Da Silva, A.P.A.; Moulin, L.S. Confidence intervals for neural network based short-term load forecasting. IEEE Trans. Power Syst. 2000, 15, 1191–1196. [Google Scholar] [CrossRef]
  28. Khosravi, A.; Nahavandi, S.; Creighton, D.; Atiya, A.F. Lower upper bound estimation method for construction of neural network-based prediction intervals. IEEE Trans. Neural Netw. 2011, 22, 337–346. [Google Scholar] [CrossRef] [PubMed]
  29. Deb, K.; Agrawal, S.; Pratap, A.; Meyarivan, T. A fast elitist non-dominated sorting genetic algorithm for multi-objective optimization: NSGA-II. Parallel Probl. Solving Nat. PPSN VI 2000, 849–858. [Google Scholar] [CrossRef]
  30. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  31. Coello Coello, C.A.; Lechuga, M.S. MOPSO: A proposal for multiple objective particle swarm optimization. In Proceedings of the 2002 Congress on Evolutionary Computation, CEC 2002, Honolulu, HI, USA, 12–17 May 2002; Volume 2, pp. 1051–1056. [Google Scholar]
  32. Padhye, N. Topology Optimization of Compliant Mechanism Using Multi-objective Particle Swarm Optimization. In Proceedings of the 10th Annual Conference Companion on Genetic and Evolutionary Computation, Atlanta, GA, USA, 12–16 July 2008; pp. 1831–1834. [Google Scholar]
  33. Alaya, I.; Solnon, C.; Ghedira, K. Ant Colony Optimization for Multi-Objective Optimization Problems. In Proceedings of the 19th IEEE International Conference on Tools with Artificial Intelligence (ICTAI 2007), Patras, Greece, 29–31 October 2007; pp. 450–457. [Google Scholar] [CrossRef]
  34. Xue, F.; Sanderson, A.C.; Graves, R.J. Pareto-based multi-objective differential evolution. In Proceedings of the 2003 Congress on Evolutionary Computation, Canberra, Australia, 8–12 December 2003; Volume 2, pp. 862–869. [Google Scholar]
  35. Mirjalili, S.Z.; Mirjalili, S.; Saremi, S.; Faris, H.; Aljarah, I. Grasshopper optimization algorithm for multi-objective optimization problems. Appl. Intell. 2017. [Google Scholar] [CrossRef]
  36. Knowles, J.D.; Corne, D.W. Approximating the Nondominated Front Using the Pareto Archived Evolution Strategy. Evol. Comput. 2000, 8, 149–172. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Wang, J.; Yang, W.; Du, P.; Niu, T. A novel hybrid forecasting system of wind speed based on a newly developed multi-objective sine cosine algorithm. Energy Convers. Manag. 2018, 163, 134–150. [Google Scholar] [CrossRef]
  38. Du, P.; Wang, J.; Guo, Z.; Yang, W. Research and application of a novel hybrid forecasting system based on multi-objective optimization for wind speed forecasting. Energy Convers. Manag. 2017, 150, 90–107. [Google Scholar] [CrossRef]
  39. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef] [Green Version]
  40. Service, T.C. A No Free Lunch theorem for multi-objective optimization. Inf. Process. Lett. 2010, 110, 917–923. [Google Scholar] [CrossRef]
  41. Rodriguez, P.; Wiles, J.; Elman, J.L. A Recurrent Neural Network that Learns to Count. Conn. Sci. 1999, 11, 5–40. [Google Scholar] [CrossRef] [Green Version]
  42. Chandra, R.; Zhang, M. Cooperative coevolution of Elman recurrent neural networks for chaotic time series prediction. Neurocomputing 2012, 86, 116–123. [Google Scholar] [CrossRef]
  43. Cacciola, M.; Megali, G.; Pellicanó, D.; Morabito, F.C. Elman neural networks for characterizing voids in welded strips: A study. Neural Comput. Appl. 2012, 21, 869–875. [Google Scholar] [CrossRef]
  44. Wang, J.J.; Zhang, W.; Li, Y.; Wang, J.J.; Dang, Z. Forecasting wind speed using empirical mode decomposition and Elman neural network. Appl. Soft Comput. 2014, 23, 452–459. [Google Scholar] [CrossRef]
  45. Huang, N.; Shen, Z.; Long, S.; Wu, M.; Shih, H.; Zheng, Q.; Yen, N.; Tung, C.; Liu, H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  46. Wu, Z.; Huang, N.E. Ensemble Empirical Mode Decomposition: A Noise-Assisted Data Analysis Method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
  47. Yeh, J.R.; Shieh, J.S.; Huang, N.E. Complementary ensemble empirical mode decomposition: A novel noise enhanced data analysis method. Adv. Adapt. Data Anal. 2010, 2, 135–156. [Google Scholar] [CrossRef]
  48. Torres, M.E.; Colominas, M.A.; Schlotthauer, G.; Flandrin, P. A complete ensemble empirical mode decomposition with adaptive noise. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing, Prague, Czech Republic, 22–27 May 2011; Volume 7, pp. 4144–4147. [Google Scholar] [CrossRef]
  49. Elman, J.L. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  50. Quan, H.; Srinivasan, D.; Khosravi, A. Uncertainty handling using neural network-based prediction intervals for electrical load forecasting. Energy 2014, 73, 916–925. [Google Scholar] [CrossRef]
  51. Coello Coello, C.A. Evolutionary multi-objective optimization: Some current research trends and topics that remain to be explored. Front. Comput. Sci. China 2009, 3, 18–30. [Google Scholar] [CrossRef]
  52. Branke, J.; Kaußler, T.; Schmeck, H. Guidance in evolutionary multi-objective optimization. Adv. Eng. Softw. 2001, 32, 499–507. [Google Scholar] [CrossRef] [Green Version]
  53. Deb, K. Advances in Evolutionary Multi-objective Optimization. In Search Based Software Engineering; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–26. ISBN 978-3-642-33119-0. [Google Scholar]
  54. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  55. Wang, J.; Niu, T.; Lu, H.; Guo, Z.; Yang, W.; Du, P. An analysis-forecast system for uncertainty modeling of wind speed: A case study of large-scale wind farms. Appl. Energy 2018, 211, 492–512. [Google Scholar] [CrossRef]
  56. Eckmann, J.-P.; Kamphorst, S.O.; Ruelle, D. Recurrence Plots of Dynamical Systems. Europhys. Lett. 1987, 4, 973–977. [Google Scholar] [CrossRef]
  57. Marwan, N.; Wessel, N.; Meyerfeldt, U.; Schirdewan, A.; Kurths, J. Recurrence-plot-based measures of complexity and their application to heart-rate-variability data. Phys. Rev. E Stat. Phys. Plasmas Fluids Relat. Interdiscip. Top. 2002, 66. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Shu, F.; Luonan, C. Short-term load forecasting based on an adaptive hybrid method. IEEE Trans. Power Syst. 2006, 21, 392–401. [Google Scholar]
Figure 1. The structure of the lower bound and upper bound estimation (LUBE) based on the Elman neural network.
Figure 1. The structure of the lower bound and upper bound estimation (LUBE) based on the Elman neural network.
Energies 11 01561 g001
Figure 2. Forecasting flowchart of the proposed hybrid model.
Figure 2. Forecasting flowchart of the proposed hybrid model.
Energies 11 01561 g002
Figure 3. Data description of experiments. (a) Location of sample sites; (b) Division of train set and test set; (c) Structure of input set and output set; and (d) Entropy of each IMF).
Figure 3. Data description of experiments. (a) Location of sample sites; (b) Division of train set and test set; (c) Structure of input set and output set; and (d) Entropy of each IMF).
Energies 11 01561 g003
Figure 4. Boxplot of the entire set of data samples.
Figure 4. Boxplot of the entire set of data samples.
Energies 11 01561 g004
Figure 5. Performance of different samples with the width coefficient 0.05.
Figure 5. Performance of different samples with the width coefficient 0.05.
Energies 11 01561 g005
Figure 6. Performance of different samples with the width coefficient 0.025.
Figure 6. Performance of different samples with the width coefficient 0.025.
Energies 11 01561 g006
Figure 7. Interval prediction plot of partial samples in NSW.
Figure 7. Interval prediction plot of partial samples in NSW.
Energies 11 01561 g007
Figure 8. Recurrence plot of the samples obtained from the four quarters in VIC.
Figure 8. Recurrence plot of the samples obtained from the four quarters in VIC.
Energies 11 01561 g008aEnergies 11 01561 g008b
Table 1. Related parameters in hybrid model.
Table 1. Related parameters in hybrid model.
Submodels and ParametersValue
Elman Neural Network (ENN)
Inputnum6
Hiddennum13
Outputnum3
Train.epoch500
Train.lr0.1
Train.func“Adam”
Complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN)
Nstd0.2
NR200
Maxiter100
Multi-objective salp swarm algorithm (MOSSA)
Dim754
Lb−2
Ub2
Obj_no2
Pop_num50
Table 2. Results of different models with α = 0.05 for sample in New South Wales (NSW) and Tasmania (TAX).
Table 2. Results of different models with α = 0.05 for sample in New South Wales (NSW) and Tasmania (TAX).
ModelsCriterion (%)
NSWFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE72.6612.1312.900.03089.2217.2018.010.02075.8816.9717.140.02272.3612.2313.850.085
E-LUBE72.7512.0012.920.03786.0817.3718.200.02374.0016.1416.410.02587.1712.1013.580.021
PO-E-LUBE79.6711.8012.580.01092.4217.2517.830.00782.7514.9815.290.04989.7511.2612.860.033
IO-E-LUBE83.0011.7212.500.03296.5017.1917.680.00383.4214.9615.250.11590.2510.6211.650.015
CEEMDAN-E-LUBE83.2511.6712.360.02197.8317.1017.900.00291.2514.7515.280.00792.0810.4811.230.022
CEEMDAN-PO-E-LUBE86.0811.5212.240.09997.8316.6617.260.00298.3314.4714.990.00593.9210.4610.970.043
CEEMDAN-IO-E-LUBE93.5811.3612.240.016100.0016.2716.660.000100.0014.1815.720.00095.758.2810.260.007
TAXFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE88.9320.8421.250.62477.6818.7519.570.53396.8818.9319.210.02194.8528.3328.710.015
E-LUBE94.1720.6321.100.02078.0817.4518.970.12797.5818.1718.800.00896.0025.0425.520.016
PO-E-LUBE94.7520.1220.630.00978.5016.5517.610.13697.6717.9418.450.01197.2525.1325.480.007
IO-E-LUBE95.9220.2920.650.00778.3314.3815.740.25797.4217.9618.600.00697.3324.6724.980.008
CEEMDAN-E-LUBE98.5019.9820.360.00585.8316.6416.930.03299.0017.3518.820.00198.9223.4823.690.002
CEEMDAN-PO-E-LUBE99.0019.9220.310.00187.9216.0116.300.03499.7517.1118.560.00199.5824.7925.040.002
CEEMDAN-IO-E-LUBE99.0819.9820.300.00688.1717.2517.750.04699.2517.3719.040.00199.2524.2024.490.001
Table 3. Results of different models with α = 0.05 for sample in Queensland (QLD) and Victoria (VIC).
Table 3. Results of different models with α = 0.05 for sample in Queensland (QLD) and Victoria (VIC).
ModelsCriterion (%)
QLDFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE94.8219.4720.120.01296.5820.8621.750.00492.7819.5820.110.00299.3319.6920.470.001
E-LUBE95.5017.2918.380.00796.5020.9621.540.00395.1719.3219.850.00199.3319.1019.800.000
PO-E-LUBE98.8317.5818.420.00196.6720.5721.070.00397.8319.0719.540.00199.7519.3820.240.001
IO-E-LUBE99.1717.0317.770.00696.7820.1821.860.00397.4218.7919.270.00299.7518.5319.040.000
CEEMDAN-E-LUBE99.7518.1718.880.00099.8320.6921.600.00099.9219.3219.860.00099.8319.4920.100.000
CEEMDAN-PO-E-LUBE99.5018.2618.970.00199.8320.1921.710.00099.9219.4219.930.00099.8319.2619.800.000
CEEMDAN-IO-E-LUBE99.2516.9617.670.002100.0020.1721.660.00099.9219.1719.530.00099.9218.4218.900.001
VICFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE70.859.7710.620.04290.2517.8518.370.35183.4213.6114.290.01370.227.587.900.018
E-LUBE72.089.4110.290.02091.9216.6617.410.12785.3313.2514.020.00873.677.047.870.016
PO-E-LUBE76.009.4110.400.00992.7515.8216.340.13685.1713.0914.120.01176.506.947.840.007
IO-E-LUBE78.839.1810.180.00795.7516.6117.320.25785.5812.8213.910.00676.257.048.220.008
CEEMDAN-E-LUBE78.429.2910.720.00598.8315.4016.330.03288.1712.5813.300.00180.927.238.010.002
CEEMDAN-PO-E-LUBE82.679.2210.430.00198.7517.1218.120.03491.6712.7013.570.00180.507.077.930.002
CEEMDAN-IO-E-LUBE83.259.2110.130.00299.9217.0018.010.03694.0813.1713.790.00182.086.907.780.001
Table 4. Results of different models with α = 0.025 for sample in NSW and TAX.
Table 4. Results of different models with α = 0.025 for sample in NSW and TAX.
NSWFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE58.206.566.920.36070.669.089.600.16344.857.828.750.53160.427.5211.830.158
E-LUBE58.585.766.500.28670.839.239.820.12046.927.608.260.27660.757.3712.100.146
PO-E-LUBE67.506.006.510.06173.678.418.790.02752.337.828.640.51576.085.666.600.126
IO-E-LUBE67.675.916.510.13873.088.588.380.06251.007.758.020.23177.425.446.070.161
CEEMDAN-E-LUBE69.505.806.670.14982.507.778.070.05767.257.898.650.04977.925.375.910.120
CEEMDAN-PO-E-LUBE69.755.836.130.11894.257.848.860.09586.757.287.750.08485.674.935.400.069
CEEMDAN-IO-E-LUBE70.505.686.240.08596.177.968.600.00287.257.487.640.01486.674.635.130.083
TAXFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE70.6111.5211.830.18745.788.8410.800.53476.879.7810.850.12672.3112.6114.040.120
E-LUBE72.0810.1410.390.16351.428.679.250.77474.679.329.860.07376.5812.8313.650.097
PO-E-LUBE73.0810.2911.000.09953.507.658.870.18876.429.049.330.08578.7512.1112.400.050
IO-E-LUBE74.679.6510.320.09052.678.509.400.27879.088.909.750.13778.2512.6013.270.065
CEEMDAN-E-LUBE88.0810.1110.510.03672.759.279.710.17990.008.538.840.01186.2511.5711.790.018
CEEMDAN-PO-E-LUBE88.0810.0410.240.01972.008.578.820.19491.508.608.890.00990.7511.5611.840.013
CEEMDAN-IO-E-LUBE88.089.7810.210.01574.258.468.970.17193.428.798.300.01391.0811.4911.840.018
Table 5. Results of different models with α = 0.025 for sample in QLD and VIC.
Table 5. Results of different models with α = 0.025 for sample in QLD and VIC.
ModelsCriterion (%)
QLDFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE70.559.9410.820.15872.7510.4410.980.06973.369.9710.350.07585.399.8310.760.009
E-LUBE74.759.2910.000.02474.4210.1310.430.06577.589.579.870.05887.179.7710.010.008
PO-E-LUBE80.428.719.100.06478.0810.8211.270.04980.759.679.890.04590.4210.1110.570.008
IO-E-LUBE86.838.269.710.30978.3310.1611.610.07280.509.509.750.04689.179.6610.110.003
CEEMDAN-E-LUBE82.759.549.880.15694.5010.7911.230.01291.589.7010.180.00393.429.2510.200.004
CEEMDAN-PO-E-LUBE91.008.729.010.08595.5810.5411.010.00898.759.7710.210.00295.179.079.390.003
CEEMDAN-IO-E-LUBE91.428.418.770.04895.2510.1311.170.00499.759.649.910.00295.338.528.840.003
VICFirst QuarterSecond QuarterThird QuarterFourth Quarter
CPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWDCPPINAWPINRWAWD
LUBE50.165.396.830.88270.038.258.800.53768.916.557.030.62250.063.554.100.486
E-LUBE56.755.106.590.92372.338.108.630.59670.676.637.040.69250.003.474.140.575
PO-E-LUBE62.754.835.770.31276.258.348.910.08170.676.717.310.60950.253.534.040.127
IO-E-LUBE66.754.635.470.80579.257.828.790.13170.006.547.030.01349.503.445.220.226
CEEMDAN-E-LUBE60.004.755.280.43375.838.048.310.05973.426.647.040.03854.503.684.450.262
CEEMDAN-PO-E-LUBE65.834.645.350.12383.508.669.270.12778.926.987.320.03353.254.004.930.473
CEEMDAN-IO-E-LUBE68.254.565.640.08194.087.938.230.04485.756.186.700.04369.333.444.450.161
Table 6. Sensitivity analysies results of different hyper-parameters.
Table 6. Sensitivity analysies results of different hyper-parameters.
MetricsThe Number of Salp Populations in MOSSA
102030405060708090100
CP95.2195.3297.0196.6698.6998.3398.5097.8597.4698.10
PINAW17.3417.6313.5213.8913.1013.0213.6414.0513.8213.72
PINRW18.2818.0514.1814.3513.8413.7514.3614.9214.6714.53
Time(s)42545247252454859366873486910.45
MetricsThe Initial Threshold of Parameters
[−0.5, 0.5][−1, 1][−2, 2][−3, 3][−5, 5]
CP97.6598.8499.0098.2696.89
PINAW14.3612.8212.8013.1213.68
PINRW15.3213.4613.4213.9414.36
Time(s)433450461453484
Table 7. Consistency analysis results of some basic models and new datasets.
Table 7. Consistency analysis results of some basic models and new datasets.
ModelsNSW-2018-NEWVIC-2018-NEW
CPPINAWPINRWAWDTime(s)CPPINAWPINRWAWDTime(s)
Proposed89.5815.5116.580.023593.2989.0811.5012.660.065564.55
LSSVM78.6715.9517.640.677495.3286.6712.1613.010.026486.85
FITNET72.0816.2517.240.043405.5274.3311.6612.950.087300.72
LSTM44.005.475.920.3821199.0459.835.285.790.250947.78
ModelsNSW-2017-2QVIC-2017-4Q
CPPINAWPINRWAWDTime(s)CPPINAWPINRWAWDTime(s)
Proposed100.0016.2716.660.000543.2082.086.907.780.001526.39
LSSVM94.4216.6717.120.038409.2471.587.6010.590.097435.50
FITNET94.3315.8317.290.012402.1274.337.888.940.076504.31
LSTM70.676.016.370.101753.6065.333.103.550.248732.21

Share and Cite

MDPI and ACS Style

Wang, J.; Gao, Y.; Chen, X. A Novel Hybrid Interval Prediction Approach Based on Modified Lower Upper Bound Estimation in Combination with Multi-Objective Salp Swarm Algorithm for Short-Term Load Forecasting. Energies 2018, 11, 1561. https://doi.org/10.3390/en11061561

AMA Style

Wang J, Gao Y, Chen X. A Novel Hybrid Interval Prediction Approach Based on Modified Lower Upper Bound Estimation in Combination with Multi-Objective Salp Swarm Algorithm for Short-Term Load Forecasting. Energies. 2018; 11(6):1561. https://doi.org/10.3390/en11061561

Chicago/Turabian Style

Wang, Jiyang, Yuyang Gao, and Xuejun Chen. 2018. "A Novel Hybrid Interval Prediction Approach Based on Modified Lower Upper Bound Estimation in Combination with Multi-Objective Salp Swarm Algorithm for Short-Term Load Forecasting" Energies 11, no. 6: 1561. https://doi.org/10.3390/en11061561

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop