1. Introduction
Power load forecasting is to study the law of power development and build models between power demand and characteristics based on historical data, and then forecast future load [
1]. Many operations in power systems rely heavily on precise load forecasting, for instance, operation, maintenance, and planning of power systems. Therefore, improving the accuracy of load forecasting can bring benefits to power systems.
Short-term load forecasting (STLF) usually predicts load from one hour to several weeks [
2]. In reality, utility companies pay more attention to short-term load forecasting, as it plays a significant role in the control of spinning reserve, optimum of unit commitment, and evaluation of sales contracts. Therefore, researchers have made great efforts in short-term load forecasting. The various approaches developed for STLF can be divided into three categories: (1) Traditional statistical methods, for instance, ARMAX [
3], ARIMA [
4], and autoregressive based time varying model [
5]; (2) the artificial intelligence (AI) methods, such as support vector regression (SVR) [
6], artificial neural networks (ANN) [
7], and gradient boosting [
8]; and (3) the hybrid method, such as hybridizing extended Kalman Filter (EKF) and ELM [
9], a hybrid STLF approach integrating linear regression and neural network [
10], and wavelet neural network [
11].
In the past decade, the artificial neural network (ANN) has been widely concerned in a variety fields, such as the electronic and finance industry, biomedical applications, image processing, natural language processing, etc. With the continuous improvement of prediction accuracy requirements, ANN-based approaches have become a hot topic in recent years. A novel pooling-based deep recurrent neural network (PDRNN) is proposed for household load forecasting. The outperformance of PDRNN is confirmed by experiments on 920 smart metered customers from Ireland [
12]. However, the main disadvantage of RNN is that the activation function of RNN uses chain rules to operate the gradient descent algorithm, which will lead to the problem of gradient vanishing. To alleviate the problem, researchers have proposed long short-term memory (LSTM) [
13], which changes the internal structure and transfers the state of hidden layers through the concept of gates. Thus, LSTM can efficiently mitigate the gradient vanishing of RNN. Recently, LSTM and gated recurrent units (GRUs) have achieved good results in long-term horizon forecasting [
14,
15,
16]. By using an LSTM-based method to exploit the long-term dependencies of electric load time series, the prediction accuracy of load forecasting is improved [
17]. The experiment results show that the method has a good effect in complex electrical load forecasting. A STLF method using gated recurrent unit neural networks with multi-source data has been developed [
18]. The proposed method is superior to other existing methods, such as RNNs, BPNNs, Stacked Auto Encoders (SAEs), and LSTM.
Except for the aforementioned representative methods, the encoder–decoder network is getting popular in the field of prediction because of its success in machine translation. The main idea is to encode the source sequence as a fixed-length vector and use the decoder to generate a target sequence. The encoder–decoder network has achieved great success in various prediction applications, e.g., vehicle trajectory prediction [
19], crowd density estimation [
20], and human trajectory prediction [
21]. The major problem with encoder–decoder networks is that their performance deteriorates rapidly with the increase of input sequence length. To resolve the issue, attention mechanisms [
22] are used to model dependencies in sequences without regard for their actual distances in the sequence. As attention-based encoder–decoder networks have shown their efficiency for machine translation, urban air quality inference [
23], and diagnosis prediction in healthcare [
24], it is reasonable to exploit their usage in load forecasting.
Hybrid technology is a combination of two or more algorithms. For load forecasting, past studies have shown that hybrid technologies often outperform the individual forecasts. The paper proposes a short-term load forecast model by similar shape functional time series [
25]. An approach for short-term load forecasting was developed by integrating a regression model with a seasonal exponential adjustment method [
26]. To improve forecasting accuracy, the author adopted an algorithm integrating support vector regression and differential evolution for short-term load forecast [
27]. By integrating time series multi-feature regression with seasonal and trend adjustment, a transfer learning approach for cross-building energy forecasting was developed [
28]. The seasons of the year and human activities may induce unique seasonal patterns. Similarly, trends may vary from location-to-location and need to be taken into consideration. Therefore, to gain better load forecasting results, it is necessary to consider seasonality differences and trend of the data.
To this end, we developed a hybrid approach for short-term load forecasting, which combines the seasonal and trend adjustment technique and multi-head attention-based encoder–decoder framework. We named this framework seasonal and trend adjustment attention encoder–decoder (STA–AED). The seasonal and trend decomposing technique was employed to decompose the raw electric load series into three components: Seasonal component, trend component, and irregular component. Then, an encoder–decoder network based on multi-head attention mechanism was employed on every sub-component. The proposed hybrid prediction model can achieve the expected result and improve the forecasting accuracy extensively.
The contributions of this paper are summarized as follows:
We developed a novel hybrid prediction framework: A hybrid STA–AED framework was developed to predict electric load. Instead of processing the original electric load series directly, the proposed approach splits data into three components by the seasonal and trend decomposing technique first.
Based on the multi-head attention mechanism, we developed an attention-based encoder–decoder architecture for power load forecasting.
The model was implemented in univariate load series data only. For other forecasting approaches, a variety variables were considered as predictive model inputs to improve the accuracy of prediction, such as holiday arrangement, weather, and economic environment. However, our proposed model was implemented without utilizing other input features and gained better prediction results still.
Our approach was evaluated on a real-world dataset. Compared with other counterpart models, our approach achieved the best prediction accuracy in 14 out of 15 experiments.
The remaining sections are organized as follows: In
Section 2, we elaborate on the proposed STA–AED framework for short-term load forecasting. In
Section 3, details about the dataset, the experimental setting, and the selected counterpart models are illustrated. In addition, analysis and comparisons are provided. Finally, the conclusion is shown in
Section 4.
2. The Proposed Method
2.1. The Overview of the Proposed Framework
As shown in
Figure 1, the major procedures of the novel hybrid STA–AED approach are as follows:
Step 1: Decomposing original load series data into three sub-components: Seasonal component, trend component, and irregular component.
Step 2: Build encoder–decoder models based on multi-head attention mechanism for the decomposed seasonal, trend, and irregular components, separately.
Step 3: Final forecast of load series are obtained by linearly adding all of sub predictions from step 2.
2.2. Seasonal and Trend Adjustment
For time series regression, seasonal and trend adjustment is a process to improve the properties of the parameter estimates. Seasonal and trend adjustment often helps better understand time series data. For instance, if electric consumption in June is up 20% from May. By decomposed data, it is easier to find out that rising of electric consumption is mainly caused by seasonal effects associated with weather. Therefore, to gain better power forecasting results, it is important to decompose original electric series first.
Denote as the actual time series value at period t. The time series seasonal and trend adjustment approach decomposes into three components: A trend component, a seasonal component, and an irregular component. These three components are combined by an additive or a multiplicative model.
An additive decomposition model is defined as:
And a multiplicative decomposition model is defined as:
where
= trend value at time period
t = seasonal index at time period
t = irregular index at time period
t.The multiplicative model is suitable for situations where the seasonal fluctuations change over time. When the seasonal fluctuations do not depend on the level of the time series, it is good to use the additive decomposition model.
Other than above two basic decomposition models, the decomposition model can also be defined as:
when the seasonal component is changing slowly, we can use model (3) to decompose time series. As a result, the non-seasonal forecasting model can be applied to a combination of trend and irregular components. A similar model can be applied to the multiplicative model as well:
In this paper, we used the electric load data from GEFCOM2012 [
29] to explain the implementation of the hybrid model.
Figure 2 shows the seasonal and trend decomposition process for load demand of zone 1. As
Figure 2 shows, the electric load pattern shows randomness. However, it can be seen from the analysis of weekly electricity seasonal decomposition results that the seasonal and trend components have less volatility, which makes it easier to predict future seasonal and trend components. Although regularity of the irregular component is not obvious, the magnitude of the irregular component is much smaller than the magnitude of the original electric load series. As a result, the prediction accuracy of the method applying the prediction model on three sub-components separately can be better than that of the method making a prediction on original electric load series directly.
2.3. Attention-Based Models
The attention-based model proposed in this paper is depicted in
Figure 3. Basically, the model consists of three components: An encoder network, a multi-head attention model, and a decoder network. Composed of a deep recurrent neural network, the encoder network reads the sequence of electric load series x and calculates a sequence of encoded features
. The multi-head attention mechanism generalizes the output of the encoder according to the current state of the decoder to calculate a context vector
, as depicted in
Section 2.4. The decoder network is comprised of a deep recurrent neural network modeling an output distribution over the sequence of previous prediction.
2.4. Encoder with Multi-Head Attention
For a regular encoder–decoder architecture, an encoder transforms the input signal
into a representation vector
.
where
is a hidden state at time
t. The framework of an encoder varies in different applications. For instance, a recurrent neural network (RNN) is widely chosen as an encoder in machine translation. In the field of image caption, it is natural to use a convolutional neural network (CNN) as the encoder. While applied to load forecasting, LSTM is the most appropriate algorithm, since LSTM can efficiently mitigate the problem caused by the gradient vanishing of RNN and transfers the state of hidden layers through forget gate, input gate, and output gate.
By encoding the necessary information in a sequence, bidirectional long short-term memory (BiLSTM) brings outstanding performance [
30]. Therefore, BiLSTM was selected as an encoder to take the temporal relation of electric load into account. BiLSTM splits a normal LSTM neuron into two directions, one is the forward states, and the other is backward states. In addition, the output of these two states is not related. By using two directions, the future and past sequence information of the current electric load can be applied.
Figure 4 is the flow diagram of BiLSTM. First, the forward LSTM reads the forward input sequence (from
to
) and computes the forward hidden states (
). At the same time, the backward LSTM reads the sequence in the opposite direction to obtain a series of backward hidden states (
). By concatenating the forward hidden state
and the backward hidden state
, an annotation
for each
is obtained. As a result, the annotation
has the knowledge of both the preceding and the following electric loads.
Multi-headed attention (MHA) was first applied in machine translation [
31]. Now, we explore this work in load forecasting. In MHA architecture, each head can produce a different distribution of attention. Therefore, each head plays a different role in attending the encoder output, which makes it easier for the decoder to retrieve information from the encoder. In detail, the model applies
M independent attention heads. Each calculates attention values,
, for
:
Then, each attention value is converted into a soft attention weight by a softmax operation, which is employed to calculate a summary of encoder features,
:
Finally, the individual summaries are concatenated together to calculate the overall context vector: .
2.5. Decoder
In the decoder, we combine the weighted summed context vector at a future step with the last output of decoder . Then, the hidden state of decoder is updated with , where represents an LSTM unit used in the decoder.
Eventually, the final prediction can be calculated as below:
where
is a concatenation of the decoder hidden state and the context vector.
The parameters and map the concatenation to the size of the decoder hidden states. We use a linear function with weights and bias to produce the final prediction output.
3. Experiments and Results
To verify the forecasting accuracy of the proposed STA–AED model, a real-world dataset from the United States was used for validation purposes in this section, and the experiments are described and analyzed elaborately. The comparisons with random forest, gradient boosting decision tree (GBDT), GRU, encoder–decoder, encoder–decoder with multi-head attention, and the proposed STA–AED model are also analyzed in detail.
3.1. Dataset Description
This paper used an electric load dataset from GEFCOM2012 [
29], which includes hour loads from 20 zones in the United States. We used load data from 1 January 2007 to 30 June 2008 in this paper. The data sampling interval was one hour. All test data consisted of 13,128 observations. In this study, the load data from 1 January 2007 to 31 May 2008 were used as a training set, and the load data from 1 June 2008 to 20 June 2008, which are the data of the last month in the original dataset, were used as a testing set. As an example, hourly load data in zone 12 and zone 15 from 1 May 2008 to 30 June 2008 are shown in
Figure 5.
In the original dataset, there were 20 geographical zones in all. According to analysis in [
29], data from zones 2, 3, 4, 8, and zone 10 were not used because of data duplication. Therefore, we only used data from the remaining 15 zones in this paper.
In this paper, we used the additive model to decompose the electric load series, and the decomposing interval was set as weekly.
Figure 6 shows a sample of the seasonal and trend decomposition process for load demand of zone 12. It demonstrates that trend and seasonal components can be separated from original load demand data by seasonal and trend decomposition. Both trend and seasonal components exhibit a clear pattern.
3.2. Model Evaluation Indexes
The mean absolute percentage error (MAPE) and the root mean square error (RMSE) were calculated to evaluate forecasting accuracy. The smaller the values of MAPE and RMSE, the better the forecasting accuracy. The error measures are defined as follows:
where
N is the forecasting period, and
and
are the
predicted and actual values, respectively.
3.3. Method Comparison
To evaluate the effectiveness of the proposed approach in this paper, two traditional machine learning methods and another three deep learning approaches were implemented as counterparts for compassion purpose. The following are simple introductions for these five counterparts:
- (1)
Random forest regressor: Random forest regressor is one of the most widely used traditional machine learning methods.
- (2)
Gradient boosting decision tree (GBDT): GBDT is one of the commonly used machine learning algorithms. It is very popular in load forecasting because of its excellent automatic feature combination ability and efficient operation.
- (3)
Gated recurrent units (GRUs): GRUs are a widely used variant of the recurrent neural network (RNN). GRUs are similar to long short-term memory (LSTM) but with fewer parameters than LSTM.
- (4)
Encoder–decoder: This approach is based on encoder–decoder architecture without applying the attention mechanism.
- (5)
Encoder–decoder with multi-head attention: The only difference with the method proposed in this paper is that the input data are not processed by the seasonal and trend decomposing technique.
To make a fair comparison, we also applied seasonal and trend adjustment on all counterpart methods. We use the prefix “STA”, which stands for seasonal and trend adjustment—for instance, STA–GBDT.
3.4. The Detailed Exprimental Setting
In this paper, the proposed method was used to predict the load of the next 24 h, and the past 3 × 24 h load data were used as the input variable of the model. The learning rate was 0.001 and the batch size for both the encoder and decoder were 128. The hidden neuron of the LSTM module was set as 128. The rectified linear unit (ReLU) function was used as the active function. The Adam Optimizer was used to optimize the parameters of the model by performing mini-batch stochastic gradient descent (SGD). The model was trained by the objective function of the standard mean square error. We implemented seasonal and trend decomposition in statsmodels and deep learning method in the PyTorch framework, and ran all the experiments with one GTX 1080Ti GPU.
3.5. Experimental Results and Analysis
The proposed method was used to forecast the load data from 1 June 2008 to 30 June 2008. The prediction results of zone 7 are depicted in
Figure 7. As
Figure 7 shows, the prediction values are very close to the actual values, except the peak and bottom part.
Table 1 and
Table 2 summarize the experimental results for all 15 zones, in terms of MAPE and RMSE.
Table 1 and
Table 2 show that the MAPE and RMSE of the random forest regressor are the largest in 12 out of 15 zones. The performance of GBDT is slightly better than the random Forest regressor. It is obvious that the performance of the deep neural networks is much superior than GBDT and the random forest regressor. The results of the GRU modules are a little better than GBDT and the random forest regressor, and the results of the encoder–decoder architecture are better than GRU. The prediction accuracy of two attention-based modules are much better than the rest of the modules. From
Table 1 and
Table 2, we observe that in 14 out of 15 zones, the solution generated by the proposed model is associated with the lowest RMSE and MAPE values. In fact, in 13 out of 15 zones, MAPE obtained by the STA–AED model is less than 5%, which shows model superiority in the field of short-term load forecasting. When applying seasonal and trend adjustment on both GBDT and the random forest regressor, the prediction accuracy got worse in all 15 experiments—while for all deep learning approaches, the prediction accuracy could be improved in 14 out 15 experiments by applying the seasonal and trend adjustment technique. We think it is because seasonal and trend adjustment takes better effect in deep learning environment. In conclusion, our proposed STA–AED model can make a more accurate forecast by integrating the seasonal and trend decomposing technique and an encoder–decoder network with a multi-head attention mechanism.
Meanwhile, our proposed model is more stable than the other forecasting approaches.
Figure 8 and
Figure 9 show the comparison results for all 6 approaches in all 15 zones, in terms of MAPE and RMSE, respectively.
Figure 8 and
Figure 9 show that the curves representing the proposed STA–AED model are approximately the smallest in all zones. Although the encoder–decoder network with multi-head attention shows the best prediction result in zone 11, it is just slightly better than the proposed STA–AED model. To summarize, the proposed STA–AED model achieved good prediction results in all 15 zones, which demonstrates that the proposed STA–AED model can provide better load forecasting ability and can improve the stability of short-term load forecasting.