Forecasting of Short-Term Load Using the MFF-SAM-GCN Model
Abstract
:1. Introduction
- (1)
- An MFF and GCN based SLF model is suggested, in which MFF is used to extract spatiotemporal correlation characteristics of load data and GCN is to elicit external non-numerical information.
- (2)
- A novel MFF structure is developed, which employs the 1D-CNN based on the SAM to extract spatially relevant load data features. Moreover, the use of Bi-LSTM network can help to extract temporally relevant load data features.
- (3)
- A BO algorithm is used to find the optimal model parameters, and the proposed prediction model is experimentally compared with currently popular prediction models to investigate the role of each module and the effect of hyperparameter settings and busbars on prediction performance via ablation experiments.
2. Related Work
3. Method
3.1. Overall Network
3.2. Multi-Feature Fusion Framework
3.2.1. Bi-Directional Long Short-Term Memory Network
- (1)
- Forget gate: The forget gate utilizes the sigmoid function to pick some of the information from the output at the moment of and the input at the moment of t. The output of the forgotten gate is given by:
- (2)
- Input gate: The and functions are used to choose certain information from the output at the moment of and the input at the moment of t, respectively. Two outputs of input gate are:
- (3)
- Information update: The output of the forget gate at the moment of t is multiplied by the state vector at the moment of and added to the product of the two output values of the input gate. The updated information is represented as:
- (4)
- Output gate: on the one hand, it utilizes the function to pick some information from the output at the moment of and the input at the moment of t. Use the function, on the other hand, to filter the updated state vector selection and then multiply the two to obtain the output of the present instant that is:In the above description, the calculated weight matrix , and bias vector , as described previously, are the parameters that the LSTM network must learn during the training phase.
3.2.2. One-Dimensional Convolutional Neural Network
3.2.3. Self-Attention Mechanism
3.3. Graph Convolutional Network
3.4. Bayesian Optimization
4. Experiments and Evaluation
4.1. Training Planning
4.1.1. Information Sources
4.1.2. Pre-Processing of Data
4.1.3. Establishment of a Training Environment
4.1.4. Indicators of Evaluation
4.2. Test Results Comparison
4.3. Experiments with Ablation
4.4. Extensive Research
4.4.1. The Effect of Hyperparameter Values
4.4.2. The Impact of the Bus
5. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
AM | Attention Mechanism |
Bi-LSTM | Bi-directional Long Short-Term Memory |
BO | Bayesian Optimization |
CNN | Convolutional Neural Network |
1D-CNN | One-Dimensional Convolutional Neural Network |
DT | Decision Tree |
DSR | Demand Side Response |
EEMD | Ensemble Empirical Mode Decomposition |
GCN | Graph Convolutional Network |
GRU | Gated Recurrent Unit |
KFCM | Kernel Fuzzy C-Means |
KNN | K-Nearest Neighbor |
LLF | Long-term Load Forecasting |
LR | Linear Regression |
LSTM | Long Short Term Memory |
MBE | Mean Bias Error |
MFF | Multi-Feature Fusion |
MLF | Medium-term Load Forecasting |
MSE | Mean Square Error |
PSO | Particle Swarm Optimization |
RMSE | Root Mean Square Error |
RNN | Recurrent Neural Network |
SAM | Self-Attention Mechanism |
SLF | Short-term Load Forecasting |
SMAPE | Symmetric Mean Absolute Percentage Error |
TCN | Time Convolution Network |
References
- Fang, X.; Misra, S.; Xue, G.; Yang, D. Smart Grid-The New and Improved Power Grid: A Survey. Commun. Surv. Tutor. IEEE 2012, 14, 950–956. [Google Scholar] [CrossRef]
- Ekonomou, L.; Christodoulo, C.A.; Mladenov, V. A short-term load forecasting method using artificial neural networks and wavelet analysis. Int. J. Power Syst. 2016, 1, 64–68. [Google Scholar]
- Javed, F.; Arshad, N.; Wallin, F.; assileva, I.V.; Dahlquist, E. Forecasting for demand response in smart grids: An analysis on use of anthropologic and structural data and short term multiple loads forecasting. Appl. Energy 2012, 96, 150–160. [Google Scholar] [CrossRef]
- Xia, C.; Wang, J.; Mcmenemy, K. Short, medium and long term load forecasting model and virtual load forecaster based on radial basis function neural networks. Int. J. Electr. Power Energy Syst. 2010, 32, 743–750. [Google Scholar] [CrossRef] [Green Version]
- Wei, L.; Zhang, Z.-G. Based on Time Sequence of ARIMA Model in the Application of Short-Term Electricity Load Forecasting. In Proceedings of the International Conference on Research Challenges in Computer Science IEEE, Shanghai, China, 28–29 December 2009. [Google Scholar]
- Song, K.-B.; Baek, Y.-S.; Hong, D.H.; Jang, G. Short-term load forecasting for the holidays using fuzzy linear regression method. IEEE Trans. Power Syst. 2005, 20, 96–101. [Google Scholar] [CrossRef]
- Al-Hamadi, H.M.; Soliman, S.A. Short-term electric load forecasting based on Kalman filtering algorithm with moving window weather and load model. Electr. Power Syst. Res. 2004, 68, 47–59. [Google Scholar] [CrossRef]
- Jain, A.; Satish, B. Clustering based Short Term Load Forecasting using Support Vector Machines. In Proceedings of the 2009 IEEE Bucharest Powertech, Bucharest, Romania, 28 June–2 July 2009. [Google Scholar]
- Hambali, A.O.J.; Akinyemi, M.; Jyusuf, N. Electric power load forecast using decision tree algorithms. Comput. Inf. Syst. Dev. Inform. Allied Res. J. 2016, 7, 29–42. [Google Scholar]
- Dudek, G. Short-term load forecasting using random forests. In Intelligent Systems; Springer: Cham, Switzerland, 2014; pp. 821–828. [Google Scholar]
- Ceperic, E.; Ceperic, V.; Baric, A. A Strategy for Short-Term Load Forecasting by Support Vector Regres-sion Machines. IEEE Trans. Power Syst. 2013, 28, 4356–4364. [Google Scholar] [CrossRef]
- Çevik, H.H.; Çunkaş, M. Short-term load forecasting using fuzzy logic and ANFIS. Neural Comput. Appl. 2015, 26, 1355–1367. [Google Scholar] [CrossRef]
- Raza, M.Q.; Khosravi, A. A review on artificial intelligence based load demand forecasting techniques for smart grid and buildings. Renew. Sustain. Energy Rev. 2015, 50, 1352–1372. [Google Scholar] [CrossRef]
- Shi, H.; Xu, M.; Li, R. Deep Learning for Household Load Forecasting—A Novel Pooling Deep RNN. IEEE Trans. Smart Grid 2017, 9, 5271–5280. [Google Scholar] [CrossRef]
- Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans. Smart Grid 2017, 10, 841–851. [Google Scholar] [CrossRef]
- Dong, X.; Qian, L.; Huang, L. Short-term load forecasting in smart grid: A combined CNN and K-means clustering ap-proach. In Proceedings of the IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju, Korea, 13–16 February 2017; pp. 119–125. [Google Scholar]
- Lang, C.; Steinborn, F.; Steffens, O.; Lang, E.W. Electricity Load Forecasting—An Evaluation of Simple 1D-CNN Network Structures. arXiv 2019, arXiv:1911.11536. [Google Scholar]
- Zhou, F.; Hu, P.; Yang, S.; Wen, C. A multimodal feature fusion-based deep learning method for online fault diagnosis of ro-tating machinery. Sensors 2018, 18, 3521. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zhang, Y.; Qi, P.; Manning, C.D. Graph convolution over pruned dependency trees improves relation extraction. arXiv 2018, arXiv:1809.10185. [Google Scholar]
- Chen, B.J.; Chang, M.W.; Lin, C.J. Load Forecasting Using Support Vector Machines: A Study on EUNITE Compe-tition 2001. IEEE Trans. Power Syst. 2004, 19, 1821–1830. [Google Scholar] [CrossRef] [Green Version]
- Rafi, S.H.; Deeba, S.R.; Hossain, E. A Short-Term Load Forecasting Method Using Integrated CNN and LSTM Network. IEEE Access 2021, 99, 32436–32448. [Google Scholar] [CrossRef]
- Goh, H.H.; He, B.; Liu, H.; Zhang, D.; Dai, W.; Kurniawan, T.A.; Goh, K.C. Multi-Convolution Feature Extraction and Recurrent Neural Network Dependent Model for Short-Term Load Forecasting. IEEE Access 2021, 9, 118528–118540. [Google Scholar] [CrossRef]
- Farsi, B.; Amayri, M.; Bouguila, N.; Eicker, U. On short-term load forecasting using machine learning techniques and a novel parallel deep LSTM-CNN approach. IEEE Access 2021, 9, 31191–31212. [Google Scholar] [CrossRef]
- Wang, Y.; Chen, J.; Chen, X.; Zeng, X.; Kong, Y.; Sun, S.; Liu, Y. Short-Term Load Forecasting for Industrial Customers Based on TCN-LightGBM. IEEE Trans. Power Syst. 2021, 36, 1984–1997. [Google Scholar] [CrossRef]
- Fahim, A.; Tan, Q.; Mazzi, M.; Sahabuddin, M.; Naz, B.; Ullah, B.S. Hybrid LSTM Self-Attention Mechanism Model for Forecasting the Reform of Scientific Research in Morocco. Comput. Intell. Neurosci. 2021, 2021, 6689204. [Google Scholar] [CrossRef] [PubMed]
- Miao, K.; Hua, Q.; Shi, H. Short-Term load forecasting based on CNN-BiLSTM with Bayesian optimization and attention mechanism. In Proceedings of the International Conference on Parallel and Distributed Computing: Applications and Technologies, Guangzhou, China, 17–19 December 2020; pp. 116–128. [Google Scholar]
- Azam, M.F.; Younis, S. Multi-Horizon Electricity Load and Price Forecasting using an Interpretable Multi-Head Self-Attention and EEMD-Based Framework. IEEE Access 2021, 9, 85918–85932. [Google Scholar] [CrossRef]
- Zang, H.; Xu, R.; Cheng, L.; Ding, T.; Liu, L.; Wei, Z.; Sun, G. Residential load forecasting based on LSTM fusing self-attention mechanism with pooling. Energy 2021, 229, 120682. [Google Scholar] [CrossRef]
- Shang, C.; Gao, J.; Liu, H.; Liu, F. Short-Term Load Forecasting Based on PSO-KFCM Daily Load Curve Clustering and CNN-LSTM Model. IEEE Access 2021, 9, 50344–50357. [Google Scholar] [CrossRef]
- Shang, C.; Gao, J.; Liu, H.; Liu, F. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Trans. Intell. Transp. Syst. 2019, 99, 1–11. [Google Scholar]
- Wang, R.; Huang, C.; Wang, X. Global relation reasoning graph convolutional networks for human pose estimation. IEEE Access 2020, 8, 38472–38480. [Google Scholar] [CrossRef]
- Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
- Wang, S.; Wang, X.; Wang, S.; Wang, D. Bi-directional long short-term memory method based on attention mechanism and rolling update for short-term load forecasting. Int. J. Electr. Power Energy Syst. 2019, 109, 470–479. [Google Scholar] [CrossRef]
- Shaw, P.; Uszkoreit, J.; Vaswani, A. Self-attention with relative position representations. arXiv 2018, arXiv:1803.02155. [Google Scholar]
- Chen, Z.M.; Wei, X.S.; Wang, P.; Guo, Y. Multi-label image recognition with graph convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Snoek, J.; Larochelle, H.; Adams, R.P. Practical bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. 2012, 25, 2960–2968. [Google Scholar]
- Liang, Y.; Niu, D.; Hong, W.C. Practical bayesian optimization of machine learning algorithms. Short Term Load Forecast. Based Feature Extr. Improv. Gen. Regres. Neural Netw. Model 2019, 166, 653–663. [Google Scholar]
Method | RMSE | SMAPE | MBE | Runtimes (Company: s) | |
---|---|---|---|---|---|
LR | 0.037 | 12.561% | 0.036 | 0.952 | 189.431 |
KNN | 0.042 | 15.243% | 0.040 | 0.947 | 159.563 |
DT | 0.045 | 16.655% | 0.042 | 0.939 | 209.767 |
BiLSTM | 0.034 | 11.046% | 0.033 | 0.964 | 317.269 |
CNN | 0.035 | 11.784% | 0.033 | 0.957 | 306.436 |
CNN-LSTM(S2S) | 0.032 | 10.542% | 0.031 | 0.974 | 324.584 |
GCN-MTF | 0.028 | 9.453% | 0.025 | 0.989 | 337.063 |
Basic | Attention | GCN | RMSE | SMAPE | MBE | Runtimes (Company: s) | |
---|---|---|---|---|---|---|---|
- | - | 0.037 | 11.485% | 0.035 | 0.965 | 316.453 | |
CNN-LSTM | ✓ | - | 0.035 | 11.236% | 0.032 | 0.972 | 322.536 |
(MFF) | - | ✓ | 0.032 | 10.256% | 0.031 | 0.979 | 326.461 |
✓ | ✓ | 0.028 | 9.453% | 0.025 | 0.989 | 337.063 |
Hyperparameter | Search Scope |
---|---|
drop_out_rate | (0.1, 0.2, 0.3) |
learning_rate | (0.1, 0.2, 0.3) |
Bi-LSTM_hidden size | (32–128) |
CNN1_number of filter | (16, 32, 64) |
CNN2_number of filter | (16, 32, 64) |
CNN1_kernel size | (2, 3, 4) |
CNN1_kernel size | (2, 3, 4) |
epochs | (50, 100, 200) |
Batch size | (15,000, 20,000, 25,000, 30,000) |
Method | RMSE | SMAPE | MBE | Runtimes (Company: s) | |
---|---|---|---|---|---|
method 1 | 0.028 | 9.286% | 0.025 | 0.990 | 353.985 |
method 2 | 0.028 | 9.453% | 0.025 | 0.989 | 337.063 |
method 3 | 0.026 | 9.232% | 0.023 | 0.993 | 2980.231 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zou, Y.; Feng, W.; Zhang, J.; Li, J. Forecasting of Short-Term Load Using the MFF-SAM-GCN Model. Energies 2022, 15, 3140. https://doi.org/10.3390/en15093140
Zou Y, Feng W, Zhang J, Li J. Forecasting of Short-Term Load Using the MFF-SAM-GCN Model. Energies. 2022; 15(9):3140. https://doi.org/10.3390/en15093140
Chicago/Turabian StyleZou, Yongqi, Wenjiang Feng, Juntao Zhang, and Jingfu Li. 2022. "Forecasting of Short-Term Load Using the MFF-SAM-GCN Model" Energies 15, no. 9: 3140. https://doi.org/10.3390/en15093140
APA StyleZou, Y., Feng, W., Zhang, J., & Li, J. (2022). Forecasting of Short-Term Load Using the MFF-SAM-GCN Model. Energies, 15(9), 3140. https://doi.org/10.3390/en15093140