Optimizing Short-Term Electrical Demand Forecasting with Deep Learning and External Influences †
Abstract
1. Introduction
- Enhanced forecasting accuracy through the integration of multiple external factors.
- A structured deep learning framework for energy demand prediction.
- Empirical validation demonstrating the effectiveness of combining CNN and long short term memory (LSTM) models.
2. Modeling Methodology
2.1. Pipeline for Building Electricity Forecasting Models
2.2. Key Aspects of Forecasting Methodology
- Number of Input Nodes: It should correspond to the relevant variables influencing demand [14].
- Hidden Layers: Networks with hidden layers are more capable of learning complex relationships; typically, one or two layers suffice [16].
- Neurons in Hidden Layers: An excessive number can lead to overfitting, with guidelines for determining the appropriate number [17].
- Hyperparameters: Model training is influenced by hyperparameters like activation functions, learning rates, and batch sizes. A GridSearchCV method is used to optimize these [18].
- Training Algorithm: The backpropagation method minimizes error by adjusting network weights to improve predictions [14].
3. Implementation and Results
3.1. Dataset
3.2. Case Study
- o
- Among the variable selection algorithms tested here, the most notable ones were correlation-based feature subset evaluation (CFS subset evaluation), classifier attribute evaluation, and relief. These algorithms selected the five or six variables that were most strongly correlated with the target variable, yielding predictions with error rates similar to or lower than those obtained when all variables in the database were considered (as indicated in the last line of Table 3). This underscores the relevance of the selected variables.
- o
- Regarding the performance of models with different prediction algorithms, LSTM stands out as having the highest computational cost. Its processing time was 10 times higher than for the DNN, five times more than for the CNN, and 2.5 times more than for the combined CNN + LSTM model. However, the models based on the LSTM and a combination of CNN + LSTM achieved the highest accuracy.
- o
- From comparing the performance of shallow and deep models, it was observed that the latter incurred a higher computational cost. Nevertheless, in most scenarios, they demonstrated superior accuracy compared to shallow models.
- o
- The distinguishing factor between the tables is the prediction horizon. It is evident that increasing the prediction horizon leads to a reduction in accuracy.
- o
- Processing involving only the target variable (Sl), with or without external factors, gave the biggest errors. This implies that deep learning algorithms face limitations in terms of their generalization capacity when operating with a very limited number of input variables, resulting in elevated error rates.
- o
- From a comparison of the predictions based on only the target variable (Sl), the selected variables (Fs), and all variables in the dataset (Av), it is notable that errors in the Fs and Av columns are very similar. This suggests that despite the reduction in the number of input variables from 23 to 6 or 7 in the variable selection step, this did not lead to an increase in the prediction error.
4. Conclusions and Future Research
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Merce, R.A.; Grover-Silva, E.; Le Conte, J. Load and Demand Side Flexibility Forecasting. In Proceedings of the International Conference on Smart Grids, Green Communications and IT Energy-aware Technologies, Lisbon, Portugal, 27 September–1 October 2020; pp. 1–6. [Google Scholar]
- Khan, K.A.; Quamar, M.M.; Al-Qahtani, F.H.; Asif, M.; Alqahtani, M.; Khalid, M. Smart Grid Infrastructure and Renewable Energy Deployment: A Conceptual Review of Saudi Arabia. Energy Strateg. Rev. 2023, 50, 101247. [Google Scholar] [CrossRef]
- Mystakidis, A.; Koukaras, P.; Tsalikidis, N.; Ioannidis, D.; Tjortjis, C. Energy Forecasting: A Comprehensive Review of Techniques and Technologies. Energies 2024, 17, 1662. [Google Scholar] [CrossRef]
- Wang, X.; Yao, Z.; Papaefthymiou, M. A Real-Time Electrical Load Forecasting and Unsupervised Anomaly Detection Framework. Appl. Energy 2023, 330, 120279. [Google Scholar] [CrossRef]
- Khatoon, S.; Ibraheem; Singh, A.K. Priti Effects of Various Factors on Electric Load Forecasting: An Overview. In Proceedings of the 2014 6th IEEE Power India International Conference (PIICON), Delhi, India, 5–7 December 2014; pp. 1–5. [Google Scholar]
- Peñaloza, A.K.A.; Balbinot, A.; Leborgne, R.C. Review of Deep Learning Application for Short-Term Household Load Forecasting. In Proceedings of the 2020 IEEE PES Transmission & Distribution Conference and Exhibition-Latin America (T&D LA), Montevideo, Uruguay, 28 September–2 October 2020; pp. 1–6. [Google Scholar]
- Ren, D.; Liu, W.; Li, J.; Huang, R. Research on Aggregate Load Prediction Methods. In Proceedings of the 2023 IEEE 7th Conference on Energy Internet and Energy System Integration (EI2), Hangzhou, China, 15–18 December 2023; pp. 3564–3567. [Google Scholar]
- Aybar-Mejía, M.; Villanueva, J.; Mariano-Hernández, D.; Santos, F.; Molina-García, A. A Review of Low-Voltage Renewable Microgrids: Generation Forecasting and Demand-Side Management Strategies. Electronics 2021, 10, 2093. [Google Scholar] [CrossRef]
- Bunn, D.W. Short-Term Forecasting: A Review of Procedures in the Electricity Supply Industry. J. Oper. Res. Soc. 1982, 33, 533–545. [Google Scholar] [CrossRef]
- Min, H.; Hong, S.; Song, J.; Son, B.; Noh, B.; Moon, J. SolarFlux Predictor: A Novel Deep Learning Approach for Photovoltaic Power Forecasting in South Korea. Electronics 2024, 13, 2071. [Google Scholar] [CrossRef]
- Manzolini, G.; Fusco, A.; Gioffrè, D.; Matrone, S.; Ramaschi, R.; Saleptsis, M.; Simonetti, R.; Sobic, F.; Wood, M.J.; Ogliari, E.; et al. Impact of PV and EV Forecasting in the Operation of a Microgrid. Forecasting 2024, 6, 591–615. [Google Scholar] [CrossRef]
- Sengupta, S.; Basak, S.; Saikia, P.; Paul, S.; Tsalavoutis, V.; Atiah, F.; Ravi, V.; Peters, A. A Review of Deep Learning with Special Emphasis on Architectures, Applications and Recent Trends. Knowl.-Based Syst. 2020, 194, 105596. [Google Scholar] [CrossRef]
- Curry, B.; Morgan, P.H. Model Selection in Neural Networks: Some Difficulties. Eur. J. Oper. Res. 2006, 170, 567–577. [Google Scholar] [CrossRef]
- Zhang, G.; Eddy Patuwo, B.; Hu, M.Y. Forecasting with Artificial Neural Networks: The State of the Art. Int. J. Forecast. 1998, 14, 35–62. [Google Scholar] [CrossRef]
- Petropoulos, F.; Apiletti, D.; Assimakopoulos, V.; Babai, M.Z.; Barrow, D.K.; Ben Taieb, S.; Bergmeir, C.; Bessa, R.J.; Bijak, J.; Boylan, J.E.; et al. Forecasting: Theory and Practice. Int. J. Forecast. 2022, 38, 705–871. [Google Scholar] [CrossRef]
- Lippmann, R. An Introduction to Computing with Neural Nets. IEEE ASSP Mag. 1987, 4, 4–22. [Google Scholar] [CrossRef]
- Sheela, K.G.; Deepa, S.N. Review on Methods to Fix Number of Hidden Neurons in Neural Networks. Math. Probl. Eng. 2013, 2013, 425740. [Google Scholar] [CrossRef]
- Chollet, F. Keras: The Python Deep Learning API. 2015. Available online: https://keras.io/ (accessed on 23 March 2021).
- Ribeiro, A.M.N.C.; do Carmo, P.R.X.; Endo, P.T.; Rosati, P.; Lynn, T. Short- and Very Short-Term Firm-Level Load Forecasting for Warehouses: A Comparison of Machine Learning and Deep Learning Models. Energies 2022, 15, 750. [Google Scholar] [CrossRef]
- Gnanambal, S.; Thangaraj, M.; Meenatchi, V.; Gayathri, V. Classification Algorithms with Attribute Selection: An Evaluation Study Using WEKA. Int. J. Adv. Netw. Appl 2018, 9, 3640–3644. [Google Scholar]
- Cogollo, M.R.; Velasquez, J.D. Methodological Advances in Artificial Neural Networks for Time Series Forecasting. IEEE Lat. Am. Trans. 2014, 12, 764–771. [Google Scholar] [CrossRef]
- Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation; Defense Technical Information Center: Fort Belvoir, VA, USA, 1985. [Google Scholar]
- Amaral, L.S.; de Araújo, G.M. An Expanded Study of the Application of Deep Learning Models in Energy Consumption. In Data and Information in Online Environments: Third EAI International Conference, DIONE 2022, Virtual, 28–29 July 2022; Proceedings; Springer Nature: Berlin/Heidelberg, Germany, 2022; Volume 452, p. 150. [Google Scholar]
- Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H. The WEKA Data Mining Software: An Update. SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar] [CrossRef]
Technique | Source | Selected Variables | Magnitude | DNN | CNN | LSTM | CNN Plus LSTM | ||||
---|---|---|---|---|---|---|---|---|---|---|---|
Shallow | Deep | Shallow | Deep | Shallow | Deep | Shallow | Deep | ||||
CFS subset evaluation | WEKA | RT demand, DACC, DA MLC, RT MLC, MIN_5MIN_RSP, MAX_5MIN_RSP | MAPE | 0.22 | 0.18 | 1.80 | 0.74 | 0.37 | 0.24 | 0.21 | 0.18 |
t(s) | 152 | 185 | 276 | 340 | 887 | 1798 | 739 | 848 | |||
Classifier attribute evaluation | MAX_5MIN_RSP, DA EC, DA CC, RT demand, DA LMP | MAPE | 0.27 | 0.23 | 1.60 | 0.48 | 0.21 | 0.16 | 0.26 | 0.24 | |
t(s) | 153 | 187 | 342 | 384 | 994 | 1813 | 761 | 790 | |||
Principal components | RT LMP, RT EC, DA LMP, DA EC, RT MLC | MAPE | 7.80 | 6.84 | 9.67 | 6.50 | 10.95 | 9.79 | 8.06 | 7.57 | |
t(s) | 87 | 130 | 150 | 219 | 594 | 912 | 996 | 386 | |||
Relief | RT demand, DA demand, DA EC, DA LMC, reg. service price | MAPE | 0.52 | 0.33 | 3.82 | 0.22 | 0.23 | 0.16 | 0.16 | 0.17 | |
t(s) | 112.8 | 123 | 274 | 339 | 906 | 1792 | 692 | 823 | |||
Mutual information | Python | DA MLC, DA LMP, MIN_5MIN_RSP, DA EC, dew point | MAPE | 5.90 | 5.76 | 8.15 | 5.03 | 10.21 | 7.82 | 7.23 | 6.11 |
t(s) | 155 | 186 | 270 | 339 | 581 | 1827 | 725 | 792 | |||
- | - | All | MAPE | 0.86 | 0.38 | 2.27 | 1.05 | 0.29 | 0.20 | 0.25 | 0.20 |
t(s) | 118 | 123 | 158 | 153 | 591 | 920 | 695 | 879 |
Technique | Model | Sl (1) | Sl + W (2) | Sl + S + C + Id (3) | Δ (%) (4) | Fs (5) | Fs + S + C + Id (6) | Δ (%) (7) | Av (8) | Av + S + C + Id (9) | Δ (%) (10) |
---|---|---|---|---|---|---|---|---|---|---|---|
DNN | Sh | 11.30 | 6.76 | 5.38 | 52.40 | 0.29 | 0.19 | 34.50 | 0.19 | 0.18 | 5.30 |
D | 10.85 | 5.98 | 8.57 | 44.90 | 0.37 | 0.17 | 54.10 | 0.22 | 0.17 | 22.70 | |
CNN | Sh | 12.13 | 8.46 | 8.00 | 34 | 2.85 | 2.23 | 21.8 | 3.40 | 2.26 | 33.50 |
D | 11.04 | 4.95 | 4.29 | 61.10 | 0.18 | 0.18 | 0.00 | 0.26 | 0.16 | 38.50 | |
LSTM | Sh | 12.37 | 8.31 | 7.86 | 36.5 | 0.65 | 0.25 | 61.50 | 0.47 | 0.17 | 63.80 |
D | 11.74 | 7.59 | 9.00 | 23.30 | 14.10 | 7.07 | 49.80 | 14.96 | 10.45 | 30.10 | |
CNN plus LSTM | Sh | 10.38 | 6.56 | 5.67 | 45.40 | 0.31 | 0.21 | 32.2 | 0.42 | 0.19 | 54.80 |
D | 11.11 | 5.09 | 7.60 | 31.60 | 0.25 | 0.15 | 40.00 | 0.25 | 0.24 | 4.00 |
Technique | Model | Sl (1) | Sl + W (2) | Sl + S + C + Id (3) | Δ (%) (4) | Fs (5) | Fs + S + C + Id (6) | Δ (%) (7) | Av (8) | Av + S + C + Id (9) | Δ (%) (10) |
---|---|---|---|---|---|---|---|---|---|---|---|
DNN | Sh | 33.20 | 23.44 | 27.98 | 15.70 | 36.58 | 29.77 | 18.60 | 29.80 | 29.76 | 0.13 |
D | 48.10 | 27.63 | 37.01 | 23.10 | 45.03 | 32.4 | 28.10 | 44.0 | 40.58 | 7.77 | |
CNN | Sh | 12.01 | 15.37 | 4.92 | 59.00 | 4.91 | 4.83 | 1.60 | 4.92 | 4.68 | 4.90 |
D | 10.60 | 9.19 | 3.77 | 64.40 | 4.79 | 3.96 | 17.30 | 4.39 | 3.63 | 17.70 | |
LSTM | Sh | 13.00 | 14.36 | 9.82 | 24.50 | 4.39 | 4.30 | 2.00 | 4.29 | 3.66 | 14.70 |
D | 15.58 | 15.65 | 15.27 | 2.00 | 15.8 | 11.34 | 4.46 | 15.4 | 15.15 | 1.62 | |
CNN plus LSTM | Sh | 15.00 | 8.04 | 4.58 | 69.50 | 4.42 | 3.93 | 11.00 | 4.12 | 4.09 | 0.70 |
D | 11.24 | 8.65 | 10.70 | 4.80 | 15.94 | 10.82 | 32.10 | 15.6 | 15.36 | 1.50 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Amaral, L.S.; Araújo, G.M.d.; Moraes, R. Optimizing Short-Term Electrical Demand Forecasting with Deep Learning and External Influences. Eng. Proc. 2025, 101, 16. https://doi.org/10.3390/engproc2025101016
Amaral LS, Araújo GMd, Moraes R. Optimizing Short-Term Electrical Demand Forecasting with Deep Learning and External Influences. Engineering Proceedings. 2025; 101(1):16. https://doi.org/10.3390/engproc2025101016
Chicago/Turabian StyleAmaral, Leonardo Santos, Gustavo Medeiros de Araújo, and Ricardo Moraes. 2025. "Optimizing Short-Term Electrical Demand Forecasting with Deep Learning and External Influences" Engineering Proceedings 101, no. 1: 16. https://doi.org/10.3390/engproc2025101016
APA StyleAmaral, L. S., Araújo, G. M. d., & Moraes, R. (2025). Optimizing Short-Term Electrical Demand Forecasting with Deep Learning and External Influences. Engineering Proceedings, 101(1), 16. https://doi.org/10.3390/engproc2025101016