You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

21 August 2023

Deep Learning Model Performance and Optimal Model Study for Hourly Fine Power Consumption Prediction

,
,
,
and
1
Department of ICT Convergence System Engineering, Chonnam National University, 77, Yongbong-ro, Buk-gu, Gwangju 500757, Republic of Korea
2
Graduate School of Data Science, Chonnam National University, Gwangju 61186, Republic of Korea
*
Authors to whom correspondence should be addressed.
This article belongs to the Section Artificial Intelligence

Abstract

Electricity consumption has been increasing steadily owing to technological developments since the Industrial Revolution. Technologies that can predict power usage and management for improved efficiency are thus emerging. Detailed energy management requires precise power consumption forecasting. Deep learning technologies have been widely used recently to achieve high performance. Many deep learning technologies are focused on accuracy, but they do not involve detailed time-based usage prediction research. In addition, detailed power prediction models should consider computing power, such as that of end Internet of Things devices and end home AMIs. In this work, we conducted experiments to predict hourly demands for the temporal neural network (TCN) and transformer models, as well as artificial neural network, long short-term memory (LSTM), and gated recurrent unit models. The study covered detailed time intervals from 1 to 24 h with 1 h increments. The experimental results were analyzed, and the optimal models for different time intervals and datasets were derived. The LSTM model showed superior performance for datasets with characteristics similar to those of schools, while the TCN model performed better for average or industrial power consumption datasets.

1. Introduction

Global energy consumption, including electricity usage, has been increasing consistently since the Industrial Revolution. This increase is attributed to global population growth, leading to increased demand in households, businesses, and industries. In addition, the number of electronic devices that depend on electricity has increased because of the increase in infrastructure and transportation facilities due to expansion of urban areas and development of technologies, such as large-scale data centers. Increased use of air conditioning and heating systems due to extreme climate changes, such as heat waves in some regions, has also contributed to increasing power consumption [1,2,3]. Such increase in power consumption causes many problems; sudden spikes in power demand can strain the energy infrastructure, resulting in blackouts or shortages. Electricity generation relies heavily on fossil fuels, such as coal, natural gas, and oil. This increase in power generation implies increasing greenhouse gas emissions that cause air pollution and environmental problems. Effective energy management is therefore essential to address these issues. Accurate power consumption forecasting is becoming more important for energy management; however, existing statistical methods have difficulties in predicting power consumption accurately owing to the irregular and complex patterns inherent in power data that are affected by various spatial and temporal factors [4]. Recently, artificial intelligence technologies have been used to predict irregular patterns in various areas, such as power consumption and power generation. In particular, deep learning models based on neural networks play an important role because they make it easy to understand trends or characteristics of data, so research on applying deep learning technologies to the electric power field has been progressing steadily [5,6,7,8,9]. Accurately predicting the demand and supply of electrical energy or reducing its usage can reduce energy wastage. Artificial intelligence can be used to identify and predict current usage to prevent unnecessary power use or production increase by adjusting the supply based on usage prediction. Detailed power consumption predictions are hence required for efficient power management [10,11]. In many previous studies, research on deep learning models for power usage prediction was conducted. However, many studies have not made detailed predictions of power consumption by time, such as predicting power consumption after 1 h or additionally predicting power consumption after 5 h or 10 h [12,13,14,15]. Fine-grained power usage forecasts offer detailed insights into power consumption patterns throughout the day, enabling better understanding of how the demand fluctuates at different times. By predicting the hourly power consumption, it is possible to manage power generation efficiently by identifying the peak power demand periods and responding accordingly. These advantages provide insights into many areas, including load planning, demand forecasting, renewable energy utilization, energy trading, and infrastructure planning. In this study, we conducted experiments to predict power consumption based on time using a deep learning model. This model used not only the basic artificial neural network (ANN) model, but also the recurrent neural network (RNN)-type long short-term memory (LSTM), gated recurrent unit (GRU), and temporal neural network (TCN) models, which use the convolutional neural network (CNN) as well as the recently used transformer model. In addition, an optimal model was derived by comparing the power consumption prediction performances over time.
The aim of this work is threefold:
-
We derived a long-term and short-term power prediction deep learning model that can predict power consumption from 1 h to 24 h using power data.
-
We derived a benchmark using a dataset acquired from Chonnam National University in Korea and an actual dataset acquired from a factory in Gyeonggi-do, Korea.
-
We derived the optimal model when targeting maximum performance and average performance according to data patterns in schools and factories, which are non-residential data.

3. Methodology

In this section, we define a time-series prediction problem for power consumption prediction. In addition, the dataset and performance indicators used for the experiments are summarized.

3.1. Dataset

Two datasets were used for the experiments in this study. The first dataset was actually collected from Chonnam National University (CNU) continuously at 1 h intervals for approximately 1 year and 3 months (1 January 2021 to 14 January 2022). The dataset contained power consumption data at 90 locations within CNU; there were 11,232 data for each location, and among these, the total power consumed at CNU building No. 7 was used for power consumption prediction. The second dataset was a high-voltage dataset continuously collected from factories at 1 h intervals for 2 years (1 August 2020 to 31 July 2022). Table 1 below shows the status of each dataset, which was subjected to min-max normalization for learning. The min-max regularization was performed as expressed in Equation (5). The dataset was constructed using the slide window method so that 168 data could be input during model learning.
x s c a l e d = x x m i n x m a x x m i n
Table 1. Dataset list.

3.2. Performance Metrics

To assess the experimental performance in this work, the mean-squared error (MSE) and mean absolute error (MAE) performance indexes were used. The MSE is a common metric used to measure the mean-squared differences between the predicted and actual values in a dataset. Equation (6) represents the expression of the MSE performance index. The MAE is another commonly used metric to evaluate the performance of a predictive model and can be expressed as in Equation (7). Unlike the MSE, instead of squaring the differences between the predicted and actual values, the MAE uses the absolute value of the difference, which makes it more robust against outliers because it does not amplify the effects of the errors like the MSE. The choice of MSE and MAE as the performance indicators depends on the specific problem, nature of the data, and goals of the model. Therefore, it is important to select a model based on appropriate performance indicators according to the goal.
M e a n   S q u a r e   E r r o r ( M S E ) = 1 n i = 1 n Y i ^ Y i 2
M e a n   A b s o l u t e   E r r o r ( M A E ) = 1 n i = 1 n Y i ^ Y i

4. Experimental Results and Comparison

4.1. Deep Learning Model Setup

The deep learning models used in this study were the ANN, LSTM, GRU, TCN, and transformer models. The ANN model consists of two hidden layers, where the first layer has 32 nodes and second layer has 8 nodes. The LSTM model consists of two layers, where the first tier has 200 nodes and second tier has 150 nodes. The GRU model consists of two layers, and each layer has 103 nodes with the final layer consisting of linear unit activation functions. The TCN model is composed of one layer; the window size of the 1D convolution is 3, with the number of filters being 64 and kernel size being 3. The activation function of all models used the ReLU function. The transformer model has the advantage of good performance but has the disadvantage of being heavy owing to the large number of parameters. Table 2 is a table showing the network model configuration. In particular, the Transformer model used a transformer model composed of three layers and eight layers, as shown in Table 2, to compensate for the problem of transformers.
Table 2. Model construction of network.

4.2. Training Methods

In this work, i was set to 168 and n was set from 0 to 23 to conduct the experiments. Based on time-series prediction, the problem can be expressed as Equation (8). The prediction problem was a problem of continuously predicting data from 1 h to 24 h. For example, in the problem of predicting 10 h, 168 time data were input, and the predicted value from 1 h to 10 h was later compared with the correct value. The implementation of the proposed method was achieved using the Keras library with TensorFlow backend. All models were trained and tested on four NVIDIA Quadro RTX A5000 24 GB GPUs.
y ^ 0 , , y ^ n = f x 0 , , x 167 ,   n = 0 ,   1 ,   2 , ,   23

4.3. Experimental Results and Analysis

The experiments proceeded with predictions according to values set from 1 to 24 h for the data entered for 168 h. The results with the best performances are indicated in bold red. The lower the MSE and MAE performance indexes, the better the results. Table 3 below shows the MSE performance of the CNU dataset. Table 4 below shows the MAE performance of the CNU dataset. Similarly, Table 5 and Table 6 below show the MSE and MAE performances of the high-voltage dataset.
Table 3. Deep learning model MSE performance of the CNU dataset.
Table 4. Deep learning model MAE performance of the CNU dataset.
Table 5. Deep learning model MSE performance of the high-voltage dataset.
Table 6. Deep learning model MAE performance of the high-voltage dataset.

4.4. Summary of Experimental Results

Section 4.3 showed the experimental results of the deep learning models for the CNU and high-voltage datasets. In this section, the experimental results are summarized, and an optimal model for predicting power consumption based on time is presented. Table 7 below shows the experimental results of the CNU dataset; as shown, when the MSE performance indicators were summarized, the TCN model was the optimal model in 16 time zones. In the problem of predicting 1 h later, the LSTM was a joint optimal model. In addition, in the 5 h prediction problem, the ANN model was a co-optimal model; the LSTM model in three time zones and ANN model in seven time zones were the overall optimal models. When the MAE performance indicators were summarized, the TCN model was the optimal model in sixteen time zones, the ANN model was optimal in four time zones, the LSTM model was optimal in three time zones, and the TF-8 model was optimal in the 17 h prediction problem. In the MSE and MAE performances, the optimal models were different for prediction problems at 1 h, 5 h, 17 h, and 22 h; this suggests that there is a problem with outliers in the CNU dataset.
Table 7. Comparison of performances over time for the CNU dataset and derivation of the optimal model.
Table 8 below shows the experimental results of the high-voltage dataset; when the MSE and MAE performance indicators were summarized, the TCN model was overwhelmingly the optimal model in terms of numbers over the entire time period. It was found that the TCN model based on the convolutional layer was able to better solve the data characteristics and problem complexity within the high-voltage dataset.
Table 8. Comparison of performances over time for the high-voltage dataset and derivation of the optimal model.
In this section, the results of the experiments are summarized, and the optimal model for predicting power consumption is presented for each dataset. Based on the dataset, the model with the best performance over the entire time period and the best model on aver-age were compared. The optimal model is indicated in bold text. Table 9 below is organized based on the experimental results of the CNU dataset. In the results of the MSE experiment, it is seen that the LSTM and TCN models were the best when the best performance was required, and the TCN model was the best on average. In the results of the MSE experiment, the LSTM and TCN models, which both show maximum performances, produced 10–15% better performances than the ANN and GRU models, and it is seen that the average performance was good from at least 10% for a maximum of four times. In the results of the MAE experiment, it is seen that the LSTM model produced the best performance and that the TCN model was good on average. The LSTM model, which shows the maximum performance in the MAE experiment results, has similar performance to the TCN model, but shows improved performance by 4–12% compared to the ANN and GRU models and by 11–388% compared to the transformer model. On average, the TCN model outperformed the other models by up to 4–221%.
Table 9. CNU dataset performance comparisons by model and optimal model arrangement.
Table 10 below is organized based on the MSE results of the high-voltage dataset. For both the MSE and MAE experimental results, the TCN model is the optimal model. In the results of the MSE experiment, the TCN model with the highest performance showed twice the performance of the ANN, LSTM, or GRU models. On average, it was seen that the performance was about 2–3 times better. In the MAE experiment results, the TCN model shows good performance from a minimum of 49% to a maximum of 650%. On average, it is seen that the performance was good, ranging from a minimum of 21% to a maximum of 303%.
Table 10. High-voltage dataset performance comparisons by model and optimal model arrangement.
Based on these experimental results, the characteristics of the dataset were analyzed. The CNU dataset has a special characteristic of being from a school; conversely, the high-voltage dataset has the specificity of being from a factory. Figure 5 below is a visualization of each dataset. From Figure 5a, the CNU dataset has periodic characteristics, whereas the high-voltage dataset (Figure 5b) has periodic peaks owing to situations such as facility operation in the factory. The X-axis of Figure 5 represents power consumption, and the horizontal axis represents the order of time. A total of 2000 data were visualized in a random part within the dataset.
Figure 5. Visualization of the (a) CNU and (b) high-voltage datasets.

5. Conclusions

In this study, we experimented with the ANN, LSTM, and GRU as well as TCN and transformer models for two sets of data, namely the CNU and high-voltage datasets. The transformer model had the disadvantage of being heavy, so a lightweight three-layer model was applied. In addition, detailed power predictions were carried out in 1 h increments from 1 to 24 h. Thus, it was possible to derive an optimal model for each time period and an optimal model for the maximum and average performances. For the dataset showing periodic patterns, the LSTM model was the best at predicting information 1 h later, and for the dataset containing peak states, the TCN model was the best. It was found that the TCN model was the best on average, from 1 to 24 h. The datasets used in this study involve special environments, such as schools and factories, rather than general households. Non-residential datasets are an important area for energy management and energy efficiency improvement. In this paper, we can suggest the direction of the optimal model in the non-residential environment dataset. The problem that can derive the maximum performance in this paper is the prediction problem after 1 h. In the future, through hourly hyper parameter optimization, it will be possible to derive a more accurate performance not only after 1 h, but also from 2 h to 24 hours’ prediction problems compared to prediction problems after 1 h. In the future, we will research technologies that can improve energy management efficiency through research using non-residential data household electricity datasets and research using datasets including external factors that affect power consumption, such as temperature and humidity.

Author Contributions

Conceptualization, S.O. (Seungmin Oh) and S.O. (Sangwon Oh); methodology, S.O. (Seungmin Oh) and H.S.; software, S.O. (Seungmin Oh); validation, S.O. (Seungmin Oh) and T.-w.U.; formal analysis, J.K.; investigation, J.K.; resources, S.O. (Seungmin Oh); data curation, S.O. (Seungmin Oh); writing—original draft preparation, S.O. (Seungmin Oh); writing—review and editing, J.K.; visualization, S.O. (Sangwon Oh); supervision, J.K.; project administration, J.K.; funding acquisition, J.K.; project administration, J.K.; funding acquisition, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No.2021-0-02068, Artificial Intelligence Innovation Hub). This research was also supported by the MSIT(Ministry of Science and ICT), Korea, under the Innovative Human Resource Development for Local Intellectualization support program(IITP-2023-RS-2022-00156287) supervised by the IITP(Institute for Information & Communications Technology Planning & Evaluation).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tvaronavičienė, M.; Baublys, J.; Raudeliūnienė, J.; Jatautaitė, D. Chapter 1—Global energy consumption peculiarities and energy sources: Role of renewables. In Energy Transformation towards Sustainability; Tvaronavičienė, M., Ślusarczyk, B., Eds.; Elsevier: Amsterdam, The Netherlands, 2020; pp. 1–49. ISBN 9780128176887. [Google Scholar]
  2. Kober, T.; Schiffer, H.-W.; Densing, M.; Panos, E. Global energy perspectives to 2060—WEC’s World Energy Scenarios 2019. Energy Strat. Rev. 2020, 31, 100523. [Google Scholar] [CrossRef]
  3. Bilgen, S. Structure and environmental impact of global energy consumption. Renew. Sustain. Energy Rev. 2014, 38, 890–902. [Google Scholar] [CrossRef]
  4. IIris, Ç.; Lam, J.S.L. A review of energy efficiency in ports: Operational strategies, technologies and energy management systems. Renew. Sustain. Energy Rev. 2019, 112, 170–182. [Google Scholar] [CrossRef]
  5. Desislavov, R.; Martínez-Plumed, F.; Hernández-Orallo, J. Compute and energy consumption trends in deep learning inference. arXiv 2021, arXiv:2109.05472. [Google Scholar]
  6. Zhou, K.; Yang, S. Understanding household energy consumption behavior: The contribution of energy big data analytics. Renew. Sustain. Energy Rev. 2016, 56, 810–819. [Google Scholar] [CrossRef]
  7. Somu, N.; MR, G.R.; Ramamritham, K. A deep learning framework for building energy consumption forecast. Renew. Sustain. Energy Rev. 2021, 137, 110591. [Google Scholar] [CrossRef]
  8. Khalil, M.; McGough, A.S.; Pourmirza, Z.; Pazhoohesh, M.; Walker, S. Machine Learning, Deep Learning and Statistical Analysis for forecasting building energy consumption—A systematic review. Eng. Appl. Artif. Intell. 2022, 115, 105287. [Google Scholar] [CrossRef]
  9. Chou, J.-S.; Tran, D.-S. Forecasting energy consumption time series using machine learning techniques based on usage patterns of residential householders. Energy 2018, 165, 709–726. [Google Scholar] [CrossRef]
  10. Liang, F.; Yu, A.; Hatcher, W.G.; Yu, W.; Lu, C. Deep learning-based power usage forecast modeling and evaluation. Procedia Comput. Sci. 2019, 154, 102–108. [Google Scholar] [CrossRef]
  11. García-Martín, E.; Rodrigues, C.F.; Riley, G.; Grahn, H. Estimation of energy consumption in machine learning. J. Parallel Distrib. Comput. 2019, 134, 75–88. [Google Scholar] [CrossRef]
  12. Jiang, W. Deep learning based short-term load forecasting incorporating calendar and weather information. Internet Technol. Lett. 2022, 5, e383. [Google Scholar] [CrossRef]
  13. Xu, A.; Tian, M.-W.; Firouzi, B.; Alattas, K.A.; Mohammadzadeh, A.; Ghaderpour, E. A New Deep Learning Restricted Boltzmann Machine for Energy Consumption Forecasting. Sustainability 2022, 14, 10081. [Google Scholar] [CrossRef]
  14. Hong, Y.; Wang, D.; Su, J.; Ren, M.; Xu, W.; Wei, Y.; Yang, Z. Short-Term Power Load Forecasting in Three Stages Based on CEEMDAN-TGA Model. Sustainability 2023, 15, 11123. [Google Scholar] [CrossRef]
  15. Chen, Z.; Wang, C.; Lv, L.; Fan, L.; Wen, S.; Xiang, Z. Research on Peak Load Prediction of Distribution Network Lines Based on Prophet-LSTM Model. Sustainability 2023, 15, 11667. [Google Scholar] [CrossRef]
  16. Makala, B.; Bakovic, T. Artificial Intelligence in the Power Sector; International Finance Corporation: Washington, DC, USA, 2020. [Google Scholar]
  17. Wu, Q.; Ren, H.; Shi, S.; Fang, C.; Wan, S.; Li, Q. Analysis and prediction of industrial energy consumption behavior based on big data and artificial intelligence. Energy Rep. 2023, 9, 395–402. [Google Scholar] [CrossRef]
  18. Alzoubi, A. Machine learning for intelligent energy consumption in smart homes. Int. J. Comput. Inf. Manuf. (IJCIM) 2022, 2, 62–75. [Google Scholar] [CrossRef]
  19. Kim, T.; Cho, S. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  20. Khan, N.; Haq, I.U.; Khan, S.U.; Rho, S.; Lee, M.Y.; Baik, S.W. DB-Net: A novel dilated CNN based multi-step forecasting model for power consumption in integrated local energy systems. Int. J. Electr. Power Energy Syst. 2021, 133, 107023. [Google Scholar] [CrossRef]
  21. Ryu, S.; Noh, J.; Kim, H. Deep Neural Network Based Demand Side Short Term Load Forecasting. Energies 2017, 10, 3. [Google Scholar] [CrossRef]
  22. Amarasinghe, K.; Marino, D.L.; Manic, M. Deep neural networks for energy load forecasting. In Proceedings of the 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE), Edinburgh, UK, 19–21 June 2017; pp. 1483–1488. [Google Scholar]
  23. Yazdan, M.M.S.; Khosravia, M.; Saki, S.; Mehedi, M.A.A. Forecasting Energy Consumption Time Series Using Recurrent Neural Network in Tensorflow. Preprints 2022, 2022090404. [Google Scholar]
  24. Sachin, M.M.; Baby, M.P.; Ponraj, A.S. Analysis of Energy Consumption Using RNN-LSTM and ARIMA Model. J. Phys. Conf. Ser. 2020, 1716, 012048. [Google Scholar] [CrossRef]
  25. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, X.; Zhao, T.; Liu, H.; He, R. Power consumption predicting and anomaly detection based on long short-term memory neural network. In Proceedings of the 2019 IEEE 4th International Conference on Cloud Computing and Big Data Analysis (ICCCBDA), Chengdu, China, 12–15 April 2019; pp. 487–491. [Google Scholar]
  27. Torres, J.F.; Martínez-Álvarez, F.; Troncoso, A. A deep LSTM network for the Spanish electricity consumption forecasting. Neural Comput. Appl. 2022, 5, 10533–10545. [Google Scholar] [CrossRef] [PubMed]
  28. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  29. Mahjoub, S.; Chrifi-Alaoui, L.; Marhic, B.; Delahoche, L. Predicting Energy Consumption Using LSTM, Multi-Layer GRU and Drop-GRU Neural Networks. Sensors 2022, 22, 4062. [Google Scholar] [CrossRef] [PubMed]
  30. Li, X.; Zhuang, W.; Zhang, H. Short-Term Power Load Forecasting Based on Gate Recurrent Unit Network and Cloud Computing Platform. In Proceedings of the 4th International Conference on Computer Science and Application Engineering, Sanya, China, 20 October 2020; pp. 1–6. [Google Scholar]
  31. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  32. Cai, C.; Li, Y.; Su, Z.; Zhu, T.; He, Y. Short-Term Electrical Load Forecasting Based on VMD and GRU-TCN Hybrid Network. Appl. Sci. 2022, 12, 6647. [Google Scholar] [CrossRef]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
  34. Saoud, L.S.; Al-Marzouqi, H.; Hussein, R. Household Energy Consumption Prediction Using the Stationary Wavelet Transform and Transformers. IEEE Access 2022, 10, 5171–5183. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.