Next Article in Journal
Phase Behavior of Carbon Dioxide + Isobutanol and Carbon Dioxide + tert-Butanol Binary Systems
Previous Article in Journal
Technology-Agnostic Assessment of Wave Energy System Capabilities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Step Time-Series Clustering-Based Seq2Seq LSTM Learning for a Single Household Electricity Load Forecasting

Department of Electrical Engineering, Chonnam National University, Gwangju 61186, Korea
*
Author to whom correspondence should be addressed.
Energies 2022, 15(7), 2623; https://doi.org/10.3390/en15072623
Submission received: 14 March 2022 / Revised: 25 March 2022 / Accepted: 29 March 2022 / Published: 3 April 2022

Abstract

:
The deep learning (DL) approaches in smart grid (SG) describes the possibility of shifting the energy industry into a modern era of reliable and sustainable energy networks. This paper proposes a time-series clustering framework with multi-step time-series sequence to sequence (Seq2Seq) long short-term memory (LSTM) load forecasting strategy for households. Specifically, we investigate a clustering-based Seq2Seq LSTM electricity load forecasting model to undertake an energy load forecasting problem, where information input to the model contains individual appliances and aggregate energy as historical data of households. The original dataset is preprocessed, and forwarded to a multi-step time-series learning model which reduces the training time and guarantees convergence for energy forecasting. Furthermore, simulation results show the accuracy performance of the proposed model by validation and testing cluster data, which shows a promising potential of the proposed predictive model.

1. Introduction

The smart grid (SG) refers to the electric grid as an intelligent network of generation, transmission, and distribution. In the SG network, one of the benefits for the consumer is to generate energy using renewable resources while sustaining consumption and production with demand. The SG technologies are made possible by two-way communication, which involves data processing and control systems. The smart meter placed near a proximity network connects consumers with the grid and transmits data using advanced metering infrastructure (AMI) [1,2,3,4,5]. The AMI data help to analyze the distribution network and provide bidirectional communications to the consumer [2], which consolidates pricing information and demand response (DR). The recent advancement in the SG with effective demand-side management (DSM) allows big data into substantial benefits for the utilities and the customers. As reported [6], a smart meter can transfer energy information usage every 15 min, thus every million meters can generate 96 million reads in a day. However, due to the machine learning (ML) limitations [7,8,9,10,11], there has been a significant increase in deep learning (DL) models to exhibit complex correlations from a large dataset with numerous formats and replacements of manual feature extraction [12]. As reported [13], the expansion of AMI and wide-area monitoring systems (WAMS) in the SG network has increased the essential of DL techniques to deal along with the massive data. Moreover, a DL-based energy forecasting model is proposed to learn the correlation among distinct consumption behaviors for short-term load forecasting (STLF) [14,15,16]. The growth of literature on STLF for the individual household has demanded more practical analysis in the past decade. Specifically, the most commonly used DL techniques to address STLF problem include recurrent neural network (RNN), long short-term memory (LSTM), gated recurrent unit (GRU), convolutional neural network (CNN), reinforcement learning, autoencoders (AEs), and restricted Boltzmann machines (RBM). For example in [17], two state-of-art deep neural networks (DNN) architectures, deep feed neural network (FNN) and deep-RNN have been proposed for the STLF. The RNN is suitable for time-series data, where the network’s output is fed back to the input. The major drawback of using RNNs is that it is likely to vanish and explode gradient problems. For SG applications, this problem is overcome with memory gated structures that are LSTM and GRU.
In the recent past, it has been shown that load clustering [18,19,20] and multi-step load forecasting [21,22,23,24] techniques have been used to uncover power usage patterns established on specific measures of power data. The electric load clustering determines the power consumption patterns for a sustainable SG energy network with the demand-side response (DSR). However, based on the cluster data, it allows utilities to improve their policy and infrastructure planning techniques. Recently, Izonin et al. [25] and Tkachenko [26] proposed a non-iterative geometric transformations model (GTM), a neural-like structure discussed for solving time-series prediction tasks. Considering time-series forecasting problems, Ribeiro et al. [27] proposed a self-adaptive decomposed heterogeneous shallow model, predicting commercial and industrial electricity prices with multi-step-ahead commercial and industrial electricity prices. The optimal DL models drive more accurate forecasting and outperform the results of shallow structures models due to the limitations. For example, for the energy consumption forecasting problem, an LSTM-RNN-based univariate model [28] has been proposed, which forecast short- and medium-term time horizon few days up to weeks. On the other hand [29], a hybrid multivariate model has been proposed which combines CNN and LSTM for short- and long-term residential electricity forecasting. An increasing number of studies have found that home energy management systems (HEMS) and battery energy management systems (BEMS) have redirected the interest from aggregate loads towards data disaggregation of individual household loads. A novel clustering, classification, and forecasting (CCF) method has been proposed [30], outperforming the conventional smart meter-based model (SMBM) of individual household loads. The research activities for electric load clustering have increased continuously, and DL-based techniques for big data have recently demonstrated forecasting techniques for the SG network. The present information published to date has a significant viewpoint for DL models of aggregated residential loads that thoroughly improves the household load profiles and DR in the SG network. However, individual household load forecasting problems still need significant improvements and are required to be investigated in detail. The ENERTALK dataset is the first publicly available Korean dataset on electricity load consumption of 22 households with a sampling rate of 15 Hz. In this paper, we used the publicly available ENERTALK dataset [31]. The main contributions of this paper are as follows:
  • This paper proposes deep learning-based multi-step time-series Seq2Seq LSTM framework for the electricity load forecasting.
  • This paper takes on a new vision which combines the Seq2Seq LSTM and clustering to improve the efficiency of the DR program and provides multistep lookback analysis of a single household.
  • Different from the aggregated residential load, in this paper, a multi-step time-series electric load clustering and forecasting for a single household is proposed, which deals load forecasting to a DR program for supply and demand control.
We believe that our work presented in this paper will promote information communication technology (ICT) and artificial intelligence (AI) energy in SG networks.
This paper is organized as follows: Section 2 gives a brief overview of the proposed DL-based multi-step time-series Seq2Seq LSTM system model. In Section 3, we analyze load clustering and propose a multi-step time-series forecasting. Simulation results are presented in Section 4, followed by the conclusion in Section 5.

2. System Model

Electricity load forecasting is a challenging task for electricity utilities due to households’ different energy patterns and load characteristics. Nevertheless, it improves the utilities’ operational cost and estimates the electricity supply and demand to their customers. Figure 1 shows a proposed system model for a single household with a multi-step time-series clustering-based energy load forecasting strategy. In the system model, we use the 2.1 M samples of household 01 from the ENERTALK dataset. In the first stage, incomplete data such as noise, duplicates, and imbalances are handled by cleansing and normalization. Specifically, the data input size is reduced by extracting meaningful information for the later clustering and forecasting stage. Next, electric load clustering is essential for adopting an appropriate clustering algorithm for household data, where N represents the total number of clusters. Thus, each cluster is trained by the Seq2Seq LSTM in the final stage for better load dispatch and energy transfer scheduling.

Data Preprocessing

It is known that data preprocessing is one integral part of the DL, which affects the learning ability of the DL model. This dataset contains the electricity consumption of 22 households from 1 September 2016 to 30 April 2017. Figure 2 shows the 122 days of energy consumption pattern from household 01, 02, and 03, where each appliance can be differentiated from a particular periodicity pattern to another. The active power recorded for appliances: refrigerator, rice cooker, washing machine, water purifier, and television. Considering regional features, kimchi-refrigerator [32] is commonly used in Korea, and also impacts global energy consumption.
The available dataset is large in length, which was originally sampled at 15 Hz. Therefore, we downsampled the electricity load of 122 days of household at every single minute into a multi-step time-series household consumption load profile. As a result, the downsampled pattern of electricity consumption behavior can be observed for the total household load of all appliances of 122 days in Figure 3. Specifically, considerable attention has been taken in advance to the SG dataset before applying it to the clustering and forecasting stage. Firstly, a null hypothesis was test enacted, which confirms a sample comes from a normal distribution [33]. For the distribution and normality test, s and k represent the skew and kurtosis for the dataset, respectively. It has been observed that both s and k hold greater than zero, which confers statistical distribution is moderately skewed. The electricity load data is different for each appliance of all 122 days that can form an overfitting model. This statistical overfitting error can be minimized by partitioning the available data. The proposed system model overcomes this problem by adopting multi-step time-series clustering, which partitions the load data into parallel training and test data.

3. Proposed Framework

3.1. Multi-Step Time-Series Electric Load Clustering

The clustering aims to find patterns of the load curves, which decide shifts in the load demand, specifically the groups of sample curves of original data values and their derivatives. We use the K-means algorithm, which finds centroids and groups of sample curves based on the nearest centroid value. The muti-step time-series electricity load patterns of a single household are summarized in Figure 4. It shows cluster centroids and related sample groups, which is obtained iteratively to reduce the sum of the Euclidean distances. Furthermore, we evaluate the goodness of the number of clusters for our multi-step time-series clustering algorithm by using the Silhouette score.
The Silhouette score evaluates intercluster distance for each sample within a cluster and intracluster distances among all clusters. It can be observed that the K-means algorithm with optimal silhouette score has four different clusters groups. The cluster groups 1 and 4 show the hourly load curve when consumption is low for all the 122 days. The red load curves have high peaks and later keep a steady load curve the rest of the time. This prediction is noteworthy when employees mostly stay at home, which also includes weekends and special breaks. Cluster 3 has a high peak, but the rise is not steady most of the time compared to cluster 2, which can be observed as the fit load curve for business and school.

3.2. Forecast Multiload Profiles

The basic RNN model works well for the sequence of data, and if valuable information from the cluster is modified or neglected, it will reduce the model’s accuracy [34]. Additionally, derivation during the backpropagation produces the vanishing gradient problem. Therefore, to overcome this problem for our forecasting model, we select one of LSTM’s architecture. It is known that LSTM has an artificial RNN chain structure, and the basic architecture of an LSTM model is composed of an input gate, output gate, and a forget gate. The function of these gates is to regulate the information through the LSTM network. Recently, research has shown the LSTM encoder and decoder architecture commonly used for language translation, where it processes the sequence from one domain to another. Furthermore, encoder–decoder architectures are conditional autoregressive models, generating a sequence from one domain to another. Therefore, we use the Seq2Seq model, which deals with the sequence of data to forecast our multi-step time-series problem.
This paper adopts the Seq2Seq LSTM model for a more substantial analysis of our multi-step time-series load forecasting problem. The architecture of a Seq2Seq LSTM model is shown in Figure 5, where each rectangular block holds an LSTM cell. Furthermore, each LSTM cell contains a hidden state, h t , and cell state, c t , at timestep, t. The architecture is mainly divided into three parts: the encoder that is the input to the model, the decoder, which is the model’s output, and the encoder state vector. It can be seen that the encoder part is stacked with LSTM cells, and each of the cells allow a single element from the sequence as input.
The mathematical equations for the model are represented from (1) to (3). x = { x 1 , x 2 , , x T } and y = { y 1 , y 2 , , y T } are the input time sequence and targeted output sequence of the forecasting model, respectively. It is important to note that the input sequence length T may differ from the target sequence length T . h t and h t represent the encoder (1) and decoder (2) state from each LSTM cell at time t, respectively. Specifically, x t is the historical time-series data input to the LSTM cell at each timestep t. Furthermore, the LSTM cell forwards the collected information from each cluster for training to the next LSTM cell. The output produced by each LSTM cell at time t is represented by h t , which is the encoder vector, the final hidden state produced by the encoder. The vector forwards the information for all input elements to the decoder for the predictions. y t represents the target output sequence at a timestep t. The LSTM encoder finds a conditional probability (3) by obtaining a hidden state vector v from the input sequence x. The hidden state vector for the LSTM decoder learns conditional probability [35] for sequences of y. Therefore, by utilizing two different LSTM input and output sequences, learning is improved at the minimal cost of increased computational cost. Specifically, it benefits the network model to learn multi-step time-series simultaneously. The importance of the Seq2Seq model is that it can outline sequences of diverse time-series clustering lengths to each other. Next, the proposed framework forecast the load data for each cluster based on the train and test subsets.
h t = LSTM e n c o d e r x t , h t 1 ,
h t = LSTM d e c o d e r ( y t , h t 1 ) ,
p y 1 , y 2 , , y T | x 1 , x 2 , , x T = t = 1 T p y t | v , y 1 , y 2 , , y t 1 .

4. Numerical Analysis

Finally, the evaluation results are obtained from the proposed multistep time-series learning model. Table 1 shows the experiment parameters settings for all scenarios. We used LSTM, RNN, GRU, BiLSTM, and our proposed multi-step time-series Seq2Seq LSTM learning in these scenarios. The tests produce with Adam optimizer [36] on GPU enabled in a single NVidia to accelerate the computations. The platform adapts to develop the Seq2Seq LSTM learning in TensorFlow and Keras environment. In the simulation settings, we used timestamp as the time-series index. Each cluster datum is divided into training and testing subsets with a proportion of 67 % and 33 % , respectively. A further 25 % of the training data are used to confirm the experiment for validation. Furthermore, we used different combinations of lookback periods, which find how many previous timesteps have been used to predict the subsequent timestep. For example, 60 lookback period shows the timestep at t 60 , t 59 , , t 1 , and t has been used to predict the value at time t + 1 . Additionally, the dropout regularization [37] is set to 0.2 , and MinMax scaling [38] is set between 1 to 1 for the normalization of the dataset. Another hyperparameter is the batch size, set to 16, 64, and 128 combinations in the experiment. Moreover, for the result comparisons, we use the mean absolute error (MAE), mean squared error (MSE), and root mean squared error (RMSE) metric evaluation.
Figure 6 confirms that the cluster centroids and load curves are well apart and distinguished. It shows cluster mapping of 122 days electricity load data of a single household, where each data point is grouped into one of four different clusters. The closer data points are grouped as one cluster and represent similar load profiles. It can be seen that data points are close to the neighboring cluster, which confers the clustering is well molded.
For the comparative analysis, Table 2 shows the noteworthy results of the proposed forecasting model, and Figure 7 shows the performance comparison by using different lookback periods. The results of several learning approaches have been shown to improve the load forecasting with the same electricity load data. For the comparison, we used the existing implementation of state-of-the-art learning models, LSTM in [39], GRU in [40], and BiLSTM in [41]. However, we use the same parameter settings as in Table 1 for all learning models in this paper to make it comparable to the proposed multi-step time-series Seq2Seq LSTM learning. All learning models’ performance are observed using MAE, MAPE, and RMSE evaluation metrics. It is important to note that the results shown are adequate for the multi-step time-series SeqSeq LSTM model with 60-, 120-, and 180-step periods, which are obtained after the clustering algorithm. It has also been observed that when the step size increased, the MAE, MAPE, and RMSE increased slightly, which indicates that the benefits of our multi-step time series exhibit more stable results with a large step size. The performance of our proposed multi-step Seq2Seq LSTM model is better than LSTM, GRU, RNN, and BiLSTM, and shows a stable improvement in Figure 7. Furthermore, it has been observed from the metrics performance that the simple RNN model’s performance is the worst out of all the tested models. However, the RNN model is not useful for our proposed multi-step time-series clustering-based electricity forecasting model. These results confirm the effectiveness of the proposed encoder–decoder LSTM approach for the multi-step time-series electricity load forecast compared to outputting a vector directly into learning model.
This work finds different combinations of epochs and batch sizes to analyze the convergence Seq2Seq LSTM. The training performance of each architecture is evaluated by validation samples, including different numbers of epochs and batch sizes; to avoid overfitting, dropout has been used. Moreover, the model’s learning ability improved with the combination of time-series clustering and multi-step encoder–decoder sequences. Figure 8 shows the convergence of the loss function for the training and validation of the proposed model. It reveals that both training and validation loss decreases, and the proposed model obtained the convergence approximately after 20 epochs. In order to confirm that our proposed model also works well for the multi-step lookback periods, the training and testing have been carried out ranging from 1 to 200 lookback timesteps. Figure 9 shows the training, validation, and prediction curves for the 60-steps lookback periods. Furthermore, for the multi-step time-series forecasting, repeated executions of the proposed model have been performed by varying the number of lookback periods. For example, the actual load curves of a single household and its validation and predictive curves are shown in Figure 9 with the 60-steps forecasting. In correspondence with the testing data, the accuracy increases and the validation over the target data is achievable with the increase in size. Thus, the overall prediction of our proposed clustering-based Seq2Seq2 LSTM is the most suitable, and a very small portion of the data has a weak correlation.

5. Conclusions

This paper proposed a multi-step time-series clustering-based Seq2Seq LSTM learning model in order to forecast a single household’s electricity load. The proposed framework confirms multi-step time-series Seq2Seq-LSTM learning and compares it with other models (LSTM, RNN, GRU, and BiLSTM). The results demonstrate that the proposed model better adapts with the combination of clustering and 60-, 120-, and 180-step time series Seq2Seq load forecasting, which shows the best performance based on MAE, MAPE, and RMSE evaluation metrics. Furthermore, the simulation results showed that cluster-based multi-step time-series Seq2Seq LSTM learning significantly improves single household load forecasting. This confirms that the cluster based multi-step time-series learning is a reliable approach for the future load forecasting of households. Furthermore, the limitation of the univariate analysis can be extended to the multivariate with multi-step load forecasting in future work. This research will open further challenges for applying DL techniques to the SG network in the future.

Author Contributions

The research was carried out successfully with contribution from all authors. The main research idea and manuscript preparation were contributed by Z.M. and Y.C.; R.G. and A. contributed to the manuscript preparation and gave several suggestions from industrial perspectives. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2019R1I1A3A01060631).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ghosal, A.; Conti, M. Key management systems for smart grid advanced metering infrastructure: A survey. IEEE Commun. Surv. Tutor. 2019, 21, 2831–2848. [Google Scholar] [CrossRef] [Green Version]
  2. Desai, S.; Alhadad, R.; Chilamkurti, N.; Mahmood, A. A survey of privacy preserving schemes in IoE enabled smart grid advanced metering infrastructure. Clust. Comput. 2019, 22, 43–69. [Google Scholar] [CrossRef]
  3. Zainab, A.; Ghrayeb, A.; Syed, D.; Abu-Rub, H.; Refaat, S.S.; Bouhali, O. Big data management in smart grids: Technologies and challenges. IEEE Access 2021, 9, 73046–73059. [Google Scholar] [CrossRef]
  4. Masood, Z.; Ardiansyah; Choi, Y. Energy-Efficient Optimal Power Allocation for SWIPT Based IoT-Enabled Smart Meter. Sensors 2021, 21, 7857. [Google Scholar] [CrossRef]
  5. Chi, H.R.; Tsang, K.F.; Chui, K.T.; Chung, H.S.H.; Ling, B.W.K.; Lai, L.L. Interference-mitigated ZigBee-based advanced metering infrastructure. IEEE Trans. Ind. Inform. 2016, 12, 672–684. [Google Scholar] [CrossRef]
  6. Daki, H.; El Hannani, A.; Aqqal, A.; Haidine, A.; Dahbi, A. Big Data management in smart grid: Concepts, requirements and implementation. J. Big Data 2017, 4, 1–19. [Google Scholar] [CrossRef] [Green Version]
  7. Kotsiopoulos, T.; Sarigiannidis, P.; Ioannidis, D.; Tzovaras, D. Machine Learning and Deep Learning in Smart Manufacturing: The Smart Grid Paradigm. Comput. Sci. Rev. 2021, 40, 100341. [Google Scholar] [CrossRef]
  8. Janiesch, C.; Zschech, P.; Heinrich, K. Machine learning and deep learning. Electron. Mark. 2021, 31, 685–695. [Google Scholar] [CrossRef]
  9. Gantassi, R.; Gouissem, B.B.; Othmen, J.B. Routing protocol LEACH-K using K-means algorithm in wireless sensor network. In Proceedings of the Workshops of the International Conference on Advanced Information Networking and Applications, Caserta, Italy, 15–17 April 2020; pp. 299–309. [Google Scholar] [CrossRef]
  10. Gantassi, R.; Ben Gouissem, B.; Cheikhrouhou, O.; El Khediri, S.; Hasnaoui, S. Optimizing quality of service of clustering protocols in large-scale wireless sensor networks with mobile data collector and machine learning. Sec. Commun. Netw. 2021, 2021, 5531185. [Google Scholar] [CrossRef]
  11. Nguyen, G.; Dlugolinsky, S.; Bobák, M.; Tran, V.; García, Á.L.; Heredia, I.; Malík, P.; Hluchỳ, L. Machine learning and deep learning frameworks and libraries for large-scale data mining: A survey. Artif. Intell. Rev. 2019, 52, 77–124. [Google Scholar] [CrossRef] [Green Version]
  12. Hong, Y.; Zhou, Y.; Li, Q.; Xu, W.; Zheng, X. A deep learning method for short-term residential load forecasting in smart grid. IEEE Access 2020, 8, 55785–55797. [Google Scholar] [CrossRef]
  13. Shobol, A.; Ali, M.H.; Wadi, M.; TüR, M.R. Overview of big data in smart grid. In Proceedings of the 2019 8th International Conference on Renewable Energy Research and Applications (ICRERA), Brasov, Romania, 3–6 November 2019; pp. 1022–1025. [Google Scholar] [CrossRef]
  14. Li, N.; Wang, L.; Li, X.; Zhu, Q. An effective deep learning neural network model for short-term load forecasting. Concurr. Comput. Pract. Exp. 2020, 32, e5595. [Google Scholar] [CrossRef]
  15. Kim, S.H.; Lee, G.; Kwon, G.Y.; Kim, D.I.; Shin, Y.J. Deep learning based on multi-decomposition for short-term load forecasting. Energies 2018, 11, 3433. [Google Scholar] [CrossRef] [Green Version]
  16. Choi, H.; Ryu, S.; Kim, H. Short-term load forecasting based on ResNet and LSTM. In Proceedings of the 2018 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm), Aalborg, Denmark, 29–31 October 2018; pp. 1–6. [Google Scholar] [CrossRef]
  17. Mohammad, F.; Kim, Y.C. Energy load forecasting model based on deep neural networks for smart grids. Int. J. Syst. Assur. Eng. Manag. 2020, 11, 824–834. [Google Scholar] [CrossRef]
  18. Jeong, D.; Park, C.; Ko, Y.M. Short-term electric load forecasting for buildings using logistic mixture vector autoregressive model with curve registration. Appl. Energy 2021, 282, 116249. [Google Scholar] [CrossRef]
  19. Syed, D.; Abu-Rub, H.; Ghrayeb, A.; Refaat, S.S.; Houchati, M.; Bouhali, O.; Bañales, S. Deep learning-based short-term load forecasting approach in smart grid with clustering and consumption pattern recognition. IEEE Access 2021, 9, 54992–55008. [Google Scholar] [CrossRef]
  20. Zhou, B.; Meng, Y.; Huang, W.; Wang, H.; Deng, L.; Huang, S.; Wei, J. Multi-energy net load forecasting for integrated local energy systems with heterogeneous prosumers. Int. J. Electr. Power Energy Syst. 2021, 126, 106542. [Google Scholar] [CrossRef]
  21. Masum, S.; Liu, Y.; Chiverton, J. Multi-step time series forecasting of electric load using machine learning models. In Proceedings of the International Conference on Artificial Intelligence and Soft Computing, Zakopane, Poland, 3–7 June 2018; pp. 148–159. [Google Scholar] [CrossRef] [Green Version]
  22. Yang, Y.; Shang, Z.; Chen, Y.; Chen, Y. Multi-objective particle swarm optimization algorithm for multi-step electric load forecasting. Energies 2020, 13, 532. [Google Scholar] [CrossRef] [Green Version]
  23. Nugraha, G.D.; Musa, A.; Cho, J.; Park, K.; Choi, D. Lambda-based data processing architecture for two-level load forecasting in residential buildings. Energies 2018, 11, 772. [Google Scholar] [CrossRef] [Green Version]
  24. Deng, Z.; Wang, B.; Xu, Y.; Xu, T.; Liu, C.; Zhu, Z. Multi-scale convolutional neural network with time-cognition for multi-step short-term load forecasting. IEEE Access 2019, 7, 88058–88071. [Google Scholar] [CrossRef]
  25. Izonin, I.; Tkachenko, R.; Kryvinska, N.; Tkachenko, P. Multiple linear regression based on coefficients identification using non-iterative SGTM neural-like structure. In Proceedings of the International Work-Conference on Artificial Neural Networks, Gran Canaria, Spain, 12–14 June 2019; pp. 467–479. [Google Scholar] [CrossRef]
  26. Tkachenko, R.; Izonin, I. Model and principles for the implementation of neural-like structures based on geometric data transformations. In Proceedings of the International Conference on Computer Science, Engineering and Education Applications, Kiev, Ukraine, 18–20 January 2018; pp. 578–587. [Google Scholar] [CrossRef]
  27. Ribeiro, M.H.D.M.; Stefenon, S.F.; de Lima, J.D.; Nied, A.; Mariani, V.C.; Coelho, L.d.S. Electricity price forecasting based on self-adaptive decomposition and heterogeneous ensemble learning. Energies 2020, 13, 5190. [Google Scholar] [CrossRef]
  28. Bouktif, S.; Fiaz, A.; Ouni, A.; Serhani, M.A. Optimal deep learning LSTM model for electric load forecasting using feature selection and genetic algorithm: Comparison with machine learning approaches. Energies 2018, 11, 1636. [Google Scholar] [CrossRef] [Green Version]
  29. Kim, T.Y.; Cho, S.B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  30. Yildiz, B.; Bilbao, J.I.; Dore, J.; Sproul, A. Household electricity load forecasting using historical smart meter data with clustering and classification techniques. In Proceedings of the 2018 IEEE Innovative Smart Grid Technologies-Asia (ISGT Asia), Singapore, 22–25 May 2018; pp. 873–879. [Google Scholar] [CrossRef]
  31. Shin, C.; Lee, E.; Han, J.; Yim, J.; Rhee, W.; Lee, H. The ENERTALK dataset, 15 Hz electricity consumption data from 22 houses in Korea. Sci. Data 2019, 6, 1–13. [Google Scholar] [CrossRef] [PubMed]
  32. Ayub, M.; El-Alfy, E.S.M. Impact of Normalization on BiLSTM Based Models for Energy Disaggregation. In Proceedings of the 2020 International Conference on Data Analytics for Business and Industry: Way Towards a Sustainable Economy (ICDABI), Sakheer, Bahrain, 26–27 October 2020; pp. 1–6. [Google Scholar] [CrossRef]
  33. Anderson, D.R.; Burnham, K.P.; Thompson, W.L. Null hypothesis testing: Problems, prevalence, and an alternative. J. Wildl. Manag. 2000, 64, 912–923. [Google Scholar] [CrossRef]
  34. Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  35. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. Adv. Neural Inform. Process. Syst. 2014, 27, 1–9. [Google Scholar]
  36. Bock, S.; Weiß, M. A proof of local convergence for the Adam optimizer. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar] [CrossRef]
  37. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  38. Ahsan, M.M.; Mahmud, M.; Saha, P.K.; Gupta, K.D.; Siddique, Z. Effect of data scaling methods on machine learning algorithms and model performance. Technologies 2021, 9, 52. [Google Scholar] [CrossRef]
  39. Memarzadeh, G.; Keynia, F. Short-term electricity load and price forecasting by a new optimal LSTM-NN based prediction algorithm. Electr. Power Syst. Res. 2021, 192, 106995. [Google Scholar] [CrossRef]
  40. Veeramsetty, V.; Reddy, K.R.; Santhosh, M.; Mohnot, A.; Singal, G. Short-term electric power load forecasting using random forest and gated recurrent unit. Electr. Eng. 2022, 104, 307–329. [Google Scholar] [CrossRef]
  41. Mughees, N.; Mohsin, S.A.; Mughees, A.; Mughees, A. Deep sequence to sequence Bi-LSTM neural networks for day-ahead peak load forecasting. Expert Syst. Appl. 2021, 175, 114844. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed multi-step time-series electricity clustering and load forecasting system model. Time-series data preprocessing of aggregate load and downsampling of a single household. Before electricity load forecasting, each multi-step cluster datum is fed into a multi-step time-series Seq2Seq LSTM learning model.
Figure 1. The framework of the proposed multi-step time-series electricity clustering and load forecasting system model. Time-series data preprocessing of aggregate load and downsampling of a single household. Before electricity load forecasting, each multi-step cluster datum is fed into a multi-step time-series Seq2Seq LSTM learning model.
Energies 15 02623 g001
Figure 2. Periodic patterns of home appliance of household ID = ‘01’, ‘02’, and ‘03’. Energy consumption behavior of ON state for households’ appliances washing machine, kimchi-refrigerator, refrigerator, water purifier, television, rice cooker, microwave, and the total power watts.
Figure 2. Periodic patterns of home appliance of household ID = ‘01’, ‘02’, and ‘03’. Energy consumption behavior of ON state for households’ appliances washing machine, kimchi-refrigerator, refrigerator, water purifier, television, rice cooker, microwave, and the total power watts.
Energies 15 02623 g002
Figure 3. Time-series daily load profiles without clustering (122 days from 1 October 2016 to 31 January 2017).
Figure 3. Time-series daily load profiles without clustering (122 days from 1 October 2016 to 31 January 2017).
Energies 15 02623 g003
Figure 4. Load characteristics curve of multi-step time-series clusters with number of clusters N = 4.
Figure 4. Load characteristics curve of multi-step time-series clusters with number of clusters N = 4.
Energies 15 02623 g004
Figure 5. A Seq2Seq LSTM network model for the time-series load forecasting. At each timestep, the encoder takes one series of data x t at time t, and its previous state h t 1 and produces an output vector h t state and cell state C t . The next decoder generates an output sequence y t , at each step taking at time t, the previous state, and a weighted combination of all the encoder outputs (i.e., encoder state vector).
Figure 5. A Seq2Seq LSTM network model for the time-series load forecasting. At each timestep, the encoder takes one series of data x t at time t, and its previous state h t 1 and produces an output vector h t state and cell state C t . The next decoder generates an output sequence y t , at each step taking at time t, the previous state, and a weighted combination of all the encoder outputs (i.e., encoder state vector).
Energies 15 02623 g005
Figure 6. Results of time-series data points of K-means clustering.
Figure 6. Results of time-series data points of K-means clustering.
Energies 15 02623 g006
Figure 7. Example performance evaluation of our proposed model with other learning models with 60 timesteps.
Figure 7. Example performance evaluation of our proposed model with other learning models with 60 timesteps.
Energies 15 02623 g007
Figure 8. Results of convergence of the loss function of the proposed multi-step time-series Seq2Seq LSTM learning. Train and validation loss with number of epochs and mean square error.
Figure 8. Results of convergence of the loss function of the proposed multi-step time-series Seq2Seq LSTM learning. Train and validation loss with number of epochs and mean square error.
Energies 15 02623 g008
Figure 9. The actual load curves of household ID = ‘01’ of 122 days together with its validation and predictive curves from our multi-step time-series forecasting method.
Figure 9. The actual load curves of household ID = ‘01’ of 122 days together with its validation and predictive curves from our multi-step time-series forecasting method.
Energies 15 02623 g009
Table 1. Paramters settings for the multi-step time-series learning model.
Table 1. Paramters settings for the multi-step time-series learning model.
ParametersValue
Forecasting modelGRU, RNN, BiLSTM, Seq2Seq LSTM
Training dataset 67 % of the total input
Testing dataset 33 % of the total input
Validation split 25 % of the train dataset
Minmax normalization 1 to 1
Regularizationdropout 0.2 each layer
Number of sequence64 units
Number of lookback60, 120, and 180 periods
Nnumber of maximal epochs100
Optimization algorithmAdam
Testing evaluation metricsMAE, MAPE, RMSE
Table 2. Comparison of the value of MAE, MAPE, and RMSE associated to all scenarios with the proposed multi-step time-series Seq2Seq LSTM forecasting model, specifically shows three different lookback periods at 60, 120, and 180 timesteps.
Table 2. Comparison of the value of MAE, MAPE, and RMSE associated to all scenarios with the proposed multi-step time-series Seq2Seq LSTM forecasting model, specifically shows three different lookback periods at 60, 120, and 180 timesteps.
Forecasting ModelsMAEMAPERMSE
LSTM 60 timesteps44.311.9392.91
RNN 60 timesteps55.717.99102.37
GRU 60 timesteps39.212.2084.97
BiLSTM 60 timesteps40.312.3288.41
Seq2Seq LSTM 60 timesteps (proposed)35.110.9382.75
LSTM 120 timesteps50.916.2298.35
RNN 120 timesteps61.220.10109.28
GRU 120 timesteps36.610.8782.71
BiLSTM 120 timesteps47.213.2391.63
Seq2Seq LSTM 120 timesteps (proposed)46.512.2286.50
LSTM 180 timesteps48.913.0999.50
RNN 180 timesteps67.322.47113.92
GRU 180 timesteps42.913.2588.60
BiLSTM 180 timesteps41.611.1989.75
Seq2Seq LSTM 180 timesteps (proposed)38.513.3288.65
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Masood, Z.; Gantassi, R.; Ardiansyah; Choi, Y. A Multi-Step Time-Series Clustering-Based Seq2Seq LSTM Learning for a Single Household Electricity Load Forecasting. Energies 2022, 15, 2623. https://doi.org/10.3390/en15072623

AMA Style

Masood Z, Gantassi R, Ardiansyah, Choi Y. A Multi-Step Time-Series Clustering-Based Seq2Seq LSTM Learning for a Single Household Electricity Load Forecasting. Energies. 2022; 15(7):2623. https://doi.org/10.3390/en15072623

Chicago/Turabian Style

Masood, Zaki, Rahma Gantassi, Ardiansyah, and Yonghoon Choi. 2022. "A Multi-Step Time-Series Clustering-Based Seq2Seq LSTM Learning for a Single Household Electricity Load Forecasting" Energies 15, no. 7: 2623. https://doi.org/10.3390/en15072623

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop