Next Article in Journal
A Neuro-Predictive Controller Scheme for Integration of a Basic Wind Energy Generation Unit with an Electrical Power System
Next Article in Special Issue
A Review of Auto-Regressive Methods Applications to Short-Term Demand Forecasting in Power Systems
Previous Article in Journal
Influence of a Type of Rock Mass on the Stability of Headings in Polish Underground Copper Mines Based on Boundary Element Method
Previous Article in Special Issue
Medium-Term Forecasts of Load Profiles in Polish Power System including E-Mobility Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Use of Singular Spectrum Analysis and K-Means Clustering-Based Bootstrap to Improve Multistep Ahead Load Forecasting

by
Winita Sulandari
1,*,
Yudho Yudhanto
2 and
Paulo Canas Rodrigues
3
1
Study Program of Statistics, Universitas Sebelas Maret, Surakarta 57126, Indonesia
2
Informatics Engineering, Vocational School, Universitas Sebelas Maret, Surakarta 57126, Indonesia
3
Department of Statistics, Federal University of Bahia, Salvador 40170-110, Brazil
*
Author to whom correspondence should be addressed.
Energies 2022, 15(16), 5838; https://doi.org/10.3390/en15165838
Submission received: 12 July 2022 / Revised: 5 August 2022 / Accepted: 8 August 2022 / Published: 11 August 2022
(This article belongs to the Special Issue Intelligent Forecasting and Optimization in Electrical Power Systems)

Abstract

:
In general, studies on short-term hourly electricity load modeling and forecasting do not investigate in detail the sources of uncertainty in forecasting. This study aims to evaluate the impact and benefits of applying bootstrap aggregation in overcoming the uncertainty in time series forecasting, thereby increasing the accuracy of multistep ahead point forecasts. We implemented the existing and proposed clustering-based bootstrapping methods to generate new electricity load time series. In the proposed method, we use singular spectrum analysis to decompose the series between signal and noise to reduce the variance of the bootstrapped series. The noise is then bootstrapped by K-means clustering-based generation of Gaussian normal distribution (KM.N) before adding it back to the signal, resulting in the bootstrapped series. We apply the benchmark models for electricity load forecasting, SARIMA, NNAR, TBATS, and DSHW, to model all new bootstrapped series and determine the multistep ahead point forecasts. The forecast values obtained from the original series are compared with the mean and median across all forecasts calculated from the bootstrapped series using the Malaysian, Polish, and Indonesian hourly load series for 12, 24, and 36 steps ahead. We conclude that, in this case, the proposed bootstrapping method improves the accuracy of multistep-ahead forecast values, especially when considering the SARIMA and NNAR models.

1. Introduction

Electricity load forecasting plays a critical role in controlling the balance between power demand and supply. Sometimes, the energy demand exceeds the energy supply and vice versa, which results in financial losses. An important aspect of a smart grid system is determining an accurate load forecasting model. Electricity load forecasting provides information that will simplify the work of planning consumption, generation, distribution, and other essential tasks of the smart grid system [1,2].
Much work has been performed to develop models and strategies to improve the electricity load forecasting accuracy. Generally, an hourly load series shows three relationships, i.e., between the observations for consecutive hours on a particular day, between the observations for the same hour on consecutive days, and between the observations for the same hour on the same day in successive weeks. In certain countries, the hourly load series may become more complex due to calendar variations [3]. The effect of calendar variation is usually considered by including a dummy variable in the model [4,5,6,7]. In countries with four seasons, the temperature is often included in the load forecasting model [8,9]. For countries with two seasons, such as Malaysia, it is also possible to include temperature information to improve the forecasts’ accuracy [10]. Many models, from simple to complex, have been proposed and developed by researchers and practitioners around the world to improve the accuracy of electricity load forecasting, e.g., regression and seasonal autoregressive integrated moving average (SARIMA) models [4,11,12], exponential smoothing [3,6,13,14,15], neural network (NN) [16,17,18,19], singular spectrum analysis (SSA) [20,21,22], wavelets [23,24], fuzzy systems [10,25,26], support vector machine [1,21,27,28], among others. However, the most suitable model for electricity load forecasting in a given country may not be the best to model the data in another country because of different consumption and behavioral characteristics.
In this study, we discuss the implementation of the bootstrap aggregating method to improve the accuracy of multistep ahead load forecast. Bootstrap aggregation, which is known by the acronym “bagging”, was proposed by [29] to reduce the variance of the predictor. It works by generating replicated bootstrap samples of the training data and using them to obtain the aggregated predictor. Bagging aims at improving the point forecast by considering sources of uncertainty, namely, the parameter estimates, the appropriate model determination, and the noise. In 2016, [30] successfully applied the development of this method in the field of time series forecasting by using the moving block bootstrap (MBB). Further, [31] explored how bagging improves point forecasts and showed that model selection as a solution to model uncertainty was the most influential on the success of bagging in time series. As described in [30], MBB bagging methods first apply the Box–Cox transform to the original series and then decompose it into a trend, seasonal, and noise using STL (Seasonal and Trend decomposition using Loess). STL is a decomposition method developed by [32]. In MBB, the noise is bootstrapped and added back to the trend and seasonal components. The new transformed bootstrapped series are then inverted and modeled. However, MBB is more appropriate for bootstrapping stationary time series. When the original data are not stationary, the bootstrapped series may be very noisy and do not fluctuate as the original series [30].
Recently, [33] proposed three clustering-based bootstrap aggregating methods, i.e., Smoothed MBB (S.MBB), K-means clustering based (KM), and K-means clustering based-generated from Gaussian normal distribution (KM.N), which perform better under noisy and fluctuating data. In adapting the fluctuating data, the S.MBB method smooths the noise using simple exponential smoothing before applying MBB. Meanwhile, KM and KM.N methods adapt to the noisy series by first implementing the K-means cluster. The original series clusters into K groups and then creates new time series based on the clusters. The difference between KM and KM.N is in how they generate the bootstrap series. In KM, a new time series is created directly by sampling values of clusters, while in KM.N, it is created by generating values based on the parameters of the Gaussian normal distribution of clusters. Both the KM and KM.N methods succeeded in making the bootstrapped time series have low variance between each other. Based on the experimental study of the electricity load series with multiple seasonal and calendar effects, KM.N performed better than the KM method [33]. However, this method creates a bootstrapped series based on the original data without sorting out signal and noise. Thus, in more complex series where the calendar effect may not be visible clearly in the original data, it will produce a noisier bootstrapped series at specific points, especially at times affected by calendar variation.
Inspired by [30,33], this study proposes an SSA–clustering-based method named SSA.KM.N as a modification of KM.N. Our proposed method combines singular spectrum analysis (SSA) as an alternative to the STL method in MBB and KM.N to generate new series from the remainder of the SSA decomposition. Literature shows that SSA is powerful in decomposing time series with complex seasonal patterns ([5,20]). SSA plays a role in breaking down time series, which have trends, multiple seasonal components, and are affected by calendar variation, into signal and noise, which generally contain extreme values representing calendar effects in more detail. By taking advantage of the unique strengths of SSA and KMN, our methodology can better adapt to fluctuating time series related to the effects of calendar variation. Bootstrapping the noise using KM.N and adding it to the signal is expected to produce a bootstrapped time series with low variance and values around the original series.
In this work, the proposed method compares with KM.N in its application to bootstrap two Malaysian load time series with different sample sizes and different time periods; one from Poland and another from Indonesia. We evaluate the impact and benefits of applying SSA.KM.N and KM.N in overcoming the source of uncertainty in time series forecasting and their success in increasing the accuracy of multistep ahead point forecast obtained by standard models such as SARIMA, NNAR, TBATS, and DSHW models.
The rest of the paper is organized as follows. Section 2 describes the methods used in this paper, starting from forecasting methods, decomposition, bagging, and ensemble methods. We also present the procedure of our proposed approach in this section. Section 3 reports the application of KM.N and SSA.KM.N to the four electricity load time series and shows the error evaluation for 12, 24, and 36 steps-ahead point forecasts obtained from SARIMA, NNAR, DSHW, and TBATS for further investigation and assessment. Conclusions are found in Section 4.

2. Materials and Methods

This section contains a brief overview of the methods used for time series modeling and forecasting, the decomposition method, the ensemble learning, and the proposed approach.

2.1. Forecasting Methods

SARIMA, and exponential smoothing (i.e., TBATS and DSHW), are popular approaches to forecast trend and seasonal time series. On the other hand, NNAR is a powerful method for capturing nonlinear relationships in time series data. These four methods are frequently used in modeling load series, and their forecast accuracy is used as benchmarks for other proposed methods [4,10,22]. For example, the Spanish Transmission System Operator uses autoregressive (AR) and NN models [7].
The seasonal ARIMA model, notated as SARIMA p , d , q P , D , Q S , is an extension of ARIMA model that accommodates the seasonal component of the time series [34,35], and can be written as follows:
ϕ p B ψ P B S d S D z t = θ q B ϑ Q B S a t
where z t is observation at time t , S is seasonal period, p , P , q , and Q are the orders of autoregressive, seasonal autoregressive, moving average, and seasonal moving average, respectively. Superscript d and D notate the regular and seasonal differentiation, while S = 1 B S is a backshift operator, and ϕ p B , and θ q B are polynomials in B of degree p and q , respectively. Notations ψ P B S , and ϑ Q B S are polynomials in B S of degrees P and Q , and a t is white noise. The orders of p , q , P , and Q can be determined from the correlogram and partial correlogram. Oftentimes, the identification of these orders is not an easy task and the user experience is required [36]. The automatic algorithms discussed in [37] with the “auto.arima” function of the R software can be used to help handle this problem [31]. However, other researchers may prefer to estimate the parameters manually instead of using automated packages [38]. In this case, we use “Arima” function included in the package “forecast” in the software R [39].
NNAR is a feedforward neural network that consists of lagged input neurons, one hidden layer with nonlinear function, and one output neuron [40,41]. NNAR p , P , k S model can be represented as in Equation (2),
z t = b 1 + j = 1 k v j f j b 0 + i = 1 p z t i w i j + m = 1 P z t m S w m j
where b 0 and b 1 are biased, k is the number of neurons in the hidden layer, p is the order of the non-seasonal component, while P is the order of the seasonal component. The sigmoid function at the jth neuron in the hidden layer, f j , is defined in Equation (3),
f j u = 1 1 + exp u ,
where u = b 0 + i = 1 p z t i w i j + m = 1 P z t m S w m j . In its implementation we use the “nnetar” function in the R package “forecast” ([39]). Later in the experimental study, P is set to be 1, p is the optimal number of lags for the linear model fitted to the seasonally adjusted data, and k is determined by the rounded value of P + p / 2 . The final forecast values are obtained by averaging 20 networks with different random starting weights.
TBATS and DSHW are modifications of the exponential smoothing to handle trends and multiple seasonal patterns in time series forecasting [42]. DSHW, proposed by [13], accommodates two seasonal patterns where one cycle may be nested within another. Meanwhile, TBATS, proposed by [3], can handle a more complex seasonal pattern in time series forecasting. The term “complex” means that the time series has a trend, and multiple seasonal patterns with integer or non-integer periods, and this may include dual-calendar seasonal effects. Success studies of the use of DSHW and TBATS in modeling electricity load time series can be found in [3,13,43]. Detail of these models can be found in [3,13,42,44]. In this paper, TBATS and DSHW are fitted by using the “tbats” and “dshw” functions included in the R package “forecast” ([39]).

2.2. SSA Decomposition Method

SSA is a technique in the field of time series analysis that has a vast range of applicability in decomposition [20], missing value imputation [45], and forecasting [46]. In this study, we focus on the use of the SSA algorithm to decompose time series into the following two components: signal and unstructured noise. SSA consists of four steps, namely, embedding, singular value decomposition (SVD), grouping, and diagonal averaging [47,48].
In embedding, we transform a time series z t   =   { z 1 , z 2 , , z n tr } into a trajectory matrix as in (4) as follows:
Z = z 1 z 2 z 3 z c 2 z 2 z 3 z 4 z c 2 + 1 z c 1 z c 1 + 1 z c 1 + 2 z n t r
where c 1 is the window length, n t r the number of training data sets, c 2 = n t r c 1 + 1 and c 1 c 2 .
The matrix Z is then decomposed by SVD and expressed as follows:
Z = Z 1 + Z 2 = l = 1 r 1 λ l u l v l + l = r 1 + 1 r 2 λ l u l v l
where λ l are the eigenvalues of matrix Z , u l and v l are left and right singular vectors of the matrix Z corresponding to the eigenvalues λ l , respectively. We need to determine r 1 , the number of signal components used for reconstruction in the grouping stage. Finally, we can obtain signal and noise by the diagonal averaging procedure, i.e., the anti-diagonals that map the matrices of the signal and noise components back to time series. The original time series can then be expressed as follows:
z t = z ˜ t 1 + z ˜ t 2
where z ˜ t 1 is the signal and z ˜ t 2 is the noise.

2.3. Bagging and Ensemble Methods

The K-means clustering-based bagging method, proposed by [33], clusters a univariate time series into K groups by the K-means method. Each cluster has its average value and standard deviation as parameters of the cluster. A Gaussian normal distribution according to these parameters is then used to generate a random number as a new value of the bootstrap time series. For example, suppose z t is a value of the original time series at time t that belongs to the ith cluster. In that case, we can obtain the new value of the bootstrap time series at time t by generating a random value based on the Gaussian normal distribution of the particular cluster. The code for bootstrapping time series by K-means clustering-based-generated from Gaussian normal distribution (KM.N) can be found in [49].

2.4. Proposed Approach

The proposed SSA.KM.N bagging method is a modification of KM.N where the first stage decomposes the original time series using SSA (Figure 1; blue cells). As shown in [20,50], SSA can be used to decompose complex time series into several simple pattern components.
Step 1.
Divide the series into the following two parts: training and testing datasets;
Step 2.
Generate B new series from the original training data by the KM.N method;
Step 3.
Generate B new series from the original training data by the SSA.KM.N method;
a.
Apply SSA to the original training data to define the signal and the irregular component;
b.
Generate B new series from the irregular component obtained in Step 3a by the KM.N method;
c.
Sum each new irregular component obtained from Step 3b with the signal component so that we obtain B bootstrapped series of the original training data;
Step 4.
Model the original training data, and each bootstrapped series by SARIMA, NNAR, TBATS, and DSHW;
Step 5.
Calculate up to M-steps-ahead forecast values by each model obtained in Step 4;
a.
Define the M-steps-ahead forecast values from the SARIMA, NNAR, TBATS, and DSHW models obtained from the original training data series;
b.
Apply mean and median to calculate the final forecast of the bootstrap series determined in Step 2 and Step 3 for the first n B bootstrap series;
Step 6.
Evaluate the forecast accuracy based on root mean square error (RMSE) and mean absolute percentage error (MAPE).
The two accuracy measures considered in this study can be defined as follows. MAPE, calculated by Equation (7), is frequently used in evaluating load forecasting accuracy since it is a scale-independent error that may compare forecast performance between different data sets [44,51]. Meanwhile, RMSE, calculated using Equation (8), is a scale-dependent error that can be used to compare the accuracy performance of several models on the same data set [51].
MAPE = 100 % H h = 1 H | e n t r + h / y n t r + h |
RMSE = h = 1 H e n t r + h 2 H 1 / 2
In addition, we also evaluate the model using mean bias error (MBE) as defined in Equation (9). It provides information whether there is a positive or negative bias [52]. We can calculate MBE by the following:
MBE = h = 1 H e n t r + h H
where
e n t r + h = y n t r + h y ^ n t r + h
y ^ n t r + h and y n t r + h are the predicted value and the actual value at time ( n t r + h ) , respectively. H is the number of observations included in the calculation and n t r is the size of training data.
In this study, each bootstrapped series is modeled separately, and the final forecast for time t is obtained by the following two ensemble methods: the mean and the median across all forecast values at time t, calculated from the bootstrap series. In this study, we obtain the mean and the median of the predicted values generated from the first n B (between 10 to 100) bootstrapped series to investigate whether the number of generated time series affects the accuracy of the forecast results.

3. Results and Discussion

In this section, we discuss two hourly electricity loads in Johor, Malaysia, and two other electricity load datasets from Poland and Indonesia. We decided to use these four data sets to show the generality of our work for electricity load forecasting.

3.1. Application to the Hourly Electricity Load of Johor, Malaysia

This subsection focuses on short-term forecasting of hourly electricity load with application to Malaysian data. We consider two datasets with different sizes that can be accessed in [53]. The first is the hourly load series from 1 January to 31 December 2009, and the second is the hourly load series from 1 January to 31 July 2010, which are depicted in Figure 2 (see Figure 2a,b, respectively).
The period from 1 January, time 00:00, to 30 November 2009, time 23:00, and the period from 1 January, time 00:00, to 30 June 2010, time 23:00, were used for estimation purposes as the training data. The remainder was used to evaluate the forecast performance of the models. These periods are summarized in Table 1.
Our analysis generated 100 bootstrap time series using the KM.N and the SSA.KM.N methods. Note that the original series is included in those 100-bootstrap series. Figure 3 shows the original time series and a realization of the bootstrap series by each of the two methods.
From Figure 3, we can see that both the KM.N and SSA.KM.N methods produce bootstrap series (in red) with almost the same pattern as the original series (in black). Even in certain parts, where the data have lower or higher values than other times as a result of the calendar variation, the SSA.KM.N can generate series closer to the original time series than the KM.N, visually. As illustrations, we zoom the load on the time period influenced by the Prophet’s birthday (Figure 4a) and Eid al-Fitr (Figure 4b). Figure 4 shows that the variance of bootstrapped series obtained by KM.N (left) is larger than those obtained by SSA.KM.N (right).
The performance of the one-step ahead forecast accuracy of SARIMA, NNAR, TBATS, and DSHW is shown in Table 2. All these calculations were performed in the R software. From the analysis of the correlogram and the partial correlogram, the model SARIMA(2,0,3)(2,1,2)24 was chosen for the first data set and the SARIMA(2,0,0)(3,1,0)24 for the second data set. The most appropriate NNAR, TBATS, and DSHW models were reconstructed and chosen automatically by the “nnetar”, “tbats”, and “dshw” functions in R. Based on Table 2, we can see that both for the first and second data sets, the NNAR and the DSHW produce smaller RMSE and MAPE than SARIMA and TBATS in the case of one-step ahead forecasting.
Furthermore, we investigate how these four models work for multistep ahead load forecasting with and without bagging implementation. The comparative values of RMSE and MAPE for 12, 24, and 36 steps ahead for SARIMA, NNAR, TBATS, and DSHW are presented in Table 3, Table 4, Table 5 and Table 6, respectively. We also present in Table 3, Table 4, Table 5 and Table 6 the RMSE and MAPE obtained from the forecast values of each model with four different numbers of bootstrap time series, to infer whether the number of bootstrap samples interferes with the accuracy of the forecasts.
Based on Table 3, we can see that the SSA.KM.N performed better than the KM.N in reducing the RMSE and MAPE of forecasts for 24- and 36 steps-ahead, respectively, obtained by the SARIMA model. The green cells in Table 3 represent the RMSE and MAPE values of the SARIMA model obtained from the bootstrap series, which are lower than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Moreover, in Table 3, it can be seen that for the first dataset, SARIMA provided high accuracy values for forecasting one day ahead (next 24 h), indicated by the MAPE values of less than two. For the second dataset, the MAPE values were between two and three. For each bagging method, there is no significant difference between the forecast results obtained by the mean and the median ensemble.
Based on the analysis for multistep-ahead forecasting by the SARIMA model with the bagging methods, it cannot be concluded that the more bootstrapped series used in the calculation, the more accurate the forecasting results will be. As we can see, the values in bold (see Table 3) are not in the n B = 100 row, being some in the n B = 25 row.
The comparative forecast results obtained by the NNAR model reconstructed from the original and bootstrap series are presented in Table 4. Based on the analysis of this table, the NNAR tends to produce larger RMSE than SARIMA. This is not in line with the results for predicting one step ahead (see Table 2). However, both the KM.N and SSA.KM.N methods can improve the accuracy performance of the forecasts. The 36-steps-ahead forecast values obtained from the bootstrap series using SSA.KM.N produced a larger RMSE than those obtained from the original time series, but the MAPE value showed the opposite direction. Similar to the results shown in Table 3, in this case, a greater number of bagging samples does not necessarily result in a better performance in terms of forecasting accuracy.
Table 5 shows that, for the first dataset, bagging did not improve the forecast accuracy obtained by the TBATS model. For the second dataset, the SSA.KM.N enhanced the performance of forecasting accuracy, but this did not apply to the KM.N. Based on the RMSE, the forecast values calculated from the TBATS model were more accurate than those obtained from the NNAR model.
Contrary to the results shown in Table 5, bagging implementation improved the forecasting accuracy of the DSHW model for the first dataset but not for the second dataset (see Table 6). In this case, the KM.N bagging performed better than the SSA.KM.N in reducing the forecasting error. Table 6 shows that the MAPE values obtained by the DSHW model for the first dataset are on average 2–3 times higher than those obtained by the DSHW model for the second dataset. However, in this case, the application of KM.N was able to reduce the MAPE value for 12-step ahead by approximately 36%.
By implementing the SSA.KM.N in the hourly load forecasting of Malaysia up to 36 steps ahead, the RMSE was able to be reduced by 4.97% and 40% when using SARIMA and NNAR, respectively. Meanwhile, KM.N was able to reduce the RMSE value by up to 3.8% for SARIMA and up to 35.43% for NNAR. Furthermore, although in one case, the SSA.KM.N bagging implementation for predicting up to 36 steps ahead using TBATS and DSHW can decrease the RMSE by more than 10%, in another case, it may behave differently. Similar conclusions were obtained when analyzing the MAPE.
For further evaluation, we consider MBE to see the direction of the models and present the results in Table 7 and Table 8. The MBE is supposed to provide information on the long-term performance of the model. Based on Table 7 and Table 8, the interpretation of the model performance is consistent with that based on RMSE and MAPE. The directions of the bias generated by the models with and without bagging are the same, except for the forecasting of 12 steps ahead by SARIMA (see Table 7). It may be related to the weakness of MBE, where the positive and negative errors can cancel each other, and high individual errors can result in low MBE values. However, we can see that bagging methods, both KM.N and SSA.KM.N, reduces MBE values obtained from the NNAR model compared with those obtained from the original time series. In the case of modeling the second data set by TBATS, SSA, KM.N yields lower MBEs than the KM.N bagging method.

3.2. Application to the Hourly Electricity Load of Poland

Figure 5 shows the hourly electricity load of Poland, in Megawatts (Mwh), from 26 October, at 01:00 to 16 December 2020 at 00:00. The data were accessed from https://www.pse.pl/obszary-dzialalnosci/krajowy-system-elektroenergetyczny/zapotrzebowanie-kse (accessed on 21 January 2021). This data set contains the linear trend and multiple seasonal patterns with daily and weekly periods. There was a slight pattern change around time index 400 (11 November 2020) due to the influence of the National Independence Day holiday (shown by the orange rectangle in Figure 5). We fit the model using the first 1212 observations and evaluated the forecasting accuracy performance using the last 36 observations.
In this experimental study, we generate 50 bootstrapped series from the original electricity load of Poland using KM.N and SSA.KM.N. We then model each generated time series by SARIMA, NNAR, TBATS, and NNAR, in the same way as for the Malaysian data. The accuracy evaluation was based on RMSE, MAPE, and MBE for 12, 24, and 36 steps ahead, and is summarized in Table 9, Table 10 and Table 11.
Based on Table 9 and Table 10, we can see that SSA.KM.N can improve the accuracy of 24 and 36 steps ahead of electricity load forecasting for the Polish data using the NNAR model, while KM.N fails to improve forecasting accuracy for this model. On the other hand, the KM.N bagging method works well on the DSHW model, while SSA.KM.N does not perform so well. However, both of them succeeded in increasing the accuracy of forecasting for the ARIMA model.
We can see from Table 10 that implementing the bagging method on Polish data does not reduce the MAPE values in the case of the TBATS model. Still, it makes the MBEs smaller (in absolute values) than those obtained from the original data (Table 11). Furthermore, although the RMSE and MAPE values of the DSHW model decreased with bagging, the results were the opposite when analyzing the MBE values. SSA.KM.N gives better outcomes for the NNAR model, while KM.N is better for the DSHW model.

3.3. Application to the Hourly Electricity Load of Java-Bali, Indonesia

To show the generality of the implementation of bagging methods in electricity load forecasting, we also discuss the hourly electricity load of Java-Bali, Indonesia. The data consists of 1464 observations, from 1 October to 30 November 2015. Figure 6 shows that the data has no trend but has double seasonal patterns. It is relatively stable except at time points around 312–336 (14 October 2015) due to the influence of the Hijriyah New Year holiday (shown by the orange rectangle in Figure 6). Moreover, this data set was also discussed in [54].
We generate 50 bootstrapped time series for this case based on 1428 observations (1 October at 01.00 to 29 November at 12.00). The error evaluation in terms of RMSE, MAPE, and MBE obtained from the SARIMA, NNAR, TBATS, and DSHW models are summarized in Table 12, Table 13 and Table 14, respectively. The overall results shown in Table 12, Table 13 and Table 14 for this application are similar to those of the previous applications for Malaysian and Polish electricity load data.
Table 13 shows that SSA.KM.N produces lower MAPE than KM.N for the NNAR model. Compared with that obtained from the original data, the MAPE of the NNAR model was able to be reduced by up to 31.38%, 24.27%, and 17% for 12, 24, and 36 steps ahead of forecast values, respectively. Meanwhile, KM.N failed to lower the MAPE value for 12 steps-ahead, and it only declined approximately 8.74% and 11% for 24 and 36 steps-ahead forecast values, respectively.
In addition, for the NNAR model, MBE presented in Table 14 shows the application of the SSA.KM.Ns bagging method provides less bias for 12 and 24 steps-ahead forecast values than without bagging. However, this does not apply to KM.N.
Based on the experimental findings of the four data sets, bagging implementation can work well to improve the forecasting accuracy of the SARIMA and NNAR models. However, the TBATS and DSHW did not yield the same behavior. The success of this implementation is thought to be influenced by the uncertainty of the models. In this experimental study, we found that some bootstrapped series failed to be modeled by TBATS and DSHW, affecting the final forecast results calculated based on the mean and median across all the forecast values.
The number of bootstrap series does not seem to affect the forecasting accuracy calculated by the mean and median ensemble. In some cases, the SSA.KM.N was able to improve the multistep-ahead forecasting accuracy, but in other cases, the KM.N provided better results.
In this case, the selection of the model is an important step to be considered. The application of bagging with the right forecasting model will increase the accuracy of multistep ahead forecast values. Further development of the hybrid model, i.e., FFORMA [55] and exponential smoothing-neural network [56] or other combinations depending on the pattern of the data, can be considered to help overcome the uncertainty of the models [30].

4. Conclusions

In this study, we evaluated the impact and benefit of applying the existing KM.N and our proposed clustering-based bootstrap method, SSA.KM.N, in overcoming the uncertainty in time series multistep-ahead point forecasts. We focused on time series with a trend, seasonal, and affected by calendar variation and considered two Malaysian, one Polish, and one Indonesian electricity load time series as illustrative examples.
KM.N is considered an appropriate method for bootstrapping data with complex seasonal patterns, such as electrical load data. In the proposed method, we combined SSA and KM.N with the hope of producing bootstrap values that are more similar to the original data. We considered the SSA method to decompose the load series into signal and noise. By SSA, the observed values influenced by the calendar variation appear more clearly in the noise component than in the original data. Bootstrapping this residual value and adding it to the signal will result in the bootstrap series values around the original data.
Furthermore, we applied the following four models, usually used as benchmark models in forecasting electricity load time series: SARIMA, NNAR, TBATS, and DSHW. These four models are applied to all bootstrapped series to obtain up to 36-steps ahead of forecast values. The final forecast at time t is obtained by the following two ensemble methods: the mean and the median across all forecast values at time t. Based on the experimental results, we note that the number of bootstrapped series does not seem to affect the forecasting accuracy calculated by the mean and the median ensemble. We also found that the model suitable for the original series is not necessarily good for all bootstrapped series. We note that the accuracy of multiple-step-ahead forecasting values can be improved when the model, with different parameters, is appropriate for both the original and bootstrap data. Thus, combining several models and ensemble learning methods can be the direction of future research.

Author Contributions

Conceptualization, methodology, analysis, writing-original draft preparation, W.S.; investigation, software, visualization, and project administration, W.S. and Y.Y.; writing-review and editing, supervision, validation, W.S. and P.C.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Education, Culture, Research, and Technology Indonesia with source from the DIPA DIKTI RISTEK (Direktorat Riset, Teknologi, dan Pengabdian Kepada Masyarakat, Direktorat Jenderal Pendidikan Tinggi, Riset, dan Teknologi) 2022, Number SP DIPA-023.17.1.690523/2022 (second revision on 22 April 2022), in the scheme of National Competitive Basic Research (Penelitian Dasar Kompetitif Nasional) with Contract Number 096/E5/PG.02.00.PT/2022.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data were obtained from [53] and are available at https://data.mendeley.com/datasets/f4fcrh4tn9/1 (accessed on 5 March 2021).

Acknowledgments

We thank the three anonymous reviewers for their valuable comments and suggestions to improve the quality of this paper. W. Sulandari and Y. Yudhanto acknowledge support from LPPM (Lembaga Penelitian dan Pengabdian kepada masyarakat) Universitas Sebelas Maret and thanks Subanar for his guidance. P.C. Rodrigues acknowledges financial support from the Brazilian national council for scientific and technological development (CNPq) grant “bolsa de produtividade PQ-2” 305852/2019-1.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

AbbreviationDefinition
DSHWDouble Seasonal Holt-Winters
KMK-means Clustering-Based
KM.NK-means Clustering-Based Generated from Gaussian Normal Distribution
MAPEMean Absolute Percentage Error
MBEMean Bias Error
MBBMoving Block Bootstrap
NNNeural Network
NNARNeural Network Autoregressive
RMSERoot Mean Square Error
SARIMASeasonal Autoregressive Integrated Moving Average
S.MBBSmoothed Moving Block Bootstrap
SSASingular Spectrum Analysis
SSA.KM.NSingular Spectrum Analysis, K-means Clustering-Based Generated from Gaussian Normal Distribution
STLSeasonal and Trend decomposition using Loess
SVDSingular Value Decomposition
TBATSTrigonometric, Box–Cox transform, ARMA errors, Trend, and Seasonal Components

References

  1. Ahmad, W.; Ayub, N.; Ali, T.; Irfan, M.; Awais, M.; Shiraz, M.; Glowacz, A. Towards Short Term Electricity Load Forecasting Using Improved Support Vector Machine and Extreme Learning Machine. Energies 2020, 13, 2907. [Google Scholar] [CrossRef]
  2. Ayub, N.; Irfan, M.; Awais, M.; Ali, U.; Ali, T.; Hamdi, M.; Alghamdi, A.; Muhammad, F. Big Data Analytics for Short and Medium-Term Electricity Load Forecasting Using an AI Techniques Ensembler. Energies 2020, 13, 5193. [Google Scholar] [CrossRef]
  3. De Livera, A.M.; Hyndman, R.J.; Snyder, R.D. Forecasting Time Series with Complex Seasonal Patterns Using Exponential Smoothing. J. Am. Stat. Assoc. 2011, 106, 1513–1527. [Google Scholar] [CrossRef]
  4. Soares, L.J.; Medeiros, M.C. Modeling and Forecasting Short-Term Electricity Load: A Comparison of Methods with an Ap-plication to Brazilian Data. Int. J. Forecast. 2008, 24, 630–644. [Google Scholar] [CrossRef]
  5. Sulandari, W.; Subanar, S.; Suhartono, S.; Utami, H.; Lee, M.H.; Rodrigues, P.C. SSA-Based Hybrid Forecasting Models and Applications. Bull. Electr. Eng. Inform. 2020, 9, 2178–2188. [Google Scholar] [CrossRef]
  6. Bernardi, M.; Petrella, L. Multiple Seasonal Cycles Forecasting Model: The Italian Electricity Demand. Stat. Methods Appl. 2015, 24, 671–695. [Google Scholar] [CrossRef]
  7. López, M.; Valero, S.; Sans, C.; Senabre, C. Use of Available Daylight to Improve Short-Term Load Forecasting Accuracy. Energies 2021, 14, 95. [Google Scholar] [CrossRef]
  8. Hong, T.; Wang, P.; White, L. Weather Station Selection for Electric Load Forecasting. Int. J. Forecast. 2015, 31, 286–295. [Google Scholar] [CrossRef]
  9. Luo, J.; Hong, T.; Fang, S.-C. Benchmarking Robustness of Load Forecasting Models under Data Integrity Attacks. Int. J. Forecast. 2018, 34, 89–104. [Google Scholar] [CrossRef]
  10. Sadaei, H.J.; de Lima e Silva, P.C.; Guimarães, F.G.; Lee, M.H. Short-Term Load Forecasting by Using a Combined Method of Convolutional Neural Networks and Fuzzy Time Series. Energy 2019, 175, 365–377. [Google Scholar] [CrossRef]
  11. Cabrera, N.G.; Gutiérrez-Alcaraz, G.; Gil, E. Load Forecasting Assessment Using SARIMA Model and Fuzzy Inductive Rea-soning. In Proceedings of the 2013 IEEE International Conference on Industrial Engineering and Engineering Management, Bangkok, Thailand, 10–13 December 2013; pp. 561–565. [Google Scholar] [CrossRef]
  12. Chikobvu, D.; Sigauke, C. Regression-SARIMA Modelling of Daily Peak Electricity Demand in South Africa. J. Energy South. Afr. 2012, 23, 23–30. [Google Scholar] [CrossRef]
  13. Taylor, J.W. Short-Term Electricity Demand Forecasting Using Double Seasonal Exponential Smoothing. J. Opera. Tional. Res. Soc. 2003, 54, 799–805. [Google Scholar] [CrossRef]
  14. Taylor, J.W. Triple Seasonal Methods for Short-Term Electricity Demand Forecasting. Eur. J. Oper. Res. 2010, 204, 139–152. [Google Scholar] [CrossRef]
  15. Arora, S.; Taylor, J.W. Short-Term Forecasting of Anomalous Load Using Rule-Based Triple Seasonal Methods. IEEE Trans. Power Syst. 2013, 28, 3235–3242. [Google Scholar] [CrossRef]
  16. Bakirtzis, A.G.; Petridis, V.; Kiartzis, S.J.; Alexiadis, M.C.; Maissis, A.H. A Neural Network Short Term Load Forecasting Model for the Greek Power System. IEEE Trans. Power Syst. 1996, 11, 858–863. [Google Scholar] [CrossRef]
  17. Charytoniuk, W.; Chen, M.-S. Very Short-Term Load Forecasting Using Artificial Neural Networks. IEEE Trans. Power Syst. 2000, 15, 263–268. [Google Scholar] [CrossRef]
  18. Mordjaoui, M.; Haddad, S.; Medoued, A.; Laouafi, A. Electric Load Forecasting by Using Dynamic Neural Network. Int. J. Hydrogen Energy 2017, 42, 17655–17663. [Google Scholar] [CrossRef]
  19. Leite Coelho da Silva, F.; da Costa, K.; Canas Rodrigues, P.; Salas, R.; López-Gonzales, J.L. Statistical and Artificial Neural Networks Models for Electricity Consumption Forecasting in the Brazilian Industrial Sector. Energies 2022, 15, 588. [Google Scholar] [CrossRef]
  20. Sulandari, W.; Subanar, S.; Suhartono, S.; Utami, H. Forecasting Time Series with Trend and Seasonal Patterns Based on SSA. In Proceedings of the 2017 3rd International Conference on Science in Information Technology (ICSITech) Theory and Applicattion of IT for Education, Industry and Society in Big Data Era, Bandung, Indonesia, 25–26 October 2017; pp. 694–699. [Google Scholar] [CrossRef]
  21. Zhang, X.; Wang, J.; Zhang, K. Short-Term Electric Load Forecasting Based on Singular Spectrum Analysis and Support Vector Machine Optimized by Cuckoo Search Algorithm. Electr. Power Syst. Res. 2017, 146, 270–285. [Google Scholar] [CrossRef]
  22. Sulandari, W.; Subanar, S.; Lee, M.H.; Rodrigues, P.C. Time Series Forecasting Using Singular Spectrum Analysis, Fuzzy Systems and Neural Networks. MethodsX 2020, 7, 101015. [Google Scholar] [CrossRef]
  23. Tao, D.; Xiuli, W.; Xifan, W. A Combined Model of Wavelet and Neural Network for Short Term Load Forecasting. In Proceedings of the International Conference on Power System Technology, Kunming, China, 13–17 October 2002; Volume 4, pp. 2331–2335. [Google Scholar] [CrossRef]
  24. Moazzami, M.; Khodabakhshian, A.; Hooshmand, R. A New Hybrid Day-Ahead Peak Load Forecasting Method for Iran’s National Grid. Appl. Energy 2013, 101, 489–501. [Google Scholar] [CrossRef]
  25. Pandian, S.C.; Duraiswamy, K.; Rajan, C.C.A.; Kanagaraj, N. Fuzzy Approach for Short Term Load Forecasting. Electr. Power Syst. Res. 2006, 76, 541–548. [Google Scholar] [CrossRef]
  26. Al-Kandari, A.M.; Soliman, S.A.; El-Hawary, M.E. Fuzzy Short-Term Electric Load Forecasting. Int. J. Electr. Trical. Power Energy Syst. 2004, 26, 111–122. [Google Scholar] [CrossRef]
  27. Chen, Y.; Yang, Y.; Liu, C.; Li, C.; Li, L. A Hybrid Application Algorithm Based on the Support Vector Machine and Artificial Intelligence: An Example of Electric Load Forecasting. Appl. Math. Model. 2015, 39, 2617–2632. [Google Scholar] [CrossRef]
  28. Li, Y.; Fang, T.; Yu, E. Study of Support Vector Machines for Short-Term Load Forecasting. Proc. CSEE 2003, 23, 55–59. [Google Scholar]
  29. Breiman, L. Bagging Predictors | SpringerLink. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  30. Bergmeir, C.; Hyndman, R.J.; Benítez, J.M. Bagging Exponential Smoothing Methods Using STL Decomposition and Box–Cox Transformation. Int. J. Forecast. 2016, 32, 303–312. [Google Scholar] [CrossRef]
  31. Petropoulos, F.; Hyndman, R.J.; Bergmeir, C. Exploring the Sources of Uncertainty: Why Does Bagging for Time Series Fore-casting Work? Eur. J. Oper. Res. 2018, 268, 545–554. [Google Scholar] [CrossRef]
  32. Laurinec, P.; Lóderer, M.; Lucká, M.; Rozinajová, V. Density-Based Unsupervised Ensemble Learning Methods for Time Series Forecasting of Aggregated or Clustered Electricity Consumption. J. Intell. Inf. Syst. 2019, 53, 219–239. [Google Scholar] [CrossRef]
  33. Cleveland, R.B.; Cleveland, W.S.; Terpenning, I. STL: A Seasonal-Trend Decomposition Procedure Based on Loess. J. Off. Stat. 1990, 6, 3. [Google Scholar]
  34. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  35. Wei, W.W.-S. Time Series Analysis: Univariate and Multivariate Methods, 2nd ed.; Pearson Addison-Wesley: Boston, MA, USA, 2006. [Google Scholar]
  36. Huang, S.-J.; Shih, K.-R. Short-Term Load Forecasting via ARMA Model Identification Including Non-Gaussian Process Con-siderations. IEEE Trans. Power Syst. 2003, 18, 673–679. [Google Scholar] [CrossRef]
  37. Hyndman, R.J.; Khandakar, Y. Automatic Time Series Forecasting: The Forecast Package for R. JSS 2008, 27, 1–23. [Google Scholar] [CrossRef]
  38. Mohamed, N.; Ahmad, M.H.; Ismail, Z. Improving Short Term Load Forecasting Using Double Seasonal Arima Model. World Appl. Sci. J. 2011, 15, 223–231. [Google Scholar]
  39. Hyndman, R.J.; Athanasopoulos, G.; Bergmeir, C.; Caceres, G.; Chhay, L.; O’Hara-Wild, M.; Petropoulos, F.; Razbash, S.; Wang, E.; Yasmeen, F. Forecast: Forecasting Functions for Time Series and Linear Models. In R Package Version 8.15. 2021. Available online: https://pkg.robjhyndman.com/forecast/ (accessed on 18 December 2021).
  40. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice, 2nd ed.; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  41. Zhang, G.; Patuwo, B.E.; Hu, M.Y. Forecasting with Artificial Neural Networks: The State of the Art. Int. J. Forecast. 1998, 14, 35–62. [Google Scholar] [CrossRef]
  42. Sulandari, W.; Suhartono; Subanar; Rodrigues, P.C. Exponential Smoothing on Modeling and Forecasting Multiple Seasonal Time Series: An Overview. Fluct. Noise Lett. 2021, 20, 2130003. [Google Scholar] [CrossRef]
  43. Sulandari, W.; Subanar, S.; Suhartono, S.; Utami, H. Forecasting Electricity Load Demand Using Hybrid Exponential Smoothing-Artificial Neural Network Model. Int. J. Adv. Intell. Inform. 2016, 2, 131–139. [Google Scholar] [CrossRef]
  44. Hyndman, R.; Koehler, A.B.; Ord, J.K.; Snyder, R.D. Forecasting with Exponential Smoothing: The State Space Approach; Springer Science & Business Media: Berlin, Germany, 2008. [Google Scholar]
  45. Golyandina, N.; Osipov, E. The “Caterpillar”-SSA Method for Analysis of Time Series with Missing Values. J. Stat. Plan. Inference 2007, 137, 2642–2653. [Google Scholar] [CrossRef]
  46. Golyandina, N.; Korobeynikov, A. Basic Singular Spectrum Analysis and Forecasting with R. Comput. Stat. Data Anal. 2014, 71, 934–954. [Google Scholar] [CrossRef]
  47. Golyandina, N.; Zhigljavsky, A. Singular Spectrum Analysis for Time Series, 2nd ed.; Springer Briefs in Statistics; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar] [CrossRef]
  48. Golyandina, N.; Korobeynikov, A.; Zhigljavsky, A. Singular Spectrum Analysis with R; Springer: Berlin, Germany, 2018. [Google Scholar]
  49. PetoLau/petolau.github.io. GitHub. Available online: https://github.com/PetoLau/petolau.github.io (accessed on 19 April 2021).
  50. Hassani, H. Singular Spectrum Analysis: Methodology and Comparison. J. Data Sci. 2007, 5, 239–257. [Google Scholar] [CrossRef]
  51. Hyndman, R.J.; Koehler, A.B. Another Look at Measures of Forecast Accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef]
  52. Stone, R.J. Improved Statistical Procedure for the Evaluation of Solar Radiation Estimation Models. Sol. Energy 1993, 51, 289–291. [Google Scholar] [CrossRef]
  53. Guimaraes, F.; Javedani Sadaei, H. Data for: Short-Term Load Forecasting by Using a Combined Method of Convolutional Neural Networks and Fuzzy Time Series. Mendeley Data 2019, Version 1. Available online: https://data.mendeley.com/datasets/f4fcrh4tn9/1 (accessed on 5 March 2021).
  54. Sulandari, W.; Subanar; Lee, M.H.; Rodrigues, P.C. Indonesian Electricity Load Forecasting Using Singular Spectrum Analysis, Fuzzy Systems and Neural Networks. Energy 2020, 190, 116408. [Google Scholar] [CrossRef]
  55. Montero-Manso, P.; Athanasopoulos, G.; Hyndman, R.J.; Talagala, T.S. FFORMA: Feature-Based Forecast Model Averaging. Int. J. Forecast. 2020, 36, 86–92. [Google Scholar] [CrossRef]
  56. Smyl, S. A Hybrid Method of Exponential Smoothing and Recurrent Neural Networks for Time Series Forecasting. Int. J. Forecast. 2020, 36, 75–85. [Google Scholar] [CrossRef]
Figure 1. Procedure for generating bootstrapped time series by the SSA-KM.N method.
Figure 1. Procedure for generating bootstrapped time series by the SSA-KM.N method.
Energies 15 05838 g001
Figure 2. Hourly load series of Johor, Malaysia: (a) 1 January time 00:00, to 30 November 2009, time 23:00; (b) 1 January, time 00:00, to 30 June 2010, time 23:00.
Figure 2. Hourly load series of Johor, Malaysia: (a) 1 January time 00:00, to 30 November 2009, time 23:00; (b) 1 January, time 00:00, to 30 June 2010, time 23:00.
Energies 15 05838 g002
Figure 3. The original series (in black) and the bootstrap time series (in red) obtained by the (a) KM.N method; (b) SSA.KM.N method.
Figure 3. The original series (in black) and the bootstrap time series (in red) obtained by the (a) KM.N method; (b) SSA.KM.N method.
Energies 15 05838 g003
Figure 4. The original series (in black) and the bootstrap time series (in red) obtained by the KM.N method (left) and SSA.KM.N (right) (a) for time period influenced by the Prophet’s birthday; (b) for time period influenced by the Aid al-Fitr.
Figure 4. The original series (in black) and the bootstrap time series (in red) obtained by the KM.N method (left) and SSA.KM.N (right) (a) for time period influenced by the Prophet’s birthday; (b) for time period influenced by the Aid al-Fitr.
Energies 15 05838 g004
Figure 5. The hourly electricity load of Poland between 26 October and 16 December 2020.
Figure 5. The hourly electricity load of Poland between 26 October and 16 December 2020.
Energies 15 05838 g005
Figure 6. The hourly electricity load of Indonesia between 1 October and 30 November 2015.
Figure 6. The hourly electricity load of Indonesia between 1 October and 30 November 2015.
Energies 15 05838 g006
Table 1. Training and testing datasets of hourly load series used in the experimental study.
Table 1. Training and testing datasets of hourly load series used in the experimental study.
1st Data Set2nd Data Set
Training01/01/2009 00:00–30/11/2009 23:0001/01/2010 00:00–30/06/2010 23:00
Testing (for up to h-step ahead):
h = 101/12/2009 00:00–31/12/2009 23:0001/07/2010 00:00–31/07/2010 23:00
h = 1201/12/2009 00:00–01/12/2009 11:0001/07/2010 00:00–01/07/2020 11:00
h = 2401/12/2009 00:00–01/12/2009 23:0001/07/2010 00:00–01/07/2020 23:00
h = 3601/12/2009 00:00–01/12/2009 11:0001/07/2010 00:00–02/07/2020 11:00
Table 2. RMSE and MAPE for one-step-ahead forecasts for the testing data of the two data sets of hourly electricity load in Malaysia, obtained by SARIMA, NNAR, TBATS, and DSHW.
Table 2. RMSE and MAPE for one-step-ahead forecasts for the testing data of the two data sets of hourly electricity load in Malaysia, obtained by SARIMA, NNAR, TBATS, and DSHW.
MethodTesting of 1st Data Set
(1–31 December 2009)
Testing of 2nd Data Set
(1–31 July 2010)
RMSEMAPERMSEMAPE
SARIMA1168.021.901719.682.54
NNAR749.921.211182.681.75
TBATS1584.472.911466.332.37
DSHW839.181.291035.141.46
Table 3. RMSE and MAPE of h -step ahead forecast obtained by SARIMA model from the original series, KM.N and SSA.KM.N bootstrap series.
Table 3. RMSE and MAPE of h -step ahead forecast obtained by SARIMA model from the original series, KM.N and SSA.KM.N bootstrap series.
1st Data Set2nd Data Set
RMSEMAPERMSEMAPE
h 122436122436122436122436
original466.511408.202118.070.881.943.131139.081268.581324.192.272.482.97
Bagging + Mean
KM.N
n B = 25512.571404.682035.790.971.943.001345.291405.541366.412.252.572.87
n B = 50530.991419.452049.640.991.963.031320.971393.691366.422.282.592.91
n B = 75544.231429.022053.971.021.993.051323.971394.821367.792.292.602.92
n B = 100556.731438.172062.041.092.033.081329.881397.271371.852.322.612.93
SSA. KM.N
n B = 25490.521368.942029.141.041.942.961067.121234.851286.552.112.402.87
n B = 50494.711357.472018.931.051.942.951069.561235.921288.942.132.412.88
n B = 75498.231356.262015.311.071.942.951068.791235.361288.802.132.412.88
n B = 100495.761360.242018.481.061.952.951072.371236.661289.222.142.412.88
Bagging + Median
KM.N
n B = 25548.321415.382045.611.031.983.041330.611388.671348.252.122.492.80
n B = 50543.851424.972055.741.021.983.051306.421383.851355.372.182.532.86
n B = 75549.991430.502056.881.052.003.061322.841390.971360.402.222.552.88
n B = 100566.061444.642066.111.092.043.091325.921392.351365.122.272.582.90
SSA.KM.N
n B = 25501.401372.292029.041.101.982.981072.911239.461290.892.132.412.88
n B = 50514.831355.262012.781.131.972.951076.471239.431292.002.152.422.89
n B = 75514.081355.952011.371.131.972.951069.091234.931289.342.132.412.88
n B = 100508.081365.382018.371.111.972.961074.031238.001290.432.152.422.88
Green cells represent the RMSE and MAPE values of the SARIMA model obtained from the bootstrap series lower than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Table 4. RMSE and MAPE of h -step ahead forecast obtained by NNAR model from the original series, KM.N and SSA.KM.N bootstrap series.
Table 4. RMSE and MAPE of h -step ahead forecast obtained by NNAR model from the original series, KM.N and SSA.KM.N bootstrap series.
1st Data Set2nd Data Set
RMSEMAPERMSEMAPE
h 122436122436122436122436
Original1761.722365.264393.212.123.185.812474.072797.865179.863.624.246.15
Bagging + Mean
KM.N
n B = 25960.091723.562881.121.102.354.012367.772062.454128.653.503.325.19
n B = 50964.931730.892894.541.122.374.032433.152096.234154.823.583.345.22
n B = 75996.611747.332949.971.162.394.082358.622044.604084.193.483.285.13
n B = 1001007.271766.482972.191.162.424.122388.402067.724120.693.513.305.17
SSA. KM.N
n B = 25973.691661.792756.241.112.293.932279.291958.575344.053.413.156.04
n B = 50954.971618.732701.081.112.233.842349.541997.225451.243.523.186.14
n B = 75971.001634.742732.761.112.243.862325.341980.905443.803.473.166.13
n B = 100968.261628.322716.571.112.243.852311.581974.445452.813.453.166.13
Bagging + Median
KM.N
n B = 25913.181695.232836.711.072.343.962364.002069.784029.043.523.355.14
n B = 50912.721690.942861.261.052.323.982469.392110.044136.533.633.365.21
n B = 75948.081697.712883.831.132.354.012399.132068.574035.983.533.315.11
n B = 100964.861713.662892.471.122.364.022418.192080.954088.093.553.325.15
SSA.KM.N
n B = 25978.421654.412785.531.132.283.952248.301952.465361.383.373.146.05
n B = 50979.691633.472657.161.132.263.812270.891954.885411.123.393.146.09
n B = 75975.651642.362683.661.122.263.832315.221980.655475.963.443.166.15
n B = 100959.001612.462631.461.132.233.782307.141977.875438.963.433.156.12
Green cells represent the RMSE and MAPE values of the SARIMA model obtained from the bootstrap series lower than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Table 5. RMSE and MAPE of h -step ahead forecast obtained by TBATS model from the original series, KM.N and SSA.KM.N bootstrap series.
Table 5. RMSE and MAPE of h -step ahead forecast obtained by TBATS model from the original series, KM.N and SSA.KM.N bootstrap series.
1st Data Set2nd Data Set
RMSEMAPERMSEMAPE
h 122436122436122436122436
original1613.362057.212146.924.033.724.081574.782771.372685.713.875.545.54
Bagging + Mean
KM.N
n B = 251740.642333.252360.333.993.983.911410.823122.422864.933.505.935.86
n B = 501668.462245.732282.533.893.863.841296.393045.172805.203.185.675.63
n B = 751664.152246.182290.123.863.853.841280.053047.592806.133.145.675.62
n B = 1001732.412325.222370.243.913.933.901255.273002.582775.603.105.595.55
SSA. KM.N
n B = 251966.142547.932560.874.614.444.261371.731974.322286.353.434.034.61
n B = 502023.972609.352634.364.604.484.281447.051939.142316.253.513.994.68
n B = 752011.412580.082615.164.604.454.281503.401861.192247.763.563.874.61
n B = 1001994.502572.822597.214.594.454.271468.421879.362259.163.523.904.60
Bagging + Median
KM.N
n B = 251937.562611.752574.414.194.274.161336.973256.252953.003.306.035.91
n B = 501719.882336.302337.813.863.903.871297.493240.042933.743.195.965.83
n B = 751818.782406.982410.714.014.034.001297.063239.032934.093.195.965.83
n B = 1002010.252642.922639.854.194.314.211287.173203.362909.663.165.905.78
SSA.KM.N
n B = 251980.322604.852589.404.564.464.161594.162542.152846.893.995.135.72
n B = 502021.702674.562675.424.514.504.191482.302157.742548.263.714.465.16
n B = 752043.602661.862678.064.594.534.271476.012070.702462.003.654.285.01
n B = 1002040.192629.412651.924.594.504.251415.002012.092390.523.544.144.88
Green cells represent the RMSE and MAPE values of the SARIMA model obtained from the bootstrap series lower than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Table 6. RMSE and MAPE of h -step ahead forecast obtained by DSHW model from the original series, KM.N and SSA.KM.N bootstrap series.
Table 6. RMSE and MAPE of h -step ahead forecast obtained by DSHW model from the original series, KM.N and SSA.KM.N bootstrap series.
1st Data Set2nd Data Set
RMSEMAPERMSEMAPE
h 122436122436122436122436
original3448.134802.004207.314.196.976.391458.591231.291148.922.772.052.00
Bagging + Mean
KM.N
n B = 252183.443644.903236.112.675.395.321809.371902.751718.993.293.062.75
n B = 502230.953640.883233.822.835.465.401890.922012.371801.603.433.242.89
n B = 752194.683590.453183.482.815.405.341918.152065.241842.823.473.332.96
n B = 1002268.223660.823241.712.915.505.401913.902052.311828.073.493.322.96
SSA. KM.N
n B = 253168.624611.094018.483.766.625.992361.262782.862349.973.684.333.82
n B = 503157.424596.484006.133.746.605.972552.002945.082475.154.064.563.95
n B = 753156.844598.524007.593.746.605.972544.552946.872487.054.034.594.05
n B = 1003153.794597.774005.713.736.605.972521.972911.702455.594.024.553.99
Bagging + Median
KM.N
n B = 252556.484295.013778.133.186.275.931860.061967.001771.433.333.122.79
n B = 502542.874273.273758.953.126.235.901922.972026.091817.493.473.232.88
n B = 752574.914263.893750.113.156.215.881934.912060.741843.214.483.302.93
n B = 1002658.874320.543801.113.216.285.941935.222046.941830.223.523.292.92
SSA.KM.N
n B = 253184.944621.304025.463.776.635.992372.522855.732427.663.584.504.06
n B = 503183.664607.604014.753.776.625.982602.982938.162472.204.114.463.93
n B = 753178.474606.994014.943.766.625.982517.252841.472404.814.014.383.91
n B = 1003184.524610.004016.523.776.625.982521.242812.682374.884.034.333.84
Green cells represent the RMSE and MAPE values of the SARIMA model obtained from the bootstrap series lower than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Table 7. MBEs of h -step ahead forecast for the first data set obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 7. MBEs of h -step ahead forecast for the first data set obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original145.99907.891446.04751.371574.982308.25−69.78598.89607.952233.793829.693311.54
Bagging + Mean
KM.N
n B = 25167.58905.801390.65443.291232.641968.24290.351200.321246.451356.392890.412618.15
n B = 50214.87937.028881.84441.631236.061269.93384.091287.13470.021447.272924.451012.76
n B = 75230.20950.191425.00460.411246.862003.83371.321252.581318.351436.042891.672610.35
n B = 100257.88969.211441.16473.651266.202025.98348.741242.511298.321490.212951.162654.25
SSA. KM.N
n B = 25−137.62728.931264.851411.071170.291895.15269.681112.431132.331921.133601.543093.45
n B = 50−150.17714.81869.88386.291130.321210.20221.891042.36334.781919.843593.181010.49
n B = 75−164.26705.981243.53402.641142.611863.60238.151053.851069.631919.893595.023087.34
n B = 100−154.71714.481250.64391.241133.871851.27323.681139.931162.731913.613591.933083.82
Bagging + Median
KM.N
n B = 25198.26925.211407.78413.571213.511938.24447.491340.141307.191533.133350.102965.00
n B = 50230.70946.821426.14413.631209.421948.60297.711130.411117.381553.863350.172963.54
n B = 75234.08951.991428.00420.361207.341954.58375.901203.431178.551593.233355.652964.73
n B = 100260.12972.471444.30440.881227.321970.68558.251415.631394.911645.623403.653003.21
SSA.KM.N
n B = 25−152.57722.761258.23387.981150.801894.71343.251289.541343.091932.043610.623097.90
n B = 50−172.75699.001237.00378.601130.511820.84431.021376.171441.361935.883603.033092.51
n B = 75−177.18696.201233.93382.151136.421835.54421.091348.281416.421934.803603.843094.52
n B = 100−168.17708.371244.94367.051113.811801.46413.721325.611396.901939.263606.193095.18
Green cells represent the MBE values obtained from the bootstrap series for the first data set lower (in absolute values) than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Table 8. MBEs of h -step ahead forecast for the second data set obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 8. MBEs of h -step ahead forecast for the second data set obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original287.06−228.44−591.452319.171146.122903.47−1266.71−2379.66−1457.27−1178.78−805.60−534.08
Bagging + Mean
KM.N
n B = 25616.40−3.69−389.821733.21472.401818.49−1191.85−2641.61−1815.55−1496.71−1538.48−935.39
n B = 50568.38−38.61−582.741784.60530.701375.18−1071.33−2541.56−431.21−1568.82−1638.4010.61
n B = 75566.62−40.03−423.901729.60476.551807.17−1041.47−2524.44−1688.93−1591.59−1687.65−1044.28
n B = 100559.38−43.89−428.301744.63493.211827.51−1009.37−2480.44−1640.25−1592.14−1679.05−1047.71
SSA. KM.N
n B = 25291.41−224.24−571.591672.25481.192374.84−97.56−869.941−314.26−1783.46−2256.37−1713.42
n B = 50283.98−225.54−606.411728.23557.901976.58−149.71−914.66−18.19−1952.89−2375.25−335.45
n B = 75283.69−226.25−574.951708.84537.572457.88−125.375−832.99−348.08−1943.21−2389.85−1849.06
n B = 100285.39−222.95−571.831695.97518.812448.55−209.48−940.32−418.47−1931.83−2363.00−1804.90
Bagging + Median
KM.N
n B = 25653.5510.43−376.271730.47422.541737.09−1138.90−2723.60−1888.97−1529.41−1578.47−957.24
n B = 50600.71−16.70−403.81811.03532.831586.62−1102.34−2699.53−1860.87−1591.57−1636.36−995.47
n B = 75604.67−19.03−405.871751.48487.221788.34−1095.80−2690.14−1849.37−1600.02−1674.19−1024.61
n B = 100578.09−35.40−419.541763.27504.381817.09−1083.54−2654.90−1816.41−1609.89−1663.54−1023.59
SSA.KM.N
n B = 25289.37−221.65−571.521652.67470.112375.43−3.50−755.85−200.04−1746.97−2341.34−1867.79
n B = 50285.59−223.41−574.001668.23499.542419.34−227.46−1074.75−432.69−1977.57−2334.53−1784.67
n B = 75283.23−224.74−575.071694.84517.582458.64−267.19−1001.17−421.91−1921.88−2271.71−1797.42
n B = 100284.76−221.12−571.491690.07514.032440.18−350.40−1103.52−506.35−1925.08−2246.15−1737.53
Green cells represent the MBE values obtained from the bootstrap series for the first data set lower (in absolute values) than those obtained from the original series. Bold values represent the lowest value in a column of each bagging method with green cells.
Table 9. RMSEs of h -step ahead forecast for the hourly electricity load of Poland obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 9. RMSEs of h -step ahead forecast for the hourly electricity load of Poland obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
Original2120.282436.072540.46107.18308.45382.81259.27437.82466.29351.51289.03284.54
Bagging + Mean
KM.N
n B = 10298.01254.32229.17133.53343.25416.31309.15456.14468.23323.29258.16254.41
n B = 20276.62237.69220.48135.79319.32390.97307.57457.72467.18326.77255.35255.16
n B = 25266.20231,54218.07132.76305.93371.25308.52459.61469.26305.33240.89240.29
n B = 50300.58257.61230.87146.32316.44383.40304.79455.58467.42318.51246.98243.43
SSA. KM.N
n B = 10361.10304.85256.78130.42254.16290.95291.24433.47441.28365.75298.55291.71
n B = 20356.84302.42254.80128.66228.89262.25296.73439.51444.58369.77298.33291.90
n B = 25353.73299.78252.83132.14234.19268.70296.91442.01447.02370.41300.02291.14
n B = 50353.14299.24252.50131.97231.66267.86299.30435.43437.14374.79300.84290.49
Bagging + Median
KM.N
n B = 10323.16270.39243.86127.12330.51401.92319.28463.28470.74324.95261.77253.61
n B = 20287.77243.37222.76135.81312.82385.15312.11461.49470.19330.34257.77256.02
n B = 25265.14227.19213.62136.70298.21370.86312.14461.90470.74316.96249.81245.81
n B = 50302.26257.68228.44150.87313.86381.42310.11460.61470.76336.98261.51257.76
SSA.KM.N
n B = 10364.58306.80258.27137.93286.20326.78296.29448.13448.54366.90303.00297.67
n B = 20362.98307.99259.25132.48245.93287.50297.94449.62451.81376.13303.83298.76
n B = 25364.04308.75260.17137.58247.58284.35300.09453.73456.32373.93304.04295.95
n B = 50306.64258.381.44133.83232.00267.26299.78450.28450.92378.22305.12294.39
Green cells represent the RMSE values obtained from the bootstrap series for the Polish data set lower than those obtained from the original series.
Table 10. MAPEs of h -step ahead forecast for the hourly electricity load of Poland obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 10. MAPEs of h -step ahead forecast for the hourly electricity load of Poland obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original7.099.9910.200.360.841.191.011.521.591.341.131.11
Bagging + Mean
KM.N
n B = 101.140.990.860.420.981.291.151.891.841.230.980.97
n B = 201.020.900.820.430.911.211.131.891.841.190.910.93
n B = 250.970.880.810.410.861.401.131.901.851.130.870.89
n B = 501.120.990.860.480.921.201.141.881.841.140.860.87
SSA. KM.N
n B = 101.441.220.860.420.750.911.101.761.711.371.161.13
n B = 201.421.210.860.420.690.831.101.801.731.381.151.12
n B = 251.411.210.860.430.710.841.101.811.751.381.161.12
n B = 501.411.200.860.430.700.841.101.791.711.391.151.11
Bagging + Median
KM.N
n B = 101.221.050.910.390.931.241.181.931.871.221.010.98
n B = 201.060.920.830.4110.901.201.151.911.861.180.900.92
n B = 250.960.860.790.410.841.141.151.911.861.140.890.90
n B = 501.120.980.840.490.931.201.161.911.861.200.910.92
SSA.KM.N
n B = 101.451.230.960.460.831.011.101.841.751.381.191.17
n B = 201.441.240.960.440.740.901.091.851.771.401.171.15
n B = 251.451.240.970.460.740.891.091.871.791.391.171.14
n B = 501.441.220.860.440.700.841.091.851.761.401.181.13
Green cells represent the MAPE values obtained from the bootstrap series for the Polish data set lower than those obtained from the original series.
Table 11. MBEs of h -step ahead forecast for the hourly electricity load of Poland obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 11. MBEs of h -step ahead forecast for the hourly electricity load of Poland obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original1561.97411.79962.1286.73198.18286.42−117.2679.96170.71135.940.5140.35
Bagging + Mean
KM.N
n B = 10273.31174.8376.82104.49227.21309.50−120.87−39.7166.39154.1844.3267.89
n B = 20244.56147.7453.40104.45211.54289.94−129.99−49.5753.46187.9284.30107.71
n B = 25232.10136.4642.7097.04198.57270.37−132.55−51.0153.80168.6672.6095.20
n B = 50268.19174.3781.24118.23215.67287.11−118.26−35.5366.27195.4892.75110.85
SSA. KM.N
n B = 10344.43247.28163.0399.88163.18202.15−150.01−41.6445.34171.1815.6958.17
n B = 20340.06234.96158.87101.06147.35178.70−150.37−52.8833.91178.9823.5065.33
n B = 25337.43241.26155.77105.18152.51183.96−147.70−54.2133.98179.5623.7763.67
n B = 50336.97240.48154.86104.43150.84183.92−156.03−64.7116.50189.3625.2065.02
Bagging + Median
KM.N
n B = 10290.28194.9291.5792.97214.79−288.01−136.15−61.4147.86140.7333.8350.60
n B = 20253.80158.1966.38100.62207.48−228.01−131.90−56.8148.96195.8591.18109.55
n B = 25231.23138.5614.3497.07193.80−228.01−129.80−53.9750.75180.3176.9996.29
n B = 50267.03173.0285.13121.50216.16−228.01−120.45−44.3059.69206.92101.26120.60
SSA.KM.N
n B = 10347.60248.26162.68111.64189.09231.51−146.64−61.8528.45154.73−3.7745.28
n B = 20346.04249.75164.58107.36160.93201.89−145.21−62.1230.76174.0915.2561.85
n B = 25347.75250.10166.04112.56162.74198.52−147.18−67.2727.96173.7418.0459.33
n B = 50345.41248.56162.74106.97151.23182.94−151.43−72.1118.52187.0219.2761.72
Green cells represent the MBE values obtained from the bootstrap series for the Polish data set lower (in absolute values) than those obtained from the original series.
Table 12. RMSE of h -step ahead forecast for the hourly electricity load of Indonesia obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 12. RMSE of h -step ahead forecast for the hourly electricity load of Indonesia obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original1236.722440.283119.10193.12216.95234.57611.23702.52609.34167.51265.62351.18
Bagging + Mean
KM.N
n B = 10140.46148.61182.96192.32200.15211.26581.86681.02593.49138.18231.80339.36
n B = 20137.65148.79182.42201.59201.79207.81581.67678.01592.20137.31224.45318.91
n B = 25136.02149.93184.64197.91198.36204.31581.49677.87591.87139.20227.65327.53
n B = 50136.02149.93184.64194.08200.23208.27580.00677.27590.68147.81245.81362.49
SSA. KM.N
n B = 10181.30178.87194.99136.43177.76208.49605.75704.10608.36170.08270.61364.78
n B = 20186.56184.79199.14135.42177.59208.45600.50700.37605.79167.54277.51385.43
n B = 25185.58183.79198.45134.86178.64209.78597.76698.29604.43167.74277.13383.10
n B = 50193.68189.13201.59146.47184.91213.14602.30699.76605.76150.49260.36370.83
Bagging + Median
KM.N
n B = 10139.51144.76183.00190.81199.99210.46581.67680.87593.92153.44245.70352.49
n B = 20138.10146.75183.39198.30200.90205.40581.72680.30593.77162.20247.44344.32
n B = 25139.17150.71187.80195.26196.90202.43584.21681.03594.27160.89242.23344.98
n B = 50136.72146.71184.83192.91199.18206.05585.00682.61595.21161.88258.46372.81
SSA.KM.N
n B = 10180.44178.15194.23135.73174.27206.11605.40700.65606.85166.57268.12358.32
n B = 20187.25184.52198.52132.52176.76207.01604.35700.16606.58168.90267.43360.99
n B = 25186.72183.84198.25135.54178.50208.79598.77696.99604.16168.29268.13366.47
n B = 50193.01188.59200.85149.27186.48213.25607.15701.46607.49155.88258.63359.15
Green cells represent the RMSE values obtained from the bootstrap series for the Indonesian data set lower than those obtained from the original series.
Table 13. MAPE of h -step ahead forecast for the hourly electricity load of Indonesia obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 13. MAPE of h -step ahead forecast for the hourly electricity load of Indonesia obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original5.8310.8713.450.951.031.002.783.252.590.771.101.40
Bagging + Mean
KM.N
n B = 100.670.700.760.960.960.922.673.152.530.640.921.29
n B = 200.650.710.761.010.980.922.663.152.530.610.901.22
n B = 250.630.710.770.990.960.912.663.142.530.630.911.25
n B = 500.630.710.770.970.960.912.653.142.530.680.961.37
SSA. KM.N
n B = 100.870.860.850.660.780.842.763.242.580.791.131.45
n B = 200.910.900.880.660.780.832.733.232.570.781.121.51
n B = 250.900.890.880.660.780.842.713.222.560.791.121.50
n B = 500.940.920.890.730.830.872.733.222.570.681.031.43
Bagging + Median
KM.N
n B = 100.670.690.750.950.960.922.663.152.540.710.981.35
n B = 200.650.700.760.990.970.912.663.262.540.751.001.33
n B = 250.660.720.780.970.940.892.683.162.540.740.971.32
n B = 500.650.710.760.960.950.902.683.172.550.761.021.43
SSA.KM.N
n B = 100.870.860.850.670.780.832.733.242.580.771.121.43
n B = 200.910.900.880.650.780.832.753.232.580.761.111.43
n B = 250.900.890.870.670.790.842.723.222.570.771.101.44
n B = 500.950.890.920.750.840.882.763.232.580.701.061.41
Green cells represent the MAPE values obtained from the bootstrap series for the Indonesian data set lower than those obtained from the original series.
Table 14. MBE of h -step ahead forecast for the hourly electricity load of Indonesia obtained by SARIMA, NNAR, TBATS, and DSHW models.
Table 14. MBE of h -step ahead forecast for the hourly electricity load of Indonesia obtained by SARIMA, NNAR, TBATS, and DSHW models.
SARIMANNARTBATSDSHW
h 122436122436122436122436
original760.771914.502658.76−171.70−73.39−7.27−247.66−394.84−210.2813.7753.09184.57
Bagging + Mean
KM.N
n B = 10−110.53−107.27−74.87−176.59−90.30−17.23−196.70−361.33−186.3430.9863.46194.51
n B = 20−107.48−109.13−77.58−185.87−104.32−29.48−192.40−354.91−178.9214.4146.96170.86
n B = 25−104.71−110.04−79.76−181.50−104.71−30.37−192.71−355.37−179.7125.9158.99184.20
n B = 50−104.71−110.04−79.76−176.85−99.07−24.80−199.37−358.92−184.3052.4389.04224.39
SSA. KM.N
n B = 10−157.22−147.24−104.51−121.14−33.9026.25−242.95−399.82−220.519.2953.06193.08
n B = 20−163.43−154.65−111.92−120.07−32.9727.19−237.07−394.13−214.8812.7868.00216.31
n B = 25−162.16−153.24−110.25−119.85−32.3128.96−232.08−390.43−211.7914.1068.99215.21
n B = 50−170.78−158.66−113.53−133.25−41.0322.80−241.80−395.38−215.158.5560.63206.89
Bagging + Median
KM.N
n B = 10−109.90−105.26−73.05−174.03−91.36−18.07−197.11−361.31−185.1036.1172.84204.91
n B = 20−109.00−107.40−75.75−182.41−105.51−31.60−196.47−359.60−182.8029.8362.95191.74
n B = 25−108.90−111.56−80.54−177.65−105.33−31.90−199.31−360.76−183.5432.2271.68199.89
n B = 50−105.43−105.82−73.99−174.42−100.32−26.06−203.93−364.34−187.1357.5999.10234.13
SSA.KM.N
n B = 10−156.02−146.64−104.14−119.95−35.5925.79−239.17−392.13−211.835.8846.37185.00
n B = 20−164.52−153.32−109.83−118.08−32.9927.89−239.19−391.83−211.09−1.9746.70186.84
n B = 25−162.60−152.57−109.06−121.50−34.0228.08−231.34−387.20−207.831.6052.27193.65
n B = 50−171.08−158.53−113.30−136.04−42.5421.33−242.47−393.92−212.834.1948.66190.21
Green cells represent the MBE values obtained from the bootstrap series for the Indonesian data set lower (in absolute values) than those obtained from the original series.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sulandari, W.; Yudhanto, Y.; Rodrigues, P.C. The Use of Singular Spectrum Analysis and K-Means Clustering-Based Bootstrap to Improve Multistep Ahead Load Forecasting. Energies 2022, 15, 5838. https://doi.org/10.3390/en15165838

AMA Style

Sulandari W, Yudhanto Y, Rodrigues PC. The Use of Singular Spectrum Analysis and K-Means Clustering-Based Bootstrap to Improve Multistep Ahead Load Forecasting. Energies. 2022; 15(16):5838. https://doi.org/10.3390/en15165838

Chicago/Turabian Style

Sulandari, Winita, Yudho Yudhanto, and Paulo Canas Rodrigues. 2022. "The Use of Singular Spectrum Analysis and K-Means Clustering-Based Bootstrap to Improve Multistep Ahead Load Forecasting" Energies 15, no. 16: 5838. https://doi.org/10.3390/en15165838

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop