Next Article in Journal
Geothermal Boreholes in Poland—Overview of the Current State of Knowledge
Previous Article in Journal
Influence of Debris on Water Intake Gratings in Small Hydroelectric Plants: An Experimental Study on Hydraulic Parameters
Previous Article in Special Issue
A Single Scalable LSTM Model for Short-Term Forecasting of Massive Electricity Time Series
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Importance of the Long-Term Seasonal Component in Day-Ahead Electricity Price Forecasting Revisited: Parameter-Rich Models Estimated via the LASSO

by
Arkadiusz Jędrzejewski
,
Grzegorz Marcjasz
and
Rafał Weron
*
Department of Operations Research and Business Intelligence, Wrocław University of Science and Technology, 50-370 Wrocław, Poland
*
Author to whom correspondence should be addressed.
Energies 2021, 14(11), 3249; https://doi.org/10.3390/en14113249
Submission received: 28 March 2021 / Revised: 21 May 2021 / Accepted: 27 May 2021 / Published: 2 June 2021

Abstract

:
Recent studies suggest that decomposing a series of electricity spot prices into a trend-seasonal and a stochastic component, modeling them independently, and then combining their forecasts can yield more accurate predictions than an approach in which the same parsimonious regression or neural network-based model is calibrated to the prices themselves. Here, we show that significant accuracy gains can also be achieved in the case of parameter-rich models estimated via the least absolute shrinkage and selection operator (LASSO). Moreover, we provide insights as to the order of applying seasonal decomposition and variance stabilizing transformations before model calibration, and propose two well-performing forecast averaging schemes that are based on different approaches for modeling the long-term seasonal component.

1. Introduction

The trend-seasonal pattern of electricity spot prices, which is also known as the long-term seasonal component (LTSC), has always attracted the attention of energy analysts [1,2,3,4,5,6,7]. This is especially so when modeling the average daily prices in the medium- or the long-term. On the other hand, the short-term electricity price forecasting (EPF) literature has generally ignored it and considered models with only intra-day and intra-week periodicities, as the LTSC was believed to add unnecessary complexity. Nowotarski and Weron [8] only recently introduced the seasonal component (SC) approach and the seasonal component autoregressive (SCAR) models that decompose the electricity spot price series into a trend-seasonal and stochastic component, predict them independently, and then combine their day-ahead forecasts. The seasonal component approach works well for autoregressive (AR) [7,9] as well as non-linear autoregressive (NARX) neural network-type models [10], in the context of point and probabilistic predictions [11]. However, the studies that have been published to date may be criticized for only utilizing parsimonious structures with a relatively small number of explanatory variables or features. Additionally, these are known to underperform when compared to parameter-rich models with hundreds of regressors that are estimated via the least absolute shrinkage and selection operator (LASSO) [12,13,14,15,16]. To the best of our knowledge, only two studies have treated the LTSC in the context of LASSO-estimated models [17,18]. However, no comparisons have been made between different variants of the LTSC or with analogous models that do not utilize seasonal decomposition. An open question remains as to whether the SC approach is also beneficial in the case of parameter-rich LASSO-estimated models and to what extent.
To this end, we perform an extensive empirical study that involves:
  • Six-year-long electricity price and fundamental variables time series from two distinct power markets—Nord Pool and PJM Interconnection, providing two three-year-long test periods of hourly-resolution.
  • Two commonly used approaches for modeling the LTSC of electricity price series—one based on wavelet smoothing [2,3,4,5,7] and one on the Hodrick-Prescott (HP) filter [4,10,19,20].
  • A parameter-rich, LASSO-estimated autoregressive model with nearly 130 regressors, after [21] called the LEAR model.
  • The area hyperbolic sine (asinh) variance stabilizing transformation (VST) [22,23], which has been found to perform well when negative or close to zero electricity prices are analyzed [15,24,25,26].
  • Two methods of combining point forecasts—one that selects the best combination of a pool of models (dubbed the Best Combination, BC) and one inspired by Bayesian Model Averaging (BMA) [27]. The latter weighs combinations by the inverse (of the) root mean squared error (iRMSE; similarly as in [28,29,30]), instead of the posterior probabilities that were originally proposed in [27]. We refer to it as BMA for notational convenience.
  • Model validation in terms of the robust relative mean absolute errors (rMAE) and relative root mean squared errors (rRMSE) [21,31], and the Giacomini and White [32] test for significant differences in conditional predictive ability (CPA).
Because we utilize both seasonal decomposition (via the HP or the wavelet filter) and variance stabilization, a question arises regarding the order in which they should be applied. In some studies, the VST comes first [2,4,8,9,10,11], whereas in others—seasonal decomposition [18]. Not having found clear recommendations in the literature, we compare both approaches.
Our contribution is threefold. Firstly, we show that significant—as measured by the CPA test [32]—accuracy gains can be achieved by applying the seasonal component approach [8] not only to parsimonious autoregressive [7,9] or neural network [10] structures, but also to LASSO-estimated AR (LEAR) models. Provided that the performance of the latter is nearly on par with that of Deep Neural Networks (DNN) [21], this finding provides valuable insights in constructing state-of-the-art models, which do not require hyperparameter optimization and offer orders of magnitude faster execution than DNNs. Secondly, we provide insights as to the order of applying seasonal decomposition and variance stabilizing transformations before model calibration. Thirdly, we propose two well-performing forecast averaging schemes that are based on different approaches to modeling the LTSC.
The remainder of the paper is structured, as follows. In Section 2, we briefly present the datasets. Subsequently, in Section 3, we describe the methodology: the forecasting framework, the asinh transformation, seasonal decomposition, a parsimonious autoregressive model used as a benchmark, the LEAR model, and the two combination schemes. In Section 4, we measure the forecast accuracy in terms of rMAE and rRMSE, evaluate the conditional predictive ability, and comment on the computational complexity of the proposed methods. Finally, in Section 5, we wrap up the results and provide conclusions.

2. Datasets

We evaluate the considered models using datasets from two major power markets. Nord Pool is the first one, a renewable energy source dominated market in the Northern Europe, exhibiting long-term, weather-dependent fluctuations in the price levels. The second is PJM Interconnection, the world’s largest competitive wholesale electricity market covering Northeastern United States, with a coal–gas–nuclear generation mix.
The Nord Pool (NP) dataset, as depicted in Figure 1, comprises three time series at hourly resolution: day-ahead market system prices in EUR/MWh, day-ahead system load forecasts (called consumption prognosis) for four Nordic countries (Denmark, Finland, Norway, and Sweden), and day-ahead wind power generation forecasts for Denmark. The data are freely available on the Nord Pool website www.nordpoolspot.com (accessed on 15 January 2021) The PJM dataset, as depicted in Figure 2, also comprises three time series at an hourly resolution: day-ahead market prices in the Commonwealth Edison (COMED; located in the state of Illinois) zone in USD/MWh, and two day-ahead load forecasts series—the system load and the COMED zonal load. The data are freely available on the PJM website www.pjm.com (accessed on 18 January 2021). Both of the datasets span the same six-year-long time period—from 1 January 2013 to 24 December 2018 (exactly 2184 = 6 × 364 days or 52,416 = 6 × 364 × 24 h). It is noteworthy that, while the NP price is more volatile in the three-year-long evaluation window than in the model selection window, the PJM price exhibits the opposite behavior, see Table 1 with summary statistics for the analyzed time series. This will allow us to compare the models under different market conditions. Additionally, it is noteworthy that the same datasets were used in [21].

3. Methodology

We implement the multivariate modeling framework, as defined in [15], in which the predictions are separately performed for each hour. To represent our variables, we use ‘day × hour’ matrix-like structures. The day-ahead forecasts p ^ d , h | d 1 p ^ d , h of the electricity price p d , h for day d and hour h are computed on day d 1 using all of the information known up to that point in time. We utilize a rolling window scheme, i.e., our models are calibrated to data in a 364-day-long window and, once the 24 forecasts for all hours of the next day are made, the window is rolled forward 24 h. Subsequently, the models are calibrated again, and forecasts are computed for the next 24 h. This procedure continues until the predictions for the last day in the sample are made.
Our models involve day-ahead electricity prices and two exogenous variables, i.e., the system load and wind power forecasts for NP and system and zonal load forecasts for PJM, see Section 2 and Figure 1 and Figure 2. All three time series undergo a variance stabilizing transformation (VST; see Section 3.1, below) and seasonal decomposition (see Section 3.2, below) prior to model calibration; we consider two orders of applying these transformations—VST first, then seasonal decomposition, and vice versa. The only exceptions are the benchmark models for which seasonal decomposition is not performed. Although we use a multivariate modelling framework, both of the transformations are carried out in the calibration window for the original time series at an hourly resolution, just like in [8]. Moreover, because the forecasts of the exogenous variables are known one day in advance, in their case seasonal decomposition and variance stabilization are applied to a window that is one day longer, i.e., which covers 365 days and includes the day for which the price forecasts are made.
When it comes to forecasting, like in [8,10], the LTSC of the price time series is assumed to persist into the future (see Section 3.2, below), whereas the stochastic component is predicted based on two models with an autoregressive structure. The first one is a parsimonious ordinary least squares (OLS) estimated model, which wsa built on some prior knowledge of experts (see Section 3.3 below), whereas the second one is a parameter-rich LASSO-estimated model (see Section 3.4, below).
Finally, in Section 3.5, we propose two methods of averaging/selecting forecasts. We construct our combined predictions that are based on performance in a two-year-long model selection (or validation) window and compare them—like all the considered models—in the three-year-long evaluation (or test) window, see Figure 1 and Figure 2.

3.1. Data Transformation

A number of transformations can be used to reduce the variation in data and allow handling close to zero or negative electricity prices. Here, following [15,18,22], we resort to the area hyperbolic sine (asinh), which is a simple, but well performing, variance stabilizing transformation (VST). Before applying it, we normalize each series x d , h , i.e., prices or load/wind forecasts, using its median Med τ and median absolute deviation MAD τ in the calibration window τ :
y d , h = 1 z 0.75 x d , h Med τ MAD τ ,
where z 0.75 is the 75th percentile of the standard normal distribution and it ensures asymptotically normal consistency to the standard deviation. Such a normalization is more robust to outliers than the one based on the mean and standard deviation and, thus, it is preferred for spiky data [23]. Having normalized our variables, we apply the VST:
z d , h = asinh ( y d , h ) = log y d , h + y d , h 2 + 1 .
The forecasts are obtained for models that were calibrated to the VST-transformed series, after applying the inverse transformation:
x d , h = MAD τ z 0.75 sinh z d , h + Med τ ,
where sinh is the hyperbolic sine. Because the latter is a nonlinear function, given random variable X, E sinh ( X ) does not have to equal sinh ( E X ) . Hence, from the probabilistic point of view, Equation (3) is not the correct inverse transformation. Although this problem is generally ignored in the literature, Narajewski and Ziel [33] argue that using the correct transformation yields better forecasts. On the other hand, when the forecasts are averaged, the differences between the two approaches seem to vanish [34]. Because we eventually consider forecast combinations, to keep it simple, we use Equation (3) as the inverse transformation.

3.2. Seasonal Decomposition

Seasonal decomposition is a term that generally refers to representing a signal as a sum or product of a periodic component and the remaining variability (also known as the stochastic component). Recent studies have shown that the seasonal decomposition of day-ahead prices and separate treatment of the seasonal and stochastic components can result in more accurate forecasts being generated by OLS-estimated [10,11], as well as LASSO-estimated [18], models. The approach that was proposed by Nowotarski and Weron [8] relies on an additive decomposition that splits the original time series, Y t , into the long-term trend-seasonal component (LTSC), T t , and the remaining stochastic component with short-term periodicity, X t :
Y t = T t + X t .
In our study, depending on the order of applying seasonal decomposition and the variance stabilizing transformation, Y t may denote asinh-transformed or raw data. In the latter case, the VST is only applied to X t . Both of the components are modeled independently and their predictions are combined to yield price forecasts. Like in [8,9,10,11], we consider persistent forecasts, i.e., assume that the last 24 hourly observations of the LTSC will repeat the next day:
T ^ d * + 1 , h = T d * , h ,
where d * is the last day in the calibration window and the single and double time indexes satisfy t = 24 d + h .
We consider two well-performing methods of extracting and modeling the LTSC—one that is based on wavelet smoothing [2,3,4,5,7] and one on the Hodrick–Prescott (HP) filter [4,10,19,20]. In wavelet smoothing [1,35], the original time series is decomposed using the discrete wavelet transform into a sum of an approximation series capturing the general trend, S k , and a number of detail series, D k , representing higher-frequency components: S k + D k + D k 1 + + D 1 , where k is the smoothing level. The LTSC is then approximated by S k . We consider the Daubechies family of wavelets, which is frequently used in EPF studies [8,9,10,11,36,37,38,39]. More precisely, we utilize Daubechies wavelets of order 4 (that are denoted by ‘db4’), and the smoothing level k ranges from 6 to 14. Readers are directed to the left panels in Figure 3 for an illustration.
While wavelet smoothing removes layers of details, D 1 , D 2 , etc., the Hodrick–Prescott [40] filter simply returns T t , which minimizes squared deviations from Y t (first term) and squared fluctuations of the smoothed series itself (the second term):
min T t t = t 0 τ ( Y t T t ) 2 + λ t = t 0 + 1 τ 1 ( T t + 1 T t ) ( T t T t 1 ) 2 ,
where t 0 and τ are, respectively, the beginning and end of the calibration window in the single time index notation, and λ is the smoothing parameter. Here, we consider nine values of the latter: λ = 10 k with k = 5 , 6 , , 13 . The larger the λ , the smoother the resulting series, see the right panels shown in Figure 3.

3.3. The OLS-Estimated Benchmark

As a benchmark, we consider the well-performing expert DoW , nl model that was proposed in [15], which was extended to include two exogenous variables. We denote it by ARX to reflect its autoregressive structure with exogenous variables. Within this model, the transformed price on day d and hour h is given by:
X d , h = β h , 1 X d 1 , h + β h , 2 X d 2 , h + β h , 3 X d 7 , h autoregressive effects + β h , 4 X d 1 , m i n + β h , 5 X d 1 , m a x non-linear effects + β h , 6 X d 1 , 24 midnight value + β h , 7 C 1 , d , h + β h , 8 C 2 , d , h exogenous variables + i = 1 7 β h , 8 + i D i , d weekday dummies + ϵ d , h ,
where X d 1 , m i n and X d 1 , m a x are the previous day’s minimum and maximum observations, X d 1 , 24 is the previous day’s midnight value, i.e., the last known observation in the day-ahead market, C 1 , d , h and C 2 , d , h are two exogenous variables (here, day-ahead forecasts of price drivers that are relevant for a given dataset, see Section 2), D 1 , d , , D 7 , d are weekday dummies, and ϵ d , h is the noise term. The parameters of the model are estimated using OLS, independently for each hour.

3.4. The LEAR Model

The structure of the main predictive model that is considered in this study is a natural extension of Equation (7), which includes all 24 hourly observations for the considered three days, i.e., yesterday, two days ago, and a week ago, and predictions of the exogenous variables for all 24 h of the day:
X d , h = i = 1 24 β h , i X d 1 , i + β h , 24 + i X d 2 , i + β h , 48 + i X d 7 , i + β h , 73 X d 1 , m i n + β h , 74 X d 1 , m a x + i = 1 24 β h , 74 + i C 1 , d , i + β h , 98 + i C 2 , d , i + i = 1 7 β h , 122 + i D i , d + ϵ d , h .
We estimate the 129 regressors via the least absolute shrinkage and selection operator (LASSO) [41]:
β ^ h = arg min β RSS + α i = 1 129 β h , i ,
where RSS is the residual sum of squares and α 0 is a tuning parameter. For α = 0 , we get the standard OLS estimator, for large values of α , all coefficients become zeros, while, for intermediate α ’s, the LASSO shrinks some of them to zero and, thus, performs variable selection. There are several techniques for optimizing the tuning parameter [42]. Here, we use cross-validation with seven folds since the number of observations in the calibration window is divisible by 7. Following [21], we refer to this LASSO-estimated autoregressive model as the LEAR model, although the LEAR model that is defined in [21] is a richer structure (it includes nearly 250 regressors, most notably lagged fundamental variables).

3.5. Forecast Averaging Schemes

The best-performing seasonal decomposition is not known in advance. We utilize forecast averaging to address this problem [28,29,30]. We generate a pool of forecasts from a single model (here, ARX or LEAR) calibrated to data decomposed with different filters (either wavelet- or HP-based) and then combine them into a single value, as suggested in [8]. We consider two methods of combining point forecasts—one that selects the best combination of a pool of models and one that is inspired by Bayesian Model Averaging [27].
Moreover, because our results (see Section 4) do not provide a clear indication for the optimal order of applying transformations—seasonal decomposition first and variance stabilization second, i.e., VST[SD(×)], or variance stabilization first and seasonal decomposition second, i.e., and SD[VST(×)]—we also consider combining the forecasts from both approaches. Overall, the combined predictions are constructed from individual forecasts: either 18 wavelet-based (nine smoothing levels and two orders of applying transformations) or 18 HP filter-based (nine values of λ and two orders of applying transformations).

3.5.1. Best Combination

The simpler averaging scheme selects the best performing combination of a pool of models. Hence, we refer to it as the Best Combination and denote it by BC. Overall, given 18 individual forecasts (→ columns in the left rectangle in Figure 4), there are 2 18 1 possible combinations (→ rows in the same rectangle). The first combination is just the forecast for the first LTSC, the second is the forecast for the second LTSC, …, the 19th is the combination of forecasts for the first and second LTSCs, the 20th is the combination of forecasts for the first and third LTSCs, …, and the last is the combination of forecasts for all of the LTSCs.
In each row, the predictions are averaged using the arithmetic mean. Subsequently, in order to choose the best-performing combination, we calculate the root mean squared error (RMSE) for all considered combinations of forecasts in the model selection window (i.e., the two-year period directly preceding the three-year test period, see Figure 1 and Figure 2):
RMSE = 1 728 × 24 d = 365 1092 h = 1 24 ( p d , h p ^ d , h c ) 2 ,
where p ^ d , h c is one of the 2 18 1 combined forecasts and p d , h is the actual price for day d and hour h. We choose the combination that returns the lowest RMSE, as seen in the green square in Figure 4. This combination is eventually evaluated in the model evaluation window (i.e., the three-year test period) and compared to other combined and individual models.
In Table 2, we present the five best-preforming combinations in the model selection window that was obtained for both datasets and the LEAR model with wavelet-based filtering. Squares indicate individual predictions that are averaged in a given combination. For instance, for the NP dataset, the best combination is composed of forecasts for three LTSCs that were obtained by first applying seasonal decomposition, and then the VST: S 8 , S 10 , and S 13 , and one LTSC being obtained by first applying the VST, then seasonal decomposition: S 10 . The RMSE of this combination in the model selection window is 2.2984, being nearly the same as that of the second best combination. The last row in each part of the table represents averaging over all individual forecasts; for NP, the RMSE of 2.3485 is ca. 2% higher than the best-performing combination.

3.5.2. BMA-Type Averaging

The more complex averaging scheme, instead of selecting one forecast combination out of 2 18 1 possibilities, weighs all of the combinations:
p ^ d , h BMA = i = 1 2 18 1 w i p ^ d , h c , i ,
where p ^ d , h c , i is the ith of the 2 18 1 combined forecasts. The weights are computed based on past performance:
w i = iRMSE i i = 1 2 18 1 iRMSE i ,
where iRMSE i is the inverse of the root mean squared error of the ith combination in Equation (10), see Figure 4. In the last column of Table 2, we report iRMSE values for the five best-preforming combinations that were obtained for both datasets and the LEAR model with wavelet-based filtering. As can be seen, because of the small differences in RMSE values, the weights are nearly identical for the top performing combinations and very similar to the weight for the average across all individual forecasts.
This idea is similar in spirit to Bayesian Model Averaging [27], which weighs combinations by posterior probabilities. The latter has not performed very well in an extensive EPF study [30], thus inspiring our idea to replace posterior probabilities by iRMSE weights, similarly to what was done in [28,29]. Although our approach is only inspired by Bayesian Model Averaging, for simplicity of notation, we denote it by BMA.

4. Results

4.1. Forecast Evaluation

Forecast accuracy is assessed in the three-year or 3 × 364 = 1092 -day model evaluation window, using two error measures. Following the recommendations put forward in [21], both are relative metrics. This allows us to compare the model performance across different datasets. The relative mean absolute error (rMAE) and the relative root mean squared error (rRMSE) are defined as:
rMAE = d = 1093 2184 h = 1 24 p d , h p ^ d , h d = 1093 2184 h = 1 24 p d , h p ^ d , h naive , rRMSE = d = 1093 2184 h = 1 24 p d , h p ^ d , h 2 d = 1093 2184 h = 1 24 p d , h p ^ d , h naive 2 ,
where p d , h denotes the actual (observed) price for day d and hour h, p ^ d , h is the corresponding prediction that is obtained using the model under the evaluation, and p ^ d , h naive is a similar-day prediction [43]:
p ^ d , h naive = p d 7 , h for Monday , Saturday , and Sunday , p d 1 , h otherwise .
Figure 5 and Figure 6 present the results for the individual and combined ARX and LEAR models for both datasets. They all lead to similar conclusions.
Firstly, we can see that seasonal decomposition can improve the accuracy of forecasts that are generated not only by simple autoregressive models, but also by parameter-rich models with automated variable selection via the LASSO. More considerable improvements can be achieved in the NP market, see Figure 5. Moreover, in this case, most of the individual models with seasonal decomposition return lower error scores than the benchmark model without it (as denoted by noLTSC). For the PJM dataset, the improvements are smaller, and they occur more frequently for ARX models, as seen in Figure 6. The only problem with such individual models is that not all of them beat the benchmark model without seasonal decomposition. Thus, it is difficult to choose the optimal LTSC ex-ante.
Secondly, there is no general answer to the question of which order of applying data transformations performs better. It seems to depend on the considered model, market, type of the LTSC, or error measure. For instance, in the top row of Figure 6, where the rMAE is presented for the PJM dataset, the best performing approach changes depending on the filtering type considered. For wavelet-based filters, it is generally more effective to apply the seasonal decomposition first, whereas, for HP-based filters, the is best applied VST first. On the contrary, for the NP dataset and rRMSE (see the bottom row of Figure 5), the performance strongly depends on the chosen parameters that are used to extract the LTSC, i.e., the smoothing level k or the smoothing parameter λ .
We consider two averaging schemes to overcome these difficulties, see Section 3.5. In Table 3, we collect all of the error scores for the combined and benchmark models. It turns out that, for both datasets, the BC and BMA averaging schemes always return lower rMAE and rRMSE scores than the benchmark model without seasonal decomposition (denoted by noLTSC). Moreover, there are generally only a few individual models that can outperform the combined models. Hence, we can conclude that forecast averaging solves the problem with the ex-ante selection of the best performing LTSC and the order of data transformations. Furthermore, all of the combined LEAR models outperform their ARX counterparts, so the use of LASSO-estimated models additionally increases the accuracy of the combined forecasts.
Finally, let us compare the averaging schemes and the methods of modeling the LTSC. It turns out that the BMA approach returns lower errors than BC in nine out of 16 cases. When it comes to the best-performing approach for extracting the LTSC, we can observe that the one that is based on HP filters yields lower error scores in all of the cases except one (i.e., LEAR, BC averaging, NP market; see Table 3).

4.2. Testing for Conditional Predictive Ability

The obtained forecasts are also compared using the conditional predictive ability (CPA) test of Giacomini and White [32]. Here, following [15], we only focus on the results of the multivariate version of the test, which, for each pair of models, returns a single p-value for all 24 h. The tests were separately conducted for the absolute and squared losses.
Following [10,15,21], we present the CPA test results as chessboards with a color-coded p-value, where the color ranges from dark red (higher p-values) to dark green (lower p-values). A colored square in the chessboards indicates that the predictions of the model on the x-axis are significantly better than the predictions of the model on the y-axis. The greener this field (the lower the p-value is), the more statistically significant the difference. A black square means that there is no statistically significant difference in the predictive accuracy at the 10% level.
In Figure 7 and Figure 8, we illustrate the results for the absolute and squared losses, respectively. As can be seen, the combined forecasts significantly outperform the predictions of the benchmark models without seasonal decomposition in 22 out of 32 cases (at the 10% level). When we only consider the LEAR models, the improvement is statistically significant in 11 out of 16 cases. Only for the PJM market (both averaging schemes for the squared losses, see Figure 8, and LEAR BMA db4 for linear losses, see Figure 7), the improvements over the benchmark are not statistically significant. Moreover, the outperformance of the forecasts of the combined LEAR models over their ARX counterparts is significant in 14 out of 16 cases. Thus, the CPA tests confirm that seasonal decomposition improves the accuracy of combined forecasts that are generated by the LASSO-estimated models.
Regarding the method of combining forecasts, although the BMA approach returns lower errors than the BC approach in nine out of 16 cases (see Table 3), the CPA tests indicate that outperformance is only significant in two cases. On the other hand, the BC approach significantly outperforms the BMA approach in six cases. Given that the differences in the error scores between both of the averaging schemes are small, we can recommend the simpler approach—BC averaging.
Finally, although the combined models based on HP filters return lower error scores in almost all cases, the advantages of using one LTSC function over the other are not so clear from the perspective of the CPA test. The combined forecasts that are based on HP filters are only significantly better than those based on wavelet filters in six cases, at the 10% level, whereas the remaining 10 cases show no significant differences. When considering the 5% level and only the combined LEAR models, all of the differences become statistically insignificant.

4.3. Computational Complexity

The models that are considered in this study are computationally feasible, even in time-constrained scenarios. It is notable that all of the times reported in this subsection reflect a single-threaded task run on an AMD Threadripper 1950X processor.
The average computation time per forecast day amounts to 8–9 s for the LEAR models and 0.03–0.05 s for the ARX models. The listed times reflect the whole process, including data preprocessing, such as the application of the LTSC or the VST. Nevertheless, model estimation is the most time-consuming task. To be more precise, the ARX model without the LTSC is, on average, computed in 0.027 s, whereas the computation of the same model with the LTSC takes 0.03 s for db4 wavelets and 0.05 s for HP filters. For the LEAR model, the input data impact the convergence rate. The smallest time in this case was measured for the model with the LTSC based on db4 wavelets—8 s per day. On the other hand, for the model with the LTSC that was based on HP filters or without the LTSC, the computations took closer to 9 s.
The averaging schemes require additional computational time. One has to consider the generation of forecasts to determine the best averaging scheme. The sequential computation of the set of forecasts for one LTSC variant and a 728-day selection window takes around 33 h. Computing the errors of all 2 18 1 possible combinations of the forecasts takes 101 s. However, the above operations are only done once.

5. Conclusions

To address the question of whether the seasonal component approach is beneficial in the case of parameter-rich models, we have performed an extensive empirical study that involved a well-performing, LASSO-estimated autoregressive (LEAR) model with 129 regressors, two approaches to modeling the LTSC (wavelet smoothing and the Hodrick-Prescott filter), and the area hyperbolic sine transformation. Given that our initial results did not provide a clear indication for the optimal choice of the LTSC or the optimal order of applying transformations—seasonal decomposition first and then variance stabilization second, or vice versa—we have introduced two averaging schemes. The first one, dubbed Best Combination (BC), selects the best combination of a pool of models. The second is inspired by Bayesian Model Averaging, but it weighs combinations by the inverse (of the) root mean squared error (iRMSE), instead of the originally proposed posterior probabilities; for notational convenience, we refer to it as BMA.
Our results indicate that seasonal decomposition can significantly—as measured by the conditional predictive ability (CPA) test—improve the accuracy of the forecasts that are generated not only by simple autoregressive, but also by parameter-rich models with automated variable selection via the LASSO. Moreover, for both datasets, the averaging schemes always return lower error scores than the benchmark model without seasonal decomposition. At the same time, there are only a few individual models that can outperform the combined models. Hence, we can conclude that forecast averaging solves the problem with the ex-ante selection of the best performing LTSC and order of data transformations. Furthermore, all of the combined LEAR models outperform their parsimonious ARX counterparts, so the use of LASSO-estimated models additionally increases the accuracy of the combined forecasts. Although the BMA approach generally returns slightly lower errors than BC, the CPA tests indicate that outperformance is, in most cases, insignificant. Hence, we recommend the simpler approach—BC averaging. Interestingly, this is consistent with the recommendations put forward in the forecasting literature, i.e., that instead of combining the full set of forecasts, it may be advantageous to discard the models with the worst performance [29,44].
Finally, let us note that forecasts, no matter how accurate, are of limited use if they cannot be utilized to yield profits [16]. However, measuring the economic value of reducing electricity price forecasting errors—although desirable—is a difficult task, as argued in [45,46]. A model that yields lower errors may not always lead to better trading decisions. Because there are no golden standards in this respect, instead of devising an artificial strategy, we encourage practitioners to evaluate our models in an actual trading environment. Our codes are available upon request.

Author Contributions

Conceptualization, A.J., G.M. and R.W.; Investigation, A.J. and G.M.; Software, A.J. and G.M.; Validation, R.W.; Writing—original draft, A.J. and G.M.; Writing—review & editing, A.J., G.M. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Science Center (NCN, Poland) through MAESTRO grant No. 2018/30/A/HS4/00444 (to A.J. and R.W.) and the Ministry of Science and Higher Education (MNiSW, Poland) through Diamond Grant No. 0219/DIA/2019/48 (to G.M.).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Weron, R. Modeling and Forecasting Electricity Loads and Prices: A Statistical Approach; John Wiley & Sons: Chichester, UK, 2006. [Google Scholar]
  2. Janczura, J.; Trück, S.; Weron, R.; Wolff, R. Identifying spikes and seasonal components in electricity spot price data: A guide to robust modeling. Energy Econ. 2013, 38, 96–110. [Google Scholar] [CrossRef] [Green Version]
  3. Nowotarski, J.; Tomczyk, J.; Weron, R. Robust estimation and forecasting of the long-term seasonal component of electricity spot prices. Energy Econ. 2013, 39, 13–27. [Google Scholar] [CrossRef] [Green Version]
  4. Lisi, F.; Nan, F. Component estimation for electricity prices: Procedures and comparisons. Energy Econ. 2014, 44, 143–159. [Google Scholar] [CrossRef]
  5. Afanasyev, D.; Fedorova, E. The long-term trends on the electricity markets: Comparison of empirical mode and wavelet decompositions. Energy Econ. 2016, 56, 432–442. [Google Scholar] [CrossRef]
  6. Lisi, F.; Pelagatti, M.M. Component estimation for electricity market data: Deterministic or stochastic? Energy Econ. 2018, 74, 13–37. [Google Scholar] [CrossRef]
  7. Grossi, L.; Nan, F. Robust forecasting of electricity prices: Simulations, models and the impact of renewable sources. Technol. Forecast. Soc. Chang. 2019, 141, 305–318. [Google Scholar] [CrossRef]
  8. Nowotarski, J.; Weron, R. On the importance of the long-term seasonal component in day-ahead electricity price forecasting. Energy Econ. 2016, 57, 228–235. [Google Scholar] [CrossRef] [Green Version]
  9. Afanasyev, D.O.; Fedorova, E.A. On the impact of outlier filtering on the electricity price forecasting accuracy. Appl. Energy 2019, 236, 196–210. [Google Scholar] [CrossRef]
  10. Marcjasz, G.; Uniejewski, B.; Weron, R. On the importance of the long-term seasonal component in day-ahead electricity price forecasting with NARX neural networks. Int. J. Forecast. 2019, 35, 1520–1532. [Google Scholar] [CrossRef]
  11. Uniejewski, B.; Marcjasz, G.; Weron, R. On the importance of the long-term seasonal component in day-ahead electricity price forecasting: Part II—Probabilistic forecasting. Energy Econ. 2019, 79, 171–182. [Google Scholar] [CrossRef] [Green Version]
  12. Gaillard, P.; Goude, Y.; Nedellec, R. Additive models and robust aggregation for GEFCom2014 probabilistic electric load and electricity price forecasting. Int. J. Forecast. 2016, 32, 1038–1050. [Google Scholar] [CrossRef]
  13. Uniejewski, B.; Nowotarski, J.; Weron, R. Automated variable selection and shrinkage for day-ahead electricity price forecasting. Energies 2016, 9, 621. [Google Scholar] [CrossRef] [Green Version]
  14. Ziel, F. Forecasting electricity spot prices using LASSO: On capturing the autoregressive intraday structure. IEEE Trans. Power Syst. 2016, 31, 4977–4987. [Google Scholar] [CrossRef] [Green Version]
  15. Ziel, F.; Weron, R. Day-ahead electricity price forecasting with high-dimensional structures: Univariate vs. multivariate modeling frameworks. Energy Econ. 2018, 70, 396–420. [Google Scholar] [CrossRef] [Green Version]
  16. Uniejewski, B.; Weron, R. Regularized quantile regression averaging for probabilistic electricity price forecasting. Energy Econ. 2021, 95, 105121. [Google Scholar] [CrossRef]
  17. Ziel, F.; Steinert, R.; Husmann, S. Efficient modeling and forecasting of electricity spot prices. Energy Econ. 2015, 47, 89–111. [Google Scholar] [CrossRef] [Green Version]
  18. Uniejewski, B.; Weron, R. Efficient forecasting of electricity spot prices with expert and LASSO models. Energies 2018, 11, 2039. [Google Scholar] [CrossRef] [Green Version]
  19. Weron, R.; Zator, M. A note on using the Hodrick-Prescott filter in electricity markets. Energy Econ. 2015, 48, 1–6. [Google Scholar] [CrossRef] [Green Version]
  20. Caldana, R.; Fusai, G.; Roncoroni, A. Electricity forward curves with thin granularity: Theory and empirical evidence in the hourly EPEXspot market. Eur. J. Oper. Res. 2017, 261, 715–734. [Google Scholar] [CrossRef]
  21. Lago, J.; Marcjasz, G.; De Schutter, B.; Weron, R. Forecasting day-ahead electricity prices: A review of state-of-the-art algorithms, best practices and an open-access benchmark. Appl. Energy 2021, 293, 116983. [Google Scholar] [CrossRef]
  22. Schneider, S. Power spot price models with negative prices. J. Energy Mark. 2011, 4, 77–102. [Google Scholar] [CrossRef] [Green Version]
  23. Uniejewski, B.; Weron, R.; Ziel, F. Variance stabilizing transformations for electricity spot price forecasting. IEEE Trans. Power Syst. 2018, 33, 2219–2229. [Google Scholar] [CrossRef] [Green Version]
  24. De Lagarde, C.M.; Lantz, F. How renewable production depresses electricity prices: Evidence from the German market. Energy Policy 2018, 117, 263–277. [Google Scholar] [CrossRef]
  25. Zhou, Y.; Scheller-Wolf, A.; Secomandi, N.; Smith, S. Managing wind-based electricity generation in the presence of storage and transmission capacity. Prod. Oper. Manag. 2019, 28, 970–989. [Google Scholar] [CrossRef]
  26. Marcjasz, G. Forecasting electricity prices using deep neural networks: A robust hyper-parameter selection scheme. Energies 2020, 13, 4605. [Google Scholar] [CrossRef]
  27. Hoeting, J.; Madigan, D.; Raftery, A.; Volinsky, C. Bayesian model averaging: A tutorial. Stat. Sci. 1999, 14, 382–401. [Google Scholar]
  28. Bates, J.M.; Granger, C.W.J. The combination of forecasts. Oper. Res. Q. 1969, 20, 451–468. [Google Scholar] [CrossRef]
  29. Bordignon, S.; Bunn, D.W.; Lisi, F.; Nan, F. Combining day-ahead forecasts for British electricity prices. Energy Econ. 2013, 35, 88–103. [Google Scholar] [CrossRef] [Green Version]
  30. Nowotarski, J.; Raviv, E.; Trück, S.; Weron, R. An empirical comparison of alternate schemes for combining electricity spot price forecasts. Energy Econ. 2014, 46, 395–412. [Google Scholar] [CrossRef]
  31. Hyndman, R.; Koehler, A. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef] [Green Version]
  32. Giacomini, R.; White, H. Tests of conditional predictive ability. Econometrica 2006, 74, 1545–1578. [Google Scholar] [CrossRef] [Green Version]
  33. Narajewski, M.; Ziel, F. Econometric modelling and forecasting of intraday electricity prices. J. Commod. Mark. 2020, 19, 100107. [Google Scholar] [CrossRef] [Green Version]
  34. Marcjasz, G.; Uniejewski, B.; Weron, R. Beating the Naïve—Combining LASSO with Naïve Intraday Electricity Price Forecasts. Energies 2020, 13, 1667. [Google Scholar] [CrossRef] [Green Version]
  35. Percival, D.B.; Walden, A.T. Wavelet Methods for Time Series Analysis; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  36. Conejo, A.J.; Plazas, M.A.; Espinola, R.; Molina, A.B. Day-ahead electricity price forecasting using the wavelet transform and ARIMA models. IEEE Trans. Power Syst. 2005, 20, 1035–1042. [Google Scholar] [CrossRef]
  37. Amjady, N.; Keynia, F. Short-term load forecasting of power systems by combination of wavelet transform and neuro-evolutionary algorithm. Energy 2009, 34, 46–57. [Google Scholar] [CrossRef]
  38. Tan, Z.; Zhang, J.; Wang, J.; Xu, J. Day-ahead electricity price forecasting using wavelet transform combined with ARIMA and GARCH models. Appl. Energy 2010, 87, 3606–3610. [Google Scholar] [CrossRef]
  39. Li, S.; Wang, P.; Goel, L. A novel wavelet-based ensemble method for short-term load forecasting with hybrid neural networks and feature selection. IEEE Trans. Power Syst. 2015, 31, 1788–1798. [Google Scholar] [CrossRef]
  40. Hodrick, R.J.; Prescott, E.C. Postwar U.S. business cycles: An empirical investigation. J. Money Credit Bank. 1997, 29, 1–16. [Google Scholar] [CrossRef]
  41. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  42. Hastie, T.; Tibshirani, R.; Wainwright, M. Statistical Learning with Sparsity: The Lasso and Generalizations; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  43. Weron, R. Electricity price forecasting: A review of the state-of-the-art with a look into the future. Int. J. Forecast. 2014, 30, 1030–1081. [Google Scholar] [CrossRef] [Green Version]
  44. Stock, J.H.; Watson, M.W. Combination forecasts of output growth in a seven-country data set. J. Forecast. 2004, 23, 405–430. [Google Scholar] [CrossRef]
  45. Zareipour, H.; Canizares, C.A.; Bhattacharya, K. Economic impact of electricity market price forecasting errors: A demand-side analysis. IEEE Trans. Power Syst. 2010, 25, 254–262. [Google Scholar] [CrossRef]
  46. Hong, T.; Pinson, P.; Wang, Y.; Weron, R.; Yang, D.; Zareipour, H. Energy forecasting: A review and outlook. IEEE Open Access J. Power Energy 2020, 7, 376–388. [Google Scholar] [CrossRef]
Figure 1. Day-ahead system prices in the Nord Pool electricity market (NP; top) and two forward looking series: day-ahead system load forecasts (middle) and day-ahead wind power generation forecasts for Denmark (middle). The vertical dashed lines indicate the beginnings of the model selection (31 December 2013) and model evaluation (29 December 2015) windows.
Figure 1. Day-ahead system prices in the Nord Pool electricity market (NP; top) and two forward looking series: day-ahead system load forecasts (middle) and day-ahead wind power generation forecasts for Denmark (middle). The vertical dashed lines indicate the beginnings of the model selection (31 December 2013) and model evaluation (29 December 2015) windows.
Energies 14 03249 g001
Figure 2. Day-ahead market prices in the Commonwealth Edison zone (COMED; top) and two forward looking series: the day-ahead system (middle) and COMED zone (middle) load forecasts. The vertical dashed lines indicate the beginnings of the model selection (31 December 2013) and model evaluation (29 December 2015) windows.
Figure 2. Day-ahead market prices in the Commonwealth Edison zone (COMED; top) and two forward looking series: the day-ahead system (middle) and COMED zone (middle) load forecasts. The vertical dashed lines indicate the beginnings of the model selection (31 December 2013) and model evaluation (29 December 2015) windows.
Energies 14 03249 g002
Figure 3. Illustration of the long-term trend-seasonal component (LTSC) based on wavelets S 6 , S 9 , and S 14 (left column) and HP filters with λ = 10 5 , 10 9 , and 10 13 (right column) fitted to Nord Pool prices in the initial calibration window, see Figure 1. Only the last eight weeks are displayed.
Figure 3. Illustration of the long-term trend-seasonal component (LTSC) based on wavelets S 6 , S 9 , and S 14 (left column) and HP filters with λ = 10 5 , 10 9 , and 10 13 (right column) fitted to Nord Pool prices in the initial calibration window, see Figure 1. Only the last eight weeks are displayed.
Energies 14 03249 g003
Figure 4. An illustration of the BC and BMA-type averaging schemes. For both approaches, all possible combinations ( 2 18 1 rows in the left rectangle) of individual forecasts for the different variants of the LTSC ( 2 × 9 = 18 columns representing two orders of applying transformations, VST[SD(×)] and SD[VST(×)], and nine wavelet levels or nine HP filter’s λ s; see also Table 2) are generated. Subsequently, the predictions of all models for a given combination are averaged using the arithmetic mean and either the best-performing combination in terms of RMSE in the model selection window is selected (→ BC) or all combined forecasts are iRMSE -weighted (→ BMA).
Figure 4. An illustration of the BC and BMA-type averaging schemes. For both approaches, all possible combinations ( 2 18 1 rows in the left rectangle) of individual forecasts for the different variants of the LTSC ( 2 × 9 = 18 columns representing two orders of applying transformations, VST[SD(×)] and SD[VST(×)], and nine wavelet levels or nine HP filter’s λ s; see also Table 2) are generated. Subsequently, the predictions of all models for a given combination are averaged using the arithmetic mean and either the best-performing combination in terms of RMSE in the model selection window is selected (→ BC) or all combined forecasts are iRMSE -weighted (→ BMA).
Energies 14 03249 g004
Figure 5. Relative mean absolute errors (rMAE; top row) and relative root mean squared errors (rRMSE; bottom row) for individual and combined forecasts obtained for the ARX (left panels) and LEAR (right panels) models and the NP dataset. Circles represent the performance of individual models where seasonal decomposition is executed first, then the VST is applied, while triangles are in the opposite order. Models without seasonal decomposition are indicated by the black solid line (labeled noLTSC), whereas the combined models by colored solid lines with markers at the ends.
Figure 5. Relative mean absolute errors (rMAE; top row) and relative root mean squared errors (rRMSE; bottom row) for individual and combined forecasts obtained for the ARX (left panels) and LEAR (right panels) models and the NP dataset. Circles represent the performance of individual models where seasonal decomposition is executed first, then the VST is applied, while triangles are in the opposite order. Models without seasonal decomposition are indicated by the black solid line (labeled noLTSC), whereas the combined models by colored solid lines with markers at the ends.
Energies 14 03249 g005
Figure 6. Relative mean absolute errors (rMAE; top row) and relative root mean squared errors (rRMSE; bottom row) for individual and combined forecasts obtained for the ARX (left panels) and LEAR (right panels) models and the PJM dataset. Markers and line styles used are the same as in Figure 5.
Figure 6. Relative mean absolute errors (rMAE; top row) and relative root mean squared errors (rRMSE; bottom row) for individual and combined forecasts obtained for the ARX (left panels) and LEAR (right panels) models and the PJM dataset. Markers and line styles used are the same as in Figure 5.
Energies 14 03249 g006
Figure 7. Results of the CPA test [32] for absolute losses. A colored square (other than black) indicates that the predictions of the model on the x-axis are significantly better than the predictions of the model on the y-axis, at a given significance level.
Figure 7. Results of the CPA test [32] for absolute losses. A colored square (other than black) indicates that the predictions of the model on the x-axis are significantly better than the predictions of the model on the y-axis, at a given significance level.
Energies 14 03249 g007
Figure 8. Results of the CPA test [32] for squared losses. A colored square (other than black) indicates that the predictions of the model on the x-axis are significantly better than the predictions of the model on the y-axis, at a given significance level.
Figure 8. Results of the CPA test [32] for squared losses. A colored square (other than black) indicates that the predictions of the model on the x-axis are significantly better than the predictions of the model on the y-axis, at a given significance level.
Energies 14 03249 g008
Table 1. The summary statistics for the considered time series: the minimum value, the 1st quartile (Q1), the 2nd quartile (Q2; i.e., the median), the mean, the 3rd quartile (Q3), the maximum value, and the standard deviation. For each series, three rows report values in the initial calibration window (top), the model selection window (middle), and the model evaluation window (middle).
Table 1. The summary statistics for the considered time series: the minimum value, the 1st quartile (Q1), the 2nd quartile (Q2; i.e., the median), the mean, the 3rd quartile (Q3), the maximum value, and the standard deviation. For each series, three rows report values in the initial calibration window (top), the model selection window (middle), and the model evaluation window (middle).
Time SeriesWindowMin.Q1Q2MeanQ3Max.Std.
Nord Pool market
Day-ahead priceinitial1.3834.3037.4538.1340.90109.556.94
selection1.1421.3826.2725.3430.6169.948.01
evaluation2.1725.9630.2833.2739.92199.9711.17
System load forecastinitial26.2236.5542.6343.6450.2269.119.00
selection26.3736.9542.2343.1349.0267.188.07
evaluation26.8737.3843.1944.4050.9870.588.94
Wind power forecastinitial0.030.430.921.241.874.201.00
selection0.020.551.251.532.384.451.14
evaluation0.000.591.261.552.315.051.17
PJM market
Day-ahead priceinitial 3.24 25.8930.6432.3836.39302.1413.91
selection 6.98 23.8029.4234.1337.56839.3028.80
evaluation 3.62 20.8025.5127.3832.29184.4810.79
System load forecastinitial57.8880.1089.8591.00100.30157.1615.88
selection58.2479.5589.4191.44102.56149.9316.26
evaluation58.3579.4288.3990.7299.98153.3116.29
Zonal load forecastinitial7.4610.0011.3711.5712.6722.242.26
selection7.259.8011.1711.3312.5220.302.08
evaluation7.359.7311.0311.3312.3421.272.25
Table 2. Root mean squared errors (RMSE) and their inverse values (iRMSE) calculated in the two-year model selection window (see Figure 1 and Figure 2) for the five best-performing combined forecasts obtained from the LEAR models with wavelet-based filtering. Squares indicate individual forecasts that are averaged in a given combination. For comparison, the last row represents averaging over all individual forecasts. Individual models where seasonal decomposition is executed first, and then the VST is applied, are labeled by VST[SD(×)], whereas models with the opposite order are labeled by SD[VST(×)].
Table 2. Root mean squared errors (RMSE) and their inverse values (iRMSE) calculated in the two-year model selection window (see Figure 1 and Figure 2) for the five best-performing combined forecasts obtained from the LEAR models with wavelet-based filtering. Squares indicate individual forecasts that are averaged in a given combination. For comparison, the last row represents averaging over all individual forecasts. Individual models where seasonal decomposition is executed first, and then the VST is applied, are labeled by VST[SD(×)], whereas models with the opposite order are labeled by SD[VST(×)].
Nord Pool market
VST[SD(×)]SD[VST(×)]
S 6 S 7 S 8 S 9 S 10 S 11 S 12 S 13 S 14 S 6 S 7 S 8 S 9 S 10 S 11 S 12 S 13 S 14 RMSE iRMSE
1 2.29840.4351
2 2.29840.4351
3 2.29850.4351
4 2.29910.4350
5 2.30020.4347
2.34850.4258
PJM market
VST[SD(×)]SD[VST(×)]
S 6 S 7 S 8 S 9 S 10 S 11 S 12 S 13 S 14 S 6 S 7 S 8 S 9 S 10 S 11 S 12 S 13 S 14 RMSE iRMSE
1 19.05960.0525
2 19.12810.0523
3 19.14230.0522
4 19.16330.0522
5 19.20040.0521
20.49950.0488
Table 3. The relative mean absolute errors (rMAE) and relative root mean squared errors (rRMSE) calculated in the model evaluation window for the combined and the baseline ARX and LEAR models. The lowest score in a row is emphasized in bold, independently for the ARX and LEAR models, whereas the lowest score in a row across all of the models has a green background.
Table 3. The relative mean absolute errors (rMAE) and relative root mean squared errors (rRMSE) calculated in the model evaluation window for the combined and the baseline ARX and LEAR models. The lowest score in a row is emphasized in bold, independently for the ARX and LEAR models, whereas the lowest score in a row across all of the models has a green background.
ErrorARXLEAR
noLTSCLTSCBCBMAnoLTSCLTSCBCBMA
Nord Pool market
rMAE0.7817HP0.66150.6607 0.7062HP0.60160.6063
db40.66260.6658db40.59440.6074
rRMSE0.7541HP0.71540.71900.7153HP0.64240.6503
db40.73210.7312db40.65630.6744
PJM market
rMAE0.7273HP0.67610.6814 0.6713HP0.66030.6520
db40.72630.7076db40.66340.6596
rRMSE0.7343HP0.70260.69880.6959HP0.68790.6765
db40.72360.7157db40.69230.6792
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jędrzejewski, A.; Marcjasz, G.; Weron, R. Importance of the Long-Term Seasonal Component in Day-Ahead Electricity Price Forecasting Revisited: Parameter-Rich Models Estimated via the LASSO. Energies 2021, 14, 3249. https://doi.org/10.3390/en14113249

AMA Style

Jędrzejewski A, Marcjasz G, Weron R. Importance of the Long-Term Seasonal Component in Day-Ahead Electricity Price Forecasting Revisited: Parameter-Rich Models Estimated via the LASSO. Energies. 2021; 14(11):3249. https://doi.org/10.3390/en14113249

Chicago/Turabian Style

Jędrzejewski, Arkadiusz, Grzegorz Marcjasz, and Rafał Weron. 2021. "Importance of the Long-Term Seasonal Component in Day-Ahead Electricity Price Forecasting Revisited: Parameter-Rich Models Estimated via the LASSO" Energies 14, no. 11: 3249. https://doi.org/10.3390/en14113249

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop