3. Comparing Methodologies: Yearly vs. Monthly
In this section, we compare the two different methodologies—yearly and monthly—that were proposed in [
18,
19], respectively. For this purpose, we considered a dataset spanning four years of weather and irradiation information between 1 January 2019 and 31 December 2022. Furthermore, we retrained all the models of both methodologies in the same way as specified in
Section 3 of [
18] and [
19], respectively. In order to have one coherent and consistent dataset (which was not the case for [
18,
19]), we developed and implemented a database, the details of which are covered in
Section 4.
For this paper, as we acquired one and two years of additional information, respectively, in regard to [
18,
19], we went one step further and calculated for each methodology (i.e., yearly and monthly) and the 36 best models (12 months and each month the three best models of SARIMAX, SARIMA, and ARIMAX) the averages of all four performance metrics (RMSE, MAE, ME, and MPE) for four different years (2019 to 2022). Note that in calculating the overall averages, we attributed the same weight of 0.25 to all four considered performance metrics.
Table 1 and
Table 2 show the optimized three models of SARIMAX, SARIMA, and ARIMAX for each month of the year, based on the two different methodologies of [
18,
19], respectively. Those tables also show for each month the best ARIMA-based model, in bold text. It is important to note that the numbers within the brackets are each model’s parameters, the details of which can be found in Section A.2 of [
18]. It can be seen that the yearly based methodology generated five ARIMAX, six SARIMAX, and one SARIMA best models, whereas the monthly based methodology produced three ARIMAX, six SARIMAX, and three SARIMA best models. As mentioned above, the best model for each month was the one that provided the best average accuracy for the considered four performance metrics (RMSE, MAE, ME, and MPE) and for the four studied years of 2019 to 2022.
Figure 1 and
Figure 2 illustrate the four considered performance metrics—RMSE, MAE, ME, and MPE—of the yearly based methodology. Note that the horizontal dotted lines present the average value of the corresponding performance metric, by considering the results of four years. It can be seen from
Figure 1 that out of 48 months, approximately 20 had values above the average RMSE and MAE of 10.6 and 8.7, respectively, occurring mostly between April and August. Analyzing the ME and MPE values, it can be seen from
Figure 2 that out of 48 months, about 22 of them had values exceeding the average ME and MPE of 6.4 and 15.4, respectively, occurring mostly during January, April, June, November, and December. Thus, we concluded that for the month of June, all four performance metrics had values above their corresponding averages.
Figure 3 and
Figure 4 demonstrate the four performance metrics—RMSE, MAE, ME, and MPE—of the monthly based methodology, where the horizontal dotted lines have the same meaning as in the previous two figures. We can see in
Figure 3 that in the course of the 48 months, both the RMSE and the MAE exceeded their corresponding averages of 10.36 and 8.35, respectively, 25 times, mostly between March and July. Regarding the ME and MPE values, it can be seen in
Figure 4 that in the course of the 48 months, their values were greater than the corresponding averages of 5.99 and 14.8, respectively, about 25 times, mostly during April (only for ME), July, and from September to November. Thus, we can conclude that for the monthly based methodology, all four performance metrics were above their averages during the month of July.
Figure 5 shows boxplots of the yearly based, the monthly based, and the hybrid methodologies. The three boxplots were created by considering four years (2019 to 2022) and 12 months (January to December), which resulted in 48 data points. Furthermore, each data point presents the average of the four performance metrics: ME, MPE, RMSE, and MAE. It can be seen in
Figure 5 that the monthly based methodology obtained, in general, lower error rates than the yearly based methodology: the former had median, minimum, and maximum error rates of 10%, 7% (reported as an outlier), and 12.49%, respectively, whereas the latter had 10.49%, 6.64%, and 13.71%, respectively. Those results led us to carry out a further detailed comparison between the two methodologies, from which we derived the following conclusions: neither methodology outperformed absolutely the other, as there were certain months when monthly based models performed better than yearly based ones and vice versa. Consequently, in this paper, we propose the hybrid methodology. We selected the best model for each month out of the monthly based and yearly based methodologies, and the results are shown in
Table 3. The improvements, in terms of performance metrics, are noticeable from the boxplot in
Figure 5 of the hybrid methodology. The hybrid methodology took the best part of the monthly based and yearly based methodologies and had median, minimum, and maximum error rates of 9.7%, 6.64% (reported as an outlier), and 12.3% (reported as an outlier), respectively. Note that the outliers (i.e., the circles in
Figure 5) were due to the fact that most of the error rates were within the interquartile range (IQR), while the rest (the outliers) lay outside of this range.
6. Conclusions
Forecasting accurately the generation of large-scale renewable energy sources is challenging, due to the underlying stochastic nature of the problem. In this paper, we considered the two different methodologies—yearly based and monthly based—proposed in our respective previous works [
1,
2]. To compare the accuracy of those two methodologies, unlike in our previous works, we assumed a consistent dataset spanning four years (1 January 2019 to 31 December 2022). Furthermore, we computed four different performance metrics (RMSE, MAE, ME, and MPE) for the best ARIMA-based model of each month and methodology. Using the equal weights of those four metrics, we then derived a new set of optimal models for the two considered methodologies, and we compared them to each other. We found that neither methodology was absolutely superior to the other: consequently, we proposed a hybrid of the two. Our hybrid methodology considered the best ARIMA-based model for each month of the year. By means of boxplots (i.e., minimum, maximum, mean, and median values), we showed that our proposed hybrid methodology improved accuracy (i.e., reduced error rates) compared to the yearly based and monthly based methodologies. Furthermore, we implemented those new best models in an open-source REN4KAST platform, and we provided the implementation details. Note that such a platform provides services (data retrieval, forecasting, evaluation), with respect to the percentage of renewables in Germany. We then carried out experimental analysis, by considering four years of data from Germany. We compared the observed (i.e., real data) to the forecast (i.e., models derived from our proposed hybrid methodology) values, and showed that the average annual RMSE for the four years was about 10.5%. To provide evidence of those improvements, we showed the distribution of the error rates for each month of the year. Our results showed that, in general, the months of May to September had higher error rates than the other months of the year, with the month of June suffering the highest error rates, of about 16%: we will reserve, for our future work, investigation of the reasons. Also, as a future work, we will investigate the impact of anomalies (i.e., outliers and missing values) in our dataset on the accuracy of the derived ARIMA-based forecasting models.