Next Article in Journal
Towards Cognition-Aligned Visual Language Models via Zero-Shot Instance Retrieval
Previous Article in Journal
Improving Real-Time Performance of Micro-ROS with Priority-Driven Chain-Aware Scheduling
Previous Article in Special Issue
Robust Nonsingular Terminal Sliding Mode Control of a Buck Converter Feeding a Constant Power Load
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Two-Stage Neural Network Optimization for Robust Solar Photovoltaic Forecasting

1
School of Electrical Engineering, Korea University, Seoul 02841, Republic of Korea
2
Department of ICT Convergence, Soonchunhyang University, Asan 31538, Republic of Korea
3
Department of AI and Big Data, Soonchunhyang University, Asan 31538, Republic of Korea
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(9), 1659; https://doi.org/10.3390/electronics13091659
Submission received: 5 March 2024 / Revised: 11 April 2024 / Accepted: 23 April 2024 / Published: 25 April 2024

Abstract

:
Neural networks (NNs) have shown outstanding performance in solar photovoltaic (PV) power forecasting due to their ability to effectively learn unstable environmental variables and their complex interactions. However, NNs are limited in their practical industrial application in the energy sector because the optimization of the model structure or hyperparameters is a complex and time-consuming task. This paper proposes a two-stage NN optimization method for robust solar PV power forecasting. First, the solar PV power dataset is divided into training and test sets. In the training set, several NN models with different numbers of hidden layers are constructed, and Optuna is applied to select the optimal hyperparameter values for each model. Next, the optimized NN models for each layer are used to generate estimation and prediction values with fivefold cross-validation on the training and test sets, respectively. Finally, a random forest is used to learn the estimation values, and the prediction values from the test set are used as input to predict the final solar PV power. As a result of experiments in the Incheon area, the proposed method is not only easy to model but also outperforms several forecasting models. As a case in point, with the New-Incheon Sonae dataset—one of three from various Incheon locations—the proposed method achieved an average mean absolute error (MAE) of 149.53 kW and root mean squared error (RMSE) of 202.00 kW. These figures significantly outperform the benchmarks of attention mechanism-based deep learning models, with average scores of 169.87 kW for MAE and 232.55 kW for RMSE, signaling an advance that is expected to make a significant contribution to South Korea’s energy industry.

1. Introduction

Global climate change and sustainable development are two of the most pressing issues related to energy production and are receiving a great deal of attention around the world [1]. While energy consumption continues to grow, most energy production still relies on fossil fuels, including natural gas, coal, and oil [2]. The excessive use of these fuels is increasing greenhouse gas (GHG) emissions and causing climate anomalies around the world [3]. If our reliance on fossil fuels continues, GHG emissions are expected to increase by 30% over the next 20 years, increasing the potential for energy crises and environmental risks [4]. In response to these challenges, researchers and governments are intensifying their exploration of renewable energy resources [5]. Given the high rate of GHG emissions, renewable energy is considered a viable alternative for future energy management [4,5]. The use of renewable energy sources such as solar, wind, and biomass is considered a sustainable solution from an environmental perspective [1,2].
Due to its non-polluting and sustainable/renewable characteristics, photovoltaic (PV) energy has received considerable attention as a solution to climate change, efficiently producing energy through the harmonious operation of its main components, as shown in Table 1 [6]. In this context, solar PV power generation is an innovative technology that directly converts the energy of sunlight into electricity [7]. It is based on the principle of converting the flow of electrons generated by solar panels through the photoelectric effect into electric current, and this direct current needs to be converted into alternating current that can be used in our daily lives [6,7]. In addition, solar PV power generation has the advantage of being easy to install, having infinite resources, being environmentally friendly, and being able to generate electricity without noise or pollutant emissions compared to other renewable energy sources [8]. Because of these characteristics, the proportion of solar PV generation and use is expected to continue to increase in the future [7,8].
Predicting solar PV power generation is very important for the operation of a sustainable energy system, including maintaining a balance between energy production and consumption and optimal utilization of energy resources [9]. However, compared to traditional energy sources, the dependence of solar PV power generation on environmental factors (e.g., sensitivity to changes in weather conditions) introduces significant variability. This variability, coupled with the dynamic nature of these environmental influences over time, makes the task of accurately predicting solar PV power generation extremely challenging [8,9]. For a complete understanding of the terms and acronyms used in this context, please refer to Appendix A, which provides definitions for key acronyms and variables discussed herein. From the perspective of technological progress, solar PV power generation forecasting has evolved significantly over time [10,11].
Initially, traditional forecasting methods such as simple statistical and physical models were widely used. While effective at capturing linear relationships, these methods were often inadequate for modeling the complex, nonlinear patterns that characterize solar energy production [11]. Subsequently, with the gradual introduction of machine learning (ML) techniques, the ability to learn and predict complex data patterns has improved. For example, advances in ML and the subsequent integration of deep learning (DL) algorithms have significantly surpassed the capabilities of traditional forecasting methods, enabling more accurate and reliable predictions. These advances have played a key role in improving the accuracy of solar PV power generation forecasting, thereby contributing to more effective energy management and stable operation of the power grid [12,13,14,15,16,17,18,19].
Qazi et al. [12] developed an artificial neural network (ANN)-based solar irradiance prediction method based on weather data analysis. Wang et al. [13] proposed a forecasting scheme using nonlinear time series analysis to predict solar power in smart grids. Van der Meer et al. [14] used two computational methods, autoregressive integrated moving average (ARIMA) models and mixed integer linear programming, for PV power forecasting and optimal electric vehicle (EV) charging management, respectively. Voyant et al. [15] developed and compared several ML algorithms, such as nearest neighbor and bootstrap aggregation, to accurately predict solar irradiation. Almonacid et al. [16] used ANNs to demonstrate prediction performance metrics of concentrator PV technology. Wang et al. [17] proposed two ML methods based on support vector machine (SVM) and adaptive learning to predict solar irradiance with high accuracy. Zhang et al. [18] developed a Bayesian network-based prediction scheme to perform spatio-temporal analysis for advanced short-term prediction of PV power generation. Al-Dahidi et al. [19] proposed a day-ahead solar PV power generation prediction method based on an ensemble of diversified ANNs optimized with bagging and bootstrapping techniques.
In particular, DL models have received much attention due to their structure and learning method, which are excellent for processing data with complex patterns [20,21,22,23,24,25]. Abdel-Nasser and Mahmoud [20] developed a PV power prediction model using deep long short-term memory (LSTM) networks, which captures temporal dynamics with high accuracy. Patel [21] proposed a solar irradiance prediction based on the exploration of LSTM and convolutional neural networks (CNNs) to overcome the challenges of model convergence and accuracy. Wen et al. [22] developed a novel multistep-ahead prediction scheme for PV power rate control using deep CNNs and stacked sky images, bypassing the need for additional time series models and exogenous variables for industrial suitability. Jiao et al. [23] proposed a novel DL architecture for solar irradiance prediction that integrates convolutional graph neural networks with LSTM to improve accuracy and reliability for distributed PV systems. Zameer et al. [24] proposed two DL models based on bidirectional LSTM (Bi-LSTM) and gated recurrent unit (GRU) to achieve superiority in short-term solar PV prediction over traditional methods such as lasso, ridge, elastic net, and SVM, highlighting the robustness and precision of DL. Rocha et al. [25] conducted an in-depth analysis using LSTM, Bi-LSTM, and temporal convolutional network (TCN) for predicting solar PV generation on a 1320 watt-peak (Wp) amorphous plant, demonstrating the superior performance of TCN in terms of accuracy for both short-term (15-minute) and long-term (24-h) predictions.
However, the reliability of current solar PV power generation forecasting is often compromised by several factors, leading to inaccuracies. The application of advanced DL techniques as baseline models is often hampered by the need for high-performance computing resources, such as graphics processing units (GPUs) [26]. In addition, the practicality of implementing such advanced models is further limited in environments where energy managers may find the complexity of ML and DL daunting, and where energy companies or government agencies do not always have the budget or infrastructure to support high-end computing resources. DL techniques require the management of multiple hyperparameters, which adds complexity to the optimization process and can hinder model learning [27]. This complexity may limit practical applications in the energy sector, where DL expertise is not widely available. Identifying the optimal configuration of neural networks (NNs), such as the appropriate number of hidden layers, is a significant challenge and research gap that affects prediction accuracy [27,28]. Simply increasing the number of hidden layers does not guarantee improved model performance [28], and what works in one scenario may not be effective in another [29]. Consequently, researchers have attempted to improve prediction performance by integrating multiple deep neural network (DNN) models with different numbers of hidden layers, a strategy that merits further exploration [29]. In the face of these challenges, stacking ensemble learning emerges as a powerful solution that is particularly adept at overcoming the twin challenges of technical complexity and resource constraints [29,30,31,32].
Bâra and Oprea [30] incorporated weather forecast errors into the forecasting method for various PV systems, including grid-tied, hybrid, and off-grid setups. This study demonstrated that stacking ensemble learning could adjust PV forecasts based on these errors, leading to significant improvements in forecast accuracy across different PV system sizes and connectivity. Another investigation [31] proposed a stacking ensemble learning method that combines ML algorithms such as random forest (RF) and extreme gradient boosting with DL techniques such as DNNs and GRUs. This approach used weather data and system performance to predict PV generation for both on-grid and off-grid systems and showed that incorporating battery load and state of charge significantly improved the model’s ability to accurately estimate potential power output. Further exploration [32] developed a stacking ensemble learning method that integrates deterministic and stochastic models to address the complexity of PV system forecasting. This method proved to be particularly effective in reducing error rates and improving forecast reliability for various PV system types, demonstrating the ability of stacking ensemble learning to mitigate the limitations of individual forecasting models.
In addition, the effectiveness of traditional time-series forecasting models tends to decrease with the time since the last training, indicating the need for innovative approaches such as online learning [33]. To overcome these problems, we propose a novel online learning-based solar PV power generation forecasting model using two-stage NN optimization. This model aims to facilitate accurate and fast multistep-ahead forecasting in computing environments without GPU support, with a focus on Incheon’s smart city development [34]. This study raises important research questions regarding the effectiveness of stacking ensemble models in improving prediction accuracy and their performance in environments with limited computing resources. The proposed method addresses key issues and makes a significant contribution to the field with an online learning-based prediction model that improves both accuracy and computational efficiency, as detailed in the main contributions of this paper:
  • We use Optuna for hyperparameter optimization within scikit-learn’s MLPRegressor and implement fivefold cross-validation to ensure DNN model development with sufficient training data, even on low-performance systems.
  • We develop an online learning approach for the meta-regression model using a Ranger-based RF model, which ensures accurate predictions over time despite the increasing gap between training and evaluation periods.
  • By interpreting variable importance from the RF, we highlight the contributions of specific DNN models, enhancing the practical understanding for energy managers and the applicability of the proposed method.
  • We demonstrate the superiority of the proposed method through 11-point multistep-ahead forecasting, minimizing uncertainty and outperforming various statistical and DL models in mean absolute error (MAE) and root mean square error (RMSE).
The remainder of this paper is organized as follows. Section 2 introduces the dataset composition and preprocessing, as well as the methodology and experimental setup. Section 3 presents the results of the experiments, which demonstrate the effectiveness of the proposed method. Section 4 discusses the limitations of the study and future directions. Finally, Section 5 summarizes the main results and contributions of the work.

2. Materials and Methods

2.1. Dataset Description

2.1.1. Dataset Collection

Incheon [35], also known as Incheon Metropolitan City, is a major international metropolis in South Korea that has established itself as a pioneer in global exchange and creative urban planning ideas. Since opening as an international port in 1883, Incheon has grown into a world-class city with a population of more than three million people. Connected to various parts of the world through places such as Incheon International Airport and Incheon Port, Incheon serves as a transportation hub in Northeast Asia. Incheon also hosted the 2014 Asian Games and is known as a smart city. The city, along with Seoul and Busan, received a high rating in the 2022 Smart City Index released [36]. This rating shows how much Incheon cares about smart city management and infrastructure services. In particular, Incheon is making efforts to actively utilize new and renewable energy, especially solar power generation [37].
Therefore, we used solar PV power generation data from the Incheon area located on the west coast of South Korea in this study. We also decided to use the solar PV power generation dataset from Incheon for the following reasons: First, Incheon’s location on the west coast of South Korea makes it the perfect place for high solar irradiance and relatively stable sunshine, especially in the summer. Second, Incheon has a temperate climate, with cold, dry winters and hot, humid summers, so a wide range of seasonal variations can be taken into account when analyzing and optimizing the performance of solar PV systems. In addition, Incheon’s strong commitment to renewable energy and smart city implementation ensures excellent reliability in terms of data quality and coverage. This is essential for this study into the benefits of solar energy and its function in smart cities.
In this paper, we used solar PV power generation data collected from three solar power plants in Incheon: New-Incheon Sonae, New-Incheon Observatory, and Incheon Fishery Water Purification Plant. These data are publicly available data collected through the Korea Public Data Portal [38] and were collected at 11 data points per day from 8:00 a.m. to 6:00 p.m. at hourly intervals from 1 January 2018, to 31 December 2019. We divided the data in a 5:5 ratio from 1 January 2018, to 31 December 2018, as the training set and from 1 January 2019, to 31 December 2019, as the test set. Table 2 shows the statistical analysis of the data collected from each location.
The Incheon Fishery Water Purification Plant appears to have the most efficient PV system of the three, based on data showing that it has the highest average solar energy output throughout the training and test sets. The New-Incheon Sonae, on the other hand, shows reduced energy output, with a significant drop in the test set, suggesting possible unpredictability or seasonal effects on performance. Nevertheless, the energy production statistics appear to be fairly evenly distributed with minimal fluctuations, as indicated by the observed kurtosis and skewness across all installations, indicating stable operating conditions.

2.1.2. Dataset Preprocessing

In order to construct a forecasting model using collected data, it is essential to consider various external variables that have a significant impact on solar PV power generation. In this paper, we considered time stamps and weather information as input variables to develop a robust solar PV power generation forecasting model. These factors are very important because they directly affect the efficiency and performance of solar PV systems, and they are variables that can be realistically considered in real industrial applications in South Korea [39].
Time data consists of month, day, and hour data. Time data are a one-dimensional form of data, that can be challenging to use because it is difficult to accurately reflect its periodicity. For example, 30 November and 1 December are actually adjacent dates and times, but in the one-dimensional form of day data, there is a numerical difference of 29. In this paper, we applied Equations (1)–(4) to augment the one-dimensional month and day data with continuous data in two-dimensional space to reflect the periodicity of the data [39,40].
MonthX = sin(360°/12 × Month),
MonthY = cos(360°/12 × Month),
DayX = sin(360°/LDMMonth × Day),
DayY = cos(360°/LDMMonth × Day),
where LDMMonth refers to the last day of the month to which a given day belongs. For example, if it is June, the LDMMonth is 30. If it is July, the LDMMonth is 31. This information is useful for determining what month a given day belongs to and how many days are in that month.
However, for the time period from sunset to sunrise, 0 is recorded regardless of any weather value. Therefore, this study considered the time period from 8:00 a.m. to 6:00 p.m. It is difficult to reflect the periodicity when reduced to 11 variables (8–18) and converted into two dimensions; therefore, 11 variables were generated by one-hot encoding.
The weather data essential for this study, especially for the prediction of solar PV power generation at three locations in Incheon, were obtained from the Incheon Meteorological Station, located at 37.47772 latitude and 126.6249 longitude, operated by the Korea Meteorological Administration (KMA). These data, which are critical to the input variables in DNN models, were obtained from the Weather Data Service’s Meteorological Data Portal [41]. The KMA provides different types of weather forecasts, such as ultra-short-term forecasts, short-term forecasts, medium-term forecasts, and long-term forecasts. The ultra-short-term forecast predicts weather elements such as temperature, precipitation, and lightning strikes at one-hour intervals for up to four hours ahead, providing instant weather insights, while the short-term forecast extends predictions for variables such as temperature, humidity, wind speed, and precipitation up to three days ahead, with updates every three hours [42].
Since a solar PV power generation forecasting model needs to have forecasts at least one day ahead for effective power utilization planning [43], we specifically used daily average temperature, daily minimum temperature, daily maximum temperature, air temperature, humidity, wind speed, and precipitation data from the KMA short-term forecast [39]. Although the forecasts are available three days in advance, for the purpose of this study, we only focus on using the forecasts provided one day in advance to improve the accuracy and reliability of the day-ahead forecasts. The solar PV power generation data, which are collected in 1-h increments, contrast with the KMA short-term forecast data, which are initially available in 3-h increments. For these purposes, these data have been cleverly converted to hourly intervals using linear interpolation, which improves the temporal resolution and applicability for solar PV power forecasting.
A key challenge for KMA short-term forecasts is the lack of solar irradiance data, which are critical for predicting solar PV power generation. To address this, we adopted a creative strategy to indirectly simulate the effects of solar irradiance on the proposed method. By examining how solar power generation changes with different weather conditions, such as season and time of day, we could estimate the role of solar irradiance in power generation. In addition, to compensate for the lack of direct solar irradiance data and to improve the predictive accuracy of our model, we developed multiple DNN models and applied the stacking ensemble method. This innovative two-step strategy enhances the ability of our model to mimic the effects of solar irradiance and significantly improves the prediction accuracy. As a result, this approach allowed us to build a comprehensive model with 22 carefully selected input variables, as shown in Table 3.
Our dataset includes only external variables, such as weather conditions and time stamps, and intentionally omits internal factors. This choice ensures that the dataset is broadly applicable across South Korea, without being limited by the specific details of individual solar PV systems. This universal applicability is key to developing effective energy harvesting strategies that can be easily adapted to different solar PV installations in South Korea. Thus, the dataset configuration can provide great value by supporting the creation of predictive models that are both powerful and widely applicable, helping power producers improve their operations and increase the environmental friendliness of their power generation.
Min–max normalization can make all input variables similar by putting them on a scale from 0 to 1 [44]. This technique allows the model to learn more accurately and overcome any shortcomings in the data. We used Equation (5) to train the consistency of the DL-based solar PV power generation prediction model by applying min–max normalization to all input variables, including time stamps and weather information.
XNorm = (XXMin)/(XMaxXMin),
where XNorm represents the new value of a number in the set of observations X after normalization. The minimum and maximum values of X are represented by the variables XMin and XMax. We applied them to the training set before applying the XNorm values from the training set to the test set. As a result, the minimum and maximum values of an input variable can be zero and one.
Figure 1 shows a synchronized overview of normalized meteorological variables and solar PV power generation at three different locations in Incheon for July 2019. Each variable—temperature, humidity, wind speed, precipitation, and PV power generation for three solar power plants in Incheon—is represented by a min–max normalization to ensure an equitable scale of 0 to 1. This normalization technique allows us to identify patterns and correlations in the fluctuating climate of Incheon’s monsoon season and its subsequent impact on PV power generation. The hourly time index represents a sequential timeline that captures the intricacies of solar energy variability in a region characterized by strong climatic transitions. This graphical representation not only illustrates the temporal distribution of each parameter but also serves as a powerful visual reinforcement of the variability inherent in solar energy systems, especially during periods of climatic extremes such as the onset of the East Asian Monsoon [45].
We ensured that the processed datasets were designed to enhance replicability and further research. To this end, the preprocessed datasets are publicly available for those who wish to use them in model training or additional studies. Researchers can find these datasets in the Supplementary Material of this paper, specifically in Tables S1–S3. This step is taken to underscore our commitment to openness and the advancement of scientific inquiry in the energy field.

2.2. Model Construction

Figure 2 shows the overall architecture of the proposed solar PV power generation forecasting method. First, we divided the preprocessed dataset into a training set (2018) and a test set (2019) in a 5:5 ratio as described earlier. We performed two stages of optimization techniques on nine DNN models and derived final predictions. The details of this study are described in the following subsections.

2.2.1. Stage 1: Optimized Deep Neural Network Architecture with Optuna

A multilayer perceptron (MLP), also known as a typical NN configuration, allows computers to learn tasks by example, as shown in Figure 3 [46]. It is similar to a schematic that guides the computer in formulating a solution to a problem. The framework of an MLP consists of individual elements, called perceptrons, organized in layers. An MLP conventionally includes three types of layers: the input layer, which captures data that the computer uses to make a prediction; one or more hidden layers, which relevant pertinent information from the input layer to refine the prediction; and the output layer, which provides the final result based on the synthesized information from the hidden layers [46,47]. We used Equation (6), which integrates the input data, neural connections (weights), and an activation function, to compute the output values of the neurons in the hidden layer of an MLP.
yj = φ(∑wij × xi + bi)
The input layer of an MLP consists of neurons, each of which is represented by xi. Synaptic weights, abbreviated as wij, reflect the connections between neurons in the input and hidden layers. The activation function (φ) controls the output of the hidden layer neurons. When an MLP contains more than two hidden layers, it is referred to as a DNN, indicating a deeper level of computational complexity [47,48]. Biases (bi) are added to the inputs of the activation function to help the network make more accurate predictions. Weights serve as important links not only between the input and hidden layers, but also between multiple hidden layers and between the last hidden layer and the output layer, and they play an important role in information processing and modeling. A feature of MLPs is their ability to process information in a distributed manner, with each neuron acting independently. This makes the system fault-tolerant because it can continue to learn and process data even if some neurons fail [48].
MLPRegressor is one of the components of scikit-learn, the widely used ML package in Python. It is a DL model for regression tasks that is adept at predicting numerical outputs from given input features [49]. While DNN models are typically constructed using frameworks such as TensorFlow, Keras, or PyTorch, which excel in GPU environments, MLPRegressor provides an effective benchmark for utilities and government organizations managing solar PV systems, particularly in scenarios where investment in advanced computing infrastructure is not feasible. MLPRegressor facilitates efficient model deployment on lower-capacity systems, underscoring its role as a practical solution in resource-constrained environments.
In addition, MLPRegressor can produce excellent results on structured datasets [29,48]. Its performance can be significantly improved by tuning key hyperparameters: hidden layer sizes, activation functions, batch size, learning rate, and maximum iterations. Therefore, to address the challenge of solar PV power generation prediction on low-power systems, we decided to use scikit-learn’s MLPRegressor and adjust these hyperparameters for optimal results. This approach balances the need for cost efficiency and effective use of computational resources, ensuring that our model is widely applicable and can operate effectively in constrained computing environments, making it an excellent benchmark model.
The configuration of the hidden layer has a significant impact on the efficiency of the network, highlighting the importance of considering the number of layers and nodes, as well as the activation function, when designing the network [28]. Although increasing the number of hidden layers often improves network efficiency, it also increases the possibility of overfitting [28,29]. Identifying the optimal number of hidden layers to prevent overfitting can be particularly challenging, especially for those not specialized in ML and DL, such as energy experts. This complexity arises because the development of DL models can yield different results based on different training conditions, and is particularly resource-intensive [29]. To mitigate these challenges, we first built several DNN models with different hidden layer structures. For our prediction efforts, we created nine DNN models with the number of layers ranging from 4 to 12 within the training set.
Traditional hyperparameter tuning techniques have been extended to include gradient-based optimization, evolutionary algorithms, simulated annealing, particle swarm optimization, hyperband, gradient-based optimization, random search, and Bayesian optimization to improve DL models [50,51]. While these methods expand the range of tools available to improve model performance, they often share the same drawbacks: high computational requirements, significant time commitment, and a complexity that can be intimidating to practitioners not trained in ML, such as those working in the energy industry. Under severe time and computational resource constraints, these approaches may not provide the best model efficiency and scalability due to their thorough and often random search through hyperparameter spaces.
Optuna [48,52], a versatile Python toolkit, provides a unified interface for refining hyperparameters in ML constructs to increase their effectiveness. Hyperparameters are the pre-training configurations of algorithms, as opposed to the parameters that are adjusted during the training process itself. Within the architecture of an NN, critical hyperparameters include the number of hidden layers, the learning rate, and the intensity of regularization. Designed to streamline the hyperparameter optimization process, Optuna extends an intuitive and customizable application programming interface for architecting ML models and defining the scope of hyperparameter exploration. This suite uses a variety of methods to traverse the hyperparameter terrain, including the Tree-structured Parzen Estimator, as well as random and grid searches, to find the best hyperparameters that improve metrics such as accuracy or F1 score.
Optuna is unique in its ability to synchronize simultaneous trials, which are different evaluations of a model configured with different hyperparameter combinations. This feature drives Optuna’s ability to traverse the hyperparameter landscape with increased speed and precision. A key feature of Optuna is its early termination mechanism, which allows the optimization effort to be halted when the likelihood of further trials refining the results is minimal, thereby optimizing resource utilization and mitigating the risk of overfitting. In addition, Optuna includes a tracking API that provides insight into the trajectory of the optimization and detailed records of each trial, essential for the analysis and refinement of the optimization strategy. We defined the hyperparameter search space as shown in Table 4 and set the number of iterations in Optuna to 1000. We then used Optuna through fivefold cross-validation on the training set to determine the optimal hyperparameter values for each DNN model, as shown in Figure 4.
In this paper, we constructed nine DNN models with 2 to 10 hidden layers in the training set and carefully determined the optimal hyperparameter settings for each model using Optuna’s fivefold cross-validation. This preliminary step ensured that we identified the most effective configurations to improve the predictive accuracy of our models.
In addition, we took a strategic approach to the challenge of constructing a sufficient validation set within a limited timeframe, especially when a separate validation set may not have been feasible due to the constraints of not having an extended training period. We performed another round of fivefold cross-validation on the training set based on these optimized hyperparameters to prepare adequate training data for our meta-learning model within the stacking ensemble framework. This method allowed us to ensure robust learning from even a one-year training dataset, effectively overcoming potential limitations in the availability of validation datasets. In parallel, we used the entire training dataset, now fully trained with the optimized hyperparameters, to make preliminary solar PV power generation predictions on the test set. This comprehensive approach was able to maximize the predictive capabilities of our models under the specific constraints of the study.

2.2.2. Stage 2: Ranger-Based Online Learning Model for Multistep-Ahead Prediction

Even with DNN models optimized by Optuna, the high computational load of DNNs for online learning poses a challenge, especially for real-time energy management. To address this issue, our stacking ensemble method, which integrates different DNNs with an RF meta-regression model, leverages the predictive capabilities of DNNs while mitigating their computational demands. This approach provides significant benefits to energy managers by enabling efficient, real-time prediction of solar PV power generation, which is critical for optimizing energy distribution and demand forecasting. By combining the analytical depth of DNNs with the operational efficiency of RF, the proposed method can provide a practical solution for advancing energy management practices, highlighting the originality and applicability of our approach in the energy sector.
RF [53,54] is a prime example of a bagging ensemble technique. It constructs multiple decision trees (DTs) trained on randomly selected subsets of variables (features) and outputs the result by averaging (for regression tasks) or majority voting (for classification tasks) over the predictions of all DTs [54]. Efficient for online learning and requiring minimal hyperparameter tuning, RF works efficiently on large datasets and shows excellent performance in areas such as bioinformatics and time series prediction because it can handle large numbers of input variables without the need for variable elimination [55,56]. Its robust performance and ability to handle multicollinearity make it suitable for handling highly correlated inputs from DNN models. In addition, RF generally requires less hyperparameter tuning than other ML methods and often provides satisfactory performance even with default hyperparameter settings [57]. We focus on two commonly adjusted RF hyperparameters for model training: the number of trees to grow (n_estimators for Python and nTree for R) and the number of variables considered for partitioning at each tree node (max_features for Python and mTry for R) [56,57].
We selected the RF algorithm based on its adaptability to the specific needs and constraints of this study, rather than on claims of universal superiority [58,59]. The implementation of online learning, which selects optimal hyperparameters at each time point for model training and prediction, poses significant challenges in real-world industrial settings due to its computational and time requirements [33]. RF’s adaptability and ease of tuning are particularly valuable in these settings. RF is characterized by robust performance with minimal need for hyperparameter optimization, in contrast to DL models that often require careful tuning of hyperparameters to achieve peak performance. Furthermore, its effectiveness as a meta-regression model in a stacking ensemble highlights its versatility. In addition, RF correctly handles multicollinearity, ensuring model robustness for highly correlated input variables [60]. This feature allows for more effective learning and final prediction of solar PV power generation estimation and prediction values in the training and test sets derived from DNN models. Therefore, the adoption of RF was based on its balanced learning capabilities and conformity with the requirements and limitations of the study [61,62].
The RandomForestRegressor implementation in Python and the randomForest package in R are widely appreciated for their rich functionality. However, they are not ideally optimized for high-dimensional datasets. Furthermore, a major drawback is the slow execution speed of the model. The Ranger, an innovative software package, has been developed to overcome these limitations and increase efficiency for high-dimensional data analysis [63]. Efficient memory management is achieved by minimizing data duplication, storing node information in compact data structures, and freeing memory in a timely manner. The Ranger package introduces two innovative partitioning methods. The first method sorts feature values by index before partitioning, while the second method splits raw values directly and organizes them afterward. These methods optimize the runtime for the main and minor nodes used in the first and second scenarios, respectively. These runtime optimization methods make Ranger a powerful tool for RF implementations in complex scenarios.
The essence of stacking ensemble learning is to orchestrate a diverse collection of models, such as DNNs, to work in parallel and synergize their predictions through a meta-regression framework, such as RF, to refine the overall prediction accuracy [29,64]. However, because many of these weak learners are based on ML or DL methods that rely heavily on data characteristics, stacking ensemble learning may have limitations in improving solar PV power generation predictions. As a result, ML or DL methods may struggle to achieve high accuracy on out-of-sample (unseen) datasets that differ significantly from the in-sample (training) data. For example, if the observed PV power in the test dataset significantly exceeds that in the training dataset, a model calibrated to the latter may not effectively predict escalation [61]. In addition, achieving satisfactory predictive performance with insufficient data becomes difficult, especially when the available training data are scarce [61,65].
We used R Ranger to construct an online learning model that uses time series cross-validation (TSCV) to address these challenges [65]. TSCV emphasizes the evaluation of multiple prediction horizons within each training set. Additional training sets were used according to the planned time period, each containing one observation not included in the previous training set. We designed the test set to range from one hour to one day (up to 11 h) after the current period to allow for multistep-ahead prediction of solar PV power generation, as shown in Figure 5. The prediction accuracy was calculated at each time point and the results were averaged to evaluate the effectiveness of the prediction model. To train the Ranger model, nine input variables were selected, encapsulating predictions from DNNs with 2 to 10 hidden layers. In addition, in accordance with best practices and to tailor our model specifically for regression tasks, we set the number of trees to 128 and the number of features to 3, reflecting one-third of our nine input variables for optimized regression [57,66].
The proposed method excels in predicting solar PV power output by combining ensemble learning with TSCV. This approach could overcome the limitations of static DL algorithms such as DNN, LSTM, Bi-LSTM, GRU, and one-dimensional CNN (1D-CNN) by progressively incorporating fresh data to gradually increase the prediction accuracy. The proposed method captures complicated nonlinear interactions in the data by using a two-stage NN optimization that incorporates different DNNs in the first stage and an RF algorithm as a meta-regression model in the second stage. It also uses incremental learning to adapt to new trends and patterns as more data becomes available. Most importantly, this approach is optimized for central processing unit (CPU) usage, ensuring broad application in real-world environments with limited computing resources. As a result, even with a small initial dataset, the proposed method ensures robust learning and accurate prediction, which is important for ML applications that collect data over time.

3. Results

3.1. Experimental Setup

In this research, we established an experimental framework to test our proposed approach for solar PV power generation forecasting 11 steps (one day) in Incheon, South Korea. The experimental hardware consisted of an Intel(R) Core(TM) i7-9700 CPU @3.00 GHz and 64 GB RAM. Software tools included Anaconda 22.9.0 and Python 3.8.0 for data preprocessing, and RStudio version R-4.2.2 along with R version 4.2.2 (2022-10-31 UCRT) for additional data processing and development of the Ranger-based online learning model. This setup was critical for evaluating and improving the accuracy of the solar PV power generation prediction model. We set the random state and the random seed to 42 and 1234 in Python and R, respectively, to ensure the reproducibility of the experiment and to allow direct comparison of models trained under identical conditions.
Two statistical measures were used to evaluate the accuracy of the prediction models: MAE, which is calculated as the average of the absolute differences between the predicted and actual values, and RMSE, which measures the square root of the average squared differences between the predicted and actual values, defined by the following Equations (7) and (8), respectively:
MAE = ∑|PtAt|/n,
RMSE = √(∑(PtAt)2)/n,
where At and Pt denote the actual and predicted values at time t, n is the number of observations.

3.2. Experimental Results

In this study, we meticulously conducted experiments using the New-Incheon Sonae, New-Incheon Observatory, and Incheon Fishery Water Purification Plant test sets (2019) from Incheon, South Korea as unseen data to evaluate the performance of our proposed model. Our main objective was to compare our proposed model with a set of benchmark models, which had previously demonstrated remarkable performance in predicting solar PV power generation. First, we constructed an RF model using Ranger as a benchmark model for multistep-ahead forecasting, using TSCV with 22 input variables. Then, we constructed DL models for multistep-ahead forecasting, starting with robust time series forecasting models from the RNN family, including LSTM, Bi-LSTM, and GRU models [23]. In addition, we developed seven state-of-the-art models that incorporate attention mechanisms (Att) [67,68,69,70,71,72,73]. These steps allowed us to comprehensively compare the established benchmark models, known for their effectiveness in various domains, with our proposed model. To create a comparable environment for multistep-ahead solar PV power generation forecasting, we configured the time steps from 1 to 11 using a many-to-many approach. We adhered to the core architectural frameworks of the models as described in the research articles, and fine-tuned the hyperparameters using Optuna to ensure optimal performance.
The benchmark models include:
  • Att-LSTM: Zhou et al. [67] developed the Att-LSTM model using a hybrid ensemble DL framework with two LSTM NNs focused on temperature and power output for short-term PV power generation. The incorporation of an attention mechanism improved the prediction accuracy by prioritizing significant input features.
  • Att-Bi-LSTM: He et al. [68] constructed the Att-Bi-LSTM model, which combines a Bi-LSTM architecture and attention mechanisms for predicting PV power generation from solar PV systems, and demonstrated superior accuracy by selecting environmental factors using the Pearson coefficient.
  • Att-GRU: Jung et al. [69] introduced the Att-GRU-based energy forecasting model, an attention-based GRU approach for multistep-ahead electrical load forecasting, which outperforms other models by emphasizing crucial variables in response to sudden changes in power consumption.
  • Att-1D-CNN: Wu et al. [70] proposed the Att-1D-CNN model, which uses a CNN and informer approach, and achieves accurate solar PV power generation prediction by analyzing data correlations, significantly improving the prediction accuracy over traditional models.
  • Att-TCN: Ren et al. [71] introduced the Att-TCN model, a novel dual-channel TCN that combines a multi-head attention mechanism with TCN to extract spatiotemporal features, and demonstrated superior prediction performance for solar PV power generation forecasting.
  • Att-GRU-TCN: Xiao et al. [72] proposed the Att-GRU-TCN model for probabilistic power flow in multi-microgrids, which integrates GRU and TCN with an attention mechanism to improve reliability and efficiency in networks with incomplete information.
  • Att-Bi-GRU-TCN: Zhou et al. [73] developed the Att-Bi-GRU-TCN model for short-term solar PV power generation forecasting, which uses a two-step approach with TCN and Bi-GRU to capture both short-term and long-term dependencies, ensuring robust forecasting accuracy.
In Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10, we observed the exceptional performance of the proposed method. This model showed excellent predictive capabilities with the smallest MAE and RMSE values among the multistep-ahead predictions. For example, the proposed method showed an average MAE of 149.53 kW and RMSE of 202.00 kW for the New-Incheon Sonae dataset, which is significantly lower than the average MAE of up to 171.11 kW and RMSE of up to 233.66 kW of the benchmark models in Table 5. The model’s effective implementation of online learning and a sophisticated stacking ensemble approach can achieve a high level of accuracy in predicting solar PV power generation for the New-Incheon Sonae, New-Incheon Observatory, and Incheon Fishery Water Purification Plant datasets. The stacking ensemble technique could effectively combine the predictive strengths of different DNN models, resulting in a consistent and fine-tuned prediction.
Notably, both our model and the Ranger model, which incorporates online learning, routinely outperform typical DL models. The results may demonstrate the effectiveness of online learning in adapting to new, real-time data, which is an important feature for models operating in dynamic situations. However, it is critical to understand the constraints experienced. The DL models struggled to adapt to a training set spanning one year, indicating a weakness in their ability to learn from large amounts of temporal data in a short period of time. In addition, the lack of solar irradiance as an input variable limits the ability to improve prediction accuracy. Since solar irradiance is such an important aspect of PV power generation, its inclusion in future models could lead to significant improvements in prediction accuracy.
To evaluate our proposed model for single-day-ahead forecasting in Incheon, we compared it with several benchmark models. For a thorough evaluation, we analyzed the performance over different time periods, including quarterly and overall metrics. We evaluated the 11-step-ahead (day-ahead) forecast values of the proposed method by measuring the forecast accuracy. First, we adopted a statistical method from the persistence model, which relies on data from the corresponding time of the previous day. For example, to forecast solar PV power generation at 14:00 on June 8, we used data from 14:00 on June 7. We also used the hold-out technique to construct the Ranger-based RF model with 22 input variables. In addition, we evaluated the prediction values of nine DNN models configured for one-step prediction, whose performance was enhanced by hyperparameter optimization using Optuna.
In the extended analysis, we found a striking trend: over time, our model with online learning widens the performance gap in prediction accuracy over other models, as shown in Table 11, Table 12 and Table 13. In particular, the proposed method showed a consistent improvement in accuracy for day-ahead solar PV power generation forecasting, with a significant reduction in MAE and RMSE values. In the New-Incheon Sonae dataset, this model maintained a lower average MAE and RMSE of 151.66 kW and 204.72 kW, respectively, compared to other DNN models with higher layers, which showed MAEs and RMSEs increasing to 185.54 kW and 255.43 kW, respectively. This pattern indicates that the proposed method’s online learning capability is not only effective in adapting to immediate data trends but also improves its prediction accuracy over time, outperforming benchmark models. We concluded that the model’s ability to continuously learn and integrate new information plays a critical role in its sustained superior performance.
Table 14, Table 15 and Table 16 present the results of a comprehensive evaluation of the day-ahead forecasting models, showing their varying effectiveness under different weather conditions. Notably, the proposed method consistently outperformed not only the benchmark Persistence model but also other models, such as Ranger and various DNN configurations, in terms of accuracy under clear weather conditions by demonstrating lower MAE and RMSE. Specifically, for the New-Incheon Sonae dataset, the proposed method achieved an MAE of 141.17 kW and an RMSE of 189.96 kW, surpassing the Persistence model’s MAE and RMSE of 148.04 kW and 226.16 kW, respectively, as well as those of the Ranger and DNN models.
Conversely, the performance of the proposed method under rainfall conditions dropped below that of the Persistence model, which proved to be more accurate, as reflected in the MAE and RMSE metrics. For example, for the Incheon Fishery Water Purification Plant dataset during rain events, our model recorded an MAE of 162.19 kW, while the Persistence model showed a lower MAE of 132.16 kW. This discrepancy highlights the current limitation of our dataset, which spans two years, with one year allocated to the training set. The limited representation of rainy conditions in the dataset likely led to the observed performance gap, although our model still outperformed the Ranger and DNN models even under these conditions.
In Figure 6, we revisit the timeline first introduced in Figure 1, contrasting projected and actual solar PV power generation data for July 2019. While the model effectively reflects general trends in energy production, it occasionally shows shortcomings in identifying sudden declines in power generation.
The reasons for these discrepancies vary. Despite its robust design, the model’s internal workings may not fully capture the nuanced complexities of PV operations. In addition, the potential scarcity of solar irradiance metrics could undermine the accuracy of the forecast. Furthermore, location differences between our data points and Incheon’s meteorological readings may have resulted in less accurate representations of local weather patterns, which are influenced by subtle microclimatic variations.
To address these findings, subsequent research efforts will integrate finer-grained weather data and refine the model’s mechanisms to more accurately reflect the dynamic nature of PV production. The expected improvements from incorporating advanced DL strategies aim to significantly refine the predictive accuracy of the proposed method. The expected improvements go beyond creating a reliable tool for forecasting PV generation; they will improve grid management and energy distribution strategies amidst the variability inherent in renewable energy sources.

3.3. Model Interpretability

Ensuring the accuracy and trustworthiness of predictive models is paramount. As such, methods to clarify how models make decisions have gained prominence. The goal of this research was to identify the most important influencing factors in the Ranger-based online learning model. The Ranger package calculates variable importance by assessing how much the model’s accuracy decreases when the data for each variable is shuffled. This approach quantifies the importance of each variable to the model’s predictive performance, with a higher score indicating that a variable is more important in making accurate predictions. To present these results, we used the ‘vip’ package, a notable tool in R’s extensive visualization ecosystem, to illustrate the importance of variables [74].
Figure 7 and Figure 8 show the importance of the input variables in models with fewer hidden layers, providing an essential foundation for the detailed analysis in Figure 9. In these preliminary figures, the DNN_HL02 model, which is characterized by a lower number of hidden layers, emerges as consistently significant. This level of importance persists across different model complexities, as shown in the heatmaps. Figure 9 further illustrates that while models with a moderate number of hidden layers, specifically DNN_HL05 and DNN_HL07, show varying degrees of influence on variable importance, this does not eclipse the persistent significance of the simpler DNN_HL02 model. Such observations suggest that models of lesser complexity retain their importance alongside more complex models.
Furthermore, DNN_HL02 not only retains its relevance across a spectrum of models but also sees an increase in its impact, suggesting an enhanced role when analyzing larger datasets. Similarly, Figure 9 shows the increasing importance of DNN_HL05 and DNN_HL07, highlighting their improved ability to decipher the complexity of the data. This reflects the complexity of the modeling process and the need to continuously explore different model frameworks to gain deeper insights. Variations observed may signal the dynamic characteristics of the subject matter at each site examined. Consistent color intensity across models indicates general agreement on key variables, while variances reveal detailed assessments of variable importance as model complexity is adjusted over time.
Our study provides an analytical view of variable importance through the lens of different DNN models, focusing on their internal evaluation rather than direct predictive performance. This calibrated view of the heatmaps provides us with a broader understanding of how different DNN configurations weigh variable importance, which aids in the strategic selection and tuning of models for the New-Incheon Sonae, New-Incheon Observatory, and Incheon Fishery Water Purification Plant datasets.

4. Discussion

In this paper, we proposed a two-stage DNN optimization method for solar PV power generation prediction. Through comparative analysis with well-known benchmark models, including Att-LSTM, Att-Bi-LSTM, Att-GRU, Att-1D-CNN, Att-TCN, Att-GRU-TCN, and Att-Bi-GRU-TCN, the proposed method highlighted the superior performance of the proposed model in terms of MAE and RMSE on three datasets: New-Incheon Sonae, New-Incheon Observatory, and Incheon Fishery Water Purification Plant. Although these results are promising, our investigation revealed limitations that guide the direction of future research to improve prediction performance.
  • To address the observed forecast discrepancies under different weather conditions, we will consider incorporating additional independent variables to more accurately capture the effects of precipitation, or adopting a hybrid forecasting approach. For example, during expected rain events, we plan to leverage the predictive power of the persistence model to improve the accuracy of the proposed method.
  • Solar PV power generation is strongly correlated with solar irradiation. However, none of the environmental factors used as input variables in this study were used. Therefore, in the direction of future research, it is expected that more innovative prediction performance can be developed by performing solar irradiation prediction through existing input variables and using them together as input variables.
  • The hybrid models used as benchmarks are currently widely used techniques with good performance in time series forecasting. However, the training data used in this study has a period of one year, which is a very small amount of data in the field of time series forecasting, and we believe that the following performance is due to this. Therefore, if the amount of data is increased by data augmentation techniques, such as generative adversarial networks (GANs) and diffusions, the prediction performance of the models used as benchmarks will be improved, and the method proposed in this study will also be improved.
  • We aim to explore the frontiers of ML techniques, including advanced DL and ensemble learning methods. These techniques have demonstrated the potential to identify complex patterns within meteorologically influenced datasets. We are optimistic that such approaches will refine the sensitivity of the proposed method to minute variations in weather patterns and, by extension, solar PV power generation.
  • We will focus on the incorporation of hyper-local weather data, possibly by establishing dedicated weather monitoring stations at PV installation sites. An extension of the model’s parameter set is also planned to better capture the dynamic behavior of PV systems and the nonlinear interaction of solar irradiance [75].
  • Solar energy, as an important renewable energy source, is being utilized globally, but this research has been limited to the Incheon region of South Korea. Expanding our data sources to a wider range of regions will facilitate the verification of the universality of our proposed model and its ability to ensure grid stability with clean energy.
These strategic initiatives are expected to result not only in a more reliable tool for forecasting solar PV power generation but also to contribute significantly to optimizing grid management and strategic energy distribution in an era of increasingly variable renewable energy sources.

5. Conclusions

Achieving robust solar PV power generation prediction on low-performance computing systems is critical due to the need to balance cost and efficient use of computational resources. In this context, constructing the optimal DNN model for such prediction becomes a formidable challenge, requiring not only significant computational resources but also a depth of knowledge that may be beyond the reach of many energy experts. To address these issues, we proposed a stacking ensemble-based solar PV power generation prediction model that uses actual data from solar PV systems in Incheon combined with weather forecasts from the KMA. In the first stage, we constructed DNN models with 2 to 10 hidden layers using scikit-learn’s MLPRegressor optimized by Optuna, which simplified the development process and ensured the accessibility of the model even on low-performance systems.
In the next phase, we focused on preparing the training dataset for our meta-regression model, a Ranger-based RF. Specifically, we used the predictions from each DNN model as input features for the meta-regression model. A key innovation was the use of fivefold cross-validation to ensure sufficient training data for our meta-regression model without the need for a separate validation set, demonstrating the effectiveness of the model despite data limitations. In addition, the Ranger-based RF online learning model effectively learns from current trends and patterns in real time, taking into account the divergence between training and evaluation periods, thereby improving prediction accuracy over time.
Experimental results demonstrated the model’s superior performance in minimizing uncertainty and improving MAE and RMSE for multistep-ahead predictions. The use of RF’s variable importance output clarified the contributions of different DNN models, increasing the transparency and applicability of our prediction method. Despite its strengths, the model’s exclusion of solar irradiance as an input was identified as a limitation, suggesting an area for future improvement. Our study can provide robust solar PV power generation predictions on low-power systems, thus simplifying the construction of optimal DNN models for energy experts while balancing cost and computational efficiency.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/electronics13091659/s1, Table S1: New-Incheon Sonae_MinMax_Norm, Table S2: New-Incheon Observatory_MinMax_Norm, Table S3: Incheon Fishery Water Purification Plant_MinMax_Norm.

Author Contributions

Conceptualization, J.O. and D.S.; methodology, J.O.; software, J.O. and D.S.; validation, J.J., N.K. and J.M.; formal analysis, J.M.; investigation, D.S.; resources, J.M.; data curation, J.J. and N.K.; writing—original draft preparation, J.O. and D.S.; writing—review and editing, J.M. and E.H.; visualization, J.O. and J.M.; supervision, E.H.; project administration, E.H.; funding acquisition, J.M. and E.H. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by BK21 FOUR (Fostering Outstanding Universities for Research) (No. 5199990514663) and the Soonchunhyang University Research Fund.

Data Availability Statement

The datasets presented in this study are publicly available in the following public repositories: The original contributions created for this study are available at the Korea Open Data Portal, accessible at https://www.data.go.kr/index.do (accessed on 1 October 2023), and the Korea Meteorological Administration Data Portal, accessible at https://data.kma.go.kr/cmmn/main.do (accessed on 1 October 2023). These resources were used in the preparation of the data supporting the findings of this study. For those interested in directly using the preprocessed datasets for model training or further analysis, the datasets have been carefully prepared and can be found in the Supplementary Material of this article, specifically in Tables S1–S3. This preparation ensures that the data are in an appropriate format for immediate use in research efforts.

Acknowledgments

We would like to express our deepest gratitude to Hwimyeong Ha of LG Energy Solution, Ltd. (Seoul, Korea). His expert insights into the nuanced considerations and specific conditions within South Korea have been invaluable in paving the way for the potential integration of our technology into photovoltaic systems. We would also like to express our sincere gratitude to the reviewers and editors for their insightful and valuable feedback, which has helped us improve our work.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

This section provides a comprehensive list of acronyms used throughout this document, along with their corresponding definitions. The left column lists the acronyms for quick reference, while the right column provides a detailed explanation of each term to ensure clarity and enhance the reader’s understanding of the technical terminology relevant to this study:
ACalternating current
ANNartificial neural network
ARIMAautoregressive integrated moving average
Attattention mechanism
Att-1D-CNNattention-based one-dimensional convolutional neural network
Att-Bi-GRU-TCNattention-based bidirectional gated recurrent unit with temporal convolutional network
Att-Bi-LSTMattention-based bidirectional long short-term memory
Att-GRUattention-based gated recurrent unit
Att-GRU-TCNattention-based gated recurrent unit with temporal convolutional network
Att-LSTMattention-based long short-term memory
Att-TCNattention-based temporal convolutional network
Bi-LSTMbidirectional long short-term memory
CNNconvolutional neural network
CPUcentral processing unit
DCdirect current
DLdeep learning
DNNdeep neural network
DTdecision tree
EVelectric vehicle
GANgenerative adversarial network
GHGgreenhouse gas
GPUgraphics processing unit
GRUgated recurrent unit
KMAKorea Meteorological Administration
LSTMlong short-term memory
MAEmean absolute error
MLmachine learning
MLPmultilayer perceptron
NNneural network
PVphotovoltaic
RFrandom forest
RMSEroot mean square error
RNNrecurrent neural network
SVMsupport vector machine
TCNtemporal convolutional network
TSCVtime series cross-validation
Wpwatt-peak

References

  1. Olabi, A.G.; Abdelkareem, M.A. Renewable Energy and Climate Change. Renew. Sustain. Energy Rev. 2022, 158, 112111. [Google Scholar] [CrossRef]
  2. Wang, J.; Azam, W. Natural resource scarcity, fossil fuel energy consumption, and total greenhouse gas emissions in top emitting countries. Geosci. Front. 2024, 15, 101757. [Google Scholar] [CrossRef]
  3. Yoro, K.O.; Daramola, M.O. CO2 Emission Sources, Greenhouse Gases, and the Global Warming Effect. In Advances in Carbon Capture; Woodhead Publishing: Sawston, UK, 2020; pp. 3–28. [Google Scholar]
  4. Khan, Z.A.; Hussain, T.; Haq, I.U.; Ullah, F.U.M.; Baik, S.W. Towards efficient and effective renewable energy prediction via deep learning. Energy Rep. 2022, 8, 10230–10243. [Google Scholar] [CrossRef]
  5. Wen, J.; Okolo, C.V.; Ugwuoke, I.C.; Kolani, K. Research on influencing factors of renewable energy, energy efficiency, on technological innovation. Does trade, investment and human capital development matter? Energy Policy 2021, 160, 112718. [Google Scholar] [CrossRef]
  6. Aghaei, M.; Kumar, N.M.; Eskandari, A.; Ahmed, H.; de Oliveira, A.K.V.; Chopra, S.S. Solar PV Systems Design and Monitoring. In Photovoltaic Solar Energy Conversion; Academic Press: Cambridge, MA, USA, 2020; pp. 117–145. [Google Scholar]
  7. Ahmad, L.; Khordehgah, N.; Malinauskaite, J.; Jouhara, H. Recent advances and applications of solar photovoltaics and thermal technologies. Energy 2020, 207, 118254. [Google Scholar] [CrossRef]
  8. Park, S.; Kim, D.; Moon, J.; Hwang, E. Zero-Shot Photovoltaic Power Forecasting Scheme Based on a Deep Learning Model and Correlation Coefficient. Int. J. Energy Res. 2023, 2023, 1–14. [Google Scholar] [CrossRef]
  9. Ma, W.; Chen, Z.; Zhu, Q. Ultra-Short-Term Forecasting of Photo-Voltaic Power via RBF Neural Network. Electronics 2020, 9, 1717. [Google Scholar] [CrossRef]
  10. Ahmed, R.; Sreeram, V.; Mishra, Y.; Arif, M.D. A review and evaluation of the state-of-the-art in PV solar power forecasting: Techniques and optimization. Renew. Sustain. Energy Rev. 2020, 124, 109792. [Google Scholar] [CrossRef]
  11. Iheanetu, K.J. Solar Photovoltaic Power Forecasting: A Review. Sustainability 2022, 14, 17005. [Google Scholar] [CrossRef]
  12. Qazi, A.; Fayaz, H.; Wadi, A.; Raj, R.G.; Rahim, N.; Khan, W.A. The artificial neural network for solar radiation prediction and designing solar systems: A systematic literature review. J. Clean. Prod. 2015, 104, 1–12. [Google Scholar] [CrossRef]
  13. Wang, Y.; Cao, G.; Mao, S.; Nelms, R.M. Analysis of solar generation and weather data in smart grid with simultaneous inference of nonlinear time series. In Proceedings of the IEEE INFOCOM 2015—IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Hong Kong, China, 26 April 2015–1 May 2015; pp. 600–605. [Google Scholar]
  14. Van der Meer, D.; Mouli, G.R.C.; Mouli, G.M.E.; Elizondo, L.R.; Bauer, P. Energy Management System with PV Power Forecast to Optimally Charge EVs at the Workplace. IEEE Trans. Ind. Inform. 2016, 14, 311–320. [Google Scholar] [CrossRef]
  15. Voyant, C.; Notton, G.; Kalogirou, S.; Nivet, M.-L.; Paoli, C.; Motte, F.; Fouilloy, A. Machine learning methods for solar radiation forecasting: A review. Renew. Energy 2017, 105, 569–582. [Google Scholar] [CrossRef]
  16. Almonacid, F.; Fernandez, E.F.; Mellit, A.; Kalogirou, S. Review of techniques based on artificial neural networks for the electrical characterization of concentrator photovoltaic technology. Renew. Sustain. Energy Rev. 2017, 75, 938–953. [Google Scholar] [CrossRef]
  17. Wang, Y.; Shen, Y.; Mao, S.; Cao, G.; Nelms, R.M. Adaptive Learning Hybrid Model for Solar Intensity Forecasting. IEEE Trans. Ind. Inform. 2018, 14, 1635–1645. [Google Scholar] [CrossRef]
  18. Zhang, R.; Ma, H.; Hua, W.; Saha, T.K.; Zhou, X. Data-Driven Photovoltaic Generation Forecasting Based on a Bayesian Network With Spatial–Temporal Correlation Analysis. IEEE Trans. Ind. Inform. 2019, 16, 1635–1644. [Google Scholar] [CrossRef]
  19. Al-Dahidi, S.; Ayadi, O.; Alrbai, M.; Adeeb, J. Ensemble Approach of Optimized Artificial Neural Networks for Solar Photovoltaic Power Prediction. IEEE Access 2019, 7, 81741–81758. [Google Scholar] [CrossRef]
  20. Abdel-Nasser, M.; Mahmoud, K. Accurate photovoltaic power forecasting models using deep LSTM-RNN. Neural Comput. Appl. 2017, 31, 2727–2740. [Google Scholar] [CrossRef]
  21. Patel, H.K. Solar Radiation Prediction Using LSTM and CNN. Ph.D. Thesis, California State University, Sacramento, CA, USA, 2021. [Google Scholar]
  22. Wen, H.; Du, Y.; Chen, X.; Lim, E.; Wen, H.; Jiang, L.; Xiang, W. Deep Learning Based Multistep Solar Forecasting for PV Ramp-Rate Control Using Sky Images. IEEE Trans. Ind. Inform. 2020, 17, 1397–1406. [Google Scholar] [CrossRef]
  23. Jiao, X.; Li, X.; Lin, D.; Xiao, W. A Graph Neural Network Based Deep Learning Predictor for Spatio-Temporal Group Solar Irradiance Forecasting. IEEE Trans. Ind. Inform. 2021, 18, 6142–6149. [Google Scholar] [CrossRef]
  24. Zameer, A.; Jaffar, F.; Shahid, F.; Muneeb, M.; Khan, R.; Nasir, R. Short-term solar energy forecasting: Integrated computational intelligence of LSTMs and GRU. PLoS ONE 2023, 18, e0285410. [Google Scholar] [CrossRef]
  25. Rocha, H.R.O.; Fiorotti, R.; Fardin, J.F.; Garcia-Pereira, H.; Bouvier, Y.E.; Rodríguez-Lorente, A.; Yahyaoui, I. Application of AI for Short-Term PV Generation Forecast. Sensors 2023, 24, 85. [Google Scholar] [CrossRef]
  26. Mittal, S.; Vaishay, S. A survey of techniques for optimizing deep learning on GPUs. J. Syst. Arch. 2019, 99, 101635. [Google Scholar] [CrossRef]
  27. Yang, L.; Shami, A. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing 2020, 415, 295–316. [Google Scholar] [CrossRef]
  28. Moon, J.; Park, S.; Rho, S.; Hwang, E. A comparative analysis of artificial neural network architectures for building energy consumption forecasting. Int. J. Distrib. Sens. Netw. 2019, 15. [Google Scholar] [CrossRef]
  29. Moon, J.; Jung, S.; Rew, J.; Rho, S.; Hwang, E. Combination of short-term load forecasting models based on a stacking ensemble approach. Energy Build. 2020, 216, 109921. [Google Scholar] [CrossRef]
  30. Bâra, A.; Oprea, S. Embedding the weather prediction errors (WPE) into the photovoltaic (PV) forecasting method using deep learning. J. Forecast. 2024, 1–26. [Google Scholar] [CrossRef]
  31. Oprea, S.-V.; Bâra, A. On-grid and off-grid photovoltaic systems forecasting using a hybrid meta-learning method. Knowl. Inf. Syst. 2024, 66, 2575–2606. [Google Scholar] [CrossRef]
  32. Oprea, S.-V.; Bâra, A. A stacked ensemble forecast for photovoltaic power plants combining deterministic and stochastic methods. Appl. Soft Comput. 2023, 147, 110781. [Google Scholar] [CrossRef]
  33. Moon, J.; Park, S.; Rho, S.; Hwang, E. Robust Building Energy Consumption Forecasting Using an Online Learning Approach with R Ranger. J. Build. Eng. 2022, 47, 103851. [Google Scholar] [CrossRef]
  34. Smart City Korea. Incheon-Type Smart City, Inform the World. Report; Smart City Comprehensive Portal: Smart City Korea, 2022. Available online: https://smartcity.go.kr/en/ (accessed on 4 April 2023).
  35. Wikipedia. Incheon. Available online: https://en.wikipedia.org/wiki/Incheon (accessed on 4 April 2023).
  36. SMART CITIES INDEX REPORT. Available online: https://smartcitiesindex.org/smartcitiesindexreport2022 (accessed on 4 April 2023).
  37. Smart City Korea. Renewable Energy Business Brand Smart Energy Factory Loan Support. Report; Smart City Comprehensive Portal: SMART CITY KOREA, 2020. Available online: https://smartcity.go.kr/en/ (accessed on 4 April 2023).
  38. Korea Public Data Portal. Available online: https://www.data.go.kr/en/index.do (accessed on 4 April 2023).
  39. Park, S.; Jung, S.; Moon, J.; Hwang, E. Explainable Photovoltaic Power Forecasting Scheme Using BiLSTM. KIPS Trans. Softw. Data Eng. 2022, 11, 339–346. [Google Scholar]
  40. Yao, X.; Fu, X.; Zong, C. Short-term load forecasting method based on feature preference strategy and LightGBM-XGboost. IEEE Access 2022, 10, 75257–75268. [Google Scholar] [CrossRef]
  41. Korea Meteorological Administration. KMA Weather Data Service’s MET Data Portal. Available online: https://data.kma.go.kr/resources/html/en/aowdp.html (accessed on 4 April 2023).
  42. Korea Meteorological Administration. Weather Forecast. Available online: https://www.kma.go.kr/eng/biz/forecast_01.jsp (accessed on 4 April 2023).
  43. Visser, L.; AlSkaif, T.; van Sark, W. Operational day-ahead solar power forecasting for aggregated PV systems with a varying spatial distribution. Renew. Energy 2021, 183, 267–282. [Google Scholar] [CrossRef]
  44. Thanh, P.N.; Cho, M.Y.; Chang, C.L.; Chen, M.J. Short-term three-phase load prediction with advanced metering infrastructure data in smart solar microgrid based convolution neural network bidirectional gated recurrent unit. IEEE Access 2022, 10, 68686–68699. [Google Scholar] [CrossRef]
  45. Korea Meteorological Administration. Climate Information. Available online: https://www.kma.go.kr/eng/biz/climate_01.jsp (accessed on 4 April 2023).
  46. Bâra, A.; Oprea, S.-V. Machine Learning Algorithms for Power System Sign Classification and a Multivariate Stacked LSTM Model for Predicting the Electricity Imbalance Volume. Int. J. Comput. Intell. Syst. 2024, 17, 1–22. [Google Scholar] [CrossRef]
  47. Kazmi, M.; Khan, H.R.; Lubaba; Bin Khalid, M.H.; Qazi, S.A. Threefold Optimized Forecasting of Electricity Consumption in Higher Education Institutions. Comput. Mater. Contin. 2022; 73, 2351–2370. [Google Scholar] [CrossRef]
  48. Jang, J.; Jeong, W.; Kim, S.; Lee, B.; Lee, M.; Moon, J. RAID: Robust and Interpretable Daily Peak Load Forecasting via Multiple Deep Neural Networks and Shapley Values. Sustainability 2023, 15, 6951. [Google Scholar] [CrossRef]
  49. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Duchesnay, É. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  50. Roy, S.; Mehera, R.; Pal, R.K.; Bandyopadhyay, S.K. Hyperparameter optimization for deep neural network models: A comprehensive study on methods and techniques. Innov. Syst. Softw. Eng. 2023, 1–12. [Google Scholar] [CrossRef]
  51. Öztürk, M.M. Initializing Hyper-Parameter Tuning with a Metaheuristic-Ensemble Method: A Case Study Using Time-Series Weather Data. Evol. Intell. 2023, 16, 1019–1031. [Google Scholar] [CrossRef]
  52. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-Generation Hyperparameter Optimization Framework. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 25 July 2019; pp. 2623–2631. [Google Scholar]
  53. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  54. Bâra, A.; Oprea, S.-V.; Tudorică, B.G. From the East-European Regional Day-Ahead Markets to a Global Electricity Market. Comput. Econ. 2023, 1–33. [Google Scholar] [CrossRef]
  55. Tyralis, H.; Papacharalampous, G. Variable Selection in Time Series Forecasting Using Random Forests. Algorithms 2017, 10, 114. [Google Scholar] [CrossRef]
  56. Qi, Y. Random Forest for Bioinformatics. In Ensemble Machine Learning: Methods and Applications; Springer: New York, NY, USA, 2012; pp. 307–323. [Google Scholar]
  57. Dudek, G. A Comprehensive Study of Random Forest for Short-Term Load Forecasting. Energies 2022, 15, 7547. [Google Scholar] [CrossRef]
  58. Chou, J.-S.; Liu, C.-Y.; Prayogo, H.; Khasani, R.R.; Gho, D.; Lalitan, G.G. Predicting nominal shear capacity of reinforced concrete wall in building by metaheuristics-optimized machine learning. J. Build. Eng. 2022, 61, 105046. [Google Scholar] [CrossRef]
  59. Antoniadis, A.; Lambert-Lacroix, S.; Poggi, J.-M. Random forests for global sensitivity analysis: A selective review. Reliab. Eng. Syst. Saf. 2020, 206, 107312. [Google Scholar] [CrossRef]
  60. Cao, Y.; Li, H.; Yang, Y. Combining random forest and multicollinearity modeling for index tracking. Commun. Stat.-Simul. Comput. 2022; 1–12. [Google Scholar] [CrossRef]
  61. Leem, S.; Oh, J.; Moon, J.; Kim, M.; Rho, S. Enhancing multistep-ahead bike-sharing demand prediction with a two-stage online learning-based time-series model: Insight from Seoul. J. Supercomput. 2023, 80, 4049–4082. [Google Scholar] [CrossRef]
  62. Bellahsen, A.; Dagdougui, H. Aggregated short-term load forecasting for heterogeneous buildings using machine learning with peak estimation. Energy Build. 2021, 237, 110742. [Google Scholar] [CrossRef]
  63. Wright, M.N.; Ziegler, A. Ranger: A fast implementation of random forests for high dimensional data in C++ and R. J. Stat. Softw. 2017, 77, 1–17. [Google Scholar] [CrossRef]
  64. Alkhammash, E.H.; Hadjouni, M.; Elshewey, A.M. A Hybrid Ensemble Stacking Model for Gender Voice Recognition Approach. Electronics 2022, 11, 1750. [Google Scholar] [CrossRef]
  65. So, D.; Oh, J.; Leem, S.; Ha, H.; Moon, J. A Hybrid Ensemble Model for Solar Irradiance Forecasting: Advancing Digital Models for Smart Island Realization. Electronics 2023, 12, 2607. [Google Scholar] [CrossRef]
  66. Oshiro, T.M.; Perez, P.S.; Baranauskas, J.A. How Many Trees in a Random Forest? In Machine Learning and Data Mining in Pattern Recognition: 8th International Conference, MLDM 2012, Berlin, Germany, 13–20 July 2012; Proceedings 8; Springer: Berlin/Heidelberg, Germany, 2012; pp. 154–168. [Google Scholar]
  67. Zhou, H.; Zhang, Y.; Yang, L.; Liu, Q.; Yan, K.; Du, Y. Short-Term Photovoltaic Power Forecasting Based on Long Short Term Memory Neural Network and Attention Mechanism. IEEE Access 2019, 7, 78063–78074. [Google Scholar] [CrossRef]
  68. He, B.; Ma, R.; Zhang, W.; Zhu, J.; Zhang, X. An Improved Generating Energy Prediction Method Based on Bi-LSTM and Attention Mechanism. Electronics 2022, 11, 1885. [Google Scholar] [CrossRef]
  69. Jung, S.; Moon, J.; Park, S.; Hwang, E. An Attention-Based Multilayer GRU Model for Multi-Step-Ahead Short-Term Load Forecasting. Sensors 2021, 21, 1639. [Google Scholar] [CrossRef]
  70. Wu, Z.; Pan, F.; Li, D.; He, H.; Zhang, T.; Yang, S. Prediction of Photovoltaic Power by the Informer Model Based on Convolutional Neural Network. Sustainability 2022, 14, 13022. [Google Scholar] [CrossRef]
  71. Ren, X.; Zhang, F.; Sun, Y.; Liu, Y. A Novel Dual-Channel Temporal Convolutional Network for Photovoltaic Power Forecasting. Energies 2024, 17, 698. [Google Scholar] [CrossRef]
  72. Xiao, H.; Pei, W.; Wu, L.; Ma, L.; Ma, T.; Hua, W. A novel deep learning based probabilistic power flow method for Multi-Microgrids distribution system with incomplete network information. Appl. Energy 2023, 335, 120716. [Google Scholar] [CrossRef]
  73. Zhou, H.; Wang, J.; Ouyang, F.; Cui, C.; Li, X. A Two-Stage Method for Ultra-Short-Term PV Power Forecasting Based on Data-Driven. IEEE Access 2023, 11, 41175–41189. [Google Scholar] [CrossRef]
  74. Greenwell, B.M.; Boehmke, B.C.; Gray, B. Variable Importance Plots-An Introduction to the vip Package. R J. 2020, 12, 343. [Google Scholar] [CrossRef]
  75. Tudorică, B.-G.; Bucur, C.; Panait, M.; Oprea, S.-V.; Bâra, A. Energetic Equilibrium: Optimizing renewable and non-renewable energy sources via particle swarm optimization. Util. Policy 2024, 87, 101722. [Google Scholar] [CrossRef]
Figure 1. Min–max normalized meteorological and solar photovoltaic (PV) power data in Incheon for July 2019. The hourly time index runs from 1 a.m. on the 1st to midnight on the 31st.
Figure 1. Min–max normalized meteorological and solar photovoltaic (PV) power data in Incheon for July 2019. The hourly time index runs from 1 a.m. on the 1st to midnight on the 31st.
Electronics 13 01659 g001
Figure 2. Procedure for multistep-ahead solar PV power generation forecasting.
Figure 2. Procedure for multistep-ahead solar PV power generation forecasting.
Electronics 13 01659 g002
Figure 3. Multilayer perceptron architecture.
Figure 3. Multilayer perceptron architecture.
Electronics 13 01659 g003
Figure 4. Optimization of a deep neural network (DNN) model with fivefold cross-validation using Optuna. NMAE represents negative mean absolute error.
Figure 4. Optimization of a deep neural network (DNN) model with fivefold cross-validation using Optuna. NMAE represents negative mean absolute error.
Electronics 13 01659 g004
Figure 5. Time series cross-validation for multistep-ahead solar PV power generation forecasting.
Figure 5. Time series cross-validation for multistep-ahead solar PV power generation forecasting.
Electronics 13 01659 g005
Figure 6. Comparative analysis of actual and predicted solar PV power generation for July 2019 in Incheon. (a) New-Incheon Sonae; (b) New-Incheon Observatory; (c) Incheon Fishery Water Purification Plant.
Figure 6. Comparative analysis of actual and predicted solar PV power generation for July 2019 in Incheon. (a) New-Incheon Sonae; (b) New-Incheon Observatory; (c) Incheon Fishery Water Purification Plant.
Electronics 13 01659 g006aElectronics 13 01659 g006b
Figure 7. Variable importance heatmap of the Ranger model for New-Incheon Sonae.
Figure 7. Variable importance heatmap of the Ranger model for New-Incheon Sonae.
Electronics 13 01659 g007
Figure 8. Variable importance heatmap of the Ranger model for New-Incheon Observatory.
Figure 8. Variable importance heatmap of the Ranger model for New-Incheon Observatory.
Electronics 13 01659 g008
Figure 9. Variable importance heatmap of the Ranger model for Incheon Fishery Water Purification Plant.
Figure 9. Variable importance heatmap of the Ranger model for Incheon Fishery Water Purification Plant.
Electronics 13 01659 g009
Table 1. Key components and functions of the solar photovoltaic (PV) system.
Table 1. Key components and functions of the solar photovoltaic (PV) system.
ComponentFunction
Solar cell panelThe core component that converts sunlight into electrical energy
InverterConverts direct current (DC) into alternating current (AC) that can be used
BatteryStores electricity
Charge controllerControls the flow of electricity to prevent battery overcharging
WiringConnects the components of the system
Table 2. Statistical insights and parameters of Incheon’s solar PV installations (unit: kW).
Table 2. Statistical insights and parameters of Incheon’s solar PV installations (unit: kW).
StatisticsNew-Incheon
Sonae
New-Incheon
Observatory
Incheon Fishery Water
Purification Plant
Training SetTest SetTraining SetTest SetTraining SetTest Set
Mean317.26286.56309.96295.22330.91323.40
Standard error3.723.893.703.494.013.82
Median286.25232.50285.88264.52296.75288.00
Mode000000
Standard deviation235.67246.43234.24221.27254.25242.36
Variance55,538.3460,727.8154,867.2448,958.3464,645.4258,737.13
Kurtosis−1.10−1.01−1.21−1.11−1.10−1.08
Skewness0.310.520.280.350.340.36
Range927.93889.00849.37797.01970.87904.79
Minimum000000
Maximum927.93889.00849.37797.01970.87904.79
Sum1,273,793.951,150,536.241,244,471.641,185,312.061,328,589.981,298,457.14
Number of
observations
401540154015401540154015
LocationNamdong-gu, IncheonSeo-gu, IncheonSeo-gu, Incheon
Capacity (MW)0.9980.2001.742
Year of
completion
Nov. ‘11Sep. ‘12Dec. ‘17
RemarksRPARPSRPS
MW, megawatt; RPA, Renewable Portfolio Standard Act (assumed context based on the content); RPS, Renewable Portfolio Standard.
Table 3. Input variables for solar PV power generation forecasting modeling.
Table 3. Input variables for solar PV power generation forecasting modeling.
No.Input Variable Variable TypeNo.Input Variable Variable Type
1MonthXNumeric123:00 p.m.Binary
2MonthYNumeric134:00 p.m.Binary
3DayXNumeric145:00 p.m.Binary
4DayYNumeric156:00 p.m.Binary
58:00 a.m.Binary16Avg. TemperatureNumeric
69:00 a.m.Binary17Min. TemperatureNumeric
710:00 a.m.Binary18Max. TemperatureNumeric
811:00 a.m.Binary19TemperatureNumeric
912:00 p.m.Binary20HumidityNumeric
101:00 p.m.Binary21Wind SpeedNumeric
112:00 p.m.Binary22PrecipitationNumeric
Table 4. List of hyperparameters for optimal deep neural network (DNN) model construction.
Table 4. List of hyperparameters for optimal deep neural network (DNN) model construction.
HyperparameterDefinitionRange
hidden_layer_sizesNumber of neurons in each hidden layer, for 2 to 10 layers2–10 hidden layers, each with 1–40 (Step: 1) [48]
activationActivation function for the hidden layerrelu [29]
solverThe solver for weight optimizationadam [29]
alphaL2 penalty (regularization term) parameter0.0001–0.001 (Step: 0.0001) [48]
batch_sizeSize of minibatches for stochastic optimizer5–100 (Step: 1) [48]
learning_rateLearning rate schedule for weight updatesconstant, adaptive [48]
learning_rate_initInitial learning rate0.0001–0.1 (Step: 0.0001) [48]
max_iterMaximum number of iterations100–2000 (Step: 10) [48]
Table 5. Mean absolute error (MAE) comparison across multistep-ahead prediction models for the New-Incheon Sonae dataset (unit: kW).
Table 5. Mean absolute error (MAE) comparison across multistep-ahead prediction models for the New-Incheon Sonae dataset (unit: kW).
StepRangerLSTM [24]Bi-LSTM [24]GRU [24]Att-LSTM [67]Att-Bi-LSTM [68]Att-GRU [69]Att-1D-CNN [70]Att-TCN [71]Att-GRU-TCN [72]Att-Bi-GRU-TCN [73]Ours
1152.15170.97173.56171.32172.29165.32163.85169.03165.04170.55166.99142.87
2153.82165.58173.97161.89172.05165.79165.37172.68165.80172.81168.75146.58
3155.07166.01172.23163.85172.46165.51168.67169.40172.06172.30170.14148.71
4155.32166.27173.12166.56171.47165.63168.48166.29174.60172.12170.61149.52
5155.62167.39172.25165.86172.97165.00167.51167.12176.10171.37172.09150.47
6156.08171.34171.81164.26173.20167.55170.65168.62176.29171.83171.10150.64
7156.52177.69169.89164.92173.21166.89167.48170.93176.95173.85172.34150.94
8156.13173.54168.21165.60171.94169.11165.94171.11177.94172.23171.12151.07
9156.29170.43168.32165.00170.34168.09165.17171.35176.90172.98169.81151.23
10156.39173.03168.66162.81165.75165.13165.03169.80177.73171.85167.15151.11
11156.55180.00163.19160.60163.45162.26169.09169.98176.16168.63164.99151.66
Avg.155.45171.11170.47164.79170.83166.02167.02169.66174.14171.87169.55149.53
Table 6. Root mean squared error (RMSE) comparison across multistep-ahead prediction models for the New-Incheon Sonae dataset (unit: kW).
Table 6. Root mean squared error (RMSE) comparison across multistep-ahead prediction models for the New-Incheon Sonae dataset (unit: kW).
StepRangerLSTM [24]Bi-LSTM [24]GRU [24]Att-LSTM [67]Att-Bi-LSTM [68]Att-GRU [69]Att-1D-CNN [70]Att-TCN [71]Att-GRU-TCN [72]Att-Bi-GRU-TCN [73]Ours
1204.24226.73236.26225.94235.27226.33222.45229.64229.44236.28229.76192.52
2206.53222.35237.06219.66236.22227.36226.45234.44230.51239.73232.96197.94
3208.05226.82234.84224.09236.41225.69229.53230.40237.60238.70235.03201.16
4208.42227.68235.80226.63235.78226.84229.76224.21240.84237.19235.25202.04
5208.87226.87235.60226.79236.94226.69227.79225.44242.42236.46235.55203.30
6209.61227.04235.00226.46237.15230.58228.38229.34242.05236.57234.64203.63
7209.96229.70232.59227.50237.34229.70224.41233.84242.56238.90236.76203.81
8209.54228.46231.02228.68235.15232.15223.28234.60244.05237.32235.10204.25
9209.86226.48231.05227.13233.73228.75222.50233.76243.54239.02234.28204.29
10210.00232.52230.14222.13226.25225.56219.84230.65244.64236.89231.69204.36
11210.20247.36222.70215.02219.97223.93220.78231.85242.25232.58228.75204.72
Avg.208.66229.28232.91224.55233.66227.60225.01230.74239.99237.24233.62202.00
Table 7. MAE comparison across multistep-ahead prediction models for the New-Incheon Observatory dataset (unit: kW).
Table 7. MAE comparison across multistep-ahead prediction models for the New-Incheon Observatory dataset (unit: kW).
StepRangerLSTM [24]Bi-LSTM [24]GRU [24]Att-LSTM [67]Att-Bi-LSTM [68]Att-GRU [69]Att-1D-CNN [70]Att-TCN [71]Att-GRU-TCN [72]Att-Bi-GRU-TCN [73]Ours
1126.63146.56141.96191.38144.77145.45145.98154.40138.39143.33144.29118.73
2128.40139.88143.10174.92145.17147.26150.80153.41136.64143.37144.92122.98
3129.40142.76143.14163.30144.14145.85149.78155.18137.78143.48144.53124.59
4130.06143.70143.28158.55142.43145.92153.24156.50137.56143.40146.69125.74
5130.41142.99143.27157.53141.06147.70150.45156.86139.61142.66146.39126.16
6130.78144.08143.16155.06141.35146.56145.64164.68139.75142.57146.01126.71
7130.78144.76143.37148.62141.14146.58145.38164.83140.24142.23143.79126.57
8130.98143.90142.63148.32142.35146.77144.11159.99140.56144.31142.67126.76
9130.77140.44141.23147.79142.55144.57143.02153.87142.03144.39141.90126.70
10130.72136.36140.78144.18140.10141.78142.18148.90142.76144.07139.21126.94
11130.83133.12138.21143.37135.78139.20139.44160.12142.43143.87137.90127.20
Avg.129.98141.69142.19157.55141.89145.24146.37157.16139.80143.42143.48125.37
Table 8. RMSE comparison across multistep-ahead prediction models for the New-Incheon Observatory dataset (unit: kW).
Table 8. RMSE comparison across multistep-ahead prediction models for the New-Incheon Observatory dataset (unit: kW).
StepRangerLSTM [24]Bi-LSTM [24]GRU [24]Att-LSTM [67]Att-Bi-LSTM [68]Att-GRU [69]Att-1D-CNN [70]Att-TCN [71]Att-GRU-TCN [72]Att-Bi-GRU-TCN [73]Ours
1176.54203.94205.31245.31211.65206.83212.93205.23200.80206.97207.98156.64
2179.27200.98206.21230.58212.86212.44217.08204.29200.31207.53210.68163.03
3180.54204.45207.56218.74212.09210.36217.09205.52202.18209.09209.92165.41
4181.43203.38208.95213.52210.11211.39221.39209.32203.46208.93212.78167.02
5182.25202.38208.92212.33207.99212.53217.06209.54205.70207.77212.90167.85
6182.65203.52207.80210.77208.18211.96210.09210.93205.67207.13211.67168.52
7182.70204.84209.12206.93206.92211.91207.51211.28205.53206.75208.79168.56
8183.04203.45208.31205.01206.79211.18206.23206.25206.05207.54206.55168.69
9182.73198.70208.02204.08207.10207.22203.16201.53207.19208.87204.72168.81
10182.57191.73207.01203.73202.21202.93203.39200.14208.61208.39199.84168.74
11183.00189.31202.51203.75195.36198.34196.87211.18207.70208.15196.32169.24
Avg.181.52200.61207.25214.07207.39208.83210.25206.84204.84207.92207.47166.59
Table 9. MAE comparison across multistep-ahead prediction models for the Incheon Fishery Water Purification Plant dataset (unit: kW).
Table 9. MAE comparison across multistep-ahead prediction models for the Incheon Fishery Water Purification Plant dataset (unit: kW).
StepRangerLSTM [24]Bi-LSTM [24]GRU [24]Att-LSTM [67]Att-Bi-LSTM [68]Att-GRU [69]Att-1D-CNN [70]Att-TCN [71]Att-GRU-TCN [72]Att-Bi-GRU-TCN [73]Ours
1140.26147.83143.76179.15156.70165.53152.46151.25152.65159.05155.23134.30
2142.06148.27145.05163.19158.16163.33154.90150.67153.34161.40156.06139.08
3142.98147.53147.99159.11158.94167.27155.12154.25156.64160.22156.29140.63
4143.27146.01148.47159.82158.15166.84156.73155.43158.63161.80157.30141.80
5143.68147.00147.38154.41157.28167.10153.75156.82164.01162.62156.75142.77
6144.22149.24146.16153.63157.16167.89153.60155.36164.52164.39157.73143.18
7144.25151.27145.19166.88157.57167.25152.89155.40165.77168.45158.31143.65
8144.44154.03143.37172.64158.12163.71152.79155.79166.31168.32158.46143.31
9144.50153.78143.74167.57157.95160.42150.29154.76167.54172.15159.16143.64
10144.45153.34144.01161.81152.69154.48146.53153.74167.42171.12157.51143.58
11144.47152.60145.95163.11148.41157.31149.34151.17168.10168.09156.13144.76
Avg.143.51150.08145.55163.76156.47163.74152.58154.06162.27165.24157.18141.88
Table 10. RMSE comparison across multistep-ahead prediction models for the Incheon Fishery Water Purification Plant dataset (unit: kW).
Table 10. RMSE comparison across multistep-ahead prediction models for the Incheon Fishery Water Purification Plant dataset (unit: kW).
StepRangerLSTM [24]Bi-LSTM [24]GRU [24]Att-LSTM [67]Att-Bi-LSTM [68]Att-GRU [69]Att-1D-CNN [70]Att-TCN [71]Att-GRU-TCN [72]Att-Bi-GRU-TCN [73]Ours
1191.17199.32198.02223.15222.31227.75213.35209.16218.36225.33221.23180.06
2193.58202.74204.28209.48224.15226.52218.43209.35218.42229.17221.65186.61
3194.71201.54205.54203.77225.12230.32220.53215.95222.55228.46222.53189.35
4195.33200.84205.01202.72223.25230.15223.18217.49225.26230.92223.74191.21
5196.04204.18203.33199.51222.76233.76219.63218.88232.78232.25221.73192.33
6196.69207.98202.03204.37223.28234.27219.32218.51232.86233.87223.38193.02
7196.64209.91202.12220.68223.32231.16218.15219.19234.09239.09223.95193.81
8196.95211.28200.49228.41222.49227.35217.97220.03234.38238.62224.20193.29
9197.08209.77201.02221.48222.12222.39212.89219.16236.33243.50224.28193.41
10196.97206.69200.95213.13215.67214.71207.69218.26236.41241.91221.63193.87
11196.84205.10201.74214.04206.18222.28201.86211.45237.45237.20219.51194.71
Avg.195.64205.40202.23212.80220.97227.33215.73216.13229.90234.57222.53191.06
Table 11. MAE and RMSE comparisons across day-ahead prediction models for the New-Incheon Sonae dataset (unit: kW). HL indicates the number of hidden layers.
Table 11. MAE and RMSE comparisons across day-ahead prediction models for the New-Incheon Sonae dataset (unit: kW). HL indicates the number of hidden layers.
ModelJan.–Mar.Apr.–Jun.Jul.–Sep.Oct.–Dec.Avg.
MAERMSEMAERMSEMAERMSEMAERMSEMAERMSE
Persistence155.98228.90145.03223.95152.23217.00137.54231.26147.66225.33
Ranger146.67203.00163.98226.73172.26227.08178.09244.83165.35226.02
DNN (HL: 2)153.22216.03172.61241.38181.96239.90179.98251.56172.05237.68
DNN (HL: 3)152.82214.98167.46234.08170.42224.52175.35245.54166.58230.13
DNN (HL: 4)149.26210.65169.05237.00179.26237.94176.94247.87168.73233.88
DNN (HL: 5)151.42212.71171.57239.27184.31242.80178.59250.63171.58236.90
DNN (HL: 6)156.61219.92174.89241.31175.93232.83177.55247.88171.32235.79
DNN (HL: 7)149.89212.75168.68236.05179.25235.15174.57246.72168.19233.09
DNN (HL: 8)160.53230.99193.44262.10185.83245.41184.02264.24181.03251.13
DNN (HL: 9)157.31223.52181.86249.99185.94244.11178.54252.16176.00242.79
DNN (HL: 10)165.88240.03197.50264.04193.75254.01184.72262.70185.54255.43
Ours144.79199.57147.52203.86151.31194.65162.75219.76151.66204.72
Table 12. MAE and RMSE comparisons across day-ahead prediction models for the New-Incheon Observatory dataset (unit: kW).
Table 12. MAE and RMSE comparisons across day-ahead prediction models for the New-Incheon Observatory dataset (unit: kW).
ModelJan.–Mar.Apr.–Jun.Jul.–Sep.Oct.–Dec.Avg.
MAERMSEMAERMSEMAERMSEMAERMSEMAERMSE
Persistence115.34172.45149.96228.21145.82205.65115.98175.94131.82196.93
Ranger122.89174.81162.69227.35161.80221.24107.43160.86138.72198.20
DNN (HL: 2)131.66190.30169.57237.91173.02234.15111.46169.69146.44210.04
DNN (HL: 3)135.99194.08170.45237.15168.46226.68110.12167.41146.25208.15
DNN (HL: 4)134.87194.99169.32237.75174.55237.38113.04173.29147.96212.70
DNN (HL: 5)140.14200.92178.78251.04177.20239.97117.81180.11153.48219.90
DNN (HL: 6)136.75195.61172.91243.59173.21232.90112.74171.48148.90212.86
DNN (HL: 7)133.38194.05173.52247.44173.57235.79112.42171.19148.23214.36
DNN (HL: 8)137.65200.24173.35243.62173.48233.39112.83170.10149.33213.79
DNN (HL: 9)193.89220.54210.35244.40180.27205.69191.60214.61193.98221.72
DNN (HL: 10)141.13204.00184.18259.97181.91245.97120.08183.07156.83225.41
Ours119.14158.90150.21196.17142.11179.9997.32135.77127.20169.24
Table 13. MAE and RMSE comparisons across day-ahead prediction models for the Incheon Fishery Water Purification Plant dataset (unit: kW).
Table 13. MAE and RMSE comparisons across day-ahead prediction models for the Incheon Fishery Water Purification Plant dataset (unit: kW).
ModelJan.–Mar.Apr.–Jun.Jul.–Sep.Oct.–Dec.Avg.
MAERMSEMAERMSEMAERMSEMAERMSEMAERMSE
Persistence161.06234.32154.87234.35154.51214.23131.65198.87150.45220.83
Ranger160.58218.36173.66240.66162.65218.01115.96171.33153.11213.49
DNN (HL: 2)177.65243.16196.83270.23176.47233.18127.11187.26169.39235.22
DNN (HL: 3)172.46238.68193.61264.08174.41231.51131.18192.19167.82232.92
DNN (HL: 4)172.92239.06200.84275.78175.70232.54123.82184.64168.20235.11
DNN (HL: 5)175.31242.71199.05276.23178.21237.30128.26192.01170.10238.82
DNN (HL: 6)174.68241.96200.25277.26177.83234.69126.17189.93169.62237.86
DNN (HL: 7)172.36239.23207.54282.71182.23241.33125.95191.57171.92240.77
DNN (HL: 8)169.13236.70189.34261.31176.14236.41131.86195.42166.54233.56
DNN (HL: 9)178.06247.03205.57282.08181.84241.26121.96186.51171.73241.50
DNN (HL: 10)180.50248.94197.63276.87175.72233.06125.17187.41169.62238.61
Ours159.48215.19160.07213.17151.87191.57108.27153.48144.76194.71
Table 14. Comparative analysis of day-ahead prediction models during dry conditions and precipitation events for the New-Incheon Sonae dataset (unit: kW).
Table 14. Comparative analysis of day-ahead prediction models during dry conditions and precipitation events for the New-Incheon Sonae dataset (unit: kW).
ModelDry ConditionsPrecipitation Events
MAERMSEMAERMSE
Persistence148.04226.16132.18194.43
Ranger162.22221.27222.66299.37
DNN (HL: 2)168.30232.31242.62320.51
DNN (HL: 3)162.97224.86233.85311.10
DNN (HL: 4)165.36229.01231.96310.00
DNN (HL: 5)168.15232.02235.51313.18
DNN (HL: 6)167.92231.03233.56309.99
DNN (HL: 7)164.75228.15231.55309.72
DNN (HL: 8)178.31247.60232.74309.61
DNN (HL: 9)173.04238.68231.76309.26
DNN (HL: 10)183.18252.40231.25307.26
Ours141.17189.96172.68232.81
Table 15. Comparative analysis of day-ahead prediction models during dry conditions and precipitation events for the New-Incheon Observatory dataset (unit: kW).
Table 15. Comparative analysis of day-ahead prediction models during dry conditions and precipitation events for the New-Incheon Observatory dataset (unit: kW).
ModelDry ConditionsPrecipitation Events
MAERMSEMAERMSE
Persistence132.09197.70121.07174.92
Ranger133.16190.05238.99309.68
DNN (HL: 2)140.60201.77253.37324.57
DNN (HL: 3)140.39199.95252.94321.57
DNN (HL: 4)142.53205.43246.84316.11
DNN (HL: 5)148.46213.37245.68315.69
DNN (HL: 6)143.37205.17249.96321.25
DNN (HL: 7)142.74207.06249.26318.61
DNN (HL: 8)144.00206.59247.43316.80
DNN (HL: 9)194.34222.04189.13217.54
DNN (HL: 10)151.89219.14248.70318.62
Ours117.37154.93142.40183.84
Table 16. Comparative analysis of day-ahead prediction models during dry conditions and precipitation events for the Incheon Fishery Water Purification Plant dataset (unit: kW).
Table 16. Comparative analysis of day-ahead prediction models during dry conditions and precipitation events for the Incheon Fishery Water Purification Plant dataset (unit: kW).
ModelDry ConditionsPrecipitation Events
MAERMSEMAERMSE
Persistence151.03222.30132.16180.59
Ranger148.93208.00230.29296.06
DNN (HL: 2)165.23230.28246.55312.32
DNN (HL: 3)164.12228.32238.08305.78
DNN (HL: 4)164.56230.84236.95303.54
DNN (HL: 5)166.36234.68239.89305.54
DNN (HL: 6)165.84233.65239.88305.45
DNN (HL: 7)167.91236.25245.97312.43
DNN (HL: 8)163.20229.54229.92298.67
DNN (HL: 9)167.80237.19244.91310.70
DNN (HL: 10)165.46233.84246.67313.66
Ours132.70177.92162.19213.98
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Oh, J.; So, D.; Jo, J.; Kang, N.; Hwang, E.; Moon, J. Two-Stage Neural Network Optimization for Robust Solar Photovoltaic Forecasting. Electronics 2024, 13, 1659. https://doi.org/10.3390/electronics13091659

AMA Style

Oh J, So D, Jo J, Kang N, Hwang E, Moon J. Two-Stage Neural Network Optimization for Robust Solar Photovoltaic Forecasting. Electronics. 2024; 13(9):1659. https://doi.org/10.3390/electronics13091659

Chicago/Turabian Style

Oh, Jinyeong, Dayeong So, Jaehyeok Jo, Namil Kang, Eenjun Hwang, and Jihoon Moon. 2024. "Two-Stage Neural Network Optimization for Robust Solar Photovoltaic Forecasting" Electronics 13, no. 9: 1659. https://doi.org/10.3390/electronics13091659

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop