Next Article in Journal
Implementation of the Bio-Inspired Metaheuristic Firefly Algorithm (FA) Applied to Maximum Power Point Tracking of Photovoltaic Systems
Previous Article in Journal
Application of Generation Adequacy Analysis for Reliability Evaluation of a Floating Production Storage and Offloading Platform Power System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessing the Impact of Features on Probabilistic Modeling of Photovoltaic Power Generation

1
Department of Electrical Engineering, Graduate School of Science and Technology, Tokyo University of Science, 2641 Yamazaki, Noda 278-8510, Chiba, Japan
2
Faculty of Engineering, Information and Systems, University of Tsukuba, Tsukuba 305-8573, Ibaraki, Japan
*
Author to whom correspondence should be addressed.
Energies 2022, 15(15), 5337; https://doi.org/10.3390/en15155337
Submission received: 18 May 2022 / Revised: 29 June 2022 / Accepted: 13 July 2022 / Published: 22 July 2022

Abstract

:
Photovoltaic power generation has high variability and uncertainty because it is affected by uncertain factors such as weather conditions. Therefore, probabilistic forecasting is useful for optimal operation and risk hedging in power systems with large amounts of photovoltaic power generation. However, deterministic forecasting is the mainstay of photovoltaic generation forecasting; there are few studies on probabilistic forecasting and feature selection from weather or time-oriented features in such forecasting. In this study, prediction intervals were generated by the lower upper bound estimation (LUBE) using neural networks with two outputs to make probabilistic modeling for predictions. The objective was to improve prediction interval coverage probability (PICP), mean prediction interval width (MPIW), continuous ranked probability score (CRPS), and loss, which is the integration of PICP and MPIW, by removing unnecessary features through feature selection. When features with high gain were selected by random forest (RF), in the modeling of 14.7 kW PV systems, loss improved by 1.57 kW, CRPS by 0.03 kW, PICP by 0.057 kW, and MPIW by 0.12 kW on average over two weeks compared to the case where all features were used without feature selection. Therefore, the low gain features from RF act as noise and reduce the modeling accuracy.

1. Introduction

Renewable energy sources, including photovoltaic (PV) generation, are being developed in many countries as the need for clean energy increases [1]. In particular, the number of PV installations has increased significantly in recent years due to the low cost of the modules, no carbon dioxide emissions, and the ease of installing the panels. However, PV power generation is highly variable and uncertain, as it is affected by weather conditions and other uncertain factors. This variability and uncertainty significantly impact the operation of power systems in which large amounts of PV power generation is installed. As a countermeasure to this problem, it is thought that the impact on the power systems can be mitigated by conducting PV power output forecasting, which is then considered in the operational plans of thermal power, hydroelectric power, and other power sources whose output can be adjusted. The probabilistic forecast for PV generation is especially useful for battery systems [2,3,4] in the distribution network. The operator of the battery systems needs to consider the risk of the peak as probability to determine the charge or discharge with a specific amount.
PV forecasting models can be divided into three main categories: physical models, statistical models, and hybrid models [5]. Physical models are constructed using numerical weather prediction (NWP) and satellite imagery; Miyazaki et al. [6] used optical flow to estimate the geographical motion of PV output lump related to the cloud motion. Saint-Drenan et al. [7] probabilistically predicted PV output from reference PV output derived from NWP and meteorological data. The statistical model, including machine learning, is a data-driven forecasting model, which constructs a forecast model based on historical data. Compared to physical models, statistical models rely solely on measured data and require no prior knowledge, making them easy to implement and highly adaptable. Therefore, research on predicting PV power output using statistical models, mainly machine learning, is becoming increasingly popular. In PV forecasting, autoregressive moving average methods, represented by time series models [8], and machine learning-based methods, such as neural network (NN) [9], support vector regression (SVR) [10,11], and long short-term memory (LSTM) [12], are used. Hybrid models are methods that improve prediction accuracy by combining statistical models or combining physical and statistical models to complement information that cannot be compensated by single models. Kodaira et al. [13] constructed an ensemble model of NN, k-means, and LSTM and used particle swarm optimization to optimize the weights of each predictive model to predict PV output. Yang et al. [14] classified training data using self-organizing map and learning vector quantization, trained them using SVR, and used phage inference to select the model to use on the forecast date to predict PV output.
Machine learning predictions can be broadly classified into two types: predictions based on deterministic point estimates and predictions based on interval estimates that account for uncertainty in a probabilistic manner. Compared to deterministic forecasting, probabilistic forecasting has more information because it accounts for the error. Therefore, probabilistic forecasts are useful for optimal operation and risk hedging in grids with significant PV generation. However, in conventional PV forecasting, forecasts are mainly based on deterministic point estimates. In probabilistic forecasting, Gaussian process [15] and quantile regression (QR) [16,17], natural gradient boosting (NGboost) [18] and others have been used to forecast PV power generation. Recently, lower upper bound estimation (LUBE) has attracted attention as a more direct method of outputting prediction intervals (PIs) without any special assumptions on the distribution [19]. LUBE has been applied in various fields of engineering. Khosravi et al. [20] used the LUBE to predict wind power generation, and Quan et al. [21] uses LUBE to forecast electrical loads. Ni et al. [22] used LUBE to forecast PV generation.
In machine learning, the features used in the prediction model have a direct impact on prediction accuracy. Incorporating unimportant features into the forecast model can lead to complex computational processing and reduced forecast accuracy due to learning unnecessary noise [23]. In general, weather variables such as temperature and solar radiation are used as features in PV forecasting. De Giorgi et al. [24] evaluated forecasting accuracy using NN for (1) a data set consisting of power generation only, (2) a data set consisting of power generation and solar radiation, and (3) a data set consisting of (2) plus module temperature and ambient temperature, and concluded that (3) was superior based on NRMSE, standard deviation, and other evaluation indicators. Zhong et al. [25] evaluated the prediction accuracy of NN, LSTM, etc., in one-hour-ahead forecasting assuming 11 sets of feature patterns from variables such as solar radiation, precipitation, temperature, water vapor mixing ratio, sea surface temperature, wind speed, etc. In NN, all meteorological variables plus accumulated solar radiation were superior. In LSTM, forecasts using features with added variables showed signs of over-learning, and forecasts that included only solar radiation and total precipitation as features were superior. C. Tao et al. [12] performed feature selection based on LightGBM from 29 features, including weather and time-oriented features. They selected 16 features, such as solar radiation and time-oriented features. With feature selection, there was a reduction in LSTM training time, a 10% improvement in RMSE, and a 20% improvement in MAE. In probabilistic forecasting, G. Mitrentsis et al. [18] evaluated the importance of the features to the PIs by Shapley additive explanation (SHAP) [26]. Of the 11 weather and time-oriented features, the lag features for solar radiation and output were highly important, while temperature, precipitation, and wind speed were less important. In probabilistic forecasting with NGboost using only important features, RMSE improved by 6 percent and continuous ranked probability score (CRPS) [27] by 10 percent compared to forecasting using all features. As shown in the cited papers on features, there are likely to be variations in the required features depending on the forecasting model and data set. Still, in the case of features with a strong relationship with PV output, such as solar radiation, they are likely to be required in any forecasting model. Similarly, there are features that are not needed in any forecast model. Therefore, in feature selection, it is important to select features that are considered necessary in most forecast models, and at the same time, it is important to determine features that are unnecessary in most forecast models.
The authors investigated studies on feature selection for deterministic forecasts in machine learning [12,20,21,22]. However, there are still few studies on feature selection for probabilistic forecasts. In this study, we used random forest (RF) for feature selection and LUBE and QR as probabilistic modeling to generate 24 h PIs for a PV power system. As a precaution, the temperature, solar radiation, and other features used in this study are not predictions, but actual measurements, so this is probabilistic modeling, not probabilistic forecasting. However, by examining which features are important, it can be applied to forecasting as well. The PV power system is in Tokyo, Japan, and is rated at 14.7 kW. Data of 14 features, including weather variables, were measured at a nearby weather station. The training data are approximately 10 months old, and the test period is 2 weeks. The objective of this study is to improve prediction interval coverage probability (PICP), mean prediction interval width (MPIW), CRPS, and Loss [28], which is the integration of PICP and MPIW, by removing unnecessary features through feature selection. The contributions of this study are as follows:
(1)
The effects of the 14 variables were evaluated by RF. Features were selected according to their gains, and PIs were generated by LUBE and QR.
(2)
When features with high gain were selected, there was an improvement in the accuracy of PIs in both QR and LUBE compared to when all features were used. However, features with low gain rarely contribute to or worsen the accuracy of PIs. According to Loss, the low gain features reduced the accuracy of PIs in LUBE, especially when PV output fluctuations were large.
The remainder of the paper is organized as follows: Section 2 describes feature selection with RF. Section 3 describes LUBE. Section 4 evaluates the accuracy of PIs by LUBE and QR based on feature selection by RF. Section 5 concludes the study.

2. Feature Selection Using Random Forest

Because PV power output fluctuates due to weather factors, modeling accuracy varies greatly depending on the features used in the model. If features with additional variables that are effective in modeling are used, the modeling accuracy improves; otherwise, the modeling accuracy deteriorates. It also increases the complexity of the model and computational processing. In this study, feature selection was performed by RF, which is an ensemble model that is obtained by modeling a forecasting model that repeats binary classification on a tree called a “decision tree” for each of the multiple samples generated using the bootstrap method, and averaging the forecasting accuracy [29]. RF has been proved useful in forecasting renewable energy sources such as solar power and wind power, which are sensitive to environmental factors [30,31]. The conceptual diagram of feature selection by RF is shown in Figure 1 and the following procedure is used [32,33].
(i).
From the training data consisting of n sets of P predictors and corresponding target variables, n sets are extracted, allowing for overlap. The extraction is repeated to generate K bootstrap samples. When generating bootstrap samples, approximately two-thirds of the training data are extracted at least once from the sample and about one-third are never extracted. The group of samples that are not extracted is referred to as out-of-bag (OOB).
(ii).
A decision tree is modeled for each of the K bootstrap samples and the mean square error (MSE) is obtained using OOB as test data. MSE OOB t represents the MSE when OOB t is the test data. Y is the target variable and the hat symbol represents the predicted value.
MSE OOB t = 1 n i = 1 n Y i Y i , t ^ 2
(iii).
One arbitrary feature from the OOB features is selected, randomly permuted, and the MSE is obtained again. This is repeated for all features.
MSE OOB t X j p e r m u t e d = 1 n i = 1 n Y i Y i , t ^ X j p e r m u t e d 2
where X j denotes the feature and j = 1 ,   2 ,   ,   P , MSE OOB t X j p e r m u t e d denotes the MSE at OOB t when the feature X j is permuted.
(iv).
The changes in MSE before and after permuting is obtained and the K results are averaged. Normalization is applied so that the sum of the importance of each feature is 1. If a feature is important to the accuracy of the forecast, permuting will significantly reduce the accuracy of the forecast. For unimportant features, the accuracy of the prediction is not affected.
MSE d i f X j = 1 K t = 1 K MSE OOB t X j p e r m u t e d MSE OOB t
g a i n X j = MSE d i f X j j = 1 P MSE d i f X j

3. Lower Upper Bound Estimation (LUBE)

LUBE is a nonparametric method that directly generates PIs using NNs with two outputs corresponding to the upper and lower bounds of the PIs [19]. Traditional methods such as the Delta and Bayesian methods are parametric methods that first perform point estimation and then generate PIs by assuming a distribution over the data [34,35]. LUBE directly generates PIs, which is simple, fast, and without special assumptions about the distribution or a large amount of computational work [21]. In LUBE, PIs are evaluated using the following three indicators. The conceptual diagram of the neural network in LUBE is shown in Figure 2.

3.1. Prediction Interval Coverage Probability (PICP)

PICP is an index that evaluates the percentage of measured values that fall within the interval of PIs and is one of the important evaluation indicators. The predicted lower and upper PI bounds are y ^ L i , y ^ U i . A vector, k , of length n represents whether each data point has been captured by the estimated PIs.
k i = 1 , i f   y ^ L i     y i     y ^ U i       0 ,                 e l s e
We define the total number of data points captured as c .
c   i = 1 n k i
PICP is defined by Equation (7).
PICP   c n

3.2. Mean Prediction Interval Width (MPIW)

MPIW represents the mean interval width and is an important metric for evaluating PIs; even if all measured values are within the interval and PICP is 100%, a too-wide MPIW implies high uncertainty and is meaningless as a forecast.
MPIW   1 c i = 1 n y ^ U i y ^ L i · k i

3.3. Loss

The two key metrics, PICP and MPIW, need to be evaluated simultaneously when generating PIs using LUBE. However, if the width of PIs is narrowed, PICP is likely to decrease because some measured values will drop out of the PIs. In LUBE, the two trade-off indicators are evaluated simultaneously using Loss [28].
Loss = MPIW + λ n α 1 α m a x 0 , 1 α PICP 2
where n is the number of data samples and α is a measure of the confidence level of the PIs. For example, when α = 0.05, 1 − α is 95%. At this time, the qualitative understanding in the equation is that if PICP does not exceed 0.95, a penalty is imposed by λ. In this case, Loss is the sum of MPIW, and the PICP term including λ. λ is the parameter that imposes a penalty and, at the same time, a tuning parameter that relates MPIW to PICP. If it exceeds 0.95, Loss is equal to MPIW. In PIs, when PICP is much lower than the established confidence level, the PIs can be considered lacking validity [14]. Therefore, in determining λ, there is need to adjust PICP and then consider the combination with MPIW.

3.4. Continuous Ranked Probability Score (CRPS)

In LUBE, the neural network learns to minimize Loss. Since the ratio of PICP and MPIW changes depending on the value of λ in Loss, the evaluation of PIs was also performed using CRPS [36]. The CRPS measures the difference between the predicted and observed cumulative distribution functions and is expressed by
c r p s F , y = F t H t y 2   d t
CRPS = 1 n i = 1 n c r p s F i , y i
where F is the predictive cumulative distribution function (CDF) and y is the verifying observation, H t y denotes the Heaviside function and takes the value 0 when t < y and the value 1 otherwise, and n is total points.
In Equation (11), when F follows a Gaussian distribution with mean μ and dispersion σ 2 , it is expressed by
c r p s N μ , σ 2 , y =   σ { y μ σ 2 Φ y μ σ 1 + 2 φ y μ σ 1 π }
where φ y μ σ and Φ y μ σ denote probability density functions (PDF) and the CDF, respectively, of the normal distribution with mean 0 and variance 1 evaluated at the normalized prediction error. To calculate the (12) expression in LUBE, we used proper scoring [37], an external python library. G. Mitrentsis et al. [18] used proper scoring to calculate CRPS for LUBE. However, since LUBE outputs PIs directly without assuming a distribution, care must be taken when evaluating calculations assuming a Gaussian distribution.

4. Case Study

For the probabilistic modeling of solar power generation, excluding unnecessary features improves modeling accuracy, but missing necessary features reduces modeling accuracy. In particular, PV modeling is difficult when output fluctuation is large, and if unnecessary features are included, the modeling accuracy is likely to be even lower. In this case study, we evaluated the importance of each feature for the probabilistic modeling. We first used RF to evaluate the gain of 14 features, consisting of weather and time-oriented features, for modeling PV power output. Next, PIs were generated using LUBE with a confidence level of 95%. In LUBE, we used 14 pairs of features added in order from variables with the highest gain. The 2-week average modeling accuracy of the PIs from each feature was evaluated, and the optimal features were considered. We also compared PIs by QR using the gradient boosted regression trees method with those of LUBE. See L. Massidda’s study [38] for details on the principle of QR. We also evaluated the robustness of the modeling by setting the number of simulations to 85 in LUBE. Finally, in relation to output fluctuation, we considered each day to evaluate when modeling was successful and when they failed, with and without feature selection.
The PV power plant subject of this study is in Tokyo, Japan, and has a capacity of 14.7 kW. PV power generation is observed every 30 min for 24 h. In addition to power generation data, atmospheric temperature, precipitation, degree of cloudiness, solar radiation, wind speed, and humidity were obtained from nearby meteorological observatories. The corresponding year, month, day, and hour were also added to the data set. The time data were expressed as a trigonometric function to account for periodicity. For example, the trigonometric representation of hour considers 24 h as one cycle, and daily considers the number of days in a month as one cycle. Azimuth, zenith, and declination [39] are also considered effective features, but are not included in this study. All features were standardized for uniformity of scale. Figure 3 shows how the data is divided. The observed data are for the period from 15 August 2013 to 14 June 2014. The feature selection by RF used data from 15 August 2013 to 31 May 2014. Training data for PIs generation by the LUBE and QR used data from 15 August 2013 to 16 May 2014. Validation data were from 17 May 2014 to 31 May 2014, and test data were from 1 June 2014 to 14 June 2014. Add data were from 1 June 2014 to the day before the test day. Adding the most recent data to the training and validation data was expected to improve modeling accuracy. Figure 4 shows a flowchart of LUBE using feature selection by random forest. We used NN consisting of an input layer, an intermediate layer with 100 neurons, and an output layer in LUBE. Regarding the NN parameters, from the combinations of nλ included in Equation (9), learning rate, and epochs listed in Table 1, the parameters with high PICP and narrow MPIW were adopted when all features were used. In QR, a grid search was performed from the parameters of Table 1 so that the CRPS would be small.

4.1. Result of the Feature Selection by Random Forest

Table 2 shows the results of feature selection using RF. The gain for solar radiation is 0.764, indicating that it is an extremely important feature. The hour sine and hour cosine have a combined gain of 0.172, and the annual cosine has a gain of 0.013. The gains for the other features are almost zero, indicating that they are of low importance.

4.2. Prediction Intervals by Lower Upper Boundary Estimation and Quantile Regression

4.2.1. Evaluation of Prediction Intervals at 2-Week Average

Figure 5 shows the correspondence between the number of features and the two-week average of Loss in LUBE and total gain. In Figure 5, the green triangle means median, and the small circle means outliers. LUBE(1) includes only the solar radiation, which is the highest gain. LUBE(4) includes solar radiation, hour sine, hour cosine, and annual cosine. LUBE(14) includes all features. The same applies to QR(1), (4), (14), etc. Table 3 shows the Loss statistics for LUBE(1), (4), and (14). Figure 6 shows a histogram of the results of 85 simulations of Loss for the 2-week average of LUBE(1), (4), and (14).
Figure 5 shows that features with a gain of nearly zero act as noise, reducing the accuracy and robustness of the modeling in LUBE. LUBE(4) has the smallest Loss, increases thereafter, and at LUBE(14) has the largest. The distribution of Loss in the box-and-whisker diagram is expanding from LUBE(4) to LUBE(14). To evaluate the modeling accuracy and robustness, we compared LUBE(14) with LUBE(1), which includes only solar radiation, and LUBE(4), which has the lowest Loss. In Table 3, LUBE(14) is significantly worse than LUBE(1) and LUBE(4) in all statistics. At the median, Loss is 1.38 kW greater than LUBE(1) and 1.57 kW greater than LUBE(4). In addition, as shown by the standard deviations in Table 3 and histograms in Figure 6, the distribution of LUBE(14) is wide, and the modeling accuracy is varied. These results indicate that features that are low in gain should be removed because they reduce not only modeling accuracy but also robustness.
We then similarly evaluated PICP and MPIW. Figure 7 and Figure 8 show the 2-week average PICP and MPIW in LUBE and QR, respectively. In Figure 7 and Figure 8, the green triangle means median, and the small circle means outliers. Table 4 and Table 5 show the PICP and MPIW statistics for LUBE(1), (4), (14), and QR(2), respectively; Figure 9 and Figure 10 show the histograms for LUBE(1), (4), and (14).
Figure 7 and Figure 8 show that the features after LUBE(4) work as noise in LUBE, because the PICP tends to decrease and MPIW tends to increase after LUBE(4). In QR, they also show features after QR(4) do not contribute to the prediction accuracy, because there is almost no change in PICP and MPIW after QR(4). Compared to LUBE, QR has wider MPIW and higher PICP. In Table 4 and Table 5, we evaluated the impact of features in LUBE. In Table 4, comparing median values, PICP is 0.072 and 0.057 higher for LUBE(1), and LUBE(4), respectively, compared to LUBE(14). In Table 5, MPIW is 0.12 kW narrower for LUBE(4) than for LUBE(14). However, in LUBE(1), MPIW is 0.42 kW wider than in LUBE(14). This means that LUBE(1) is highly uncertain because it contains only one feature. LUBE(4) is the narrowest MPIW in Table 5 because it contains enough features with high gain. Therefore, LUBE(4) outperforms LUBE(14) in MPIW and PICP as well. From the standard deviations in Table 4 and Table 5 and the histograms in Figure 9 and Figure 10, PICP and MPIW, as well as Loss, features with low gain reduce the robustness of the modeling. These results indicate that in PICP and MPIW, as in Loss, the features after LUBE(4) are noise.
Finally, using CRPS, we evaluated LUBE and QR. Figure 11 shows the correspondence between the number of features and the two-week average of CRPS in LUBE and QR, and total gain. In Figure 11, the green triangle means median, and the small circle means outliers. Table 6 shows the statistics of CRPS for the 2-week average of LUBE(1), (4), (14) and QR(2).
Figure 11 shows that both LUBE(1) and QR(1) have poor modeling accuracy because they only include solar radiation as a feature. In Figure 11, the CRPS after LUBE(4) and QR(4) is almost constant, indicating that features with low gain do not contribute to modeling accuracy. LUBE has the smallest CRPS in LUBE(4). However, QR has the smallest CRPS in QR(2), not QR(4). This is because QR has PICP close to 95% for any number of features in Figure 7, and QR(2) has the narrowest MPIW in Figure 8. Therefore, we can conclude that QR(2) and LUBE(4) use enough features as needed and have high modeling accuracy, and features with low gain do not contribute to improving modeling accuracy.

4.2.2. Accuracy Evaluation by Forecast Target Date

To evaluate the relationship between output fluctuation and modeling accuracy according to features, we analyze LUBE(14), which includes all features, LUBE(1), LUBE(4), and QR(2) by day as shown in Table 7 and Table 8. In Table 7 and Table 8, out of 85 simulations in LUBE, the median of Loss and corresponding MPIW, PICP, and CRPS for each day are shown. In Table 7 and Table 8, we evaluated the modeling accuracy using an index named Fluctuation, which is the average output fluctuations at 30-min intervals over a day. This study defines Fluctuation as indicated in Equation (13).
Fluctuation = i = 1 48 v t i v t i 1 48
where v t i represents the power output at time t i . Since output power is observed at 30-min intervals, there are 48 i in a day. For example, when Fluctuation is 0 kW, the daily generation output is constant. In Table 7 and Table 8, the threshold for Fluctuation is 0.5 kW, which is the average of the output fluctuations for all days, and days when the output fluctuations exceed 0.5 kW are highlighted in gray. On each day, the highest PICP and the lowest MPIW, Loss, and CRPS are shown in bold among LUBE(1), (4), (14), and QR(2). For example, on 2 June, Fluctuation is 0.582 kW, QR(2) has the highest PICP of 0.958 and LUBE(14) has the narrowest MPIW of 1.40 kW. LUBE(4) has the lowest Loss and CRPS at 2.30 kW and 0.48 kW, respectively. Table 9 shows the correlation coefficients between output fluctuations at 30-min intervals and PICP, MPIW, Loss, and CRPS for LUBE(1), (4), (14), and QR(2).
Based on the strong correlation between Loss, CRPS and 30-min output fluctuation for all methods in Table 9, it is easy to be modeling when the output fluctuation is small and difficult when it is large. When Fluctuation does not exceed 0.5 kW in Table 7 and Table 8, LUBE(4) is the best. From Average (Fluctuation < 0.5 kW) in Table 7 and Table 8, PICP, MPIW, Loss, and CRPS over 7 days is best for LUBE(4). When Fluctuation exceeds 0.5 kW, LUBE(4) or QR(2) is suitable for modeling. As indicated by the strong correlation between MPIW and output fluctuation for QR(2) in Table 9, QR(2) has a wide MPIW and high uncertainty on days when Fluctuation is large. However, PIs in QR(2) include many real values and have the largest average PICP in Table 7. In Table 8 Average (Fluctuation > 0.5 kW), CRPS is smallest for LUBE(4), but there are several days when it is more than 0.1 below the confidence level; if the PICP is much below the confidence level, the PIs are not considered valid, so LUBE(4) or QR(2) is appropriate for modeling. Throughout the 14 days, the CRPS and Loss of LUBE (14) are larger than LUBE (4) on all days. In particular, according to Loss, features with low gain on days with large output fluctuation act as noise in LUBE. Therefore, we can conclude that features with low gain should not be used.

4.2.3. Days of Maximum and Minimum Output Fluctuation

Figure 12a shows the PIs for LUBE(1), (4), (14) and QR (2) for the day with the lowest Fluctuation of 0.193 kW in Table 7 and Table 8. PICP is above 0.9 in all methods in Table 7. LUBE(1) has the smallest MPIW and Loss, and LUBE(4) is second in Table 7 and Table 8. CRPS is the smallest for LUBE(1), (4) in Table 8. Therefore, LUBE(1) is appropriate, followed by LUBE(4).
Figure 12b shows the PIs for LUBE(1), (4), (14) and QR (2) for the day with the largest Fluctuation of 0.968 kW in Table 7 and Table 8. LUBE(1) or QR (2) is best for 9 June. CRPS is the smallest in LUBE(4). However, PICP of LUBE(4) is 0.833, which is more than 0.1 below the confidence level. The CRPS of LUBE(1) and QR(2) is 0.70, the second smallest, and the PICP is about 0.875. Therefore, LUBE(1) or QR(2) is best.

5. Conclusions

The objective of this study was to improve the PICP, MPIW, Loss, and CRPS of PIs using LUBE and QR by removing unnecessary features. We considered the change in PI accuracy by incorporating, in order, the features with high gain evaluated by RF into the features used for LUBE and QR. In the number of features 1, which includes only solar radiation evaluated as the most important feature, there was high uncertainty and wide MPIW, high PICP, and large CRPS in LUBE and QR. In the number of features 14 without feature selection, features with a gain of nearly zero led to a decrease in the accuracy of PIs in LUBE and QR. In particular, according to Loss, features with low gain on days with large output fluctuation act as noise in LUBE. In QR, the number of features 2, which includes solar radiation and hour sine, and in LUBE, the number of features 4, which adds hour cosine and annual cosine to the above two features, contained enough features as needed and had the lowest CRPS and Loss. In LUBE, the number of features 4 improved Loss by 1.57 kW, CRPS by 0.03 kW, PICP by 0.057, and MPIW by 0.12 kW compared to the number of features 14 on a 14.7 kW PV system. However, on days with large output fluctuations, there are days when the PICP is more than 0.1 lower than the confidence level in LUBE. Therefore, the challenge remains to generate PIs with PICP close to the confidence level even on days with strong fluctuations.

Author Contributions

Conceptualization, H.Y. and D.K.; methodology, H.Y.; software, H.Y.; validation, H.Y.; formal analysis, H.Y.; investigation, H.Y.; resources, J.K.; data curation, H.Y.; writing—original draft preparation, H.Y.; writing—review and editing, D.K. and J.K.; visualization, H.Y.; supervision, D.K. and J.K.; project administration, D.K.; funding acquisition, D.K. and J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by JSPS KAKENHI, Grant Number JP21K14150.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank S. Ogata and M. Ohashi at OMRON Social Solutions Corporation for offering the output power data of the 14.7 kW PV system.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ahmed, R.; Sreeram, V.; Mishra, Y.; Arif, M.D. A Review and Evaluation of the State-of-the-Art in PV Solar Power Forecasting: Techniques and Optimization. Renew. Sustain. Energy Rev. 2020, 124, 109792. [Google Scholar] [CrossRef]
  2. Kodaira, D.; Jung, W.; Han, S. Optimal Energy Storage System Operation for Peak Reduction in a Distribution Network Using a Prediction Interval. IEEE Trans. Smart Grid 2020, 11, 2208–2217. [Google Scholar] [CrossRef]
  3. Rowe, M.; Yunusov, T.; Haben, S.; Singleton, C.; Holderbaum, W.; Potter, B. A Peak Reduction Scheduling Algorithm for Storage Devices on the Low Voltage Network. IEEE Trans. Smart Grid 2014, 5, 2115–2124. [Google Scholar] [CrossRef]
  4. Hafiz, F.; Awal, M.A.; De Queiroz, A.R.; Husain, I. Real-Time Stochastic Optimization of Energy Storage Management Using Deep Learning-Based Forecasts for Residential PV Applications. IEEE Trans. Ind. Appl. 2020, 56, 2216–2226. [Google Scholar] [CrossRef]
  5. Dolara, A.; Leva, S.; Manzolini, G. Comparison of Different Physical Models for PV Power Output Prediction. Sol. Energy 2015, 119, 83–99. [Google Scholar] [CrossRef] [Green Version]
  6. Miyazaki, Y.; Kameda, Y.; Kondoh, J. A Power-Forecasting Method for Geographically Distributed PV Power Systems Using Their Previous Datasets. Energies 2019, 12, 4815. [Google Scholar] [CrossRef] [Green Version]
  7. Saint-Drenan, Y.M.; Good, G.H.; Braun, M. A Probabilistic Approach to the Estimation of Regional Photovoltaic Power Production. Sol. Energy 2017, 147, 257–276. [Google Scholar] [CrossRef] [Green Version]
  8. Huang, R.; Huang, T.; Gadh, R.; Li, N. Solar Generation Prediction Using the ARMA Model in a Laboratory-Level Micro-Grid. In Proceedings of the 2012 IEEE 3rd International Conference on Smart Grid Communications, SmartGridComm 2012, Tainan, Taiwan, 5–8 November 2012; pp. 528–533. [Google Scholar]
  9. Liu, J.; Fang, W.; Zhang, X.; Yang, C. An Improved Photovoltaic Power Forecasting Model with the Assistance of Aerosol Index Data. IEEE Trans. Sustain. Energy 2015, 6, 434–442. [Google Scholar] [CrossRef]
  10. Theocharides, S.; Theristis, M.; Makrides, G.; Kynigos, M.; Spanias, C.; Georghiou, G.E. Comparative Analysis of Machine Learning Models for Day-Ahead Photovoltaic Power Production Forecasting. Energies 2021, 14, 1081. [Google Scholar] [CrossRef]
  11. Jang, H.S.; Bae, K.Y.; Park, H.S.; Sung, D.K. Solar Power Prediction Based on Satellite Images and Support Vector Machine. IEEE Trans. Sustain. Energy 2016, 7, 1255–1263. [Google Scholar] [CrossRef]
  12. Tao, C.; Lu, J.; Lang, J.; Peng, X.; Cheng, K.; Duan, S. Short-Term Forecasting of Photovoltaic Power Generation Based on Feature Selection and Bias Compensation–Lstm Network. Energies 2021, 14, 3086. [Google Scholar] [CrossRef]
  13. Generations, P.; Kodaira, D.; Tsukazaki, K.; Kure, T.; Kondoh, J. Improving Forecast Reliability for Geographically Distributed. Energies 2021, 14, 7340. [Google Scholar]
  14. Yang, H.T.; Huang, C.M.; Huang, Y.C.; Pai, Y.S. A Weather-Based Hybrid Method for 1-Day Ahead Hourly Forecasting of PV Power Output. IEEE Trans. Sustain. Energy 2014, 5, 917–926. [Google Scholar] [CrossRef]
  15. Sheng, H.; Xiao, J.; Cheng, Y.; Ni, Q.; Wang, S. Short-Term Solar Power Forecasting Based on Weighted Gaussian Process Regression. IEEE Trans. Ind. Electron. 2018, 65, 300–308. [Google Scholar] [CrossRef]
  16. Agoua, X.G.; Girard, R.; Kariniotakis, G. Probabilistic Models for Spatio-Temporal Photovoltaic Power Forecasting. IEEE Trans. Sustain. Energy 2019, 10, 780–789. [Google Scholar] [CrossRef] [Green Version]
  17. Almeida, M.P.; Muñoz, M.; de la Parra, I.; Perpiñán, O. Comparative Study of PV Power Forecast Using Parametric and Nonparametric PV Models. Sol. Energy 2017, 155, 854–866. [Google Scholar] [CrossRef] [Green Version]
  18. Mitrentsis, G.; Lens, H. An Interpretable Probabilistic Model for Short-Term Solar Power Forecasting Using Natural Gradient Boosting. Appl. Energy 2022, 309, 118473. [Google Scholar] [CrossRef]
  19. Khosravi, A.; Nahavandi, S.; Creighton, D.; Atiya, A.F. Lower Upper Bound Estimation Method for Construction of Neural Network-Based Prediction Intervals. IEEE Trans. Neural Netw. 2011, 22, 337–346. [Google Scholar] [CrossRef]
  20. Khosravi, A.; Nahavandi, S.; Creighton, D. Prediction Intervals for Short-Term Wind Farm Power Generation Forecasts. IEEE Trans. Sustain. Energy 2013, 4, 602–610. [Google Scholar] [CrossRef]
  21. Quan, H.; Srinivasan, D.; Khosravi, A. Uncertainty Handling Using Neural Network-Based Prediction Intervals for Electrical Load Forecasting. Energy 2014, 73, 916–925. [Google Scholar] [CrossRef]
  22. Ni, Q.; Zhuang, S.; Sheng, H.; Kang, G.; Xiao, J. An Ensemble Prediction Intervals Approach for Short-Term PV Power Forecasting. Sol. Energy 2017, 155, 1072–1083. [Google Scholar] [CrossRef]
  23. Raza, M.Q.; Nadarajah, M.; Ekanayake, C. On Recent Advances in PV Output Power Forecast. Sol. Energy 2016, 136, 125–144. [Google Scholar] [CrossRef]
  24. De Giorgi, M.G.; Congedo, P.M.; Malvoni, M. Photovoltaic Power Forecasting Using Statistical Methods: Impact of Weather Data. IET Sci. Meas. Technol. 2014, 8, 90–97. [Google Scholar] [CrossRef]
  25. Zhong, Y.J.; Wu, Y.K. Short-Term Solar Power Forecasts Considering Various Weather Variables. In Proceedings of the 2020 International Symposium on Computer, Consumer and Control, IS3C 2020, Taichung City, Taiwan, 13–16 November 2020; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2020; pp. 432–435. [Google Scholar]
  26. Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.-I. From Local Explanations to Global Understanding with Explainable AI for Trees. Nat. Mach. Intell. 2020, 2, 56–67. [Google Scholar] [CrossRef] [PubMed]
  27. Lauret, P.; David, M.; Pinson, P. Verification of Solar Irradiance Probabilistic Forecasts. Sol. Energy 2019, 194, 254–271. [Google Scholar] [CrossRef] [Green Version]
  28. Pearce, T.; Zaki, M.; Brintrup, A.; Neely, A. High-Quality Prediction Intervals for Deep Learning: A Distribution-Free, Ensembled Approach. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Volume 9, pp. 6473–6482. [Google Scholar]
  29. Breiman, L. Random Forests; Springer: Berlin/Heidelberg, Germany, 2001; Volume 45. [Google Scholar]
  30. Kim, S.G.; Jung, J.Y.; Sim, M.K. A Two-Step Approach to Solar Power Generation Prediction Based on Weather Data Using Machine Learning. Sustainability 2019, 11, 1501. [Google Scholar] [CrossRef] [Green Version]
  31. Lahouar, A.; Ben Hadj Slama, J. Hour-Ahead Wind Power Forecast Based on Random Forests. Renew. Energy 2017, 109, 529–541. [Google Scholar] [CrossRef]
  32. Grömping, U. Variable Importance Assessment in Regression: Linear Regression versus Random Forest. Am. Stat. 2009, 63, 308–319. [Google Scholar] [CrossRef]
  33. Svetnik, V.; Liaw, A.; Tong, C.; Christopher Culberson, J.; Sheridan, R.P.; Feuston, B.P. Random Forest: A Classification and Regression Tool for Compound Classification and QSAR Modeling. J. Chem. Inf. Comput. Sci. 2003, 43, 1947–1958. [Google Scholar] [CrossRef]
  34. Khosravi, A.; Nahavandi, S.; Creighton, D.; Atiya, A.F. Comprehensive Review of Neural Network-Based Prediction Intervals and New Advances. IEEE Trans. Neural Netw. 2011, 22, 1341–1356. [Google Scholar] [CrossRef]
  35. Panamtash, H.; Zhou, Q.; Hong, T.; Qu, Z.; Davis, K.O. A Copula-Based Bayesian Method for Probabilistic Solar Power Forecasting. Sol. Energy 2020, 196, 336–345. [Google Scholar] [CrossRef]
  36. Gneiting, T.; Raftery, A.E.; Westveld, A.H.; Goldman, T. Calibrated Probabilistic Forecasting Using Ensemble Model Output Statistics and Minimum CRPS Estimation. Mon. Weather. Rev. 2005, 133, 1098–1118. [Google Scholar] [CrossRef]
  37. Stephan Hoyer Properscoring. Available online: https://github.com/TheClimateCorporation/properscoring.git (accessed on 17 May 2022).
  38. Massidda, L.; Marrocu, M. Quantile Regression Post-Processing Ofweather Forecast for Short-Term Solar Power Probabilistic Forecasting. Energies 2018, 11, 1763. [Google Scholar] [CrossRef] [Green Version]
  39. Markovics, D.; Mayer, M.J. Comparison of Machine Learning Methods for Photovoltaic Power Forecasting Based on Numerical Weather Prediction. Renew. Sustain. Energy Rev. 2022, 161, 112364. [Google Scholar] [CrossRef]
Figure 1. Feature selection by RF.
Figure 1. Feature selection by RF.
Energies 15 05337 g001
Figure 2. LUBE.
Figure 2. LUBE.
Energies 15 05337 g002
Figure 3. Lower upper bound estimation method with random forest.
Figure 3. Lower upper bound estimation method with random forest.
Energies 15 05337 g003
Figure 4. Flowchart of LUBE with feature selection by random forest.
Figure 4. Flowchart of LUBE with feature selection by random forest.
Energies 15 05337 g004
Figure 5. Number of features, 2-week average Loss in LUBE, and total gain.
Figure 5. Number of features, 2-week average Loss in LUBE, and total gain.
Energies 15 05337 g005
Figure 6. Histograms of 2-week average Loss for LUBE(1), (4), and (14).
Figure 6. Histograms of 2-week average Loss for LUBE(1), (4), and (14).
Energies 15 05337 g006
Figure 7. Number of features and 2-week average PICP in LUBE and QR.
Figure 7. Number of features and 2-week average PICP in LUBE and QR.
Energies 15 05337 g007
Figure 8. Number of features and 2-week average MPIW in LUBE and QR.
Figure 8. Number of features and 2-week average MPIW in LUBE and QR.
Energies 15 05337 g008
Figure 9. Histograms of 2-week average PICP for LUBE (1), (4), and (14).
Figure 9. Histograms of 2-week average PICP for LUBE (1), (4), and (14).
Energies 15 05337 g009
Figure 10. Histograms of 2-week average MPIW for LUBE(1), (4), and (14).
Figure 10. Histograms of 2-week average MPIW for LUBE(1), (4), and (14).
Energies 15 05337 g010
Figure 11. Number of features, 2-week average CRPS in LUBE and QR, and total gain.
Figure 11. Number of features, 2-week average CRPS in LUBE and QR, and total gain.
Energies 15 05337 g011
Figure 12. Prediction intervals for LUBE(1), (4), (14), and QR(2): (a) 6 June, the day with the smallest output fluctuation in two weeks.; (b) 9 June, the day of the largest output fluctuation in two weeks.
Figure 12. Prediction intervals for LUBE(1), (4), (14), and QR(2): (a) 6 June, the day with the smallest output fluctuation in two weeks.; (b) 9 June, the day of the largest output fluctuation in two weeks.
Energies 15 05337 g012
Table 1. Grid search for neural network and quantile regression.
Table 1. Grid search for neural network and quantile regression.
ModelParameterSearch ValuesSelected Value
LUBE3, 5, 75
learning rate0.01, 0.001, 0.00010.001
epochs100, 500, 1000, 30003000
QRlearning rate0.05, 0.02, 0.010.05
maximum depth of trees3, 4, 53
minimum number of samples in leaf3, 5, 77
number of estimates100, 200, 500500
Table 2. Evaluating the importance of features using random forest.
Table 2. Evaluating the importance of features using random forest.
FeatureGainFeatureGain
solar radiation0.7644atmospheric temperature0.0065
hour sine0.0961monthly cosine0.0062
hour cosine0.0757wind speed0.0057
annual cosine0.0130daily sine0.0038
degree of cloudiness0.0076daily cosine0.0037
annual sine0.0069monthly sine0.0025
humidity0.0069precipitation0.0003
Table 3. Statistics of 2-week average Loss.
Table 3. Statistics of 2-week average Loss.
Number of FeaturesMinMaxMeanMedianStandard Deviation
LUBE(1)2.282.372.322.320.0193
LUBE(4)1.952.302.132.130.0861
LUBE(14)2.656.033.743.700.5967
Table 4. Statistics of 2-week average PICP.
Table 4. Statistics of 2-week average PICP.
Number of FeaturesMinMaxMeanMedianStandard Deviation
LUBE(1)0.9070.9210.9160.9170.00231
LUBE(4)0.8860.9160.9030.9020.00537
LUBE(14)0.7950.8860.8450.8450.0160
QR(2)---0.942-
Table 5. Statistics of 2-week average MPIW.
Table 5. Statistics of 2-week average MPIW.
Number of FeaturesMinMaxMeanMedianStandard Deviation
LUBE(1)2.022.132.072.070.0151
LUBE(4)1.471.661.531.530.0340
LUBE(14)1.531.781.641.650.0536
QR(2)---2.51-
Table 6. Statistics of 2-week average CRPS.
Table 6. Statistics of 2-week average CRPS.
Number of FeaturesMinMaxMeanMedianStandard Deviation
LUBE(1)0.530.540.530.530.0024
LUBE(4)0.390.430.410.410.0056
LUBE(14)0.420.460.440.440.0064
QR(2)---0.51-
Table 7. Evaluation for LUBE(1), (4), (14) and QR(2) by PICP and MPIW from 1 June to 14 June.
Table 7. Evaluation for LUBE(1), (4), (14) and QR(2) by PICP and MPIW from 1 June to 14 June.
PICPMPIW [kW]
   
DayFluctuation [kW]LUBE(1)LUBE(4)LUBE(14)QR(2)LUBE(1)LUBE(4)LUBE(14)QR(2)
10.4890.9580.9790.9790.9382.451.552.022.91
20.5820.8960.8640.7910.9582.501.521.403.01
30.5050.9160.8750.8311.0002.501.611.582.88
40.5220.8950.8530.8340.9382.411.491.592.86
50.3500.9790.9790.9791.0001.741.662.032.13
60.1930.9160.9840.9111.0000.851.441.611.48
70.2330.9370.9841.0001.0001.251.461.921.86
80.4120.8940.9170.8750.8961.641.651.502.25
90.9680.8740.8330.7650.8752.731.631.472.90
100.9090.8350.8490.8150.9382.531.631.662.86
110.3080.9570.8790.8760.8751.431.101.441.69
120.2710.9370.9560.9580.9581.921.581.762.34
130.8070.8940.8080.7390.8962.561.361.533.03
140.5130.9370.9740.9370.9172.431.601.682.90
Average
(Fluctuation > 0.5 kW)
0.6870.8920.8650.8160.9322.521.551.562.92
Average
(Fluctuation < 0.5 kW)
0.3220.9400.9540.9400.9521.611.491.752.09
Average
(All days)
0.5040.9160.9100.8780.9422.071.521.662.51
Table 8. Evaluation for LUBE (1), (4), (14) and QR (2) by Loss and CRPS from June 1 to 14.
Table 8. Evaluation for LUBE (1), (4), (14) and QR (2) by Loss and CRPS from June 1 to 14.
Loss [kW]CRPS [kW]
   
DayFluctuation [kW]LUBE(1)LUBE(4)LUBE(14)LUBE(1)LUBE(4)LUBE(14)QR(2)
10.4892.451.552.020.620.260.270.63
20.5822.812.304.050.750.480.490.55
30.5052.622.213.070.580.480.530.43
40.5222.732.493.000.610.450.470.38
50.3501.741.662.030.260.260.310.31
60.1930.971.441.770.200.200.220.28
70.2331.271.461.920.200.150.220.30
80.4121.961.762.090.380.350.410.40
90.9683.333.085.070.870.640.700.70
100.9093.932.703.570.910.690.720.73
110.3081.431.642.010.290.420.440.51
120.2711.931.581.760.310.240.240.36
130.8072.893.506.220.870.730.760.89
140.5132.451.601.690.620.280.290.73
Average
(Fluctuation > 0.5 kW)
0.6872.972.553.810.740.540.570.63
Average
(Fluctuation < 0.5 kW)
0.3221.681.581.940.320.270.300.40
Average
(All days)
0.5042.322.072.880.530.400.430.51
Table 9. Correlation coefficient with 30-min output fluctuation for LUBE (1), (4), (14), and QR (2).
Table 9. Correlation coefficient with 30-min output fluctuation for LUBE (1), (4), (14), and QR (2).
MetricPICPMPIW
MethodLUBE(1)LUBE(4)LUBE(14)QR(2)LUBE(1)LUBE (4)LUBE(14)QR(2)
Correlation coefficient−0.726−0.758−0.762−0.510.8180.258−0.3550.862
MetricLossCRPS
MethodLUBE(2)LUBE(4)LUBE(14)QR(2)LUBE(1)LUBE(4)LUBE(14)QR(2)
Correlation coefficient0.9320.8790.821-0.9490.8940.8910.807
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yamamoto, H.; Kondoh, J.; Kodaira, D. Assessing the Impact of Features on Probabilistic Modeling of Photovoltaic Power Generation. Energies 2022, 15, 5337. https://doi.org/10.3390/en15155337

AMA Style

Yamamoto H, Kondoh J, Kodaira D. Assessing the Impact of Features on Probabilistic Modeling of Photovoltaic Power Generation. Energies. 2022; 15(15):5337. https://doi.org/10.3390/en15155337

Chicago/Turabian Style

Yamamoto, Hiroki, Junji Kondoh, and Daisuke Kodaira. 2022. "Assessing the Impact of Features on Probabilistic Modeling of Photovoltaic Power Generation" Energies 15, no. 15: 5337. https://doi.org/10.3390/en15155337

APA Style

Yamamoto, H., Kondoh, J., & Kodaira, D. (2022). Assessing the Impact of Features on Probabilistic Modeling of Photovoltaic Power Generation. Energies, 15(15), 5337. https://doi.org/10.3390/en15155337

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop