Next Article in Journal
Two-Feeder Dynamic Voltage Restorer for Application in Custom Power Parks
Next Article in Special Issue
Extended Flow-Based Security Assessment for Real-Sized Transmission Network Planning
Previous Article in Journal
Reactive Power Optimization of a Distribution System Based on Scene Matching and Deep Belief Network
Previous Article in Special Issue
Long-Term Demand Forecasting in a Scenario of Energy Transition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improvement of Short-Term BIPV Power Predictions Using Feature Engineering and a Recurrent Neural Network

1
Department of Architectural Engineering, Hanyang University, Seoul 04763, Korea
2
Department of Architectural Engineering, Cheongju University, Cheongju 28503, Korea
3
Department of Architecture, Cheongju University, Cheongju 28503, Korea
*
Author to whom correspondence should be addressed.
Energies 2019, 12(17), 3247; https://doi.org/10.3390/en12173247
Submission received: 28 June 2019 / Revised: 5 August 2019 / Accepted: 20 August 2019 / Published: 23 August 2019
(This article belongs to the Special Issue Electricity for Energy Transition)

Abstract

:
The time resolution and prediction accuracy of the power generated by building-integrated photovoltaics are important for managing electricity demand and formulating a strategy to trade power with the grid. This study presents a novel approach to improve short-term hourly photovoltaic power output predictions using feature engineering and machine learning. Feature selection measured the importance score of input features by using a model-based variable importance. It verified that the normative sky index in the weather forecasted data had the least importance as a predictor for hourly prediction of photovoltaic power output. Six different machine-learning algorithms were assessed to select an appropriate model for the hourly power output prediction with onsite weather forecast data. The recurrent neural network outperformed five other models, including artificial neural networks, support vector machines, classification and regression trees, chi-square automatic interaction detection, and random forests, in terms of its ability to predict photovoltaic power output at an hourly and daily resolution for 64 tested days. Feature engineering was then used to apply dropout observation to the normative sky index from the training and prediction process, which improved the hourly prediction performance. In particular, the prediction accuracy for overcast days improved by 20% compared to the original weather dataset used without dropout observation. The results show that feature engineering effectively improves the short-term predictions of photovoltaic power output in buildings with a simple weather forecasting service.

Graphical Abstract

1. Introduction

The development of technologies to harvest renewable forms of energy, such as solar photovoltaics and wind power generators, is in fact one of the key drivers for their implementation in microgrids, interconnected with the grid, to trade generated electricity [1]. Although clean technologies address the need for sustainable energy, their inherent variability and dependence on weather conditions complicate the management of the power network [2]. Consequently, it is only prudent to require that earlier methods, which have always been used to balance power generation and consumption, be updated to incorporate these complications [3]. For example, to address the variability in the power demand on microgrids, the first step must be to predict the short-term load demand [4,5]. This is necessary to obtain the amount of energy that will be consumed to efficiently balance the supply and demand sides of a system with power generators. The second step is to predict the energy generated by renewable energy sources, in order to continuously update the status of the power network. For example, solar power generation is known to be directly related to onsite weather conditions [2]. Therefore, solar power is not only diurnal in nature but also varies throughout the day with changes in the solar irradiation [6]. Photovoltaic (PV) power output predictions require precursor knowledge of the weather parameters.
Different models and parameters that are known to have an influence, such as cloudiness and solar irradiance, have been used for PV power output prediction [7]. Previous research has suggested that the prediction of renewable energy generation is an approach to balance supply and demand in electricity energy management [1,8,9].
In recent years, numerous prediction methods for PV power systems have been developed with the use of statistical models and machine-learning algorithms. Researchers have attempted to overcome many difficulties to present accurate methods to predict PV power output. Several prediction methods have had different levels of success in improving the accuracy and reducing the complexity in terms of computational cost. These methods can be categorized into direct and indirect.
Indirect prediction methods estimate the PV power output by using or predicting the solar irradiance over different time scales. They can use numerical weather forecast information, image processing, and statistical and hybrid artificial neural networks to predict solar irradiance [10]. Several hybrid models have been proposed to predict PV power output based on the indirect method. Filipe et al. [11] suggested a combination of hybrid solar power prediction methods, consisting of both statistical and physical methods. In other words, an electrical model of a PV system is combined with a statistical model that converts numerical weather forecasted data into solar power with a short-term predictive horizon for the physical model. Dong et al. [12] developed a hybrid prediction method for hourly solar irradiance by integrating self-organizing map, support vector regression, and particle swarm optimization approaches. Alternatively, PV power can be predicted using the indirect method with the aid of simulation tools, such as TRNSYS, PVFORM, and HOMER [13]. Although many indirect approaches have been developed, the prediction accuracy depends on the accuracy of solar irradiation predictions. However, the prediction performance of solar irradiation is still not good because of the quality of forecasted weather data [14]. It also has limited applicability for typical buildings that do not have solar meters to measure the various forms of solar irradiation, such as direct normal irradiance (DNI) or global horizontal irradiance (GHI).
Direct prediction methods employ historical PV power generation data and forecasted weather conditions that generally do not include solar irradiation. Several studies have reviewed the methods that enhance building-integrated photovoltaic (BIPV) power prediction using ensemble methods. Wan et al. [15] analyzed different PV power and solar irradiance prediction techniques. Raza et al. [16] focused on PV power prediction models. Gandoman et al. [17] evaluated the prediction performance of PV power output based on the short-term influences of cloud cover. Machine learning was used to develop direct prediction methods. Shi et al. [18] developed a 1-day-ahead PV power output prediction based on a support vector machine (SVM) and the features of weather categorization at a PV station in China. Mellit et al. [19] proposed a short-term prediction of meteorological time-series parameters using the least squares SVM. Although numerous studies have suggested the prediction of PV power output using various prediction algorithms and hybrid models based on the direct method, conventional direct methods have limited ability to maintain a high hourly prediction performance of short-term PV power output because these models mainly depend on forecasted weather data, which does not include solar irradiance. In addition, prior studies have focused on the type of algorithm and configuration of the particular models to improve the prediction accuracy. Therefore, both input features, such as weather information, and the prediction model configuration must be considered to improve the hourly prediction accuracy.
This study proposes an improved method for short-term prediction of BIPV power generation with simple weather forecast data using feature engineering and machine learning. First, feature selection is performed by using the weather forecast parameters to identify the impact of input variables used for BIPV prediction by taking into account the importance of each variable. The BIPV prediction performance of several machine-learning algorithms are then compared to select an appropriate model.
This study is organized as follows: Section 2 introduces the framework of the prediction models and performance improvement techniques. Section 3 characterizes the weather features by variable importance. Section 4 discusses the prediction performance with several machine-learning models. The main results of the improvement techniques are presented in Section 5 and the conclusions appear in Section 6.

2. Methodology

The proposed short-term prediction model for BIPV power systems is a combination of a feature engineering technique and machine-learning model. Figure 1 shows the framework, which is composed of three steps: data acquisition, the selection of a machine-learning model for BIPV power output prediction, and a technique to improve the prediction performance. First, training (historical) and testing (forecasted) datasets were constructed by collecting data from the online weather service and historical hourly PV power outputs. Second, the lowest correlated feature in the weather forecast data was selected by feature selection, then the most optimal BIPV prediction model was selected by comparing six conventional machine-learning models using the original input data set. Finally, this study suggests a method to improve prediction accuracy by using the dropout observation algorithm, one of the feature engineering approaches.

2.1. Characteristics of the BIPV

Photovoltaic power generation systems were installed on the roof of a medium-sized office building in South Korea at a latitude 37°31′N and longitude 127°14′E, as shown in Figure 2. The capacity of this photovoltaic generation system is approximately 50 kW and the specifications of the BIPV system are listed in Table 1. The BIPV power was measured in minutes and then aggregated into a relational database management system (RDBMS). For the purpose of short-term prediction, this study collected hourly BIPV power measurements from June 2017 to August 2018.

2.2. Description of Predictors

The forecasted weather dataset containing the historical weather data was obtained from the Korea Meteorological Administration (KMA) and Korea Astronomy and Space Science Institute (KASI). Seven weather variables were used as predictors: outdoor air temperature (OT), relative humidity (RH), wind direction (WD), wind speed (WS), normative sky index (NSI), precipitation (PP), and solar altitude (SA). The entire set of weather variables consists of data collected at 3-hour intervals. In this study, these variables, with 3-hour resolution, were pre-processed by converting them to an hourly resolution to perform the short-term prediction for the hourly BIPV power system.

2.3. Prediction Algorithms

BIPV power output mainly depends on the input variables (i.e., weather data and system characteristics) and prediction model. The performance of the prediction model is affected by the model configuration and type. This study assessed six machine-learning models, i.e., an artificial neural network (ANN), support vector machine (SVM), classification and regression tree (CART), chi-square automatic interaction detection (CHAID), random forest (RF), and recurrent neural network (RNN), with the objective of selecting an appropriate prediction model. Each prediction model is implemented with MATLAB (version R2016b, MathWorks, USA) with Windows 10.
The ANN model, which is effective when solving nonlinear problems, has a structure that consists of an input layer, hidden layers, and one output layer. Learning occurs by assigning weights and bias to each layer [20]. Various neural network models exist, which differ in terms of their learning process, model structure, and other features. Among these models, a feed-forward neural network (FFNN) with back-propagation was selected for this study because of its simple model configuration and computational speed. FFNN functions by relocating the information in the former layer to the next layer in a forward direction [21]. In addition, FFNN has a hyperbolic tangent function, which is used to transfer the trained information from the hidden layer to the output layer.
The SVM has been widely applied for pattern recognition (e.g., classification and regression problems). SVM uses a hyperplane and margin to identify the support vectors of the datasets. In particular, the original datasets are mapped into higher dimensional space to classify them using kernel functions capable of effectively solving nonlinear classification problems [22]. Here, a radial basis function (RBF) is applied, which has been proven to outperform kernel functions due to its computational efficiency, simplicity, and reliability compared with other functions [23].
Among the various algorithms based on the decision trees, the CART and CHAID models are the most popular algorithms. These nonparametric methods can be used to process various data types, such as nominal, ordinal, and continuous variables [24]. The two models differ significantly. The CART model grows the next tree by splitting two child nodes. The main predictor is generated by measuring an impurity or diversity. CHAID constructs multiple split nodes and performs statistical tests when each node is split [25].
The RF is an ensemble-learning algorithm that solves both classification and regression problems. Previous studies, especially in recent years, have shown that this algorithm is capable of solving real-world problems in many fields. In particular, RF is a more robust model than models based on a single decision tree, such as CART and CHAID, because several decision tree models are used and then combined using a bootstrap. Furthermore, each tree with the lowest error rate is replaced in the original dataset before the next tree is grown [26,27].
The RNN [28] is a modification of the ANN and has a circulated connection between nodes of layers to maintain an inner state, compared with feed forward connections. The circulated connection allows the network to act on information from previous steps in the computational sequence; thus, it exhibits dynamic temporal behavior. This characteristic makes RNNs advantageous to use for the analysis of sequential data. Applying dropout to a neural network amounts to sampling a “thinned” network from it. As a recurrent neural network, long short-term memory (LSTM) cells were used, which solve the problem of a vanishing gradient during training in this study. LSTM cells contain an explicit state. The value stored in this state is then regulated by gates within the cell. These gates have specific rules that define when to store, update, or forget the value in the internal state [29]. Considering model constitution, RNN is an appropriate model for time series prediction.
To determine the optimal conditions for each machine-learning model, cross-validation was implemented with different parameters based on [30] in MATLAB. The obtained configurations of each prediction model are listed in Table 2.

2.4. Variable Importance

Variable importance analysis is required to determine which variables contain noise that would prevent accurate prediction. Variable importance analysis is a key pre-processing step for any appropriate prediction method. It is important to identify the potential predictors that have a high impact on the function response. The use of SVM for variable importance analysis was applied for this study.
Among the model-based variable importance approaches such as linear models, random forests, and bagged/boosted trees, SVM has been reported as a high performance in many classification and prediction problems due to its strong generalization ability and robust feature selection method [31,32,33]. Variable importance analysis with SVM uses gradient descent to minimize generalization bounds. The SVM develops an optimal hyperplane in the dimensional space of x     R n into a high (possibly infinite) dimensional space and constructs an optimal hyperplane in this space [34]. The decision function of the SVM classification for a new sample of x is defined by Equation (1).
f ( x ) = W · ( x ) + b = i α i y i K ( x i ,   x ) + b
where ( x ) is the maximal margin function from the hyperplane, W is the weight, and b is the bias of the SVM decision function. K(x) is the kernel function, and α i denotes Lagrange multipliers corresponding to primal constraints. x i denotes training vectors, and y i denotes the corresponding labels of ± 1 . The SVM attempts to find the function from the dataset, by solving Equation (2) with Lagrange multipliers, it can reduce to maximizing the following optimization problem.
W 2 ( α ) =   i = 1 l α i 1 2 i , j = 1 l α i α j y i y j K ( x i ,   x j )
The SVM is used for analyzing the input variables and output variables by changing the cost function. The SVM identifies the variable importance by using the gradient of weight (∆W) between the weight with the specific variable and that without the variable given by Equation (3). A larger change in the weight illustrates that the specific variable impacts the result more.
W = | W W witout   specific   variable |

2.5. Observation Dropout

The most popular method for network regularization is dropout, which randomly selects certain activation portions in networks. Dropout regularizes data by removing layers or nodes in the neural network [35].
Different dropout models were initially developed to improve the manner in which strong networks can facilitate the interpretation of decisions, e.g., the assignment of a particular label to an image similar to the one in question [36,37,38]. A number of studies have highlighted the use of various dropout methods as a way to improve model performance [39]. Kingma et al. [40] proposed the Bayes form of interpretation [41] by compressing the architectural model [42]. These proposals were aimed at transforming the rate of dropout to optimize model performance. The units that remain after the dropout are those that form the network. Dropout is commonly used for regularization to enhance the performance. The dropout observation method was introduced to analyze which features are relevant for a predicted target variable.
Observation dropout is used in the algorithm proposed in this study based on feature dropout [43] of the neural network models. When using feature dropout, specific feature vectors are allocated noise with the dropout procedure for each training instance, where the noise is randomly allocated to data that has a specific probability or zero as its value. The effects of feature dropout may maximize the performance contribution of other features based on the proposed neural network model [44].
However, in this study, compared with the original feature dropout technique, several observations with specific probabilities in the specific feature for each training instance were removed, as shown in Figure 3. The removal of certain data in the specific feature implies that the addition of noise to unrelated data improves the performance, which has an identical effect as using the dropout in the neural network methods. SVM identified the lowest correlated feature by considering the importance of each variable. Then, a regularization method was applied to the specific feature such that its value became zero in the training and prediction process with a probability of p.
These units, n, can be used for a hyper parameter to define the number of neural nets. In the presentation of each training case, a thinned network is trained and sampled. This indicates that, when a neural network is trained with dropout, it is equivalent to training 2n networks that are thinned and share the weight. In doing so, thinned networks are largely prevented from becoming strained. After a certain time, clearly estimating the different models that were thinned would become unsustainable. Nevertheless, a simple approach to estimation can be effective. It is logical to use one neural network at a time in the absence of dropout. The pressure exerted by these network forms is a representation of strained weights. In the event that a given unit is strained with a possibility, p, at the time of training, the outgoing weights for that unit are increased by p at the time of the prediction. This ensures that any output that was hidden is similar to the exact output at the time of the prediction. When scaling is implemented, 2n nets, whose weights are shared, are collected to form one neural network for the prediction time. Training networks with dropout, as well as the use of proper averaging procedures at the prediction time, have been found to reduce the number of errors due to generalization, for numerous classification challenges when compared with the use of other methods of training regularization [35].
In this methodology, during training, random probabilities are multiplied with a specific feature of the input training data to achieve dropout, which yields an approximate averaging method of the regularization at the prediction time using probability, p, and that without the regularization (1 − p) as shown in Figure 4.

3. Feature Selection by Variable Importance

As described in Section 2.4, the SVM measures the importance scores by the prediction performance, incorporating the correlation structure between the forecasted weather variables. Table 3 provides the importance score of each weather feature based on the historical dataset for the four types of daily sky conditions. SA had the highest importance score for all sky conditions among the variables. Cloudier skies correspond to higher variable importance scores for SA due to the correlation between BIPV power generation and the angle of solar irradiation.
Although NSI indicates clearness of the sky, it was the least important factor with a variable importance score of 0.01 in clear and slightly cloudy days and 0.00 in cloudy and overcast days. This is mainly because of the characteristics of NSI data. KMA provides the float numbers for OT, RH, WS, and the degrees for WD (0~360) and SA (−90 ~ +90), whereas NSI has just four step integers: 1-no cloud, 2-intermittant cloud, 3-cloudy, and 4-overcast at intervals of 3 h with a geological resolution of 5 km by 5 km. As shown in Figure 5, as an example, the integer value of NSI is almost constant over a day, but the PV power output varies depending on the time and the SA. This illustrates the reason for the low correlation between NSI and PV power outputs. This feature plays a role in increasing the complexity of the prediction model and disrupts the prediction accuracy of the BIPV power output.

4. Model Selection

This section presents the proposed BIPV prediction model, which was used to compare six machine-learning models: ANN, SVM, CART, CHAID, RF, and RNN.
The hourly error of the prediction models was compared for evaluating model performance by the coefficient of variation of the root mean square error (CV(RMSE)), the mean absolute deviation (MAD), and the mean absolute percentage error (MAPE). Each index was calculated using the following equations.
CV ( RMSE ) = 1 n i = 1 n ( A i P i ) 2 A i ¯
MAD = 1 n i = 1 n | A i P i |
MAPE = 1 n i = 1 n | A i P i A i | × 100 ( % )
where n is the number of data values, A i is the actual BIPV power output, and P i is the predicted value of the BIPV power outputs.
Figure 6 shows the results of the hourly BIPV predictions obtained using the six different machine-learning models with the four types of typical daily sky conditions. The results for each model show that the RNN outperformed the other models with an average hourly MAPE of 48.68% for four days, followed by RF (MAPE = 61.88%), CART (MAPE = 62.17%), SVM (MAPE = 66.63%), ANN (MAPE = 70.93%), and CHAID (MAPE = 98.52%). Table 4 lists the average hourly CV(RMSE), MAD, and MAPE of each machine-learning model for prediction for 64 days from May 2018 to August 2018 without missing data. By using the original weather forecast data set, the prediction performance was different for each model. In addition, all models had better prediction performance on clear and slightly cloudy days than on cloudy and overcast days. The RNN had the lowest average hourly error at 0.51, 1.39, 23.79% of CV(RMSE), MAD, and MAPE, respectively, for all periods.
Figure 7 illustrates the daily PV power output (left) by summation of hourly actual measurements and model predictions. The means and deviations of daily total PV power output (right) varied depending on sky conditions and the model. The difference in the actual daily total PV output was 72.7 kW, 153.41 kW, 191.44 kW, 211.78 kW in clear, slightly cloudy, cloudy, and overcast sky conditions, respectively. The standard deviation ranged from 22.10 (clear) to 68.74 (overcast). All models performed better on clear and slightly cloudy days than on cloudy and overcast days, which was similar to the hourly prediction results. The large variance in PV power output on cloudy and overcast days affected the model performance.
As shown in Figure 7, the RNN outperformed the other models in estimating the daily total PV power output for the period. Although RNN predicted the hourly BIPV power output with less than 10% of MAPE and 0.20 of CV(RMSE) on clear and slightly cloudy days, the prediction performance decreased on cloudy and overcast days. In particular, the MAPE value of the daily total power output was higher than 50% on overcast days, which was similar to the values for the other models. This may imply that the prediction performance for overcast sky conditions does not depend on the model but, rather, the correlation between the input feature and the BIPV power generation.
Section 5 presents an investigation of the weather features that affect the hourly BIPV prediction performance on overcast days. Subsequently, this study proposes using the RNN with feature engineering, which can improve the prediction accuracy on overcast days.

5. Performance Improvement by Dropout Observation

This section presents techniques to improve the BIPV prediction performance on overcast days by using the feature engineering method based on the RNN model.

5.1. Effect of Dropout Observation

The variable importance indicated that the NSI is a redundant variable that contains high levels of noise, which prevents accurate predictions.
Section 4 demonstrated that the RNN is the most promising model but its performance on the 18 overcast days was not good as the clear or slightly cloudy days. In general, the NSI was 4 for the entire day on overcast days but the hourly BIPV output fluctuated with time, as shown in Figure 5. In addition, on several days, the onsite weather was clear, but the weather forecast was overcast or vice-versa.
Figure 8 shows the results of the hourly BIPV power prediction on overcast days combined with different levels of observation dropout using RNN. In Figure 8a, the onsite weather was overcast but the NSI was 2 (slightly cloudy) for the entire day. In the other (Figure 8b) case, the onsite weather was slightly cloudy, whereas the NSI was 4, i.e., overcast. The hourly average MAPE varied by different values of p. The prediction results of p = 0.0 (without the dropout observation) and that of p = 1.0 (without all the NSI data for the training and predicting process) illustrates that the dropout observation is necessary to improve the prediction performance. The hourly predicted value, at a probability of 0.3, was closer to the actual BIPV power output than for the other probability variations.
The average hourly CV(RMSE), MAD, and MAPE on a day when both cases occurred was 2.54, 6.35, and 160.53%, respectively, with exposure to the default condition, but with a p of 0.3, the hourly prediction result of the BIPV power provided a more accurate fitting at 1.33, 3.00, and 103.52%, respectively. In particular, the hourly performance during the peak period from 11 a.m. to 3 p.m. was significantly better in all cases.
The daily predictions of the BIPV power validate the daily effect of the dropout observation at a p of 0.3 on 18 overcast days. Figure 9 compares the average hourly MAPE on a day characterized by the best-performing p probability (= 0.3) and the original RNN at p = 0.0 over the same period. The effect of the dropout observation provides a more optimal fit for the 5 days during which the unstable weather conditions prevailed. The average hourly MAPE per day at p = 0.3 was 68.16%, compared with the MAPE for the default condition, which was 107.13% for 5 days. For 18 days, the daily prediction performance with the dropout observation produced a minimum MAPE of 39.67%, which was an improvement on the performance compared with the original configuration of 50.48% MAPE. Therefore, the dropout observation method improves the hourly prediction accuracy of the RNN model by more than 20% for overcast sky conditions, when the forecasted weather differs from the onsite weather.

5.2. Application and Integration of the Feature Engineered RNN Model

Direct models can predict BIPV power output with a simple environment that has historical PV power outputs and weather forecasted data. However, the prediction performance of the direct model depends on the accuracy of forecasted weather data. This study presents a machine-learning model, RNN with feature engineering, to include the benefits and reduce the drawbacks of the direct model approach. By feature engineering, the prediction performance of the model has been improved on cloudy and overcast days, in which the model accuracy decreases because of the high uncertainty of the sky condition as described in earlier sections.
The simple structure of the proposed model using an online weather forecasting service enables us to predict the hourly power outputs for the onsite BIPV without the substantial sensor network or data management system, which are not applicable for small or micro distributed energy resources (DER). This helps a building energy manager balance the supply and demand of electricity for the building and make a strategy for purchasing electricity from the grid.
In addition, the aggregators for DERs in the smart grid, virtual power plant, or energy cloud can simply estimate the hourly PV power output of several DERs in a day-ahead. This allows the smart grid operators make an optimal decision on whether the generated power should be traded into the electricity market or be stored, to maximize the efficiency of the BIPV application and the benefits of the electricity storage system.

6. Conclusions

The short-term BIPV power output prediction method is essential not only to manage the hourly electricity energy balance in a building but also to transfer the generated power from the BIPV into the grid. Although the indirect method can yield a higher prediction accuracy depending on the prediction accuracy of the solar radiation, it requires solar radiation data, DNI or GHI at a particular site. However, the direct method is able to predict the PV power output with conventional forecasted weather data, not including solar irradiation, but the accuracy depends on differences between the forecasted and actual on-site weather conditions.
This study introduced a novel approach to improve the prediction performance of the short-term BIPV power output with a direct method using feature engineering and machine learning. Feature selection (based on variable importance with SVM) was used with seven variables from the weather forecast data to identify the factor that prevents improvements in short-term BIPV power prediction. Among the forecasted weather variables, NSI has the lowest correlation scores to the PV power output in most sky conditions because of the variable characteristics from the weather forecast service. NSI proved to be redundant for the training and prediction process.
Six machine-learning algorithms, namely the ANN, SVM, CART, CHAID, RF, and RNN algorithms, were compared to identify an appropriate prediction model using historical weather data exposed to four typical types of weather conditions. The results demonstrated that the RNN provides the best predictive performance. The RNN outperformed the other methods not only in terms of hourly predictive accuracy but also with regard to long-term daily prediction performance for 64 days. Although the RNN exhibited high accuracy on clear and slightly cloudy days, the prediction performance on overcast days was higher than 50% of the average hourly MAPE.
Dropout observation with different probabilities of NSI, the lowest correlated variables, was then applied to the RNN model training and predictions to remove additional noise in the neural net for PV power output prediction on overcast days. The results showed that the prediction performance improved by more than 20% compared with the RNN for 18 overcast days without feature engineering. Therefore, the observation dropout of the feature engineering method appears to be capable of providing reliable prediction performance for both hourly intervals and daily BIPV power output for short-term periods.
Although we suggested a new approach to improve the hourly prediction performance of BIPV power output with the direct method, this study has room for improvement with respect to investigating the prediction performance with different model configurations, different variables for dropout observation, and input levels. It is necessary to optimize the short-term trade schedule for the predicted PV power output into the grid, one day ahead, in the microgrid. This can be used to minimize the purchased electricity from the grid by incorporation into the energy storage system (ESS).

Author Contributions

Conceptualization, D.-K.L. and Y.-T.C.; data curation, D.-K.L. and J.-H.J.; investigation, D.-K.L., J.-H.J. and Y.-T.C.; methodology, D.-K.L. and Y.-T.C.; supervision, Y.-T.C.; validation, D.-K.L. and J.-H.J.; visualization, D.-K.L. and J.-H.J.; writing—original draft, D.-K.L. and J.-H.J.; writing—review and editing, D.-K.L., S.-H.Y. and Y.-T.C.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, R.; Huang, T.; Gadh, R.; Li, N. Solar generation prediction using the ARMA model in a laboratory-level micro-grid. In Proceedings of the IEEE 3rd International Conference on Smart Grid Communications, Tainan, Taiwan, 5–8 November 2012; pp. 528–533. [Google Scholar]
  2. Mishra, A.K.; Ramesh, L. Application of neural networks in wind power (generation) prediction. In Proceedings of the International Conference on Sustainable Power Generation and Supply, Nanjing, China, 6–7 April 2009; pp. 1–5. [Google Scholar]
  3. Sharma, N.; Sharma, P.; Irwin, D.; Shenoy, P. Predicting solar generation from weather forecasts using machine learning. In Proceedings of the IEEE 2nd International Conference on Smart Grid Communications, Brussels, Belgium, 17–20 October 2011; pp. 528–533. [Google Scholar]
  4. Hernández, L.; Baladrón, C.; Aguiar, J.M.; Calavia, L.; Carro, B.; Sánchez-Esguevillas, A.; Sanjuán, J.; González, Á.; Lloret, M. Improved short-term load forecasting based on two-stage predictions with artificial neural networks in a microgrid environment. Energies 2013, 6, 4489–4507. [Google Scholar] [CrossRef]
  5. Rana, M.; Koprinska, I. Forecasting electricity load with advanced wavelet neural networks. Neurocomputing 2016, 182, 118–132. [Google Scholar] [CrossRef]
  6. Rodríguez, F.; Fleetwood, A.; Galarza, A.; Fontán, L. Predicting solar energy generation through artificial neural networks using weather forecasts for microgrid control. Renew. Energy 2018, 126, 855–864. [Google Scholar] [CrossRef]
  7. Lara-Fanego, V.; Ruiz-Arias, A.J.; Pozo-Vázquez, D.; Santos-Alamillos, F.J.; Tovar-Pescador, J. Evaluation of the WRF model solar irradiance forecasts in Andalusia (southern Spain). Sol. Energy 2018, 86, 2200–2217. [Google Scholar] [CrossRef]
  8. Chen, C.; Duan, S.; Cai, T.; Liu, B.; Hu, G. Smart energy management system for optimal microgrid economic operation. IET Renew. Power Gener. 2011, 5, 258–267. [Google Scholar] [CrossRef]
  9. Zhao, P.; Wang, J.; Xia, J.; Dai, Y.; Sheng, Y.; Yue, J. Performance evaluation and accuracy enhancement of a day-ahead wind power forecasting system in China. Renew. Energy 2012, 43, 234–241. [Google Scholar] [CrossRef]
  10. Tanaka, K.; Uchida, K.; Ogimi, K.; Goya, T.; Yona, A.; Senjyu, T.; Funabashi, T.; Kim, C.H. Optimal operation by controllable loads based on smart grid topology considering insolation forecasted error. IEEE Trans. on Smart Grid 2011, 2, 438–444. [Google Scholar] [CrossRef]
  11. Filipe, J.M.; Bessa, R.J.; Sumaili, J.; Tomé, R.; Sousa, J.N. A hybrid short-term solar power forecasting tool. In Proceedings of the International Conference on Intelligent System Application to Power Systems (ISAP), Porto, Portugal, 11–16 September 2015; Volume 18, pp. 11–16. [Google Scholar]
  12. Dong, Z.; Dazhi, Y.; Reindl, T.; Walsh, W.M. A novel hybrid approach based on self-organizing maps, support vector regression and particle swarm optimization to forecast solar irradiance. Energy 2015, 82, 570–577. [Google Scholar] [CrossRef]
  13. Dalton, G.J.; Lockington, D.A.; Baldock, T.E. Feasibility analysis of renewable energy supply options for a grid-connected large hotel. Renew. Energy 2009, 34, 955–964. [Google Scholar] [CrossRef]
  14. Fentis, A.; Bahatti, L.; Mestari, M.; Tabaa, M.; Jarrou, A.; Chouri, B. Short-term PV power forecasting using support vector regression and local monitoring data. In Proceedings of the 4th IEEE International Renewable and Sustainable Energy Conference, Marrakech, Morocco, 14–17 November 2016; pp. 2380–7393. [Google Scholar]
  15. Wan, C.; Zhao, J.; Song, Y.H.; Xu, Z.; Lin, J.; Hu, Z. Photovoltaic and solar power forecasting for smart grid energy management. CSEE J. Power and Energy Syst. 2015, 1, 38–46. [Google Scholar] [CrossRef]
  16. Muhammad, Q.R.; Mithulananthan, N.; Chandima, E. On recent advances in PV output power forecast. Sol. Energy 2016, 136, 125–144. [Google Scholar]
  17. Gandoman, F.H.; Raeisi, F.; Ahmadi, A. A literature review on estimating of PV-array hourly power under cloudy weather conditions. Renew. Sustain. Energy Rev. 2016, 36, 579–592. [Google Scholar] [CrossRef]
  18. Shi, J.; Lee, W.J.; Liu, Y.; Yang, Y.; Wang, P. Forecasting power output of photovoltaic system based on weather classification and support vector machine. In Proceedings of the IEEE Industry Applications Society Annual Meeting, Orlando, FL, USA, 9–13 October 2011; pp. 1–6. [Google Scholar]
  19. Mellit, A.; Pavan, A.M.; Benghanem, M. Least squares support vector machine for short-term prediction of meteorological time series. Theoret. Appl. Climatol. 2018, 111, 297–307. [Google Scholar] [CrossRef]
  20. Esfe, M.H.; Saedodin, S.; Sina, S.; Afrand, M.; Rostami, S. Designing an artificial neural networks to predict thermal conductivity and dynamic viscosity of ferromagnetic nanofluid. Intern. Commun. Heat Mass Transf. 2015, 68, 50–57. [Google Scholar] [CrossRef]
  21. Erkaymaz, O.; Ozer, M.; Perc, M. Performance of small-world neural networks for the diagnosis of diabetes. Appl. Math. Comput. 2017, 311, 22–28. [Google Scholar] [CrossRef]
  22. Pandarakone, S.E.; Mizuno, Y.; Nakamura, H. A comparative study machine learning algorithm and artificial intelligence neural network in detecting minor bearing fault of induction motors. Chin. J. Mech. Eng. 2017, 30, 1347–1356. [Google Scholar]
  23. Olatomiwa, L.; Mekhilef, S.; Shamshirband, S.; Mohammadi, K.; Petkovic, D.; Sudheer, C. A support vector machine-firefly algorithm-based model for global solar radiation prediction. Sci. Direct Sol. Energy 2015, 115, 632–644. [Google Scholar] [CrossRef]
  24. Olfaz, M.; Tirink, C.; Önder, H. Use of cart and algorithms in karayaka sheep breeding. Kafkas Univ Vet Fak Derg 2019, 25, 105–110. [Google Scholar]
  25. Alkhasawneh, M.S.; Ngah, U.K.; Tay, L.T.; Isa, N.A.M.; Al-Batah, M.S. Modeling and testing landside hazard using decision tree. J. Appl. Math. 2014, 2014, 1–9. [Google Scholar] [CrossRef]
  26. Qiu, X.; Zhang, L.; Suganthan, P.N.; Amaratunga, G.A.J. Oblique random forest via least square estimation for time series forecasting. Infor. Sci. 2017, 420, 249–262. [Google Scholar] [CrossRef]
  27. Huang, N.; Lu, G.; Xu, D. A permutation importance-based feature selection method for short-term electricity load forecasting using random forest. Energies 2016, 9, 767. [Google Scholar] [CrossRef]
  28. Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. Available online: https://arxiv.org/abs/1412.3555 (accessed on 7 August 2019).
  29. Graves, A.; Mohamed, A.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  30. Schratz, P.; Muenchow, J.; Iturritxa, E.; Richter, J.; Brenning, A. Performance evaluation and hyperparameter tuning of statistical and machine-learning models using spatial data. arXiv 2018, 1830–11266. Available online: https://arxiv.org/abs/1803.11266 (accessed on 7 August 2019).
  31. Gregorutti, B.; Michel, B.; Saint-Pierre, P. Correlation and variable importance in random forests. Stat. Comput. 2017, 27, 659–678. [Google Scholar] [CrossRef]
  32. Date, Y.; Kikuchi, J. Application of a deep neural network to metabolomics studies and its performance in determining important variables. Anal. Chem. 2017, 90, 1805–1810. [Google Scholar] [CrossRef] [PubMed]
  33. Lin, S.W.; Lee, Z.J.; Chen, S.C.; Tseng, T.Y. Parameter determination of support vector machine and feature selection using simulated annealing approach. Appl. Soft Comput. 2008, 8, 1505–1512. [Google Scholar] [CrossRef]
  34. Chapelle, O.; Vapnik, V.; Bousquet, O.; Mukherjee, S. Choosing multiple parameters for support vector machines. Mach. Learn. 2002, 46, 131–159. [Google Scholar] [CrossRef]
  35. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskelever, I.; Ruslan, S. Dropout: A simple way to prevent neural networks from overfitting. JMLP 2014, 15, 1929–1958. [Google Scholar]
  36. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv 2014, arXiv:1312.6034. [Google Scholar]
  37. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. arXiv 2013, arXiv:1311.2901. Available online: https://arxiv.org/abs/1312.6034 (accessed on 7 August 2019).
  38. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. arXiv 2016, arXiv:1602.04938. Available online: https://arxiv.org/abs/1602.04938 (accessed on 7 August 2019).
  39. Gal, Y.; Hron, M.; Kendall, A. Concrete dropout. arXiv 2017, arXiv:1705.07832. Available online: https://arxiv.org/abs/1705.07832 (accessed on 7 August 2019).
  40. Kingma, D.P.; Salimans, T.; Welling, M. Variational dropout and the local reparameterization trick. arXiv 2015, arXiv:1506.02557v2. Available online: https://arxiv.org/abs/1506.02557 (accessed on 7 August 2019).
  41. Maeda, S.I. A Bayesian encourages dropout. arXiv 2014, arXiv:1412.7003. Available online: https://arxiv.org/abs/1412.7003 (accessed on 7 August 2019).
  42. Molchanv, D.; Ashukha, A.; Vetrov, D. Variational dropout sparsifies deep neural networks. arXiv 2017, arXiv:1701.05369. Available online: https://arxiv.org/abs/1701.05369 (accessed on 7 August 2019).
  43. Neelakantan, A.; Vilnis, L.; Le, Q.V.; Sutskever, I.; Kaiser, L.; Kurach, K.; Martens, J. Adding gradient noise improves learning for very deep networks. arXiv 2015, arXiv:1511.06807. Available online: https://arxiv.org/abs/1511.06807 (accessed on 7 August 2019).
  44. Gulcehre, C.; Moczulski, M.; Denil, M.; Bengio, Y. Noisy Activation Functions. International Conference on Machine Learning. 2016. 48. Available online: https://arxiv.org/abs/1603.00391 (accessed on 7 August 2019).
Figure 1. The overall framework for the short-term BIPV power output prediction.
Figure 1. The overall framework for the short-term BIPV power output prediction.
Energies 12 03247 g001
Figure 2. The BIPV generation system installed on a medium-sized office building.
Figure 2. The BIPV generation system installed on a medium-sized office building.
Energies 12 03247 g002
Figure 3. A comparison of the network structures among the standard network, dropout network, and observation dropout.
Figure 3. A comparison of the network structures among the standard network, dropout network, and observation dropout.
Energies 12 03247 g003
Figure 4. (a) The unit training time step that is present with a probability of p; (b) the use of output with a probability of p during the prediction time step.
Figure 4. (a) The unit training time step that is present with a probability of p; (b) the use of output with a probability of p during the prediction time step.
Energies 12 03247 g004
Figure 5. Characteristics of seven weather features and BIPV power output for 7 days.
Figure 5. Characteristics of seven weather features and BIPV power output for 7 days.
Energies 12 03247 g005
Figure 6. Hourly prediction results of the BIPV power output for typical sky conditions using the six machine-learning algorithms.
Figure 6. Hourly prediction results of the BIPV power output for typical sky conditions using the six machine-learning algorithms.
Energies 12 03247 g006
Figure 7. Prediction results of the daily total BIPV power output with typical sky conditions based on six machine-learning models.
Figure 7. Prediction results of the daily total BIPV power output with typical sky conditions based on six machine-learning models.
Energies 12 03247 g007aEnergies 12 03247 g007b
Figure 8. A comparison of the prediction results of the BIPV power output for specific days using the observation dropout technique.
Figure 8. A comparison of the prediction results of the BIPV power output for specific days using the observation dropout technique.
Energies 12 03247 g008
Figure 9. The daily MAPE for the BIPV power prediction with dropout observation for overcast days.
Figure 9. The daily MAPE for the BIPV power prediction with dropout observation for overcast days.
Energies 12 03247 g009
Table 1. Specifications of the BIPV generation system.
Table 1. Specifications of the BIPV generation system.
ElementParameters
Module typeSM-250PG8 (Grid-connected)
Connection type10 series × 20 parallel
Nominal module efficiency15.03%
Table 2. The specified conditions for each of the six machine-learning models.
Table 2. The specified conditions for each of the six machine-learning models.
ModelConfiguration ParametersRange/TypeSelected Parameter
Artificial neural network (ANN)Activation functionSigmoid, Hyperbolic tangent, Rectified linear unit Hyperbolic tangent
Number of hidden neurons7 ~ 1510
Data division (training/testing)80%/20%,
70%/30%
70%/30%
Support vector machine (SVM)Kernel typePolynomial,
RBF, Sigmoid
RBF
RBF Gamma−0.5 ~ +0.5+0.1
Gamma−3.0 ~ +3.0+1.0
Regression precision(epsilon)−0.5 ~ +0.50.1
Classification and decision tree
(CART)
Tree depth1 ~ 115
Minimum change in impurity1 ~ 0.00010.0001
Impurity measureEntropy, GiniGini
Overfit prevention set20% ~ 30%30%
Chi-square automatic interaction detection (CHAID)Tree depth1 ~ 114
Minimum change in expected cell frequencies1 ~ 0.0010.001
Maximum iterations for convergence50 ~ 150100
Random forest
(RF)
Number of nodes10 ~ 10,00010,000
Tree depth1 ~ 1110
Child node size55
Recurrent neural
network(RNN)
Activation functionSigmoid, Hyperbolic tangent, Rectified linear unitHyperbolic tangent
Number of hidden neurons7 ~ 1510
Data division (training/testing)80%/20%,
70%/30%
70%/30%
Table 3. The variable importance score of weather variables based on SVM.
Table 3. The variable importance score of weather variables based on SVM.
Daily Sky Conditions
Clear Days
(15 Days)
Slightly Cloudy Days
(14 Days)
Cloudy Days
(17 Days)
Overcast Days
(18 Days)
OT0.070.040.030.03
RH0.200.220.160.18
WS0.010.040.010.02
WD0.020.010.000.01
NSI0.010.010.000.00
PP0.000.000.040.01
SA0.690.650.790.75
Summation of the importance score with measured values is 1.
Table 4. Comparison of model performance in terms of BIPV power output (64 days) for four types of sky conditions.
Table 4. Comparison of model performance in terms of BIPV power output (64 days) for four types of sky conditions.
DaysIndexPrediction Models
ANNSVMCARTCHAIDRFRNN
Clear days
(15 days)
CV(RMSE)0.180.390.300.300.220.16
MAD1.182.691.982.021.540.80
MAPE(%)8.9028.6617.4715.5310.716.89
Slightly cloudy days
(14 days)
CV(RMSE)0.290.490.360.350.270.24
MAD1.853.062.022.161.651.44
MAPE(%)13.5524.6218.5319.2313.8712.02
Cloudy days
(17 days)
CV(RMSE)0.770.720.900.910.860.41
MAD2.752.733.173.283.021.47
MAPE(%)32.4535.1243.8842.8734.6019.21
Overcast days
(18 days)
CV(RMSE)1.601.571.322.121.461.10
MAD2.462.482.503.102.401.77
MAPE(%)71.2863.4172.1199.4866.8551.34
Mean
(64 days)
CV(RMSE)0.760.830.760.980.750.51
MAD2.102.722.452.692.201.39
MAPE(%)33.7239.2740.0847.2133.5423.79

Share and Cite

MDPI and ACS Style

Lee, D.; Jeong, J.; Yoon, S.H.; Chae, Y.T. Improvement of Short-Term BIPV Power Predictions Using Feature Engineering and a Recurrent Neural Network. Energies 2019, 12, 3247. https://doi.org/10.3390/en12173247

AMA Style

Lee D, Jeong J, Yoon SH, Chae YT. Improvement of Short-Term BIPV Power Predictions Using Feature Engineering and a Recurrent Neural Network. Energies. 2019; 12(17):3247. https://doi.org/10.3390/en12173247

Chicago/Turabian Style

Lee, Dongkyu, Jinhwa Jeong, Sung Hoon Yoon, and Young Tae Chae. 2019. "Improvement of Short-Term BIPV Power Predictions Using Feature Engineering and a Recurrent Neural Network" Energies 12, no. 17: 3247. https://doi.org/10.3390/en12173247

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop