Next Article in Journal
Triple Collocation-Based Assessment of Satellite Soil Moisture Products with In Situ Measurements in China: Understanding the Error Sources
Next Article in Special Issue
A Computational Workflow for Generating A Voxel-Based Design Approach Based on Subtractive Shading Envelopes and Attribute Information of Point Cloud Data
Previous Article in Journal
Deep Learning with Open Data for Desert Road Mapping
Previous Article in Special Issue
Characterizing Geological Heterogeneities for Geothermal Purposes through Combined Geophysical Prospecting Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multistep-Ahead Solar Radiation Forecasting Scheme Based on the Light Gradient Boosting Machine: A Case Study of Jeju Island

School of Electrical Engineering, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul 02841, Korea
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(14), 2271; https://doi.org/10.3390/rs12142271
Submission received: 16 June 2020 / Revised: 9 July 2020 / Accepted: 13 July 2020 / Published: 15 July 2020
(This article belongs to the Special Issue Assessment of Renewable Energy Resources with Remote Sensing)

Abstract

:
Smart islands have focused on renewable energy sources, such as solar and wind, to achieve energy self-sufficiency. Because solar photovoltaic (PV) power has the advantage of less noise and easier installation than wind power, it is more flexible in selecting a location for installation. A PV power system can be operated more efficiently by predicting the amount of global solar radiation for solar power generation. Thus far, most studies have addressed day-ahead probabilistic forecasting to predict global solar radiation. However, day-ahead probabilistic forecasting has limitations in responding quickly to sudden changes in the external environment. Although multistep-ahead (MSA) forecasting can be used for this purpose, traditional machine learning models are unsuitable because of the substantial training time. In this paper, we propose an accurate MSA global solar radiation forecasting model based on the light gradient boosting machine (LightGBM), which can handle the training-time problem and provide higher prediction performance compared to other boosting methods. To demonstrate the validity of the proposed model, we conducted a global solar radiation prediction for two regions on Jeju Island, the largest island in South Korea. The experiment results demonstrated that the proposed model can achieve better predictive performance than the tree-based ensemble and deep learning methods.

Graphical Abstract

1. Introduction

Due to the serious problems caused by the use of fossil fuels, much attention has been focused on renewable energy sources (RESs) and smart grid technology to reduce greenhouse gas emissions [1,2]. Smart grid technology incorporates information and communication technology into the existing power grid using diverse smart sensors [3]. Smart grid technology can optimize the energy supply and demand by exchanging power production and consumption information between consumers and suppliers [4]. In particular, many countries, including smart islands, are replacing fossil fuels with RESs for energy self-sufficiency and carbon-free energy generation [5,6,7]. Two representative RESs are wind and global solar radiation. Although wind power has a smaller installation area and better power production than solar power, it suffers from higher maintenance costs and more noise. For example, due to various support policies of the Korean government related to renewable energies and smart grid technologies [8], the demand for photovoltaics (PV) is rapidly increasing in South Korea [9]. PV are best known as a method of generating electric power using solar cells to convert energy from the sun into a flow of electrons using the PV effect. Moreover, PV power system is based on an ecofriendly and infinite resource, and is cheaper to build than other power generation systems [10].
Various meteorological factors influence the PV system, and global solar radiation is the most crucial factor in the PV system [11,12]. Therefore, accurate global solar radiation forecasting is essential for the optimal operation of PV systems [13]. Recently, artificial neural network (ANN)-based global solar radiation forecasting models, such as the shallow neural network (SNN), deep neural network (DNN), and long short-term memory (LSTM) network, have been constructed to handle the nonlinearity and fluctuation of global solar radiation [14,15,16,17,18,19,20]. In addition, many studies have been conducted to predict global solar radiation accurately based on an ensemble learning technique that combines several weak models. For instance, in [21], the authors constructed two global solar radiation forecasting models based on the ANN and random forest (RF) methods. Then, they demonstrated that RF, which is an ensemble learning technique, exhibited better prediction performance than the ANN. In [22], the authors proposed four global solar radiation forecasting models based on the bagging and boosting techniques and analyzed the excellence and feature importance of the ensemble learning techniques.
Because global solar radiation is affected by diverse factors, such as season, time, and weather variables, predicting global solar radiation is challenging in the time domain [13]. The ensemble learning technique can avoid the overfitting problem and perform a more accurate prediction than the single model [23]. In this paper, we propose a novel forecasting model for multistep-ahead (MSA) global solar radiation predictions based on the light gradient boosting machine (LightGBM), which is a tree-based ensemble learning technique. The LightGBM can perform learning and prediction very quickly, which reduces the time needed for MSA prediction and performs more accurate predictions. Our forecasting model uses the meteorological information provided by the Korea Meteorological Administration (KMA) for global solar radiation prediction. In addition, to handle the uncertainty of PV scheduling, our MSA forecasting scheme makes hourly solar forecasts from 8 a.m. to 6 p.m. for 24 h from the current time. Usually, the farther the prediction point is from the learning point, the higher the probability that various changes will occur during the trend and pattern of the meteorological conditions and global solar radiation. To address this issue, we used time-series cross-validation (TSCV). We conducted rigorous experiments to compare the performance of LightGBM, various tree-based ensembles, and deep learning methods. Finally, we used the feature importance of the proposed model to provide interpretable forecasting results.
The contributions of this paper are as follows:
  • We proposed an MSA forecasting scheme for the efficient PV system operation.
  • We proposed an interpretable forecasting model based on feature importance analysis.
  • We increased the accuracy of global solar radiation forecasting using TSCV.
This paper is organized as follows. In Section 2, we describe the overall process for constructing a LightGBM-based forecasting model for MSA global solar radiation forecasting. In Section 3, we analyze the experimental results and describe the interpretable forecasting results of our proposed model. Lastly, we discuss in Section 4 the conclusions and some future research directions.

2. Materials and Methods

2.1. Data Collection and Preprocessing

In this paper, we used the date/time, meteorological data, and historical global solar radiation data provided by the KMA as input variables to construct a global solar radiation forecasting model. We considered two regions located on Jeju Island. Jeju is the largest island in South Korea and is implementing various measures to change into a smart island. For instance, it is enforcing diverse energy policies that encourage a shift from conventional fossil fuels to RESs. The two regions that we selected for validating prediction performance are Ildo-1 dong (latitude: 33.51411 and longitude: 126.52969) and Gosan-ri (latitude: 33.29382 and longitude: 126.16283). The data collection period is from 8 a.m. to 6 p.m. for a total of eight years from 2011 to 2018, and the collected data include temperature, humidity, wind speed, and global solar radiation. The meteorological observation data provided by the KMA include extra data, such as soil temperature, total cloud volume, ground-surface temperature, and sunshine amount. However, because the sky condition (also known as weather observation), temperature, humidity, and wind speed are provided by KMA’s short-term weather forecasts, as shown in Figure 1, we only considered these factors [24].
In the meteorological data we collected, about 0.1% of the total data for each category were missing, and the missing values were indicated as −1. Because the temperature, humidity, wind speed, and global solar radiation have continuous data characteristics, missing values can be estimated using linear interpolation. The sky condition data were presented as categorical values from 1 to 4, and missing values were approximated using logistic regression for similarity with the adjacent data.
For the date, to reflect the periodicity, one-dimensional data were augmented with continuous data in two-dimensional space using Equations (1) and (2) [25]. In the equations, end-of-month (EoM) indicates the last day of the month. The equations converted each Julian date into a value from 1 to 365. For instance, the Julian date of January 1 is converted to 1, and December 31 is converted to 365. In the case of leap years, 366 was used instead of 365 in the equations. Figure 2 illustrates an example of preprocessing the date data.
D a t e X = sin ( 360 ° × ( 1 M o n t h 1 E o M + D a y ) / 365 )
D a t e Y = cos ( 360 ° × ( 1 M o n t h 1 E o M + D a y ) / 365 )
The cloud amount is provided by the KMA. Of the two popular methods for representing cloud amount, which are meteorology 1/8 and climatology 1/10, the KMA uses the second method. Hence, the cloud amount is represented by eleven scales (i.e., from 0 for a clear sky to 10 for an overcast sky). The sky condition data have four interval scales [26,27]: 1 for clear (0 ≤ cloud amount ≤ 2), 2 for partly cloudy (3 ≤ cloud amount ≤ 5), 3 for mostly cloudy (6 ≤ cloud amount ≤ 8), and 4 for cloudy (9 ≤ cloud amount ≤ 10). Because we represent the sky condition data using one-hot encoding, a value of 1 is placed in the binary variable for a specific sky condition, and 0 is used for the other sky conditions. Time data were also represented by interval scales. Global solar radiation is highest during the day from 12 to 2 p.m. To assess these variables more effectively, we used one-hot encoding to represent time intervals.
In addition, to reflect the recent trends in global solar radiation, we used the sky condition, temperature, humidity, wind speed, and global solar radiation of the day before the forecast point as input variables. We considered 30 input variables to construct our prediction model, as shown in Table 1. As our goal is to perform MSA (all time points for the next 24 h) forecasting, we needed all the input variables for 11 prediction time points. Therefore, we used 330 input variables (i.e., 30 input variables × 11 prediction time points) with 32,143 tuples for the MSA forecasting model construction, as shown in Figure 3.

2.2. Forecasting Model Construction

The purpose of our model is to predict global solar radiation for the next 11 time points from the current time. To construct a global solar radiation forecasting model, we used LightGBM, a gradient boosting machine (GBM)-based model. The LightGBM model [28] is based on a gradient boosting decision tree (GBDT) applying gradient-based one-side sampling and exclusive feature bundling technologies. Unlike the conventional GBM tree splitting method, a leafwise method is used to create complex models to achieve higher accuracy; hence, it is useful for time-series forecasting. Because of the GBDT and leafwise method, LightGBM has the advantages of reduced memory usage and faster training speed. The LightGBM contains various hyperparameters to be tuned. Among them, the learning rate, number of iterations, and number of leaves are closely related to the prediction accuracy. In addition, overfitting can be prevented by adjusting the colsample by tree and subsample hyperparameters. Moreover, LightGBM also can use different algorithms for its learning iterations. In this paper, we constructed two LightGBM models using two boosting types: GBDT and dropouts meet multiple additive regression trees (DART) [29] for comparison. Both models perform predictions on multiple outputs using the MultiOutputRegressor module in scikit-learn (v. 0.22.1).
In general, to evaluate a forecasting method, we first divide a dataset into training and test sets. Then, we construct the forecasting model using the training set. Finally, we evaluate the performance of the forecasting model using the test set. A greater time interval between training and forecasting lowers the prediction performance [30]. To solve this problem, we applied TSCV, which is popularly used when data exhibit time-series characteristics and are focused on a single forecast of the dataset [6]. The TSCV uses all data before the prediction point as a training set and predicts the next forecasting point by setting it as a test set, iteratively.
However, if TSCV is performed at every point, it requires a considerable amount of time to train and forecast. To reduce this overhead, we conducted monthly TSCV, as shown in Figure 4. In addition, for interpretable global solar radiation forecasting, we analyzed the variable importance changes for the 30 input variables by obtaining the feature importance using LightGBM.

2.3. Baseline Models

To demonstrate the performance of our model, we constructed various forecasting models based on the tree-based ensemble and deep learning methods.
In the case of tree-based ensemble learning methods, because they combine several weak models effectively, they usually exhibit better prediction performance than a single model. In the experiment, we considered RF, GBM, and extreme gradient boosting (XGBoost) ensemble methods to construct MSA global solar radiation forecasting models. The RF method trains each tree independently by using a randomly selected sample of the data. As the RF method tends to reduce the correlation between trees, it provides a robust model for out-of-sample forecasting [31]. In addition, the GBM is a forward-learning ensemble method that obtains predictive results using gradually improved estimations [32]. The adjusted model is built by applying the residuals of the previous model, and this procedure is repeated N times to build a robust model. We constructed two GBM models by considering the quantile regression and Huber loss functions, respectively. The XGBoost method is an algorithm that can prevent overfitting by reducing the tree correlation using the shrinkage method [33]. Moreover, it can perform parallel processing by applying the column subsampling method. The XGBoost method constructs a weak model and evaluates the consistency using the training set. After that, the method constructs an adjusted prediction model with the explanatory variable for the gradient in the direction in which consistency increases using the gradient descent method. This procedure is repeated N times to build a robust model [34]. We constructed two XGBoost models by applying two boosting types (i.e., GBDT and DART). To predict multiple outputs, we constructed an RF model using the MultivariateRandomForest [35] package in R (v. 3.5.1) and the GBM and XGBoost models using the MultiOutputRegressor module in scikit-learn (v. 0.22.1) [36].
For deep learning-based MSA global solar radiation forecasting models, we considered the SNN, DNN, LSTM network, and attention-based LSTM (ATT-LSTM) network. These models require a sufficient amount of training data for accurate predictive performance, and the models can overfit the data if the training data are insufficient [37]. A typical ANN consists of an input layer, one or more hidden layers, and an output layer, and each layer consists of one or more hidden nodes [38,39]. The ANNs have various hyperparameters that affect prediction performance [38]. These hyperparameters include the number of hidden layers, number of hidden nodes, an activation function, and so on. In addition, the SNN has one hidden layer, and the DNN has two or more hidden layers [39]. The LSTM network [40] is a model that can solve the long-term dependency problem of the existing recurrent neural network. The LSTM network is useful for training sequence data in the time-series forecasting method. Nevertheless, although the length of the input variable is long, the forecasting accuracy of the sequence-to-sequence model suffers due to focusing on all input variables. To solve this problem, an attention mechanism [41] has been developed in the field of machine translation. The attention mechanism comprises an encoder that builds a vector from the input variable and a decoder that outputs a dependent variable using the vector output by the encoder as input. The decoder part performs the model training focused on data representing high similarity by indicating the similarity with the encoder as a value; hence, it can exhibit accurate forecasting performance. Applying the attention mechanism to the LSTM described above focuses the model on specific vectors so that it obtains more accurate forecasting results [42].
In our previous work [7], we constructed several deep learning models for MSA global solar radiation forecasting in the same experimental environment. We used the dropout method to control the weight of the hidden layers to prevent overfitting. To do this, we found optimal hyperparameter values for each deep learning model, as indicated in Table 2.

3. Results and Discussion

In the experiments, we used two global solar radiation datasets collected from two regions from 2011 to 2018. The two regions are Ildo-1 and Gosan-ri on Jeju island. We divided each dataset into two parts at a ratio of 75:25: a training set (in-sample) spanning 2011 to 2016, and a test set (out-of-sample) spanning 2017 to 2018. Table 3 lists various statistical analysis for the datasets by considering the training and test sets. The statistical analysis was performed by using Excel’s Descriptive Statistics data analysis tool. Figure 5 represents the boxplots of the global solar radiation data for each region.
For continuous data, such as humidity, wind speed, temperature, and historical global solar radiation, we performed standardization using Equation (3). In the equation, xi and x denote the input variable and original data, respectively. In addition, μ and σ denote the average of the original data and the standard deviation, respectively.
x i = x μ σ
To evaluate the prediction performance of the models, we used four metrics: mean biased error (MBE), mean absolute error (MAE), root mean square error (RMSE), and normalized root mean square error (NRMSE), as shown in Equations (4)–(7). Here, At and Ft represent the actual and forecasted values, respectively, at time t, n indicates the number of observations, and A ¯ represents the average of the actual values.
M B E = 1 n t = 1 n ( A t F t )
M A E = 1 n t = 1 n | F t A t |
R M S E = t = 1 n ( F t A t ) 2 n
N R M S E = t = 1 n ( F t A t ) 2 n A ¯ × 100
We implemented an RF-based forecasting model using R (v. 3.5.1) and all other forecasting models using Python (v. 3.6). We found optimal values for the hyperparameters of the tree-based ensemble learning models via GridSearchCV in scikit-learn (v. 0.22.1), as displayed in Table 4. Because the two regions are close together, we obtained the same hyperparameter values for the two regions.
Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 and Table 13 and Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13 demonstrate that our model could achieve lower RMSE and MAE values than all other forecasting models that we considered, except the XGBoost model. In addition, tree-based ensemble models exhibited better performance than deep learning-based models. Moreover, the TSCV scheme demonstrated better prediction performance than the holdout scheme, as presented in Table 13. The XGBoost and LightGBM methods exhibited a similar prediction performance. However, regarding the aspect of the training and testing time, LightGBM took 220 s, whereas XGBoost took 3798 s. That is, LightGBM is 17 times faster than XGBoost. Hence, LightGBM has a clear advantage in terms of accuracy and time. In the forecasting results of LightGBM, we observed that the MAE and RMSE values were lowest at the first time point, and as the distance increased, these values increased.
Feature importance is a measure of variable importance when data have obtained a subset of all features. Feature importance can be determined from logistic regression or tree-based models. We determined the feature importance of our model, LightGBM-DART (TSCV), at each test point (one month) according to the TSCV cycle. Figure 14 and Figure 15 present a heat map graph that reveals the feature importance of the input variables mentioned in Table 2 for both regions. The variable importance values are exhibited in the range of 0 to 1 using minimum–maximum normalization to help readers understand. From the table, we confirmed that the day number of the year ( D a t e X and D a t e y ) consistently exhibited high feature importance, and the temperature, humidity, and wind speed, among the meteorological information, presented high feature importance. In particular, the importance of humidity increased over time.

4. Conclusions

In this paper, we proposed an MSA global solar radiation forecasting method based on LightGBM. To do this, we first configured 330 input variables considering the time and weather information provided by KMA to forecast the global solar radiation at multiple time points over the next 24 h. Then, we constructed a LightGBM-based forecasting model with DART boosting. To evaluate the performance of our model, we implemented diverse ensemble-based models and deep learning-based models and compared their performance using global solar radiation data from Jeju Island. From the comparison, we confirmed that our model exhibited better forecasting performance than other methods. We plan to conduct a forecasting model using only historical global solar radiation data in the future to provide accurate global solar radiation forecasting in regions where meteorological information is not provided. We will also conduct smart grid scheduling based on photovoltaic forecasting.

Author Contributions

Conceptualization, J.P. and J.M.; methodology, J.P.; software, J.P. and S.J.; validation, J.P. and S.J.; formal analysis, J.M.; investigation, S.J.; data curation, J.P. and S.J.; writing—original draft preparation, J.P. and J.M.; writing—review and editing, E.H.; visualization, J.M.; supervision, E.H.; project administration, E.H.; funding acquisition, E.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Korea Electric Power Corporation (grant number: R18XA05) and in part by Energy Cloud R&D Program (grant number: 2019M3F2A1073179) through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moon, J.; Kim, K.-H.; Kim, Y.; Hwang, E. A Short-Term Electric Load Forecasting Scheme Using 2-Stage Predictive Analytics. In Proceedings of the IEEE International Conference on Big Data and Smart Computing (BigComp), Shanghai, China, 15–17 January 2018; pp. 219–226. [Google Scholar] [CrossRef]
  2. Abdel-Nasser, M.; Mahmoud, K. Accurate photovoltaic power forecasting models using deep LSTM-RNN. Neural Comput. Appl. 2019, 31, 2727–2740. [Google Scholar] [CrossRef]
  3. Son, M.; Moon, J.; Jung, S.; Hwang, E. A Short-Term Load Forecasting Scheme Based on Auto-Encoder and Random Forest. In Proceedings of the International Conference on Applied Physics, System Science and Computers, Dubrovnik, Croatia, 26–28 September 2018; pp. 138–144. [Google Scholar] [CrossRef]
  4. Moon, J.; Park, J.; Hwang, E.; Jun, S. Forecasting power consumption for higher educational institutions based on machine learning. J. Supercomput. 2018, 74, 3778–3800. [Google Scholar] [CrossRef]
  5. Wang, H.Z.; Li, G.Q.; Wang, G.B.; Peng, J.C.; Jiang, H.; Liu, Y.T. Deep learning based ensemble approach for probabilistic wind power forecasting. Appl. Energy 2017, 188, 56–70. [Google Scholar] [CrossRef]
  6. Moon, J.; Kim, Y.; Son, M.; Hwang, E. Hybrid Short-Term Load Forecasting Scheme Using Random Forest and Multilayer Perceptron. Energies 2018, 11, 3283. [Google Scholar] [CrossRef] [Green Version]
  7. Jung, S.; Moon, J.; Park, S.; Hwang, E. A Probabilistic Short-Term Solar Radiation Prediction Scheme Based on Attention Mechanism for Smart Island. KIISE Trans. Comput. Pract. 2019, 25, 602–609. [Google Scholar] [CrossRef]
  8. Park, S.; Moon, J.; Jung, S.; Rho, S.; Baik, S.W.; Hwang, E. A Two-Stage Industrial Load Forecasting Scheme for Day-Ahead Combined Cooling, Heating and Power Scheduling. Energies 2020, 13, 443. [Google Scholar] [CrossRef] [Green Version]
  9. Lee, M.; Lee, W.; Jung, J. 24-Hour photovoltaic generation forecasting using combined very-short-term and short-term multivariate time series model. In Proceedings of the 2017 IEEE Power & Energy Society General Meeting, Chicago, IL, USA, 16–20 July 2017; pp. 1–5. [Google Scholar] [CrossRef]
  10. Khosravi, A.; Koury, R.N.N.; Machado, L.; Pabon, J.J.G. Prediction of hourly solar radiation in Abu Musa Island using machine learning algorithms. J. Clean. Prod. 2018, 176, 63–75. [Google Scholar] [CrossRef]
  11. Meenal, R.; Selvakumar, A.I. Assessment of SVM, empirical and ANN based solar radiation prediction models with most influencing input parameters. Renew. Energy 2018, 121, 324–343. [Google Scholar] [CrossRef]
  12. Aggarwal, S.; Saini, L. Solar energy prediction using linear and non-linear regularization models: A study on AMS (American Meteorological Society) 2013–14 Solar Energy Prediction Contest. Energy 2014, 78, 247–256. [Google Scholar] [CrossRef]
  13. Yadav, A.K.; Chandel, S. Solar radiation prediction using Artificial Neural Network techniques: A review. Renew. Sustain. Energy Rev. 2014, 33, 772–781. [Google Scholar] [CrossRef]
  14. Rao, K.D.V.S.K.; Premalatha, M.; Naveen, C. Analysis of different combinations of meteorological parameters in predicting the horizontal global solar radiation with ANN approach: A case study. Renew. Sustain. Energy Rev. 2018, 91, 248–258. [Google Scholar] [CrossRef]
  15. Cornaro, C.; Pierro, M.; Bucci, F. Master optimization process based on neural networks ensemble for 24-h solar irradiance forecast. Sol. Energy 2015, 111, 297–312. [Google Scholar] [CrossRef]
  16. Dahmani, K.; Dizene, R.; Notton, G.; Paoli, C.; Voyant, C.; Nivet, M.L. Estimation of 5-min time-step data of tilted solar global irradiation using ANN (Artificial Neural Network) model. Energy 2014, 70, 374–381. [Google Scholar] [CrossRef]
  17. Leva, S.; Dolara, A.; Grimaccia, F.; Mussetta, M.; Sahin, E. Analysis and validation of 24 hours ahead neural network forecasting for photovoltaic output power. Math. Comput. Simul. 2017, 131, 88–100. [Google Scholar] [CrossRef] [Green Version]
  18. Amrouche, B.; le Pivert, X. Artificial neural network based daily local forecasting for global solar radiation. Appl. Energy 2014, 130, 333–341. [Google Scholar] [CrossRef]
  19. Kaba, K.; Sarıgül, M.; Avcı, M.; Kandırmaz, H.M. Estimation of daily global solar radiation using deep learning model. Energy 2018, 162, 126–135. [Google Scholar] [CrossRef]
  20. Qing, X.; Niu, Y. Hourly day-ahead solar irradiance prediction using weather forecasts by LSTM. Energy 2018, 148, 461–468. [Google Scholar] [CrossRef]
  21. Benali, L.; Notton, G.; Fouilloy, A.; Voyant, C.; Dizene, R. Solar radiation forecasting using artificial neural network and random forest methods: Application to normal beam, horizontal diffuse and global components. Renew. Energy 2019, 132, 871–884. [Google Scholar] [CrossRef]
  22. Lee, J.; Wang, W.; Harrou, F.; Sun, Y. Reliable solar irradiance prediction using ensemble learning-based models: A comparative study. Energy Conv. Manag. 2020, 208, 112582. [Google Scholar] [CrossRef] [Green Version]
  23. Rew, J.; Cho, Y.; Moon, J.; Hwang, E. Habitat Suitability Estimation Using a Two-Stage Ensemble Approach. Remote Sens. 2020, 12, 1475. [Google Scholar] [CrossRef]
  24. Kim, J.; Moon, J.; Hwang, E.; Kang, P. Recurrent inception convolution neural network for multi short-term load forecasting. Energy Build. 2019, 194, 328–341. [Google Scholar] [CrossRef]
  25. Jung, S.; Moon, J.; Park, S.; Rho, S.; Baik, S.W.; Hwang, E. Bagging Ensemble of Multilayer Perceptrons for Missing Electricity Consumption Data Imputation. Sensors 2020, 20, 1772. [Google Scholar] [CrossRef] [Green Version]
  26. Kim, K.H.; Oh, J.K.-W.; Jeong, W. Study on Solar Radiation Models in South Korea for Improving Office Building Energy Performance Analysis. Sustainability 2016, 8, 589. [Google Scholar] [CrossRef] [Green Version]
  27. Lee, M.; Park, J.; Na, S.-I.; Choi, H.S.; Bu, B.-S.; Kim, J. An Analysis of Battery Degradation in the Integrated Energy Storage System with Solar Photovoltaic Generation. Electronics 2020, 9, 701. [Google Scholar] [CrossRef]
  28. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T. LightGBM: A highly efficient gradient boosting decision tree. In Advances in Neural Information Processing Systems; Morgan Kaufmann Publishers: San Mateo, CA, USA, 2017; pp. 3148–3156. [Google Scholar]
  29. Rashmi, K.V.; Gilad-Bachrach, R. DART: Dropouts meet Multiple Additive Regression Trees. In Proceedings of the AISTATS, San Diego, CA, USA, 9–12 May 2015. [Google Scholar]
  30. De Livera, A.M.; Hyndman, R.J.; Snyder, R.D. Forecasting time series with complex seasonal patterns using exponential smoothing. J. Am. Stat. Assoc. 2011, 106, 1513–1527. [Google Scholar] [CrossRef] [Green Version]
  31. Breiman, L.E.O. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  32. Natekin, A.; Knoll, A. Gradient Boosting Machines: A Tutorial. Front. Neurorobot. 2013, 7, 21. [Google Scholar] [CrossRef] [Green Version]
  33. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  34. Park, S.; Moon, J.; Hwang, E. 2-Stage Electric Load Forecasting Scheme for Day-Ahead CCHP Scheduling. In Proceedings of the IEEE International Conference on Power Electronics and Drive System (PEDS), Toulouse, France, 9–12 July 2019. [Google Scholar] [CrossRef]
  35. Rahman, R.; Otridge, J.; Pal, R. IntegratedMRF: Random forest-based framework for integrating prediction from different data types. Bioinformatics 2017, 33, 1407–1410. [Google Scholar] [CrossRef] [PubMed]
  36. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  37. Feng, C.; Cui, M.; Hodge, B.-M.; Zhang, J. A data-driven multi-model methodology with deep feature selection for short-term wind forecasting. Appl. Energy 2017, 190, 1245–1257. [Google Scholar] [CrossRef] [Green Version]
  38. Moon, J.; Park, S.; Rho, S.; Hwang, E. A comparative analysis of artificial neural network architectures for building energy consumption forecasting. Int. J. Distrib. Sens. Netw. 2019, 15. [Google Scholar] [CrossRef] [Green Version]
  39. Moon, J.; Jung, S.; Rew, J.; Rho, S.; Hwang, E. Combination of short-term load forecasting models based on a stacking ensemble approach. Energy Build. 2020, 216, 109921. [Google Scholar] [CrossRef]
  40. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  41. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  42. Li, H.; Shen, Y.; Zhu, Y. Stock Price Prediction Using Attention-based Multi-Input LSTM. In Proceedings of the 10th Asian Conference on Machine Learning (ACML 2018), Beijing, China, 14–16 November 2018; pp. 454–469. [Google Scholar]
  43. Klambauer, G.; Unterthiner, T.; Mayr, A.; Hochreiter, S. Self-Normalizing Neural Networks. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 971–980. [Google Scholar]
  44. Moon, J.; Kim, J.; Kang, P.; Hwang, E. Solving the Cold-Start Problem in Short-Term Load Forecasting Using Tree-Based Methods. Energies 2020, 13, 886. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Short-term weather forecast by the Korea Meteorological Administration.
Figure 1. Short-term weather forecast by the Korea Meteorological Administration.
Remotesensing 12 02271 g001
Figure 2. Example of date data preprocessing.
Figure 2. Example of date data preprocessing.
Remotesensing 12 02271 g002
Figure 3. Input variable configuration for multistep-ahead (MSA) global solar radiation forecasting.
Figure 3. Input variable configuration for multistep-ahead (MSA) global solar radiation forecasting.
Remotesensing 12 02271 g003
Figure 4. Example of monthly time-series cross-validation.
Figure 4. Example of monthly time-series cross-validation.
Remotesensing 12 02271 g004
Figure 5. Boxplots by region ( MJ / m 2 ).
Figure 5. Boxplots by region ( MJ / m 2 ).
Remotesensing 12 02271 g005
Figure 6. Average mean bias error for each model of Ildo-1 ( MJ / m 2 ).
Figure 6. Average mean bias error for each model of Ildo-1 ( MJ / m 2 ).
Remotesensing 12 02271 g006
Figure 7. Average mean absolute error for each model of Ildo-1 ( MJ / m 2 ).
Figure 7. Average mean absolute error for each model of Ildo-1 ( MJ / m 2 ).
Remotesensing 12 02271 g007
Figure 8. Average root mean square error for each model of Ildo-1 ( MJ / m 2 ).
Figure 8. Average root mean square error for each model of Ildo-1 ( MJ / m 2 ).
Remotesensing 12 02271 g008
Figure 9. Average normalized root mean square error for each model of Ildo-1 ( % ).
Figure 9. Average normalized root mean square error for each model of Ildo-1 ( % ).
Remotesensing 12 02271 g009
Figure 10. Average mean bias error for each model for Gosan-ri ( MJ / m 2 ).
Figure 10. Average mean bias error for each model for Gosan-ri ( MJ / m 2 ).
Remotesensing 12 02271 g010
Figure 11. Average mean absolute error for each model for Gosan-ri ( MJ / m 2 ).
Figure 11. Average mean absolute error for each model for Gosan-ri ( MJ / m 2 ).
Remotesensing 12 02271 g011
Figure 12. Average root mean square error for each model for Gosan-ri ( MJ / m 2 ).
Figure 12. Average root mean square error for each model for Gosan-ri ( MJ / m 2 ).
Remotesensing 12 02271 g012
Figure 13. Average normalized root mean square error for each model for Gosan-ri ( % ).
Figure 13. Average normalized root mean square error for each model for Gosan-ri ( % ).
Remotesensing 12 02271 g013
Figure 14. Result of feature importance via time-series cross-validation using the input variables in Table 2 for Ildo-1. A cooler color indicates a lower feature importance value, whereas a warmer color indicates a higher feature importance value.
Figure 14. Result of feature importance via time-series cross-validation using the input variables in Table 2 for Ildo-1. A cooler color indicates a lower feature importance value, whereas a warmer color indicates a higher feature importance value.
Remotesensing 12 02271 g014
Figure 15. Result of feature importance via time-series cross-validation using the input variables in Table 2 for Gosan-ri. A cooler color indicates a lower feature importance value, whereas a warmer color indicates a higher feature importance value.
Figure 15. Result of feature importance via time-series cross-validation using the input variables in Table 2 for Gosan-ri. A cooler color indicates a lower feature importance value, whereas a warmer color indicates a higher feature importance value.
Remotesensing 12 02271 g015
Table 1. List of input variables (IV) for the proposed model.
Table 1. List of input variables (IV) for the proposed model.
IV #Input Variable (Feature)IV #Input Variable (Feature)
IV01 D a t e X (numeric)IV16 M o s t l y   c l o u d y (binary)
IV02 D a t e Y (numeric)IV17 C l o u d y (binary)
IV03 8   a . m . (binary)IV18 T e m p e r a t u r e (numeric)
IV04 9   a . m . (binary)IV19 H u m i d i t y (numeric)
IV05 10   a . m . (binary)IV20 W i n d   s p e e d (numeric)
IV06 11   a . m . (binary)IV21 D a t e X 1 day before (numeric)
IV07 12   p . m . (binary)IV22 D a t e Y 1 day before (numeric)
IV08 1   p . m . (binary)IV23 C l e a r 1 day before (binary)
IV09 2   p . m . (binary)IV24 P a r t l y   c l o u d y 1 day before (binary)
IV10 3   p . m . (binary)IV25 M o s t l y   c l o u d y 1 day before (binary)
IV11 4   p . m . (binary)IV26 C l o u d y 1 day before (binary)
IV12 5   p . m . (binary)IV27 T e m p e r a t u r e 1 day before (numeric)
IV13 6   p . m . (binary)IV28 H u m i d i t y 1 day before (numeric)
IV14 C l e a r (binary)IV29 W i n d   s p e e d 1 day before (numeric)
IV15 P a r t l y   c l o u d y (binary)IV30 G l o b a l   s o l a r   r a d i a t i o n 1 day before (numeric)
Table 2. Selected optimal hyperparameters for each deep learning model.
Table 2. Selected optimal hyperparameters for each deep learning model.
ModelsSelected Hyperparameters
SNNNumber of hidden layers: 1
Number of hidden nodes: 14
Activation function: sigmoid
Loss function: mean squared error
Optimizer: Adam
DNNNumber of hidden layers: 7
Activation function: ReLU, SELU [43]
Remaining hyperparameters are the same as those for the SNN model
LSTM networkSequence length: 11
Number of hidden layers: 2
Activation function: ReLU, SELU [43]
Loss function: Huber loss
Optimizer: RMSProp
Batch size: 11
Learning rate: 0.000001
Epoch: 5,000
ATT-LSTM networkNumber of attention layers: 1
Remaining hyperparameters are the same as those for the LSTM network model
Notes: SNN: shallow neural network; DNN: deep neural network; LSTM: long short-term memory; ATT-LSTM: attention-based LSTM; ReLU: rectified linear unit; SELU: scaled exponential linear unit.
Table 3. Statistical analysis of global solar radiation data by region ( MJ / m 2 ).
Table 3. Statistical analysis of global solar radiation data by region ( MJ / m 2 ).
StatisticsIldo-1Gosan-ri
Training SetTest SetTraining SetTest Set
Mean1.1881.2581.1791.044
Standard error0.0060.0110.0060.009
Median0.9101.0100.9100.840
Mode0000
Standard deviation0.9951.0000.9890.842
Sample variance0.9901.0000.9790.710
Kurtosis−0.756−0.869−0.400−0.419
Skewness0.6590.5680.7640.713
Range3.7503.7204.1303.550
Minimum0000
Maximum3.7503.7204.1303.550
Sum28,656.910,097.928,422.88383.9
Count24,112803024,1128030
Table 4. Selected hyperparameters for each ensemble learning model. Selected values are bold.
Table 4. Selected hyperparameters for each ensemble learning model. Selected values are bold.
ModelsPackage or ModuleSelected Hyperparameters
Random forestMultivariateRandomForestNumber of trees: 128 [44]
Number of features: 110 [44]
GBMQuantile regressionGradientBoostingRegressor
GridSearchCV
Learning rate: 0.01, 0.05, 0.1
Number of iterations: 100, 250, 500
Maximum depth of the tree: 5, 10
Huber lossLearning rate: 0.01, 0.05, 0.1
Number of iterations: 100, 250, 500
Maximum depth of the tree: 5, 10
XGBoostGBDTXGBoost 1.0.2
GridSearchCV
Learning rate: 0.01, 0.05, 0.1
Number of iterations: 250, 500, 1000
Maximum depth of the tree: 6, 8, 10
Subsample: 0.5, 0.75, 1.0
Colsample by tree: 0.5, 0.75, 1.0
Colsample by level: 0.5, 0.75, 1.0
Colsample by node: 0.5, 0.75, 1.0
DARTLearning rate: 0.01, 0.05, 0.1
Number of iterations: 250, 500, 1000
Maximum depth of the tree: 6, 8, 10
Subsample: 0.5, 0.75, 1.0
Colsample by tree: 0.5, 0.75, 1.0
Colsample by level: 0.5, 0.75, 1.0
Colsample by node: 0.5, 0.75, 1.0
LightGBMGBDTLightGBM 2.3.1
GridSearchCV
Learning rate: 0.01, 0.05, 0.1
Number of iterations: 1000, 1500
Number of leaves: 64
Subsample: 0.5
Colsample by tree: 1.0
DART (our model)Learning rate: 0.01, 0.05, 0.1
Number of iterations: 1000, 1500
Number of leaves: 64
Subsample: 0.5
Colsample by tree: 1.0
Notes: GBM: gradient boosting machine; XGBoost: extreme gradient boosting; LightGBM: light GBM; GBDT: gradient boosting decision tree; DART: dropouts meet multiple additive regression trees.
Table 5. Mean bias error (MBE) distribution for each model for Ildo-1 ( MJ / m 2 ).
Table 5. Mean bias error (MBE) distribution for each model for Ildo-1 ( MJ / m 2 ).
ModelsPoints
1234567891011
SNN (Dropout O)−0.05−0.04−0.03−0.03−0.03−0.05−0.06−0.07−0.07−0.06−0.04
SNN (Dropout X)−0.04−0.01−0.04−0.05−0.05−0.05−0.04−0.04−0.04−0.04−0.04
DNN-ReLU (Dropout O)−0.040−0.06−0.07−0.08−0.09−0.10−0.12−0.13−0.13−0.14
DNN-ReLU (Dropout X)−0.06−0.06−0.07−0.07−0.08−0.08−0.08−0.08−0.08−0.07−0.06
DNN-SELU (Dropout O)0.050.080.02−0.03−0.04−0.05−0.030−0.03−0.02−0.02
DNN-SELU (Dropout X)−0.06−0.06−0.09−0.10−0.10−0.10−0.10−0.10−0.10−0.10−0.11
LSTM-ReLU (Dropout O)−0.100.05−0.02−0.06−0.06−0.05−0.04−0.02−0.02−0.02−0.23
LSTM-ReLU (Dropout X)−0.19−0.01−0.02−0.06−0.10−0.12−0.12−0.11−0.10−0.09−0.08
LSTM-SELU (Dropout O)−0.09−0.02−0.06−0.07−0.06−0.04−0.03−0.03−0.02−0.03−0.03
LSTM-SELU (Dropout X)−0.10−0.03−0.06−0.09−0.10−0.09−0.09−0.09−0.09−0.09−0.09
ATT−LSTM-RELU (Dropout O)−0.16−0.14−0.09−0.07−0.07−0.010.020.030.01−0.05−0.09
ATT−LSTM-RELU (Dropout X)−0.14−0.11−0.11−0.0600.090.070.090.120.01−0.02
ATT−LSTM-SELU (Dropout O)−0.040.02−0.060.010.04−0.06−0.1000.03−0.02−0.03
ATT−LSTM-SELU (Dropout X)0.130.130.08−0.05−0.11−0.18−0.25−0.18−0.29−0.30−0.20
RF (TSCV)−0.03−0.04−0.05−0.06−0.07−0.07−0.07−0.07−0.07−0.06−0.06
RF (Holdout)−0.04−0.05−0.07−0.08−0.09−0.09−0.09−0.09−0.08−0.08−0.08
GBM-Huber (TSCV)−0.02−0.03−0.04−0.05−0.05−0.05−0.05−0.04−0.04−0.04−0.03
GBM-Huber (Holdout)−0.03−0.04−0.05−0.07−0.07−0.07−0.07−0.06−0.06−0.05−0.04
GBM-Quantile (TSCV)0.240.320.350.380.390.410.420.430.450.450.46
GBM-Quantile (Holdout)0.240.310.340.360.380.390.400.420.440.450.46
XGBoost-GDBT (TSCV)−0.02−0.03−0.05−0.05−0.05−0.05−0.05−0.05−0.05−0.04−0.04
XGBoost-GDBT (Holdout)−0.03−0.05−0.06−0.07−0.08−0.07−0.07−0.07−0.07−0.06−0.06
XGBoost-DART (TSCV)−0.02−0.03−0.04−0.05−0.05−0.05−0.05−0.05−0.05−0.04−0.04
XGBoost-DART (Holdout)−0.03−0.05−0.06−0.07−0.08−0.07−0.07−0.07−0.06−0.06−0.06
LightGBM-GDBT (TSCV)−0.02−0.03−0.04−0.04−0.05−0.05−0.05−0.05−0.04−0.05−0.05
LightGBM-GDBT (Holdout)−0.03−0.05−0.06−0.07−0.07−0.08−0.08−0.07−0.07−0.08−0.07
LightGBM-DART (TSCV)−0.03−0.04−0.05−0.06−0.06−0.06−0.06−0.05−0.05−0.05−0.05
LightGBM-DART (Holdout)−0.04−0.06−0.07−0.08−0.08−0.08−0.08−0.07−0.07−0.07−0.06
Table 6. Mean absolute error (MAE) distribution for each model for Ildo-1. A cooler color indicates a lower MAE value, whereas a warmer color indicates a higher MAE value ( MJ / m 2 ).
Table 6. Mean absolute error (MAE) distribution for each model for Ildo-1. A cooler color indicates a lower MAE value, whereas a warmer color indicates a higher MAE value ( MJ / m 2 ).
ModelsPoints
1234567891011
SNN (Dropout O)0.4450.4060.3940.3900.3860.3850.3850.3860.3850.3850.385
SNN (Dropout X)0.4130.3900.3930.3920.3900.3880.3870.3870.3880.3910.395
DNN-ReLU (Dropout O)0.4190.3840.3890.3860.3820.3790.3800.3820.3850.3880.391
DNN-ReLU (Dropout X)0.4450.4010.3980.3870.3840.3850.3830.3810.3810.3810.381
DNN-SELU (Dropout O)0.3490.3270.3430.3540.3630.3680.3750.3820.3950.3950.405
DNN-SELU (Dropout X)0.4080.3830.3800.3780.3760.3740.3730.3730.3740.3760.378
LSTM-ReLU (Dropout O)0.3650.3880.4090.4120.4200.4270.4360.4360.4460.4580.460
LSTM-ReLU (Dropout X)0.3800.3990.4170.4260.4310.4420.4560.4670.4680.4700.470
LSTM-SELU (Dropout O)0.3180.3320.3500.3600.3720.3770.3710.3730.3780.3790.370
LSTM-SELU (Dropout X)0.3570.3790.4000.4120.4130.4380.4460.4670.4880.5010.517
ATT-LSTM-RELU (Dropout O)0.2910.3240.3290.3400.3470.3480.3570.3630.3680.3820.394
ATT-LSTM-RELU (Dropout X)0.2720.2990.3240.3320.3390.3510.3490.3600.3750.3780.392
ATT-LSTM-SELU (Dropout O)0.2380.2910.3110.3300.3330.3460.3680.3480.3580.3710.381
ATT-LSTM-SELU (Dropout X)0.2610.3010.3070.3190.3430.3760.4140.3860.4390.4520.415
RF (TSCV)0.2150.2720.3060.3300.3470.3630.3750.3880.4010.4140.428
RF (Holdout)0.2230.2800.3150.3400.3580.3710.3830.3960.4090.4230.435
GBM-Huber (TSCV)0.1860.2510.2920.3180.3370.3510.3590.3680.3740.3820.385
GBM-Huber (Holdout)0.1890.2570.3000.3300.3500.3590.3700.3770.3830.3920.394
GBM-Quantile (TSCV)0.2880.3760.4180.4450.4620.4760.4840.4980.5110.5180.525
GBM-Quantile (Holdout)0.2870.3750.4130.4450.4610.4700.4820.4970.5100.5240.530
XGBoost-GDBT (TSCV)0.1940.2570.2960.3240.3390.3480.3560.3660.3750.3840.390
XGBoost-GDBT (Holdout)0.1970.2630.3050.3330.3490.3580.3670.3740.3830.3900.396
XGBoost-DART (TSCV)0.1840.2490.2890.3170.3330.3460.3530.3620.3690.3770.382
XGBoost-DART (Holdout)0.1880.2550.2980.3280.3460.3570.3640.3720.3790.3860.393
LightGBM-GDBT (TSCV)0.1890.2530.2920.3190.3340.3480.3550.3630.3710.3770.381
LightGBM-GDBT (Holdout)0.1930.2610.3050.3310.3460.3590.3680.3730.3820.3910.394
LightGBM-DART (TSCV)0.1890.2520.2900.3180.3330.3440.3500.3580.3650.3740.380
LightGBM-DART (Holdout)0.1930.2570.3000.3280.3440.3550.3590.3690.3770.3840.389
Table 7. Root mean square error (RMSE) distribution for each model for Ildo-1. A cooler color indicates a lower RMSE value, whereas a warmer color indicates a higher RMSE value ( MJ / m 2 ).
Table 7. Root mean square error (RMSE) distribution for each model for Ildo-1. A cooler color indicates a lower RMSE value, whereas a warmer color indicates a higher RMSE value ( MJ / m 2 ).
ModelsPoints
1234567891011
SNN (Dropout O)0.5880.5460.5370.5340.5280.5260.5240.5250.5240.5250.528
SNN (Dropout X)0.5450.5320.5340.5330.5290.5260.5240.5240.5260.5290.534
DNN-ReLU (Dropout O)0.5450.5280.5290.5250.5190.5140.5130.5150.5190.5230.527
DNN-ReLU (Dropout X)0.5980.5480.5450.5250.5200.5180.5150.5140.5130.5140.515
DNN-SELU (Dropout O)0.3920.4410.4650.4790.4870.4990.5060.5080.5220.5190.532
DNN-SELU (Dropout X)0.5410.5260.5230.5120.5150.5110.5090.5090.5100.5110.514
LSTM-ReLU (Dropout O)0.5440.5470.5450.5410.5400.5400.5470.5510.5590.5700.636
LSTM-ReLU (Dropout X)0.5450.5500.5430.5440.5000.5430.5500.5510.5580.5670.637
LSTM-SELU (Dropout O)0.4080.4480.4620.4730.4840.4940.5060.5160.5170.5180.519
LSTM-SELU (Dropout X)0.4310.4670.4810.4920.5060.5070.5210.5310.5510.5510.558
ATT-LSTM-RELU (Dropout O)0.3810.4300.4460.4640.4780.4810.4910.4990.5050.5150.528
ATT-LSTM-RELU (Dropout X)0.3830.4280.4580.4660.4750.4960.4980.5120.5290.5300.543
ATT-LSTM-SELU (Dropout O)0.3290.3950.4310.4550.4640.4790.5020.4930.5080.5190.528
ATT-LSTM-SELU (Dropout X)0.3570.4150.4310.4500.4750.5090.5570.5380.5860.5980.561
RF (TSCV)0.3020.3780.4210.4500.4710.4900.5040.5160.5280.5400.554
RF (Holdout)0.3080.3860.4310.4620.4840.5020.5150.5270.5380.5500.564
GBM-Huber (TSCV)0.2850.3690.4170.4510.4720.4900.4980.5070.5130.5180.520
GBM-Huber (Holdout)0.2880.3760.4270.4650.4900.5010.5140.5210.5260.5300.533
GBM-Quantile (TSCV)0.4120.5270.5940.6380.6630.6820.6950.7140.7320.7390.749
GBM-Quantile (Holdout)0.4090.5230.5840.6370.6600.6710.6890.7060.7270.7390.751
XGBoost-GDBT (TSCV)0.2890.3710.4170.4520.4710.4840.4940.5040.5110.5170.521
XGBoost-GDBT (Holdout)0.2900.3760.4270.4660.4850.4980.5080.5140.5200.5230.528
XGBoost-DART (TSCV)0.2800.3630.4100.4440.4660.4810.4910.4990.5040.5090.514
XGBoost-DART (Holdout)0.2830.3690.4210.4600.4830.4980.5080.5130.5190.5220.527
LightGBM-GDBT (TSCV)0.2850.3680.4150.4490.4690.4880.4960.5070.5120.5150.520
LightGBM-GDBT (Holdout)0.2890.3770.4310.4660.4850.5040.5140.5190.5260.5320.537
LightGBM-DART (TSCV)0.2840.3660.4110.4440.4640.4790.4870.4960.5020.5080.514
LightGBM-DART (Holdout)0.2880.3700.4210.4590.4820.4940.5020.5110.5180.5230.525
Table 8. Normalized root mean square error (NRMSE) distribution for each model for Ildo-1. A cooler color indicates a lower NRMSE value, whereas a warmer color indicates a higher NRMSE value ( % ).
Table 8. Normalized root mean square error (NRMSE) distribution for each model for Ildo-1. A cooler color indicates a lower NRMSE value, whereas a warmer color indicates a higher NRMSE value ( % ).
ModelsPoints
1234567891011
SNN (Dropout O)46.643.342.642.441.941.741.641.641.541.641.8
SNN (Dropout X)42.941.741.441.140.840.540.340.340.440.540.7
DNN-ReLU (Dropout O)43.641.942.041.641.140.840.740.941.141.441.8
DNN-ReLU (Dropout X)47.443.542.441.741.241.040.840.740.740.740.8
DNN-SELU (Dropout O)31.235.036.938.038.739.740.240.441.541.342.2
DNN-SELU (Dropout X)43.242.242.442.242.041.741.541.541.741.942.4
LSTM-ReLU (Dropout O)48.243.542.442.141.841.441.140.940.841.041.2
LSTM-ReLU (Dropout X)55.948.747.246.546.145.945.545.845.345.445.6
LSTM-SELU (Dropout O)46.743.142.341.641.240.940.840.941.141.341.5
LSTM-SELU (Dropout X)50.545.044.344.043.943.743.143.343.543.743.9
ATT-LSTM-RELU (Dropout O)30.234.235.436.937.938.239.039.640.140.941.9
ATT-LSTM-RELU (Dropout X)30.434.036.337.037.739.439.540.742.042.143.1
ATT-LSTM-SELU (Dropout O)26.131.434.236.136.838.039.839.140.341.242.0
ATT-LSTM-SELU (Dropout X)28.332.934.235.837.740.444.242.746.647.644.5
RF (TSCV)24.030.033.435.737.438.940.141.042.043.044.1
RF (Holdout)24.530.734.336.738.539.941.041.942.843.844.8
GBM-Huber (TSCV)22.629.333.135.837.538.939.640.340.841.241.4
GBM-Huber (Holdout)22.829.933.936.939.039.840.941.441.842.142.3
GBM-Quantile (TSCV)32.741.947.250.752.754.255.256.758.258.859.5
GBM-Quantile (Holdout)32.541.646.450.652.453.454.856.257.858.859.7
XGBoost-GDBT (TSCV)22.929.433.135.937.438.539.340.140.641.141.4
XGBoost-GDBT (Holdout)23.129.933.937.038.539.640.340.841.441.642.0
XGBoost-DART (TSCV)22.328.932.635.337.038.239.039.740.140.540.8
XGBoost-DART (Holdout)22.529.333.536.538.439.640.440.841.341.541.9
LightGBM-GDBT (TSCV)22.729.333.035.737.238.839.540.340.740.941.4
LightGBM-GDBT (Holdout)23.029.934.237.038.540.040.941.241.842.342.7
LightGBM-DART (TSCV)22.529.132.735.336.938.138.739.439.940.440.9
LightGBM-DART (Holdout)22.929.433.536.538.339.339.940.641.141.641.8
Table 9. Mean bias error (MBE) distribution for each model of Gosan-ri ( MJ / m 2 ).
Table 9. Mean bias error (MBE) distribution for each model of Gosan-ri ( MJ / m 2 ).
ModelsPoints
1234567891011
SNN (Dropout O)−0.04−0.01−0.02−0.02−0.02−0.02−0.02−0.02−0.02−0.02−0.02
SNN (Dropout X)0.010.01−0.01−0.02−0.02−0.02−0.01000.010.02
DNN-ReLU (Dropout O)−0.0100.010.020.020.020.020.020.020.030.03
DNN-ReLU (Dropout X)0.040.070.050.040.030.030.040.040.040.040.04
DNN-SELU (Dropout O)0.040.060.070.060.040.050.050.060.06−0.01−0.01
DNN-SELU (Dropout X)−0.020.020.01−0.02−0.03−0.04−0.04−0.04−0.05−0.05−0.05
LSTM-ReLU (Dropout O)−0.130.080.080.050.02−0.01−0.01−0.010.010.030.04
LSTM-ReLU (Dropout X)−0.160.020.030−0.03−0.05−0.05−0.05−0.40−0.03−0.02
LSTM-SELU (Dropout O)00.030−0.02−0.04−0.04−0.04−0.04−0.04−0.04−0.04
LSTM-SELU (Dropout X)−0.050.040.020−0.02−0.03−0.03−0.03−0.02−0.02−0.02
ATT-LSTM-RELU (Dropout O)0.02−0.07−0.010.030.040.130.160.10−0.020.060
ATT-LSTM-RELU (Dropout X)−0.05−0.09−0.08−0.16−0.24−0.20−0.25−0.15−0.24−0.28−0.25
ATT-LSTM-SELU (Dropout O)0.050.050.01−0.02−0.010.01−0.08−0.11−0.14−0.06−0.02
ATT-LSTM-SELU (Dropout X)−0.04−0.06−0.01−0.02−0.04−0.02−0.020−0.010.03−0.01
RF (TSCV)00000000.010.010.020.02
RF (Holdout)0000000.010.010.020.020.03
GBM-Huber (TSCV)−0.01−0.01−0.01−0.01−0.01000000
GBM-Huber (Holdout)−0.01−0.01−0.01−0.01−0.0100.010.010.010.010.01
GBM-Quantile (TSCV)0.370.460.530.580.600.620.630.650.660.670.68
GBM-Quantile (Holdout)0.260.320.360.390.420.410.430.440.450.470.46
XGBoost-GDBT (TSCV)−0.01−0.01−0.01−0.01−0.01−0.01−0.010−0.0100
XGBoost-GDBT (Holdout)−0.01−0.01−0.01−0.01−0.01−0.0100000
XGBoost-DART (TSCV)0−0.01−0.01−0.010000000
XGBoost-DART (Holdout)0−0.010000.010.010.010.010.010.01
LightGBM-GDBT (TSCV)−0.01−0.01000000000
LightGBM-GDBT (Holdout)0−0.01000.010.010.010.010.010.010.01
LightGBM-DART (TSCV)−0.01−0.02−0.01−0.01−0.01−0.01−0.010−0.0100
LightGBM-DART (Holdout)−0.01−0.02−0.01−0.01−0.0100.010000
Table 10. Mean absolute error (MAE) distribution for each model of Gosan-ri. A cooler color indicates a lower MAE value, whereas a warmer color indicates a higher MAE value ( MJ / m 2 ).
Table 10. Mean absolute error (MAE) distribution for each model of Gosan-ri. A cooler color indicates a lower MAE value, whereas a warmer color indicates a higher MAE value ( MJ / m 2 ).
ModelsPoints
1234567891011
SNN (Dropout O)0.4260.3990.3910.3870.3810.3770.3730.3740.3750.3740.373
SNN (Dropout X)0.3880.3750.3740.3700.3640.3600.3580.3570.3570.3560.355
DNN-ReLU (Dropout O)0.4090.3810.3770.3730.3700.3690.3670.3670.3690.3670.374
DNN-ReLU (Dropout X)0.4210.3990.3870.3940.3840.3790.3710.3880.4010.3990.389
DNN-SELU (Dropout O)0.2780.3060.3250.3430.3550.3650.3670.3790.3880.3870.397
DNN-SELU (Dropout X)0.4090.3810.3800.3750.3700.3670.3640.3630.3650.3670.367
LSTM-ReLU (Dropout O)0.3450.3580.3590.3620.3700.3770.3860.3760.3760.3780.380
LSTM-ReLU (Dropout X)0.3770.3990.4010.4080.4150.4200.4220.4210.4210.4210.420
LSTM-SELU (Dropout O)0.2980.3120.3200.3400.3520.3670.3710.3830.3880.3890.390
LSTM-SELU (Dropout X)0.3400.3510.3520.3570.3600.3610.3690.3720.3710.3710.371
ATT-LSTM-RELU (Dropout O)0.2210.2650.2890.3170.3320.3650.3770.3650.3660.3710.373
ATT-LSTM-RELU (Dropout X)0.2400.2950.3110.3460.3850.3830.4090.3840.4160.4450.442
ATT-LSTM-SELU (Dropout O)0.2190.2620.2890.3160.3270.3400.3540.3630.3760.3780.383
ATT-LSTM-SELU (Dropout X)0.2310.2730.2970.3100.3270.3380.3470.3590.3670.3760.380
RF (TSCV)0.1760.2270.2560.2770.2930.3070.3190.3310.3430.3530.363
RF (Holdout)0.1800.2300.2600.2800.2970.3100.3240.3350.3460.3560.364
GBM-Huber (TSCV)0.1640.2240.2590.2840.3020.3140.3230.3300.3400.3470.352
GBM-Huber (Holdout)0.1650.2260.2640.2900.3060.3160.3240.3340.3400.3500.355
GBM-Quantile (TSCV)0.2890.3590.4100.4530.4680.4810.4920.5050.5100.5220.530
GBM-Quantile (Holdout)0.2970.3670.4180.4570.4840.4830.5040.5110.5230.5410.534
XGBoost-GDBT (TSCV)0.1680.2220.2540.2800.2950.3060.3140.3240.3340.3410.346
XGBoost-GDBT (Holdout)0.1700.2240.2580.2850.2990.3080.3160.3250.3340.3430.349
XGBoost-DART (TSCV)0.1590.2160.2490.2750.2910.3020.3120.3210.3290.3390.345
XGBoost-DART (Holdout)0.1610.2200.2550.2810.2960.3070.3160.3250.3330.3410.348
LightGBM-GDBT (TSCV)0.1670.2250.2590.2830.3000.3120.3200.3310.3360.3430.353
LightGBM-GDBT (Holdout)0.1690.2280.2650.2900.3050.3160.3250.3360.3430.3480.355
LightGBM-DART (TSCV)0.1630.2180.2520.2760.2930.3040.3140.3220.3320.3410.347
LightGBM-DART (Holdout)0.1650.2220.2590.2830.2970.3110.3180.3280.3360.3440.352
Table 11. Root mean square error (RMSE) distribution for each model for Gosan-ri. A cooler color indicates a lower RMSE value, whereas a warmer color indicates a higher RMSE value ( MJ / m 2 ).
Table 11. Root mean square error (RMSE) distribution for each model for Gosan-ri. A cooler color indicates a lower RMSE value, whereas a warmer color indicates a higher RMSE value ( MJ / m 2 ).
ModelsPoints
1234567891011
SNN (Dropout O)0.5590.5330.5280.5310.5300.5270.5250.5260.5230.5220.521
SNN (Dropout X)0.5180.5180.5200.5170.5120.5060.5020.5010.4990.4970.497
DNN-ReLU (Dropout O)0.5220.5190.5190.5120.5060.5000.4960.4950.4950.4960.497
DNN-ReLU (Dropout X)0.5520.5220.5230.5220.5200.5200.5210.5210.5210.5220.526
DNN-SELU (Dropout O)0.3660.4070.4310.4520.4690.4840.4900.5030.5160.5140.524
DNN-SELU (Dropout X)0.5320.5190.5210.5170.5140.5110.5090.5090.5100.5110.511
LSTM-ReLU (Dropout O)0.4000.4200.4370.4420.4510.4620.4730.4880.5010.5110.511
LSTM-ReLU (Dropout X)0.4550.4620.4770.4840.4910.5000.5110.5310.5470.5670.580
LSTM-SELU (Dropout O)0.3750.3800.4100.4440.4670.4860.4920.5100.5260.5330.538
LSTM-SELU (Dropout X)0.4060.4170.4510.4710.4900.4960.5000.5080.5220.5470.555
ATT-LSTM-RELU (Dropout O)0.3060.3700.4030.4350.4540.4850.4970.4950.5080.5150.517
ATT-LSTM-RELU (Dropout X)0.3380.4160.4450.4900.5410.5390.5690.5270.5660.5860.581
ATT-LSTM-SELU (Dropout O)0.2970.3600.4010.4360.4520.4730.5000.5150.5290.5240.526
ATT-LSTM-SELU (Dropout X)0.3220.3800.4150.4320.4560.4700.4850.4990.5110.5220.524
RF (TSCV)0.2570.3220.3590.3870.4090.4260.4410.4570.4700.4810.490
RF (Holdout)0.2600.3260.3640.3910.4130.4310.4490.4640.4750.4850.494
GBM-Huber (TSCV)0.2510.3240.3680.4000.4250.4420.4570.4660.4770.4860.491
GBM-Huber (Holdout)0.2520.3270.3730.4080.4300.4460.4570.4710.4790.4890.491
GBM-Quantile (TSCV)0.3880.4800.5490.6010.6250.6440.6590.6820.6870.7040.711
GBM-Quantile (Holdout)0.3970.4880.5560.6070.6410.6480.6730.6850.7060.7230.720
XGBoost-GDBT (TSCV)0.2520.3200.3620.3960.4160.4300.4450.4580.4690.4960.480
XGBoost-GDBT (Holdout)0.2530.3210.3670.4030.4230.4350.4480.4610.4720.4800.483
XGBoost-DART (TSCV)0.2440.3140.3570.3910.4140.4280.4430.4570.4650.4740.480
XGBoost-DART (Holdout)0.2460.3180.3640.3990.4190.4340.4480.4620.4720.4790.484
LightGBM-GDBT (TSCV)0.2510.3230.3670.3990.4220.4400.4540.4680.4740.4820.493
LightGBM-GDBT (Holdout)0.2540.3280.3750.4080.4290.4480.4610.4770.4860.4890.495
LightGBM-DART (TSCV)0.2470.3180.3600.3910.4140.4310.4450.4580.4700.4770.482
LightGBM-DART (Holdout)0.2490.3220.3670.4000.4200.4380.4520.4650.4750.4840.490
Table 12. Normalized root mean square error (NRMSE) distribution for each model for Gosan-ri. A cooler color indicates a lower NRMSE value, whereas a warmer color indicates a higher NRMSE value ( % ).
Table 12. Normalized root mean square error (NRMSE) distribution for each model for Gosan-ri. A cooler color indicates a lower NRMSE value, whereas a warmer color indicates a higher NRMSE value ( % ).
ModelsPoints
1234567891011
SNN (Dropout O)53.451.050.550.750.750.450.250.349.949.949.8
SNN (Dropout X)49.549.549.749.448.948.448.047.847.747.547.4
DNN-ReLU (Dropout O)52.749.950.049.949.749.749.849.749.749.850.2
DNN-ReLU (Dropout X)49.949.749.648.948.347.847.447.347.347.347.4
DNN-SELU (Dropout O)35.139.041.343.244.946.346.948.249.449.250.2
DNN-SELU (Dropout X)50.849.649.749.349.148.948.648.648.748.848.7
LSTM-ReLU (Dropout O)56.650.650.649.748.948.748.448.348.148.148.0
LSTM-ReLU (Dropout X)60.653.754.053.152.452.452.252.551.852.052.0
LSTM-SELU (Dropout O)49.749.148.949.049.249.449.649.849.849.749.6
LSTM-SELU (Dropout X)54.451.351.651.551.251.451.451.551.151.051.2
ATT-LSTM-RELU (Dropout O)29.335.438.641.643.546.447.647.448.649.349.5
ATT-LSTM-RELU (Dropout X)26.933.135.338.943.042.845.241.945.046.646.2
ATT-LSTM-SELU (Dropout O)28.534.538.441.743.345.347.849.350.750.150.3
ATT-LSTM-SELU (Dropout X)30.836.439.841.343.645.046.447.748.850.050.1
RF (TSCV)24.630.834.437.139.140.842.343.845.046.147.0
RF (Holdout)24.931.234.837.439.641.343.044.445.546.547.3
GBM-Huber (TSCV)24.031.035.238.340.742.343.744.645.746.547.0
GBM-Huber (Holdout)24.131.335.839.041.242.743.745.145.846.947.1
GBM-Quantile (TSCV)37.246.052.657.659.961.763.165.365.867.568.2
GBM-Quantile (Holdout)38.146.753.258.161.462.064.465.667.669.268.9
XGBoost-GDBT (TSCV)24.130.634.637.939.941.242.643.944.945.646.0
XGBoost-GDBT (Holdout)24.230.835.138.640.541.742.944.245.245.946.3
XGBoost-DART (TSCV)23.430.134.237.539.641.042.443.744.645.446.0
XGBoost-DART (Holdout)23.630.434.838.240.141.642.944.245.245.846.4
LightGBM-GDBT (TSCV)24.131.035.238.240.442.243.544.845.446.247.2
LightGBM-GDBT (Holdout)24.331.435.939.141.142.944.145.646.646.847.4
LightGBM-DART (TSCV)23.630.434.537.439.641.242.643.845.045.746.1
LightGBM-DART (Holdout)23.830.935.238.340.242.043.344.545.546.346.9
Table 13. Average mean bias error, mean absolute error, root mean square error, and normalized root mean square error comparison according to the forecasting models.
Table 13. Average mean bias error, mean absolute error, root mean square error, and normalized root mean square error comparison according to the forecasting models.
ModelsIldo-1Gosan-ri
MBEMAERMSENRMSEMBEMAERMSENRMSE
SNN (Dropout O)−0.0480.3940.53542.4−0.0210.5300.38550.6
SNN (Dropout X)−0.0400.3920.53141.0−0.0020.5080.36548.5
DNN-ReLU (Dropout O)−0.0870.3880.52341.50.0170.5050.37550.1
DNN-ReLU (Dropout X)−0.0720.3920.53041.90.0400.5250.39248.3
DNN-SELU (Dropout O)−0.0060.3690.48638.60.0420.4690.35444.9
DNN-SELU (Dropout X)−0.0930.3790.51642.1−0.0290.5150.37349.2
LSTM-ReLU (Dropout O)−0.0520.4230.55642.20.0150.4630.37049.6
LSTM-ReLU (Dropout X)−0.0910.4390.55347.1−0.0680.5100.41153.3
LSTM-SELU (Dropout O)−0.0440.3620.48641.9−0.0240.4690.35549.4
LSTM-SELU (Dropout X)−0.0840.4380.50944.4−0.0150.4880.36151.6
ATT-LSTM-RELU (Dropout O)−0.0560.3490.47437.70.0400.4530.33143.4
ATT-LSTM-RELU (Dropout X)−0.0050.3430.48338.4−0.1800.5090.36940.4
ATT-LSTM-SELU (Dropout O)−0.0190.3340.46436.8−0.0280.4560.32843.6
ATT-LSTM-SELU (Dropout X)−0.1110.3650.49839.5−0.0190.4560.32843.6
RF (TSCV)−0.0590.3490.46937.20.0050.4090.29539.2
RF (Holdout)−0.0760.3580.47938.10.0080.4140.29839.6
GBM-Huber (TSCV)−0.0400.3280.45836.4−0.0020.4170.29439.9
GBM-Huber (Holdout)−0.0550.3360.47037.30.0010.4200.29740.2
GBM-Quantile (TSCV)0.3910.4550.65051.60.5860.6120.45658.6
GBM-Quantile (Holdout)0.3810.4540.64551.30.4020.6220.46559.6
XGBoost-GDBT (TSCV)−0.0440.3300.45736.3−0.0070.4110.28939.2
XGBoost-GDBT (Holdout)−0.0630.3380.46737.1−0.0050.4130.29239.6
XGBoost-DART (TSCV)−0.0430.3240.45135.90.0000.4060.28538.9
XGBoost-DART (Holdout)−0.0620.3330.46436.90.0040.4110.28939.4
LightGBM-GDBT (TSCV)−0.0430.3260.45736.3−0.0010.4160.29439.8
LightGBM-GDBT (Holdout)−0.0660.3370.47137.40.0050.4230.29840.5
LightGBM-DART (TSCV)−0.0510.3230.45035.8−0.0090.4080.28739.1
LightGBM-DART (Holdout)−0.0690.3320.46336.8−0.0040.4150.29239.7
Notes: MBE: mean bias error; MAE: mean absolute error; RMSE: root mean square error; NRMSE: normalized RMSE.

Share and Cite

MDPI and ACS Style

Park, J.; Moon, J.; Jung, S.; Hwang, E. Multistep-Ahead Solar Radiation Forecasting Scheme Based on the Light Gradient Boosting Machine: A Case Study of Jeju Island. Remote Sens. 2020, 12, 2271. https://doi.org/10.3390/rs12142271

AMA Style

Park J, Moon J, Jung S, Hwang E. Multistep-Ahead Solar Radiation Forecasting Scheme Based on the Light Gradient Boosting Machine: A Case Study of Jeju Island. Remote Sensing. 2020; 12(14):2271. https://doi.org/10.3390/rs12142271

Chicago/Turabian Style

Park, Jinwoong, Jihoon Moon, Seungmin Jung, and Eenjun Hwang. 2020. "Multistep-Ahead Solar Radiation Forecasting Scheme Based on the Light Gradient Boosting Machine: A Case Study of Jeju Island" Remote Sensing 12, no. 14: 2271. https://doi.org/10.3390/rs12142271

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop