Next Article in Journal
Activation of Low-Quality Coal Gangue Using Suspension Calcination for the Preparation of High-Performance Low-Carbon Cementitious Materials: A Pilot Study
Previous Article in Journal
Optimizing Pneumonia Diagnosis Using RCGAN-CTL: A Strategy for Small or Limited Imaging Datasets
Previous Article in Special Issue
Atomic Orbital Search Algorithm for Efficient Maximum Power Point Tracking in Partially Shaded Solar PV Systems
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:

Optimizing Short-Term Photovoltaic Power Forecasting: A Novel Approach with Gaussian Process Regression and Bayesian Hyperparameter Tuning

Md. Samin Safayat Islam
Puja Ghosh
Md. Omer Faruque
Md. Rashidul Islam
Md. Alamgir Hossain
Md. Shafiul Alam
4 and
Md. Rafiqul Islam Sheikh
Department of Electrical & Electronic Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
Department of Electrical & Electronic Engineering, Dhaka University of Engineering & Technology, Gazipur 1707, Bangladesh
Queensland Micro- and Nanotechnology Centre, Griffith University, Nathan, QLD 4111, Australia
Applied Research Center for Environment & Marine Studies, King Fahd University of Petroleum & Minerals (KFUPM), Dhahran 31261, Saudi Arabia
Author to whom correspondence should be addressed.
Processes 2024, 12(3), 546;
Submission received: 23 January 2024 / Revised: 24 February 2024 / Accepted: 4 March 2024 / Published: 11 March 2024
(This article belongs to the Special Issue Recent Advances in Sustainable Electrical Energy Technologies)


The inherent volatility of PV power introduces unpredictability to the power system, necessitating accurate forecasting of power generation. In this study, a machine learning (ML) model based on Gaussian process regression (GPR) for short-term PV power output forecasting is proposed. With its benefits in handling nonlinear relationships, estimating uncertainty, and generating probabilistic forecasts, GPR is an appropriate approach for addressing the problems caused by PV power generation’s irregularity. Additionally, Bayesian optimization to identify optimal hyper-parameter combinations for the ML model is utilized. The research leverages solar radiation intensity data collected at 60-min and 30-min intervals over periods of 1 year and 6 months, respectively. Comparative analysis reveals that the data set with 60-min intervals performs slightly better than the 30-min intervals data set. The proposed GPR model, coupled with Bayesian optimization, demonstrates superior performance compared to contemporary ML models and traditional neural network models. This superiority is evident in 98% and 90% improvements in root mean square errors compared to feed-forward neural network and artificial neural network models, respectively. This research contributes to advancing accurate and efficient forecasting methods for PV power output, thereby enhancing the reliability and stability of power systems.

1. Introduction

The global demand for energy has been steadily increasing over the last few decades, leading to substantial growth in electricity production worldwide [1]. Traditionally, conventional fossil fuels have served as the primary energy source for electricity generation. However, the sustainability of these traditional fossil fuel resources is now uncertain due to their rapid depletion [2]. Moreover, the combustion of fossil fuels is a major contributor to greenhouse gas emissions, exacerbating severe global warming and posing a threat to the planet’s health [1,3].
In response to these challenges, there has been a significant shift in focus from fossil fuels to renewable energy sources (RESs) for electricity generation. This transition involves harnessing solar energy, wind energy, tidal energy, and biomass energy [4,5]. In the context of growing energy development, the incorporation of renewable energy sources (RES) into contemporary power systems is increasingly reliant on hydrogen. Among the various RESs, solar energy stands out as the most widespread and promising resource for cleaner power generation. The Earth’s surface receives average solar radiation of approximately 1367 W/m2, capable of generating a staggering 1.8 × 1011 MW annually [6]. This immense and pervasive energy supply is more than sufficient to effortlessly meet the world’s energy needs. The distribution of solar potential, as illustrated in Figure 1, showcases enormous opportunities to harness solar energy on Earth.
The International Energy Agency (IEA) recognizes photovoltaics (PV) as one of the fastest-growing and most cost-effective renewable energy sources. Currently, PV forecasting research is considered a central focus among the major areas of prediction analytics. The production of PV power is highly influenced by meteorological variables, including solar radiation, clouds, humidity, pressure, temperature, etc. The inherently intermittent nature of these factors renders PV power production unpredictable, thereby impacting grid stability and the cost-effectiveness of power system operation [7]. Therefore, the accurate prediction of PV power output is crucial to eliminate these uncertainties and fluctuations.
Forecasting PV power on an hourly to day-ahead basis offers numerous advantages, including optimized power dispatching, cost efficiency, and a dependable power supply. Accurate PV power prediction contributes to appropriate power dispatching, cost-effectiveness, and a reliable power supply. With precise forecasting, companies can avoid penalties, schedule supply offerings to the market as efficiently as possible, and enhance revenues [8,9]. Additionally, accurate forecasting can be used to make decisions regarding unit commitment, reserve needs, and maintenance planning to achieve the lowest possible operating costs. Considering these factors, accurate PV forecasting has been acknowledged as one of the major issues in power systems [10].
Solar power forecasting can be categorized into physical, statistical, and machine-learning models. Numerous methods have been developed using both physical models and numerical approaches. Numerical weather prediction (NWP) is established based on structural and physical hypotheses [11]. The chaotic nature of weather patterns and atmospheric uncertainties makes NWP computationally challenging and necessitates a greater operating cost. Additionally, these models must be developed for specific locations. The foundation of statistical models is the use of conventional regressive mathematical models like linear regression (LR) and autoregressive integrated moving average (ARIMA) models. Since LR models linearly map the target power and inputs, they are unable to effectively depict the nonlinear relationship between solar power’s input characteristics and outputs. Furthermore, the accuracy of these models declines as the forecasting horizon lengthens. To resolve these difficulties, several machine learning (ML) techniques have been developed.
Artificial neural networks (ANNs) and support vector machines (SVMS) are the most common ML models with strong nonlinear approximation capabilities [12]. Since the 1980s, researchers have increasingly used ANNs. ANNs have gained popularity among researchers over statistical models due to a higher degree of accuracy and the nonlinearity of meteorological data. The most frequently used ANNs include the multilayer perceptron (MLP), feed-forward neural networks (FFNNs), radial basis neural networks (RBNNs), recurrent neural networks (RNNs), backpropagation neural networks (BPNNs), general regression neural network (GRNNs), and adaptive neuro-fuzzy interface systems (ANFISs) [13]. A good comparison can be found in the work of Donnelly et al., which utilized Gaussian process (GP) models to emulate a hydraulic inundation model, enabling real-time flood depth predictions with uncertainty quantification [14]. Their framework accurately reproduces water depths and the extent of inundation, surpassing a neural network-based emulator and offering significant computational speedups compared to the original simulator.
MLP is a type of supervised feed-forward ANN with one or more layers between its input and output layers. These layers, sometimes known as hidden layers, can be adjusted to fit the difficulty of a problem. FFNN is a less sophisticated ANN architecture in which information simply passes from the input layer to the output layer [15]. On the other hand, an RBNN is considered a two-layer ANN, and the learning process can be divided into two distinct stages based on the weights of the synaptic connections [16]. It performs well but requires time for the learning process. An RNN is effective at learning various computational structures and intricate interactions, making it a suitable tool for forecasting time-series data.
According to Yona et al., forecasting errors are significantly reduced with an RNN compared to an FFNN [17]. BPNNs have garnered considerable attention among other ANN-related techniques due to their superior non-linear mapping function. An LSTM model, incorporating a clearness index method and k-means classification of the weather, has been introduced to enhance prediction accuracy on overcast days [18]. Furthermore, a framework for partial daily pattern prediction (PDPP) has recently been proposed to improve time correction modification (TCM) in the LSTM-RNN model [19]. However, the primary drawbacks of ANNs are overfitting and the substantial amount of data needed for training, increasing the complexity of implementation and costs in reality.
Support vector regressors are a type of machine learning model that can establish nonlinear relationships. Solar power forecasts have been conducted using support vector regression (SVR), considering various weather factors such as cloud cover and sun exposure [20]. A hybrid technique combining genetic algorithms and an SVM has been developed to enhance accuracy compared to a simplified SVM [21]. However, these models rely on the predetermined nonlinear mapping and parameters, making it challenging to fully comprehend the underlying nonlinear relationship between the inputs and the intended values [22,23].
In the present study, various research gaps were identified. These gaps encompass the need for a deeper understanding of the GPR model’s ability to capture underlying patterns when faced with noisy or outlier-laden data. Additionally, the manual tweaking of hyperparameters in the Bayesian optimizer, recognized as a time-intensive process, constitutes another area for exploration. It presents a methodical approach to applying Bayes’s theorem in order to update observed data with the prior understanding [24,25]. The study also suggests investigating the model’s adaptability to diverse time intervals and delving into the evaluation of objective functions in the presence of noise. To overcome these challenges, this study contributes to the following areas.
  • Enhanced Forecasting Accuracy: The proposed GPR model with Bayesian optimization demonstrates a significant enhancement (98% and 90%) in forecasting accuracy compared to alternative models like the feed-forward neural network and ANN. This highlights a substantial contribution to improving prediction outcomes.
  • Robustness and Flexibility: The integration of GPR with Bayesian optimization contributes to the model’s robustness and flexibility, especially in handling noisy and limited data. This adaptability ensures that the model remains relevant and useful in evolving environments.
  • Time and Resource Savings: Bayesian optimization is noted for its efficiency in systematically exploring the hyperparameter space, leading to potential time and resource savings. This contribution is particularly relevant in scenarios involving complex models and large data sets.
  • Handling Multivariate Analysis: The suitability of GPR for multivariate analysis is highlighted, showcasing its effectiveness in handling many input variables. This contribution is essential for addressing the complexities of associations between predictor variables and the target variable, such as in solar energy forecasting.
The subsequent sections of this paper are presented as follows: A literature review is provided in Section 2; Data and data preparation are described in Section 3 and Section 4, respectively; An overview of the Bayesian optimizer is discussed in Section 5. The proposed methodology is presented in Section 6. Results and analysis with a comparison are discussed in Section 7 and Section 8 summarizes the conclusions and future work.
Figure 1. World solar resources map [26].
Figure 1. World solar resources map [26].
Processes 12 00546 g001

2. Literature Review

There are many forecasting algorithms presented in the literature on solar power forecasting. Dhillon et al. [27] compared three models (feed-forward neural networks, backpropagation neural networks, and model-averaged neural networks) for forecasting solar energy in the field of precision agriculture to predict power for wireless sensor networks (WSNs). The model was developed using data from NREL’s (the National Renewable Laboratory’s) four-month period. Eight variables, including pressure, temperature, relative humidity, the dew point, wind speed, the hour of the day, the zenith angle, and historical solar intensity readings, were utilized to predict GHI for the next 24 h. The outcome showed that a feed-forward neural network provided the best result with the least memory needed (20 KBytes), and the values of the RMSE and the coefficient of correlation ( R 2 ) were, respectively, 98.052 and 56.61. Elizabeth et al. [2] proposed a unique multi-step CNN-stacked LSTM model with a dropout layer for the prediction of short-term solar irradiance and POA irradiance. The actual solar data were obtained from the Abu Dhabi-based Sweihan Photovoltaic Independent Power Project. The proposed model offered the best RMSE and R 2 values of 0.36 and 0.98 for solar irradiance prediction and 61.24 with R 2 0.96 for POA prediction, which also showed better performance compared to ML techniques such as LR, SVR, and ANNs with the same data set. To confirm the suggested architecture’s robustness, noise-injected data were also tested.
Khan et al. [28] suggested that the most effective forecasting model for PV power output is recurrent neural networks. The data were provided from Quaid-e-Azam Solar Park in Bahawalpur, Pakistan, a 100 MW solar power plant. The study concluded that the bidirectional long short-term memory (LSTM) RNN framework outperformed the SARIMA model under all weather conditions (sunny, cloudy, rainy, partially cloudy, dusty, and foggy), particularly under cloudy weather conditions where the root mean square error (RMSE) was found to be the lowest at 0.0025, R 2 was at 0.99, and the coefficient of variation of the root mean square error (CV-RMSE) was observed at 0.0095%. C.-H. Liu et al. [13] proposed an ML methodology architecture based on LSTM and MLP with different train data set sizes from two PV sites, KHH (Kaohsiung) and MFU (Thailand), respectively. The result of the machine learning operations for data processing, model fitting, cross-validation, metrics evaluation, and hyperparameters tweaking demonstrated that the suggested simplified LSTM model performed better than the MLP model. The average RMSE was 0.512, which is highly encouraging and suggests that the proposed technique and architecture will work well for applications that require short-term solar power forecasting.
Vanderstar et al. [29] proposed a method to anticipate two-hour-ahead solar irradiance levels using real-time solar irradiance measured both locally and at distant monitoring stations using an ANN at a site in Northwestern Alberta, Canada. The results of this study show that it is possible to use as few as five remote monitoring stations to achieve near-peak forecasting accuracy from the algorithm and that it is preferable to give the remote monitoring sites adequate geospatial separation around the target site, rather than clustering the sites in the strictly upwind directions. The RMSE was 10.8%, which is a comparatively low value.
Taoufik et al. [30] proposed a solar energy predictor for communicating sensors or SEPCS. This prediction model makes short-term predictions about future energy availability based on historical energy observations. The authors used a database that identified the evolution of solar radiation over the course of a year to evaluate the performance of the suggested algorithm. Then, a performance comparison revealed that the SEPCS predictor greatly outperformed the most advanced energy predictors by lowering the average prediction error from 28 to 6.5%.
Zang et al. [31] introduced two brand-new deep convolutional neural networks (CNNs), namely the Residual Network (ResNet) and the Dense Convolutional Network (DenseNet), as the main forecasting models. To build input feature maps for the two novel CNNs, a data preprocessing method involving historical PV power series, meteorological components, and numerical weather prediction was proposed. A meta-learning strategy based on a multi-loss-function network was introduced to train the two deep networks, ensuring high robustness of the extracted convolutional features with superb nonlinear representation abilities consisting of more than ten layers. The point forecasting mean absolute errors (MAEs) of the hybrid ResNet and DenseNet were 0.152 kW and 0.180 kW, respectively, according to a case study using a 5 kW capacity PV array. The coverage error rate of the two models in probabilistic forecasting was less than 5%, notably with the error rate of ResNet being less than 1%.
Perveen et al. [32] used fuzzy logic and neural networks to construct their short-term PV power forecasting technique using a 15-year data set across India. Regression models and intelligent models were compared using statistical measures. The suggested model was used to forecast short-term PV power under a variety of climatic circumstances. The Adaptive Neuro-Fuzzy Inference System (ANFIS) model outperformed other models, according to simulation findings, for PV power forecasting. The greatest root mean square error (RMSE) for the suggested model was 0.5% on foggy days.
Ahmed et al. [33] conducted a statistical comparison among seven brain models to ascertain the global solar radiation (GSR) in Qena, Egypt. In contrast to other models, the study showed that the artificial neural network (ANN) model produced good outcomes. Comparing models and assessing the importance of the anticipated data are both possible using the test of the mean values of the mean bias error (MBE), RMSE, mean percentage error (MPE), Nash and Sutcliffe efficiency (NSE), and R 2 . The hydraulic system’s calibration criteria included the NSE metric, which provides a better comparative outcome.
Asl et al. [34] introduced solar energy prediction in Dezful, Iran (positioned at 32.1648° N and 48.25° E) using a linear transferable Levenberg–Marquardt (LM) method based on a multi-layer perceptron neural network (MPL NN) and taking the input parameters as the relative humidity, the daily mean air temperature, evaporation, the soil temperature, the day of the year, daylight hours, and the wind speed. The MPL has three neurons in the first hidden layer and two in the second hidden layer with the sigmoid transferred function, which produces the best prediction outcomes. The findings of the testing data showed that the mean absolute percentage error (MAPE) equaled 6.08%, while R 2 equaled 99.03%, and the training data showed that the mean square error (MSE) equaled 0.0042 and the standard error of estimate (SEE) equaled 5.9278, which is suitable for estimating global solar radiation (GSR).
DURRANI et al. [35] proposed an irradiance forecasting system based on multiple feed-forward neural networks for predicting the PV yield. Five years’ worth of historical meteorological data were used to train the proposed irradiance forecast model’s neural networks. For Stuttgart, the mean absolute percentage inaccuracy of the global horizontal irradiance forecast was 3.4% on sunny days and 23% on gloomy days. Sunlight, cloudiness, and temperature have a greater impact on the Global Horizontal Irradiance (GHI) neural network forecast model than other inputs. The research concluded that this system’s PV power forecast surpassed that of the PV persistence power forecast model.
Al-Shamisi et al. [36] used the artificial neural network (ANN), multi-layer perceptron (MPL), and radial basis function (RBF) methodologies with the input parameters of sunshine hours, the maximum temperature, the mean relative humidity, and the mean wind speed to perform a research study in the UAE’s Al-Ain city (located at 24.1302° N and 55.8023° E). A 3-year sample of data (2005–2007) was used for testing, while the statistical values for the 10-year period (1995–2004) used to train the network were R 2 , the root mean square error (RMSE), and the mean bias error (MBE). The findings of 11 models with varied input layer parameters demonstrated that, in the majority of circumstances, the RBF technique was superior to the MPL technique.
Wang et al. [37] introduced a brand-new hybrid model with adaptive learning for forecasting solar intensity. It is an online forecasting technique, and as data volumes rise over time, the system’s effectiveness gets better. The UMASS Trace Repository data set, which captures solar intensity in watts/m2, and meteorological data regarding four weather parameters (temperature, humidity, the dew point, wind speed, and precipitation), was used to test the suggested technique. The authors claimed that, after six months, the average monthly RMSE fell from 28.38 to 14.72 W/m2.
Jensona et al. [38] worked with two data sets: solar radiation data (SoDa) and the National Center for Environmental Predictions (NCEP) to predict the average daily solar irradiance using two models (a radial basis function network (RBFN) and a backpropagation neural network (BPN)). Data from two years were used to train the network, while data from one year were utilized to test it. For the two data sets, 3.12 MJ/m2 and 3.212 MJ/m2, respectively, the RMSE values were shown to be lower for BPN.
Manjili et al. [39] established an adaptive system for day-ahead solar intensity forecasting using data from NREL. The proposed approach was distinctive in that it combined statistical and artificial intelligence techniques. Time, temperature, relative humidity, zenith angle, azimuth angle, total cloud cover, opaque cloud cover, and pressure were the eight weather characteristics that were fed into the model. The method’s root mean square error (RMSE) was determined to be 149.29 W/m2.
Zhang et al. [40] presented different deep convolutional neural networks (CNNs) using high-resolution weather forecast data, investigating various temporal and geographical connectivities to capture the cloud movement pattern and its impact on forecasting solar energy generation for solar farms. It was able to lower the error rate from the persistent model’s 21% to 15.1% from the support vector regression (SVR) models and 11.8% from the convolutional neural networks compared to the state-of-the-art forecast error rate. Due to the use of high-resolution weather data in the training data set, the research found that deep learning models performed better in predicting solar energy generation than the earlier physical models.
Haixiang et al. [41] proposed the hybrid method of variational mode decomposition (VMD) and a deep convolutional neural network (CNN) with multiple input factors, which was capable of enhancing the accuracy of short-term PV power forecasting. Compared to other 1D VMD-based forecasting approaches, the model’s prediction accuracy was higher because it was trained on correlations in both hours and days. When developing new CNN models, transfer learning is crucial. The study came to the conclusion that deep learning models beat conventional physical models when it comes to predicting solar energy outputs.
Lee et al. [42] proposed a novel deep neural network that can be trained to estimate the amount of solar energy that would be produced the next day using time series data gathered from photovoltaic inverters and national weather centers. The study refers to integrating two CNNs with filters of various sizes in order to effectively identify local short-term characteristics and efficiently capture long-term features. With roughly estimated weather data, the suggested method accurately predicts solar power and operates reliably without sophisticated preprocessing to weed out outliers. The network surpasses a number of conventional regressors as well as a cutting-edge deep neural network-based solar power prediction algorithm.
Table 1 summarizes insights from various works in the literature on key research aspects. It provided a vital foundation for our planned investigation by pointing out patterns and gaps in the literature. Recognizing fundamental limitations, it became our starting point for our research, shaping our contributions. Through this synthesis, we aimed to build upon existing knowledge and address identified gaps in the current state of research.

3. Data Description

This study utilizes meteorological data obtained from Solcast (Solar Api and Weather Forecasting Tool) in the geographical coordinates of Canberra, Australia ( 35 . 2802 S and 149 . 1310 E, as shown in Figure 2), in order to create the proposed model ( In order to evaluate the model’s performance, it was subjected to testing using two distinct data sets originating from an identical geographical location. The data sets encompassed the period spanning from 1 September 2021 to 1 September 2022, with the temporal resolution of 60 min and 30 min. The analysis utilized a data set consisting of 8785 samples for a 60-min interval and 17,615 samples for a 30-min interval, which was partitioned into two distinct subsets.
For the 60-min-interval data set, the training data set comprised a total of 7057 samples, but the testing data set had only 1757 samples, and for the 30-min-interval data set, the training data set comprised a total of 14,091 samples, but the testing data set had only 3524 samples available for use in judging the accuracy of the models. Every single data sample included eight different characteristics, which were as follows: the ambient temperature, cloud capacity, dew point, snow depth, pressure, wind velocity, zenith, and global horizontal irradiance (GHI). The aforementioned elements exert a substantial influence on the precision of the GHI forecast. The meteorological observations are linked to the global horizontal irradiance and impact the prediction of the GHI. Figure 3 and Figure 4 present plots of the various input variables, together with the GHI values for both 30-min and 60-min intervals, respectively. The plots were generated using the mean value for each 24-h cycle in order to improve comprehension.

4. Data Preparation

4.1. Data Scaling

Data scaling is an important step in developing a solar forecasting model because solar irradiance data have a large number of variables, and since they have so many dimensions, they must be scaled before being input into an ML network. The process of rescaling data from their original range so that all values fall between [0 and 1] or [−1 and 1] is known as data scaling. It is required to know or precisely estimate the minimum and maximum sampled values, which can be calculated using available recorded data. Data scaling is based on (1).
So, each signal was scaled between 0 and 1 or −1 and 1, corresponding to the minimum and maximum values obtained across the whole data set. In this context, the min–max scaler was employed for the purpose of data scaling.
x s c a l e n = x n x m a x x m i n
where the following points apply:
  • x n : Specific sample.
  • x s c a l e n : New value of x n .
  • x m a x : Highest value in the feature column.
  • x m i n : Lowest value in the feature column.

4.2. Feature Extraction

The process of selecting or transforming raw data into a set of useful and informative characteristics that can be utilized as inputs for predictive models is referred to as feature extraction. To extract the best features from the data set, the heat map was used to highlight signal correlation. The heat map can ensure a high correlation between the signal for prediction and other important system indications. Figure 5 and Figure 6 illustrate the heat map that was generated from the data sets utilized in this research with the 60-min interval and 30-min interval, respectively. Here, the ambient temperature, cloud opacity, wind velocity, and zenith showed a high correlation with GHI for both data sets. Since the other variables did not exhibit a convenient correlation score with GHI, they were not utilized to train the suggested or comparison models. In Figure 5 and Figure 6 the following abbreviations are used: AT—ambient temperature, CO—cloud opacity, DP—dew point, SD—snow depth, WV—wind velocity, Pr—pressure, Zn—zenith, and GHI—global horizontal irradiance.

5. An Overview of Bayesian Optimizer

Hyperparameters play a crucial role in ML algorithms since they directly dictate the behavior of training algorithms and significantly impact the performance of the resulting models. Various techniques have been developed and successfully applied in specific domains, although their implementation often requires specialized knowledge and expert experience. In some cases, brute-force search methods are employed due to the complexity of the problem. Consequently, the development of an efficient hyperparameter optimization algorithm capable of optimizing any given machine learning method would greatly enhance the overall efficiency of the learning process. This research aims to establish a relationship between the performance of ML models and their respective hyperparameters through the utilization of Gaussian processes. By doing so, the problem of hyperparameter tuning can be abstracted as an optimization problem, and Bayesian optimization is employed to address it. Bayesian optimization operates on the principles of the Bayesian theorem, in which a prior distribution is established over the optimization function, and information from previous samples is used to update the posterior distribution of the function [43]. A utility function is then employed to select the next sample point that maximizes the optimization function. It consists of three fundamental components: a search space that encompasses potential parameter samples, an objective function, and a surrogate. It constructs a probabilistic model of the objective function, utilizing this model to identify the most auspicious hyperparameters for assessment against the genuine objective function.
x = a r g max x ϵ A f x
The above equation reflects the core concept of Bayesian optimization. Here, A′ represents x’s search space. The fundamental purpose of Bayesian optimization is to integrate the a priori distribution of the function f(x) with the empirical evidence in order to derive the posterior distribution of said function. Subsequently, this posterior information is utilized to identify the optimal location of the function f(x) based on a specified criterion. This criterion is enacted via a utility function, u, which is also known as the acquisition function.
P ( M | V ) P ( V | M ) P ( M )
The core idea of Bayesian optimization is deduced from Bayes’s theorem. For instance, if V is the presented evidence data, the probability of posterior function P(M|V) for model M is equal to the product of the prior probability P(M) and P(V|M). Here, P(V|M) explains how well the model explains the observed data, in which (V) is the observing data for a given particular model, M.
x t = a r g m a x x u x | D 1 : t 1
y t = f x t
For t = 1 , 2 , 3 , , x t can be found optimizing the acquisition function u using function f. D 1 : t 1 = { x i , y i } i = 1 t 1 denotes the training data set consisting of t 1 samplings of function f. Then, observe the objective function y t = f ( x t ) . This process is repeated until either the maximum number of iterations is reached or the difference between the current value and the ideal value acquired thus far is less than a predetermined threshold. It can be noted that, compared to other optimization techniques, Bayesian optimization does not require the explicit definition of function f.

6. Methodology

6.1. Gaussian Process Regression

Gaussian process regression (GPR) is a machine learning technique based on Bayesian learning theory. But GPR is quite different from other ML models. Where other models provide a single predicted value, GPR provides a probability distribution over a probable target value for each input data point for which a level of uncertainty can be considered in the model. Through the Gaussian process, the underlying connection between the input features and the target is modeled via GPR using a bunch of random variable s. In this model, the probable functions each represent a mapping from the input feature to the target feature, and the Gaussian process specifies a distribution over functions. The distribution over functions is accomplished using a mean and covariance function (kernel function). The mean function signifies the expected behavior of the functions in the Gaussian process. It creates a baseline for making predictions. The mean function is chosen based on the given problem and the preceding data set. The covariance function is used to compare the input points and find the measure of similarity between them. It takes two input points and finds the correlation between the specified input points. Different covariance functions can be used to identify various relationships in the data. Since the data set used here is continuous and dense, shows a bit of noise, and shows a smooth relationship between the input features and target features, the radial basis function (RBF) is used. The RBF kernel assumes the decreasing correlation of input points with the increase of distance between the input points. The model finds the base prediction and correlation of input points to help the distribution over the target value. Since GPR can consider multiple functions, and each function is considered a map from the input feature to the target feature, the covariance of the input points helps the model decide how much ripple can be considered for the prediction. This ripple allows the consideration of uncertainty in the prediction, which provides an upper hand over other ML models. This unique feature often gives this model an upper hand over neural networks as well.
The basic form of the GPR equation is
y i = f x + ε
where the following points apply:
  • y is the observed PV output variable.
  • x is the input variable.
  • f(x) represents the unknown function that relates the input and output variables.
  • ε is the noise term assumed to be normally distributed with a mean of zero and variance σ n 2 .
To predict the future PV output f * , considering the weighted function and noise with a Bayesian linear model, the distribution can be shown as
p ( f * | x , X , y ) = p ( f * | x , ω ) p ( ω | X , y ) d ω

6.2. Prior Distribution

Prior distribution in GPR represents the assumption about a target variable before sampling any data with a mean function and covariance function illustrated in Figure 7a. Mathematically, it is denoted with
P ( f ) = N ( m ( x ) , K ( x , x ) )
where the following points apply:
  • P ( f ) denotes the prior distribution for the target variable f.
  • N represents the multivariate Gaussian distribution.
  • m ( x ) is a mean function that can be set as either constant or a function, depending on the input variable x.
  • K ( x , x ) is the covariance or kernel function that measures covariance between different input functions.
Figure 7. (a) Prior samples, (b) Posterior distribution with observations. Solid line: mean forecast; dashed lines: samples. Shaded zone: twice the standard deviation at each input value of x.
Figure 7. (a) Prior samples, (b) Posterior distribution with observations. Solid line: mean forecast; dashed lines: samples. Shaded zone: twice the standard deviation at each input value of x.
Processes 12 00546 g007

6.3. Posterior Distribution

The posterior distribution in GPR is obtained after seeing the training data. It takes into account both the previous distribution and the actual data, allowing us to make predictions and estimate uncertainty. This is illustrated in Figure 7b. The posterior distribution is also a Gaussian process, and it can be expressed as follows:
P ( f | X , y ) = N ( m * ( x ) , K * ( x , x ) )
where the following points apply:
  • P ( f | X , y ) denotes the posterior distribution over the target variable f, given the input data X and target values y.
  • N represents the multivariate Gaussian distribution.
  • m * ( x ) is the posterior mean function that predicts the value of f at input x based on the observed data.
  • K * ( x , x ) represents the posterior covariance function, which measures the uncertainty in the predictions.
The posterior distribution enables predictions for fresh input values to be made by sampling from the distribution or computing the mean or median estimate. The posterior covariance function may be used to quantify the uncertainty in the predictions.

6.4. Equations for Mean and Covariance Prediction

The predictive mean m * can be calculated as follows: m * = K ( X * , X ) · [ k ( X , X ) ] + σ 2 I ] 1 · y where σ 2 is the noise variance, and I is the identity matrix.
The mathematical from of covariance can be represented as
k * = K ( X * , X ) K ( X * , X ) · [ k ( X , X ) + σ 2 I ] 1 · K ( X * , X )

6.5. Radial Basis Function (RBF) Kernel

The RBF kernel, also called the squared exponential kernel, is commonly used in GPR because it can capture smooth and continuous interactions between inputs and the target variable. It is also a positive definite kernel; therefore, the resultant covariance matrix is a positive semidefinite. The hyperparameters of the RBF kernel (i.e., σ 2 and I) of the GPR model are learned from data during training using a maximum likelihood estimation or other optimization approaches. The GPR model adjusts to the peculiarities of the individual data set and captures the underlying patterns in the data by maximizing these hyperparameters. The RBF kernel can be mathematically expressed as follows:
K ( x , x ) = σ 2 exp 1 2 ( x x l ) 2
where the following points apply:
  • K ( x , x ) is the covariance between input variables x and x .
  • σ 2 is the variance parameter.
  • ( x x ) is the Euclidean distance between x and x .
  • l is the lengthscale parameter that determines the smoothness and length scale of the covariance function.
The integration of Bayesian optimization techniques with GPR architecture allows for the combination of the versatile Gaussian process models with Bayesian optimization methods. The employed approach utilizes a probabilistic model, commonly incorporating the RBF kernel, in order to effectively capture intricate correlations within the data set. Bayesian optimization serves as a guiding framework for the hyperparameter tweaking of the model, enabling it to efficiently adapt and enhance its predictive capabilities. Initially, raw weather data undergo preprocessing, followed by feature extraction, to prepare them for the GPR model. During model training, the Bayesian optimizer selects informative data points to query, maximizing the acquisition function and improving model performance. After each iteration, the model’s performance is evaluated using a testing set, and adjustments to the model parameters and data points are made accordingly. The detailed architecture is displayed in Figure 8. The utilization of GPR in conjunction with Bayesian optimization presents a formidable approach for the modeling and optimization of intricate functions that are characterized by uncertain or noisy data.

7. Result and Analysis

The primary objectives of this work were to analyze the performance of various ML-based forecasting techniques, such as LSTM, GPR, and a proposed GPR with a Bayesian optimizer, to predict the GHI. Moreover, some conventional machine learning models were also evaluated and compared with the proposed techniques. To assess the feasibility of this study, two different solar data sets with different sampling times from Canberra, Australia, provided by Solcast, were used. Forecasting performance was measured using various performance evaluation metrics. The data were divided into a percentages of 70–30 for the training set and the testing set. Simulations and programming were conducted in a Python environment using Google Colabratory.

7.1. Performance Evaluation Matrices

As discussed above, the following performance measuring metrics were used in this study to assess the performance of the forecasted model: the root mean square error (RMSE), mean square error (MSE), the determination of correlation coefficient ( R 2 ), and the mean absolute error (MAE). These metrics were considered to evaluate the model’s performance. Lower values of these metrics indicate better predictive results. In the following discussion, n represents the total number of samples, Y act represents the actual values of forecasting variables, and Y pred represents the predicted solar power.
(a) Root mean square error (RMSE): The RMSE is the square root of the average of the squared differences between actual vs. predicted values:
R M S E = 1 n i = 1 n ( Y act Y pred ) 2
(b) Coefficient of determination ( R 2 ): R 2 denotes the percentage of variance in actual values that can be explained using expected values. It has a scale from 0 to 1, with 1 indicating a perfect fit:
R 2 = 1 i = 1 n ( Y act Y pred ) 2 i = 1 n ( Y act Y mean ) 2
where Y mean is the mean of actual solar power.
(c) Mean square error (MSE): It is the average of the squared differences between the predicted and actual values:
M S E = 1 n i = 1 n ( Y act Y pred ) 2
(d) Mean absolute error (MAE): The absolute value of the difference between the actual value and predicted values:
M A E = 1 n i = 1 n | Y act Y pred |
(e) Average absolute error (AAE): AAE calculates the average absolute difference between each predicted and actual value:
AAE = 1 n i = 1 n Y pred Y act Y act × 100 %
(f) Average biased error (ABE): ABE calculates the average difference between each predicted and actual value:
ABE = 1 n i = 1 n Y pred Y act Y act × 100 %
In this section, the collected solar irradiance data are used to test the performance of the proposed methodology. This study employed a multivariate analysis approach, utilizing two distinct data sets that exhibited varying time intervals between each sample. One of the experimental conditions involved a 60-min gap between each sample, whereas the other condition involved a 30-min interval. Under both conditions, we used four columns as input variables to estimate GHI. The comparison between the observed solar irradiance values and the anticipated values is illustrated in figures for seven different models: RF, MLP, SVR, regression, LSTM, GPR, and the suggested GPR with a Bayesian optimizer. It provides insight into the performance of all the methods discussed so far. The test performance accuracy and flexibility are discussed in Table 2 using six error matrices (RMSE, MSE, MAE, R 2 , AAE, and ABE), as described previously [44,45]. The value of the R 2 closeness to 1 indicates a more accurate model. The RF, MLP, SVR, and regression models all performed well in terms of computing time, finishing their calculations in less than 60 s. On the other hand, the GPR model required a significantly longer processing time—roughly 368 s on average. In a comparable manner, the suggested model and LSTM model showed higher computing demands, averaging 406 and 476 s, respectively. The aforementioned results highlight the greater computational load linked to GPR, LSTM, and the proposed model in contrast to RF, MLP, SVR, and conventional regression techniques. From Table 2, it can be seen that the GPR model, LSTM model, and GPR with a Bayesian optimizer outperformed all the other models. The GPR model for the 60-min interval performed the prediction task with the following error metrics: RMSE = 1.80, MSE = 3.24, MAE = 1.06, R 2 = 0.97, AAE = 0.279%, and ABE = 0.018%. And for the 30-min interval, the obtained error metrics were as follows: RMSE = 1.89, MSE = 3.57, MAE = 1.13, R 2 = 0.96, AAE = 0.576%, and ABE = −0.549%. However, the LSTM architecture performed better than GPR for both time intervals with the following error metrics: RMSE = 1.28, MSE = 1.64, MAE = 0.94, R 2 = 0.99, AAE = 0.706%, and ABE = −0.509% for the 60-min interval and RMSE = 1.17, MSE = 1.37, MAE = 1.15, R 2 = 0.99, AAE = 0.619%, and ABE = −0.615% for the 30-min interval. Lastly, it can be observed that the proposed GPR with a Bayesian optimizer model outperformed other previously stated deep learning and ML models for all performance matrices. The performance results are noted as RMSE = 1.09, MSE = 1.18, MAE = 0.49, and R 2 = 0.99 for the 60-min interval, and for the 30-min interval, the obtained performance metrics were RMSE = 1.12, MSE = 1.25, MAE = 1.25, and R 2 = 0.99. The advantages of GPR, LSTM, and the suggested model exceeded the disadvantages even though their computing times were longer. On the other hand, even if certain models have quicker computation times, their poor efficiency makes them inappropriate for practical usage. This emphasizes the trade-off between predicted accuracy and computing efficiency, with the latter being more crucial when choosing a model.
It can be shown that, of the seven models utilized under both the 30-min interval and the 60-min interval circumstances, GPR with a Bayesian optimizer achieved the best performance measure scores. The performance metrics of each tested model are presented in Table 2, and the used model parameters and optimal values are shown in Table 3. The best-performing hyperparameters for LSTM and GPR were chosen through manual testing. On the other hand, a Bayesian optimizer was employed to find the optimal parameters for our proposed model. A predefined range was provided to the Bayesian optimizer to explore it in order to identify the optimal hyperparameters. The optimizer successfully suggested the optimal hyperparameters to optimize model performance. This approach improved the predictive accuracy and automated hyperparameter selection.

7.2. Comparative Analysis

Here, a comprehensive analysis is carried out on each model that underwent testing. Furthermore, the outcomes derived from our data sets are shown. Each graph in this section consists of the first 100 samples that were obtained from our test data.
The figures indicate that the GPR model achieved comparable performance for both data sets depicted in Figure 9a for a 60-min interval and Figure 10a for a 30-min period. Despite being the least efficient among the three models, GPR exhibited acceptable performance in terms of usability. Due to the minimal difference between the actual and anticipated data, identifying any distinctions in the plots became challenging. The difference between the observed and forecasted values is shown in Figure 9b and Figure 10b, representing the 60-min and 30-min intervals, respectively. While the majority of the observed data points exhibited differences of less than 2 for 60-min intervals and 1.5 for 30-min intervals, it is worth noting that a significant number of samples displayed larger disparities. Some of them almost achieved as high a difference as 5 (30 min) and 3 (60 min).
The similarity between the LSTM model and the GPR model was also noted. The LSTM model exhibited superior performance compared to the GPR model, with an R 2 score of 0.99 in both instances. This demonstrates that the model performed flawlessly. The data seen and anticipated for various time intervals are depicted in Figure 11a and Figure 12a. In a similar manner, the distinctions between the observed and anticipated values are graphically shown for each sample in Figure 11b and Figure 12b.
In Figure 11b, it is observed that, even though the LSTM model kept the difference low for the 60-min-interval data set, it failed to maintain the same performance for the 30-min-interval data set.
The GPR model with a Bayesian optimizer demonstrated superior performance compared to the other two models, which is clearly indicated in the graphical representations. The comparison between the observed and predicted results for the 60-min and 30-min intervals using GPR with a Bayesian optimizer is illustrated in Figure 13a and Figure 14a, respectively. Given the model’s impressive performance metrics, the figures presented in Figure 13b and Figure 14b illustrate the differences between the actual and predicted values for data sets with 60-min and 30-min intervals, respectively. In these figures, it can be observed that the model predicted very close values to the observed data. Here, it is also noticeable that the model worked better for the 60-min-interval data set compared to the 30-min interval data set. Figure 13c and Figure 14c present the zoomed-in version of the observed and predicted values for data sets with 60-min and 30-min intervals, respectively.
Here, the proposed methodology is compared with other models from the literature review, as detailed in Table 3.

7.3. Discussions

The comparative advantage of GPR with Bayesian optimization over alternative neural network models, such as long short-term memory (LSTM) and other used models in the context of this forecasting task, can be related to a number of crucial variables. Bayesian optimization is a highly effective method for optimizing hyperparameters, particularly in the context of GPR models. This approach addresses the natural challenge of tuning hyperparameters in neural networks that possess a large number of such parameters. Additionally, the built-in characteristics of GPR, such as its ability to resist overfitting and provide uncertainty estimates, contribute to its robustness as a suitable option for forecasting applications. Moreover, the capability of GPR to effectively handle multivariate data, along with the systematic method of Bayesian optimization, significantly contributes to its effectiveness in this particular situation. The selection between GPR and neural networks is dependent upon the unique attributes of the job, the quantity of the data set, and the modeling prerequisites. In this regard, GPR presents itself as a persuasive alternative for accurate prediction, and in our case, it outperformed neural network models like LSTM.
In Figure 15 and Figure 16, all three models’ prediction and the actual data are displayed utilizing line charts for 60-min interval data and 30-min interval data, respectively. To understand the actual value and predicted value of the models, a box plot of 60-min-interval data and 30-min-interval data is also included in the figures. Here, for the box plot, all the test data and their corresponding predicted data were used.

7.3.1. Gaussian Process Regression (GPR)

The GPR model is a model based on probability that presumes a Gaussian process to describe the connection between input and output data. The present work employed GPR with the utilization of the RBF kernel, which includes hyperparameters such as the length scale and alpha. The initial configuration for GPR entailed the selection of the RBF kernel with certain hyperparameters, namely a length scale of 1 and an alpha value of 0.01. The number of optimizer restarts (five) was selected to ensure that the model underwent optimization iterations in order to refine the hyperparameters.
The strength resides in its capacity to effectively represent complex linkages and uncertainties. The selection of the kernel and hyperparameters has a significant impact on the model’s ability to accurately fit the given data and effectively generalize to unseen samples. However, when considering the used data sets, it was observed that the link between the inputs and outputs was intricate, resulting in difficulties for the GPR model in accurately capturing it.
Another crucial aspect to consider is the process of hyperparameter tweaking. The performance of the GPR model is heavily influenced by its hyperparameters. When the Bayesian optimization process is not employed to tune the GPR model, it may not be optimized with the most suitable hyperparameters. Consequently, this can result in inferior performance metrics compared to the other three models that were tested. The GPR model requires effective feature engineering; yet, in our particular scenario, the data exhibited complexity, and the inputs shown limited the correlation with the output. Despite the model’s somewhat lower performance among the three, it nonetheless exhibited favorable performance metrics, indicating its usability.

7.3.2. Long Short-Term Memory (LSTM) Networks

LSTM is a recurrent neural network (RNN) that is well suited to managing sequential data. As our data set contained time-series data, LSTM effectively incorporated these dependencies, which may not have been possible for the other models. LSTM is a deep learning model that can automatically acquire hierarchical, complex features from data. It can model complex relationships between input characteristics and the objective variable. As a probabilistic model, GPR may not be able to capture such complex patterns. LSTM can simultaneously learn feature representations and temporal dependencies, which is advantageous when the input–output relationships are not expressly known and must be learned from the data. In addition, LSTM allows for parameter tuning during training, which improves model performance. LSTM can acquire relevant features automatically from the data, reducing the need for manual feature engineering. In coping with our complex data sets, this was a significant benefit. Because of these benefits, LSTM outperformed GPR and demonstrated superior performance metrics. The result is notable when observed in the plots shown in the previous section.

7.3.3. GPR with Bayesian Optimizer

GPR with a Bayesian optimizer performed the best among the three models used for analysis, while GPR performed marginally worse than LSTM and the proposed model. The Bayesian optimizer is the deciding factor in this case. The results can be seen in the case of both data sets in the observed and predicted data given below for the 60-min interval and the 30-min interval, respectively, in Figure 15 and Figure 16.
Bayesian optimization represents a sophisticated and effective approach to systematically investigating the hyperparameter space of ML models. In contrast to computationally expensive methods such as brute force, grid search, or random search, Bayesian optimization employs a proactive strategy by constantly selecting hyperparameter combinations that are more likely to produce improved outcomes. The previously discussed efficiency has the potential to result in substantial time and resource savings, particularly in situations involving complex models and large data sets. It also allows the model to adapt to the specific characteristics of the data. The method begins by establishing an initial collection of hyperparameters and afterwards iteratively improving them to correspond with the observed performance of the model. The capacity to adapt is of the highest priority when working with data sets that exhibit diverse time intervals, as it enables the model to flexibly change its behavior in order to accurately capture the underlying patterns. Bayesian optimization automates the process of hyperparameter tuning. It intelligently chooses the next set of hyperparameters to evaluate based on past observations, guided by an acquisition function. This automation simplifies the model development process and reduces the need for manual trial-and-error tuning.
GPR is a highly suitable method for conducting multivariate analysis due to its ability to successfully handle many input variables. In the present study, the input factors were demonstrative of a complex association between the predictor variables and the target variable, which is the GHI. GPR is particularly suited to collecting data on complex and non-linear interactions like these. The model is probabilistic in nature, so it includes both predictable predictions and the ability to measure the level of uncertainty associated with these predictions. The ability to evaluate the certainty of a model’s predictions is of significant importance in the context of forecasting work. The utilization of Bayesian optimization, when combined with GPR, enables the optimization of hyperparameters by taking into account both the accuracy of predictions and the associated uncertainty. This approach ultimately results in the development of robust models. The intrinsic smoothing capabilities of GPR contribute to its robustness against overfitting. The attribute of robustness holds significant importance in the field of forecasting, as highly complex models have a tendency to exhibit poor performance when applied to data that have not been previously seen. The integration of Bayesian optimization can provide an additional means of tackling the problem of overfitting by simplifying the selection of hyperparameters that achieve a balance between the complexity of the model and its ability to accurately predict outcomes.
In summary, the integration of GPR with Bayesian optimization presents numerous benefits for the problem of prediction. The algorithm effectively examines the hyperparameter space, adjusts to diverse data patterns, facilitates multivariate analysis, estimates uncertainty, and exhibits robust performance. These qualities make it an ideal contender for handling complex and ever-changing information, such as that encountered in solar forecasting.

8. Conclusions and Future Work

This research has introduced a robust approach to predicting solar energy variables using the GPR model with a Bayesian optimizer. The proposed methodology achieved a significant improvement of 98% and 90%, respectively, in forecasting accuracy compared to the feed-forward neural network and ANN models, which were considered alternative models. A key advantage of integrating GPR with a Bayesian optimizer lies in its capacity to handle noisy and limited data.
The Bayesian optimizer’s active search for informative samples, guided by uncertainty estimates from the GPR model, enhances the exploitation of available data while mitigating noise effects. This dynamic approach results in improved model convergence and accuracy, offering robustness and adaptability to changing environments. The flexibility to update the GPR model with additional data ensures its relevance and usefulness as the underlying system evolves, particularly in the context of time-varying objective functions, for which sequential optimization can provide optimal solutions.
Future studies in the field of forecasting offer several promising avenues. One potential direction involves delving into hyper-parameter optimization and real-time data integration, exploring the handling capacity of noisy objective functions that present substantial opportunities for further exploration. Additionally, the application of these forecasting techniques in real-world scenarios, such as precision agriculture, could be a fruitful area for in-depth discussion. These possibilities pave the way for future research and advancements in the field.

Supplementary Materials

The supporting information can be downloaded at:

Author Contributions

Conceptualization, M.S.S.I., M.O.F. and M.R.I.; Methodology, P.G.; Validation, M.S.S.I. and M.O.F.; Formal analysis, M.O.F.; Investigation, M.R.I., M.A.H., M.S.A. and M.R.I.S.; Writing—original draft, M.S.S.I., P.G. and M.A.H.; Writing—review & editing, M.A.H., M.S.A. and M.R.I.S. All authors have read and agreed to the published version of the manuscript.


This research received no external funding.

Data Availability Statement

Data are contained within the article and Supplementary Materials.

Conflicts of Interest

The authors declare no conflicts of interest.


  1. Kumari, P.; Toshniwal, D. Deep learning models for solar irradiance forecasting: A comprehensive review. J. Clean. Prod. 2021, 318, 128566. [Google Scholar] [CrossRef]
  2. Elizabeth Michael, N.; Mishra, M.; Hasan, S.; Al-Durra, A. Short-term solar power predicting model based on multi-step CNN stacked LSTM technique. Energies 2022, 15, 2150. [Google Scholar] [CrossRef]
  3. Faruque, M.O.; Rabby, M.A.J.; Hossain, M.A.; Islam, M.R.; Rashid, M.M.U.; Muyeen, S. A comparative analysis to forecast carbon dioxide emissions. Energy Rep. 2022, 8, 8046–8060. [Google Scholar] [CrossRef]
  4. Islam, M.R.; Hasan, J.; Islam, M.R.; Kouzani, A.Z.; Mahmud, M.P. Transient Performance Augmentation of DFIG Based Wind Farms by Nonlinear Control of Flux-Coupling-Type Superconducting Fault Current Limiter. IEEE Trans. Appl. Supercond. 2021, 31, 1–5. [Google Scholar] [CrossRef]
  5. Abuella, M.; Chowdhury, B. Solar power probabilistic forecasting by using multiple linear regression analysis. In Proceedings of the SoutheastCon 2015, Fort Lauderdale, FL, USA, 9–12 April 2015; pp. 1–5. [Google Scholar]
  6. Shah, A.S.B.M.; Yokoyama, H.; Kakimoto, N. High-precision forecasting model of solar irradiance based on grid point value data analysis for an efficient photovoltaic system. IEEE Trans. Sustain. Energy 2015, 6, 474–481. [Google Scholar] [CrossRef]
  7. Aslam, M.; Lee, S.J.; Khang, S.H.; Hong, S. Two-stage attention over LSTM with Bayesian optimization for day-ahead solar power forecasting. IEEE Access 2021, 9, 107387–107398. [Google Scholar] [CrossRef]
  8. Bird, L.; Cochran, J.; Wang, X. Wind and Solar Energy Curtailment: Experience and Practices in the United States; Technical report; National Renewable Energy Lab. (NREL): Golden, CO, USA, 2014. [Google Scholar]
  9. Azim, M.I.; Islam, M.T.; Rakib, J.H.; Islam, M.R.; Ali, L.; Alzahrani, S.; Masrur, H.; Muyeen, S. Coalition game theoretic P2P trading in a distribution network integrity-ensured local energy market. Sustain. Energy Grids Netw. 2023, 36, 101186. [Google Scholar] [CrossRef]
  10. Kariniotakis, G. Renewable Energy Forecasting: From Models to Applications; Woodhead Publishing: Sawston, UK, 2017. [Google Scholar]
  11. Liu, J.; Vitelli, V.; Zio, E.; Seraoui, R. A novel dynamic-weighted probabilistic support vector regression-based ensemble for prognostics of time series data. IEEE Trans. Reliab. 2015, 64, 1203–1213. [Google Scholar] [CrossRef]
  12. Liu, J.; Fang, W.; Zhang, X.; Yang, C. An improved photovoltaic power forecasting model with the assistance of aerosol index data. IEEE Trans. Sustain. Energy 2015, 6, 434–442. [Google Scholar] [CrossRef]
  13. Liu, C.H.; Gu, J.C.; Yang, M.T. A simplified LSTM neural networks for one day-ahead solar power forecasting. IEEE Access 2021, 9, 17174–17195. [Google Scholar] [CrossRef]
  14. Donnelly, J.; Abolfathi, S.; Pearson, J.; Chatrabgoun, O.; Daneshkhah, A. Gaussian process emulation of spatio-temporal outputs of a 2D inland flood model. Water Res. 2022, 225, 119100. [Google Scholar] [CrossRef]
  15. Malki, H.A.; Karayiannis, N.B.; Balasubramanian, M. Short-term electric power load forecasting using feedforward neural networks. Expert Syst. 2004, 21, 157–167. [Google Scholar] [CrossRef]
  16. Gupta, M.; Jin, L.; Homma, N. Static and Dynamic Neural Networks: From Fundamentals to Advanced Theory; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
  17. Tomonobu, S. Application of recurrent neural network to short-term-ahead generating power forecasting for wind power generator. In Proceedings of the International Conference on Electrical Engineering (ICEE), YongPyong, Republic of Korea, 9–13 July 2006. [Google Scholar]
  18. Yu, Y.; Cao, J.; Zhu, J. An LSTM short-term solar irradiance forecasting under complicated weather conditions. IEEE Access 2019, 7, 145651–145666. [Google Scholar] [CrossRef]
  19. Wang, F.; Xuan, Z.; Zhen, Z.; Li, K.; Wang, T.; Shi, M. A day-ahead PV power forecasting method based on LSTM-RNN model and time correlation modification under partial daily pattern prediction framework. Energy Convers. Manag. 2020, 212, 112766. [Google Scholar] [CrossRef]
  20. VanDeventer, W.; Jamei, E.; Thirunavukkarasu, G.S.; Seyedmahmoudian, M.; Soon, T.K.; Horan, B.; Mekhilef, S.; Stojcevski, A. Short-term PV power forecasting using hybrid GASVM technique. Renew. Energy 2019, 140, 367–379. [Google Scholar] [CrossRef]
  21. Gigoni, L.; Betti, A.; Crisostomi, E.; Franco, A.; Tucci, M.; Bizzarri, F.; Mucci, D. Day-ahead hourly forecasting of power generation from photovoltaic plants. IEEE Trans. Sustain. Energy 2017, 9, 831–842. [Google Scholar] [CrossRef]
  22. Das, U.K.; Tey, K.S.; Seyedmahmoudian, M.; Mekhilef, S.; Idris, M.Y.I.; Van Deventer, W.; Horan, B.; Stojcevski, A. Forecasting of photovoltaic power generation and model optimization: A review. Renew. Sustain. Energy Rev. 2018, 81, 912–928. [Google Scholar] [CrossRef]
  23. Sheng, H.; Xiao, J.; Cheng, Y.; Ni, Q.; Wang, S. Short-term solar power forecasting based on weighted Gaussian process regression. IEEE Trans. Ind. Electron. 2017, 65, 300–308. [Google Scholar] [CrossRef]
  24. Oakley, J.E.; O’Hagan, A. Probabilistic sensitivity analysis of complex models: A Bayesian approach. J. R. Stat. Soc. Ser. Stat. Methodol. 2004, 66, 751–769. [Google Scholar] [CrossRef]
  25. Batsch, F.; Daneshkhah, A.; Cheah, M.; Kanarachos, S.; Baxendale, A. Performance Boundary Identification for the Evaluation of Automated Vehicles using Gaussian Process Classification. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 419–424. [Google Scholar] [CrossRef]
  26. The World Bank. Global Solar Atlas 2.0. Source: Global Solar Atlas 2.0, 2020. Solar Resource Data: Solargis. Available online: (accessed on 24 February 2024).
  27. Dhillon, S.; Madhu, C.; Kaur, D.; Singh, S. A solar energy forecast model using neural networks: Application for prediction of power for wireless sensor networks in precision agriculture. Wirel. Pers. Commun. 2020, 112, 2741–2760. [Google Scholar] [CrossRef]
  28. Khan, S.; Shaikh, F.; Siddiqui, M.M.; Hussain, T.; Kumar, L.; Nahar, A. Hourly forecasting of solar photovoltaic power in Pakistan using recurrent neural networks. Int. J. Photoenergy 2022, 2022, 7015818. [Google Scholar] [CrossRef]
  29. Vanderstar, G.; Musilek, P.; Nassif, A. Solar forecasting using remote solar monitoring stations and artificial neural networks. In Proceedings of the 2018 IEEE Canadian Conference on Electrical & Computer Engineering (CCECE), Quebec, QC, Canada, 13–16 May 2018; pp. 1–4. [Google Scholar]
  30. Bouguera, T.; Diouris, J.F.; Andrieux, G.; Chaillout, J.J.; Jaouadi, R. A novel solar energy predictor for communicating sensors. IET Commun. 2018, 12, 2145–2149. [Google Scholar] [CrossRef]
  31. Zang, H.; Cheng, L.; Ding, T.; Cheung, K.W.; Wei, Z.; Sun, G. Day-ahead photovoltaic power forecasting approach based on deep convolutional neural networks and meta learning. Int. J. Electr. Power Energy Syst. 2020, 118, 105790. [Google Scholar] [CrossRef]
  32. Perveen, G.; Rizwan, M.; Goel, N. Comparison of intelligent modelling techniques for forecasting solar energy and its application in solar PV based energy system. IET Energy Syst. Integr. 2019, 1, 34–51. [Google Scholar] [CrossRef]
  33. Ahmed, E.A. Statistical comparison between empirical models and artificial neural network method for global solar radiation at Qena, Egypt. J. Multidiscip. Eng. Sci. Technol. 2015, 2, 1899–1906. [Google Scholar]
  34. Asl, S.F.Z.; Karami, A.; Ashari, G.; Behrang, A.; Assareh, A.; Hedayat, N. Daily global solar radiation modeling using multi-layer perceptron (MLP) neural networks. Int. J. Environ. Ecol. Eng. 2011, 5, 403–405. [Google Scholar]
  35. Durrani, S.P.; Balluff, S.; Wurzer, L.; Krauter, S. Photovoltaic yield prediction using an irradiance forecast model based on multiple neural networks. J. Mod. Power Syst. Clean Energy 2018, 6, 255–267. [Google Scholar] [CrossRef]
  36. Al-Shamisi, M.; Assi, A.; Hejase, H. Estimation of global solar radiation using artificial neural networks in Abu Dhabi city, United Arab Emirates. J. Sol. Energy Eng. 2014, 136, 024502. [Google Scholar] [CrossRef]
  37. Wang, Y.; Shen, Y.; Mao, S.; Cao, G.; Nelms, R.M. Adaptive learning hybrid model for solar intensity forecasting. IEEE Trans. Ind. Inform. 2018, 14, 1635–1645. [Google Scholar] [CrossRef]
  38. Praynlin, E.; Jensona, J. Ida. Solar radiation forecasting using artificial neural network. In Proceedings of the 2017 Innovations in Power and Advanced Computing Technologies (i-PACT), Vellore, India, 21–22 April 2017; pp. 1–7. [Google Scholar]
  39. Manjili, Y.S.; Vega, R.; Jamshidi, M.M. Data-analytic-based adaptive solar energy forecasting framework. IEEE Syst. J. 2017, 12, 285–296. [Google Scholar] [CrossRef]
  40. Zhang, R.; Feng, M.; Zhang, W.; Lu, S.; Wang, F. Forecast of solar energy production-A deep learning approach. In Proceedings of the 2018 IEEE International Conference on Big Knowledge (ICBK), Singapore, 17–18 November 2018; pp. 73–82. [Google Scholar]
  41. Zang, H.; Cheng, L.; Ding, T.; Cheung, K.W.; Liang, Z.; Wei, Z.; Sun, G. Hybrid method for short-term photovoltaic power forecasting based on deep convolutional neural network. IET Gener. Transm. Distrib. 2018, 12, 4557–4567. [Google Scholar] [CrossRef]
  42. Lee, W.; Kim, K.; Park, J.; Kim, J.; Kim, Y. Forecasting solar power using long-short term memory and convolutional neural networks. IEEE Access 2018, 6, 73068–73080. [Google Scholar] [CrossRef]
  43. Wu, J.; Chen, X.Y.; Zhang, H.; Xiong, L.D.; Lei, H.; Deng, S.H. Hyperparameter optimization for machine learning models based on Bayesian optimization. J. Electron. Sci. Technol. 2019, 17, 26–40. [Google Scholar]
  44. Qian, X.; Lee, S.; Soto, A.m.; Chen, G. Regression model to predict the higher heating value of poultry waste from proximate analysis. Resources 2018, 7, 39. [Google Scholar] [CrossRef]
  45. Sheng, C.; Azevedo, J. Estimating the higher heating value of biomass fuels from basic analysis data. Biomass Bioenergy 2005, 28, 499–507. [Google Scholar] [CrossRef]
Figure 2. Solar resources map, Australia (SOLARGIS, 2020) [26].
Figure 2. Solar resources map, Australia (SOLARGIS, 2020) [26].
Processes 12 00546 g002
Figure 3. Data set from Solcast for 60-min interval.
Figure 3. Data set from Solcast for 60-min interval.
Processes 12 00546 g003
Figure 4. Data set from Solcast for 30-min interval.
Figure 4. Data set from Solcast for 30-min interval.
Processes 12 00546 g004
Figure 5. Input correlation analysis using heat map for 60-min interval.
Figure 5. Input correlation analysis using heat map for 60-min interval.
Processes 12 00546 g005
Figure 6. Input correlation analysis using heat map for 30-min interval.
Figure 6. Input correlation analysis using heat map for 30-min interval.
Processes 12 00546 g006
Figure 8. Architecture of GPR with Bayesian optimizer.
Figure 8. Architecture of GPR with Bayesian optimizer.
Processes 12 00546 g008
Figure 9. Observation for 60-min interval with GPR: (a) actual vs. predicted and (b) absolute difference.
Figure 9. Observation for 60-min interval with GPR: (a) actual vs. predicted and (b) absolute difference.
Processes 12 00546 g009
Figure 10. Observation for 30-min interval with GPR: (a) actual vs. predicted and (b) absolute difference.
Figure 10. Observation for 30-min interval with GPR: (a) actual vs. predicted and (b) absolute difference.
Processes 12 00546 g010
Figure 11. Observation for 60-min interval with LSTM: (a) actual vs. predicted and (b) absolute difference.
Figure 11. Observation for 60-min interval with LSTM: (a) actual vs. predicted and (b) absolute difference.
Processes 12 00546 g011
Figure 12. Observation for 30-min interval with LSTM: (a) actual vs. predicted and (b) absolute difference.
Figure 12. Observation for 30-min interval with LSTM: (a) actual vs. predicted and (b) absolute difference.
Processes 12 00546 g012
Figure 13. Observation for 60-min interval with GPR with a Bayesian optimizer: (a) actual vs. predicted, (b) absolute difference, and (c) zoomed-in version for samples 70–90 of the observed and predicted results.
Figure 13. Observation for 60-min interval with GPR with a Bayesian optimizer: (a) actual vs. predicted, (b) absolute difference, and (c) zoomed-in version for samples 70–90 of the observed and predicted results.
Processes 12 00546 g013
Figure 14. Observation for 30-min interval with GPR with a Bayesian optimizer: (a) actual vs. predicted, (b) absolute difference, and (c) zoomed-in version for samples 70–90 of the observed and predicted results.
Figure 14. Observation for 30-min interval with GPR with a Bayesian optimizer: (a) actual vs. predicted, (b) absolute difference, and (c) zoomed-in version for samples 70–90 of the observed and predicted results.
Processes 12 00546 g014
Figure 15. Comparison among the tested models for the 60-min interval (a) using a box plot and (b) using line charts.
Figure 15. Comparison among the tested models for the 60-min interval (a) using a box plot and (b) using line charts.
Processes 12 00546 g015
Figure 16. Comparison among the tested models for the 30-min interval (a) using a box plot and (b) using line charts.
Figure 16. Comparison among the tested models for the 30-min interval (a) using a box plot and (b) using line charts.
Processes 12 00546 g016
Table 1. Literature review summary.
Table 1. Literature review summary.
Reference No.Author NameGeneralizabilityComputational EfficiencyLong-Term ForecastingUncertainty AnalysisReal-World Application StudiesModel Robustness
[27]Dhillon et al.XXXX
[2]Elizabeth et al.XXXX
[28]Khan et al.XXX
[13]C.-H. Liu et al.XXX
[29]Vanderstar et al.XXX
[30]Taoufik et al.XXX
[31]Zang et al.XXXX
[32]Perveen et al.XXX
[33]Ahmed et al.XXXX
[34]Asl et al.XX
[35]DURRANI et al.XXX
[36]Al-Shamisi et al.XXX
[37]Wang et al.XXX
[38]Jensona et al.X
[40]Manjili et al.XXX
[41]Zhang et al.XXX
[42]Lee. et al.XXXX
This study
Table 2. Performance metrics for different models and data sets.
Table 2. Performance metrics for different models and data sets.
RF60-min. int.7.0349.453.690.993.52−0.03
RF30-min. int.8.2568.084.080.994.08−0.38
MLP60-min. int.13.72188.499.170.999.10−0.85
MLP30-min. int.23.52553.3115.970.9915.052.38
SVR60-min. int.40.281622.4639.020.8739.023.88
SVR30-min. int.49.652465.1248.760.8648.524.01
Regression60-min. int.68.734723.8165.380.8264.774.57
Regression30-min. int.74.655572.6072.980.8073.044.87
GPR60-min. int.1.803.241.060.970.270.01
GPR30-min. int.1.893.571.130.960.57−0.54
LSTM60-min. int.1.281.640.940.990.70−0.50
LSTM30-min. int.1.171.371.150.990.61−0.61
GPR with Bay. Opt.60-min. int.−0.03
GPR with Bay. Opt.30-min. int.
Table 3. Model parameters and optimal values (top) and comparison of proposed methodology with other models from the literature review (bottom).
Table 3. Model parameters and optimal values (top) and comparison of proposed methodology with other models from the literature review (bottom).
ModelParametersOptimal Value
LSTMNo. of layers3
LSTM units per layer32
Dense units16
Learning rate0.001
Batch size32
GPRKernelRBF (length scale = 1, length scale bounds 1 × 10−2 and 1 × 103)
Number of restarts for the optimizer5
GPR with Bayesian optimizerKernelRBF
Length scale bounds(1, 10)
Alpha bounds(1 × 10−5, 1)
Number of restarts for the optimizer5
Number of Iterations for Bayesian optimizer100
Number of initial points5
No.AuthorArchitecture usedRMSE R 2
[27]Dhillon et al.Feed-forward neural network56.610.98
[2]Elizabeth et al.CNN-stacked LSTM61.240.96
[29]Vanderstar et al.ANN10.80.912
[34]Asl et al.MLP6.080.9903
[35]DURRANI et al.Neural network8.60 (Clear sky)
37.06 (Partly cloudy)
38.52 (Cloudy)
Proposed modelGPR with Bayesian optimizer 1.09 0.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Islam, M.S.S.; Ghosh, P.; Faruque, M.O.; Islam, M.R.; Hossain, M.A.; Alam, M.S.; Islam Sheikh, M.R. Optimizing Short-Term Photovoltaic Power Forecasting: A Novel Approach with Gaussian Process Regression and Bayesian Hyperparameter Tuning. Processes 2024, 12, 546.

AMA Style

Islam MSS, Ghosh P, Faruque MO, Islam MR, Hossain MA, Alam MS, Islam Sheikh MR. Optimizing Short-Term Photovoltaic Power Forecasting: A Novel Approach with Gaussian Process Regression and Bayesian Hyperparameter Tuning. Processes. 2024; 12(3):546.

Chicago/Turabian Style

Islam, Md. Samin Safayat, Puja Ghosh, Md. Omer Faruque, Md. Rashidul Islam, Md. Alamgir Hossain, Md. Shafiul Alam, and Md. Rafiqul Islam Sheikh. 2024. "Optimizing Short-Term Photovoltaic Power Forecasting: A Novel Approach with Gaussian Process Regression and Bayesian Hyperparameter Tuning" Processes 12, no. 3: 546.

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop