1. Introduction
Solving the challenges posed by the massive integration of solar energy into electricity grids is a key issue for reducing the carbon footprint of power generation. Indeed, due to the inherent variability and lack of predictability of solar energy, a high share of solar energy in the electricity mix makes it more complicated to manage the supply-demand balance and increases the vulnerability of the grid. One of the strategies to reduce the effect of solar variability is to predict future solar irradiance and corresponding solar power for short-term horizons ranging from 1 min to several days in advance. Many techniques have been developed to predict solar irradiance [
1,
2,
3]. Numerical weather prediction (NWP) is suitable for horizons longer than 6 h. Forecasts derived from geostationary meteorological satellite images are effective for a horizon ranging from about 1 h to 6 h. It is also important to consider the target application to provide relevant forecasts. For instance, for common PV plants, one should forecast the global solar irradiance in the plane of the array, which includes both direct and diffuse components [
3]. For the particular case of CSP plants, which have a narrow view angle facing the sun [
4], reliable forecasts of the beam’s normal irradiance are required. Therefore, to forecast the solar energy for CSP plants and for very short-term horizons of less than 1 h, the approach based on all sky imagers (ASI) seems to be a promising technique [
5,
6].
Regarding the very short-term horizon, Ajith and Martínez-Ramón [
7] compared three categories of solar irradiance forecasting methods: time series, sky camera images, and hybrid models combining infrared images with radiation time series. The authors showed that the normalized root mean square error (nRMSE) varied from 30 to 53% in terms of forecasting error. In the literature on ASI, most of the works dealing with the prediction of solar irradiance or cloudiness propose deterministic forecasts [
8,
9,
10]. To improve forecasts derived from ASI, different approaches have been carried out. For instance, Paletta et al. [
11] evaluated deterministic and probabilistic predictions based on ASI for different weather conditions (clear, cloudy, and overcast skies). In their work, the probabilistic approach demonstrates a richer operational forecasting framework by facilitating uncertainty quantification in cloudy conditions and for long-term horizons. With the same aim, Nouri et al. [
12] used a statistical analysis of historical data to add uncertainty to ASI deterministic forecasts. Their original approach allows the creating of a probabilistic forecast of the DNI and GHI. However, very few recent methods were developed to generate probabilistic forecasts from sky camera images.
It is well-known that weather forecasts are uncertain because the evolution of the weather and consequently solar irradiance are chaotic processes. Thus, in decision-making operations that use solar forecasts, such as the management of power plants, probabilistic forecasts are crucial. Indeed, probabilistic forecasts assign probability levels to future events and allow their users to assess the associated risks. Research works dealing with probabilistic solar forecasts are relatively recent but numerous works have been released on the topic in the last 10 years [
3,
13,
14]. However, as mentioned previously, very few works concerning ASI proposed a method to generate a probabilistic solar forecast. This work will contribute to filling this gap in the literature.
Usually, solar forecasting systems provide the future level of solar irradiance or PV generation as a continuous variable [
3]. But, for particular applications, such as the management of concentrated solar plants (CSP) with tracking devices, the operator needs to anticipate the achievement of a solar irradiance threshold to start or to stop their system [
15]. In this case, an accurate binary forecast is more relevant. In the wide domains of meteorology or economy, numerous works propose discrete choice models to generate binary forecasts [
16,
17]. However, in the field of solar energy, very few works propose binary forecasts. One of the rare works on the topic is proposed by Alonso and Batlles [
18] who developed a method to forecast the cloudiness from a sequence of images given by the MeteoSat Second Generation (MSG) satellite MSG or an ASI. Their model generates deterministic binary forecasts of cloud presence (i.e.,
cloudy and
clear sky) corresponding to a direct normal irradiance (DNI) above 400 W/m
. The forecasts were tested over 2 years for the city of Almería in the south-east of Spain. The success rate of the forecasts derived from the ASI is 83% for the first 15 min and drops to 60% for a 3 h horizon. However, no work proposes a probabilistic approach to generate discrete solar forecasts. Indeed, the probabilistic approaches developed previously in the field of solar energy derive an irradiance level, also called quantile, from a fixed probability. As a consequence, they do not allow estimating the probability of reaching an arbitrary chosen irradiance threshold. Therefore, the main challenge of this work is to propose a methodology and suitable models able to precisely forecast the probability of crossing a defined irradiance level.
In light of the two main lacks of the literature underlined above, the main objective of this work is to propose a novel methodology to generate probabilistic solar forecast as a binary event for horizons ranging from 1 to 30 min using an all sky imager. The developed approach will combine a state-of-the-art ASI method and discrete choice models proposed in other domains, such as economy or meteorology. In a first step, a model based on the detection of cloud motions will use sequences of images from an ASI to generate binary deterministic forecasts of the cloudiness. This deterministic prediction is focused on the operation of concentrating solar thermal power plants, mainly central tower plants, since the appearance of clouds is a conditioning factor that can affect the operation and integrity of different components of the facilities, such as the central receivers [
19]. Then, in a second step, binary choice models will be used to convert the deterministic discrete forecasts into probability levels of cloud presence. Finally, we will assess the quality of the generated forecasts on a case study to evaluate the added value of the proposed method.
The remainder of the paper is organized as follows.
Section 2 presents the methodology used to develop and evaluate the proposed model.
Section 3 gives a brief overview of the state-of-the-art model used to generate the deterministic forecasts. Then,
Section 4 details the discrete choice models used to forecast the probability of cloud presence.
Section 5 depicts the case study and the corresponding data. Results are presented and discussed in
Section 6. Finally,
Section 7 gives our concluding remarks.
2. Overall Methodology and Forecasts Evaluation
The probabilistic forecasts are generated in two steps as presented in
Figure 1. First, we generated deterministic forecasts of cloud presence using the method proposed by [
18] and briefly presented in
Section 3. The results are discrete forecasts (1 = no cloud and 0 = presence of clouds) with a time resolution of 1 min and horizons of forecast up to 30 min. The second step is a post-processing of the deterministic forecasts with a probabilistic model. In this work, we compared three different models, described in
Section 4, to post-process the deterministic forecasts. The final probabilistic forecasts have the same temporal resolution and the same horizons as the deterministic forecasts. After these two steps, the generated forecasts are probabilities, in the interval
, that gives a level of confidence or risk associated with the future presence of clouds. Compared to deterministic forecasts, this additional information may help the user in decision-making.
To evaluate the forecasts done with the methodology described above, we propose to use a simple concept of detection of clouds that mask the sun based on DNI measurements and CSP operation. Thus, cloud detection was carried out following the methodology presented in [
20] obtaining a cloud identification (clouds which attenuate the DNI below 400 Wm
) based on the optimal operating value for CSP plants, as the case of the Gemasolar plant, which used this irradiance level, like the appropriate for producing electricity [
18].
In this work, both deterministic and probabilistic forecasts will be evaluated. If comprehensive frameworks have been proposed to evaluate forecast quality of the solar irradiance as a continuous variable [
21,
22,
23], no previous work details the evaluation of discrete solar forecasts. However, specific error metrics have been designed in the field of meteorology to assess the quality of binary forecasts. Let us recall that the quality of a forecasting system evaluates the agreement between the forecasts and the corresponding observations [
24]. Interested readers may refer to the web page published by the Joint Working Group on Forecast Verification Research to have an extended overview of weather forecast verification [
25].
Regarding binary deterministic forecasts, the most common metrics are derived from the confusion matrix presented in
Figure 2. In our case, a “yes” event corresponds to a clear sky (no clouds) and
N is the total number of observation/forecast pairs used for the verification. The confusion matrix is a useful tool for classifying the types of errors. A perfect forecast system would generate only hits and correct negatives, and no misses or false alarms. Numerous metrics are derived from the four cells in the confusion matrix, such as fraction correct, probability of detection (POD) or its opposite probability of false detection (PODF), success rate (SR), or false alarm ratio (FAR) [
26]. Each metric describes a different aspect of forecast performance. In this work we will focus on the accuracy, defined in Equation (
1). Accuracy ranges from 0 to 1, with 1 the perfect score. The accuracy also called fraction correct gives the fraction of correct forecasts. It is simple and intuitive but, in case of very rare events, this indicator may lead to confusion [
25].
Regarding the verification of probabilistic forecasts, two main attributes define the quality: the reliability and the resolution. Reliability refers to the statistical consistency between the forecasts and the observations. In other words, the forecast probability should be equal to the observed probability of the event (e.g., 20% of the events should happen for a forecast probability of 20%). The reliability is a crucial prerequisite as non-reliable forecasts would lead to a systematic bias in subsequent decision-making processes [
27]. The most used visual tool to assess reliability is the reliability diagram [
28]. Resolution refers to the ability of a forecasting system to generate case-dependent forecasts. For example, the climatology model, which predicts the average probability of the event (i.e., always the same probability regardless of the horizon or the weather conditions) has no resolution. Similarly to the reliability diagram, the ROC (relative operating characteristic) diagram [
29] provides a visual assessment of the resolution. It plots the hit rate (Equation (
2)) against the false alarm rate (Equation (
3)), for a set of increasing probabilities (e.g., 0.1, 0.2, 0.3, etc.). It is a measure of the ability of the forecasting system to discriminate between two different outcomes.
Only a few metrics, also called scores, exist to quantitatively evaluate the quality of probabilistic forecasts of binary events. For this work, we propose to use the Brier Score (BS) [
24], formulated as follows:
where
N is the number of observation/forecast pairs,
the forecast probability, and
the observation. If the event did occur
, and if it did not occur
. The BS measures the mean square probability error. This global proper score is appealing because it includes the two basic skills of a probabilistic forecast (i.e., reliability and resolution) and it corresponds to
for a deterministic forecast. The BS ranges between 0 and 1 with 0 the perfect score.
Skill scores (SS), derived from the above-mentioned metrics are also commonly proposed to evaluate forecast quality [
21,
22]. SS quantify the improvement of a proposed method compared to a reference model. They are relevant for comparing forecasts generated for different sites or time periods. In this work, we will compute a SS based on the BS with the climatology as a reference, as given in the following equation:
3. Deterministic Forecasts of Cloud Presence with a Sky Camera
To issue a forecast, a sequence of consecutive sky camera images is required. Following the methodology proposed by [
18], three consecutive images are enough to make a prediction. The next step consists of identifying the cloud presence in these three consecutive scenarios. Clouds identification follows the methodology proposed by Alonso et al. [
30]. In this work, the authors used the RGB and HSV color spaces and radiometric data to identify clouds. RGB represents red, green, and blue colors, whereas HSV represents hue, saturation, and value. The clouds identified are those that mainly attenuate DNI below 400 Wm
. The correlation between these three consecutive images makes it possible to establish the behavioral pattern of cloud movement at a given time.
The processing of images consists in defining different value combinations from RGB and HSV color spaces. Cloud identification follows a sequence of combinations. The first one is the detection of pixels that represent the cloudless sky parts in the image. The process involves complex and numerous combinations and ratios between all image channels, which can be analyzed in [
30]. When the algorithm finalizes, a sky representation is obtained, taking the position of clouds if they exist.
Figure 3 represents a typical cloud detection in a sky cam image, where the A image represents the raw image, whereas B represents the processed image where the blue pixels represent the clear sky and white pixels the clouds.
In order to study cloud movement, the following steps are taken [
18]:
The picture taken with the sky camera is divided into the different sectors given in
Figure 4, since the movement of the clouds will depend on the sector covered by the sky camera.
The cloud motion vector (CMV) is calculated for each sector by applying the maximum cross-correlation method.
Different quality tests are applied to ensure the correct determination of the cloud motion.
The CMV is applied to the last image received and re-applied to the result obtained. This process is repeated up to 30 times (prediction for 30 min ahead), obtaining the movement of the pixels from the minute in which the image was taken to the 30th minute in the future. Therefore, each application of the CMV is 1 min of forecasting. Finally, the prediction of cloud presence consists of checking if the new position of the clouds masks the future sun’s path.
According to the sequence of sky cam images used, one-minute forecasting is considered an appropriate interval to study the cloud movement. Having a lower interval (i.e., a lower time between scenes and lower temporal horizons in the forecasting) will result in an important computational time. On the other hand, increasing the interval to several minutes leads to a loss of information because of the rapid movement of clouds.
4. Post-Processing with Binary Probabilistic Models
In the literature, three main categories of statistical models are proposed to generate probabilistic forecasts of binary events [
31]. The first family of models, called parametric, assumes that the probability of the event follows a known distribution law, such as Gaussian or logistic. Conversely, the second type of models, called non-parametric, is not based on underlying distributions. Predictions are learned from a sample of data and, obviously, machine learning techniques dominate this second family of models. The last category, called semi-parametric, is a mix of the two previous ones. In this work, we proposed to test two parametric models and one non-parametric model to post-process the deterministic forecasts.
4.1. Parametric Approach
The first approach proposed here is based on the very well-known statistical models logit and probit used for decision-making problems involving binary or categorical choices in various domains such as economy [
16] or meteorology [
32]. These two parametric models belong to the generalized linear models (GLM) [
33]. Their aim is to model the probabilities of a random response variable
Y as a function of some explanatory variables. The model combines two functions. First, a function, called index function or systematic component, associates independent explanatory variables within a linear or non-linear model. Second, a link function links the systematic component with the random response variable.
For the logit and probit models, the index function
Z is a linear combination of independent explanatory variables (
, …,
) and corresponding regression coefficients (
, …,
), written as follows:
Used alone, this linear function is not able to provide suitable probability levels. Indeed, linear functions are not bounded in the range
. To overcome this issue, link functions have been proposed to transform the result of the index function
Z into probabilities ranging between 0 (i.e., cloud) and 1 (i.e., no cloud). In our case, only the link function differentiates the logit and probit models. For the logit model, the link function is the following logistic function (also called sigmoid function):
For the probit model, the link function is the cumulative standard normal distribution function given below:
The input variables of these models can be either continuous, binary, or categorical. This feature is very important in our case because the explanatory variables available to generate probabilistic forecasts of the cloud presence are binary (i.e., the deterministic forecasts) and continuous (i.e., measured irradiance, solar zenith angle, hour of the day, etc.). The main difference between these two parametric models is the shape of the link function. The logistic function produces heavier tails than the standard normal distribution function. To implement the logit and probit models, we used the “glm” function of the package “stats” that is part of R [
34], which is based on the maximum likelihood approach to estimate the coefficients.
4.2. Non-Parametric Approach
As most real-life phenomena do not follow a known distribution law, non-parametric models have been developed. Non-parametric binary choice models have been initially developed for economic applications [
35]. A set of non-parametric regressions, designed for continuous variables and also suitable for binary events probability, are available in the literature [
36,
37]. The main challenge for these regression models is to combine continuous, categorical, and binary data as input [
38], like in our work.
Decision trees (DT) and, by extension, random forests (RF), which belong to the supervised machine learning methods, are appealing non-parametric models to predict discrete choice. Indeed, they can predict either a numerical value (regression tree), a class, or a discrete choice (classification tree). They can use either continuous, categorical, or binary variables as input. They require less computational effort than classical non-parametric regression methods. Indeed, the computation time of regression methods increases exponentially with the number of variables which is not the case for DT and RF.
The characteristics of the RF used in this work are introduced by [
39]. The readers may refer to [
40] for a general presentation. A classification tree is a decision tool that estimates the most likely class of a categorical or a binary variable to predict when the input variables are known. Decision trees are simple models that partition the features (or inputs) space into subsets [
40]. An iterative algorithm is used to split the input space. At each step or node, the data are divided into two subsets, applying an “If, Then” rule to one of the input variables. At each step, the selected input is chosen to provide the best possible separation of the classes to predict. The aim is to generate the optimal sequences of rules to predict the different possible classes [
39].
An RF is a set of trees that are built on bootstrapped training subsets. Several decision trees are therefore trained. When RF are used as classifiers, the probabilities of the predicted classes are averaged from the answers of the individual trees, as illustrated in
Figure 5. In the RF, the strengths (and weaknesses) of each tree are aggregated. A cross-validation was done on the number of trees and a good trade-off was obtained for 500 trees. In this work, we used the RF classifier algorithm implemented in the R package “randomForest” based on [
41].
4.3. Implementation
As previously introduced and presented in
Figure 1, the forecasting process has two main steps: generation of deterministic forecast from the sky imager and post-processing with the probabilistic models. The first step is briefly detailed in
Section 3. Here, we will focus on the implementation of the probabilistic models. The simplest approach is to use only the discrete cloud forecast
as input of the probabilistic model. However, numerous works show that the addition of inputs, such as past observations or solar path variables, can significantly improve the quality of solar forecasts generated by time series models [
42] or post-processing methods [
43,
44]. Thus, to improve the performance of the post-processing step, we tested the addition of easy-to-compute variables as input to the three tested probabilistic models. The tested additional variables are: solar zenith angle, current and past global horizontal irradiances, beam normal irradiances and clear sky indices, and mean and variability over past observed clear sky indices. The best combination of inputs, based on the BS, is as follows:
The deterministic forecast of cloud presence ;
The current clear sky index ;
The mean over the five past clear sky indices .
Finally, we created one post-processing model by forecast horizon. Considering a time resolution of 1 min and horizons up to 30 min, we trained 30 different models for each of the three probabilistic methods presented above.
5. Case Study and Data
In this study, images from a sky camera with a rotational shadow band (TSI-880 model) have been used to provide a hemispheric vision of sky (fish-eye vision). Additionally, the measurements of diffuse and global irradiance from two CMP Kipp & Zonen pyranometers and direct irradiance from a CH1 Kipp & Zonen pyrheliometer were used, and all the instruments were installed on a two-axis solar tracker. The testing facility is located at the Center of Research of Solar Energy (CIESOL) at the University of Almería in a region in southern Spain. The facility has a Mediterranean climate with a large presence of maritime aerosols and is located at 36.8
N latitude and 2.4
W longitude at sea level. Data are collected every minute, as this was proposed to be a suitable frequency [
3]. Appropriate maintenance was performed on the sensors and sky camera. The sensors are cleaned with ethyl alcohol every day. The sky camera mirror is cleaned using a soft rag with distilled water three times a week.
Figure 6 shows the sky cam installed on the roof of CIESOL.
Images were taken with 352 × 288 color pixel resolution, which corresponds to 24 bits in JPEG format. They have three different channels that represent the red, green, and blue levels. Each pixel of the image is represented by 8 bits, with values between 0 and 255.
For the cloud now casting, data from 2010 and 2011 were used, for moments where the solar altitude degree was higher than 5. For 2010, a total of 137,794 moments were analyzed for each interval of prediction (1 to 30 min) independently, whereas for 2011, 134,993 predictions were processed, also for each forecast interval. The year 2010 was used to train the post-processing models and the year 2011 was used to test them.
It should be noted that the number of observation/forecast pairs and the ratio of observed cloudless skies in the test set (2011) are not identical for the different forecast horizons. Indeed, the ASI fails to predict the presence of clouds for long horizons when the cloud speed is high and/or when the cloud base height is low. Specifically, under these conditions, predicted cloud locations have a high probability of leaving the ASI’s field of view before a horizon of 30 min. As a consequence, the total number of observation/forecast pairs decreases while the ratio of observed cloudless skies in the test set slightly increases with the forecast horizon as presented in
Figure 7. As the clear skies are easier to forecast, this pattern will impact the assessment of the model’s accuracy.
6. Results and Discussion
The post-processing of the deterministic forecasts gives a probability level of the possible future cloud presence. But, it can also be seen as a calibration of the deterministic forecasts based on the training set statistics. Furthermore, it is common to transform the probability level resulting from the discrete choice models in new binary and deterministic forecasts. Indeed, considering a perfectly calibrated probabilistic forecast (i.e., for a forecast probability of x%, we observed x% of yes event), a threshold set at a probability of 50% to make a decision (event or no event) maximizes the accuracy of the resulting deterministic forecast. This assumption is discussed in
Appendix B. Therefore, to convert a probabilistic forecast into a deterministic one, we assume that a probability above 0.5 (>50%) corresponds to a “yes” event, i.e., in our case a clear sky, whereas a probability below 0.5 (<50%) is a “no” event corresponding to the presence of clouds. To assess the ability of the selected discrete choice models to improve the deterministic forecast, this transformation was applied to the probabilistic forecasts. Thus, the evaluation of the generated forecasts will be performed in two steps. First, we will evaluate the improvement of the quality of the deterministic forecasts before and after the post-processing step. Second, we will assess the quality of the probabilistic forecasts and the improvement compared to the corresponding deterministic forecasts.
Table A1 provided in
Appendix A, gives the detailed numeric results used to plot the graphs evaluating the quality of the forecasts.
6.1. Deterministic Forecasts Quality
Figure 8 shows the evaluation of the quality of the deterministic forecasts before and after the post-processing with the three discrete choice models tested in this work. Surprisingly, with longer horizons, the accuracy of the initial forecasts done with the sky imager increases (solid black line). This observation results from the share of clear and cloudy skies available in the test sets presented previously in
Section 5 and
Figure 7. Indeed, for longer horizons, the share of clear skies, which are easier to forecast when there is no cloud in the field of view of the ASI, is more important. We could have homogenized the test sets of the different horizons to cancel this effect. However, removing conditions with fast-moving clouds from the shorter horizons would have biased the analysis of the improvement brought by the probabilistic approach which is more interesting when forecasting becomes more uncertain. Even if this effect does not influence significantly the results of this work, the reader must keep in mind that for the longest horizons, the test sets lead to a higher share of situations that are easier to forecast [
3,
45].
As expected, the post-processing with the discrete choice models improves significantly the accuracy of the forecasting system. For horizons from 1 to 15 min, the accuracy resulting from the three models decreases. Above a 15 min horizon, as for the original ASI forecasts, the accuracy increases slightly. Among the three tested models, the RF model, which is a non-parametric method, shows the best improvement with an accuracy of 93.4% and 90.3% for horizons of 1 min and 30 min respectively. Compared to the initial ASI forecasts, this improvement corresponds to a gain of 11.6 percentage points for the shortest horizon (i.e., 1 min) and 7.5 percentage points for the longest one (i.e., 30 min). Regarding the two parametric techniques, the logit model, which has an accuracy close to the RF, clearly outperforms the probit model. However, both of them show a significant improvement over the ASI original forecasts. Given the simplicity and the low computational efforts of the logit model, the former offers a very good trade-off for the study case selected in this work.
6.2. Probabilistic Forecasts Quality
As previously discussed in
Section 2, the reliability of the probabilistic forecast is the first attribute to verify.
Figure 9 gives the reliability diagrams of the three discrete choice models and of the climatology model for all the horizons of forecast (i.e., overall reliability). The climatology is a very simple model used as a reference, which forecasts the average probability of the event whatever the weather conditions and the horizon. Here, the average probability of having a clear sky computed from the test set is 74.8%. The reliability diagram is a visual tool that gives a qualitative assessment of the reliability. It plots the correspondence between the forecast probability (x axis) and the observed frequency of the event (y axis). A perfectly reliable model should result in a reliability curve that sticks to the diagonal. Here, none of the three tested models presents perfect reliability. Conversely to the RF models, the probit and logit models never generate forecast probabilities of 0 and 1. As a consequence, their reliability curves do not reach the lower and upper limits of the diagram. The important deviations from the diagonal of the probit and logit models indicate a peak of underconfidence for a forecast probability of 0.5 and overconfidence for forecast probabilities ranging between 0.75 and 0.9. In other words, when these two models issue a forecast probability of 0.5 (i.e., 50% probability of a clear sky), the actual observed frequency is higher than 0.75. The RF model shows a better overall reliability than the two parametric models with a high reliability when it forecasts a clear sky with forecast probabilities above 0.5.
The second interesting attribute to verify is the resolution, which corresponds to the ability of a forecasting system to issue case dependent forecasts.
Figure 10 gives the ROC curves of the four probabilistic models evaluated in this work. A perfect resolution will lead to a curve that sticks to the y axis, i.e., a false alarm rate equal to zero, and then follows the line with a hit rate equal to 1. Thus, the higher the area under the ROC curve, the better the resolution. The climatology, which always gives the same forecast, has no resolution and follows the diagonal. The probit and the logit models have exactly the same resolution. The quality of their forecast depends only on their reliability. Finally, the RF model shows a better resolution than the other methods. As the RF model also has better reliability, its overall quality evaluated with a proper score should be the best.
In addition to the reliability and resolution assessment, the BS and the corresponding SS provide quantitative information on the quality of the forecast. The BS is negatively oriented and a lower value indicates better quality.
Figure 11 shows the BS of the original ASI forecasts, of the climatology model, based on the training set, and of the three discrete choice models. For the original ASI forecasts (solid black line), which are deterministic, the BS is derived from the accuracy as detailed in
Section 2. First, we can observe that the quality of the climatology increases slightly with the horizon. Again, this trend results from the increased share of clear skies for the longer horizons in the test set. However, the SS given in
Table A1, which gives the relative improvement over the climatology, shows that the relative efficiency of the three tested models decreases while the horizon of forecast increases. Second, the BS of the probit and logit models is almost the same regardless of the horizon. This result, which differs from that obtained with their deterministic counterparts, highlights that the information included in a probabilistic forecast cannot be translated into a deterministic forecast. Finally, the RF model clearly outperforms the two parametric models. The good performance of this non-parametric model comes from several advantages. Indeed, RF is able to map non-linear relationships between inputs and outputs. It is designed to handle different types of variables, which can be binary, categorical, or continuous. Finally, unlike probit and logit models, RF issues probability forecasts of 0 and 1 with high reliability.
In this work, we propose an alternative method to evaluate the probability associated with a solar forecast. Until now, the developed forecasting models are calibrated on arbitrarily fixed probability levels. The resulting outputs are the associated quantiles of the future available solar energy. In this work, the calibration works in the opposite way. We defined a solar irradiance level (here 400 Wm) and our models give the probability of overtaking this threshold. This approach, based on a single threshold, is well suited to CSP plant operation, which requires anticipating the starting and stopping conditions. However, one can imagine applying the same approach to evaluate the probabilities of exceeding a set of arbitrarily chosen solar energy thresholds. As no other work proposes a similar binary solar forecast, it is currently impossible to compare our results with previously developed models. With the aim of making our results suitable for future comparison, we proposed a comprehensive testing procedure. Indeed, the selected tools and metrics, which come from other domains such as meteorology and economy, allow evaluation of the main attributes of the forecast quality.
7. Conclusions
This work is the first attempt, in the field of solar energy, to propose a methodology to generate very short-term probabilistic forecasts as a binary event. The objective is to anticipate the moment when the direct normal irradiance is higher than a defined threshold, suitable to the operation of concentrated solar power plants. The proposed approach combines binary forecast based on a sky imager with discrete choice models commonly used in various decision-making problems to generate a probability forecast of cloud presence. Two parametric (probit and logit) models and one non-parametric (RF) discrete choice model have been tested in this work. The RF clearly outperforms the widely used probit and logit models. Beyond a better quality assessed with the reliability diagram, the ROC, and the BS, the RF provides better features, like the ability to forecast probability levels of 0 or 1 with high reliability.
As this work is the first one on the topic, no comparison with other models or approaches is possible and it is difficult to evaluate the actual performance of the proposed method. However, the generated forecasts show a good quality. Indeed, the accuracy of the deterministic forecasts derived from the probability level is above 90% with an improvement ranging from 7.5 to 11.6 percentage points compared to the original ASI forecasts. Regarding the probability forecasts obtained with the three tested models, their BS is below 0.1, regardless of the horizon of the forecast.