Next Article in Journal
A Gap Nonlinearity Compensation Strategy for Non-Direct-Drive Servo Systems
Previous Article in Journal
Control of Grid-Connected and Standalone Microhydraulic Turbine Using a Six-Phase Induction Generator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Quantum Neural Network for Solar Irradiance Forecasting: A Case Study Using the Folsom Dataset, California

by
Victor Oliveira Santos
1,
Felipe Pinto Marinho
2,
Paulo Alexandre Costa Rocha
1,3,
Jesse Van Griensven Thé
4 and
Bahram Gharabaghi
1,*
1
School of Engineering, University of Guelph, Guelph, ON N1G 2W1, Canada
2
Department of Teleinformatics Engineering, Technology Center, Federal University of Ceará, Fortaleza 60020-181, CE, Brazil
3
Department of Mechanical Engineering, Technology Center, Federal University of Ceará, Fortaleza 60020-181, CE, Brazil
4
Lakes Environmental Research Inc., 170 Columbia St W, Waterloo, ON N2L 3L3, Canada
*
Author to whom correspondence should be addressed.
Energies 2024, 17(14), 3580; https://doi.org/10.3390/en17143580
Submission received: 28 June 2024 / Revised: 14 July 2024 / Accepted: 19 July 2024 / Published: 21 July 2024
(This article belongs to the Section A: Sustainable Energy)

Abstract

:
Merging machine learning with the power of quantum computing holds great potential for data-driven decision making and the development of powerful models for complex datasets. This area offers the potential for improving the accuracy of the real-time prediction of renewable energy production, such as solar irradiance forecasting. However, the literature on this topic is sparse. Addressing this knowledge gap, this study aims to develop and evaluate a quantum neural network model for solar irradiance prediction up to 3 h in advance. The proposed model was compared with Support Vector Regression, Group Method of Data Handling, and Extreme Gradient Boost classical models. The proposed framework could provide competitive results compared to its competitors, considering forecasting intervals of 5 to 120 min ahead, where it was the fourth best-performing paradigm. For 3 h ahead predictions, the proposed model achieved the second-best results compared with the other approaches, reaching a root mean squared error of 77.55 W/m2 and coefficient of determination of 80.92% for global horizontal irradiance forecasting. The results for longer forecasting horizons suggest that the quantum model may process spatiotemporal information from the input dataset in a manner not attainable by the current classical approaches, thus improving forecasting capacity in longer predictive windows.

1. Introduction

Recent developments in quantum computing have enabled the application of machine learning paradigms within a quantum framework [1,2]. This innovative field leverages quantum physics principles, like entanglement, superposition, and interference, to execute computational loads [3,4]. With these properties, it is theorized that quantum computers can leverage computational time in critical areas, such as financial transaction decisions, new materials and drug discovery, cryptography, and so on [5]. Unlike classical computers, quantum computers process information using quantum bits (qubits). The qubit’s state, represented by its orientation in a high-dimensional complex Hilbert ℂ⊗n space (with n being the number of qubits), is managed by quantum gates. These unitary operators perform reversible operations on qubits, altering their orientation within the Hilbert space. These operations can implement superposition, entanglement, and interference properties to the quantum circuit, composed of qubits, allowing computers to generate unique and more complex patterns than those produced by classical computational systems. In this way, it is expected that quantum machine learning (QML) models will surpass the predictive performance of traditional ML approaches, reaching what is known as the quantum advantage, where a complex problem can be efficiently tackled with a quantum framework [1,6].
Despite being a newly developed field, it is possible to find many QML models, with their applications being broad and promising [7]. One of the possible applications lies within the renewable energy area, an important player for containing the adverse effects of climate change caused by the emission of greenhouse gases. The majority of greenhouse gases come from electricity production operations [8,9]. Their collateral effects have been linked to damage to ecological areas [10,11] and the occurrence of extreme weather events [12,13], among other negative consequences [14,15]. The adverse effects of pollutant gas emissions also negatively impact the global economy, causing the loss of billions of USD worldwide [16,17,18,19].
Within this context, renewable energies serve as a significant player in mitigating the harmful effects of climate change. Recently, there has been a global movement towards developing greener and renewable energy sources. One can highlight the importance of solar power, which has led to global investments in renewable energies [20,21].
While renewable energies offer significant benefits in cleaner energy production, their inherently stochastic and intermittent nature presents challenges. Therefore, developing precise and reliable tools for predicting energy output from these sources is crucial. In this field, there are three major approaches for solar irradiance forecasting: numerical weather prediction (NWP), statistical, and time-series models [22,23]. The NWP uses a physical-based approach involving weather attributes to determine future values of solar irradiance [24]. Statistical models rely on temporal correlation to forecast solar irradiance, being able to handle non-stationary data [25,26]. Finally, it is also possible to determine future values of solar irradiance using time-series data, where machine learning models can be implemented [22,23].
Machine learning (ML) has emerged as a valuable tool in this regard, demonstrating its ability to identify complex relationships within time-series data and enhancing the accuracy of renewable energy forecasting. Numerous works have employed ML to assess the behavior of key environmental indicators [27,28,29,30], including applications in the renewable energy area [31,32,33,34,35]. However, quantum machine learning (QML), remains in its early stages, with an absence of studies exploring its application in time-series scenarios and, consequently, in the modeling of renewable energy sources [36,37].
To address the knowledge gap in QML applications within renewable energy, this study proposes the development of a simulated quantum neural network (QNN) framework for short-term forecasting [22] of solar irradiance up to 3 h in advance. The framework utilizes the benchmarking dataset from Folsom, as introduced by Pedro et al. [38]. The main objectives of this work are as follows:
  • To address the information gap in the scientific literature regarding solar irradiance estimation using QML predictive models.
  • To assess the feasibility of a QML model designed for forecasting solar irradiance up to 3 h in advance.
  • To evaluate different configurations for the QNN framework, investigating several parameters composing its structure and their impact on the model’s output.
  • To investigate different optimization strategies and their impact on the model’s output.
  • To evaluate the QNN performance compared to classical ML counterparts.

2. Materials and Methods

2.1. Solar Irradiance

Solar irradiance touching the Earth’s surface can be categorized into three distinct types. The first type, Direct Normal Irradiance (DNI), includes all solar irradiance reaching the Earth’s surface directly without dispersion. When dispersion occurs, it is referred to as diffuse irradiance (DI). The third type, Global Horizontal Irradiance (GHI), encompasses the total solar energy reaching the Earth’s surface from all directions, i.e., including DNI and DI [39].
Accurate forecasting of GHI and DNI is essential to the cost-efficient implementation of photovoltaic projects and the proper development of public policies or plans related to solar energy [40,41,42].

2.2. The Folsom Dataset

The Folsom dataset spans 2014 to 2016, containing meticulously checked minute-by-minute measurements of GHI and DNI from Folsom, California, USA. This region holds elevated interest in solar irradiance, given its importance in the US economy and its status as a technological HUB housing millions of people [32,38]. The Folsom dataset was conceived to be used as a benchmark dataset to promote the development of solar forecasting models [38], containing ground-based sky images, satellite data, and forecasts from numerical weather prediction models. The present project’s dataset was limited to GHI and DNI measurements and sky images.
As mentioned previously, solar irradiance is stochastic and intermittent. Additionally, its forecasting depends on many factors, such as the celestial body motion, weather, geographic location, cloud coverage, and albedo, to cite a few [31,43]. A way to mitigate these effects, along with daily and seasonal variances of solar irradiance [44], is using the normalization by the modeling of the clear-sky index, k t , c s , a dimensionless number between 0 and 1, defined as below:
k t , c s = I I c s
In Equation (1), the term I usually refers to GHI or DNI. The denominator I c s is a modeled value of clear-sky irradiance, representing the maximum incident irradiance reaching the Earth’s surface. For this dataset, the clear sky model developed by Ineichen and Perez [45] was implemented, taking into account the characteristics of the study location.
In the original work by Pedro and colleagues [38], from the calculated k t , c s clear-sky index value, three feature vectors were engineered, namely Backward average (Bi), Lagged average values (Li), and clear-sky Variability (Vi). It is important to highlight that these features are calculated within the time window preceding the desired forecasting interval. Additionally, the sky images were collected from all sky-dome pixels, using 8-bit color images from a camera pointed toward the sky. The pixels were converted to a floating-point vector representing three different color channels: Red (R), Green (G), and Blue (B). Furthermore, the vectors relating the red-to-blue and the normalized red-to-blue ratios were also calculated. Finally, for each of the previous features, the mean value, standard deviation, and entropy were determined, resulting in fifteen features related to the sky images [38]. The total number of features is 18, including the engineered feature vectors. Table 1 compiles the 18 attributes used by Pedro et al. [38].
At this point, it is essential to make clear that the solar irradiance forecasting in this project was made upon the prediction of the clear-sky index value for each irradiance, i.e., k t , c s , G H I and k t , c s , D N I , which are later used to determine the values for GHI and DNI, respectively. Such an approach was used in the original work of Pedro et al. [38] and other solar irradiance forecasting works [8,31,32]. This approach is preferred over direct GHI and DNI forecasting, as the clear-sky index is a normalized and non-seasoned solar irradiance measurement, which may improve the determination of future values [46]. Statistical analysis and data visualization for the solar irradiance under forecast can be found in the Supplementary Materials file.

2.3. Classical Machine Learning Models

Traditional ML methodologies process data through their structure. To this end, different strategies have been developed in the previous decades, each with specific characteristics. For the present study, three classical methodologies were implemented.
The first one was the Support Vector Regression (SVR). This ML model is a variation of the Support Vector Machine (SVM) but designed specifically for regression scenarios. Similarly to the SVM, the SVR uses kernel functions to extrapolate the dataset to higher dimensions, allowing it to learn complex non-linear interactions within the dataset [47,48]. Furthermore, the SVR also uses convex optimization to improve its computational performance, mitigating the error during the training phase [49,50,51].
Another approach under consideration in this work was the Group Model of Data Handling (GMDH). Its configuration resembles a multilayer perceptron (MLP), and it has a self-organizing structure that allows it to select the best configuration for a given dataset [52,53,54]. The output from the GMDH is a polynomial approximation of the target variable, resulting from a combination of the dependent and independent attributes [30,55,56].
Another classical ML model implemented in this study was the Extreme Gradient Boost (XGBoost), a tree-based structure [57]. In essence, the XGBoost consists of sampling the outputs of smaller trees composing it, ultimately combining them into a larger and more complex structure. This unique feature categorizes it as an ensemble method, given that its structure aggregates multiple models. The ensemble strategy employed by the XGBoost model enhances its robustness and reduces its sensitivity to data variance. Additionally, it serves as a regularization technique, effectively mitigating the risk of overfitting [30,48].

2.4. Qubits and Quantum Gates

Quantum bits are the paramount basis of quantum information processing [58]. They serve much like classical bits, being able to store information depending on their state [59]. However, qubits have a unique property that sets them apart from their classical counterpart: they can assume different states simultaneously. Mathematically, qubits can be understood as vectors where 0 = 1   0 T and 1 = 0   1 T for the excited and non-excited states, respectively, considering only one qubit. For a general n-qubit system, the resulting quantum state can be described by Equation (2) as follows:
Ψ = j = 0 n 1 α j j n
In Equation (2), the quantum state Ψ results from a linear combination of j n quantum states for a n-qubit system [60]. Additionally, the term α j , where α j   C , corresponds to the amplitude of a given quantum state and has a property such that α j = 1 . The absolute value amplitude can be understood as the probability for the qubit to assume a specific quantum state. Note that the quantum system allows superposition and entanglement of the qubit states.
In quantum computing, the information carried by the qubits is processed using quantum gates. Quantum gates are fundamental operators that apply unitary transformations in the qubit, describing its system’s evolution [3]. These operators can be mathematically represented by unitary matrices, defined as follows:
U 1 = U
Equation (3) describes that a unitary quantum gate operator has its inverse equal to its complex conjugate. This property is paramount to quantum information processing, as it reverses the operation and keeps the vector distance unchanged when applied to multiple qubits [3,60].
Applying one or more quantum gates in the qubits composing a quantum system is called a quantum circuit. The quantum circuits composing the QNN implemented in this work are discussed in the following subsection.

2.5. Quantum Machine Learning

The integration between classical ML and the newer quantum computer framework can be achieved in diverse ways, as initially proposed by Aimeur et al. [61], and this can be visualized in Figure 1 [3].
From Figure 1, we observe that the four combinations are based on the origin of the data (classical “C” or quantum “Q”) and its processing methodology (also classical or quantum). The top-left represents the classical machine learning approach, where a classical computer evaluates classical information. On the bottom-right of the chart, the quantum data is being processed by a quantum machine. In the context of this project, the “CQ” approach was adopted since the Folsom dataset, which was obtained through classical means, was processed using a quantum computer.
It is important to note that a simulated quantum computer was employed for the quantum processing component of this project. The use of simulated quantum circuits is a common practice given the current state of quantum computers, categorized as Noisy Intermediate-Scale Quantum (NISQ) machines. Although quantum computers have advanced to the point where a significantly large number of qubits is feasible, the real quantum hardware in NISQ machines still cannot provide sufficiently long coherence times (i.e., the qubit state is not maintained long enough) and is not entirely fault tolerant. Consequently, these devices are currently unable to deliver the theorized quantum advantage [62,63,64].
Additionally, the availability of real quantum hardware is minimal, restricted to a few enterprises that invest in such technology. These limitations hamper a deeper investigation of QML methodologies, as using the real NISQ devices requires monetary investments while offering limited computational power in its current state and may require an extended queue time in order to use them [3,62,63,65,66]. Therefore, to overcome these adversities, many researchers opt to simulate their quantum circuits, an approach adopted by the present study.

2.5.1. QNN Structure

As previously mentioned, quantum computers utilize qubits to process information, a fundamental distinction between quantum and classical ML frameworks. This results in an entirely new structure for the QML approach. In the context of QML, the QNN model consists of three basic components: a Feature Map, an Ansatz, and Measurement, which are constituents of quantum circuits where data are processed. A generic QNN architecture is depicted in Figure 2 [67].
In Figure 2, it is possible to visualize the structures composing the QNN model (encapsulated by the grey rectangle). The data coming from the n-qubits ( q 0 to q n 1 ) enter the QNN structure from left to right, passing through the feature map, the ansatz, and finally, the measurement stage, where the output is mapped to a classical bit configuration.

2.5.2. Feature Map

The feature map is a fundamental component of the QNN circuit. Since the Folsom dataset is of classical origin, it must be encoded in a way that allows it to be processed by a quantum framework. The feature map serves this purpose by mapping the classical inputs into a quantum space defined within a highly complex Hilbert space, enabling a quantum computer to process the data [68,69,70]. This step is arguably the most critical part of QML design, as it transforms the data structure deeply and non-linearly, achieving alterations that are not feasible through classical methods.
The current literature registers two methods for transforming classical data into a quantum format. The first one is called implicit, and it is related to directly applying a quantum circuit to implement the mapping strategy, defined as follows [69]:
ϕ : x ϕ x = U ϕ x 000 0
In Equation (4), a generic feature map ϕ takes the classical data vector x , where x R m × p , and encodes it into a quantum framework by using a parameterized quantum circuit U ϕ x containing n-qubits with initial states 0 . The second mapping strategy, called explicit, is related to using variational quantum circuits to determine the best mapping configuration. This study employs the implicit approach to facilitate the validation of proper encoding.
Considering the implicit feature mapping strategy, several distinct encoding strategies can be applied. In the context of this project, two feature mapping approaches were investigated. The first one is called angle encoding. This approach is straightforward, applying rotational quantum gates to all n-qubits in the quantum circuit. However, it still offers a powerful way to transform classical data into a quantum paradigm [3,59]. Each rotational gate is parameterized, taking as input the numerical input vector x from the classical dataset. Later, the input information is passed as the rotation angles for the quantum gates [59]. Therefore, the angle encoding uses the number of quantum qubits in its circuit with the same dimensions in the dataset, i.e., the total of predictors used to forecast solar irradiance [71]. The angle encoding mapping strategy can be represented mathematically by the following equation:
ϕ x = i = 0 n 1 R ϕ ( x ) ϕ
Equation (5) is such that, for a n-qubit system, the rotational gate R ϕ is applied to each one of the input parameters of the original classical dataset, resulting in the quantum state ϕ x . Figure 3 represents a generic angle encoding mapping for n-qubits at an initial 0 state.
In this approach, R ϕ can assume any Pauli’s gate form, i.e., X, Y, Z, or I. Detailed information regarding these quantum gates can be found in dedicated references such as in Nielsen and Chuang [58].
The second feature mapping strategy investigated in this project was the ZZ feature map. It is the second-order Pauli expansion circuit [65]. This approach maps classical data to a quantum framework using a quantum circuit that transforms input data as follows [72]:
U ϕ x = exp i S I ϕ S x i S P i
In Equation (6), the mapping quantum circuit U ϕ x results from a relationship between the data-mapping term ϕ S x and the Pauli gates, P i = { X ,   Y ,   Z ,   I } , having a diagonal form [73]. Furthermore, S represents a set of qubit indices in the feature map, and I is a set containing all these index sets. Commonly, the term ϕ S x is defined as
ϕ S x = x i ,                                                     i f   S =     i j S P i π x j ,       i f   | S | > i  
The ZZ feature map is also a parameterized circuit, where the input values from its gates come from the classical dataset. The ZZ feature map circuit is presented in Figure 4 [59].
From Figure 4, the ZZ feature map initiates its circuit by applying Hadamard gates to the initial 0 state of the n-qubit system. It then proceeds to apply a rotation, where P i = Z . After that, the circuit utilizes controlled-NOT gates pairwise, i.e., it takes {j, k} ⊆ {q0, q1, …, qn−1} for j < k, to implement entangle states in the system [59].

2.5.3. Ansatz

The second component of the QNN structure is the ansatz. In quantum computing, ansatz refers to the initial guess for a quantum state, which will be used as a starting point for the computation. In the context of the QNN model, similarly to classical ML, it assumes the role of the training weights of the QML. To this end, the ansatz is implemented as a parameterized variational quantum circuit. The ansatz has parameters trained for each iteration through a classical optimization algorithm, reckoned by a classical computer [74]. A generic variational quantum circuit configuration is depicted as follows [3].
From Figure 5, showing the quantum circuit’s parameters, the ansatz takes as input the previously mapped classical data. Then, its parameters θ are updated in each iteration after consulting the cost function value. This process is usually performed by an iterative classical optimization algorithm, which aims to reduce its value in each iteration. Note that, at this point, the measurement shown in Figure 5 refers to the ansatz’s optimization process, not the final measurement performed at the end of the QNN structure (Figure 2).
For this project, three different ansatz were investigated. The first is a simple Ry ansatz, similar to the parameterized quantum circuit previously discussed for the feature map. It consists of applying parameterized Pauli Y quantum gates. The second strategy was the Two Local ansatz. This ansatz is composed of alternating rotations and entanglement layers using parameterized gates [75,76]. The rotational gates are applied individually for each qubit, while the entanglement is issued over two qubits according to the desired entanglement strategy. Two entanglement strategies were evaluated in this project: linear and full entanglement. The former consists of entangling an i-th qubit with its successor i + 1 for all qubits where i { 0,1 , , n 2 } for a n-qubit system. The latter entanglement strategy consists of entangling all qubits with each other. The qubit entanglement may be repeated. An example of the Two Local quantum circuit ansatz with linear entanglement and two repetitions is depicted as follows [77].
Figure 6 depicts a Two Local Ansatz with linear entanglement and two repetitions. It is possible to visualize that the rotational gates R ϕ accept parameterized values θ as inputs. These values represent the total change in qubit orientation, measured in radians. The total of parameters to be trained will depend on the total number of repetitions in the ansatz circuit. In the given example, since the quantity of qubits is the same as the number of dimensions in the classical dataset, defined as R m × p , there is a total of p parameters in the first rotational layer (i.e., p = n − 1). Afterward, linear entanglement is implemented using controlled-NOT gates in a pairwise fashion. Later, the first repetition takes place, doubling the total count of parameters that can be trained. This is followed by another linear entanglement, which precedes the second repetition. In this example, the optimization algorithm needs to determine a total of 3p trainable weights.
The final ansatz used in this project was the Quantum Approximate Optimization Algorithm (QAOA). It is a hybrid ansatz approach combining classical and quantum frameworks [78]. The QAOA is a more specific form of variational quantum eigensolver (VQE) quantum circuits, which aim to find the lowest eigenvalue of the Hamiltonian of a system [59,79]. To this end, the parameterized ansatz circuit prepares the quantum state such that | Ψ ( θ ) = U ( θ ) , where U ( θ ) represents the parameterized ansatz. In this context, this relationship can be rewritten as follows [3]:
C θ = Ψ θ | H Ψ θ
In Equation (8), the cost function, C θ , is calculated using the expected values of the system’s Hamiltonians. The cost function is minimized using classical optimizer algorithms [3,80]. Unlike the VQE, the QAOA approach uses a specially optimized ansatz, consisting of unique parameterizations of the problem Hamiltonian and rotational gates. The main configuration of QAOA is determined by a single integer parameter, which sets the depth of the ansatz circuit and directly influences the quality of the approximated result [78]. A generic quantum circuit representing the QAOA ansatz is presented as follows [3].
In Figure 7, C and B define the ansatz state starting with a uniform superposition, after implementing the Hadamard gates upon all the qubits. The term C is defined as C z = k = 1 K C k z , where z = { 0 ,   1 } n   represents a string satisfying a set of C k , and the diagonal element of the Hamiltonian of a system, H i , i , is defined as H i , i = C(z). The term B represents the sum of Pauli X operators in the quantum circuit B = i = 1 n ( P a u l i X ) i . Finally, U B , β = e i β B and U C , γ = e i γ B [3].
Beyond the number of qubits and the quantum gates composing them, the number of circuit repetitions also significantly characterizes both feature maps and the ansatz. This repeats the circuit in sequence m-times to the same qubits [3]. This results in an embedded circuit in the form U x U x U x = U x m . Considering the ansatz structure, the repetition of its quantum circuit can be compared to classical layers in a classical neural network [81]. Works on this subject have shown that repeating circuits for feature maps can encode data to come up with universal approximation theorems of QML algorithms when m   , improving their output [3]. In this project, repetitions of 1 and 2 were investigated when using the Two Local ansatz, as previously discussed.

2.5.4. Optimization Algorithms

One popular strategy adopted by researchers investigating the implementation of QML models is to not rely exclusively on a quantum paradigm but rather split it into quantum and classical architectures. The motivation behind this approach is to ease the computational efforts during the quantum circuit simulations. A possible implementation of this approach is to process the classical data using quantum circuits (feature map and ansatz), while the training phase is optimized using classical algorithms [3]. There are several strategies to optimize the cost function. In the context of this project, three optimization algorithms were investigated.
The first is the Constrained Optimization BY Linear Approximation (COBYLA) algorithm. It is a numerical optimization method used for problems with constraints, where the objective function’s derivative is unknown [82]. It minimizes a scalar function of one or more variables using linear approximation over a simplex, not requiring the cost function gradient to converge, thus mitigating errors related to gradient evaluation [83,84,85]. Its usage in QML models has been investigated in previous studies, showing that it requires a reduced number of iterations and provides good accuracy [85,86].
The second optimizer investigated was the Limited-memory Broyden-Fletcher-Goldfarb-Shanno Bound (L-BFGS-B). This is a quasi-Newton algorithm, meaning it does not require a second derivative of the objective function to minimize the value of a differentiable scalar function. The L-BFGS-B is an interactive model that solves unconstrained, non-linear optimization problems, updating the estimated value of the objective function with each iteration [87,88]. Application of this optimization strategy for quantum problems can be found in the literature, achieving good convergence speed and accuracy [89,90].
Finally, the last approach investigated in this project was the Quantum Natural Simultaneous Perturbation Stochastic Approximation (QN-SPSA) optimizer. Different from the previous optimizers, this algorithm is a gradient-based approach to minimize the cost function value. This is done by taking the derivative of an objective function, f(x), and changing the x-value to minimize it for each step. Considering the case for multiple inputs, f(x1, …, xn), the optimization must consider the partial derivatives of the function concerning each one of its inputs. In this case, one must calculate the gradient of the function x i f x . Thus, for the latter case, the new value of x that minimizes f(x) for an arbitrary time step can be calculated as follows [47,49]:
x t + 1 = x t ϵ x f x t
where, in Equation (9), the updated values of the attribute x t + 1 result from the variation of the previous x t by an amount ϵ of the gradient of the function x f x t . Here, ϵ is the learning rate, ϵ 0 ,   1 , and it regulates the total of the change attribute to the new vector. The learning rate, a hyperparameter of the gradient descent optimizer, directly influences the convergence of the cost function. Slight values for the learning rate may require more steps, t, to achieve convergence. Conversely, larger learning rate amounts can cause the gradient change to become too unstable, even preventing the function from converging toward its minima [47].
The QN-SPSA optimizer is based on the SPSA one [91], but seeks to achieve convergence by sampling the natural gradient instead of the first-order one. This approach resembles the Monte Carlo method, where the higher the number of samplings, the more accurate the model. This may lead to an exact analytic result for the objective function [92]. The QN-SPSA is well suited for use in large-scale population models, adaptive modeling, and simulation optimization. This results in the QN-SPSA approach speeding up the natural gradient calculation by decreasing its accuracy [92].
The QN-SPSA implementation in Qiskit requires fidelity and perturbation values when a specific learning rate number is passed to its function. The fidelity refers to the measurement of the closeness between two quantum states, which in this case are the Hessian of the fidelity matrix defined by the ansatz and its real value. The perturbation refers to the magnitude of the fluctuations in the parameters of the quantum circuit for the finite difference approximation of the gradients. Perturbation and learning rate can be set to have the same values when implementing the QN-SPSA optimizer [91]. It achieves this by approximating the Hessian of the fidelity of the ansatz circuit.

2.5.5. Measurement

The final step of the QNN structure is measuring the achieved quantum state after previous data processing by the feature map and ansatz quantum. Differently from quantum gates, which apply reversible operations, the measurement of the qubit state is not reversible. The measurement of a qubit’s state makes it collapse into one of its possible states defined in Equation (2) [60]. As previously mentioned, the larger the amplitude value, the more probable the wave function is to collapse into a specific quantum state.

3. Results

The classical forecasting models were implemented using the Scikit-Learn library for Python [93], while the QNN approach was built and simulated using Qiskit [94]. The hyperparameters of the classical ML models were selected using the GridSearch strategy, but for GMDH, which is a self-organizing model. The dataset was split in the 80/20 Pareto fashion for the training and testing phases, without shuffling the dataset. The computer used to build and run the models had a 13th generation i7 processor, 32 Gb of RAM, and an RTX 4080 GPU.
The ML and QNN models’ performances were evaluated using the root mean square error (RMSE), the coefficient of determination (R2), and the forecast skill (FS). These are some of the commonly used metrics to analyze the performance of predictive models.

3.1. Parameter Selection

The best parameter subset from the eighteen total was selected for each forecasting horizon. To this end, the recursive feature elimination (RFE) was implemented [48]. It consists of selecting the best features, starting with the whole set of predictors and selecting the most relevant ones according to a previously selected ML model. The present study implemented the RFE using a random forest (RF) model to select the best subset of attributes. Subsequently, they were given importance after all the attributes were used. Finally, the least important predictor was removed, and the RFE restarted. This process was repeated until a desired number of features was achieved.
Different feature values were evaluated. It was observed that, for smaller numbers, RFE would return only attributes related to one type of data information. That is, the selected parameters would be temporal (Bi, Vi, and Li) or spatial (R, B, G, and their derivates). Finally, the total of attributes was set to 5. The motivation behind this is twofold:
  • To keep both information from statistical features, related to temporal information, and sky image features, related to spatial data.
  • To keep the problem complexity low for quantum machine learning simulation.
From the RFE output, it was observed that for forecasting intervals of 5 and 10 min, most of the predictors were tied to the dataset’s temporal dimension, encompassing elements B, V, and L, along with two spatial predictors. However, as the forecasting window expanded, the balance shifted. This time, three out of five predictors were associated with the dataset’s spatial dimension, i.e., sky images, leaving only two tied to the temporal aspect. This suggests a dynamic interaction between spatial and temporal factors depending on the length of the forecasting horizon, where spatial information conveys more relevant information for solar irradiance forecasting.
As previously mentioned, the engineered temporal attributes were calculated considering the time interval prior to the desired prediction window. In this study’s methodology, we opted to select the time-series attributes considering the information for the desired forecasting horizon. Figure 8 illustrates the attribute selection process for a 5 min forecasting horizon, and the selected predictors are presented in Table 2 for each prediction interval.

3.2. Feature Map Quantum Circuit Configuration Selection

For the feature map quantum circuit, different rotations were assessed for the angle encoding strategy. The 5 min prediction window and global horizontal irradiance were used to benchmark the best feature map strategy. During this phase, we kept the ansatz constant, applying a simple rotational Pauli Y gate to each qubit in the circuit. We also used the COBYLA algorithm for optimization. It is worth noting that we did test different ansatz rotations and optimization algorithms during this phase. However, these alternatives did not yield better results than the chosen configuration. Due to its significant computational time requirements, the Two Local ansatz was not considered during this phase. The Pauli Y rotation yielded an RMSE of 56.91 W/m2 and R2 of 95.99% for the selected feature map circuit configuration. At this point, Pauli X and Pauli Z gates were discarded as angle encoding strategies for the feature map quantum circuit. The graph showing the convergence of this approach is presented in the following figure.
The second feature map investigated was the ZZ feature map. Again, using the forecasting horizon of 5 min for GHI as a benchmarking dataset, with the Pauli Y ansatz and COBYLA optimizer, the ZZ data encoding output RMSE was 447.78 W/m2, and the R2 was −1.48. The metrics elucidate that the ZZ feature map yielded inferior values to the ones provided by the Ry angle encoding. The graph showing the convergence of this approach is presented as follows.
Figure 9 shows that the feature map for the given configuration presented downward fashion from iteration one until around iteration twenty. Afterward, the results showed convergence. Figure 9 and Figure 10 show that the objective function achieved convergence at around forty iterations in both cases. However, the model’s convergence was near zero when Pauli Y was implemented as a feature map quantum circuit. At the same time, in the ZZ approach, it occurred not long past the 0.65 value, indicating poorer performance than its predecessor. This can be observed by analyzing the metric values, where the first approach had a greater RMSE, by approximately 155%.

3.3. Ansatz Quantum Circuit Configuration Selection

For the ansatz quantum circuit, the configuration was first investigated using the previously selected Pauli Y as feature map angle encoding, with the Two Local ansatz with the Pauli Y gate for rotation, the controlled-x gate for entanglement, and COBYLA as the optimizer for the 5 min forecasting horizon dataset for GHI. Note that the Pauli Y rotation used as the ansatz had already been evaluated in the previous subsection (Figure 9).
Considering the Two Local approach, two different entanglements were investigated: linear and full. The former presented better results than the latter. This configuration returned an RMSE of 58.08 W/m2 and R2 of 95.82%. The convergence for this model can be visualized in the following figure.
Changing the rotation angle encoding to ZZ feature map and using the Two Local ansatz with the COBYLA optimizer, the QNN model returned subpar results, with an RMSE of 402.04 W/m2 and R2 of −1.00. Its convergence is presented in the following figure.
In Figure 11, the convergence occurs after around 40 iterations and is near to a zero value for the objective function, after the bumpy behavior of its objective function value. Figure 12 presents a smoother objective function value, with the convergence occurring only after 100 iterations and well before the zero value of the objective function, indicated in the image’s y-axis. The superior performance of the angle encoding approach for the feature map circuit is again observed from the given results. Considering the ZZ feature map results presented in Figure 10 and Figure 12, this mapping circuit was discarded as an encoding strategy for the QNN model.
Finally, the QAOA was evaluated using a similar configuration, considering only the Pauli Y as a feature map quantum circuit. For this ansatz, the circuit depth was set to two. The convergence of this ansatz circuit is presented in the following figure.
Figure 13 shows that QAOA convergence is similar to the one in Figure 11, which occurs around 30 iterations. Assessing the values present in the y-axis, the QAOA ansatz does not converge to a value near zero, indicating its suboptimal performance. The QAOA ansatz for the given configuration returned an RMSE of 498.86 W/m2 and R2 of −2.08, being the worst-performing approach so far.

3.4. Optimization Algorithm

The selection of the best-suited optimizer was benchmarked using the previously selected Pauli Y feature map approach for 5 min GHI forecasting. The Pauli Y, Two Local, and QAOA ansatz were also evaluated during this phase. The QN-SPSA optimizer used a learning rate and perturbation values of 0.5 for the Pauli Y ansatz, 0.1 for the Two Local ansatz, and 0.3 for the QAOA ansatz. Different values were tested for QN-SPSA, and when applied to each ansatz were not able to surpass the results from the selected learning rates. Table 3 compiles the results for selecting the best optimization algorithm for each evaluated configuration.
Table 3 shows that the best configuration for the QNN model is composed by the Pauli Y angle encoding as a feature map, using the Two Local ansatz and the L-BFGS-B optimizer. The QN-SPSA optimizer achieved better results than those output by COBYLA and L-BFGS-B when used together with the Ry ansatz quantum circuit. However, when the Two Local approach was implemented, it was the worst-performing approach among the ones investigated. The QAOA ansatz also did not provide good outcomes for solar irradiance forecasting using the benchmarking configuration. It had a similar behavior as the one observed using the ZZ feature map, where the convergence occurred far from the zero value (Figure 10 and Figure 12). The following graph shows the convergence of the best-performing configuration present in the following figure.
Figure 14 showcases the convergence of the best-performing configuration for QNN. Compared with Figure 9, which uses COBYLA as an optimizer, it is noticeable that the convergence with the L-BFGS-B optimizer occurs faster, after around five iterations, having almost no oscillations along its path. The final configuration improved, by almost 15%, upon the output yielded by the second-best performing configuration, considering the RMSE metric (Table 3).

3.5. Number of Repetitions for the Ansatz

Additionally, the Two Local ansatz was implemented using two repetitions for its configuration, as shown in Figure 6, using controlled-NOT gates for entanglement. To this end, the best QNN configuration, previously defined and presented in Figure 14, was used. Using the Two Local ansatz with two repetitions yielded superior output metric values for 5 min GHI forecasting, achieving an RMSE of 39.74 W/m2 and R2 of 98.04%, meaning an enhancement of over 23% compared to the second-best configuration. However, its processing time proved prohibitive, as it took eight times longer (42 h) than the configuration presented in Table 3 (5 h). Therefore, the Two Local ansatz with two repetitions was not selected for the final QNN configuration. The selected QNN configuration is presented in the following Figure 15.

3.6. Solar Irradiance Forecasting Results

The best configurations for both classical and quantum neural networks were selected, and the models were implemented for solar irradiance forecasting up to 3 h ahead. Their outputs are compiled in Table 4 and Table 5 for GHI and DNI predictions, respectively. They are discussed in the follow-up section. The unit for the RMSE metric is W/m2.

4. Discussion

4.1. QML Models Found in the Literature

As previously stated, this study successfully implemented a simulated QNN model utilizing the Qiskit library in Python. Existing studies have applied simulated quantum machine learning to renewable energy, whether using PennyLane [95], Qiskit, or other available libraries.
In the work by Correa-Jullian et al. [96], a QML model using SVM was implemented for wind turbine fault detection. Their work proved that the quantum approach could improve the results over traditional ML methodologies, being able to outperform some of them. Another quantum application in the renewable energy area can be found in the work by Sagingalieva et al. [97]. There, the authors forecasted the photovoltaic energy output using a hybrid model, combining both classical and quantum paradigms. Their methodology achieved excellent results, proving the feasibility of this architecture in solar power prediction. The usage of QML strategies for microgrid control was investigated in the work of Jing and colleagues [98]. In their work, the authors used a hybrid QML model based on the Harrow Hassidim Lloyd (HHL) algorithm for microgrid management, which proved to yield remarkable results for such a task.
Considering that QML is still in its nascent phase, few works addressing solar irradiance forecasting can be found. In work by Senekane and Taele [99], the authors implemented a quantum SVM model for solar irradiation forecasting. Their work proved the efficacy of using QML for solar irradiance forecasting. In a more recent work, the authors Sushmit and Mahbubul [81] explored the implementation of a hybrid QML methodology for solar irradiance forecasting. Their methodology was tested across various locations and demonstrated its reliability as a forecasting tool. The structure employed in the current study mirrors the approach taken by Sushmit and Mahbubul. They utilized a parameterized quantum circuit with angle encoding as a feature map and an ansatz closely resembling the Two Local circuit used in this study. Their findings indicated that the QNN alone could not surpass the hybrid configurations examined, yielding RMSE and R2 values comparable to those in Table 4. Another study published by Yu et al. [70] investigated the combination of a quantum circuit embedded in a recurrent neural network for 1 h GHI forecasting. Their methodology also holds similarities to the one presented in this study, like the use of angle encoding. On the other hand, their ansatz circuit was somewhat distinct, presenting entanglement and rotation operations over the qubits. Their results for the annual prediction performance returned average results for RMSE of 61.76 W/m2 and 94.6% for R2, showing a significant improvement over the classical ML evaluated.
The present study yields similar outcomes to those of the published studies in QML for solar irradiance forecasting. Notably, references [70,81] utilize hybrid QML configurations to enhance their solar irradiance predictions [100]. Nevertheless, based on the evaluation metrics, our study matches these predictions effectively.

4.2. GHI and DNI Forecasting Results

From Table 4 and Table 5, the superior performance of the XGBoost approach is noticeable when compared to that of its competitors. It managed to overcome each of the other models for every assessed time horizon, especially for GHI, where it reached the best values of 28.74 and 98.98% for RMSE and R2, respectively. This is not a surprise, given that the XGBoost is a well-established ML model, and its superiority for solar energy forecasting has been reported in several studies [31,101,102,103]. The remarkable XGBoost performance can be explained as twofold:
  • The parameter selection was performed by a random forest model, which is a tree-based model and therefore may have added some bias toward the improved performance of the XGBoost.
  • The inherent capacity of tree-based models to outperform other ML and deep learning approaches on tabular data has been noted, as investigated in [104,105].
Considering the forecasting interval between 5 and 120 min, the models SVR and GMDH could also provide similar results to those of XGBoost, though differing slightly. For this same time window, the QNN model was the fourth-best-performing approach, also able to provide competitive results for GHI and DNI forecasting. The superiority of the classical models over the QNN approach is due to their much more mature implementation in a classical device, whilst QML software remains in its early stages [106], employing a very limited number of qubits. Additionally, classical ML approaches explicitly for a target task often overcome non-hybrid QML structures. Nevertheless, the quantum paradigm still can provide competitive results, as reported in [96,106].
Additional investigation of Table 4 and Table 5 reveals that for the longer forecasting horizon of 3 h, the QNN approach managed to reach the second-best performance among the assessed models. Commonly, in time-series forecasting of spatiotemporal environmental indicators, the predictive performance of the models is expected to decrease as the forecasting horizon expands. This behavior can be explained by the lack of additional information from the input parameters, causing the model’s performance to deteriorate for longer horizons [27,107]. This phenomenon occurs for the assessed models, as it is possible to observe for the forecasting comprising the interval of 60 to 180 min. However, the QNN approach presents a much more robust performance than the classical ones. For instance, in the 60 to 120 min interval for GHI forecasting, the QNN model exhibits less performance deterioration, as indicated by the performance metrics. This pattern does not hold for DNI forecasting, where the QNN approach displays a significantly more volatile behavior relative to other models. Remarkably, when forecasting both types of irradiances for a 3 h period, the QNN model secured the second-best performance among the evaluated paradigms. This fact can be an indication that the quantum model may identify and retrieve relevant spatiotemporal information from the input dataset in such a manner not attainable by the current classical approaches. This capability significantly enhances its performance over extended prediction windows.
Finally, from the results in Table 4 and Table 5, it is also possible to observe that the overall performance for all the evaluated models decreased considerably for DNI predictions. Such a behavior is common among predictive models. The GHI represents the total radiation that arrives at the Earth’s surface from all directions. On the other hand, the DNI is a highly directional measurement, where even minor obstacles can prevent sun rays from reaching the Earth’s surface, resulting in inaccuracies in its modeling [31,108].
Although the outcomes are suboptimal for shorter forecasting windows, i.e., from 5 to 120 min in advance, when compared to conventional ML frameworks, the quantum paradigm provides a compelling perspective on its utility in forecasting solar irradiance, more specifically for longer prediction intervals. This approach could potentially contribute significantly to advancing investigations in the domain of renewable energies. As the current quantum computing technology evolves, the application of quantum circuit models using real quantum hardware is expected to become more tangible [7]. In this context, the implementation of QML models on real quantum devices will lead to faster processing times with superior accuracy to that currently possible. This would ultimately allow the achievement of quantum supremacy [5].

5. Conclusions

This study investigated the potential of applying quantum machine learning to develop a forecasting model for the GHI and DNI to optimize solar energy production. The widely recognized Folsom dataset served as the benchmark for this investigation. In this analysis, the most effective subset of predictors, encompassing both spatial and temporal attributes, were chosen for each forecasting horizon ranging from 5 to 180 min. The attribute selection was made using the RFE approach. Different configurations for the quantum model were examined, and the best-performing one was selected. The QML results were compared to classical ML algorithms, namely SVR, XGBoost, and GMDH.
Results for the parameter selection showed that the influence of spatial-related predictors increased as the time window escalated. For the QML, the best configuration for the QNN model was achieved using Pauli Y as an angle encoding for the feature map, the Two Local ansatz with Pauli Y, and controlled-NOT gates, containing one repetition, with the L-BFGS-B optimizer. The overall best results for irradiance forecasting were reached using the XGBoost model for all investigated time horizons. The QNN model, however, was able to provide competitive results for horizons spanning from 5 to 120 min, especially for forecasting GHI. However, the expansion of the forecasting horizon to 180 min revealed that the QNN model managed to overcome the classical approaches of SVR and GMDH, indicating a superior capacity for identifying and extracting spatiotemporal information from the dataset compared to its competitors. The overall performance of the models for the DNI forecasting was inferior when compared to the GHI results.
For future works, the proposed QNN methodology could be evaluated for other renewable energy applications, such as wind speed forecasting for wind energy production. In addition, a more complex ansatz configuration containing more than one repetition could be investigated. Finally, it is expected that further investigation of the QNN model applied to real quantum hardware may allow the usage of deeper quantum circuits. This would permit the implementation of more complex quantum circuit configurations, such as larger repetitions for the ansatz, promoting a more expressive quantum machine learning model, while having improved processing time and more accurate outcomes.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/en17143580/s1, Figure S1: Boxplot for GHI considering all the forecasting horizons.; Figure S2: Boxplot for DNI considering all the forecasting horizons.; Figure S3: Histograms for GHI considering all the forecasting horizons.; Figure S4: Histograms for DNI considering all the forecasting horizons.; Table S1: Statistical summary for GHI for each forecasting horizon. The unit is W/m2.; Table S2: Statistical summary for DNI for each forecasting horizon. The unit is W/m2.

Author Contributions

Conceptualization, J.V.G.T. and B.G.; methodology, P.A.C.R., J.V.G.T. and B.G.; software, V.O.S., F.P.M. and P.A.C.R.; validation, P.A.C.R., J.V.G.T. and B.G.; formal analysis, P.A.C.R.; investigation, P.A.C.R., J.V.G.T. and B.G.; resources, J.V.G.T. and B.G.; data curation, J.V.G.T. and B.G.; writing—original draft preparation, V.O.S., F.P.M. and P.A.C.R.; writing—review and editing, V.O.S., F.P.M., P.A.C.R., J.V.G.T. and B.G.; visualization, V.O.S., F.P.M. and P.A.C.R.; supervision, J.V.G.T. and B.G.; project administration, J.V.G.T. and B.G.; funding acquisition, B.G. and J.V.G.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research study was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) Alliance, grant No. 401643, in association with Lakes Environmental Software Inc., and by the Conselho Nacional de Desenvolvimento Científico e Tecnológico—Brasil (CNPq), grant No. 303585/2022-6.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Conflicts of Interest

The author Jesse Van Griensven Thé is employed by the company Lakes Environmental. The remaining authors declare that this research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

References

  1. Biamonte, J.; Wittek, P.; Pancotti, N.; Rebentrost, P.; Wiebe, N.; Lloyd, S. Quantum Machine Learning. Nature 2017, 549, 195–202. [Google Scholar] [CrossRef] [PubMed]
  2. Sachdeva, N.; Harnett, G.S.; Maity, S.; Marsh, S.; Wang, Y.; Winick, A.; Dougherty, R.; Canuto, D.; Chong, Y.Q.; Hush, M.; et al. Quantum Optimization Using a 127-Qubit Gate-Model IBM Quantum Computer Can Outperform Quantum Annealers for Nontrivial Binary Optimization Problems. arXiv 2024, arXiv:2406.01743. [Google Scholar]
  3. Schuld, M.; Petruccione, F. Machine Learning with Quantum Computers; Springer Nature: Berlin/Heidelberg, Germany, 2021; ISBN 978-3-030-83098-4. [Google Scholar]
  4. Cerezo, M.; Verdon, G.; Huang, H.-Y.; Cincio, L.; Coles, P.J. Challenges and Opportunities in Quantum Machine Learning. Nat. Comput. Sci. 2022, 2, 567–576. [Google Scholar] [CrossRef]
  5. Brooks, M. The Race to Find Quantum Computing’s Sweet Spot. Available online: https://www.nature.com/articles/d41586-023-01692-9.pdf (accessed on 13 June 2024).
  6. Alchieri, L.; Badalotti, D.; Bonardi, P.; Bianco, S. An Introduction to Quantum Machine Learning: From Quantum Logic to Quantum Deep Learning. Quantum. Mach. Intell. 2021, 3, 28. [Google Scholar] [CrossRef]
  7. Guan, W.; Perdue, G.; Pesah, A.; Schuld, M.; Terashi, K.; Vallecorsa, S.; Vlimant, J.-R. Quantum Machine Learning in High Energy Physics. Mach. Learn. Sci. Technol. 2021, 2, 011003. [Google Scholar] [CrossRef]
  8. Vidal Bezerra, F.D.; Pinto Marinho, F.; Costa Rocha, P.A.; Oliveira Santos, V.; Van Griensven Thé, J.; Gharabaghi, B. Machine Learning Dynamic Ensemble Methods for Solar Irradiance and Wind Speed Predictions. Atmosphere 2023, 14, 1635. [Google Scholar] [CrossRef]
  9. Osman, A.I.; Chen, L.; Yang, M.; Msigwa, G.; Farghali, M.; Fawzy, S.; Rooney, D.W.; Yap, P.-S. Cost, Environmental Impact, and Resilience of Renewable Energy under a Changing Climate: A Review. Environ. Chem. Lett. 2023, 21, 741–764. [Google Scholar] [CrossRef]
  10. Walther, G.-R.; Post, E.; Convey, P.; Menzel, A.; Parmesan, C.; Beebee, T.J.C.; Fromentin, J.-M.; Hoegh-Guldberg, O.; Bairlein, F. Ecological Responses to Recent Climate Change. Nature 2002, 416, 389–395. [Google Scholar] [CrossRef] [PubMed]
  11. Duffy, P.B.; Field, C.B.; Diffenbaugh, N.S.; Doney, S.C.; Dutton, Z.; Goodman, S.; Heinzerling, L.; Hsiang, S.; Lobell, D.B.; Mickley, L.J.; et al. Strengthened Scientific Support for the Endangerment Finding for Atmospheric Greenhouse Gases. Science 2019, 363, eaat5982. [Google Scholar] [CrossRef]
  12. Zhang, W.; Zhou, T. Significant Increases in Extreme Precipitation and the Associations with Global Warming over the Global Land Monsoon Regions. J. Clim. 2019, 32, 8465–8488. [Google Scholar] [CrossRef]
  13. Ebi, K.L.; Vanos, J.; Baldwin, J.W.; Bell, J.E.; Hondula, D.M.; Errett, N.A.; Hayes, K.; Reid, C.E.; Saha, S.; Spector, J.; et al. Extreme Weather and Climate Change: Population Health and Health System Implications. Annu. Rev. Public Health 2021, 42, 293–315. [Google Scholar] [CrossRef] [PubMed]
  14. Bezirtzoglou, C.; Dekas, K.; Charvalos, E. Climate Changes, Environment and Infection: Facts, Scenarios and Growing Awareness from the Public Health Community within Europe. Anaerobe 2011, 17, 337–340. [Google Scholar] [CrossRef] [PubMed]
  15. Khoury, C.K.; Brush, S.; Costich, D.E.; Curry, H.A.; Haan, S.; Engels, J.M.M.; Guarino, L.; Hoban, S.; Mercer, K.L.; Miller, A.J.; et al. Crop Genetic Erosion: Understanding and Responding to Loss of Crop Diversity. New Phytol. 2022, 233, 84–118. [Google Scholar] [CrossRef] [PubMed]
  16. Dechezleprêtre, A.; Rivers, N.; Stadler, B. The Economic Cost of Air Pollution: Evidence from Europe; OECD Economics Department Working Papers; OECD: Paris, France, 2019; Volume 1584. [Google Scholar]
  17. Errigo, I.M.; Abbott, B.W.; Mendoza, D.L.; Mitchell, L.; Sayedi, S.S.; Glenn, J.; Kelly, K.E.; Beard, J.D.; Bratsman, S.; Carter, T.; et al. Human Health and Economic Costs of Air Pollution in Utah: An Expert Assessment. Atmosphere 2020, 11, 1238. [Google Scholar] [CrossRef]
  18. Fisher, S.; Bellinger, D.C.; Cropper, M.L.; Kumar, P.; Binagwaho, A.; Koudenoukpo, J.B.; Park, Y.; Taghian, G.; Landrigan, P.J. Air Pollution and Development in Africa: Impacts on Health, the Economy, and Human Capital. Lancet Planet. Health 2021, 5, e681–e688. [Google Scholar] [CrossRef]
  19. Chen, Z.; Wang, F.; Liu, B.; Zhang, B. Short-Term and Long-Term Impacts of Air Pollution Control on China’s Economy. Environ. Manag. 2022, 70, 536–547. [Google Scholar] [CrossRef]
  20. IRENA. Renewable Capacity Statistics 2024. 2024. Available online: https://www.irena.org/Publications/2024/Mar/Renewable-capacity-statistics-2024 (accessed on 30 May 2024).
  21. IEA. World Energy Investment 2023. 2023. Available online: https://www.iea.org/reports/world-energy-investment-2023 (accessed on 30 May 2024).
  22. Voyant, C.; Notton, G.; Kalogirou, S.; Nivet, M.-L.; Paoli, C.; Motte, F.; Fouilloy, A. Machine Learning Methods for Solar Radiation Forecasting: A Review. Renew. Energy 2017, 105, 569–582. [Google Scholar] [CrossRef]
  23. Rajagukguk, R.A.; Ramadhan, R.A.A.; Lee, H.-J. A Review on Deep Learning Models for Forecasting Time Series Data of Solar Irradiance and Photovoltaic Power. Energies 2020, 13, 6623. [Google Scholar] [CrossRef]
  24. Chu, Y.; Wang, Y.; Yang, D.; Chen, S.; Li, M. A Review of Distributed Solar Forecasting with Remote Sensing and Deep Learning. Renew. Sustain. Energy Rev. 2024, 198, 114391. [Google Scholar] [CrossRef]
  25. Yang, D.; Kleissl, J.; Gueymard, C.A.; Pedro, H.T.C.; Coimbra, C.F.M. History and Trends in Solar Irradiance and PV Power Forecasting: A Preliminary Assessment and Review Using Text Mining. Sol. Energy 2018, 168, 60–101. [Google Scholar] [CrossRef]
  26. Wang, W.; Du, Y.; Chau, K.; Chen, H.; Liu, C.; Ma, Q. A Comparison of BPNN, GMDH, and ARIMA for Monthly Rainfall Forecasting Based on Wavelet Packet Decomposition. Water 2021, 13, 2871. [Google Scholar] [CrossRef]
  27. Costa Rocha, P.A.; Oliveira Santos, V.; Scott, J.; Van Griensven Thé, J.; Gharabaghi, B. Application of Graph Neural Networks to Forecast Urban Flood Events: The Case Study of the 2013 Flood of the Bow River, Calgary, Canada. Int. J. River Basin Manag. 2024, 1–18. [Google Scholar] [CrossRef]
  28. Oliveira Santos, V.; Costa Rocha, P.A.; Scott, J.; Van Griensven Thé, J.; Gharabaghi, B. Spatiotemporal Air Pollution Forecasting in Houston-TX: A Case Study for Ozone Using Deep Graph Neural Networks. Atmosphere 2023, 14, 308. [Google Scholar] [CrossRef]
  29. Oliveira Santos, V.; Costa Rocha, P.A.; Thé, J.V.G.; Gharabaghi, B. Graph-Based Deep Learning Model for Forecasting Chloride Concentration in Urban Streams to Protect Salt-Vulnerable Areas. Environments 2023, 10, 157. [Google Scholar] [CrossRef]
  30. Oliveira Santos, V.; Guimarães, B.M.D.M.; Neto, I.E.L.; de Souza Filho, F.d.A.; Costa Rocha, P.A.; Thé, J.V.G.; Gharabaghi, B. Chlorophyll-a Estimation in 149 Tropical Semi-Arid Reservoirs Using Remote Sensing Data and Six Machine Learning Methods. Remote Sens. 2024, 16, 1870. [Google Scholar] [CrossRef]
  31. Rocha, P.A.C.; Santos, V.O. Global Horizontal and Direct Normal Solar Irradiance Modeling by the Machine Learning Methods XGBoost and Deep Neural Networks with CNN-LSTM Layers: A Case Study Using the GOES-16 Satellite Imagery. Int. J. Energy Environ. Eng. 2022, 13, 1271–1286. [Google Scholar] [CrossRef]
  32. Marinho, F.P.; Rocha, P.A.C.; Neto, A.R.R.; Bezerra, F.D.V. Short-Term Solar Irradiance Forecasting Using CNN-1D, LSTM, and CNN-LSTM Deep Neural Networks: A Case Study With the Folsom (USA) Dataset. J. Sol. Energy Eng. 2023, 145, 041002. [Google Scholar] [CrossRef]
  33. Carneiro, T.C.; Rocha, P.A.C.; Carvalho, P.C.M.; Fernández-Ramírez, L.M. Ridge Regression Ensemble of Machine Learning Models Applied to Solar and Wind Forecasting in Brazil and Spain. Appl. Energy 2022, 314, 118936. [Google Scholar] [CrossRef]
  34. Oliveira Santos, V.; Costa Rocha, P.A.; Scott, J.; Van Griensven Thé, J.; Gharabaghi, B. Spatiotemporal Analysis of Bidimensional Wind Speed Forecasting: Development and Thorough Assessment of LSTM and Ensemble Graph Neural Networks on the Dutch Database. Energy 2023, 278, 127852. [Google Scholar] [CrossRef]
  35. Abdullah-Vetter, Z.; Dwivedi, P.; Ekins-Daukes, N.J.; Trupke, T.; Hameiri, Z. Automated Analysis of Internal Quantum Efficiency Measurements of GaAs Solar Cells Using Machine Learning. In Proceedings of the 2023 IEEE 50th Photovoltaic Specialists Conference (PVSC), San Juan, Puerto Rico, 11–16 June 2023; pp. 1–3. [Google Scholar]
  36. Liu, H.; Tang, W. Quantum Computing for Power Systems: Tutorial, Review, Challenges, and Prospects. Electr. Power Syst. Res. 2023, 223, 109530. [Google Scholar] [CrossRef]
  37. Yousef, L.A.; Yousef, H.; Rocha-Meneses, L. Artificial Intelligence for Management of Variable Renewable Energy Systems: A Review of Current Status and Future Directions. Energies 2023, 16, 8057. [Google Scholar] [CrossRef]
  38. Pedro, H.T.C.; Larson, D.P.; Coimbra, C.F.M. A Comprehensive Dataset for the Accelerated Development and Benchmarking of Solar Forecasting Methods. J. Renew. Sustain. Energy 2019, 11, 036102. [Google Scholar] [CrossRef]
  39. Duffie, J.A.; Beckman, W.A.; Blair, N. Solar Engineering of Thermal Processes, Photovoltaics and Wind; John Wiley & Sons: Hoboken, NJ, USA, 2020; ISBN 978-1-119-54028-1. [Google Scholar]
  40. El Boujdaini, L.; Mezrhab, A.; Moussaoui, M.A. Artificial Neural Networks for Global and Direct Solar Irradiance Forecasting: A Case Study. Energy Sources Part A Recovery Util. Environ. Eff. 2021, 1–21. [Google Scholar] [CrossRef]
  41. Cavaco, A.; Canhoto, P.; Collares Pereira, M. Procedures for Solar Radiation Data Gathering and Processing and Their Application to DNI Assessment in Southern Portugal. Renew. Energy 2021, 163, 2208–2219. [Google Scholar] [CrossRef]
  42. Gupta, R.; Yadav, A.K.; Jha, S.; Pathak, P.K. Long Term Estimation of Global Horizontal Irradiance Using Machine Learning Algorithms. Optik 2023, 283, 170873. [Google Scholar] [CrossRef]
  43. Larson, D.P.; Li, M.; Coimbra, C.F.M. SCOPE: Spectral Cloud Optical Property Estimation Using Real-Time GOES-R Longwave Imagery. J. Renew. Sustain. Energy 2020, 12, 026501. [Google Scholar] [CrossRef]
  44. Garcia-Gutierrez, L.; Voyant, C.; Notton, G.; Almorox, J. Evaluation and Comparison of Spatial Clustering for Solar Irradiance Time Series. Appl. Sci. 2022, 12, 8529. [Google Scholar] [CrossRef]
  45. Ineichen, P.; Perez, R. A New Airmass Independent Formulation for the Linke Turbidity Coefficient. Sol. Energy 2002, 73, 151–157. [Google Scholar] [CrossRef]
  46. Lauret, P.; Alonso-Suárez, R.; Le Gal La Salle, J.; David, M. Solar Forecasts Based on the Clear Sky Index or the Clearness Index: Which Is Better? Solar 2022, 2, 432–444. [Google Scholar] [CrossRef]
  47. Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2019; ISBN 978-1-09-812246-1. [Google Scholar]
  48. James, G.; Witten, D.; Hastie, T.; Tibshirani, R.; Taylor, J. An Introduction to Statistical Learning: With Applications in Python; Springer International Publishing: Berlin/Heidelberg, Germany, 2023; ISBN 978-3-031-38746-3. [Google Scholar]
  49. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; ISBN 978-0-262-03561-3. [Google Scholar]
  50. Chollet, F. Deep Learning with Python, 2nd ed.; Simon and Schuster: Shelter Island, NY, USA, 2021; ISBN 978-1-63835-009-5. [Google Scholar]
  51. Tanveer, M.; Rajani, T.; Rastogi, R.; Shao, Y.H.; Ganaie, M.A. Comprehensive Review on Twin Support Vector Machines. Ann. Oper. Res. 2022. [Google Scholar] [CrossRef]
  52. Azimi, H.; Bonakdari, H.; Ebtehaj, I.; Gharabaghi, B.; Khoshbin, F. Evolutionary Design of Generalized Group Method of Data Handling-Type Neural Network for Estimating the Hydraulic Jump Roller Length. Acta Mech. 2018, 229, 1197–1214. [Google Scholar] [CrossRef]
  53. Elkurdy, M.; Binns, A.D.; Bonakdari, H.; Gharabaghi, B.; McBean, E. Early Detection of Riverine Flooding Events Using the Group Method of Data Handling for the Bow River, Alberta, Canada. Int. J. River Basin Manag. 2022, 20, 533–544. [Google Scholar] [CrossRef]
  54. Zaji, A.H.; Bonakdari, H.; Gharabaghi, B. Reservoir Water Level Forecasting Using Group Method of Data Handling. Acta Geophys. 2018, 66, 717–730. [Google Scholar] [CrossRef]
  55. Stajkowski, S.; Hotson, E.; Zorica, M.; Farghaly, H.; Bonakdari, H.; McBean, E.; Gharabaghi, B. Modeling Stormwater Management Pond Thermal Impacts during Storm Events. J. Hydrol. 2023, 620, 129413. [Google Scholar] [CrossRef]
  56. Bonakdari, H.; Ebtehaj, I.; Gharabaghi, B.; Vafaeifard, M.; Akhbari, A. Calculating the Energy Consumption of Electrocoagulation Using a Generalized Structure Group Method of Data Handling Integrated with a Genetic Algorithm and Singular Value Decomposition. Clean Techn. Environ. Policy 2019, 21, 379–393. [Google Scholar] [CrossRef]
  57. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, San Francisco, CA, USA, 13 August 2016; pp. 785–794. [Google Scholar]
  58. Nielsen, M.A.; Chuang, I.L. Quantum Computation and Quantum Information, 10th ed.; Cambridge University Press: Cambridge, MA, USA; New York, NY, USA, 2010; ISBN 978-1-107-00217-3. [Google Scholar]
  59. Combarro, E.F.; Gonzalez-Castillo, S.; Meglio, A.D. A Practical Guide to Quantum Machine Learning and Quantum Optimization: Hands-on Approach to Modern Quantum Algorithms; Packt Publishing Ltd.: Birmingham, UK, 2023; ISBN 978-1-80461-830-1. [Google Scholar]
  60. Sutor, R.S. Dancing with Qubits: How Quantum Computing Works and How It May Change the World; Expert insight; Packt: Birmingham, UK; Mumbai, India, 2019; ISBN 978-1-83882-736-6. [Google Scholar]
  61. Aïmeur, E.; Brassard, G.; Gambs, S. Machine Learning in a Quantum World. In Proceedings of the Advances in Artificial Intelligence; Tawfik, A.Y., Goodwin, S.D., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; Volume 3060, pp. 431–442. [Google Scholar]
  62. Lau, J.W.Z.; Lim, K.H.; Shrotriya, H.; Kwek, L.C. NISQ Computing: Where Are We and Where Do We Go? AAPPS Bull. 2022, 32, 27. [Google Scholar] [CrossRef]
  63. Li, G.; Ding, Y.; Xie, Y. Tackling the Qubit Mapping Problem for NISQ-Era Quantum Devices. In Proceedings of the Twenty-Fourth International Conference on Architectural Support for Programming Languages and Operating Systems, ACM, Providence, RI, USA, 4 April 2019; pp. 1001–1014. [Google Scholar]
  64. Ullah, M.H.; Eskandarpour, R.; Zheng, H.; Khodaei, A. Quantum Computing for Smart Grid Applications. IET Gener. Trans. Dist. 2022, 16, 4239–4257. [Google Scholar] [CrossRef]
  65. Orazi, F.; Gasperini, S.; Lodi, S.; Sartori, C. Hybrid Quantum Technologies for Quantum Support Vector Machines. Information 2024, 15, 72. [Google Scholar] [CrossRef]
  66. Zheng, K.; Van Griensven, J.; Fraser, R. A Quantum Machine Learning Approach to Spatiotemporal Emission Modelling. Atmosphere 2023, 14, 944. [Google Scholar] [CrossRef]
  67. Quantum Neural Networks—Qiskit Machine Learning 0.7.2. Available online: https://qiskit-community.github.io/qiskit-machine-learning/tutorials/01_neural_networks.html (accessed on 10 June 2024).
  68. Schuld, M.; Killoran, N. Quantum Machine Learning in Feature Hilbert Spaces. Phys. Rev. Lett. 2019, 122, 040504. [Google Scholar] [CrossRef] [PubMed]
  69. Mengoni, R.; Di Pierro, A. Kernel Methods in Quantum Machine Learning. Quantum Mach. Intell. 2019, 1, 65–71. [Google Scholar] [CrossRef]
  70. Yu, Y.; Hu, G.; Liu, C.; Xiong, J.; Wu, Z. Prediction of Solar Irradiance One Hour Ahead Based on Quantum Long Short-Term Memory Network. IEEE Trans. Quantum. Eng. 2023, 4, 1–15. [Google Scholar] [CrossRef]
  71. Sierra-Sosa, D.; Telahun, M.; Elmaghraby, A. TensorFlow Quantum: Impacts of Quantum State Preparation on Quantum Machine Learning Performance. IEEE Access 2020, 8, 215246–215255. [Google Scholar] [CrossRef]
  72. PauliFeatureMap. Available online: https://docs.quantum.ibm.com/api/qiskit/qiskit.circuit.library.PauliFeatureMap (accessed on 12 June 2024).
  73. Havlicek, V.; Córcoles, A.D.; Temme, K.; Harrow, A.W.; Kandala, A.; Chow, J.M.; Gambetta, J.M. Supervised Learning with Quantum Enhanced Feature Spaces. Nature 2019, 567, 209–212. [Google Scholar] [CrossRef]
  74. Variational Circuits—PennyLane. Available online: https://pennylane.ai/qml/glossary/variational_circuit/ (accessed on 3 June 2024).
  75. NLocal. Available online: https://docs.quantum.ibm.com/api/qiskit/qiskit.circuit.library.NLocal (accessed on 12 June 2024).
  76. TwoLocal. Available online: https://docs.quantum.ibm.com/api/qiskit/qiskit.circuit.library.TwoLocal (accessed on 12 June 2024).
  77. Funcke, L.; Hartung, T.; Heinemann, B.; Jansen, K.; Kropf, A.; Kühn, S.; Meloni, F.; Spataro, D.; Tüysüz, C.; Chinn Yap, Y. Studying Quantum Algorithms for Particle Track Reconstruction in the LUXE Experiment. J. Phys. Conf. Ser. 2023, 2438, 012127. [Google Scholar] [CrossRef]
  78. Farhi, E.; Goldstone, J.; Gutmann, S. A Quantum Approximate Optimization Algorithm. arXiv 2014, arXiv:1411.4028. [Google Scholar]
  79. VQE. Available online: https://docs.quantum.ibm.com/api/qiskit/0.26/qiskit.algorithms.VQE (accessed on 12 June 2024).
  80. QAOA. Available online: https://docs.quantum.ibm.com/api/qiskit/0.26/qiskit.algorithms.QAOA (accessed on 12 June 2024).
  81. Sushmit, M.M.; Mahbubul, I.M. Forecasting Solar Irradiance with Hybrid Classical–Quantum Models: A Comprehensive Evaluation of Deep Learning and Quantum-Enhanced Techniques. Energy Convers. Manag. 2023, 294, 117555. [Google Scholar] [CrossRef]
  82. COBYLA. Available online: https://docs.quantum.ibm.com/api/qiskit/0.26/qiskit.algorithms.optimizers.COBYLA (accessed on 12 June 2024).
  83. Quick, J.; Annoni, J.; King, R.; Dykes, K.; Fleming, P.; Ning, A. Optimization Under Uncertainty for Wake Steering Strategies. J. Phys.: Conf. Ser. 2017, 854, 012036. [Google Scholar] [CrossRef]
  84. Zhang, Z.; Zhang, C.; Qiao, Y.; Zhou, Y.; Wang, S. Design and Mass Optimization of Numerical Models for Composite Wind Turbine Blades. JMSE 2023, 11, 75. [Google Scholar] [CrossRef]
  85. Miki, T.; Tsukayama, D.; Okita, R.; Shimada, M.; Shirakashi, J. Variational Parameter Optimization of Quantum-Classical Hybrid Heuristics on Near-Term Quantum Computer. In Proceedings of the 2022 IEEE International Conference on Manipulation, Manufacturing and Measurement on the Nanoscale (3M-NANO), Tianjin, China, 8 August 2022; pp. 415–418. [Google Scholar]
  86. Miháliková, I.; Friák, M.; Pivoluska, M.; Plesch, M.; Saip, M.; Šob, M. Best-Practice Aspects of Quantum-Computer Calculations: A Case Study of the Hydrogen Molecule. Molecules 2022, 27, 597. [Google Scholar] [CrossRef]
  87. Dalvand, Z.; Hajarian, M. Solving Generalized Inverse Eigenvalue Problems via L-BFGS-B Method. Inverse Probl. Sci. Eng. 2020, 28, 1719–1746. [Google Scholar] [CrossRef]
  88. L_BFGS_B. Available online: https://docs.quantum.ibm.com/api/qiskit/0.37/qiskit.algorithms.optimizers.L_BFGS_B (accessed on 12 June 2024).
  89. Wilson, M.; Stromswold, R.; Wudarski, F.; Hadfield, S.; Tubman, N.M.; Rieffel, E.G. Optimizing Quantum Heuristics with Meta-Learning. Quantum Mach. Intell. 2021, 3, 13. [Google Scholar] [CrossRef]
  90. Li, J.; Alam, M.; Saki, A.A.; Ghosh, S. Hierarchical Improvement of Quantum Approximate Optimization Algorithm for Object Detection: (Invited Paper). In Proceedings of the 2020 21st International Symposium on Quality Electronic Design (ISQED), Santa Clara, CA, USA, 25–26 March 2020; pp. 335–340. [Google Scholar]
  91. Gacon, J.; Zoufal, C.; Carleo, G.; Woerner, S. Simultaneous Perturbation Stochastic Approximation of the Quantum Fisher Information. Quantum 2021, 5, 567. [Google Scholar] [CrossRef]
  92. QNSPSA. Available online: https://docs.quantum.ibm.com/api/qiskit/0.40/qiskit.algorithms.optimizers.QNSPSA (accessed on 12 June 2024).
  93. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  94. Javadi-Abhari, A.; Treinish, M.; Krsulich, K.; Wood, C.J.; Lishman, J.; Gacon, J.; Martiel, S.; Nation, P.D.; Bishop, L.S.; Cross, A.W.; et al. Quantum Computing with Qiskit. 2024. [Google Scholar]
  95. Bergholm, V.; Izaac, J.; Schuld, M.; Gogolin, C.; Ahmed, S.; Ajith, V.; Alam, M.S.; Alonso-Linaje, G.; AkashNarayanan, B.; Asadi, A.; et al. PennyLane: Automatic Differentiation of Hybrid Quantum-Classical Computations. arXiv 2018, arXiv:1811.04968, 4968. [Google Scholar]
  96. Correa-Jullian, C.; Cofre-Martel, S.; San Martin, G.; Lopez Droguett, E.; De Novaes Pires Leite, G.; Costa, A. Exploring Quantum Machine Learning and Feature Reduction Techniques for Wind Turbine Pitch Fault Detection. Energies 2022, 15, 2792. [Google Scholar] [CrossRef]
  97. Sagingalieva, A.; Komornyik, S.; Senokosov, A.; Joshi, A.; Sedykh, A.; Mansell, C.; Tsurkan, O.; Pinto, K.; Pflitsch, M.; Melnikov, A. Photovoltaic Power Forecasting Using Quantum Machine Learning. arXiv 2023, arXiv:2312.16379. [Google Scholar]
  98. Jing, H.; Li, Y.; Brandsema, M.J.; Chen, Y.; Yue, M. HHL Algorithm with Mapping Function and Enhanced Sampling for Model Predictive Control in Microgrids. Appl. Energy 2024, 361, 122878. [Google Scholar] [CrossRef]
  99. Senekane, M.; Taele, B.M. Prediction of Solar Irradiation Using Quantum Support Vector Machine Learning Algorithm. SGRE 2016, 7, 293–301. [Google Scholar] [CrossRef]
  100. Ayoade, O.; Rivas, P.; Orduz, J. Artificial Intelligence Computing at the Quantum Level. Data 2022, 7, 28. [Google Scholar] [CrossRef]
  101. Li, X.; Ma, L.; Chen, P.; Xu, H.; Xing, Q.; Yan, J.; Lu, S.; Fan, H.; Yang, L.; Cheng, Y. Probabilistic Solar Irradiance Forecasting Based on XGBoost. Energy Rep. 2022, 8, 1087–1095. [Google Scholar] [CrossRef]
  102. Phan, Q.-T.; Wu, Y.-K.; Phan, Q.-D. Short-Term Solar Power Forecasting Using XGBoost with Numerical Weather Prediction. In Proceedings of the 2021 IEEE International Future Energy Electronics Conference (IFEEC), Taipei, Taiwan, 16 November 2021; pp. 1–6. [Google Scholar]
  103. Didavi, A.B.K.; Agbokpanzo, R.G.; Agbomahena, M. Comparative Study of Decision Tree, Random Forest and XGBoost Performance in Forecasting the Power Output of a Photovoltaic System. In Proceedings of the 2021 4th International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris/Créteil, France, 8 December 2021; pp. 1–5. [Google Scholar]
  104. Grinsztajn, L.; Oyallon, E.; Varoquaux, G. Why Do Tree-Based Models Still Outperform Deep Learning on Typical Tabular Data?
  105. Uddin, S.; Lu, H. Confirming the Statistically Significant Superiority of Tree-Based Machine Learning Algorithms over Their Counterparts for Tabular Data. PLoS ONE 2024, 19, e0301541. [Google Scholar] [CrossRef] [PubMed]
  106. Ajagekar, A.; You, F. Quantum Computing for Energy Systems Optimization: Challenges and Opportunities. Energy 2019, 179, 76–89. [Google Scholar] [CrossRef]
  107. Oliveira Santos, V.; Costa Rocha, P.A.; Scott, J.; Van Griensven Thé, J.; Gharabaghi, B. A New Graph-Based Deep Learning Model to Predict Flooding with Validation on a Case Study on the Humber River. Water 2023, 15, 1827. [Google Scholar] [CrossRef]
  108. Atmani, H.; Bouzgou, H.; Gueymard, C. Intra-Hour Forecasting of Direct Normal Solar Irradiance Using Variable Selection with Artificial Neural Networks; Springer: Cham, Switzerland, 2018. [Google Scholar]
Figure 1. Different ways to combine classical and quantum computing strategies to perform quantum machine learning.
Figure 1. Different ways to combine classical and quantum computing strategies to perform quantum machine learning.
Energies 17 03580 g001
Figure 2. Generic QNN architecture.
Figure 2. Generic QNN architecture.
Energies 17 03580 g002
Figure 3. Generic angle encoding mapping approach. The rotational gate is applied to each one of the dimensions of the dataset.
Figure 3. Generic angle encoding mapping approach. The rotational gate is applied to each one of the dimensions of the dataset.
Energies 17 03580 g003
Figure 4. Generic ZZ feature map encoding approach.
Figure 4. Generic ZZ feature map encoding approach.
Energies 17 03580 g004
Figure 5. Generic parameterized variational quantum circuit for the ansatz structure.
Figure 5. Generic parameterized variational quantum circuit for the ansatz structure.
Energies 17 03580 g005
Figure 6. A Two Local Ansatz with linear entanglement and two repetitions.
Figure 6. A Two Local Ansatz with linear entanglement and two repetitions.
Energies 17 03580 g006
Figure 7. A generic QAOA ansatz quantum circuit.
Figure 7. A generic QAOA ansatz quantum circuit.
Energies 17 03580 g007
Figure 8. Pipeline for the parameter selection, using as an example the 5 min forecasting horizon. (a) The complete set of attributes. (b) The random forest strategy to select the best subset for the 5 min forecasting horizon. (c) The selected subset of attributes is then used to train and validate the models (d).
Figure 8. Pipeline for the parameter selection, using as an example the 5 min forecasting horizon. (a) The complete set of attributes. (b) The random forest strategy to select the best subset for the 5 min forecasting horizon. (c) The selected subset of attributes is then used to train and validate the models (d).
Energies 17 03580 g008
Figure 9. Graph showing the progress of the objective function value versus the iteration step using Pauli Y as feature map angle encoding strategy, Pauli Y as ansatz, and COBYLA as optimizer.
Figure 9. Graph showing the progress of the objective function value versus the iteration step using Pauli Y as feature map angle encoding strategy, Pauli Y as ansatz, and COBYLA as optimizer.
Energies 17 03580 g009
Figure 10. Graph showing the progress of the objective function value versus the iteration step using ZZ feature map strategy, Pauli Y as ansatz, and COBYLA as optimizer.
Figure 10. Graph showing the progress of the objective function value versus the iteration step using ZZ feature map strategy, Pauli Y as ansatz, and COBYLA as optimizer.
Energies 17 03580 g010
Figure 11. Graph showing the progress of the objective function value versus the iteration step using Pauli Y feature map strategy, Two Local as ansatz, and COBYLA as optimizer.
Figure 11. Graph showing the progress of the objective function value versus the iteration step using Pauli Y feature map strategy, Two Local as ansatz, and COBYLA as optimizer.
Energies 17 03580 g011
Figure 12. Graph showing the progress of the objective function value versus the iteration step using ZZ feature map strategy, Two Local as ansatz, and COBYLA as optimizer.
Figure 12. Graph showing the progress of the objective function value versus the iteration step using ZZ feature map strategy, Two Local as ansatz, and COBYLA as optimizer.
Energies 17 03580 g012
Figure 13. Graph showing the progress of the objective function value versus the iteration step using Pauli Y feature map strategy, QAOA as ansatz, and COBYLA as optimizer.
Figure 13. Graph showing the progress of the objective function value versus the iteration step using Pauli Y feature map strategy, QAOA as ansatz, and COBYLA as optimizer.
Energies 17 03580 g013
Figure 14. Graph showing the progress of the objective function value versus the iteration step using the best QNN configuration achieved.
Figure 14. Graph showing the progress of the objective function value versus the iteration step using the best QNN configuration achieved.
Energies 17 03580 g014
Figure 15. The quantum circuit selected for the QNN model. The feature map is composed of Pauli Y angle encoding, using the Two Local ansatz with Pauli Y rotations, controlled-NOT gates for linear entanglement, and one repetition, with L-BFGS-B as the optimizer.
Figure 15. The quantum circuit selected for the QNN model. The feature map is composed of Pauli Y angle encoding, using the Two Local ansatz with Pauli Y rotations, controlled-NOT gates for linear entanglement, and one repetition, with L-BFGS-B as the optimizer.
Energies 17 03580 g015
Table 1. The attributes and their symbols as used in the present study.
Table 1. The attributes and their symbols as used in the present study.
Attribute NameAttribute Symbol
Backward AverageBi
Lagged AverageLi
Clear-sky VariabilityVi
Red ChannelR
Green ChannelG
Blue ChannelB
Red-to-Blue RatioRB
Normalized Red-to-Blue RatioNRB
AverageAVG
Standard DeviationSTD
EntropyENT
Table 2. Selected parameters using RFE.
Table 2. Selected parameters using RFE.
Selected Features
Forecasting Horizon
(min)
GHIDNI
5Bi, Vi, Li, ENT (R), ENT (RB)Bi, Vi, Li, ENT (G), ENT (RB)
10Bi, Vi, Li, ENT (G), ENT (RB)Bi, Vi, Li, ENT (G), ENT (RB)
15Bi, Vi, AVG (G), ENT (B), ENT (RB)Bi, Vi, ENT (R), AVG (G), ENT (RB)
20Bi, Vi, AVG (G), ENT (G), ENT (RB)Bi, Vi, ENT (R), AVG (G), ENT (RB)
25Bi, Vi, ENT (R), AVG (G), ENT (RB)Bi, Vi, ENT (R), AVG (G), ENT (RB)
30Bi, Vi, ENT (R), AVG (G), ENT (RB)Bi, Vi, ENT (R), AVG (G), ENT (RB)
60Bi, Vi, AVG (R), ENT (G), ENT (RB)Bi, Vi, AVG (G), ENT (G), ENT (RB)
120Bi, Vi, AVG (G), ENT (G), ENT (RB)Vi, ENT (R), AVG (G), AVG (B), ENT (RB)
180Bi, Li, AVG (G), ENT (B), ENT (RB)Vi, Li, AVG (R), ENT (R), ENT (RB)
Table 3. Results for different configurations of the QNN model using Pauli Y as feature map angle encoding and different ansatz and optimizers for 5 min GHI forecasting.
Table 3. Results for different configurations of the QNN model using Pauli Y as feature map angle encoding and different ansatz and optimizers for 5 min GHI forecasting.
Pauli Y Feature Map
Ansatz, OptimizerRMSE
(W/m2)
R2
Ry, COBYLA 56.9195.99%
Ry, L BFGS B56.7396.01%
Ry, QN-SPSA 56.5596.04%
Two Local, COBYLA 58.0895.82%
Two Local, L-BFGS-B48.9897.03%
Two Local, QN-SPSA 63.9094.94%
QAOA, COBYLA 498.86−2.08
QAOA, L-BFGS-B 498.85−2.08
QAOA, QN-SPSA 501.21−2.11
Table 4. Results for GHI forecasting by the classical and quantum ML models for all the forecasting horizons.
Table 4. Results for GHI forecasting by the classical and quantum ML models for all the forecasting horizons.
GHI
Forecasting Time
(min)
AlgorithmRMSE
(W/m2)
R2
(%)
Forecast Skill
(%)
5SVR35.9498.4092.33
XGBoost28.7498.9893.87
GMDH34.8298.5092.58
QNN48.9897.0389.55
10SVR43.0097.7190.82
XGBoost32.20 98.7293.13
GMDH45.5697.4390.28
QNN54.4896.3388.38
15SVR50.1796.9089.29
XGBoost40.6897.9691.31
GMDH51.6296.7288.98
QNN65.56 94.7186.00
20SVR52.69 96.6088.74
XGBoost42.69 97.7790.88
GMDH54.15 96.4188.42
QNN65.4094.7686.02
25SVR55.11 96.3088.20
XGBoost44.23 97.6290.53
GMDH57.57 95.9787.68
QNN63.09 95.1686.50
30SVR57.0096.0887.80
XGBoost45.7097.4890.20
GMDH58.30 95.9087.50
QNN63.97 95.0686.30
60SVR55.6396.3288.00
XGBoost34.7698.5692.50
GMDH61.0895.5686.82
QNN61.2095.5486.80
120SVR69.9194.9984.18
XGBoost54.6796.9387.63
GMDH74.9194.2683.05
QNN66.2495.5085.01
180SVR81.9494.0079.84
XGBoost58.9396.9085.50
GMDH81.9094.0179.85
QNN77.5594.6380.92
Table 5. Results for DNI forecasting by the classical and quantum ML models for all the forecasting horizons.
Table 5. Results for DNI forecasting by the classical and quantum ML models for all the forecasting horizons.
DNI
Forecasting Time
(min)
AlgorithmRMSE
(W/m2)
R2
(%)
Forecast Skill
(%)
5SVR80.3193.3387.16
XGBoost61.9696.0390.09
GMDH72.0894.6388.47
QNN103.0389.0283.52
10SVR95.1790.7084.75
XGBoost77.0893.9087.65
GMDH92.4991.2285.18
QNN117.2185.8981.22
15SVR105.2588.7783.08
XGBoost87.7692.1985.90
GMDH105.8688.6482.99
QNN127.1583.6179.57
20SVR112.3187.4381.88
XGBoost93.6991.2584.89
GMDH114.7286.8881.49
QNN130.7682.9678.91
25SVR117.5886.4980.95
XGBoost98.1990.5884.09
GMDH121.7885.5180.27
QNN133.7082.5378.34
30SVR122.8485.5480.00
XGBoost102.3389.9783.34
GMDH127.7984.3679.20
QNN136.3482.1977.81
60SVR123.7785.6979.53
XGBoost84.3393.3586.05
GMDH134.9682.9877.68
QNN157.2376.9173.99
120SVR172.0577.3169.11
XGBoost107.2491.1980.74
GMDH168.1678.3369.80
QNN243.0656.3554.72
180SVR192.5874.8161.46
XGBoost145.5485.6170.87
GMDH177.8878.5164.40
QNN168.8180.6466.22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Oliveira Santos, V.; Marinho, F.P.; Costa Rocha, P.A.; Thé, J.V.G.; Gharabaghi, B. Application of Quantum Neural Network for Solar Irradiance Forecasting: A Case Study Using the Folsom Dataset, California. Energies 2024, 17, 3580. https://doi.org/10.3390/en17143580

AMA Style

Oliveira Santos V, Marinho FP, Costa Rocha PA, Thé JVG, Gharabaghi B. Application of Quantum Neural Network for Solar Irradiance Forecasting: A Case Study Using the Folsom Dataset, California. Energies. 2024; 17(14):3580. https://doi.org/10.3390/en17143580

Chicago/Turabian Style

Oliveira Santos, Victor, Felipe Pinto Marinho, Paulo Alexandre Costa Rocha, Jesse Van Griensven Thé, and Bahram Gharabaghi. 2024. "Application of Quantum Neural Network for Solar Irradiance Forecasting: A Case Study Using the Folsom Dataset, California" Energies 17, no. 14: 3580. https://doi.org/10.3390/en17143580

APA Style

Oliveira Santos, V., Marinho, F. P., Costa Rocha, P. A., Thé, J. V. G., & Gharabaghi, B. (2024). Application of Quantum Neural Network for Solar Irradiance Forecasting: A Case Study Using the Folsom Dataset, California. Energies, 17(14), 3580. https://doi.org/10.3390/en17143580

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop