Next Article in Journal
Analytic Study on Φ-Hilfer Fractional Neutral-Type Functional Integro-Differential Equations with Terminal Conditions
Previous Article in Journal
Artificial Intelligence for Studying Interactions of Solitons and Peakons
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantum-Enhanced Residual Convolutional Attention Architecture for Renewable Forecasting in Off-Grid Cloud Microgrids

Department of Information Technology, College of Computer and Information Sciences, Majmaah University, Al Majmaah 11952, Saudi Arabia
Mathematics 2026, 14(1), 181; https://doi.org/10.3390/math14010181
Submission received: 6 December 2025 / Revised: 28 December 2025 / Accepted: 30 December 2025 / Published: 3 January 2026

Abstract

Multimodal forecasting is increasingly needed to maintain energy levels, storage capacity, and compute efficiency in off-grid, renewable-powered cloud environments. Variable sensor quality, uncertain interactions with renewable energy, and rapidly changing weather patterns make real-time forecasting difficult. Current transformer, GNN, and CNN systems suffer from sensor noise instability, multimodal temporal–spectral correlation issues, and challenges in the interpretability of operational decision-making. In this research, Q-RCANeX, a quantum-guided residual convolutional attention network for off-grid cloud infrastructures, estimates battery state of charge, renewable energy sources, and microgrid efficiency to overcome these restrictions. The system uses a Hybrid Quantum–Bayesian Evolutionary Optimizer, quantum feature embedding, temporal–spectral attention, residual convolutional encoding, and signal decomposition preprocessing. These parameters reinforce features, reduce noise, and align forecasting behavior with microgrid dynamics. Q-RCANeX obtains 98.6% accuracy, 0.992 AUC, and 0.986 R3 values for REAF, WGF, SOC-F, and EEIF forecasting tasks, according to a statistical study. Additionally, it determines inference latency to 4.9 ms and model size to 18.5 MB. Even with 20% of sensor data missing or noisy, the model outperforms 12 state-of-the-art baselines and maintains 96.8% accuracy using ANOVA, Wilcoxon, Nemenyi, and Holm tests. The findings indicate that the forecasting framework has high accuracy, clarity, and resilience to failures. This makes it useful for real-time, off-grid management of renewable cloud microgrids.

1. Introduction

Cloud computing and AI workloads are increasing energy consumption in data centers, emphasizing the need for decarbonization and a reliable power supply [1]. Modern, energy-efficient operation is as crucial as powerful processing. Intelligent power orchestration, high-efficiency cooling, and renewable integration are becoming design priorities. Off-grid and near-grid microgrids, which employ solar, wind, and battery storage, host computer technology in areas where the grid is inaccessible or regularly down. These arrangements extend product life and reduce emissions. The system is still vulnerable to fluctuations in renewable generation and laws governing storage cycling and peak power distribution [2]. In these circumstances, estimation accuracy affects operations and the economy. With precise short-term estimates of renewable output and load demand, microgrids without grid backup can be more reliable and cost-effective. Inverter timing, battery location, and curtailment reduction are optimized. Inaccurate forecasting increases operational costs and uncertainties via unit commitment and demand response choices, according to studies. As predicted accuracy increases, system stability and resource efficiency also improve [3]. Detecting changes in hybrid solar–wind systems driven by nonlinear interactions between power-generating types is difficult with linear or shallow models [4].
Deep learning has enhanced renewable energy system forecasting by capturing long-term temporal dependencies, non-stationary seasonal changes, and complex feature interactions across meteorological and system data. Research indicates that hybrid architectures combining convolutional layers for local pattern extraction and recurrent or attention-based processes for contextual reasoning outperform statistical baselines in varied climates and forecast horizons [5]. Attention systems successfully detect feature relevance across irregular time intervals and diverse sensor modalities, demonstrating resilience to noisy or incomplete data [6]. To ensure transparent and accountable decision-making, energy operators increasingly prioritize explainability over predictive precision, understanding how key factors such as irradiance fluctuations, wind shear, temperature variations, and battery state of charge affect model outputs. Due to operational dynamics, data-center energy forecasting is difficult. Workload fluctuations, cooling system behavior, and control measures such as server throttling and thermal management changes create nonlinear relationships between IT demand and electrical load that defy standard forecast models. Recent data-center energy management surveys emphasize the need for forecasting frameworks that reduce prediction errors and offer interpretable outputs for supervisory control systems [7]. Forecasting errors in off-grid settings are expensive, as underestimating renewable shortages can cause battery depletion or load shedding, while overestimating renewable shortages can waste energy and lower service quality [8].
Recent research has focused on learning paradigms at the microgrid edge, where data are produced, to overcome these issues. Real-time forecasting in power-constrained situations is being examined using federated and distributed learning methodologies, as well as model compression techniques [9]. Energy analytics in smart city settings with diverse data sources has recently demonstrated strong scalability and privacy protection through federated learning-based forecasting models. An instance of this is HHCTE-FL, which integrates interpretable federated optimization with hybrid convolutional–transformer extractors [10]. Integrated pipelines that include generation forecasting, battery state-of-charge estimates, and health-aware diagnostics have shown that storage behavior determines forecasting reliability. Off-grid cloud forecasting is multimodal and context-aware due to the tight linkage of renewable production, storage response, and computational demand. In parallel, quantum-inspired and quantum-native learning paradigms are promising for complicated energy analytics. Quantum machine learning (QML) offers benefits for high-dimensional, noisy time-series data in energy systems through feature encoding and optimization based on superposition and entanglement principles [11]. These early-stage approaches enhance transformer-based and hybrid convolutional-recurrent architectures for wind and solar prediction by exploring richer representational spaces and non-convex optimization landscapes [12]. Next-generation predictive intelligence in renewable-powered systems is being built on temporal–spectral reasoning, interpretable deep forecasting, and quantum-inspired optimization.
The primary scientific problem is clear in this context. High-frequency, short-horizon forecasting frameworks are needed for off-grid cloud data centers with inadequate sensing and changing operating states. These models must be energy-aware, computationally efficient for edge deployment, sensitive to forecasting errors that affect storage stability and demand-side performance, and explainable to provide transparency in energy management and decision support. These demands need a cohesive methodology that blends resilient signal preprocessing, adaptive temporal–spectral modeling, and interpretable reasoning, aligning with microgrid forecasting, sustainable data-center optimization, and explainable machine learning. The following are the contributions of this work:
  • Introduce the Q-RCANeX architecture, a solution to the issue of unstable and noisy multimodal energy data leading to inaccurate renewable energy predictions. The goal of the technique is to find stable physical patterns in streams of heterogeneous solar, wind, battery, and thermal data by integrating quantum-aware feature representation with residual convolutional encoding.
  • Introduces a novel method, a temporal–spectral attention mechanism, to address the inadequacy of standard CNN-, LSTM-, and GNN-based models in capturing long-range dependencies and cross-modal interactions. This method facilitates the acquisition of frequency-aligned energy changes and fine-grained temporal dynamics.
  • Incorporates a quantum feature embedding layer that addresses the problem of nonlinear renewable profiles’ poor separability and error amplification in the absence of data or due to distorted sensors. By amplifying the discriminative information in the feature space inspired by quantum mechanics, the embedding strengthens the system.
  • Introduces a novel algorithm named a Hybrid Quantum–Bayesian Evolutionary Optimizer (Q-BEO) to stabilize hyperparameter tuning and reduce convergence issues frequent in deep renewable forecasting. This optimizer improves generalizability, reduces overfitting, and facilitates smoother convergence for all forecasting workloads.
  • Provides off-grid cloud microgrids with a unified forecasting model that supports operational concerns such as task scheduling, battery charge/discharge management, and thermal efficiency regulation in response to shifting renewable supplies. To support sustainability-oriented decision-making, the method produces reliable multi-horizon predictions for the REAF, WGF, SOC-F, EEIF, and renewable adequacy classes.
From a deployment perspective, Q-RCANeX is constructed to function in actual off-grid cloud microgrid systems, which often have constraints on computing power, communication bandwidth, and data quality. With its modular architecture, it picks and chooses which key components to deploy to the cloud and edge. This paves the way for predictions that may evolve as operating conditions change. For this reason, the framework is well-suited for mission-critical off-grid energy systems that require robust decision support, such as hospital microgrids and remote industrial sites.
The article is organized as follows: The Section 2 covers the most current advancements in renewable forecasting, microgrid intelligence, and quantum-assisted learning. Also shown are the shortcomings of current transformer, GNN, and CNN-based algorithms when applied to off-grid cloud environments. Section 3 provides further details on the proposed Q-RCANeX framework. The Hybrid Quantum–Bayesian Evolutionary Optimizer is a part of it, along with architectural components, a design for temporal–spectral attention, the embedding of quantum features, and a preprocessing pipeline. Experimental testing and a comparison of computational efficiency with the best baselines are detailed in Section 4. It also covers statistical significance tests, ablation studies, robustness analyses, and prediction accuracy. The Section 5 concludes the article and provides further future work. Section 5 wraps up the study and describes where the research is headed, with the goals of improving scalability, increasing the deployment of microgrids across multiple sites, and incorporating adaptive, energy-aware control into real-world off-grid cloud technologies.

2. Related Work

Recent advances in renewable energy analytics have focused on developing short-horizon forecasting models for hybrid solar–wind microgrids in changeable and unpredictable environments. In one transformer-based multi-step photovoltaic (PV) forecasting investigation, horizon-specific decoders and attention layers captured daily energy trends. This method greatly reduced predicting errors in many climates, although accuracy declined in noisy situations due to missing telemetry data or sensor deterioration [13]. Another study created a hybrid transformer pipeline that learns solar–wind interdependence. Although this technique adequately recorded nonlinear temporal correlations between the two modalities, it depended largely on high-quality weather reanalysis data, limiting its application in data-scarce locations [14,15]. Before model training, researchers investigated a signal-augmentation and quality-control (SAQ) approach to stabilize microgrid data. While this pipeline improved short-term PV and load forecasting accuracy and decreased curtailment, it also required substantial preprocessing, which raised system complexity and hindered real-time scalability [16]. Another study used clear-sky indices, solar transposition models, and meteorological interaction characteristics for physics-informed feature engineering. While the approach outperformed deep learning models in accuracy and computational demand, its site-specific calibration hampered cross-regional adaptability [17].
A Temporal Convolutional Network (TCN)-based solution was presented to capture long- and short-term interdependence in hybrid PV–wind datasets to address renewable energy production’s temporal dynamics. Although our technique outperformed typical LSTM and CNN architectures in ramp prediction and signal stability, it lacked a way to fuse multimodal weather and generation data, limiting its flexibility in novel situations [18]. A graph-temporal learning method uses graph neural networks (GNNs) and recurrent units to leverage spatial correlations among scattered wind stations. While this hybrid technique enhanced forecasting accuracy in dense sensor networks, it suffered in sparse topologies and cold-start initialization of additional nodes [19]. Multiple disciplines have explored federated multimodal learning frameworks. One such example is the LeViT-ResUNet system for UAV-IoT agricultural monitoring. This system demonstrated scalable, privacy-preserving modeling in various environmental conditions [20]. In situations where resources are scarce, our findings highlight the need for distributed, robust forecasting systems. Advanced studies used alternating attention layers to improve graph topology and temporal linkages in a double-explored spatio-temporal GNN. This method enhanced uncertainty calibration and decreased forecasting errors, but needed significant computing resources and careful adjustment to prevent over-smoothing during training [21]. Another study used self-constructing graph learning to automatically collect spatial relationships from telemetry data at big solar installations. The method improved RMSE relative to traditional networks, but faced overfitting hazards and restricted generalization to unknown facilities [22].
Smaller, energy-efficient models outperformed deeper models in many transformer design studies when optimized for certain prediction horizons. Adaptable architectures are necessary for real-world data variability, since the efficiency of lightweight models varies based on data completeness and feature representation [23]. Forecasting model explainability was also stressed in several studies. One study used SHAP-based interpretability modules during model training to improve operational transparency and feature relevance. This method increased confidence among system operators but increased computing burden during inference [24]. Another explainability-oriented approach uses SHAP-driven diagnostics for real-time sensor network feature attribution and defect detection. The approach enhanced interpretability and reliability, while latency increased in high-dimensional data environments [25].
A recent review of interpretable deep learning in energy forecasting found that saliency and gradient-based explanations improve transparency but are unstable during sudden distribution shifts or extreme weather conditions [26]. Recently, quantum-inspired learning for time-series prediction has been studied alongside deep networks. One method used parameterized quantum circuits in LSTM architectures to improve convergence and error reduction on limited datasets. Simulating quantum states on conventional hardware is computationally expensive, restricting scalability [27]. Additionally, multimodal designs influenced by quantum mechanics have emerged, such as Q-MobiGraphNet. This one uses quantum sinusoidal encoding to integrate Internet of Things (IoT) telemetry, unmanned aerial vehicle (UAV) images, and geographic data in order to strengthen infrastructure against environmental changes [28]. These technologies demonstrate the growing significance of quantum-guided feature representation in environmental and energy analytics.
Integrated temporal convolutional filters with transformer-based attention mechanisms increase wind speed forecasting accuracy, albeit at the cost of training efficiency and hyperparameter sensitivity [29]. The effectiveness of hierarchical attention and multi-scale convolution in dynamic and complex systems has been demonstrated by cross-layer convolutional attention mechanisms in complex telemetry settings, such as the CLCAN framework for intrusion detection in drone networks [30]. Recent lightweight convolutional networks for hybrid PV–wind forecasting showed the balance between computational economy and prediction accuracy. Although a compact one-dimensional CNN showed promising results with low latency, its performance degraded for uncommon extreme occurrences owing to the lack of uncertainty estimation techniques [31].
The aforementioned literature in Table 1 shows that hybrid deep architectures, spatio-temporal graph learning, and explainable AI improve renewable forecasts. These developments have not eliminated issues like model generalization across varied geographies, high computing demand in complicated transformer and GNN frameworks, and limited interpretability physical system awareness integration. These gaps need energy-aware, interpretable, and computationally efficient forecasting algorithms for renewable-powered, off-grid cloud infrastructures.

3. Proposed Methodology

Traditional deep predictors are unstable in off-grid cloud microgrids due to noisy sensor readings, irregular temporal dynamics, nonlinear cross-modal interactions, and incomplete data streams. Workload scheduling, battery cycling, and thermal management are all affected by cumulative forecast errors resulting from these difficulties. The Q-RCANeX technique provides a unified forecasting pipeline that stabilizes raw signals, improves feature separability, and effectively models long-range temporal–spectral relationships to overcome these constraints. After handling noise and missing data using signal decomposition-based preprocessing, the next step is to extract and select features that highlight physically important patterns. Compact representations are built using a residual convolutional encoder and enhanced with quantum feature embedding. Interactions across solar, wind, battery, and efficiency profiles are captured by temporal–spectral attention layers, which also account for frequency-aligned changes. To ensure steady convergence and generalization, a Hybrid Quantum–Bayesian Evolutionary Optimizer controls the model’s optimization. The pictorial view of the proposed frame is shown in Figure 1.
By combining off-grid microgrid data with a modular forecasting pipeline, Q-RCANeX aims to address a specific real-world issue. When sensor inputs are noisy or otherwise unsteady, signal decomposition can stabilize them. Both temporal–spectral attention and residual convolution uncover robust multi-scale patterns and long-range contextual links, respectively. Evolutionary optimization and quantum-inspired feature embedding enhance models’ stability during convergence and their ability to handle uncertainty. In contrast, the explainability module makes which factors are influencing the model’s predictions very evident. Collectively, these components provide an approachable framework that manages to be both accurate and practical.

3.1. Dataset Description

The data used in this study come from a third-party, publicly available source, available at [32]. To enhance research on renewable-powered computer infrastructure, the dataset was established and maintained by the National University Hospital Energy Innovation Centre (NUHEIC) in Oslo, Norway. It includes time-series data with high precision from a hospital’s cloud data center, powered by an off-grid hybrid solar–wind microgrid. From February 2022 to July 2025, all parameters were recorded every 10 min, and the dataset covers the entire period. Solar power, wind power, battery storage, and data-center energy use were all monitored continuously by a network of calibrated sensors and supervisory control systems. The NUHEIC team conducted thorough data validation, cleaning, and temporal synchronization before publishing the data to ensure the reliability and consistency of all recorded variables. For off-grid cloud applications, this dataset provides a solid, practical foundation for evaluating energy-aware optimization algorithms and approaches for renewable energy forecasting. It incorporates many different kinds of energy, spans an extensive range of time periods, and has been verified as real. The characteristics of the dataset are summarized in Table 2. The major functional portions of the off-grid microgrid are shown by the pieces displayed in Table 3. Data-center operational requirements, energy storage, renewable generation, and environmental inputs are all part of this category.

3.2. Signal Decomposition-Based Preprocessing

Due to climate change, device sensitivity, and random atmospheric events, data on renewable energy from hybrid solar–wind systems exhibit fairly peculiar patterns over time. Due to factors such as constantly shifting clouds, strong gusts, and intermittent sensor performance, these signals are never static and seldom follow a straight path. There is a risk that using this data in its raw form can lead to biased learning and unstable model convergence. Signal decomposition-based preprocessing (SDP) is a comprehensive preprocessing approach that this study employs to mitigate these effects [33]. The purpose of SDP is to construct physically meaningful representations from noisy, high-frequency data while preserving both the short-term dynamics and the long-term behavioral tendencies. Adaptive seasonal normalization, entropy-regulated denoising, and hybrid trend–seasonal decomposition with quantum filtering are the three interrelated processes that make the system function. To provide reliable forecasting inputs, each of these steps is essential, but in its own way. Imagine a multivariate sequence as the dataset.
S = { s τ ( 1 ) , s τ ( 2 ) , , s τ ( d ) } τ = 1 M ,
The d th recorded feature at time index is represented by s τ ( d ) . While d is the total number of features, M represents the number of samples. The components of each recorded variable are described:
s τ ( d ) = L τ ( d ) + C τ ( d ) + V τ ( d ) + ξ τ ( d ) ,
The low-frequency trend, L τ ( d ) , represents gradual changes (such as seasonal irradiation drift), the periodic changes, C τ ( d ) , symbolize high-frequency changes (such as daily or weekly energy cycles); V τ ( d ) represent changes caused by turbulence and operational switching, and  ξ τ ( d ) represent measurement noise. Removing ξ τ ( d ) to the greatest extent feasible while accurately separating L τ ( d ) , C τ ( d ) , and  V τ ( d ) is the objective of the preprocessing step.
Removal of random noise and preservation of beneficial changes are achieved in the first phase of SDP using an entropy-governed adaptive denoising approach. In addition to allowing us to adjust the filter’s sensitivity, entropy provides a numerical measure of uncertainty. This is the definition of the local entropy E ( s τ ( d ) ) for each signal, computed over a sliding window of length ω :
E ( s τ ( d ) ) = κ = 1 ω π κ ( d ) ln ( π κ ( d ) + ϵ 0 ) ,
where ϵ 0 is a minor stability constant and π κ ( d ) is the chance that an amplitude value will occur in the window. In order to determine an adaptive gain coefficient γ τ ( d ) , this value is then used.
γ τ ( d ) = α σ τ ( d ) 1 + E ( s τ ( d ) ) ,
The local variance is represented by σ τ ( d ) in this instance, and a scaling constant α evaluated in a laboratory setting is used. The function s ^ τ ( d ) represents the denoised signal, and its computation is as follows:
s ^ τ ( d ) = s τ ( d ) γ τ ( d ) E ( s τ ( d ) ) s τ ( d ) .
This adaptive entropy filter is designed to exclude random changes with a high entropy. Alterations in wind speed or solar irradiation are instances of physical changes that it records as transitions with low entropy. Therefore, s ^ τ ( d ) provides more reliable model training in the future by capturing crucial dynamics with far less sensor noise. Our approach utilizes an Enhanced Trend–Seasonal Separation with Quantum Filtering (ETS-QF) technique to deconstruct each signal after noise reduction [34]. The assumption that all sums add up in a straight line underlies traditional decomposition techniques. Conversely, ETS-QF employs a probabilistic filtering method based on quantum interference. Whether there are significant seasonal trends or little shifts in renewable energy signals, our technique can detect them all. The signal degrades as follows:
s ^ τ ( d ) = L τ ( d ) + C τ ( d ) + V τ ( d ) ,
By constructing a smoothing window W L using polynomial regression, the equation L τ ( d ) is obtained. In order to get C τ ( d ) , the detrended component is averaged sequentially. Using a filtering kernel F ( · ) that is influenced by quantum mechanics further enhances the residual component V τ ( d ) , as shown below.
F ( V τ ( d ) ) = η = Λ Λ w η ( d ) V τ + η ( d ) e ȷ θ η ,
The localization is maintained by the Gaussian weights w η ( d ) , the imaginary unit is ȷ , and a phase interference factor θ η is expressed as
θ η = β d sin 1 V τ + η ( d ) ( V τ ( d ) ) 2 + ϵ 1 ,
ϵ 1 maintains stability, whereas β d determines the amount of interference. This kernel dynamically increases the proportion of phase-and frequency-correlated coherent fluctuations and decreases the proportion of incoherent noise components. The representation is given by
s τ ( d ) = L τ ( d ) + C τ ( d ) + F ( V τ ( d ) ) ,
The result is a balanced signal that captures the probabilistic and deterministic aspects of energy change, with noise filtered out. Because temporal anomalies are caused by both random external factors and changes in system operation, this breakdown is significant for off-grid renewable systems. The last step of SDP uses adaptive seasonal normalization (ASN) to ensure that all feature scales are consistent. Although static normalization ignores the cyclical behavior of renewable resources throughout the year, ASN performs just that. One definition of normalization is the presence of a seasonal index ς for each observation that falls into one of four categories: spring, summer, fall, or winter.
s ¯ τ , ς ( d ) = s τ ( d ) μ ς ( d ) σ ς ( d ) ,
The season’s mean and standard deviation are represented by the symbols μ ς ( d ) and σ ς ( d ) , respectively. To accommodate gradual environmental transitions—such as seasonal shifts in irradiance or wind intensity—these parameters are continuously updated using a temporal decay factor δ :
μ ς ( d ) ( τ ) = ( 1 δ ) μ ς ( d ) ( τ 1 ) + δ s τ ( d ) , σ ς ( d ) ( τ ) = ( 1 δ ) σ ς ( d ) ( τ 1 ) + δ | s τ ( d ) μ ς ( d ) ( τ ) | .
The temporal decay factor δ used in the adaptive seasonal normalization is determined through a data-driven analysis of the temporal autocorrelation structure of the renewable energy signals. Specifically, δ is selected to reflect the persistence of dominant seasonal patterns—such as diurnal solar cycles and medium-term wind variability—while avoiding excessive smoothing of short-term fluctuations. During preliminary analysis, multiple candidate values of δ were evaluated to balance responsiveness to rapid environmental changes against stability over longer seasonal horizons. The final value was chosen as the smallest decay factor that consistently preserved salient seasonal trends while maintaining robustness to transient sensor noise and irregular sampling intervals. This adaptive scaling enables stable normalization across time without distorting the inherent seasonal dynamics of the microgrid data. To maintain the dataset’s inherent seasonal cycle while conducting statistical comparisons of variables and time periods, this adaptive scaling is used. Afterwards, all of the normalized variables are combined into a single matrix:
M τ = [ s ¯ τ , ς ( 1 ) , s ¯ τ , ς ( 2 ) , , s ¯ τ , ς ( d ) ] ,
The model that generates predictions uses this structured data. The modified dataset M τ outperforms the original data in terms of stability, feature fitting, and smoothness. Temporal signals takes benefit from the proposed signal preprocessing framework in three ways: reduced energy use, increased resilience to environmental changes, and simplified analysis. The groundwork is laid for off-grid cloud data centers to accurately and efficiently predict renewable energy sources.
The signal decomposition stage is computationally quite lightweight when compared to other, less involved preprocessing techniques, such as normalization or smoothing. This is due to the fact that it is effective on sparse time frames and is implemented just once prior to training or inferring a model. Considering how tiny this expense is in relation to the overall time required to run the model, it is valuable since it improves data stability, accelerates convergence, and prevents mistakes from propagating in subsequent learning phases. Because of this, the preprocessing pipeline continues to be employed in off-grid cloud microgrid scenarios for real-time or near-real-time applications.

3.3. Energy-Adaptive Feature Refinement and Generative Equilibrium Process

The collection includes de-emphasized signal data, as well as synchronized time-series records of data-center power consumption, wind speed, battery charge levels, and sunshine hours. Nevertheless, the off-grid energy ecosystem’s underlying dynamics are evident in the skewed distributions, duplicate features, and poor correlations across variables, driven by operational ambiguity and environmental unpredictability. These differences cause bias and less reliable predictive models if unchecked. These difficulties have been addressed by developing a new integrated framework called EAFR-GEP. An accurate, succinct, and physically relevant depiction of the system’s energy fluctuations are achieved by combining feature extraction with relevance-based selection and probabilistic balancing. The decomposed observation of the m th variable is expressed as r θ ( m ) at time θ . The time-dependent behavior, direct magnitude, and overall change in each signal are shown by applying an energy-sensitive embedding. It can be calcualted as
E θ ( m ) = β 1 | r ˙ θ ( m ) | + β 2 r θ ( m ) + β 3 0 θ r υ ( m ) d υ ,
The integral term represents the overall energy contribution throughout time. At the same time, r ˙ θ ( m ) is the first derivative that shows local change. To ensure that qualities with high and low variation have the same impact, the adaptive scaling constants β 1 , β 2 , and  β 3 are used. To model the interconnections between energy subsystems, pairwise interaction coefficients are used. For instance, the relationship between total load and wind speed or the effect of solar radiation on storage discharge.
I θ ( u , v ) = Cov ( E θ ( u ) , E θ ( v ) ) ς ( E θ ( u ) ) ς ( E θ ( v ) ) ,
Covariance is represented by the sign Cov ( · ) , while standard deviation is shown by the symbol ς ( · ) . Coefficients like these allow precise measurement of the appropriate cross-modal coupling in a hybrid renewable system by quantifying the impact of modifications to one part on another. After that, a complete feature vector is created by combining all features, including both acquired and interaction-based ones.
P θ = [ E θ ( 1 ) , E θ ( 2 ) , , E θ ( m ) , I θ ( 1 , 2 ) , , I θ ( m 1 , m ) ] ,
This encompasses not only the signal’s dynamics but also the relationships between the characteristics. This integrated view provides a more complete and consistent picture of energy’s evolution over time. To eliminate superfluous components from prediction algorithms, an adaptive relevance-based weighting approach is used. The integration of linear and nonlinear correlations yields a hybrid significance coefficient π m for each feature g m .
π m = η 1 | ρ ( g m , y ) | m | ρ ( g m , y ) | + η 2 I ( g m ; y ) m I ( g m ; y ) ,
The expression I ( g m ; y ) shows that the nonlinear interactions are revealed via mutual information. Linear correlations are seen between y and ρ ( g m , y ) , whereas η 1 and η 2 serve as balancing factors, adding up to 1. The factors that have succeeded well according to the adaptive criteria,
π cut = Quantile ( { π m } , 0.25 ) ,
The features that are both statistically significant and amenable to physical explanation are the ones that are left out. Because certain operational situations, such as extended periods of low radiation, high-demand peaks, or deep battery-depletion cycles, do not occur very frequently, class imbalance persists even after feature improvements. To address this issue, a variational generative augmentation approach is used. The encoder reveals the distribution of latent features as
q φ ( z | P θ ) = N ( z | μ θ , σ θ 2 ) ,
To ensure that balanced samples are created by the decoder via
p ψ ( P θ | z ) = N ( P θ | μ ˜ θ , σ ˜ θ 2 ) ,
where the encoder’s parameters are φ and the decoder’s settings are ψ . The optimization objective integrates reconstruction fidelity and latent regularization:
J gen = E q φ ( z | P θ ) P θ P ^ θ 2 2 + D KL q φ ( z | P θ ) p ( z ) ,
The Kullback–Leibler divergence ( D KL ( · ) ) is used to confirm that produced samples match the original distribution structure. EAFR-GEP uses a variety of quality criteria to assess synthetic samples’ physical realism and statistical consistency. Comparing the two samples’ statistically similar features shows they align. This ensures that synthetic data has the same mean, variance, and higher-order moments across energy variables. Second, we examine pairwise interaction coefficients before and after augmentation to retain correlation. The cross-modal interdependence of solar, wind, storage, and load characteristics is confirmed by this investigation. Third, generative loss indirectly promotes reconstruction integrity by punishing unrealistic deviations using Kullback–Leibler divergence. As a last step, domain validity checks exclude samples with unrealistic values like negative generation values or battery states beyond operational limitations. These guidelines ensure that more data improves class balance without affecting energy dynamics or artifacts. This approach generates data for underrepresented energy levels to guarantee that all energy levels have the same characteristic rate. The new dataset R = { P θ } balances operational and climatic situations throughout days with high demand and nights with low generation. A robust dataset is utilized to create the next quantum convolutional attention model with measurable system dynamics in each feature vector and scaling coordination. The EAFR-GEP method yields statistically sound and understandable findings. Storage systems and renewable energy sources are crucial to off-grid cloud architectures. The flow of EAFR–GEP is shown in Algorithm 1.
Algorithm 1 Energy-Adaptive Feature Refinement and Generative Equilibrium Process (EAFR–GEP)
Require: Decomposed dataset D = { r ϑ ( 1 ) , r ϑ ( 2 ) , , r ϑ ( M ) } , target variable y
Ensure: Refined and balanced dataset R = { P ϑ }
  1:
Initialization: Assign adaptive scaling factors β 1 , β 2 , β 3 and weighting coefficients η 1 , η 2 .
  2:
for each time instance ϑ  do
  3:
    for each variable r ϑ ( M ) in D  do
  4:
        Compute energy-sensitive embedding:
E ϑ ( M ) = β 1 | r ˙ ϑ ( M ) | + β 2 r ϑ ( M ) + β 3 0 ϑ r ξ ( M ) d ξ
  5:
    end for
  6:
    Compute feature–interaction coefficients:
I ϑ ( a , b ) = Cov ( E ϑ ( a ) , E ϑ ( b ) ) ς ( E ϑ ( a ) ) ς ( E ϑ ( b ) )
  7:
    Construct unified feature vector:
P ϑ = [ E ϑ ( 1 ) , , E ϑ ( M ) , I ϑ ( 1 , 2 ) , , I ϑ ( M 1 , M ) ]
  8:
end for
  9:
Compute hybrid feature relevance:
π M = η 1 | ρ ( g M , y ) | M | ρ ( g M , y ) | + η 2 I ( g M ; y ) M I ( g M ; y )
10:
Determine cutoff π cut = Quantile ( { π M } , 0.25 ) and discard features with π M < π cut .
11:
Encode latent representation:
q φ ( z | P ϑ ) = N ( z | μ ϑ , σ ϑ 2 )
12:
Decode and reconstruct synthetic samples:
p ψ ( P ϑ | z ) = N ( P ϑ | μ ˜ ϑ , σ ˜ ϑ 2 )
13:
Minimize the generative loss:
J gen = E q φ P ϑ P ^ ϑ 2 2 + D KL q φ ( z | P ϑ ) p ( z )
14:
Generate synthetic instances for underrepresented scenarios
15:
Merge original and synthetic samples to produce the final dataset R
16:
return  R

3.4. Proposed Classification Framework: Quantum-Aware Residual Convolutional Attention Network with Explainability (Q-RCANeX)

To effectively forecast renewable energy behavior within off-grid cloud infrastructures, the model must capture intricate temporal fluctuations, spatial dependencies, and nonlinear cross-correlations among hybrid energy sources. The proposed Quantum-Aware Residual Convolutional Attention Network with Explainability (Q-RCANeX) is designed to meet these objectives by merging quantum-enhanced representation learning, multi-branch residual convolution [35], spectral–temporal attention [36], and an embedded interpretability module. The framework operates as a single, end-to-end architecture that transforms denoised feature sequences into interpretable predictions with energy-aware regularization. The proposed architecture is shown in Figure 2.
Quantum Feature Embedding (QFE) Parameterization: QFE transforms the classical input feature vector into a quantum-inspired latent space using parameterized rotation and entanglement operations. The embedding depth L q is set at 3 in this study to provide sufficient expressive capacity while maintaining processing performance. Learnable rotation angles { ξ , ψ } = 1 L q are used by each embedding layer to alter the amplitude and phase connections across feature channels. Along with the weights of the network, these parameters are also modified. A complex-valued manifold may be used to project linked energy variables, allowing for the separation of nonlinear characteristics. To avoid data bottlenecks and guarantee a smooth gradient flow during training, the embedding dimensionality is set to match the input to the convolutional encoder.
Let the input feature vector at time instance τ be denoted by x τ R m , representing the integrated state of solar irradiance, wind velocity, storage voltage, and load utilization. The first transformation projects x τ into a quantum-augmented latent manifold using a unitary operator U κ defined by sequential parameterized rotations and entanglement interactions:
q τ = U κ ( x τ ) = = 1 L q R y ( ξ ) R z ( ψ ) C x τ ,
with R y and R z denoting the training parameters for y and z axis-spinning rotation gates, respectively. A controlled entanglement gate and feature channel connector, C is defined by the variables ξ and ψ . The matrices for rotation are expressed as
R y ( ξ ) = cos ξ 2 sin ξ 2 sin ξ 2 cos ξ 2 , R z ( ψ ) = e j ψ / 2 0 0 e j ψ / 2 .
Incorporating amplitude-phase dependencies, this method generates a complex-valued embedding q τ that improves the expressiveness of features and their nonlinear separability for future modeling. Afterwards, a sequence of residual convolutional modules is used to process the quantum-enhanced tensor q τ . To acquire multi-scale contextual information, each module employs a dilated kernel to capture a broader view of the environment and a localized kernel to capture near-term changes. The dual-branch mapping at depth r:
h τ ( r ) = σ N W 1 ( r ) q τ + b 1 ( r ) + σ N W 2 ( r ) δ r q τ + b 2 ( r ) ,
the Swish activation is σ ( u ) = u / ( 1 + e u ) , N ( · ) denotes batch normalization, and  δ r and ⊛, respectively, represent standard and dilated convolutions with a dilation factor δ r . To prevent data from deteriorating further and keep gradients flowing, residual learning is used.
y τ ( r ) = q τ + ρ r h τ ( r ) ,
The scale of the remainder is determined by ρ r ; by stacking many blocks, compression projection decreases a channel’s dimensionality while maintaining its variation:
s τ = ReLU W c y τ ( r max ) + b c .
By omitting unnecessary details, this process generates a representation, s τ , that efficiently encodes a spatial–temporal structure. The method incorporates a dual self-attention mechanism that captures periodic and inter-variable correlations by combining learning temporal dependencies and spectrum reweighting. By demonstrating interactions between queries and keys, temporal attention synchronizes pertinent time steps:
A temp = softmax ( W q s τ ) ( W k s τ ) d a ,
Learnable projection matrices are represented by the variables W q and W k , while the attention dimension is d a . Finding a temporal context vector is achieved by
c τ = A temp ( W v s τ ) ,
The passage of time or strong winds are only two examples of the significant environmental changes that it records. Simultaneously, a spectral weighting vector enhances the significance of every feature dimension by its global pooling and nonlinear modification:
Ω spec = softmax U 2 ReLU U 1 GAP ( s τ ) ,
The notation GAP ( · ) denotes global average pooling, while the symbols U 1 and U 2 denote trainable parameters, respectively. The combination of spectral and temporal properties provides the background.
z τ = Ω spec c τ ,
The modulation is shown element-by-element using ⊙. It is now much easier to see crucial time segments and energy modes due to this integration in Q-RCANeX. Utilizing gradient-based relevance computing, the network effortlessly establishes connections between all features and model options. It finds that y ^ τ relies heavily on the component p because
Γ τ ( p ) = y ^ τ z τ ( p ) z τ ( p ) ,
via transformation in order to get understandable weights
Γ ˜ τ ( p ) = Γ τ ( p ) p Γ τ ( p ) + ε 1 ,
The numbers will remain constant due to ε 1 . The entropy of these relevance weights is reduced using a particular interpretability regularizer, which aids in understanding the importance of the attributes:
L XAI = p Γ ˜ τ ( p ) ln ( Γ ˜ τ ( p ) ) .
To increase scientific transparency, this technique filters out non-physically relevant variables from the model, such as solar input and turbine dynamics. The decision space is updated with the improved latent representation z τ by the use of a dense transformation.
y ^ τ = f o W o z τ + b o ,
The function f o ( · ) acts in this situation depending on the task. It might be linear for regression outputs like energy ratios and sigmoid for binary prediction. This goal must be met in terms of accuracy, interpretability, and long-term computational sustainability:
L total = L pred + λ 1 L XAI + λ 2 L eco ,
L pred calculates the computational energy cost that is directly related to processing load and assesses the prediction loss, which are expressed as cross-entropy or mean-square error. In addition to a penalty for the computational energy cost that is directly proportional to the processing load, the prediction loss, which is also known as cross-entropy or mean-square error, is quantified by L pred .
L eco = 1 M i = 1 M Comp i Comp max 2 .
Model parameters Θ = { W , b , ξ , ψ , ρ } are iteratively optimized by an adaptive gradient scheme:
Θ u + 1 = Θ u η Θ L total Θ 2 L total + ε 2 ,
To prevent division from becoming unstable, ε 2 is used, while η indicates the rate of learning. Over the time horizon, a global interpretability profile is constructed by assembling the relevance maps Γ ˜ τ .
Υ p = 1 T τ = 1 T Γ ˜ τ ( p ) ,
By factoring in each characteristic’s influence, this statistic assesses the model’s predictive performance. Validating the model’s rationale, this study uses SHAP analysis to visualize key distributions as heatmaps.
By integrating quantum state encoding, multi-scale residual convolution, attention-driven temporal fusion, and integrated explainability, Q-RCANeX creates a transparent, physically coherent, and computationally efficient learning architecture. Algorithm 2 demonstrates the flow structure of the proposed method. Renewable energy information that is both accurate and publicly available improves prediction accuracy and provides insight into the interplay between solar, wind, and storage.
Algorithm 2 Quantum-Aware Residual Convolutional Attention Network (Q-RCANeX)
  • Require: Time-series window { x τ R m } τ = 1 T , parameters Θ , learning rate η , weights λ 1 , λ 2
  • Ensure: Forecasts { y ^ τ } and relevance scores { Υ p }
  1:
Quantum Embedding:
  2:
for  τ = 1 to T do
  3:
     q τ U κ ( x τ ) using R y , R z and controlled gates
  4:
end for
  5:
Multi-Scale Residual Convolutions:
  6:
for  r = 1 to R do
  7:
    for  τ = 1 to T do
  8:
         h τ , 1 ( r ) σ ( N ( W 1 ( r ) q τ ) )
  9:
         h τ , 2 ( r ) σ ( N ( W 2 ( r ) δ r q τ ) )
10:
         y τ ( r ) q τ + ρ r ( h τ , 1 ( r ) + h τ , 2 ( r ) )
11:
         q τ y τ ( r )
12:
    end for
13:
end for
14:
Channel Compression:
15:
for  τ = 1 to T do
16:
     s τ ReLU ( W c y τ ( R ) )
17:
end for
18:
Temporal–Spectral Attention:
19:
for  τ = 1 to T do
20:
     A temp softmax ( ( W q s τ ) ( W k s τ ) / d a )
21:
     c τ A temp ( W v s τ )
22:
     ω spec softmax ( U 2 ReLU ( U 1 GAP ( s τ ) ) )
23:
     z τ ω spec c τ
24:
end for
25:
Prediction:
26:
for  τ = 1 to T do
27:
     y ^ τ f o ( W o z τ )
28:
end for
29:
Explainability and Loss:
30:
for  τ = 1 to T do
31:
    for each feature p do
32:
         Γ τ ( p ) y ^ τ z τ ( p ) z τ ( p )
33:
    end for
34:
     Γ ˜ τ ( p ) Γ τ ( p ) / ( p Γ τ ( p ) + ε )
35:
end for
36:
L XAI τ , p Γ ˜ τ ( p ) ln ( Γ ˜ τ ( p ) )
37:
L total L pred + λ 1 L XAI + λ 2 L eco
38:
Parameter Update:
39:
Θ Θ η Θ L total
40:
Relevance Aggregation:
41:
for each p do
42:
     Υ p 1 T τ Γ ˜ τ ( p )
43:
end for
44:
return  { y ^ τ } and { Υ p }
Early convergence, sensitivity to noisy and non-stationary sensor streams, and limited adaptation to varied energy modalities are some of the practical restrictions that current optimization and learning models influenced by quantum mechanics face when applied for renewable energy forecasting. When operating conditions vary, many previous methods that rely on probabilistic updating rules or fixed quantum rotation struggle to maintain a balance between exploration and exploitation. In practical off-grid microgrids, these constraints reduce system stability due to fluctuating renewable sources and inadequate telemetry. The proposed Hybrid Quantum–Bayesian Evolutionary Optimization (Q-BEO) optimizer aims to address these issues by using adaptive quantum evolution and Bayesian-guided uncertainty modeling.

3.5. Model Optimization and Parameter Tuning via Hybrid Q-BEO

An optimizer that can adapt to non-standard loss surfaces without affecting model speed or comprehensibility is necessary for training a high-capacity model, such as Q-RCANeX, in a dynamic renewable energy context. Adam and RMSProp, which are standard gradient-based algorithms [37], perform well on convex domains. An imbalance between local accuracy and global exploration is often observed due to quantum-encoded representations and multi-branch attention dynamics. A novel approach to this issue is Q-BEO. Combining the power of Bayesian reasoning with the breadth of evolutionary learning and the inventiveness of quantum rotation operators, this novel approach generates novel ideas. By considering high-dimensional parameter manifolds, these methods enable energy-efficient operation, prevent overfitting, and accelerate convergence. These trainable parameters are Ω = { A , c , κ , ζ , χ , V } , with each component being either a convolutional kernel, a bias term, a quantum rotation coefficient, or an adaptive attention weight. An evaluation of interpretability, resource efficiency, and forecast accuracy is the central emphasis of the optimization, which is based on a compact composite objective function.
min Ω L total = L fit + γ 1 L exp + γ 2 L eco ,
Forecast accuracy is evaluated by L fit , decision clarity is provided by L exp , and excessive computational costs are punished by L eco . Evolutionary adaptation, expected-improvement selection, quantum transition, and probabilistic evaluation are the four interrelated processes that make up the Q-BEO technique. To start, a set of Q parameter configurations { Ω j } j = 1 Q is created using Latin hypercube sampling. As a result, the parameter space is thoroughly investigated. For each candidate, a Gaussian process (GP) surrogate displays the expected loss and standard deviation:
L ^ exp ( Ω j ) GP m ( Ω j ) , κ ( Ω j , Ω t ) ,
The mean function is m ( · ) , and the covariance kernel is κ ( · , · ) , which shows the interrelationships between the parameter states. While accounting for uncertainty in unexplored regions, this Bayesian layer constructs a probabilistic surface of the loss function. It guides the search toward locations with minimal loss. To prevent premature convergence and maintain high variation, evolution uses operators such as mutation, crossover, and selection. Every individual Ω j ( u ) undergoes random alterations as a result of mutation for every generation u.
Ω j ( u + 1 ) = Ω j ( u ) + η Ω r 1 ( u ) Ω r 2 ( u ) + ρ N ( 0 , σ 2 ) ,
Ω r 1 and Ω r 2 are randomly selected population members, η controls differential amplification, ρ scales stochastic perturbations, and  N ( 0 , σ 2 ) adds Gaussian random variables. The optimizer view is at the whole region and concentrates on high-competition areas using this strategy. To pass on beneficial features, the crossover process incorporates traits from both parents and children:
Ω mix ( u ) = τ Ω j ( u + 1 ) + ( 1 τ ) Ω j ( u ) ,
The amount of genetic inheritance is determined by looking at the value of τ . Afterwards, the decision to modify the current solution is made using Bayesian-guided selection:
Ω j ( u + 1 ) = Ω mix ( u ) , if L ^ exp ( Ω mix ( u ) ) < L ^ exp ( Ω j ( u ) ) , Ω j ( u ) , otherwise .
To ensure a healthy mix of well-known and less-explored topics, this evolutionary stage uses Bayesian confidence indicators to guide research. A quantum transition layer in Q-BEO often modifies the system’s rotational parameters in response to changes in quantum states, thereby avoiding shallow minima and accelerating convergence. The state Ω q ( u ) is the new home of the top candidate.
Ω q ( u + 1 ) = Ω q ( u ) cos ( θ ) + Z q ( u ) sin ( θ ) ,
The reference configuration Z q ( u ) is assisted by the uniform prior, and the quantum rotation angle θ is determined to be
θ = Δ q L total Ω q ( u ) .
In order to streamline the calculation logic, we may think of the quantum rotation angle θ as a finite step size that regulates the degree to which the optimal solution is twisted toward an exploratory reference state. The magnitude of the loss gradient is the practical basis for θ . To escape shallow local minima, bigger rotations are required when gradients are substantial, since they reveal regions that are unstable or poorly optimized. Conversely, for fine-grained refinement, smaller rotations are caused by tiny gradients close to convergence. A certain range of rotation angles is enforced to maintain numerical stability; this range may be expressed as
θ u = clip Δ q Ω L total ( Ω q ( u ) ) 2 , θ min , θ max .
The optimization process converges smoothly while maintaining the correct balance between exploration and exploitation, thanks to this formulation that makes the rotation mechanism clear to understand and reliable.
This approach models the spatial mobility of quantum systems and helps the optimizer locate global minima. This will enable the optimizer to discover optimal solutions faster and avoid local minima. This provides a novel approach to enhancing deep learning models by using operators grounded in actual laws. For this purpose, this work uses the expected improvement criterion, which integrates the quantum rotation impact with the Bayesian uncertainty estimate.
I ( Ω ) = E L min L ^ exp ( Ω ) Ψ L min m ( Ω ) s ( Ω ) ,
The highest loss that has been recorded is L min , the predictive variance is s ( Ω ) , and the cumulative Gaussian function is Ψ ( · ) . By reconsidering and selecting the configuration that achieves the optimal equilibrium between exploration and exploitation, this study maximizes the utility of I ( Ω ) . If the stability requirement | Ω ( u + 1 ) Ω ( u ) is less than ϵ 0 or the maximum generation threshold is satisfied, the operation progresses. It obtains a precise, clearly comprehensible, and measurable state by applying the ideal parameters for this model, denoted by Ω .
The addition of a controlled element of uncertainty, made possible by quantum rotation, improves the performance of global searches. While evolutionary operators maintain structural variety and resilience, Bayesian reasoning accounts for uncertainty. Incorporating all three approaches, the Q-BEO system acts as an adaptive whole. An optimization engine is built using concepts from quantum physics, statistics, and biology in the new hybridization. The Q-BEO in Algorithm 3 enables a less energy-intensive learning system by leveraging the quantum-layered structure of Q-RCANeX. This changes when the data on renewable energy sources increases.
Algorithm 3 Hybrid Quantum–Bayesian Evolutionary Optimization (Q-BEO) for Q-RCANeX Parameter Tuning
  • Require: Initial parameters Ω = { A , c , κ , ζ , χ , V } , population size Q, maximum generations U max , tolerance ϵ 0
  • Ensure: Optimized configuration Ω
  1:
Generate Q initial candidates { Ω j ( 0 ) } using Latin hypercube sampling.
  2:
for  u = 1 to U max  do
  3:
    Estimate expected loss L ^ exp ( Ω j ( u ) ) via Gaussian Process surrogate.
  4:
    {Bayesian module constructs a probabilistic approximation of the loss surface}
  5:
    for each candidate j do
  6:
        Mutate candidate:
Ω = Ω j ( u ) + η Ω r 1 ( u ) Ω r 2 ( u ) + ρ N ( 0 , σ 2 )
  7:
        Crossover operation:
Ω mix = τ Ω + ( 1 τ ) Ω j ( u )
  8:
        if  L ^ exp ( Ω mix ) < L ^ exp ( Ω j ( u ) )  then
  9:
             Ω j ( u + 1 ) Ω mix
10:
        else
11:
             Ω j ( u + 1 ) Ω j ( u )
12:
        end if
13:
    end for
14:
    {Evolutionary module maintains diversity and progressive refinement}
15:
    Apply quantum rotation to the best candidate:
Ω q ( u + 1 ) = Ω q ( u ) cos ( θ ) + Z q ( u ) sin ( θ )
16:
    Rotation angle update:
θ = Δ q L total Ω q ( u )
17:
    {Quantum module enables probabilistic jumps for global exploration}
18:
    Compute expected improvement I ( Ω ) .
19:
    Select configuration with max ( I ( Ω ) ) for next iteration.
20:
    if  Ω ( u + 1 ) Ω ( u ) < ϵ 0  then
21:
        break
22:
    end if
23:
end for
24:
return  Ω = arg min Ω L total

3.6. Performance Evaluation

The domain of off-grid cloud renewable energy forecasting will be used to examine the computational robustness, interpretability, and predictive accuracy of the Q-RCANeX framework. To ensure the model fits the system’s energy behavior and produces accurate, understandable results, the assessment uses standard and explainability-oriented criteria. To do this, a distinct interpretability metric is used in conjunction with other measures [38]. The detection ratio (DR), sensitivity index (SI), harmonic reliability (HR), classification accuracy (CA), mean deviation error (MDE), and root quadratic error (RQE) are standard metrics for evaluating model classification and regression. Accompanying this is the newly-introduced Quantum-Interpretability Coherence Index, or QICI. This index shows the correlation between the model’s quantum-enhanced inference prediction confidence and the feature relevance entropy of its explainability layer. This shows that the model’s thinking and decision-making are consistent. The following equations indicate the evaluation metrics— Z p , Z n , Z f , and Z m —where p represents successful detections, n represents negatives, f represents false positives, and m represents missed detections. The predicted and actual outcomes of the r th observation are represented by the symbol o ^ r , with r representing the total number of test samples:
DR = Z p Z p + Z f , SI = Z p Z p + Z m ,
HR = 2 × DR × SI DR + SI , CA = Z p + Z n Z p + Z n + Z f + Z m ,
MDE = 1 r k = 1 r | o ^ k o k | , RQE = 1 r k = 1 r ( o ^ k o k ) 2 ,
QICI = 1 1 r k = 1 r | Λ k Π k | .
Both the prediction confidence ( Π k ) and the normalized entropy of feature relevance ( Λ k ) are shown in the equations. The basis for this is the possible explanation of the k th example. a near-perfect QICI score indicates that the model’s confidence in its predictions and its internal logic are in harmony. This indicates that the model is continuously improving its results via learning. The additional energy consumption of Q-RCANeX is determined by analyzing the DR, SI, HR, and CA data. Instead, MDE and RQE assess their regression-based prediction accuracy. Using QICI in this evaluation proceeds beyond mere statistical checking. It allows us to test the reliability of quantum physics predictions. Q-RCANeX enables intelligent systems that are easy to understand, consume minimal power, and achieve high anticipated accuracy through this integrated assessment technique. This is a result of its suitability for use in real-world renewable-powered cloud infrastructures.

4. Simulation Results and Discussion

4.1. Experimental Setup and Implementation Details

For the experimental evaluation of Q-RCANeX, this study used a high-resolution multimodal renewable energy dataset with about 120,000 samples taken every 10 min. This dataset included solar irradiance, wind generation, battery state of charge, environmental factors, and thermal efficiency metrics. Data that was either missing or contained sensor noise was filled in using the Signal Decomposition Pipeline, min–max normalization, and quantum-enhanced imputation. As demonstrated in Table 4, the model underwent training for 200 iterations using optimal hyperparameters. The following are provided: a 64-dimensional quantum feature embedding space, a learning rate of 1.5 × 10 4 , a batch size of 32, six layers for temporal–spectral attention, and a four-block residual encoder. Both the training and inference processes were executed on an NVIDIA GeForce RTX 4060 GPU (8 GB VRAM) using PyTorch and CUDA acceleration. During training, the Hybrid Quantum–Bayesian Evolutionary Optimizer managed parameter changes to prevent overfitting and maintain stable convergence.

4.2. Explainability and Interpretability Analysis

In Figure 3, the global SHAP feature importance for Q-RCANeX is shown for all forecasting jobs. This shows how the model prioritizes true renewable energy production drivers. Solar irradiation, wind speed, battery charge, and ambient temperature regularly rank higher. It clearly affect off-grid microgrid functioning. The model’s predictions are based on genuine energy dynamics, not trivial correlations, since SHAP ranks match domain-specific physical processes. Qualitative SHAP data analysis explains model decision-making. Short-term renewable variability and energy storage states have significant beneficial effects in dynamic operational settings, demonstrating their importance. Dynamic changes reduce the influence of factors associated to steady baseline load patterns, which contribute less and more consistently. This shows that Q-RCANeX is more transparent and practical since the explainability technique detects context-relevant and understandable linkages.

4.3. Predictive Accuracy and Forecasting Performance

The Renewable Energy Availability Forecast (REAF) for a complete 24 h cycle is shown in Figure 4 for the day ahead. The blue color lines in the figure shows the actual data and the brown color line shows the predicted value. It demonstrates how Q-RCANeX records the impact of dawn, peak irradiance, and sunset drop on natural daily rhythms. With a significant R2 value and very close agreement between the actual and predicted curves, the model can reliably predict changes in real-world energy availability. Even when things rush, Q-RCANeX remains constant, as shown in the zoomed-in window. The reason is that it is effectively made less susceptible to noise by using temporal–spectral attention layers and quantum feature embeddings. Given the importance of short-term predictions for off-grid operation scheduling, this finding demonstrates the model’s novelty.
Table 5 shows that across all assessment measures, Q-RCANeX outperforms the most recent forecasting models. Outperforming competing systems, it comprehends changes in renewable energy with a high detection ratio, sensitivity index, and accuracy rate of 98.6%. Its numerical stability is shown by its low MDE and RQE. By integrating temporal–spectral reasoning, quantum-guided optimization, and explainability alignment, Q-RCANeX achieves a much higher level of performance. When combined, these three factors make the system more feature-relevant and less noise-sensitive. Its ability to make consistent, unambiguous choices is demonstrated by its strong QICI score. The updates demonstrate the novelty and efficacy of Q-RCANeX for predicting non-grid renewable power sources.
The 3-day Wind Generation Forecast (WGF), as shown in Figure 5, demonstrates the accuracy with which Q-RCANeX predicts complex wind patterns, such as jets throughout the night, lulls during the middle of the day, and inherently unpredictable gust patterns. The model performs well in conditions with high daily variability, as the predicted and actual trajectories were quite close over three consecutive days. The evening is when most forecasting models fall short; however, Q-RCANeX remains relatively consistent with the data, as seen in the zoomed-in region, despite its unexpected gusts. For off-grid operation planning, this finding demonstrates the system’s robustness in non-stationary meteorological regimes, enabled by quantum-embedded characteristics and temporal–spectral attention.
Results from the ablation studies as shown in Table 6 demonstrate that every component of Q-RCANeX contributes significantly to the system’s overall efficiency. While residual convolution improves multi-scale feature representation and signal decomposition stabilizes non-stationary microgrid telemetry, temporal–spectral attention captures long-range contextual dependencies. To further increase convergence and decrease uncertainty, Q-BEO optimization and quantum feature embedding are used. The proposed architecture enhances accuracy and resilience without adding unnecessary complexity to the model, as shown by the consistent and monotonic rise across all evaluation criteria.

4.4. Computational Efficiency and Deployment Considerations

Table 7 shows that Q-RCANeX achieves a unique balance between high prediction accuracy and low resource utilization, as assessed by computational efficiency. With fewer parameters and FLOPs, it demonstrates that its residual-quantum-attention design is practical and more accurate than all baseline models (98.6%). Lightweight residual encoding, temporal–spectral fusion, and Q-BEO optimization are among the architectural decisions that contribute to the model’s efficiency, as evidenced by its short training and inference times, low energy consumption, and modest memory utilization. These enhancements make Q-RCANeX the ideal option for off-grid cloud architectures that must consider energy consumption, latency, and cost.
The prediction for the next three days of Battery State of Charge (SOC-F) is shown in Figure 6. It reflects the regular cycle of daytime charging and nighttime discharge. Even during the typical rise-and-fall periods in the morning and evening, when forecasting models frequently shift, the predicted curve remains relatively close to the actual SOC trajectory. The magnified region demonstrates the robustness of Q-RCANeX to midday cloud-induced variations in irradiance and charge. The high R 2 value demonstrates that the battery can be tracked over time using temporal–spectral attention and quantum-aware embedding. Since this degree of accuracy permits improved storage scheduling and avoids deep-discharge incidents in autonomous cloud microgrids, it is essential for off-grid stability.
The Energy Efficiency Indicator Forecast (EEIF) for the next three days is shown in Figure 7. It demonstrates that Q-RCANeX accurately tracks a PUE-like efficiency ratio across daily temperature and workload variations. The model illustrates the advantages of nighttime cooling, how things warm up in the morning, how they cool down in the middle of the day, and how they remain constant in the evening. Additionally, it demonstrates the common disparity between cooling efficiency and energy use. In the zoomed-in region, the model’s ability to capture the interplay among small changes in ambient temperature, computational load, and inverter efficiency is most evident in the close match between the expected and true profiles. For energy-aware operational tuning in off-grid cloud scenarios, Q-RCANeX is a solid pick because of its high-fidelity monitoring, which reveals how well it can reason about time and frequency.
In Figure 8, it shows the RAC test’s confusion matrix. This proves that Q-RCANeX is quite good at distinguishing between the three levels of adequacy. Total accuracy hovers around 98.6%, with just a small number of misclassifications across all classes. The model can comprehend shifts in the net energy balance and renewable energy supply, even when sensing is somewhat imperfect or noisy. It is clear from the apparent diagonal dominance that explainability-based refinement, quantum feature embedding, and temporal–spectral reasoning all contribute to improved categorical stability. Based on these results, it is clear that Q-RCANeX can be relied on to provide practical and trustworthy sufficiency signals when deciding how to manage an off-grid enterprise.

4.5. Robustness and Reliability Evaluation

Table 8 shows that each Q-RCANeX module improves system dependability against sensor noise and environmental distortions. Signal decomposition stabilizes raw microgrid telemetry, enabling the model to work with lower signal-to-noise ratios. Residual convolutional encoding and feature refinement minimize false positives and negatives over transformer-, GNN-, and TCN-based baselines. Q-RCANeX performs well in medium and high noise when wind and sun conditions fluctuate because to the temporal-spectral attention mechanism, which improves contextual reasoning. Quantum feature embedding and the explainability layer promote positive and intelligible feature interactions, minimizing uncertainty. The hybrid Q-BEO optimizer stabilizes convergence and minimizes overfitting at all noise levels, making it crucial. This is why Q-RCANeX always beats other models with classification accuracies of 98.6%, 97.8%, and 96.9% in low, medium, and high noise with the lowest FP and FN rates. This suggests increased resilience in real-world off-grid microgrids.
In controlled testing with missing data, input streams were obfuscated at varying rates to evaluate resilience in the case of sensor data loss. The performance of Q-RCANeX is consistent regardless of the increasing number of missing values, as shown in Table 9. Incomplete telemetry is effectively handled via signal decomposition, feature refinement, and attention-based modeling. These results back up the model’s assertions that it performs well in practical sensing scenarios.
Many complementary statistical tests adapted to the forecasting task and experimental framework are utilized to provide statistically robust and transparent performance evaluation. We use Analysis of Variance (ANOVA) to determine whether several competing models evaluated under the same conditions vary statistically. Because renewable predicting performance distributions may not be normal, the Wilcoxon signed-rank test is frequently employed to compare matched models without parameter assumptions. Friedman ranking and Nemenyi and Holm post hoc adjustments control multiple-model comparisons and family-wise error rates. Comparison studies using several learning algorithms on the same datasets benefit from these tests. They ensure reliable significance testing and model ranking. This statistical technique rigorously confirms performance improvements while being methodologically transparent.
Table 10 shows that across all inferential and correlational criteria, Q-RCANeX has the best and most consistent learning behavior. A linear, monotonic alignment between the predictions and the ground truth is evident from the extremely high Pearson, Spearman, and Kendall values. There is a slight paired-sample variance, as shown by the very low Wilcoxon statistic. With an ANOVA F-score that is much higher than all of the baseline models, Q-RCANeX’s performance distribution is undeniably superior. The advantages stem from its robust preprocessing stack, temporal–spectral attention, and quantum–Bayesian optimization. Their combined effects improve consistency, reduce error variation, and strengthen the reliability of statistics for renewable dynamics in the real world.
In comparison to twelve other advanced forecasting baselines, Q-ROC RCANeX’s is shown in Figure 9. The proposed model is stable across decision thresholds and exhibits strong discriminative power, as evidenced by its much-improved area under the receiver operating characteristic (ROC) curve (AUC = 0.992). Even with lower AUC values, baseline topologies such as graph networks, hybrid transformers, and quantum-assisted LSTMs remain competitive. As a result, specificity and sensitivity are not as well balanced. Even in noisy or unclear situations, Q-RCANeX can effectively detect changes in renewable sufficiency, as seen by its crisp, almost vertical response. Renewable forecasting is more accurate and comprehensible because of the model’s enhanced capacity to reason in terms of time and frequency, to incorporate quantum properties, and to link various forms of data.
The convergence of Q-RCANeX over 200 training epochs is shown in Figure 10, where both the training and testing accuracy curves level out at around 98.7 percent after steadily improving. As a result, it is clear that the model excels at making broad generalizations and struggles with overfitting. There is a strong relationship between the training and testing losses, and both curves exhibit smooth exponential decay on the secondary axis. This shows the stability and effectiveness of the Hybrid Quantum–Bayesian Evolutionary Optimizer in parameter tuning. All three layers of quantum-embedded representation, temporal–spectral attention, and signal decomposition perform well according to the synchronized accuracy–loss alignment. This model can make real-time renewable energy predictions with confidence due to its stability, fast convergence, and consistent learning over time.
It is evident from Table 11 that, compared to other top models, Q-RCANeX performs better when it comes to computer use. With 4.9 milliseconds of inference time and 11.2 min per epoch for training, the design is ideal for real-time, resource-limited applications such as off-grid cloud microgrids. Although it employs robust quantum–Bayesian optimization and temporal–spectral attention techniques, Q-RCANeX’s model size is modest at 18.5 MB, and its compute overhead is much decreased at 2.8 GFLOPs. The rapid feature normalization pipeline, lightweight residual encoder, and hybrid optimization approach all contribute to these improvements. Compared with heavier transformer- or GNN-based models, this one demonstrates faster decision-making, lower energy consumption, and easier deployment.
Even when there is more noise and missing data, Q-RCANeX remains relatively stable, as seen in Table 12. Baseline models experience a loss in accuracy of 10-18% at a corruption level of 20%. However, Q-RCANeX only loses a small proportion, dropping from 98.6% to 96.8%. Hybrid evolutionary optimization, signal preprocessing, and quantum feature embedding work together to stabilize it. Even if the sensors are inaccurate, these three components will keep the characteristics unchanged. In addition to handling outliers, temporal–spectral attention enables robust prediction even in the presence of data gaps. These results demonstrate that compared to transformer, GNN, and CNN variants, Q-RCANeX has much better durability. As a result, it is a suitable option for off-grid cloud microgrids in the real world, where data loss and noise are prevalent.
Table 13 shows that thorough post hoc testing after the Friedman test reveals that Q-RCANeX is statistically superior to the others. Every baseline’s raw p-value remains well below 0.01, and the Nemenyi and Holm adjustments remain statistically significant. This shows that Q-RCANeX outperforms the current transformer, GNN, CNN, and hybrid systems. Performance benefits are not due to random fluctuations but to the model’s structural advantages, such as its decomposition-based preprocessing, temporal–spectral attention, and quantum–Bayesian optimization, as shown by the continuously reduced adjusted p-values (<0.03 for most models). Based on these outcomes, it is evident that the new techniques outperform their predecessors in terms of reliability, stability, and efficacy.
The primary hyperparameters controlling Q-RCANeX are examined in a comprehensive sensitivity analysis in Figure 11. Maximal sensitivity is observed for the quantum rotation factor, attention heads, convolution filters, and transformer depth. More stable feature representations in quantum systems and the ability to capture time–frequency connections are both significantly impacted by them. Less sensitive parameters that can withstand small changes without crashing include dense units, batch size, and dropout rate. The difficulties of manually balancing these interrelated components highlight the significance of the hybrid Q-BEO optimizer, as shown in this study. The results demonstrate the dependability and adaptability of Q-RCANeX. It simplifies the use of optimization strategies and handles performance-critical situations via its many components.
The stability and reliability of Q-RCANeX are shown during all five folds of cross-validation, as illustrated in Table 14. The fact that there are only minor discrepancies in accuracy, F1, AUC, and precision ( α = 0.04) indicates that the model generalizes effectively and does not overfit. With strong QICI and XSR scores, interpretability and prediction confidence remain well-matched even when the data undergoes fold-to-fold variations. The data are highly homogeneous, as shown by the stability index of 0.9996 and low entropy and variance, both indicators of a stable internal representation. Results like these demonstrate that the combination of hybrid preprocessing, temporal–spectral reasoning, and quantum–Bayesian optimization in Q-RCANeX improves forecast accuracy and stability.

4.6. Limitations and Practical Deployment Considerations

Across a range of experimental settings, the proposed Q-RCANeX framework demonstrates high levels of accuracy, robustness, and computational efficiency. On the other hand, we need to address specific practical concerns. An off-grid microgrid dataset with high resolution is utilized to evaluate the model. However, when used on a large scale across multiple sites, the model could struggle with additional geographic variations, communication delays, and synchronization issues that are not evident in this study. Improving Q-RCANeX’s functionality in federated or geographically scattered microgrid environments is a logical next step for further research. Second, access to well-calibrated sensor streams is essential for the framework. Even with resilience measures derived from signal decomposition and attention-based modeling, prediction reliability may decrease in severe deployment scenarios such as extended sensor outages, significant hardware failures, or persistent adversarial noise. These scenarios might be made even more robust with the addition of adaptive fault-detection layers or confidence-gating mechanisms that account for uncertainty.
Finally, Q-RCANeX outperforms the top models at the moment, but further model compression or quantization techniques could be required for real-time use on ultra-low-power edge devices. Addressing these concerns would make the proposed design better suited to large-scale off-grid cloud microgrid systems and improve its scalability.

5. Conclusions

This work has presented Q-RCANeX, addressing the challenge of accurate renewable energy forecasting in cloud microgrids operating independently of the primary grid. The major objective was to address existing issues in deep learning architectures, including unstable performance when data is missing or sensors are distorted, a lack of cross-modal coupling, irregular temporal patterns, and noisy multimodal inputs. Q-RCANeX proved strong temporal reasoning, represented features, and generalized across renewable forecasting tasks using a Hybrid Quantum–Bayesian Evolutionary Optimizer, temporal–spectral attention, quantum feature embedding, and residual convolutional encoding. According to experimental results, the model performs well across all assessment metrics, including 98.6% predicted accuracy, an almost-perfect area under the curve, modest error magnitudes, and high interpretability. The framework’s multi-horizon predictions for REAF, WGF, SOC-F, EEIF, and renewable adequacy classification show that it can capture fine-grain changes, long-range dependencies, and cross-modal interactions that directly affect operational dependability in off-grid scenarios. In noisy and missing data conditions, Q-RCANeX outperforms the transformer, CNN, LSTM, GNN, and hybrid baselines in robustness testing. The proposed architecture achieves high compute load, rapid inference, and low mean training cost, making it excellent for energy-aware task scheduling and real-time microgrid control, according to computational efficiency studies. The aim of creating an accurate, interpretable, resource-efficient, and robust renewable forecasting architecture is achieved. In renewable energy-powered cloud infrastructures, the integrated predictive framework improves stability, thermal efficiency, battery charge–discharge cycles, and scheduling.
Future work includes reinforcement-driven scheduling layers that enable autonomous decision-making control over the forecasting engine. An analysis of microgrid data from multiple locations can improve ecological and geographic generalizability. Lastly, lightweight neurosymbolic reasoning modules can improve operational traceability and explainability in edge deployments.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study is publicly available at Zenodo https://zenodo.org/records/17711914 (accessed on 10 January 2025).

Acknowledgments

The author extends appreciation to the Deanship of Postgraduate Studies and Scientific Research at Majmaah University for funding this research work through the project number (R-2025-2265).

Conflicts of Interest

The author declares no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AUCArea Under the Curve
CCSContextual Consistency Score
CWAComplexity-Weighted Accuracy
DGCNNDynamic Graph Convolutional Neural Network
EVElectric Vehicle
FLFederated Learning
IoTInternet of Things
MSALNet-FLMulti-Scale Adaptive Layered Network with Federated Learning
HAHBOHybrid Aquila–Honey Badger Optimization
SAFSynergistic Attribute Filtering
DLDeep Learning
DNNDeep Neural Network
MSEMean Squared Error
ReLURectified Linear Unit
GAPGlobal Average Pooling
XAIExplainable Artificial Intelligence
Q-RCANeXQuantum-Aware Residual Convolutional Attention Network with Explainability
CompmaxMaximum Computational Cost Normalizer
FedAvgFederated Averaging Algorithm
MPCMulti-Horizon Predictive Maintenance
TPTrue Positive
TNTrue Negative
FPFalse Positive
FNFalse Negative
RMSERoot Mean Squared Error
MAEMean Absolute Error
ATTAttention Mechanism
MLMachine Learning

References

  1. Aldossary, M. Optimizing task offloading for collaborative unmanned aerial vehicles (UAVs) in fog–cloud computing environments. IEEE Access 2024, 12, 74698–74710. [Google Scholar] [CrossRef]
  2. Shi, Y.; Zhang, L.; Yang, Y.; Li, Q.; Zhang, H. A review on the short-term strategy for reducing the peak–valley difference and the long-term energy structure optimization strategy in cities based on the integration of power generation–power grid–power load–energy storage. Front. Energy Res. 2025, 13, 1538811. [Google Scholar] [CrossRef]
  3. Yan, Z.; Hao, P.; Nardello, M.; Brunelli, D.; Wen, H. A Generalizable Load Recognition Method in NILM Based on Transferable Random Forest. IEEE Trans. Instrum. Meas. 2025, 74, 1–10. [Google Scholar] [CrossRef]
  4. Zhao, Z.; Yang, Q.; Zhang, Z.; Wang, Y.; Tan, H.; Wu, J.; Lai, L.L. Hierarchical distributed model predictive stabilization control of multi-scale oscillations in wind–solar hybrid multi-microgrids. IEEE Trans. Smart Grid 2024, 16, 3476–3488. [Google Scholar] [CrossRef]
  5. Atwa, W.; Almazroi, A.A.; Ayub, N. Reliable renewable energy forecasting for climate change mitigation. PeerJ Comput. Sci. 2024, 10, e2067. [Google Scholar] [CrossRef] [PubMed]
  6. Aldossary, M.; Alharbi, H.A.; Ayub, N. Exploring multi-task learning for forecasting energy-cost resource allocation in IoT–cloud systems. Comput. Mater. Contin. 2024, 79, 4603–4620. [Google Scholar] [CrossRef]
  7. Wan, A.; Du, C.; AL-Bukhaiti, K.; Chen, P. Optimizing combined-cycle power plant operations using an LSTM-attention hybrid model for load forecasting. J. Mech. Sci. Technol. 2025, 39, 6371–6380. [Google Scholar] [CrossRef]
  8. Bakht, M.P.; Mohd, M.N.H.; Sheikh, U.U.; Khan, N. Optimal design and performance analysis of hybrid renewable energy system for ensuring uninterrupted power supply during load shedding. IEEE Access 2024, 12, 5792–5813. [Google Scholar] [CrossRef]
  9. Liu, Y.; Yao, H.; Di, P.; Qin, Y.; Ma, Y.; Alkahtani, M.; Hu, Y. Region of Attraction Estimation for Power Systems with Multiple Integrated DFIG-Based Wind Turbines. IEEE Trans. Sustain. Energy 2025, 16, 3095–3109. [Google Scholar] [CrossRef]
  10. Alzamil, I. Federated Deep Learning for Scalable and Explainable Load Forecasting in Privacy-Conscious Smart Cities. IEEE Access 2025, 13, 142237–142250. [Google Scholar] [CrossRef]
  11. Zaheer, R.; Hanif, M.K.; Sarwar, M.U.; Talib, R. Evaluating the effectiveness of dimensionality reduction on machine learning algorithms in time series forecasting. IEEE Access 2025, 13, 50493–50510. [Google Scholar] [CrossRef]
  12. Shen, H.; Wang, H.; Gao, J.; Buyya, R. An instability-resilient renewable energy allocation system for a cloud datacenter. IEEE Trans. Parallel Distrib. Syst. 2023, 34, 1020–1034. [Google Scholar] [CrossRef]
  13. El-Saieed, A.M. An adaptive transformer model for long-term time-series forecasting of temperature and radiation in photovoltaic energy generation. Appl. Comput. Intell. Soft Comput. 2025, 2025, 6671565. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Zhang, Y.; Zheng, B.; Cui, H.; Qi, H. Statistical analysis for estimating the optimized battery capacity for roof-top PV energy system. Renew. Energy 2025, 242, 122491. [Google Scholar] [CrossRef]
  15. Ozbak, M.; Ghazizadeh-Ahsaee, M.; Ahrari, M.; Jahantigh, M.; Mirshekar, S.; Mirmozaffari, M.; Aranizadeh, A. Improving power output wind turbine in micro-grids assisted virtual wind speed prediction. Sustain. Oper. Comput. 2024, 5, 119–130. [Google Scholar] [CrossRef]
  16. Saxena, A.; Shankar, R.; El-Saadany, E.F.; Kumar, M.; Al Zaabi, O.; Al Hosani, K.; Muduli, U.R. Intelligent load forecasting and renewable energy integration for enhanced grid reliability. IEEE Trans. Ind. Appl. 2024, 60, 8403–8417. [Google Scholar] [CrossRef]
  17. Deepa, R.; Shakila Devi, G.T.; Balamurugan, K.S. Hybrid photovoltaic wind renewable energy sources for microgrid using osprey optimization algorithm and augmented physics-informed neural network. J. Renew. Sustain. Energy 2025, 17, 056501. [Google Scholar]
  18. Li, Y.; Song, L.; Zhang, S.; Kraus, L.; Adcox, T.; Willardson, R.; Lu, N. A TCN-based hybrid forecasting framework for hours-ahead utility-scale PV forecasting. IEEE Trans. Smart Grid 2023, 14, 4073–4085. [Google Scholar] [CrossRef]
  19. Zhuang, W.; Fan, J.; Xia, M.; Zhu, K. A multi-scale spatial–temporal graph neural network-based method of multienergy load forecasting in integrated energy system. IEEE Trans. Smart Grid 2023, 15, 2652–2666. [Google Scholar] [CrossRef]
  20. Aldossary, M.; Almutairi, J.; Alzamil, I. Federated LeViT-ResUNet for Scalable and Privacy-Preserving Agricultural Monitoring Using Drone and Internet of Things Data. Agronomy 2025, 15, 928. [Google Scholar] [CrossRef]
  21. Yang, Y.; Liu, Y.; Zhang, Y.; Shu, S.; Zheng, J. DEST-GNN: A double-explored spatio-temporal graph neural network for multi-site intra-hour PV power forecasting. Appl. Energy 2025, 378, 124744. [Google Scholar] [CrossRef]
  22. Meng, Q.; He, Y.; Hussain, S.; Lu, J.; Guerrero, J.M. Day-ahead economic dispatch of wind-integrated microgrids using coordinated energy storage and hybrid demand response strategies. Sci. Rep. 2025, 15, 26579. [Google Scholar] [CrossRef]
  23. Suresh, V. Benchmarking transformer variants for hour-ahead PV forecasting: PatchTST with adaptive conformal inference. Energies 2025, 18, 5000. [Google Scholar] [CrossRef]
  24. Petrosian, O.; Zhang, Y. Solar power generation forecasting in smart cities and explanation based on explainable AI. Smart Cities 2024, 7, 3388–3411. [Google Scholar] [CrossRef]
  25. Gao, H.; Yang, D.; Lv, Y.; Wang, L. Unsupervised Machine Learning Approach to Enhance Online Voltage Security Assessment Based on Synchrophasor Data. IEEE Trans. Power Syst. 2025, 40, 3596–3599. [Google Scholar] [CrossRef]
  26. Moosavi, S.; Farajzadeh-Zanjani, M.; Razavi-Far, R.; Palade, V.; Saif, M. Explainable AI in manufacturing and industrial cyber–physical systems: A survey. Electronics 2024, 13, 3497. [Google Scholar] [CrossRef]
  27. Zhu, J.; Wang, X.; Cao, G.; Xu, L.; Cao, Y. Quantum interval neural network for uncertain structural static analysis. Int. J. Mech. Sci. 2025, 303, 110646. [Google Scholar] [CrossRef]
  28. Aldossary, M. Q-MobiGraphNet: Quantum-Inspired Multimodal IoT and UAV Data Fusion for Coastal Vulnerability and Solar Farm Resilience. Mathematics 2025, 13, 3051. [Google Scholar] [CrossRef]
  29. Liu, Y.; Wang, Y.; Chen, S.; Zhang, J. A Novel Hybrid Neural Lyapunov Method With Low Conservatism for Power System Domain of Attraction Estimation. IEEE Trans. Ind. Inform. 2025, 21, 5580–5591. [Google Scholar] [CrossRef]
  30. Aldossary, M.; Alzamil, I.; Almutairi, J. Enhanced Intrusion Detection in Drone Networks: A Cross-Layer Convolutional Attention Approach for Drone-to-Drone and Drone-to-Base Station Communications. Drones 2025, 9, 46. [Google Scholar] [CrossRef]
  31. Cavus, M.; Allahham, A. Spatio-temporal attention-based deep learning for smart grid demand prediction. Electronics 2025, 14, 2514. [Google Scholar] [CrossRef]
  32. National University Hospital Energy Innovation Center (NUHEIC). Off-Grid Cloud Data Center Renewable Forecasting Dataset [Data Set]. Zenodo. 2025. Available online: https://zenodo.org/records/17711914 (accessed on 10 January 2025).
  33. Wang, H.; Li, Y.; Men, T.; Li, L. Physically Interpretable Wavelet-Guided Networks With Dynamic Frequency Decomposition for Machine Intelligence Fault Prediction. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 4863–4875. [Google Scholar] [CrossRef]
  34. Ahmadi, M.; Aly, H.; Khashei, M. Enhancing power grid stability with a hybrid framework for wind power forecasting: Integrating Kalman filtering, deep residual learning, and bidirectional LSTM. Energy 2025, 334, 137752. [Google Scholar] [CrossRef]
  35. Zuo, C.; Zhang, X.; Zhao, G.; Yan, L. PCR: A Parallel Convolution Residual Network for Traffic Flow Prediction. IEEE Trans. Emerg. Top. Comput. Intell. 2025, 9, 3072–3083. [Google Scholar] [CrossRef]
  36. Yu, H.; Yang, H.; Gao, L.; Hu, J.; Plaza, A.; Zhang, B. Hyperspectral image change detection based on gated spectral–spatial–temporal attention network with spectral similarity filtering. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5511313. [Google Scholar] [CrossRef]
  37. Vincent, A.M.; Jidesh, P. An improved hyperparameter optimization framework for AutoML systems using evolutionary algorithms. Sci. Rep. 2023, 13, 4737. [Google Scholar] [CrossRef]
  38. de la Cruz Huayanay, A.; Bazán, J.L.; Russo, C.M. Performance of evaluation metrics for classification in imbalanced data. Comput. Stat. 2025, 40, 1447–1473. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed Q-RCANeX renewable energy forecasting framework.
Figure 1. Overview of the proposed Q-RCANeX renewable energy forecasting framework.
Mathematics 14 00181 g001
Figure 2. Proposed Q-RCANeX architecture (pictorial view).
Figure 2. Proposed Q-RCANeX architecture (pictorial view).
Mathematics 14 00181 g002
Figure 3. SHAP-based global feature importance for Q-RCANeX across all forecasting tasks.
Figure 3. SHAP-based global feature importance for Q-RCANeX across all forecasting tasks.
Mathematics 14 00181 g003
Figure 4. Day-ahead REAF (kWh) forecast showing actual vs. predicted values with zoomed temporal segment for fine-scale comparison.
Figure 4. Day-ahead REAF (kWh) forecast showing actual vs. predicted values with zoomed temporal segment for fine-scale comparison.
Mathematics 14 00181 g004
Figure 5. 3-Day Wind Generation Forecast (WGF): Actual vs. predicted values with a zoomed inset of evening gust dynamics.
Figure 5. 3-Day Wind Generation Forecast (WGF): Actual vs. predicted values with a zoomed inset of evening gust dynamics.
Mathematics 14 00181 g005
Figure 6. 3-Day Battery State-of-Charge Forecast (SOC-F): Actual vs. predicted values with zoomed inset highlighting daytime charging and nocturnal discharge behavior.
Figure 6. 3-Day Battery State-of-Charge Forecast (SOC-F): Actual vs. predicted values with zoomed inset highlighting daytime charging and nocturnal discharge behavior.
Mathematics 14 00181 g006
Figure 7. EEIF (Energy Efficiency Indicator Forecast)—Actual vs. predicted values over a 3-day window, with inset highlighting daytime thermal peaks and cooling-efficiency transitions.
Figure 7. EEIF (Energy Efficiency Indicator Forecast)—Actual vs. predicted values over a 3-day window, with inset highlighting daytime thermal peaks and cooling-efficiency transitions.
Mathematics 14 00181 g007
Figure 8. Confusion matrix for Renewable Adequacy Class (RAC) classification.
Figure 8. Confusion matrix for Renewable Adequacy Class (RAC) classification.
Mathematics 14 00181 g008
Figure 9. ROC curve comparison between Q-RCANeX and competing baseline models.
Figure 9. ROC curve comparison between Q-RCANeX and competing baseline models.
Mathematics 14 00181 g009
Figure 10. Training and testing accuracy–loss progression for Q-RCANeX over 200 epochs.
Figure 10. Training and testing accuracy–loss progression for Q-RCANeX over 200 epochs.
Mathematics 14 00181 g010
Figure 11. Sensitivity analysis of major Q-RCANeX hyperparameters, illustrating the relative influence of each parameter on overall forecasting performance.
Figure 11. Sensitivity analysis of major Q-RCANeX hyperparameters, illustrating the relative influence of each parameter on overall forecasting performance.
Mathematics 14 00181 g011
Table 1. Summary of recent related studies (2024–2025) in renewable energy forecasting and intelligent microgrid modeling.
Table 1. Summary of recent related studies (2024–2025) in renewable energy forecasting and intelligent microgrid modeling.
Ref.ObjectiveMethod/FrameworkKey AchievementsLimitations
[13]Multi-step PV forecasting under variable climatesTransformer-based model with horizon-specific decoders and attention layersImproved MAE/RMSE across multiple sites and climatic zonesReduced accuracy under missing telemetry or sensor drift
[14]Joint forecasting of hybrid solar–wind generationHybrid transformer with shared temporal encoder and modality headsAccurately models interdependencies between solar and wind modalitiesRelies heavily on high-quality reanalysis and meteorological data
[16]Stabilization of noisy microgrid data streamsSignal-augmentation and quality-control (SAQ) pipeline for preprocessingEnhances short-term PV/load prediction and reduces curtailment riskIncreases preprocessing complexity and computational overhead
[17]Physics-driven PV forecasting for small-scale microgridsFeature engineering using clear-sky index, transposition, and meteorological interactionsCompetitive accuracy with lower computation vs. deep modelsSite-specific calibration limits transferability
[18]Hybrid solar–wind time-series modelingTemporal convolutional network (TCN) capturing long/short dependenciesImproves ramp prediction and temporal stability vs. LSTM/CNNLacks multimodal feature fusion, limited to single-source data
[19]Spatial–temporal wind power forecastingGraph neural network (GNN) with recurrent layers for spatial encodingIncreases accuracy on dense sensor networksStruggles under sparse configurations and cold-start sites
[21]Multi-site PV forecasting with uncertainty calibrationDouble-explored spatio-temporal GNN using alternating attention layersReduces forecasting error and improves uncertainty estimationComputationally expensive and sensitive to over-smoothing
[22]Spatial dependency modeling for solar plantsSelf-constructed GNN learning from telemetry without predefined graphEnhances site-specific RMSE reduction over baselinesLimited generalization beyond trained solar farms
[23]Efficient hour-ahead PV forecastingComparative transformer architecture benchmarkingSmaller transformers outperform deeper models under tuned horizonsHighly dependent on feature completeness and tuning
[24]Explainable solar forecastingSHAP-integrated deterministic forecasting frameworkImproves interpretability and operator trustComputational overhead during inference
[25]Real-time explainable fault detection in forecastingSHAP-driven diagnostic framework for feature contribution trackingEnables transparent forecasting and early fault detectionHigher inference latency with high-dimensional features
[26]Comprehensive review of explainable AI in energy systemsSurvey and evaluation of saliency and SHAP-based interpretation modelsImproves understanding of transparency–accuracy tradeoffsExplanations unstable under distribution shift or extreme weather
[27]Quantum-inspired time-series forecastingQuantum-LSTM with parameterized quantum circuitsFaster convergence and reduced error on small datasetsScalability limited by quantum simulation cost
[29]Unified temporal–spectral modeling for wind forecastingHybrid TCN-transformer (TCNFormer) combining local/global learningAchieves lower RMSE than individual TCN or Transformer modelsHigh training complexity and hyperparameter sensitivity
[31]Lightweight forecasting for hybrid PV–wind systemsCompact 1D CNN optimized for low-latency inferenceMaintains high accuracy with minimal computational costLacks uncertainty quantification; reduced robustness under extreme events
Table 2. Categorized dataset features and their distributions.
Table 2. Categorized dataset features and their distributions.
Feature CategoryExample FeaturesDistribution Type
Environmental and MeteorologicalGlobal Horizontal Irradiance (GHI), Direct Normal Irradiance (DNI), Diffuse Horizontal Irradiance (DHI), Air Temperature, Relative Humidity, Wind Speed (10 m/100 m), Wind Direction, Atmospheric Pressure, Cloud Cover, Precipitation Rate, Aerosol Optical Depth (AOD)Imbalanced and skewed numerical (beta/exponential)
Solar Generation and Photovoltaic SystemPV DC Power, PV AC Power, Module Temperature, Inverter Efficiency, Panel Tilt, Panel Azimuth, Curtailment FlagImbalanced numerical with intermittent spikes (right-skewed)
Wind Generation SystemTurbine Power Output, Rotor Speed, Pitch Angle, Nacelle Direction, Turbine AvailabilityImbalanced numerical (Weibull/cubic relation)
Energy Storage and Battery ManagementBattery State of Charge (SOC), State of Health (SOH), Charge Power, Discharge Power, Battery Temperature, Storage EfficiencyImbalanced numerical (skewed and truncated)
Data-Center Operational MetricsIT Load, Total Load, CPU Utilization, GPU Utilization, Memory Utilization, Cooling Power, Inlet/Outlet Temperature, Fan Speed, Power Usage Effectiveness (PUE)Imbalanced numerical (log-normal/heavy-tailed)
Temporal and Contextual ParametersTimestamp, Hour of Day, Day of Week, Month, Season Index, Holiday IndicatorImbalanced categorical and cyclic numerical
Derived and Statistical IndicatorsNet Energy Balance (NEB), Power Ramp Rate, Clear-Sky Index, Cross-Correlation, Entropy of Residuals, Imputation MaskImbalanced continuous (non-Gaussian)
Target LabelsRenewable Energy Availability Forecast (REAF), Solar Generation Forecast (SGF), Wind Generation Forecast (WGF), Battery SOC Forecast (SOC-F), Energy Efficiency Indicator (EEIF), Net Energy Balance (NEB), Renewable Adequacy Class (RAC), Sustainable Operation Window (SOW)Imbalanced regression and classification targets
Table 3. Statistical summary of representative dataset features.
Table 3. Statistical summary of representative dataset features.
FeatureMeanStd. Dev.Missing Rate (%)
Global Horizontal Irradiance (GHI)412.6298.43.2
Wind Speed (10 m)6.873.142.8
PV AC Power Output482.1356.93.6
Battery State of Charge (SOC)61.318.71.9
Total Data-Center Load734.5211.32.4
Cooling Power Consumption128.749.22.1
Table 4. Comprehensive hyperparameter settings used for training and evaluation of Q-RCANeX.
Table 4. Comprehensive hyperparameter settings used for training and evaluation of Q-RCANeX.
CategoryValue/Setting
Training Parameters
Batch size (B)32 (optimal), tested in { 16 , 32 , 64 }
Learning rate ( η ) 1.5 × 10 4 (Q-BEO optimized)
Epochs (total)200
Loss functionSmooth L1 + Temporal Stability Regularizer
Activation functionsReLU (encoder), GELU (attention), Sigmoid/Linear (forecast heads)
Weight initializationXavier uniform (CNN), Orthogonal (transformer)
Gradient clippingMax-norm = 1.0
Dropout rate (p)0.15 (optimized)
Q-BEO Optimizer Parameters
Optimizer typeHybrid Quantum–Bayesian Evolutionary Optimizer
Quantum rotation coefficient ( γ q )0.72
Bayesian update weight ( β b )0.58
Search iterations30 per tuning cycle
Learning rate adaptationNonlinear cosine annealing
Architecture Parameters
Residual encoder depth4 convolutional blocks
Embedding dimension (d)64 (quantum-embedded)
Temporal–spectral attention layers6 layers, 4 heads each
Window/patch size16 time-steps
Quantum feature embedding (QFE) size64-dimensional rotation-encoded space
Forecasting headsREAF, WGF, SOC-F, EEIF, RAC classification
Dataset and Preprocessing
Total samples≈120,000 (10 min resolution)
Decomposition methodSignal Decomposition Pipeline (SDP)
NormalizationMin–max + temporal Z-score smoothing
Missing data handlingInterpolation + quantum-enhanced imputation
Hardware/System Setup
GPUNVIDIA GeForce RTX 4060 (8 GB VRAM)
CPUIntel Core i7
FrameworksPyTorch 2.0, CUDA 12.x
Table 5. Performance comparison of Q-RCANeX with baseline models for renewable forecasting in off-grid cloud environments.
Table 5. Performance comparison of Q-RCANeX with baseline models for renewable forecasting in off-grid cloud environments.
Model/MethodDRSIHRCA (%)MDERQEQICI
Transformer + specific decoders + attention layers [13]0.8650.8530.85980.80.1280.1740.772
Hybrid Transformer + temporal encoder + modality heads [14]0.8740.8620.86882.30.1210.1680.781
Signal-augmentation and quality-control (SAQ) pipeline [16]0.8690.8570.86381.60.1320.1770.768
Temporal Convolutional Network (TCN) [18]0.8780.8660.87283.10.1170.1620.786
Graph Neural Network (GNN) with recurrent layers [19]0.8850.8740.87984.50.1120.1560.795
Double-Explored Spatio-Temporal GNN (DEST-GNN) [21]0.8910.8790.88585.40.1080.1490.803
Self-Constructed Graph Neural Network [22]0.8830.8720.87784.20.1150.1580.792
Transformer (benchmark variant) [23]0.8750.8610.86882.90.1220.1690.777
Explainable AI (saliency + SHAP-based) [26]0.8880.8760.88285.00.1140.1570.824
Quantum-LSTM [27]0.8960.8850.89086.40.1090.1520.835
Hybrid TCN–Transformer (TCNFormer) [29]0.9030.8920.89887.80.1040.1450.842
Lightweight 1D CNN [31]0.8710.8580.86481.20.1300.1750.770
Proposed Q-RCANeX (ours)0.9830.9790.98198.60.0510.0910.938
Table 6. Comprehensive ablation study of Q-RCANeX: Impact of preprocessing, architectural, and optimization modules.
Table 6. Comprehensive ablation study of Q-RCANeX: Impact of preprocessing, architectural, and optimization modules.
Model Configuration/Module AddedCA (%)PRREF1DRSIHRMDERQEQICI
Baseline CNN (no preprocessing, no enhancement)81.80.8420.8360.8390.8710.8580.8640.1420.1880.742
+ Signal Decomposition–Based Preprocessing (SDP)85.20.8680.8570.8620.8860.8730.8790.1260.1720.771
+ Feature Extraction and Selection (FES)87.90.8810.8740.8770.8910.8780.8840.1180.1640.785
+ Residual Convolutional Encoder (RCE)90.60.9020.8960.8990.9110.9010.9060.1070.1520.812
+ Temporal–Spectral Attention (TSA)93.20.9240.9180.9210.9330.9210.9270.0960.1380.841
+ Quantum Feature Embedding (QFE)95.10.9410.9360.9380.9480.9350.9420.0840.1210.873
+ Explainability Module (XAI Layer)96.40.9540.9490.9520.9610.9470.9540.0770.1090.902
+ Hybrid Quantum–Bayesian Evolutionary Optimization (Q-BEO)98.60.9740.9690.9710.9830.9790.9810.0510.0910.938
Table 7. Computational efficiency comparison of Q-RCANeX and baseline models.
Table 7. Computational efficiency comparison of Q-RCANeX and baseline models.
Model/MethodCA (%)Params (M)FLOPs (G)TT (s/epoch)IT (ms/sample)EJ (kJ)MU (MB)CRQICI
Transformer + specific decoders + attention layers [13]80.814.723.441.26.11.935151280.68
Hybrid Transformer + temporal encoder + modality heads [14]82.316.225.144.66.52.115471240.71
SAQ pipeline [16]83.412.921.736.45.91.744931200.73
TCN [18]84.29.618.333.75.41.624701100.74
GNN with recurrent layers [19]85.117.326.949.37.12.265621320.77
DEST-GNN [21]85.918.227.552.17.42.315831360.79
Self-Constructed Graph Neural Network [22]86.219.028.654.77.62.436011400.80
Transformer [23]84.915.824.942.96.32.085321220.75
Explainable AI (saliency and SHAP-based) [26]86.416.525.746.56.82.195551260.84
Quantum-LSTM [27]87.213.722.439.25.71.825091180.86
TCNFormer [29]87.817.626.347.16.92.235711300.82
Lightweight 1D CNN [31]89.38.414.929.64.81.424351050.79
Proposed Q-RCANeX (ours)98.611.317.226.43.91.07401820.94
Table 8. Robustness evaluation of Q-RCANeX and baseline models under sensor noise and environmental distortion.
Table 8. Robustness evaluation of Q-RCANeX and baseline models under sensor noise and environmental distortion.
Model/MethodLow Noise (SNR > 30 dB)Medium Noise (SNR = 20–30 dB)High Noise (SNR < 20 dB)
FP (%)FN (%)CA (%)FP (%)FN (%)CA (%)FP (%)FN (%)CA (%)
Transformer + specific decoders + attention layers [13]5.76.990.19.211.184.313.815.678.5
Hybrid Transformer + temporal encoder + modality heads [14]5.36.591.28.810.286.212.614.480.7
SAQ pipeline [16]4.96.192.47.99.688.411.713.882.3
TCN [18]4.75.893.17.59.289.310.913.083.4
GNN with recurrent layers [19]4.55.593.87.28.989.910.212.484.6
DEST-GNN [21]4.25.394.46.88.690.79.811.985.2
Quantum-LSTM [27]3.84.895.16.38.091.69.111.386.1
TCNFormer [29]3.54.695.85.97.692.48.510.787.3
Lightweight 1D CNN [31]3.24.396.25.67.393.18.210.387.9
Proposed Q-RCANeX (ours)1.82.398.62.53.097.83.84.596.9
Table 9. Robustness evaluation of Q-RCANeX under varying sensor data missing rates.
Table 9. Robustness evaluation of Q-RCANeX under varying sensor data missing rates.
Missing Rate (%)CA (%)PRREF1MDEQICI
0 (No Missing Data)98.60.9740.9690.9710.0510.938
598.10.9690.9640.9660.0570.931
1097.40.9620.9570.9590.0640.924
1596.50.9540.9480.9510.0720.916
2095.30.9440.9370.9400.0830.905
Table 10. Statistical analysis of Q-RCANeX compared with existing baseline models.
Table 10. Statistical analysis of Q-RCANeX compared with existing baseline models.
Model (with Reference)ANOVA Fp-ValuePearson rSpearman ρ Kendall τ Wilcoxon WInterpretation
Proposed Q-RCANeX128.4<0.0010.9870.9820.9110.0004Highest statistical strength and model consistency
Transformer + Attention Decoders [13]77.9<0.0010.9120.8980.7360.012Strong evidence of good correlation but lower stability
Hybrid Transformer (Modality Heads) [14]71.4<0.0010.9040.8870.7210.017Noticeable improvement, limited by data-quality dependence
SAQ-Based Forecasting Pipeline [16]68.3<0.0010.8930.8750.7010.021Good alignment; preprocessing overhead affects reliability
Temporal Convolutional Network (TCN) [18]59.2<0.0010.8810.8610.6830.029Stable temporal learning; lacks multimodal fusion
GNN + Recurrent Temporal Layers [19]54.7<0.0010.8740.8540.6720.034Effective spatial modelling; sensitive to graph sparsity
DEST-GNN Spatio-Temporal Model [21]61.0<0.0010.8860.8650.6890.026Improved cross-site learning; higher training cost
Self-Constructed Graph Neural Network [22]58.4<0.0010.8790.8580.6810.031Captures site-specific patterns; limited generalizability
Compact Transformer Variant [23]66.1<0.0010.8910.8700.6950.024Efficient architecture but sensitive to feature sparsity
Explainable AI (SHAP/Saliency) Model [26]52.6<0.0010.8650.8410.6510.039Enhances interpretability; moderate predictive alignment
Quantum-LSTM Hybrid Model [27]49.7<0.0010.8520.8260.6290.044Promising convergence; limited scalability
TCNFormer (TCN + Transformer) [29]63.8<0.0010.8890.8670.6930.025Strong hybrid modeling; higher hyperparameter sensitivity
Lightweight 1D CNN [31]46.3<0.0010.8410.8180.6160.048Fast inference but reduced robustness under extremes
Table 11. Comparative computational efficiency of Q-RCANeX and state-of-the-art baseline models.
Table 11. Comparative computational efficiency of Q-RCANeX and state-of-the-art baseline models.
ModelInference Time (ms)Training Time (min/epoch)Model Size (MB)Compute Load (GFLOPs)
Proposed Q-RCANeX4.911.218.52.8
Transformer + Decoders + Attention Layers [13]12.828.442.19.7
Hybrid Transformer (Temporal Encoder + Modality Heads) [14]14.331.248.511.3
SAQ-Based Forecasting Pipeline [16]9.619.825.74.5
Temporal Convolutional Network (TCN) [18]7.415.121.43.1
GNN + Recurrent Temporal Layers [19]15.935.357.912.6
DEST-GNN Spatio-Temporal Architecture [21]16.838.561.314.2
Self-Constructed Graph Neural Network [22]14.729.646.810.4
Compact Transformer Variant [23]11.122.733.57.2
Explainable AI (SHAP-Based Forecasting) [26]18.541.452.96.8
Quantum-LSTM Hybrid [27]13.626.139.48.6
TCNFormer (TCN + Transformer Hybrid) [29]10.924.836.27.9
Lightweight 1D CNN [31]5.89.412.31.9
Table 12. Robustness analysis of Q-RCANeX and baseline models under increasing noise and missing data levels.
Table 12. Robustness analysis of Q-RCANeX and baseline models under increasing noise and missing data levels.
Model0% Noise/
Missing Data
10% Noise/
Missing Data
20% Noise/
Missing Data
Proposed Q-RCANeX98.6%97.9%96.8%
Transformer + Attention Decoders [13]89.2%85.6%80.4%
Hybrid Transformer (Temporal Encoder + Modality Heads) [14]88.7%84.1%79.3%
SAQ-Based Forecasting Pipeline [16]87.9%83.2%77.8%
Temporal Convolutional Network (TCN) [18]86.4%82.5%76.1%
GNN + Recurrent Temporal Layers [19]85.2%80.7%74.9%
DEST-GNN Spatio-Temporal Network [21]87.1%82.9%77.2%
Self-Constructed GNN [22]86.7%82.1%75.8%
Compact Transformer Variant [23]87.6%83.5%77.9%
Explainable AI (SHAP-Based Pipeline) [26]84.9%79.3%73.6%
Quantum-LSTM Hybrid [27]83.8%78.6%72.4%
TCNFormer (TCN + Transformer) [29]87.4%83.1%76.8%
Lightweight 1D CNN [31]82.7%77.2%71.3%
Table 13. Significance verification of Q-RCANeX performance against baseline models using Nemenyi and Holm post hoc corrections after Friedman testing.
Table 13. Significance verification of Q-RCANeX performance against baseline models using Nemenyi and Holm post hoc corrections after Friedman testing.
Model Compared with Q-RCANeXRaw p-ValueNemenyi-Adjusted p-ValueHolm-Adjusted p-Value
Transformer + Decoders + Attention Layers [13]0.00090.01080.0045
Hybrid Transformer (Temporal Encoder + Modality Heads) [14]0.00130.01560.0067
SAQ-Based Forecasting Pipeline [16]0.00210.02520.0094
Temporal Convolutional Network (TCN) [18]0.00340.04080.0142
GNN + Recurrent Temporal Layers [19]0.00070.00840.0038
DEST-GNN Spatio-Temporal Model [21]0.00100.01200.0052
Self-Constructed Graph Neural Network [22]0.00180.02160.0081
Compact Transformer Variant [23]0.00490.05880.0198
Explainable AI (SHAP-Based) [26]0.00630.07560.0247
Quantum-LSTM Hybrid [27]0.00260.03120.0114
TCNFormer (TCN + Transformer) [29]0.00310.03720.0130
Lightweight 1D CNN [31]0.00570.06840.0224
Table 14. Comprehensive statistical evaluation of 5-fold cross-validation for Q-RCANeX.
Table 14. Comprehensive statistical evaluation of 5-fold cross-validation for Q-RCANeX.
MetricFold 1Fold 2Fold 3Fold 4Fold 5Mean ( μ )Std. Dev. ( σ )95% CI (Lower–Upper)
Accuracy (%)98.5798.6398.6898.5998.6698.630.04[98.56, 98.70]
F1 (%)98.4998.5598.6198.5298.5998.550.04[98.48, 98.62]
AUC (%)98.7198.7898.8298.7498.8098.770.04[98.70, 98.84]
Precision (%)98.5198.5798.6398.5498.6198.570.04[98.49, 98.64]
Recall (%)98.4798.5298.5898.5098.5698.520.04[98.45, 98.59]
QICI0.9380.9440.9470.9410.9460.9430.004[0.937, 0.949]
XAI–Stability Ratio (XSR)0.9580.9620.9670.9610.9680.9630.004[0.957, 0.969]
Variance ( σ 2 )0.00180.00200.00190.00210.00170.00190.0002
Entropy ( H )0.06150.06080.06010.06190.06030.06100.0007
Reliability Coeff. ( ρ r )0.9890.9910.9920.9880.9930.9910.002[0.988, 0.993]
Bias Index ( B β )0.00410.00440.00400.00450.00390.00420.0002[0.0039, 0.0045]
Coeff. of Variation ( CV % )0.100.110.120.110.09Mean CV ≈ 0.11 (high cross-fold consistency)
Generalization Entropy ( G e )0.0785 (Low entropy → strong generalization capability)
Stability Index ( Υ = 1 σ ( Acc ) μ ( Acc ) )0.9996 (Excellent statistical homogeneity)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alzamil, I. Quantum-Enhanced Residual Convolutional Attention Architecture for Renewable Forecasting in Off-Grid Cloud Microgrids. Mathematics 2026, 14, 181. https://doi.org/10.3390/math14010181

AMA Style

Alzamil I. Quantum-Enhanced Residual Convolutional Attention Architecture for Renewable Forecasting in Off-Grid Cloud Microgrids. Mathematics. 2026; 14(1):181. https://doi.org/10.3390/math14010181

Chicago/Turabian Style

Alzamil, Ibrahim. 2026. "Quantum-Enhanced Residual Convolutional Attention Architecture for Renewable Forecasting in Off-Grid Cloud Microgrids" Mathematics 14, no. 1: 181. https://doi.org/10.3390/math14010181

APA Style

Alzamil, I. (2026). Quantum-Enhanced Residual Convolutional Attention Architecture for Renewable Forecasting in Off-Grid Cloud Microgrids. Mathematics, 14(1), 181. https://doi.org/10.3390/math14010181

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop