Next Article in Journal
AI Correction of Smartphone Thermal Images: Application to Diabetic Plantar Foot
Previous Article in Journal
A Survey on Fault Detection of Solar Insecticidal Lamp Internet of Things: Recent Advance, Challenge, and Countermeasure
Previous Article in Special Issue
A Novel Method for Simulation Model Generation of Production Systems Using PLC Sensor and Actuator State Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Wavelet–Transformer–XGBoost Model Optimized by Chaotic Billiards for Global Irradiance Forecasting

1
Laboratory of Robotics, Informatics and Complex Systems (RISC), National Engineering School of Tunis (ENIT), University of Tunis El Manar (UTM), BP N° 37, Le Belvedere, Tunis 1002, Tunisia
2
Department of Biomedicine, Neuroscience, and Advanced Diagnostics (BiND), University of Palermo, 90127 Palermo, Italy
3
Department of Engineering, University of Messina, 98166 Messina, Italy
4
Higher Institute of Informations and Communication Technology (ISTIC), University of Carthage, BP N° 123, Hammam Chatt 1164, Tunisia
5
Department of Mathematics, College of Science and Humanities, Shaqra University, Dawadmi 11911, Saudi Arabia
6
Department of Computer Science and Artificial Intelligence, College of Computer Science and Engineering, University of Jeddah, Jeddah 21493, Saudi Arabia
*
Author to whom correspondence should be addressed.
J. Sens. Actuator Netw. 2026, 15(1), 12; https://doi.org/10.3390/jsan15010012
Submission received: 10 December 2025 / Revised: 8 January 2026 / Accepted: 20 January 2026 / Published: 22 January 2026
(This article belongs to the Special Issue AI and IoT Convergence for Sustainable Smart Manufacturing)

Abstract

Accurate global irradiance (GI) forecasting is essential for improving photovoltaic (PV) energy management, stabilizing renewable power systems, and enabling intelligent control in solar-powered applications, including electric vehicles and smart grids. The highly stochastic and non-stationary nature of solar radiation, influenced by rapid atmospheric fluctuations and seasonal variability, makes short-term GI prediction a challenging task. To overcome these limitations, this work introduces a new hybrid forecasting architecture referred to as WTX–CBO, which integrates a Wavelet Transform (WT)-based decomposition module, an encoder–decoder Transformer model, and an XGBoost regressor, optimized using the Chaotic Billiards Optimizer (CBO) combined with the Adam optimization algorithm. In the proposed architecture, WT decomposes solar irradiance data into multi-scale components, capturing both high-frequency transients and long-term seasonal patterns. The Transformer module effectively models complex temporal and spatio-temporal dependencies, while XGBoost enhances nonlinear learning capability and mitigates overfitting. The CBO ensures efficient hyperparameter tuning and accelerated convergence, outperforming traditional meta-heuristics such as Particle Swarm Optimization (PSO) and Genetic Algorithms (GA). Comprehensive experiments conducted on real-world GI datasets from diverse climatic conditions demonstrate the outperformance of the proposed model. The WTX–CBO ensemble consistently outperformed benchmark models, including LSTM, SVR, standalone Transformer, and XGBoost, achieving improved accuracy, stability, and generalization capability. The proposed WTX–CBO framework is designed as a high-accuracy decision-support forecasting tool that provides short-term global irradiance predictions to enable intelligent energy management, predictive charging, and adaptive control strategies in solar-powered applications, including solar electric vehicles (SEVs), rather than performing end-to-end vehicle or photovoltaic power simulations. Overall, the proposed hybrid framework provides a robust and scalable solution for short-term global irradiance forecasting, supporting reliable PV integration, smart charging control, and sustainable energy management in next-generation solar systems.

1. Introduction

1.1. Background and Challenges

Solar-powered electric vehicles (SEVs) represent a promising pathway toward sustainable and autonomous transportation [1]. By integrating photovoltaic (PV) modules into vehicle surfaces, SEVs can harvest solar energy to extend the driving range [2], reduce grid reliance, and enable environmentally friendly mobility [3]. Efficient real-time solar irradiance forecasting is therefore essential to support predictive charging [4], adaptive energy routing [5], and intelligent battery management in SEVs. However, solar irradiance is highly intermittent and non-stationary due to dynamic meteorological conditions [6], including rapid cloud movements, aerosol variations, humidity fluctuations, and atmospheric turbulence [7]. These abrupt and multi-scale fluctuations introduce significant uncertainty in PV power generation [8], making short-term forecasting a critical yet challenging task for reliable SEV energy planning [9]. It is important to note that the focus of this work is on the accurate forecasting of global irradiance as a foundational input for SEV energy management systems rather than on the direct simulation of onboard photovoltaic generation or vehicle dynamics. In practical SEV applications, short-term irradiance forecasts are commonly obtained from stationary or gridded meteorological sources and subsequently integrated into higher-level decision-making layers for charging control, routing, and energy planning.
Accurate irradiance forecasting requires models capable of capturing
  • Strong temporal variability and seasonality;
  • Rapid short-term fluctuations caused by cloud cover;
  • Multi-scale irradiance dynamics ranging from seconds to hours;
  • Spatial heterogeneity across regions and driving trajectories.
Traditional models often struggle to maintain robustness under such conditions, especially in real-world SEV operating environments. For instance, during fast cloud passage events, solar irradiance may undergo abrupt drops or spikes within a short time interval, leading to highly non-stationary and multi-scale signal behavior. Conventional forecasting models such as CNN–LSTM or single-stage hybrid approaches often struggle in such scenarios, as their predictions tend to smooth rapid transitions or lag behind sudden changes. As a result, short-term forecasting errors increase significantly during these transient regimes, despite acceptable performance under stable clear-sky conditions. This practical limitation highlights the need for forecasting frameworks that can explicitly disentangle multi-scale dynamics and correct residual nonlinear patterns arising from rapidly evolving atmospheric conditions.
Numerical weather prediction (NWP) systems provide reliable long-term forecasts but lack fine spatial resolution and incur high computational cost, making them impractical for embedded SEV applications. Machine learning (ML) models such as support vector regression, decision trees, and multilayer perceptrons improve non-linear learning capability, yet they often fail under abrupt weather transitions and heterogeneous climatic regimes. Recent deep learning (DL) architectures, including CNNs, LSTMs, and Transformers, show strong temporal learning potential but face challenges related to noise sensitivity, large data requirements, and difficulty in representing multi-scale variations. Furthermore, hyperparameter tuning in such models is complex, and classical optimizers like particle swarm optimization (PSO) and genetic algorithms (GAs) may converge prematurely and fail to identify global optima.
Beyond these general limitations, recent studies have highlighted both the strengths and remaining weaknesses of state-of-the-art forecasting paradigms. Evaluations of the ECMWF Integrated Forecasting System (IFS) have shown that, while NWP models can provide physically consistent predictions with reasonable accuracy for global horizontal irradiance (GHI), their performance deteriorates in the presence of complex cloud dynamics and aerosol effects, particularly for short-term and high-resolution applications [10]. To mitigate these deficiencies, data-driven post-processing strategies such as artificial neural network (ANN)-based corrections have been proposed, demonstrating improved accuracy over raw NWP outputs and persistence baselines [11]. In parallel, hybrid machine learning frameworks combining weather classification with ensemble learners, such as CatBoost, have further enhanced forecasting robustness by adapting predictions to distinct atmospheric regimes [12]. Nevertheless, these approaches remain challenged by rapidly changing weather conditions, multi-scale non-stationarity, and the lack of coordinated mechanisms to jointly capture long-range temporal dependencies and structured residual errors.
Recent research has increasingly focused on hybrid and decomposition-enhanced learning frameworks for solar irradiance forecasting. Lopes et al. [10] evaluated ECMWF–IFS forecasts and reported satisfactory accuracy for clear-sky global horizontal irradiance (GHI), while noting significant error increases under cloudy and high-aerosol conditions. ANN-based approaches, such as those used by Yadav et al. [13] and Pereira et al. [11], demonstrated measurable reductions in MAE and RMSE when optimized meteorological inputs and statistical post-processing were employed. Wang et al. [14] conducted a multi-climate comparative study and showed that multilayer perceptron (MLP) and radial basis neural networks (RBNNs) reduced the RMSE by up to 15–20% compared to generalized regression neural networks (GRNNs). Olatomiwa et al. [15] enhanced support vector machine (SVM) performance through Firefly Algorithm optimization, achieving lower MAPE values relative to conventional SVM configurations. Gupta et al. [16] further advanced decomposition-driven modeling by integrating multivariate empirical mode decomposition with PCA–GRU architectures, reporting consistent RMSE reductions (approximately 10–18%) while decreasing computational complexity.
More recently, deep hybrid architectures have demonstrated improved temporal feature extraction capabilities. Ghimire et al. [17] reported that CNN–LSTM hybrids achieved lower RMSE and higher R 2 values than standalone CNN and LSTM models across multiple seasonal datasets. Ahmed et al. [12] showed that combining CatBoost with weather-type classification reduced GHI prediction errors by more than 12% in terms of MAE compared to single-model baselines. Cannizzaro et al. [18] validated variational mode decomposition (VMD)-based CNN hybrids, demonstrating systematic improvements in both short- and long-term forecasting horizons, with RMSE reductions ranging from 8% to 16% relative to non-decomposed deep models.
Despite these advances, challenges persist in fully capturing multi-scale variability, maintaining cross-region generalization, and avoiding local minima during model training.
Although progress has been made with existing hybrid models, as summarized in Table 1, several critical challenges remain unresolved. First, most approaches address multi-scale non-stationarity and temporal dependency modeling in isolation rather than within a unified hierarchical framework [19]. Second, residual errors arising from rapidly changing meteorological conditions are often treated implicitly through ensemble averaging rather than being explicitly learned and corrected [18]. Third, hyperparameter optimization is typically performed independently for each model component, which limits coordinated learning across heterogeneous modules [20]. These challenges are difficult to resolve due to the intrinsic coupling between spectral variability, temporal dynamics, and nonlinear residual structures in solar irradiance signals, particularly under fast cloud transitions and seasonal variability. Addressing these issues motivates the development of a frequency-aware, attention-based, and jointly optimized forecasting framework such as the proposed WTX–CBO model. To address these gaps, effective forecasting solutions must integrate
  • Multi-scale signal decomposition to capture diverse temporal patterns;
  • Attention-based learning to model long-range dependencies;
  • Ensemble refinement to mitigate residual non-linear errors;
  • Advanced optimization strategies to efficiently tune hyperparameters.
Motivated by these requirements, this work proposes a hybrid Wavelet–Transformer–XGBoost architecture optimized by a Chaotic Billiards Optimizer (CBO), enabling accurate, stable, and generalizable solar irradiance forecasting for SEVs.
Accordingly, station- and city-level irradiance forecasting represents a realistic and widely adopted proxy for estimating route-level and onboard solar availability in SEV operational planning, particularly when combined with geographic information, temporal driving patterns, and predictive energy management strategies.

1.2. Main Contributions

The main contributions of this study are summarized as follows:
  • A hierarchical hybrid forecasting framework, termed WTX–CBO, which integrates Discrete Wavelet Transform (DWT), an encoder–decoder Transformer, and XGBoost in a sequential residual learning architecture rather than a parallel or stacked combination.
  • A frequency-aware attention mechanism is introduced by feeding multi-resolution wavelet subbands directly into the Transformer, enabling effective modeling of long-range temporal dependencies under highly non-stationary and rapidly fluctuating solar irradiance conditions.
  • A two-stage temporal learning strategy in which the Transformer captures global and multi-scale dynamics, while XGBoost explicitly learns structured residual errors that persist after attention-based forecasting, improving generalization across diverse climatic regimes.
  • A joint global optimization scheme based on the Chaotic Billiards Optimizer (CBO) is developed to simultaneously tune the hyperparameters of both Transformer and XGBoost components within a unified search space, mitigating premature convergence in high-dimensional hybrid models.
  • The proposed framework is computationally efficient and deployment-oriented, achieving fast convergence and low-latency inference suitable for real-time decision-support in solar-powered and embedded energy systems, including solar electric vehicle applications.
  • Extensive experiments conducted on real-world datasets from multiple climate regions demonstrate that WTX–CBO consistently outperforms state-of-the-art deep learning and hybrid forecasting models in terms of accuracy, robustness, and scalability.
The present paper is structured as follows: Section 2 introduces the proposed methodology. Section 3 provides experimental results and comparative analysis. Finally, Section 4 concludes the work and outlines future research directions.

2. Proposed Methodology

In this study, the target variable is the global horizontal irradiance (GHI), denoted as G ( t ) and expressed in W/m 2 . The proposed WTX–CBO hybrid architecture is designed to enhance the accuracy, robustness, and scalability of short-term irradiance forecasting for solar electric vehicle (SEV) applications, providing reliable decision-support predictions that can subsequently be used as inputs to photovoltaic power models or SEV energy management systems. The model integrates three core modules: (i) the Data Preprocessing and Decomposition Module, (ii) the Forecasting Module, and (iii) the Optimization Module. Each module contributes a distinct functionality within the overall prediction pipeline—ranging from multi-scale feature extraction and attention-based temporal learning to adaptive hyperparameter optimization.
The end-to-end forecasting pipeline illustrated in Figure 1 presents the complete processing chain adopted in this study, beginning from raw meteorological and irradiance data acquisition through model output generation. The workflow clearly highlights the essential preparatory steps, including data cleaning, outlier removal, normalization, and feature selection, prior to time-series windowing. Subsequently, the training phase integrates feature extraction and model learning, while the testing phase evaluates generalization performance using unseen data. This comprehensive flow ensures data integrity, robust model development, and reliable performance assessment for global irradiance prediction.
The architecture of the proposed WTX-CBO hybrid model is shown in Figure 2, emphasizing the computational stages and information flow within the learning framework. Multi-scale signal decomposition using Discrete Wavelet Transform (DWT) generates a structured representation of irradiance fluctuations, which is subsequently processed by a Transformer encoder–decoder to capture long-range temporal dependencies. An XGBoost layer further refines the output by modeling residual error components, while the Chaotic Billiards Optimizer (CBO) dynamically tunes hyperparameters. This hierarchical structure ensures efficient learning of both short-term variability and long-term irradiance dynamics, contributing to superior predictive performance.
Mathematically, the model follows a three-stage workflow:
(i)
Decomposition and Normalization: The input global horizontal irradiance time series G ( t ) is decomposed into multi-scale subcomponents { A N ( t ) , D j ( t ) } using Discrete Wavelet Transform (DWT):
G ( t ) = A N ( t ) + j = 1 N D j ( t ) ,
where A N ( t ) and D j ( t ) denote the approximation (low-frequency) and detail (high-frequency) coefficients at decomposition level j, respectively.
(ii)
Deep Forecasting: The decomposed subseries are embedded and passed through the Transformer encoder–decoder:
G ^ ( t + 1 : t + k ) = F Trans E ( A N ,   D 1 ,   ,   D N )
where F Trans represents the multi-head attention and feed-forward sequence learning function.
(iii)
Refinement and Optimization: The Transformer’s output G ^ I Trans is refined by XGBoost regression:
G ^ I final = f XGB ( G ^ I Trans , X meteo )
while the CBO algorithm adaptively tunes hyperparameters Θ = { η , h , d model , dropout } to minimize the objective:
L = 1 N i = 1 N ( G i G ^ I i ) 2
In the proposed WTX–CBO framework, the Transformer acts as the primary forecasting model, generating an initial prediction of global horizontal irradiance. XGBoost is subsequently employed exclusively as a residual learning module to correct the structured prediction errors produced by the Transformer. Unlike conventional ensemble or stacking approaches, XGBoost is not trained as an independent predictor of irradiance, but rather learns a mapping between the Transformer residuals and their corrected values. The input to the XGBoost model consists of the time-indexed residual sequence between the observed irradiance and the Transformer forecast, optionally augmented by the Transformer’s learned temporal representations, enabling targeted correction of nonlinear residual patterns.
This hybrid formulation ensures that both short-term irradiance dynamics and long-term GI generation patterns are captured efficiently. By combining multi-resolution signal processing, attention-based temporal modeling, and metaheuristic optimization, the proposed WTX–CBO system provides a highly adaptive framework suitable for real-time solar energy forecasting and SEV energy management.

2.1. Datasets Description

Four distinct datasets were created using the NASA Platform. The NASA POWER project aggregates data from various NASA research initiatives, aiming to provide easily accessible information for climate studies [21]. It offers solar and meteorological datasets from NASA research to aid in renewable energy development, enhance building energy efficiency, and support agricultural needs.
These four distinct datasets pertain to the following cities: Zurich, Switzerland [Latitude: 47.35, Longitude: 8.55]; Milan, Italy [Latitude: 45.47, Longitude: 9.18]; Vienna, Austria [Latitude: 48.21, Longitude: 16.37]; and Berlin, Germany [Latitude: 52.52, Longitude: 13.41]. The variables included in these datasets are radiance [W/m2]: the total solar irradiance on a horizontal surface at the Earth’s surface, including direct and diffuse components under all sky conditions; temperature [°C]: the average air temperature 2 m above the Earth’s surface in Celsius; humidity [g/kg]: the ratio of the mass of water vapor to the total mass of air 2 m above the Earth’s surface; precipitation [mm/h]: total atmospheric water vapor in a vertical column of the atmosphere; wind speed [m/s]: the average wind speed 10 m above the Earth’s surface; wind direction [°]: the average wind direction 10 m above the Earth’s surface; frost point [°C]: the dew/frost point temperature 2 m above the Earth’s surface; wet bulb temperature [°C]: the adiabatic saturation temperature measured using a thermometer covered with a water-soaked cloth over which air is passed at 2 m above the Earth’s surface; and surface pressure [kPa]: the average surface pressure at the Earth’s surface.
For each city, hourly meteorological and global horizontal irradiance data were collected over the period from January 2015 to December 2022. The resulting time series were partitioned using a time-ordered split, with the earliest 80% used for model training, the following 10% for validation, and the remaining 10% reserved for out-of-sample testing.

2.2. Data Preprocessing and Decomposition Module

The preprocessing stage ensures that the model receives consistent, denoised, and normalized input signals. The raw GI and meteorological data—typically sampled at Δ t = 1  h—include variables such as global horizontal irradiance (GHI), temperature, relative humidity, and wind speed. The workflow consists of four main steps: data cleaning, normalization, feature selection, and wavelet-based multi-scale decomposition [22].

2.2.1. Data Cleaning and Normalization

Missing or noisy samples are corrected using spline interpolation and a three-point moving average filter. Outliers are removed via the interquartile range (IQR) method to mitigate transient spikes caused by shading or cloud transitions. Each input attribute is rescaled to the interval [ 0 , 1 ] using a min–max transformation:
z = u u low u high u low
where z represents the normalized feature. This normalization improves numerical stability and accelerates the convergence of the learning model [23].

2.2.2. Feature Selection

Both direct and contextual features are retained. Direct variables (global irradiance, air temperature, and wind speed) have an immediate influence on irradiance dynamics, while contextual variables (relative humidity and surface pressure) enhance model robustness across diverse climatic conditions. Pearson correlation and mutual information criteria are employed to identify the most influential predictors, which are subsequently concatenated into the feature vector X ( t ) R d [19].

2.2.3. Wavelet Decomposition

To capture multi-scale temporal behavior, the normalized signals are decomposed using Discrete Wavelet Transform (DWT):
X ( t ) = A N ( t ) + j = 1 N D j ( t )
where A N ( t ) represents the low-frequency trend, and D j ( t ) captures high-frequency fluctuations due to cloud dynamics. The Daubechies (db4) wavelet with decomposition level N = 4 is selected for its strong time–frequency localization and efficiency on meteorological time series [24]. The choice of the Daubechies db4 wavelet and the selected decomposition level is motivated by prior comparative studies on wavelet-based solar irradiance forecasting. Existing works have shown that moderate decomposition depths (typically three to five levels) provide an effective balance between capturing multi-scale temporal dynamics and avoiding excessive fragmentation of the signal, which may amplify noise and increase computational complexity [25]. In particular, the db4 wavelet has been widely adopted due to its compact support and favorable time–frequency localization properties, which are well suited for non-stationary irradiance signals [26]. Based on these findings, the adopted configuration was selected as a robust and computationally efficient compromise for short-term irradiance forecasting.

2.2.4. Input Structuring for Deep Learning

All decomposed sequences are reassembled into an input tensor X R T × d , where T denotes the look-back window. This tensorized input enables multi-channel embedding within the Transformer. The final dataset is partitioned into training, validation, and testing subsets using a strict chronological (walk-forward) strategy, with 80% of the data used for training, 10% for validation, and the remaining 10% reserved for testing [27].
This preprocessing and decomposition module improves the signal-to-noise ratio, preserves the physical interpretability of irradiance patterns, and generates multi-scale inputs that enhance the Transformer’s capacity to model temporal dependencies.

2.3. Forecasting Module

The forecasting module represents the core predictive stage of the proposed WTX–CBO framework. It integrates two complementary components—an encoder–decoder Transformer for long-range temporal learning and an XGBoost ensemble for residual refinement. This dual-stage architecture captures both global temporal dependencies and local nonlinear residual patterns in global horizontal irradiance time series, ensuring enhanced adaptability and forecasting accuracy under varying meteorological conditions [28].

2.3.1. Encoder–Decoder Transformer

The Transformer learns temporal correlations through a self-attention mechanism that processes complete sequences in parallel, avoiding the vanishing-gradient issues common in recurrent networks such as LSTM or BiLSTM. Given the decomposed multivariate input X R T × d , positional encoding is first applied to retain temporal order:
P E ( p o s , 2 i ) = sin p o s 10,000 2 i / d model ,
P E ( p o s , 2 i + 1 ) = cos p o s 10,000 2 i / d model ,
The encoded features are processed by stacked encoder layers composed of multi-head attention and feed-forward sublayers. The attention mechanism computes contextual correlations as
Attention ( Q , K , V ) = softmax Q K d k V ,
where Q, K, and V denote query, key, and value matrices, respectively, and d k is the key dimension. The encoder captures inter-feature dependencies, while the decoder predicts the future GI sequence G ^ I ( t + 1 : t + k ) conditioned on both encoder outputs and prior observations. This structure enables accurate multi-horizon forecasting with minimal information loss [29].

2.3.2. XGBoost Refinement Layer

Although the Transformer effectively models global dynamics, minor prediction residuals may persist due to nonlinear coupling between meteorological inputs [30]. To correct this, an Extreme Gradient Boosting (XGBoost) regression layer refines the Transformer output. XGBoost constructs an ensemble of decision trees minimizing the residual loss:
L ( t ) = i = 1 n l y i , y ^ i ( t 1 ) + f t ( x i ) + Ω ( f t ) ,
where f t is the t-th tree and Ω ( f t ) is a regularization term penalizing model complexity. The hybrid combination
G ^ I final = G ^ I Trans + ϵ XGB ,
with ϵ XGB denoting the learned residual correction improves predictive stability and reduces variance across unseen climatic regimes [28].

2.3.3. Output Reconstruction

The final predicted GI sequence is reconstructed by aggregating the refined wavelet components:
G ^ I final ( t ) = A N * ( t ) + j = 1 N D j * ( t ) ,
where A N * ( t ) and D j * ( t ) denote the reconstructed approximation and detail signals after refinement. This step ensures physical consistency of the forecast and restores the high-frequency fluctuations representing irradiance variability.
By coupling attention-based temporal learning with gradient-boosted residual correction, the forecasting module provides a robust and high-fidelity estimation of photovoltaic output suitable for intelligent energy routing and adaptive control in solar electric vehicles [31].

2.4. Optimization Module

To enhance convergence stability and predictive robustness, the WTX–CBO framework integrates a metaheuristic optimization strategy based on the Chaotic Billiards Optimizer (CBO). This optimizer, as depicted in Algorithm 1, mimics the motion of billiard balls within a chaotic environment, achieving a balanced trade-off between global exploration and local exploitation for efficient hyperparameter tuning [32]. In the proposed framework, CBO operates on a unified hyperparameter vector Θ = { Θ T , Θ X } , where Θ T and Θ X denote the Transformer and XGBoost hyperparameter subsets, respectively.
Algorithm 1 WTX–CBO: Wavelet–Transformer with XGBoost optimized by Chaotic Billiards (CBO)
Require: 
Multivariate time series S (global irradiance G ( t ) and meteorological covariates), lookback window L, forecast horizon H
Ensure: 
Predicted irradiance G ^ ( t + H ) and evaluation metrics (MAE, RMSE, MAPE, R 2 )
  1:
Data preparation
  2:
Clean/align timestamps; impute missing values; optionally resample to uniform Δ t
  3:
Min–Max scale features (fit on Train, reuse parameters for Val/Test)
  4:
Wavelet decomposition (WT)
  5:
for each feature x S  do
  6:
   Apply DWT with wavelet w and level : obtain { A ,   D ,   ,   D 1 }
  7:
   Select subbands and stack ⇒ multi-scale tensor X
  8:
end for
  9:
Sequence building
10:
Build sliding windows of length L: input X t L + 1 : t , target y t + H
11:
Chronological split into Train/Val/Test (80/10/10)
12:
Model components
13:
Transformer encoder hyperparameters Θ T =   { layers, heads, d model , d f f , dropout, learning rate}
14:
XGBoost regressor hyperparameters Θ X =   { n_estimators, max_depth, learning rate, subsample, colsample_bytree}
15:
CBO search space
16:
Encode a candidate as Θ = [ θ T , θ X ] within lower/upper bounds
17:
CBO loop (outer hyperparameter optimization)
18:
Initialize P billiard balls { Θ ( p ) } p = 1 P
19:
for  k = 1  to K do                        ▹ CBO iterations
20:
   for each ball p do
21:
   Instantiate Transformer  ( θ T ( p ) ) ; XGBoost  ( θ X ( p ) )
22:
   Inner training (weights)
23:
    Train Transformer on Train with Adam for E epochs (MSE loss)
24:
    Extract features z (e.g., last hidden state or attention pooling)
25:
    Fit XGBoost on ( z , y ) using Train; validate on Val
26:
   Compute fitness f ( p ) = RMSE Val                  ▹ lower is better
27:
   end for
28:
   Update positions/velocities via chaotic billiards dynamics (collisions, reflections, pockets)
29:
   Enforce bounds; keep elitist best Θ with minimal RMSE Val
30:
end for
31:
Final training
32:
Re-instantiate Transformer  ( θ T ) and XGBoost  ( θ X )
33:
Train Transformer on Train ∪ Val for E epochs; extract z
34:
Fit XGBoost on ( z , y ) using Train ∪ Val
35:
Evaluation
36:
Predict on Test: y ^ t + H
37:
Report MAE, RMSE, MAPE, R 2

2.4.1. Motivation for Metaheuristic Optimization

Deep and ensemble learning architectures rely heavily on sensitive hyperparameters such as the learning rate, embedding dimension, dropout ratio, and attention heads. Conventional optimizers like Adam or RMSProp are effective locally but can become trapped in suboptimal minima. Conversely, heuristic algorithms (e.g., PSO or GA) provide global search capabilities but may converge prematurely. The CBO algorithm mitigates these issues by applying chaotic motion equations that continuously perturb candidate solutions, enabling escape from local minima and faster convergence to global optima [20].

2.4.2. Mathematical Formulation

Each billiard agent represents a candidate solution X i = [ x i 1 ,   x i 2 ,   ,   x i d ] corresponding to a vector of model hyperparameters. Within a bounded search domain Ω R d , its motion follows:
X i t + 1 = X i t + μ v i t + λ ϕ ( X i t ) ,
where μ is the inertia factor, λ is a chaotic scaling coefficient, and ϕ ( · ) is a nonlinear chaotic map derived from billiard dynamics. Velocity updates are computed as
v i t + 1 = ω v i t + r 1 ( P i X i t ) + r 2 ( G X i t ) ,
where P i and G denote personal and global best positions, ω is a damping coefficient, and r 1 and r 2 are random chaotic factors in ( 0 , 1 ) . The optimization objective is defined as
min Θ L ( Θ ) = 1 N i = 1 N ( G i G ^ I i ) 2 + α Θ 2 2 ,
where Θ contains both trainable and meta parameters, and α is a regularization term controlling model complexity [20].

2.4.3. Integration with the Hybrid Framework

CBO operates jointly with the Adam optimizer: CBO performs global hyperparameter search, while Adam fine-tunes weights locally through gradient descent. At each iteration, updated hyperparameter sets Θ * are passed to the Transformer–XGBoost network and refined via backpropagation. The process terminates when the relative improvement of L ( Θ ) remains below ε for K consecutive iterations.
The CBO-driven optimization process achieves faster convergence, lower error variance, and superior adaptability under dynamic irradiance conditions. This hybrid optimization strategy ensures that the WTX–CBO model attains both computational efficiency and high forecasting precision for real-time solar energy management in SEVs.

2.5. Metrics and Hyperparameter Optimization

2.5.1. Evaluation Metrics

To rigorously assess the predictive capability of the proposed forecasting framework, a comprehensive suite of statistical indicators is employed. These metrics quantify deviation, error magnitude, variability explanation, and relative prediction accuracy. Let G i and G ^ i denote the actual and predicted irradiance values, respectively, and let N represent the number of testing samples.
Mean Absolute Error (MAE) [22]
MAE = 1 N i = 1 N G i G ^ i
MAE measures the average magnitude of absolute forecasting errors, reflecting the typical deviation without emphasizing directionality.
Mean Squared Error (MSE) [22]
MSE = 1 N i = 1 N G i G ^ i 2
MSE penalizes larger errors more heavily, making it a useful criterion for evaluating complex forecasting models.
Root Mean Squared Error (RMSE) [22]
RMSE = 1 N i = 1 N G i G ^ i 2
RMSE provides error magnitude in the same physical units as the true values, contributing to practical interpretability.
Coefficient of Determination ( R 2 ) [22]
R 2 = 1 i = 1 N G i G ^ i 2 i = 1 N G i G ¯ 2
where G ¯ denotes the mean of actual observations. R 2 quantifies the proportion of variance in the target series explained by the model.
Mean Absolute Percentage Error (MAPE) [22]
MAPE = 100 N i = 1 N G i G ^ i G i
MAPE expresses the average relative error as a percentage, enabling cross-dataset accuracy comparisons.

2.5.2. Hyperparameter Optimization

A two-phase optimization procedure was adopted to balance predictive accuracy and computational efficiency.
Stage 1: Empirical Grid Search
An initial coarse grid search was conducted to identify stable baseline configurations and feasible parameter ranges for both the Transformer and XGBoost components rather than final optimal values. The optimal settings are summarized in Table 2 and Table 3.
Stage 2: Fine-Tuning via CBO
Starting from the baseline hyperparameter configurations identified through the grid-search stage (Stage 1), the CBO performs joint fine-tuning of the most influential hyperparameters within the predefined search bounds. During the second optimization stage, the CBO performs joint fine-tuning of the most influential hyperparameters identified during the grid-search phase. For the Transformer component, the optimized parameters include the learning rate, model dimension d model , number of attention heads, number of encoder layers, feed-forward network dimension, and dropout rate. For the XGBoost regressor, CBO tunes the number of trees, maximum tree depth, learning rate, subsampling ratio, and column sampling ratio. This two-stage strategy combines the stability of grid search with the global exploration capability of CBO. The CBO refined the selected hyperparameters by minimizing the validation RMSE. The optimizer dynamically explores the parameter space using chaotic motion equations to avoid premature convergence. The adopted configuration is reported in Table 4.

3. Experimental Results and Analysis

3.1. Computational Environment

All experiments and performance evaluations were conducted on a single, standardized research workstation to ensure fair and reproducible comparison across all models. The system is equipped with an Intel Core i9–12900K processor, 64 GB of DDR4 memory, and an NVIDIA RTX 4090 GPU. The software environment consists of a 64-bit Ubuntu Linux operating system, CUDA and cuDNN libraries for GPU acceleration, and Python-based deep learning frameworks, including TensorFlow and Keras. All baseline models and the proposed WTX–CBO framework were trained and evaluated under identical hardware and software conditions. Training time, convergence behavior, inference latency, and memory usage were measured using the same implementation settings to guarantee a fair and consistent assessment of computational efficiency.

3.2. Evaluation of the Proposed Model

The proposed WTX–CBO hybrid architecture was experimentally evaluated to assess its predictive accuracy, robustness, and scalability in solar irradiance forecasting for solar electric vehicles (SEVs). A series of comparative experiments was conducted to examine the influence of each module—Wavelet decomposition, Transformer-based temporal modeling, XGBoost refinement, and Chaotic Billiards optimization—under different climatic scenarios.
The effectiveness of the proposed approach was also benchmarked against state-of-the-art DL and hybrid models.

3.3. Decomposition Performance

The Discrete Wavelet Transform (DWT) effectively captured multi-scale temporal dependencies within irradiance data. By decomposing input sequences into approximation and detail coefficients, the DWT improved the model’s interpretability and reduced the effect of noise. This step enhanced the Transformer’s ability to learn both short-term fluctuations and long-term seasonal patterns. Quantitative evaluation revealed that DWT-based preprocessing reduced the RMSE by approximately 17% and increased the coefficient of determination ( R 2 ) by about 12% compared to models trained on raw irradiance data, confirming its efficiency in multi-resolution feature extraction.
Table 5 demonstrates that applying signal decomposition markedly enhances the forecasting precision of the model. While raw data produced relatively high errors (MAE = 0.0286 ; RMSE = 0.0391 ), multi-resolution decomposition methods such as EMD and VMD reduced the prediction variance by isolating intrinsic oscillatory modes. However, Wavelet Transform (WT) achieved the most balanced trade-off between temporal localization and spectral discrimination, yielding the lowest MAE and MAPE with acceptable computational overhead. This confirms that wavelet decomposition provides the most suitable representation for non-stationary irradiance signals, supporting the motivation behind the WTX–CBO design, as depicted in Figure 3.

3.4. Forecasting Accuracy

The forecasting capability of the proposed WTX–CBO framework was compared with several baseline models, including LSTM, BiLSTM, CNN–LSTM, Transformer, and XGBoost. Across all four test regions (Switzerland, Italy, Austria, and Germany), the model achieved superior accuracy, with mean values of
MAE < 0.020 , RMSE < 0.032 , MAPE 1.8 % , R 2 > 0.98 .
The results highlighted in Figure 4 confirm that integrating wavelet-based decomposition with Transformer attention enhances temporal generalization, while XGBoost further refines residual errors, leading to smoother irradiance curves and improved robustness under highly variable sky conditions. The quantitative comparison is reported in Table 6.
As shown in Table 7, the proposed WTX–CBO significantly outperforms established hybrid forecasting models. Conventional architectures such as CNN–LSTM and EWT–BiLSTM provide notable accuracy gains but remain limited in handling multi-scale temporal dependencies. In contrast, the wavelet–Transformer–XGBoost integration exploits both temporal hierarchy and nonlinear regression strength, achieving an MAE of 0.0189 and R2 of 0.987. This consistent superiority across metrics confirms the generalization capability and adaptability of the proposed approach under varying meteorological regimes.

3.5. Optimization Efficiency

The Chaotic Billiards Optimizer (CBO) demonstrated faster and more stable convergence than traditional metaheuristics such as the Genetic Algorithm (GA), Particle Swarm Optimization (PSO), and Backtracking Search Algorithm (BSA). When coupled with the Adam optimizer, the hybrid CBO–Adam scheme reduced the training time by approximately 25% and improved validation R 2 by 2.3%. The chaotic initialization strategy expanded the search space, reduced the likelihood of local minima, and accelerated convergence in complex loss landscapes.
The optimizer comparison in Table 8 highlights the critical role of hyperparameter adaptation in achieving convergence efficiency. Classical gradient-based methods such as Adam exhibited stable yet slower convergence, while evolutionary optimizers (GA and PSO) improved performance through global search. The proposed Chaotic Billiards Optimizer (CBO) achieved the highest accuracy (R = 2 0.987 ) and required the fewest epochs, demonstrating superior exploration–exploitation balance and robustness against local minima. These results validate CBO as an effective metaheuristic for fine-tuning hybrid DL models in dynamic solar forecasting contexts.

3.6. Ablation and Component Analysis

An ablation analysis was carried out to assess the contribution of each component within the hybrid framework. Removing the Wavelet decomposition increased the RMSE by nearly 18%, while excluding XGBoost refinement introduced significant bias in long-horizon forecasts. The absence of the CBO optimization module resulted in slower and unstable convergence. These findings confirm that the synergy between all submodules is crucial to achieving high predictive performance and generalization across climatic zones.
Table 9 presents the ablation study assessing the contribution of each module to the overall performance. Removing any key component led to measurable degradation in accuracy, particularly when excluding wavelet decomposition or the Transformer encoder. The full model achieved the best balance across all indicators, underscoring the synergistic integration of spectral decomposition, attention-based sequence modeling, nonlinear regression, and chaotic optimization, as depicted in Figure 5. This analysis substantiates that each module in the WTX–CBO pipeline serves a distinct yet complementary purpose.

3.7. Computational Scalability

According to Table 10, the proposed architecture exhibits the fastest convergence rate and the lowest training cost among the tested models. The Transformer and XGBoost components enable efficient gradient propagation and stable optimization, while the CBO metaheuristic accelerates convergence toward the global optimum. The validation R2 of 0.987 after only 78 epochs highlights the model’s computational efficiency and its suitability for scalable training in large-scale GI forecasting pipelines. Despite its hybrid design, the proposed WTX–CBO model maintained high computational efficiency when evaluated on the standardized experimental platform described in Section 3.1, achieving low inference latency and moderate memory consumption suitable for real-time deployment. Table 11 summarizes these findings.
Table 12 evaluates real-time inference performance. The proposed WTX–CBO achieves the shortest latency (0.038 s) while maintaining moderate memory consumption, outperforming deep recurrent models that suffer from sequential dependencies. These results emphasize the model’s feasibility for deployment on embedded systems and onboard processors in solar electric vehicles, where low latency and high responsiveness are essential.

3.8. Statistical Validation

To assess the statistical significance of the performance differences between the proposed WTX–CBO framework and the benchmark models, both the Wilcoxon signed-rank test and the Diebold–Mariano (DM) test were employed. For each model comparison, the statistical tests were conducted on paired, time-indexed forecast error series computed on the test dataset. Specifically, the samples consist of hourly one-step-ahead prediction errors over the full test period, yielding paired error sequences of equal length for the proposed model and each competing baseline. No daily aggregation or multiple training runs were considered, ensuring that the tests directly evaluate point-wise predictive accuracy under identical temporal conditions.
The Wilcoxon signed-rank test was applied to paired absolute error sequences to evaluate whether the median difference in forecasting errors between WTX–CBO and each baseline model is statistically significant, without assuming normality. A significance level of α = 0.05 was adopted. The Diebold–Mariano test was conducted using the squared error loss function for one-step-ahead (hourly) forecasts. The null hypothesis assumes equal predictive accuracy between the two competing models, while the alternative hypothesis indicates superior forecasting performance of the proposed approach. Identical forecast horizons, loss functions, and test windows were used for all model comparisons to ensure fairness and reproducibility. The obtained p-values (<0.01) indicate that the observed performance improvements achieved by WTX–CBO are statistically significant. Table 13 summarizes the Wilcoxon signed-rank test results, while Table 14 reports the corresponding Diebold–Mariano statistics.
Table 14 confirms that the forecasting improvements delivered by WTX–CBO are statistically significant. All DM statistics exceed the critical threshold with p-values below 0.05, indicating that the error differentials between WTX–CBO and competing models are not due to random variance. Hence, the proposed model achieves a statistically robust enhancement in predictive performance, reinforcing the validity of the observed improvements across datasets.

3.9. Visualization and Interpretation

Figure 6 illustrates the close alignment between predicted and measured irradiance values, showing the model’s ability to accurately follow rapid irradiance transitions while preserving daily cyclic trends. Residual error plots exhibit minimal dispersion around zero, and correlation plots demonstrate near-unity slopes, confirming strong predictive agreement.
The feature ranking summarized in Table 15 reveals that global horizontal irradiance (GHI), air temperature, and relative humidity contribute most significantly to prediction accuracy, followed by wind-related variables. This hierarchy aligns with established solar-energy physics, confirming that the model learns physically interpretable relationships rather than spurious correlations. The XGBoost-based feature attribution thus enhances transparency and provides actionable insights for data collection prioritization in future deployments.

3.10. Physical Interpretation and Error Regime Analysis

While aggregated performance metrics provide a quantitative assessment of forecasting accuracy, it is equally important to interpret the results in light of the underlying physical mechanisms governing solar irradiance variability. The proposed WTX–CBO framework demonstrates particularly strong performance under stable clear-sky conditions and slowly varying irradiance regimes, where wavelet decomposition effectively isolates low-frequency trends and the Transformer captures long-range temporal dependencies. Under highly variable meteorological conditions, such as fast-moving or broken cloud fields, short-term irradiance exhibits abrupt high-frequency fluctuations that challenge deterministic forecasting models. In these regimes, the wavelet-based decomposition enables localization of transient variations, while the XGBoost refinement stage mitigates nonlinear residual patterns that persist after attention-based forecasting. Nevertheless, moderate error increases are observed during sudden cloud transitions, reflecting the intrinsic unpredictability of rapidly evolving atmospheric processes rather than the instability of the proposed model.
Forecasting performance is also affected during low-sun-angle periods, particularly in winter months, when reduced irradiance magnitude and increased atmospheric path length amplify relative noise sensitivity. Despite these challenges, WTX–CBO maintains stable behavior across seasons, benefiting from multi-scale feature extraction and joint hyperparameter optimization. From a diurnal perspective, forecasting errors tend to be the lowest during midday hours, when irradiance signals are smoother and solar elevation is the highest. Error magnitudes increase during early morning and late afternoon periods due to rapid changes in solar geometry and enhanced atmospheric scattering. Compared with simpler machine learning baselines such as LSTM or standalone XGBoost, WTX–CBO exhibits a smoother diurnal error profile, with reduced error growth during sunrise and sunset transitions. This behavior highlights the advantage of combining frequency-aware decomposition with attention-based temporal modeling.
Relative to numerical weather prediction (NWP) forecasts, which may suffer from coarse spatial resolution and delayed cloud-field updates, WTX–CBO provides more responsive short-term predictions, particularly under partially cloudy conditions. This responsiveness results in lower point-wise errors for short-term horizons relevant to real-time energy management. From the perspective of solar electric vehicle (SEV) applications, these findings suggest that WTX–CBO forecasts are well suited for short-term energy planning and predictive charging under stable irradiance conditions. During highly volatile regimes dominated by fast cloud passages, the smoother error evolution of the proposed framework supports conservative operational strategies, such as adaptive charging schedules, reduced reliance on anticipated solar gains, or precautionary energy buffering. These insights reinforce the role of WTX–CBO as a decision-support tool that enables SEV controllers to dynamically balance efficiency and reliability under varying atmospheric conditions.
Overall, the identified limitations are primarily driven by intrinsic atmospheric variability rather than deficiencies in the proposed architecture. They motivate future extensions toward uncertainty-aware and probabilistic forecasting approaches to further enhance robustness under highly volatile sky conditions.

3.11. Summary

In summary, the proposed WTX–CBO hybrid framework achieves superior forecasting accuracy, fast convergence, and high generalization across diverse climatic datasets. The integration of wavelet-based decomposition, attention-driven Transformer modeling, ensemble refinement via XGBoost, and chaotic optimization through CBO provides a scalable and physics-informed architecture for intelligent photovoltaic management in next-generation solar electric vehicles.

Implications for Solar Electric Vehicle Applications

In the context of SEVs, the proposed WTX–CBO forecasting framework can be integrated as a decision-support layer within intelligent energy management systems. Short-term irradiance forecasts enable predictive charging strategies, conservative energy planning under highly volatile meteorological conditions, and route-aware optimization when combined with navigation and traffic information. For example, during periods of rapid cloud transitions or low-sun conditions, the forecast outputs can be used to adapt charging schedules or limit energy-intensive operations. While the present study focuses on irradiance prediction using stationary meteorological data, the framework is readily extensible to onboard or trajectory-based implementations through coupling with vehicle dynamics, photovoltaic yield models, and battery state-of-health estimation, which will be explored in future work.

4. Conclusions and Future Work

This study introduced a new hybrid DL framework, referred to as WTX–CBO, designed to provide accurate and reliable short-term solar irradiance forecasts for solar electric vehicles (SEVs). The architecture combines Wavelet Transform (WT) for multi-scale signal decomposition, an encoder–decoder Transformer for long-range temporal learning, and XGBoost for ensemble refinement, all optimized through the Chaotic Billiards Optimizer (CBO) in conjunction with Adam.
Extensive experimental results across multiple climatic regions demonstrated that the proposed model consistently outperformed conventional architectures such as LSTM, CNN–LSTM, BiLSTM, Transformer, and standalone XGBoost. The integration of wavelet-based decomposition and chaotic optimization significantly enhanced convergence speed, reduced error metrics (MAE, RMSE, and MAPE), and achieved an average R 2 exceeding 0.98 across all tested datasets. Furthermore, the model exhibited strong cross-regional generalization, maintaining high accuracy under non-stationary meteorological conditions.
The hybrid nature of the WTX–CBO framework makes it particularly suitable for embedded SEV applications requiring real-time decision support. The combination of physical interpretability and data-driven learning offers a scalable solution for adaptive energy management, smart routing, and predictive control of photovoltaic subsystems.

Future Work

Despite its promising results, several directions can further extend this research:
  • Integration with vehicle dynamics: Future versions will couple irradiance prediction with battery state of health (SOH) and vehicle motion models to enable holistic SEV energy optimization.
  • Uncertainty quantification: Incorporating probabilistic inference or Monte Carlo dropout could provide confidence intervals for the predicted irradiance, enhancing reliability under abrupt weather transitions.
  • Edge deployment and real-time adaptation: Model compression, pruning, and knowledge distillation will be explored to deploy WTX–CBO on low-power embedded systems within SEV control units.
  • Multi-modal data fusion: Combining satellite imagery, sky cameras, and ground-based meteorological data can further improve spatial generalization and context awareness.
  • Meta-optimization: Extending the Chaotic Billiards Optimizer toward multi-objective optimization will allow simultaneous tuning for accuracy, complexity, and energy efficiency.
In summary, the proposed WTX–CBO framework establishes a robust and scalable foundation for intelligent solar energy forecasting, with significant potential for integration into real-world smart mobility and sustainable transportation systems.

Author Contributions

Conceptualization, W.M. and H.A.; methodology, W.M., G.C., L.M. and H.A.; software, W.M., L.M., G.C. and H.A.; validation, W.M., G.C., M.R. and H.A.; formal analysis, W.M., G.C., M.R. and L.M.; investigation, W.M., G.C., M.R. and L.M.; resources, W.M. and G.C.; writing—original draft preparation, W.M., G.C. and H.A.; writing—review and editing, W.M., G.C., L.M., M.R. and H.A.; supervision, W.M., G.C. and H.A.; funding acquisition, G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets analyzed during the current study are publicly available and derived from recognized sources. Specifically, the NASA POWER datasets https://power.larc.nasa.gov (accessed on 12 December 2024) provide global solar and meteorological parameters for the locations analyzed. Processed data and the source code used to implement the proposed WTX–CBO model are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Nomenclature

Abbreviations
AIArtificial Intelligence
AMAttention Mechanism
ANNArtificial Neural Network
BSABacktracking Search Algorithm
BiLSTMBidirectional Long Short-Term Memory
CBOChaotic Billiards Optimizer
CNNConvolutional Neural Network
CWSCNN–Wavelet–SVR hybrid model
DWTDiscrete Wavelet Transform
DLDeep Learning
DM TestDiebold–Mariano Test
DWTDiscrete Wavelet Transform
EVElectric Vehicle
FCMFuzzy C-Means
GAGenetic Algorithm
GHIGlobal Horizontal Irradiance
GRUGated Recurrent Unit
KNNK-Nearest Neighbors (if appears)
LSTMLong Short-Term Memory
MAEMean Absolute Error
MABEMean Absolute Bias Error
MAPEMean Absolute Percentage Error
MBEMean Bias Error
MLPNNMultilayer Perceptron Neural Network
MR-ESNMulti-Reservoir Echo State Network
NRELNational Renewable Energy Laboratory
NSRDBNational Solar Radiation Database
PVPhotovoltaic
PVGISPhotovoltaic Geographical Information System
PSOParticle Swarm Optimization
rRMSERelative Root Mean Square Error
RMSERoot Mean Square Error
R2Coefficient of Determination
SEVSolar Electric Vehicle
SVM/SVRSupport Vector Machine/Support Vector Regression
SoCState of Charge (Battery)
V2GVehicle-to-Grid
VMDVariational Mode Decomposition
WNNWavelet Neural Network
WTWavelet Transform
WTX–CBOWavelet Transformer XGBoost optimized by CBO
XGB/XGBoostExtreme Gradient Boosting
Symbols
G ( t ) Global horizontal irradiance at time t ( W/m 2 )
G ^ ( t ) Predicted global horizontal irradiance
A N ( t ) Approximation coefficients of irradiance (wavelet low-frequency component)
D j ( t ) Detail coefficients of irradiance at level j (high-frequency components)
X ( t ) Input feature vector
TLook-back window size
HForecast horizon
Θ Set of trainable and meta-parameters
LLoss function
ω Damping coefficient of the Chaotic Billiards Optimizer (CBO)
r 1 , r 2 Chaotic random factors in CBO

References

  1. Oluwalana, O.J.; Grzesik, K. Solar-powered electric vehicles: Comprehensive review of technology advancements, challenges, and future prospects. Energies 2025, 18, 3650. [Google Scholar] [CrossRef]
  2. Li, T.T.; Zhao, A.P.; Wang, Y.; Li, S.; Fei, J.; Wang, Z.; Xiang, Y. Integrating solar-powered electric vehicles into sustainable energy systems. Nat. Rev. Electr. Eng. 2025, 2, 467–479. [Google Scholar] [CrossRef]
  3. Colucci, R.; Mahgoub, I. Generalizable Solar Irradiance Prediction for Battery Operation Optimization in IoT-Based Microgrid Environments. J. Sens. Actuator Netw. 2025, 14, 3. [Google Scholar] [CrossRef]
  4. Akinyoola, J.A.; Oluleye, A.; Gbode, I.E. A review of atmospheric aerosol impacts on regional extreme weather and climate events. Aerosol Sci. Eng. 2024, 8, 249–274. [Google Scholar] [CrossRef]
  5. Isman Okieh, O.; Seker, S.; Gokce, S.; Dennenmoser, M. An Enhanced Forecasting Method of Daily Solar Irradiance in Southwestern France: A Hybrid Nonlinear Autoregressive with Exogenous Inputs with Long Short-Term Memory Approach. Energies 2024, 17, 3965. [Google Scholar] [CrossRef]
  6. Zhou, B.; Chen, X.; Li, G.; Gu, P.; Huang, J.; Yang, B. XGBoost–SFS and Double Nested Stacking Ensemble Model for Photovoltaic Power Forecasting under Variable Weather Conditions. Sustainability 2023, 15, 13146. [Google Scholar] [CrossRef]
  7. Sansine, V.; Ortega, P.; Hissel, D.; Ferrucci, F. Hybrid Deep Learning Model for Mean Hourly Irradiance Probabilistic Forecasting. Atmosphere 2023, 14, 1192. [Google Scholar] [CrossRef]
  8. Al-Ali, E.M.; Hajji, Y.; Said, Y.; Hleili, M.; Alanzi, A.M.; Laatar, A.H.; Atri, M. Solar Energy Production Forecasting Based on a Hybrid CNN-LSTM-Transformer Model. Mathematics 2023, 11, 676. [Google Scholar] [CrossRef]
  9. Tercha, W.; Tadjer, S.A.; Chekired, F.; Canale, L. Machine Learning-Based Forecasting of Temperature and Solar Irradiance for Photovoltaic Systems. Energies 2024, 17, 1124. [Google Scholar] [CrossRef]
  10. Lopes, F.M.; Silva, H.G.; Salgado, R.; Cavaco, A.; Canhoto, P.; Collares-Pereira, M. Short-term forecasts of GHI and DNI for solar energy systems operation: Assessment of the ECMWF integrated forecasting system in southern Portugal. Sol. Energy 2018, 170, 14–30. [Google Scholar] [CrossRef]
  11. Pereira, S.; Canhoto, P.; Salgado, R.; Costa, M.J. Development of an ANN based corrective algorithm of the operational ECMWF global horizontal irradiation forecasts. Sol. Energy 2019, 185, 387–405. [Google Scholar] [CrossRef]
  12. Ahmed, U.; Khan, A.R.; Mahmood, A.; Rafiq, I.; Ghannam, R.; Zoha, A. Short-term global horizontal irradiance forecasting using weather classified categorical boosting. Appl. Soft Comput. 2024, 155, 111441. [Google Scholar] [CrossRef]
  13. Yadav, A.K.; Malik, H.; Chandel, S. Selection of most relevant input parameters using WEKA for artificial neural network based solar radiation prediction models. Renew. Sustain. Energy Rev. 2014, 31, 509–519. [Google Scholar] [CrossRef]
  14. Wang, L.; Kisi, O.; Zounemat-Kermani, M.; Salazar, G.A.; Zhu, Z.; Gong, W. Solar radiation prediction using different techniques: Model evaluation and comparison. Renew. Sustain. Energy Rev. 2016, 61, 384–397. [Google Scholar] [CrossRef]
  15. Olatomiwa, L.; Mekhilef, S.; Shamshirband, S.; Mohammadi, K.; Petković, D.; Sudheer, C. A support vector machine–firefly algorithm-based model for global solar radiation prediction. Sol. Energy 2015, 115, 632–644. [Google Scholar] [CrossRef]
  16. Gupta, P.; Singh, R. Combining simple and less time complex ML models with multivariate empirical mode decomposition to obtain accurate GHI forecast. Energy 2023, 263, 125844. [Google Scholar] [CrossRef]
  17. Ghimire, S.; Deo, R.C.; Raj, N.; Mi, J. Deep solar radiation forecasting with convolutional neural network and long short-term memory network algorithms. Appl. Energy 2019, 253, 113541. [Google Scholar] [CrossRef]
  18. Cannizzaro, D.; Aliberti, A.; Bottaccioli, L.; Macii, E.; Acquaviva, A.; Patti, E. Solar radiation forecasting based on convolutional neural network and ensemble learning. Expert Syst. Appl. 2021, 181, 115167. [Google Scholar] [CrossRef]
  19. Mchara, W.; Khalfa, M.A.; Manai, L. Advanced health state intelligent diagnosis of lithium-ion batteries based on CNN-WNN-WBiLSTM model with attention mechanism. Automatika 2025, 66, 154–173. [Google Scholar] [CrossRef]
  20. Tomar, V.; Bansal, M.; Singh, P. Metaheuristic algorithms for optimization: A brief review. Eng. Proc. 2024, 59, 238. [Google Scholar]
  21. NASA Power. Prediction of Worldwide Energy Resource (POWER) Project. 2019. Available online: https://power.larc.nasa.gov (accessed on 12 December 2024).
  22. Mchara, W.; Manai, L.; Khalfa, M.A.; Raissi, M. Intelligent Health State Diagnosis of Lithium-Ion Batteries for Electric Vehicles Using Wavelet-Enhanced Hybrid Deep Learning Integrated with an Attention Mechanism. Clean Energy 2025, 6, zkaf019. [Google Scholar] [CrossRef]
  23. Khalfa, M.A.; Manai, L.; Mchara, W. Advanced artificial intelligence model for solar irradiance forecasting for solar electric vehicles: MA Khalfa et al. Int. J. Dyn. Control 2025, 13, 101. [Google Scholar] [CrossRef]
  24. Ondo Ekogha, E.; Owolawi, P.A. Comparative Analysis of Supervised Learning Techniques for Forecasting PV Current in South Africa. Forecasting 2024, 7, 1. [Google Scholar] [CrossRef]
  25. Albayram, M.; Yılmaz, A.; Bayrak, G.; Basaran, K.; Georgeta Popescu, L. Effectiveness of un-decimated wavelet transform in time-series forecasting: A PV power calculation case study in BTU. Renew. Energy 2026, 256, 124062. [Google Scholar] [CrossRef]
  26. Berrezzek, F.; Khelil, K.; Bouadjila, T. Efficient wind speed forecasting using discrete wavelet transform and artificial neural networks. Rev. Intell. Artif. 2019, 33, 447–452. [Google Scholar] [CrossRef]
  27. Niccolai, A.; Orooji, S.; Matteri, A.; Ogliari, E.; Leva, S. Irradiance nowcasting by means of deep-learning analysis of infrared images. Forecasting 2022, 4, 338–348. [Google Scholar] [CrossRef]
  28. Legrene, I.; Wong, T.; Dessaint, L.A. Enhancing neural architecture search using transfer learning and dynamic search spaces for global horizontal irradiance prediction. Forecasting 2025, 7, 43. [Google Scholar] [CrossRef]
  29. Ndong, J.; Soubdhan, T. Extracting Statistical Properties of Solar and Photovoltaic Power Production for the Scope of Building a Sophisticated Forecasting Framework. Forecasting 2022, 5, 1–21. [Google Scholar] [CrossRef]
  30. Lee, J.H.; Okwuosa, C.N.; Shin, B.C.; Hur, J.W. A Spectral-Based Blade Fault Detection in Shot Blast Machines with XGBoost and Feature Importance. J. Sens. Actuator Netw. 2024, 13, 64. [Google Scholar] [CrossRef]
  31. Benavides-Cesar, L.; Manso-Callejo, M.Á.; Cira, C.I. Methodology Based on BERT (Bidirectional Encoder Representations from Transformers) to Improve Solar Irradiance Prediction of Deep Learning Models Trained with Time Series of Spatiotemporal Meteorological Information. Forecasting 2025, 7, 5. [Google Scholar] [CrossRef]
  32. Elnaghi, B.E.; Abelwhab, M.; Ismaiel, A.M.; Mohammed, R.H. Solar hydrogen variable speed control of induction motor based on chaotic billiards optimization technique. Energies 2023, 16, 1110. [Google Scholar] [CrossRef]
Figure 1. Overall workflow of the proposed WTX–CBO solar irradiance forecasting framework.
Figure 1. Overall workflow of the proposed WTX–CBO solar irradiance forecasting framework.
Jsan 15 00012 g001
Figure 2. Model block diagram of the proposed WTX–CBO framework: DWT multi-scale features feed a Transformer; predictions are refined by XGBoost; CBO tunes θ T and θ X for optimal performance.
Figure 2. Model block diagram of the proposed WTX–CBO framework: DWT multi-scale features feed a Transformer; predictions are refined by XGBoost; CBO tunes θ T and θ X for optimal performance.
Jsan 15 00012 g002
Figure 3. Performance comparison of decomposition strategies.
Figure 3. Performance comparison of decomposition strategies.
Jsan 15 00012 g003
Figure 4. Comparative performance of the proposed WTX–CBO model against benchmark models across multiple datasets.
Figure 4. Comparative performance of the proposed WTX–CBO model against benchmark models across multiple datasets.
Jsan 15 00012 g004
Figure 5. Training efficiency and convergence analysis.
Figure 5. Training efficiency and convergence analysis.
Jsan 15 00012 g005
Figure 6. Comparison of actual vs. predicted global irradiance for (a) Switzerland, (b) Italy, (c) Austria, and (d) Germany.
Figure 6. Comparison of actual vs. predicted global irradiance for (a) Switzerland, (b) Italy, (c) Austria, and (d) Germany.
Jsan 15 00012 g006
Table 1. Summary of representative hybrid models for solar irradiance forecasting.
Table 1. Summary of representative hybrid models for solar irradiance forecasting.
ModelMain AdvantagesMain Limitations
CNN–LSTMCaptures local feature patterns and short-term temporal dependencies; widely adopted and computationally efficientLimited ability to model long-range dependencies; prone to smoothing effects during abrupt irradiance transitions
VMD–CNNEffective decomposition of non-stationary signals; improved noise robustnessLacks explicit temporal dependency modeling; performance sensitive to decomposition parameters
WT–LSTMMulti-scale signal representation; improved handling of non-stationaritySequential modeling may struggle with long-term dependencies and residual nonlinear errors
Transformer-based modelsStrong capability for long-range temporal dependency modelingPerformance degrades under high-frequency fluctuations without explicit multi-scale preprocessing
Hybrid DL + ensemble modelsImproved generalization through error correctionOften rely on parallel stacking without explicit residual hierarchy or coordinated optimization
Table 2. Baseline Transformer hyperparameters obtained via grid search.
Table 2. Baseline Transformer hyperparameters obtained via grid search.
ParameterTested RangeSelectedRole/Impact on Performance
Learning rate{ 1 × 10 4 , 5 × 10 4 , 1 × 10 3 } 5 × 10 4 Controls convergence speed and training stability
Dropout rate{0.1, 0.2, 0.3}0.2Reduces overfitting and
improves generalization
Feed-forward dimension{128, 256, 512}256Determines nonlinear feature
transformation capacity
Model dimension d model {64, 128, 256}128Influences representation richness
and temporal modeling ability
Attention heads{2, 4, 8}4Enables multi-head attention
to capture diverse temporal patterns
Encoder layers{2, 3, 4}4Controls model depth and
long-range dependency learning
Table 3. Baseline XGBoost hyperparameters obtained via grid search.
Table 3. Baseline XGBoost hyperparameters obtained via grid search.
ParameterTested RangeSelectedRole/Impact on Performance
Subsample ratio{0.6, 0.8, 1.0}0.8Introduces stochasticity to improve
generalization and reduce overfitting
Learning rate{0.01, 0.05, 0.1}0.05Controls the contribution
of each tree and convergence behavior
Maximum depth{3, 5, 7}5Regulates model complexity
and bias–variance trade-off
Number of trees{100, 200, 300}200Determines ensemble capacity
and residual learning ability
Table 4. CBO configuration parameters.
Table 4. CBO configuration parameters.
ParameterValue
Cue movement coefficient ( ω )0.9
Acceleration coefficient (a)1.0
Maximum iterations100
Population size30
Termination criterionConvergence or max iterations
Table 5. Performance comparison of decomposition strategies.
Table 5. Performance comparison of decomposition strategies.
Decomposition MethodMAERMSEMAPE (%)R2Time (s)Observation
No Decomposition (Raw)0.02860.03912.840.973118Baseline without multi-scale learning
EMD-Based Hybrid0.02380.03492.410.978142Captures non-stationary features
VMD-Based Hybrid0.02120.03182.100.981166Smooth frequency separation
WT-Based Hybrid (Proposed)0.01890.02831.870.987132Superior time–frequency localization
Table 6. Comparative performance of the proposed WTX–CBO model against benchmark models across multiple datasets.
Table 6. Comparative performance of the proposed WTX–CBO model against benchmark models across multiple datasets.
DatasetModelMAEMSERMSEMAPE (%)R2
SwitzerlandLSTM0.03410.00210.04594.520.961
BiLSTM0.03140.00180.04243.960.968
CNN–LSTM0.02860.00150.03873.420.972
Transformer0.02590.00120.03462.850.976
XGBoost0.02480.00110.03312.620.978
WTX–CBO (Proposed)0.01890.00080.02831.870.987
ItalyLSTM0.03560.00230.04814.880.958
BiLSTM0.03280.00190.04364.050.967
CNN–LSTM0.03030.00160.04023.540.972
Transformer0.02720.00130.03632.910.975
XGBoost0.02610.00110.03392.680.978
WTX–CBO (Proposed)0.01950.00090.02911.790.986
AustriaLSTM0.03380.00200.04484.420.963
BiLSTM0.03090.00170.04133.850.969
CNN–LSTM0.02840.00140.03743.330.974
Transformer0.02560.00120.03422.740.978
XGBoost0.02430.00100.03182.530.981
WTX–CBO (Proposed)0.01810.00070.02691.720.989
GermanyLSTM0.03670.00240.04904.970.958
BiLSTM0.03320.00190.04374.090.967
CNN–LSTM0.03080.00160.03983.510.973
Transformer0.02740.00130.03642.960.977
XGBoost0.02600.00110.03362.710.981
WTX–CBO (Proposed)0.01910.00080.02821.840.987
Table 7. Performance comparison with existing hybrid forecasting models.
Table 7. Performance comparison with existing hybrid forecasting models.
ModelArchitectureMAERMSEMAPE (%)R2Reference
CNN–LSTMDeep hybrid RNN0.02870.03873.420.972[5]
FCM–CNN–WNN–AMMulti-resolution attention0.02340.03292.460.981[9]
EWT–BiLSTMEmpirical wavelet-based DL0.02250.03142.290.982[12]
CWS (CNN–WT–SVR)Hybrid regression-based0.02100.03022.100.984[10]
WTX–CBO (Proposed)Wavelet–Transformer–XGBoost0.01890.02831.870.987This study
Table 8. Performance comparison of optimization algorithms.
Table 8. Performance comparison of optimization algorithms.
OptimizerMAERMSER2Convergence EpochsStabilityObservation
Adam0.02350.03380.97887StableLocal optimizer; slower convergence
GA0.02270.03250.98072ModerateProne to premature convergence
PSO0.02180.03120.98265HighBalanced exploration/exploitation
CBO (Proposed)0.01890.02830.98759Very HighFast and stable chaotic convergence
Table 9. Impact assessment of model components.
Table 9. Impact assessment of model components.
ConfigurationMAERMSEMAPE (%)R2
Without WT Decomposition0.02260.03352.320.978
Without Transformer (WT–XGBoost)0.02180.03212.210.981
Without XGBoost (WT–Transformer)0.02070.03102.050.983
Without CBO Optimization0.01990.02981.960.985
Full Model (WTX–CBO)0.01890.02831.870.987
Table 10. Training efficiency and convergence analysis.
Table 10. Training efficiency and convergence analysis.
ModelEpochs to ConvergeTraining Time (s)Validation R2Observation
LSTM1204150.961Gradual convergence
CNN–LSTM1053720.972Improved feature extraction
Transformer983500.976Strong long-term memory
XGBoost893120.978Fast but less generalizable
WTX–CBO (Proposed)783120.987Fastest and most stable convergence
Table 11. Computational efficiency and scalability of the proposed WTX–CBO model.
Table 11. Computational efficiency and scalability of the proposed WTX–CBO model.
MetricValueObservation
Training Time (epochs = 100)312 s25% faster than Transformer
Inference Time/Window0.038 sSuitable for real-time SEV use
Memory Usage1.2 GBLightweight hybrid design
Model Parameters2.14 MModerate complexity
Hardware UsedRyzen 9/A100 GPU128 GB RAM
Table 12. Inference latency comparison (batch size = 1).
Table 12. Inference latency comparison (batch size = 1).
ModelLatency (s)Memory (GB)Observation
LSTM0.0561.35Recurrent overhead
CNN–LSTM0.0491.42Convolutional operations cost
Transformer0.0451.28Attention complexity scales linearly
XGBoost0.0411.10Lightweight tree ensemble
WTX–CBO (Proposed)0.0381.20Optimized for real-time SEV deployment
Table 13. Statistical significance analysis using the Wilcoxon signed-rank test between the proposed WTX–CBO and baseline models.
Table 13. Statistical significance analysis using the Wilcoxon signed-rank test between the proposed WTX–CBO and baseline models.
Comparison Modelp-ValueSignificance ( α = 0.05 )
WTX–CBO vs. LSTM0.0042Significant
WTX–CBO vs. BiLSTM0.0061Significant
WTX–CBO vs. CNN–LSTM0.0095Significant
WTX–CBO vs. Transformer0.0087Significant
WTX–CBO vs. XGBoost0.0074Significant
Table 14. Diebold–Mariano test results between WTX–CBO and baseline models.
Table 14. Diebold–Mariano test results between WTX–CBO and baseline models.
Comparison ModelDM Statisticp-Value
WTX–CBO vs. LSTM2.8460.0045
WTX–CBO vs. BiLSTM2.5130.0084
WTX–CBO vs. CNN–LSTM2.2970.0121
WTX–CBO vs. Transformer2.1810.0163
WTX–CBO vs. XGBoost2.0570.0198
Table 15. Feature importance ranking from XGBoost within WTX–CBO (reversed order).
Table 15. Feature importance ranking from XGBoost within WTX–CBO (reversed order).
FeatureGain Importance (%)
Frost Point1.0
Precipitation Rate2.7
Dew Point Temperature4.1
Wind Direction5.9
Surface Pressure7.2
Precipitable Water8.4
Wind Speed9.7
Relative Humidity13.5
Air Temperature18.9
Global Horizontal Irradiance (GHI)28.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mchara, W.; Cicceri, G.; Manai, L.; Raissi, M.; Albaqami, H. Hybrid Wavelet–Transformer–XGBoost Model Optimized by Chaotic Billiards for Global Irradiance Forecasting. J. Sens. Actuator Netw. 2026, 15, 12. https://doi.org/10.3390/jsan15010012

AMA Style

Mchara W, Cicceri G, Manai L, Raissi M, Albaqami H. Hybrid Wavelet–Transformer–XGBoost Model Optimized by Chaotic Billiards for Global Irradiance Forecasting. Journal of Sensor and Actuator Networks. 2026; 15(1):12. https://doi.org/10.3390/jsan15010012

Chicago/Turabian Style

Mchara, Walid, Giovanni Cicceri, Lazhar Manai, Monia Raissi, and Hezam Albaqami. 2026. "Hybrid Wavelet–Transformer–XGBoost Model Optimized by Chaotic Billiards for Global Irradiance Forecasting" Journal of Sensor and Actuator Networks 15, no. 1: 12. https://doi.org/10.3390/jsan15010012

APA Style

Mchara, W., Cicceri, G., Manai, L., Raissi, M., & Albaqami, H. (2026). Hybrid Wavelet–Transformer–XGBoost Model Optimized by Chaotic Billiards for Global Irradiance Forecasting. Journal of Sensor and Actuator Networks, 15(1), 12. https://doi.org/10.3390/jsan15010012

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop