Next Article in Journal
Utilizing Cob–Earth and Sand-Filled Plastic Bottles to Address Environmental Challenges: A Sustainable Construction Solution
Previous Article in Journal
Multi-Scale Ecosystem Service Supply–Demand Dynamics and Driving Mechanisms in Mainland China During the Last Two Decades: Implications for Sustainable Development
Previous Article in Special Issue
The Lithium-Ion Battery Temperature Field Prediction Model Based on CNN-Bi-LSTM-AM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Metaheuristic Optimized Extreme Learning Machine for Sustainability Focused CO2 Emission Prediction Using Globalization-Driven Indicators

by
Mahmoud Almsallti
,
Ahmad Bassam Alzubi
and
Oluwatayomi Rereloluwa Adegboye
*
Business Administration Department, Institute of Graduate Research and Studies, University of Mediterranean Karpasia, Mersin-10, TR-10 Mersin, Northern Cyprus, Lefkosa 99010, Turkey
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(15), 6783; https://doi.org/10.3390/su17156783
Submission received: 10 June 2025 / Revised: 23 July 2025 / Accepted: 23 July 2025 / Published: 25 July 2025

Abstract

The escalating threat of climate change has intensified the global urgency to accurately predict carbon dioxide (CO2) emissions for sustainable development, particularly in developing economies experiencing rapid industrialization and globalization. Traditional Extreme Learning Machines (ELMs) offer rapid learning but often yield unstable performance due to random parameter initialization. This study introduces a novel hybrid model, Red-Billed Blue Magpie Optimizer-tuned ELM (RBMO-ELM) which harnesses the intelligent foraging behavior of red-billed blue magpies to optimize input-to-hidden layer weights and biases. The RBMO algorithm is first benchmarked on 15 functions from the CEC2015 test suite to validate its optimization effectiveness. Subsequently, RBMO-ELM is applied to predict Indonesia’s CO2 emissions using a multidimensional dataset that combines economic, technological, environmental, and globalization-driven indicators. Empirical results show that the RBMO-ELM significantly surpasses several state-of-the-art hybrid models in accuracy (higher R2) and convergence efficiency (lower error). A permutation-based feature importance analysis identifies social globalization, GDP, and ecological footprint as the strongest predictors underscoring the socio-economic influences on emission patterns. These findings offer both theoretical and practical implications that inform data-driven Artificial Intelligence (AI) and Machine Learning (ML) applications in environmental policy and support sustainable governance models.

1. Introduction

The accumulation of greenhouse gases in the atmosphere resulting from massive emissions has inflicted significant and potentially irreversible harm on the Earth’s natural ecosystem [1]. Between 2011 and 2020, the average global surface temperature rose by approximately 1.1 °C compared to pre-industrial levels (1850–1900). This ongoing environmental degradation has made climate change a major concern shared by nations. Despite increasing global awareness, greenhouse gas emissions in general, and CO2 emissions in particular, have continued to rise [2]. These emissions are driven by unsustainable practices in energy production, land use, and consumption [3]. According to the Intergovernmental Panel on Climate Change (IPCC), global warming has led to more frequent and intense heatwaves, droughts, floods, and sea-level rise [4].
These changes threaten ecosystems, food and water security, human health, and economic stability, particularly in regions with limited capacity to adapt. Without significant reductions in greenhouse gas emissions, many of these effects risk becoming irreversible, further widening social inequalities and obstructing sustainable development [5]. In response to the escalating threat, the global community has undertaken over the years a series of international initiatives and agreements. One of the earliest milestones was the 1992 United Nations Conference on Environment and Development, which laid the foundation for future climate action [6]. In 1994, the United Nations Framework Convention on Climate Change (UNFCCC) came into effect as a non-binding treaty aimed at encouraging voluntary efforts to address greenhouse gas emissions. The first legally binding international accord was the Kyoto Protocol, adopted in 1997 and entered into force in 2005 [7]. The 2015 Paris Agreement marked another significant global commitment, aiming to limit global temperature rise to well below 2 °C above pre-industrial levels, and preferably to 1.5 °C [8]. Recently, at the United Nations’ COP26 conference on climate change held in 2021, strategies for mitigating greenhouse gas emissions were explored across different thematic areas. Among these areas, particular attention was given to the application of machine learning in forecasting greenhouse gas emissions, emphasizing its potential to support informed decision-making in efforts to reduce and manage emissions effectively [9]. This highlights a critical implication: the success of climate action initiatives does not rely only on political actions but also on effective scientific support. Mitigating emissions requires accurate models to predict greenhouse gas levels and identify key contributing factors at the individual, community, and national levels.
The literature on CO2 emissions forecasting generally categorizes existing methodologies into three main groups: conventional statistical models, standalone machine learning models, and hybrid approaches. Conventional statistical models comprise techniques such as System Dynamics (SD), Vector Auto-Regression (VAR), Auto-Regressive Integrated Moving Average (ARIMA), and Gray Models (GM), which have long been applied. Feng et al., using an SD model, forecasted Beijing’s CO2 emissions and concluded its increase to 169.67 million tones CO2 equivalent by 2030 [10]. Similarly, Bao et al. used an SD model to simulate carbon emissions in China’s thermal power sector, predicting a peak of 4.228 billion tones by 2026 under current development trends [11]. Hernandez et al. applied VAR and ARIMA models to forecast energy consumption and CO2 emissions in a smart industry context, concluding that the ARIMA model demonstrated superior predictive accuracy [12]. Hamzacebi & Karakurt applied a GM to estimate Turkey’s energy-related CO2 emissions from 1965 to 2025 [13]. While these studies show the use of statistical methods in CO2 emission forecasting, they also reveal notable limitations, as highlighted by Hu et al. [14]. A major shortcoming is their inability to capture complex, nonlinear relationships among influencing variables. Another key drawback lies in their reliance on predefined functional forms and assumptions about data distribution, which can oversimplify real-world dynamics and lead to reduced predictive accuracy.
The second category, standalone machine learning models, including Artificial Neural Networks (ANN) [15], Support Vector Regression (SVR) [16], and Long Short-Term Memory Networks (LSTM) [17], responds to the increasing availability of high-dimensional, real-time environmental data for emission forecasting [18]. Machine learning algorithms are particularly effective at detecting complex patterns and correlations within large datasets. As a result, they hold substantial promise for enhancing the accuracy of CO2 emission predictions. However, they often require intensive tuning and can suffer from overfitting, leading to the rise in hybrid approaches.
Hybrid approaches integrate machine learning algorithms with metaheuristic optimization techniques or combine multiple models to improve predictive accuracy. In this category, Saqr et al. proposed a hybrid model combining Greylag Goose Optimization (GGO) with a Multi-Layer Perceptron (MLP) to improve the accuracy of electric vehicle (EV) emissions forecasting [19]. Khajavi and Rastgoo developed a hybrid model using Random Forest, Support Vector Regression, and optimization algorithms to predict CO2 emissions in 30 Chinese cities [20]. The Random Forest model optimized with the Slime Mold Algorithm achieved the best test accuracy, making it the most effective predictor. Alhussan et al. introduced a CO2 forecasting framework using Bidirectional Gated Recurrent Unit (BIGRU) networks optimized with GGBERO, a novel algorithm that fuses Greylag Goose Optimization and Al-Biruni Earth Radius [21]. The model achieved a very low MSE. Phatai and Luangrungruang developed a predictive model combining a Backpropagation Neural Network (BPNN) with Particle Swarm Optimization (PSO) to estimate CO2 emissions based on energy consumption data [22]. The results showed that the PSO-optimized machine learning model achieved high prediction accuracy.
The ELM, a single-hidden-layer feedforward neural network, has gained recognition for addressing certain limitations in traditional machine learning models, particularly in terms of computational speed and efficiency [23]. ELM is distinguished by its rapid training process and strong generalization ability, achieved by randomly assigning weights in the hidden layer and analytically determining the output weights. This streamlined architecture allows ELM to handle large datasets with low computational cost. However, like many machine learning models, ELM’s performance is highly sensitive to the random initialization of input weights and biases, which can result in variability and reduced prediction accuracy [24]. Furthermore, its generalization performance tends to degrade when dealing with high-dimensional or noisy datasets. These challenges have motivated ongoing research into optimization strategies aimed at improving ELM’s reliability and predictive power [25]. Pradhan et al. proposed a software defect prediction model combining ELM with Improved JAYA (IMJAYA) to optimize weights and biases [26]. Van Thieu et al. proposed hybrid models combining ELM with Pareto-like Sequential Sampling (PSS), Weighted Mean of Vectors (INFO), and Runge–Kutta Optimizer (RUN) to improve streamflow prediction [27]. Among the models tested, PSS-ELM achieved the best performance with high accuracy and robust convergence. Abba et al. applied ELM models optimized with Particle Swarm Optimization (PSO), Genetic Algorithm (GA), Biogeography-Based Optimization (BBO), and a BBO-PSO hybrid to predict treated water quality parameters in Nigeria [28]. The BBO-PSO-ELM model achieved superior accuracy for pH, total dissolved solids, and hardness, while BBO-ELM performed best for turbidity. Saleh et al. addressed the prediction of Suspended Sediment Load (SSL), by ELM with four metaheuristic algorithms: PSO, Henry Gas Solubility Optimization (HGSO), Electromagnetic Field Optimization (EFO), and Nuclear Reaction Optimization (NRO) [29]. Experiments demonstrated that the ELM-HGSO model outperformed other models. Sun et al. addressed the challenge of identifying key drivers of CO2 emissions in Hebei, China. They proposed a PSO-optimized ELM (PSO-ELM) model that incorporates factor analysis to reduce the number of influencing variables. The outcome showed that PSO-ELM outperformed standard ELM and BPNN, enhancing prediction accuracy and offering policy insights for low-carbon strategies [30]. Wei et al. tackled the issue of accurate CO2 emission prediction in Hebei, China. They proposed a hybrid MFO-RF-ELM model, where Random Forest was used for feature analysis and ELM, optimized by Moth-Flame Optimization (MFO), handled prediction. The outcome showed that the proposed model surpassed other parallel models, achieving higher accuracy in forecasting CO2 emissions [31]. Algwil and Khalifa addressed the need for accurate CO2 prediction. They proposed a GMSMFO-ELM hybrid model, where Gaussian Mutation and Shrink Mechanism-based Moth Flame Optimization enhanced ELM training. The model achieved 96.5% R2 and outperformed other hybrids across multiple error metrics, with key predictors identified as economic growth, FDI, and renewable energy [24]. Wang et al. focused on improving carbon emission forecasting in public buildings. They proposed the MIV-IHHO-DELM hybrid model, integrating Mean Impact Value (MIV) for feature selection, an Improved Harris Hawk Optimization (IHHO) for optimization, and a Deep ELM (DELM) for prediction. The model achieved a MAPE of 0.704% and RMSE of 18.01, significantly outperforming existing models and showing strong generalization, making it a powerful tool for emission reduction planning in the construction sector [32].
While these reviewed hybrid approaches have improved the ELM’s performance, challenges remain in achieving consistent accuracy, robustness in noisy datasets, and adaptability to the high-dimensional, nonlinear nature of CO2 emissions data. These limitations are especially pressing in the context of recent international climate action. Notably, the COP26 United Nations Climate Change Conference emphasized the urgent need for improved strategies in monitoring, reporting, and forecasting greenhouse gas emissions, particularly in developing economies where data gaps and limited technical infrastructure remain key challenges. To address these limitations, this study proposes a novel hybrid model, the Red-Billed Blue Magpie Optimizer-tuned Extreme Learning Machine (RBMO-ELM). The RBMO algorithm, inspired by the adaptive foraging behavior of red-billed blue magpies [33], efficiently optimizes ELM’s input to hidden layer weights and hidden layer biases. This not only reduces the impact of random initialization but also enhances stability and predictive power across varied environmental conditions. The RBMO-ELM model is thus designed to offer a more robust, adaptive, and computationally efficient solution for CO2 emission forecasting.
The remainder of the paper is structured as follows. Section 2 outlines the methodology, including the Red-Billed Blue Magpie Optimizer (RBMO), the Extreme Learning Machine (ELM), and the proposed RBMO-ELM hybrid prediction model. This is followed by the Experiment and Discussion section, which presents an evaluation of the RBMO-ELM model’s performance in Section 3. The paper concludes with a summary of the key findings and suggestions for future research in Section 4.

2. Methodology

2.1. The Red-Billed Blue Magpie Optimizer (RBMO)

The RBMO developed by Fu et al. is inspired by the Red-billed Blue Magpie (RBM), an avian species indigenous to parts of Asia. The optimization process of RBMO commences with the random generation of a population of candidate solutions within the specified bounds of the search space [33]. The initial population matrix X is expressed in Equation (1).
X = x 1,1 x 1 , j x 1 , d i m x 2,1 x 2 , j x 2 , d i m x n , 1 x n , j x n , d i m
where X R n × d i m represents the positions of all n search agents across d i m dimensions, and x i , j indicates the value of the j th   variable for the i th   agent. Each element x i , j is initialized as expressed in Equation (2).
x i , j = ( u b l b ) · R a n d 1 + l b
Here, u b and l b are the upper and lower boundaries of the decision space, and R a n d 1 U ( 0,1 ) is a uniformly distributed random number. RBMs forage either in small collaborative groups or larger clusters, adapting their strategy according to group size. This dynamic is modeled in RBMO using two separate update equations as given in Equations (3) and (4). Small group behavior (at each iteration, a group of randomly chosen individuals of size p are chosen within { 2,5 } ) as given in Equation (3).
X i t + 1 = X i t + 1 p m = 1 p   X m t X r s t · R a n d 2
Cluster-based exploration, with group size q is randomly chosen within [ 10 , n ] , and is modeled as expressed in Equation (4). The values of the two group sizes p and q are stochastically determined at each iteration to maintain diversity and dynamic role distribution among agents. This randomized group partitioning enhances the global and local search trade-off throughout the optimization process.
X i t + 1 = X i t + 1 q m = 1 q   X m t X r s t · R a n d 3
where X i t is the current position of the i t h agent at iteration t , X m t denotes a randomly selected individual from the population, X r s t is a randomly sampled solution from the current population, and R a n d 2   a n d   R a n d 3 denote a random number within 0 and 1. These equations emulate the adaptive and cooperative food-searching behavior of the RBMs. Once a food source is identified, the magpies coordinate their attack. The RBMO mirrors this through targeted movements toward promising solutions (the food location), moderated by cooperation and randomness. In small groups, the position is updated with Equation (5).
X i t + 1 = X f o o d t + C F · 1 p · m = 1 p   X m t X i t · R a n d n 1
In cluster attacks, the update rule is expressed in Equation (6).
X i t + 1 = X f o o d t + C F · 1 q · m = 1 q   X m t X i t · R a n d n 2
where X food   t is the position of the best solution, and R a n d n 1 and R a n d n 2 denote a random number with standard normal distribution. C F (Control Factor) is defined as follows in Equation (7)
C F = 1 t T 2 · t T
with T being the maximum number of iterations while t is the current iteration. This term gradually shifts the search emphasis from exploration to exploitation as the algorithm progresses. To simulate the food storing behavior of RBMs, the algorithm retains promising solutions for future exploitation. This mechanism ensures that only improvements in fitness are accepted, as expressed in Equation (8).
X i t + 1 = X i t ,   if   fitness   o l d   i >   fitness   new   i X i t + 1 ,   otherwise  
This elitism-based mechanism reinforces convergence stability and guards against performance degradation.

2.2. Extreme Learning Machine (ELM)

The ELM, introduced by Huang et al. [34], is a type of single-hidden-layer feedforward neural network (SLFN). ELM is recognized for its extremely fast training speed, minimal parameter tuning, and strong generalization capability. Unlike deep neural networks, which require large-scale datasets, extensive hyperparameter tuning, and high computational cost, ELM offers an efficient alternative for achieving high predictive accuracy. ELM addresses the limitations associated with traditional learning algorithms, such as slow convergence and entrapment in local minima, particularly by circumventing the need for iterative gradient-based optimization. This makes it well-suited for modeling highly nonlinear systems with complex activation functions [35]. The ELM architecture operates under two primary principles. Firstly, the input weights w i and biases b i of the hidden layer neurons are assigned randomly and remain fixed throughout training. The only adjustable parameter is the number of hidden nodes N . Secondly, instead of using iterative updates, the output weights β are computed analytically by transforming the learning task into a least squares problem, where the hidden layer output matrix is pseudoinverted using the Moore—Penrose method. Given an input x , the output function of ELM can be expressed Equation (9).
f ( x ) = i = 1 N   β i G w i , b i , x = β · h ( x )
where β i denotes the output weight connecting the i th   hidden node to the output, w i and b i are the input weight and bias of the i th   hidden neuron, G w i , b i , x is the activation function applied to the input, and h ( x ) = G w 1 , b 1 , x , , G w N , b N , x represents the hidden layer output vector with respect to input x . The training objective is to minimize the discrepancy between the predicted and target outputs over the training set, which leads to the following optimization problem in Equation (10)
m i n i = 1 N   β · h x i y i 2
The optimal output weight vector β is obtained through the Moore–Penrose generalized inverse of the hidden layer output matrix H , as follows in Equation (11).
β = H + Y
H +   is the Moore—Penrose inverse of H, and Y is the target output matrix. Despite its computational efficiency, the random initialization of w i and b i in ELM sometimes leads to suboptimal solutions, impairing predictive accuracy and stability, especially in complex, chaotic systems such as CO2 prediction. To mitigate this shortcoming, optimization techniques like the RBMO are often employed. The RBMO enhances the ELM by adaptively tuning the input weights and hidden layer biases, thereby leveraging the exploratory strengths of swarm intelligence to augment the learning capacity and predictive robustness of the ELM framework. The illustration of the ELM model is given in Figure 1.

2.3. Proposed Prediction Model

The flowchart in Figure 2. illustrates the operational framework of the Red-billed Blue Magpie Optimizer-Extreme Learning Machine (RBMO-ELM) model, a hybrid approach designed to enhance predictive accuracy and robustness by combining the global search capabilities of the RBMO with the rapid learning efficiency of the ELM. The process begins with the initialization of a population of candidate solutions, each representing a set of input weights and biases for the ELM model. These candidates are evaluated using the ELM, which computes the fitness of each solution based on a predefined objective function, in this research, the Root Mean Squared Error (RMSE). The main optimization loop proceeds iteratively, where in each iteration, a stochastic operation is applied based on whether small-group (Equations (3) and (5)) or cluster-based (Equations (4) and (6)) update strategies should be applied, simulating the cooperative foraging and hunting behavior of red-billed blue magpies. After updating the positions, the new solutions are again evaluated using ELM. If a new solution yields better fitness than its predecessor, it is retained; otherwise, the previous position is preserved through an elitist strategy. The global best solution is continuously tracked and updated until the maximum number of iterations is reached, after which the optimal set of ELM parameters is returned. Despite its efficiency and generalization strength, the ELM suffers from critical limitations. These include its sensitivity to the random initialization of input weights and biases, which often leads to suboptimal solutions. The RBMO component effectively mitigates these issues by optimizing the ELM’s parameters through adaptive, population-based search, enhancing both the stability and accuracy of the overall model. Thus, the RBMO-ELM hybrid provides a powerful tool for tackling nonlinear, high-dimensional prediction tasks that require both computational speed and adaptive learning capability.

3. Experiment and Discussion

In this section, two experimental frameworks are devised to comprehensively assess the efficacy of the proposed Red-billed Blue Magpie Optimizer (RBMO) and its hybridization with the Extreme Learning Machine (ELM). The first experimental phase focuses on the standalone performance of the RBMO, which is benchmarked against 15 standardized test functions from the CEC 2015 competition suite [36]. These functions encompass a broad spectrum of optimization scenarios, including unimodal, multimodal, hybrid, and composite landscapes designed to rigorously evaluate the algorithm’s search ability, convergence behavior, and adaptability across diverse and challenging environments. Each test function provides a medium for examining the RBMO’s capacity to balance exploration and exploitation effectively. In the second phase, the RBMO is integrated with the ELM to construct a hybrid predictive framework aimed at modeling carbon dioxide (CO2) emissions. This task poses significant challenges for traditional predictive models due to its nonlinear, dynamic, and chaotic nature. The performance of the RBMO-ELM model is systematically validated using a suite of evaluation metrics. These metrics offer comprehensive insights into the model’s predictive accuracy, generalization capacity, and robustness.

3.1. Evaluation of the Proposed RBMO on CEC2015 Functions

To comprehensively evaluate the optimization proficiency of the proposed RBMO algorithm, a suite of 15 benchmark functions from the CEC 2015 competition test suite is employed. These benchmark problems are deliberately developed to emulate a broad spectrum of real-world optimization challenges and span various categories, including unimodal, multimodal, hybrid, and composite functions. Each category presents distinct topological complexities, such as separability and modality, that rigorously test the algorithm’s global search behavior, convergence stability, and exploitation capability. In order to ensure a fair and unbiased comparative evaluation, RBMO is benchmarked against several state-of-the-art metaheuristic algorithms, namely the original Aquila Optimizer (AO) [37], Exponential Distribution Optimizer (EDO) [38], Harris Hawks Optimization (HHO) [39], the JAYA algorithm [40], Polar Lights Optimization (PLO) [41], Sine Cosine Algorithm (SCA) [42], and the Red-billed Blue Magpie Optimizer (RBMO). This focuses on recently developed metaheuristic optimizers that demonstrate strong performance in complex search scenarios and have seen successful integration into hybrid machine learning models. Collectively, these algorithms form a modern, competitive benchmark that enables robust assessment of standalone RBMO and RBMO-ELM’s search dynamics and convergence behavior relative to other leading metaheuristic approaches. Each algorithm is subjected to a uniform experimental procedure: the maximum number of iterations is fixed at 4000, the population size is maintained at 30, and all experiments are executed in a 30-dimensional search space. Moreover, to account for the stochastic nature of these algorithms, each test is independently repeated 30 times. The specific parameter configurations for all compared algorithms are systematically outlined in Table 1. The primary objective of this experimental framework is to quantitatively and qualitatively assess the effectiveness of RBMO in solving high-dimensional and complex optimization problems. The parameter values presented in Table 1 are adopted from the original study, where these configurations were empirically tuned and validated across multiple benchmark functions. These settings have been retained to ensure consistency and comparability with the foundational algorithm.
The comparative analysis of the RBMO against several state-of-the-art metaheuristic algorithms reveals the consistent superiority of the RBMO across the CEC 2015 benchmark suite as seen in Table 2. In the single-modal category (F1–F2), RBMO achieves significantly better performance than all other compared algorithms, demonstrating strong global convergence capabilities and high precision in optimizing functions with a single optimum. It outperforms traditional approaches such as AO and HHO, which exhibit a higher average error and variability. For multimodal functions (F3–F5), the RBMO maintains a clear advantage by avoiding premature convergence and locating better optima in F3 and F5. Compared to competitors, especially the EDO, SCA, and JAYA, the RBMO demonstrates a more consistent ability to escape local minima and achieve lower error values, which is critical in multimodal landscapes with numerous deceptive basins of attraction. The PLO demonstrated enhanced convergence in F4 and F5.
In the case of hybrid functions (F6–F8), which integrate multiple search space properties and present a higher level of complexity, the RBMO consistently delivers superior results. It records lower average objective values and reduced standard deviations compared to all other algorithms, indicating both efficiency and stability in F6 and F8. While some algorithms, such as the PLO and SCA, show competitive performance on selected functions, the RBMO leads across the board in this category. The composite functions (F9–F15) further affirm RBMO’s robustness—more specifically F10, F12, F13, F14, and F15. These functions, designed to simulate real-world optimization challenges with diverse structural traits, are effectively navigated by RBMO, which secures the top performance on nearly all of them. In contrast, algorithms like HHO and EDO struggle with convergence or exhibit higher variability, while RBMO reliably produces better and more consistent results. Its advantage is particularly pronounced in functions with complex, rugged landscapes, where other methods demonstrate increased instability. Statistical validation using the Friedman test confirms the dominance of the RBMO, which obtains the best rank across all compared algorithms. The test also indicates statistically significant differences in performance compared to the other algorithms according to the Wilcoxon Sign Rank Sum p-value, reinforcing the reliability of the observed outcomes.
The convergence plots for functions F1 through F15 in Figure 3 offer valuable insights into the dynamic behavior and optimization efficiency of the RBMO in comparison to other metaheuristic algorithms. Across the plots in Figure 3, the fitness mean values of 30 independent runs over 4000 iterations provide an illustration of convergence speed, solution quality, and algorithmic stability.
For F1 and F2 (unimodal functions), the RBMO exhibits a faster convergence rate and significantly lower final fitness values than all competing algorithms. This demonstrates its high precision in locating the global optimum in smooth search landscapes and reflects strong exploitation ability. In particular, while other algorithms exhibit slower or stagnated convergence trends, the RBMO continues to improve steadily and consistently. In F3, F4, and F5, which are multimodal functions characterized by numerous local optima, the RBMO maintains a notable advantage in convergence behavior. It not only reaches better final solutions but also demonstrates smoother and more stable convergence trajectories, F3 and F5, though slightly less than the PLO in F4. Competing algorithms such as the EDO and SCA show fluctuating or plateaued performance, suggesting premature convergence or ineffective exploration.
The convergence behavior on F6, F7, and F8, which belong to the hybrid function category, further reinforces the RBMO’s effectiveness in F6 and F8. These functions are complex due to their combined modality and landscape heterogeneity. Despite these challenges, the RBMO attains the lowest fitness values with minimal oscillations, indicating a superior balance between global exploration and local exploitation. Notably, the AO and JAYA display inconsistent convergence patterns and fail to reach comparable solution quality. For F9–F15, representing composite functions with intricate structural properties, RBMO once again demonstrates the leading performance in most functions. It converges more rapidly and achieves superior fitness outcomes, whereas most competitors converge prematurely. The gradual but decisive descent in the RBMO’s fitness curve highlights its robustness in handling non-separable, rugged, and multimodal landscapes.
The box plots presented in Figure 4 for benchmark functions F1 through F15 provide a robust statistical visualization of the distribution, stability, and robustness of the optimization results obtained during the benchmarking experiment of the RBMO against compared optimizers. These box plots graphically illustrate the median, interquartile range (IQR), and outlier behavior, offering deeper insights beyond average convergence behavior. For F1 and F2, which are unimodal functions, the RBMO demonstrates extremely compact box plots with narrow interquartile ranges and no outliers. This reflects a high degree of consistency and precision across multiple independent runs. In contrast, algorithms such as the HHO and JAYA show high variance and frequent outliers, indicating less reliable convergence and sensitivity to initialization. On multimodal functions (F3–F5), the advantage of the RBMO becomes even more evident. For instance, in F3 and F5, the RBMO exhibits lower medians and tighter distributions compared to all other optimizers. Competing algorithms, particularly the EDO and HHO, not only have wider box plots but also demonstrate multiple outliers, revealing frequent suboptimal convergence. While the PLO shows a relatively compact distribution in some cases, the RBMO still consistently records lower fitness values and greater robustness.
In the case of hybrid functions (F6–F8), which integrate diverse modalities and impose more complex search space dynamics, the RBMO continues to exhibit excellent performance. Its box plots are notably compact with consistently lower median values, suggesting strong adaptability and efficient search behavior in irregular and composite landscapes. Conversely, algorithms like the EDO and HHO reveal substantial variance and outliers, implying the less effective handling of complex landscapes. For composite functions such as F9 and F15, the RBMO maintains superior distributional characteristics. The box plots for these functions show that the RBMO not only achieves lower central tendency but also exhibits reduced dispersion compared to the other methods. Particularly in F10, while the HHO and SCA show high variance and numerous extreme values, the RBMO remains stable with a tight clustering of results, affirming its robustness and convergence reliability under challenging conditions. Finally, the box plot analysis corroborates the RBMO’s exceptional stability, reliability, and resistance to outliers across a wide range of benchmark functions. Its ability to produce consistently high-quality solutions with minimal performance variation highlights its potential as a dependable optimizer for both theoretical and practical applications.

3.2. Evaluation of RBMO-ELM Model

3.2.1. Data

The dataset utilized in this study comprises quarterly observations from the first quarter of 1990 (1990 Q1) to the fourth quarter of 2022 (2022 Q4), representing a comprehensive time span of 33 years for Indonesia. The data encompasses a range of environmental, technological, and socio-economic indicators relevant to the analysis of carbon emissions. Specifically, the variables include the following: carbon emissions (CO2) as the dependent variable, and six independent features—ecological footprint (EF), economic growth (GDP), green innovation (GINO), medium and high-tech industries (MHT), renewable energy (REN), and social globalization (SG). These variables were sourced from reputable international databases, including the World Bank [38], OECD Green Growth Indicators [39], Global Footprint Network [40], the British Petroleum Database (BP), and the KOF Swiss Economic Institute Database, to ensure data reliability. Prior to analysis, the dataset underwent a series of preprocessing steps. All features were normalized using the Min-Max scaling technique to bring them into the [0, 1] range, using the minimum and maximum values of each feature, thereby ensuring comparability and improving model convergence. The dataset was subsequently divided into training and testing subsets using a 70:30 split with randomized selection to avoid order-based bias. Additionally, the correlation heatmap and distribution plots of variables, are illustrated in Figure 5 and Figure 6. These preprocessing procedures ensured the dataset was clean, well-structured, and analytically robust for the subsequent application of machine learning and optimization algorithms. For data completeness, only observations with full availability across all selected indicators were used, resulting in a clean and consistent dataset. Stationarity testing, typically required for statistical models like ARIMA, was not applied here, as EML and related machine learning models can handle non-stationary inputs due to their flexible nonlinear approximation capabilities.

3.2.2. Evaluation Metrics

Evaluation metrics are fundamental to machine learning, providing a consistent and quantitative basis for evaluating the predictive performance of models. These metrics not only measure the reliability of model outputs in approximating actual observations but also serve as critical tools in guiding model refinement, comparative assessment, and the selection of optimal predictive frameworks. In this study, a comprehensive suite of standard performance indicators is utilized to evaluate the accuracy and reliability of the proposed model. The selected metrics include the coefficient of determination R 2 , root mean squared error (RMSE), mean squared error (MSE), maximum error (ME), and relative absolute error (RAE) [43]. The respective mathematical definitions of these metrics are provided below in Equations (12)–(16).
R M S E = 1 n i = 1 n   y i y ˆ i 2
M S E = 1 n i = 1 n   y i y ˆ i 2
R A E = i = 1 n     y i y ˆ i i = 1 n     y i y
R 2 = 1 i = 1 n     y i y ˆ i 2 i = 1 n     y i y 2
M E = m a x y i y ˆ i
In the equations above, n denotes the number of data instances, y i represents the actual observed values, y ˆ i denotes the predicted values, and y is the mean of the observed data. Collectively, these metrics enable a rigorous evaluation of the model’s predictive accuracy, robustness, and generalization capability, thereby supporting a reliable comparison among the compared models.

3.2.3. Evaluation of Models and Discussion

In this study, the number of hidden neurons N in the ELM was fixed at 10 for all optimizer-tuned models and the baseline ELM. This design choice ensures that performance differences across models stem solely from the quality of the optimized weights and biases, rather than variation in network architecture. The range of optimization for each weight and bias was set between [−10,10], consistent with common practice in ELM-based hybrid models. The experimental evaluation, conducted using 5-fold cross validation, rigorously compares the performance of the proposed RBMO-ELM against several state-of-the-art hybrid models, including the AO-ELM, EDO-ELM, HHO-ELM, JAYA-ELM, PLO-ELM, SCA-ELM, and the baseline ELM model. This study utilized recent optimizers due to their enhanced exploration–exploitation balance, dynamic control mechanisms, and adaptive strategies, making them more suitable for benchmarking the proposed RBMO-ELM. This focus ensures a fair and modern comparative framework aligned with current algorithmic advancements in optimization research. The results span both training and testing phases, evaluated using five regression metrics as expressed in Equations (12)–(16).
The RBMO-ELM consistently outperforms all benchmarked models in the training phase across all five folds. Its R2 values range between 0.9957 and 0.9979, indicating that more than 99.5% of the variance in the training data is accounted for by the model predictions, as seen in Table 3. In contrast, most other hybrid models display R2 values ranging between 0.88 and 0.95, suggesting higher bias and less predictive precision. The RBMO-ELM also exhibits substantially lower RMSE and MSE values across all folds. For instance, in Fold 1, the model achieves an RMSE of 5.7070 × 10−3 and an MSE of 3.3000 × 10−5, in contrast to the AO-ELM and EDO-ELM, which yield higher RMSE and MSE values. These reductions in error imply a more accurate model with minimal residual variance. The maximum error (ME) achieved by the RBMO-ELM is also the lowest in every fold, ranging from 1.6825 × 10−2 to 2.3126 × 10−2, which highlights the model’s ability to maintain prediction reliability even in the worst-case scenario. In comparison, other methods such as the HHO-ELM and AO-ELM often produce higher ME values, indicating higher sensitivity to outlier predictions. Furthermore, the RBMO-ELM demonstrates remarkable performance in terms of the RAE, while the PLO-ELM and SCA-ELM though competitive still exhibit higher RAE in most folds. Models like the AO-ELM and HHO-ELM again perform the worst, with RAE values reflecting a poor normalization of absolute errors.
The robustness and generalization capability of the RBMO-ELM are most evident during the testing phase in Table 4. Across all five folds, the RBMO-ELM maintains the highest generalization accuracy, with R2 values ranging from 0.9878 to 0.9981. In contrast, traditional hybrid models like the AO-ELM and HHO-ELM exhibit performance degradation during testing, with R2 dropping as low as 0.84, reflecting a tendency to overfit during training. In terms of the RMSE and MSE, the RBMO-ELM continues to yield the most accurate predictions. This significant reduction in prediction error confirms that the RBMO-ELM not only achieves low training error but also generalizes well on unseen data. Moreover, the ME for the RBMO-ELM remains the lowest across the majority of the test folds, ranging from 1.2794 × 10−2 to 4.3513 × 10−2, again substantially outperforming other models. The RAE values of the RBMO-ELM during testing remain impressively low, ranging from 2.7561 × 10−2 to 6.5896 × 10−2, further reinforcing the model’s strong predictive generalization and consistent reduction in absolute errors relative to baseline predictions. Competing optimizers like the JAYA-ELM and EDO-ELM frequently record high RAE values, indicating relative inefficiency.
The statistical analysis of the average, standard deviation, and best results of the training and testing outcomes across 20 independent runs confirms the superior performance of the RBMO-ELM model relative to its competitors. In the training phase, as detailed in Table 5, the RBMO-ELM achieved the highest average R2 of 0.997847, significantly surpassing all peer optimizers, including other strong performers like the PLO-ELM (0.984792) and SCA-ELM (0.984936). This high R2 value indicates that the RBMO-ELM model captures nearly all the variance in the data, thus exhibiting exceptional predictive power. Furthermore, the RBMO-ELM attained the lowest average MSE of 3.5750 × 10−5 and RMSE of 5.9545 × 10−3, reflecting the model’s high precision in approximating true values. The minimal ME of 1.9722 × 10−2 and RAE of 2.7684 × 10−2 further substantiate its accuracy and robustness. These performance metrics are accompanied by very low standard deviations across all indicators, underscoring the model’s stability and consistent convergence across runs.
In the testing phase from Table 6, the RBMO-ELM continued to outperform other hybrid models. It achieved the highest average R2 value of 0.989096, confirming its superior generalization ability on unseen data. In comparison, the PLO-ELM and SCA-ELM, though competitive, attained lower average R2 values of 0.980525 and 0.979919, respectively. The RBMO-ELM also yielded the lowest average RMSE of 1.0178 × 10−2, MSE of 1.0515 × 10−4, and ME of 2.5465 × 10−2, alongside the smallest RAE of 4.8434 × 10−2. Notably, the RBMO-ELM model exhibited the lowest variability, evidenced by the smallest standard deviations, which further confirms its robustness and optimization stability. Traditional optimizers such as the AO-ELM, HHO-ELM, and JAYA-ELM showed inferior results, with notably higher error magnitudes and standard deviations, indicating less reliable performance across runs. The proposed RBMO-ELM model demonstrates substantial improvements in predictive accuracy, reliability, and consistency due to the enhanced exploitation–exploration balance introduced by the RBMO strategy. These findings validate the algorithm’s efficacy for the CO2 prediction task and underscore its superiority in optimizing the ELM model when compared to both classical and recently developed metaheuristic-based alternatives. To address the risk of overfitting, this study adopted a 5-fold cross-validation strategy, ensuring that the RBMO-ELM model generalizes well across different data subsets. Consistently high R2 values and low error metrics (RMSE, MSE, RAE) across both training and testing phases, as shown in Table 3, Table 4, Table 5 and Table 6, indicate that the model maintains robust predictive performance without overfitting. Additionally, the low standard deviations across 20 independent runs, along with minimal error dispersion in Figure 7, Figure 8 and Figure 9, further validate the model’s generalization capability and stability.
The scatter plots shown in Figure 7 illustrate the predictive performance of various hybrid ELM models using their best result from 20 independent runs. Each plot represents the relationship between predicted and actual CO2 emission values for both training and testing sets, with a dashed red line indicating the ideal line (perfect prediction). Among all models, the RBMO-ELM plot exhibits the most alignment of both training and testing data points along the ideal line, reflecting its superior predictive accuracy and generalization capability. The points are tightly clustered with minimal dispersion, indicating that the model captures the underlying mapping between the features and target with high fidelity. This strong fit correlates well with the model’s lowest recorded RMSE and MSE values, as well as the highest R2 from both training and testing evaluations, as given in Table 5 and Table 6.
In contrast, the AO-ELM, HHO-ELM, and JAYA-ELM plots demonstrate noticeable deviation from the ideal line, particularly in the test set points. These deviations highlight their relatively weaker generalization. The HHO-ELM and JAYA-ELM exhibit larger vertical dispersion, which is consistent with their higher MSE and ME values from the test phase. Models such as PLO-ELM and SCA-ELM show reasonably good clustering, particularly during training, but the spread of testing data points indicates mild overfitting and reduced generalization. Though they perform better than the earlier methods, their scatter plots do not exhibit the tight fit achieved by the RBMO-ELM. This is corroborated by their slightly lower R2 and higher RMSE on the test data compared to the RBMO-ELM. The baseline ELM model, lacking any optimization, displays a more dispersed pattern around the ideal line, especially in the test data. This illustrates the limitations of ELM when used without fine-tuned input weights and biases, leading to reduced prediction precision and increased error margins.
The error plot in Figure 8 provides a comparative analysis of the predictive performance and robustness of each hybrid ELM model, showcasing actual versus predicted values across data samples and the corresponding absolute error (in dark red) for both training and testing sets. Among the models, the RBMO-ELM distinctly demonstrates the most consistent and minimal prediction error throughout the entire dataset. The error curve remains tightly centered around zero, with negligible oscillations, particularly during the testing phase. This directly aligns with the model’s superior performance in Table 5 and Table 6, where it achieved the lowest RMSE, MSE, ME, and RAE. The minimal discrepancy between predicted and actual values illustrates the RBMO-ELM’s exceptional generalization capability and its ability to maintain accuracy on unseen data. In contrast, models such as the AO-ELM, HHO-ELM, and JAYA-ELM exhibit significant volatility in their error curves. The amplitude of error deviations increases notably in the testing phase, suggesting reduced stability and a tendency toward overfitting or insufficient adaptation to unseen data. Specifically, the HHO-ELM and JAYA-ELM show several spikes, indicating multiple samples with substantial prediction errors. These fluctuations are consistent with their relatively higher ME and RMSE values reported in the quantitative evaluations. The EDO-ELM performs moderately better, with a comparatively smoother error than the AO-ELM and HHO-ELM. However, it still exhibits noticeable deviations during testing, indicating moderate instability. Similarly, the SCA-ELM and PLO-ELM show improved predictive alignment in the training phase, but the test phase reveals modest error dispersion, reflective of slightly diminished generalization compared to the RBMO-ELM. The baseline ELM model, which lacks any optimization of input weights and biases, presents the most erratic error behavior. Its error plot reveals frequent and high-magnitude deviations across both training and testing samples, underlining its limited modeling capacity and susceptibility to poor convergence. This observation is consistent with its inferior statistical metrics across all performance indicators.
The convergence curve of optimizer-based ELM models during training using the RMSE as the objective function is depicted in Figure 9. The plot clearly demonstrates the optimization efficiency and learning dynamics of the RBMO-ELM in comparison to other hybrid models. The RMSE serves as the fitness function, providing insight into how effectively each optimizer minimizes the training error over 50 iterations. The RBMO-ELM curve exhibits the fastest and most stable convergence, achieving the lowest RMSE throughout the training process. This smooth and consistent descent reflects the robust exploitation and refined search balance achieved through the RBMO mechanism. By contrast, the SCA-ELM and PLO-ELM show moderately strong convergence behaviors, but with slower descent rates and higher final RMSE values than RBMO-ELM. While both models exhibit steady error reduction, they plateau at higher fitness levels, indicating premature convergence or suboptimal search capabilities relative to the RBMO-ELM.
The HHO-ELM optimizer shows early promise with a sharp initial drop in RMSE, but its progression slows significantly and stabilizes at a higher error level than the RBMO. This behavior suggests that while the HHO-ELM can initially exploit promising regions, it lacks the fine-tuned convergence behavior needed for precise optimization. On the other hand, the AO-ELM and EDO-ELM demonstrate noticeably slower and less efficient convergence. The AO-ELM exhibits a gradual descent with significant stagnation, while the EDO-ELM remains largely flat with minimal error reduction, reflecting its limited effectiveness in tuning the ELM model for this particular task. The JAYA-ELM, meanwhile, shows one of the least responsive curves, with minimal variation across iterations. Its near-horizontal trajectory suggests weak search adaptability and underperformance in minimizing the RMSE compared to the other strategies.
Permutation-based feature importance was adopted in this study due to its model-agnostic simplicity, interpretability, and effectiveness in quantifying the relative contribution of each feature to prediction accuracy. Permutation importance remains a widely accepted method for understanding model behavior, particularly in black-box settings such as hybrid metaheuristic-optimized models [44,45]. Though widely accepted, the limitation of permutation importance is that it assumes feature independence. As observed in the correlation matrix (Figure 5), certain moderate correlations exist. This study adopted the permutation importance score as a method to explore the impact of input variables on the model prediction accuracy. The permutation feature importance analysis of the RBMO-ELM model, as presented in Figure 10, provides a comprehensive ranking of input variables based on their influence on CO2 emission predictions. The Social Globalization (SG) indicator emerged as the most critical predictor, contributing the highest importance score. This underscores the significant impact of global social integration such as international communication, information exchange, and migration on national emission levels, due to changes in consumption behavior, industrial demand, and policy diffusion. Social Globalization (SG), as measured by the KOF index, significantly shapes CO2 emissions through increased interpersonal contact, information exchange, and cultural diffusion. Research by Dreher et al. [46] established that higher globalization generally corresponds with elevated CO2 output, attributable to rising consumption and industrial demand [47]. A recent ARDL study by Rasool et al. [48] found that in Indonesia, SG tends to mitigate environmental degradation, largely through the adoption of green technologies and the spread of global environmental norms. Conversely, studies from the South Asia region [49] reveal that SG drastically reduces CO2 emissions. This reflects Indonesia’s policy trajectory to harness globalization to advance green innovation while managing rising consumption and energy pressures. Consequently, policymakers should design SG-linked strategies that simultaneously promote environmental standards, green technologies, and sustainable consumption behaviors. In conclusion, the identification of SG as a significant determinant in CO2 emissions in Indonesia is consistent with findings reported in prior studies. Following SG, Economic Growth (GDP) ranks second in importance, reinforcing the well-established link between economic expansion and increased energy consumption, which often results in elevated carbon emissions. The third most influential feature is Ecological Footprint (EF), which captures environmental degradation and resource utilization efficiency. Its relatively high score signifies its role in explaining variations in emissions by reflecting unsustainable consumption and land use patterns. On the other hand, Medium and High-Tech Industry (MHT), Green Innovation (GINVO), and Renewable Energy (REN) exhibit lower importance scores.

3.2.4. Preservation of System Properties and Model Contributions

Artificial Neural Networks (ANNs), including Extreme Learning Machines (ELMs), have long been recognized for their exceptional capability in modeling nonlinear, high-dimensional, and complex systems. Their strength lies in their data-driven learning mechanism, which enables them to capture patterns and interactions that are often challenging for traditional models analytically. In environmental modeling domains such as CO2 emission prediction, these characteristics make neural networks particularly attractive due to their ability to generalize well across diverse scenarios.
However, despite their versatility, neural networks such as the RBMO-ELM face notable limitations in preserving the inherent structural properties of the systems they model. Neural networks often operate under the assumption of independent and identically distributed data and do not inherently incorporate domain-specific constraints such as temporal dependencies, physical laws, or inter-variable causal relationships. This limitation affects model interpretability and realism in systems governed by complex physical or socio-economic interactions, where preserving the system structure is crucial.
Recent advancements in structure-preserving neural architectures have aimed to bridge this gap. For instance, structure-preserving neural networks, as proposed by Hernández et al., enable the model to maintain specific properties or structures of the data while learning in a manner consistent with the laws of thermodynamics [50]. Additionally, Xiao et al. proposed a machine learning approach that employs recurrent neural networks (RNNs) and is grounded in a mathematical framework that preserves the Birkhoffian structure [51]. Physics-informed neural networks (PINNs) developed by Raissi et al. integrate partial differential equations into the loss function to enforce given laws [52]. These innovations highlight a growing emphasis on embedding domain structure preservation into the learning process of machine learning models.
Although the RBMO-ELM framework proposed in this study does not incorporate structure-preserving mechanisms to preserve the inherent properties of the CO2 emission system in a structural sense, it lacks the integration of domain-specific physical constraints, temporal dependencies, or causal relationships that characterize structure-preserving models. The inherent properties of the CO2 emission system in this study encompass several key structural characteristics. First, nonlinear relationships exist among variables. Emissions are influenced by complex, nonlinear dependencies on predictors such as GDP, social globalization, and ecological footprint, reflecting intertwined socio-economic and environmental mechanisms. Second, the system exhibits temporal dynamics, as emission levels evolve over time in response to industrialization trajectories, policy interventions, and globalization trends. Third, the system is characterized by strong interdependencies among predictors; for instance, GDP affects energy consumption, which in turn is closely linked to the ecological footprint, forming a network of causal and correlative interactions. This limitation arises because the ELM component primarily relies on randomized weights and biases, without incorporating prior knowledge of the system’s physical laws or socio-economic dynamics. The RBMO, on the other hand, focuses solely on minimizing prediction error, without enforcing structural coherence in the learned representation. Nevertheless, it effectively captures the underlying relationships through optimized learning. It contributes significantly to performance and pattern learning through the following three key strengths:
  • Optimized Parameter Tuning: The RBMO enhances ELM training by fine-tuning weights and biases, thereby improving model stability and convergence.
  • Flexible Model Architecture: The inherent simplicity and modularity of ELM allow rapid training and adaptation, enabling the model to approximate complex mappings.
  • Interpretability: Feature importance analysis, especially highlighting social globalization, provides valuable insights for policy framing.
These strengths are evidenced in the empirical results. As illustrated in the predicted vs. actual plots (Figure 7), Error plots in Figure 8, and the sensitivity analysis (Figure 10), the RBMO-ELM demonstrates the capacity to learn consistent patterns within the dataset. The close alignment between predicted and observed CO2 levels suggests effective internalization of the underlying relationships, even in the absence of formal structure-preserving constraints. Preserving the inherent properties of complex systems necessitates model architectures that extend beyond conventional feedforward approaches such as Extreme Learning Machines (ELMs). Advanced frameworks such as recurrent neural networks specifically designed to capture temporal dependencies, and manifold regularization techniques that preserve topological or relational structures offer promising avenues in this regard. Future experiments will explore the integration of these structure-preserving methodologies to enhance model fidelity and interpretability.

4. Conclusions

This study proposed a novel hybrid predictive framework, the RBMO-ELM, which integrates the RBMO with the ELM to improve carbon emission prediction accuracy in Indonesia. The RBMO algorithm, inspired by the cooperative and adaptive foraging behaviors of red-billed blue magpies, demonstrated a strong exploration–exploitation balance and was employed to optimally tune the weights and biases of the ELM model. Comprehensive evaluations were conducted on benchmark functions (CEC2015) to assess the optimization capabilities of RBMO. A quarterly dataset consisting of economic, technological, environmental, and globalization indicators was used to predict CO2 Emissions using the RBMO-ELM model. Empirical results from multiple experiments, including 5-fold cross-validation and 20 independent runs, consistently demonstrate that the RBMO-ELM outperforms other hybrid counterparts in terms of R2, RMSE, MSE, ME, and RAE metrics. The RBMO-ELM also achieved the fastest and most stable convergence profile, highlighting its superior learning and generalization capabilities. Permutation feature importance analysis further identified Social Globalization, Economic Growth, and Ecological Footprint as the most influential drivers of CO2 emissions in Indonesia, providing valuable policy implications regarding the socio-economic dimensions of environmental sustainability. The limitations of this work include the use of a single-country dataset, which limits the generalizability of the model to other countries. Moreover, while the ELM offers rapid learning, it remains sensitive to initial parameter selection, motivating the use of metaheuristics but also introducing algorithmic complexity. Future research will extend this study by applying the RBMO-ELM framework to multi-country or sector-specific datasets, incorporating dynamic or adaptive control strategies within the RBMO for enhanced scalability. Future research will expand the benchmarking framework by incorporating more models, including advanced deep learning architectures such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU), as well as ensemble learning methods like XGBoost and Random Forest. Future work will also focus on embedding structure-preserving techniques within the RBMO-ELM framework. Potential directions will include incorporating domain constraints via regularization, exploring socio-economic informed ELM variants, or leveraging graph-based feature encoders to maintain system preservation.

Author Contributions

M.A.: Conceptualization, Methodology, Formal Analysis, Original Draft. A.B.A.: Supervision, Resources, Editing. O.R.A.: Supervision, Resources, Editing, Methodology, Original Draft. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data obtained through the experiments are available upon request from the corresponding author.

Conflicts of Interest

The authors declare that there are no conflicts of interest.

References

  1. Sun, W.; Huang, C. Predictions of carbon emission intensity based on factor analysis and an improved extreme learning machine from the perspective of carbon emission efficiency. J. Clean. Prod. 2022, 338, 130414. [Google Scholar] [CrossRef]
  2. Khaliq, M.A.; Alsudays, I.M.; Alhaithloul, H.A.S.; Rizwan, M.; Yong, J.W.H.; Ur Rahman, S.; Sagir, M.; Bashir, S.; Ali, H.; Hongchao, Z. Biochar impacts on carbon dioxide, methane emission, and cadmium accumulation in rice from Cd-contaminated soils; A meta-analysis. Ecotoxicol. Environ. Saf. 2024, 274, 116204. [Google Scholar] [CrossRef]
  3. Calvin, K.; Dasgupta, D.; Krinner, G.; Mukherji, A.; Thorne, P.W.; Trisos, C.; Romero, J.; Aldunce, P.; Barrett, K.; Blanco, G.; et al. IPCC, 2023: Climate Change 2023: Synthesis Report. Contribution of Working Groups I, II and III to the Sixth Assessment Report of the Intergovernmental Panel on Climate Change; Core Writing Team, Lee, H., Romero, J., Eds.; Intergovernmental Panel on Climate Change (IPCC): Geneva, Switzerland, 2023. [Google Scholar] [CrossRef]
  4. Saleh, C.; Dzakiyullah, N.R.; Nugroho, J.B. Carbon dioxide emission prediction using support vector machine. IOP Conf. Ser. Mater. Sci. Eng. 2016, 114, 012148. [Google Scholar] [CrossRef]
  5. Jha, M.K.; Dev, M. Impacts of Climate Change. In Smart Internet of Things for Environment and Healthcare; Azrour, M., Mabrouki, J., Alabdulatif, A., Guezzaz, A., Amounas, F., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 139–159. [Google Scholar] [CrossRef]
  6. Thomas, C. The United Nations Conference on Environment and Development (UNCED) of 1992 in Context. Available online: https://www.tandfonline.com/doi/abs/10.1080/09644019208414053 (accessed on 9 June 2025).
  7. Mor, S.; Aneja, R.; Madan, S.; Ghimire, M. Kyoto Protocol and Paris Agreement: Transition from Bindings to Pledges—A Review. Millenn. Asia 2024, 15, 690–711. [Google Scholar] [CrossRef]
  8. Dwivedi, Y.K.; Hughes, L.; Kar, A.K.; Baabdullah, A.M.; Grover, P.; Abbas, R.; Andreini, D.; Abumoghli, I.; Barlette, Y.; Bunker, D.; et al. Climate change and COP26: Are digital technologies and information management part of the problem or the solution? An editorial reflection and call to action. Int. J. Inf. Manag. 2022, 63, 102456. [Google Scholar] [CrossRef]
  9. Emami Javanmard, M.; Ghaderi, S.F. A Hybrid Model with Applying Machine Learning Algorithms and Optimization Model to Forecast Greenhouse Gas Emissions with Energy Market Data. Sustain. Cities Soc. 2022, 82, 103886. [Google Scholar] [CrossRef]
  10. Feng, Y.Y.; Chen, S.Q.; Zhang, L.X. System dynamics modeling for urban energy consumption and CO2 emissions: A case study of Beijing, China. Ecol. Model. 2013, 252, 44–52. [Google Scholar] [CrossRef]
  11. Bao, X.; Xie, T.; Huang, H. Prediction and Control Model of Carbon Emissions from Thermal Power Based on System Dynamics. Pol. J. Environ. Stud. 2021, 30, 5465–5477. [Google Scholar] [CrossRef]
  12. Hernandez, R.M.; Castillo, J.J.; Ragasa, A.A.; Mendez, C.; Agdon, F. Energy Consumption Forecasting for Smart Industry Using Auto-Regressive Integrated Moving Average (ARIMA) and Vector Auto-Regression (VAR) Model. In SIET ’23, Proceedings of the 8th International Conference on Sustainable Information Engineering and Technology, Bali, Indonesia, 24–25 October 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 129–136. [Google Scholar] [CrossRef]
  13. Hamzacebi, C.; Karakurt, I. Forecasting the Energy-related CO2 Emissions of Turkey Using a Grey Prediction Model. Energy Sources Part Recovery Util. Environ. Eff. 2015, 37, 1023–1031. [Google Scholar] [CrossRef]
  14. Hu, Y.; Wang, B.; Yang, Y.; Yang, L. An Enhanced Particle Swarm Optimization Long Short-Term Memory Network Hybrid Model for Predicting Residential Daily CO2 Emissions. Sustainability 2024, 16, 8790. [Google Scholar] [CrossRef]
  15. Altikat, S. Prediction of CO2 emission from greenhouse to atmosphere with artificial neural networks and deep learning neural networks. Int. J. Environ. Sci. Technol. 2021, 18, 3169–3178. [Google Scholar] [CrossRef]
  16. Wen, L.; Cao, Y. Influencing factors analysis and forecasting of residential energy-related CO2 emissions utilizing optimized support vector machine. J. Clean. Prod. 2020, 250, 119492. [Google Scholar] [CrossRef]
  17. Moon, T.; Choi, H.Y.; Jung, D.H.; Chang, S.H.; Son, J.E. Prediction of CO2 Concentration via Long Short-Term Memory Using Environmental Factors in Greenhouses. Hortic. Sci. Technol. 2020, 38, 201–209. [Google Scholar] [CrossRef]
  18. Singh, P.K.; Pandey, A.K.; Ahuja, S.; Kiran, R. Multiple forecasting approach: A prediction of CO2 emission from the paddy crop in India. Environ. Sci. Pollut. Res. 2022, 29, 25461–25472. [Google Scholar] [CrossRef]
  19. Saqr, A.E.-S.; Saraya, M.S.; El-Kenawy, E.-S.M. Enhancing CO2 emissions prediction for electric vehicles using Greylag Goose Optimization and machine learning. Sci. Rep. 2025, 15, 16612. [Google Scholar] [CrossRef]
  20. Khajavi, H.; Rastgoo, A. Predicting the carbon dioxide emission caused by road transport using a Random Forest (RF) model combined by Meta-Heuristic Algorithms. Sustain. Cities Soc. 2023, 93, 104503. [Google Scholar] [CrossRef]
  21. Alhussan, A.A.; Metwally, M.; Towfek, S.K. Predicting CO2 Emissions with Advanced Deep Learning Models and a Hybrid Greylag Goose Optimization Algorithm. Mathematics 2025, 13, 1481. [Google Scholar] [CrossRef]
  22. Phatai, G.; Luangrungruang, T. Modeling Energy-Related CO2 Emissions with Backpropagation and Metaheuristics. In Proceedings of the 2024 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Osaka, Japan, 19–22 February 2024; pp. 429–433. [Google Scholar] [CrossRef]
  23. Cambria, E.; Huang, G.-B.; Kasun, L.L.C.; Zhou, H.; Vong, C.M.; Lin, J.; Yin, J.; Cai, Z.; Liu, Q.; Li, K.; et al. Extreme Learning Machines [Trends & Controversies]. IEEE Intell. Syst. 2013, 28, 30–59. [Google Scholar] [CrossRef]
  24. Algwil, A.R.A.; Khalifa, W.M.S. An enhanced moth flame optimization extreme learning machines hybrid model for predicting CO2 emissions. Sci. Rep. 2025, 15, 11948. [Google Scholar] [CrossRef]
  25. Van Thieu, N.; Houssein, E.H.; Oliva, D.; Hung, N.D. IntelELM: A python framework for intelligent metaheuristic-based extreme learning machine. Neurocomputing 2025, 618, 129062. [Google Scholar] [CrossRef]
  26. Pradhan, D.; Muduli, D.; Zamani, A.T.; Yaqoob, S.I.; Alanazi, S.M.; Kumar, R.R.; Parveen, N.; Shameem, M. Refined Software Defect Prediction Using Enhanced JAYA Optimization and Extreme Learning Machine. IEEE Access 2024, 12, 141559–141579. [Google Scholar] [CrossRef]
  27. Van Thieu, N.; Nguyen, N.H.; Sherif, M.; El-Shafie, A.; Ahmed, A.N. Integrated metaheuristic algorithms with extreme learning machine models for river streamflow prediction. Sci. Rep. 2024, 14, 13597. [Google Scholar] [CrossRef] [PubMed]
  28. Abba, S.I.; Pham, Q.B.; Malik, A.; Costache, R.; Gaya, M.S.; Abdullahi, J.; Mati, S.; Usman, A.G.; Saini, G. Optimization of Extreme Learning Machine with Metaheuristic Algorithms for Modelling Water Quality Parameters of Tamburawa Water Treatment Plant in Nigeria. Water Resour. Manag. 2025, 39, 1377–1401. [Google Scholar] [CrossRef]
  29. Saleh, A.; Zulkifley, M.A. Prediction of suspended sediment load in Sungai Semenyih using extreme learning machines and metaheuristic optimization approach. J. Environ. Manag. 2025, 380, 124987. [Google Scholar] [CrossRef]
  30. Sun, W.; Wang, C.; Zhang, C. Factor analysis and forecasting of CO2 emissions in Hebei, using extreme learning machine based on particle swarm optimization. J. Clean. Prod. 2017, 162, 1095–1101. [Google Scholar] [CrossRef]
  31. Wei, S.; Yuwei, W.; Chongchong, Z. Forecasting CO2 emissions in Hebei, China, through moth-flame optimization based on the random forest and extreme learning machine. Environ. Sci. Pollut. Res. 2018, 25, 28985–28997. [Google Scholar] [CrossRef]
  32. Wang, J.; Tan, Y.; Yu, J.; Yu, H.; Wang, M.; Zhou, M. A deep hybrid prediction framework for building operational carbon emissions: Integrating enhanced extreme learning machines. Energy Rep. 2025, 13, 4126–4140. [Google Scholar] [CrossRef]
  33. Fu, S.; Li, K.; Huang, H.; Ma, C.; Fan, Q.; Zhu, Y. Red-billed blue magpie optimizer: A novel metaheuristic algorithm for 2D/3D UAV path planning and engineering design problems. Artif. Intell. Rev. 2024, 57, 134. [Google Scholar] [CrossRef]
  34. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  35. Chai, S.; Zhang, Z.; Zhang, Z. Carbon price prediction for China’s ETS pilots using variational mode decomposition and optimized extreme learning machine. Ann. Oper. Res. 2025, 345, 809–830. [Google Scholar] [CrossRef]
  36. Liang, J.J.; Qu, B.Y.; Suganthan, P.N.; Chen, Q. Problem Definitions and Evaluation Criteria for the CEC 2015 Competition on Learning-Based Real-Parameter Single Objective Optimization; Technical Report 201411A; Computational Intelligence Laboratory, Zhengzhou University: Zhengzhou, China; Nanyang Technological University: Singapore, 2014; Volume 29. [Google Scholar]
  37. Abualigah, L.; Yousri, D.; Abd Elaziz, M.; Ewees, A.A.; Al-qaness, M.A.A.; Gandomi, A.H. Aquila Optimizer: A novel meta-heuristic optimization algorithm. Comput. Ind. Eng. 2021, 157, 107250. [Google Scholar] [CrossRef]
  38. Abdel-Basset, M.; El-Shahat, D.; Jameel, M.; Abouhawwash, M. Exponential distribution optimizer (EDO): A novel math-inspired algorithm for global optimization and engineering problems. Artif. Intell. Rev. 2023, 56, 9329–9400. [Google Scholar] [CrossRef]
  39. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  40. Venkata Rao, R. Jaya: A simple and new optimization algorithm for solving constrained and unconstrained optimization problems. Int. J. Ind. Eng. Comput. 2016, 7, 19–34. [Google Scholar] [CrossRef]
  41. Yuan, C.; Zhao, D.; Heidari, A.A.; Liu, L.; Chen, Y.; Chen, H. Polar lights optimizer: Algorithm and applications in image segmentation and feature selection. Neurocomputing 2024, 607, 128427. [Google Scholar] [CrossRef]
  42. Mirjalili, S. SCA: A Sine Cosine Algorithm for solving optimization problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  43. Khajavi, H.; Rastgoo, A.; Masoumi, F. Enhanced Streamflow Forecasting for Crisis Management Based on Hybrid Extreme Gradient Boosting Model. Iran. J. Sci. Technol. Trans. Civ. Eng. 2025, 1–22. [Google Scholar] [CrossRef]
  44. 23 Permutation Feature Importance—Interpretable Machine Learning. Available online: https://christophm.github.io/interpretable-ml-book/feature-importance.html (accessed on 2 July 2025).
  45. Altmann, A.; Toloşi, L.; Sander, O.; Lengauer, T. Permutation importance: A corrected feature importance measure. Bioinformatics 2010, 26, 1340–1347. [Google Scholar] [CrossRef]
  46. Dreher, A.; Gaston, N.; Martens, P. Measuring Globalisation; Springer: New York, NY, USA, 2008. [Google Scholar] [CrossRef]
  47. Gygli, S.; Haelg, F.; Potrafke, N.; Sturm, J.-E. The KOF Globalisation Index—Revisited. Rev. Int. Organ. 2019, 14, 543–574. [Google Scholar] [CrossRef]
  48. Rasool, Y.; Jianguo, D.; Ali, K. Exploring the linkage between globalization and environmental degradation: A disaggregate analysis of Indonesia. Environ. Dev. Sustain. 2024, 26, 16887–16915. [Google Scholar] [CrossRef]
  49. Abbas, M.; Yang, L.; Lahr, M.L. Globalization’s effects on South Asia’s carbon emissions, 1996–2019: A multidimensional panel data perspective via FGLS. Humanit. Soc. Sci. Commun. 2024, 11, 1171. [Google Scholar] [CrossRef]
  50. Hernández, Q.; Badías, A.; González, D.; Chinesta, F.; Cueto, E. Structure-preserving neural networks. J. Comput. Phys. 2021, 426, 109950. [Google Scholar] [CrossRef]
  51. Xiao, S.; Chen, M.; Zhang, R.; Tang, Y. Structure-Preserving Recurrent Neural Networks for a Class of Birkhoffian Systems. J. Syst. Sci. Complex. 2024, 37, 441–462. [Google Scholar] [CrossRef]
  52. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
Figure 1. Extreme Learning Machine.
Figure 1. Extreme Learning Machine.
Sustainability 17 06783 g001
Figure 2. RBMO-ELM Flowchart.
Figure 2. RBMO-ELM Flowchart.
Sustainability 17 06783 g002
Figure 3. Convergence Plot of the RBMO and other Optimizers.
Figure 3. Convergence Plot of the RBMO and other Optimizers.
Sustainability 17 06783 g003aSustainability 17 06783 g003b
Figure 4. Box Plot of RBMO and Compared Optimizers on the CEC2015 function.
Figure 4. Box Plot of RBMO and Compared Optimizers on the CEC2015 function.
Sustainability 17 06783 g004aSustainability 17 06783 g004b
Figure 5. Correlation Matrix of Variables.
Figure 5. Correlation Matrix of Variables.
Sustainability 17 06783 g005
Figure 6. Distribution of Variables.
Figure 6. Distribution of Variables.
Sustainability 17 06783 g006
Figure 7. Scatter Plot of Predicted vs. Actual for Training and Testing.
Figure 7. Scatter Plot of Predicted vs. Actual for Training and Testing.
Sustainability 17 06783 g007
Figure 8. Error Plots of Training and Testing.
Figure 8. Error Plots of Training and Testing.
Sustainability 17 06783 g008aSustainability 17 06783 g008b
Figure 9. Convergence Curve.
Figure 9. Convergence Curve.
Sustainability 17 06783 g009
Figure 10. Feature Importance Score.
Figure 10. Feature Importance Score.
Sustainability 17 06783 g010
Table 1. Parameter Settings of Optimizers.
Table 1. Parameter Settings of Optimizers.
OptimizerHyperparameter
AOµ = 0.00565, ω = 0.005 ,
α   =   δ = 0.1
EDOSwitch Parameter = 0.5
HHO E 0 = [ 2,0 ]
JAYA-
PLO m = 100 ,   a = [ 1,1.5 ]
SCA a = 2
RBMO ϵ = 0.5
Table 2. Results of the RBMO and Compared Optimizers on CEC2015.
Table 2. Results of the RBMO and Compared Optimizers on CEC2015.
FunctionMetricAOEDOHHOJAYAPLOSCARBMO
F1AVG4.673 × 1077.707 × 1091.888 × 10106.290 × 1096.754 × 1051.290 × 10102.703 × 103
STD3.619 × 1073.837 × 1096.666 × 1092.385 × 1092.090 × 1052.860 × 1093.242 × 103
F2AVG4.132 × 1043.376 × 1045.874 × 1046.673 × 1046.358 × 1043.846 × 1042.000 × 102
STD5.106 × 1037.202 × 1032.567 × 1031.796 × 1049.265 × 1035.769 × 1031.039 × 103
F3AVG3.273 × 1023.293 × 1023.398 × 1023.343 × 1023.188 × 1023.361 × 1023.128 × 102
STD3.3932.9652.4292.9951.3272.5603.949
F4AVG4.614 × 1037.153 × 1035.344 × 1037.748 × 1033.581 × 1037.542 × 1033.662 × 103
STD5.767 × 1024.755 × 1028.097 × 1023.870 × 1023.877 × 1023.338 × 1024.489 × 102
F5AVG5.013 × 1025.030 × 1025.017 × 1025.026 × 1025.007 × 1025.027 × 1025.007 × 102
STD5.375 × 10−14.409 × 10−15.724 × 10−13.321 × 10−11.176 × 10−13.054 × 10−12.438 × 10−1
F6AVG6.007× 1026.023 × 1026.033 × 1026.011 × 1026.005 × 1026.021 × 1026.003 × 102
STD1.327× 10−11.0433.465 × 10−13.059 × 10−15.882 × 10−24.445 × 10−15.300 × 10−2
F7AVG7.006× 1027.300 × 1027.437 × 1027.181 × 1027.002 × 1027.248 × 1027.004 × 102
STD2.365 × 10−11.154 × 1016.0524.2672.913 × 10−25.1411.948 × 10−1
F8AVG8.408 × 1027.927 × 1044.176 × 1051.491 × 1048.146 × 1024.513 × 1048.050 × 102
STD9.9331.298 × 1055.129 × 1059.928 × 1031.4545.614 × 1041.152
F9AVG9.124 × 1029.134 × 1029.131 × 1029.133 × 1029.121 × 1029.133 × 1029.112 × 102
STD4.319 × 10−12.337 × 10−13.916 × 10−12.533 × 10−13.793 × 10−12.187 × 10−14.550 × 10−1
F10AVG1.852 × 1061.153 × 1062.519 × 1075.757 × 1064.941 × 1059.643 × 1062.243 × 104
STD1.224 × 1061.473 × 1061.428 × 1072.203 × 1061.927 × 1054.361 × 1061.194 × 103
F11AVG4.527 × 1031.434 × 1049.176 × 1055.276 × 1051.137 × 1038.158 × 1061.142 × 103
STD4.695 × 1031.026 × 1042.497 × 1064.552 × 1052.408 × 1017.609 × 1066.184 × 101
F12AVG4.813 × 1036.926 × 1043.498 × 1095.145 × 1053.002 × 1038.429 × 1081.383 × 103
STD1.326 × 1033.320 × 1051.516 × 10108.770 × 1054.957 × 1021.299 × 1094.260 × 102
F13AVG1.562 × 1031.656 × 1031.848 × 1031.570 × 1031.559 × 1031.585 × 1031.558 × 103
STD1.5402.948 × 1011.417 × 1029.0099.386 × 10−18.1566.216
F14AVG2.074 × 1034.818 × 1034.444 × 1032.367 × 1032.115 × 1032.853 × 1031.973 × 103
STD8.698 × 1015.004 × 1021.480 × 1031.911 × 1021.353 × 1022.297 × 1023.022 × 10−1
F15AVG2.843 × 1032.378 × 1032.753 × 1032.786 × 1032.572 × 1032.910 × 1032.365 × 103
STD4.992 × 1013.537 × 1028.056 × 1011.431 × 1028.141 × 1013.159 × 1012.663 × 102
p-value6.550 × 10−46.550 × 10−46.550 × 10−46.550 × 10−41.703 × 10−26.550 × 10−46.550 × 10−4
Friedman Mean3.274.935.934.902.175.571.23
Friedman Rank3574261
Bold means the best value.
Table 3. Training results with 5-fold Cross Validation of the RBMO-ELM.
Table 3. Training results with 5-fold Cross Validation of the RBMO-ELM.
Cross ValidationModelR2RMSEMSEMERAE
Fold 1AO-ELM0.8907314.0066 × 10−21.6050 × 10−31.1848 × 10−11.8758 × 10−1
EDO-ELM0.9757011.8427 × 10−23.4000 × 10−44.3473 × 10−28.6112 × 10−2
HHO-ELM0.9319513.1619 × 10−21.0000 × 10−37.2702 × 10−21.4803 × 10−1
JAYA-ELM0.9498192.8074 × 10−27.8800 × 10−46.1632 × 10−21.3229 × 10−1
PLO-ELM0.9833971.5618 × 10−22.4400 × 10−43.4396 × 10−27.3118 × 10−2
SCA-ELM0.9845471.5068 × 10−22.2700 × 10−43.7415 × 10−27.0541 × 10−2
RBMO-ELM0.9979275.7070 × 10−33.3000 × 10−52.3126 × 10−22.6891 × 10−2
ELM0.9397732.9746 × 10−28.8500 × 10−49.8621 × 10−21.3926 × 10−1
Fold 2AO-ELM0.911953.6210 × 10−21.3110 × 10−39.3294 × 10−21.7603 × 10−1
EDO-ELM0.8983293.7361 × 10−21.3960 × 10−39.0330 × 10−21.7812 × 10−1
HHO-ELM0.9493112.7474 × 10−27.5500 × 10−46.9879 × 10−21.3356 × 10−1
JAYA-ELM0.9382622.9150 × 10−28.5000 × 10−49.2647 × 10−21.3255 × 10−1
PLO-ELM0.9839971.5437 × 10−22.3800× 10−43.5256 × 10−27.5046 × 10−2
SCA-ELM0.9864081.4227 × 10−22.0200 × 10−42.9028 × 10−26.9161 × 10−2
RBMO-ELM0.996526.9200 × 10−34.8000 × 10−51.8310 × 10−23.1468 × 10−2
ELM0.9362393.0814 × 10−29.4900 × 10−49.3922 × 10−21.4980 × 10−1
Fold 3AO-ELM0.9300053.3130 × 10−21.0980 × 10−37.6624 × 10−21.5226 × 10−1
EDO-ELM0.9666572.2814 × 10−25.2000 × 10−44.3149 × 10−21.0841 × 10−1
HHO-ELM0.9763241.9268 × 10−23.7100 × 10−45.4520 × 10−28.8552 × 10−2
JAYA-ELM0.948442.7950 × 10−27.8100 × 10−46.8194 × 10−21.3523 × 10−1
PLO-ELM0.9875481.3974 × 10−21.9500 × 10−43.4310 × 10−26.4219 × 10−2
SCA-ELM0.9868081.4383 × 10−22.0700 × 10−43.2648 × 10−26.6101 × 10−2
RBMO-ELM0.9957078.0650 × 10−36.5000 × 10−51.8547 × 10−23.9022 × 10−2
ELM0.9487332.8354 × 10−28.0400 × 10−47.2765 × 10−21.3031 × 10−1
Fold 4AO-ELM0.8854073.7812 × 10−21.4300 × 10−31.0660 × 10−11.7353 × 10−1
EDO-ELM0.9323783.2731 × 10−21.0710 × 10−38.6088 × 10−21.5206 × 10−1
HHO-ELM0.9023053.4913 × 10−21.2190 × 10−37.7119 × 10−21.6022 × 10−1
JAYA-ELM0.9421352.8082 × 10−27.8900 × 10−47.8125 × 10−21.3091 × 10−1
PLO-ELM0.9806381.5543 × 10−22.4200 × 10−43.3148 × 10−27.1330 × 10−2
SCA-ELM0.980931.5425 × 10−22.3800 × 10−43.4529 × 10−27.0788 × 10−2
RBMO-ELM0.9964486.9570 × 10−34.8000 × 10−51.6825 × 10−23.2432 × 10−2
ELM0.9387492.7644 × 10−27.6400 × 10−41.0055 × 10−11.2687 × 10-−1
Fold 5AO-ELM0.9181093.4764 × 10−21.2090 × 10−37.3694 × 10−21.6327 × 10−1
EDO-ELM0.8800424.0124 × 10−21.6100 × 10−39.7470 × 10−21.8370 × 10−1
HHO-ELM0.8739044.3138 × 10−21.8610 × 10−31.1263 × 10−22.0260 × 10−1
JAYA-ELM0.9749521.8887 × 10−23.5700 × 10−34.7962 × 10−28.8049 × 10−2
PLO-ELM0.9869981.3852 × 10−21.9200 × 10−43.2134 × 10−26.5057 × 10−2
SCA-ELM0.9844671.5141 × 10−22.2900 × 10−43.6857 × 10−27.1108 × 10−2
RBMO-ELM0.9974835.9870 × 10−33.6000 × 10−51.9211 × 10−22.7911 × 10−2
ELM0.940442.9648 × 10−28.7900 × 10−49.0713 × 10−21.3924 × 10−1
Bold means the best value.
Table 4. Test Results with 5-fold Cross Validation of the RBMO-ELM.
Table 4. Test Results with 5-fold Cross Validation of the RBMO-ELM.
Cross ValidationModelR2RMSEMSEMERAE
Fold 1AO-ELM0.9007113.7251 × 10−21.3880 × 10−38.1431 × 10−21.7438 × 10−1
EDO-ELM0.9762321.9915 × 10−23.9700 × 10−44.0401 × 10−29.3899 × 10−2
HHO-ELM0.8783594.1231 × 10−21.7000 × 10−36.8421 × 10−21.9301 × 10−1
JAYA-ELM0.9127352.8869 × 10−28.3300 × 10−45.8517 × 10−21.3188 × 10−1
PLO-ELM0.9813001.6166 × 10−22.6100 × 10−43.7234 × 10−27.5677 × 10−2
SCA-ELM0.9815571.6055 × 10−22.5800 × 10−43.7685 × 10−27.5155 × 10−2
RBMO-ELM0.9915488.9840 × 10−38.1000 × 10−52.4345 × 10−24.1042 × 10−2
ELM0.9430232.8219 × 10−27.9600 × 10−46.7522 × 10−21.3210 × 10−1
Fold 2AO-ELM0.8873783.5240 × 10−21.2420 × 10−37.2771 × 10−21.4569 × 10−1
EDO-ELM0.9154933.8588 × 10−21.4890 × 10−38.2820 × 10−21.6928 × 10−1
HHO-ELM0.9271162.8349 × 10−28.0400 × 10−45.9033 × 10−21.1720 × 10−1
JAYA-ELM0.8840114.2803 × 10−21.8320 × 10−39.9881 × 10−22.2885 × 10−1
PLO-ELM0.9793901.5075 × 10−22.2700 × 10−43.1758 × 10−26.2323 × 10−2
SCA-ELM0.9799631.4864 × 10−22.2100 × 10−42.6537 × 10−26.1450 × 10−2
RBMO-ELM0.9930661.0465 × 10−21.1000 × 10−43.1616 × 10−25.5953 × 10−2
ELM0.9515692.3109 × 10−25.3400 × 10−45.8281 × 10−29.5537 × 10−2
Fold 3AO-ELM0.8406023.9643 × 10−21.5720 × 10−37.4635 × 10−22.0174 × 10−1
EDO-ELM0.9352822.4409 × 10−25.9600 × 10−44.1236 × 10−21.0799× 10−1
HHO-ELM0.9456442.3150 × 10−25.3600 × 10−44.8822 × 10−21.1781 × 10−1
JAYA-ELM0.9081783.0144 × 10−29.0900 × 10−45.2873 × 10−21.2573 × 10−1
PLO-ELM0.9685101.7620 × 10−23.1000 × 10−44.0261 × 10−28.9669 × 10−2
SCA-ELM0.9687461.7554 × 10−23.0800 × 10−43.0425 × 10−28.9332 × 10−2
RBMO-ELM0.9916759.0760 × 10−38.2000 × 10−51.6118 × 10−23.7857 × 10−2
ELM0.8846593.3723 × 10−21.1370 × 10−39.8628 × 10−21.7161 × 10−1
Fold 4AO-ELM0.8484985.5825 × 10−23.1160 × 10−31.2691 × 10−12.8611 × 10−1
EDO-ELM0.8578743.6160 × 10−21.3080 × 10−39.1723 × 10−21.7492 × 10−1
HHO-ELM0.9185734.0926 × 10−21.6750 × 10−39.7528 × 10−22.0975 × 10−1
JAYA-ELM0.9444263.1634 × 10−21.0010 × 10−36.7481 × 10−21.5076 × 10−1
PLO-ELM0.9897061.4552 × 10−22.1200 × 10−43.1934 × 10−27.4579 × 10−2
SCA-ELM0.9899211.4399 × 10−22.0700 × 10−43.3970 × 10−27.3795 × 10−2
RBMO-ELM0.9981435.7830 × 10−33.3000 × 10−51.2794 × 10−22.7561 × 10−2
ELM0.9406443.4942 × 10−21.2210 × 10−37.4547 × 10−21.7908 × 10−1
Fold 5AO-ELM0.9519412.5601 × 10−26.5500 × 10−45.1078 × 10−21.1833 × 10−1
EDO-ELM0.8330625.4056 × 10−22.9220 × 10−31.2144 × 10−22.8049 × 10−1
HHO-ELM0.8907433.8601 × 10−21.4900 × 10−39.5016 × 10−21.7842 × 10−1
JAYA-ELM0.9664222.2964 × 10−25.2700× 10−45.1359 × 10−21.0942 × 10−1
PLO-ELM0.9757411.8189 × 10−23.3100 × 10−43.1967 × 10−28.4070 × 10−2
SCA-ELM0.9773011.7595 × 10−23.1000 × 10−43.3996 × 10−28.1323 × 10−2
RBMO-ELM0.9878221.3829 × 10−21.9100 × 10−44.3513 × 10−2 6.5896 × 10−2
ELM0.9392542.8783 × 10−28.2800 × 10−45.7253 × 10−21.3304 × 10−1
Bold means the best value.
Table 5. Train Results 20 independent run of the RBMO-ELM.
Table 5. Train Results 20 independent run of the RBMO-ELM.
AO-ELMEDO-ELMHHO-ELMJAYA-ELMPLO-ELMSCA-ELMRBMO-ELMELM
R2AVG0.9346290.9354600.9375130.9366960.9847920.9849360.9978470.938278
STD2.9454 × 10−22.7098 × 10−22.5185 × 10−22.1303 × 10−28.8751 × 10−41.0810 × 10−34.1445 × 10−42.2204 × 10−16
BEST0.9810440.9688940.9741150.9839350.9864390.9875070.9985190.938278
RMSEAVG2.9797 × 10−22.8521 × 10−22.9284 × 10−23.1872 × 10−21.4738 × 10−21.4665 × 10−25.9545 × 10−32.9703 × 10−2
STD6.8248 × 10−35.7428 × 10−35.9716 × 10−36.0272 × 10−34.3167 × 10−45.3699 × 10−45.7322 × 10−40
BEST1.6461 × 10−22.0198 × 10−21.9236 × 10−21.6340 × 10−21.3923 × 10−41.3364 × 10−24.9610 × 10−32.9703 × 10−2
MSEAVG9.3445 × 10−48.4655 × 10−48.9325 × 10−41.0521 × 10−3 2.1735 × 10−42.1540 × 10−43.5750 × 10−58.8200 × 10−4
STD4.2106 × 10−43.5542 × 10−43.6003 × 10−43.5410 × 10−41.2615 × 10−51.5347 × 10−56.8984 × 10−65.4210 × 10−19
BEST2.7100 × 10−44.0800 × 10−43.7000 × 10−42.6700 × 10−41.9400 × 10−41.7900 × 10−42.5000 × 10−58.8200 × 10−4
MEAVG8.3745 × 10−26.9831 × 10−27.5297 × 10−28.1140 × 10−23.5662 × 10−23.3720 × 10−21.9722 × 10−29.6242 × 10−2
STD2.6277 × 10−22.2064 × 10−21.7958 × 10−22.4238 × 10−22.3099 × 10−22.1992 × 10−34.4385 × 10−32.7756 × 10−17
BEST4.1000 × 10−23.7874 × 10−23.8928 × 10−24.2437 × 10−23.0536 × 10−22.9552 × 10−21.4964 × 10−29.6242 × 10−2
RAEAVG1.4115 × 10−11.3347 × 10−11.3872 × 10−11.4818 × 10−16.9815 × 10−26.9466 × 10−22.7684 × 10−21.4071 × 10−1
STD3.2329 × 10−22.6874 × 10−22.8288 × 10−22.8022 × 10−22.0446 × 10−32.5440 × 10−32.6649 × 10−32.7756 × 10−17
BEST7.7977 × 10−29.4517 × 10−29.1121 × 10−27.5971 × 10−26.5954 × 10−26.3304 × 10−22.3064 × 10−21.4071 × 10−1
Bold means the best value.
Table 6. Test Results of 20 independent runs of the RBMO-ELM.
Table 6. Test Results of 20 independent runs of the RBMO-ELM.
AO-ELMEDO-ELMHHO-ELMJAYA-ELMPLO-ELMSCA-ELMRBMO-ELMELM
R2AVG0.9048200.9501160.9092300.8749450.9805250.9799190.9890960.943142
STD4.4333 × 10−22.4488 × 10−23.9990 × 10−25.0991 × 10−22.0611 × 10−33.3931 × 10−32.8251 × 10−32.2204 × 10−16
BEST0.9746660.9765490.9675050.9618180.9843770.9841740.9935470.943142
RMSEAVG3.6839 × 10−22.8963 × 10−23.6112 × 10−23.3952 × 10−21.7115 × 10−21.7348 × 10−21.0178 × 10−22.9284 × 10−2
STD8.8579 × 10−36.6741 × 10−38.0632 × 10−37.3024 × 10−39.0292 × 10−41.3856 × 10−31.2473 × 10−31.3878 × 10−17
BEST1.9548 × 10−22.0378 × 10−22.2138 × 10−21.9190 × 10−21.5350 × 10−21.5450 × 10−27.8890 × 10−32.9284 × 10−2
MSEAVG1.4355 × 10−38.8330 × 10−41.3690 × 10−31.2060 × 10−32.9380 × 10−43.0285 × 10−41.0515 × 10−48.5800× 10−4
STD6.6860 × 10−44.3352 × 10−46.0309 × 10−44.9182 × 10−43.1088 × 10−55.1108 × 10−52.7275 × 10−51.0842 × 10−19
BEST3.8200 × 10−44.1500 × 10−44.9000 × 10−43.6800 × 10−42.3600 × 10−42.3900 × 10−46.2000 × 10−58.5800 × 10−4
MEAVG8.4552 × 10−26.1373 × 10−27.6927 × 10−28.1366 × 10−23.8515 × 10−23.8256 × 10−22.5465 × 10−26.8495 × 10−2
STD2.6365 × 10−21.4616 × 10−21.9383 × 10−22.4915 × 10−22.1780 × 10−35.6819 × 10−35.9898 × 10−31.3878 × 10−17
BEST4.3945 × 10−23.5855 × 10−23.9776 × 10−24.1733 × 10−23.2816 × 10−23.1578 × 10−21.6918 × 10−26.8495 × 10−2
RAEAVG1.6802× 10−11.3572 × 10−11.6471 × 10−11.6156 × 10−17.8061 × 10−27.9124 × 10−24.8434 × 10−21.3357 × 10−1
STD4.0401 × 10−23.1275 × 10−23.6777 × 10−23.4749 × 10−24.1182 × 10−36.3198 × 10−35.9354 × 10−32.7756 × 10−17
BEST8.9157 × 10−29.5493 × 10−21.0097 × 10−19.1313 × 10−27.0014 × 10−27.0467 × 10−23.7541 × 10−21.3357 × 10−1
Bold means the best value.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Almsallti, M.; Alzubi, A.B.; Adegboye, O.R. Hybrid Metaheuristic Optimized Extreme Learning Machine for Sustainability Focused CO2 Emission Prediction Using Globalization-Driven Indicators. Sustainability 2025, 17, 6783. https://doi.org/10.3390/su17156783

AMA Style

Almsallti M, Alzubi AB, Adegboye OR. Hybrid Metaheuristic Optimized Extreme Learning Machine for Sustainability Focused CO2 Emission Prediction Using Globalization-Driven Indicators. Sustainability. 2025; 17(15):6783. https://doi.org/10.3390/su17156783

Chicago/Turabian Style

Almsallti, Mahmoud, Ahmad Bassam Alzubi, and Oluwatayomi Rereloluwa Adegboye. 2025. "Hybrid Metaheuristic Optimized Extreme Learning Machine for Sustainability Focused CO2 Emission Prediction Using Globalization-Driven Indicators" Sustainability 17, no. 15: 6783. https://doi.org/10.3390/su17156783

APA Style

Almsallti, M., Alzubi, A. B., & Adegboye, O. R. (2025). Hybrid Metaheuristic Optimized Extreme Learning Machine for Sustainability Focused CO2 Emission Prediction Using Globalization-Driven Indicators. Sustainability, 17(15), 6783. https://doi.org/10.3390/su17156783

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop