Next Article in Journal
Correctness of Fuzzy Inference Systems Based on f-Inclusion
Previous Article in Journal
Queues with Working Vacations: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Carbon Dioxide Emission Forecasting Using BiLSTM Network Based on Variational Mode Decomposition and Improved Black-Winged Kite Algorithm

1
School of Emergency Management, Institute of Disaster Prevention, Langfang 065201, China
2
School of Information Management, Institute of Disaster Prevention, Langfang 065201, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(11), 1895; https://doi.org/10.3390/math13111895
Submission received: 24 April 2025 / Revised: 21 May 2025 / Accepted: 30 May 2025 / Published: 5 June 2025

Abstract

With the growing severity of global climate change, forecasting and managing carbon dioxide (CO2) emissions has become one of the critical tasks in addressing climate change. To improve the accuracy of CO2 emission forecasting, an innovative framework based on variational mode decomposition (VMD), improved black-winged kite algorithm (IBKA), and BiLSTM networks is proposed. This framework aims to address the challenges associated with predicting non-stationary data and optimizing model hyperparameters. Initially, experiments were conducted on 29 benchmark functions using the IBKA algorithm, demonstrating its superior performance in highly nonlinear and complex environments. Subsequently, the BiLSTM model optimized by IBKA was employed to predict CO2 emission trends across four major industries in China, confirming its enhanced prediction accuracy. Finally, a comparative analysis with other mainstream machine learning and deep learning models revealed that the BiLSTM model consistently achieved the best predictive performance across all industries. This research proposes an efficient and practical technical pathway for intelligent carbon emission prediction under the “dual-carbon” strategic goals, offering scientific support for policy formulation and the low-carbon transition.

1. Introduction

Anthropogenic carbon dioxide (CO2) emissions have emerged as the dominant contributor to the intensification of the greenhouse effect, thereby driving global climate change. The rising global temperatures have triggered a spectrum of adverse environmental consequences, including the increased frequency of extreme weather events, accelerated glacial melt, sea level rise, and ecosystem degradation. These phenomena pose significant risks to ecological integrity and threaten the sustainable development of human societies [1]. As the world’s largest developing country and the foremost emitter of CO2, China plays a pivotal role in the global climate system. Its long-standing reliance on coal-dominated, carbon-intensive energy sources has underpinned rapid economic development but has simultaneously led to considerable environmental burdens [2]. Against this backdrop, China’s position in international climate governance has become increasingly critical. CO2, as the principal anthropogenic greenhouse gas, exerts profound and multifaceted effects by altering the climate system and exacerbating challenges related to food security, water resources, biodiversity, and public health—the urgency and complexity of climate change demand robust, science-based policy frameworks for effective mitigation [3]. In response, China has articulated an ambitious dual-carbon strategy, committing to peak carbon emissions by 2030 and to achieve carbon neutrality by 2060. This strategic vision reflects China’s proactive approach to climate governance and its commitment to long-term ecological sustainability. Carbon peaking constitutes a prerequisite for carbon neutrality, with both objectives fundamentally reliant on the precise quantification and management of emissions. Accurate forecasting of China’s CO2 emission trajectories is therefore critical for informing evidence-based policymaking and ensuring the effective implementation of decarbonization measures [4,5]. Furthermore, China’s dual-carbon framework offers a model for integrating ecological civilization into governance, fostering a symbiotic relationship between socioeconomic development and environmental protection, and contributing constructively to global climate mitigation efforts.
The prediction of carbon dioxide (CO2) emissions is subject to the influence of a complex interplay of multidimensional factors, including, but not limited to, macroeconomic dynamics, the structural composition of energy consumption, and seasonal variability. CO2 emission data across various sectors in China, such as aviation, industry, and residential sectors, typically exhibit pronounced non-stationary characteristics, posing a formidable challenge to the accuracy of predictive models. To address the non-stationary characteristics of the data, widely used preprocessing methods include traditional differential methods, logarithmic transformation [6], and exponential smoothing [7]. In addition, signal decomposition techniques that have recently emerged, such as Empirical Mode Decomposition (EMD) [8], Ensemble Empirical Mode Decomposition (EEMD) [9], and Variational Mode Decomposition (VMD) [10], have also provided practical approaches for data rationalization. Among these, VMD has demonstrated superior decomposition accuracy and greater stability, primarily due to its ability to mitigate the mode mixing problem that often affects EMD. Although VMD has been widely applied in areas such as water resources prediction and energy consumption, its application to CO2 emission prediction remains relatively limited [11,12]. Therefore, introducing VMD into carbon dioxide emission prediction is expected to offer new methods for improving prediction accuracy and achieving precise carbon management.
There are various methods for predicting carbon dioxide emissions. For instance, Lotfalipour et al. applied a grey system-based and autoregressive integrated moving average (ARIMA) to forecast Iran’s carbon dioxide emissions over a decade [13]; Borisova et al. utilized the ARMA model to assess the CH4 and CO2 emissions from a harmless waste landfill gas collection system [14]. However, both ARIMA and ARMA, as conventional statistical models based on linear assumptions, exhibit limited capability in effectively capturing the nonlinear dynamics inherent in CO2 emission data, consequently leading to suboptimal predictive accuracy.
With the rise of machine learning, the limitations of traditional linear models are being increasingly addressed. Wen et al. employed an Improved Chicken Swarm Optimization algorithm (ICSO) to optimize Support Vector Machine (SVM) parameters, subsequently applying the constructed ICSO-SVM model to forecast residential energy-related CO2 emissions in Shanghai [15]; Sun et al. proposed a hybrid model integrating Ensemble Empirical Mode Decomposition (EEMD) with a Particle Swarm Optimization-Backpropagation Neural Network (PSO-BPNN) to achieve short-term forecasting of carbon dioxide (CO2) emissions [16]. However, many machine learning approaches insufficiently account for the temporal dependencies and dynamic structures inherent in time series data.
In recent years, deep learning technology has received widespread attention due to its enhanced capabilities in modeling sequential data. Wang et al. employed a Recurrent Neural Network (RNN) to predict carbon dioxide emissions [17]. As a classic time series model, RNN is capable of capturing the temporal dependencies within sequences, but it tends to suffer from the vanishing or exploding gradient problem during long-term predictions, leading to the “forgetting” of early critical information. Singh et al. applied the Long Short-Term Memory (LSTM) model to estimate vehicle carbon dioxide emissions [18]; Kumari et al. utilized the LSTM model for forecasting India’s carbon dioxide emissions [19]. By introducing mechanisms of input gates, forget gates, and output gates, LSTM effectively controls the storage and transmission of information and alleviates the long-term dependency flaws of RNNs. It has demonstrated superior performance in temporal modeling tasks related to CO2 emission forecasting. However, traditional LSTMs only utilize unidirectional information, which restricts their ability to capture comprehensive temporal patterns and bidirectional dependencies within the data.
To address this issue, Xie et al. utilized a Bidirectional Long Short-Term Memory (BiLSTM) network model for temperature data imputation [20]; Siami-Namini et al. conducted a comparative analysis of LSTM and BiLSTM models in time series prediction [21]. BiLSTM consists of forward and backward LSTM units, enabling it to capture contextual dependencies from both past and future time steps, thereby significantly enhancing the model’s understanding and memory of complex sequential patterns. Studies have confirmed that BiLSTM consistently outperforms conventional LSTM models in various time series prediction tasks [22]. Therefore, this study adopts BiLSTM as the foundational model for carbon dioxide emission forecasting.
In BiLSTM, certain critical hyperparameters have a significant impact on model performance. Research has demonstrated that hyperparameters considerably impact the performance of deep learning models to a degree that may even exceed the impact of model architecture selection itself [23]. Hyperparameter optimization is finding the best combination of hyperparameters for a deep learning model, which first requires defining a search space, including the hyperparameters to be optimized and their corresponding ranges. Then, an optimization algorithm is defined to find the best solution within this search space [24,25]. Given the high dimensionality of the search space and the wide variability of individual hyperparameters, the number of possible combinations is vast, requiring extensive evaluation to determine the optimal set [26]. Consequently, hyperparameter tuning remains a major challenge in deep learning applications. When using BiLSTM to predict carbon dioxide emissions, there are two types of hyperparameters: (1) hyperparameters related to the BiLSTM structure, such as the number of units in each BiLSTM layer, the number of stacked BiLSTM layers, etc.; (2) hyperparameters related to the training of BiLSTM, such as the step size for parameter updates during model training (i.e., learning rate), the proportion of neurons randomly dropped during training (represented by the dropout rate), the maximum number of iterations allowed during training (epochs), the optimizer used in error backpropagation, loss function, etc. [27,28,29]. Due to the large number of hyperparameters and the combinatorial complexity of their search space, exhaustive optimization is computationally intensive. To address this, the three most influential hyperparameters—identified through rigorous sensitivity analysis—were selected for systematic optimization in this study, as summarized in Table 8.
In traditional research, these hyperparameters are typically manually determined by researchers based on empirical knowledge, a process that is inherently subjective and often incapable of ensuring optimal solutions [30]. As a result, automated hyperparameter optimization has increasingly emerged as a prominent area of research. Commonly used optimization techniques include Grid Search [31], Random Search [32], and Bayesian Optimization [33], among others. Grid Search is straightforward and exhaustive but has high computational costs; Random Search offers reduced computational demand but suffers from a lack of consistency and reproducibility; Bayesian Optimization, by constructing a probabilistic surrogate model, efficiently directs the search process, but it is prone to becoming trapped in local optima, particularly in high-dimensional and complex parameter spaces.
In recent years, Swarm Intelligence (SI) algorithms have been successfully applied to various fields, such as wind speed forecasting, image recognition, and medical diagnosis, due to their strong global search capability and high robustness [34,35,36]. Yet, their application in carbon dioxide emission forecasting remains relatively underexplored. Swarm Intelligence optimization algorithms represent a class of heuristic, population-based search methods inspired by the collective behavior and decentralized coordination observed in natural systems, designed to identify optimal or near-optimal solutions to complex problems [37]. Currently, mainstream Swarm Intelligence optimization algorithms include MVO, SCA, GWO, RIME, ALO, WOA, STOA, DO, and others [38,39,40,41,42,43,44,45]. Despite the extensive proliferation of SI algorithms, the “No Free Lunch” (NFL) Theorem asserts that no single optimization algorithm can consistently outperform others across all types of problem domains [46]. In other words, if an algorithm performs well in certain situations, it may not perform as well in others. Put differently, no algorithm can excel at all types of problems [47]. Therefore, continual refinement and adaptation of existing algorithms remain essential for enhancing their performance in addressing problem-specific challenges.
The Black Kite Algorithm (BKA), proposed by Jun Wang et al. in 2024 [48], is a swarm intelligence optimization algorithm characterized by strong global search capability, fast convergence speed, and high adaptability and robustness. Benchmark function experiments have shown that the BKA algorithm significantly outperforms several mainstream metaheuristic algorithms such as MVO, SCA, GWO, RIME, ALO, WOA, STOA, and DO in optimization performance. However, the original BKA algorithm has two issues: (1) it is prone to becoming trapped in local optima during the attack phase; (2) the population may prematurely converge to suboptimal solutions during the migration phase. To address these two problems, we propose an improved Black Kite Algorithm (IBKA), which integrates a Lévy flight-based prey escape and population cooperation strategy [49,50] and a nonlinear simplex strategy [51]. The improved algorithm is called IBKA. To verify the effectiveness of the proposed IBKA algorithm, we conducted experiments on multiple benchmark test functions and compared them with mainstream algorithms, including MVO, SCA, GWO, RIME, ALO, WOA, STOA, DO, and the original BKA algorithm. The results demonstrate that IBKA consistently achieves superior optimization performance across all tested functions.
The contributions of this study are as follows:
  • The variational mode decomposition method is applied to carbon dioxide emission prediction, incorporating the nonlinear characteristics of sample data, with the aim of mitigating the impact of inherent non-stationarity in raw data on forecasting accuracy. This approach has significantly enhanced prediction precision and provides a novel perspective for the exploration of carbon dioxide emission forecasting.
  • A VMD-IBKA-BiLSTM framework is proposed for carbon dioxide emission prediction, where BiLSTM is established as a deep learning model specifically designed for carbon dioxide emission forecasting, and IBKA is formulated as an enhanced BKA algorithm dedicated to hyperparameter optimization of BiLSTM.
  • The VMD-IBKA-BiLSTM model proposed in this study was comparatively evaluated against models ARMA, ARIMA, SVM, ANN, and LSTM in predicting carbon dioxide emissions from four sectors in China. The results demonstrate that significant superiority of the proposed model over the comparative models is observed.
This paper is structured as follows: Section 2 is dedicated to the introduction of the signal decomposition method VMD, the carbon dioxide emission prediction model BiLSTM, and the improved optimization algorithm IBKA. Section 3 presents the comparison experiments of IBKA, BKA, and IBKA with nine swarm intelligence optimization algorithms on 29 benchmark functions. It also introduces the application of IBKA and BiLSTM in predicting carbon dioxide emissions in China. Finally, Section 4 concludes this paper.

2. Materials and Methods

A framework for predicting and optimizing carbon dioxide emissions is proposed in this study, which is named VMD-IBKA-BiLSTM. In the framework, VMD is a signal decomposition method for handling non-stationary data, which is used for processing carbon dioxide data. IBKA is a swarm intelligence optimization algorithm that is employed to optimize three critical hyperparameters of BiLSTM. BiLSTM is a deep-learning predictive model for forecasting carbon dioxide emissions. Signal decomposition, the model, and the optimization algorithm are systematically elaborated on in this section.

2.1. VMD

Variational Mode Decomposition (VMD), introduced by Dragomiretskiy and Zosso in 2014 [10], is a signal decomposition technique based on variational optimization with frequency-domain bandwidth constraints. Unlike the EMD family, VMD avoids reliance on local extrema and envelope interpolation. It offers improved stability and precise parameter control, and it effectively mitigates mode mixing.
First, a variational problem is formulated under the assumption that the carbon dioxide data f is decomposed into k components. The constraint condition is defined as the equivalence between the summation of all modes and the original signal, leading to the corresponding constrained variational expression presented in Equation (1):
min { u k , ω k } k σ l δ ( t ) + j π t * u k ( t ) e - j ω k t 2 2 k u k = f ( t ) n
here, { u k } is the kth modal component, { ω k } is the central frequency of the kth mode, δ t is the Dirac delta function, and * denotes the convolution operator.
Then, the constructed variational problem is solved by introducing the Lagrange multiplier λ . The original constrained variational problem was converted into an unconstrained variational problem. The augmented Lagrange expression was obtained and shown in Equation (2):
L ( { u k } , { ω k } , λ ) = α k σ l δ ( t ) + j π t * u k ( t ) e - j ω k t 2 2 + f ( t ) - k u k ( t ) 2 2 + λ t , f ( t ) - k u k ( t )
where α is the quadratic penalty factor for reducing the interference of Gaussian noise, and λ t represents the Lagrange operator.
Through the alternate direction multiplier iteration algorithm and Fourier isometric transformation optimization, the modal components and central frequencies are obtained. Among them, the range of the central frequency is [0, +∞). This range is strictly controlled by the Fourier isometric transformation optimization, which operates as follows: First, leveraging the energy isometry of the Fourier transform (Parseval’s theorem), the variational optimization problem is transferred to the frequency domain, ensuring energy consistency between time and frequency domains. This allows precise constraint of each mode’s bandwidth (regulated by the penalty factor α), confining the central frequency ωk to the dominant frequency range of the actual signal components and avoiding frequency shifts caused by energy leakage. Second, due to the inherent symmetry of the Fourier transform, the transformation yields negative frequencies that lack physical significance. To address this, the Fourier isometric transformation optimization only updates the functionals for non-negative frequencies ( ω ≥ 0) and performs a “double lift” (energy concentration constraint) on these ω ≥ 0, directly filtering out non-physical negative frequencies. Therefore, for all ω ≥ 0 (non-negative central frequencies), the functionals u ^ k and ω k are updated. All ω ≥ 0 are performed a double lift. The detailed expressions can be found in Equations (3)–(5):
u ^ k n + 1 ( ω ) = f ^ ( ω ) - i k u i ( ω ) + λ ^ ( ω ) / 2 1 + 2 α ( ω - ω k ) 2
ω k n + 1 = 0 ω u ^ k n + 1 ( ω ) 2 d ω 0 u ^ k n + 1 ( ω ) 2 d ω
λ ^ k n + 1 ( ω ) = λ ^ k n ( ω ) + γ f ^ ( ω ) - k u ^ k n + 1 ( ω )
Among these variables, u ^ k n + 1 , u i ω , and f ^ ω represent the Fourier transforms of u k n + 1 t , u i t , and f t , respectively. The constraints are satisfied until the condition in Equation (6) is met:
k = 1 K u ^ k ( n + 1 ) - u ^ k n 2 2 u ^ k n 2 2 <

2.2. Carbon Dioxide Emission Forecasting Model: BiLSTM

The BiLSTM network utilized in this study is a bidirectional extension of the standard Long Short-Term Memory (LSTM) architecture. It employs three gating mechanisms—forget gate, input gate, and output gate—to regulate the flow of information and effectively capture temporal dependencies in time-series data. Specifically, the forget gate controls the retention of past cell state information, the input gate manages the incorporation of new input data into the cell state, and the output gate determines the portion of the cell state to be propagated to the hidden state. The structure of the LSTM unit is illustrated in Figure 1.
The calculation of LSTM is shown in Equations (7)–(12):
f t = σ ( w f × h t - 1 + u f × x t + b f )
i t = σ ( w i × h t - 1 + u i × x t + b i )
c ˜ t = tanh ( w c × h t - 1 + u c × x t + b c )
c t = f t c t - 1 + i t c ˜ t
o t = σ ( w o × h t - 1 + u o × x t + b o )
h t = o t tanh ( c t )
Among them, represents the forget gate, represents the input gate, represents the output gate, the new state is the additional new information learned at the current moment t, and the current information is how much information is extracted from previous moments plus how much information is extracted from the current moment.
A BiLSTM unit consists of a forward LSTM unit and a backward LSTM unit, which extract forward and backward temporal features from the sequence, respectively. BiLSTM has stronger sequence modeling capabilities than the basic LSTM; therefore, it is chosen as the basic unit of the prediction model. Its specific structure is shown in Figure 2.

2.3. IBKA

The BiLSTM model proposed in the study contains many hyperparameters that significantly influence its performance. Therefore, an optimization algorithm is needed to find the optimal hyperparameters. The Black-winged Kite Algorithm (BKA) is improved by adding prey escape and population cooperation strategies based on Lévy flight behavior and nonlinear simplex strategies. The IBKA is proposed. Then, IBKA is used to optimize BiLSTM. The original BKA and our improved IBKA are introduced separately in the section.

2.3.1. The Original BKA

The Black Kite Algorithm (BKA), proposed by Jun Wang et al. in 2024 [48], is a metaheuristic optimization algorithm inspired by the black kite’s predatory and migratory behaviors and strong hovering ability. BKA simulates these behaviors through three main stages: population initialization, predatory behavior, and migratory behavior. The following section presents each of these stages separately:
1.
Population initialization phase
B K = B K 1,1 B K 1,2 B K 1 , d B K 2,1 B K 2,2 B K 2 , d B K 3,1 B K 3,2 B K 3 , d . . . . . . . ¨ . . . . B K n , 1 B K n , 2 B K n , d
where n is the number of hyperparameters, d is the dimension size of the given problem, and is the j-th dimension of the i-th black-winged kite. The positions of each black-winged kite are distributed uniformly.
X i = B K lb + rand ( B K ub B K lb )
where represents the coordinate of the i-th Black-winged Kite in the j-th dimension, and are the lower and upper bounds of the dimension, respectively, and the rand is a random value with a range of [0, 1].
During the initialization phase, BKA selects the individual with the best fitness value in the initial population as the leader, representing the black kite’s optimal position. The mathematical formulation for identifying the initial leader (based on fitness minimization) is given as follows:
f b e s t = min ( f ( X i ) )
X L = X ( f i n d ( f b e s t = = f ( X i ) ) )
2.
Aggressive behavior
The Black-winged Kite initiates its attack by diving toward its prey at high speed. Figure 3a illustrates the kite in an attack state, while Figure 3b depicts its hovering posture. The mathematical model of the predatory behavior is presented below:
y t + 1 i , j = y t i , j + n ( 1 + sin ( r ) ) × y t i , j p < r y t i , j + n ( 2 r - 1 ) × y t i , j e l s e
n = 0.05 × e - 2 × t T 2
The following is a detailed description of the formula:
  • y t i , j and y t + 1 i , j represent the position of the i-th black-winged kite in the j-th dimension at the t-th and (t + 1)th iteration steps, respectively.
  • r is a random number between 0 and 1, while p is a constant value equal to 0.9.
  • T is the total number of iterations, and t is the number of iterations that have been completed.
  • n is the dynamic perturbation coefficient.
3.
Migration behavior
Inspired by bird migration behavior, the following hypothesis is proposed: if the fitness of the current population is lower than that of a randomly generated population, the current leader relinquishes its role and joins the migrating population, indicating its unsuitability to continue leading. Conversely, if the current population’s fitness is higher, the leader continues to guide the group toward the objective. This strategy enables dynamic leader selection to ensure effective migration. The mathematical model for the migration behavior of the Black-winged Kite is given below:
y t + 1 i j = y t i j + C ( 0 , 1 ) × y t i j - L t j , F i < F r i y t i j + C ( 0 , 1 ) × L t j - m × y t i j , e l s e
m = 2 × sin r + π 2
  • L t j represents the leading scorer in the j-th dimension for the black-winged kite at the t-th iteration so far.
  • y t i , j and represent the position of the i-th black-winged kite in the j-th dimension at the t-th and (t + 1)th iteration steps, respectively.
  • F i represents the current position of any black-winged kite in the j-th dimension at the t-th iteration.
  • F r i represents the fitness value of the random position in the j-th dimension obtained from any black-winged kite at the t-th iteration.
  • C (0,1) represents Cauchy mutation. It is defined as follows:
The one-dimensional Cauchy distribution is a continuous probability distribution with two parameters. The following equation shows the probability density function of the univariate Cauchy distribution:
f ( x , δ , μ ) = 1 π × δ δ 2 + ( x μ ) 2 , < x <
when δ = 1 , μ = 0, its probability density function will become the standard form. Here is the exact formula:
f ( x , δ , μ ) = 1 π × 1 x 2 + 1 , - < x <

2.3.2. Proposed IBKA

The original BKA algorithm exhibits significant advantages in global search, convergence speed, and robustness. However, it also has some drawbacks. For instance, it tends to fall into local optima during attack behaviors, and the population may converge prematurely to suboptimal solutions during migration behaviors. To address these two issues, we have added two mechanisms to the basic BKA algorithm. The following outlines our improvements in detail.
1.
Lévy Flight-Inspired Prey Escape and Collective Cooperation Strategies
The original BKA algorithm includes global search and various attack behaviors during the search process in attack behaviors. However, it overlooks the situation where the prey escapes, resulting in the failure to find the global optimal solution. The algorithm is prone to falling into local optimal solutions and lacks information exchange during the attack behaviors, which limits the global search capability. Inspired by the Harris hawk optimization algorithm [50] in the exploration and exploitation phases, prey escape, and group cooperation, and the Lévy flight behavior in biological migration [49], we propose a prey escape and population cooperation strategy based on Lévy flight behavior. When the prey senses a threat, it takes escape behavior to evade predators. This behavior can be considered as a mechanism to avoid falling into local optimal solutions in optimization algorithms. In the improved BKA algorithm, we define the formula for prey escape energy as follows:
E = X i - X L 2 = k = 1 d X i , k - X L , k 2
where E represents the energy of prey escape, denotes the position vector of the i-th individual in the kth dimension, and represents the position vector of the current leader individual in the kth dimension.
When abs (E) < 0.1, it indicates that the solution is close to the optimal one. Introduce a population collaboration strategy; the formula is as follows:
X n e w = X i + n × E × ( 1 + sin ( r ) ) × X i
where n refers to the dynamic disturbance coefficient, X i represents the current individual’s position vector, r is a random number between 0 and 1, and X n e w denotes the position after updating the individual.
When abs (E) ≥ 0.1, it indicates that the current individual is far from the optimal solution, posing a risk of falling into a local optimum or search stagnation. We have introduced a Lévy flight global perturbation strategy, and its updated formula is as follows:
X n e w = X + r × L e v y ( d ) × ( X - X L )
where Lévy (d) is a random step size generated based on the Lévy distribution, which involves the following formula:
s t e p ~ u | v | 1 / β
where u ~ N 0 , σ 2 and v ~ N 0,1 are random variables from standard normal distributions; is the Lévy index, controlling the heavy-tailed property of the step size (we set β = 1 .5 in this paper).
2.
Nonlinear Simplex Strategy
In the migration behavior design of the original Black Kite Algorithm (BKA), although the combination of the dynamic leader selection mechanism and the Cauchy mutation strategy effectively maintains the migratory directionality of the population, the mechanism still possesses potential for further optimization. Particularly when addressing complex nonlinear optimization problems, the algorithm tends to prematurely converge to local optima rather than the global optimum. The nonlinear simplex strategy is introduced to address this issue [51]. The nonlinear simplex method is a classical direct search approach frequently employed for solving extremum problems of multivariate functions. Its core idea is to gradually approximate the optimal solution of the function by continuously adjusting the positions of the simplex vertices. In the improved BKA algorithm, the nonlinear simplex strategy is applied during the migration behavior phase. Specifically, a portion of individuals with poorer fitness values are selected from the current population as the simplex vertices to be optimized. Then, the adjustment mechanism of the nonlinear simplex method is used. The positions of these vertices are updated in the hope of finding better solutions iteratively. During the updating process, operations such as reflection, expansion, contraction, and shrinking are employed to adjust the shape and position of the simplex. Through multiple iterations, when the shape and position of the simplex tend to stabilize, the algorithm has converged to the optimal or near-optimal solution within the current search space. In this paper, the reflection operation is selected to adjust the shape and position of the simplex, and its reflection formula is as follows:
X r = c + r ( t ) × ( c - X w ) × cos ( θ )
where c is the simplex’s optimal vertex, is the simplex’s worst vertex, and θ is a random angle in U [0, 2π].
The final proposed IBKA flowchart is shown in Figure 4, where the water-green section represents our improvements.

3. Results and Discussion

3.1. IBKA Performance Verification Experiments

In this section, experiments were conducted to evaluate the optimization performance of the proposed IBKA, including its development, exploration, and local optimum avoidance capabilities. These included ablation studies and comparisons with nine state-of-the-art metaheuristic algorithms. All algorithms were tested under identical conditions on a system with a 12th Gen Intel i5-12500H processor and NVIDIA GeForce RTX 4060 GPU (8 GB VRAM), using 500 iterations and a population size of 30.

3.1.1. Benchmark Functions

To evaluate the performance of IBKA, 29 different benchmark functions are selected from the CEC2017 test suite for comparative experiments [52]. Four types are categorized into 29 benchmark functions. Among them, F1–F3 are unimodal benchmark functions used to test the development ability of the algorithm (after the official deletion of the original F2, it was adjusted to F1–F3); F4–F10 are multimodal benchmark functions used to test the exploration ability of the algorithm; F11¬–F20 are hybrid functions used to test the adaptability in heterogeneous problems; F21–F30 are composition functions used to assess the algorithm’s ability to avoid local optima. The range represents the boundaries of the variables, and F (min) is the optimal value. The CEC2017 benchmark functions are shown in Table 1.
In this section, each algorithm is executed 30 times on each benchmark function to mitigate the impact of random factors. In this paper, the Friedman test method is used to rank and evaluate the results of all algorithms on the benchmark functions.

3.1.2. Ablation Analysis of the IBKA

We integrate prey escape and population cooperation strategies based on Lévy flight behavior, as well as a nonlinear simplex strategy, into the basic BKA algorithm and propose the IBKA algorithm. This section verifies the performance enhancement of the basic BKA through ablation experiments. Table 2 shows the BKA variant algorithms with one or more integrated mechanisms, where “1” indicates that the mechanism is added, and “0” indicates the opposite. These BKA variant algorithms are compared and tested on the previously mentioned 29 benchmark functions.
The experimental results are given in Table 2, where Avg represents the Friedman test average ranking of the algorithm on 29 benchmark functions. The lower the average value, the better the algorithm’s performance. “+/−/=” represents that IBKA is better, worse, or equal to other BKA variants; rank represents the final ranking of the algorithm.
From Table 3, the performance of the basic BKA is the worst. Whether adding a prey escape and population cooperation strategy based on Lévy flight behavior or a nonlinear simplex strategy, the performance of BKA can be improved. IBKA ranks first, which means adding both the prey escape and population cooperation strategy based on Lévy flight behavior and the nonlinear simplex strategy to BKA can effectively enhance its optimization capability.

3.1.3. Comparison with Other Algorithms

A comprehensive performance evaluation of the proposed IBKA was conducted by comparing it with eight state-of-the-art algorithms, along with the original BKA algorithm. The comparison algorithms used in this study are as follows:
  • Multi-verse optimization algorithm (MVO) [38]
  • Sine cosine algorithm (SCA) [39]
  • Grey wolf optimization (GWO) [40]
  • Rime optimization algorithm (RIME) [41]
  • Ant lion optimization (ALO) [42]
  • The whale optimization algorithm (WOA) [43]
  • Sooty tern optimization algorithm (STOA) [44]
  • Dandelion optimization (DO) [45]
  • Black-winged kites algorithm (BKA) [48]
The parameter settings of the aforementioned comparison algorithm are shown in Table 4:
All algorithms are tested under the same environment with a maximum of 500 iterations and a population size of 30. To minimize the impact of randomness on the experimental results, each algorithm was independently executed 30 times on each benchmark function, and the average (Aver) and standard deviation (Std) of the 30 independent runs were calculated for each algorithm on each test function (presented in Table 5, with the best test results highlighted in bold). A smaller average indicates better performance.
Based on Table 5, it is evident that IBKA ranks first among the 29 benchmark functions in the CEC2017 test set for the most part. All algorithms are ranked using the Friedman test method based on their fitness on the benchmark functions, as shown in Table 5. The results of the Friedman test are presented in Figure 5, where the lowest average rank indicates the best performance. Obviously, IBKA has the lowest average rank on the cec2017 test functions. It suggests that the optimization performance of IBKA outperforms the competing algorithms, whether on unimodal, multimodal, hybrid, or composition functions.
Furthermore, the Wilcoxon rank-sum test is used to illustrate the significant differences between IBKA and the competing algorithms. Table 6 reports the p-values of the Wilcoxon rank-sum test, where a p-value less than 0.05 indicates that IBKA has a statistically significant advantage over its competitors (results with p-values greater than 0.05 are displayed in bold). Significant results are highlighted in bold. Obviously, most p-values are below 0.05, which indicates that the superiority of IBKA on the benchmark functions is statistically significant.
The experimental results demonstrate that IBKA consistently achieves solutions closest to the theoretical optimum on most benchmark functions. To further illustrate its convergence behavior, Figure 6 shows the convergence curves of IBKA and the compared algorithms on six representative functions (F3, F10, F15, F18, F22, and F30) with a dimension of 30. For F3, IBKA performs similarly to other algorithms in the early stages but converges faster and continues improving toward the end. On F10, IBKA converges more slowly than RIME but achieves higher final accuracy than BKA. For F15, although initially inferior to BKA, IBKA surpasses it as iterations progress. On F18, IBKA maintains both high accuracy and fast convergence throughout. For F22, IBKA outperforms all others in convergence speed and final precision. On F30, its early performance is comparable to BKA, but the final accuracy is better. These results confirm that IBKA avoids premature convergence and demonstrates strong global exploration and local exploitation capabilities.

3.2. VMD-IBKA-BiLSTM Framework for CO2 Emission Forecasting

Previously, experiments on benchmark functions were completed. In this section, we apply the proposed IBKA algorithm to practical optimization tasks, constructing a CO2 emission prediction framework named VMD-IBKA-BiLSTM. Within this framework, BiLSTM serves as the deep learning model for CO2 emission prediction, while IBKA is utilized to optimize five hyperparameters of the BiLSTM.
In hyperparameter optimization for deep learning models, the optimization space is characterized by its combinatorial and non-convex nature; traditional grid search, bayesian optimization, and random search methods have limitations when dealing with such problems. The IBKA algorithm proposed in this study is particularly suitable for deep learning model optimization. The key reason is that the nonlinear simplex strategy avoids premature convergence to suboptimal solutions through continuous searching during migration behavior. Additionally, the prey escapes and population cooperation strategies based on Lévy flight behavior in the attack behavior exhibit the capability to enhance the algorithm’s search range and avoid falling into local optimal solutions. The BKA algorithm, which combines these two strategies, can accelerate the convergence of the model training process and find the global optimal solution.

3.2.1. CO2 Emission Data

The dataset used in this work is the CO2 emissions from various departments in China from 2019 to June 2023. The experiment selected CO2 emissions data from four major sectors: domestic aviation, international aviation, industry, and residents. The time range of the data is from 1 January 2019 to 31 June 2023. During this period, each sector contains 1612 sampling points. Figure 7 shows all the data used in this section. The x-axis represents the date sequence starting from 1 January 2019, and the y-axis represents the amount of carbon dioxide emissions measured in millions of tons (MM·T−1).

3.2.2. VMD Parameters

The VMD algorithm is utilized to decompose complex nonlinear and non-stationary carbon dioxide data into a series of Intrinsic Mode Function (IMF) components with clear physical significance. These components reflect the intrinsic characteristics of the data and effectively separate the high-frequency and low-frequency information of carbon dioxide emissions. Accurately extracting these features helps to construct more accurate prediction models. VMD exhibits the capability to adaptively match the optimal central frequency and bandwidth for each mode, precisely capturing the dynamic characteristics of the data. It is crucial for revealing the essential characteristics of the data and lays a solid foundation for subsequent predictions. The parameter settings for VMD are shown in Table 7:

3.2.3. Sample Making

In this study, we adopt historical data from a continuous 12-day period to predict the next day’s carbon dioxide emissions. The data samples are divided using a sliding window method, where each sample consists of 12 data points as input and 1 data point as output (as shown in Figure 8), with green data points representing the input and red data points representing the output. We obtained a total of 1600 training samples. The first 80% of the samples are used as the training set, and the remaining 20% as the test set. For model hyperparameter tuning, we perform cross-validation on the training set, with each fold using 20% of the training data as the validation subset. The final model is evaluated independently on the test set.

3.2.4. VMD-IBKA-BiLSTM Flowchart

When utilizing the BiLSTM model for CO2 emission prediction, the following three hyperparameters have the most significant impact on its predictive performance, including (1) the number of units in each BiLSTM layer (unit); (2) the parameter update step size during model training (learning_rate); (3) the maximum number of training epochs for the model (max_epochs).
The improved IBKA is employed to optimize the aforementioned three hyperparameters. First, it is necessary to define the search space boundaries for each hyperparameter with specific value ranges, as shown in Table 8.
Table 8. The value range of the hyperparameter.
Table 8. The value range of the hyperparameter.
Hyperparameter NameDescriptionLower BoundsUpper Bounds
unitThe number of units in the BiLSTM layer50300
learning_rateParameter update step during model training0.0010.01
max_epochsMaximum number of cycles for model training50300
Secondly, initialize IBKA. Among them, the maximum number of iterations is 200, the dimension d is 3, and the population size n is 30. Then, the loss of BiLSTM is used as the fitness function (the loss function is RMSE in the study). The solver of BiLSTM is set to AdaGrad. Finally, the IBKA algorithm is used to search for the optimal hyperparameters of BiLSTM. Figure 9 shows the flowchart of the entire VMD-IBKA-BiLSTM framework.

3.2.5. Evaluation Metrics

Three indicators are adopted to evaluate model performance: Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and Mean Absolute Percentage Error (MAPE) in this section. The calculation formulas are as follows:
M A E = 1 n i = 1 n y i y ^ i
R M S E = 1 n i = 1 n y i y ^ i 2
M A P E = 1 n i = 1 n y i y ^ i y i × 100 %
where n is the number of test samples, is the actual measured value of the sample data, is the predicted value of the sample data, and is the average measured value of the sample data.

3.2.6. Comparison of BiLSTM Optimized by Various Algorithms

In addition to IBKA, we also employed 12 optimization algorithms to optimize our proposed CO2 emission prediction model, BiLSTM, and compared the optimization performance of different algorithms in this section. These comparative algorithms include three traditional optimization algorithms (namely Grid Search, Random Search, and Bayesian Optimization Algorithm), eight state-of-the-art swarm intelligence optimization algorithms (namely MVO, SCA, GWO, RIME, ALO, WOA, STOA, DO), and the original BKA algorithm. To reduce the error caused by experimental randomness, each set of comparative experiments was repeated 30 times.
The study utilizes CO2 emission data from four Chinese government agencies, with results displayed in Table 9, Table 10, Table 11 and Table 12. Among them, the best results are displayed in bold. The terms unit, lr, and mp represent the optimal hyperparameters obtained through algorithmic optimization, corresponding to the number of network units, learning rate, and maximum training epochs, respectively, as defined in Table 8.
According to the results in Table 9, Table 10, Table 11 and Table 12, it is evident that the performance of traditional optimization algorithms (such as grid search, random search, and Bayesian optimization) is significantly lower than that of swarm intelligence optimization algorithms. Among the 10 swarm intelligence optimization algorithms, our improved IBKA achieved the best performance.
Furthermore, different optimization algorithms significantly enhance the performance of the same prediction model, BiLSTM. This indicates that model optimization plays a crucial role and may be even more critical than model selection in deep learning applications.

3.2.7. Comparison with Other Models

In the previous section, the performance of BiLSTM was analyzed under different optimization algorithms, and it was found that the model optimized by IBKA performed the best. In this section, we compare the optimized BiLSTM with five other machine learning and deep learning models, including SVM, ANN, and LSTM, as well as with the two statistical models, ARMA and ARIMA.
To ensure fairness, all models were optimized using IBKA. Each experiment was repeated 30 times, and Table 13 summarizes the average evaluation metric values from the 30 experiments. Table 13 also displays the experimental results of each model (with the best results in bold), where Rank indicates the final ranking of the model. The results demonstrate that BiLSTM significantly outperforms the comparative models in all four evaluation domains.
Figure 10, Figure 11, Figure 12 and Figure 13 visually display the comparison between the predicted values and actual values of various models in the test set, covering four different domains. Here, the x-axis represents time, and the y-axis represents CO2 emissions. From these charts, it is evident that the proposed BiLSTM model has the best predictive performance, with the most minor error, further proving the superiority of this model.

4. Conclusions

In the context of “dual carbon” (carbon peak and carbon neutrality), accurate CO2 emission forecasting plays a crucial role in formulating emission reduction strategies and addressing climate change. To tackle this issue, this study proposes an innovative deep learning framework—VMD-IBKA-BiLSTM—to enhance the accuracy of CO2 emission forecasting. The framework integrates the BiLSTM model with the IBKA optimization algorithm to improve the predictive ability for complex carbon emission data. Initially, experiments are conducted on 29 benchmark test functions to verify the performance of the IBKA optimization algorithm on different datasets. Through ablation experiments and comparison with nine mainstream metaheuristic optimization algorithms, the experimental results indicate that IBKA outperforms existing methods in all 29 benchmark tasks, demonstrating its advantages in nonlinear and high-complexity environments. Subsequently, the BiLSTM model is used to predict the CO2 emission trends of four industries in China and compare the performance of IBKA with other optimization algorithms (including nine metaheuristic optimization algorithms and three traditional optimization algorithms) during this process. The results show that the BiLSTM optimized by IBKA exhibited the best predictive effect in all experiments, significantly improving the accuracy of CO2 emission forecasting. Furthermore, we also compared the optimized BiLSTM with five mainstream machine learning and deep learning models, namely ARMA, ARIMA, ANN, SVM, and LSTM. The result indicates that BiLSTM exhibits the best performance in predicting all industries. In summary, the VMD-IBKA-BiLSTM framework demonstrates outstanding performance in dealing with complex carbon emission data, providing strong technical support for China’s carbon emission monitoring and policymaking. This study not only provides a scientific approach to achieving Chinese “dual carbon” goals but also contributes a scalable Chinese solution to global climate governance, promoting the construction of an ecological civilization system and aiding in achieving the sustainable development vision of harmonious coexistence between humanity and nature.

Author Contributions

Conceptualization, J.G.; data curation, Y.Y. and H.L.; formal analysis, Y.Y., S.L. and J.G.; funding acquisition, Y.Y.; investigation, S.L.; methodology, H.L.; project administration, J.G.; resources, S.L.; software, S.L.; supervision, Y.Y. and J.G; validation, Y.Y., S.L. and H.L.; visualization, S.L.; writing—original draft, Y.Y., S.L. and H.L.; writing—review and editing, Y.Y., S.L. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to express their sincere gratitude to Yang Yueqiao from the Institute of Disaster Prevention Science and Technology for her valuable assistance in language editing and research guidance. Additionally, we would like to thank Liu Haijun and Guo Jidong for their support and help in data analysis, experimental design, and other aspects. We also appreciate all the individuals involved in this research and the peer experts who provided valuable suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alizadeh, O. A review of ENSO teleconnections at present and under future global warming. Wiley Interdiscip. Rev. Clim. Change 2024, 15, e861. [Google Scholar] [CrossRef]
  2. Mehmood, K.; Tauseef Hassan, S.; Qiu, X.; Ali, S. Comparative analysis of CO2 emissions and economic performance in the United States and China: Navigating sustainable development in the climate change era. Geosci. Front. 2024, 15, 101843. [Google Scholar] [CrossRef]
  3. Zhang, Z.; Wang, G. China’s role in global climate governance. Clim. Policy 2017, 17 (Suppl. S1), 32–47. [Google Scholar]
  4. IEA. An Energy Sector Roadmap to Carbon Neutrality in China. Energy Technology Policy Division; IEA: Paris, France, 2021. [Google Scholar]
  5. MEE. List of Chemicals Under Priority Control (Second Batch); Ministry of Ecology and Environment of the People’s Republic of China: Beijing, China, 2020. [Google Scholar]
  6. Benoit, K. Linear Regression Models with Logarithmic Transformations; London School of Economics: London, UK, 2011; Volume 22, pp. 23–36. [Google Scholar]
  7. De Livera, A.M.; Hyndman, R.J.; Snyder, R.D. Forecasting time series with complex seasonal patterns using exponential smoothing. J. Am. Stat. Assoc. 2011, 106, 1513–1527. [Google Scholar] [CrossRef]
  8. Ge, H.; Chen, G.; Yu, H.; Chen, H.; An, F. Theoretical analysis of empirical mode decomposition. Symmetry 2018, 10, 623. [Google Scholar] [CrossRef]
  9. Faysal, A.; Ngui, W.K.; Lim, M.H. Noise eliminated ensemble empirical mode decomposition for bearing fault diagnosis. J. Vib. Eng. Technol. 2021, 9, 2229–2245. [Google Scholar] [CrossRef]
  10. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  11. Ahmadi, F.; Tohidi, M.; Sadrianzade, M. Streamflow prediction using a hybrid methodology based on variational mode decomposition (VMD) and machine learning approaches. Appl. Water Sci. 2023, 13, 135. [Google Scholar] [CrossRef]
  12. Qin, Y.; Zhao, M.; Lin, Q.; Li, X.; Ji, J. Data-driven building energy consumption prediction model based on VMD-SA-DBN. Mathematics 2022, 10, 3058. [Google Scholar] [CrossRef]
  13. Lotfalipour, M.R.; Falahi, M.A.; Bastam, M. Prediction of CO2 emissions in Iran using grey and ARIMA models. Int. J. Energy Econ. Policy 2013, 3, 229–237. [Google Scholar]
  14. Borisova, D.; Kostadinova, G.; Petkov, G.; Dospatliev, L.; Ivanova, M.; Dermendzhieva, D.; Beev, G. Assessment of CH4 and CO2 Emissions from a Gas Collection System of a Regional Non-Hazardous Waste Landfill, Harmanli, Bulgaria, Using the Interrupted Time Series ARMA Model. Atmosphere 2023, 14, 1089. [Google Scholar] [CrossRef]
  15. Wen, L.; Cao, Y. Influencing factors analysis and forecasting of residential energy-related CO2 emissions utilizing optimized support vector machine. J. Clean. Prod. 2020, 250, 119492. [Google Scholar] [CrossRef]
  16. Sun, W.; Ren, C. Short-term prediction of carbon emissions based on the EEMD-PSOBP model. Environ. Sci. Pollut. Res. 2021, 28, 56580–56594. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, H.; Liang, W.; Liang, S.; Chen, B. Research on Carbon Dioxide Concentration Prediction Based on RNN Model in Deep Learning. Highlights Sci. Eng. Technol. 2023, 48, 281–287. [Google Scholar] [CrossRef]
  18. Singh, M.; Dubey, R.K. Deep learning model based CO2 emissions prediction using vehicle telematics sensors data. IEEE Trans. Intell. Veh. 2021, 8, 768–777. [Google Scholar] [CrossRef]
  19. Kumari, S.; Singh, S.K. Machine learning-based time series models for effective CO2 emission prediction in India. Environ. Sci. Pollut. Res. 2023, 30, 116601–116616. [Google Scholar] [CrossRef]
  20. Xie, C.; Huang, C.; Zhang, D.; He, W. BiLSTM-I: A deep learning-based long interval gap-filling method for meteorological observation data. Int. J. Environ. Res. Public Health 2021, 18, 10321. [Google Scholar] [CrossRef]
  21. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The performance of LSTM and BiLSTM in forecasting time series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; IEEE: Los Angeles, CA, USA, 2019; pp. 3285–3292. [Google Scholar]
  22. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. A comparative analysis of forecasting financial time series using arima, lstm, and bilstm. arXiv 2019, arXiv:1911.09512. [Google Scholar]
  23. Khalid, R.; Javaid, N. A survey on hyperparameters optimization algorithms of forecasting models in smart grid. Sustain. Cities Soc. 2020, 61, 102275. [Google Scholar] [CrossRef]
  24. Bergstra, J.; Yamins, D.; Cox, D. Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures. In Proceedings of the International Conference on Machine Learning, PMLR, Atlanta, GA, USA, 17–19 June 2013; pp. 115–123. [Google Scholar]
  25. Probst, P.; Boulesteix, A.L.; Bischl, B. Tunability: Importance of hyperparameters of machine learning algorithms. J. Mach. Learn. Res. 2019, 20, 1934–1965. [Google Scholar]
  26. Bergstra, J.; Komer, B.; Eliasmith, C.; Yamins, D.; Cox, D.D. Hyperopt: A python library for model selection and hyperparameter optimization. Comput. Sci. Discov. 2015, 8, 014008. [Google Scholar] [CrossRef]
  27. Reimers, N.; Gurevych, I. Optimal hyperparameters for deep lstm-networks for sequence labeling tasks. arXiv 2017, arXiv:1707.06799. [Google Scholar]
  28. Zhang, B.; Rajan, R.; Pineda, L.; Lambert, N.; Biedenkapp, A.; Chua, K.; Hutter, F.; Calandra, R. On the importance of hyperparameter optimization for model-based reinforcement learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics, San Diego, CA, USA, 13–15 April 2021; PMLR: Cambridge, MA, USA, 2021; pp. 4015–4023. [Google Scholar]
  29. Wei, Y.; Chen, Z.; Zhao, C.; Tu, Y.; Chen, X.; Yang, R. A BiLSTM hybrid model for ship roll multi-step forecasting based on decomposition and hyperparameter optimization. Ocean Eng. 2021, 242, 110138. [Google Scholar] [CrossRef]
  30. Kaur, S.; Aggarwal, H.; Rani, R. Hyper-parameter optimization of deep learning model for prediction of Parkinson’s disease. Mach. Vis. Appl. 2020, 31, 32. [Google Scholar] [CrossRef]
  31. Sidana, S. Grid Search Optimized Machine Learning based Modeling of CO2 Emissions Prediction from Cars for Sustainable Environment. Int. J. Curr. Sci. Res. Rev. 2024, 7. [Google Scholar] [CrossRef]
  32. Toro-Molina, C.; Rivera-Tinoco, R.; Bouallou, C. Hybrid adaptive random search and genetic method for reaction kinetics modelling: CO2 absorption systems. J. Clean. Prod. 2012, 34, 110–115. [Google Scholar] [CrossRef]
  33. Bian, C.; Zhang, S.; Yang, J.; Liu, H.; Zhao, F.; Wang, X. Bayesian optimization design of inlet volute for supercritical carbon dioxide radial-flow turbine. Machines 2021, 9, 218. [Google Scholar] [CrossRef]
  34. Tao, H.; Salih, S.Q.; Saggi, M.K.; Dodangeh, E.; Voyant, C.; Al-Ansari, N.; Yaseen, Z.M.; Shahid, S. A newly developed integrative bio-inspired artificial intelligence model for wind speed prediction. IEEE Access 2020, 8, 83347–83358. [Google Scholar] [CrossRef]
  35. Dimov, D.T. Rotation-invariant NCC for 2D color matching of arbitrary shaped fragments of a fresco. Pattern Recognit. Lett. 2020, 138, 431–438. [Google Scholar] [CrossRef]
  36. Eshtay, M.; Faris, H.; Obeid, N. Improving extreme learning machine by competitive swarm optimization and its application for medical diagnosis problems. Expert Syst. Appl. 2018, 104, 134–152. [Google Scholar] [CrossRef]
  37. Bacanin, N.; Bezdan, T.; Tuba, E.; Strumberger, I.; Tuba, M. Optimizing convolutional neural network hyperparameters by enhanced swarm intelligence metaheuristics. Algorithms 2020, 13, 67. [Google Scholar] [CrossRef]
  38. Mirjalili, S.; Jangir, P.; Mirjalili, S.Z.; Saremi, S.; Trivedi, I.N. Optimization of problems with multiple objectives using the multi-verse optimization algorithm. Knowl.-Based Syst. 2017, 134, 50–71. [Google Scholar] [CrossRef]
  39. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  40. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  41. Su, H.; Zhao, D.; Heidari, A.A.; Liu, L.; Zhang, X.; Mafarja, M.; Chen, H. RIME: A physics-based optimization. Neurocomputing 2023, 532, 183–214. [Google Scholar] [CrossRef]
  42. Mirjalili, S. The ant lion optimizer. Adv. Eng. Softw. 2015, 83, 80–98. [Google Scholar] [CrossRef]
  43. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2015, 95, 51–67. [Google Scholar] [CrossRef]
  44. Dhiman, G.; Kaur, A. STOA: A bio-inspired based optimization algorithm for industrial engineering problems. Eng. Appl. Artif. Intell. 2019, 82, 148–174. [Google Scholar] [CrossRef]
  45. Zhao, S.; Zhang, T.; Ma, S.; Chen, M. Dandelion Optimizer: A nature-inspired metaheuristic algorithm for engineering applications. Eng. Appl. Artif. Intell. 2022, 114, 105075. [Google Scholar] [CrossRef]
  46. Wolpert, D.H. What is important about the no free lunch theorems? In Black Box Optimization, Machine Learning, and No-Free Lunch Theorems; Springer International Publishing: Cham, Switzerland, 2021; pp. 373–388. [Google Scholar]
  47. Adam, S.P.; Alexandropoulos, S.A.N.; Pardalos, P.M.; Vrahatis, M.N. No Free Lunch Theorem: A Review. Approximation and Optimization: Algorithms, Complexity and Applications; Springer: Berlin/Heidelberg, Germany, 2019; pp. 57–82. [Google Scholar]
  48. Wang, J.; Wang, W.C.; Hu, X.X.; Qiu, L.; Zang, H.-F. Black-winged kite algorithm: A nature-inspired meta-heuristic for solving benchmark functions and engineering problems. Artif. Intell. Rev. 2024, 57, 98. [Google Scholar] [CrossRef]
  49. Liu, Y.; Cao, B. A novel ant colony optimization algorithm with Levy flight. IEEE Access 2020, 8, 67205–67213. [Google Scholar] [CrossRef]
  50. Mohammad, S.; Ibrahim, M.; Zaid, M.; Yousef, S.M.K.; Anas, A.L.B.; Saja, A.D.; Laith, A. Harris Hawks Optimization Algorithm: Variants and Applications. Arch. Comput. Methods Eng. 2022, 7, 5579–5603. [Google Scholar]
  51. Martínez, S.Z.; Montano, A.A.; Coello, C.A.C. A nonlinear simplex search approach for multi-objective optimization. In Proceedings of the 2011 IEEE Congress of Evolutionary Computation (CEC), New Orleans, LA, USA, 5–8 June 2011; IEEE: New Orleans, LA, USA, 2011; pp. 2367–2374. [Google Scholar]
  52. Wu, G.; Mallipeddi, R.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2017 Competition on Constrained Real-Parameter Optimization; Technical Report; National University of Defense Technology: Changsha, China; Kyungpook National University: Daegu, Republic of Korea; Nanyang Technological University: Singapore, 2017. [Google Scholar]
Figure 1. LSTM state gating structure.
Figure 1. LSTM state gating structure.
Mathematics 13 01895 g001
Figure 2. BiLSTM model structure diagram.
Figure 2. BiLSTM model structure diagram.
Mathematics 13 01895 g002
Figure 3. Black-winged kites have two attack strategies: (a) hovering in the air, waiting to attack, and (b) hovering in the air, searching for prey.
Figure 3. Black-winged kites have two attack strategies: (a) hovering in the air, waiting to attack, and (b) hovering in the air, searching for prey.
Mathematics 13 01895 g003
Figure 4. Flowchart of IBKA.
Figure 4. Flowchart of IBKA.
Mathematics 13 01895 g004
Figure 5. Friedman rankings of different algorithms on 29 benchmark functions.
Figure 5. Friedman rankings of different algorithms on 29 benchmark functions.
Mathematics 13 01895 g005
Figure 6. Comparison of Fitness Curves Among Algorithms During Iteration.
Figure 6. Comparison of Fitness Curves Among Algorithms During Iteration.
Mathematics 13 01895 g006aMathematics 13 01895 g006b
Figure 7. CO2 emissions from different sectors.
Figure 7. CO2 emissions from different sectors.
Mathematics 13 01895 g007
Figure 8. Sample making map.
Figure 8. Sample making map.
Mathematics 13 01895 g008
Figure 9. VMD-IBKA-BiLSTM framework.
Figure 9. VMD-IBKA-BiLSTM framework.
Mathematics 13 01895 g009
Figure 10. Error curves of different models in the domestic aviation sector.
Figure 10. Error curves of different models in the domestic aviation sector.
Mathematics 13 01895 g010aMathematics 13 01895 g010b
Figure 11. Error curves of different models in the international aviation sector.
Figure 11. Error curves of different models in the international aviation sector.
Mathematics 13 01895 g011
Figure 12. Error curves of different models in the industry sector.
Figure 12. Error curves of different models in the industry sector.
Mathematics 13 01895 g012aMathematics 13 01895 g012b
Figure 13. Error curves of different models in the resident sector.
Figure 13. Error curves of different models in the resident sector.
Mathematics 13 01895 g013
Table 1. CEC2017 Benchmark Functions.
Table 1. CEC2017 Benchmark Functions.
TypeNo.FunctionsFi * = Fi (x*)
Unimodal
Functions
1Shifted and Rotated Bent Cigar Function100
3Shifted and Rotated Zakharov Function200
Simple
Multimodal
Functions
4Shifted and Rotated Rosenbrock’s Function300
5Shifted and Rotated Rastrigin’s Function400
6Shifted and Rotated Expanded Scaffer’s F6 Function500
7Shifted and Rotated Lunacek Bi _ Rastrigin Function600
8Shifted and Rotated Non-Continuous Rastrigin’s Function700
9Shifted and Rotated Levy Function800
10Shifted and Rotated Schwefel’s Function900
Hybrid
Functions
11Hybrid Function 1 (N = 3)1000
12Hybrid Function 2 (N = 3)1100
13Hybrid Function 3 (N = 3)1200
14Hybrid Function 4 (N = 4)1300
15Hybrid Function 5 (N = 4)1400
16Hybrid Function 5 (N = 4)1500
17Hybrid Function 6 (N = 5)1600
18Hybrid Function 6 (N = 5)1700
19Hybrid Function 6 (N = 5)1800
20Hybrid Function 6 (N = 6)1900
Composition
Functions
21Composition Function 1 (N = 3)2000
22Composition Function 2 (N = 3)2100
23Composition Function 3 (N = 4)2200
24Composition Function 4 (N = 4)2300
25Composition Function 5 (N = 5)2400
26Composition Function 6 (N = 5)2500
27Composition Function 7 (N = 6)2600
28Composition Function 8 (N = 6)2700
29Composition Function 9 (N = 3)2800
30Composition Function 10 (N = 3)2900
Search Range: [−100, 100]D
Table 2. Various BKA variants with two mechanisms.
Table 2. Various BKA variants with two mechanisms.
ModelLFPECCS 1NSS 2
IBKA11
LBKA10
NBKA01
BKA00
1 LFPECCS: Lévy Flight-Inspired Prey Escape and Collective Cooperation Strategies.2 NSS: Nonlinear simplex strategy.
Table 3. The results of the ablation experiment.
Table 3. The results of the ablation experiment.
AlgorithmRank+/−/=Avg
IBKA1~1.6552
LBKA218/0/122.0690
NBKA325/0/52.5862
BKA428/0/23.6897
Table 4. Parameter settings of comparison algorithms.
Table 4. Parameter settings of comparison algorithms.
AlgorithmParameterValue
MVOCoefficient of wormhole expansion ww∈(0.1–0.5)
SCAConvergence parameter spiral factor aa = 2
GWOArea vector a, random vector r1, r2A ∈ [0, 2], r1 ∈ [0, 1], r2 ∈ [0, 1]
RIMEIce crystal growth rate ααϵ(0.5–2.0)
ALOWandering step decay rate ccϵ(1E−5–1E−2)
WOAConvergence factor decay rate aaϵ(2→0)
STOAMigration step factorαϵ(0.5–2.0)
DOEnveloping contraction factor γ, Track the step factor αγϵ(1.0–3.0), αϵ(0.5–2.0)
BKAHover step factor α, Diving intensity factor γγϵ(1.0–3.0), αϵ(0.5–2.0)
Table 5. Experimental results on 29 benchmark functions (the best are in bold).
Table 5. Experimental results on 29 benchmark functions (the best are in bold).
FunF1F3F4
AverStdAverStdAverStd
IBKA8.1830E+037.0422E+031.8912E+031.0926E+035.0512E+021.8824E+01
MVO2.0834E+065.5248E+051.7653E+048.1720E+035.0848E+022.8785E+01
SCA2.1779E+103.7625E+098.0469E+041.4987E+042.9913E+039.6940E+02
GWO3.0973E+092.0276E+096.5206E+041.3765E+046.6954E+029.4817E+01
RIME3.9959E+061.5350E+065.6550E+042.0593E+045.2317E+023.0755E+01
ALO1.8263E+041.4310E+042.2269E+057.2692E+045.5684E+023.5674E+01
WOA5.1528E+092.2079E+092.6804E+056.6301E+041.3270E+033.3620E+02
STOA1.1517E+103.6744E+096.9199E+041.1507E+041.0874E+033.7627E+02
DO9.0487E+055.1165E+053.7217E+041.4747E+045.2275E+023.0270E+01
BKA9.3339E+099.5028E+093.2213E+041.4535E+042.4960E+033.3167E+03
F5F6F7
AverStdAverStdAverStd
IBKA6.7015E+024.6326E+016.3754E+021.5282E+011.0409E+031.0912E+02
MVO6.2880E+024.8084E+016.3446E+021.4620E+018.7723E+023.7373E+01
SCA8.2968E+022.665E+016.6372E+026.6056E+001.2586E+036.6517E+01
GWO6.3001E+024.4073E+016.1339E+024.1681E+009.0428E+025.5862E+01
RIME6.2020E+013.8822E+016.1278E+026.0731E+008.7318E+024.4449E+01
ALO6.7634E+024.9636E+016.4684E+027.4114E+001.1289E+038.7380E+01
WOA8.7229E+025.1075E+016.8024E+021.3970E+011.3262E+037.7212E+01
STOA7.3505E+023.1691E+016.4909E+027.7178E+001.1410E+036.6825E+01
DO6.7471E+024.2709E+016.4089E+021.4148E+011.0196E+038.4731E+01
BKA7.4849E+025.0098E+016.6304E+021.0978E+011.2209E+035.0611E+01
F8F9F10
AverStdAverStdAverStd
IBKA9.3957E+023.2733E+013.9205E+031.0900E+034.7605E+037.8266E+02
MVO9.2897E+023.3092E+016.5787E+033.6856E+034.8820E+035.8507E+02
SCA1.0954E+032.8364E+018.1511E+031.6601E+038.7252E+034.2182E+02
GWO9.0659E+022.4623E+012.5647E+031.1602E+035.5303E+031.6354E+03
RIME9.1213E+022.4865E+012.8693E+031.5267E+034.8791E+035.3497E+02
ALO9.4792E+023.5205E+014.5019E+031.2755E+035.6546E+037.1469E+02
WOA1.0904E+036.5874E+011.0858E+044.1747E+037.6357E+035.7563E+02
STOA9.9967E+023.0121E+016.5067E+031.7392E+037.4005E+036.6993E+02
DO9.6499E+023.9626E+016.1139E+031.9411E+035.2652E+036.1188E+02
BKA9.7942E+025.2221E+015.2284E+039.7625E+025.4685E+031.0673E+03
F11F12F13
AverStdAverStdAverStd
IBKA1.2740E+034.8137E+012.1344E+062.0313E+063.6575E+044.4484E+04
MVO1.3492E+037.3312E+011.7126E+071.9260E+071.4184E+058.7242E+04
SCA3.8055E+039.6441E+022.6769E+091.0277E+091.2457E+094.5075E+08
GWO2.4925E+031.0861E+031.2794E+081.4104E+084.2027E+079.0808E+07
RIME1.3430E+036.4494E+011.5445E+071.2556E+072.0367E+052.1819E+05
ALO1.6072E+032.9033E+022.9105E+072.7789E+071.1196E+054.8421E+04
WOA1.1173E+044.3832E+035.2975E+083.5710E+081.2064E+071.0695E+07
STOA2.9989E+039.8445E+027.2588E+085.5008E+081.6408E+081.3323E+08
DO1.2507E+035.0890E+011.0984E+076.2977E+067.8362E+043.7814E+04
BKA1.4823E+032.7425E+023.4050E+081.4271E+091.4742E+085.6204E+08
F14F15F16
AverStdAverStdAverStd
IBKA9.7122E+038.5039E+037.5731E+037.8311E+032.6968E+033.2905E+02
MVO3.4510E+042.7848E+046.9820E+045.2125E+042.8851E+032.6801E+02
SCA9.2015E+057.0195E+056.0567E+074.9517E+074.2423E+033.1656E+02
GWO5.3408E+056.3622E+051.1281E+061.8568E+062.6205E+033.4678E+02
RIME9.1925E+045.6751E+041.7103E+041.1875E+042.6640E+033.4723E+02
ALO3.3436E+053.7126E+055.0703E+044.6511E+043.1785E+033.0629E+02
WOA2.5124E+062.1772E+066.4599E+069.1716E+064.4090E+034.8521E+02
STOA8.1628E+058.1028E+052.9379E+073.0940E+073.1430E+033.3971E+02
DO1.0419E+051.2960E+056.1593E+044.4668E+042.8216E+033.0688E+02
BKA5.2316E+042.0990E+052.5011E+051.1217E+063.0542E+033.4306E+02
F17F18F19
AverStdAverStdAverStd
IBKA2.1869E+031.7382E+021.7548E+051.8114E+052.4472E+047.1039E+04
MVO2.2538E+032.2586E+029.6241E+056.5250E+052.5265E+062.2711E+06
SCA2.7904E+031.8499E+021.4122E+075.5519E+069.2923E+065.1945E+07
GWO2.0888E+031.4027E+022.3394E+064.0224E+062.8350E+067.5725E+06
RIME2.1895E+031.9572E+021.4803E+061.3970E+062.2062E+041.8394E+04
ALO2.5163E+032.4436E+021.3267E+061.1844E+064.8498E+064.1078E+06
WOA2.7627E+033.4280E+029.3162E+067.5396E+062.2632E+071.7285E+07
STOA2.4367E+032.9294E+024.5471E+066.1653E+061.9592E+072.9041E+07
DO2.2896E+032.5552E+021.3907E+061.7331E+061.6694E+051.6972E+05
BKA2.3278E+032.4580E+028.7398E+053.0872E+064.1854E+058.5443E+05
F20F21F22
AverStdAverStdAverStd
IBKA2.5150E+032.1790E+022.4553E+035.3725E+014.1723E+032.1115E+03
MVO2.5910E+032.3265E+022.4087E+032.4179E+015.6648E+031.5282E+03
SCA2.9201E+031.3859E+022.6055E+033.3046E+019.7246E+031.6830E+03
GWO2.4847E+031.7324E+022.4064E+032.3014E+015.1814E+032.1510E+03
RIME2.6049E+032.2393E+022.4114E+033.1850E+014.9643E+031.8501E+03
ALO2.7646E+032.0719E+022.4579E+033.6581E+015.4810E+032.1202E+03
WOA2.9553E+032.5455E+022.6568E+036.8036E+018.2539E+031.9474E+03
STOA2.8332E+031.9580E+022.5038E+032.4933E+018.7053E+031.2683E+03
DO2.7178E+031.9581E+022.4670E+034.2004E+016.0385E+031.7841E+03
BKA2.6231E+032.0607E+022.5513E+034.9758E+016.9847E+031.6217E+03
F23F24F25
AverStdAverStdAverStd
IBKA2.8503E+038.7286E+012.9399E+033.6752E+012.9006E+032.3501E+01
MVO2.7769E+033.9708E+013.0205E+039.5203E+012.9148E+032.4447E+01
SCA3.0867E+034.9441E+013.2546E+033.7694E+013.6973E+032.3620E+02
GWO2.7832E+034.8088E+012.9813E+036.6322E+013.0294E+038.5565E+01
RIME2.7931E+035.0184E+012.9554E+033.7237E+012.9272E+033.2391E+01
ALO2.8834E+036.4267E+013.0330E+036.9358E+012.9724E+033.0384E+01
WOA3.1521E+031.3010E+023.2817E+031.3662E+023.1981E+038.5220E+01
STOA2.8996E+034.5481E+013.0322E+033.2068E+013.2076E+031.3902E+02
DO2.9202E+038.0655E+013.0863E+035.7420E+012.9127E+031.8573E+01
BKA3.1457E+031.9363E+023.2771E+031.1239E+023.1276E+032.1325E+02
F26F27F28
AverStdAverStdAverStd
IBKA5.6976E+031.6824E+033.2511E+032.6541E+013.2634E+034.3976E+01
MVO4.9731E+036.8718E+023.2362E+032.9722E+013.2662E+033.2919E+01
SCA7.9205E+034.1810E+023.5862E+039.2203E+014.5569E+033.3449E+02
GWO5.0397E+034.9222E+023.2744E+034.0910E+013.5002E+031.5594E+02
RIME4.7702E+037.7115E+023.2455E+031.7650E+013.2993E+034.2249E+01
ALO5.6808E+039.2647E+023.4391E+031.1451E+023.3617E+033.9583E+01
WOA8.6925E+031.2937E+033.5157E+031.6520E+023.8686E+032.4676E+02
STOA6.1952E+034.3746E+023.3364E+035.8234E+015.1750E+031.3275E+03
DO5.9975E+031.0339E+033.3035E+035.2064E+013.2731E+032.9028E+01
BKA7.9174E+031.4532E+033.4438E+031.3972E+023.9948E+039.9809E+02
F29F30
AverStdAverStd
IBKA4.0294E+032.1901E+025.2818E+051.4079E+06
MVO4.0716E+032.4553E+025.1215E+063.5195E+06
SCA5.2651E+033.0705E+021.8984E+085.5931E+07
GWO3.8933E+031.6979E+021.1620E+071.1248E+07
RIME4.0837E+032.2735E+026.2485E+055.5801E+05
ALO4.8292E+034.1913E+021.0533E+077.3270E+06
WOA5.3572E+035.6067E+027.7638E+078.0022E+07
STOA4.6813E+033.2943E+025.3184E+073.9854E+07
DO4.1918E+032.7007E+021.7537E+069.3124E+05
BKA4.7009E+033.7168E+023.5877E+079.0955E+07
Table 6. p-values obtained from the Wilcoxon rank-sum test (F1–F30).
Table 6. p-values obtained from the Wilcoxon rank-sum test (F1–F30).
FunIBKA
vs.
MVO
IBKA
vs.
SCA
IBKA
vs.
GWO
IBKA
vs.
RIME
IBKA
vs.
ALO
IBKA
vs.
WOA
IBKA
vs.
STOA
IBKA
vs.
DO
IBKA
vs.
BKA
F13.01E−113.01E−113.01E−113.01E−115.87E−043.01E−113.01E−113.01E−113.01E−11
F33.33E−113.01E−113.01E−113.01E−113.01E−113.01E−113.01E−113.01E−113.01E−11
F43.77E−043.01E−115.18E−079.70E−019.06E−033.01E−113.33E−113.71E−011.46E−10
F58.66E−053.01E−113.14E−022.38E−045.99E−014.07E−118.19E−071.76E−028.48E−09
F68.76E−017.38E−104.18E−091.01E−085.09E−064.07E−116.28E−062.15E−033.49E−09
F73.49E−092.22E−091.49E−061.07E−093.56E−041.20E−104.11E−064.20E−011.15E−07
F81.45E−013.01E−118.31E−031.37E−012.32E−023.01E−113.35E−083.87E−013.52E−07
F97.65E−051.20E−103.18E−044.42E−031.18E−014.97E−116.04E−072.59E−059.51E−06
F107.28E−013.01E−111.76E−013.04E−011.83E−028.99E−115.49E−119.62E−024.22E−03
F111.10E−063.01E−113.01E−111.24E−046.12E−103.01E−113.01E−112.92E−023.35E−08
F121.58E−043.01E−112.37E−106.76E−052.57E−073.01E−113.01E−113.83E−056.52E−07
F132.78E−073.01E−114.61E−101.38E−063.35E−083.01E−113.01E−116.52E−075.46E−09
F143.25E−073.01E−116.69E−116.72E−108.15E−113.01E−113.01E−112.37E−101.95E−01
F151.07E−093.01E−114.19E−102.53E−043.96E−083.01E−113.33E−117.69E−082.87E−06
F166.52E−015.49E−118.23E−027.95E−015.09E−064.07E−113.25E−071.76E−022.26E−03
F176.30E−011.10E−083.51E−023.25E−013.00E−041.60E−065.08E−031.80E−011.22E−02
F185.96E−093.01E−113.49E−098.89E−104.31E−084.97E−113.33E−113.49E−091.45E−01
F193.01E−113.01E−111.07E−091.17E−033.01E−113.01E−113.01E−112.66E−094.19E−10
F208.41E−011.54E−097.48E−026.52E−016.35E−051.10E−082.12E−041.91E−027.39E−01
F216.66E−033.15E−104.63E−035.82E−039.70E−018.15E−119.79E−051.76E−025.09E−08
F226.14E−024.18E−091.85E−011.95E−015.55E−021.06E−073.15E−104.84E−026.09E−03
F239.51E−061.46E−102.49E−031.32E−021.27E−024.50E−114.08E−051.04E−043.15E−10
F243.15E−053.68E−112.23E−021.37E−032.28E−011.61E−101.02E−016.54E−042.66E−09
F258.23E−023.01E−114.97E−118.18E−012.13E−053.01E−113.01E−113.32E−013.68E−11
F262.17E−019.75E−101.29E−012.28E−019.88E−032.22E−091.07E−022.32E−029.06E−08
F274.05E−023.82E−103.18E−038.88E−011.41E−094.19E−106.52E−077.69E−043.49E−09
F286.20E−013.01E−116.06E−113.67E−032.37E−073.01E−113.01E−111.80E−016.69E−11
F292.51E−013.33E−111.85E−037.06E−019.53E−075.49E−119.06E−087.28E−016.52E−07
F303.01E−113.01E−113.68E−114.99E−093.33E−113.01E−113.01E−111.77E−103.68E−11
Table 7. VMD Parameters.
Table 7. VMD Parameters.
Mode
Number K
Penalty Factor α Noise
Tolerance τ
Convergence Tolerance tolDC Component
8180001E−70
Table 9. Comparison of various models in the domestic aviation sector.
Table 9. Comparison of various models in the domestic aviation sector.
Optimization
Algorithm
Optimal Solution from the Optimization AlgorithmEvaluation Indicators
unitlrmpMAE (MM·T−1)RMSE (MM·T−1)MAPE
Grid Search3000.0090500.00290.00391.96%
Random Search810.00941570.00300.00412.09%
Bayesian730.00613000.00280.00371.93%
MVO2830.00911270.00270.00361.88%
SCA1850.00582990.00260.00351.83%
GWO1510.00242980.00280.00371.90%
RIME1580.00762990.00250.00351.74%
ALO2980.00463000.00270.00361.86%
WOA1510.00992960.00280.00401.90%
STOA920.00343000.00270.00351.84%
DO800.00462990.00270.00351.86%
BKA1170.00652980.00260.00371.83%
IBKA1260.00822960.00230.00331.59%
Table 10. Comparison of various models in the international aviation sector.
Table 10. Comparison of various models in the international aviation sector.
Optimization
Algorithm
Optimal Solution from the Optimization AlgorithmEvaluation Indicators
unitlrmpMAE (MM·T−1)RMSE (MM·T−1)MAPE
Grid Search2240.00492630.00190.00226.03%
Random Search2590.00312940.00160.00195.52%
Bayesian1090.00272370.00140.00175.02%
MVO550.00152560.00080.00102.86%
SCA2030.0015670.00110.00133.90%
GWO890.00292680.00100.00123.41%
RIME670.00102970.00080.00092.71%
ALO540.00192750.00090.00113.31%
WOA510.00332780.00110.00133.86%
STOA560.00232680.00090.00113.21%
DO2590.00302940.00120.00153.95%
BKA550.00122720.00080.00102.81%
IBKA1240.00112210.00070.00082.41%
Table 11. Comparison of various models in the industry sector.
Table 11. Comparison of various models in the industry sector.
Optimization
Algorithm
Optimal Solution from the Optimization AlgorithmEvaluation Indicators
unitlrmpMAE (MM·T−1)RMSE (MM·T−1)MAPE
Grid Search3000.0060500.10750.14280.89%
Random Search820.00113000.09360.12600.79%
Bayesian510.00492400.09110.12270.76%
MVO2080.00352820.08560.11940.72%
SCA2640.00991270.09030.12230.75%
GWO560.00623000.08570.12080.72%
RIME1490.00452990.08280.11810.71%
ALO1870.00162810.08310.11260.70%
WOA1200.00322560.08450.11320.72%
STOA770.00242650.08960.11750.72%
DO1800.00171180.08650.11370.75%
BKA1250.00992990.08390.11480.71%
IBKA3000.01003000.08160.11170.68%
Table 12. Comparison of various models in the resident sector.
Table 12. Comparison of various models in the resident sector.
Optimization
Algorithm
Optimal Solution from the Optimization AlgorithmEvaluation Indicators
unitlrmpMAE (MM·T−1)RMSE (MM·T−1)MAPE
Grid Search3000.0040500.07490.07653.87%
Random Search2910.00691350.06430.08303.24%
Bayesian2620.0038670.05420.76532.41%
MVO2790.00962270.05340.07412.15%
SCA2480.0024580.07360.10143.22%
GWO1430.00452990.04630.06352.01%
RIME2270.00312820.04220.05801.85%
ALO2430.00113000.04520.06422.04%
WOA540.0096720.05950.05942.47%
STOA560.00322430.04690.06502.04%
DO520.00492680.04560.06381.91%
BKA3000.01003000.05520.07392.30%
IBKA1740.00352830.03850.05501.76%
Table 13. Comparison of different models in four sectors.
Table 13. Comparison of different models in four sectors.
SectorModelMAERMSEMAPERank
Aver (30 Times)
Aviation (domestic aviation)ARMA0.00480.00653.1701%6
ARIMA0.00390.00502.3482%5
SVM0.00320.00472.2108%4
ANN0.00310.00461.9845%2
LSTM0.00290.00402.0343%3
BiLSTM0.00250.00341.5995%1
Aviation (international aviation)ARMA0.00150.00185.2502%6
ARIMA0.00090.00103.0177%4
SVM0.00080.00092.7802%3
ANN0.00110.00133.7620%5
LSTM0.00080.00092.6425%2
BiLSTM0.00070.00082.4182%1
IndustryARMA0.21030.28141.7709%6
ARIMA0.14230.22861.2075%5
SVM0.09010.13460.7571%3
ANN0.12970.19941.1108%4
LSTM0.08650.11980.7220%2
BiLSTM0.08330.11490.6979%1
ResidentARMA0.09770.13234.1540%6
ARIMA0.07890.10643.3183%5
SVM0.05590.07572.3712%4
ANN0.05540.07222.2670%3
LSTM0.04620.06331.9967%2
BiLSTM0.03920.05541.7638%1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Li, S.; Liu, H.; Guo, J. Carbon Dioxide Emission Forecasting Using BiLSTM Network Based on Variational Mode Decomposition and Improved Black-Winged Kite Algorithm. Mathematics 2025, 13, 1895. https://doi.org/10.3390/math13111895

AMA Style

Yang Y, Li S, Liu H, Guo J. Carbon Dioxide Emission Forecasting Using BiLSTM Network Based on Variational Mode Decomposition and Improved Black-Winged Kite Algorithm. Mathematics. 2025; 13(11):1895. https://doi.org/10.3390/math13111895

Chicago/Turabian Style

Yang, Yueqiao, Shichuang Li, Haijun Liu, and Jidong Guo. 2025. "Carbon Dioxide Emission Forecasting Using BiLSTM Network Based on Variational Mode Decomposition and Improved Black-Winged Kite Algorithm" Mathematics 13, no. 11: 1895. https://doi.org/10.3390/math13111895

APA Style

Yang, Y., Li, S., Liu, H., & Guo, J. (2025). Carbon Dioxide Emission Forecasting Using BiLSTM Network Based on Variational Mode Decomposition and Improved Black-Winged Kite Algorithm. Mathematics, 13(11), 1895. https://doi.org/10.3390/math13111895

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop