Next Article in Journal
On Solution Set Associated with a Class of Multiple Objective Control Models
Previous Article in Journal
Binary Secretary Bird Optimization Algorithm for the Set Covering Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rainfall Forecasting Using a BiLSTM Model Optimized by an Improved Whale Migration Algorithm and Variational Mode Decomposition

1
School of Emergency Management, Institute of Disaster Prevention, Langfang 065201, China
2
School of Information Management, Institute of Disaster Prevention, Langfang 065201, China
3
China International Engineering Consulting Corporation, Ecological Technical Research Institute (Beijing) Co., Ltd., Beijing 100048, China
4
School of Urban Economics and Management, Beijing University of Civil Engineering and Architecture, Beijing 102616, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(15), 2483; https://doi.org/10.3390/math13152483
Submission received: 6 July 2025 / Revised: 26 July 2025 / Accepted: 29 July 2025 / Published: 1 August 2025

Abstract

The highly stochastic nature of rainfall presents significant challenges for the accurate prediction of its time series. To enhance the prediction performance of non-stationary rainfall time series, this study proposes a hybrid deep learning forecasting framework—VMD-IWMA-BiLSTM—that integrates Variational Mode Decomposition (VMD), Improved Whale Migration Algorithm (IWMA), and Bidirectional Long Short-Term Memory network (BiLSTM). Firstly, VMD is employed to decompose the original rainfall series into multiple modes, extracting Intrinsic Mode Functions (IMFs) with more stable frequency characteristics. Secondly, IWMA is utilized to globally optimize multiple hyperparameters of the BiLSTM model, enhancing its ability to capture complex nonlinear relationships and long-term dependencies. Finally, experimental validation is conducted using daily rainfall data from 2020 to 2024 at the Xinzheng National Meteorological Observatory. The results demonstrate that the proposed framework outperforms traditional models such as LSTM, ARIMA, SVM, and LSSVM in terms of prediction accuracy. This research provides new insights and effective technical pathways for improving rainfall time series prediction accuracy and addressing the challenges posed by high randomness.

1. Introduction

Multiple intertwined factors, including climate change, topographical heterogeneity, and seasonal variability, influence rainfall forecasting. As a result, rainfall time series often exhibit significant non-stationary characteristics, posing a serious challenge to improving the accuracy of prediction models. To address this issue, existing preprocessing techniques can generally be categorized into two major types: traditional statistical methods and emerging signal decomposition approaches. The traditional methods include differencing [1], logarithmic transformation [2], and exponential smoothing [3]. In contrast, modern signal decomposition techniques, such as Empirical Mode Decomposition (EMD) [4], Ensemble Empirical Mode Decomposition (EEMD) [5], and Variational Mode Decomposition (VMD) [6], have recently gained traction for their ability to transform non-stationary signals into components with more stable characteristics. Among them, VMD has demonstrated significant advantages in both decomposition stability and precision by overcoming the mode mixing problem inherent in EMD through an optimized variational framework. While VMD has been widely applied in fields such as seismic signal processing and energy consumption analysis [7,8], its application in rainfall forecasting remains in the early exploratory stage. Integrating VMD into the preprocessing pipeline for rainfall prediction holds promise for improving forecasting accuracy.
Various methods have been developed for rainfall prediction. For instance, Masum et al. employed an ARIMA model to forecast rainfall and temperature in Chittagong, Bangladesh, from 1953 to 2070 [9], while Karthikeyan et al. utilized an ARMA model to predict non-stationary rainfall data [10]. However, both ARIMA and ARMA are traditional statistical approaches based on linear assumptions, which limit their ability to effectively capture the inherent nonlinear patterns in rainfall data. As a result, these models often suffer from insufficient prediction accuracy when applied to complex hydrological time series.
With the rise of machine learning, the limitations of traditional linear models have gradually been overcome. For example, Hussein et al. applied Support Vector Machines (SVM) for regional rainfall prediction by classifying large-scale precipitation maps [11]. Ma et al. proposed an XGBoost-based flash flood risk assessment method by integrating two input strategies and validating its performance using Least Squares Support Vector Machine (LSSVM), demonstrating its effectiveness [12]. Reddy et al. combined Singular Spectrum Analysis (SSA) as a data preprocessing technique with supervised learning models such as Least Squares Support Vector Regression (LSSVR) and Random Forest (RF) to enhance rainfall prediction reliability [13]. However, these machine learning approaches often fail to capture the dynamic characteristics inherent in time series data adequately.
In recent years, deep learning techniques have garnered significant attention due to their powerful capabilities in sequence modeling. Ma et al. employed a multimodal recurrent neural network (MM-RNN) for rainfall prediction [14]. As a classical time series model, RNN can capture temporal dependencies within sequences; however, it is prone to gradient vanishing or exploding in long-term predictions, leading to the “forgetting” of critical earlier information. To address this, Poornima et al. utilized the Long Short-Term Memory (LSTM) model for rainfall estimation [15], and Salehin et al. applied LSTM to assess rainfall and improve crop yield forecasting [16]. By introducing input, forget, and output gates, LSTM effectively regulates the flow and retention of information, thereby mitigating the long-term dependency issues of RNNs and demonstrating superior temporal modeling performance in rainfall forecasting tasks. Nevertheless, conventional LSTM networks rely solely on unidirectional information flow, which limits their ability to fully capture global sequence trends and variation patterns.
To address these limitations, Xie et al. employed a Bidirectional Long Short-Term Memory (BiLSTM) model to impute temperature data [17]. At the same time, Siami-Namini et al. conducted a comparative analysis of LSTM and BiLSTM models in time series forecasting tasks [18]. BiLSTM integrates both forward and backward LSTM units, allowing it to capture information from both past and future contexts within the sequence. This dual-directional structure significantly enhances the model’s ability to comprehend and retain complex sequence patterns. Empirical studies have demonstrated that BiLSTM outperforms conventional LSTM models across various time series prediction scenarios [19]. Accordingly, this study adopts BiLSTM as the baseline model for rainfall forecasting.
In BiLSTM models, hyperparameters have a substantial impact on predictive performance. Studies have shown that the influence of hyperparameters on deep learning models often exceeds that of the model architecture itself [20]. The core objective of hyperparameter optimization is to define a search space comprising the tunable hyperparameters and their value ranges, and then use an optimization algorithm to explore this space in search of the optimal configuration [21]. However, due to the large number of hyperparameters and the broad range of possible values for each, the search space contains an enormous number of parameter combinations that must be evaluated individually to identify the best solution [22]. This renders hyperparameter optimization a challenging task in deep learning applications. In the context of BiLSTM-based rainfall forecasting, hyperparameters can generally be categorized into two groups: (1) structural hyperparameters, such as the number of units per BiLSTM layer and the number of stacked BiLSTM layers, and (2) training-related hyperparameters, such as the learning rate (i.e., the step size for parameter updates during training), dropout rate (which determines the proportion of neurons randomly dropped during training), maximum number of epochs, optimization algorithm, and loss function [23,24,25]. Due to the vast number of hyperparameters and the size of their respective search spaces, optimizing all of them simultaneously is computationally expensive. Therefore, this study focuses on optimizing the three most critical hyperparameters that have the most significant impact on BiLSTM performance.
In traditional studies, hyperparameters are often manually configured based on the researcher’s experience, a highly subjective approach that makes it challenging to ensure globally optimal solutions [26]. Against this backdrop, automated hyperparameter optimization methods have gradually become a research focus. Currently, the mainstream approaches include Grid Search [27], Random Search [28], and Bayesian Optimization [29]. Grid Search is straightforward and intuitive but suffers from high computational costs. Random Search reduces computational burden but lacks stability and consistency. Bayesian Optimization, on the other hand, constructs a probabilistic model to guide the search process effectively. However, it is prone to getting trapped in local optima in high-dimensional and complex search spaces.
In recent years, Swarm Intelligence (SI) optimization algorithms have garnered attention due to their robust global search capabilities and strong performance. They have been successfully applied in domains such as wind speed forecasting, image recognition, and medical diagnosis [30,31,32]. However, their application in rainfall prediction remains relatively limited. SI algorithms are a class of heuristic optimization methods inspired by collective behaviors in nature—such as bird flocking and fish schooling—and are designed to find global or near-global solutions through collaborative search strategies [33]. Currently, mainstream SI algorithms include the Harris hawks optimization (HHO), Sine cosine algorithm (SCA), Grey wolf optimizer (GWO), Elk herd optimization (EHO), Butterfly optimization algorithm (BOA), Whale optimization algorithm (WOA), Particle swarm optimization (PSO), and Differential evolution (DE) [34,35,36,37,38,39,40,41]. Although a wide range of SI algorithms has been developed, the “No Free Lunch” theorem suggests that no single optimization algorithm is universally superior for all types of problems [42]. In other words, an algorithm that performs well on one class of problems may not perform as well on others [43]. Therefore, enhancing existing algorithms to improve their effectiveness in solving specific types of problems remains a valuable direction for research.
The Whale Migration Algorithm (WMA), proposed by Ghasemi et al. in 2025 [44], is inspired by the cooperative and migratory behavior of humpback whales. It demonstrates strong global search capability, fast convergence speed, and excellent adaptability and robustness. Previous studies have shown that WMA outperforms several mainstream metaheuristic algorithms, including HHO, SCA, GWO, EHO, BOA, WOA, PSO, and DE, across various benchmark test functions. Despite its competitive performance, WMA tends to encounter issues such as “clustering” or “biased distribution” during the initial population generation phase, which may cause the search process to fall into local optima. To enhance population diversity and global exploration ability, this study incorporates a logistic map-based chaotic initialization strategy [45], which exhibits ergodic and stochastic properties. Furthermore, the original WMA lacks an explicit elitism mechanism, which may result in the loss of the current best solution during iterative updates. To improve convergence stability and solution accuracy, an elitism preservation strategy [46] is also introduced. Based on these enhancements, we propose an Improved Whale Migration Algorithm (IWMA). To validate the effectiveness of IWMA, comprehensive experiments were conducted on a suite of standard benchmark test functions, and the performance was compared against nine representative optimization algorithms, including HHO, SCA, GWO, EHO, BOA, WOA, PSO, DE, and the original WMA.
The paper proposes a novel hybrid deep learning prediction framework—VMD-IWMA-BiLSTM—which comprehensively utilizes Variational Mode Decomposition (VMD), Improved Whale Migration Algorithm (IWMA), and Bidirectional Long Short-Term Memory Network (BiLSTM) to achieve accurate modeling of rainfall data.

2. Materials and Methods

This study proposes a novel VMD-IWMA-BiLSTM framework for rainfall forecasting and optimization. The framework consists of three core modules. First, Variational Mode Decomposition (VMD) is introduced as a signal decomposition technique to handle non-stationary data, which serves as a preprocessing step for the rainfall series. Second, the Bidirectional Long Short-Term Memory (BiLSTM) network is employed as the central deep learning model to perform the forecasting task. Third, the Improved Whale Migration Algorithm (IWMA) is utilized as a swarm intelligence-based optimization tool to optimize three key hyperparameters of the BiLSTM network. The following sections provide a detailed introduction to the signal decomposition method, the forecasting model, and the optimization algorithm, respectively.

2.1. VMD

Variational Mode Decomposition (VMD) is a novel signal decomposition method proposed by Dragomiretskiy and Zosso in 2014 [6]. In contrast to Empirical Mode Decomposition (EMD) and its variants, Variational Mode Decomposition (VMD) eliminates the dependence on local extrema and envelope interpolation by formulating a constrained optimization problem in the frequency domain. Through a variational optimization framework, VMD enables robust and adaptive signal decomposition. This approach provides enhanced parameter controllability and algorithmic stability, effectively overcoming the mode mixing issue frequently encountered in conventional decomposition methods.
First, a variational model is constructed by assuming that the rainfall data are decomposed into K intrinsic mode components. The model imposes a constraint that the sum of all components equals the original signal. The corresponding constrained variational formulation is given in Equation (1):
min { u k , ω k } k σ l δ ( t ) + j π t u k ( t ) e - j ω k t 2 2 k u k = f ( t ) n
Here, { u k } is the kth modal component, { ω k } is the central frequency of the kth mode, δ t is the Dirac delta function, and * denotes the convolution operator.
Subsequently, the formulated variational problem is addressed by introducing Lagrange multipliers λ , thereby converting the original constrained problem into an unconstrained variational form. The corresponding augmented Lagrangian expression is presented in Equation (2):
L ( { u k } , { ω k } , λ ) = α k σ l δ ( t ) + j π t u k ( t ) e - j ω k t 2 2 + f ( t ) - k u k ( t ) 2 2 + λ t , f ( t ) - k u k ( t )
where α is the quadratic penalty factor for reducing the interference of Gaussian noise and λ t represents the Lagrange operator.
The modal components and their corresponding center frequencies are obtained through the Alternating Direction Method of Multipliers (ADMM) and Fourier isometric transformation optimization. Among them, the center frequency is constrained within the range [0,+∞), which ensures strict compliance with the requirements of isometric transformation optimization in VMD. The main procedure is as follows: Leveraging the energy-equidistant property of isometric transformation, the VMD optimization problem is mapped onto a uniformly distributed frequency space, thereby ensuring minimal spectral overlap among modes. This improves the accuracy of mode separation by mitigating mode aliasing and ensures that each ω k is confined to the dominant frequency band of the actual signal, avoiding energy leakage and spectral redundancy. To this end, for low-frequency modes ( ω 0 ), the center frequency update function ω k     ω k adopts a “weighted dual-boosting” strategy to penalize non-physical negative frequency values, ensuring energy-concentrated characteristics. Accordingly, when ω   > 0 (non-zero center frequency), the update proceeds along the regular optimization direction, whereas for ω 0 , a reconstruction operation is introduced, as detailed in Equations (3)–(5).
u ^ k n + 1 ( ω ) = f ^ ( ω ) - i k u i ( ω ) + λ ^ ( ω ) / 2 1 + 2 α ( ω - ω k ) 2
ω k n + 1 = 0 ω u ^ k n + 1 ( ω ) 2 d ω 0 u ^ k n + 1 ( ω ) 2 d ω
λ ^ k n + 1 ( ω ) = λ ^ k n ( ω ) + γ f ^ ( ω ) - k u ^ k n + 1 ( ω )
Among these variables, u ^ k n + 1 , u i ω , and f ^ ω represent the Fourier transforms of u k n + 1 t , u i t , and f t , respectively. The constraints are applied until the constraint conditions are satisfied in Equation (6) is met:
k = 1 K u ^ k ( n + 1 ) - u ^ k n 2 2 u ^ k n 2 2 <

2.2. Rainfall Prediction Model: BiLSTM

The BiLSTM network used in this study is a variant of the Long Short-Term Memory (LSTM) model. It introduces three gating mechanisms—forget gate, input gate, and output gate—to regulate the flow of information, thereby enabling effective modeling of time series data. Specifically, the forget gate determines how much of the previous cell state should be discarded, the input gate controls how much of the current input should be added to the cell state, and the output gate decides how much information from the cell state should be passed to the current hidden state. The structure of the LSTM unit is illustrated in Figure 1.
The calculation of LSTM is shown in Equations (7)–(12):
f t = σ ( w f × h t - 1 + u f × x t + b f )
i t = σ ( w i × h t - 1 + u i × x t + b i )
c ˜ t = tanh ( w c × h t - 1 + u c × x t + b c )
c t = f t c t - 1 + i t c ˜ t
o t = σ ( w o × h t - 1 + u o × x t + b o )
h t = o t tanh ( c t )
Among them, represents the forget gate, represents the input gate, represents the output gate, the new state is the additional new information learned at the current moment t, and the current information is how much information is extracted from previous moments plus how much information is extracted from the current moment.
The BiLSTM unit is composed of a forward LSTM and a backward LSTM, which extract temporal features from the input sequence in the forward and reverse time directions, respectively. Compared with the standard LSTM, BiLSTM can capture temporal dependencies more comprehensively through its bidirectional information flow mechanism, thereby significantly enhancing its sequence modeling capability. Therefore, BiLSTM is adopted in this study as the core unit of the prediction model. Its specific structure is shown in Figure 2.

2.3. IWMA

Since the BiLSTM model involves several critical hyperparameters that significantly influence its performance, it is necessary to employ optimization algorithms to automatically determine the optimal parameter combination. To this end, this study introduces an improved Whale Migration Algorithm (IWMA) to optimize the BiLSTM model. Based on the original Whale Migration Algorithm (WMA), IWMA integrates a chaotic initialization strategy using logistic mapping and an elite retention mechanism, thereby enhancing search diversity and convergence stability. The remainder of this section introduces the basic principles of the original WMA and details the improvements incorporated in the proposed IWMA.

2.3.1. The Original WMA

The Whale Migration Algorithm (WMA) was proposed by Ghasemi et al. in 2025 [44]. Inspired by the migratory behavior of humpback whale groups, the algorithm innovatively introduces a “leader–follower” dynamic mechanism to replace the individual hunting-based optimization strategy used in conventional algorithms such as the Whale Optimization Algorithm (WOA). As illustrated in Figure 3, the fundamental mechanism of WMA includes a “move toward the nearest solution” strategy for inexperienced individuals and a “guided exploration” strategy for experienced individuals. During the iteration process, the algorithm determines the update method for each individual based on its experience level, thereby achieving an optimization dynamic that balances “local exploitation” and “global exploration”, enhancing both convergence accuracy and search efficiency. The overall WMA process can be divided into three stages: population initialization, the follower phase for inexperienced individuals, and the exploration phase for experienced individuals. The strategies for each stage are detailed as follows.
  • Population initialization phase
In the initialization phase, WMA first generates an initial population consisting of n whale individuals, where each individual is represented as a D-dimensional solution vector. The initialization is carried out using a uniform random strategy:
W i   =   L   +   r a n d 1 , D     U     L , i   =   1,2 , , N p o p
where L and U denote the lower and upper bounds of the search space, respectively; rand(1, D) represents uniformly distributed random numbers in the interval [0, 1]; and denotes the Hadamard product. This step is intended to provide diverse initial solutions for the algorithm.
2.
Following Phase for Inexperienced Whales
In this stage, two key mechanisms are employed: movement of less-experienced whales (juveniles) toward the nearest solution, and guided movement under the leadership of more experienced individuals. Specifically, in each iteration, the top-performing individuals are selected based on fitness values to form the leader set. The central position of this leader group is defined as follows:
W m e a n   =   1 N L i   =   1 N L W i
where W i denotes the position of the i-th leader individual, and N L represents the number of leader individuals.
For individuals ranked lower in the population (i.e., non-leader individuals), an imitation mechanism is adopted for the update strategy: their positions are updated by imitating the position of the previous whale, the mean position of the leaders W m e a n , and the current best position W B e s t . The update formula is as follows:
W i n e w   = W m e a n   +   r a n d 1 , D     ( W i 1     W i ) +   r a n d 1 , D     W B e s t W m e a n , i   =   N L   +   1 , , N p o p
  • W i n e w denotes the new position of the i-th non-leader individual;
  • rand(1,D) represents a uniformly distributed random number within the interval [0, 1];
  • denotes the Hadamard product;
  • W B e s t is the current best solution in the population;
  • N p o p is the total population size, and N L is the number of leader individuals.
3.
Exploration by Experienced Whales
The top N L whales in the ranking are considered “leaders”. They guide the entire population to explore new regions through perturbation-based wide-range search. The position update is defined as follows:
W i n e w   = W i   +   r 1     L +   r 1     r 2     U     L ,         i   = 1,2 , , N L
  • r 1 , r 2 : random vectors with values in the interval [0, 1];
  • L , U : the lower and upper bounds of the search space.

2.3.2. Proposed IWMA

Although the original Whale Migration Algorithm (WMA) demonstrates notable advantages in terms of global search capability, convergence speed, and robustness, it also presents several limitations. Specifically, the algorithm is prone to issues such as “clustering” or “biased distribution” during the initial population generation phase, which may cause the search process to fall into local optima. Moreover, the original WMA lacks an explicit elitism mechanism, making it possible to lose the current best solutions during iterative updates. To address these issues, we incorporate two enhancement mechanisms into the original WMA. The details of these improvements are presented below.
  • Chaotic Mapping Mechanism
To enhance the population diversity and solution space coverage capability of the Whale Migration Algorithm (WMA) during the early stages of the search process, this study adopts a logistic chaotic sequence for population initialization. Compared to the traditional uniform random initialization, which may result in uneven distribution and premature convergence, chaotic mapping—owing to its properties of ergodicity, sensitivity to initial conditions, and pseudo-randomness—can effectively improve the global exploration ability in the initial phase of the algorithm. In this work, the adopted logistic map is formulated as follows:
x n + 1   =   r     x n 1     x n ,         r   =   4 ,         x n     ( 0,1 )
During the initialization of the population, the chaotic sequence generated by the logistic map is mapped to the search space:
W i   =   L   +   x i     U     L ,         i   =   1,2 , , N
where x i represents the chaotic sequence generated by the logistic map, and L and U denote the lower and upper bounds of the decision variables, respectively. Experimental results show that applying chaotic initialization enables a more uniform population coverage, significantly improving the quality of early-stage searches and providing a better convergence foundation for subsequent iterations.
2.
Elite Retention Mechanism
To further enhance the algorithm’s ability to maintain the global optimum and improve convergence accuracy, this study introduces an elite retention strategy. Specifically, in each iteration, the currently best-performing individual (referred to as the elite) is preserved and used to guide the update of other individuals during the evolutionary process. The detailed implementation is as follows: during each population update, the individual with the best fitness value is embedded into the update formula for non-leader individuals, serving as a guide to direct their evolution. The specific update strategy is given by the following:
W i n e w   =   W m e a n   +   r 1     W i     1     W i   +   r 2     W E l i t e     W m e a n ,         i   >   N L
  • W m e a n : the mean position of the current leader individuals.
  • W E l i t e : the position of the globally optimal individual.
  • r 1 , r 2 [ 0,1 ] D : random vectors within the range [0, 1] in D-dimensional space.
The proposed strategy not only enhances the stability of the algorithm but also significantly improves its local exploitation capability in the middle and later stages of the search. By preventing the optimal solution from being forgotten or replaced by inferior ones, it further accelerates the convergence speed and improves the solution accuracy.
The final proposed IWMA flowchart is illustrated in Figure 4, where the components highlighted in purple represent our improvements.

3. Results and Discussion

3.1. IWMA Performance Verification Experiments

In this section, a series of experiments are conducted to evaluate the optimization performance of the proposed Improved Whale Migration Algorithm (IWMA), focusing on its exploitation capability, exploration capability, and ability to avoid local optima. The experimental design includes both ablation studies and comparative analyses with mainstream metaheuristic algorithms. All experiments are performed under the same hardware environment to ensure a fair comparison. Specifically, the experiments are conducted on a computer equipped with a 12th Gen Intel® Core™ i5-12500H 12-core processor and an NVIDIA GeForce RTX 4060 GPU with 8 GB of VRAM (Manufacturer: Mechanical Revolution, City: Beijing, China, Country: China). To maintain consistency across tests, all algorithms are configured with the same parameters: the maximum number of iterations (T) is set to 500, and the population size (n) is set to 30.

3.1.1. Benchmark Functions

To comprehensively evaluate the optimization performance of the proposed Improved Whale Migration Algorithm (IWMA), we conducted comparative experiments using 29 benchmark functions selected from the CEC2017 test suite [47]. These 29 functions are categorized into four types: F1–F3 are unimodal benchmark functions used to assess the algorithm’s exploitation capability (note that the original F2 was removed by the official committee, resulting in an adjusted range of F1–F3); F4–F10 are multimodal benchmark functions designed to evaluate exploration capability; F11–F20 are hybrid functions used to test adaptability in heterogeneous problem landscapes; and F21–F30 are composition functions intended to evaluate the algorithm’s ability to avoid local optima. “Range” denotes the boundary constraints of the decision variables, and “F(min)” represents the global optimum. The CEC2017 benchmark functions are summarized in Table 1.
In this experimental section, to reduce the impact of randomness, each algorithm was independently executed 30 times on every benchmark test function, and the resulting outcomes were recorded. To systematically evaluate the overall performance of each algorithm across multiple benchmark functions, the Friedman rank test was employed to rank and statistically analyze the optimization results. This method provides a robust statistical basis for comparing the relative performance of multiple algorithms over a range of problems, enabling a comprehensive and objective assessment of their strengths and weaknesses.

3.1.2. Ablation Analysis of the IWMA

Building upon the original Whale Migration Algorithm (WMA), this study introduces a chaotic mapping strategy and an elite retention mechanism to propose an improved variant, namely IWMA. To evaluate the effectiveness of these enhancements, ablation experiments were designed to systematically assess the performance of various WMA variants incorporating different combinations of these mechanisms. Table 2 presents the WMA variants, where “1” indicates the inclusion of a specific mechanism and “0” denotes its exclusion. All algorithmic variants were executed on the 29 benchmark functions from the aforementioned CEC2017 suite. The experiments aim to comprehensively analyze the individual contributions and synergistic effects of each mechanism on the overall performance of the algorithm.
The experimental results are presented in Table 2. Here, Avg denotes the average rank value of each algorithm across the 29 benchmark functions based on the Friedman test. A lower Avg value indicates better overall optimization performance. The column “+/−/=” represents the number of functions on which the proposed IWMA achieves better, worse, or comparable performance relative to other WMA variants. Rank indicates the final comprehensive ranking of each algorithm among all compared variants.
As shown in the results of Table 3, the baseline version of the WMA algorithm performs the worst among all variants. The incorporation of either the chaotic mapping strategy or the elitism strategy can improve the optimization performance to varying degrees. Among all tested algorithms, IWMA, which integrates both strategies, ranks first overall. This indicates that the combination of chaotic initialization and elitism significantly enhances the algorithm’s global search capability and convergence performance, thereby effectively improving its overall optimization ability.

3.1.3. Comparison with Other Algorithms

To further verify the optimization performance of the proposed IWMA algorithm, this study conducted comparative experiments between IWMA, the original WMA algorithm, and eight other representative and widely used metaheuristic algorithms. The selected benchmark algorithms are as follows:
  • Harris hawks optimization (HHO) [38];
  • Sine cosine algorithm (SCA) [39];
  • Grey wolf optimization (GWO) [40];
  • Elk herd optimization (EHO) [41];
  • Butterfly optimization algorithm (BOA) [42];
  • Whale optimization algorithm (WOA) [43];
  • Particle swarm optimization (PSO) [44];
  • Differential evolution (DE) [45].
The parameter settings of the aforementioned comparison algorithm are shown in Table 4:
To ensure fairness in the experiments, all algorithms were executed under the same computational environment, with the maximum number of iterations set to 500 and a population size of 30. To mitigate the impact of randomness on the experimental outcomes, each algorithm was independently run 30 times on each benchmark test function, and the average value (Aver) and standard deviation (Std) of the results were calculated. Table 5 presents the detailed optimization results of all algorithms on the 29 CEC2017 benchmark functions. For each function, the best-performing result is highlighted in bold. In this evaluation, a smaller Aver value indicates better optimization performance on that specific function.
As clearly shown by the Aver values in Table 5, the proposed IWMA algorithm consistently ranks first on the majority of benchmark functions, demonstrating outstanding global search and local exploitation capabilities. To further validate the algorithm’s performance, the Friedman test was employed to statistically rank and evaluate the fitness performance of all algorithms across the benchmark functions. The average rank values obtained from the Friedman test are illustrated in Figure 5, where a lower average rank indicates better overall performance. The results reveal that IWMA achieves the lowest average rank on the CEC2017 benchmark suite, significantly outperforming the other competing algorithms. This strongly confirms that IWMA possesses superior adaptability and optimization performance across various types of optimization problems, including unimodal functions, multimodal functions, hybrid functions, and composite functions.
Moreover, the Wilcoxon rank-sum test was conducted to statistically evaluate the performance differences between IWMA and other competing algorithms. Table 6 reports the p-values obtained from the Wilcoxon test for IWMA versus each comparison algorithm on the 29 CEC2017 benchmark functions. A p-value less than 0.05 indicates that the performance improvement of IWMA is statistically significant compared to the corresponding algorithm. For clarity, p-values greater than 0.05 are highlighted in bold to denote non-significant differences. As shown in Table 6, the majority of p-values are below 0.05, demonstrating that the superiority of IWMA is not only numerical but also statistically significant. These findings provide further evidence of the effectiveness and robustness of the proposed enhancements in IWMA.
To further evaluate the performance of IWMA across different types of optimization problems, this study selected representative functions from the CEC2017 benchmark suite, including a unimodal function (F1), multimodal functions (F5, F7), a hybrid function (F14), and composite functions (F22, F26). The average fitness convergence curves of IWMA and nine other comparison algorithms on these functions are illustrated in Figure 6. As shown in Figure 6, IWMA consistently outperforms the other algorithms in terms of convergence behavior across all function types. On the unimodal function F1, IWMA demonstrates a very fast convergence rate and excellent exploitation ability, rapidly approaching the global optimum and maintaining stability. Although the original WMA also exhibits competitive performance, its final convergence precision is slightly inferior. For multimodal functions F5 and F7, IWMA maintains a clear advantage by effectively avoiding local optima, showing strong global exploration capabilities. Notably, on F7, IWMA continues to improve throughout the iterations and ultimately achieves the lowest fitness value, indicating superior search capability in complex landscapes. In the case of the hybrid function F14, IWMA escapes from high-fitness regions early and performs better than most competing algorithms, demonstrating good adaptability to heterogeneous problems and strong robustness. For the composite functions F22 and F26, which are more structurally complex, IWMA quickly reaches near-optimal values with smooth and stable convergence curves, showing no significant oscillations, which highlights its excellent global optimization ability and convergence stability. In summary, IWMA not only excels in the early convergence stages but also maintains stability and continues approaching the global optimum in later iterations. These results fully validate the effectiveness and superiority of IWMA across a wide range of complex optimization tasks.

3.2. VMD-IWMA-BiLSTM Framework for Rainfall Forecasting

The previous experiments have validated the performance of the proposed algorithm on benchmark functions. In this section, we apply the Improved Whale Migration Algorithm (IWMA) to a real-world optimization task by constructing a rainfall forecasting framework named VMD-IWMA-BiLSTM. In this framework, the Bidirectional Long Short-Term Memory (BiLSTM) network serves as the core deep learning model for rainfall prediction, while IWMA is employed to optimize three key hyperparameters of the BiLSTM model.
In hyperparameter optimization for deep learning models, the search space is often combinatorial and non-convex, making it challenging for traditional methods such as grid search, Bayesian optimization, and random search, which typically suffer from low efficiency and a tendency to get trapped in local optima. The Improved Whale Migration Algorithm (IWMA) proposed in this study is particularly well-suited for such optimization tasks. By integrating a chaotic initialization mechanism and an elite retention strategy into the original WMA, IWMA not only accelerates the convergence during model training but also effectively searches for the global optimum. This, in turn, significantly enhances the predictive performance of deep learning models.

3.2.1. Data Source and Missing Value Handling

The dataset used in this study was obtained from the National Centers for Environmental Information (NCEI), a subsidiary of the National Oceanic and Atmospheric Administration (NOAA) of the United States. The data consist of daily meteorological observations spanning the period from 1929 to 2024. To ensure both timeliness and representativeness in the experiments, this study selects daily rainfall data from the Xinzheng National Meteorological Observatory for the period from 1 January 2020 to 31 December 2024, comprising a total of 1827 valid records. Since the original dataset contained some missing values during the download process, and considering that Variational Mode Decomposition (VMD) requires input data to exhibit a certain degree of stationarity, linear interpolation was employed to fill in the missing values. The resulting rainfall time series after interpolation is illustrated in Figure 7, where the horizontal axis represents the year and the vertical axis denotes the daily rainfall amount in millimeters (mm).

3.2.2. VMD Decomposition

To enhance the predictive performance on non-stationary rainfall series, this study adopts the Variational Mode Decomposition (VMD) method to decompose the original rainfall sequence into five Intrinsic Mode Functions (IMFs). Subsequently, a separate predictive model is constructed for each IMF component, and the final rainfall prediction is obtained by aggregating the outputs of all individual IMF models. The decomposition results are illustrated in Figure 8. This decomposition strategy effectively decouples different frequency components within the original signal, enabling each sub-model to more accurately capture local temporal patterns, thereby improving the overall prediction accuracy. The parameter settings for the VMD decomposition are summarized in Table 7.

3.2.3. Sample Making

In this study, a sliding window approach is employed to construct sample data for short-term forecasting of the non-stationary rainfall series. Specifically, each sample consists of rainfall data from seven consecutive days as input, with the rainfall value on the eighth day as the prediction target, as illustrated in Figure 9. In the figure, green markers indicate the input data points, while red markers denote the prediction targets. A total of 1820 training samples are generated. Among them, 80% are used for training and the remaining 20% are reserved for testing. To optimize the model’s hyperparameters, cross-validation is conducted on the training set, where 20% of the training data is used as a validation subset in each fold. The final model is then evaluated independently on the test set.

3.2.4. Data Normalization Processing

To improve the training efficiency and prediction stability of the model, the original rainfall time series data were standardized prior to modeling. Due to the inconsistency in scale among different feature dimensions, directly feeding the raw data into the model may result in gradient imbalance or training non-convergence. Therefore, the z-score normalization method was adopted to normalize both the input features and the target outputs. The z-score normalization is calculated using the following formula:
x = x μ σ
where x represents the original data, μ is the mean of the training samples, σ is the standard deviation of the training samples, and x is the normalized data. This method transforms the data into a standard normal distribution with a mean of 0 and a standard deviation of 1, which facilitates faster convergence during the training phase and enhances the predictive performance of neural network models.
In practice, the input features and target outputs of the training set are standardized using the mean and standard deviation computed from the training data. For the test set, the same normalization parameters derived from the training set are applied to ensure consistency in data distribution, thereby improving the model’s generalization capability. During the prediction phase, in order to convert the model output back to the original rainfall scale, an inverse normalization process is applied. The corresponding inverse transformation is given by the following:
x = x σ + μ
Here, x denotes the predicted value from the model, and x represents the actual prediction result after inverse normalization. This post-processing step enables the direct application of the forecasted results in real-world engineering scenarios.

3.2.5. VMD-IWMA-BiLSTM Flowchart

When employing the BiLSTM model for rainfall prediction, the following three hyperparameters have the most significant impact on its predictive performance (1) the number of units in each BiLSTM layer (unit); (2) the learning rate for parameter updates during training (learning_rate); and (3) the maximum number of training epochs (max_epochs).
In this study, the improved IWMA is employed to optimize the three aforementioned hyperparameters. Prior to the optimization process, it is necessary to define the upper and lower bounds of the search space for each hyperparameter. The specific value ranges are presented in Table 8.
Subsequently, the IWMA is initialized with a maximum number of iterations set to 200, dimensionality d is 3, and population size n is 30. The BiLSTM model’s loss function—Root Mean Square Error (RMSE)—is employed as the fitness function in the optimization process. The solver for BiLSTM training is configured as AdaGrad. Finally, the IWMA algorithm is utilized to search for the optimal set of hyperparameters for the BiLSTM model. The overall flowchart of the VMD-IWMA-BiLSTM framework is illustrated in Figure 10.

3.2.6. Evaluation Metrics

This section employs four metrics to evaluate model performance: Mean Absolute Error (MAE), Root Mean Square Error (RMSE), Mean Absolute Percentage Error (MAPE), and Pearson Temporal Correlation Coefficient (PCC). Their calculation formulas are as follows:
M A E = 1 n i = 1 n y i y ^ i
R M S E = 1 n i = 1 n y i y ^ i 2
M A P E = 1 n i = 1 n y i y ^ i y i × 100 %
ρ T = t = 1 n ( y t y ¯ ) ( y ^ t y ^ ¯ ) ( y t y ¯ ) 2 ( y ^ t y ^ ¯ ) 2
where n is the number of test samples, is the actual measured value of the sample data, is the predicted value of the sample data, and is the average measured value of the sample data.

3.2.7. Comparison of BiLSTM Optimized by Various Algorithms

In this section, in addition to IWMA, twelve other optimization algorithms were employed to optimize the proposed BiLSTM-based rainfall prediction model, and their optimization performances were compared. These comparison algorithms include three conventional optimization methods—Grid Search, Random Search, and Bayesian Optimization—as well as eight widely used swarm intelligence algorithms, namely HHO, SCA, GWO, EHO, BOA, WOA, PSO, and DE, along with the original WMA algorithm. To minimize the impact of randomness on the experimental results, each set of comparative experiments was independently repeated 30 times.
The experimental dataset was obtained from the Xinzheng National Meteorological Observatory. The results are presented in Table 9, where the best-performing values are highlighted in bold. The parameters unit, lr, and mp represent the optimal hyperparameters obtained by each algorithm, corresponding to the unit number (number of units in the BiLSTM layer), learning rate, and maximum number of training epochs, respectively, as described in Table 8.
Based on the results in Table 9, it is evident that traditional optimization methods such as grid search, random search, and Bayesian optimization perform significantly worse than swarm intelligence algorithms. Among the ten swarm intelligence algorithms tested, the proposed Improved Whale Migration Algorithm (IWMA) achieved the best performance.
Moreover, the experimental results indicate that different optimization algorithms, when applied to tune the hyperparameters of the same BiLSTM prediction model, can significantly improve its performance to varying degrees. This finding suggests that in deep learning applications, the effectiveness of the optimization strategy plays a decisive role in determining the model’s final performance—sometimes even more critical than the choice of the model architecture itself. Therefore, the rational design and application of efficient optimization algorithms are of great importance for enhancing prediction accuracy.

3.2.8. Comparison with Other Models

In the previous section, we analyzed the performance of the BiLSTM model under different optimization algorithms and found that the model optimized by IWMA achieved the best results. To further validate the effectiveness and advantages of this model, this section compares it with five commonly used machine learning and deep learning models, including Support Vector Machine (SVM), Least Squares Support Vector Machine (LSSVM), Gated Recurrent Unit (GRU), and Long Short-Term Memory network (LSTM). In addition, the ARIMA statistical model is introduced as a baseline to comprehensively evaluate the overall performance of the proposed VMD-IWMA-BiLSTM framework in rainfall prediction tasks.
To ensure fairness, all models were optimized using the IWMA algorithm. Each experiment was independently repeated 30 times, and the average values of the evaluation metrics across the 30 runs are presented in Table 10. The table also reports the performance results of each model, with the best results highlighted in bold. The column “Rank” represents the final ranking of each model based on overall performance. As shown in the results, the BiLSTM model significantly outperforms all other comparison models across all evaluation metrics on the Xinzheng National Meteorological Observatory rainfall dataset, confirming its superior performance in the rainfall forecasting task.
Figure 11 visually compares the predicted and actual rainfall values on the test set for each model. The x-axis represents time, while the y-axis denotes rainfall amount. As clearly illustrated in the plots, the proposed BiLSTM model exhibits the best forecasting performance, with the smallest prediction error, further validating its superiority.

4. Conclusions

To improve the accuracy of rainfall forecasting, this study proposes a hybrid prediction framework that integrates Variational Mode Decomposition (VMD), Improved Whale Migration Algorithm (IWMA) for hyperparameter optimization, and a Bidirectional Long Short-Term Memory (BiLSTM) deep learning model. The framework is empirically validated using daily rainfall data from the Xinzheng National Meteorological Observatory for the period 2020–2024.
The contributions of this paper are as follows:
  • To address the nonlinear characteristics of the sample data, this study introduces the Variational Mode Decomposition (VMD) method into the field of rainfall prediction. This approach effectively mitigates the non-stationarity problem of the original sequence, thereby improving modeling accuracy and predictive performance. It provides a novel perspective for modeling complex meteorological time series.
  • This paper proposes a VMD-IWMA-BiLSTM framework for rainfall prediction and systematically verifies the superiority of the Improved Whale Migration Algorithm (IWMA) in hyperparameter optimization for deep learning models, as well as its significant enhancement of BiLSTM’s predictive ability.
  • In comparative experiments with models such as ARIMA, SVM, LSSVM, RNN, and LSTM, the VMD-IWMA-BiLSTM model demonstrated the best performance in both modeling capability and prediction accuracy, significantly outperforming traditional machine learning and deep learning methods. The study further confirms the effectiveness of IWMA in hyperparameter optimization and the applicability of VMD in handling complex, non-stationary time series.
In conclusion, the VMD-IWMA-BiLSTM framework demonstrates excellent performance in rainfall prediction tasks, providing viable technical pathways and methodological paradigms for improving the accuracy of rainfall time series prediction and addressing the challenges of high randomness.

Author Contributions

Conceptualization, L.Z. and T.Z.; data curation, Y.Y., S.L. and B.D.; formal analysis, Y.Y., S.L., L.Z. and X.S.; funding acquisition, Y.Y. and T.Z.; investigation, S.L., X.S. and B.D.; methodology, Y.Y. and S.L.; project administration, X.S., L.Z. and T.Z.; resources, S.L., T.Z. and B.D.; software, S.L. and B.D.; supervision, Y.Y., T.Z. and X.S.; validation, Y.Y., S.L. and L.Z.; visualization, S.L.; writing—original draft, Y.Y., S.L., T.Z. and L.Z.; writing—review and editing, Y.Y., S.L., T.Z., L.Z., X.S. and B.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to express their sincere gratitude to Yueqiao Yang from the Institute of Disaster Prevention Science and Technology for her valuable assistance in language editing and research guidance. Additionally, we would like to thank Liang Zhao for their support and help in data analysis, experimental design, and other aspects. We also wish to acknowledge Ting Zhou for his financial support and valuable assistance in paper revisions. Finally, we extend our thanks to Xiao Shi and junior sister apprentice Boni Du for their dedication and contributions during the paper revision process. We also appreciate all the individuals involved in this research and the peer experts who provided valuable suggestions.

Conflicts of Interest

Authors Ting Zhou and Xiao Shi were employed by the company China International Engineering Consulting Corporation, Ecological Technical Research Institute (Beijing) Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Liker, J.K.; Augustyniak, S.; Duncan, G.J. Panel data and models of change: A comparison of first difference and conventional two-wave models. Soc. Sci. Res. 1985, 14, 80–101. [Google Scholar] [CrossRef]
  2. West, R.M. Best practice in statistics: The use of log transformation. Ann. Clin. Biochem. 2022, 59, 162–165. [Google Scholar] [CrossRef] [PubMed]
  3. De Livera, A.M.; Hyndman, R.J.; Snyder, R.D. Forecasting time series with complex seasonal patterns using exponential smoothing. J. Am. Stat. Assoc. 2011, 106, 1513–1527. [Google Scholar] [CrossRef]
  4. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, E.H.; Zheng, Q.; Tung, C.C.; Liu, H.H. The empirical mode decomposition method and the Hilbert spectrum for non-stationary time series analysis. Proc. Roy. Soc. 1998, 454A, 903–995. [Google Scholar] [CrossRef]
  5. Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
  6. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  7. Banjade, T.P.; Liu, J.; Li, H.; Ma, J. Enhancing earthquake signal based on variational mode decomposition and SG filter. J. Seismol. 2021, 25, 41–54. [Google Scholar] [CrossRef]
  8. Qin, Y.; Zhao, M.; Lin, Q.; Li, X.; Ji, J. Data-driven building energy consumption prediction model based on VMD-SA-DBN. Mathematics 2022, 10, 3058. [Google Scholar] [CrossRef]
  9. Masum, M.H.; Islam, R.; Hossen, M.A.; Akhie, A.A. Time series prediction of rainfall and temperature trend using ARIMA model. J. Sci. Res. 2022, 14, 215–227. [Google Scholar] [CrossRef]
  10. Karthikeyan, L.; Kumar, D.N. Predictability of nonstationary time series using wavelet and EMD based ARMA models. J. Hydrol. 2013, 502, 103–119. [Google Scholar] [CrossRef]
  11. Hussein, E.; Ghaziasgar, M.; Thron, C. Regional rainfall prediction using support vector machine classification of large-scale precipitation maps. In Proceedings of the 2020 IEEE 23rd International Conference on Information Fusion (FUSION), Rustenburg, South Africa, 6–9 July 2020; IEEE: New York, NY, USA; pp. 1–8. [Google Scholar]
  12. Ma, M.; Zhao, G.; He, B.; Li, Q.; Dong, H.; Wang, S.; Wang, Z. XGBoost-based method for flash flood risk assessment. J. Hydrol. 2021, 598, 126382. [Google Scholar] [CrossRef]
  13. Reddy, P.C.S.; Yadala, S.; Goddumarri, S.N. Development of rainfall forecasting model using machine learning with singular spectrum analysis. IIUM Eng. J. 2022, 23, 172–186. [Google Scholar] [CrossRef]
  14. Ma, Z.; Zhang, H.; Liu, J. MM-RNN: A multimodal RNN for precipitation nowcasting. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4101914. [Google Scholar] [CrossRef]
  15. Poornima, S.; Pushpalatha, M. Prediction of rainfall using intensified LSTM based recurrent neural network with weighted linear units. Atmosphere 2019, 10, 668. [Google Scholar] [CrossRef]
  16. Salehin, I.; Talha, I.M.; Hasan, M.M.; Dip, S.T.; Saifuzzaman, M.; Moon, N.N. An artificial intelligence based rainfall prediction using LSTM and neural network. In Proceedings of the 2020 IEEE International Women in Engineering (WIE) Conference on Electrical and Computer Engineering (WIECON-ECE), Bhubaneswar, India, 26–27 December 2020; IEEE: New York, NY, USA; pp. 5–8. [Google Scholar]
  17. Xie, C.; Huang, C.; Zhang, D.; He, W. BiLSTM-I: A deep learning-based long interval gap-filling method for meteorological observation data. Int. J. Environ. Res. Public Health 2021, 18, 10321. [Google Scholar] [CrossRef]
  18. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The performance of LSTM and BiLSTM in forecasting time series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; IEEE: New York, NY, USA; pp. 3285–3292. [Google Scholar]
  19. Latifoğlu, L. A novel combined model for prediction of daily precipitation data using instantaneous frequency feature and bidirectional long short time memory networks. Environ. Sci. Pollut. Res. 2022, 29, 42899–42912. [Google Scholar] [CrossRef]
  20. Khalid, R.; Javaid, N. A survey on hyperparameters optimization algorithms of forecasting models in smart grid. Sustain. Cities Soc. 2020, 61, 102275. [Google Scholar] [CrossRef]
  21. Bergstra, J.; Yamins, D.; Cox, D. Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures. In Proceedings of the International Conference on Machine Learning, PMLR, Atlanta, GA, USA, 17–19 June 2013; pp. 115–123. [Google Scholar]
  22. Probst, P.; Boulesteix, A.L.; Bischl, B. Tunability: Importance of hyperparameters of machine learning algorithms. J. Mach. Learn. Res. 2019, 20, 1934–1965. [Google Scholar]
  23. El-Kenawy, E.S.M.; Abdelhamid, A.A.; Alrowais, F.; Abotaleb, M.; Ibrahim, A.; Khafaga, D.S. Al-Biruni Based Optimization of Rainfall Forecasting in Ethiopia. Comput. Syst. Sci. Eng. 2023, 46, 2885–2899. [Google Scholar] [CrossRef]
  24. Zoremsanga, C.; Hussain, J. Particle swarm optimized deep learning models for rainfall prediction: A case study in Aizawl, Mizoram. IEEE Access 2024, 12, 57172–57184. [Google Scholar] [CrossRef]
  25. Amini, A.; Dolatshahi, M.; Kerachian, R. Effects of automatic hyperparameter tuning on the performance of multi-Variate deep learning-based rainfall nowcasting. Water Resour. Res. 2023, 59, e2022WR032789. [Google Scholar] [CrossRef]
  26. Shawki, N.; Nunez, R.R.; Obeid, I.; Picone, J. On automating hyperparameter optimization for deep learning applications. In Proceedings of the 2021 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), Philadelphia, PA, USA, 4 December 2021; IEEE: New York, NY, USA; pp. 1–7. [Google Scholar]
  27. Belete, D.M.; Huchaiah, M.D. Grid search in hyperparameter optimization of machine learning models for prediction of HIV/AIDS test results. Int. J. Comput. Appl. 2022, 44, 875–886. [Google Scholar] [CrossRef]
  28. Alibrahim, H.; Ludwig, S.A. Hyperparameter optimization: Comparing genetic algorithm against grid search and bayesian optimization. In Proceedings of the 2021 IEEE Congress on Evolutionary Computation (CEC), Online, 28 June–1 July 2021; IEEE: New York, NY, USA; pp. 1551–1559. [Google Scholar]
  29. Du, L.; Gao, R.; Suganthan, P.N.; Wang, D.Z. Bayesian optimization based dynamic ensemble for time series forecasting. Inf. Sci. 2022, 591, 155–175. [Google Scholar] [CrossRef]
  30. Wang, J.; Yang, Z. Ultra-short-term wind speed forecasting using an optimized artificial intelligence algorithm. Renew. Energy 2021, 171, 1418–1435. [Google Scholar] [CrossRef]
  31. Xu, M.; Cao, L.; Lu, D.; Hu, Z.; Yue, Y. Application of swarm intelligence optimization algorithms in image processing: A comprehensive review of analysis, synthesis, and optimization. Biomimetics 2023, 8, 235. [Google Scholar] [CrossRef]
  32. Venkatesh, C.; Prasad, B.V.V.S.; Khan, M.; Babu, J.C.; Dasu, M.V. An automatic diagnostic model for the detection and classification of cardiovascular diseases based on swarm intelligence technique. Heliyon 2024, 10, e25574. [Google Scholar] [CrossRef] [PubMed]
  33. Gaspar, A.; Oliva, D.; Cuevas, E.; Zaldívar, D.; Pérez, M.; Pajares, G. Hyperparameter optimization in a convolutional neural network using metaheuristic algorithms. In Metaheuristics in Machine Learning: Theory and Applications; Springer International Publishing: Cham, Switzerland, 2021; pp. 37–59. [Google Scholar]
  34. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  35. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  36. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  37. Al-Betar, M.A.; Awadallah, M.A.; Braik, M.S.; Makhadmeh, S.; Doush, I.A. Elk herd optimizer: A novel nature-inspired metaheuristic algorithm. Artif. Intell. Rev. 2024, 57, 48. [Google Scholar] [CrossRef]
  38. Arora, S.; Singh, S. Butterfly optimization algorithm: A novel approach for global optimization. Soft Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  39. Seyedali, M.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2015, 95, 51–67. [Google Scholar]
  40. Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
  41. Storn, R.; Price, K. Differential evolution–a simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  42. Joyce, T.; Herrmann, J.M. A review of no free lunch theorems, and their implications for metaheuristic optimisation. In Nature-Inspired Algorithms and Applied Optimization; Springer: Berlin/Heidelberg, Germany, 2017; pp. 27–51. [Google Scholar]
  43. Adam, S.P.; Alexandropoulos, S.A.N.; Pardalos, P.M.; Vrahatis, M.N. No free lunch theorem: A review. In Approximation and Optimization: Algorithms, Complexity and Applications; Springer: Berlin/Heidelberg, Germany, 2019; pp. 57–82. [Google Scholar]
  44. Ghasemi, M.; Deriche, M.; Trojovský, P.; Mansor, Z.; Zare, M.; Trojovská, E.; Abualigah, L.; Ezugwu, A.E.; Mohammadi, S.K. An efficient bio-inspired algorithm based on humpback whale migration for constrained engineering optimization. Results Eng. 2025, 25, 104215. [Google Scholar] [CrossRef]
  45. Varol Altay, E.; Alatas, B. Bird swarm algorithms with chaotic mapping. Artif. Intell. Rev. 2020, 53, 1373–1414. [Google Scholar] [CrossRef]
  46. Wang, F.; Zhou, L.; Ren, H.; Liu, X. Search improvement process-chaotic optimization-particle swarm optimization-elite retention strategy and improved combined cooling-heating-power strategy based two-time scale multi-objective optimization model for stand-alone microgrid operation. Energies 2017, 10, 1936. [Google Scholar] [CrossRef]
  47. Wu, G.; Mallipeddi, R.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2017 Competition on Constrained Real-Parameter Optimization; National University of Defense Technology: Changsha, China; Kyungpook National University: Daegu, Republic of Korea; Nanyang Technological University: Singapore, 2017. [Google Scholar]
  48. Yang, Y.; Li, S.; Liu, H.; Guo, J. Carbon Dioxide Emission Forecasting Using BiLSTM Network Based on Variational Mode Decomposition and Improved Black-Winged Kite Algorithm. Mathematics 2025, 13, 1895. [Google Scholar] [CrossRef]
Figure 1. LSTM state gating structure.
Figure 1. LSTM state gating structure.
Mathematics 13 02483 g001
Figure 2. BiLSTM model structure diagram.
Figure 2. BiLSTM model structure diagram.
Mathematics 13 02483 g002
Figure 3. Workflow of the Whale Migration Algorithm inspired by humpback whale behavior.
Figure 3. Workflow of the Whale Migration Algorithm inspired by humpback whale behavior.
Mathematics 13 02483 g003
Figure 4. Flowchart of IWMA.
Figure 4. Flowchart of IWMA.
Mathematics 13 02483 g004
Figure 5. The Friedman rankings of different algorithms on 29 benchmark functions.
Figure 5. The Friedman rankings of different algorithms on 29 benchmark functions.
Mathematics 13 02483 g005
Figure 6. Fitness curve comparison across algorithms during iterative processes.
Figure 6. Fitness curve comparison across algorithms during iterative processes.
Mathematics 13 02483 g006
Figure 7. Daily rainfall data from 2020 to 2024.
Figure 7. Daily rainfall data from 2020 to 2024.
Mathematics 13 02483 g007
Figure 8. Results of VMD decomposition.
Figure 8. Results of VMD decomposition.
Mathematics 13 02483 g008
Figure 9. Sample making map.
Figure 9. Sample making map.
Mathematics 13 02483 g009
Figure 10. VMD-IWMA-BiLSTM framework.
Figure 10. VMD-IWMA-BiLSTM framework.
Mathematics 13 02483 g010
Figure 11. Error curves of different models.
Figure 11. Error curves of different models.
Mathematics 13 02483 g011
Table 1. Benchmark functions of CEC 2017 [48].
Table 1. Benchmark functions of CEC 2017 [48].
TypeNo.FunctionsFi * = Fi (x*)1
Unimodal Functions1Shifted and Rotated Bent Cigar Function100
3Shifted and Rotated Zakharov Function200
Simple
Multimodal
Functions
4Shifted and Rotated Rosenbrock’s Function300
5Shifted and Rotated Rastrigin’s Function400
6Shifted and Rotated Expanded Scaffer’s F6 Function500
7Shifted and Rotated Lunacek Bi_Rastrigin Function600
8Shifted and Rotated Non-Continuous Rastrigin’s Function700
9Shifted and Rotated Levy Function800
10Shifted and Rotated Schwefel’s Function900
Hybrid
Functions
11Hybrid Function 1 (N = 3)1000
12Hybrid Function 2 (N = 3)1100
13Hybrid Function 3 (N = 3)1200
14Hybrid Function 4 (N = 4)1300
15Hybrid Function 5 (N = 4)1400
16Hybrid Function 5 (N = 4)1500
17Hybrid Function 6 (N = 5)1600
18Hybrid Function 6 (N = 5)1700
19Hybrid Function 6 (N = 5)1800
20Hybrid Function 6 (N = 6)1900
Composition
Functions
21Composition Function 1 (N = 3)2000
22Composition Function 2 (N = 3)2100
23Composition Function 3 (N = 4)2200
24Composition Function 4 (N = 4)2300
25Composition Function 5 (N = 5)2400
26Composition Function 6 (N = 5)2500
27Composition Function 7 (N = 6)2600
28Composition Function 8 (N = 6)2700
29Composition Function 9 (N = 3)2800
30Composition Function 10 (N = 3)2900
Fi *: the optimal solution function valued.
Table 2. Different variants of WMA incorporating two mechanisms.
Table 2. Different variants of WMA incorporating two mechanisms.
ModelCM 1ERM 2
IWMA11
CWMA10
EWMA01
WMA00
1 CM: Chaotic Mapping Strategy. 2 ERM: Elite Retention Strategy.
Table 3. Ablation experiment results.
Table 3. Ablation experiment results.
AlgorithmRank+/−/=Avg
IWMA1~2.3103
CWMA213/12/12.3793
EWMA312/16/12.6207
WMA413/12/12.6897
Table 4. Parameter configurations of comparison algorithms.
Table 4. Parameter configurations of comparison algorithms.
AlgorithmParameterValue
HHOEscape energy (E), Random number (q, r)E ∈ (−1–1); q, r ∈ (0,1)
SCAConvergence parameter spiral factor (a)a = 2
GWOArea vector (a), random vector (r1, r2)A ∈ [0, 2], r1 ∈ [0, 1], r2 ∈ [0, 1]
EHOMale ratio (br)br ϵ (0–1)
BOAPerception factor (c)c ϵ (0.01–0.1)
WOAConvergence factor decay rate (a)a ϵ (2 → 0)
PSOPopulation size of the particle swarmα ϵ (20–1000)
DEScaling factor (F), Crossover probability (CR)F ϵ (0–2], CR ϵ [0–1]
WMAPopulation size (N), Number of leaders (NL)N ϵ (20–100), NL ϵ (10–50)
Table 5. Experimental results on 29 benchmark functions (optimal values are shown in boldface).
Table 5. Experimental results on 29 benchmark functions (optimal values are shown in boldface).
FunF1F3F4
AverStdAverStdAverStd
IWMA3.8335E+034.2917E+034.5676E+041.3383E+044.9064E+022.1045E+01
HHO4.6203E+082.8176E+085.6934E+047.3345E+036.8264E+027.5893E+01
SCA2.0577E+103.1811E+098.7978E+042.0971E+043.5622E+031.2615E+03
GWO2.3986E+091.5824E+096.3014E+041.2302E+047.1221E+023.0209E+02
EHO2.7460E+046.2361E+042.2317E+081.3346E+055.1184E+024.2300E+01
BOA3.6896E+107.5074E+091.4765E+054.8241E+041.2544E+042.6969E+03
WOA6.5862E+092.8032E+092.7092E+056.0351E+041.3645E+033.8047E+02
PSO9.6025E+081.1254E+096.6689E+043.0171E+046.9417E+023.3474E+02
DE1.1514E+085.0998E+071.9832E+052.3148E+045.6293E+022.3962E+01
WMA8.2757E+037.8386E+036.1679E+042.1791E+044.8691E+021.6544E+01
F5F6F7
AverStdAverStdAverStd
IWMA6.1811E+025.3252E+01 6.2289E+021.0344E+018.9391E+026.4040E+01
HHO7.7268E+022.8107E+016.6673E+026.5391E+001.3244E+036.1795E+01
SCA8.2804E+022.6270E+016.6291E+025.7395E+001.2241E+035.3127E+01
GWO6.1629E+023.2591E+016.1262E+026.0461E+008.9891E+026.3787E+01
EHO6.3614E+026.3391E+016.1997E+029.3679E+009.3030E+027.8652E+01
BOA8.4816E+022.8787E+016.8019E+029.3301E+001.2631E+037.9684E+01
WOA8.6193E+028.0098E+016.8101E+028.2975E+001.3248E+038.0341E+01
PSO6.2439E+023.1687E+016.1707E+028.2960E+008.7893E+024.3155E+01
DE7.3265E+021.7138E+016.0717E+021.0704E+009.8986E+021.7707E+01
WMA6.1073E+025.1076E+016.2055E+026.7917E+009.2189E+025.9488E+01
F8F9F10
AverStdAverStdAverStd
IWMA9.2511E+025.0289E+013.1146E+031.4248E+036.6263E+031.4733E+03
HHO9.8562E+022.1213E+018.5077E+031.1114E+036.1629E+037.8103E+02
SCA1.1009E+032.3298E+018.1538E+031.4530E+038.9109E+033.7081E+02
GWO8.9907E+021.9437E+012.6117E+039.8134E+025.7143E+031.5692E+03
EHO9.2112E+025.4542E+014.6601E+032.6534E+038.1383E+031.6801E+03
BOA1.0875E+033.4643E+018.4363E+031.8711E+038.2623E+032.6840E+02
WOA1.0754E+034.6075E+011.1753E+043.9656E+037.6494E+037.5031E+02
PSO9.1193E+022.3463E+013.5110E+032.2083E+034.4821E+037.2567E+02
DE1.0397E+031.4250E+015.1060E+031.0360E+038.1185E+033.6845E+02
WMA9.3006E+026.2443E+013.0402E+031.3284E+037.6698E+034.1130E+02
F11F12F13
AverStdAverStdAverStd
IWMA1.2482E+035.5580E+015.2486E+051.1992E+061.1842E+041.2110E+04
HHO1.5881E+031.8010E+028.8699E+077.5546E+071.0541E+065.9189E+05
SCA4.3086E+031.4446E+032.6397E+096.8462E+081.2108E+098.5282E+08
GWO2.7959E+031.1159E+031.3578E+081.3592E+081.7521E+076.6045E+07
EHO1.2920E+031.4219E+021.5906E+061.7021E+061.6083E+041.5214E+04
BOA1.4092E+049.0656E+038.1566E+093.0805E+095.5045E+093.8426E+09
WOA1.1471E+044.9141E+034.4957E+082.6414E+081.3698E+071.3192E+07
PSO1.2856E+035.3667E+014.3042E+079.9985E+077.2437E+073.7998E+08
DE1.8917E+032.3892E+021.4808E+085.4445E+078.2392E+062.9812E+06
WMA1.2859E+035.2266E+013.1658E+053.2888E+052.0172E+041.8563E+04
F14F15F16
AverStdAverStdAverStd
IWMA2.5566E+042.4109E+045.3260E+034.2708E+032.6455E+033.7316E+02
HHO1.4314E+061.7982E+061.2411E+055.5410E+043.7253E+035.0875E+02
SCA9.4795E+059.1080E+055.5602E+073.2588E+074.0964E+032.2868E+02
GWO6.4772E+056.3478E+052.6830E+066.9853E+062.6477E+032.7702E+02
EHO5.1852E+047.4068E+049.3466E+031.0258E+043.3032E+036.4961E+02
BOA5.5652E+065.7188E+061.0258E+041.0258E+045.1279E+031.0749E+03
WOA2.5985E+062.0213E+064.6108E+064.0688E+064.3018E+036.1214E+02
PSO4.5817E+044.9343E+049.3432E+039.9290E+032.7077E+032.6751E+02
DE4.6123E+053.2696E+051.1444E+067.7283E+053.2785E+031.6600E+02
WMA3.9975E+043.0426E+041.5217E+041.1414E+042.5404E+033.4748E+02
F17F18F19
AverStdAverStdAverStd
IWMA2.2747E+031.9667E+022.1521E+051.9037E+056.9415E+034.1947E+03
HHO2.7611E+033.3641E+025.4493E+065.3706E+062.1051E+061.5549E+06
SCA2.7980E+031.9634E+021.2451E+076.8237E+061.1225E+087.7834E+07
GWO2.0944E+032.0920E+022.5848E+065.2115E+062.4152E+066.9286E+06
EHO2.4811E+033.3588E+025.0061E+062.0602E+066.0849E+035.5869E+03
BOA4.3826E+033.3059E+034.9864E+073.6979E+071.1974E+081.1917E+08
WOA2.8598E+032.5644E+021.9953E+071.4790E+072.8107E+072.5189E+07
PSO2.1635E+031.8680E+021.7294E+063.6252E+068.9578E+041.9737E+05
DE2.3102E+031.2255E+026.1156E+063.9196E+062.8107E+072.8107E+07
WMA2.2430E+032.2677E+023.2512E+053.1613E+051.5112E+041.4898E+04
F20F21F22
AverStdAverStdAverStd
IWMA2.4919E+031.6450E+022.4066E+033.8466E+012.5278E+031.2319E+03
HHO2.7479E+031.8716E+022.6019E+034.9930E+017.2860E+031.6917E+03
SCA2.9492E+031.1744E+022.5943E+032.2758E+019.9041E+031.3108E+03
GWO2.4845E+031.4931E+022.4172E+034.2145E+015.5441E+032.3771E+03
EHO2.8337E+034.7266E+022.4092E+034.8439E+016.4281E+033.6584E+03
BOA2.8521E+031.4929E+022.6599E+033.9958E+017.4068E+032.1326E+03
WOA2.9693E+032.0264E+022.6484E+035.0793E+018.2888E+031.5766E+03
PSO2.5124E+032.4595E+022.4203E+033.0635E+013.7765E+031.7634E+03
DE2.6176E+031.3078E+022.5281E+031.6652E+017.9579E+031.3318E+03
WMA2.4771E+032.0009E+022.4228E+035.8878E+015.0660E+033.4448E+03
F23F24F25
AverStdAverStdAverStd
IWMA2.7823E+032.8817E+012.9334E+034.8514E+012.8938E+031.4374E+01
HHO3.2580E+031.4151E+023.5244E+031.2935E+023.0032E+032.6020E+01
SCA3.0910E+035.1821E+013.2525E+034.7233E+013.6980E+033.1486E+02
GWO2.7746E+034.0570E+012.9816E+037.5686E+013.0228E+035.7992E+01
EHO2.7803E+035.6466E+012.9535E+037.5145E+012.9141E+032.6943E+01
BOA3.4089E+031.7738E+024.1588E+032.9684E+024.8421E+036.2663E+02
WOA3.1917E+031.0472E+023.2791E+031.1495E+023.2267E+036.9736E+01
PSO2.9199E+037.5239E+013.1152E+031.0319E+022.9113E+033.3957E+01
DE2.8671E+031.1840E+013.0541E+031.2370E+012.9637E+032.0632E+01
WMA2.7725E+033.1533E+012.9500E+035.7793E+012.8944E+031.5054E+01
F26F27F28
AverStdAverStdAverStd
IWMA4.9091E+031.1808E+033.2751E+034.7705E+013.2253E+033.1848E+01
HHO8.1650E+039.1076E+023.5248E+031.2926E+023.4832E+038.7642E+01
SCA8.0558E+034.9639E+023.5452E+039.1623E+014.5054E+033.6902E+02
GWO4.9371E+035.2822E+023.2784E+033.4752E+013.5233E+032.5514E+02
EHO5.2198E+037.3582E+023.2000E+035.9215E−053.3000E+036.4242E−05
BOA9.8109E+037.4426E+024.2218E+033.4908E+026.6211E+036.9803E+02
WOA8.5364E+031.0797E+033.4755E+031.3712E+023.8131E+032.5593E+02
PSO5.2878E+031.2512E+033.2925E+035.9154E+013.3032E+031.2242E+02
DE5.9038E+031.1456E+023.2437E+037.9643E+003.3778E+034.3239E+01
WMA5.1406E+034.1734E+023.2840E+034.9737E+013.2314E+032.3790E+01
F29F30
AverStdAverStd
IWMA3.9872E+032.4230E+028.9112E+038.6313E+03
HHO4.9993E+035.0915E+028.6770E+066.1255E+06
SCA5.2185E+033.3745E+022.0164E+088.6977E+07
GWO3.9589E+031.8399E+021.3657E+071.1548E+07
EHO3.8611E+034.3740E+021.5969E+041.0573E+04
BOA8.8982E+038.6935E+037.2774E+085.0417E+08
WOA5.4822E+035.1093E+026.4434E+076.2238E+07
PSO3.8672E+032.5080E+021.3000E+052.7154E+05
DE4.5387E+031.6933E+021.0018E+064.5699E+05
WMA4.0958E+032.5660E+021.6543E+049.0353E+03
Table 6. Wilcoxon rank-sum test-generated p-values (F1–F30).
Table 6. Wilcoxon rank-sum test-generated p-values (F1–F30).
FunIWMA
VS.
HHO
IWMA
VS.
SCA
IWMA
VS.
GWO
IWMA
VS.
EHO
IWMA
VS.
BOA
IWMA
VS.
WOA
IWMA
VS.
PSO
IWMA
VS.
DE
IWMA
VS.
WMA
F13.02E−113.02E−113.02E−112.71E−013.02E−113.02E−113.02E−113.02E−117.62E−01
F31.68E−033.08E−085.56E−043.02E−113.69E−113.02E−115.56E−043.02E−118.50E−02
F43.02E−113.02E−113.69E−116.74E−013.02E−113.02E−113.37E−053.02E−114.55E−01
F51.33E−103.34E−115.40E−014.55E−013.02E−113.02E−119.94E−012.61E−108.77E−02
F63.02E−113.02E−116.01E−089.33E−023.02E−113.02E−111.78E−045.49E−115.55E−02
F73.02E−113.02E−115.37E−021.54E−013.02E−113.02E−114.84E−021.56E−084.97E−02
F81.52E−033.02E−111.19E−013.48E−013.69E−111.33E−101.15E−019.76E−102.92E−02
F93.02E−113.02E−113.51E−029.47E−013.02E−113.02E−119.05E−025.09E−067.06E−01
F109.82E−014.20E−101.76E−026.77E−058.29E−063.15E−028.68E−038.88E−064.21E−02
F114.62E−103.02E−113.69E−115.99E−013.02E−113.02E−112.46E−013.02E−111.62E−01
F123.02E−113.02E−113.02E−114.44E−073.02E−113.02E−111.87E−073.02E−118.19E−01
F133.02E−113.02E−113.02E−112.58E−013.02E−113.02E−111.20E−083.02E−116.10E−03
F141.46E−103.34E−112.39E−081.75E−053.34E−113.34E−112.68E−063.69E−111.38E−02
F153.02E−113.02E−113.02E−119.33E−023.02E−113.02E−111.17E−033.02E−114.86E−03
F164.62E−104.98E−117.98E−024.03E−033.02E−118.16E−112.97E−019.83E−084.83E−01
F173.52E−071.41E−095.26E−047.39E−011.31E−086.01E−086.10E−011.67E−012.28E−01
F183.20E−093.02E−111.07E−071.44E−033.34E−111.78E−105.86E−063.02E−112.71E−01
F193.02E−113.02E−113.02E−119.71E−013.02E−113.02E−114.83E−013.02E−112.61E−02
F202.03E−092.15E−103.71E−011.03E−025.97E−094.99E−094.64E−014.46E−049.94E−01
F211.78E−104.08E−114.38E−011.67E−014.50E−115.49E−112.92E−025.46E−092.17E−01
F223.02E−113.02E−113.02E−112.03E−073.02E−113.02E−113.02E−113.02E−119.03E−04
F233.02E−113.02E−111.96E−015.69E−013.02E−113.69E−116.52E−091.87E−075.55E−02
F243.02E−113.02E−115.75E−029.59E−013.02E−113.34E−114.20E−101.41E−096.95E−01
F253.02E−113.02E−117.39E−112.24E−023.02E−113.02E−117.73E−021.96E−106.41E−01
F263.02E−113.02E−118.42E−014.03E−033.02E−114.62E−102.92E−022.61E−102.81E−02
F274.50E−113.02E−118.07E−013.02E−113.02E−117.38E−102.07E−023.92E−025.75E−02
F283.02E−113.02E−113.02E−113.02E−113.02E−113.02E−114.08E−113.02E−113.37E−04
F298.10E−103.34E−112.81E−026.20E−043.02E−114.08E−111.30E−033.08E−089.71E−01
F303.02E−113.02E−113.02E−112.00E−053.02E−113.02E−114.43E−033.02E−117.17E−01
Table 7. VMD parameters.
Table 7. VMD parameters.
Mode Number (K)Penalty Factor (α)Noise Tolerance (τ)Convergence Tolerance (tol)DC Component
5180001E−70
Table 8. Range of values for the hyperparameter.
Table 8. Range of values for the hyperparameter.
Hyperparameter NameDescriptionLower BoundsUpper Bounds
unitThe number of units in the BiLSTM layer50300
learning_rateParameter update step during model training0.0010.01
max_epochsMaximum number of cycles for model training50300
Table 9. Comparison of various models.
Table 9. Comparison of various models.
Optimization AlgorithmOptimal Solution from the Optimization AlgorithmEvaluation Indicators
UnitlrmpMAERMSEMAPEPCC
Grid Search3000.0080501.51593.420534.57%0.8823
Random Search2250.00312981.12832.446428.59%0.8932
Bayesian2350.00372871.23862.622625.82%0.8964
HHO500.0010500.55961.21789.95%0.9786
SCA2800.0085500.38890.83836.23%0.9821
GWO500.00681720.40500.88996.82%0.9875
EHO530.0011640.47701.07987.99%0.9863
BOA750.00732860.43560.99607.56%0.9833
WOA760.00541140.37080.79566.14%0.9842
PSO500.00641770.39480.85606.90%0.9882
DE710.00403000.37850.83806.81%0.9898
WMA3000.00103000.34270.71095.47%0.9932
IWMA2880.00102580.22510.38963.76%0.9962
Table 10. Comparative analysis of diverse models.
Table 10. Comparative analysis of diverse models.
Detection StationModelMAERMSEMAPEPCCRank
Aver (30 Times)
Xinzheng Testing Station, Zhengzhou CityARIMA1.48983.571537.58%0.89126
SVM1.18722.611126.91%0.94945
LSSVM0.42390.81625.19%0.97443
GRU0.33680.69395.09%0.98692
LSTM0.36320.75535.84%0.98614
BiLSTM0.22510.38963.76%0.99621
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Li, S.; Zhou, T.; Zhao, L.; Shi, X.; Du, B. Rainfall Forecasting Using a BiLSTM Model Optimized by an Improved Whale Migration Algorithm and Variational Mode Decomposition. Mathematics 2025, 13, 2483. https://doi.org/10.3390/math13152483

AMA Style

Yang Y, Li S, Zhou T, Zhao L, Shi X, Du B. Rainfall Forecasting Using a BiLSTM Model Optimized by an Improved Whale Migration Algorithm and Variational Mode Decomposition. Mathematics. 2025; 13(15):2483. https://doi.org/10.3390/math13152483

Chicago/Turabian Style

Yang, Yueqiao, Shichuang Li, Ting Zhou, Liang Zhao, Xiao Shi, and Boni Du. 2025. "Rainfall Forecasting Using a BiLSTM Model Optimized by an Improved Whale Migration Algorithm and Variational Mode Decomposition" Mathematics 13, no. 15: 2483. https://doi.org/10.3390/math13152483

APA Style

Yang, Y., Li, S., Zhou, T., Zhao, L., Shi, X., & Du, B. (2025). Rainfall Forecasting Using a BiLSTM Model Optimized by an Improved Whale Migration Algorithm and Variational Mode Decomposition. Mathematics, 13(15), 2483. https://doi.org/10.3390/math13152483

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop