Next Article in Journal
Heavy Metal Accumulation in Maize and Wheat in Acidic Soil: A Comparative Study
Previous Article in Journal
Optimizing UAV Spraying for Sustainability: Different System Spray Drift Control and Adjuvant Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Combined Forecasting Model Based on a Modified Pelican Optimization Algorithm for Ultra-Short-Term Wind Speed

by
Lei Guo
1,2,
Chang Xu
3,*,
Xin Ai
2,
Xingxing Han
3 and
Feifei Xue
3
1
College of Water Conservancy and Hydropower Engineering, Hohai University, Nanjing 210098, China
2
School of Electrical Engineering, Nanchang Institute of Technology, Nanchang 330099, China
3
School of Renewable Energy, Hohai University, Nanjing 210098, China
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(5), 2081; https://doi.org/10.3390/su17052081
Submission received: 18 January 2025 / Revised: 21 February 2025 / Accepted: 25 February 2025 / Published: 27 February 2025

Abstract

:
Ultra-short-term wind speed forecasting is crucial for ensuring the safe grid integration of wind energy and promoting the efficient utilization and sustainable development of renewable energy sources. However, due to the arbitrary, intermittent, and volatile nature of wind speed, achieving satisfactory forecasts is challenging. This paper proposes a combined forecasting model using a modified pelican optimization algorithm, variational mode decomposition, and long short-term memory. To address issues in the current combination model, such as poor optimization and convergence performance, the pelican optimization algorithm is improved by incorporating tent map-based population initialization, Lévy flight strategy, and classification optimization concepts. Additionally, to obtain the optimal parameter combination, the modified pelican optimization algorithm is used to optimize the parameters of variational mode decomposition and long short-term memory, further enhancing the model’s predictive accuracy and stability. Wind speed data from a wind farm in China are used for prediction, and the proposed combined model is evaluated using six indicators. Compared to the best model among all compared models, the proposed model shows a 10.05% decrease in MAE, 4.62% decrease in RMSE, 17.43% decrease in MAPE, and a 0.22% increase in R2. The results demonstrate that the proposed model has better accuracy and stability, making it effective for wind speed prediction in wind farms.

1. Introduction

Due to the worldwide consensus to attain carbon neutrality, expanding renewable energy sources has increasingly become the norm. Renewable energy like solar, wind, and hydro power produce much lower carbon emissions than traditional fossil fuel-based energy. Among renewable options, wind power stands out as having especially prominent low-carbon attributes since it generates fewer emissions than alternatives. Additionally, wind resources are widely available in large quantities, yielding significant reserves of wind power that can be tapped for energy production [1]. Therefore, wind power has been increasingly valued by countries worldwide. In recent years, the new installed capacity of global wind power has shown a rapid upward trend, rising from 50.7 GW in 2018 to 93 GW in 2021. And as of 2021, the cumulative installed global wind power capacity had arrived at 837 GW. Moreover, according to the forecast of the Global Wind Energy Council, over the next five years (2022–2026), there will be an increase of 557 GW in installed wind power capacity globally, with a compound annual growth rate of 6.6%. By 2026, the new installed capacity of wind power globally will reach 128.8 GW [2]. However, wind energy exhibits intermittency and stochastic fluctuations, where changes in WS (wind speed) directly affect the active power and reactive power of wind farms, resulting in unstable power output. With the rapid development of wind power generation, the issues caused by the unstable output power of wind farms, such as difficulties in power system dispatching and operation, as well as significant wind curtailment, have become increasingly prominent. These directly impact the safe, stable, and economically efficient operation of the power grid [3,4]. The impact of WS prediction on wind power generation is significant, as accurate forecasting allows for the better planning and management of wind farms. By predicting WS with precision, operators can optimize the operation of turbines, adjust the power output accordingly, and anticipate fluctuations in energy production. This leads to the improved efficiency, reduced costs, and increased reliability of wind power generation systems. Additionally, accurate WS prediction enables grid operators to integrate wind energy more effectively into the overall energy mix, contributing to a more sustainable and reliable power supply. Therefore, accurate WS prediction can effectively and economically solve problems caused by the instability of wind power and wind power integration problems.
WS forecasting can be classified from different angles. According to the forecasted time horizon, WS forecasts can be categorized into ultra-short-term, short-term, medium-term, and long-term forecasts [5]. And the ultra-short term is a few seconds to 30 min, the short term is 30 min to 6 h, the medium term is 6 h to a day, and the long term is a day to a few weeks. In addition, WS forecasts can also be categorized into physical models, statistical models, and artificial intelligence models according to different prediction models [6]. The physical model is based on computational fluid dynamics and is built using detailed information such as meteorological information, the topographic features of the location of the wind farm, and surface roughness. Therefore, the NWP (numerical weather prediction) model is a typical physical model. Feng et al. [7] proposed a physics-based prediction method using WS as input data for numerical weather forecasting. Moreover, Liu et al. [8] greatly enhanced the precision of wind power prediction by improving the accuracy of NWP. And Li et al. [9] proposed a correction strategy for NWP to further enhance the precision of wind power prediction. The biggest advantage of a physical model is that it does not require historical operation data from wind farms, and it is suitable for newly built farms or those with incomplete data. However, physical models rely on rich and accurate meteorological, terrain, and other environmental data, and need to simulate the changes in meteorological factors such as WS and direction under the local effects of wind farms. The modeling complexity of physical models is high, and there are many uncertain links. Due to the limitations of model accuracy and simulation ability, systematic deviations may occur. Because of the high computational cost of numerical weather forecasting, the physical model based on numerical weather forecasting is not suitable for short-term WS forecasting, so it is mostly used for medium- and long-term WS forecasting [10]. Statistical models are established by studying the mapping relationship between input data and output data in historical data, and their main processes include model identification, parameter estimation, and model validation. Traditional statistical models include the Kalman filter model [11], AR (autoregressive) model [12], MA (moving average) model [13], ARMA (autoregressive moving average) model [14], and ARIMA (autoregressive integrated moving average) model [15]. Compared to physical models, statistical models have simple methods and use a single type of data, but their ability to process sudden changes is poor. At the same time, this model type requires a large amount of historical operating data, so it is not suitable for newly built or incomplete wind farms. In addition, the performance of statistical models deteriorates as the prediction time increases, and it is difficult to obtain excellent forecasting results for nonlinear and unstable WS data [16]. Moreover, the artificial intelligence model is derived from statistical methods and is built through the repeated learning and training of data relationships based on large amounts of historical data. And common artificial intelligence models include the SVM (support vector machine) [17], BP (backpropagation) neural network [18], deep learning model [19,20,21], DBN (deep belief network) [22], reinforcement learning model [23], and broad learning models [24]. Nevertheless, a single artificial intelligence model has the disadvantages of being easily trapped in local minima, overfitting, underfitting, etc., and cannot fully capture all information about WS sequences. In order to overcome these shortcomings and make full use of the advantages of a single model and various methods, building a combined model to predict WS has gradually become a popular topic of research [25].
At present, the most common combined forecasting methods are the combination of different models and the combination of models and algorithms. In terms of different model combinations, Neto et al. [26] constructed a hybrid model that combines linear statistical and artificial intelligence (AI) forecasters, which can effectively overcome the shortcomings of single models. Moreover, Cheng et al. [16] combined the four models of BP, a random vector functional link network, ENN (evolutionary neural network), and GRNN (generalized regression neural network) to form a new hybrid model, which can better utilize the characteristics of each model and make up for the inadequacy of a single model. Moreover, Shahzad et al. [27] proposed a new combined model based on ARAR (autoregressive–autoregressive) and ANN (artificial neural network) models, which was experimentally demonstrated to better predict WS time series datasets. And Liang et al. [28] combined the RNN (recurrent neural network) and CNN (convolutional neural network) to propose a new method. On the other hand, there are also many relevant studies on the combination of models and algorithms. According to the different functions of algorithms, they can be divided into two categories: optimization algorithms and decomposition methods. Because of their strong versatility, high efficiency, and simple implementation, swarm intelligence optimization algorithms have become a popular topic of research in the field of artificial intelligence and intelligent computing [29,30] and have been widely used in the parameter optimization of WS prediction models. For instance, there are some common optimization algorithms that have been combined with models, which include PSO (particle swarm optimization) [31], the WOA (whale optimization algorithm) [32], the DA (dragonfly algorithm) [33], GWO (grey wolf optimization) [34], etc. WS data have the characteristics of strong nonlinearity and non-stationarity. Signal decomposition methods have been gradually introduced into WS prediction in recent years because of their outstanding role in denoising and extracting signal features in complex signal processing. They are mainly used in combination with various models for data preprocessing. In WS prediction, common decomposition methods include EMD (empirical mode decomposition) [35], VMD (variational mode decomposition) [36], EEMD (ensemble empirical mode decomposition) [37], CEEMD (complementary ensemble empirical mode decomposition) [38], etc. Ibrahim et al. [39] proposed a combined prediction model based on AD-PSO, Guided WOA, and LSTM (long short-term memory); the experiment showed that the LSTM optimized by AD-PSO and Guided WOA has strong predictive performance. Altan et al. [40] proposed a combined prediction model based on a single model, namely, LSTM, a decomposition method, namely, ICEEMDAN (improved complementary ensemble empirical mode decomposition with adaptive noise), and an advanced optimization algorithm, namely, GWO; the combined model could better capture the nonlinear characteristics of WS data and had good prediction accuracy.
While combination models integrate multiple individual models and techniques to leverage their respective strengths and compensate for limitations, existing combination models still have some issues. First, determining optimal model integration and weighting is challenging when combining multiple model types. Simply combining various models often does not significantly improve combination model performance. Currently, there is a lack of comparative analysis on the roles of various types of techniques in composite model research. The theoretical construction of composite models needs further improvement. Second, when combining models with optimization algorithms, most existing combination models directly introduce the algorithm without improvements, limiting the algorithm’s functionality. Third, when combining models with decomposition methods, decomposition parameter settings significantly impact results. Existing models often use experiential parameter settings rather than effectively optimizing parameter combinations. Finally, combination models’ complex structures and numerous parameters generally cause long prediction times. These issues directly affect model prediction performance.
Based on the above analysis, the primary factors that impact the performance of a WS prediction model are as follows: First, WS data preprocessing is inadequate. WS data sources contain much noise and interference, making complete and precise data collection challenging. Existing preprocessing cannot effectively reduce noise and extract time series information. Second, model parameter optimization is insufficient. The many parameters in combination models often make conventional intelligent optimization algorithms inaccurate and slow in application, negatively affecting model prediction. Optimization algorithm convergence, search efficiency, and parameter settings need further development and improvement. Finally, combination model construction is unreasonable. WS characteristics like strong randomness, intermittency, and volatility require models with robust adaptive abilities. Existing models lack integrated advanced data mining and parameter optimization technologies for performance optimization. Therefore, it is evident that existing models have certain limitations and fail to meet the increasing demand for WS prediction performance, especially in terms of selecting and optimizing individual techniques within combination models, leaving significant room for improvement. To address these issues, this paper proposes a hybrid ultra-short-term WS forecasting model based on a modified pelican optimization algorithm. Firstly, targeting the common problems of swarm intelligence optimization algorithms, such as being easily trapped in local optima and insufficient optimization precision, an MPOA (modified pelican optimization algorithm) is introduced. This algorithm enhances the diversity of the initial population by incorporating chaotic mapping, improves the algorithm’s local exploration capability by introducing Lévy flight strategy, and dynamically adjusts critical parameters to enhance adaptability, thereby achieving overall performance enhancement in algorithm optimization. Secondly, to address the challenges of parameter abundance and the difficulty in combination optimization in VMD, a combined VMD key parameter optimization method using the MPOA and sample entropy is proposed, enabling the deep exploration of WS’s nonlinear variation characteristics. Finally, to tackle the problem of numerous hyperparameters in LSTM and the difficulty in achieving parameter optimization manually, the MPOA is applied to optimize the hyperparameters of LSTM, resulting in a significant improvement in LSTM’s predictive performance. Additionally, this paper conducts comparative simulation experiments between the proposed combination model and other models using WS data from a wind farm in eastern China, validating the predictive performance of the model.
The primary contributions of this paper include the following:
  • A modified pelican optimization algorithm is proposed. First, the tent map is applied for population initialization and the Lévy flight factor is introduced. Second, algorithm applications to different dimensional problems are classified and optimized. Finally, improved algorithm optimization performance is verified through test functions. The proposed modified pelican optimization algorithm not only effectively enhances the algorithm’s global and local search capabilities but also provides a new avenue for optimizing swarm intelligence algorithms;
  • The original WS data are modally decomposed by VMD. Through the use of sample entropy as an evaluation metric to determine K (the modal decomposition number) for VMD, and of RMSE (root mean square error) as the objective function for the MPOA, the alpha (moderate bandwidth constraint) and tau (noise tolerance) of the other two parameters of VMD are optimized. Applying VMD with the optimized parameter combination for WS data decomposition significantly reduces the nonlinearity and instability characteristics of the original WS data, effectively exploring the features of WS data. By introducing the composite evaluation metrics of sample entropy and RMSE in the parameter optimization of the WS signal decomposition strategy VMD, the data preprocessing technique effectively removes noise while retaining the main features of the original signal sequence, significantly improving the effectiveness of WS prediction;
  • The MPOA is applied to optimize LSTM model hyperparameters including the iteration number, first LSTM layer neuron number, second LSTM layer neuron number, and learning rate. Through hyperparameter optimization, the MPOA significantly strengthens LSTM predictive capabilities;
  • A combined ultra-short-term WS forecasting model based on the modified pelican optimization algorithm is proposed, combining the MPOA, VMD, and LSTM. The WS prediction process based on the combined model is constructed. The proposed VMD-MPOA-LSTM model provides a new effective method to improve ultra-short-term WS prediction model performance;
  • A comprehensive and effective evaluation of the developed combined model’s performance is carried out. The evaluation system uses three experiments and six performance indicators to effectively assess prediction accuracy and stability. Experimental analysis specifically determines optimization algorithm and data decomposition method impacts on the prediction model, providing a theoretical and practical basis for combined model construction.
The rest of this paper is organized as follows: In Section 2, the methodology is introduced, including optimization algorithms, data preprocessing techniques, and LSTMs and their combined models. And in Section 3, the experimental procedures of the proposed combined model are illustrated, and the comparative experimental results of different models are shown. Additionally, the precision and validity of the proposed model are discussed and validated in Section 4. Concluding remarks are then provided in Section 5.

2. Methodology

This section details the methodologies used to build the hybrid forecasting model, including the modified pelican optimization algorithm, variational mode decomposition, LSTM neural network, and framework of the proposed combined model. The modified pelican optimization algorithm is applied to optimize the parameters of VMD and LSTM, enhancing the capabilities of the hybrid forecasting model.

2.1. Modified Pelican Optimization Algorithm

2.1.1. Pelican Optimization Algorithm

The pelican optimization algorithm is a novel swarm intelligence algorithm proposed by Trojovský et al. [41], which simulates the natural behavior of pelicans during hunting. The procedural steps for optimization using the algorithm are detailed below.
(1)
Initialization
Initially, the population is randomly generated in a certain range according to Equation (1).
x i , j = l j + r a n d ( u j l j ) , i = 1 , 2 , , N , j = 1 , 2 , , m
where x i , j represents the value of the jth variable for the ith pelican in the population. The population size is N, and the dimension is m. rand denotes a randomly generated number between 0 and 1. l j and u j represent the lower and upper bounds, respectively.
In the proposed POA, the population members consisting of pelicans are represented by a population matrix, defined in Equation (2).
X = X 1 X i X N N × m = x 1 , 1 x 1 , j x 1 , m   x i , 1 x i , j x i , m   x N , 1 x N , j x N , m N × m
where X is the population matrix of pelicans and X i is the ith pelican member of the population.
In the POA, every constituent of the population represents a potential solution in the form of a pelican. Thus, the objective function for the given problem can be evaluated for each candidate solution. The resulting objective function values for all candidates are contained in a vector called the objective function vector, defined in Equation (3).
F = F 1 F   i F   N N × 1 = F ( X   1 ) F ( X   i ) F ( X   N ) N × 1
where F denotes the objective function vector comprising the objective function values computed for each prospective solution. F i refers to the objective function value of the ith potential solution.
(2)
Location Update
The location update process in the POA simulates the foraging behavior of pelicans, which is divided into two phases: exploration and exploitation.
In the exploration phase, pelicans first locate potential prey and then move toward areas where prey is detected. The prey location is arbitrarily created inside the search field, allowing the POA to better probe the problem domain. This increases the algorithm’s ability to thoroughly traverse the problem resolution landscape. The pelican’s movement strategy toward the prey position is mathematically modeled as follows:
x i , j P 1 = x i , j + r a n d ( p i I x i , j ) , F p < F i x i , j + r a n d ( x i , j p j ) , e l s e
where p j is the position of the prey in the jth dimension, and F p represents the objective function value of the prey. I is a random number taking either 1 or 2 as its value. When I = 2 , it allows each individual pelican to increase its movement distance and enter new areas of the search space for further exploration.
After the exploration phase, the objective function value at the pelican’s new position is compared to the objective function value at the prey location. The position with the lower objective function value is selected, whether it be the prey location or the pelican’s new explored position. This comparison helps guide the pelicans toward better solutions within the search space. The update method is as follows:
X i = X i P 1 , F i P 1 < F i X i   , e l s e
where X i P 1 is the new position of the ith pelican after phase 1 exploration and F i P 1 is its corresponding objective function value.
In the exploitation phase, pelicans utilize a different foraging strategy. After reaching the target area located in phase 1, pelicans beat their wings over the water to steer prey together and capture it in their throat pouches. Mathematically modeling this behavior increases the POA’s local search abilities and resource utilization within identified promising regions. More prey can be captured within the attack area using this exploitation strategy. The pelican hunting process is formally simulated using the following mathematical model:
x i , j P 2 = x i , j + R 1 t T 2 r a n d 1 x i , j
where X i , j P 2 represents the new position of the pelican i in dimension j after exploitation in phase 2. R is a constant at 0.2, and R · ( 1 t / T ) defines the proximity searching radius of x i , j , where t refers to the ongoing iteration and T is the final iteration.
During this phase, Equation (7) models an effective acceptance–rejection process for the new pelican position determined by local exploitation.
X i = X i P 2 ,   F i P 2 < F i X i   , e l s e
where X i P 2 and F i P 2 represent the new position and corresponding objective value of the ith pelican after applying the phase 2 local exploitation strategy.
(3)
Step Repetition
Once all population members have undergone the exploration and exploitation phases to update their positions, the optimal solution found thus far is refreshed based on the new population distribution and corresponding objective function values. The algorithm then iterates, recycling through the stipulated POA steps as per Equations (4)–(7). This iterative process continues until the full execution is complete. Ultimately, the best candidate solution discovered over all the iterations is offered as a suboptimal solution to the optimization problem. By progressively improving the candidate solutions over multiple iterations in this manner, the POA aims to converge toward the most suitable solution.

2.1.2. Improvement in the Algorithm

(1)
Population initialization based on tent map
During population initialization, the POA adopts a similar approach to other metaheuristic optimization algorithms, employing a randomly generated method. This approach results in a lack of diversity within the population. The diversity in the population directly influences the convergence accuracy and speed of the optimization algorithm. A well-diversified population is advantageous for the algorithm to quickly identify the global optimal solution. Therefore, introducing an appropriate population initialization method that covers a broader range of the solution space is essential. A chaotic sequence has the characteristics of regularity, randomness, and ergodicity. Fusing the POA’s initial population with chaos provides better diversity compared to full randomization. Chaos mainly generates a chaotic sequence through mapping the relationship of (0, 1), and then it converts the map to the population individual space. Many types of chaos exist, with logistic mapping being most common. However, logistic mapping has poor ergodicity and sensitivity to initial parameters, with high mapping point density at the edges and lower density in the middle region. Tent mapping has better uniformity than logistic mapping [42,43]. Therefore, tent mapping was selected to initialize the population in this research. The tent map is a piecewise linear function with parameter μ defined as follows:
x n + 1 = f μ ( x n ) = μ x n f o r   x n < 1 / 2 μ ( 1 x n )   f o r   1 / 2 x n
where μ ( 0 , 2 is the chaos parameter, which is proportional to the chaos.
(2)
Classification value of I value
It can be known from Equation (4) that I is a random number, representing the displacement of individuals in the group. Thus, it represents the performance of entering a new search space. When the value of I is small, individuals can effectively explore local regions, thus benefiting the later-stage optimization of the population and making it more suitable for low-dimensional functions. When the value of I is large, the individuals’ movement space expands, allowing for a broader exploration range and escaping local optima. Consequently, this is advantageous for the early-stage optimization of the population and is more suitable for high-dimensional functions with multiple local optima. In the POA, I is a fixed value, preventing it from achieving optimal optimization performance across different types of functions. Therefore, selecting different values of I for different function types can effectively enhance the optimization performance of the POA. In this paper, I is classified and selected based on the function’s dimension as follows:
I = 2 , D i m > 10 ; R a n d o m   v a l u e [ 1 , 2 )   ,   e l s e
(3)
Introducing the Lévy flight strategy
In the basic POA, especially for high-dimensional function optimization, pelican individuals can “clump” as iterations increase, causing the algorithm to fall into local extrema. A Lévy flight involves searching the solution space through random walks. The step length of a Lévy flight follows a Lévy stable distribution, enabling it to conduct a wide-ranging search in the solution space and making the movement of individuals more agile. The Lévy flight can break this “clump” phenomenon by increasing population diversity and improving local exploration. Therefore, Lévy flights are introduced into the POA’s position update to obtain
x i , j P 2 = x i , j + [ 0.05 × L e v y ( λ ) 1 ] x i , j ; D i m > 10 x i , j + R ( 1 t / T ) ( 2 r a n d 1 ) x i , j , e l s e
L e v y ( λ ) is the Lévy random search path and satisfies
L e v y ~ u = t λ 1 < λ < 3
where u represents a random number following a normal distribution, and t represents the flight time.
Figure 1 depicts a flowchart of the improved POA. The modifications are highlighted using red text boxes.

2.1.3. The Test of the MPOA

The CEC (Congress on Evolutionary Computation) benchmark test functions are commonly used to test swarm intelligence algorithms, comprising 23 classical functions divided into three categories: unimodal, multimodal, and low-dimensional multimodal. Unimodal functions evaluate an algorithm’s convergence rate, precision, and local search capabilities in high-dimensional spaces. Multimodal functions, containing multiple local optima, assess an algorithm’s ability to find the global optimum through broad search space exploration. Low-dimensional multimodal functions specifically examine algorithm performance under low-dimensional conditions. Using both unimodal and multimodal benchmark functions of varying dimensions allows the comprehensive testing of an algorithm’s convergence properties, local refinement capabilities, and global exploration tendencies. Nine test functions were selected from the CEC benchmarks for this research, representing the most commonly used typical functions across the three categories, with variable dimensions ranging from 2 to 30. This provides strong test performance to comprehensively evaluate optimization algorithm capabilities. Table 1, Table 2 and Table 3 show details for each function.
To validate the efficacy of the MPOA, this study selected 9 benchmark functions from the classic benchmark functions according to actual needs and used Matlab®(2021b) to optimize the calculation. Then, the optimization results of the MPOA and several popular swarm intelligence optimization algorithms (PSO, MFO, WOA) were compared and analyzed. The parameters selected for each algorithm are listed in Table 4. All optimization calculations in Matlab® were performed on a desktop computer with 8 GB memory, a Ryzen 8-core processor, and a main frequency of 3.4 GHz.
All algorithms were limited to a specified maximum number of iterations based on their convergence speed. For functions F1–F6, the limit was 1000 iterations, and for F7–F9, it was 100 iterations. Each algorithm independently ran 30 times on each benchmark function to obtain statistically valid results. The performance metrics reported include the mean, median, standard deviation, best, and worst objective function values over the 30 runs, detailed in Table 5, Table 6 and Table 7. Convergence curves plotting optimization progress over iterations were also generated for each algorithm and function, as illustrated in Figure 2.
The results in Table 5, Table 6 and Table 7 and Figure 2 show that the MPOA demonstrated significantly better convergence accuracy and speed compared to the other four algorithms when optimizing F1–F6. For F7–F9, its convergence accuracy and speed were slightly better than the POA but still substantially improved over the other three algorithms, with superior stability.

2.2. MPOA-VMD

2.2.1. VMD

VMD is an adjustable signal processing approach developed by Dragomiretskiy et al. [44] that concurrently decomposes a multivariate signal into several univariate signals, including amplitude modulation (AM) and frequency modulation (FM) signals. This avoids issues like end effects and spurious components that can arise with iterative decomposition methods. Additionally, VMD capably handles nonlinear and non-stationary signals. Unlike traditional EMD, the IMFs in VMD are reconceptualized as AM or FM signals and expressed as
u k = A k t cos ( φ k ( t ) )
where the instantaneous amplitude is represented by A k ( t ) and the instantaneous phase by φ k ( t ) . VMD is based on a variational problem framework, which achieves the adaptive decomposition of signals by determining the best solution to a constrained variational model. The constrained variational model is expressed as
m i n { u k } , { w k } k = 1 k 𝜕 t σ ( t ) + j π t * u k ( t ) e j w k t 2 2 s . t . k = 1 K u k = f ( t )
where u k = u 1 , , u k is the K components obtained after decomposition; ω k = ω 1 , , ω k is the center frequency of each component; and is the convolution symbol.
Because the constrained variational problem is more difficult to solve directly, it is necessary to solve the constrained variational problem in Equation (13) by introducing quadratic penalty factor α and Lagrange multiplication operator λ   ( t ) . The extended Lagrangian function expression is as follows:
L u k , w k , λ = α K 𝜕 t σ ( t ) + j π t * u k ( t ) e j w k t 2 2 + x ( t ) k = 1 K u k 2 2 + λ t , x ( t ) k = 1 K u k ( t )
The optimal solution in Equation (13) can be obtained by finding the saddle point in Equation (4) through the alternating direction multiplier algorithm.

2.2.2. VMD Based on MPOA

As seen in the VMD steps, the appropriate values of K, alpha, and tau need to be set before signal decomposition. Setting K too high can cause over-decomposition, while setting it too low may lead to under-decomposition. Similarly, a large alpha can lose frequency band information, while a small one gives redundant information. Tau also impacts decomposed signal fidelity, so its value must be appropriately set to avoid losing details. Hence, determining the optimal (K, alpha, and tau) combination is key for VMD. In response, many scholars have proposed using a single indicator like energy loss or entropy as the objective function [45,46], obtaining the best parameter combination through an optimization algorithm. However, the effect is not optimal. Therefore, this study proposes a composite evaluation using sample entropy fused with RMSE. To avoid subjective influence, K is first determined by the sample entropy trending to a minimum stable value. RMSE is then used as the objective function, with the MPOA automatically tuning alpha and tau to the optimal combination.

Sample Entropy

Sample entropy [47] is a measure that can be used to characterize the level of randomness or unpredictability within a time series. It indicates the chaotic or irregular nature of the data. The greater the probability of a new pattern occurring in a sequence over time, the larger the sample entropy value, which indicates a higher complexity of the sequence. When the sample entropy value is smaller, the complexity of the sequence is lower, and the signal component is simpler. Sample entropy can provide a reliable measure of complexity or regularity unaffected by dataset scale and shows good consistency between calculations. Therefore, it is extensively utilized for time series signal evaluation.
The steps to compute sample entropy are as follows: Similarly to the first step of approximate entropy, the time series x(n) of length N is incorporated into an m-dimensional vector. The difference is that i goes from 1 to N m , not N m + 1 , which eliminates errors from vector number differences between dimensions m and m + 1 . In addition, sample entropy also sets the similarity tolerance distance r to obtain d i j m .
When the sample entropy algorithm calculates the probability distribution C m ( i ) , the Heaviside function is used to remove self-matching between vectors:
C r m i = N m i N m 1 j 1 , j i N m θ d ij m r , i = 1 , 2 , , N m
θ z = 1 , Z 0 0 , Z > 0
Obtaining sample entropy also necessitates the calculation of the mean value of the probability distribution of all vectors. When the dimension is m, the calculation formula for the ϕ m ( r ) of sample entropy is as follows:
ϕ m ( r ) = 1 N m i = 1 N m C i m ( r )
When the dimension is increased from m to m + 1 , the corresponding ϕ m + 1 ( r ) can be obtained.
Sample entropy is defined as
S a m p E n r , m , N = l n φ m + 1 r φ m r
The sequence with the smallest value of sample entropy is the trend item of the decomposed sequence. In this paper, the minimum sample entropy of each sequence after VMD decomposition is used as an indicator to measure the decomposition effect. Specifically, a smaller entropy value indicates a more regular periodic item sequence and better decomposition. When K is small, causing under-decomposition, other interference terms are mixed into the trend term, increasing its sample entropy. However, with an appropriate K value, the trend term’s sample entropy decreases. As K continues to increase, sample entropy levels off and stabilizes. In this study, sample entropy is considered stable when the maximum deviation over three consecutive iterations does not exceed 5%. Therefore, in order to prevent over-decomposition, the point where sample entropy reaches stability is used to determine the VMD component number.

RMSE

RMSE is a commonly used metric for evaluating the accuracy of prediction models. It measures the degree of deviation between predicted values and true values by calculating the square root of the sum of squared differences between predicted and true values. The range of RMSE is from 0 to positive infinity, where smaller values indicate smaller prediction errors and stronger predictive capabilities of the model. In practical applications, RMSE is not only useful for evaluating the predictive accuracy of regression models but also commonly employed to assess the fidelity of data after decomposition, given its ability to effectively reflect the degree of deviation between two sets of data. To consider the fidelity of VMD, the IMFs obtained from the decomposition are cumulatively reconstructed into a sequence m. The RMSE between this reconstructed sequence m and the original sequence M is then calculated. This RMSE is used as an index to measure the information integrity after decomposition, calculated as
RMSE = 1 n i = 1 n ( M t m t ) 2
where M t represents the original value at time t , m t denotes the reconstructed value at time t , and n is the sequence length.

Objective Function

Fusing sample entropy and RMSE reflects both the completeness of the decomposed sequence information and the decomposition effect on the sequence. Therefore, sample entropy is first used to identify the optimal K value. Then, RMSE is employed as the objective function to optimize the two parameters alpha and tau. The fitness function expression is
F i t n e s s = RMSE

VMD Optimization Steps

The specific steps for optimizing K, alpha, and tau in VMD are as follows:
Step 1: since “K” equals 1 as the VMD number, compute the sample entropy of the trend term under different “K” values, and take the “K” value corresponding to when the sample entropy of the trend term tends to be stable as the optimal value.
Step 2: initialize MPOA input parameters.
Step 3: initialize all individuals in the population, and randomly generate a series of parameter combinations (alpha, tau).
Step 4: perform VMD on the WS time series signal under the value of the current individual, calculate the RMSE between the original and reconstructed sequences, and calculate the fitness value of each individual according to Equation (20).
Step 5: update the optimal value of the individual and groups according to the fitness value of the individual, and update the population position according to Equation (8).
Step 6: recycle through steps 4 and 5 until the maximal iterations are attained, and output the individual position and fitness value at this time as the optimal solution.

2.3. LSTM

The LSTM neural network is an efficient architecture based on the RNN model, proposed by Hochreiter et al. [48]. LSTM represents the development of an RNN with much better applicability. It is a special RNN form, replacing the hidden layer with a memory unit and adding a state unit Ct to store the long-term network state. Accuracy is improved by introducing new storage units and gate control. Currently, the LSTM neural network has achieved success in fields like image resolution, audio processing, timing prediction, and big data analysis [49,50,51]. They have found extensive application across research domains including traffic analysis, meteorology, network security, and other relevant fields [52,53,54]. The structure of LSTM is shown in Figure 3.
As presented in Figure 3, LSTM networks introduce three “gate” control units—the forgetting gate f t , the input gate i t , and the output gate o t — and an input unit h t 1 . The three gate structures constitute LSTM’s unique “forgetting” process to regulate information flow, which is an advantage over traditional RNNs. This structure allows LSTM to recall long-distance information, effectively overcoming the vanishing gradient and making LSTM powerful for handling nonlinear sequence data with long- and short-term dependencies. In addition, compared to other machine learning methods, LSTM uses backpropagation through time, operating sequentially to better handle time series data. Since WS prediction involves temporal dependencies, LSTM can capture these relationships to accurately forecast future wind speeds over short time horizons. Consequently, LSTM has increasingly been applied to WS prediction tasks in recent years with good results. However, the performance of the LSTM model largely depends on the selection of its hyperparameters. Optimizing LSTM hyperparameters is a crucial step in applying the LSTM model. Choosing appropriate hyperparameters can effectively improve the model’s performance and generalization ability. Therefore, this paper will optimize the key parameters of LSTM, including the number of iterations Titer, learning rate LR, number of first-layer neurons L1, and number of second-layer neurons L2, using the MPOA with mean square error as the objective function.

2.4. VMD-MPOA–LSTM Combined Model

This section introduces a combined WS forecasting system of VMD-MPOA-LSTM based on time series preprocessing technology, a swarm intelligence optimization algorithm, and deep learning. It uses VMD to obtain the trend traits of the original WS sequence and MPOA to optimize LSTM hyperparameters. The framework of the VMD-MPOA-LSTM approach is illustrated in Figure 4, with the key steps described below.
Step 1: Apply MPOA-VMD to decompose the original WS time series data. This takes sample entropy and RMSE as evaluation indexes, using the MPOA to obtain the optimal combination of K, alpha, and tau, improving reconstruction precision after decomposition. MPOA-VMD decomposes the time series into K IMFs while determining the optimal parameter group.
Step 2: use LSTM to predict each IMF component.
(1)
Split each of the K IMFs from decomposition into separate training and testing datasets;
(2)
Use the mean squared error between predicted and actual values as the objective function to optimize four LSTM hyperparameters;
(3)
Apply the optimized LSTM model to generate WS forecasts and the most accurate predicted values.
Step 3: aggregate the predicted values from each IMF through linear combination to yield the overall forecast for the original WS time series, obtaining the final result.
Step 4: evaluate the predictive performance of the proposed hybrid model using several evaluation indexes.

3. Experiment and Analysis Results

This section provides details on the dataset used, the evaluation metrics used to assess forecast accuracy, and the experimental setup and results for the application of the proposed combined VMD-MPOA-LSTM model.

3.1. Datasets

The original WS time series used in the experiments is from an eastern China wind farm. A total of 2600 data points were selected from 1 April 2021 at 0:00 to 28 April 2021 at 24:00 as the test data. The dataset was divided, with 2300 points allocated to the training set and 300 points to the testing set. The WS sampling interval was 15 min. Data from April were chosen because historical data show that average wind speeds in April are highest with frequent changes, better suiting model effectiveness validation. A 15 min resolution was selected because higher-resolution data have more peaks and complex noise, useful for verifying model performance, meeting China’s ultra-short-term forecasting requirements. Although 15 min sampling makes the data more unstable, decreasing prediction performance, it demonstrates data generality and makes the prediction more meaningful. Additionally, this study primarily focuses on ultra-short-term WS prediction. Due to the short sampling interval, WS data collected every 15 min can avoid the influence of seasonal factors. Therefore, the WS dataset no longer needed to be sampled separately for each season. Figure 5 displays a sample of the ultra-short-term WS data used in the experiments.

3.2. Performance Metrics

To thoroughly assess the model’s predictive capabilities, six common performance indicators were selected: MAE, RMSE, MAPE, R2, TIC, and EC. MAE, RMSE, and MAPE evaluate prediction accuracy. R2, TIC, and EC gauge model fit. The smaller the values for MAE, RMSE, MAPE, and TIC, and the larger the values for R2 and EC, the better the prediction performance. The specific expressions are as follows:
MAE = 1 N t = 1 N u a c t u a l t u p r e d i c t t
RMSE = 1 N t = 1 N ( u a c t u a l t u p r e d i c t t ) 2
MAPE = 1 N t = 1 N u a c t u a l t u p r e d i c t t u a c t u a l t
R 2 = t = 1 N ( u a c t u a l t u a c t m e a n ) × ( u p r e d i c t t u p r e m e a n ) t = 1 N ( u a c t u a l t u a c t m e a n ) 2 × t = 1 N u p r e d i c t t u p r e m e a n 2 2
TIC = 1 N t = 1 N ( u a c t u a l t u p r e d i c t t ) 2 1 N t = 1 N ( u a c t u a l t ) 2 + 1 N t = 1 N ( u p r e d i c t t ) 2
EC = 1 1 N t = 1 N ( u a c t u a l t u p r e d i c t t ) 2 1 N t = 1 N ( u a c t u a l t u a c t m e a n ) 2
where u a c t u a l t represents the actual WS value at time t ; u p r e d i c t l t represents the forecast value at time t ; u a c t m e a n t represents the average actual value; u p r e m e a n t is the average forecast value; and N refers to the total number of data points in the time series u t .
Furthermore, the DM test [55] was employed to perform the validity analysis of the proposed model in this paper. The DM test utilizes the output statistic DM to judge whether there is a significant difference in the predictive capabilities of the two models.

3.3. Parameters Setting

In this paper, BP, SVM, ELM, LSTM, and their forecasting methods combined with some optimization algorithms are used for comparison with the proposed improved combined forecasting method. The parameter settings of PSO, MFO, WOA, POA, and MPOA are the same as in Table 4, except the population size is changed to eight and number of iterations to 50. The remaining model-related parameter settings are shown in Table 8. The models mentioned in this paper are named according to the primary methods adopted by each model. Models such as BP, ELM, and SVM are single models that do not use signal decomposition or optimization algorithms. Models like EMD-LSTM, VMD-LSTM, and EEMD-LSTM are combination models composed of various signal decomposition methods and LSTM. Models such as VMD-WOA-SVM, VMD-MPOA-SVM, VMD-WOA-LSTM, etc., are combination models that integrate the signal decomposition method VMD with swarm intelligence algorithms and single models.

3.4. Different Experiments and Result Analysis

Three experiments were conducted comparing the proposed model against other models. An analysis of the comparative results is provided in this section.

3.4.1. Experiment I: Comparison with Models Using Different Data Preprocessing Methods

To verify the effectiveness of the proposed MPOA-VMD signal decomposition method for WS prediction, it is compared with two common methods, EMD and EEMD. Figure 6 shows the MPOA-VMD WS signal decomposition results.
Figure 7 displays the prediction results for the five models in this experiment. The proposed model is the improved VMD-MPOA-LSTM combined forecasting model studied in this paper. Table 9 shows the specific evaluation indicators for the prediction results of each combined model. Figure 8 depicts histograms comparing each evaluation index across models. Table 10 provides the DM test values between each model and the proposed model.
According to the WS prediction curves in Figure 7, the proposed model’s WS prediction values are closer to the actual values. This is reflected in its R2 performance index being the highest among the models. Moreover, Table 9 and Figure 8 show that the proposed model has better values across all six evaluation indicators compared to EMD-LSTM and EEMD-LSTM. This indicates that VMD is significantly more effective than EMD and EEMD for WS decomposition. Compared to VMD-LSTM, MPOAVMD-LSTM has slightly better results across the six indicators. This shows that MPOA optimization of VMD improves decomposition, enhancing model performance. The DM value represents the difference in the prediction sequence between the proposed combination model and other models, and its size represents the significance of the difference. All models in Table 10 exhibit DM values at a significance level of 10%, indicating a significant difference in the predictive performance of the proposed model compared to other models. The improvement in the model has led to a noticeable enhancement in predictive effectiveness. Overall, the examination of the five combined model results clearly demonstrates that the proposed model outperforms the others.
Remark: Examining the prediction results across six indicators shows that the proposed model demonstrates superior capabilities compared to the others. The DM test also reveals significant differences between the proposed model’s prediction sequences and those of the other models. It is evident that the introduction of decomposition methods can significantly improve the predictive performance of the ensemble model, and different decomposition methods yield distinctly different results in wind speed signal decomposition. Therefore, selecting an appropriate decomposition method is crucial for the wind speed prediction ensemble model. Additionally, the parameter settings of the decomposition method directly impact the quality of signal decomposition, thereby influencing the predictive performance of the model. Introducing optimization algorithms into the decomposition method for parameter optimization is beneficial for further enhancing the predictive performance of the ensemble model.

3.4.2. Experiment II: Comparison with Models Using Different Optimization Algorithms

To validate the efficacy of the MPOA for WS prediction, it is compared against combined models using conventional optimization algorithms, namely, PSO, WOA, MFO, and POA. Figure 9 displays the prediction results for the five different prediction models in this experiment. Table 11 provides the prediction results for each combined model. Figure 10 uses histograms to depict the values of each evaluation index applied to the prediction results across models.
In the above experiment, all five models used LSTM but with different optimization algorithms. Upon examination of the six evaluation index values presented in Table 11 and Figure 10, it can be observed that the prediction performance of the four models—WOA-LSTM, POA-LSTM, MFO-LSTM, and PSO-LSTM—decreases sequentially, and their MAPE values are 17.21%, 17.68%, 18.07%, and 19.76%, respectively. The prediction effect of WOA-LSTM is superior to that of the other three optimization algorithms. Furthermore, the six indicators of the combined model composed of MPOA and LSTM are better than those of WOA-LSTM, and its MAPE value is 16.92%. Therefore, it is proved that the MPOA not only further improves the optimization performance of POA, but also can enhance the prediction capabilities of LSTM more effectively than the other optimization algorithms.
Remark: Compared with the six prediction performance indicators of all models, it can be clearly seen that the optimization algorithm can enhance the prediction capabilities of the models to some degree, but the degree of improvement is very limited. In addition, it can be seen that the performance improvement in the POA for LSTM is not the best compared with other optimization algorithms. However, the optimization performance of the improved POA is significantly improved, which makes the prediction performance of MPOA-LSTM superior to that of other models.

3.4.3. Experiment Ⅲ: Comparison with Other Common Single and Combined Models

To further validate the efficacy of the proposed model, it was compared with five classical single forecasting models (ARIMA, BP, ELM, SVM, and LSTM) and six combined models (WOA-SVM, VMD-MPOA-BP, VMD-MPOA-ELM, VMD-WOA-SVM, VMD-MPOA-SVM, and VMD-WOA-LSTM). Figure 11 shows the prediction results of all predictive models in this experiment. Table 12 presents the specific evaluation index values of each combination model. Figure 12 depicts histograms of each evaluation index, comparing the prediction results of the different combined models. Table 13 indicates the DM values between each model and the proposed model.
As shown in Table 12 and Figure 12, the evaluation indicators of the seven combined prediction models (WOA-SVM, VMD-MPOA-BP, VMD-MPOA-ELM,VMD-WOA-SVM,VMD-MPOA-SVM, VMD-WOA-LSTM, and the proposed model) are significantly better than those of the five single prediction models (ARIMA, BP, ELM, SVM, and LSTM). Moreover, the six combined forecasting models, VMD-MPOA-BP, VMD-MPOA-ELM, VMD-WOA-SVM, VMD-MPOA-SVM, VMD-WOA-LSTM, and the proposed model, which incorporate the signal decomposition method, an optimization algorithm, and machine learning, all demonstrate good prediction performance. Specifically, their MAPE values are 7.38%, 7.96%, 9.01%, 8.63%, 7.86%, and 6.49%, while their R2 values are 0.9726, 0.9585, 0.9756, 0.9779, 0.9781, and 0.9803, respectively. In conjunction with Figure 11, these results indicate that the proposed model possesses the best predictive capabilities among all the examined models.

4. Discussion

In this section, the predictive performance of each model type, the efficacy of the proposed model, and the enhancements in predictive accuracy as well as its performance in multi-step forecasting are discussed.

4.1. Comparison of Prediction Effects of Various Models

At present, the common combined model is a combination of an algorithm and prediction model, where the algorithm can be divided into decomposition methods and optimization algorithms. These can be paired with models individually or simultaneously. A decomposition method or optimization algorithm may be coupled with a model, or both algorithm types can be combined with a model concurrently. Based on the results of the three experiments presented in the previous section, the combined model demonstrates significantly better predictive performance compared to the single models. However, in a comparison of the experimental results of Experiment I and Experiment II, the results illustrate that while optimization algorithms are able to enhance a model’s prediction performance to some degree, the improvement is far less than that in the decomposition methods. Specifically, the evaluation indices of each model in Experiment I are markedly superior to those in Experiment II. For instance, the R2 values of each model in Experiment II were above 0.9, much higher than the R2 values of each model in Experiment I. Furthermore, the MAPE values in Experiment II were all below 13%, much lower than the MAPE values in Experiment I. This is because in AI-based WS prediction models, the processing of the original WS data has become one of the most crucial factors impacting the prediction capabilities of the model. Additionally, through a comparison of the results of Experiment III with those of Experiment I and Experiment II, it can be concluded that the combined model composed of decomposition methods, optimization algorithms, and a prediction model demonstrates the optimal predictive performance among all combination approaches examined. This combination approach can effectively utilize the advantages of decomposition methods and optimization algorithms to boost the performance of a single model. Overall, the model prediction effects of the various combination techniques are ranked in descending order as follows: (1) decomposition method, optimization algorithm, and single model; (2) decomposition method and single model; (3) optimization algorithm and single model; and (4) single model.

4.2. Effectiveness of the Proposed Model

The DM test was utilized in this paper to validate the efficacy of the proposed model, where the DM value signifies the divergence in predicted series between the proposed model and other models. Table 10 and Table 13 illustrate that only the VMD-WOA-SVM combined model has a DM value of 1.8611, which is at the 10% significance level, while the other models exhibit DM values at the 1% significance level. Therefore, the above results demonstrate that the prediction sequence generated by the proposed model significantly diverges from the sequences produced by the alternative models. This signifies that the predictive performance of the proposed model is markedly superior to that of the other examined models.

4.3. Improvement in the Prediction Performance of the Proposed Model

To evaluate the enhancement in predictive performance attained by the proposed model in a more comprehensive and intuitive manner, the six evaluation indicators mentioned in the previous section were utilized for an extensive comparison. The six indicator values for the proposed model were compared against those of all models across the three experiments, and the growth percentages of the evaluation index values between the proposed model and the other models are presented in Table 14. Taking PMAE as an example, its definition is
P MAE = MAE 1 MAE 2 MAE 1
where MAE1 is the MAE value of the proposed model, and MAE2 is the MAE value of the comparison model.
As depicted in Table 14, the six indicators of the proposed model outperform those of the other 20 prediction models examined in this paper. Among these, the maximum values of PMAE, PRMSE, PMAPE, PR2, PTIC, and PEC are 74.73%, 73.56%, 70.95%, 35.06%, 75.33%, and 65.41%, respectively, while their lowest values are 10.05%, 4.62%, 11.90%, 0.22%, 0.62%, and 0.02%, respectively. Therefore, the proposed model demonstrates outstanding predictive capabilities and robustness compared to all other models. Relative to the other models, the proposed model is markedly superior in each index, owing to the introduction of the improved POA. The enhanced POA can exert prominent optimization effects during the optimization of key VMD parameters and LSTM hyperparameter optimization. This enables the effective integration of the POA, VMD, and LSTM.

4.4. Effectiveness of Multi-Step Prediction

To examine the multi-step prediction capabilities of the proposed model, this section conducts simulation experiments of one-step-, two-step-, and three-step-ahead predictions for representative models discussed in Section 3. The experimental results are presented in Table 15. The values in bold in Table 15 denote the lowest MAE, RMSE, and MAPE values among the models. Additionally, the results in Table 15 illustrate that the one-step prediction performance is superior to the two-step and three-step predictions, indicating that prediction accuracy decreases as the number of predicted steps increases. Compared to the other four models, the proposed model demonstrates the optimal prediction performance at all step sizes. This signifies that the proposed model possesses superior multi-step prediction capabilities.

5. Conclusions

Wind power generation technology has been rapidly advancing, leading to increases in installed wind power capacity and its contribution to overall power supply. Hence, the impact of wind farms on reliable grid operation is becoming increasingly pronounced. However, the strong time variation in, nonlinearity of, and complexity of WS render the prediction effect unsatisfactory. Therefore, this paper proposes a combined ultra-short-term WS forecasting model utilizing a modified pelican optimization algorithm, composed of the WS decomposition methods VMD, MPOA, and LSTM. Firstly, techniques including tent map-based population initialization and Lévy flight strategy are introduced to form an improved POA, whose superior optimization performance is verified by comparison with other popular optimization algorithms. Secondly, given the difficulty in selecting the objective function in current VMD parameter optimization, the optimal number of decompositions in VMD is determined by the minimum sample entropy value. Concurrently, taking RSME as the objective function, the optimal parameter combination of k, alpha, and tau in VMD is obtained through the MPOA, effectively enhancing the accuracy of VMD signal reconstruction after WS signal decomposition. Additionally, the MPOA is applied to LSTM hyperparameter optimization to attain the optimal combination. Finally, single-month WS data from an eastern China wind farm were selected, where WS fluctuations during this month were most frequent and drastic compared to during other months of the year. The proposed combined prediction model is compared, analyzed, and evaluated against other models through the aforementioned three experiments and six evaluation indicators. The experiments focus on the quantitative analysis of utilizing optimization algorithms and decomposition methods to enhance the prediction capabilities of the combined model, providing a theoretical basis and practical guidance for constructing hybrid models moving forward. The experimental results demonstrate that the combined model achieves a MAPE of 6.49% and R2 of 0.9803, with optimal values on other metrics. This signifies that the combined model possesses superior predictive capabilities compared to conventional single models and earlier combined models. Additionally, through further, concurrent optimization of the parameters of both the decomposition method and the single model, the optimization algorithm, decomposition method, and single model can be organically integrated, markedly improving the WS predictive capabilities of the combined model. This study provides an effective approach to further enhance the predictive capabilities of combined models.
In future work, the following issues should be studied further:
(1)
The combination of an optimization algorithm, decomposition method, and neural network needs to be further studied based on the mechanism, and effective methods to realize the in-depth integration of different technologies need to be explored. At the same time, it is necessary to integrate more and more advanced technologies into the combination forecasting model to further enhance the predictive capabilities of the combination model;
(2)
Despite experimental verification of its superior predictive performance, a key limitation of the current model is its prolonged computation time. Reducing calculation time and improving forecasting efficiency should be priority research directions moving forward. Parallel computing, algorithm optimization, model simplification, and data preprocessing will likely be effective solutions to reduce computation time and improve the computational efficiency of the combined model;
(3)
The model predicts WS through historical data and only considers the magnitude of WS. In fact, meteorological parameters like wind orientation, temperature, moisture, and barometric pressure impact the precision of predictions. In the future, it is necessary to consider combining these external factors to establish a multi-input prediction model for further enhancing WS forecasting accuracy;
(4)
The proposed model will be tested in additional actual wind farm scenarios and compared with more prediction models to further validate its robustness and adaptability;
(5)
For the MPOA, although the improvement methods and experimental validation results are detailed, there is a lack of theoretical analysis of the improved algorithm. A mathematical proof of the convergence of the improved algorithm and the analysis of the mechanisms that enhance the model’s prediction performance will be important directions for future research.

Author Contributions

Conceptualization, L.G. and C.X.; methodology, L.G.; validation, X.A. and X.H.; formal analysis, L.G.; investigation, C.X.; resources, L.G. and X.H.; data curation, X.A.; writing—original draft preparation, L.G.; writing—review and editing, F.X.; supervision, X.A.; project administration, X.H.; funding acquisition, L.G., X.H. and C.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 52406234), the China Postdoctoral Science Foundation (No. 2024M760739), the Fundamental Research Funds for the Central Universities (No. B240201171), the Changzhou Science and Technology Plan—Applied Basic Research Special Project (No. CJ20240094), and the National Key R&D Plan of China (No. 2024YFE0212200).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support this manuscript are available from Chang Xu upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
WSwind speed
NWPnumerical weather prediction
MAmoving average
ARMAautoregressive moving average
ARIMAautoregressive integrated moving average
POApelican optimization algorithm
MPOAmodified pelican optimization algorithm
PSOparticle swarm optimization
MFOmoth–flame optimization
WOAwhale optimization algorithm
DAdragonfly algorithm
GWOgrey wolf optimization
EMDempirical mode decomposition
EEMDensemble empirical mode decomposition
VMDvariational mode decomposition
CEEMDcomplementary ensemble empirical mode decomposition
LSTMlong short-term memory
RNNrecurrent neural network
BPbackpropagation
ELMextreme learning machine
SVMsupport vector machine
CNNconvolutional neural network
RNNrecurrent neural network
DBNdeep belief network
AIartificial intelligence
ENNevolutionary neural network
GRNNgeneralized regression neural network
ARARautoregressive–autoregressive
ANNartificial neural network
IMFintrinsic mode functions
MAEmean absolute error
RMSEroot mean square error
MAPEmean absolute percentile error
TICTheil’s inequality coefficient
ECefficiency coefficient
DMDiebold–Mariano

References

  1. Yang, B.; Zhong, L.; Wang, J.; Shu, H.; Zhang, X.; Yu, T.; Sun, L. State-of-the-art one-stop handbook on wind forecasting technologies: An overview of classifications, methodologies, and analysis. J. Clean. Prod. 2021, 283, 124628. [Google Scholar] [CrossRef]
  2. GWEC. Global Wind Energy Council. Global Wind Report; GWEC: Brussels, Belgium, 2021. [Google Scholar]
  3. Badal, F.R.; Das, P.; Sarker, S.K.; Das, S.K. A survey on control issues in renewable energy integration and microgrid. Protect. Contr. Mod. Power Syst. 2019, 4, 6008. [Google Scholar] [CrossRef]
  4. Yang, B.; Zhang, X.; Yu, T.; Shu, H.; Fang, Z. Grouped grey wolf optimizer for maximum power point tracking of doubly-fed induction generator based wind turbine. Energy Convers. Manag. 2017, 133, 427–443. [Google Scholar] [CrossRef]
  5. Yuan, X.; Chen, C.; Yuan, Y.; Huang, Y.; Tan, Q. Short-term wind power prediction based on LSSVM-GSA model. Energy Convers. Manag. 2015, 101, 393–401. [Google Scholar] [CrossRef]
  6. Liu, T.; Liu, S.; Heng, J.; Gao, Y. A new hybrid approach for wind speed forecasting applying support vector machine with ensemble empirical mode decomposition and cuckoo search algorithm. Appl. Sci. 2018, 8, 1754. [Google Scholar] [CrossRef]
  7. Feng, S.; Wang, W.; Liu, C.; Dai, H. Short term wind speed predictionbase on physical principle. Acta Energiae Solaris Sin. 2011, 32, 611–616. [Google Scholar]
  8. Liu, Y.; Wang, Y.; Li, L.; Han, S.; Infield, D. Numerical weather prediction wind correction methods and its impact on computational fluid dynamics based wind power forecasting. J. Renew. Sustain. Energy 2016, 8, 033302. [Google Scholar] [CrossRef]
  9. Li, Y.; Tang, F.; Gao, X.; Zhang, T.; Qi, J.; Xie, J.; Li, X.; Guo, Y. Numerical Weather Prediction Correction Strategy for Short-Term Wind Power Forecasting Based on Bidirectional Gated Recurrent Unit and XGBoost. Front. Energy Res. 2022, 9, 836144. [Google Scholar] [CrossRef]
  10. Liu, Y.; Yi, S. Optimization of a wind power time series prediction model based on a two-parameter least squares support vector machine. J. Beijing Univ. Chem. Technol. 2019, 46, 97–102. [Google Scholar]
  11. Zheng, P.; Yu, L.; Hou, S.; Wang, C. Multi-Step Wind Power Forecasting based on Kalman Filter Modification. J. Eng. Therm. Energy Power 2020, 35, 235–241. [Google Scholar]
  12. Karakus, O.; Kuruoglu, E.E.; Altinkaya, M.A. One-day ahead wind speed/power prediction based on polynomial autoregressive model. IET Renew. Power Gen. 2017, 11, 1430–1439. [Google Scholar] [CrossRef]
  13. Riahy, G.H.; Abedi, M. Short term wind speed forecasting for wind turbine applications using linear prediction method. Renew. Energy 2008, 33, 35–41. [Google Scholar] [CrossRef]
  14. Do, D.P.N.; Lee, Y.; Choi, J. Hourly Average wind speed Simulation and Forecast Based on ARMA Model in Jeju Island, Korea. J. Electr. Eng. Technol. 2016, 11, 1548–1555. [Google Scholar] [CrossRef]
  15. Zuluaga, C.D.; Alvarez, M.A.; Giraldo, E. Short-term wind speed prediction based on robust Kalman filtering: An experimental comparison. Appl. Energy 2015, 156, 321–330. [Google Scholar] [CrossRef]
  16. Cheng, Z.; Wang, J. A new combined model based on multi-objective salp swarm optimization for wind speed forecasting. Appl. Soft Comput. 2020, 92, 106294. [Google Scholar] [CrossRef]
  17. Liu, M.; Cao, Z.; Zhang, J.; Wang, L.; Huang, C.; Luo, X. Short-term wind speed forecasting based on the Jaya-SVM model. Int. J. Electr. Power 2020, 121, 106056. [Google Scholar] [CrossRef]
  18. Arjun, B.; Kumar, G.; Sudhansu, K.M. A review of short term load forecasting using artifificial neural network models. Procedia Comput. Sci. 2015, 48, 121–125. [Google Scholar]
  19. Mehrkanoon, S. Deep shared representation learning for weather elements forecasting. Knowl. Based Syst. 2019, 179, 120–128. [Google Scholar] [CrossRef]
  20. Shi, Z.; Liang, H.; Dinavahi, V. Direct interval forecast of uncertain wind power based on recurrent neural networks. IEEE Trans. Sustain. Energy 2018, 9, 1177–1187. [Google Scholar] [CrossRef]
  21. Zhang, C.; Chen, C.; Gan, M.; Chen, L. Predictive deep boltzmann machine for multiperiod wind speed forecasting. IEEE Trans. Sustain. Energy 2015, 6, 1416–1425. [Google Scholar] [CrossRef]
  22. Wang, K.; Qi, X.; Liu, H.; Song, J. Deep belief network based k-means cluster approach for short-term wind power forecasting. Energy 2018, 165, 840–852. [Google Scholar] [CrossRef]
  23. Zhao, J.; Guo, Y.; Lin, Y.; Zhao, Z.; Guo, Z. A novel dynamic ensemble of numerical weather prediction for multi-step wind speed forecasting with deep reinforcement learning and error sequence modeling. Energy 2024, 32, 131787. [Google Scholar] [CrossRef]
  24. Zhu, L.; Lian, C.; Zeng, Z.; Su, Y. A broad learning system with ensemble and classification methods for multi-step-ahead wind speed prediction. Cogn. Comput. 2019, 12, 654–666. [Google Scholar] [CrossRef]
  25. Tascikaraoglu, A.; Uzunoglu, M. A review of combined approaches for prediction of short-term wind speed and power. Renew. Sustain. Energy Rev. 2014, 34, 243–254. [Google Scholar] [CrossRef]
  26. Neto, P.S.G.D.; Oliveira, J.F.L.; Santos, D.S.D.; Siqueira, H.V.; Marinho, M.H.N.; Madeiro, F. A hybrid nonlinear combination system for monthly wind speed forecasting. IEEE Access 2020, 8, 191365–191377. [Google Scholar] [CrossRef]
  27. Shahzad, M.N.; Kanwal, S.; Hussanan, A. A new hybrid ARAR and neural network model for multi-step ahead wind speed forecasting in three regions of pakistan. IEEE Access 2020, 8, 199382–199392. [Google Scholar] [CrossRef]
  28. Liang, C.; Liu, Y.; Zhou, J.; Yan, J.; Lu, Z. Wind speed prediction at multi-locations based on combination of recurrent and convolutional neural networks. Power Syst. Technol. 2021, 45, 534–542. [Google Scholar]
  29. Altan, A. Performance of metaheuristic optimization algorithms based on swarm intelligence in attitude and altitude control of unmanned aerial vehicle for path following. In Proceedings of the 4th International Symposium on Multidisciplinary Studies and Innovative Technologies, Istanbul, Turkey, 22–24 October 2020. [Google Scholar]
  30. Ozcelik, Y.B.; Altan, A. Overcoming nonlinear dynamics in diabetic retinopathy classification: A robust AI-based model with chaotic swarm intelligence optimization and recurrent long short-term memory. Fractal Fract. 2023, 7, 589. [Google Scholar] [CrossRef]
  31. Bou-Rabee, M.; Lodi, K.A.; Ali, M.; Ansari, M.F.; Tariq, M.; Sulaiman, S.A. One-month-ahead wind speed forecasting using hybrid AI model for coastal locations. IEEE Access 2020, 8, 198482–198493. [Google Scholar] [CrossRef]
  32. Hu, R.; Qiao, J.; Li, Y.; Sun, Y.; Wang, B. Medium and long term wind power forecast based on WOA-VMD-SSA-LSTM. Acta Energiae Solaris Sin. 2024, 45, 549–556. [Google Scholar]
  33. Li, L.; Zhao, X.; Tseng, M.L.; Tan, R. Short-term wind power forecasting based on support vector machine with improved dragonfly algorithm. J. Clean. Prod. 2020, 242, 118447. [Google Scholar] [CrossRef]
  34. Phan, Q.B.; Nguyen, T.T. Enhancing wind speed forecasting accuracy using a GWO-nested CEEMDAN-CNN-BiLSTM model. ICT Express 2024, 10, 485–490. [Google Scholar] [CrossRef]
  35. Xiong, Z.; Yao, J.; Huang, Y.; Yu, Z.; Liu, Y. A wind speed forecasting method based on EMD-MGM with switching QR loss function and novel subsequence superposition. Appl. Energy 2024, 353, 122248. [Google Scholar] [CrossRef]
  36. Parri, S.; Teeparthi, K. VMD-SCINet: A hybrid model for improved wind speed forecasting. Earth SCI Inform. 2024, 17, 329–350. [Google Scholar] [CrossRef]
  37. Shang, Z.; Chen, Y.; Chen, Y.; Guo, Z.; Yang, Y. Decomposition-based wind speed forecasting model using causal convolutional network and attention mechanism. Expert Syst. Appl. 2023, 223, 119878. [Google Scholar] [CrossRef]
  38. Chen, J.; Guo, Z.; Zhang, L.; Zhang, S. Short-term wind speed prediction based on improved Hilbert-Huang transform method coupled with NAR dynamic neural network model. Sci. Rep. 2024, 14, 617. [Google Scholar] [CrossRef]
  39. Ibrahim, A.; Mirjalili, S.; El-Said, M.; Ghoneim, S.S.M.; Al-Harthi, M.M.; Ibrahim, T.F.; El-Kenawy, E.M. Wind speed ensemble forecasting based on deep learning using adaptive dynamic optimization algorithm. IEEE Access 2021, 9, 125787–125804. [Google Scholar] [CrossRef]
  40. Altan, A.; Karasu, S.; Zio, E. A new hybrid model for wind speed forecasting combining long short-term memory neural network, decomposition methods and grey wolf optimizer. Appl. Soft Comput. 2021, 100, 106996. [Google Scholar] [CrossRef]
  41. Trojovsky, P.; Dehghani, M. Pelican Optimization algorithm: A novel nature-inspired algorithm for engineering applications. Sensors 2022, 22, 855. [Google Scholar] [CrossRef]
  42. Kaur, G.; Arora, S. Chaotic whale optimization algorithm. J. Comput. Des. Eng. 2018, 5, 275–284. [Google Scholar] [CrossRef]
  43. Arora, S.; Sharma, M.; Anand, P. A Novel Chaotic Interior Search Algorithm for Global Optimization and Feature Selection. Appl. Artif. Intell. 2020, 34, 292–328. [Google Scholar] [CrossRef]
  44. Dragomiretskiy, K.; Zosso, D. Variational Mode Decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  45. Gai, J.; Shen, J.; Hu, Y.; Wang, H. An integrated method based on hybrid grey wolf optimizer improved variational mode decomposition and deep neural network for fault diagnosis of rolling bearing. Measurement 2020, 162, 107901. [Google Scholar] [CrossRef]
  46. Xu, H.; Lu, T.; Montillet, J.P.; He, X. An improved adaptive IVMD-WPT-based noise reduction algorithm on GPS height time series. Sensors 2022, 21, 8295. [Google Scholar] [CrossRef] [PubMed]
  47. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, 2039–2049. [Google Scholar] [CrossRef]
  48. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  49. Alamri, N.M.H.; Packianather, M.; Bigot, S. Optimizing the Parameters of Long Short-Term Memory Networks Using the Bees Algorithm. Appl. Sci. 2023, 13, 2536. [Google Scholar] [CrossRef]
  50. Subbiah, S.S.; Paramasivan, S.K.; Arockiasamy, K.; Senthivel, S.; Thangavel, M. Deep learning for wind speed forecasting using bi-lstm with selected features. Intell. Autom. Soft Comput. 2023, 35, 3824–3829. [Google Scholar]
  51. Wang, Y.; Vijayakumar, P.; Gupta, B.B.; Alhalabi, W.; Sivaraman, A. An improved entity recognition approach to cyber social knowledge provision of intellectual property using a CRF-LSTM model. Pattern Recogn. Lett. 2022, 163, 145–151. [Google Scholar] [CrossRef]
  52. Li, S.; Qin, D.; Wu, X.; Li, J.; Li, B.; Han, W. False Alert Detection Based on Deep Learning and Machine Learning. Int. J. Semant. Web Inf. Syst. 2022, 18, 297035. [Google Scholar] [CrossRef]
  53. Akl, N.A.; Khoury, J.E.; Mansour, C. Trip-based prediction of hybrid electric vehicles velocity using artifificial neural networks. In Proceedings of the 2021 IEEE 3rd International Multidisciplinary Conference on Engineering Technology (IMCET), Beirut, Lebanon, 8–10 December 2021; pp. 60–65. [Google Scholar]
  54. Ehteram, M.; Ahmed, A.N.; Khozani, Z.S.; El-Shafie, A. Graph convolutional network-Long short term memory neural network- multi layer perceptron- Gaussian progress regression model: A new deep learning model for predicting ozone concertation. Atmos. Pollut. Res. 2023, 14, 101766. [Google Scholar] [CrossRef]
  55. Diebold, F.X.; Mariano, R.S. Comparing predictive accuracy. J. Bus. Econ. Stat. 1995, 13, 134–144. [Google Scholar] [CrossRef]
Figure 1. Flowchart of MPOA.
Figure 1. Flowchart of MPOA.
Sustainability 17 02081 g001
Figure 2. Convergence characteristic curves for MPOA, POA, PSO, MFO, and WOA in solving the test function.
Figure 2. Convergence characteristic curves for MPOA, POA, PSO, MFO, and WOA in solving the test function.
Sustainability 17 02081 g002
Figure 3. LSTM structure diagram.
Figure 3. LSTM structure diagram.
Sustainability 17 02081 g003
Figure 4. The flowchart of VMD-MPOA-LSTM.
Figure 4. The flowchart of VMD-MPOA-LSTM.
Sustainability 17 02081 g004
Figure 5. Ultra-short-term WS data samples.
Figure 5. Ultra-short-term WS data samples.
Sustainability 17 02081 g005
Figure 6. VMD diagram of the original WS data.
Figure 6. VMD diagram of the original WS data.
Sustainability 17 02081 g006
Figure 7. The forecasting results of different models in Experiment I.
Figure 7. The forecasting results of different models in Experiment I.
Sustainability 17 02081 g007
Figure 8. Histogram of prediction results of different models in Experiment I.
Figure 8. Histogram of prediction results of different models in Experiment I.
Sustainability 17 02081 g008
Figure 9. The forecasting results of different models in Experiment II.
Figure 9. The forecasting results of different models in Experiment II.
Sustainability 17 02081 g009
Figure 10. Histogram of prediction results of different models in Experiment II.
Figure 10. Histogram of prediction results of different models in Experiment II.
Sustainability 17 02081 g010
Figure 11. The forecasting results of different models in Experiment III.
Figure 11. The forecasting results of different models in Experiment III.
Sustainability 17 02081 g011
Figure 12. Histogram of prediction results of different models in Experiment III.
Figure 12. Histogram of prediction results of different models in Experiment III.
Sustainability 17 02081 g012
Table 1. Unimodal test functions.
Table 1. Unimodal test functions.
Function IDNameFunctionDRange f o p t
F1Sphere f 1 x = i = z n x i 2 30 [ 100 ,   100 ] D 0
F2Schwefel 2.22 f 2 x = i = 1 n x i + i = 1 n x i 30 [ 10 ,   10 ] D 0
F3Schwefel 1.2 f 3 x = i = 1 n j = 1 i x j 2 30 [ 100 ,   100 ] D 0
F4Schwefel 2.21 f 4 x = max i x i , 1 i n 30 [ 100 ,   100 ] D 0
Table 2. Multimodal test functions.
Table 2. Multimodal test functions.
Function IDNameFunctionDRange f o p t
F5Rastrigin f 5 x = i = 1 n ( x i 2 10 cos ( 2 π x i ) + 10 ) 2 30 [ 5 . 12 ,   5 . 12 ] D 0
F6Griewank f 6 x = 1 4000 i = 1 n x i 100 2 i = 1 n cos x i 100 i + 1 30 [ 600 ,   600 ] D 0
Table 3. Low-dimension multimodal test functions.
Table 3. Low-dimension multimodal test functions.
Function IDNameFunctionDRange f o p t
F7Six Hump Camel f 7 ( x ) = 4 x 1 2 2.1 x 1 4 + 1 3 x 1 6 + x 1 x 2 4 x 2 2 + 4 x 2 4 2 [ 5 ,   5 ] D −1.0316
F8Hartman 6 f 8 x = i = 1 4 exp j = 1 6 a i j x j p i j 2 6 [ 0 ,   1 ] D −3.322
F9Shekel 10 f 9 x = i = 1 10 x i a i x i a i T + c i 1 4 [ 0 ,   10 ] D −10.5364
Table 4. Parameter values used in MPOA, POA, PSO, MFO, and WOA.
Table 4. Parameter values used in MPOA, POA, PSO, MFO, and WOA.
AlgorithmParameterValue
MPOAPopulation size30
R0.2
POAPopulation size30
R0.2
PSOPopulation size30
Inertia weight0.9
Cognitive coefficient2
Social coefficient2
MFOPopulation size30
Logarithmic spiral shape constant1
WOAPopulation size30
Control parameter 2 ( 1 t / max _ t )
Table 5. A comparison of different methods in solving the unimodal test functions in Table 1 at 30 dimensions.
Table 5. A comparison of different methods in solving the unimodal test functions in Table 1 at 30 dimensions.
FStatisticsMPOAPOAPSOMFOWOA
F1Best01.8765 × 10−23117.312392787.25274 × 10−341.3236 × 10−169
Worst01.3643 × 10−208343.1260393.16648 × 10−293.8854 × 10−145
Average04.5568 × 10−21094.12268654.48706 × 10−301.2951 × 10−146
Median03.0295 × 10−22186.660746468.94626 × 10−311.8246 × 10−156
Std.0075.898119297.3453 × 10−307.0937 × 10−146
F2Best02.7192 × 10−1183.6995479482.05632 × 10−201.7078 × 10−113
Worst08.2946 × 10−10585.698930536.44423 × 10−183.6239 × 10−101
Average02.8073 × 10−10627.650893958.30673 × 10−191.2397 × 10−102
Median02.253 × 10−11022.473492692.98559 × 10−194.1401 × 10−108
Std.01.5137 × 10−10520.408063671.3558 × 10−186.6122 × 10−102
F3Best01.1709 × 10−232492.594232.07228 × 10−11425.2057722
Worst07.6844 × 10−20429,869.916921.87939 × 10−540,417.33133
Average02.5833 × 10−2056185.8737178.69467 × 10−723,019.36643
Median02.6954 × 10−2193386.6451854.84279 × 10−825,775.38504
Std.007591.6491863.42518 × 10−610,583.178
F4Best01.9984 × 10−1132.2155666294.33564 × 10−60.806799875
Worst03.0929 × 10−1049.67960543112.3129093792.54653222
Average02.4403 × 10−1056.13323032.21840869947.9021703
Median01.5409 × 10−1096.0088529780.11077977846.6429587
Std.07.7213 × 10−1051.9090895163.79595887829.40424297
Table 6. A comparison of different methods in solving the multimodal test functions in Table 2 at 30 dimensions.
Table 6. A comparison of different methods in solving the multimodal test functions in Table 2 at 30 dimensions.
FStatisticsMPOAPOAPSOMFOWOA
F5Best0071.218884172.9848771710
Worst00207.606701360.763267120
Average00123.822184621.867708860
Median00123.551323716.914288860
Std.0033.1160173214.990495990
F6Best001.1586713860.0590896110
Worst002.1466167740.3717984710.064446821
Average001.5951051210.1572438790.003780121
Median001.4918650570.1218577730
Std.000.3049026540.08731310.014528786
Table 7. A comparison of different methods in solving the low-dimension multimodal test functions in Table 3.
Table 7. A comparison of different methods in solving the low-dimension multimodal test functions in Table 3.
FStatisticsMPOAPOAPSOMFOWOA
F7Best−1.03162845−1.03162845−1.03162845−1.03162845−1.03162845
Worst−1.03162845−1.03162845−1.03162845−1.03162845−1.03162186
Average−1.03162845−1.03162845−1.03162845−1.03162845−1.03162801
Median−1.03162845−1.03162845−1.03162845−1.03162845−1.03162842
Std.01.20021× 10−102.0215× 10−1001.27858× 10−6
F8Best−3.32197715−3.32197824−3.32199509−3.32199512−3.3176645
Worst−3.32128798−3.28912155−2.80400500−3.19746545−2.92189763
Average−3.32180723−3.32071225−3.16013744−3.25419394−3.14570194
Median−3.32185384−3.32184033−3.13764150−3.20310192−3.13507813
Std.0.0001550270.005968650.150602820.060307460.11223570
F9Best−10.5363955−10.5363807−10.5364098−10.5364098−10.5091620
Worst−10.3922192−5.12832512−2.42171064−2.42733518−1.83567256
Average−10.5267121−10.3512593−7.26983991−8.60482899−4.63411539
Median−10.5330304−10.5327092−10.5159525−10.5364076−4.09295008
Std.0.0257872320.9864771063.6196544273.2960455762.478088566
Table 8. The parameter values of each model.
Table 8. The parameter values of each model.
Model NameParameter NameParameter Value
BPLearning rate 0.05
Epoch100
SVMKernel function coefficient Gamma0.01
Penalty coefficient C80
ELMNode number20
LSTMLearning rate0.01
The maximum number of iterations100
The number of nodes in the first layer8
The number of nodes in the second layer8
Optimization algorithm + LSTMLower bound lb1, 1, 0.0001, 1
Upper bound ub100, 100, 0.01, 100
Table 9. Evaluation index values of predicted results of different models in Experiment I.
Table 9. Evaluation index values of predicted results of different models in Experiment I.
ModelMAE (m/s)RMSE (m/s)MAPE (%)R2TICEC
EMD-LSTM0.79331.00412.950.91710.06030.9029
EEMD-LSTM0.67990.844310.160.94380.05050.9313
VMD-LSTM0.58610.76298.260.96390.04580.9439
MPOAVMD-LSTM0.5450.71297.670.97250.04280.951
Proposed model0.40470.5396.490.98030.03190.972
Table 10. The DM values of different models in Experiment I.
Table 10. The DM values of different models in Experiment I.
ModelEMD-LSTMEEMD-LSTMVMD-LSTMMPOAVMD-LSTM
DM4.8888 *5.7897 *3.5223 *3.0579 *
* denotes 1% significance level.
Table 11. Evaluation index values of predicted results of different models in Experiment II.
Table 11. Evaluation index values of predicted results of different models in Experiment II.
ModelMAE (m/s)RMSE (m/s)MAPE (%)R2TICEC
PSO-LSTM1.33791.689319.760.735040.101940.7153
WOA-LSTM1.17321.49317.210.793370.0897170.7851
MFO-LSTM1.24751.573918.070.790720.0957550.7612
POA-LSTM1.22861.551117.680.791940.0941340.7681
MPOA-LSTM1.15711.48417.010.797950.0893470.7877
Table 12. Evaluation index values of predicted results of different models in Experiment III.
Table 12. Evaluation index values of predicted results of different models in Experiment III.
ModelMAE (m/s)RMSE (m/s)MAPE (%)R2TICEC
ARIMA1.60152.038722.340.72580.12930.5876
BP1.20931.552618.130.78980.09420.7677
ELM1.15631.501417.320.79090.09020.7828
SVM1.12481.430617.060.80340.08480.8028
LSTM1.18581.510917.670.78850.09080.78
WOA-SVM1.11651.425416.930.80650.08490.8042
VMD-MPOA-BP0.54630.70387.380.97260.04240.9523
VMD-MPOA-ELM0.52280.67447.960.95850.03970.9493
VMD-WOA-SVM0.49840.62289.010.97560.0360.9702
VMD-MPOA-SVM0.46350.57688.630.97790.0330.971
VMD-WOA-LSTM0.44990.56517.860.97810.03210.9718
Proposed model0.40470.5396.490.98030.03190.972
Table 13. The DM values of different models in Experiment III.
Table 13. The DM values of different models in Experiment III.
ModelARIMABPELMSVMLSTMWOA-SVMVMD-MPOA-BPVMD-MPOA-ELMVMD-WOA-SVMVMD-MPOA-SVMVMD-WOA-LSTM
DM8.66 *6.32 *7.27 *7.69 *7.16 *7.73 *2.99 *2.78 *2.49 **2.09 **1.81 ***
*, **, and *** denote 1%, 5%, and 10% significance levels, respectively.
Table 14. Improvement percentage of the proposed model.
Table 14. Improvement percentage of the proposed model.
ModelPMAEPRMSEPMAPEPR2PTICPEC
ARIMA74.73%73.56%70.95%35.06%75.33%65.41%
BP66.53%65.28%64.18%24.12%66.10%26.62%
ELM65.00%64.10%62.49%23.95%64.59%24.18%
SVM64.02%62.32%61.91%22.02%62.35%21.08%
LSTM65.87%64.32%63.22%24.32%64.81%24.62%
PSO-LSTM69.75%68.09%67.11%33.36%68.66%35.87%
WOA-LSTM65.50%63.90%62.24%23.56%64.39%23.80%
MFO_LSTM67.56%65.75%64.04%23.97%66.64%27.69%
POA-LSTM67.06%65.25%63.24%23.78%66.06%26.54%
MPOA-LSTM65.02%63.68%61.82%22.85%64.25%23.39%
WOA-SVM63.75%62.18%61.64%21.55%62.38%20.87%
EMD-LSTM48.98%46.31%49.81%6.89%47.04%7.66%
EEMD-LSTM40.47%36.16%36.08%3.86%36.71%4.37%
VMD-LSTM30.95%29.34%21.33%1.70%30.30%2.98%
MPOAVMD-LSTM25.74%24.39%15.33%0.81%25.42%2.21%
VMD-MPOA-BP25.91%23.42%11.90%0.78%24.66%2.07%
VMD-MPOA-ELM22.59%20.08%18.48%2.27%19.65%2.39%
VMD-WOA-SVM18.80%13.46%27.93%0.48%11.32%0.19%
VMD-MPOA-SVM12.68%6.55%24.79%0.24%3.33%0.1%
VMD-WOA-LSTM10.05%4.62%17.43%0.22%0.62%0.02%
Table 15. The simulation results of further step forecasting.
Table 15. The simulation results of further step forecasting.
ModelMAE (m/s)RMSE (m/s)MAPE (%)
1-Step2-Step3-Step1-Step2-Step3-Step1-Step2-Step3-Step
BP1.20931.43121.57931.55261.86572.123018.1320.6723.17
ELM1.15631.29311.58071.50141.60171.956217.3219.5622.86
SVM1.12481.23611.30931.43061.62651.793317.0618.6620.30
LSTM1.18581.30171.49571.51091.71131.965717.6718.5622.91
WOA-SVM1.11651.21791.28761.42541.58781.773716.9317.6819.01
EMD-LSTM0.79330.93170.98561.00402.12362.856212.9514.6515.11
EEMD-LSTM0.67990.81290.84790.84430.91780.986610.1611.5612.65
VMD-LSTM0.58610.68260.70310.76290.85210.91638.269.0410.56
VMD-MPOA-BP0.54630.60120.65320.703820.77210.83127.388.179.31
VMD-MPOA-ELM0.52280.59570.66090.67440.78350.80097.968.379.56
VMD-WOA-SVM0.49840.55650.60260.62280.68190.77969.0110.8211.78
VMD-MPOA-SVM0.46350.53270.58910.57680.65620.78868.639.5611.04
VMD-WOA-LSTM0.44990.49830.55370.56510.64490.73127.867.839.16
Proposed model0.40470.45390.50110.53900.60010.65216.497.117.65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, L.; Xu, C.; Ai, X.; Han, X.; Xue, F. A Combined Forecasting Model Based on a Modified Pelican Optimization Algorithm for Ultra-Short-Term Wind Speed. Sustainability 2025, 17, 2081. https://doi.org/10.3390/su17052081

AMA Style

Guo L, Xu C, Ai X, Han X, Xue F. A Combined Forecasting Model Based on a Modified Pelican Optimization Algorithm for Ultra-Short-Term Wind Speed. Sustainability. 2025; 17(5):2081. https://doi.org/10.3390/su17052081

Chicago/Turabian Style

Guo, Lei, Chang Xu, Xin Ai, Xingxing Han, and Feifei Xue. 2025. "A Combined Forecasting Model Based on a Modified Pelican Optimization Algorithm for Ultra-Short-Term Wind Speed" Sustainability 17, no. 5: 2081. https://doi.org/10.3390/su17052081

APA Style

Guo, L., Xu, C., Ai, X., Han, X., & Xue, F. (2025). A Combined Forecasting Model Based on a Modified Pelican Optimization Algorithm for Ultra-Short-Term Wind Speed. Sustainability, 17(5), 2081. https://doi.org/10.3390/su17052081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop