Improved MNet-Atten Electric Vehicle Charging Load Forecasting Based on Composite Decomposition and Evolutionary Predator–Prey and Strategy
Abstract
1. Introduction
- (1)
- The load data are denoised by the composite decomposition algorithm to eliminate the pseudo components, which are based on the characteristics of electric vehicle charging load data. Meanwhile, the intrinsic mode functions and residual components with different amplitudes and frequencies are obtained.
- (2)
- The MNet-Atten prediction model is designed and implemented. The convolutional module is employed to capture local dependencies within the sequences, while recurrent and skip-connection modules are utilized to extract both long term and short-term temporal features, thereby enhancing the inherent predictability of the data.
- (3)
- By employing the EPPS algorithm, which possesses strong optimization capabilities and rapid convergence speed, to mine the intrinsic features of the MNet-Atten network, the learning rate is optimized and performs superposition calculations for load forecasting, thereby enhancing prediction accuracy.
2. Composite Decomposition
2.1. Singular Value Decomposition Model Based on Matrix Operations
2.2. Complementary Ensemble Empirical Mode Decomposition
3. Modeling of MNet-Atten Network
3.1. Convolution Module
3.2. Recurrent and Recurrent-Skip Module
3.3. Modeling of Temporal Pattern Attention
3.4. Modeling of Autoregressive Module
4. MNet-Atten Based on Optimization of Evolutionary Predator and Prey Strategy
5. Model and Case Validation
5.1. Data Description and Data Pre-Processing
5.2. Data Decomposition Based on Composite Decomposition
5.3. Comparative Study and Analysis of Different Load Decomposition Methods
5.4. Analysis of Simulation Results
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Wang, C.; Wang, Y.; Song, F. Research on Electric Vehicle Charging Load Forecasting Method Based on Improved LSTM Neural Network. World Electr. Veh. J. 2025, 16, 265. [Google Scholar] [CrossRef]
- Ma, P.; Cui, S.; Chen, M.; Zhou, S.; Wang, K. Review of family-level short-term load forecasting and its application in household energy management system. Energies 2023, 16, 5809. [Google Scholar] [CrossRef]
- Habbak, H.; Mahmoud, M.; Metwally, K.; Fouda, M.; Ibrahem, M. Load forecasting techniques and their applications in smart grids. Energies 2023, 16, 1480. [Google Scholar] [CrossRef]
- Dewangan, F.; Abdelaziz, A.Y.; Biswal, M. Load forecasting models in smart grid using smart meter information: A review. Energies 2023, 16, 1404. [Google Scholar] [CrossRef]
- Tarmanini, C.; Sarma, N.; Gezegin, C.; Ozgonenel, O. Short term load forecasting based on ARIMA and ANN approaches. Energy Rep. 2023, 9, 550–557. [Google Scholar] [CrossRef]
- Wan, A.; Chang, Q.; Khalil, A.; He, J. Short-term power load forecasting for combined heat and power using CNN-LSTM enhanced by attention mechanism. Energy 2023, 282, 128274. [Google Scholar] [CrossRef]
- Wei, N.; Yin, C.; Yin, L.; Tan, J.; Liu, J.; Wang, S.; Qiao, W.; Zeng, F. Short-term load forecasting based on WM algorithm and transfer learning model. Appl. Energy 2024, 353, 122087. [Google Scholar] [CrossRef]
- Sun, Y.; Tian, X.; Bao, W.; Qu, S.; Li, Q.; Chen, Y.; Shi, P. Improving the forecast performance of hydrological models using the cubature Kalman filter and unscented Kalman filter. Water Res. Res. 2023, 59, e2022WR033580. [Google Scholar] [CrossRef]
- Junior, M.; Freire, R.; Seman, L.; Stefenon, S.; Mariani, V.; Coelho, L. Optimized hybrid ensemble learning approaches applied to very short-term load forecasting. Int. J. Electr. Power Energy Syst. 2024, 155, 109579. [Google Scholar] [CrossRef]
- Bartwal, D.; Sindhwani, R.; Vaidya, O. Improving forecast accuracy for seasonal products in FMCG industry: Integration of SARIMA and regression model. Int. J. Ind. Sys. Eng. 2024, 46, 259–279. [Google Scholar] [CrossRef]
- Zhang, J. Utilizing gaussian process regression model enhanced by metaheuristic algorithms to forecast undrained shear strength. J. Appl. Sci. Eng. 2025, 28, 741–752. [Google Scholar] [CrossRef]
- Bacanin, N.; Stoean, C.; Zivkovic, M. On the benefits of using metaheuristics in the hyperparameter tuning of deep learning models for energy load forecasting. Energies 2023, 16, 1434. [Google Scholar] [CrossRef]
- Abedinia, O.; Amjady, N.; Shafie-Khah, M. Electricity price forecast using combinatorial neural network trained by a new stochastic search method. Energy Convers. Manag. 2015, 105, 642–654. [Google Scholar] [CrossRef]
- Huang, N.; Wang, S.; Wang, R.; Cai, G.; Liu, Y.; Dai, Q. Gated spatial-temporal graph neural network based short-term load forecasting for wide-area multiple buses. Int. J. Electr. Power Energy Syst. 2023, 145, 108651. [Google Scholar] [CrossRef]
- Ullah, K.; Ahsan, M.; Hasanat, S.; Haris, M.; Yousaf, H.; Raza, S. Short-term load forecasting: A comprehensive review and simulation study with CNN-LSTM hybrids approach. IEEE Access 2024, 12, 111858–111881. [Google Scholar] [CrossRef]
- Wazirali, R.; Yaghoubi, E.; Abujazar, M.; Rami, A.; Vakili, A. State-of-the-art review on energy and load forecasting in microgrids using artificial neural networks, machine learning, and deep learning techniques. Electr. Power Syst. Res. 2023, 225, 109792. [Google Scholar] [CrossRef]
- Lotfipoor, A.; Patidar, S.; Jenkins, D. Deep neural network with empirical mode decomposition and Bayesian optimisation for residential load forecasting. Expert Syst. Appl. 2024, 237, 121355. [Google Scholar] [CrossRef]
- Zhang, Q.; Wu, J.; Ma, Y.; Li, G.; Ma, J.; Wang, C. Short-term load forecasting method with variational mode decomposition and stacking model fusion. Sus. Energy Grids Netw. 2022, 30, 100622. [Google Scholar] [CrossRef]
- Liu, Q.; Cao, J.; Zhang, J.; Zhong, Y.; Ba, T.; Zhang, Y. Short-term power load forecasting in FGSM-Bi-LSTM networks based on empirical wavelet transform. IEEE Access 2023, 11, 105057–105068. [Google Scholar] [CrossRef]
- Shi, J.; Teh, J. Load forecasting for regional integrated energy system based on complementary ensemble empirical mode decomposition and multi-model fusion. Appl. Energy 2024, 353, 122146. [Google Scholar] [CrossRef]
- Li, K.; Duan, P.; Cao, X.; Cheng, Y.; Zhao, B.; Xue, Q.; Feng, M. A multi-energy load forecasting method based on complementary ensemble empirical model decomposition and composite evaluation factor reconstruction. Appl. Energy 2024, 365, 123283. [Google Scholar] [CrossRef]
- Xiang, Y.; Zhuang, X.H. Application of ARIMA model in short-term prediction of international crude oil price. Adv. Mater. Res. 2013, 798, 979–982. [Google Scholar] [CrossRef]
- Murat, A.; Tokat, E. Forecasting oil price movements with crack spread futures. Energy Econ. 2009, 31, 85–90. [Google Scholar] [CrossRef]
- Nugroho, D.B.; Setiawan, A.; Morimoto, T. Modelling and Forecasting Financial Volatility with Realized GARCH Model: A Comparative Study of Skew-t Distributions Using GRG and MCMC Methods. Econometrics 2025, 13, 33. [Google Scholar] [CrossRef]
- Cavicchioli, M. Forecasting Markov switching vector autoregressions: Evidence from simulation and application. J. Forecast. 2025, 44, 136–152. [Google Scholar] [CrossRef]
- Pukach, A.; Teslyuk, V.; Lysa, N.; Sikora, L.; Fedyna, B. Information-Cognitive Concept of Predicting Method for HCI Objects’ Perception Subjectivization Results Based on Impact Factors Analysis with Usage of Multilayer Perceptron ANN. Appl. Sci. 2025, 15, 9763. [Google Scholar] [CrossRef]
- Jedrzejczyk, A.; Firek, K.; Rusek, J. Prediction of damage intensity to masonry residential buildings with convolutional neural network and support vector machine. Sci. Rep. 2024, 14, 16256. [Google Scholar] [CrossRef]
- Li, X.; Wang, J. Predicting Constitutive Behaviour of Idealized Granular Soils Using Recurrent Neural Networks. Appl. Sci. 2025, 15, 9495. [Google Scholar] [CrossRef]
- Jang, M.; Joo, S.-K. Pattern-Aware BiLSTM Framework for Imputation of Missing Data in Solar Photovoltaic Generation. Energies 2025, 18, 4734. [Google Scholar] [CrossRef]
- Jha, A.; Ray, D.; Sarkar, D.; Prakash, T.; Dewangan, N. Bidirectional long-short-term memory-based fractional power system stabilizer: Design, simulation, and real-time validation. Int. J. Numer. Model. Electron. Netw. Devices Fields 2024, 37, 3300. [Google Scholar] [CrossRef]
- Chen, J.; Wu, Q.; Ji, T.; Wu, P.; Li, M. Evolutionary predator and prey strategy for global optimization. Inf. Sci. 2016, 327, 217–232. [Google Scholar] [CrossRef]
- Wang, L.; Xie, M.; Pan, M.; He, F.; Yang, B.; Gong, Z.; Wu, X.; Shang, M.; Shan, K. Improved Deep Learning Predictions for Chlorophyll Fluorescence Based on Decomposition Algorithms: The Importance of Data Preprocessing. Water 2023, 15, 4104. [Google Scholar] [CrossRef]
- Chen, L.Y.; Wang, C.; Xiao, X.Y.; Ren, C.; Zhang, D.J. Denoising in SVD-based ghost imaging. Opt. Express 2022, 30, 6248–6257. [Google Scholar] [CrossRef]
- Zeng, S.; Cui, J.; Luo, D.; Lu, N. Bridge Damage Identification Using Time-Varying Filtering-Based Empirical Mode Decomposition and Pre-Trained Convolutional Neural Networks. Sensors 2025, 25, 4869. [Google Scholar] [CrossRef]
- Guo, Y.; Si, J.; Wang, Y. Ensemble-Empirical-Mode-Decomposition (EEMD) on SWH prediction: The effect of decomposed IMFs, continuous prediction duration, and data-driven models. Ocean Eng. 2025, 324, 120755. [Google Scholar] [CrossRef]
- Xu, Y.; Li, H.; Meng, X.; Chen, J.; Zhang, X.; Peng, T. An Energy System Modeling Approach for Power Transformer Oil Temperature Prediction Based on CEEMD and Robust Deep Ensemble RVFL. Processes 2025, 13, 2487. [Google Scholar] [CrossRef]
- Yang, X.; Meng, P.; Jiang, Z.; Zhou, L. Deep siamese residual support vector machine with applications to disease prediction. Comput. Biol. Med. 2025, 196, 110693. [Google Scholar] [CrossRef] [PubMed]
- Peng, Y.; Kai, J.; Yu, X.; Zhang, Z.; Li, Q.J.; Yang, G.; Kong, L. Pavement Friction Prediction Based Upon Multi-View Fractal and the XGBoost Framework. Lubricants 2025, 13, 391. [Google Scholar] [CrossRef]
- Fauzan, A.N.; Assahari, M.S.; Jainun, A.R.; Somantri. Backpropagation Neural Network Algorithm for Optimizing Network Bandwidth Allocation Based on User Access Patterns. Eng. Proc. 2025, 107, 56. [Google Scholar] [CrossRef]
- Han, H.; Peng, J.; Ma, J.; Liu, H.; Liu, S. Research on Load Forecasting Prediction Model Based on Modified Sand Cat Swarm Optimization and SelfAttention TCN. Symmetry 2025, 17, 1270. [Google Scholar] [CrossRef]
- Li, H.; Li, S.; Li, H.; Bai, L. Ultra-Short-Term Wind Power Prediction Based on Fused Features and an Improved CNN. Processes 2025, 13, 2236. [Google Scholar] [CrossRef]
- Sun, R.; Huang, Z.; Liang, X.; Zhu, S.; Li, H. A GRU-Enhanced Kolmogorov–Arnold Network Model for Sea Surface Temperature Prediction Derived from Satellite Altimetry Product in South China Sea. Remote Sens. 2025, 17, 2916. [Google Scholar] [CrossRef]
Model | Reference | Category | Advantages | Disadvantages |
---|---|---|---|---|
AutoRegressive Integrated Moving Average Model (ARIMA) | [22] | Traditional Statistical Model | 1. Solid theoretical foundation, clear statistical properties, and strong interpretability. 2. Excellent ability to capture linear relationships, especially suitable for stationary time series with obvious trends and seasonality. 3. Model parameters have clear statistical meanings. 4. Serves as a benchmark model for handling univariate time series. | 1. Only applicable to univariate time series and cannot directly utilize multivariate information. 2. Requires the series to be stationary or stationary after differencing, demanding high-quality data preprocessing. 3. Assumes a linear relationship between series, making it ineffective in capturing complex nonlinear patterns. 4. The process of model identification and order determination requires certain experience and statistical knowledge. |
Random Walk (RW) | [23] | Traditional Statistical Model | 1. Extremely simple model with no need for parameter estimation. 2. Acts as an effective benchmark model for many financial time series and is often difficult to outperform. 3. Embodies the philosophical idea that “the future is unpredictable” and serves as a crucial tool for testing the effectiveness of forecasting models. | 1. Extremely low practicality except when used as a benchmark. It simply uses the current value as the best prediction for the next moment. 2. Predictions form a horizontal line, making it unable to forecast trends and fluctuations. 3. The variance of prediction errors increases infinitely over time. |
Generalized Autoregressive Conditional Heteroskedasticity Model (GARCH) | [24] | Traditional Statistical Model | 1. Specifically designed to characterize and forecast volatility (variance), rather than the series values themselves. 2. Can accurately capture volatility clustering (large fluctuations followed by large fluctuations, small fluctuations followed by small fluctuations) in financial time series. 3. Provides a key tool for fields such as risk management and option pricing. | 1. Mainly used for volatility modeling and usually needs to be combined with models such as ARIMA (ARIMA-GARCH) to simultaneously forecast mean and variance. 2. Sensitive to model parameters and has a relatively complex setup. 3. Also based on the linear assumption, which may lead to insufficient forecasting of extreme events. |
Vector Autoregression Model (VAR) | [25] | Traditional Statistical Model | 1. Naturally handles multivariate time series and can capture bidirectional dynamic relationships between variables. 2. Concise model form, where all variables are treated equally, facilitating estimation and understanding. 3. Can be used to analyze Granger causality and impulse response between variables. | 1. A large number of parameters. When the number of variables is K and the lag order is p, the number of parameters is K2p, which easily leads to overfitting (the “curse of dimensionality”). 2. Also requires the system to be stationary and relies heavily on the linear relationship assumption. 3. Highly sensitive to the selection of predictive variables. |
Artificial Neural Network (ANN) | [26] | Machine Learning Model | 1. Follows the universal approximation theorem, enabling it to approximate any complex nonlinear mapping relationship with arbitrary precision. 2. Has loose requirements on data distribution assumptions, without the need for strict conditions such as stationarity required by traditional models. 3. Strong robustness to noisy data. | 1. A “black-box” model with extremely poor interpretability, making it difficult to understand its internal decision-making logic. 2. Requires a large amount of data for training and is prone to overfitting when the data volume is small. 3. The training process involves large computational costs, and parameter tuning (network structure, learning rate, etc.) is complex, requiring numerous experiments. |
Support Vector Machine (SVM) | [27] | Machine Learning Model | 1. Based on the principle of structural risk minimization, it has strong generalization ability, performs well on small-sample datasets, and is not prone to overfitting. 2. Can efficiently handle nonlinear problems through kernel tricks. 3. Capable of finding global optimal solutions rather than local optimal ones. | 1. Training speed decreases sharply with the increase in data volume, and it consumes a great deal of memory, making it unsuitable for ultra-large-scale datasets. 2. Sensitive to the selection of parameters, so parameter tuning is crucial. 3. Essentially a binary classification model; its performance and efficiency decline when used for regression (SVR) and multi-class classification tasks. 4. The interpretability of results remains relatively poor. |
Recurrent Neural Network (RNN) | [28] | Deep Learning Model | 1. Specifically designed for sequence data, with memory functionality, enabling it to make predictions using information from all historical time steps. 2. Parameter sharing mechanism significantly reduces the number of model parameters. 3. Theoretically capable of capturing temporal dependencies of any length. | 1. Suffers from the problem of gradient vanishing/explosion, making it difficult to effectively learn long term dependencies (e.g., when information from a long time ago has a significant impact on current predictions, it cannot capture this effectively). 2. The training and computation process is sequential, making parallelization difficult and resulting in slow training speed. |
LSTM (Long Short-Term Memory Network) | [29] | Deep Learning Model | 1. Perfectly solves the gradient vanishing problem of RNN, enabling effective learning of long term dependencies and serving as a milestone model in time series forecasting. 2. Precisely controls the flow and memory of information through the “gating” mechanism (input gate, forget gate, output gate). | 1. The model is highly complex with a large number of parameters, leading to high training costs (in terms of time and computing resources). 2. Requires an extremely large amount of data to train an effective model; otherwise, severe overfitting will occur. |
BiLSTM (Bidirectional LSTM) | [30] | Deep Learning Model | BiLSTM can simultaneously utilize contextual information from the past and future, resulting in stronger feature extraction capabilities and often higher accuracy. | 1. The model is highly complex with a large number of parameters, leading to high training costs (in terms of time and computing resources). 2. Requires an extremely large amount of data to train an effective model; otherwise, severe overfitting will occur. |
1: Initialize model parameters; set the initialization parameters for the EPPS algorithm. |
2: Train the MNet-Atten network. |
3: Return fitness values to EPPS. |
4: Utilize the search mechanism to update population individuals. |
5: Reach the maximum number of iterations or the fitness function. |
6: Enter the optimal learning rate. |
7: Extracts both short- and long term dependencies within the data. |
8: Employ a temporal attention mechanism to focus on key sequences while filtering out distracting elements. |
9: Extract the linear features of the sequence from the autoregressive model. |
10: Combine the results of the nonlinear and linear components to obtain the forecast outcome. |
11: Calculate the error between the output value and the target value. |
12: Meet the requirements of the error and obtain the forecast results. |
Method | MAPE | RMSE | PCC |
---|---|---|---|
Without the decomposition method | 2.1344 | 6.7538 | 95.3642 |
EMD decomposition method | 1.7927 | 5.4967 | 96.3758 |
EEMD decomposition method | 1.6086 | 4.5416 | 97.6602 |
CEEMD decomposition method | 1.4655 | 3.9805 | 98.5126 |
The proposed decomposition method | 0.9295 | 1.8219 | 99.2387 |
Method | MAPE | RMSE | PCC |
---|---|---|---|
SVM | 3.4681 | 6.5304 | 96.5514 |
XGBoost | 2.7236 | 5.6927 | 97.0046 |
BPNN | 2.4015 | 4.3651 | 97.2893 |
TCN | 1.9122 | 3.2798 | 98.0904 |
CNN | 2.2467 | 5.3286 | 97.5033 |
GRU | 1.6839 | 2.8803 | 98.2237 |
The proposed method | 0.9295 | 1.8219 | 99.2387 |
Model | MAPE | RMSE | PCC |
---|---|---|---|
MNet-Atten | 1.8012 | 5.1258 | 98.2326 |
EPPS-MNet-Atten | 1.4326 | 4.9864 | 98.4579 |
CEEMD-MNet-Atten | 1.2435 | 3.8615 | 98.5744 |
CEEMD-SVD-MNet-Atten | 1.0611 | 3.4127 | 98.7803 |
CEEMD-EPPS-MNet-Atten | 0.9809 | 2.4591 | 99.0897 |
CEEMD-SVD-EPPS-MNet-Atten | 0.9295 | 2.1586 | 99.2387 |
Model | MAPE | RMSE | PCC |
---|---|---|---|
CEEMD-SVD-PSO-MNet-Atten | 0.9498 | 2.1979 | 98.8721 |
CEEMD-SVD-GA-MNet-Atten | 0.9376 | 2.1781 | 99.1417 |
CEEMD-SVD-GSO-MNet-Atten | 0.9333 | 2.1784 | 99.2109 |
CEEMD-SVD-EPPS-MNet-Atten | 0.9295 | 2.1586 | 99.2387 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Published by MDPI on behalf of the World Electric Vehicle Association. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wei, X.; Jiang, Q.; Xia, H.; Kong, X. Improved MNet-Atten Electric Vehicle Charging Load Forecasting Based on Composite Decomposition and Evolutionary Predator–Prey and Strategy. World Electr. Veh. J. 2025, 16, 564. https://doi.org/10.3390/wevj16100564
Wei X, Jiang Q, Xia H, Kong X. Improved MNet-Atten Electric Vehicle Charging Load Forecasting Based on Composite Decomposition and Evolutionary Predator–Prey and Strategy. World Electric Vehicle Journal. 2025; 16(10):564. https://doi.org/10.3390/wevj16100564
Chicago/Turabian StyleWei, Xiaobin, Qi Jiang, Huaitang Xia, and Xianbo Kong. 2025. "Improved MNet-Atten Electric Vehicle Charging Load Forecasting Based on Composite Decomposition and Evolutionary Predator–Prey and Strategy" World Electric Vehicle Journal 16, no. 10: 564. https://doi.org/10.3390/wevj16100564
APA StyleWei, X., Jiang, Q., Xia, H., & Kong, X. (2025). Improved MNet-Atten Electric Vehicle Charging Load Forecasting Based on Composite Decomposition and Evolutionary Predator–Prey and Strategy. World Electric Vehicle Journal, 16(10), 564. https://doi.org/10.3390/wevj16100564