Next Article in Journal
High-Precision Low-Speed Measurement for Permanent Magnet Synchronous Motors Using an Improved Extended State Observer
Previous Article in Journal
Technical Architecture and Control Strategy for Residential Community Orderly Charging Based on an Active Reservation Mechanism for Unconnected Charging Pile
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Machine Learning Techniques for Battery State of Health Prediction: A Comparative Review

1
Department of Electrical Engineering, Nelson Mandela University, P.O. Box 77000, Gqeberha 6031, South Africa
2
Department of Mechanical Engineering, Nelson Mandela University, P.O. Box 77000, Gqeberha 6031, South Africa
*
Authors to whom correspondence should be addressed.
World Electr. Veh. J. 2025, 16(11), 594; https://doi.org/10.3390/wevj16110594
Submission received: 8 September 2025 / Revised: 17 October 2025 / Accepted: 25 October 2025 / Published: 28 October 2025
(This article belongs to the Section Storage Systems)

Abstract

Accurate estimation of the state of health (SOH) of lithium-ion batteries is essential for the safe and efficient operation of electric vehicles (EVs). Conventional approaches, including Coulomb counting, electrochemical impedance spectroscopy, and equivalent circuit models, provide useful insights but face practical limitations such as error accumulation, high equipment requirements, and limited applicability across different conditions. These challenges have encouraged the use of machine learning (ML) methods, which can model nonlinear relationships and temporal degradation patterns directly from cycling data. This paper reviews four machine learning algorithms that are widely applied in SOH estimation: support vector regression (SVR), random forest (RF), convolutional neural networks (CNNs), and long short-term memory networks (LSTMs). Their methodologies, advantages, limitations, and recent extensions are discussed with reference to the existing literature. To complement the review, MATLAB-based simulations were carried out using the NASA Prognostics Center of Excellence (PCoE) dataset. Training was performed on three cells (B0006, B0007, B0018), and testing was conducted on an unseen cell (B0005) to evaluate cross-battery generalisation. The results show that the LSTM model achieved the highest accuracy (RMSE = 0.0146, MAE = 0.0118, R2 = 0.980), followed by CNN and RF, both of which provided acceptable accuracy with errors below 2% SOH. SVR performed less effectively (RMSE = 0.0457, MAPE = 4.80%), reflecting its difficulty in capturing sequential dependencies. These outcomes are consistent with findings in the literature, indicating that deep learning models are better suited for modelling long-term battery degradation, while ensemble approaches such as RF remain competitive when supported by carefully engineered features. This review also identifies ongoing and future research directions, including the use of optimisation algorithms for hyperparameter tuning, transfer learning for adaptation across battery chemistries, and explainable AI to improve interpretability. Overall, LSTM and hybrid models that combine complementary methods (e.g., CNN-LSTM) show strong potential for deployment in battery management systems, where reliable SOH prediction is important for safety, cost reduction, and extending battery lifetime.

1. Introduction

The growing dependence on batteries for electric vehicles—with global electric vehicle (EV) sales projected to reach 30 million annually by 2030 [1]—has made it crucial to precisely measure and forecast battery health. The state of health (SOH) is usually defined as the ratio (percentage) of the battery’s current available capacity to the initial rated capacity [2]. The precise measurement of SOH is vital to achieving maximum battery efficiency and maximising the life and safety of all its applications.
Multiple factors influence battery SOH, including temperature, charge/discharge cycles, depth of discharge (DoD), and operating conditions [3]. High temperatures within a battery accelerate chemical reactions such as electrolyte decomposition and electrode material degradation, which results in capacity reduction. Battery cells are subjected to mechanical stress along with loss of active material as a result of regular charging and discharging operations, especially at high discharge depths. Furthermore, battery degradation worsens when they are exposed to overcharging and high current rates and when stored under improper conditions [4]. These complex interrelated factors make the prediction of SOH challenging as they vary between battery chemistries and use cases. Therefore, accurate SOH prediction is critical to ensure safety, optimise performance, and reduce costs.
Currently, the SOH of the battery is estimated on the basis of direct measurements and physical models. Common methods for direct measurement include Coulomb counting [5], which tracks the flow of charge to estimate the reduction in capacity, and electrochemical impedance spectroscopy (EIS) [6], which analyses changes in impedance. Physics-based approaches include equivalent circuit models (ECMs) [7] and empirical models, which simulate battery behaviour or are based on laboratory test data. These methods rely on measurable parameters such as voltage, current, temperature, and impedance to estimate the SOH. Despite their widespread use, these traditional SOH estimation methods have significant limitations. Coulomb counting is prone to cumulative errors over time, especially under dynamic operating conditions. EIS requires specialised equipment and controlled environments, making it impractical for real-time applications, whereas ECM and empirical models often simplify complex degradation mechanisms, leading to reduced precision for batteries outside of the tested conditions [7]. Additionally, these methods struggle to account for the nonlinear and interdependent effects of ageing factors, resulting in poor generalisability across different battery types and usage scenarios. For these reasons, there exists a need for more reliable means of battery SOH estimation. This brings about the utilisation of machine learning for the purpose of battery SOH prediction.
Machine learning, a subfield of artificial intelligence, involves algorithms that learn patterns from data to make predictions or decisions without being explicitly programmed. Machine learning (ML) algorithms, such as neural networks [8], support vector machines [9], and decision trees [10], can learn intricate patterns from large datasets comprising voltage, current, temperature, and cycle data. Therefore, they are able to model battery degradation more accurately and in a scalable data-driven manner across diverse operational conditions. For example, Xu et al. [11] used a hybrid machine learning model that uses both convolution neural networks (CNNs) and a long short-term memory (LSTM) algorithm to estimate the state of health of lithium-ion batteries. Validation of the approach was performed on NASA and Oxford battery datasets and the results showed improved accuracy and robustness, with RMSE below 0.004. Such results underscore the potential of ML-based models to overcome the limitations of traditional methods by capturing both spatial and temporal dependencies in battery performance data.
Despite their promise, ML models are not without limitations. Challenges include data dependency, model interpretability, and the need for large high-quality training datasets [12]. Nonetheless, ML provides a powerful framework for generalising across battery types and operational conditions, which are areas where traditional methods fall short. This article provides a review of four widely used ML techniques in SOH estimation, critically evaluating their merits, demerits, and advancement. Additionally, a simulation of the four techniques is carried out on a sample dataset to analyse and compare the performance of these ML algorithms.
The contributions of this paper are as follows:
i.
A comprehensive literature review of four widely used ML algorithms in SOH prediction, namely support vector regression (SVR), long short-term memory (LSTM), convolution neural networks (CNNs), and random forest (RF), including relevant mathematical formulations.
ii.
An evaluation of the strengths, limitations, and recent advancements of each algorithm.
iii.
A MATLAB R2025a-based simulation to assess and compare the performance of the four selected ML models in SOH prediction.

2. Machine Learning Algorithms in State of Health Prediction

Several recent studies have applied ML methods to batteries in EV contexts. This section explores the application of four commonly used ML algorithms, namely support vector regression (SVR), convolutional neural network (CNN), long short-term memory (LSTM), and random forest (RF). It examines their methodologies, strengths, and challenges, highlighting their potential to improve battery health prediction.

2.1. Support Vector Regression

SVR is a supervised learning algorithm designed for regression tasks. SVR differs from standard linear regression in that it only calculates losses when prediction errors exceed a defined tolerance ε while optimising model performance through the maximum interval bandwidth and minimum total loss [13]. The regression problem requires finding an optimal hyperplane that positions all sample points at their closest distance from the hyperplane, as shown in Figure 1 [14].
For a linear hard-spaced SVR, the optimisation objective is as follows:
1 2 | |   w | | 2 + C i = 1 n ξ i + ξ i * ,
which are subject to the following constraints:
y i w , x i b ε   + ξ i w , x i + b y i ε   + ξ i * , ξ i , ξ i * 0 .
The constraints ensure that the predictions, given by
f ( x i ) = ( w , x i ) + b ,
lie within an ε-tube around the true targets y i while allowing some flexibility through the slack variables. The parameter ε allows the predicted value to have no loss within the range of ε and reduces the impact of noise.
Figure 1. Illustration of 1D linear SVR, showing the hyperplane, support vectors, and tolerance ε, adapted from ref. [14].
Figure 1. Illustration of 1D linear SVR, showing the hyperplane, support vectors, and tolerance ε, adapted from ref. [14].
Wevj 16 00594 g001
To solve this constrained optimisation, SVR is typically formulated in its dual form using Lagrange multipliers. The dual problem involves optimising the Lagrange multipliers α i ,   α i * , leading to the regression function:
f ( x ) = i = 1 n ( α i α i * ) K ( x i , x ) + b
where w is the weight vector, C is the regularisation parameter, ε is the tube width, ξ i and ξ i * are slack variables, α i and α i * are Lagrange multipliers, and K ( x i , x ) is the kernel function. The regularisation parameter, C , controls the trade-off between model complexity and error tolerance. The greater the C , the heavier the penalty for errors that exceed the range of ε.
Kernels enable SVR to handle nonlinear relationships by implicitly mapping data into higher-dimensional spaces. Common kernels include linear, polynomial, sigmoid, and radial basis functions and are given in Table 1.
One of the key merits of SVR is its robustness to outliers and its high prediction accuracy, which is particularly valuable in SOH prediction, where battery data often exhibit noise due to varying operating conditions. Zhao et al. [15] employed SVR to estimate SOH based on capacity fade data from lithium-ion batteries, using features derived from charge/discharge cycles. Their model achieved a mean absolute error (MAE) of less than 2% across multiple datasets, leveraging SVR’s epsilon-insensitive loss function to mitigate the impact of outliers on voltage and capacity measurements. This robustness ensured reliable predictions despite irregularities in battery degradation patterns. Another study [16] reported mean R2 values of up to 0.962 for test sets using full discharge data. For partial discharge windows, the model maintained strong performance across most voltage ranges, except for high-voltage intervals. The study highlights the robustness, computational efficiency, and applicability of SVRs to real-world scenarios, particularly with partial data, although it notes limitations with outlier batteries exhibiting unique degradation patterns.
The flexibility of SVRs in modelling nonlinear relationships through kernel functions is another significant advantage. This is evident in the work by Chen et al. in which they employed a radial basis function kernel in an SVR-based model using partial charge voltage and current data to estimate SOH. The model achieved a mean squared error (MSE) < 0.00052 and a mean absolute relative error (MARE) < 0.93% on two test cells. In [17], Feng et al. constructed a hybrid kernel function of the polynomial kernel function and radial basis function, compensating for the shortcomings of single kernel functions. This improved the model’s generalisation ability and learning ability and enhanced the nonlinear modelling ability of the model, effectively avoiding overfitting the training set.
Recent researchers have used optimisation algorithms to tune hyperparameters such as C and ε to improve accuracy and SVR’s model stability. In [18], the authors employed the particle swarm optimisation (PSO) algorithm to estimate the SVR kernel parameter. The improved PSO–SVR model showed great robustness when the training data contained noise and measurement outliers. Recent work by Vedhanayaki and Indragandhi [19] introduced a Bayesian optimised support vector regression (BO-SVR) framework with a Gaussian kernel for SOH estimation. Their study employed the equivalent charging voltage difference interval (ECVDI) and equivalent discharging voltage difference interval (EDVDI) as inputs, achieving a remarkably low RMSE of 0.0082, outperforming both standard kernel SVR and Gaussian process regression models. Xia et al. [20] proposed an impedance-based SOH estimation framework that combines feature selection and an improved support vector regression model. The sine sparrow search algorithm (Sine-SSA) was applied for hyperparameter tuning of SVR. Using a large commercial EIS dataset across multiple temperatures, the model achieved a maximum error of 2.58%, outperforming LSTM and conventional SVRs.
The grey wolf optimisation (GWO) algorithm, which features a basic structure, minimal adjustable parameters, and straightforward implementation, optimised both the kernel and penalty parameters of kernel SVM in [21]. This implementation led to a longer runtime as the data increased compared with a non-optimal KSVM. Other algorithms that have been used include the ant lion optimisation algorithm [22], ant colony optimisation [23], and dung beetle optimisation [24].
Despite the improvements in accuracy and model stability with the use of optimisation algorithms, SVR models may become computationally expensive for large datasets and require careful feature selection to avoid overfitting. The data dependency of SVRs also limits their applicability to emerging chemistries such as sodium-ion or solid-state batteries, a key gap in the scope of the research [25].

2.2. Long Short-Term Memory

Long short-term memory (LSTM) networks are a specialised type of recurrent neural network (RNN) designed to model and predict sequential data with long-term dependencies, making them ideal for time series tasks. Unlike traditional RNNs, which often suffer from vanishing or exploding gradients, LSTMs introduce a memory cell and three gates (input, forget, and output) to selectively remember or forget information over extended time periods [26].
Figure 2 shows the core structure of an LSTM [27]. The architecture includes:
  • A memory cell that maintains a cell state C t , which carries information across time steps, enabling long-term memory.
  • Gates that control the flow of information. The gates are as follows:
    Forget gate: Decides what to discard from the previous cell state.
    Input gate: Determines what new information to store.
    Output gate: Selects what to output at the current time step.
  • Each gate uses sigmoid (σ) and tanh activations to regulate information.
Figure 2. General architecture of an LSTM unit, showing memory cell and gates, adapted from ref. [27].
Figure 2. General architecture of an LSTM unit, showing memory cell and gates, adapted from ref. [27].
Wevj 16 00594 g002
The forget gate is given by the following equation:
f t = σ ( W f     h t 1   ,   x t +   b f )
where f t [ 0,1 ] decides what to forget from C t 1 , W f is the weight matrix of the forget gate, b f is the bias term of the forget gate, and σ is the sigmoid activation function
The input gate consists of two stages. The first stage, the sigmoid layer, determines which value to update, and, in the second part, a t a n h layer determines which value should be added to the state. The output of the first part is:
i t = σ ( W i [ h t 1 , x t ] + b i ) ,
where   W i is the weight matrix of the input gate, h t 1 is the output value at time t −1, x t is the input at time t , b i is the bias term of the input gate, and σ is the sigmoid activation function.
The output of the second part is given by:
C ~ t = tanh W C h t 1 , x t + b C
where i t   selects new information, t a n h is the activation function, C ~ t is the candidate cell state,   W C is the weight matrix of the output gate, h t 1 is the output value at time t −1, x t is the input at time t , and b C is the bias term of the output gate.
The cell state is then updated by:
C t = f t C t 1 + i t C ~ t .
The output gate has a sigmoid layer, which determines what information reaches the output. It is given by the following equation:
o t = σ W o h t 1 , x t + b o ,
where W o is the weight matrix of the output gate, h t 1 is the output value at time t − 1, x t is the input at time t , b o is the bias term of the output gate, and σ is the sigmoid activation function. The output is given by:
  h t = o t tanh ( C t )
where h t is the output at time t , o t is the output value of the output gate at time t , and C t is the long-term state of the unit at time t .
Several recent studies have explored LSTM-based methods for SOH prediction, addressing challenges such as capacity regeneration, hyperparameter optimisation, and data scarcity. Optimisation algorithms have been employed to highlight the importance of hyperparameter tuning in improving LSTM convergence and generalisation. For instance, Gong et al. [27] developed an LSTM model optimised by particle swarm optimisation (PSO) using four health indicators (HIs) selected through grey relational analysis. PSO was used to tune hyperparameters, such as neurone numbers, to enhance convergence and prevent overfitting via RMSProp and dropout. The model achieved a 5% accuracy improvement over classical LSTM in experimental datasets. However, this model may not generalise well to batteries with different chemistries or operating conditions. Wang et al. [28] proposed an attention mechanism-enhanced LSTM (AM-LSTM) model. A moving average filter reduced noise in capacity data, and the attention mechanism weighted critical temporal features, improving prediction accuracy. Validated on NASA and University of Maryland datasets, the model demonstrated robustness across diverse battery types and discharge rates. The downside was increased model complexity, which could lead to higher computational costs and longer training times.
Yang and Chen [29] proposed a BiLSTM framework for SOH prediction that integrates time-varying filter empirical mode decomposition (TVF-EMD) and a sliding window to extract nonlinear features and preserve temporal patterns. Hyperparameters were optimised using Bayesian optimisation with hyperband (BOHB), resulting in RMSE < 0.0016 and R2 > 0.999 on NASA datasets (B0005–B0007). The study demonstrated that combining advanced signal processing with bidirectional LSTM significantly improves accuracy compared with LSTM-, SVR-, and CNN-based models. Peng et al. [30] proposed an SOH estimation method that integrates an improved grey wolf optimisation (IGWO) algorithm with an LSTM network to optimise hyperparameters and reduce overfitting. The model used only partial discharging health features (HFs), significantly lowering data requirements and computation time while maintaining accuracy. Experimental results on lithium-ion batteries achieved MAE, RMSE, and MAPE values all within 1%, showing that the IGWO-LSTM framework provided efficient and highly accurate SOH prediction.
Transfer learning further enables application across diverse battery types with limited data. In [31], the author developed a transfer learning-based LSTM approach for SOH and prediction of cycle life across different battery types (NCA and NCM). The model was pre-trained on NCA battery data and adapted to NCM batteries with limited data, reducing the need for extensive retraining. Using NASA and custom datasets, the study achieved high accuracy in capacity fade prediction, with RMSE metrics indicating robust performance. The effectiveness of transfer learning depends on the similarity between source and target battery types. Significant differences in degradation patterns could reduce accuracy. The model also requires pre-training on large datasets, which may not always be available for new battery chemistries. Yang et al. [32] proposed a GAN-LSTM-TL framework for SOH estimation, where a generative adversarial network (GAN) generated synthetic training data to reduce overfitting and enhance learning. The LSTM network models temporal dependencies, and transfer learning (TL) improves adaptability across different datasets, including NASA and CALCE cells. Experiments showed that GAN-LSTM-TL achieved higher accuracy than standalone LSTM or GAN-LSTM, maintaining prediction errors below 3% even with limited training data.
Hybrid models, such as Hu et al.’s complementary ensemble empirical mode decomposition (CEEMD)–transformer LSTM [33] and Liu et al.’s CNN-BiLSTM [34], improve robustness under diverse operating conditions. Although two datasets were used in [33], both are still within the lab context, field data validation is missing. Also, integrating CEEMD, transformer, and LSTM makes the framework complex and harder to implement or maintain. The use of raw electrochemical impedance spectroscopy (EIS) data in [34] can be resource intensive during the training process because the data are high-dimensional. The model also struggled to predict abrupt capacity drops (e.g., at cycle 190), which may pose safety concerns. Xiang et al. [35] introduced a hybrid model that combines variational mode decomposition (VMD), an extended LSTM (xLSTM), and a frequency-enhanced channel attention mechanism (FECAM) for SOH prediction. The approach used VMD to separate long-term trends from noise, while xLSTM and FECAM jointly captured temporal and frequency-domain features, improving robustness against capacity regeneration and disturbances. Experiments on CALCE and NASA datasets showed the proposed model outperformed LSTM, GRU, BiLSTM, transformer, and CNN-LSTM, achieving MAE as low as 0.0049 and RMSE of 0.0085.
In summary, despite the ability of LSTMs to eliminate the problem of vanishing or exploding gradients, the increased computational complexity requires significant resources, making it less practical for resource-constrained environments. The need for large datasets can be a barrier, and overfitting is a risk without proper regularisation [36]. LSTM also struggles with abrupt changes in battery behaviour, leading to larger errors at high cycle numbers or during rapid degradation. Its black box nature limits interpretability, which can be a major concern in safety-critical battery applications.

2.3. Convolution Neural Network (CNN)

It is a classical feed-forward deep neural network that is used in image recognition, computer vision, and natural language processing, among others. Typically, the network is made up of an input layer, several convolutional layers, subsampling layers, fully connected layers, and an output layer [37]. Figure 3 shows a general hierarchical structure of a CNN.
Unlike LSTMs, which excel at sequential modelling, CNNs leverage convolutional layers to extract local patterns, reducing computational complexity and capturing spatial or temporal features, enabling effective feature extraction with lower computational complexity.
The convolutional layers apply filters to the input data to extract features. Each filter slides over the input, computing dot products to produce feature maps. For a 1D input sequence x R T , a convolutional layer applies K filters of size F yielding:
z i , k = j = 0 F 1 w j , k x i + j + b k
where w k is the k -th filter’s weights, b k is the bias, and z i , k is the output at position i for filter k .
ReLU activation introduces nonlinearity given by:
a i , k =   max ( 0 ,   z i , k )
The pooling layers reduce spatial dimensions to lower complexity and prevent overfitting. Figure 4 shows a schematic representation of a pooling layer in CNN [38]. Max pooling over a window of size P is given by:
p i , k = max ( a i , k ,   a i + 1 , k a i + P 1 , k )
The fully connected layer aggregates features for regression or classification. For regression, the final layer output is given by:
y ^ = W h + b
where h is the final flattened feature vector, W is the weight matrix, and b is the bias term.
Figure 4. Pooling layer in CNN, adapted from ref. [38].
Figure 4. Pooling layer in CNN, adapted from ref. [38].
Wevj 16 00594 g004
Recent work highlights a clear progression in CNN-based SOH estimation methods, moving from handcrafted feature design to automated and quantum-enhanced frameworks. Lu et al. [39] demonstrated that extracting and fusing intra-cycle and inter-cycle features from partial voltage curves significantly improves prediction accuracy, with MAE below 0.0028 and MAPE under 0.32%. Bockrath et al. proposed a temporal convolutional network (TCN) for SOH estimation using raw sensor data from partial discharge profiles across different SOC ranges. The TCN processes current and time data directly, avoiding complex preprocessing, and was optimised via Bayesian hyperparameter tuning with stratified cross-validation. Results on NASA randomised battery usage data showed an overall RMSE of 1.0%. Chen et al. [40] simplified feature engineering by applying a CNN directly to partial constant-voltage charging data, showing that only the first 1000 s of CV data was sufficient for highly accurate predictions across different chemistries. Building on these foundations, Liang et al. [41] proposed a quantum CNN (QCNN) with automated feature fusion using quantum encoding, which achieved R2 > 96% across diverse chemistries and operating conditions while reducing parameter requirements. Together, the evolution reflects a shift towards more practical, generalisable, and computationally efficient SOH prediction methods.
CNNs are often combined with other architectures for SOH prediction. Some studies have combined CNN with LSTM, BiLSTM, or attention mechanisms to improve temporal modelling and accuracy. Xing et al. in [42] combined BiLSTM and the attention mechanism with CNN. In [43], the authors used a combination of CNN with LSTM and deep neural networks (DNNs). In the study by Yao et al. [44], it features a hybrid model of CNN, wavelet neural network (WNN), and wavelet LSTM (WLSTM) that inherits both the fast convergence and robust stability of the WNN. In [45], the author combines CNN with the Kolmogorov–Arnold network (KAN). Peng et al. [46] combined CNN-based feature extraction with a probability-sparse self-attention mechanism, reducing computational complexity while capturing both local and long-range degradation features. Tested on cycle life data, including voltage, current, and temperature, the model significantly outperformed traditional models in both accuracy and computational efficiency, showing promise for remote and real-time SOH prediction. These hybrid approaches leverage the strengths of each model, such as robust stability, fast convergence, and improved sequence modelling, to provide better SOH estimation.
Hyperparameter optimisation algorithms have been integrated with CNNs to enhance the performance of SOH prediction. Wu et al. [47] used a multilayer CNN optimised by the Kepler optimisation algorithm (KOA). The model outperformed CNNs without hyperparameter optimisation by up to 58.97% (MAE). It was also noted that increasing CNN depth beyond two layers degraded performance. Despite the high accuracy, KOA adds computational overhead during training, which may not be suitable for systems with limited resources. In [48], the authors propose a hybrid CNN-BiGRU–attention model, optimised using an improved gray wolf optimisation (IGWO) algorithm [49], a tree-structured Parzen estimation (TPE) algorithm is used to optimise the model parameters. The study also used a MC-CNN–TimesNet model where the TimesNet converts 1D time series into a collection of 2D tensors, leveraging multiple cycles to capture intra- and inter-timescale relationships and dependencies and CNN captures deeper spatial. The results indicate that SOH and RUL can be predicted with an average RMSE within 1.5% of features by convolutional operations on the input features. A Bayesian optimisation algorithm was used in [50] to optimise CNN and LSTM and prevent the combined network model from converging to local optima. This model achieved an RMSE of <1% in the joint validation involving the NASA public dataset and the laboratory self-collected dataset
In summary, CNN is a powerful tool that offers high accuracy and the ability to automatically extract features from complex data. Despite challenges such as data dependency, computational complexity, and limited interpretability, recent advances in hybrid models, multi-modal fusion, and transfer learning enhance CNN’s applicability in different applications.

2.4. Random Forest

RF is an ensemble machine learning algorithm that uses multiple decision trees to handle complex battery data, offering robust predictions against noise [51]. RF generates multiple subsets of the dataset through bootstrap sampling (random sampling with replacement). Each subset, typically of the same size as the original dataset, contains a random selection of data points, allowing some samples to be repeated while others are excluded. This process creates diverse training sets for each decision tree, reducing overfitting and improving robustness.
For each bootstrap sample, a decision tree is constructed. At each node of the tree, a random subset of features (e.g., a subset of voltage, current, or temperature features) is selected to determine the best split, reducing the correlation between trees. The splitting criterion for regression is typically the minimisation of variance, measured by mean squared error (MSE):
M S E = 1 n i = 1 n ( y i y ^ i ) 2
where y i is the actual SOH and y ^ i is the predicted SOH.
Each decision tree independently predicts the SOH for a given input. For regression, the final SOH prediction is the average of all tree outputs:
y ^ = 1 T t = 1 T y ^ t
where T is the number of trees and y ^ t is the prediction of the t -th tree.
An RF model was built to estimate battery SOH, whose number of regression trees and the number of input variables per node are optimised by the out-of-bag (OOB) estimation [52]. The model used partial data from typical on-board charging conditions but was tested only on two cells from a public dataset. Another study by [53] used RF to predict the state of health of lithium-ion 21,700—5 Ah cells. A grid search was employed to optimise hyperparameters and the model was compared with SVR. The results achieved an R 2 of 0.92 and an RMSE of 0.06. However, the model relies entirely on manually extracted features, which may require extensive domain knowledge. Liang et al. [54] developed an enhanced RF model optimised through PSO to predict the SOH of LiFePO4 battery modules. The model relies heavily on manual feature extraction from voltage, current, and capacity during charge/discharge cycles, which may limit automation and portability to different datasets. Moreover, even with PSO tuning, RF does not capture temporal dependencies or sequence effects across cycles, thus limiting its ability to track gradual degradation trends. This makes it less suitable for predictive degradation modelling over long-time horizons compared with deep recurrent models.
Lamprecht et al. [55] proposed a novel method that leverages active charge balancing (ACB) behaviour as a degradation indicator and applies RF regression for prediction. Using a modular simulation framework replicating battery aging and balancing processes, their approach achieved high accuracy (1.94% for capacity SOH and 4.28% for resistance SOH), offering one of the few models capable of accurately estimating both capacity fade and resistance growth.
Several recent studies demonstrate the growing effectiveness of hybrid and ensemble learning approaches for SOH estimation. Wang et al. [56] combined RF with gated recurrent u (GRU) to leverage on temporal or spatial features. This led to an improvement in the estimation accuracy and computational efficiency by at least 15.84% The model, however, depends heavily on the quality and consistency of the input data. The study in [57] used a combined RF–CNN model, which achieved a 34–46% reduction in MAE. The model was validated on specific datasets, so generalisation to different battery chemistries, formats, or use cases (e.g., solid-state, fast-charging) is not guaranteed. Retraining or transfer learning might be needed for new conditions. Lu et al. [58] introduced a transfer learning framework that integrates multi-attention mechanisms, CNN, and RF regression. By using only partial charging segments, freezing CNN feature extraction layers, and replacing dense layers with RF optimised via Bayesian TPE, the method achieved very low RMSE (0.46% for LFP and 1.61% for NCA), reducing estimation errors by up to 80.9%. In [59], RF was used to capture nonlinear relationships and select robust features from voltage, current, temperature, and cycle data; then, the prediction was refined using an artificial neural network (ANN) to extract deeper patterns. The hybrid architecture, however, does not explicitly model time series dependencies or degradation trends over many cycles.
In summary, RF remains a powerful and interpretable that should be paired with robust feature engineering or hybrid architectures for SOH prediction. Its inability to model sequential dependencies and reliance on handcrafted inputs pose challenges for broader deployment in real-time battery health monitoring systems. Future research may focus on combining RF with temporal deep learning models, automated feature extraction, and transfer learning to improve scalability, adaptability, and predictive accuracy across diverse battery types and operating conditions.

3. Discussion

This comparative review highlights the growing importance of ML algorithms in improving the accuracy, adaptability, and robustness of battery SOH prediction. The four models investigated—viz. SVR, LSTM, CNN, and random Fforest—each offer unique strengths and face specific limitations, influencing their suitability in different scenarios. Table 2 shows a summary of the papers used in this literature review.
SVR stands out for its strong performance on smaller datasets and its robustness to outliers, making it suitable for early-stage SOH estimation and low-resource environments [25]. However, its inability to capture temporal dependencies and its sensitivity to kernel parameter tuning limit its application in long-term or dynamic battery use cases. Advances in hybrid kernel functions and optimisation algorithms (e.g., PSO, GWO) have improved its performance, yet scalability remains a concern.
LSTM networks are well-suited for sequential modelling, particularly in capturing long-term dependencies in degradation trends. However, they are computationally intensive, require large volumes of labelled data, and suffer from interpretability issues [26]. Hybrid models that combine LSTM with attention mechanisms or signal decomposition techniques (e.g., CEEMD) have demonstrated superior accuracy, though at the cost of increased model complexity.
CNNs offer an effective framework for extracting spatial features and handling complex high-dimensional input data. When applied to voltage, current, or EIS signals, CNNs can automatically derive salient features that contribute to accurate SOH prediction. However, they struggle with temporal dependencies when used in isolation. Consequently, CNNs are most effective when integrated with temporal models such as LSTM or BiLSTM, enabling robust spatiotemporal learning [60].
Random forest remains a reliable and interpretable model, particularly useful in scenarios where data are limited or noisy. It is computationally efficient and easy to implement but lacks the ability to model degradation trends over time [61]. To overcome this, RF is often combined with recurrent models (e.g., GRU) or deep networks (e.g., CNN or ANN) to improve its temporal modelling capability and generalisability across battery chemistries and usage patterns.
In general, hybrid models, such as CNN-LSTM, RF-GRU, and CNN-KAN, have shown promise by combining the strengths of individual algorithms. These architectures have achieved better generalisation, robustness, and accuracy in SOH estimation. However, they introduce trade-offs in terms of complexity, computational cost, and training time. Therefore, the choice of model should be guided by specific application requirements, including the data availability, the target hardware constraints, and the desired prediction horizon. Newer studies have focused on newer hybrid frameworks and optimised-assisted models.
While this study only compared SVR, RF, CNN, and LSTM, some of the recent literature has demonstrated the potential of transformer-based architectures and extreme learning machine algorithms for SOH estimation. For instance, Duan et al. [62] developed a variable forgetting factor online sequential extreme learning machine (VFOS-ELM), optimised using an improved whale optimisation algorithm (IWOA) for SOH estimation and remaining useful life prediction. The hybrid framework achieved RMSE below 0.15% and exhibited strong adaptability across NASA, Oxford, and MIT datasets. Shu et al. [63] introduced a voltage-segment transformer framework, achieving sub-2% errors using only partial charging data, thus making the approach practical for EV scenarios. Chen et al. [64] further demonstrated the adaptability of transformers in an edge–cloud collaborative SOH framework. By coupling a variational mode decomposition (VMD) step with MEWOA optimisation and a pre-LN transformer, they achieved MAE < 0.6% and MSE < 0.008% across NASA, CALCE, MIT, and EV field data. Their ability to model both short-term fluctuations and long-term degradation trends suggests that future BMS implementations may increasingly adopt transformer variants, particularly in edge–cloud collaborative frameworks.
Interpretability, data scarcity, and transferability remain critical challenges. Many high-performing models such as CNNs and LSTMs function as black boxes, limiting their adoption in safety-critical applications such as electric vehicles. Additionally, most studies rely on public laboratory datasets, which may not reflect real-world operating conditions. Efforts to integrate physics-informed features, enable online learning, and apply transfer learning are key to improving model robustness and deployment across diverse battery chemistries, formats, and environmental conditions. Some of those efforts can be seen in studies such as [65], which proposed an enhanced 1D-CNN framework that uses variable-length charging segments to mimic user random charging behaviour and integrates transfer learning for adaptability across different chemistries.
To address the challenge of explainability, SHAP (SHapley Additive exPlanations) and LIME (local interpretable model-agnostic explanations) [66] that are both popular explainable AI (XAI) have been explored. Jafari and Byun [67] applied SHAP to a CNN–LSTM–ConvLSTM fusion model, showing that SOH predictions can be decomposed into physically meaningful contributions from capacity, voltage, current, and temperature. Similarly, M.K. B et al. [68] introduced the CART-GX hybrid model, where SHAP gradient explainer revealed consistent feature importance across NASA and CALCE datasets. Notably, they argued that SHAP is more robust than LIME, whose explanations may vary with small perturbations. Together, these studies show that SHAP offers both interpretability and stability, bridging the gap between high-performing deep models and practical BMS requirements.

4. MATLAB Simulations

4.1. Introduction to the Datasets

This article uses the NASA battery dataset for the simulations [69]. It consists of four lithium-ion battery datasets labelled as B0005, B0006, B0007, and B0018 with a rated capacity of 2 Ah and a rated voltage of 3.7 V. These four batteries were subjected to charge/discharge experiments at 24 °C using a constant current–constant voltage charging protocol at 1.5 A until 4.2 V, followed by a constant voltage phase until the current dropped below 20 mA. Discharge was performed under a constant current of 2.0 A down to 2.7 V (B0005), 2.5 V (B0006), 2.2 V (B0007), and 2.5 V (B0018). The time series cycle measurements of voltage (V), current (I), temperature (T), and capacity (Ah) were recorded until end-of-life, which is defined as the point when the measured capacity drops below 70% of nominal capacity.
Figure 5 illustrates the capacity degradation curves of the four batteries. For each battery, the discharge capacity decreases progressively with cycling. This degradation is used to construct the SOH labels. Specifically, SOH is calculated as
S O H = C i C o × 100 %   ,
where C i   is the discharge capacity at the current cycle and C o   is the initial nominal capacity. Figure 6 shows the raw SOH degradation curves. The degradation curves highlight several important aspects of lithium-ion battery ageing. First, the degradation is nonlinear, with an initial period of gradual decline followed by accelerated capacity loss near end of life. This behaviour reflects the combined effects of electrode degradation, electrolyte decomposition, and loss of active material. Second, variation across the cells can be observed, even under identical cycling conditions, underscoring the influence of inherent manufacturing differences and stochastic degradation mechanisms.
These trends justify the use of machine learning approaches, as traditional models often struggle to capture such nonlinear and cell-specific ageing patterns. The inclusion of B0005 as the independent test set further emphasises the need for generalisable models, to ensure fair comparison, since predictive accuracy must extend across batteries with similar but not identical degradation trends. For this paper, B006, B0007, and B0018 were used as the training data and B0005 was used as the testing data.

4.2. Health Features Extraction

The extraction of health features from discharge cycles followed a standardised process to achieve model comparison fairness.
i.
Resampling to a fixed length
The raw dataset contains discharge cycles that have varying time steps because of different sampling rates and cycle lengths. To make the cycles comparable, all signals were resampled to 120 uniformly spaced points along a normalised time axis (0 to 1). The normalisation process maintains the original degradation patterns while removing the effects of varying cycle lengths.
ii.
Measurement signals
From the resampled signals, six synchronised sequences were constructed: the raw measurements of voltage (V), current (I), and temperature (T), together with their first-order derivatives with respect to time (dV/dt, dI/dt, dT/dt). This produced a 6 × 120 representation per cycle, capturing both the original signal values and their time-dependent variations.
iii.
Use across models
  • CNN and LSTM: The full 6 × 120 sequences were provided directly as inputs, enabling the networks to learn both local patterns (via convolution) and temporal dependencies (via recurrent layers).
  • SVR and RF: Since these models require vectorised features rather than full sequences, summary statistics were computed from the same 6 × 120 representation. These features are shown in Table 3, which gives a summary of the health features that were used in the SVR and RF models.
All the extracted features were retained to preserve physical interpretability and ensure a fair and consistent basis for model comparison. No feature selection was carried out. This design ensured that all models learned from the same physical information, with differences arising only from the modelling approach rather than from variations in input features. The objective was to capture both the absolute values of key electrochemical signals and their dynamic behaviour, while ensuring consistency across cycles of different lengths.

4.3. Model Training

  • CNN: A 1D convolutional neural network with three convolutional layers, batch normalisation, global average pooling, and fully connected layers, trained for 100 epochs with early stopping on validation loss
  • LSTM: A stacked bi-directional LSTM network, followed by fully connected and dropout layers, trained for 60 epochs with similar validation protocol.
  • SVR: The RBF kernel was used with grid search performed over the hyperparameters (C, ε, kernel scale) tuned via 5-fold cross validation, and the best model refit on full training data.
  • RF-TreeBagger regression with hyperparameters (number of trees, minimum leaf size) chosen via grid search and out-of-bag RMSE minimisation.
Both CNN and LSTM models were optimised using the Adam optimiser with an initial learning rate of 5 × 10−4 and a regularisation coefficient of 10−4. B0005 was held out entirely for testing, ensuring cross-battery generalisation.
Validation is performed using a subset of the training data, with early stopping for deep models to avoid overfitting. Classical models are trained on the complete training set. A summary of the search space used and the final values of the hyperparameters is given in Table 4

4.4. Evaluation

All models were tested on the independent test battery (B0005). Performance was assessed using metrics such as R-squared (R2), mean absolute error (MAE), mean absolute percentage error (MAPE), and root mean squared error (RMSE). These metrics provide separate statistical measures to evaluate model accuracy and the strength of the relationships between predictor variables and target outcomes.
(a)
Mean Absolute Error (MAE)
The mean absolute error (MAE) measures the average discrepancy between forecasted and actual values to show how much predictions differ from real outcomes.
The MAE can be formulated as follows:
M A E = 1 n   i = 1 n ( y i y ^ i )
where y i   and y ^ i are the actual and predicted values, respectively. A lower MAE score signifies improved model performance. Mean absolute percentage error (MAPE) expresses the average magnitude of error as a percentage of the actual values.
(b)
R-squared (R2)
The R-squared value, also known as the coefficient of determination, represents the proportion of variance in the dependent variable that is predictable from the independent variables. It is given by,
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ i ) 2
(c)
Root Mean Squared Error (RMSE)
The root mean squared error (RMSE) measures the square root of the average of squared differences between predicted and actual values. It is widely a key metric for assessing regression-based ML models. Generally, RMSE and MAE metrics indicate the variability of residuals, while R2 values represent how well the model fits the data:
R M S E = 1 n   i = 1 n ( y i y ^ i ) 2

4.5. SOH Estimation Results and Discussion

4.5.1. Performance Metrics

Table 5 summarises the predictive accuracy of the four models on the independent test battery (B0005). The LSTM model achieved the best overall performance, with the lowest RMSE, MAE, and MAPE, and the highest R2, indicating superior accuracy in SOH estimation. The CNN model also performed very well, slightly below LSTM but outperforming the RF and SVR models. Random forest (RF) performed comparably to CNN, demonstrating the effectiveness of tree-based ensembles with engineered features. Support vector regression (SVR) showed significantly higher error rates and lower R2, suggesting a weaker capacity for capturing nonlinearities and sequence-dependent patterns in the data.

4.5.2. Predicted vs. Actual SOH Across the Different ML Models

Figure 7 illustrates the predicted versus actual SOH trajectories for each model on the independent test battery (B0005). Across all models, the overall degradation trend was captured with varying levels of accuracy. The random forest (RF) model aligned well with the actual SOH during mid-life cycles but diverged in later stages, consistently underestimating the SOH. This behaviour reflects RF’s strength in modelling global nonlinear relationships, but also its limitation in representing sequential dependencies inherent in degradation processes.
The support vector regression (SVR) model failed to capture sharper declines in SOH during late-life cycles, leading to systematic overestimation near end of life. While SVR remains effective for regression tasks, its static formulation makes it less suitable for long-term time-dependent battery dynamics.
In contrast, the deep learning models achieved stronger performance. The convolutional neural network (CNN) showed close agreement with the actual SOH, particularly during mid-life cycles, and effectively extracted local degradation features. However, its reliance on convolutional filters constrained its ability to capture long-range temporal dependencies, resulting in deviations in later cycles. The long short-term memory (LSTM) network provided the closest alignment with the true SOH across the entire cycle range. The success of LSTM is not only because of its sequence modelling capabilities but also because it captures the inherent time-related physical–chemical processes during battery aging (such as SEI film growth and lithium deposition), which have strong memory effects and timing dependence.
By exploiting its memory mechanism and its ability to model sequential dependencies, LSTM produced smooth and accurate predictions, even during sharp capacity declines. This indicates strong robustness in both short- and long-term SOH tracking. In comparison, RF demonstrated the utility of ensemble methods when combined with engineered statistical features, though it lagged behind CNN and LSTM. SVR struggled to generalise across different batteries, reflecting its sensitivity to kernel selection and its limited flexibility in modelling degradation trajectories.
Quantitatively, LSTM achieved the lowest error values (RMSE and MAE, see Figure 8), followed by CNN, RF, and SVR. These findings reinforce that models explicitly designed for sequential data offer superior performance compared with classical machine learning techniques in battery prognostics.
From a practical perspective, accurate SOH prediction is essential for the safe and efficient use of lithium-ion batteries in electric vehicles and stationary energy storage. LSTM’s superior accuracy suggests strong potential for integration into battery management systems (BMSs), where reliable long-term forecasting is critical for maintenance planning, failure prevention, and extended service life. However, their high computational and memory demands make direct deployment on embedded BMS hardware challenging. Efficiency can be improved through lightweight strategies such as pruning, quantisation, knowledge distillation, or hybrid offline–online frameworks. Nonetheless, models such as RF and SVR remain relevant in resource-constrained applications due to their lower computational requirements.

4.5.3. Training Time

The evaluation process also included an assessment of computational efficiency. As expected, the training times for SVR and RF were significantly shorter than for deep learning models, as shown in Table 6. The SVR and RF models trained very quickly, typically within seconds to a minute, on a typical workstation equipment. The training duration for CNN and BiLSTM models took longer than the other models due to their larger parameter space and iterative optimisation, though this additional cost was compensated by higher predictive accuracy. Despite these differences in training, all models in this evaluation demonstrate sufficient real-time SOH estimation capabilities.

5. Conclusions

This review compared machine learning methods for lithium-ion battery SOH estimation, supported by MATLAB simulations on the NASA B0005, B0006, B0007, and B0018 dataset. The results showed that LSTM achieved the highest accuracy (RMSE = 0.0145, MAE = 0.0118, R2 = 0.980, MAPE = 1.39%), followed by and CNN and RF, while SVR lagged behind (RMSE = 0.0457, MAE = 0.0392, R2 = 0.800, MAPE = 4.80%). These findings are consistent with the literature, where sequence-based models such as LSTM and BiLSTM demonstrate superior ability to capture temporal degradation, while RF and CNN remain competitive. Support vector regression (SVR) showed weaker generalisation and struggled with sequence-dependent degradation patterns
Across published studies, optimisation algorithms (PSO, SSA, Bayesian methods) and hybrid models (CNN-LSTM, CNN-WNN-WLSTM, IGWO-LSTM) further improve accuracy, often reducing prediction errors below 1%. Emerging methods such as transformers, transfer learning, TCNs, and QCNNs indicate a shift toward models that combine local feature extraction, long-range dependency modelling, and computational efficiency.
The findings confirm that deep learning approaches, particularly those designed for sequential data, outperform classical machine learning methods in SOH estimation. Nevertheless, classical algorithms such as RF and SVR remain useful in contexts where computational efficiency and simplicity are prioritised. The results of both the literature review and simulation highlight the potential of hybrid models, which integrate complementary strengths of individual models to improve accuracy, robustness, and efficiency.
Future research should address several open challenges in the application of ML for battery SOH estimation:
  • Interpretability and Trust: Deep models such as LSTM and CNN remain black box in nature. Incorporating explainable AI techniques (e.g., SHAP, LIME) [70] can enhance transparency and facilitate adoption in safety-critical applications.
  • Model Robustness and Real-World Applicability: The simulations in this study used only the controlled NASA PCoE dataset (25 °C, fixed charge/discharge protocols), which does not fully represent real-world EV conditions such as fluctuating temperatures, dynamic load profiles, and irregular charging. Future work should validate these models on field datasets to assess robustness. Transfer learning and domain adaptation should be further explored to improve generalisation across chemistries and usage conditions [71].
  • Hybrid and Physics-Informed Models: Combining machine learning with physics-informed features [72] or models can improve accuracy while reducing data requirements, creating more generalisable frameworks.
  • Resource-Constrained Deployment: Simplified models (e.g., RF or lightweight neural networks) should be optimised for embedded deployment in battery management systems (BMSs), especially in low-cost EVs and grid storage applications.
  • Long-Term Prognostics: Future work should extend beyond short-term SOH estimation to include reliable prediction of remaining useful life (RUL) [62], which is critical for lifecycle optimisation and predictive maintenance.

Author Contributions

Conceptualisation, L.M. and K.R.; methodology, L.M. and K.R.; software, L.M.; validation, L.M., K.R. and A.B.; formal analysis, L.M.; investigation, L.M.; resources, K.R. and A.B.; writing—original draft preparation, L.M.; writing—review and editing, K.R. and A.B.; visualisation, L.M.; supervision K.R. and A.B.; project administration, K.R. and A.B.; funding acquisition, K.R. and A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the European Union’s Intra-Africa Academic Mobility Scheme under the grant agreement No. 101144276—NeDMEV. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Education and Culture Executive Agency. Neither the European Union nor the granting authority can be held responsible for them.

Data Availability Statement

The data supporting the results of this study were obtained from the NASA Prognostics Center of Excellence (PCoE). Available online: https://www.nasa.gov/intelligent-systems-division/discovery-and-systems-health/pcoe/pcoe-data-set-repository/ (accessed on 1 August 2025).

Acknowledgments

During the preparation of this manuscript, the authors used Chatgpt-4.5 for the purposes of refining the manuscript’s language, improving clarity, and enhancing overall readability. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Global EV Outlook 2025 Expanding Sales in Diverse Markets. Available online: www.iea.org (accessed on 15 August 2025).
  2. Padder, S.G.; Ambulkar, J.; Banotra, A.; Modem, S.; Maheshwari, S.; Jayaramulu, K.; Kundu, C. Data-Driven Approaches for Estimation of EV Battery SoC and SoH: A Review. IEEE Access 2025, 13, 35048–35067. [Google Scholar] [CrossRef]
  3. Wu, Z.; Zhang, Y.; Wang, H. Battery degradation diagnosis under normal usage without requiring regular calibration data. J. Power Sources 2024, 608, 234670. [Google Scholar] [CrossRef]
  4. Li, G.; West, A.C.; Preindl, M. Characterizing degradation in lithium-ion batteries with pulsing. J. Power Sources 2023, 580, 233328. [Google Scholar] [CrossRef]
  5. Lee, J.; Won, J. Enhanced Coulomb Counting Method for SoC and SoH Estimation Based on Coulombic Efficiency. IEEE Access 2023, 11, 15449–15459. [Google Scholar] [CrossRef]
  6. Bourelly, C.; Vitelli, M.; Milano, F.; Molinara, M.; Fontanella, F.; Ferrigno, L. EIS-Based SoC Estimation: A Novel Measurement Method for Optimizing Accuracy and Measurement Time. IEEE Access 2023, 11, 91472–91484. [Google Scholar] [CrossRef]
  7. Merrouche, W.; Lekouaghet, B.; Bouguenna, E.; Himeur, Y. Parameter estimation of ECM model for Li-Ion battery using the weighted mean of vectors algorithm. J. Energy Storage 2024, 76, 109891. [Google Scholar] [CrossRef]
  8. Kara, A. A data-driven approach based on deep neural networks for lithium-ion battery prognostics. Neural Comput. Appl. 2021, 33, 13525–13538. [Google Scholar] [CrossRef]
  9. Wang, R.; Xu, X.; Zhou, Q.; Zhang, J.; Wang, J.; Ye, J.; Wu, Y. State of Health Estimation for Lithium-Ion Batteries Using Enhanced Whale Optimization Algorithm for Feature Selection and Support Vector Regression Model. Processes 2025, 13, 158. [Google Scholar] [CrossRef]
  10. Zhang, Z.; Li, L.; Li, X.; Hu, Y.; Huang, K.; Xue, B.; Wang, Y.; Yu, Y. State-of-health estimation for the lithium-ion battery based on gradient boosting decision tree with autonomous selection of excellent features. Int. J. Energy Res. 2022, 46, 1756–1765. [Google Scholar] [CrossRef]
  11. Xu, H.; Wu, L.; Xiong, S.; Li, W.; Garg, A.; Gao, L. An improved CNN-LSTM model-based state-of-health estimation approach for lithium-ion batteries. Energy 2023, 276, 127585. [Google Scholar] [CrossRef]
  12. Khaleghi, S.; Hosen, M.S.; Van Mierlo, J.; Berecibar, M. Towards machine-learning driven prognostics and health management of Li-ion batteries. A comprehensive review. Renew. Sustain. Energy Rev. 2024, 192, 114224. [Google Scholar] [CrossRef]
  13. Awad, M.; Khanna, R. Support Vector Regression. In Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers; Awad, M., Khanna, R., Eds.; Apress: Berkeley, CA, USA, 2015; pp. 67–80. [Google Scholar] [CrossRef]
  14. Wilson, M.D. Support Vector Machines. In Encyclopedia of Ecology, Five-Volume Set; Jørgensen, S.E., Fath, B.D., Eds.; Academic Press: Cambridge, MA, USA, 2008; Volume 1–5, pp. 3431–3437. [Google Scholar] [CrossRef]
  15. Zhao, Q.; Qin, X.; Zhao, H.; Feng, W. A novel prediction method based on the support vector regression for the remaining useful life of lithium-ion batteries. Microelectron. Reliab. 2018, 85, 99–108. [Google Scholar] [CrossRef]
  16. Si, Q.; Matsuda, S.; Yamaji, Y.; Momma, T.; Tateyama, Y. Data-Driven Cycle Life Prediction of Lithium Metal-Based Rechargeable Battery Based on Discharge/Charge Capacity and Relaxation Features. Adv. Sci. 2024, 11, 2402608. [Google Scholar] [CrossRef]
  17. Feng, R.; Wang, S.; Yu, C.; Hai, N.; Fernandez, C. High precision state of health estimation of lithium-ion batteries based on strong correlation aging feature extraction and improved hybrid kernel function least squares support vector regression machine model. J. Energy Storage 2024, 90, 111834. [Google Scholar] [CrossRef]
  18. Chen, J.; Hu, Y.; Zhu, Q.; Rashid, H.; Li, H. A novel battery health indicator and PSO-LSSVR for LiFePO4 battery SOH estimation during constant current charging. Energy 2023, 282, 128782. [Google Scholar] [CrossRef]
  19. Vedhanayaki, S.; Indragandhi, V. A Bayesian Optimized Deep Learning Approach for Accurate State of Charge Estimation of Lithium Ion Batteries Used for Electric Vehicle Application. IEEE Access 2024, 12, 43308–43327. [Google Scholar] [CrossRef]
  20. Xia, X.; Chen, Y.; Shen, J.; Liu, Y.; Zhang, Y.; Chen, Z.; Wei, F. State of health estimation for lithium-ion batteries based on impedance feature selection and improved support vector regression. Energy 2025, 326, 136135. [Google Scholar] [CrossRef]
  21. Liu, S.; Fang, L.; Zhao, X.; Wang, S.; Hu, C.; Gu, F.; Ball, A. State-of-health estimation of lithium-ion batteries using a kernel support vector machine tuned by a new nonlinear gray wolf algorithm. J. Energy Storage 2024, 102, 114052. [Google Scholar] [CrossRef]
  22. Li, Q.; Li, D.; Zhao, K.; Wang, L.; Wang, K. State of health estimation of lithium-ion battery based on improved ant lion optimization and support vector regression. J. Energy Storage 2022, 50, 104215. [Google Scholar] [CrossRef]
  23. Stighezza, M.; Bianchi, V.; De Munari, I. FPGA Implementation of an Ant Colony Optimization Based SVM Algorithm for State of Charge Estimation in Li-Ion Batteries. Energies 2021, 14, 7064. [Google Scholar] [CrossRef]
  24. Zhu, T.; Wang, S.; Fan, Y.; Hai, N.; Huang, Q.; Fernandez, C. An improved dung beetle optimizer-hybrid kernel least square support vector regression algorithm for state of health estimation of lithium-ion batteries based on variational model decomposition. Energy 2024, 306, 132464. [Google Scholar] [CrossRef]
  25. Valizadeh, A.; Amirhosseini, M.H. Machine Learning in Lithium-Ion Battery: Applications, Challenges, and Future Trends. SN Comput. Sci. 2024, 5, 717. [Google Scholar] [CrossRef]
  26. Van Houdt, G.; Mosquera, C.; Nápoles, G. A review on the long short-term memory model. Artif. Intell. Rev. 2020, 53, 5929–5955. [Google Scholar] [CrossRef]
  27. Gong, Y.; Zhang, X.; Gao, D.; Li, H.; Yan, L.; Peng, J.; Huang, Z. State-of-health estimation of lithium-ion batteries based on improved long short-term memory algorithm. J. Energy Storage 2022, 53, 105046. [Google Scholar] [CrossRef]
  28. Wang, F.K.; Amogne, Z.E.; Chou, J.H.; Tseng, C. Online remaining useful life prediction of lithium-ion batteries using bidirectional long short-term memory with attention mechanism. Energy 2022, 254, 124344. [Google Scholar] [CrossRef]
  29. Yang, N.C.; Chen, W.C. State of health prediction method for battery cells using bidirectional long short-term memory neural network with time-varying filter empirical mode decomposition. J. Power Sources 2025, 656, 238007. [Google Scholar] [CrossRef]
  30. Peng, S.; Wang, Y.; Tang, A.; Jiang, Y.; Kan, J.; Pecht, M. State of health estimation joint improved grey wolf optimization algorithm and LSTM using partial discharging health features for lithium-ion batteries. Energy 2025, 315, 134293. [Google Scholar] [CrossRef]
  31. Wang, Y.; Zhu, J.; Cao, L.; Gopaluni, B.; Cao, Y. Long Short-Term Memory Network with Transfer Learning for Lithium-ion Battery Capacity Fade and Cycle Life Prediction. Appl. Energy 2023, 350, 121660. [Google Scholar] [CrossRef]
  32. Yang, G.; Ma, Q.; Sun, H.; Zhang, X. State of Health Estimation Based on GAN-LSTM-TL for Lithium-ion Batteries. Int. J. Electrochem. Sci. 2022, 17, 221128. [Google Scholar] [CrossRef]
  33. Hu, W.Y.; Zhang, C.L.; Luo, L.J.; Jiang, S.H. Integrated Method of Future Capacity and RUL Prediction for Lithium-Ion Batteries Based on CEEMD-Transformer-LSTM Model. Energy Sci. Eng. 2024, 12, 5272–5286. [Google Scholar] [CrossRef]
  34. Liu, Z.; Sun, Y.; Li, Y.; Liu, Y.; Chen, Y.; Zhang, Y. Lithium-ion battery health prognosis via electrochemical impedance spectroscopy using CNN-BiLSTM model. J. Mater. Inform. 2024, 4, 9. [Google Scholar] [CrossRef]
  35. Xiang, N.; Zhang, T.; Liu, Y. State of health prediction for Lithium-Ion batteries using extended long Short-Term memory network and frequency enhanced channel attention mechanism with Variational mode Decomposition. Measurement 2025, 249, 117084. [Google Scholar] [CrossRef]
  36. Krichen, M.; Mihoub, A. Long Short-Term Memory Networks: A Comprehensive Survey. AI 2025, 6, 215. [Google Scholar] [CrossRef]
  37. Mchara, W.; Manai, L.; Khalfa, M.A.; Raissi, M. Intelligent health state diagnosis of lithium-ion batteries for electric vehicles using wavelet-enhanced hybrid deep learning integrated with an attention mechanism. Clean Energy 2025, 9, 64–79. [Google Scholar] [CrossRef]
  38. Liang, M.; Gan, Y.; Chang, Z.; Wan, Z.; Schlangen, E.; Šavija, B. Microstructure-informed deep convolutional neural network for predicting short-term creep modulus of cement paste. Cem. Concr. Res. 2022, 152, 106681. [Google Scholar] [CrossRef]
  39. Lu, Z.; Fei, Z.; Wang, B.; Yang, F. A feature fusion-based convolutional neural network for battery state-of-health estimation with mining of partial voltage curve. Energy 2024, 288, 129690. [Google Scholar] [CrossRef]
  40. Chen, S.Z.; Liang, Z.; Yuan, H.; Yang, L.; Xu, F.; Fan, Y. A novel state of health estimation method for lithium-ion batteries based on constant-voltage charging partial data and convolutional neural network. Energy 2023, 283, 129103. [Google Scholar] [CrossRef]
  41. Liang, C.; Tao, S.; Huang, X.; Wang, Y.; Xia, B.; Zhang, X. Stochastic state of health estimation for lithium-ion batteries with automated feature fusion using quantum convolutional neural network. J. Energy Chem. 2025, 106, 205–219. [Google Scholar] [CrossRef]
  42. Xing, Q.K.; Sun, X.W.; Fu, Y.P.; Wang, K. Lithium-ion battery health estimate based on electrochemical impedance spectroscopy and CNN-BiLSTM-Attention. Ionics 2025, 31, 1389–1403. [Google Scholar] [CrossRef]
  43. Zraibi, B.; Okar, C.; Chaoui, H.; Mansouri, M. Remaining Useful Life Assessment for Lithium-Ion Batteries Using CNN-LSTM-DNN Hybrid Method. IEEE Trans. Veh. Technol. 2021, 70, 4252–4261. [Google Scholar] [CrossRef]
  44. Yao, Q.; Song, X.; Xie, W. State of health estimation of lithium-ion battery based on CNN–WNN–WLSTM. Complex Intell. Syst. 2024, 10, 2919–2936. [Google Scholar] [CrossRef]
  45. Zhang, Z.; Liu, X.; Zhang, R.; Liu, X.M.; Chen, S.; Sun, Z.; Jiang, H. Lithium-ion battery SOH estimation method based on multi-feature and CNN-KAN. Front. Energy Res. 2024, 12, 1494473. [Google Scholar] [CrossRef]
  46. Peng, C.; Wang, M.; Li, C.; Lv, Y. A speedily-accurately prediction model based on convolutional neural network enhanced feature extraction for lithium-ion batteries remaining life. J. Energy Storage 2025, 133, 118055. [Google Scholar] [CrossRef]
  47. Wu, M.; Zhang, X.; Wang, Z.; Tan, C.; Wang, Y.; Wang, L. State of health estimation of lithium-ion batteries based on the Kepler optimization algorithm-multilayer-convolutional neural network. J. Energy Storage 2025, 122, 116644. [Google Scholar] [CrossRef]
  48. Liu, P.; Liu, C.; Wang, Z.P.; Wang, Q.S.; Han, J.L.; Zhou, Y.P. A Data-Driven Comprehensive Battery SOH Evaluation and Prediction Method Based on Improved CRITIC-GRA and Att-BiGRU. Sustainability 2023, 15, 15084. [Google Scholar] [CrossRef]
  49. Li, Y.; Qin, X.; Chai, M.; Wu, H.; Zhang, F.; Jiang, F.; Wen, C. SOH evaluation and RUL estimation of lithium-ion batteries based on MC-CNN-TimesNet model. Reliab. Eng. Syst. Saf. 2025, 261, 111125. [Google Scholar] [CrossRef]
  50. Ding, P.; Xia, M.; Wang, X.; Pan, H.; Gao, Q.; Guo, W.; Shi, P.; Min, Y. Advanced lithium-ion battery health state estimation using a Bayesian optimization hybrid neural network model. J. Energy Storage 2025, 123, 116562. [Google Scholar] [CrossRef]
  51. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  52. Chen, Z.; Sun, M.; Shu, X.; Shen, J.; Xiao, R. On-board state of health estimation for lithium-ion batteries based on random forest. In Proceedings of the 2018 IEEE International Conference on Industrial Technology (ICIT), Lyon, France, 20–22 February 2018; pp. 1754–1759. [Google Scholar] [CrossRef]
  53. Amamra, S.A. Random Forest-Based Machine Learning Model Design for 21,700/5 Ah Lithium Cell Health Prediction Using Experimental Data. Physchem 2025, 5, 12. [Google Scholar] [CrossRef]
  54. Liang, Z.; Wang, R.; Zhan, X.; Li, Y.; Xiao, Y. Lithium-Ion Battery State-of-Health Prediction for New-Energy Electric Vehicles Based on Random Forest Improved Model. Appl. Sci. 2023, 13, 11407. [Google Scholar] [CrossRef]
  55. Lamprecht, A.; Riesterer, M.; Steinhorst, S. Random Forest Regression of Charge Balancing Data: A State of Health Estimation Method for Electric Vehicle Batteries. In Proceedings of the 2020 International Conference on Omni-layer Intelligent Systems (COINS), Virtually, 31 August–2 September 2020; pp. 1–6. [Google Scholar] [CrossRef]
  56. Wang, X.; Hu, B.; Su, X.; Xu, L.; Zhu, D. State of Health estimation for lithium-ion batteries using Random Forest and Gated Recurrent Unit. J. Energy Storage 2024, 76, 109796. [Google Scholar] [CrossRef]
  57. Yang, N.; Song, Z.; Hofmann, H.; Sun, J. Robust State of Health estimation of lithium-ion batteries using convolutional neural network and random forest. J. Energy Storage 2022, 48, 103857. [Google Scholar] [CrossRef]
  58. Lu, D.; Cui, N.; Li, C. A Novel Transfer Learning Framework Combining Attention Mechanisms and Random Forest Regression for State of Health Estimation of Lithium-Ion Battery with Different Formulations. IEEE Trans. Ind. Appl. 2024, 60, 5726–5736. [Google Scholar] [CrossRef]
  59. Garse, K.M.; Bairwa, K.N.; Roy, A. Hybrid Random Forest Regression and Artificial Neural Networks for Modelling and Monitoring the State of Health of Li-Ion Battery. J. Electr. Syst. 2024, 20, 2231–2243. [Google Scholar] [CrossRef]
  60. Chen, J.; Kollmeyer, P.; Ahmed, R.; Emadi, A. Battery state-of-health estimation using CNNs with transfer learning and multi-modal fusion of partial voltage profiles and histogram data. Appl. Energy 2025, 391, 125923. [Google Scholar] [CrossRef]
  61. Sulaiman, M.H.; Mustaffa, Z. State of charge estimation for electric vehicles using random forest. Green Energy Intell. Transp. 2024, 3, 100177. [Google Scholar] [CrossRef]
  62. Duan, W.; Song, S.; Xiao, F.; Chen, Y.; Peng, S.; Song, C. Battery SOH estimation and RUL prediction framework based on variable forgetting factor online sequential extreme learning machine and particle filter. J. Energy Storage 2023, 65, 107322. [Google Scholar] [CrossRef]
  63. Shu, X.; Yang, H.; Liu, X.; Feng, R.; Shen, J.; Hu, Y.; Chen, Z.; Tang, A. State of health estimation for lithium-ion batteries based on voltage segment and transformer. J. Energy Storage 2025, 108, 115200. [Google Scholar] [CrossRef]
  64. Chen, Y.; Huang, X.; He, Y.; Zhang, S.; Cai, Y. Edge–cloud collaborative estimation lithium-ion battery SOH based on MEWOA-VMD and Transformer. J. Energy Storage 2024, 99, 113388. [Google Scholar] [CrossRef]
  65. Yang, Z.; Li, Y.; Yang, D.; Feng, X.; Sun, Y.; Zhao, Y.; Wu, C.; Pan, R. Enhanced CNN-based state-of-health estimation framework for lithium-ion batteries using variable-length charging segments and transfer learning. J. Energy Storage 2025, 128, 117214. [Google Scholar] [CrossRef]
  66. Arbaoui, S.; Samet, A.; Ayadi, A.; Mesbahi, T.; Boné, R. Data-driven strategy for state of health prediction and anomaly detection in lithium-ion batteries. Energy AI 2024, 17, 100413. [Google Scholar] [CrossRef]
  67. Jafari, S.; Byun, Y.C. Interpretable AI for explaining and predicting battery state of health using PSO-enhanced deep learning models. Energy Rep. 2025, 14, 1779–1798. [Google Scholar] [CrossRef]
  68. Karthikeyan, M.; Anirudh, N.; Columbus, C.; Aravind, C.K. Optimizing battery health monitoring in electric vehicles using interpretable CART–GX model. Results Eng. 2025, 27, 106043. [Google Scholar] [CrossRef]
  69. Saha, B.; Goebel, K. Battery Data Set; NASA Ames Research Center Moffett Field: Mountain View, CA, USA, 2007. Available online: https://www.nasa.gov/intelligent-systems-division/discovery-and-systems-health/pcoe/pcoe-data-set-repository/ (accessed on 3 October 2025).
  70. Etem, T. Interpretable machine learning for battery health insights: A LIME and SHAP-based study on EIS-derived features. Bull. Pol. Acad. Sci. Tech. Sci. 2025, 73, e155033. [Google Scholar] [CrossRef]
  71. Cao, Z.; Gao, W.; Fu, Y.H.; Kurdkandi, N.V.; Mi, C. A general framework for lithium-ion battery state of health estimation: From laboratory tests to machine learning with transferability across domains. Appl. Energy 2025, 381, 125086. [Google Scholar] [CrossRef]
  72. Tang, A.; Xu, Y.; Hu, Y.; Tian, J.; Nie, Y.; Yan, F.; Tan, Y.; Yu, Q. Battery state of health estimation under dynamic operations with physics-driven deep learning. Appl. Energy 2024, 370, 123632. [Google Scholar] [CrossRef]
Figure 3. General hierarchy of a convolutional neural network, adapted from ref. [37].
Figure 3. General hierarchy of a convolutional neural network, adapted from ref. [37].
Wevj 16 00594 g003
Figure 5. Raw capacity degradation curves.
Figure 5. Raw capacity degradation curves.
Wevj 16 00594 g005
Figure 6. Raw SOH degradation curves.
Figure 6. Raw SOH degradation curves.
Wevj 16 00594 g006
Figure 7. Predicted vs. actual SOH across the different ML models.
Figure 7. Predicted vs. actual SOH across the different ML models.
Wevj 16 00594 g007
Figure 8. Bar graph showing RMSE and MAE across all the models.
Figure 8. Bar graph showing RMSE and MAE across all the models.
Wevj 16 00594 g008
Table 1. Common kernels and their equations.
Table 1. Common kernels and their equations.
Linear Kernel K x i , x = ( x i , x )
Polynomial Kernel K x i , x = ( γ ( x i , x ) + r ) d
Radial Basis Function Kernel (Gaussian Kernel) K x i , x = exp ( γ x i x 2 )
Sigmoid Kernel K x i , x = tanh ( γ ( x i , x ) + r )
where γ is the scaling factor, r is the bias, and d is the degree of polynomial. Proper selection and tuning of kernel parameters are critical to model performance and generalisation.
Table 2. Summary of research showing author, ML algorithm, dataset used and key findings.
Table 2. Summary of research showing author, ML algorithm, dataset used and key findings.
1Zhao et al. (2018) [15]SVR (RBF)NASA + CALCEMAE < 2% SOH; epsilon-insensitive loss helped robustness to outliers/noisy voltage–capacity data.
2Si et al. (2024) [16]SVR Lab datasetsR2 up to 0.962 on test sets with full discharge; strong performance on partial windows except at high-voltage ranges.
3Feng et al. (2024) [17]Hybrid-kernel LSSVR (poly + RBF)CALCEHybrid kernel improved learning/generalisation and helped avoid overfitting vs. single-kernel SVMs.
4Chen et al. (2023) [18]PSO-LSSVR (LiFePO4, CC charging)Partial charge V/I; 2 test cellsReported MSE < 0.00052, MARE < 0.93%; PSO improved kernel/penalty tuning; robustness to noise/outliers.
5Liu et al. (2024) [21]KSVM tuned by nonlinear GWOCambridge Cavendish Lab dataset (>20,000 EIS spectra, 12 LR2032 LIBs, 25–45 °C)GWO-tuned KSVM improved accuracy/stability; longer runtime with larger data.
6Li et al. (2022) [22]Improved ant lion optimisation + SVRNASAMeta-heuristic tuning enhanced SVR SOH accuracy.
7Stighezza et al. (2021) [23]ACO-based SVM (FPGA)Panasonic 18650PF dataset (NN driving cycle + US06)Focus is SoC (not SOH); demonstrates hardware feasibility of ACO-SVM.
8Zhu et al. (2024) [24]Improved dung beetle optimiser + hybrid-kernel LSSVR (+ VMD)NASA + CALCEHybrid LSSVR + decomposition achieved high SOH accuracy; robust modelling.
9Gong et al. (2022) [27]PSO-optimised LSTM (RMSProp + dropout)Experimental Li-ion; 4 HIs via GRA~5% accuracy improvement vs. vanilla LSTM; better convergence/overfitting control.
10Wang, Amogne, Chou, Tseng (2022) [28]BiLSTM + attentionNASA + Univ. MarylandNoise-reduced capacity + attention to critical time steps ⇒ improved online RUL/SOH robustness across battery types; higher compute cost.
11Wang et al. (2023) [31]Transfer-learning LSTM (NCA→NCM)NASA + custom (NCA/NCM)Accurate cross-chemistry SOH/cycle life prediction; depends on source-target similarity and pretraining data size.
12Hu et al. (2024) [33]CEEMD–transformer-LSTMTwo lab datasetsRobust accuracy across operating conditions; framework complexity; lacks field validation.
13Liu et al. (2024) [34]CNN-BiLSTM (EIS)Raw EISOutperformed GPR, CNN, and LSTM; achieved R2 up to 0.89 for SOH estimation; enabled early-life RUL prediction from first 50 cycles.
15Chen et al. (2018) [52]RF (OOB-tuned)Sandia National Lab public LFP datasetRF SOH feasible on partial on-board data; optimisation via OOB; tested on two cells.
16Amamra (2025) [53]RF (grid search)21,700/5 Ah cellsR2 = 0.92, RMSE = 0.06; compared against SVR.
17Liang et al. (2023) [54]PSO-optimised RF (LiFePO4 modules)Charge/discharge featuresImproved SOH, but heavy manual feature extraction; limited automation/portability.
18Wang et al. (2024) [56]RF + GRU (hybrid)Not specifiedHybrid improved estimation accuracy by ≥15.84% vs. standalone RF.
19Yang et al. (2022) [57]CNN + RF (hybrid)Public datasets34–46% MAE reduction vs. baselines; cautions on generalisation across chemistries/form factors.
20Bairwa & Roy (2024) [59]RF + ANN (hybrid)V/I/T + cycle featuresRF for nonlinear features + ANN refinement; does not model long-range sequence dynamics.
22Xing et al. (2025) [42]CNN-BiLSTM–attention (EIS)EISAttention + BiLSTM with CNN improves temporal modelling and accuracy under varied conditions.
23Zraibi et al. (2021) [43]CNN-LSTM-DNN (hybrid)NASA + CALCEHybrid deep network for RUL/SOH achieved strong performance.
24Yao et al. (2024) [44]CNN-WNN-WLSTM (hybrid)NASA battery datasets (No. 5, 6, 7)Combines WNN’s fast convergence/stability with WLSTM; robust SOH estimation.
25Zhang et al. (2024) [45]CNN-KAN (hybrid)Lab-generated CC–CV battery cycling data (4 charge rates)CNN fused with KAN for SOH; improved performance with multi-feature inputs.
26Wu et al. (2025) [47]Multilayer CNN + Kepler optimisation (KOA)Lab datasetsKOA-tuned CNN cut MAE by up to 58.97%; depth > 2 layers degraded accuracy.
27Liu et al. (2023) [48]Att-BiGRU (improved GWO)Real-world EV operational dataAttention-BiGRU with improved GWO delivered comprehensive SOH evaluation/prediction. (GRU noted in CNN section as related.)
28Li et al. (2025) [49]MC-CNN-TimesNet + TPEMulti-cycle time series (2D tensors)Avg. RMSE within 1.5% by leveraging inter-timescale dependencies.
29Ding et al. (2025) [50]Bayesian-optimised hybrid NN (CNN + LSTM)NASA + labBO prevented local minima; RMSE < 1% in joint validation.
30Chen, Kollmeyer, Ahmed, Emadi (2025) [60]CNN with TL + multi-modal fusionPartial voltage profiles + histogramsTransfer learning + multi-modal fusion improved SOH; supports re-use across datasets.
31Vedhanayaki & Indragandhi (2025) [19]Bayesian optimised SVR (BO-SVR) with Gaussian kernelNASA B0005Achieved RMSE = 0.0082, outperforming standard SVR and GPR; Bayesian optimisation improved efficiency.
32Peng et al. (2025) [46]Convolutional-ProbSparse-transformer (CPT)NCM (BN-74, BN-100) + public datasets3.2× faster training vs. transformer; accuracy ↑ 112%; robust for long-term prediction
33Chen et al. (2023) [40]CNN with partial CV charging + TLMulti-chemistry datasetOnly first 1000 s of CV data needed; accurate SOH prediction; transfer learning improved generalisation
34Lu et al. (2024) [39]Feature fusion CNN (capacity–voltage + derivatives)18 batteries, 3 datasetsMAE ≤ 0.0028, MAPE ≤ 0.32%; intra-cycle + inter-cycle features improved accuracy.
35Yang et al. (2022) [32]GAN-LSTM–transfer learningNASA+ CALCEErrors < 3%; GAN mitigated data scarcity, TL improved cross-dataset adaptation
36Xiang et al.(2025) [35]VMD + xLSTM + FECAM hybridCALCE+ NASAMAE = 0.0049, RMSE = 0.0085, outperformed LSTM, GRU, transformer.
37Yang and Cheng (2025) [29]TVF-EMD + BiLSTM with BOHB tuningNASA B0005-B0007RMSE < 0.0016, R2 > 0.999, better than CNN-LSTM, SVR; robust across datasets.
38Xia et al. (2025) [20]Sine-SSA optimised SVR with EIS-based feature selectionCommercial EIS dataset (>20 k samples, 25–45 °C)Max error = 2.58%, outperforming LSTM, GPR; robust under varying temperatures.
39Lamprecht et al. (2020) [55]RF regressionSimulated Tesla Model S battery pack (96S74P, 85 kWh) with aging + charge balancing frameworkAchieved 1.94% error for capacity SOH and 4.28% for resistance SOH, outperforming other ML methods.
40Lu et al. (2024) [58]CNN + attention + random forest with transfer learningOxford dataset (LCO pouch cells), Sandia dataset (LFP, NCA 18,650 cells)Reduced estimation errors by 80.9% (LCO), 41.3% (LFP), and 25.6% (NCA). Achieved very low RMSE (0.46% for LFP).
Table 3. Summary of the extracted health features and their definitions.
Table 3. Summary of the extracted health features and their definitions.
Feature NameMathematical ExpressionDefinition of Symbols
Mean x ¯ = 1 T x t x t   :     signal   at   time   t :   for   each   signal   x { V , I , T , d V , d I , d T } ;
T: number of samples
Standard deviation σ x = 1 T 1 ( x t x ˉ ) 2 x ˉ : mean of signal
10th percentile x 10 = p e r c e n t i l e ( x ,   10 ) 10 %   quantile   of   all   x t values
90th percentile x 90 = p e r c e n t i l e ( x ,   90 ) 90 %   quantile   of   all   x t values
Quartile slopes V 50 V 25 0.25 ,   V 75 V 50 0.25 V 25 ,   V 50 ,   V 75 : voltages at 25%, 50%, and 75% of discharge
Overall voltage slope V 90 V 10 0.8 V 10 ,   V 90 : voltages at 10% and 90% of discharge
Maximum rate of change | d V | m a x = m a x d V t d V t = V t V t 1 : voltage derivative
Peak-to-peak derivative amplitude d V p t p = m a x ( d V t ) m i n ( d V t ) d V p t p : peak-to-peak derivative voltage
Correlation (V, I, T) c o v V , I σ V σ I , c o v V , T σ V σ T c o v I , T σ I σ T   c o v (   , ) : covariance between signals
Table 4. Summary of the hyperparameters used for each ML model.
Table 4. Summary of the hyperparameters used for each ML model.
ML ModelHyperparametersHyperparameters Search SpaceFinal Value Used
Random Forest (RF)Number of trees{300, 600, 800}600
Minimum leaf size{1, 2, 4}2
Support Vector Regression (SVR)C{1, 10, 100}10
ε{0.01, 0.03, 0.1}0.03
Kernel{0.3, 1, 3}1
CNNConvolution filtersConv filters: [64 (k = 7),Conv1D(64,7)
128 (k = 5), 128 (k = 3)];Conv1D(128,5)
Conv1D(128,3)
FC layer size{128, 256} 256
Dropout{0.2, 0.3}0.3
Epochs{80, 100, 120}100
Batch size{16, 32}32
BiLSTMHidden units{64, 128}biLSTM(128, sequence)
biLSTM(64, last)
Layers{1, 2}2
Dropout{0.25, 0.3}0.25/0.3
FC layer size{64, 128}128
Epochs{50, 60, 80}60
Batch size{8, 16}16
Table 5. Performance metrics across the ML models.
Table 5. Performance metrics across the ML models.
ModelRMSEMAER2MAPE (%)
CNN0.01940.01600.9641.84
LSTM0.01460.01180.9801.39
SVR0.04570.03920.8004.80
RF0.02070.01720.9592.15
Table 6. The training time of the four machine learning models.
Table 6. The training time of the four machine learning models.
ML ModelTraining Time (s)
CNN57.95
SVR5.29
RF26.11
LSTM201.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mbagaya, L.; Reddy, K.; Botes, A. Machine Learning Techniques for Battery State of Health Prediction: A Comparative Review. World Electr. Veh. J. 2025, 16, 594. https://doi.org/10.3390/wevj16110594

AMA Style

Mbagaya L, Reddy K, Botes A. Machine Learning Techniques for Battery State of Health Prediction: A Comparative Review. World Electric Vehicle Journal. 2025; 16(11):594. https://doi.org/10.3390/wevj16110594

Chicago/Turabian Style

Mbagaya, Leila, Kumeshan Reddy, and Annelize Botes. 2025. "Machine Learning Techniques for Battery State of Health Prediction: A Comparative Review" World Electric Vehicle Journal 16, no. 11: 594. https://doi.org/10.3390/wevj16110594

APA Style

Mbagaya, L., Reddy, K., & Botes, A. (2025). Machine Learning Techniques for Battery State of Health Prediction: A Comparative Review. World Electric Vehicle Journal, 16(11), 594. https://doi.org/10.3390/wevj16110594

Article Metrics

Back to TopTop