1. Introduction
Lithium-ion batteries are currently the most used energy storage system in electric vehicles (EVs) due to their higher performance in comparison with other energy storage systems: high power and energy density, slow self-discharge rate, and fast response time [
1,
2]. However, as with any electrochemical system, lithium-ion batteries suffer from different irreversible reactions that lead to their degradation and the loss of their characteristics [
3]. Ageing mechanisms mainly lead to battery capacity loss and internal impedance increase [
4]. Therefore, EV batteries’ prognosis and health management (PHM) are necessary for efficient and secure battery system management. Regarding vehicular applications, the PHM methods focus on monitoring and forecasting the battery capacity evolution to assess battery health status [
5]. The state of health (SoH) prediction methods can be model-based or data-driven.
The model-based methods consist of fitting the parameters of the degradation model using available battery measurements. The SoH prediction is obtained by extrapolating the fitted ageing model to the upcoming cycles [
6]. Usually, the battery prognostic is done with particle filters (PF) to fit and update model parameters. The study presented in [
7] used a generalised Dakin degradation model to predict the capacity of LFP and LTO cells. The model parameters are optimised using PF. The predicting error is less than 6.64%. In [
8], the authors used an exponential model to predict the capacity of LCO cells while the model parameters were fitted with an improved particle filter (PF) algorithm. The study in [
9] used a grey model that combines the exponential, linear, and polynomial models to predict the battery capacity. The model’s parameters are fitted using PF. The accuracy of the method is verified using the NASA Public Lithium-Ion Battery Data Set. The results show that combining three models achieves higher prediction accuracy than single-model results.
Data-driven approaches to battery prognosis proposed in the literature use sequences of previous capacity values to predict their values over future cycles [
10]. Among the different approaches, machine learning techniques such as Gaussian Process Regression, Support Vector Machine, and deep-learning methods such as recurrent neural networks (RNNs) are widely referenced [
10,
11,
12,
13,
14,
15,
16,
17,
18]. The study in [
12] proposed and evaluated an optimised LSTM model for capacity prediction. This method was evaluated based on experimental tests of battery cycling of NCA cells under two different temperatures. The results showed that the capacity was predicted with a mean absolute error (MAE) of less than 0.012Ah. In [
13], the authors used the GPR model to predict the capacity of NCA cells from calendar ageing data. The RMSE is less than 0.0105Ah. In [
14], the authors used an LSTM model for battery prognosis. First, the capacity trends are decomposed using complete ensemble empirical mode decomposition and principal component analysis. Second, the LSTM model based on transfer learning used the decomposed signal to predict the cell’s capacity. The relative error of the prediction is less than 1.9%.
Finally, although these prediction methods show good accuracy and robustness for different battery chemistry capacity predictions, the data used to design the models are generally obtained from experimental laboratory tests. This limits their use in electric vehicles because the tests needed to measure capacity are time-consuming, expensive, and require specific conditions [
15,
16]. In addition, the ageing behaviour varies between cells due to several factors, such as working conditions, position of the cells in the module, and driver behaviour [
17]. Therefore, the prognosis methods should take into account the ageing trajectory of each cell to predict its future evolution. The prognosis frameworks proposed in the literature require significant computing resources and usually combine several algorithms and mechanisms for SoH prediction [
18,
19], making them difficult to implement for on-board training. A practical method of predicting the state of health (SoH) of batteries in electric vehicles should be based on a history of capacity estimated on-board to take into consideration the ageing trajectory of each cell.
Concerning data-driven prediction methods, recurrent neural networks such as LSTM, GRU, and their variants have shown their efficiency for different time-series prediction tasks, particularly for lithium-ion prognostics [
10,
11,
12,
13,
14,
15,
16,
17,
18,
19]. The GPR model has proven its efficiency for predicting battery ageing across various studies, due to its low computational cost and minimal data requirements [
13,
20]. Recently, echo state network (ESN) has emerged as an alternative to gradient descent training-based approaches and has been successfully applied for time-series prediction [
21]; the main advantages of ESN are the capacity to learn from a small database and the simplicity of its training process (linear regression problem), which makes it convenient for applications where computational sources are limited [
22]. In the following, the four prediction approaches previously mentioned are evaluated and compared: LSTM, GRU, ESN, and GPR. The objective is to assess and compare simple architectures for capacity prediction to select algorithms that achieve higher accuracy with less computational power for electric vehicles. These methods are first trained with measured capacity values to test their ability to predict the battery ageing trend. In the second step, the prediction models are trained using the estimated capacity. This step allows for assessing their accuracy and robustness to the estimation errors in the training data, which is close to the actual conditions under which vehicle battery capacity is estimated. Finally, the prediction models are evaluated using training data at three ageing stages of the cells (initial, medium, and advanced), and they allow the assessment of long-term, medium-term, and short-term prediction performance. The model training and validation data are generated experimentally from battery cycling of two cell chemistries (LFP and NMC) with two different ageing trends. Results are compared and evaluated in terms of prediction accuracy and learning time. The main contributions of this paper are:
- −
The assessment and comparison of four machine learning methods (LSTM, GRU, GPR, and ESN) for the capacity prediction of lithium-ion cells for vehicular applications.
- −
The prediction models are compared by considering two scenarios: the capacity values are measured or estimated, and different percentages of the training data are considered.
- −
The models are evaluated and compared using data from two different ageing trends extracted from the cycling of LFP and NMC cells.
The rest of the paper is organised as follows:
Section 2 details the prediction structure, the experimental tests and the training data, and the prognosis models.
Section 3 presents the prediction results and their analysis. A conclusion closes the paper.
3. Results
3.1. NMC Cells
Figure A1,
Figure A2,
Figure A3,
Figure A4,
Figure A5 and
Figure A6 present the evolution of the predicted and measured capacity using measured and estimated data for all NMC cells.
Table 4 and
Table 5 present the MAPE of capacity prediction of each cell, using experimental and estimated training data respectively. The mean and standard deviation of the MAPE values are also provided.
Figure 6 and
Figure 7 present the evolution of the mean MAPE and the standard deviation for NMC cells for each prediction model when the prediction models are trained with measured and estimated data, respectively.
In the model trained with measured capacity, the best results are obtained with ESN (MAPE average value lower than 1% even with only 40% of the training dataset), followed by LSTM and GRU (MAPE average value around 1%). Regarding GPR, the results show that non-uniformity in the ageing trend significantly increases the prediction error, which is the case for cells #3 and #4:
- -
At initial ageing (40% of the dataset), the error is no greater than 2% for LSTM and GRU and 1.5% for ESN. At the same time, the error reaches 25% for GPR.
- -
At average ageing (60% of the dataset), the error is lower than 2% for LSTM and 2.2% for GRU. These errors are higher than those of the initial ageing due to the higher prediction error of cell #4. Regarding ESN, the MAPE reaches 1.07%. In comparison, the error reaches 8% for GPR.
- -
In advanced ageing conditions (80% of the dataset), the four prediction models present a lower prediction error than the first two ageing cases. For LSTM and ESN models, the error is lower than 0.91%, less than 0.51% for GRU, and 1.45% for GPR.
Regarding the performance of the prediction models trained with estimated capacity values, the prediction errors are higher than in the previous case because they include the capacity estimation errors. The predicted capacities converge to the target values with maximum errors of 2.73% for LSTM, 3.27% for GRU, 2.12% for ESN, and 32.74% for GPR. Regarding the evolution of prediction accuracy, with the increasing of training percentages, the following conclusions can be drawn:
- -
At initial ageing (40% of the dataset), the maximum error is no greater than 1.60% for LSTM, 2.51% for GRU, and 1.07% for ESN. The error reaches 32.74% for GPR.
- -
At average ageing (60% of the dataset), the prediction error reaches 1.28%, 2.10%, 1.74%, and 14.32% for LSTM, GRU, ESN, and GPR, respectively.
- -
At advanced ageing conditions (80% of the dataset), the highest prediction errors for LSTM, GRU, and ESN increase to 2.73%, 3.27%, and 2.12%, respectively. Meanwhile, the mean error is drastically reduced to 0.88% with GPR.
3.2. LFP Cells
Figure A7,
Figure A8,
Figure A9,
Figure A10,
Figure A11 and
Figure A12 present the evolution of the predicted and measured capacity when the prediction models are trained using measured and estimated data.
Table 6 and
Table 7 display the MAPE of all models trained with estimated and measured data, and the mean and the standard deviation are also provided.
Figure 8 and
Figure 9 present the evolution of the average MAPE and the standard deviation for LFP cells, in the cases of measured and training sets, respectively.
In the case of using experimental data as a training set, the LSTM, GRU, and ESN present a maximum prediction error of up to 2% for all study cases. Regarding the GPR model, the prediction error varies from 13.19% to 0.43% when 40% or 80% of the dataset is used for training, respectively. The errors are more significant for the cells whose ageing evolution is non-uniform, as for cell #2.
Regarding the model’s performance when trained with experimental data, the following conclusions can be drawn:
- −
At initial ageing (40% of the dataset), the three recurrent networks (LSTM, GRU, and ESN) achieve good results: the error is less than 1.93% for LSTM, 1.06% for GRU, and 0.64% for ESN. The GPR prediction error is not higher than 3% for cells #1, #3, and #4, while for cell #2, it reaches 13.19%.
- −
At average ageing (60% of the dataset), the prediction accuracy increases: the error is not higher than 0.42% for LSTM, 0.87% for GRU, and 0.37% for ESN. Regarding GPR, the prediction error is still higher for cell #2 (less than 4%), while it is not higher than 0.7% for the other cells.
- −
The four prediction models perform better at advanced ageing conditions (80% of the dataset). The error is lower than 1% for LSTM and GRU and lower than 0.5% for ESN and GPR.
Regarding the models’ performance when they are trained with estimated data, the prediction results of LFP are in agreement with those of NMC cells. The prediction errors are higher when the models are trained with measured capacity (
Table 6). The maximum MAPE is less than 3.15% for LSTM, 3.35% for GRU, and 2.31% for ESN. The GPR is the most affected by the estimated capacity errors, with an MAPE as high as 22.54%:
- -
At initial ageing (40% of the cycles), the three recurrent networks (LSTM, GRU, and ESN) achieve good results. The maximum error is less than 2.72% for LSTM, 2.59% for GRU, and 2.31% for ESN. Meanwhile, the prediction error for GPR varies between 16.75% and 22.54%.
- -
At average ageing (60% of the dataset), the prediction errors decrease to 1.22% and 1.99% for LSTM and ESN, respectively. Meanwhile, for GRU, the MAPE increases to 3.34%. GPR still presents a higher prediction error, ranging from 3.15% to 7.4%, compared to the three recurrent networks.
- -
At advanced ageing (80% of the dataset), the prediction errors increase to 3.15% and 3.35% for LSTM and GRU, respectively. The MAPE for ESN is less than 1.92%, while for GPR, the prediction error is less than 2.03%.
3.3. Training Time
Finally,
Table 8 displays the training times and the memory usage during training of each prediction model, considering the three dataset percentages (40%, 60%, and 80%) for NMC and LFP cells. The prediction algorithms are developed in a Python[DP1] [AH2] 3.10.19 environment using TensorFlow 2.20.0 and GPflow 2.10.0 libraries. The algorithms are designed and evaluated in a Windows 10
® environment. The computer has an i-5 processor with 2.50 (GHz) and 16 GB of RAM.
As expected, the results show that the more the data percentage increases, the more the required training time increases. The comparative evaluation shows that LSTM has the longest training duration, followed by GRU and GPR. ESN has a shorter training time (less than 0.15s), following the description of the model’s architecture. Regarding memory usage, ESN exhibits the lowest memory requirement, followed by GPR, LSTM, and GRU. The GRU requires more memory than the LSTM because it uses two hidden layers (
Table 2).
The evaluation of the prognosis models on both NMC and LFP cells shows that the three recurrent networks (LSTM, GRU, and ESN) exhibit good prediction accuracy at the three ageing stages (40%, 60%, and 80%). The results also highlight that the prediction errors differ from one cell to another. These differences are related to the ageing evolution of each cell. Regarding GPR, the evaluation results indicate its limitations for long-term prediction and its sensitivity to the non-uniformity of the ageing trend [
42]. The limitation of the GPR for long-term prediction is related to using the local kernel. As presented in
Figure 10, the higher the cycle number, the higher the distance
; thus, the kernel function decays. Consequently, the prediction converges to the mean value of the function that was set to zero [
33,
34].
Regarding LSTM and GRU networks, the results confirm the efficiency of the gating mechanisms for filtering the training data (Equations (2)–(4), (8) and (9)) [
43]. The high accuracy of the LSTM is due to the superiority of the gating mechanisms compared with GRU. The results obtained for the ESN also confirm its efficiency in handling time-series problems and capacity to learn from noisy data, as reported in several studies [
44,
45].
The results of the four prediction models, when using the measured dataset, show that the higher the percentage of cycles used for model fitting, the better the prediction accuracy, and the greater the increase in prediction accuracy. When the prediction models are trained using estimated capacity values, the results show that the prediction error increases due to the estimation errors and outliers in the training data. However, the mean error for LSTM, GRU, and ESN is always less than 2% (between 1 and 1.5% for LSTM and ESN). Finally, comparing the prognosis models, ESN outperforms the other models’ accuracy and computational time, making it the most suitable for vehicular applications [
46].
Table 9 presents the MAPE of the proposal compared with other prognostic models from the literature. The results show that even when the prediction models (LSTM and ESN) are trained with estimated capacity values, the prediction accuracy is comparable to that obtained with models trained on measured capacity values.