Next Article in Journal
Cutting Performance of TiN/DLC-Coated Cemented Carbide Tool in Dry Cutting of Laser-Clad Cr-Ni-Based Steel
Previous Article in Journal
Multi-Factor Analysis of Cutting Parameters for Bamboo Milling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on the SOH of Lithium Batteries Based on the TCN–Transformer–BiLSTM Hybrid Model

School of Energy and Power Engineering, North University of China, Taiyuan 030051, China
*
Author to whom correspondence should be addressed.
Coatings 2025, 15(10), 1149; https://doi.org/10.3390/coatings15101149
Submission received: 27 August 2025 / Revised: 20 September 2025 / Accepted: 29 September 2025 / Published: 2 October 2025

Abstract

Lithium-ion batteries are widely used in energy storage and power systems due to their high energy density, long cycle life, and stability. Accurate prediction of the state of health (SOH) of batteries is critical to ensuring their safe and reliable operation. However, the prediction task remains challenging due to various complex factors. This paper proposes a hybrid TCN–Transformer–BiLSTM prediction model for battery SOH estimation. The model is first validated using the NASA public dataset, followed by further verification with dynamic operating condition simulation experimental data. Health features correlated with SOH are identified through Pearson analysis, and comparisons are conducted with existing LSTM, GRU, and BiLSTM methods. Experimental results demonstrate that the proposed model achieves outstanding performance across multiple datasets, with root mean square error (RMSE) values consistently below 2% and even below 1% in specific cases. Furthermore, the model maintains high prediction accuracy even when trained with only 50% of the data.

Graphical Abstract

1. Introduction

Lithium-ion batteries have outstanding advantages such as high energy density, long cycle life, low self-discharge rate, no memory, and good stability [1,2], and have achieved a high status in the field of power batteries as well as various types of energy storage devices, and are widely used in major industries. As the battery use increases, the maximum capacity of the battery will decay accordingly, and when the current remaining maximum capacity of the battery is reduced to 80% of the initial value, the battery is considered to reach the end of life [3]. Due to the many influencing factors of power battery life degradation, it is a big challenge to give an accurate assessment of the battery’s health status. Battery management system (BMS) can detect the actual usage of the battery, provide timely feedback, and issue early warnings to ensure the long-term safe and stable operation of the battery [4]. The state of health (SOH) of the battery is an important indicator in the BMS to characterize the aging degree of the battery. Therefore, it is of great significance to accurately predict the state of health of batteries.
Currently, scholars at home and abroad have conducted a lot of research on SOH estimation methods for power batteries [5], and the existing estimation methods can be categorized into three types: model-based methods, data-driven methods [6], and fusion-based methods [7]. Model-based human methods have been widely used to predict battery performance [8]. Hosseininasab et al. proposed a SOH estimation method based on a reduced-order electrochemical model for joint estimation of capacity and resistance [9]. Nejad et al. applied a dual extended Kalman filtering algorithm to update the parameters of a second-order equivalent circuit battery model so that an accurate estimation of the state of charge can be obtained in real time [10]. However, their computational complexity limits their applicability in real-time scenarios. Instead of relying on physical and chemical models inside the battery, data-driven methods utilize a large amount of experimental or real-world usage data as well as machine learning techniques to build statistical models. Typical data-driven methods include correlation vector machine (RVM) [11], support vector regression (SVR) [12], support vector machine (SVM) [13], and neural network (NN) [14]. Traditional RNNs can effectively utilize a priori information to solve nonlinear problems, but the performance of RNNs is limited by gradient explosion or gradient vanishing during training. Therefore, researchers have created some variants of RNN to solve these problems, such as Long Short-Term Memory (LSTM), Bidirectional LSTM (BiLSTM), GRU, etc. [15]. Traditional neural network models are not well suited to dealing with spatial patterns in data. Therefore, convolutional neural networks (CNNs) have become an effective solution to capture spatial patterns and short-term dependencies in time series data. For example, Chen et al. proposed a CV charging stage-based CNN for SOH estimation of lithium batteries, which extracts features from the early charging stage, avoids complex feature extraction engineering, and achieves SOH prediction by extracting features at the battery charging stage [16]. CNNs are good at recognizing spatial patterns and short-term temporal dependencies. Based on CNN, temporal convolutional networks (TCNs) are introduced to capture long-term dependencies while reducing computational complexity. TCNs lack the ability to model global features dynamically, and to further enhance the temporal modeling capability. Transformers have a unique advantage in capturing global dependencies through a self-attentive mechanism, and thus have been widely adopted [17]. Fusion–model-based approaches can overcome the limitations of a single model and better utilize the available information to improve the accuracy of the prediction [18].
Traditional single models (e.g., LSTM, CNN, etc.) perform well on specific tasks, but their capabilities may be limited by the characteristics of the model architecture. For example, LSTM is strong in capturing long-time dependencies but may underperform when dealing with short-time features [19]. Transformers excel in capturing global relationships but are deficient in dealing with sequence continuity (e.g., temporal dependencies). TCN excels in extracting short-term and localized features of a time series but lacks the ability to strongly capture global and nonlinear relationships of a sequence. Lithium battery SOH prediction involves multi-scale features (short-term changes and long-term trends), which are difficult to be adequately captured by a single model. In response to the above analysis, this paper proposes a prediction model based on TCN–Transformer–BiLSTM [20]. A neural network combining TCN, Transformer, and Bi-LSTM can integrate the advantages of these three models to handle complex time series data. This architecture aims to leverage the local time-dependent modeling of TCN, the global context modeling of Transformer, and the pre- and post-context-dependent processing capabilities of Bi-LSTM to improve the overall performance of the model [21].

2. Introduction to Algorithms

2.1. Temporal Convolutional Network

Temporal convolutional networks (TCNs) extract features from time series using causal convolution, dilated convolution, and residual connections. Causal convolution ensures temporal order, but its receptive field is limited by kernel size, requiring many stacked layers to capture long-term dependencies [22]. To address this, TCNs introduce dilated convolution, which inserts gaps between kernel elements and samples inputs at intervals. For example, at the bottom layer d = 1 every input point is sampled, while at higher layers d = 2, 4, … inputs are taken at wider intervals. This design allows the receptive field to grow exponentially with depth, enabling efficient modeling of both short- and long-range dependencies [23]. The results of dilated causal convolution in TCN are illustrated in Figure 1.
TCN consists of multiple hidden layers stacked on top of each other, each with a pair of hyperparameters (k, d), where k represents the size of the convolution kernel and d is the dilation factor. Each residual block consists of two branches. One branch contains two identical dilated causal convolutional layers, and the other one realizes the width alignment of the input and output by a 1 × 1 convolution and connects it directly to the output.

2.2. Sparrow’s Algorithm to Optimize BP Neural Networks

The Transformer was proposed to overcome the limitations of RNNs and LSTMs in handling long sequences, such as poor parallelization, difficulty in capturing long-range dependencies, and gradient vanishing or explosion. It addresses these issues through the self-attention mechanism, which allows the model to assign different weights to positions in the input sequence when generating outputs [24].
The Transformer architecture consists of an encoder and a decoder. The encoder transforms input data into an internal representation, while the decoder generates outputs based on this representation. Both components employ self-attention and feed-forward neural networks, as illustrated in Figure 2.
Since the Transformer is based on a parallel computation approach under the mechanism of multi-head self-attention, it does not have the ability to handle sequence ordering as classical recurrent neural networks such as RNN and LSTM inherently bring with them, and therefore introduces positional encoding to preserve positional information [25]. A common form of positional encoding is to use a combination of sine and cosine functions that can represent sequences of arbitrary length:
P E p o s , 2 i = sin p o s 10,000 2 i / d
P E p o s , 2 i + 1 = cos p o s 10,000 2 i / d
where pos is the position, i is the dimension index, and d is the model dimension. The position encoding will be added to the word vectors so that the representation of each word contains the position information.
The attention mechanism can be viewed as a means of resource allocation in the model learning process, and in deep neural networks, the weight parameter of each feature is a resource that the attention mechanism focuses on. The attention mechanism focuses the model on important information by adaptively assigning weights to input variables [26]. The structure of multi-head attention is shown in Figure 3.
The computation process of scaled dot-product attention is as follows: compute the product of query matrix, key matrix, and obtain the score matrix scores to represent the correlation between each query and all the keys; scale the score matrix to ensure the stability of the values; mask is equivalent to shielding the useless information; normalize the scaled scores by softmax function to normalize the scaled scores to obtain the attention weight distribution; use the obtained attention weights to weigh and sum the value vectors to obtain the weighted matrix. The formula is
A t t e n t i o n Q , K , V = s o f t m a x Q K T d k V
where d k is the number of columns of the Q, K matrix, the vector dimension, and d k is the scale factor contributing to the stabilization gradient.
The drawback of the self-attention mechanism is that the model will overly focus its attention on its own position when encoding the information of the current position, and the effective information grasping ability is a little worse. Therefore, it is proposed to solve this problem by a multi-head attention mechanism. Multi-head is performed by generating multiple sets of Q, K, V using attention to obtain multiple results, and then weighted averaging of the multiple results to obtain the final result [27]. The formula for multi-head self-attention is
M u l t i H e a d Q , K , V = C o n c a t h e a d 1 , h e a d h W O
where Q, K, V denote the query vector, key vector, and value vector, respectively; h denotes the number of heads; h e a d i denotes the output of the ith head; and W O is the output transformation matrix. The output of each head h e a d i can be expressed as
w h e r e   h e a d i = A t t e n t i o n Q W i Q , K W i K , V W i V
where W i Q , W i K , W i V are the query, key, and value transformation matrices of the ith header, respectively. Attention is the attention calculation function.

2.3. Bidirectional Long and Short-Term Memory Neural Network

Long Short-Term Memory is a variant of recurrent neural networks designed to solve the gradient vanishing and gradient explosion problems of traditional RNNs when processing long sequences. Unlike the simple structure of standard RNNs, LSTMs have a more complex design containing a cellular shape and three main gating mechanisms (input gate, forgetting gate, and output gate):
Cell state: This represents the long-term memory of the network and runs through the entire chain of LSTM cells with only a small number of linear interactions. It can be thought of as a conveyor belt to carry important information [28].
Forget gate: Decides what information should be discarded from the cell state.
Input gate: Decides what new information should be added to the cell state.
Output gate: Decides the output of the next hidden state based on the cell state.
The workflow of LSTM is shown below:
(1)
Calculation of the forgetting gate value f t :
f t = σ W f h t 1 , x t + b f
where σ is the sigmoid activation function, W f is the weight matrix, b f is the bias term; h t 1 the hidden state of the previous time step; x t is the input of the current time step.
(2)
Calculate the input gate value i_t and the candidate value C ˜ t :
i t = σ W i h t 1 , x t + b i
C ˜ t = t a n   h W C h t 1 , x t + b C
where W i and W C are weight matrices; b i and b C are bias terms; and t a n   h is the hyperbolic tangent activation function.
(3)
Renewal of cellular state C t :
C t = f t C t 1 + i t C ˜ t
(4)
Compute the output gate value o t and the new hidden state h t :
o t = σ W o h t 1 , x t + b o
h t = o t t a n   h C t
Its structure is shown in Figure 4.
LSTM processes sequence data in a unidirectional manner, i.e., it processes the input data sequentially from the beginning to the end of the sequence. This approach only captures forward contextual information. Bidirectional LSTM processes sequence data in a bidirectional manner; the single-layer Bi-LSTM is a combination of two independent LSTM layers, one going forward to process the input sequence, and the other processing the sequence in reverse, and the final output is a combination of the outputs of the forward and reverse LSTM layers, so the Bidirectional LSTM is able to capture both the forward and backward contextual information of the sequence [29]. Its working principle is shown in Figure 5.

2.4. Hybrid Modeling

A neural network combining TCN, Transformer, and Bi-LSTM can integrate the strengths of all three models to process complex temporal data. This architecture aims to leverage the local time-dependent modeling of TCN, the global context modeling of Transformer, and the pre- and post-context-dependent processing capabilities of Bi-LSTM to enhance the overall performance of the model. The model structure is shown in Figure 6, which is mainly composed of input layer, TCN layer, Transformer layer, BiLSTM layer, and output layer. The model structure is described as follows:
(1)
Input layer: Extract and filter features related to battery capacity, select highly correlated factors as indirect inputs, and preprocess the data. The training set is then fed into the model.
(2)
TCN layer: Processes the input sequence to extract local temporal features. By using dilated convolution, TCN captures short- and medium-term dependencies and provides refined temporal representations for the following layers.
(3)
Transformer layer: Takes TCN outputs as input and employs multi-head self-attention and feed-forward networks to capture global contextual dependencies, generating feature representations with long-range temporal information.
(4)
BiLSTM layer: Processes Transformer features in both forward and backward directions, capturing bidirectional dependencies in the sequence.
(5)
Output layer: The BiLSTM outputs are passed to a fully connected layer, which produces the final SOH prediction using the test data.
The model is based on MATLAB2023b implementation of the model and the model parameters are shown in Table 1.
TCN has two layers; the number of temporal convolutional layers in the first layer is 16, the size of the convolutional kernel is 3, and the expansion rate is 1. The number of convolutional kernels in the second layer is 32, the size of the convolutional kernel is 3, and the expansion rate is 2. There are two encoders for the Transformer, and the number of heads of each of the multi-head self-attention mechanisms is 4. The number of layers for the Bi-LSTM is 2, and the number of hidden units in the first layer is 16. The number of hidden units in the second layer is 32.

3. Dataset Validation

3.1. NASA Dataset Validation

The data is aligned by timestamp and denoised using median/SG filtering; the normalizer is fitted with the training set only for validation/testing; samples are constructed with a fixed sliding window. The feature pool contains cycle number, constant current capacity, constant voltage duration, temperature rise, pressure drop, dV/dt, dT/dt, etc. In this paper, we filtered under three dimensions: correlation (Pearson + Spearman), redundancy, and physical interpretability, and finally adopted HI1, HI3, and HI4 in order to obtain robust combinations that can be repeatedly collected in the field.
This study uses the battery dataset provided by NASA. The batteries used in the NASA dataset are 18,650 cylindrical Li-ion batteries with a capacity of 2 Ah and termination conditions of 70% of the rated capacity, i.e., 1.4 Ah. The batteries were divided into nine groups and were subjected to charging and discharging cyclic aging tests under random conditions as well as stationary conditions at 4 °C, 24 °C, and 43 °C. The batteries were tested at 4 °C, 24 °C, and 43 °C. The batteries were then subjected to the same conditions as those in the NASA dataset. In this paper, batteries 5, 6, and 7 were selected for prediction study, and the test conditions for this group of data were as follows:
Temperature: Room temperature 24 degrees.
Charging: Charge in constant current (CC) mode at 1.5 A until the battery voltage reaches 4.2 V, then charge in constant voltage (CV) mode until the charging current drops to 20 mA.
Discharge: Discharge in constant current (CC) mode at 2 A until the voltage of batteries 5, 6, and 7 drops to 2.7 V, 2.5 V, and 2.2 V.
Termination condition: Rated capacity drops to 1.4 Ah.
The cycle times of battery No. 5, 6, and 7 are all 168 times. Among them, batteries No. 5 and No. 6 reach below the failure threshold. The selected health features are six features: the number of cycles, the time of discharge to cut-off voltage, the time of iso-current drop, the time of iso-voltage rise, and the average temperature from the initial temperature to the highest point. For the convenience of description, these six features were replaced by HI1~HI6, respectively, and they were analyzed with SOH by Pearson correlation. The results of the analysis are shown in Table 2.
The three features with the highest correlation are selected as inputs to the model, i.e., number of cycles, iso-current drop time, and iso-pressure rise time, and SOH is used as the output of the model. The data from the battery was divided into training and test sets, using 60% of the data as the training set and 40% of the data as the test set.
In order to verify the accuracy of the algorithm, the prediction is obtained by comparing the mainstream algorithms of LSTM, GRU, and Bi-LSTM, and the LSTM, GRU, and Bi-LSTM, and the proposed models named M1, M2, M3, and M4, respectively. The predicted results are shown in Table 3 as well as in Figure 7.
As can be seen from the figure, the method proposed in this paper performs SOH prediction for different batteries, and the prediction results are compared with the test set and compared with the LSTM, GRU, and Bi-LSTM algorithms, all of which embody the most accurate prediction results. The root mean square error of the prediction results for three batteries is below 2%, the No. 7 batteries are controlled to be within 1%, the absolute coefficients are all greater than 0.96, the average absolute errors are all controlled to be within 1.5% or less, and battery No. 5 and No. 7 can be controlled to about 0.5%. By comparison, the method proposed in this paper has high accuracy and can accurately predict SOH for real vehicles.

3.2. Model Migration Capability Validation

In order to further verify the effectiveness of the algorithm, data are obtained through experiments for verification. The object of this experiment is a brand of 105 Ah lithium iron phosphate square shell battery; the experimental platform is mainly composed of test batteries, charge and discharge test equipment, constant temperature and humidity test chamber, connecting wires, temperature sensors, bolts, and so on. The experimental platform is shown in Figure 8.
The experimental conditions are extracted from the operating conditions of a real vehicle, which include driving with discharge, stopping and standing, braking feedback, and stopping and charging. The experimental conditions are in line with the daily vehicle driving status and driving behavior. The capacity change in one cycle of this condition is 10.5 Ah, and in the experiment, the operating condition is set to cycle eight times for one cycle, i.e., 84 Ah. After discharge, the standard charging method in the battery specification is adopted, i.e., constant current and constant voltage (CC-CV) charging method, whereby the battery is charged with a constant current of 0.5 C up to 3.65 V, and then charged with a constant voltage until the current is reduced to 0.05 C. At the end of each cycle, the battery is left to rest for a period of time in order to recover the internal capacity of the battery and to recover its internal resistance and reduce the temperature to avoid the cumulative thermal effect on the experimental results. Subsequently, the experiment continued with the next cycle.
The experimental capacity results are shown in Figure 9.
As can be seen from the figure, the battery capacity shows a capacity rebound, which is caused by the battery resting after cycling for a period of time. When the battery is resting for a period of time, the SEI membrane may self-repair or re-stabilize, resulting in a lower internal impedance and a recovery of the battery’s performance, which in turn leads to a capacity rebound.
Health features are similarly selected as inputs to the model before SOH prediction. Specifically selected features include the number of cycles, charging start voltage, charging maximum voltage, charging start current, charging median voltage, constant current charging capacity, constant voltage charging capacity, charging start average temperature, charging cut-off average temperature, charging process average maximum temperature, and charging time. These 11 features are named as HI1 to HI11, and the correlation analysis between the selected features and SOH is also carried out using the Pearson correlation coefficient method. The results of the analysis are shown in Figure 10.
Feature screening follows the three-dimensional criterion of ‘linear correlation + monotonicity + physical interpretability’: Pearson for linear initial screening, Spearman for monotonicity checking, and ultimately the combination of HI1/HI3/HI4 based on the electrochemical mechanism, with its limited capability of nonlinear inscription. Methods such as mutual information/HSIC will be used as an extension.
The calculation shows that the number of cycles, constant current charging capacity, and charging time have a higher correlation with SOH, so these three health features are selected as inputs to the model and SOH as outputs of the model for SOH prediction study. In order to verify whether the algorithm is able to make accurate predictions despite the small amount of data, we divided the dataset of the two batteries to different degrees and used 50%, 60%, and 70% of the training set for model training, respectively. The predicted results are shown in Figure 11.
The indicators for evaluating the predicted results are shown in Table 4.
Observing the prediction results, it can be seen that with the increase in the dataset used for model training, the prediction results are gradually closer to the real values, and the MAE and RMSE are also gradually reduced. And when only 50% of the dataset is used for model training, the RMSE of the test set can be controlled at 0.00118, which indicates that the proposed model has strong prediction ability.
In this paper, RMSE is calculated as normalized SOH (0−1); its conversion to actual capacity units is RMSEAh = RMSEnorm × (SOHmax − SOHmin) × Crated. The data division follows a chronological order: k cycles of isolation are set between the training/validation/testing segments; the normalizer is only fitted by the training set and used for validation/testing; all features and labels come from the same cycle and are not spliced across cycles to avoid information leakage.

3.3. Applicable Boundaries

The above three characterization schemes are valid in this data domain and do not guarantee the optimality for extreme temperature zones/anomalous working conditions; more complex scenarios can consider nonlinear indicators such as mutual information/HSIC and richer working condition characteristics.

4. Conclusions

This paper proposes a TCN–Transformer–BiLSTM based method for SOH estimation of lithium-ion batteries. Aiming at the limitations of existing single models in capturing complex temporal features, this paper proposes a hybrid model integrating Time Convolutional Network, Transformer, and Bidirectional Long and Short-Term Memory Network with the aim of improving the prediction accuracy through multi-scale feature extraction. The model architecture captures local temporal dependencies through TCN, models global contextual information using Transformer’s self-attention mechanism, and combines BiLSTM’s bidirectional sequence processing capability to achieve a comprehensive characterization of short-term fluctuations and long-term trends during battery degradation. Firstly, based on the NASA public dataset, the number of cycles, iso-current drop time, and isobaric rise time are selected as the key feature inputs, and the mainstream methods such as LSTM, GRU, and BiLSTM are compared. The results show that the root mean square error of the proposed models is lower than 2%, and the coefficient of determination (R2) is higher than 0.96, which performs better under the established baseline and the same evaluation protocols. Secondly, the migration capability and data adaptability of the model are further verified for the dynamic working condition data of a commercial lithium iron phosphate battery. The experiments show that even when only 50% of the training data is used, the model still maintains a prediction accuracy of RMSE below 0.12%, demonstrating strong potential for engineering. Under this data partitioning and model setup, the performance improves with increasing sample proportion, and the improvement tends to slow down at higher percentages; this observation is affected by both the data domain and the model capacity, and the conclusions are not generalized.

Author Contributions

Conceptualization, S.H.; data curation, X.L.; methodology, Z.S. and X.P.; software, Z.S.; validation, X.P. and S.H.; writing—original draft, S.H.; writing—review and editing, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Fundamental Research Program of Shanxi Province (No. 202303021211161); Scientific and Technologial Innovation Programs of Higher Education Institutions in Shanxi (2024L175); Fundamental Research Program of Shanxi Province (202403021222149).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BMSBattery Management System
SOHState of Health
RVMRelevance Vector Machine
SVRSupport Vector Regression
SVMSupport Vector Machine
NNNeural Network
RMSERoot Mean Squared Error
LSTMLong Short-Term Memory
TCNTemporal Convolutional Network

References

  1. Li, X.; Yu, Y.; Zhang, Z. External characteristics of lithium-ion power batteries based on electrochemical aging degradation model. J. Phys. 2022, 71, 345–353. [Google Scholar]
  2. Wang, B.; Li, S.; Yu, J. Research on cooling performance of thermal management system of cylindrical lithium-ion battery based on liquid cooling. J. Lanzhou Coll. Arts Sci. 2020, 34, 71–77. [Google Scholar] [CrossRef]
  3. Zhou, Y.; Shi, H. Real-vehicle data-oriented prediction of battery retirement trajectories for electric vehicles. J. Sol. Energy 2022, 43, 510–517. [Google Scholar]
  4. Chu, Y.; Chen, Y.; Mi, Y. A CNN-LSTM lithium battery health state estimation based on attention mechanism. Power Supply Technol. 2022, 46, 634–637+651. [Google Scholar]
  5. Xiong, Q.; Di, Z.; Ji, S. A review of research progress on health state estimation and life prediction of lithium-ion batteries. High Volt. Technol. 2024, 50, 1182–1195. [Google Scholar]
  6. Li, X.; Su, Z.; Ding, J. Data-driven power battery failure prediction algorithm for electric vehicles. J. North Cent. Univ. 2025, 46, 293–305. [Google Scholar]
  7. Wang, Z.; Wang, Q.; Liu, P. A review of big data-driven methods for power battery health state estimation. J. Mech. Eng. 2023, 59, 151–168. [Google Scholar]
  8. Hu, X.; Xu, L.; Lin, X.; Pecht, M. Battery Lifetime Prognostics. Joule 2020, 4, 310–346. [Google Scholar] [CrossRef]
  9. Seyedmehdi, H.; Changwei, L.; Stefan, P. State-of-health estimation of lithium-ion batteries for electrified vehicles using a reduced-order electrochemical model. J. Energy Storage 2022, 52, 104684. [Google Scholar]
  10. Shahab, N.; Thomas, D.G. Online Battery State of Power Prediction Using PRBS and Extended Kalman Filter. IEEE Trans. Ind. Electron. 2020, 67, 3747–3755. [Google Scholar]
  11. Meng, X. Research on Electric Vehicle BMS Based on Functional Safety and Correlation Vector Machine. Master’s Thesis, Guilin University of Electronic Science and Technology, Guilin, China, 2024. [Google Scholar]
  12. Li, X.; Yuan, C.; Wang, Z. State of health estimation for Li-ion battery via partial incremental capacity analysis based on support vector regression. Energy 2020, 203, 117852. [Google Scholar] [CrossRef]
  13. Xiao, P.; Zhen, Z.; Jie, W. An interval prediction approach based on fuzzy information granulation and linguistic description for remaining useful life of lithium-ion batteries. J. Power Sources 2022, 542, 231750. [Google Scholar] [CrossRef]
  14. Li, Y.; Wang, L.; Feng, Y. An online state-of-health estimation method for lithium-ion battery based on linear parameter-varying modeling framework. Energy 2024, 298, 131277. [Google Scholar] [CrossRef]
  15. Chen, J.; Yu, T.; Yuan, S. State of health prediction of lithium-ion batteries based on bidirectional gated recurrent unit and transformer. Energy 2023, 285, 129401. [Google Scholar] [CrossRef]
  16. Chen, S.; Liang, Z.; Yuan, H. A novel state of health estimation method for lithium-ion batteries based on constant-voltage charging partial data and convolutional neural network. Energy 2023, 283, 129103. [Google Scholar] [CrossRef]
  17. Chen, C.; Wu, Y.; Shi, J. A parallel weighted ADTC-Transformer framework with FUnet fusion and KAN for improved lithium-ion battery SOH prediction. Control. Eng. Pract. 2025, 159, 106302. [Google Scholar] [CrossRef]
  18. Song, K.; Hu, D.; Tong, Y. Remaining life prediction of lithium-ion batteries based on health management: A review. J. Energy Storage 2023, 57, 106193. [Google Scholar] [CrossRef]
  19. Han, S.; Li, C.; Ding, J.; Gao, X.; Li, X.; Zhang, Z. An Improved PSO-Based DC Discharge Heating Strategy for Lithium-Ion Batteries at Low Temperatures. Energies 2025, 18, 2261. [Google Scholar] [CrossRef]
  20. Han, S.; Wei, T.; Wang, L.; Li, X.; Chen, D.; Jia, Z.; Zhang, R. Study of Lithium-Ion Battery Charge State Estimation Based on BP Neural Network Fusion Optimized Sparrow Algorithm. Coatings 2025, 15, 697. [Google Scholar] [CrossRef]
  21. Song, J.; Jiao, J.; Liu, H. Effect of surface state of sic fibers on their interfacial properties. Compos. Commun. 2025, 53, 102232. [Google Scholar] [CrossRef]
  22. Peng, P.; Wan, M.; Zhang, L. RUL prediction of lithium-ion batteries based on multi-scale TCN. Battery 2024, 54, 649–654. [Google Scholar]
  23. Tofigh, M.; Kharazmi, A.; Smith, J.D. Temporal dilated convolution and nonlinear autoregressive network for predicting solid oxide fuel cell performance. Eng. Appl. Artif. Intell. 2024, 136, 108994. [Google Scholar] [CrossRef]
  24. Wang, L.; Zhang, W.; Li, W. DGAT: Dynamic Graph Attention-Transformer network for battery state of health multi-step prediction. Energy 2025, 330, 136876. [Google Scholar] [CrossRef]
  25. Chen, Y.; Li, D.; Huang, X. Exploring life warning solution of lithium-ion batteries in real-world scenarios: TCN-transformer fusion model for battery pack SOH estimation. Energy 2025, 335, 138053. [Google Scholar] [CrossRef]
  26. Chen, S.; Liu, J.; Yuan, H. AM-MFF: A multi-feature fusion framework based on attention mechanism for robust and interpretable lithium-ion battery state of health estimation. Appl. Energy 2025, 381, 125116. [Google Scholar] [CrossRef]
  27. Chen, F.; Shang, D.; Zhou, G. Collaborative multiple attention mechanisms for vehicle fault prediction. Eng. Appl. Artif. Intell. 2025, 160, 111896. [Google Scholar] [CrossRef]
  28. Liu, D.; Wang, S.; Li, X. A novel extended Kalman filter-guided long short-term memory algorithm for power lithium-ion battery state of charge estimation at multiple temperatures. Energy 2025, 335, 137973. [Google Scholar] [CrossRef]
  29. Sherkatghanad, Z.; Ghazanfari, A.; Makarenkov, V. A self-attention-based CNN-Bi-LSTM model for accurate state-of-charge estimation of lithium-ion batteries. Energy Storage 2024, 88, 111524. [Google Scholar] [CrossRef]
Figure 1. Expanded causal convolution.
Figure 1. Expanded causal convolution.
Coatings 15 01149 g001
Figure 2. Transformer structure.
Figure 2. Transformer structure.
Coatings 15 01149 g002
Figure 3. Structure of the multi-head attention mechanism.
Figure 3. Structure of the multi-head attention mechanism.
Coatings 15 01149 g003
Figure 4. LSTM structure.
Figure 4. LSTM structure.
Coatings 15 01149 g004
Figure 5. BiLSTM structure.
Figure 5. BiLSTM structure.
Coatings 15 01149 g005
Figure 6. Hybrid model structure.
Figure 6. Hybrid model structure.
Coatings 15 01149 g006
Figure 7. Projected results.
Figure 7. Projected results.
Coatings 15 01149 g007
Figure 8. Experimental platforms.
Figure 8. Experimental platforms.
Coatings 15 01149 g008
Figure 9. Capacity trends.
Figure 9. Capacity trends.
Coatings 15 01149 g009
Figure 10. Pearson features.
Figure 10. Pearson features.
Coatings 15 01149 g010
Figure 11. Projected results.
Figure 11. Projected results.
Coatings 15 01149 g011aCoatings 15 01149 g011b
Table 1. Model parameters.
Table 1. Model parameters.
ParametersValue
OptimizerAdam
Loss functionRMSE
Learning rate0.01
Maximum number of training rounds100
Minimum number of iterations10
Table 2. Features.
Table 2. Features.
Health Characteristics#5#6#7
HI1−0.987−0.981−0.988
HI20.8960.185−0.041
HI30.9850.9870.986
HI40.9930.9920.994
HI50.9320.9500.874
HI6−0.018−0.1490.293
Table 3. Projected results.
Table 3. Projected results.
Battery NumberMethodologiesRMSER2MAE
#5M10.05160.20710.0477
M20.03260.68400.0305
M30.03820.56490.0355
M40.01080.96500.0059
#6M10.05030.60320.0444
M20.02740.88210.0250
M30.04440.69020.0390
M40.01550.96230.0126
#7M10.03630.44310.0330
M20.01980.83460.0184
M30.02660.70070.0236
M40.00850.96920.0052
Table 4. Evaluation indicators.
Table 4. Evaluation indicators.
Projected Starting PointMAER2RMSE
50%0.001010.935290.00118
60%0.000880.934790.00095
70%0.000830.886890.00090
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, S.; Su, Z.; Peng, X.; Wang, L.; Li, X. Research on the SOH of Lithium Batteries Based on the TCN–Transformer–BiLSTM Hybrid Model. Coatings 2025, 15, 1149. https://doi.org/10.3390/coatings15101149

AMA Style

Han S, Su Z, Peng X, Wang L, Li X. Research on the SOH of Lithium Batteries Based on the TCN–Transformer–BiLSTM Hybrid Model. Coatings. 2025; 15(10):1149. https://doi.org/10.3390/coatings15101149

Chicago/Turabian Style

Han, Shaojian, Zhenyang Su, Xingyuan Peng, Liyong Wang, and Xiaojie Li. 2025. "Research on the SOH of Lithium Batteries Based on the TCN–Transformer–BiLSTM Hybrid Model" Coatings 15, no. 10: 1149. https://doi.org/10.3390/coatings15101149

APA Style

Han, S., Su, Z., Peng, X., Wang, L., & Li, X. (2025). Research on the SOH of Lithium Batteries Based on the TCN–Transformer–BiLSTM Hybrid Model. Coatings, 15(10), 1149. https://doi.org/10.3390/coatings15101149

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop