Next Article in Journal
Research on Fuel Economy of Hydro-Mechanical Continuously Variable Transmission Rotary-Tilling Tractor
Previous Article in Journal
Receiver-Side Topologies for Wireless Power Transfer Systems: A Comprehensive Review of the Design, Challenges, and Future Trends
Previous Article in Special Issue
Deep Learning for State of Health Estimation of Lithium-Ion Batteries in Electric Vehicles: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Hybrid Deep Learning Model for Battery State of Health Estimation Using Transfer Learning

1
Department of Automotive Engineering, Shandong Vocational College of Science and Technology, Weifang 261053, China
2
Hubei Longzhong Laboratory, Hubei University of Arts and Science, Xiangyang 441000, China
3
Hubei Key Laboratory of Power System Design and Test for Electrical Vehicle, Hubei University of Arts and Science, Xiangyang 441053, China
*
Author to whom correspondence should be addressed.
Energies 2025, 18(6), 1491; https://doi.org/10.3390/en18061491
Submission received: 23 February 2025 / Revised: 12 March 2025 / Accepted: 14 March 2025 / Published: 18 March 2025

Abstract

:
Achieving accurate battery state of health (SOH) estimation is crucial, but existing methods still face many challenges in terms of data quality, computational efficiency, and cross-scenario generalization capabilities. This study proposes a hybrid deep learning framework incorporating transfer learning to address these challenges. The framework integrates inception depthwise convolution (IDC), channel reduction attention (CRA) mechanism, and staged training strategy to improve the accuracy and generalization ability of SOH estimation. The IDC module of the proposed model is capable of extracting battery degradation time series features from multiple scales while reducing the computational overhead. The CRA module effectively reduces the computational complexity and memory usage of global feature capture by compressing the channel dimensions. A well-designed pre-training/fine-tuning two-stage training strategy achieves accurate cross-scene SOH estimation by utilizing large-scale source-domain data to learn generalized aging features and then uses a small amount of new data to quickly fine-tune the base model. The proposed method is validated using two publicly available datasets, including 54 nickel cobalt manganese oxide (NCM) cells and 16 nickel manganese cobalt oxide (NMC) cells. The experimental results show that the root mean square error (RMSE) of the model on the NCM and NMC datasets is 0.522% and 0.283%, respectively, with a coefficient of determination (R2) not less than 0.98 and mean absolute percentage error (MAPE) of 0.431% and 0.22%, respectively. The proposed method not only achieves high-precision SOH estimation among the same type of batteries but also demonstrates strong generalization ability under different battery chemistries and scenarios.

1. Introduction

With the rapid development of the global economy, population growth, and the acceleration of industrialization, the global demand for natural resources has increased significantly [1,2]. To cope with these problems, lithium-ion batteries, as an efficient and environmentally friendly energy storage technology, play a key role in promoting energy transition and realizing sustainable development. With their excellent performances, such as high energy density, long cycle life, and low self-discharge rate [3,4,5], Li-ion batteries are widely used in electric transportation, mobile devices, distributed energy storage, and industrial manufacturing. However, with the charging and discharging cycles, batteries inevitably experience aging phenomena, leading to a gradual decline in their SOH, which directly affects the range, power performance, and energy recovery efficiency of electric vehicles [6,7]. Therefore, how to accurately assess the SOH of lithium-ion batteries has become a key issue to improve battery performance and extend the service life. In the field of SOH estimation for Li-ion batteries, traditional methods mainly rely on simple analysis of battery data, physical models, or means based on basic feature extraction. For example, the open-circuit voltage (OCV) method [8], the internal resistance method [9], and the capacity degradation method [10] are widely used to determine the aging degree of the battery, while the polarization voltage method [11] and the charge retention method [12,13] are used to detect the reaction degree of the active substance inside the battery and its charge retention ability. However, these traditional methods show obvious limitations in the face of data complexity and nonlinear characteristics, which make it difficult to meet the demand for high accuracy and real-time performance of modern battery management systems.
In recent years, due to the rapid development of big data and high-throughput computing technologies, data-driven methods have been widely used in the estimation of battery SOH and have demonstrated great potential for application. Among them, by utilizing large-scale historical data, machine learning methods are able to build efficient and flexible prediction models. In addition, these methods not only reduce human intervention but also show remarkable performance in dealing with complex patterns and nonlinear relationships. For example, Bin Gou et al. [14] proposed a hybrid integration method combining nonlinear autoregression (NAR) and bootstrap resampling method (Bootstrap) uncertainty management, which not only improves the accuracy of SOH estimation but also possesses the ability of online application. L Cai et al. [15] used the non-dominated sorting genetic algorithm II (NSGA-II) to optimize support vector regression (SVR) and current pulse test features to establish an efficient SOH estimator. J Qiao et al. [16] proposed the chaotic firefly–particle filtering (CF-PF) method, which combines particle filtering (PF) and extended Kalman filtering (EKF) to realize the joint estimation and bidirectional correction of SOC and SOH and dynamically adjusts the parameters by a novel battery migration model to significantly improve the estimation accuracy. T Xu et al. [17] proposed a hybrid driving model that combines electrochemical impedance spectra with integrated circuit (IC) curve features, captures the local timing and global decay trends with edited nearest neighbor (ENN) and support vector regression (SVR), respectively, and fuses the optimization through extreme learning machine (ELM) to achieve fast and accurate battery capacity estimation. K McCarthy et al. [18] investigated the different SOH correlations between equivalent circuit elements and impedance spectra of lithium-ion batteries and used curve fitting and Pearson correlation matrix analysis to identify impedance variables that are highly correlated with the battery state, which provides a key basis for impedance state modeling of real-time battery management systems. However, although these methods exhibit high accuracy in battery SOH estimation, certain limitations still exist. For example, some of the methods are not adaptable enough to deal with diverse battery materials, complex dynamic working conditions, or cross-scenario applications. In addition, the high computational complexity of some algorithms makes it difficult to meet the real-time requirements, especially in practical engineering scenarios such as battery management systems, which require high response speed and resource efficiency, and this poses a challenge to further enhance their generalizability and practical deployment capabilities.
Deep learning methods are capable of handling complex pattern recognition tasks with strong nonlinear modeling capabilities by automatically learning features and training on large-scale data. Transfer learning (TL), on the other hand, improves model performance in the presence of insufficient data by transferring knowledge learned from one task to another related task, thereby reducing training time and dependence on large-scale datasets. J Zhao et al. [19,20] used the Transformer model to enhance the ability to recognize both short-term and long-term patterns in time series data and achieved SOH by TL for real-time estimation with reduced computational cost and enhanced model adaptability. Y Ma et al. [21] combined TL with a hybrid deep belief network (DBN)-LSTM model to improve the reliability of SOH estimation through the fusion of multi-health metrics. G Ma [22] proposed a TL method for battery SOH estimation, which effectively solves the challenge of battery SOH estimation under different usage conditions by combining a convolutional neural network (CNN) and an improved maximum mean difference domain adaptive method, and the experimental results validate the superiority and generalization of the method. Y Yang [23] proposed a joint state of charge (SOC) and SOH estimation model combining deep learning and TL and achieved a dynamic SOH estimation model by taking into account the distributional differences of the training dataset, spatial characteristics, and cross-time scale characteristics; accurate and robust estimation of SOC and SOH for the whole life cycle of batteries under dynamic operating conditions was achieved. H Zhao et al. [24] further improved the consistency and performance of the model in SOH estimation by integrating multiple attention mechanisms and deep learning methods. X Shu [25] combined the long- and short-term memory neural network (LSTM) network with TL to construct the mean model and difference model, which achieved a highly accurate prediction of battery SOH with the error controlled within 3% under the significantly reduced amount of training data, effectively reducing the computational cost. S Shen [26] proposed the deep convolutional neural network-integrated TL (DCNN-ETL) approach, which combines TL and integrated learning and significantly improves the accuracy and robustness of lithium-ion battery capacity estimation under limited training data.
Hybrid models can improve prediction accuracy and enhance robustness by combining the advantages of multiple models or distributions. J Zhao [27] proposed a hybrid fusion model combining convolutional neural network (CNN) and self-attention mechanism, which utilized TL to effectively transfer the aging knowledge from LFP batteries to ternary batteries (NMC and NCA) under multiple battery materials and operating conditions to achieve low error and high diagnostic accuracy, and verified its effectiveness in battery health condition assessment. L Ren et al. [28] utilized an improved CNN combined with an LSTM network to enhance the data dimensionality through an autoencoder, supplemented with a filter to smooth the prediction results, and verified its superiority in a real dataset. Z Lv et al. [29] proposed a hybrid fusion network model based on resource-efficient artificial intelligence that reduces computational resource consumption while still maintaining highly accurate SOH estimation. Y Dai et al. [30] further improved the LSTM network by combining the CNN-LSTM with the gate-controlled recurrent unit (GRU) method, which significantly improved the SOH estimation accuracy. C Qian [31] proposed a CNN-SAM-LSTM model to improve the estimation accuracy by integrating the coupling information of SOH and, at the same time, solve the imbalance problem of multi-state parameter learning by optimizing the loss function. C Jia et al. [32] proposed a model combining the bi-directional gated recurrent unit (Bi-GRU) and Transformer in a hybrid prediction model, where they input indirect health indicators (HI) characterizing the degradation of lithium-ion batteries into Bi-GRU to learn the hidden states and extract time series features.
Although these methods show high accuracy and flexibility in battery SOH estimation, and especially the deep learning models demonstrate significant advantages in handling complex nonlinear relationships and time series features, there are still some pressing issues that need to be addressed. Specifically, these methods have high requirements on the quality and diversity of the dataset, especially in TL, where the difference in data distribution between the source and target tasks will directly affect the transfer effect of the model. On the other hand, hybrid models significantly improve prediction accuracy by combining the advantages of multiple models or algorithms, but their model structure is usually more complex, with higher computational overhead, which makes it difficult to meet the real-time requirements. In addition, most of the methods mainly model a single task or specific conditions, and the generalization ability of the models still faces certain challenges in cross-material and multi-task scenarios. Therefore, in this study, we propose a deep learning-based framework for battery SOH estimation to address the challenges of existing methods in terms of data quality, computational efficiency, and cross-scene generalization capability by combining inception depthwise convolution (IDC) [33], channel reduction attention (CRA) [34] mechanism, and TL techniques (Figure 1). Specifically, our approach innovates in the following three aspects. First, to address the high computational complexity of traditional convolutional operations, we introduce IDC, a module that significantly reduces the computational overhead through a multi-branch design and a depth-separable convolution technique while extracting time series features related to the battery health state from multiple scales and directions. Second, in order to efficiently capture global features in the battery SOH estimation task, this study employs an improved CRA mechanism. This mechanism significantly reduces the computational complexity and memory footprint of the self-attention mechanism by compressing the channel dimensions prior to the attention computation while preserving the key features of the global information. Finally, to enhance the model’s generalization ability in new scenarios, a staged training strategy is devised. During the pre-training stage, the model learns general battery aging features from large-scale source domain data. In the fine-tuning stage, the model preserves its general feature extraction ability by freezing the weight parameters of the IDC module. Meanwhile, it adjusts the weights of the CRA module to adapt to the feature distribution of the new data. This strategy not only shortens the training time in new scenarios but also substantially improves the model’s effectiveness in cross-task scenarios.

2. Dataset Overview and Feature Engineering

2.1. Sources and Description of Data

In this study, two publicly available lithium-ion battery aging datasets are used to evaluate the performance of the proposed model, namely lithium nickel/cobalt/manganese (NCM) batteries [35] and nickel/manganese/cobalt (NMC) batteries [36]. Table 1 lists the basic specifications of these two batteries, including temperature, charge/discharge protocols, and nominal specifications. Figure 2 demonstrates the degradation trends and charge/discharge characteristics of the battery datasets. There are significant differences between these two datasets in terms of operating temperature, charging and discharging protocols, electrochemical composition, and capacity degradation rate. In particular, the NCM battery dataset covers the aging data of 54 batteries with a nominal voltage of 3.6 V and a nominal capacity of 3.5 Ah. These batteries were tested by charge/discharge cycling at three different temperatures, namely, 25 °C, 35 °C, and 45 °C. The battery dataset was tested with a nominal voltage of 3.6 V and a nominal capacity of 3.5 Ah. All batteries were subjected to charge/discharge cycling tests at a charge rate of 0.5 C and a discharge rate of 1 C until the effective capacity of the batteries decreased to 80% of the original nominal capacity. The current rate was calculated based on the nominal capacity of the battery, i.e., 1 C = 3.5 A. The NMC battery dataset consists of 16 commercial Li-ions with a nominal voltage of 3.6 V and a nominal capacity of 3.0 Ah. The experiments were conducted at three ambient temperatures, including 15 °C, 25 °C, and 35 °C, and were designed to simulate the battery’s performance when operated at low, ambient, and high temperatures. All batteries were discharged using the same charging protocol, but the discharge process followed four different discharge protocols. Specifically, the battery testing experiments involved two stages: charging and discharging. During the charging phase, the batteries were first charged to 4.2 V at a current rate of 0.5 C using the constant-current (CC) mode. Subsequently, the charging mode was switched to constant-voltage (CV) charging, and this process continued until the current decreased to 0.05 C. For the discharging phase, the batteries were discharged to 2.5 V at various temperatures. The discharging was carried out in the CC mode with current rates of 0.5 C, 1 C, 2 C, or 3 C. In this study, data from 54 NCM batteries were employed to train the base model. The purpose was to assess the model’s performance under identical electrochemical compositions but with different operating conditions and degradation rates. On the other hand, the NMC cell data were utilized to verify the efficacy of the proposed method. This method aims to precisely estimate the SOH of cells with different chemical compositions.

2.2. Feature Engineering

SOH serves as a crucial metric for gauging a battery’s current health status. It quantifies the battery’s actual performance relative to its initial capabilities. Typically presented as a percentage, SOH represents the extent to which the battery’s capacity and other performance parameters have deteriorated. As the battery is used over time and undergoes multiple charge/discharge cycles, its capacity steadily diminishes, causing the SOH to decline accordingly. The calculation of SOH can be expressed by the following formula:
SOH = C a c t u a l C n o m i n a l × 100 %
where C a c t u a l is the current effective capacity (actual capacity) of the battery, usually obtained by discharge test; C n o m i n a l is the nominal capacity of the battery, i.e., the rated capacity of the battery at the factory.
In the research of battery SOH estimation, data preprocessing is a critical step to ensure model performance and prediction accuracy. It improves the quality and consistency of the data by cleaning, transforming, and organizing the raw data, thus laying a solid foundation for subsequent analysis and modeling. Our preprocessing steps include data cleaning, interpolation, feature selection, sliding window technique, normalization, and battery dataset partitioning. Specifically, data cleaning is used to deal with missing values and remove noise. Interpolation methods help fill in the gaps in the data, and feature selection is dedicated to extracting the most representative variables. Sliding window techniques are suitable for dealing with time series data, normalization or standardization ensures that different features are on the same scale, and dataset partitioning provides the necessary foundation for model training and validation. These preprocessing steps can effectively reduce errors and biases in the data and prevent the model from being affected by bad data during training. Through data preprocessing, not only can the accuracy and stability of the model be improved, but also the training speed can be accelerated and the use of resources can be optimized. In addition, good data preprocessing can help reveal potential patterns and relationships in the data, enhance the understanding of the data, and provide a more reliable basis for decision-making.
In this investigation, charging data from 80% SOC to 4.2 V cutoff voltage were selected as input to ensure the validity and representativeness of the data. Specifically, this data range can comprehensively reflect the battery’s voltage, current, and temperature variations in the high SOC region, capturing its performance fluctuations and health status under common operating conditions. However, the raw data may suffer from discontinuities and missing key values, which increase the difficulty of model processing. To address this issue, interpolation methods can be employed to fill in missing values or generate higher-resolution, smoothed sampling points, thereby enhancing the completeness and refinement of the data. For this purpose, we use the linear interpolation method to resample the data of each cycle to 100 equally spaced points so as to unify the data length and eliminate the effect of uneven data distribution on feature extraction. The formula for linear interpolation is as follows:
y = y 0 + ( x x 0 ) ( x 1 x 0 ) · ( y 1 y 0 )
where ( x 0 , y 0 ) and ( x 1 , y 1 ) are the two known data points; x are the horizontal coordinates of the difference point; and y are the vertical coordinates estimated at the interpolation point x .
The sliding window technique captures changes in the data and avoids information loss by fixing the size of the sliding window in the time series data and gradually extracting local features and temporal relationships. By extracting a subset of the data, the method is able to focus on specific time periods or data intervals, improving the relevance and effectiveness of the data. In addition, the sliding window technique helps to improve the estimation accuracy and generalization ability of the model, reduces the risk of overfitting, and balances computational efficiency and accuracy by flexibly adjusting the window size and step size. This technique not only reduces the amount of computation of the model on the data but also effectively extracts key features, adapts to dynamic data of different scales, and becomes an effective tool for processing time series data. Charge voltage directly affects energy input and internal chemical reaction rates, ensuring efficient energy conversion and storage and avoiding over or under-charging, thus extending battery life; monitoring voltage changes helps to accurately assess battery condition. Charge capacity measures the energy storage capacity, reflecting the remaining life and overall health, which decreases with the number of uses and charge/discharge cycles, revealing the degree of battery aging and helping to predict the rate of decline. Charging current determines charging speed and internal heat generation. Reasonable control of charging current helps to balance charging speed and battery life, improving the accuracy of SOH estimation. Temperature is a critical environmental factor that significantly influences battery performance. Excessively high temperatures accelerate chemical reactions and promote electrolyte decomposition, leading to capacity degradation and increased internal impedance. Conversely, excessively low temperatures slow down electrochemical reactions, reduce charge/discharge efficiency, and markedly diminish capacity. Real-time temperature monitoring helps to optimize the charging strategy, safeguard the stability and safety of the battery, and improve the reliability of SOH estimation. Therefore, we choose charging voltage, charging capacity, charging current and temperature as features in the feature extraction process. At the same time, we adopt the sliding window technique to slice the data and capture the local dynamic characteristics of the time series data. Specifically, a starting point is randomly selected in each battery cycle, and 6 windows are sequentially intercepted with a step size of 2. Each window consists of 50 data points, encompassing key parameters such as voltage, charge capacity, current, and temperature. In order to solve the influence of different feature magnitude differences on model training and to prevent certain features from dominating model training due to large values, the maximum/minimum normalization method is employed to map the feature values of voltage, charging capacity, current, and temperature into the [0, 1] interval. This normalization process brings multiple benefits. Firstly, it speeds up the convergence of the optimization algorithm, thereby enhancing the stability and accuracy of the model. Secondly, it mitigates data bias, strengthens the model’s generalization ability, and lessens the likelihood of overfitting. Moreover, normalization enables a more equitable comparison among different features. This, in turn, boosts the overall performance and prediction effectiveness of the machine learning algorithm. After normalization, the processed features are concatenated to form a multidimensional feature matrix. This matrix, with strong representational power, serves as the output of the data-processing stage. The formula for this normalization process is presented below:
x n o r m = x x l o w x u p p x l o w
where x is the original data; x l o w is the minimum value of the data; x u p p is the maximum value of the data; x n o r m is the normalized data. In order to avoid performance evaluation bias triggered by uneven data distribution or ordering problems and to ensure the adequacy and generalization ability of the model training data, we divided the NCM dataset into three subsets. As follows, 40 cells were selected as the training set for training our IDC-CRA model. Four cells were used as the validation set for further optimizing the hyperparameters of the model. The remaining 10 cells were assigned as the test set for fully evaluating the performance of the model.

3. Methodology

To achieve efficient battery health state estimation, the hybrid deep learning model proposed in this paper is constructed based on the IDC and CRA and optimized by combining the staged training strategy. As shown in Figure 3, the model architecture consists of three core modules. The IDC module is responsible for multi-scale spatiotemporal feature extraction, while the CRA module effectively captures global information. Additionally, TL enhances the model’s cross-domain generalization capability through a two-stage training process involving pre-training and fine-tuning. The hyper-parameter configuration of the model is shown in Table 2, where a higher initial learning rate is used in the pre-training phase to accelerate model convergence, and a lower learning rate is used in the fine-tuning phase, combined with an early stopping mechanism to prevent overfitting. The synergistic design of the architecture enables the model to efficiently process large-scale timing data while adapting to the feature distribution of different battery materials. By combining the optimized parameters in Table 2 with the modular structure shown in Figure 3, the model achieves a balance between computational efficiency and generalization capability while ensuring accuracy.

3.1. Inception Depthwise Convolution

In deep learning models for battery SOH estimation, the efficiency and accuracy of feature extraction is one of the key factors to ensure the performance of the model. CNNs are widely used in various tasks due to their excellent performance in image processing. The traditional convolutional operation extracts local features by performing a sliding window convolution on the input feature map. However, this operation suffers from high computational complexity and number of parameters, especially when dealing with high-dimensional features. In order to improve the computational efficiency of the model and enhance the feature extraction capability at the same time, this study employs the inception deeply differentiable convolution approach. This module combines the multi-branching design of the inception architecture with the depthwise separable convolution (DSC) technique, aiming to efficiently extract the time series features related to the battery health state and to maintain efficient performance with limited computational resources. DSC significantly reduces the computational effort by decomposing the traditional convolution operation into two steps: depth convolution and point-by-point convolution. When compared to standard convolution, depthwise separable convolution remarkably decreases the computational complexity of each channel. For a convolution kernel with a size of k × k, here is how the computational effort of depthwise separable convolution is characterized: The IDC module takes a four-dimensional tensor as its input. This tensor is defined as X B × C × H × W , where B represents the batch size, C indicates the number of input channels, and H and W, respectively, signify the height and width of the feature map. Specifically, the input tensor will be partitioned into four segments based on the number of channels. This partitioning can be defined by the following equation:
X = Split ( X , split   indexes )
Here, the split operation determines the number of channels per branch by setting the ratio r g , defined as:
g = r g × C
Here, g represents the number of channels within the branch, r g is the scaling factor for the number of channels in the branch, and C stands for the number of channels in the original input. The segmented input X h w , X w , X h , X i d undergoes processing via different convolutional branches. Horizontal and vertical convolution is realized by using depth-divisible convolution with a convolution kernel of size k × k. For each branch, the depth convolution convolves the input channels independently and then fuses the information between the channels by point-by-point convolution. For each branch, the output is processed through deeply differentiable convolution to obtain a new feature map X h w , X w , X h . The computation of these branches can be expressed as:
X h w = DWConv g g × k ( X h w ) X w = DWConv g g × k ( X w ) X h = DWConv g g × k ( X h )
where k is the size of the convolution kernel, usually defaulted to k = 3. After all branches are convolved, the final output tensor is obtained by splicing together the outputs of each convolution branch, defined as:
X = Concat ( X h w , X w , X h , X i d )
This operation splices along the channel dimensions and fuses features from different convolutional branches together, which enhances the expressive power of the model and helps it capture important features of battery capacity degradation from multiple scales and directions. In addition, depth-divisible convolution significantly reduces the computational effort by decomposing the traditional convolution operation into two steps: depth convolution and point-by-point convolution. Compared with standard convolution, depth-divisible convolution significantly reduces the computational complexity of each channel. In the case of convolution kernel size k × k, the computation of depth divisible convolution is:
FLOPs depthwise = C i n × k × k × H × W
while the computational effort for point-by-point convolution is:
FLOPs pointwise = C i n × C o u t × H × W
This computationally reduced design is particularly important in the battery SOH estimation task to efficiently process large amounts of time series data. The IDC module is designed to efficiently extract multiple features from the battery data while maintaining a low computational overhead through multiple parallel convolutional branches and deeply differentiable convolution. The module employs a splicing operation to fuse features from different branches, which enhances the expressive power of the model in battery health state estimation.

3.2. Channel Reduction Attention

In the battery SOH estimation task, accurately extracting key features of the battery performance is crucial for predicting the SOH of the battery. Among them, the computational complexity and efficiency of the model are the key factors affecting the performance and training speed. In order to optimize the computational efficiency of the battery SOH prediction model and to improve the global feature capture capability, this study employs CRA. The CRA module improves on the traditional self-attention mechanism by performing channel compression through average pooling before performing attention operations on the query (Q) and key (K) matrices. With this channel reduction operation, the computational overhead can be significantly reduced, and the global feature information can be effectively captured. Specifically, the input feature map F i C × H × W is processed into multiple heads, and the Q i , K i , and V i of each head are computed by the projection matrix, respectively, with Eq:
CRA ( F i ) = Concat ( Head 0 , , H e a d j ) W i O
Q i = F i W j Q , K i = AvgPool ( F i ) W j K , V i = AvgPool ( F i ) W j V
where W j Q C × 1 , W j K C × 1 , W j V C × 1 is the projection parameter used for the Q i , K i , and V i matrices and W O C × C is the output projection matrix. Then, the attention weight of each head is calculated, defined as:
Head j = Attn ( Q i , K i , V i )
Attn ( Q i , K i , V i ) = Softmax ( Q i K i T ) V i
The CRA module significantly reduces the computational complexity in the self-attention mechanism by reducing the channel dimensions of the query and key matrices. Compared with the traditional self-attention mechanism, CRA makes the matrix multiplication operation of each computation more efficient by compressing the number of channels, which greatly reduces the memory occupation and computation. With this design, CRA is able to efficiently capture global information while reducing unnecessary computational burden, which is especially important for the battery SOH estimation task that needs to deal with large-scale datasets.

3.3. Hyperparameter Interpretation

In terms of better understanding our model, we explain it for some of the hyperparameters. In this study, we carefully configure key parameters to optimize the model’s performance in predicting battery SOH. To ensure stable gradient updates while maintaining computational efficiency, we set the batch size to 256, which effectively reduces variance in gradient updates, leading to smoother convergence and improved generalization. Given the complexity of battery aging modeling, this setting provides a balance between computational resource utilization and prediction accuracy. The model undergoes an extensive pre-training phase of 10,000 epochs, allowing it to deeply extract key electrochemical aging features from the NCM battery dataset. To prevent overfitting, an early stopping mechanism with a patience value of 1200 is implemented, terminating training when no improvement in validation loss is observed over 1200 consecutive epochs. This ensures that the model fully learns the degradation patterns from the source dataset without unnecessary overfitting, which could otherwise hinder its generalization ability. Additionally, we enhance information integration through feature concatenation, where both the original feature representations and their differences are combined after extraction through the IDC module. This strategy enriches the input space, enabling the model to learn both absolute values and relative variations of battery characteristics. By capturing dynamic changes within charge-discharge cycles, this approach improves the model’s ability to recognize battery aging trends and enhances predictive accuracy.
To further optimize feature extraction, the input data consists of four fundamental channels—voltage, current, charge capacity, and temperature—each crucial to understanding lithium-ion battery degradation. These features are initially processed through the IDC module and then transformed via an adjust_channels layer, expanding the feature dimension from 4 to 64 channels. This transformation enhances feature aggregation within the CRA module, strengthening the model’s ability to capture long-term temporal dependencies and optimize SOH prediction accuracy. Within the IDC module, we employ depthwise convolution to perform multi-scale spatiotemporal feature extraction. Specifically, square depthwise convolution (3 × 3) enhances local spatial feature modeling, horizontal depthwise convolution (1 × 11) captures long-range dependencies along the temporal dimension, and vertical depthwise convolution (11 × 1) facilitates interactions between different input features. This design significantly reduces computational complexity while preserving detailed spatial representations. Unlike standard convolution, which applies shared filters across all input channels, depthwise convolution processes each channel independently, improving efficiency and fine-grained feature extraction. Following this, a 1 × 1 pointwise convolution is applied in the adjust_channels layer, expanding the feature dimension to 64 channels and facilitating multi-scale feature fusion. This step strengthens the model’s global attention mechanism within the CRA module, ensuring a refined representation of extracted features and ultimately improving the accuracy of SOH prediction.

3.4. Training Strategies and Transfer Learning

The model training in this study uses a standard supervised learning approach, aiming to improve the accuracy and generalization ability of the battery SOH estimation task through efficient training strategies and optimization methods. During the training process, multiple strategies are combined to ensure the stability and efficiency of the model. To measure the difference between the model prediction results and the true values, the mean square error was used as the loss function in this study. The optimization of the loss function is performed by the Adam optimizer [37], which combines the advantages of the momentum approach and the adaptive learning rate to efficiently handle high-dimensional data and accelerate the convergence of the model during the training process. On the other hand, in order to enhance the stability of the model, we also introduce weight decay and gradient shearing strategies to prevent overfitting and ensure the stability of the gradient during the training process. Additionally, early stopping is utilized to avoid overfitting, halting training when the validation performance ceases to improve over several epochs and retaining the best model weights. The learning rate of the model was adjusted in different training stages. A higher learning rate was used in the pre-training phase to accelerate the initial convergence of the model. In the fine-tuning phase, the learning rate was decremented to more finely tune the model parameters, thus ensuring the stability and adaptability of the model on new data.
To improve the generalization ability of the model in battery SOH estimation, this study employs the technique of TL. TL enables the model to utilize prior knowledge gained from a large amount of raw data so that it can adapt to new scenarios without starting from scratch. Specifically, in the fine-tuning phase, the model fine-tunes the weights of some layers using a small amount of new data, which may include features and scenarios that have never been encountered before. In the well-designed fine-tuning strategy, the IDC module is effective in capturing generic battery aging features, and thus, the weight parameters of the module are frozen to preserve the generic electrochemical pattern recognition capabilities obtained from the source domain data. The CRA module, which plays a key role in capturing global information and trends, is then weighted to adapt to the new data features. This fine-tuning strategy improves the model’s predictive accuracy and robustness to new features while retaining its essential generalized capabilities. In conclusion, by rapidly training the pre-trained model using new data, the proposed model can be effectively adapted to the battery SOH estimation task in new scenarios.

4. Results and Discussion

4.1. Model Performance on NCM Cells

In the realm of data analysis and machine learning, assessing a model’s performance is of utmost importance. This is because such an evaluation serves a dual purpose. On the one hand, it checks the model’s accuracy, making sure that it can accurately represent the patterns and relationships present in the data. On the other hand, it assesses the model’s generalization capacity, guaranteeing that it can deliver consistent results when faced with new, unencountered data. In addition, performance evaluation helps to detect overfitting or underfitting problems of the model and guides the adjustment and optimization of the model so that the optimal model can be selected and its parameters optimized. By quantifying the performance of the model, the evaluation results provide a reliable basis for business decisions and ensure that the selected model meets the actual needs and goals. At the same time, continuous performance monitoring can identify and respond to potential problems in a timely manner, ensuring the long-term effectiveness of the model in practical applications. Therefore, in order to better evaluate the accuracy and generalization ability of the proposed model performance, we introduce three evaluation indexes: root mean square error (RMSE), coefficient of determination (R²), and mean absolute percentage error (MAPE). RMSE measures the absolute size of the prediction error, and the smaller the value means the higher the accuracy of the model; R² reflects the proportion of the model explaining the variation of the data, and the closer the value is to 1, the higher the explanation ability is; MAPE indicates the relative percentage of prediction error, the smaller the value indicates the higher the prediction accuracy. The formulas for the relevant indicators are shown below:
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
M A P E = 100 % n i = 1 n | y i y ^ i y i |
where n is the sample size, y i is the i-th actual, y ^ i is the i-th forecast, and y ¯ i is the average of the actuals. Based on these evaluation metrics, a comprehensive regression evaluation of the proposed IDC-CRA hybrid model is conducted with the aim of verifying the accuracy and robustness of the model in the estimation of SOH of NCM batteries. Through a series of experimental evaluations and feature analyses, the results demonstrate that the IDC-CRA model exhibits significant advantages across different datasets, confirming its broad applicability and superior performance in various battery materials. Figure 4 further validates the excellent performance of the IDC-CRA model. In the scatter plots of capacity estimation (a and c), the predicted values after TL are highly fitted to the ideal diagonal, showing a significant improvement in prediction accuracy. This means that the IDC-CRA model can achieve highly accurate capacity estimation by predicting the battery capacity with a very small deviation between its prediction results and the actual values. The error distribution plots (b and d), on the other hand, reveal that the error range of the model is significantly narrowed on all samples and the median is close to zero, indicating that the IDC-CRA model maintains a high degree of stability and robustness in the prediction process. This error distribution characteristic indicates that the model maintains consistent prediction performance under different operating conditions and data conditions, which reduces the occurrence of extreme errors and enhances the reliability of the model.
As shown in Table 3 and Figure 5, the IDC-CRA model shows significant improvement in capacity estimation capability and generalization performance compared to before TL is applied. Specifically, the RMSE of the model is reduced from 1.13% to 0.522%, the R² is improved from 0.882 to 0.98, and the MAPE is reduced from 1.06% to 0.431%. The changes in these metrics indicate that the fitting ability of the IDC-CRA model has been significantly enhanced to more accurately capture the changing characteristics of battery capacity. Meanwhile, the model’s adaptability to complex data has also been significantly improved, and it is able to maintain a high level of prediction performance under different battery materials and variable operating conditions. These results indicate that the proposed IDC-CRA hybrid model is able to capture the characteristics of battery capacity changes more accurately, demonstrating excellent stability and robustness. The excellent performance of the model in different battery materials validates its great potential in battery SOH estimation. This hybrid model effectively migrates a priori knowledge from the NCM battery data in the source domain to the NMC battery data in the target domain through the introduction of TL, which successfully overcomes the problem of data distribution differences due to different battery materials, thus significantly improving the generalization ability and prediction accuracy of the model. In addition, the running efficiency and resource utilization of the model are optimized. Through the introduction of TL and optimal design, the IDC-CRA model maintains high computational efficiency and resource utilization under complex working conditions and diverse data conditions. This not only improves the response speed of the model in real applications but also reduces the demand for computational resources, making it more suitable for deployment and operation in resource-constrained real environments. In conclusion, this study validates the high accuracy and robustness of the IDC-CRA hybrid model in estimating the SOH of NCM batteries via regression evaluation. The experimental findings indicate that the outstanding performance of the IDC-CRA model across various datasets not only notably enhances the accuracy and stability of capacity estimation but also refines the model’s generalization ability and computational efficiency. These merits endow the IDC-CRA model with extensive application potential in real-world battery management systems. Specifically, it can effectively prolong battery lifespan, boost battery performance, and cut down on maintenance costs. With this innovative model, this study significantly improves the performance of battery capacity estimation under complex working conditions and diverse data conditions, which lays a solid foundation for its application in more battery chemistries and practical operation scenarios and promotes the intelligent and efficient development of battery management systems.

4.2. Model Performance on NMC Cells

In this study, a two-stage TL strategy, consisting of a pre-training phase and a fine-tuning phase, is adopted to enhance the model’s adaptability to different battery chemistries and operating conditions while improving the accuracy and robustness of SOH estimation. During the pre-training phase, the model is trained on a large-scale source dataset (NCM battery data) to learn the general electrochemical characteristics of battery aging. The proposed model utilizes IDC for multi-scale spatiotemporal feature extraction and incorporates CRA to enhance global feature representation. The input data consists of four channels (voltage, current, charge capacity, and temperature), which are processed through three layers of IDC to extract critical features. Each IDC layer is followed by a MaxPooling layer (MaxPool2d (2,1)), which reduces computational complexity while preserving essential temporal dependencies. After pooling, the feature maps retain four channels, which are then adjusted to 64 channels via a 1×1 convolution layer (adjust_channels) to ensure compatibility with subsequent attention mechanisms and improve feature fusion.
During training, the Adam optimizer is employed, with an initial learning rate of 8 × 10−4. A relatively high learning rate facilitates rapid feature acquisition in the early training phase, while an early stopping strategy (patience = 1200) prevents overfitting. The batch size is set to 256, which helps stabilize gradient updates, reduce variance, and improve training efficiency. The model undergoes 10,000 training epochs to fully capture battery degradation patterns. SOH prediction is formulated as a 10-dimensional output (dense_soh layer), where each dimension represents the estimated SOH at a specific charge/discharge cycle. The L2 loss function is applied to ensure stable gradient updates, with a weighting factor α = [0.1] × 10, which balances contributions across different time steps, preventing any single time step from dominating the learning process.
To accommodate potential discrepancies in the target dataset (NMC battery data) arising from different battery chemistries or operating conditions, a selective parameter update strategy is adopted in the fine-tuning phase. Specifically, the IDC module is frozen (requires_grad = False) to retain the fundamental battery degradation features learned from the source dataset, ensuring that core characteristics remain unchanged. Meanwhile, the adjust_channels layer, CRA module, and dense layers are fine-tuned to adapt to the distributional differences of the target dataset, thereby enhancing the model’s generalization to new battery chemistries. To prevent instability in small-scale datasets, the learning rate is adjusted to 6 × 10−4, slightly lower than the 8 × 10−4 used in pre-training, ensuring smooth convergence. The model is trained for 351 epochs, with early stopping (patience = 1000) to halt training once no further performance improvement is observed, thus preventing overfitting. Furthermore, batch normalization is implemented to normalize data distribution and improve generalization. Additionally, an adaptive loss weighting mechanism dynamically adjusts the contribution of different loss components based on dataset characteristics. To further mitigate overfitting, a dropout layer (0.2) is introduced after dense_2, enhancing model robustness during testing.
To validate the generalization ability of our proposed hybrid model across different battery materials, this study adopts a transfer-learning strategy. This approach effectively transfers prior knowledge from the source domain to the target domain. In this particular section, we specifically apply the feature patterns and rules gleaned from the NCM battery data in the source domain to the NMC battery data in the target domain. By doing so, the model can not only make full use of the abundant information within the NCM battery data but also address the issue of data distribution disparities in the target domain (NMC battery data) caused by the different battery materials. In implementing the TL process, we first trained the initial hybrid model on the source domain NCM battery data to ensure that the model is able to accurately capture and learn the key features and patterns in the NCM battery data. Then, the pre-trained model was applied to the target domain NCM battery data through a fine-tuning technique. During the fine-tuning process, we retained the feature extraction capabilities learned by the model in the source domain while appropriately tuning and optimizing it for the data distribution in the target domain. This process ensures that the model can quickly adapt and effectively handle different battery material properties in the new domain.
To evaluate the effect of TL, we experimentally compare and analyze the prediction performance and generalization ability of models without and with TL. The experimental results show that the hybrid model processed with TL significantly outperforms the model without TL on NMC battery data. Specifically, Figure 6 represents the comparison of the estimated capacity of the NMC battery sample with the actual capacity and the error distribution. In the fitting plots (a and c), the distribution of the fitted points of estimated capacity versus actual capacity for all samples is dense and, overall, close to the ideal diagonal, especially in the capacity range of 2.6 Ah to 2.9 Ah interval. As can be seen from the error distribution plots (b and d), the error ranges of all samples are significantly narrowed and the median is close to zero, indicating a significant increase in the robustness of the model. Notably, the samples (#03, #06, #10, and #12) have the smallest errors, which well validates the model’s ability to adapt to complex samples. As shown in Table 3, by comparing the model performance before and after TL, it is clear that the introduction of TL significantly improves the predictive ability of the model. Specifically, TL significantly reduces the root mean square error of the model from 3.63% to 0.283%, indicating that the prediction error of the model is significantly reduced; R² significantly improves from −0.149 to 0.992, indicating that the model is close to the ideal state in terms of its ability to fit the relationship between the target variables and the predicted values; in addition, the MAPE reduces from 3.19% to 0.22%, indicating that the prediction error of the model among different samples is significantly reduced. The prediction error between different samples is significantly reduced. These results show that TL is not only effective in solving the problem of pre-prediction bias due to the difference in data distribution across different batteries but also in improving battery performance. It also greatly enhances the generalization ability and prediction accuracy of the model in complex data environments. Figure 7 visualizes the difference in capacity estimation performance of NMC battery samples with and without TL in the form of a histogram. Before the introduction of TL, the model exhibits a large bias in capacity estimation, reflecting the significant impact of the difference in data distribution of different batteries on the prediction results. The reason for the poor performance of the model’s results on the NMC dataset in the absence of TL is analyzed. Firstly, due to the significant differences in the electrochemical characteristics of NCM and NMC batteries, their degradation patterns, charge/discharge behavior, and SOH evolution trends are not entirely identical. Consequently, when the model is trained solely on the NMC dataset without TL, its learning is constrained by the limited dataset size, making it difficult to fully capture the degradation features of the target domain. This results in higher prediction errors and reduced generalization capability. Secondly, the NMC dataset contains fewer samples compared to the NCM dataset, which further limits the model’s ability to comprehensively learn the SOH degradation patterns. With insufficient training data, the model is prone to overfitting, exhibiting strong performance on the training set but suffering from high prediction errors on the test set, indicating weakened generalization. Furthermore, the pre-trained model has already learned essential battery degradation patterns from the NCM dataset, utilizing IDC for multi-scale feature extraction and CRA for global feature integration, thereby enhancing the robustness of SOH representation. However, when TL is not applied, the model must learn SOH features from scratch on the NMC dataset. Given the limited size of the target domain dataset, the model’s feature extraction capability is constrained, leading to reduced prediction accuracy, increased errors, and overall performance degradation.
Taken together, these factors contribute to the model’s poor performance on the NMC dataset, as evidenced by significantly higher RMSE, lower R2 values, and larger prediction errors. This further reinforces the importance of TL in SOH estimation across different battery chemistries. To ensure the accuracy of our experimental results, we have conducted a comprehensive review of the model’s predictions, including data integrity checks and error analysis. The results in Table 3 align with our expectations and further confirm the effectiveness of TL, demonstrating that TL substantially enhances the model’s predictive performance in the target domain (NMC dataset) by compensating for the generalization limitations caused by the smaller dataset size. In contrast, with the TL strategy, the model effectively captures the distribution characteristics among different battery samples, resulting in a significant reduction of the prediction error. In addition, the TL process makes full use of the rich information in the source dataset and can still accurately extract the key features and optimize the model performance in the case of fewer target data samples. Especially in the face of significant distributional differences between different samples, the fine-tuning process is able to make more detailed hyperparameter adjustments to the target domain data, thus enhancing the robustness of the model. This approach not only significantly reduces the dependence on large-scale labeled data in the target domain but also improves the model’s ability to predict the SOH of batteries under diverse operating conditions. TL not only improves the accuracy and stability of the model but also optimizes the model’s ability to capture inter-sample differences in the high-dimensional feature space, especially on complex samples.

4.3. Comparison with Other Baseline Models

In evaluating the effectiveness of the proposed fusion modeling technique for battery health assessment, a comprehensive comparison with baseline models demonstrates significant improvements in accuracy, generalization ability, and computational efficiency, thereby validating its practical application value. In this section, comparative experiments with three baseline models—CNN, LSTM, and Transformer—are conducted to thoroughly assess the performance of the fusion model in battery capacity estimation. Specifically, these baseline models represent different types of deep learning architectures. CNN excels at extracting local features of voltage, current, and temperature, enabling accurate battery state determination, but it has limitations in handling long-term dependencies. LSTM is capable of capturing long-term dependencies in time series, improving prediction accuracy and stability; however, it is inefficient in processing large-scale, high-dimensional data and struggles with global feature modeling. In contrast, the Transformer model captures complex dependencies and global features through its global attention mechanism, exhibiting strong modeling capabilities yet requiring extensive data and computational resources, which may pose challenges in practical applications. Table 4 provides a comprehensive evaluation of model performance using three key metrics: RMSE, R², and MAPE. The results indicate that the proposed model outperforms traditional baseline models across all metrics. For instance, the RMSE of the proposed model is only 0.522%, significantly lower than the corresponding values of CNN (6.36%), LSTM (0.57%), and Transformer (1.67%), demonstrating its superior accuracy in reducing prediction errors. The R² value reaches 0.98, which is close to 1, indicating that the model can explain 98% of the variation in the target variable, whereas the R² values of the baseline models are 0.885, 0.835, and 0.801, respectively. This highlights the proposed model’s excellent goodness-of-fit. Additionally, the MAPE of the proposed model is only 0.431%, further validating its ability to maintain low relative error during the prediction process. In contrast, the MAPE values for the baseline models are 5.9%, 1.2%, and 1.51%, demonstrating the superior reliability and consistency of the proposed model’s predictions.
Furthermore, to comprehensively assess the model’s operational efficiency and resource requirements, we evaluate the inference time and floating-point operations (FLOPs) during the testing phase. The results show that the inference time of the proposed model is only 0.018 s, with 4.67 million FLOPs, demonstrating exceptional computational efficiency. Notably, compared to the computationally demanding Transformer model, the proposed model not only maintains, or even enhances, predictive performance but also significantly reduces computational resource consumption and greatly improves operational efficiency. This advantage enables the proposed model to process large-scale data more rapidly in real-world applications, making it highly suitable for battery health assessment tasks that require high real-time performance.
The outstanding performance of the proposed model regarding capacity estimation and error distribution can be more intuitively observed through the visual analysis presented in Figure 8a. It can be seen that the distribution of the fitted points of the proposed model closely aligns with the ideal diagonal. This indicates that the model boasts remarkable accuracy in predicting the battery capacity, and there is an extremely high level of agreement between the predicted values and the actual values. The error distribution plot in (b) further reveals that the error range of the proposed model is smaller, the median is closer to zero, and the number of outlier points is significantly reduced. This means that the model not only maintains a low prediction error overall but also accurately predicts the battery capacity in most cases and reduces the occurrence of extreme errors, highlighting its stability and reliability in practical applications. These experimental results fully demonstrate the significant advantages of the proposed model in terms of accuracy, generalization ability, and computational efficiency. Specifically, the proposed model effectively migrates the a priori knowledge from the NCM battery data in the source domain to the NMC battery data in the target domain through the introduction of TL, which successfully overcomes the challenges posed by the differences in data distribution and significantly improves the model’s adaptability and generalization ability in different battery materials. In addition, the optimized design enables the model to not only achieve comprehensive performance improvement but also maintain high operational efficiency and resource utilization when dealing with complex battery capacity estimation tasks. These advantages enable the proposed model to have a wide range of potential applications in real battery management systems, which can effectively extend battery life, improve battery performance, and reduce maintenance costs. In summary, through comparison experiments with CNN, LSTM, and Transformer baseline models, the proposed fusion model performs well in the battery capacity estimation task, verifying its superiority and wide applicability in different battery materials. The application of TL not only improves the prediction performance and generalization ability of the model but also optimizes the computational efficiency, which fully demonstrates its practical application value in battery health state assessment. This research result provides a solid theoretical foundation and practical support for the promotion and application of advanced modeling techniques in more battery materials and different application scenarios in the future and promotes the intelligent and efficient development of battery management systems.
To further validate the comprehensiveness and robustness of our proposed model, we systematically compare its performance against conventional machine learning techniques, specifically Gaussian process regression (GPR) and linear regression (LR), for SOH prediction in lithium-ion batteries. The experimental results in Table 4 clearly demonstrate the superior predictive capability of our model, with notable advancements across all key evaluation metrics. In terms of prediction accuracy, our model achieves an RMSE of just 0.522%, representing a 64.5% and 65.7% reduction compared to GPR (1.47%) and LR (1.52%), respectively. This substantial improvement highlights the model’s ability to provide more precise SOH estimations. Additionally, the coefficient of determination (R²) reaches 0.98, significantly outperforming GPR (0.814) by 16.6% and LR (0.675) by 45.2%, indicating its exceptional capability in capturing the nonlinear degradation trends of lithium-ion batteries. Furthermore, our model achieves a MAPE of just 0.431%, which is 63.5% and 55.1% lower than GPR (1.18%) and LR (0.959%), reinforcing its superior predictive accuracy and robust generalization ability across various operating conditions.
Beyond predictive performance, computational efficiency is a critical factor in practical applications. While traditional machine learning models, such as Gaussian process regression (GPR) and linear regression (LR), are generally known for their computational efficiency, the proposed model achieves an optimal balance between predictive accuracy and computational cost. Experimental results demonstrate that the inference time of the proposed model is 0.018 s, which, although slightly higher than GPR (0.0078 s) and LR (0.0023 s), remains well within the acceptable range for real-time BMS, ensuring its feasibility for deployment. In terms of computational complexity, GPR requires 25.6 trillion FLOPS, while LR demands only 10.3 trillion FLOPS. In contrast, the proposed model operates at 46.7 trillion FLOPS, striking a balance between maintaining a reasonable computational demand and achieving substantially higher predictive accuracy. The underlying reasons for this disparity lie in the distinct computational characteristics of these models. GPR exhibits higher complexity due to its reliance on computing and storing kernel matrices, which leads to a significant computational burden as dataset size increases. Despite its relatively short inference time, its high computational demand constrains scalability for large-scale datasets. Conversely, LR is computationally the most efficient (O(N)), involving only matrix multiplications and linear transformations, resulting in exceptionally fast inference speeds. However, its inherent linear assumptions limit its capability to effectively capture the nonlinear degradation patterns of lithium-ion batteries, thereby restricting its predictive accuracy.
In contrast, the proposed model utilizes a carefully optimized computational architecture to effectively reduce the computational overhead. This approach improves accuracy beyond traditional machine learning models while avoiding the excessive computational burden associated with deep learning architectures such as Transformers. Comparative results clearly show that the proposed model outperforms traditional machine learning approaches in terms of inference speed and computational efficiency while significantly outperforming traditional machine learning approaches in terms of prediction accuracy. These results clearly demonstrate that the proposed model not only accurately captures the complex electrochemical aging mechanism of Li-ion batteries but also provides robust SOH estimation under different operating conditions, which achieves an optimal balance between computational feasibility and prediction performance.

4.4. Ablation Experiments

To better demonstrate the role of the various parts of the model, we performed ablation experiments. The results (Table 5) demonstrate that IDC and CRA play a crucial role in improving the accuracy and stability of SOH estimation while also verifying the effectiveness of the TL strategy in enhancing the model’s generalization and adaptability across different datasets. Firstly, IDC effectively extracts the local dynamic characteristics of battery health evolution through its multi-scale convolutional kernel structure, enhancing the model’s sensitivity to SOH variations. When IDC was removed, the model’s RMSE increased to 1.72%, R² dropped to 0.789, and MAPE rose to 1.6% in the absence of TL. Although TL mitigated some of the performance degradation (reducing RMSE to 0.852%, increasing R² to 0.942, and lowering MAPE to 0.691%), there was still a noticeable gap compared to the full model. This further confirms the indispensable role of IDC in enhancing feature representation and improving the model’s adaptability to complex temporal data.
Secondly, CRA significantly contributes to suppressing redundant information and enhancing the model’s focus on critical global features, thereby improving the stability and generalization capability of SOH estimation. The ablation study shows that removing CRA had an even greater impact on model performance, with RMSE increasing to 1.94%, R² dropping to 0.526, and MAPE rising to 1.85% in the absence of TL. While TL partially compensated for the loss of CRA (reducing RMSE to 0.753%, increasing R² to 0.955, and lowering MAPE to 0.613%), the overall predictive accuracy remained inferior to the full model. This further validates that CRA plays a vital role in optimizing feature representation and enhancing the model’s robustness and predictive performance.
Additionally, the staged TL strategy proves to be critical in improving the model’s adaptability and generalization across different datasets. When transferring the model from the NCM dataset to the NMC dataset, the model exhibited limited adaptation capability without the transfer strategy, as evidenced by an RMSE of 1.13%, R² of 0.882, and MAPE of 1.06%. However, when the transfer strategy was applied, RMSE significantly decreased to 0.522%, R² improved to 0.98, and MAPE dropped to 0.431%, indicating that TL effectively leverages prior knowledge from the source dataset, allowing the model to rapidly adapt to different battery chemistries, reduce dependency on large labeled datasets, and enhance prediction accuracy. Overall, IDC, CRA, and the staged TL strategy work synergistically to improve the model’s feature learning capability, cross-dataset generalization, and predictive stability, ensuring superior SOH estimation performance under diverse operating conditions.

5. Conclusions

In this study, we propose a hybrid deep learning model that integrates IDC-CRA and TL to improve the accuracy, efficiency, and generalization ability of lithium-ion battery SOH estimation. The model utilizes IDC and CRA for effective feature extraction and comprehensive information capture, while a staged training strategy significantly enhances its adaptability to different battery chemistries and operational conditions. Experimental validation on two distinct battery datasets, NCM and NMC, confirms the model’s superior predictive accuracy and robustness across various scenarios. For the NCM dataset, TL reduces the RMSE from 1.13% to 0.522%, improves the R² from 0.882 to 0.98, and lowers the MAPE from 1.06% to 0.431%. Similarly, for the NMC dataset, TL further enhances performance, reducing RMSE from 3.63% to 0.283%, increasing R² from −0.149 to 0.992, and decreasing MAPE from 3.19% to 0.22%. This enables rapid cross-material transfer from 54 NCM cells to 16 NMC cells, demonstrating the model’s strong adaptability and high-precision predictive capability across different battery chemistries and diverse operating conditions. Moreover, its stable performance under dynamic charging and discharging protocols and varied temperature ranges highlights its potential for real-time deployment. To further validate the contribution of each component in the proposed model, we conducted an ablation study. The results demonstrate that removing either IDC or CRA leads to a notable degradation in model performance, underscoring the importance of these modules. Additionally, we compared the proposed model not only with deep learning baselines such as CNN, LSTM, and Transformer but also with traditional machine learning approaches, including GPR and LR. The results indicate that the IDC-CRA model consistently surpasses its counterparts in prediction accuracy, generalization capability, and computational efficiency. These findings highlight the IDC-CRA model’s potential for real-world applications in BMS, where precise and efficient SOH estimation is crucial for ensuring battery reliability, extending lifespan, and optimizing performance.
With the increasing demand for efficient and low-latency SOH estimation in EVs, ESS, and intelligent BMS, the scalability and computational efficiency of SOH estimation models have become critical considerations. Traditional deep learning models often rely on cloud computing platforms for high-performance computation, which may introduce communication delays and increase dependency on network infrastructure. In contrast, the proposed model, integrating IDC and CRA mechanisms, demonstrates significant computational efficiency, making it suitable for deployment in edge computing and embedded system environments.
For edge computing applications, the proposed model leverages IDC-based multi-scale feature extraction, which significantly reduces the number of trainable parameters and computational complexity compared to CNN architectures. This reduction enhances its feasibility for deployment on local BMS devices or edge servers, enabling real-time SOH estimation with minimal latency. Additionally, the CRA module’s channel reduction mechanism further alleviates computational overhead, allowing the model to operate efficiently under constrained power and processing conditions. Moreover, TL enhances the model’s adaptability to different battery chemistries and operating conditions, making it suitable for large-scale distributed BMS deployments and real-time health monitoring in heterogeneous battery systems.
For embedded system applications, the proposed model can be optimized using quantization-aware training and model pruning techniques, enabling efficient deployment on ARM-based processors, field-programmable gate arrays (FPGAs), and microcontroller units (MCUs). These optimizations minimize computational costs and memory usage while preserving predictive accuracy. Compared to LSTM or Transformer-based architectures, the proposed model demonstrates lower memory consumption and reduced computational latency, making it well-suited for onboard SOH estimation in electric vehicles (EVs), stationary energy storage systems, and industrial IoT applications. By enabling localized SOH estimation, the model reduces dependence on cloud-based computing, thereby enhancing system reliability and data security in battery-powered intelligent systems. Moreover, the integration of edge computing not only enhances system responsiveness but also decreases reliance on high-bandwidth networks, effectively lowering operational costs. Processing data directly on edge devices reduces data transmission, safeguards user privacy, and strengthens overall system security and reliability.

Author Contributions

Conceptualization, J.R.; methodology, J.R.; experimentation, J.R.; writing—original draft preparation, J.R., M.C. and D.S.; data processing, M.C.; model optimization, M.C.; validation, M.C.; supervision, D.S.; project administration, D.S.; final approval, D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [Independent Innovation Projects of the Hubei Longzhong Laboratory] grant number [2022ZZ-24] and [Hubei Superior and Distinctive Discipline Group of “New Energy Vehicle and Smart Transportation”] grant number [XKTD072023].

Data Availability Statement

The two datasets utilized in this study for modeling are openly accessible and can be found at https://doi.org/10.5281/zenodo.10963339; accessed on 10 February 2025 and https://www.batteryarchive.org/list.html; accessed on 10 February 2025, respectively.

Conflicts of Interest

The authors declare no competing interests.

References

  1. Khan, I.; Hou, F.; Le, H.P. The impact of natural resources, energy consumption, and population growth on environmental quality: Fresh evidence from the United States of America. Sci. Total Environ. 2021, 754, 142222. [Google Scholar] [CrossRef]
  2. Jie, H.; Khan, I.; Alharthi, M.; Zafar, M.W.; Saeed, A. Sustainable energy policy, socio-economic development, and ecological footprint: The economic implications of natural resources, population growth, and industrial development. Public Util. Policy 2023, 81, 101490. [Google Scholar] [CrossRef]
  3. Zheng, Y.; Wu, H.; Yi, W.; Lai, X.; Dai, H.; Han, X.; Ouyang, M. A novel classification method of commercial lithium-ion battery cells based on fast and economic detection of self-discharge rate. J. Power Sources 2020, 478, 229039. [Google Scholar] [CrossRef]
  4. Caizán-Juanarena, L.; Arnaiz, M.; Gucciardi, E.; Oca, L.; Bekaert, E.; Gandiaga, I.; Ajuria, J. Unraveling the Technology behind the Frontrunner LIC ULTIMO to Serve as a Guideline for Optimum Lithium-Ion Capacitor Design, Assembly, and Characterization. Adv. Energy Mater. 2021, 11, 2100912. [Google Scholar] [CrossRef]
  5. Fouladvand, M.; Naji, L.; Javanbakht, M.; Rahmanian, A. Electrochemical characterization of Li-ion conducting polyvinylidene fluoride/sulfonated graphene oxide nanocomposite polymer electrolyte membranes for lithium-ion batteries. J. Membr. Sci. 2021, 636, 119563. [Google Scholar] [CrossRef]
  6. Wei, H.; He, C.; Li, J.; Zhao, L. Online estimation of driving range for battery electric vehicles based on SOC-segmented actual driving cycle. J. Energy Storage 2022, 49, 104091. [Google Scholar] [CrossRef]
  7. Bai, S.; Liu, C. Overview of energy harvesting and emission reduction technologies in hybrid electric vehicles. Renew. Sustain. Energy Rev. 2021, 147, 111188. [Google Scholar] [CrossRef]
  8. Tian, J.; Xiong, R.; Shen, W.; Sun, F. Electrode ageing estimation and open circuit voltage reconstruction for lithium-ion batteries. Energy Storage Mater. 2021, 37, 283–295. [Google Scholar] [CrossRef]
  9. Lai, X.; Jin, C.; Yi, W.; Han, X.; Feng, X.; Zheng, Y.; Ouyang, M. Mechanism, modeling, detection, and prevention of the internal short circuit in lithium-ion batteries: Recent advances and perspectives. Energy Storage Mater. 2021, 35, 470–499. [Google Scholar] [CrossRef]
  10. Wang, H.; Liu, F.; Yu, R.; Xiao, Z.; Zhu, Z.; Zhou, L.; Wu, J. Co-gradient Li-rich cathode relieving the capacity decay in Lithium-ion batteries. Nano Energy 2022, 100, 107439. [Google Scholar] [CrossRef]
  11. Tian, J.; Xu, R.; Wang, Y.; Chen, Z. Capacity attenuation mechanism modeling and health assessment of lithium-ion batteries. Energy 2021, 221, 119682. [Google Scholar] [CrossRef]
  12. Na, Y.; Sun, X.; Fan, A.; Cai, S.; Zheng, C. Methods for enhancing the capacity of electrode materials in low-temperature lithium-ion batteries. Chin. Chem. Lett. 2021, 32, 973–982. [Google Scholar] [CrossRef]
  13. Wang, S.; Wu, F.; Takyi-Aninakwa, P.; Fernandez, C.; Stroe, D.I.; Huang, Q. Improved singular filtering-Gaussian process regression-long short-term memory model for whole-life-cycle remaining capacity estimation of lithium-ion batteries adaptive to fast aging and multi-current variations. Energy 2023, 284, 128677. [Google Scholar] [CrossRef]
  14. Gou, B.; Xu, Y.; Feng, X. State-of-health estimation and remaining-useful-life prediction for lithium-ion battery using a hybrid data-driven method. IEEE Trans. Veh. Technol. 2020, 69, 10854–10867. [Google Scholar] [CrossRef]
  15. Cai, L.; Meng, J.; Stroe, D.I.; Peng, J.; Luo, G.; Teodorescu, R. Multi-objective optimization of data-driven model for lithium-ion battery SOH estimation with short-term feature. IEEE Trans. Power Electron. 2020, 35, 11855–11864. [Google Scholar] [CrossRef]
  16. Qiao, J.; Wang, S.; Yu, C.; Yang, X.; Fernandez, C. A chaotic firefly-Particle filtering method of dynamic migration modeling for the state-of-charge and state-of-health co-estimation of a lithium-ion battery performance. Energy 2023, 263, 126164. [Google Scholar] [CrossRef]
  17. Xu, T.; Peng, Z.; Liu, D.; Wu, L. A hybrid drive method for capacity prediction of lithium-ion batteries. IEEE Trans. Transp. Electrif. 2021, 8, 1000–1012. [Google Scholar] [CrossRef]
  18. Mc Carthy, K.; Gullapalli, H.; Ryan, K.M.; Kennedy, T. Electrochemical impedance correlation analysis for the estimation of Li-ion battery state of charge, state of health and internal temperature. J. Energy Storage 2022, 50, 104608. [Google Scholar] [CrossRef]
  19. Zhao, J.; Wang, Z.; Wu, Y.; Burke, A.F. Predictive pretrained transformer (PPT) for real-time battery health diagnostics. Appl. Energy 2025, 377, 124746. [Google Scholar] [CrossRef]
  20. Zhao, J.; Wang, Z. Specialized convolutional transformer networks for estimating battery health via transfer learning. Energy Storage Mater. 2024, 71, 103668. [Google Scholar] [CrossRef]
  21. Ma, Y.; Shan, C.; Gao, J.; Chen, H. Multiple health indicators fusion-based health prognostic for lithium-ion battery using transfer learning and hybrid deep learning method. Reliab. Eng. Syst. Saf. 2023, 229, 108818. [Google Scholar] [CrossRef]
  22. Ma, G.; Xu, S.; Yang, T.; Du, Z.; Zhu, L.; Ding, H.; Yuan, Y. A transfer learning-based method for personalized state of health estimation of lithium-ion batteries. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 759–769. [Google Scholar] [CrossRef]
  23. Yang, Y.; Xu, Y.; Nie, Y.; Li, J.; Liu, S.; Zhao, L.; Yu, Q.; Zhang, C. Deep transfer learning enables battery state of charge and state of health estimation. Energy 2024, 294, 130779. [Google Scholar] [CrossRef]
  24. Zhao, H.; Chen, Z.; Shu, X.; Shen, J.; Lei, Z.; Zhang, Y. State of health estimation for lithium-ion batteries based on hybrid attention and deep learning. Reliab. Eng. Syst. Saf. 2023, 232, 109066. [Google Scholar] [CrossRef]
  25. Shu, X.; Shen, J.; Li, G.; Zhang, Y.; Chen, Z.; Liu, Y. A flexible state-of-health prediction scheme for lithium-ion battery packs with long short-term memory network and transfer learning. IEEE Trans. Transp. Electrif. 2021, 7, 2238–2248. [Google Scholar] [CrossRef]
  26. Shen, S.; Sadoughi, M.; Li, M.; Wang, Z.; Hu, C. Deep convolutional neural networks with ensemble learning and transfer learning for capacity estimation of lithium-ion batteries. Appl. Energy 2020, 260, 114296. [Google Scholar] [CrossRef]
  27. Zhao, J.; Qu, X.; Han, X.; Wu, Y.; Burke, A.F. Cross-material battery capacity estimation using hybrid-model fusion transfer learning. J. Power Sources 2025, 625, 235674. [Google Scholar] [CrossRef]
  28. Ren, L.; Dong, J.; Wang, X.; Meng, Z.; Zhao, L.; Deen, M.J. A data-driven auto-CNN-LSTM prediction model for lithium-ion battery remaining useful life. IEEE Trans. Ind. Inform. 2020, 17, 3478–3487. [Google Scholar] [CrossRef]
  29. Lv, Z.; Zhao, J. Resource-efficient artificial intelligence for battery capacity estimation using convolutional FlashAttention fusion networks. eTransportation 2025, 23, 100383. [Google Scholar] [CrossRef]
  30. Dai, Y.; Yu, A. CNN-LSTM GRU combination method for lithium battery SOH estimation based on health feature parameters. Energy Storage Sci. Technol. 2022, 5, 1641–1649. [Google Scholar] [CrossRef]
  31. Qian, C.; Guan, H.; Xu, B.; Xia, Q.; Sun, B.; Ren, Y.; Wang, Z. A CNN-SAM-LSTM hybrid neural network for multi-state estimation of lithium-ion batteries under dynamical operating conditions. Energy 2024, 294, 130764. [Google Scholar] [CrossRef]
  32. Jia, C.; Tian, Y.; Shi, Y.; Jia, J.; Wen, J.; Zeng, J. State of health prediction of lithium-ion batteries based on bidirectional gated recurrent unit and transformer. Energy 2023, 285, 129401. [Google Scholar] [CrossRef]
  33. Yu, W.; Zhou, P.; Yan, S.; Wang, X. Inceptionnext: When inception meets convnext. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 5672–5683. [Google Scholar]
  34. Kang, B.; Moon, S.; Cho, Y.; Yu, H.; Kang, S.J. Metaseg: Metaformer-based global contexts-aware network for efficient semantic segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 434–443. [Google Scholar]
  35. Wang, F.; Zhai, Z.; Zhao, Z.; Di, Y.; Chen, X. Physics-informed neural network for lithium-ion battery degradation stable modeling and prognosis. Nat. Commun. 2024, 15, 4332. [Google Scholar] [CrossRef]
  36. National Aeronautics and Space Administration (NASA). Lithium-Ion Battery Data Set. 2007. Available online: https://www.nasa.gov (accessed on 3 December 2024).
  37. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Battery SOH estimation architecture combining IDC-CRA hybrid deep learning model and transfer learning. (a) Shows the preprocessing steps of the data, including the partitioning of the battery aging data into training, validation, and test sets, as well as linear interpolation and feature extraction. (b) Pre-training the base model, which is a hybrid neural network using the IDC-CRA model architecture. (c) TL approach, where the fine-tuning process only adjusts the weights of a specific layer (CRA) and freezes the weights of all other layers. (d) SOH estimation results and model performance of the proposed method.
Figure 1. Battery SOH estimation architecture combining IDC-CRA hybrid deep learning model and transfer learning. (a) Shows the preprocessing steps of the data, including the partitioning of the battery aging data into training, validation, and test sets, as well as linear interpolation and feature extraction. (b) Pre-training the base model, which is a hybrid neural network using the IDC-CRA model architecture. (c) TL approach, where the fine-tuning process only adjusts the weights of a specific layer (CRA) and freezes the weights of all other layers. (d) SOH estimation results and model performance of the proposed method.
Energies 18 01491 g001
Figure 2. Degradation trends and charge/discharge characteristics analysis of battery samples. (a,b) illustrate the degradation trends of discharge capacity with cycle number for the NCM and NMC datasets. (c,d) show the variations of capacity and voltage over time under specific cycle conditions. (e) presents the change in current over time under specific cycle conditions. (f) presents the trends of charge capacity and charge voltage under specific cycle conditions. The color gradient transitions from blue to red, indicating the progression of cycle numbers.
Figure 2. Degradation trends and charge/discharge characteristics analysis of battery samples. (a,b) illustrate the degradation trends of discharge capacity with cycle number for the NCM and NMC datasets. (c,d) show the variations of capacity and voltage over time under specific cycle conditions. (e) presents the change in current over time under specific cycle conditions. (f) presents the trends of charge capacity and charge voltage under specific cycle conditions. The color gradient transitions from blue to red, indicating the progression of cycle numbers.
Energies 18 01491 g002
Figure 3. Architecture of the proposed hybrid deep learning model for battery SOH estimation. The IDC module decomposes the large kernel deep convolution into four efficient parallel branches, including 3 × 3 squares, 1 × 11 and 11 × 1 orthogonal rectangles, and a constant mapping. In the CRA module, attention is computed before the query, and key matrices reduce the channel dimensions by average pooling to improve computational efficiency while maintaining global feature capture capability.
Figure 3. Architecture of the proposed hybrid deep learning model for battery SOH estimation. The IDC module decomposes the large kernel deep convolution into four efficient parallel branches, including 3 × 3 squares, 1 × 11 and 11 × 1 orthogonal rectangles, and a constant mapping. In the CRA module, attention is computed before the query, and key matrices reduce the channel dimensions by average pooling to improve computational efficiency while maintaining global feature capture capability.
Energies 18 01491 g003
Figure 4. Comparison of estimated and actual capacities for NCM battery samples and error distribution. (a,c) present the predicted capacities versus actual capacities of the battery samples, with each plot including data from five batteries. The color gradient represents the cycle number. Subplots (b,d) illustrate the error distribution of the battery samples.
Figure 4. Comparison of estimated and actual capacities for NCM battery samples and error distribution. (a,c) present the predicted capacities versus actual capacities of the battery samples, with each plot including data from five batteries. The color gradient represents the cycle number. Subplots (b,d) illustrate the error distribution of the battery samples.
Energies 18 01491 g004
Figure 5. Comparison of performance metrics for NCM battery samples with and without transfer learning. (a) shows the RMSE values for NCM samples, where the TL method (pink bars) significantly reduces the error compared to the non-TL method (light blue bars). (b) presents the R² values, indicating higher prediction accuracy for the TL method (pink bars). (c) displays the MAPE values, highlighting the consistent error reduction achieved by the TL method (blue bars).
Figure 5. Comparison of performance metrics for NCM battery samples with and without transfer learning. (a) shows the RMSE values for NCM samples, where the TL method (pink bars) significantly reduces the error compared to the non-TL method (light blue bars). (b) presents the R² values, indicating higher prediction accuracy for the TL method (pink bars). (c) displays the MAPE values, highlighting the consistent error reduction achieved by the TL method (blue bars).
Energies 18 01491 g005
Figure 6. Comparison of estimated and actual capacities for NMC battery samples and error distribution. (a,c) present the comparison between estimated capacities and actual capacities for NMC battery samples, with each plot including data from eight batteries. The color gradient represents the cycle number, ranging from blue for the initial cycles to red for the later cycles. (b,d) illustrate the error distribution using violin plots, showing the range of data variation and central tendencies.
Figure 6. Comparison of estimated and actual capacities for NMC battery samples and error distribution. (a,c) present the comparison between estimated capacities and actual capacities for NMC battery samples, with each plot including data from eight batteries. The color gradient represents the cycle number, ranging from blue for the initial cycles to red for the later cycles. (b,d) illustrate the error distribution using violin plots, showing the range of data variation and central tendencies.
Energies 18 01491 g006
Figure 7. Comparison of performance metrics for NMC battery samples with and without transfer learning. (a) displays the RMSE values for NMC samples, where the TL method (light blue bars) significantly reduces errors compared to the non-TL method (pink bars). (b) presents the R² values, indicating that the TL method (dark blue bars) improves prediction accuracy. (c) displays the MAPE values, highlighting the consistent error reduction achieved by the TL method (pink bars).
Figure 7. Comparison of performance metrics for NMC battery samples with and without transfer learning. (a) displays the RMSE values for NMC samples, where the TL method (light blue bars) significantly reduces errors compared to the non-TL method (pink bars). (b) presents the R² values, indicating that the TL method (dark blue bars) improves prediction accuracy. (c) displays the MAPE values, highlighting the consistent error reduction achieved by the TL method (pink bars).
Energies 18 01491 g007
Figure 8. Performance comparison of different models in battery capacity estimation. (a) The proposed model’s fitting points are closest to the ideal y = x line, demonstrating the highest estimation accuracy; (b) The error distribution box plot indicates that the proposed model has a smaller error range, showing superior accuracy and stability compared to CNN, LSTM, Transformer, LR, and GPR models.
Figure 8. Performance comparison of different models in battery capacity estimation. (a) The proposed model’s fitting points are closest to the ideal y = x line, demonstrating the highest estimation accuracy; (b) The error distribution box plot indicates that the proposed model has a smaller error range, showing superior accuracy and stability compared to CNN, LSTM, Transformer, LR, and GPR models.
Energies 18 01491 g008
Table 1. Battery dataset characteristics and testing conditions.
Table 1. Battery dataset characteristics and testing conditions.
Battery DatasetNumber of
Batteries
Charge/
Discharge
Current Rate(C)
Nominal
Voltage (V)
Nominal
Capacity (Ah)
Capacity Degradation
Range
Temperature (°C)
NCM540.5/1
0.5/1
0.5/1
4.23.5100–60% or
80% SOH
25
35
45
NMC160.5/2
0.5/0.5, 0.5/2, 0.5/3
0.5/1, 0.5/2
3.63.0100–80% SOH15
25
35
Table 2. Model hyperparameter configuration.
Table 2. Model hyperparameter configuration.
HyperparametersExplanationValue
lrPre-training learning rate.8 × 10−4
Fine-tuning learning rate.6 × 10−4
batch_sizePre-training batch size.256
Fine-tuning batch size.1000
n_epochsPre-training epochs.10,000
Fine-tuning epochs.351
pool_ratioFeature maps for size reduction and information aggregation.4
in_channelsThe dimensionality of the input features.4
out_channelsThe dimensionality of the output features.64
num_headsNumber of attention heads calculated in parallel.2
dense_1Mapping convolutional layer dimension.(64,64)
dim1Number of dimensions of CRA.64
dense_sohMapping capacity estimation results.10
square_kernel_sizeSize of square convolution kernel.3
band_kernel_sizeSize of the rectangular convolution kernel.11
branch_ratioChannel allocation ratio for IDC modules.0.26
attn_dropDropout rate.0.3
patienceEpoch of early stops.1200
alphaPre-training loss weight.[0.1] × 10
Fine-tuning train_alpha.[0.1] × 9 + [0.]
Table 3. Performance metrics of battery samples with and without transfer learning.
Table 3. Performance metrics of battery samples with and without transfer learning.
Battery TypeMethodRMSE (%)R2MAPE (%)
NCMWithout transfer1.130.8821.06
With transfer0.5220.980.431
NMCWithout transfer3.63−0.1493.19
With transfer0.2830.9920.22
Table 4. Comparison of the proposed model with other base models.
Table 4. Comparison of the proposed model with other base models.
ModelRMSE (%)R2MAPE (%)Inference
Time (Seconds)
FLOPS
(Hundred Billion)
GPR1.470.8141.180.00782.56
LR1.520.6750.9590.00231.03
CNN6.360.8555.90.00721.75
LSTM0.570.8351.20.00976.79
Transformer1.670.8011.510.01314.76
Proposed0.5220.980.4310.0184.67
Table 5. Ablation study on the effectiveness of transfer learning in SOH estimation.
Table 5. Ablation study on the effectiveness of transfer learning in SOH estimation.
Battery TypeAblation ExperimentMethodRMSE (%)R2MAPE (%)
NCMCRAWithout transfer1.720.7891.6
With transfer0.8520.9420.691
IDCWithout transfer1.940.5261.85
With transfer0.7530.9550.613
IDC + CRAWithout transfer1.130.8821.06
With transfer0.5220.980.431
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ren, J.; Cai, M.; Shi, D. Efficient Hybrid Deep Learning Model for Battery State of Health Estimation Using Transfer Learning. Energies 2025, 18, 1491. https://doi.org/10.3390/en18061491

AMA Style

Ren J, Cai M, Shi D. Efficient Hybrid Deep Learning Model for Battery State of Health Estimation Using Transfer Learning. Energies. 2025; 18(6):1491. https://doi.org/10.3390/en18061491

Chicago/Turabian Style

Ren, Jinling, Misheng Cai, and Dapai Shi. 2025. "Efficient Hybrid Deep Learning Model for Battery State of Health Estimation Using Transfer Learning" Energies 18, no. 6: 1491. https://doi.org/10.3390/en18061491

APA Style

Ren, J., Cai, M., & Shi, D. (2025). Efficient Hybrid Deep Learning Model for Battery State of Health Estimation Using Transfer Learning. Energies, 18(6), 1491. https://doi.org/10.3390/en18061491

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop