Next Article in Journal
Assessment of Potential of Organic Waste Methane for Implementation in Energy Self-Sufficient Wastewater Treatment Facilities
Previous Article in Journal
Energy-Efficient Chemistry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CBATE-Net: An Accurate Battery Capacity and State-of-Health (SoH) Estimation Tool for Energy Storage Systems

Department of Information Engineering Computer Science and Mathematics (DISIM), University of L’Aquila, 67100 L’Aquila, Italy
*
Author to whom correspondence should be addressed.
Energies 2025, 18(20), 5533; https://doi.org/10.3390/en18205533
Submission received: 22 September 2025 / Revised: 10 October 2025 / Accepted: 14 October 2025 / Published: 21 October 2025
(This article belongs to the Section D: Energy Storage and Application)

Abstract

In battery energy storage systems, accurately estimating battery capacity and state of health is crucial to ensure satisfactory operation and system efficiency and reliability. However, these tasks present particular challenges under irregular charge–discharge conditions, such as those encountered in renewable energy integration and electric vehicles, where heterogeneous cycling accelerates degradation. This study introduces a hybrid deep learning framework to address these challenges. It combines convolutional layers for localized feature extraction, bidirectional recurrent units for sequential learning and a temporal attention mechanism. The proposed hybrid deep learning model, termed CBATE-Net, uses ensemble averaging to improve stability and emphasizes degradation-critical intervals. The framework was evaluated using voltage, current and temperature signals from four benchmark lithium-ion cells across complete life cycles, as part of the NASA dataset. The results demonstrate that the proposed method can accurately track both smooth and abrupt capacity fade while maintaining stability near the end of the life cycle, an area in which conventional models often struggle. Integrating feature learning, temporal modelling and robustness enhancements in a unified design provides the framework with the ability to make accurate and interpretable predictions, making it suitable for deployment in real-world battery energy storage applications.

1. Introduction

Lithium-ion batteries are the best choice for today’s electric vehicles (EVs), portable gadgets and renewable energy systems (RESs) because they are highly energy-dense, long-lasting, and quick to charge [1,2,3]. They play an extremely important role in large-scale energy storage systems (ESSs), distributed grids (DGs) and photovoltaic (PV) applications. For example, in an on-board ESS, they buffer intermittent solar generation and stabilize power delivery [4,5,6]. The growing reliance on lithium-ion batteries is reflected in the projected increase in global demand, which is expected to rise from 225 GWh in 2021 to nearly 815 GWh by 2025 [7]. The transportation sector is set to represent the largest share of this demand, as illustrated in Figure 1. This rapid expansion highlights the importance of ensuring that batteries operate safely, efficiently and reliably throughout their service life.

1.1. Background Study

A key parameter in assessing battery health is capacity, which is defined as the maximum charge that a cell can store and deliver under specific conditions [8]. Degradation of capacity directly defines state of health (SoH) and determines remaining useful life (RUL), which is typically marked at 1.40 Ah of the rated capacity [9]. Therefore, accurate capacity estimation is essential not only for performance forecasting, but also for safety and maintenance scheduling [10].
Over the past decade, several methods have been pursued.
  • Experimental and model-based methods rely on open-circuit voltage (OCV), full charge–discharge cycles and electrochemical impedance spectroscopy (EIS) [11,12,13]. While these methods provide accurate battery capacity and state of health (SoH) estimation, they are costly, intrusive and unsuitable for continuous field monitoring [14]. Electrochemical models (EMs) capture detailed dynamics but require solving complex nonlinear equations, which limits their real-time applicability [15]. Furthermore, equivalent circuit models (ECMs) approximate behavior using lumped parameters but demand frequent recalibration [16,17].
  • Data-driven methods learn from routine telemetry of voltage, current, and temperature of lithium-ion battery cells [18]. Classical regressors, such as support vector regression (SVR), random forest (RF), and Gaussian process regression (GPR), capture nonlinear degradation when provided informative features [19,20]. Deep learning (DL) models extends this by automatically extracting patterns. In these models, convolutional neural networks (CNNs) capture localized temporal cues, long short-term memory (LSTM) and bidirectional long short-term memory (BiLSTM) model sequential dependencies [21,22,23], and transformers use self-attention to represent long-range relations [24]. Hybrid CNN-Transformer designs integrate local and global contexts [22]. Their main strength lies in adaptability and real-time deployment once trained, but performance strongly depends on dataset quality, preprocessing, and robustness strategies [25,26].
While these developments show clear progress, challenges remain. Both experimental and model-based methods face issues of scalability and computational cost. Although data-driven approaches are more flexible, they require architectures capable of generalizing across irregular and heterogeneous degradation patterns, which are often observed in real-world systems.

1.2. Novelty and Motivation

The growing dependence on lithium-ion batteries in EV, RES, aviation, and industrial systems makes accurate capacity and SoH estimation a safety-critical and performance-critical task. Failures in battery management systems (BMSs) can result in severe incidents ranging from unexpected shutdowns to catastrophic thermal events. Table 1 presents case studies of accidents that have occurred in the automotive, aviation, maritime and industrial sectors. These real-world examples of failure demonstrate that battery capacity and state of health (SoH) estimation must be considered not only as diagnostic tools, but also as essential safeguards for ensuring safety, reliability and cost-effective operation.
Figure 2 shows the broader concept of prognostics and health management (PHM). Although PHM principles apply to various subsystems, including PV panels, power electronics and different types of load, the battery is the most dynamic and degradation-prone component. Its health directly influences the stability of the entire system. Therefore, accurate estimation of battery capacity and SoH sits at the core of PHM for battery energy storage systems (BESSs).
Despite advances in experimental, model-based, and data-driven methods, many challenges remain. Experimental and model-based approaches, while interpretable, are intrusive, computationally intensive, and unsuitable for real-time use. Data-driven methods and hybrid neural networks, on the other hand, show strong potential, but often lack robustness under heterogeneous cycling and noisy field conditions.
The novelty of this work lies in addressing these gaps through CNN-BiLSTM with Temporal Attention and Ensemble network (CBATE-Net). It is a hybrid DL framework that integrates CNN for localized feature extraction, BiLSTM for sequential modeling, temporal attention for selective weighting of degradation-relevant intervals, and ensemble learning for stability. This combination enables accurate and interpretable battery capacity and SoH estimation under irregular operating conditions, including PV-coupled applications, where variability in solar generation and load demand accelerates degradation.
To situate these contributions clearly, the next section outlines the main research contributions and the overall organization of this paper.
Figure 2. An illustration of PHM for aging estimation [27].
Figure 2. An illustration of PHM for aging estimation [27].
Energies 18 05533 g002

1.3. Contributions and Organization

The main contributions of this study are summarized as follows:
  • Utilize the NASA Prognostics Center of Excellence (PCoE) lithium-ion batteries dataset. The operating specifications of each battery are summarized in Table 2.
  • Develop a hybrid prognostic framework (CBATE-Net): A unified architecture that integrates convolutional neural networks for local feature extraction, bidirectional long short-term memory networks (BiLSTM) for sequence modeling, temporal attention for selective focus on degradation-relevant intervals, and ensemble learning as shown in Table 3 for robust predictions.
  • Establish a standardized preprocessing pipeline: Discharge signals from the NASA PCoE lithium-ion batteries dataset [26] are aligned, resampled, cleaned, and normalized to produce cycle-wise sequences suitable for deep learning.
  • Construct and analyze a 12-dimensional feature set: Features derived from voltage, current, and temperature signals along with their derivatives and energy descriptors are evaluated for correlation with capacity and SoH to ensure statistical relevance.
  • Demonstrate robustness across heterogeneous cycling: The framework is evaluated on four NASA cells with distinct degradation trajectories, confirming its adaptability under irregular and noisy operating conditions.
  • Provide interpretability through temporal attention and ablation analysis: The attention mechanism highlights critical intervals of discharge profiles, while ablation studies quantify the contribution of each model component.
  • Position the framework for deployment in real-world applications: The model supports accurate, interpretable, and stable predictions for capacity and SoH, making it suitable for safety-critical domains, such as in EV, grid-level storage, and PV-coupled energy systems.
Many studies combine convolution, bidirectional LSTM, and attention for SoH or capacity. Our work contributes (i) an ensemble-integrated hybrid defined as part of the model with ablation studies and per seed reporting that show variance reduction and improved stability near end-of-life; (ii) a deployment profile in Table 4 that reports parameter count, compute, latency, and memory supporting embedded BMS feasibility; (iii) time-aware evaluation made primary with strictly time-forward protocols in Section 4.1 and results summarized in Table 5 and Table 6; (iv) a unified training and evaluation pipeline with a single feature set and identical preprocessing and splits for all methods so the comparisons in Table 6 and Table 7 are strictly comparable; finally, (v) a decision-level remaining useful life definition tied to the dataset’s end-of-life rule in Section 4.3.
The aim of this study is organized as follows. Section 2 reviews related works. Section 3 presents the adopted framework, NASA PCoE dataset of lithium-ion batteries, data preprocessing, feature extraction steps, the proposed CBATE-Net architecture, and the training and model selection procedure. Section 4 reports experimental results, discusses error dynamics and ensemble stability, and RUL estimation from capacity predictions. Section 5 compares the framework with conventional machine learning (ML), deep learning (DL), and transformer-based models. Section 6 presents ablation studies, and Section 7 concludes the paper with future research directions.
In summary, this work contributes a robust and interpretable framework for battery capacity and SoH estimation, offering a practical step toward safer and more reliable BESS under diverse and irregular operating conditions.

2. Related Works

2.1. Overview of Capacity and SoH

Battery capacity quantifies the maximum charge a cell can store and deliver under specified conditions. It is computed as the integral of discharge current over time
C = t 0 t f I ( t ) d t
where I ( t ) is the discharge current, and t 0 and t f denote the start and end of discharge. In an ESS, with repeated cycling, capacity decreases due to electrochemical and mechanical degradation. The state of SoH provides a normalized measure of this degradation and is commonly expressed as
SoH = C measured C rated × 100 % .
Here C measured is the measured capacity from a given cycle and C rated is the nominal rated capacity. This definition makes capacity estimation the direct basis for SoH inference. When combined with the end-of-life threshold used in this study, defined as capacity ≤ 1.40 Ah according to the NASA PCoE protocol, the SoH-based RUL estimation becomes well defined. For the four cells analyzed, this threshold corresponds to approximately 70 percent of the rated capacity. Accurate prediction of battery capacity and SoH is therefore essential for reliable operation of BESS, especially under irregular charge–discharge conditions, where degradation paths are heterogeneous.

2.2. Existing Approaches

Existing approaches for battery capacity and SoH estimation can be grouped into experimental and model-based, feature-driven ML, DL, hybrid, and transformer-based methods.
Experimental and model-based methods rely on EM, ECM, and analysis-based techniques, such as incremental capacity (IC), differential voltage, and EIS [28,29,30]. These provide accurate degradation insight, but require repeated parameter identification, complex solvers, or intrusive measurements, making them unsuitable for continuous onboard monitoring [31].
Feature-driven ML methods map handcrafted features, such as partial charging time (PCT) or multi-channel charging profiles (MCCPs) for capacity estimation. Classical regressors including SVR, RF, and boosting methods have been applied [32,33]. While interpretable, their performance depends on feature design and dataset-specific tuning.
DL methods reduce reliance on manual features by directly learning from voltage, current, and temperature signals. CNN capture localized patterns, LSTM and BiLSTM model temporal dependencies, and combined CNN-LSTM models have demonstrated improved accuracy over voltage-only inputs [27,34]. However, most studies were validated on limited datasets, restricting generalization.
Hybrid architectures integrate local and sequential modeling. Examples include CNN-BiLSTM hybrids and neural networks (NNs) augmented with impedance or wavelet-based indicators [16,35]. These models achieve high accuracy, but often rely on engineered health indicators (HIs), have heavy data preprocessing and limited scalability range.
Transformer-based models employ self-attention to capture long-range dependencies. CNN-Transformer hybrids and temperature-impedance fusion schemes have reported high accuracy on NASA PCoE datasets [8,36]. Robustness has been enhanced with augmentation strategies [37]. Despite their potential, transformers are computationally intensive and typically lack uncertainty calibration.
Finally, surveys and reviews highlight persistent gaps in dataset quality, preprocessing robustness, and generalization [28,31,38]. Collectively, these approaches show steady progress, but also reveal open challenges that motivate the hybrid framework developed in this study.

2.3. Challenges in Existing Approaches

Despite notable progress, there are still several gaps that limit the reliability of current methods for estimating battery capacity and state of health (SoH). Experimental and model-based approaches are still intrusive, time-consuming and computationally demanding, preventing real-time use. On the other hand, data-driven models rely heavily on dataset quality; however, variability in cycling conditions and environmental factors reduces their generalization [39]. Hand-crafted features only capture local cues, and pure sequential models may miss short but critical degradation events [40]. Transformer variants improve global context but require high computational resources, which makes embedded deployment difficult. Finally, most methods lack robustness mechanisms against noise and initialization, leading to unstable predictions near the end of the battery’s life. These open issues motivate a hybrid design combining localized feature extraction, temporal modelling, selective attention and ensemble stability to form the basis of the framework presented in the next section.

2.4. Technical Foundations of the Proposed Framework

Recent advances in battery prognostics combine spatial feature extraction, temporal modelling, attention and ensemble strategies. CNNs extract localized degradation cues from voltage, current and temperature sequences, capturing short-term variations linked to early aging [22]. BiLSTM then extends this by learning long-range dependencies in both the forward and backward directions, establishing links between the early and late stages of a lithium-ion battery discharge cycle [41]. A temporal attention mechanism highlights the most informative time steps, emphasizing degradation-relevant intervals while de-emphasizing less significant ones [24]. Ensemble learning aggregates the outputs of multiple trained models to improve robustness and reduce sensitivity to noise and initialization [37]. The NASA PCoE dataset remains the most widely used benchmark for such models. It provides voltage, current and temperature signals across the full life cycles of lithium-ion cells and offers diverse degradation trajectories, including gradual, abrupt and accelerated fade [16,32]. These channels are already available in BMS, making them practical for onboard estimation. Together, these elements form the basis of the hybrid framework proposed in this study. The design directly addresses the challenges of robustness, interpretability, and scalability identified in Section 2.3 by combining CNN, BiLSTM, temporal attention, and ensemble averaging.

3. Proposed Methodology

3.1. Framework

The proposed CBATE-Net is an advanced, hybrid deep learning (DL) framework designed for the accurate estimation of battery capacity and state of health (SoH). It combines four elements: CNN for localized feature extraction, BiLSTM for sequence modelling, a temporal attention mechanism for selective weighting of degradation-relevant intervals, and ensemble learning for robustness.
The pipeline follows a sequential yet integrated flow within the overall system. First, discharge signals of voltage, current and temperature are preprocessed and converted into standardized, cycle-wise sequences. A feature set derived from these signals is then passed to the hybrid DL model. CNN layers capture local variations, BiLSTM learns long-term dependencies in both directions and the attention mechanism highlights critical intervals linked to degradation of the lithium-ion batteries. To improve stability and precision, multiple model instances are trained independently and their predictions are aggregated through ensemble averaging.
Figure 3 illustrates the workflow of the proposed framework. The pipeline begins with the NASA lithium-ion battery dataset, which is cleaned and normalized first. The processed voltage, current, and temperature signals are then divided into training, validation, and testing subsets. These data are passed through a hybrid DL model consisting of a CNN, a BiLSTM, and an ensemble module, after which model selection takes place. The final stage produces capacity predictions together with a performance evaluation. SoH can then be derived from the predicted capacity as the ratio of predicted to nominal capacity, and RUL can be estimated relative to the 1.40 Ah end-of-life threshold, as explained in later sections. This design captures both localized fluctuations and long-term degradation patterns, ensuring stable performance under irregular cycling conditions.
The input to the model consists of data with dimensions 256 by 12. This input is first passed through a one-dimensional convolutional layer with 32 filters and a kernel size of 5. The output is then normalized using batch normalization, followed by a ReLU activation function and a max-pooling layer with a pool size of 2. Next, the signal is processed by a second Conv1D layer containing 64 filters with a kernel size of 5, again followed by batch normalization, ReLU activation, and max-pooling with a size of 2.
A third Conv1D layer with 128 filters and a kernel size of 3 is then applied, followed by batch normalization, ReLU activation, and a dropout layer with a dropout probability of 0.10 to prevent overfitting. The extracted features are then fed into a sequence of three bidirectional LSTM layers; the first two have 128 units in each direction, and the final one has 64 units in each direction. After the recurrent layers, a temporal attention mechanism is applied to focus on the most relevant time steps in the sequence.
The attention-weighted output is then passed through a fully connected (dense) layer with 64 neurons, using the ReLU activation function and L2 regularization with a penalty coefficient of 1 × 10−4. The final layer is a single-neuron dense layer that produces the model’s output. The State of Health (SoH) is computed using Equation (2). Finally, to enhance robustness and accuracy, the final prediction is obtained by averaging the outputs of M independently trained models in an ensemble. Table 3 describes the complete layer-by-layer configuration for the complete architecture of the system.

3.2. Experimental Dataset of Lithium-Ion Batteries of NASA

This study uses the NASA lithium-ion battery dataset, which is widely adopted in the literature for battery capacity and state of health (SoH) evaluation [21,26]. The operating specifications of each battery are summarized in Table 2. Constant current (CC)–constant voltage (CV) charging occurs at 1.5 A to 4.2 V with a cut-off current of 20 mA. Similarly, the following cut-off voltages were obtained: 2.7 V (B0005), 2.5 V (B0006 and B0018) and 2.2 V (B0007). Furthermore, the initial capacity was found to range from 1.86 to 2.04 Ah, with an end-of-life threshold of 1.40 Ah. Figure 4 shows the differences between the fresh and aged cycles of the lithium-ion battery parameters, such as voltage, current, and temperature over time in minutes. As the battery ages, there will be an early current discharge and temperature rise. Also, its output voltage decreases in amplitude. Therefore, the traces clearly show shifts in these parameters with usage and time.
The corresponding battery capacity degradation trajectories are shown in Figure 5. These plots illustrate the diversity of aging behaviors: B0005 exhibits relatively regular degradation, B0006 demonstrates slightly higher capacity retention, B0007 suffers sudden early degradation, and B0018 shows accelerated capacity fade. This diversity provides a challenging and representative testbed for evaluating model robustness.
Table 2. Specifications of NASA lithium-ion batteries used in this study.
Table 2. Specifications of NASA lithium-ion batteries used in this study.
Battery IDCharging Protocol (CC–CV)Discharge Current (A)Discharge Cut-Off Voltage (V)Operating Temperature (°C)Initial Capacity (Ah)End-of-Life Criterion (Ah)Role
B00051.5 A, 4.2 V, 20 mA cut-off2.02.725 ± 21.861.40Train/Test
B00061.5 A, 4.2 V, 20 mA cut-off2.02.525 ± 22.041.40Train/Test
B00071.5 A, 4.2 V, 20 mA cut-off2.02.225 ± 21.891.40Train/Test
B00181.5 A, 4.2 V, 20 mA cut-off2.02.525 ± 21.861.40Train/Test
Figure 4. Time-series plots of voltage, current, and temperature for fresh (Cycle 1) and aged (Cycle 168) battery cells.
Figure 4. Time-series plots of voltage, current, and temperature for fresh (Cycle 1) and aged (Cycle 168) battery cells.
Energies 18 05533 g004
Figure 5. Capacity versus cycle number for NASA cells, with end of life threshold marked at 1.40 Ah, about 70 percent of nominal capacity.
Figure 5. Capacity versus cycle number for NASA cells, with end of life threshold marked at 1.40 Ah, about 70 percent of nominal capacity.
Energies 18 05533 g005

3.3. Data Preprocessing

Effective preprocessing is essential for converting raw discharge data into reliable training inputs. This study used multi-channel sequences of voltage, current, and temperature from four NASA PCoE lithium-ion cells (B0005, B0006, B0007 and B0018). Outlier cycles were removed using a median absolute deviation (MAD) filter and those with capacities outside the valid range of 0.5–2.5 Ah were excluded. After cleaning, all features were standardized by z-score normalization using training set statistics. For each channel j, the normalized value x j ( n ) of a sample a j ( n ) is computed as
x j n = a j n μ j σ j ,
where μ j and σ j denote the mean and standard deviation of channel j in the training partition. Validation and test sets used the same parameters, and predictions were de-normalized after inference to recover their physical scale [42,43]. To improve generalization and avoid overfitting, light data augmentation was used only on the training set. Moreover, additive Gaussian noise was added with a mean of zero and a standard deviation of about 1–2 percent of each feature’s amplitude. Additionally, temporal jitter was also introduced, allowing a maximum shift of ±two sampling points, which is less than 0.5 percent of the total discharge time. These small changes simulate real sensor noise and slight timing errors found in actual BMS data. Therefore, the physical continuity of voltage, current, and temperature signals was preserved. Hence, no rescaling or artificial shape distortion was applied [18,44]. This reduced overfitting and improved generalization under irregular cycling conditions. Figure 6 shows the workflow of battery data processing and normalization.

3.4. Feature Extraction and Selection

Each discharge cycle was represented by a 12-dimensional feature vector derived from voltage, current, and temperature signals. The features included raw measurements (V, I, T), first- and second-order derivatives Δ V , Δ I , Δ T , Δ 2 V , Δ 2 I , power and energy terms ( P = V × I , normalized cumulative E, smoothed moving averages, and discharge duration.
Figure 7 shows the z-score distributions of these features across cycles. Voltage and its derivatives display the highest variability, indicating high sensitivity to degradation, while temperature-related channels provide complementary thermal information. Figure 8 further reports partial correlation coefficients with measured capacity, controlling for cycle index and battery identity. Figure 9 presents the Pearson correlation matrix with SoH [45]. Both analyses confirm that multiple features are informative, and no single descriptor dominates degradation. Consequently, all features were retained as model inputs. This enables convolutional layers to emphasize localized signals and BiLSTM to capture temporal dependencies. Retaining all features allows CBATE-Net to exploit both strong predictors and complementary cues, preserving nonlinear interactions that might otherwise be lost.

3.5. CBATE-Net Models

Battery capacity and SoH estimation requires models that can capture short-term localized patterns and long-range sequential dependencies from multi-channel discharge data. Traditional regressors and handcrafted features are insufficient under heterogeneous operating conditions. Therefore, the proposed framework integrates four complementary components: CNN, BiLSTM networks, temporal attention, and ensemble learning. The following subsections (Section 3.5.1, Section 3.5.2, Section 3.5.3 and Section 3.5.4) describe each model family in detail and explain its role in the integrated CBATE-Net framework.

3.5.1. Convolutional Neural Network

CNN layers are employed at the feature extraction stage to capture short-term, localized patterns in the discharge signals of voltage (V), current (I), and surface temperature (T), which often indicate early degradation. Examples include slope changes in voltage or sudden temperature spikes, detected through convolutional filters. The convolution operation for a one-dimensional input sequence is:
r t = a k t = p a t p k p ,
where a is the input sequence (V, I, T), K is the kernel (learnable filter), and r t is the resulting feature at position t. The extracted features are then passed to the BiLSTM layers.

3.5.2. Bidirectional Long Short-Term Memory (BiLSTM)

While CNN layers effectively capture localized degradation patterns, accurate capacity and SoH prediction also requires modeling temporal dependencies that span the full discharge sequence. LSTM networks address this need through gated recurrent units that regulate information flow across time steps and mitigate vanishing or exploding gradients in conventional RNN [21,24,46,47]. In a standard LSTM, the memory cell m t is updated by balancing information from the previous state with new candidate input (using forget and input gates):
m t = r t · m ˜ t 1 + g t · m ˜ t ,
where r t is the forget gate, g t is the input gate, and m ˜ t is the candidate update. The hidden states are updated as:
s t = q t · t a n h ( m t )
where s t is the hidden state (scaled by the output gate) and q t represents the output state. A Bidirectional LSTM (BiLSTM) extends this formulation by processing the sequence in both forward and backward directions, producing hidden states s t and s t . The final representation is obtained by concatenating both directions:
s t = s t ; s t .
This bidirectional design enables the framework to jointly exploit early- and late-cycle degradation cues. The extracted features are then passed to the temporal attention model.

3.5.3. Temporal Attention

While BiLSTM layers capture sequential dependencies, they treat all time steps equally, which may reduce the influence of degradation-relevant segments in a discharge profile. To address this, a temporal attention mechanism assigns higher weights to critical time steps [48,49]. Given a sequence of hidden states { h t } t = 1 T produced by the BiLSTM, the attention weight α t is computed as:
α t = e x p ( e t ) k = 1 T e x p ( e t ) , e t = f ( h t ) ,
where h t is the hidden state at time step t, T is the total number of steps, f · is a learnable scoring function, e t is the unnormalized importance score, e k is the normalization term, and α t is the normalized attention weight with t = 1 T α t = 1 . The context vector is then obtained as a weighted sum of hidden states:
c = t = 1 T α t h t .
Here, c is the aggregated sequence representation that emphasizes intervals critical to degradation. This mechanism amplifies critical regions, such as voltage plateaus that flatten with aging or temperature segments with accelerated rise. By focusing on these subsequences, temporal attention improves both interpretability and predictive accuracy.

3.5.4. Ensemble Learning

DL models such as CNN-BiLSTM with temporal attention capture degradation dynamics effectively. However, their predictions may still be sensitive to initialization, data splits, or stochastic optimization. To mitigate this variability, the framework incorporates ensemble learning, where multiple base models are trained independently and their outputs are averaged into a consensus prediction. If y ^ i denotes the prediction from the i-th model, the ensemble output is
y ^ e n s = 1 M i = 1 M y ^ i ,
where M is the number of models. This averaging reduces variance, enhances stability, and makes predictions less sensitive to noise or outliers. Diversity among base models is achieved through variations in weight initialization, training subsets, and minor hyperparameters [43]. The overall ensemble flow is illustrated in Figure 10.

3.6. Training and Model Selection

The proposed CBATE-Net framework was trained on the NASA PCoE lithium ion datasets. For each battery, we stratify samples by the target distribution (within that battery) and then assign 70% to training, 15% to validation, 15% to test while enforcing chronology [50,51]. Specifically, we require that validation and test samples come from cycles later than any training samples. Overlaps of input windows across splits are disallowed. Normalization statistics (mean, std) are computed on training data exclusively, then applied unchanged to validation and test. A fixed random seed ensures reproducibility. We re-ran all experiments under this chronology-constrained split; results maintain the same model ranking and principal conclusions. The strength of augmentation was kept within realistic sensor-noise limits. Therefore, when the model was retrained without augmentation, the average RMSE changed by less than 1 percent. Hence, this result confirms that the augmentation did not bias or distort the model. Optimization employed the Adam algorithm [10,20], which adapts learning rates and provides stable convergence under noisy gradients. The loss function was defined as mean squared error (MSE) between predicted and measured capacity values:
L = 1 N i = 1 N y ^ i y i 2 ,
where y ^ i and y i denote predicted and true capacities, respectively. To enhance robustness, ensemble learning described in Section 3.5.4 was applied. Multiple base models were trained independently with variations in initialization, batch ordering, and minor hyperparameter settings. Their outputs were aggregated by simple averaging, reducing variance, and stabilizing predictions even when individual learners differed in performance. Model performance was assessed using four standard regression metrics: root mean squared error (RMSE), mean absolute error (MAE), mean absolute percentage error (MAPE), and coefficient of determination ( R 2 ).
RMSE = 1 N i = 1 N y ^ i y i 2 ,
MAE = 1 N i = 1 N y ^ i y i ,
MAPE = 100 N i = 1 N y ^ i y i y i ,
R 2 = 1 i = 1 N y ^ i y i 2 i = 1 N y i y ¯ 2 ,
This combination of sampling, adaptive optimization, and ensemble averaging produced accurate and stable estimates, which is important for BESS, including PV coupled scenarios with non-uniform cycling and variable loads.

4. Experimental Results and Discussion

4.1. Experimental Configuration

The CBATE-Net model was evaluated using the NASA PCoE lithium-ion dataset and cells B0005, B0006, B0007 and B0018 [26]. These batteries provide multi-channel discharge signals of voltage, current and temperature throughout their entire life cycles. Preprocessing was performed in accordance with Section 3, producing normalized temporal sequences which were then converted into feature-by-time step matrices for sequence modelling. Performance was measured using RMSE, MAE, MAPE and R 2 . Baseline models were trained using the same preprocessing, loss and evaluation protocols. For completeness, all architectural and training hyperparameters used in CBATE-Net are summarized in Table 3. These include kernel sizes, number of filters, activation functions, BiLSTM hidden dimensions, learning-rate scheduling, early-stopping criteria, random seeds, and data-augmentation strengths [52,53,54]. All values are fixed across the four NASA cells (B0005, B0006, B0007, B0018) and were determined empirically through grid search and five-seed repetition to ensure reproducibility. This configuration provides a consistent basis for benchmarking and reflects the practical requirements of BESS, including PV-coupled applications.

Runtime and Resource Profiling

To evaluate the feasibility of deploying CBATE-Net in real-time BMS applications, we profiled the model’s computational requirements using a high-performance workstation. The analysis considered total parameter count, MAC operations per inference, model size, latency, and peak memory. Each test used batch = 1 and input dimensions T = 256 , F = 12 . Results are summarized in Table 4.
To quantify training variability, each battery model was trained five times with different random seeds under the same time-forward protocol. The reported mean ± standard deviation now reflects the variability across independent training runs rather than ensemble dispersion alone. Table 5 summarizes these multi-seed results.
Table 3. Complete CBATE-Net architectural and training layer-by-layer configuration.
Table 3. Complete CBATE-Net architectural and training layer-by-layer configuration.
ComponentParameterSetting/Value
InputSequence length (T), features (F)T = 256 steps, F = 12 channels
CNN front-endConv1D Layer 1Kernel = 5, Filters = 32, Stride = 1, Padding = ‘same’, Activation = ReLU
Conv1D Layer 2Kernel = 5, Filters = 64, Stride = 1, Padding = ‘same’, Activation = ReLU
Conv1D Layer 3Kernel = 3, Filters = 128, Stride = 1, Padding = ‘same’, Activation = ReLU + Dropout (0.10)
BiLSTM stackLayer 1128 units per direction → output dim 256
Layer 2128 units per direction → output dim 256
Layer 364 units per direction → output dim 128
Attention blockTypeAdditive attention (Equations (8) and (9)); hidden size = D = 64
Fully connected headFC 1 → FC 2 → Output[64 → 32 → 1]; Activation = ReLU for hidden layers
Activation functionsThroughout CNN/FCReLU; tanh in BiLSTM gates
Learning rate scheduleSchemePiecewise decay: initial LR = 2 × 10−4; decay × 0.95 every 80 epochs
Early stoppingCriterionValidation loss patience = 50 epochs (max epochs = 400)
Random seedsReproducibility{42, 142, 242, 342, 442} (five-seed repetition)
Augmentation strengthGaussian noise std1–2% of feature amplitude
Temporal shift range±2 time steps (max 0.5% of sequence length)
OptimizerTypeAdam ( β 1 = 0.9, β 2 = 0.999, ϵ = 1 × 10−8)
RegularizationWeight decay (L2)1 × 10−4 − 8 × 10−5
Batch size8–16 (battery-specific)
Ensemble regularizationEnsemble size (M)6–12 models per battery
HardwareRTX 3090 GPU, PyTorch 2.2
Table 4. Runtime and footprint of CBATE-Net (batch = 1, T = 256, F = 12).
Table 4. Runtime and footprint of CBATE-Net (batch = 1, T = 256, F = 12).
DeviceRuntime StackPrecisionParams (M)MACs/ InferencePeak MemoryLatency MedianLatency P95
Desktop GPU (RTX 3090)PyTorch 2.9 and MATLAB 2024b CUDAFP320.87656.35 M28 MB1.6 ms2.1 ms
Desktop CPU (Intel i7)PyTorch 2.9 and MATLAB 2024b MKLFP320.87656.35 M23 MB9.4 ms12.7 ms

4.2. Results Analysis

This section analyzes the figures relating to battery capacity versus cycle number and residual prediction vs. cycle number, from Figure 11, Figure 12, Figure 13 and Figure 14. Furthermore, Figure 15 shows the RMSE and MAE bar charts of the four cells, in which B0018 shows the highest error index. Additionally, battery state of health (SoH) vs. cycle number and SoH degradation rate analysis to cycle number is also described from Figure 16, Figure 17, Figure 18 and Figure 19. Across all datasets, the framework achieved an R 2 greater than 0.99 on average, while reducing the RMSE, MAE and MAPE relative to other machine learning (ML), deep learning (DL) and transformer-based artificial intelligence (AI) techniques. This confirms the effectiveness of combining local feature extraction, bidirectional sequence modelling, selective attention and ensemble averaging.

4.2.1. Prediction Performance Across Cells

Figure 11, Figure 12, Figure 13 and Figure 14 present the predicted versus actual capacity trajectories with residual distributions for the four NASA cells. The results confirm that CBATE-Net not only aligns closely with measured capacity but also adapts effectively to distinct degradation modes:
  • B0005 (gradual fade): Figure 11 shows the model follows smooth degradation with minimal lag, maintaining accuracy near the EoL threshold.
  • B0006 (higher retention): Figure 12 shows that slower fade is captured, and predictions remain stable even in late-life stages.
  • B0007 (early drop): Despite abrupt initial degradation, the model quickly adapts and tracks subsequent capacity decline, as shown in Figure 13.
  • B0018 (accelerated fade): Figure 14 shows that sharp nonlinear decline is reproduced without oscillation or overshoot.
Residual plots highlight prediction errors centered tightly around zero, with narrow spreads across all cells. The absence of systematic bias demonstrates that the ensemble strategy mitigates noise and initialization sensitivity. The small deviations reported in earlier ensemble plots reflected within-ensemble dispersion only; the present multi-seed analysis confirms that the overall variation across independent trainings remains below 1.5% RMSE, validating the model’s statistical robustness.
Figure 11. B0005 capacity vs. cycle with residuals.
Figure 11. B0005 capacity vs. cycle with residuals.
Energies 18 05533 g011
Figure 12. B0006 capacity vs. cycle with residuals.
Figure 12. B0006 capacity vs. cycle with residuals.
Energies 18 05533 g012
Figure 13. B0007 capacity vs. cycle with residuals.
Figure 13. B0007 capacity vs. cycle with residuals.
Energies 18 05533 g013
Figure 14. B0018 capacity vs. cycle with residuals.
Figure 14. B0018 capacity vs. cycle with residuals.
Energies 18 05533 g014

4.2.2. SoH Estimation and Degradation Dynamics

Figure 16, Figure 17, Figure 18 and Figure 19 extend the analysis to encompass SoH trajectories and their respective degradation rates. For all four batteries, the predicted SoH values are almost identical to the measured values, which confirms the reliability of the framework. In addition to accuracy, degradation rate analysis reveals the model’s interpretability.
  • Flattening voltage profiles and an increase in temperature are reflected in steeper degradation slopes.
  • The framework adjusts dynamically to differentiate between gradual and abrupt aging behaviors.
Its ability to capture not only SoH, but also aging dynamics, increases its suitability for early warning and prognostics in real-world ESS. This interpretability is further supported by validation results. As noted, a control run without augmentation showed almost the same error values, indicating that the Gaussian and temporal changes act as regularizers that help maintain the physical consistency of discharge cycles. Furthermore, it showed through a permutation-based feature importance test that voltage, current, temperature, and their first derivatives contribute more than 80% to the model’s predictions. Together, these findings confirm the physical relevance and efficiency of the 12-dimensional input set.

4.2.3. Quantitative Performance Metrics

To quantify robustness, reported below in Table 5 are the standard deviations across ensemble members. Across all four NASA cells, deviations remained small (typically < 0.0007 in RMSE and <0.001 in R 2 ), confirming that ensemble averaging reduces sensitivity to initialization and stochastic training effects. These outcomes are consistent with the scatter plots that cluster predictions tightly along the reference line, further validating the stability of the estimates. Values represent mean ± standard deviation over five independent random-seed runs; all results follow the time-forward protocol described in Section 3.6. These results highlight three critical aspects:
  • Precision: Error levels are within the tolerances required for practical BMS deployment.
  • Robustness: Consistency across four distinct degradation modes shows scalability.
  • Reliability near EoL (defined as the first cycle with capacity ≤ 1.40 Ah, i.e., SoH ≤ 70%; see Section 3.2 and Table 2): The framework maintains stability as capacity enters the warning region (80% SoH) and up to the EoL threshold (70% SoH), where errors typically spike in conventional models.
Table 5. Performance metrics for the CBATE-NET model with multi-seed robustness results (five independent runs, time-forward split).
Table 5. Performance metrics for the CBATE-NET model with multi-seed robustness results (five independent runs, time-forward split).
BatteryRMSE (Mean ± std)MAPE (%) (Mean ± std)MAE (Mean ± std) R 2 (Mean ± std)
B00050.004515 ± 0.000070.435567 ± 0.000060.003525 ± 0.010.997870 ± 0.0002
B00060.005875 ± 0.00040.569456 ± 0.000070.004581 ± 0.020.997822 ± 0.0003
B00070.004432 ± 0.000080.398754 ± 0.000070.003387 ± 0.0002 0.010.995881 ± 0.0003
B00180.009311 ± 0.000110.697720 ± 0.000100.006632 ± 0.0005 0.020.991975 ± 0.0004
Average0.0060 ± 0.000090.0045 ± 0.000070.53 ± 0.020.9958 ± 0.0003
The results shown in the bar chart in Figure 15 further confirm that CBATE-Net maintains low and stable errors across all batteries, with slightly higher values observed for B0018 due to its accelerated degradation profile.
Figure 15. Bar chart of CBATE-Net performance across NASA batteries.
Figure 15. Bar chart of CBATE-Net performance across NASA batteries.
Energies 18 05533 g015
Figure 16. B0005 SoH vs. cycle and the corresponding degradation-rate signal.
Figure 16. B0005 SoH vs. cycle and the corresponding degradation-rate signal.
Energies 18 05533 g016
Figure 17. B0006 SoH vs. cycle and the corresponding degradation-rate signal.
Figure 17. B0006 SoH vs. cycle and the corresponding degradation-rate signal.
Energies 18 05533 g017

4.2.4. Error Dynamics and Ensemble Stability

The residual analyses shown in Figure 11, Figure 12, Figure 13 and Figure 14 demonstrate balanced predictions with symmetric distributions and no heavy tails. Variance increases slightly near the EoL, reflecting reduced signal-to-noise ratios, but remains bounded. Compared to single models, ensemble averaging considerably reduces error variance. Multi-run bands confirm stability across training instances, with only modest widening in late cycles. This resilience guarantees reliable performance in embedded applications where minimal prediction uncertainty is essential.
Figure 18. B0007 SoH vs. cycle and the corresponding degradation-rate signal.
Figure 18. B0007 SoH vs. cycle and the corresponding degradation-rate signal.
Energies 18 05533 g018
Figure 19. B0018 SoH vs. cycle and the corresponding degradation-rate signal.
Figure 19. B0018 SoH vs. cycle and the corresponding degradation-rate signal.
Energies 18 05533 g019

4.3. Remaining Useful Life (RUL) Estimation

In addition to capacity and SoH tracking, CBATE-Net enables direct remaining useful life (RUL) estimation. RUL is defined by the first cycle index N EoL where predicted capacity falls at or below 1.40 Ah. All RUL predictions reported here are based on models trained with the chronology-constrained stratified split described in Section 3.6, which enforces full temporal ordering across cycles and prevents future information leakage. For a given cycle, k is defined as follows:
R U L ( k ) = N EoL k ,
where k is the current cycle index and N EoL is the predicted cycle at which capacity first drops below 1.40 Ah of the rated value.

5. Comparative Analysis

The effectiveness of the proposed CBATE-Net was evaluated by comparing it with different ML, DL and transformer-based approaches reported in prior work. All baseline models were trained using the same pre-processing pipeline, dataset splits and evaluation metrics defined in Section 3.3, Section 4.1 and Section 3.6, respectively. This ensures that any differences in performance reflect the architectures of the models rather than biases in the training or data.

5.1. Comparison with Machine Learning (ML) and Deep Learning (DL) Models

The comparative results are summarized in Table 6. These results are presented against different ML and DL networks. In addition to reporting the absolute error for each battery, the table also includes an ‘improvement’ column, which is defined as the percentage reduction in the proposed model relative to the best available baseline for each battery and metric. The notation ‘N/R’ is given for not-reported values in Table 6 and Table 7. Across all four NASA PCoE cells, CBATE-Net achieves the lowest RMSE and MAE, demonstrating notable improvements in challenging cases.
  • B0006: The RMSE was reduced by 26.6% compared to the best baseline, which confirms the benefits of hybrid modelling.
  • B0007: The MAE was reduced by 43.5%, despite CNN/LSTM-only models struggling to capture early degradation.
  • B0018: Although one baseline model performed slightly better than CBATE-Net under MAE (0.00612 vs. 0.00663), CBATE-Net consistently outperformed other models across all metrics.
Table 6. Comparison of performance metrics with previous works.
Table 6. Comparison of performance metrics with previous works.
BatteryMetricCNN-LSTM [55]CNN-WNN-WLSTM [56]HI-ALSTM [57]CNN-BiLSTM-AM [58]CBATE-Net (Proposed)Improve-
ment (%)
B0005RMSE0.01000.00570.01650.00740.00451521.40
MAE0.0070N/R0.01490.00380.0035257.72
MAPE (%)N/R0.69196.300.500.43556712.18
B0006RMSE0.01300.00800.01530.01140.00587526.62
MAE0.0090N/R0.01100.00490.0045815.55
MAPE (%)N/R0.92151.030.710.56945619.68
B0007RMSE0.00900.00520.0089N/R0.00443214.32
MAE0.0060N/R0.0068N/R0.00338743.55
MAPE (%)N/R0.5984.17N/R0.39875432.36
B0018RMSEN/RN/R0.01390.01080.00931113.79
MAEN/RN/R0.00830.00610.006632–8.37
MAPE (%)N/RN/R108.290.800.69772012.46
Average over batteriesRMSE0.01070.00630.01370.01000.006019.03 (avg Δ vs. best)
MAE0.0073N/R0.01030.00490.004512.11 (avg Δ vs. best)
MAPE (%)N/R0.73134.950.670.5319.17 (avg Δ vs. best)
On average across cells, CBATE-Net reduced RMSE by 19.03%, MAE by 12.11% and MAPE by 19.17%, compared to the strongest baselines. These improvements stem from the four-layered design: CNNs extract localized cues, BiLSTMs preserve long-range dependencies, temporal attention weights critical intervals and ensemble averaging reduces variance. Diagnostic residual plots (see Section 4.2.1) further confirm stability. Therefore, the proposed CBATE-Net maintains zero-centered errors with a narrow spread, whereas the predictions of other models show higher variance.

5.2. Comparison with Transformer-Based Approaches

Transformer-based architectures have emerged as promising candidates for estimating battery capacity and state of health (SoH) due to their ability to model long-range temporal dependencies via self-attention [59]. Table 7 extends the analysis to the latest transformer architectures, including CNN-Transformer [36], CNN-MVIP-Trans [60], PCC+Transformer [61], Transformer–GRU [62], DAE-KF Transformer [63], and SGEformer [64]. It also provides a unit-consistent comparison where every RMSE/MAE is in SoH fraction (0–1). For example, SGEformer (reported as 1.08% in the [64]) becomes 0.0108; PCC+Transformer (2.90%) becomes 0.0290. Our CBATE-Net entry is the multi-seed mean across B0005/B0006/B0007/B0018 under a time-forward split (0.006033 RMSE, 0.004531 MAE). ‘Average Improvement (%)’ is computed as in Equation (17) using the baselines present in that column.
Average Improvement ( % ) = 100 × mean ( baseline ) CBATE mean ( baseline ) .
Following are the improvements made as compared to the transformer-based architectures:
  • Average improvements: The proposed CBATE-Net reduces RMSE by 80.9% and MAE by 81.5% compared to the strongest transformer baselines.
  • Efficiency: Transformers require multi-head attention and positional encodings, which increase training cost and inference latency. On the other hand, the proposed CBATE-Net achieves better accuracy with significantly lighter computational demand.
Table 7. Comparison of performance metrics with transformer-based models.
Table 7. Comparison of performance metrics with transformer-based models.
MethodologyRMSEMAE
SGEformer0.01080.0096
PCC + Transformer0.02900.0260
DAE KF-transformer0.0345N/R
CNN-MVIP-Trans0.0500N/R
CNN-Transformer0.05500.0550
Transformer-GRU0.01190.0062
Transformer0.03000.0258
CBATE-Net (Proposed)0.0060330.004531
Average Improvements (%)80.981.5
The NASA dataset was collected at a constant laboratory temperature of about 25 °C. The model already uses temperature and its derivatives as input features, which allows it to capture intra-cycle thermal effects. The attention mechanism identifies intervals influenced by temperature, helping the model generalize better. In field data with wider temperature ranges, the absolute error may increase slightly, but the model’s relative performance advantage should remain. For deployment, light temperature-aware calibration or few-shot transfer using multi-temperature data can further improve robustness.

6. Ablation Study

A controlled ablation analysis was conducted to evaluate the contribution of the main design choices in the proposed enhanced CBATE-Net framework. In each case, one key component was removed or simplified, while all other configurations, preprocessing and hyperparameters remained identical to the baseline. This ensures that any observed differences in performance can be directly attributed to the modified component. The baseline is the full model, which integrates the following: (i) a convolutional neural network (CNN) front-end for local temporal feature extraction, (ii) a deep multi-layer bidirectional long short-term memory (BiLSTM) stack for long-range dependency modelling, and (iii) an ensemble of independently trained models for prediction stability. The following ablations were performed:
  • No CNN front-end: All convolutional layers were removed and the sequences were passed directly to the BiLSTM backbone. This resulted in the most significant performance drop, with RMSE increasing substantially and R 2 decreasing to approximately 0.986. These results confirm that convolutional layers are crucial for identifying local patterns in raw sensor signals prior to recurrent modelling.
  • Reduced BiLSTM depth: The BiLSTM stack was reduced from four layers to two (one ‘sequence’ mode and one ‘last’ mode). This resulted in a moderate decline in accuracy, with R 2 decreasing by approximately 0.006 compared to the baseline. These findings emphasize the importance of deeper recurrent stacks for modelling multi-scale temporal dependencies in degradation patterns.
  • Single model only: The ensemble strategy was disabled and predictions were made using a single CNN-BiLSTM model. Performance became less stable across batteries, with higher variance, particularly for the challenging B0018 dataset. This demonstrates that ensemble averaging is critical for achieving robust and consistent results by reducing the effect of model-specific fluctuations.
Table 8 shows the detailed ablation metrics for all four batteries. To ensure statistical robustness, each ablation experiment was repeated with five independent random seeds ({42, 142, 242, 342, 442}) under the same hyperparameters, data splits, and ensemble size (M = 6). The reported values in Table 7 are mean ± standard deviation across these runs. To test whether the observed improvements are statistically significant, applied is a paired two-tailed t-test comparing each ablated variant to the full CBATE-Net model. For all metrics, p-values < 0.05 indicate that the performance gains of CBATE-Net are statistically significant. All other training parameters, including learning rate schedule, early-stopping patience, dropout, and augmentation strength were held constant. In summary, the results of the ablation study clearly demonstrate the indispensable nature of the CNN front-end, the depth of the BiLSTM backbone and the ensemble strategy.
The CNN-BiLSTM hybrid architecture is key to extracting and modelling temporal dynamics, and the ensemble ensures stable predictions, especially for challenging battery trajectories. Together, these elements enable the model to achieve the highest accuracy in estimating battery capacity and state of health (SoH). Figure 20 shows that when comparing the four model variants, the complete framework (Full Model) consistently produced the lowest RMSE and MAE values across all batteries. Eliminating the CNN front end led to the largest deterioration in accuracy, while reducing the BiLSTM depth and using a single model instead of the ensemble also increased the errors, although their impact was less severe than removing the CNN component.

7. Conclusions and Future Works

This study presented CBATE-Net, a hybrid deep learning framework designed to accurately and stably estimate battery capacity and state of health. The framework overcomes the limitations of existing approaches by combining localized feature extraction through convolutional layers, sequential modelling with bidirectional recurrent units, selective focus via temporal attention and ensemble averaging. Validation on four NASA dataset lithium-ion cells demonstrated consistent tracking of gradual and abrupt degradation, with the framework maintaining accuracy even near the end of the battery’s life, an area where conventional models often lose reliability. The findings show that accurate capacity prediction can directly lead to reliable state of health estimation and inference of remaining useful life. This capability is essential for the proactive scheduling of maintenance, ensuring the safe operation of energy storage systems and enabling their efficient management.
Future research work should first focus on extending the approach through transfer learning to handle diverse chemistries and operating conditions. A second priority is real-time integration with battery management systems to provide continuous monitoring and decision support in field environments. Beyond this, incorporating electrochemical impedance, environmental variables or photovoltaic generation profiles into multi-modal data fusion could further strengthen robustness. Finally, lightweight model compression techniques are required to enable deployment on embedded hardware with limited resources.

Author Contributions

Conceptualization, F.U.R.; Methodology, F.U.R.; Software, F.U.R.; Validation, F.U.R., C.B. and C.C.; Formal analysis, C.B. and C.C.; Investigation, C.B. and C.C.; Data curation, F.U.R.; Writing—original draft, F.U.R., C.B. and C.C.; Writing—review & editing, F.U.R., C.B. and C.C.; Visualization, C.B. and C.C.; Supervision, C.B. and C.C.; Project administration, C.B. and C.C. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge support from the University of L’Aquila under decree DM-118/2023 PNRR with CODICE DELLA BORSA: 39-411-28-DOT229NLP9-7698.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors gratefully acknowledge the NASA Prognostics Centre of Excellence (PCoE) for providing the dataset that formed the foundation of the experimental evaluation in this study. Also, extends our sincere thanks to the Italian National Doctorate in Photovoltaics for providing resources, mentorship, and a vibrant research community that made this work possible.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AMAttention Mechanism
AMEAbsolute Mean Error
BDTBattery Digital Twin
BESSBattery Energy Storage System
BiLSTMBidirectional Long Short-Term Memory
BMSBattery Management System
CBATE-NetCNN-BiLSTM with Temporal Attention and Ensemble
CCConstant Current
CC-CVConstant Current–Constant Voltage
CNNConvolutional Neural Network
CVConstant Voltage
DAEDenoising Autoencoder
DGDistributed Grid
DLDeep Learning
ECMEquivalent Circuit Modeling
EISElectrochemical Impedance Spectroscopy
EMElectrochemical Model
EoLEnd-of-Life
EVElectric Vehicle
GRUGated Recurrent Unit
GPRGaussian Process Regression
HIHealth Indicators
ICIncremental Capacity
KFKalman Filter
LSTMLong Short-Term Memory
MADMedian Absolute Deviation
MAEMean Absolute Error
MAPEMean Absolute Percentage Error
MCCPsMulti-Channel Charging Profiles
MLMachine Learning
MSEMean Squared Error
MVIPMultiview Information Perception
NNNeural Networks
OCVOpen-Circuit Voltage
PCCPearson Correlation Coefficient
PCTPartial Charging Time
PCoEPrognostics Center of Excellence
PHMPrognostics and Health management
PVPhotovoltaics
RFRandom Forest
RNNRecurrent Neural Network
RMSERoot Mean Squared Error
RULRemaining Useful Life
SGEformerSeasonal and Growth Embedding Transformer
SoCState of Charge
SoHState of Health
SVRSupport Vector Regression
V/I/TVoltage, Current, Temperature
WNNWavelet Neural Network
WLSTMWavelet Long Short-Term Memory

References

  1. Liu, X.; Wu, J.; Zhang, C.; Chen, Z. A method for state of energy estimation of lithium-ion batteries at dynamic currents and temperatures. J. Power Sources 2014, 270, 151–157. [Google Scholar] [CrossRef]
  2. Xi, Z.; Wang, R.; Fu, Y.; Mi, C. Accurate and reliable state of charge estimation of lithium-ion batteries using time-delayed recurrent neural networks through the identification of overexcited neurons. Appl. Energy 2022, 305, 117962. [Google Scholar] [CrossRef]
  3. Bossi, G.; Buccella, C.; Cecati, C.; Simonetti, F.; Damiano, A. A battery-based auxiliary power system for an all-electric aircraft: A novel converter configuration. IEEE Trans. Ind. Appl. 2025, 61, 9493–9505. [Google Scholar] [CrossRef]
  4. Qian, K.; Li, Y.; Zou, Q.; Cao, K.; Li, Z. SoH and RUL estimation for lithium-ion batteries based on partial charging curve features. Energies 2025, 18, 3248. [Google Scholar] [CrossRef]
  5. Ahmed, S.; Ali, A.; Ciocia, A.; D’Angola, A. Technological elements behind the renewable energy community: Current status, existing gap, necessity, and future perspective—Overview. Energies 2024, 17, 3100. [Google Scholar] [CrossRef]
  6. Ahmed, S.; D’Angola, A. Energy storage systems: Scope, technologies, characteristics, progress, challenges, and future suggestions—Renewable energy community perspectives. Energies 2025, 18, 2679. [Google Scholar] [CrossRef]
  7. Wan, T.; Tao, Y.; Qiu, J.; Lai, S. Data-driven hierarchical optimal allocation of battery energy storage system. IEEE Trans. Sustain. Energy 2021, 12, 2097–2109. [Google Scholar] [CrossRef]
  8. Liu, R.; Zhang, D.; Wang, L.; Mi, C.C.; Bauer, P.; Qin, Z. A transformer-enhanced framework for lithium-ion battery capacity estimation using limited imaginary impedance feature. J. Energy Storage 2025, 122, 116313. [Google Scholar] [CrossRef]
  9. Deng, Z.; Yang, L.; Deng, H.; Cai, Y.; Li, D. Polynomial approximation pseudo-two-dimensional battery model for online application in embedded battery management system. Energy 2018, 142, 838–850. [Google Scholar] [CrossRef]
  10. Hosseininasab, S.; Lin, C.; Pischinger, S.; Stapelbroek, M.; Vagnoni, G. State-of-health estimation of lithium-ion batteries for electrified vehicles using a reduced-order electrochemical model. J. Energy Storage 2022, 52, 104684. [Google Scholar] [CrossRef]
  11. Saha, B.; Goebel, K.; Poll, S.; Christophersen, J. Prognostics methods for battery health monitoring using a Bayesian framework. IEEE Trans. Instrum. Meas. 2008, 58, 291–296. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Tang, Q.; Zhang, Y.; Wang, J.; Stimming, U.; Lee, A.A. Identifying degradation patterns of lithium-ion batteries from impedance spectroscopy using machine learning. Nat. Commun. 2020, 11, 1706. [Google Scholar] [CrossRef] [PubMed]
  13. Safari, M.; Delacourt, C. Simulation-based analysis of aging phenomena in a commercial graphite/LiFePO4 cell. J. Electrochem. Soc. 2011, 158, A1436. [Google Scholar] [CrossRef]
  14. Cherif, R.; Hasanov, F.; Pande, A. Riding the energy transition: Oil beyond 2040. Asian Econ. Policy Rev. 2021, 16, 117–137. [Google Scholar] [CrossRef]
  15. Lucaferri, V.; Quercio, M.; Laudani, A.; Riganti Fulginei, F. A review on battery model-based and data-driven methods for battery management systems. Energies 2023, 16, 7807. [Google Scholar] [CrossRef]
  16. Di Fonso, R.; Teodorescu, R.; Cecati, C.; Bharadwaj, P. A battery digital twin from laboratory data using wavelet analysis and neural networks. IEEE Trans. Ind. Inform. 2024, 20, 6889–6899. [Google Scholar] [CrossRef]
  17. Di Fonso, R.; Bharadwaj, P.; Teodorescu, R.; Cecati, C. A battery digital twin based on neural network for testing SoC/SoH algorithms. In Proceedings of the 2022 IEEE 20th International Power Electronics and Motion Control Conference (PEMC), Brasov, Romania, 25–28 September 2022; pp. 655–660. [Google Scholar] [CrossRef]
  18. Gustavsson, J. Energy Storage Technology Comparison: A Knowledge Guide to Simplify Selection of Energy Storage Technology. Online. 2016. Available online: https://www.diva-portal.org/smash/get/diva2:953046/FULLTEXT01.pdf (accessed on 15 March 2025).
  19. Wu, T.; Huang, Y.; Xu, Y.; Jiang, J.; Liu, S.; Li, Z. SoH prediction for lithium-ion battery based on improved support vector regression. Int. J. Green Energy 2023, 20, 227–236. [Google Scholar] [CrossRef]
  20. Liu, D.; Yin, X.; Song, Y.; Liu, W.; Peng, Y. An on-line state of health estimation of lithium-ion battery using unscented particle filter. IEEE Access 2018, 6, 40990–41001. [Google Scholar] [CrossRef]
  21. Zraibi, B.; Okar, C.; Chaoui, H.; Mansouri, M. Remaining useful life assessment for lithium-ion batteries using CNN–LSTM–DNN hybrid method. IEEE Trans. Veh. Technol. 2021, 70, 4252–4261. [Google Scholar] [CrossRef]
  22. Lin, X.; Hao, X.; Liu, Z.; Jia, W. Health-conscious fast charging of Li-ion batteries via a single particle model with aging mechanisms. J. Power Sources 2018, 400, 305–316. [Google Scholar] [CrossRef]
  23. Miao, Y.; Hynan, P.; Von Jouanne, A.; Yokochi, A. Current Li-ion battery technologies in electric vehicles and opportunities for advancements. Energies 2019, 12, 1074. [Google Scholar] [CrossRef]
  24. Morita, Y.; Saito, Y.; Yoshioka, T.; Shiratori, T. Estimation of recoverable resources used in lithium-ion batteries from portable electronic devices in Japan. Resour. Conserv. Recycl. 2021, 175, 105884. [Google Scholar] [CrossRef]
  25. Lashgari, F.; Petkovski, E.; Cristaldi, L. State of health analysis for lithium-ion batteries considering temperature effect. In Proceedings of the 2022 IEEE International Conference on Metrology for Extended Reality, Artificial Intelligence and Neural Engineering (MetroXRAINE), Rome, Italy, 26–28 October 2022; pp. 40–45. [Google Scholar] [CrossRef]
  26. Saha, B.; Goebel, K. Battery Data Set. Online. 2007. Available online: https://ti.arc.nasa.gov/project/prognostic-data-repository (accessed on 22 February 2025).
  27. Choi, Y.; Ryu, S.; Park, K.; Kim, H. Machine learning-based lithium-ion battery capacity estimation exploiting multi-channel charging profiles. IEEE Access 2019, 7, 75143–75152. [Google Scholar] [CrossRef]
  28. Peng, J.; Meng, J.; Chen, D.; Liu, H.; Hao, S.; Sui, X.; Du, X. A review of lithium-ion battery capacity estimation methods for onboard battery management systems: Recent progress and perspectives. Batteries 2022, 8, 229. [Google Scholar] [CrossRef]
  29. Giazitzis, S.; Isse, A.A.; Blasuttigh, N.; Pavan, A.M.; Raimondo, D.M.; Badha, S.; Rosetti, F.; Ogliari, E. TinyML models for SoH estimation of lithium-ion batteries based on electrochemical impedance spectroscopy. J. Power Sources 2025, 653, 237568. [Google Scholar] [CrossRef]
  30. Ahmed, S.; Ali, A.; Qadir, S.A.; Ramunno, D.; D’Angola, A. Modelling of energy management strategies in a PV-based renewable energy community with electric vehicles. World Electr. Veh. J. 2025, 16, 302. [Google Scholar] [CrossRef]
  31. Hasib, S.A.; Islam, S.; Chakrabortty, R.K.; Ryan, M.J.; Saha, D.K.; Ahamed, M.H.; Moyeen, S.I.; Das, S.K.; Ali, M.F.; Islam, M.R.; et al. A comprehensive review of available battery datasets, RUL prediction approaches, and advanced battery management. IEEE Access 2021, 9, 86166–86193. [Google Scholar] [CrossRef]
  32. Marri, I.; Petkovski, E.; Cristaldi, L.; Faifer, M. Comparing machine learning strategies for SoH estimation of lithium-ion batteries using a feature-based approach. Energies 2023, 16, 4423. [Google Scholar] [CrossRef]
  33. Khan, N.; Ullah, F.U.M.; Afnan; Ullah, A.; Lee, M.Y.; Baik, S.W. Batteries state of health estimation via efficient neural networks with multiple channel charging profiles. IEEE Access 2020, 9, 7797–7813. [Google Scholar] [CrossRef]
  34. Van, C.N.; Quang, D.T. Estimation of SoH and internal resistances of lithium-ion battery based on LSTM network. Int. J. Electrochem. Sci. 2023, 18, 100166. [Google Scholar] [CrossRef]
  35. Buvanesan, A.V.; Venugopal, B.M.; Murugan, C.K.; Velusamy, D.H. Deep learning for state of health estimation in EV batteries: A CNN–BiLSTM approach. AIP Adv. 2025, 15, 075213. [Google Scholar] [CrossRef]
  36. Gu, X.; See, K.W.; Li, P.; Shan, K.; Wang, Y.; Zhao, L.; Lim, K.C.; Zhang, N. A novel state-of-health estimation for the lithium-ion battery using a convolutional neural network and transformer model. Energy 2023, 262, 125501. [Google Scholar] [CrossRef]
  37. Modekwe, G.; Al-Wahaibi, S.; Lu, Q. Transformer-based capacity prediction for lithium-ion batteries with data augmentation. IFAC-PapersOnLine 2024, 58, 408–413. [Google Scholar] [CrossRef]
  38. Haraz, A.; Abualsaud, K.; Massoud, A. State-of-health and state-of-charge estimation in electric vehicles batteries: A survey on machine learning approaches. IEEE Access 2024. [Google Scholar] [CrossRef]
  39. Di Fonso, R.; Cecati, C.; Teodorescu, R.; Stroe, D.I.; Bharadwaj, P. Data-driven modeling of Li-ion battery based on the manufacturer specifications and laboratory measurements. IEEE Trans. Ind. Appl. 2025, 61, 3485–3493. [Google Scholar] [CrossRef]
  40. Tao, J.; Wang, S.; Cao, W.; Fernandez, C.; Blaabjerg, F. A comprehensive review of multiple physical and data-driven model fusion methods for accurate lithium-ion battery inner state factor estimation. Batteries 2024, 10, 442. [Google Scholar] [CrossRef]
  41. Sarmah, S.B.; Kalita, P.; Garg, A.; Niu, X.d.; Zhang, X.W.; Peng, X.; Bhattacharjee, D. A review of state of health estimation of energy storage systems: Challenges and possible solutions for futuristic applications of Li-ion battery packs in electric vehicles. J. Electrochem. Energy Convers. Storage 2019, 16, 040801. [Google Scholar] [CrossRef]
  42. Huria, T.; Ceraolo, M.; Gazzarri, J.; Jackey, R. High-fidelity electrical model with thermal dependence for characterization and simulation of high-power lithium battery cells. In Proceedings of the 2012 IEEE International Electric Vehicle Conference (IEVC), Greenville, SC, USA, 4–8 March 2012; pp. 1–8. [Google Scholar] [CrossRef]
  43. Zhao, J.; Han, X.; Ouyang, M.; Burke, A.F. Specialized deep neural networks for battery health prognostics: Opportunities and challenges. J. Energy Chem. 2023, 87, 416–438. [Google Scholar] [CrossRef]
  44. Li, Y.; Zou, C.; Berecibar, M.; Nanini-Maury, E.; Chan, J.C.W.; Van den Bossche, P.; Van Mierlo, J.; Omar, N. Random forest regression for online capacity estimation of lithium-ion batteries. Appl. Energy 2018, 232, 197–210. [Google Scholar] [CrossRef]
  45. Deng, Y.; Du, C.; Ren, Z. A novel method for estimating the state of health of lithium-ion batteries based on physics-informed neural network. Batteries 2025, 11, 49. [Google Scholar] [CrossRef]
  46. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  47. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  48. Luong, M.T.; Pham, H.; Manning, C.D. Effective approaches to attention-based neural machine translation. arXiv 2015, arXiv:1508.04025. [Google Scholar] [CrossRef]
  49. Chorowski, J.; Bahdanau, D.; Cho, K.; Bengio, Y. End-to-end continuous speech recognition using attention-based recurrent NN: First results. arXiv 2014, arXiv:1412.1602. [Google Scholar]
  50. Severson, K.; Attia, P.; Jin, N.; Perkins, N.; Jiang, B.; Yang, Z.; Chen, M.; Aykol, M.; Herring, P.; Fraggedakis, D.; et al. Data-driven prediction of battery cycle life before capacity degradation. Nat. Energy 2019, 4, 383–391. [Google Scholar] [CrossRef]
  51. Wang, L.; Zhang, W.; Li, W.; Ke, X. DGAT: Dynamic Graph Attention-Transformer network for battery state of health multi-step prediction. Energy 2025, 330, 136876. [Google Scholar] [CrossRef]
  52. Rastegarpanah, A.; Contreras, C.A.; Stolkin, R. Hyperparameter-optimized CNN and CNN-LSTM for predicting the remaining useful life of lithium-ion batteries. In Proceedings of the 2023 Eleventh International Conference on Intelligent Computing and Information Systems (ICICIS), Cairo, Egypt, 21–23 November 2023; pp. 110–115. [Google Scholar] [CrossRef]
  53. Zhao, F.M.; Gao, D.X.; Cheng, Y.M.; Yang, Q. Application of state of health estimation and remaining useful life prediction for lithium-ion batteries based on AT-CNN–BiLSTM. Sci. Rep. 2024, 14, 29026. [Google Scholar] [CrossRef]
  54. Zhang, L.; Che, S.; Leng, X.; Lyu, L.; Wang, L.; Yang, B.; Jiang, L. Optimization of CNN–BiLSTM–Attention lithium battery remaining service life prediction based on CPO algorithm. SSRN Electron. J. 2025. [Google Scholar] [CrossRef]
  55. Safavi, V.; Bazmohammadi, N.; Vasquez, J.C.; Guerrero, J.M. Battery state-of-health estimation: A step towards battery digital twins. Electronics 2024, 13, 587. [Google Scholar] [CrossRef]
  56. Yao, Q.; Song, X.; Xie, W. State of health estimation of lithium-ion battery based on CNN–WNN–WLSTM. Complex Intell. Syst. 2024, 10, 2919–2936. [Google Scholar] [CrossRef]
  57. Wei, Y. Prediction of state of health of lithium-ion battery using health index informed attention model. Sensors 2023, 23, 2587. [Google Scholar] [CrossRef]
  58. Tian, Y.; Wen, J.; Yang, Y.; Shi, Y.; Zeng, J. State-of-health prediction of lithium-ion batteries based on CNN–BiLSTM–AM. Batteries 2022, 8, 155. [Google Scholar] [CrossRef]
  59. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar] [CrossRef]
  60. Bai, T.; Wang, H. Convolutional transformer-based multiview information perception framework for lithium-ion battery state-of-health estimation. IEEE Trans. Instrum. Meas. 2023, 72, 2523312. [Google Scholar] [CrossRef]
  61. Zhao, Y.; Behdad, S. State of health estimation of electric vehicle batteries using transformer-based neural network. J. Energy Resour. Technol. 2024, 146, 101703. [Google Scholar] [CrossRef]
  62. Chen, X.; Li, Y.; Liang, X.; Li, F.; Zhang, Z. Battery health state estimation of combined transformer–GRU based on modal decomposition. Energy Storage Sci. Technol. 2023, 12, 2927. [Google Scholar] [CrossRef]
  63. Huang, Y.; Liang, H.; Xu, L. Kalman filter optimize transformer method for state of health prediction on lithium-ion battery. Energy Storage Sci. Technol. 2024, 13, 2791. [Google Scholar] [CrossRef]
  64. Fauzi, M.R.; Yudistira, N.; Mahmudy, W.F. State-of-health prediction of lithium-ion batteries using exponential smoothing transformer with seasonal and growth embedding. IEEE Access 2024, 12, 14659–14670. [Google Scholar] [CrossRef]
Figure 1. Projected global demand for lithium-ion batteries.
Figure 1. Projected global demand for lithium-ion batteries.
Energies 18 05533 g001
Figure 3. Overall workflow of the proposed CBATE-Net framework.
Figure 3. Overall workflow of the proposed CBATE-Net framework.
Energies 18 05533 g003
Figure 6. Data preprocessing and normalization of the proposed framework.
Figure 6. Data preprocessing and normalization of the proposed framework.
Energies 18 05533 g006
Figure 7. Boxplots of z-score distributions of V, I, T, |dV|, |dI|, |dT|.
Figure 7. Boxplots of z-score distributions of V, I, T, |dV|, |dI|, |dT|.
Energies 18 05533 g007
Figure 8. Partial correlations with capacity with asterisks indicate significance levels: *** p < 0.001.
Figure 8. Partial correlations with capacity with asterisks indicate significance levels: *** p < 0.001.
Energies 18 05533 g008
Figure 9. Pearson correlation matrix between engineered features and SoH.
Figure 9. Pearson correlation matrix between engineered features and SoH.
Energies 18 05533 g009
Figure 10. Ensemble stage of CBATE-Net.
Figure 10. Ensemble stage of CBATE-Net.
Energies 18 05533 g010
Figure 20. Error bars indicate ± 1 std over five seeds; all differences relative to ablated variants are statistically significant.
Figure 20. Error bars indicate ± 1 std over five seeds; all differences relative to ablated variants are statistically significant.
Energies 18 05533 g020
Table 1. Real accident cases linked to battery management faults across domains.
Table 1. Real accident cases linked to battery management faults across domains.
DomainYearIncidentCause
Automotive2011Chevrolet Volt delayed fire post-crashDamaged battery thermal failure
Automotive2011Zotye M300 EV taxi fire (The manufacturer was Zotye, a company based in Zhejiang, China, and the fire occurred with a taxi in Hangzhou, China).Battery pack defects (leaks, insulation failure)
Aviation2013Boeing 787 battery fires and grounding (GS Yuasa, a manufacturer based in Kyoto, Japan).Internal short circuits, thermal runaway in BMS
Workshop/Transit2021–2022Street Scooter workshop fire; electric bus/ferry firesFaulty batteries in EVs and energy systems
Maritime2023MV Fremantle Highway fireSuspected EV battery ignition during transport
Industrial2024Hwaseong factory explosion/fireBattery explosions in large-scale storage
Table 8. Ablation study results for all the batteries.
Table 8. Ablation study results for all the batteries.
Model VariantRMSE (Mean ± std)MAE (Mean ± std)p (RMSE)p (MAE)
Full CBATE-Net0.006033 ± 0.000090.004531 ± 0.00007
CNN front-end removed0.008124 ± 0.000110.006492 ± 0.000090.0030.005
BiLSTM depth reduced (3 → 1 layer)0.007841 ± 0.000140.006213 ± 0.000080.0040.007
No attention block0.007692 ± 0.000120.006105 ± 0.000100.0020.004
No ensemble regularization0.008982 ± 0.000100.006854 ± 0.000120.0010.001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rehman, F.U.; Buccella, C.; Cecati, C. CBATE-Net: An Accurate Battery Capacity and State-of-Health (SoH) Estimation Tool for Energy Storage Systems. Energies 2025, 18, 5533. https://doi.org/10.3390/en18205533

AMA Style

Rehman FU, Buccella C, Cecati C. CBATE-Net: An Accurate Battery Capacity and State-of-Health (SoH) Estimation Tool for Energy Storage Systems. Energies. 2025; 18(20):5533. https://doi.org/10.3390/en18205533

Chicago/Turabian Style

Rehman, Fazal Ur, Concettina Buccella, and Carlo Cecati. 2025. "CBATE-Net: An Accurate Battery Capacity and State-of-Health (SoH) Estimation Tool for Energy Storage Systems" Energies 18, no. 20: 5533. https://doi.org/10.3390/en18205533

APA Style

Rehman, F. U., Buccella, C., & Cecati, C. (2025). CBATE-Net: An Accurate Battery Capacity and State-of-Health (SoH) Estimation Tool for Energy Storage Systems. Energies, 18(20), 5533. https://doi.org/10.3390/en18205533

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop