Next Article in Journal
Designing an AI-Supported Framework for Literary Text Adaptation in Primary Classrooms
Previous Article in Journal
Automated Deadlift Techniques Assessment and Classification Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrated Framework of LSTM and Physical-Informed Neural Network for Lithium-Ion Battery Degradation Modeling and Prediction

1
School of Computer and Information Engineering, Tianjin Normal University, Tianjin 300387, China
2
Yangtze Delta Region Institute (Quzhou), University of Electronic Science and Technology of China, Quzhou 324000, China
*
Author to whom correspondence should be addressed.
AI 2025, 6(7), 149; https://doi.org/10.3390/ai6070149
Submission received: 6 June 2025 / Revised: 29 June 2025 / Accepted: 4 July 2025 / Published: 7 July 2025

Abstract

Accurate prediction of the State of Health (SOH) of lithium-ion batteries is essential for ensuring their safe and reliable operation. However, traditional deep learning approaches often suffer from challenges such as overfitting, limited generalization capability, and suboptimal prediction accuracy. To address these issues, this paper proposes a novel framework that combines a Long Short-Term Memory (LSTM) network with a Physics-Informed Neural Network (PINN), referred to as LSTM-PINN, for high-precision SOH estimation. The proposed framework models battery degradation using state-space equations and extracts latent temporal features. These features are further integrated into a Deep Hidden Temporal Physical Module (DeepHTPM), which incorporates physical prior knowledge into the learning process. This integration significantly enhances the model’s ability to accurately capture the complex dynamics of battery degradation. The effectiveness of LSTM-PINN is validated using two publicly available datasets based on graphite cathode materials (NASA and CACLE). Extensive experimental results demonstrate the superior predictive performance of the proposed model, achieving Mean Absolute Errors (MAEs) of just 0.594% and 0.746% and Root Mean Square Errors (RMSEs) of 0.791% and 0.897% on the respective datasets. Our proposed LSTM-PINN framework enables accurate battery aging modeling, advancing lithium-ion battery SOH prediction for industrial applications.

1. Introduction

Lithium-ion batteries have become the dominant choice for rechargeable energy storage, widely deployed across portable electronics, aerospace systems, and electric vehicles. With the rapid expansion of the electric vehicle industry, the demand for high-performance lithium-ion batteries continues to rise. However, battery degradation presents significant challenges: it not only incurs high replacement costs but also has a substantial impact on vehicle safety. As a result, Battery Management Systems (BMS) have been extensively developed to monitor and manage battery health [1]. A central function of BMS is the accurate estimation of the battery’s State of Health (SOH) and the prediction of its Remaining Useful Life (RUL) [2]. Reliable SOH and RUL predictions are critical for evaluating battery condition and ensuring the safe operation of electric power systems.
Substantial research efforts [3,4,5,6] have been devoted to improving the prediction of the State of Health (SOH) and Remaining Useful Life (RUL) of lithium-ion batteries, which remains a challenging task despite significant advancements in the field. The recent progress in machine learning across various domains has sparked growing interest in applying data-driven techniques to battery health estimation. These approaches circumvent the complexities of electrochemical modeling by directly learning degradation patterns from historical operational data. This characteristic makes them particularly appealing for practical applications, especially with the emergence of advanced deep learning architectures such as Long Short-Term Memory (LSTM) networks [7], Convolutional Neural Networks (CNNs), and Recurrent Neural Networks (RNNs) [8]. While these models have demonstrated encouraging results, their generalization capability remains limited. Their performance often depends heavily on the specific features extracted from particular datasets or operating conditions, which restricts their transferability and robustness across diverse battery systems.
To bridge the gap between interpretability and flexibility, Physics-Informed Neural Networks (PINNs) have emerged as a promising hybrid approach. By embedding physical laws and constraints into the learning process, PINNs can train effectively on limited data and demonstrate stronger generalization. Recent studies have demonstrated the effectiveness of PINNs in battery state estimation [9]. For PINN-based prediction approaches, the authors in [10] proposed a physics-informed neural network framework combining a predictor and an estimator. The predictor estimates degradation trends, while the estimator corrects the predictions by replacing hard-to-obtain sensor data. The research [11] showed a physics-informed multitask learning approach for battery aging state estimation, aiming to predict the dynamic variation of lithium-ion concentration in solid particles and the electrolyte. Wang et al. [12] further introduced a PINN-based method that first extracted key features from charging data, then used them to accurately estimate SOH. However, these approaches primarily employ linear neural architectures and do not fully exploit the time-series nature of battery data. As a result, they may struggle with unstable performance, especially when training data is limited.
To overcome these limitations, we propose a novel framework that combines a Long Short-Term Memory (LSTM) network with a Physics-Informed Neural Network (PINN), referred to as LSTM-PINN, for high-precision SOH estimation of lithium-ion batteries. In our approach, an LSTM network is first employed to extract latent temporal features from sequential monitoring data. These features are then incorporated as physical priors into the Deep Hidden Temporal Physical Module (DeepHTPM) of the proposed framework, enabling it to more effectively capture the complex degradation dynamics of batteries. The overall training is constrained by a combination of three loss components: SOH prediction loss, implicit PDE loss, and PDE gradient loss. This multi-loss optimization strategy helps improve both the training stability and the generalization of the model. The main contributions of this work can be summarized as follows:
  • A novel framework is proposed, which integrates LSTM with PINN in a unified manner. The LSTM is employed to extract latent temporal features from battery monitoring sequences, which are then embedded as physical priors into the Deep Hidden Temporal Physical Module (DeepHTPM). This enables more accurate modeling of the degradation dynamics of lithium-ion batteries.
  • We propose a dynamic degradation modeling approach based on the chain rule, which leverages early-stage predictions at each LSTM step to impose additional constraints on the model. This enhances reliability, mitigates overfitting in small datasets, and strengthens generalization performance.
  • We assess the performance of our framework using the NASA and CALCE datasets. The experimental results show that LSTM-PINN consistently surpasses all baseline models, achieving Mean Absolute Errors (MAEs) of 0.594% on the NASA dataset and 0.746% on the CALCE dataset, along with Root Mean Square Errors (RMSEs) of 0.791% and 0.897%, respectively.

2. Related Works

Predicting the state of lithium-ion batteries involves constructing dynamic models that simulate the degradation process, offering valuable insights into internal battery changes over time. Existing methods for battery state prediction are typically categorized into three types: physics-based, electrochemical-based, and data-driven approaches.
Physics-based models simulate the underlying physical and chemical mechanisms governing battery behavior. While these models provide detailed insights into degradation patterns and performance under various operating conditions, they require extensive domain knowledge and precise parameterization, which are often difficult to obtain due to the complex internal reactions within lithium-ion batteries. Similarly, electrochemical-based models describe long-term battery degradation as a function of charge-discharge cycles using predefined mathematical relationships. Although these models can estimate future degradation trajectories, they struggle with real-time tracking of capacity variations and often suffer from low predictive accuracy. For example, Yang et al. [13] proposed a semi-empirical model based on Coulombic efficiency to capture degradation trends more effectively, while research in [14] combined experimental aging data with theoretical models to enhance prediction accuracy. Nevertheless, both approaches rely heavily on domain-specific expertise and require complicated modeling procedures with numerous tunable parameters, which limits their scalability and generalizability.
In contrast, data-driven models have shown significant promise in recent years, particularly for systems where physical mechanisms are poorly understood or too complex to model explicitly. These approaches extract latent patterns from historical monitoring data, without requiring explicit knowledge of the underlying physics. Data-driven methods can be broadly divided into linear regression models and time-series forecasting models [15,16], with the latter becoming the mainstream due to their superior ability to capture long-term trends and cyclical behavior. For instance, Ajami et al. [17] employed temporal neural networks to capture long-range dependencies in sequential battery data. Further enhancements were introduced in [16], where CNN and LSTM networks were combined to jointly exploit spatial and temporal features for capacity estimation. Building upon this, ref. [18] incorporated a self-attention mechanism to improve temporal feature extraction. Additionally, Mohamed et al. [19] proposed a data-driven DLinear model for SOH and RUL estimation, emphasizing the growing role of deep learning in Battery Management Systems (BMS). Despite their success, these models often underperform when applied to small, noisy datasets due to their reliance on large-scale training data and high-quality features.
To overcome these limitations, recent studies have explored PINN for battery state estimation. For example, Dev et al. [20] introduced a PINN-based approach that used an adaptive normalized loss function to estimate battery temperature. Cofré-Martel et al. [21] utilized deep neural networks in conjunction with partial differential equations (PDEs) to model complex physical degradation phenomena. However, such models often require extensive experimental measurements and physical parameters that are challenging to obtain. To address this, DeepHPM [22] was proposed to embed prior physical knowledge into neural networks using Automatic Differentiation (AutoDiff), enabling it to learn both discrete and continuous degradation dynamics—linear or nonlinear. Despite these advances, most PINN-based methods still fall short in fully exploiting the rich temporal features embedded in sequential battery data.
In this work, we propose a novel framework that overcomes these limitations by integrating LSTM with PINNs. Unlike existing methods, our approach extracts hidden temporal features from battery monitoring data and uses them as physical priors to guide the dynamic modeling process. These features are fused within a Deep Hidden Temporal Physical Module, which couples the learning capability of time-series models with the interpretability and constraints of physics-based modeling. This fusion enables the proposed framework to accurately capture the degradation behavior of lithium-ion batteries, even under data-limited conditions.

3. Problem Statement

3.1. Battery Degradation Model

SOH is a crucial concept in battery management systems. The SOH of a battery is a metric that reflects battery aging and is defined as the ratio of the battery’s current capacity to its original rated capacity. Once the SOH drops below 80%, the battery is considered to have reached the end of its first life. Since SOH serves as an indicator of the battery’s condition during each charge-discharge cycle, accurately estimating it is essential [23,24]. The SOH can be expressed as follows:
u n = SOH n = Q t Q Nom × 100 %
where Q N o m is the rated capacity, and  Q t is the available capacity after the t-th charge-discharge cycle. When a battery operates under consistent conditions, the SOH after each cycle can be denoted as follows:
u n = u ( t )
where u ( ) represents the function for predicting the SOH of the battery through t. We can derive the partial differential of the battery’s SOH through u ( ) . That is, the degradation rate of the battery’s SOH is modeled as follows:
d u ( t ) d t = F ( t , u ; θ )
Equation (3) can be interpreted as a parameterized PDE, where F ( · ) is a nonlinear function of t, capacity u, and parameter set θ . This function characterizes the battery’s degradation dynamics. However, due to the difficulty of acquiring electrochemical parameters in practice, this model is challenging to implement. To address this, literature [14] proposes a simplified aging model using regression analysis:
d u ( t ) d t = β [ u ( t ) ]
where β is a linearized aging rate affected by factors such as charge state, discharge depth, and temperature. While this model provides stable predictions, it may not capture the early-cycle degradation patterns due to variability in battery degradation modes and mechanical heterogeneity [25,26].

3.2. A Dynamic Model Based on Temporal Features

To address these limitations, deep learning approaches directly extract degradation patterns from historical data, eliminating the need for explicit electrochemical modeling. However, their generalizability is often limited by the quality of the extracted features. To improve accuracy, we propose a hybrid model that integrates LSTM with PINN, enabling dynamic and generalizable prediction of battery degradation.
To construct a data-driven battery degradation model, we define SOH as a function of the feature vector X N = [ x 1 , x 2 , , x n 1 ] T , where x k is the remaining capacity after the k-th cycle, and  N = { 1 , 2 , , n 1 } . We can extend the function of the relationship between X N and the battery SOH through the Function (2); we have the following:
u n = u ( X N )
the hidden temporal feature extracted from the input at cycle i using LSTM is given as follows:
H n = N N 1 ( X N , H n 1 )
u n = N N 2 ( H n )
Formula (6) describes the prediction of the current hidden state H n using a temporal neural network ( N N 1 ), based on the input X N and the previous hidden state H n 1 . Here, N N 1 represents the temporal network (e.g., LSTM), which captures the sequence dynamics. Formula (7) illustrates the estimation of the final output un from the current hidden state H n using a linear network ( N N 2 ). Combine Formula (5), Formula (6), and Formula (7); we have the following:
u n = u ( X N , H n 1 )
It can be understood from the composition of the differential equation that the partial derivative of SOH with respect to X N is as follows:
u n X N G X N , u n , u n H n 1 | θ
G ( · ) represents the implicit partial differential equation in the deep physical model. where θ is the parameter of nonlinear function G ( · ) , Since the explicit form of G ( · ) is difficult to obtain, we propose the DeepHTPM to implicitly model G ( · ) . We have the following:
u n X N DeepHTPM X N , u n ^ , H n 1 | Φ
where H n 1 = u n H n 1 , and  Φ denotes the trainable parameters of the DeepHTPM.

4. Framework Description

4.1. Framework of LSTM-PINN

LSTM-PINN consists of three main components: a surrogate neural network, the DeepHTPM module, and an automatic differentiator. The surrogate neural network is typically an LSTM in our framework. It takes the input sequence X n along with an initial hidden state H 0 , and processes the sequence step by step over time. At each time step, the network receives the current input x i and the hidden state H n 1 from the previous step. It then computes a new hidden state H n , which is passed to the next time step as part of the input. This iterative process continues until the final prediction u n ^ is produced. In essence, the model updates its hidden state recursively at each step based on the current input and prior hidden state, enabling it to capture temporal dependencies and generate the final output across the entire sequence. u n ^ is given as follows:
u n ^ = L S T M X N , H n 1 ; γ
To effectively capture the discrepancy between the predicted battery capacity and the actual remaining capacity, the surrogate neural network is trained by minimizing a data-driven loss function. This objective function is formulated as follows:
L u = i = 1 N L S T M ( X N , H n 1 ) u n ; γ 2
where u n ^ = L S T M ( X N , H n 1 ; γ ) denotes the output of the surrogate network parameterized by γ , and  u n is the ground truth value.
To approximate the unknown dynamic behavior of battery health evolution in the absence of an explicit physical model, the DeepHTPM architecture (illustrated in Figure 1) is introduced. DeepHTPM is designed as a multi-layer fully connected neural network. Each layer applies a non-linear activation function to its outputs, enabling the model to represent complex non-linear relationships. The hyperbolic tangent function is selected due to its smooth differentiability, which is crucial for gradient-based learning.
Inputs to the DeepHTPM module include the original sequence X N , the predicted capacity u n ^ , and the first-order partial derivatives of the hidden states with respect to their corresponding input variables. These derivatives are computed via an automatic differentiation mechanism.
The DeepHTPM parameters Φ are optimized using a loss based on a physical-inspired constraint, approximating a PDE model. The PDE consistency loss is given by the following:
L f = i = 1 N u n x k DeepHTPM X N , u n ^ , H n 1 | Φ 2
where H n 1 = [ h 1 , , h n 1 ] T comprises the required derivatives of hidden states.
To formalize this relation, we define a residual function f ( X N ; θ , Φ ) as follows:
f ( X N ; θ , Φ ) = u x ( X N ) G ( X N , u n , H n 1 | θ )
where u x = u n x 1 , , u n x n 1 T represents the vector of partial derivatives of u n with respect to each input dimension except x k .
To further enforce the smoothness and consistency of the learned residual function, we define a gradient-based PDE residual loss as follows:
L f X = i = 1 N f X ( X N ; θ , Φ ) 0 2
where f X ( X N ; θ , Φ ) = f x 1 , , f x n 1 T denotes the gradients of the residual function with respect to each input.
The overall training objective is to minimize a weighted sum of the three loss components:
L = λ u L u + λ f L f + λ f X L f X
where λ u , λ f , λ f X are weight coefficients determining the relative importance of each loss term and λ u + λ f + λ f X = 1 . Our experimental results indicate that the weighting of individual loss components plays a critical role in model performance. Given the high cost and inefficiency of manual tuning, we explore the use of Bayesian optimization as an automated approach to adjust these loss weight hyperparameters effectively.

4.2. Bayesian Optimization for Hyperparameter Tuning

Hyperparameter tuning is a critical step in training machine learning models, as the choice of hyperparameters can significantly affect performance. Traditional approaches such as grid search and random search require a large number of function evaluations and become intractable when the hyperparameter space is high-dimensional or when each evaluation (i.e., model training and validation) is costly. Bayesian optimization [27,28] offers an efficient alternative by constructing a probabilistic surrogate model of the objective function and using it to guide the search for optimal hyperparameters. The flowcharts of λ u , λ f , and  λ f x using Bayesian optimization are shown in the Figure 2. The flowchart depicts a standard Bayesian optimization process, commonly used for optimizing expensive black-box functions. The procedure begins with an initialization step, where the process starts by selecting random points to evaluate the objective function. Next, a surrogate model, typically a Gaussian process, is fitted to the data collected from these initial points. The acquisition function is then optimized to determine the next point for evaluation, balancing exploration and exploitation. The objective function is evaluated at this selected point, and the surrogate model is updated with the new data. This iterative process continues, checking for convergence criteria after each update. If the convergence criteria are met, the process stops; otherwise, it repeats the cycle of fitting the surrogate model, optimizing the acquisition function, and updating with new data.

4.2.1. Overview of Bayesian Optimization

Bayesian optimization is a sequential model-based optimization framework designed to minimize (or maximize) an expensive-to-evaluate black-box function, which is defined as follows:
f : X R ,
where X R d is the d-dimensional hyperparameter space. At each iteration t, Bayesian optimization maintains:
  • A surrogate model (often a Gaussian process) that approximates the unknown objective function f ( x ) .
  • An acquisition function α t ( x ) that quantifies the utility of evaluating f at a candidate point x X .
The general loop of Bayesian optimization is as follows:
  • Initialize: Select an initial set of n 0 hyperparameter points { x ( i ) } i = 1 n 0 (e.g., via Latin hypercube sampling) and observe y ( i ) = f x ( i ) .
  • Fit Surrogate: Train the surrogate model (e.g., a Gaussian process) on { ( x ( i ) , y ( i ) ) } i = 1 t 1 to obtain a posterior distribution p f ( x ) | D t 1 , where D t 1 = { ( x ( i ) , y ( i ) ) } i = 1 t 1 .
  • Optimize Acquisition: Find the next query point
    x t = arg max x X α t 1 ( x ) ,
    where α t 1 ( x ) is an acquisition function defined using the posterior predictive mean μ t 1 ( x ) and variance σ t 1 2 ( x ) .
  • Evaluate Objective: Compute y t = f ( x t ) (training and validating the model with hyperparameters x t ).
  • Augment Data: Update D t = D t 1 { ( x t , y t ) } .
  • Repeat: Return to Step 2 until a stopping criterion is met (e.g., budget exhausted or convergence).

4.2.2. Algorithmic Summary

Algorithm 1 summarizes the Bayesian optimization procedure for minimization. In summary, Bayesian optimization provides a principled framework for hyperparameter tuning by iteratively building a probabilistic surrogate of the objective function and using acquisition functions to balance exploration and exploitation. Its ability to find competitive hyperparameter configurations with relatively few evaluations makes it a cornerstone method in Automated Machine Learning (AutoML) pipelines.The training process of LSTM-PINN is shown in Algorithm 2, where S S E ( · ) denotes computing MSE error by using Equations (12), (13) and (15), respectively, D T r i a n = X N , u n n = 1 N is training data.
Algorithm 1 Bayesian optimization for hyperparameter tuning
Require: 
Domain X , initial points { x ( i ) } i = 1 n 0 , acquisition function α ( x ) , evaluation budget T.
  1:
Evaluate y ( i ) f ( x ( i ) ) for i = 1 , , n 0 .
  2:
D n 0 { ( x ( i ) , y ( i ) ) } i = 1 n 0 .
  3:
for  t = n 0 + 1 to T do
  4:
   Fit Gaussian Process to D t 1 .
  5:
   Select next point
x ( t ) arg max x X α x ; D t 1 .
  6:
   Evaluate y ( t ) f x ( t ) .
  7:
   Augment data: D t D t 1 { ( x ( t ) , y ( t ) ) } .
  8:
end for
  9:
Output:  x * = arg min 1 i T y ( i ) .
In summary, LSTM is a type of RNN capable of capturing long-range dependencies in sequential data. It uses gated mechanisms (input gate, forget gate, output gate) to selectively remember or discard information over time, making it effective for modeling battery SOH time series. PINNs embed physical laws, typically represented by differential equations, as constraints within the neural network’s learning process. This incorporation enhances model robustness, particularly in scenarios with limited or noisy data, thereby improving predictive reliability and accuracy. Hence, we propose a unified framework that integrates LSTM with PINN. This integration facilitates more precise modeling of the degradation dynamics of lithium-ion batteries.
Algorithm 2 The training process of LSTM-PINN
Input: 
training data D T r i a n = X N , u n n = 1 N , hyper-parameters.
Output: 
surrogate L S T M X N , H n 1 ; γ , dynamic model G X N , u n , H n 1 | θ .
  1:
initialize θ , γ , λ u , λ f , λ f x
  2:
for each i Search times do
  3:
   for each j epoch do
  4:
      u ^ u X N , H n 1 ; θ
  5:
      H ^ n 1 A u t o D i f f ( u ^ , X n )
  6:
      X ^ N A u t o D i f f ( u ^ , H n 1 )
  7:
      G ^ G X N , u n ^ , H ^ n 1 ; Φ
  8:
      f ^ X ^ N G ^
  9:
      f ^ x A u t o D i f f ( f ^ , X N )
10:
      L ^ u S S E ( u ^ , u )
11:
      L ^ f S S E ( f ^ , 0 )
12:
      L ^ f x S S E ( f x ^ , 0 )
13:
     compute loss L ^ by (18)
14:
     update θ , Φ on loss L ^
15:
   end for
16:
   update λ u , λ f , λ f x on MSE by Bayes
17:
end for
18:
return models

5. Experiments

5.1. Dataset Description

This study utilizes two publicly available datasets: the NASA lithium-ion battery dataset and the CALCE dataset from the University of Maryland. Key battery characteristics of these two datasets are presented in Figure 3.
The NASA dataset [29] provided by the Prognostics Center of Excellence at NASA Ames involves 18,650 Li-ion cells (2 Ah rated capacity) tested under accelerated aging to simulate hybrid electric vehicle conditions. Batteries were charged using a constant current (1.5 A)–constant voltage (4.2 V) protocol and discharged at 2 A until the voltage limit was reached. End of Life (EOL) was defined as a 30% reduction in capacity.
The CALCE dataset [30,31] developed by the Center for Advanced Life Cycle Engineering supports research in battery health monitoring and fault diagnostics. Similar charging and discharging protocols were used, with discharging down to 2.7 V. EOL was likewise determined when capacity fell below 70% of nominal.
We use a sliding window to extract battery features, dividing each cycle sequence into multiple sub-sequences to generate training samples [32,33]. With a window size of n and battery cycles m, the sequence is divided into X = [ X 1 , X 2 , , X i , X m n ] , where X i = [ x i , x i + 1 , x i + 2 , , x i + n 1 ] and label Y i = x i + n . This enhances data utilization and modeling of battery degradation. To mitigate variance among battery samples, SOH prediction uses SOH-based normalization for consistency and accuracy. The leave-one-out cross-validation method is used, where one battery is randomly selected as the test set and the remaining batteries are used as the training set until all batteries are used as the overtest set to train the model as a result of cross validation.

5.2. Experiment Results on the NASA and CALCE Datasets

In order to explore the performance of different models in SOH prediction, two sets of experimental cases were designed, labeled as Experiment A and Experiment B, respectively. Experiment A is for predicting the SOH of the NASA battery, and Experiment B is for predicting the SOH of the CACLE battery. In experiment A, we trained LSTM(M1), CNN-LSTM(M2), CNN-LSTM-AM(M3), swin transformer(M4), LSTM-PINN(M5), TPANet(M6),Bayes LSTM-PINN(M7) model to predict the SOH of the NASA battery. In Experiment B, these six models in Experiment A were trained to predict the SOH of the CACLE battery. LSTM [34] simultaneously considers contextual information, with the aim of enhancing the ability to understand the context. CNN-LSTM [16] combines the local feature extraction of a convolutional network and the time-dependent processing of LSTM, which can better extract data features. CNN-LSTM-AM [35] further enhances the focus on important features by adding attention mechanisms. Swin transformer [36,37] leverages a transformer architecture, enabling efficient computation and scalability while maintaining high accuracy. TPANet [38] consists of a triple parallel architecture combining convolutional neural networks and attention units, where each parallel branch extracts capacity degradation features with enhanced sensitivity to regeneration phenomena and long-term dependencies, and is preceded by an FNN-based window selection mechanism to optimize temporal feature input. LSTM-PINN and Bayes LSTM-PINN are the proposed frameworks in this study, both incorporating a surrogate neural network based on LSTM. The Bayes LSTM-PINN further integrates Bayesian optimization for hyperparameter tuning.
(1) Root Mean Square Error (RMSE). RMSE, which is applied to reveal how tightly the observed data clusters around the predicted values, is a commonly used metric to measure the difference between a model’s predicted values and the actual values. The formula of RMSE is given as follows:
R M S E = 1 N i = 1 N ( u ^ i u i ) 2
where u ^ i denotes the predicted value, u i is the observed value, and N is the number of data samples.
(2) Mean Absolute Error (MAE). MAE is a primary metric used to measure the accuracy of a model. It is measured as the average absolute difference between the predicted values and the true values. The formula of MAE is the following:
M A E = 1 N i = 1 N u ^ i u i

5.2.1. Experiment A

Experiment A verifies the SOH estimation results of the six models on the NASA batteries, with results presented in Table 1. The SOH prediction results and evaluation metrics for each model are depicted in Figure 4 and Figure 5, respectively. From Table 1, it is evident that compared to other models, the prediction errors of LSTM-PINN and Bayes LSTM-PINN are significantly reduced, with both RMSE and MAE showing marked improvement. Notably, LSTM exhibits higher errors, with RMSE and MAE values often exceeding 1.5% and 1.2%, respectively, while CNN-LSTM shows moderate improvements, typically with RMSE around 1.3% and MAE around 1.1%. The swin transformer model performs slightly better, with RMSE and MAE values generally around 1.2% and 1.0%, respectively. The prediction effect of TPANet is second only to ours, reaching the best (RMSE) 0.852% and (MAE) 0.836%. To verify the stability of the framework, a four-fold cross-validation was conducted on each model using this dataset, with experimental results shown in Figure 5. In the cross-validation of the six models, Bayes LSTM-PINN demonstrates the best prediction performance, achieving the lowest RMSE of 0.797% and MAE of 0.746%. As seen in Figure 5, compared to LSTM (with RMSE often above 1.6% and less stable performance), CNN-LSTM, swin transformer, LSTM-PINN, and TPANet, the proposed Bayes LSTM-PINN exhibits higher stability and superior performance in battery SOH prediction, likely due to its enhanced optimization and robust architecture. This performance difference can be attributed to multiple aspects. Firstly, although the traditional LSTM model has the ability of time-series modeling, it has certain limitations in capturing long-range dependencies and nonlinear features, resulting in a relatively high prediction error. CNN-LSTM and CNN-LSTM-AM enhance the ability of local feature extraction by introducing convolutional structures, but it is still difficult to model the complex temporal dynamics during the SOH degradation process. The swin transformer enhances the global modeling ability with the help of the self-attention mechanism and performs relatively better. However, the model has high complexity and strong sensitivity to the amount of data and parameters, which affects its stability. TPANet improves the model performance to a certain extent by introducing the spatio-temporal attention mechanism. However, its architecture still has problems such as parameter redundancy and difficulty in optimization, which limit the improvement of the final accuracy. The Bayes LSTM-PINN we proposed integrates physical priors in the model structure and dynamically adjusts the weights of the multitask loss terms through Bayesian optimization, achieving better generalization ability and lower prediction error. This fully demonstrates the important role of optimization strategies and physical guidance in improving the accuracy of SOH prediction.

5.2.2. Experiment B

In Experiment B, the SOH prediction for CALCE batteries was assessed using seven distinct models: LSTM, CNN-LSTM, CNN-LSTM-AM, swin transformer, LSTM-TPIPNN, TPANet, and the proposed Bayes LSTM-PINN framework. The predictive performance is summarized in Table 2 and visually represented in Figure 6, with detailed metric distributions provided in Figure 7. The LSTM model, a baseline recurrent neural network, captures temporal dependencies with RMSE and MAE values approximating 2.303% and 1.497%, respectively, across various datasets. The CNN-LSTM model integrates convolutional layers with LSTM to extract spatial features prior to temporal analysis, yielding RMSE and MAE values such as 2.85% and 1.827% for CS2_36. The CNN-LSTM-AM variant enhances this approach with an attention mechanism, yet its RMSE and MAE, e.g., 1.954% and 1.369% for CS2_37, remain inferior to the proposed model. The swin transformer employs a transformer architecture with shifted windows, achieving RMSE and MAE around 1.726% and 1.124% for CS2_35.TPANet employs a fusion model of triple parallel attention and CNN, and uses FNN technology to calculate the window size, which has significantly improved the prediction results. The RMSE has reached 0.994% and the MAE has reached 0.678% for CS2_37. The LSTM-TPIPNN combines LSTM with a physics-informed neural network, improving upon standalone LSTM with RMSE and MAE values like 1.266% and 0.964% for CS2_38. Notably, the Bayes LSTM-PINN framework outperforms all models, exhibiting the lowest RMSE of 0.897% and MAE of 0.594% on CS2_38. This superior performance is attributed to the effective temporal feature extraction via LSTM, augmented by physical model guidance, and optimized through Bayesian techniques for loss weight adjustment, thereby enhancing generalization and robustness in SOH prediction.

5.2.3. Ablation Study on Different Loss Functions

This ablation study evaluates the impact of different loss functions on battery SOH prediction, with experimental results are shown in Table 3 and Table 4, and Figure 8 and Figure 9. In the ablation experiment, we can observe that the result of using only L u on the NASA dataset is the worst, with RMSE and MAE being only 3.266% and 2.48%, respectively. After adding L f , the prediction result significantly improves, with RMSE and MAE reaching 1.917% and 1.534% respectively. Moreover, when L f x is added along with L u and L f , this enabled our framework results to reach the optimum, with RMSE and MAE reaching 1.318% and 1.078%, respectively. The results of L u on the CALCE dataset were the worst, with RMSE and MAE being only 3.671% and 3.4%, respectively. After adding L f , RMSE and MAE reached 2.744% and 2.137%, respectively. And when L f x was added along with L u and L f , RMSE and MAE reached 1.672% and 1.206%, respectively. This indicates that the L f and L f x proposed by us have a significant impact on the performance improvement of the framework. This experiment also demonstrates the effectiveness and reliability of the deep hidden physical knowledge extracted by DeepHTPM in improving predictive performance. In addition, combining LSTM with PINN significantly contributes to improving prediction accuracy.

5.3. Impact of Loss Function Weights on Prediction Performance

Since the weights of the loss components play a crucial role in balancing the model’s objectives, we employed Bayesian Optimization to tune the coefficients in the composite loss function defined as L = λ u L u + λ f L f + λ f X L f X . Figure 10 details the influence of the algorithm optimizing the loss weight. The scattered distribution indicates that the parameter space has been widely explored, and the low RMSE area is mainly concentrated within the range of λ u < 0.4 and λ f < 0.2 , suggesting that smaller weight parameters may optimize the model performance. The middle area shows a large fluctuation in RMSE, with the color changing from purple to orange, reflecting the sensitivity of the parameter combination to the error. Higher RMSE (0.9–1.1) is more common in the regions where λ u > 0.7 or λ f > 0.6 , indicating that excessive weights may lead to overfitting or performance degradation. The optimization algorithm indicates that the parameters are fine-tuned within a lower weight range to achieve the best effect, while avoiding excessive parameters to maintain the stability of the model. As shown in Figure 9, the Bayesian Optimization process quickly converged to configurations that achieved significantly lower RMSE. The final experimental results demonstrate the effectiveness of Bayesian optimization in fine-tuning the loss components in the battery performance prediction task to improve the prediction accuracy.

6. Conclusions

Since PINN provides a flexible way to integrate empirical or physical dynamics models with data-driven neural networks, a fusion framework combining an LSTM network and PINN for SOH prediction of the battery is proposed in this paper. LSTM can effectively handle long-term dependencies in time-series data and mine the hidden temporal features of each time-series state. The DeepHTPM module integrates deep hidden physics models with LSTM to accurately capture the degradation dynamics of the battery. A large number of experiments have verified the effectiveness of the proposed framework. Experimental results demonstrate that the proposed framework significantly improves both prediction performance and stability on the lithium-ion battery datasets. In the future, we plan to conduct a time consumption analysis for different models. Moreover, our next research direction will focus on extracting multiple features for various lithium-ion battery types and improving the accuracy of health state prediction based on these extracted features.

Author Contributions

Conceptualization, Y.D. and J.Z.; methodology, Y.D.; software, Y.L.; validation, M.Q., D.N. and J.Z.; formal analysis, Y.D.; investigation, Y.D.; resources, Y.D.; data curation, D.N.; writing—original draft preparation, D.N.; writing—review and editing, M.Q.; visualization, Y.L.; supervision, Y.L.; project administration, Y.D.; funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Municipal Government of Quzhou, grant No.2023D015, No.2023D014; Guiding project of Quzhou Science and Technology Bureau (2023K013 and 2023K016); Tianjin Science and Technology Program Projects, grant No.24YDTPJC00630; and the Tianjin Municipal Education Commission Research Program Project, grant No.2022KJ012.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting the findings of this study are publicly available. Battery datasets were obtained from the NASA Prognostics Center of Excellence (https://www.nasa.gov/content/prognostics-center-of-excellence-data-set-repository) and the CALCE Battery Research Group at the University of Maryland (https://web.calce.umd.edu/batteries/data.htm).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wei, Z.; Li, Y.; Cai, L. Electric Vehicle Charging Scheme for a Park-and-Charge System Considering Battery Degradation Costs. IEEE Trans. Intell. Veh. 2018, 3, 361–373. [Google Scholar] [CrossRef]
  2. Schaut, S.; Arnold, E.; Sawodny, O. Predictive Thermal Management for an Electric Vehicle Powertrain. IEEE Trans. Intell. Veh. 2023, 8, 1957–1970. [Google Scholar] [CrossRef]
  3. Sarrafan, K.; Muttaqi, K.M.; Sutanto, D. Real-Time State-of-Charge Tracking Embedded in the Advanced Driver Assistance System of Electric Vehicles. IEEE Trans. Intell. Veh. 2022, 5, 497–507. [Google Scholar] [CrossRef]
  4. Zhao, S.; Chen, S.; Yang, F.; Ugur, E.; Akin, B.; Wang, H. A Composite Failure Precursor for Condition Monitoring and Remaining Useful Life Prediction of Discrete Power Devices. IEEE Trans. Ind. Inform. 2021, 17, 688–698. [Google Scholar] [CrossRef]
  5. Rincón-Maya, C.; Guevara-Carazas, F.; Hernández-Barajas, F.; Patino-Rodriguez, C.; Usuga-Manco, O. Remaining Useful Life Prediction of Lithium-Ion Battery Using ICC-CNN-LSTM Methodology. Energies 2023, 16, 7081. [Google Scholar] [CrossRef]
  6. Zhao, S.; Blaabjerg, F.; Wang, H. An overview of artificial intelligence applications for power electronics. IEEE Trans. Power Electron. 2021, 36, 4633–4658. [Google Scholar] [CrossRef]
  7. Imen Jarraya, S.B.A.; Alahmed, F. SOH-KLSTM: A hybrid Kolmogorov-Arnold Network and LSTM model for enhanced Lithium-ion battery Health Monitoring. J. Energy Storage 2023, 122, 116541. [Google Scholar] [CrossRef]
  8. Junxiong Chen, W.C.; Zhu, Q. SOC estimation for lithium-ion battery using the LSTM-RNN with extended input and constrained output. Energy 2023, 262, 125375. [Google Scholar] [CrossRef]
  9. Zhao, S.; Peng, Y.; Zhang, Y. Parameter estimation of power electronic converters with physics-informed machine learning. IEEE Trans. Power Electron. 2022, 37, 11567–11578. [Google Scholar] [CrossRef]
  10. Qin, L.; Sun, T.; Sun, X.M.; Xia, W. Managing Battery Performance Degradation Using Physics-Informed Learning Scheme for Multiple Health Indicators. IEEE Trans. Transp. Electrif. 2025, 11, 7309–7321. [Google Scholar] [CrossRef]
  11. Zhang, S.; Liu, Z.; Xu, Y.; Su, H. A Physics-Informed Hybrid Multitask Learning for Lithium-Ion Battery Full-Life Aging Estimation at Early Lifetime. IEEE Trans. Ind. Inform. 2025, 21, 415–424. [Google Scholar] [CrossRef]
  12. Wang, F.; Zhai, Z.; Zhao, Z.; Di, Y.; Chen, X. Physics-informed neural network for lithium-ion battery degradation stable modeling and prognosis. Nat. Commun. 2024, 15, 4332. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, F.; Song, X.; Dong, G. A coulombic efficiencybased model for prognostics and health estimation of lithium-ion batteries. Energy 2019, 171, 1173–1182. [Google Scholar] [CrossRef]
  14. Xu, B.; Oudalov, A.; UIbig, A. Modeling of lithium-ion battery degradation for cell life assessment. IEEE Trans. Smart Grid 2018, 9, 1131–1140. [Google Scholar] [CrossRef]
  15. Wen, P.; Ye, Z.S.; Li, Y. Physics-Informed Neural Networks for Prognostics and Health Management of Lithium-Ion Batteries. IEEE Trans. Intell. Veh. 2024, 9, 2276–2289. [Google Scholar] [CrossRef]
  16. Xu, H.; Wu, L.; Xiong, S. An improved CNN-LSTM model-based state-of-health estimation approach for lithium-ion batteries. Energy 2023, 276, 127585. [Google Scholar] [CrossRef]
  17. Zhun, X.; Zhang, P.; Xie, M. A Joint Long Short-Term Memory and AdaBoost regression approach with application to remaining useful life estimation. Measurement 2021, 170, 108707. [Google Scholar]
  18. Sherkatghanad, Z.; Ghazanfari, A.; Makarenkov, V. A self-attention-based CNN-Bi-LSTM model for accurate state-of-charge estimation of lithium-ion batteries. Energy Storage 2024, 88, 111524. [Google Scholar] [CrossRef]
  19. Massaoudi, M.; Abu-Rub, H.; Ghrayeb, A. Advancing Lithium-Ion Battery Health Prognostics With Deep Learning: A Review and Case Study. IEEE Open J. Ind. Appl. 2024, 5, 43–62. [Google Scholar] [CrossRef]
  20. Sun, H.; Peng, L.; Huang, S. Development of a physics-informed doubly fed cross-residual deep neural network for high-precision magnetic flux leakage defect size estimation. IEee Trans. Ind. Inform. 2022, 18, 1629–1640. [Google Scholar] [CrossRef]
  21. Cofre-Martel, S.; Droguett, E.L.; Modarres, M. Remaining useful life estimation through deep learning partial differential equation models: A framework for degradation dynamics interpretation using latent variables. Shock Vib. 2021, 2021, 9937846. [Google Scholar] [CrossRef]
  22. Raissi, M. Deep hidden physics models: Deep learning of nonlinear partial differential equations. J. Mach. Learn. Res. 2018, 19, 932–955. [Google Scholar]
  23. Kong, J.Z.; Yang, F.; Zhang, X.; Pan, E.; Peng, Z.; Wang, D. Voltagetemperature health feature extraction to improve prognostics and health management of lithium-ion batteries. Energy 2021, 223, 120114. [Google Scholar] [CrossRef]
  24. Connor, W.D.; Wang, Y.; Malikopoulos, A.A.; Advani, S.G.; Prasad, A.K. Impact of connectivity on energy consumption and battery life for electric vehicles. IEEE Trans. Intell. Veh. 2021, 6, 14–23. [Google Scholar] [CrossRef]
  25. Liu, H.; Naqvi, I.H.; Li, F. An analytical model for the CC-CV charge of Li-ion batteries with application to degradation analysis. Energy Storage 2020, 29, 101342. [Google Scholar] [CrossRef]
  26. Severson, K.A.; Attia, P.M.; Jin, N.; Perkins, N. Data-driven prediction of battery cycle life before capacity degradation. Nat. Energy 2019, 4, 383–391. [Google Scholar] [CrossRef]
  27. Solís-Martín, D.; Galán-Páez, J.; Borrego-Díaz, J. CONELPABO: Composite Networks Learning via Parallel Bayesian Optimization to Predict Remaining Useful Life in Predictive Maintenance. Neural Comput. Appl. 2025, 37, 7423–7441. [Google Scholar] [CrossRef]
  28. Onorato, G. Bayesian Optimization for Hyperparameters Tuning in Neural Networks. arXiv 2024, arXiv:2410.21886. [Google Scholar]
  29. Fricke, K.; Nascimento, R.; Corbetta, M.; Kulkarni, C.; Viana, F. An Accelerated Life Testing Dataset for Lithium-Ion Batteries With Constant and Variable Loading Conditions. Int. J. Progn. Health Manag. 2023, 14, 1–12. [Google Scholar] [CrossRef]
  30. Diao, W.; Naqvi, I.H.; Pecht, M. Early Detection of Anomalous Degradation Behavior in Lithium-Ion Batteries. J. Energy Storage 2020, 32, 101710. [Google Scholar] [CrossRef]
  31. Lee, C.; Jo, S.; Kwon, D.; Pecht, M. Capacity-Fading Behavior Analysis for Early Detection of Unhealthy Li-ion Batteries. IEEE Trans. Ind. Electron. 2020, 67, 5693–5701. [Google Scholar] [CrossRef]
  32. Wu, M.; Yue, C.; Zhang, F.; Sun, R.; Tang, J.; Hu, S.; Zhao, N.; Wang, J. State of Health Estimation and Remaining Useful Life Prediction of Lithium-Ion Batteries by Charging Feature Extraction and Ridge Regression. Appl. Sci. 2024, 14, 3153. [Google Scholar] [CrossRef]
  33. Zhou, K.Q.; Qin, Y.; Lau, B.P.L.; Yuen, C.; Adams, S. Lithium-Ion Battery State of Health Estimation Based on Cycle Synchronization Using Dynamic Time Warping. arXiv 2021, arXiv:2109.13448. [Google Scholar]
  34. Mirzaee, H.; Kamrava, S. Estimation of internal states in a Li-ion battery using BiLSTM with Bayesian hyperparameter optimization. Energy Storage 2023, 74, 109522. [Google Scholar] [CrossRef]
  35. Li, W.; Li, Y.; Garg, A.; Gao, L. Enhancing real-time degradation prediction of lithium-ion battery: A digital twin framework with CNN-LSTM-attention model. Energy 2024, 286, 129681. [Google Scholar] [CrossRef]
  36. Zhang, Z.; Li, M.; Wang, X.; Liu, J. Swin Transformer for State of Charge Prediction of Lithium-Ion Batteries in Electric Aircraft. IEEE Trans. Transp. Electrif. 2025, 12, 645–647. [Google Scholar] [CrossRef]
  37. Wang, W.; Zhou, Y.; Zhang, L. MCSFormer: A Multi-Channel Swin Transformer Framework for Remaining Useful Life Prediction of Bearings. arXiv 2025, arXiv:2505.14897. [Google Scholar]
  38. Li, L.; Li, Y.; Mao, R.; Li, Y.; Lu, W.; Zhang, J. TPANet: A novel triple parallel attention network approach for remaining useful life prediction of lithium-ion batteries. Energy 2025, 309, 132890. [Google Scholar] [CrossRef]
Figure 1. The framework of LSTM-PINN.
Figure 1. The framework of LSTM-PINN.
Ai 06 00149 g001
Figure 2. The flowcharts of Bayesian optimization.
Figure 2. The flowcharts of Bayesian optimization.
Ai 06 00149 g002
Figure 3. Feature of SOH.
Figure 3. Feature of SOH.
Ai 06 00149 g003
Figure 4. The prediction result of NASA.
Figure 4. The prediction result of NASA.
Ai 06 00149 g004
Figure 5. RMSE and MAE Values of NASA.
Figure 5. RMSE and MAE Values of NASA.
Ai 06 00149 g005
Figure 6. The prediction result of CALCE.
Figure 6. The prediction result of CALCE.
Ai 06 00149 g006aAi 06 00149 g006b
Figure 7. RMSE and MAE Values of CALCE.
Figure 7. RMSE and MAE Values of CALCE.
Ai 06 00149 g007
Figure 8. The prediction result of CALCE.
Figure 8. The prediction result of CALCE.
Ai 06 00149 g008aAi 06 00149 g008b
Figure 9. The prediction result of CALCE.
Figure 9. The prediction result of CALCE.
Ai 06 00149 g009
Figure 10. The loss weight changes with the iteration of Bayes optimization.
Figure 10. The loss weight changes with the iteration of Bayes optimization.
Ai 06 00149 g010
Table 1. The RMSE and MAE of NASA.
Table 1. The RMSE and MAE of NASA.
M1M2M3M4M5M6M7
B0005RMSE1.6961.3621.150.9061.3180.8830.866
MAE1.0981.0310.8870.8051.0780.8360.746
B0006RMSE2.2741.9051.7211.461.8771.4031.353
MAE1.81.5881.4041.2161.5361.221.194
B0007RMSE1.8371.6811.0060.9061.4970.8520.797
MAE1.5771.3181.1281.0761.3451.0030.901
B0018RMSE21.7541.4471.2011.6871.1240.985
MAE1.7361.5761.2471.0591.5440.9210.822
Table 2. The RMSE and MAE of CALCE.
Table 2. The RMSE and MAE of CALCE.
M1M2M3M4M5M6M7
CS2_35RMSE2.9572.3341.9861.6692.0871.5531.438
MAE2.0241.9151.5651.4531.8801.2390.965
CS2_36RMSE3.0162.851.5861.492.7681.3861.324
MAE2.0851.8271.3661.241.7261.1151.097
CS2_37RMSE1.9511.6861.251.0151.6720.9940.897
MAE1.4551.3690.8990.7911.2390.6780.594
CS2_38RMSE2.3031.7191.5941.2821.7061.1370.964
MAE1.4971.3010.9640.8691.2060.8040.7495
Table 3. Ablation results on NASA.
Table 3. Ablation results on NASA.
ModelMetrics (%)
L u
L f
L f x
B0005RMSE3.4611.9141.318
MAE2.481.5341.078
B0006RMSE3.2662.6481.877
MAE3.0152.2241.536
B0007RMSE4.233.051.497
MAE3.842.4631.345
B0018RMSE3.4362.7351.687
MAE2.8212.231.544
Table 4. Ablation results on CALCE.
Table 4. Ablation results on CALCE.
ModelMetrics (%)
L u
L f
L f x
CS2_35RMSE3.6782.7442.087
MAE3.7942.4671.880
CS2_36RMSE4.6273.5412.768
MAE4.1373.1711.726
CS2_37RMSE3.6712.9471.672
MAE3.42.4891.239
CS2_38RMSE3.982.7641.706
MAE3.3472.1371.206
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, Y.; Zhu, J.; Liu, Y.; Ning, D.; Qin, M. Integrated Framework of LSTM and Physical-Informed Neural Network for Lithium-Ion Battery Degradation Modeling and Prediction. AI 2025, 6, 149. https://doi.org/10.3390/ai6070149

AMA Style

Ding Y, Zhu J, Liu Y, Ning D, Qin M. Integrated Framework of LSTM and Physical-Informed Neural Network for Lithium-Ion Battery Degradation Modeling and Prediction. AI. 2025; 6(7):149. https://doi.org/10.3390/ai6070149

Chicago/Turabian Style

Ding, Yan, Jinqi Zhu, Yang Liu, Dan Ning, and Mingyue Qin. 2025. "Integrated Framework of LSTM and Physical-Informed Neural Network for Lithium-Ion Battery Degradation Modeling and Prediction" AI 6, no. 7: 149. https://doi.org/10.3390/ai6070149

APA Style

Ding, Y., Zhu, J., Liu, Y., Ning, D., & Qin, M. (2025). Integrated Framework of LSTM and Physical-Informed Neural Network for Lithium-Ion Battery Degradation Modeling and Prediction. AI, 6(7), 149. https://doi.org/10.3390/ai6070149

Article Metrics

Back to TopTop