Joint Concern over Battery Health and Thermal Degradation in the Cruise Control of Intelligently Connected Electric Vehicles Using a Model-Assisted DRL Approach
Abstract
:1. Introduction
- A novel adaptive cruise control framework which actively incorporates the DRL algorithm and an electrothermal and aging battery model is proposed for autonomous EVs, by which accurate, battery model-assisted, DRL training as well as model-free online control can be realized, thereby greatly improving optimal output power comprehensively while ensuring real-time control performance.
- An advanced continuous DRL strategy based on DDPG was creatively adopted to intelligently optimize power allocation in autonomous EVs in this study, which offers accelerated convergence and improved optimization performance.
- The battery’s thermal safety and degradation considering thermal effects were creatively involved in the proposed framework through the establishment of a joint electrothermal and aging model, which realizes the accurate evaluation of a battery’s aging and the heating effectiveness of agent actors so as to provide effective guidance for DRL training.
2. System Modeling
2.1. Dynamic Modeling of an Electric Vehicle
2.2. Modeling of the Onboard Power Battery
- (1)
- Battery Energy Consumption Modeling
- (2)
- Electrothermal Modeling for LIBs
- (3)
- Aging Model of LIBs
3. Thermal- and Health-Constrained Velocity Optimization
3.1. Problem Formulation
3.2. Fundamentals of the DDPG Algorithm
Algorithm 1. Procedures of the DDPG Algorithm |
1: Initialization: critic network and actor network with weights and , target network and with weights , , memory pool R, a random process N for action exploration |
2: for episode = 1:M do |
3: get initial states: |
4: for t = 1, T do |
5: Select action according to the current policy and exploration noise |
6: Execute action , observe reward and new states |
7: Store transition in R |
8: Sample a minibatch of transitions from R with priority experience replay |
9: Set |
10: Update critic by minimizing the loss: |
11: Update the actor policy using the sampled policy gradient: |
12: Update the target networks: |
13: end for |
14: end for |
4. Results and Discussion
4.1. Conditions for Validation
4.2. Validation of the Training Process
4.3. Validation of Speed and Safety Distance
4.4. Validation of Temperature and Degradation Control
4.5. Validation of Overall Driving Cost Optimization
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Li, L.; Yang, C.; Zhang, Y.H.; Zhang, L.P.; Song, J. Correctional DP-Based Energy Management Strategy of Plug-In Hybrid Electric Bus for City-Bus Route. IEEE Trans. Veh. Technol. 2015, 64, 2792–2803. [Google Scholar] [CrossRef]
- Kim, N.; Cha, S.; Peng, H. Optimal Control of Hybrid Electric Vehicles Based on Pontryagin’s Minimum Principle. IEEE Trans. Control Syst. Technol. 2011, 19, 1279–1287. [Google Scholar]
- Hou, C.; Ouyang, M.G.; Xu, L.F.; Wang, H.W. Approximate Pontryagin’s minimum principle applied to the energy management of plug-in hybrid electric vehicles. Appl. Energy 2014, 115, 174–189. [Google Scholar] [CrossRef]
- Wang, H.; Huang, Y.J.; Khajepour, A.; He, H.W.; Cao, D.P. A novel energy management for hybrid off-road vehicles without future driving cycles as a priori. Energy 2017, 133, 929–940. [Google Scholar] [CrossRef]
- Xie, S.B.; Hu, X.S.; Xin, Z.K.; Li, L. Time-Efficient Stochastic Model Predictive Energy Management for a Plug-In Hybrid Electric Bus With an Adaptive Reference State-of-Charge Advisory. IEEE Trans. Veh. Technol. 2018, 67, 5671–5682. [Google Scholar] [CrossRef]
- Amin; Trilaksono, B.R.; Rohman, A.S.; Dronkers, C.J.; Ortega, R.; Sasongko, A. Energy Management of Fuel Cell/Battery/Supercapacitor Hybrid Power Sources Using Model Predictive Control. IEEE Trans. Ind. Inform. 2014, 10, 1992–2002. [Google Scholar] [CrossRef]
- Zhang, Q.; Deng, W.W.; Li, G. Stochastic Control of Predictive Power Management for Battery/Supercapacitor Hybrid Energy Storage Systems of Electric Vehicles. IEEE Trans. Ind. Inform. 2018, 14, 3023–3030. [Google Scholar] [CrossRef]
- Hamednia, A.; Sharma, N.K.; Murgovski, N.; Fredriksson, J. Computationally Efficient Algorithm for Eco-Driving Over Long Look-Ahead Horizons. IEEE Trans. Intell. Transp. Syst. 2022, 23, 6556–6570. [Google Scholar] [CrossRef]
- Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing atari with deep reinforcement learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
- Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
- Oh, J.; Chockalingam, V.; Lee, H. Control of memory, active perception, and action in minecraft. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 2790–2799. [Google Scholar]
- Schulman, J.; Levine, S.; Abbeel, P.; Jordan, M.; Moritz, P. Trust region policy optimization. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 1889–1897. [Google Scholar]
- Kendall, A.; Hawke, J.; Janz, D.; Mazur, P.; Reda, D.; Allen, J.-M.; Lam, V.-D.; Bewley, A.; Shah, A. Learning to drive in a day. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 8248–8254. [Google Scholar]
- Lee, H.; Kim, K.; Kim, N.; Cha, S.W. Energy efficient speed planning of electric vehicles for car-following scenario using model-based reinforcement learning. Appl. Energy 2022, 313, 118460. [Google Scholar] [CrossRef]
- Guo, Q.Q.; Angah, O.; Liu, Z.J.; Ban, X.G. Hybrid deep reinforcement learning based eco-driving for low-level connected and automated vehicles along signalized corridors. Transp. Res. Part C Emerg. Technol. 2021, 124, 102980. [Google Scholar] [CrossRef]
- Shi, J.; Qiao, F.; Li, Q.; Yu, L.; Hu, Y. Application and evaluation of the reinforcement learning approach to eco-driving at intersections under infrastructure-to-vehicle communications. Transp. Res. Rec. 2018, 2672, 89–98. [Google Scholar] [CrossRef]
- Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
- Hao, P.; Wei, Z.; Bai, Z.; Barth, M.J. Developing an Adaptive Strategy for Connected Eco-Driving under Uncertain Traffic and Signal Conditions; National Center for Sustainable Transportation: Davis, CA, USA, 2020. [Google Scholar]
- Wu, J.; Song, Z.; Lv, C. Deep Reinforcement Learning based Energy-efficient Decision-making for Autonomous Electric Vehicle in Dynamic Traffic Environments. IEEE Trans. Transp. Electrif. 2023, 10, 875–887. [Google Scholar] [CrossRef]
- Wu, J.; Zhou, Y.; Yang, H.; Huang, Z.; Lv, C. Human-guided reinforcement learning with sim-to-real transfer for autonomous navigation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 14745–14759. [Google Scholar] [CrossRef] [PubMed]
- Lin, X.; Wang, Y.Z.; Bogdan, P.; Chang, N.; Pedram, M. Reinforcement Learning Based Power Management for Hybrid Electric Vehicles. In Proceedings of the 33rd IEEE/ACM International Conference on Computer-Aided Design (ICCAD), San Jose, CA, USA, 2–6 November 2014; pp. 32–38. [Google Scholar]
- Liu, X.; Liu, Y.W.; Chen, Y.; Hanzo, L. Enhancing the Fuel-Economy of V2I-Assisted Autonomous Driving: A Reinforcement Learning Approach. IEEE Trans. Veh. Technol. 2020, 69, 8329–8342. [Google Scholar] [CrossRef]
- Li, G.Q.; Gorges, D. Ecological Adaptive Cruise Control for Vehicles With Step-Gear Transmission Based on Reinforcement Learning. IEEE Trans. Intell. Transp. Syst. 2020, 21, 4895–4905. [Google Scholar] [CrossRef]
- Xu, S.B.; Peng, H. Design and Comparison of Fuel-Saving Speed Planning Algorithms for Automated Vehicles. IEEE Access 2018, 6, 9070–9080. [Google Scholar] [CrossRef]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Tan, H.; Zhang, H.; Peng, J.; Jiang, Z.; Wu, Y. Energy management of hybrid electric bus based on deep reinforcement learning in continuous state and action space. Energy Convers. Manag. 2019, 195, 548–560. [Google Scholar] [CrossRef]
- Wu, Y.; Tan, H.; Peng, J.; Zhang, H.; He, H. Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus. Appl. Energy 2019, 247, 454–466. [Google Scholar] [CrossRef]
- Liu, L.; Guan, P. Phase-Field Modeling of Solid Electrolyte Interphase (SEI) Evolution: Considering Cracking and Dissolution during Battery Cycling. ECS Trans. 2019, 89, 101. [Google Scholar] [CrossRef]
- Shi, M.; He, H.W.; Li, J.W.; Han, M.; Jia, C.C. Multi-objective tradeoff optimization of predictive adaptive cruising control for autonomous electric buses: A cyber-physical-energy system approach. Appl. Energy 2021, 300, 117385. [Google Scholar] [CrossRef]
- Lin, X.F.; Perez, H.E.; Mohan, S.; Siegel, J.B.; Stefanopoulou, A.G.; Ding, Y.; Castanier, M.P. A lumped-parameter electro-thermal model for cylindrical batteries. J. Power Sources 2014, 257, 1–11. [Google Scholar] [CrossRef]
- Wu, J.D.; Wei, Z.B.; Li, W.H.; Wang, Y.; Li, Y.W.; Sauer, D.U. Battery Thermal- and Health-Constrained Energy Management for Hybrid Electric Bus Based on Soft Actor-Critic DRL Algorithm. IEEE Trans. Ind. Inform. 2021, 17, 3751–3761. [Google Scholar] [CrossRef]
c | 0.5 | 2 | 6 | 10 |
B(c) | 31,529 | 21,701 | 12,925 | 15,493 |
Proposed | Thermal- and Health-Neglecting | |
---|---|---|
Electricity consumption of LIBs | 2632 kW·h | 2565 kW·h |
Battery degradation | 0.78% | 1.08% |
Electricity consumption (CNY 0.96/kW·h) | CNY 2526.72 | CNY 2462.4 |
Battery degradation cost (CNY 69,800/LIB pack) | CNY 2722.2 | CNY 3769.2 |
Overall driving cost | CNY 5248.92 | CNY 6231.6 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Cheng, X.; Chen, X. Joint Concern over Battery Health and Thermal Degradation in the Cruise Control of Intelligently Connected Electric Vehicles Using a Model-Assisted DRL Approach. Batteries 2024, 10, 226. https://doi.org/10.3390/batteries10070226
Cheng X, Chen X. Joint Concern over Battery Health and Thermal Degradation in the Cruise Control of Intelligently Connected Electric Vehicles Using a Model-Assisted DRL Approach. Batteries. 2024; 10(7):226. https://doi.org/10.3390/batteries10070226
Chicago/Turabian StyleCheng, Xiangheng, and Xin Chen. 2024. "Joint Concern over Battery Health and Thermal Degradation in the Cruise Control of Intelligently Connected Electric Vehicles Using a Model-Assisted DRL Approach" Batteries 10, no. 7: 226. https://doi.org/10.3390/batteries10070226
APA StyleCheng, X., & Chen, X. (2024). Joint Concern over Battery Health and Thermal Degradation in the Cruise Control of Intelligently Connected Electric Vehicles Using a Model-Assisted DRL Approach. Batteries, 10(7), 226. https://doi.org/10.3390/batteries10070226