Energy Management Strategy Based on Reinforcement Learning and Frequency Decoupling for Fuel Cell Hybrid Powertrain
Abstract
:1. Introduction
- (I)
- The impact of filter parameters on power flow in the dynamic system is unveiled through frequency decoupling, and optimal filter parameters are determined using the reinforcement learning method. This leads to enhanced operational efficiency, improved output power stability of the fuel cell, and the attainment of superior economic and energy-saving benefits.
- (II)
- An two-layer reinforcement learning optimization framework is established for the iterative optimization of both single-step variables and global variables. This approach addresses the challenge of the reinforcement learning model struggling to assimilate all information from the data. Furthermore, the initial value for power distribution is acquired through frequency decoupling, presenting an intuitive relationship and a favorable trade-off between fuel cell and battery costs.
- (III)
- Tailored to rail transit operational scenarios, the proposed methodology conducts the model training process on a local server rather than real-time training on the train controller. This approach alleviates the burden on computing resources and exhibits a favorable power distribution effect for typical situations characterized by fixed running tracks.
2. Hybrid System Train Modeling
3. Adaptive Energy Management Strategy Combining Frequency Decoupling and Data-Driven Deep Reinforcement Learning
3.1. Frequency Decoupling
3.2. Deep Deterministic Policy Gradient
3.3. Our Energy Management Strategy for Hydrogen Fuel Cell Hybrid Train
4. Simulation Validation and Discussion
4.1. Validation of the Proposed Strategy
4.2. Impact of Different Reward Expressions on the Strategy
4.3. Discussion of the Performance about Fuel Cell and Battery
4.4. Robustness Verification against Different Driving Cycle
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Nomenclature
Abbreviations | |
EMS | Energy Management Strategy |
SOC | State of Charge |
RL | Reinforcement Learning |
DDPG | Deep Deterministic Policy Gradient |
DP | Dynamic programming |
GA | Genetic Algorithm |
PSO | Particle Swarm Optimization |
ECMS | Equivalent Consumption Minimization Strategy |
DRL | Deep Reinforcement Learning |
DQN | Deep Q-Network |
Parameters | |
inertial force | |
aerodynamic drag | |
rolling resistance | |
correction coefficient of rotating mass | |
m | mass of train |
a | acceleration |
air density | |
aerodynamic coefficient | |
A | fronted area |
g | gravity coefficient |
f | rolling resistance coefficient |
power of fuel cell | |
power of battery | |
efficiency of the fuel cell converter | |
efficiency of the battery converter | |
transmission efficiency | |
open circuit voltage | |
N | number of fuel cell monomer |
exchange current | |
dynamic response time | |
internal resistance | |
fuel cell current | |
fuel cell voltage | |
battery output power | |
battery open circuit voltage | |
battery internal resistance | |
B | inverse amplitude of the exponential region |
K | polarization constant |
battery filtration current | |
F | traction force |
radius of wheel | |
motor efficiency | |
cost of hydrogen ($/kg) | |
mass of hydrogen consumed | |
cost of turning on the fuel cell once ($) | |
operating costs of fuel cells | |
operating costs of lithium batteries ($) |
References
- Zou, Z.; Kang, J.; Hu, J. Analysis of Energy Management Strategies For Hydrogen Fuel Cell Hybrid Rail Transit. In Proceedings of the 2023 IEEE PELS Students and Young Professionals Symposium (SYPS), Shanghai, China, 27–28 August 2023; pp. 1–6. [Google Scholar] [CrossRef]
- Sharma, S.; Ghoshal, S.K. Hydrogen the future transportation fuel: From production to applications. Renew. Sustain. Energy Rev. 2015, 43, 1151–1158. [Google Scholar] [CrossRef]
- Abdelrahman, A.S.; Attia, Y.; Woronowicz, K.; Youssef, M.Z. Hybrid Fuel Cell/Battery Rail Car: A Feasibility Study. IEEE Trans. Transp. Electrif. 2016, 2, 493–503. [Google Scholar] [CrossRef]
- Hanley, E.S.; Deane, J.; Gallachóir, B.Ó. The role of hydrogen in low carbon energy futures—A review of existing perspectives. Renew. Sustain. Energy Rev. 2018, 82, 3027–3045. [Google Scholar] [CrossRef]
- Shuguang, L.; Zhenxing, Y. Modeling and Analysis of Contactless Traction Power Supply System for Urban Rail Transit. In Proceedings of the 2020 Chinese Control and Decision Conference (CCDC), Hefei, China, 22–24 August 2020; pp. 5653–5657. [Google Scholar] [CrossRef]
- Weirong, C.; Qingquan, Q.; Qi, L. Research status and development trend of fuel cell hybrid electric train (in Chinese). J. Southwest Jiaotong Univ. 2009, 44, 6. [Google Scholar]
- Yuan, X.H.; Yan, G.D.; Li, H.T.; Liu, X.; Su, C.Q.; Wang, Y.P. Research on energy management strategy of fuel cell–battery–supercapacitor passenger vehicle. Energy Rep. 2022, 8, 1339–1349. [Google Scholar] [CrossRef]
- Sun, Y.; Anwar, M.; Hassan, N.M.S.; Spiryagin, M.; Cole, C. A review of hydrogen technologies and engineering solutions for railway vehicle design and operations. Railw. Eng. Sci. 2021, 29, 212–232. [Google Scholar] [CrossRef]
- Fathy, H.K. Hybrid Electric Vehicles: Energy Management Strategies [Bookshelf]. IEEE Control. Syst. Mag. 2018, 38, 97–98. [Google Scholar] [CrossRef]
- Li, G.; Chen, J.; Zheng, X.; Xiao, C.; Zhou, S. Research on Energy Management Strategy of Hydrogen Fuel Cell Vehicles. In Proceedings of the 2020 Chinese Automation Congress (CAC), Shanghai, China, 6–8 November 2020; pp. 7604–7607. [Google Scholar] [CrossRef]
- Peng, H.; Xie, J. Energy Management Strategy for Plug-In Hybrid Electric Vehicles Based on Genetic-Fuzzy Control Strategy. In Proceedings of the 2017 International Conference on Computer Technology, Electronics and Communication (ICCTEC), Dalian, China, 19–21 December 2017; pp. 1053–1056. [Google Scholar] [CrossRef]
- Mesbahi, T.; Rizoug, N.; Bartholomeüs, P.; Sadoun, R.; Khenfri, F.; Le Moigne, P. Optimal Energy Management for a Li-Ion Battery/Supercapacitor Hybrid Energy Storage System Based on a Particle Swarm Optimization Incorporating Nelder–Mead Simplex Approach. IEEE Trans. Intell. Veh. 2017, 2, 99–110. [Google Scholar] [CrossRef]
- Schmid, R.; Buerger, J.; Bajcinca, N. Energy Management Strategy for Plug-in-Hybrid Electric Vehicles Based on Predictive PMP. IEEE Trans. Control. Syst. Technol. 2021, 29, 2548–2560. [Google Scholar] [CrossRef]
- Xie, S.; Hu, X.; Xin, Z.; Brighton, J. Pontryagin’s Minimum Principle based model predictive control of energy management for a plug-in hybrid electric bus. Appl. Energy 2019, 236, 893–905. [Google Scholar] [CrossRef]
- Li, G.; Or, S.W.; Chan, K.W. Intelligent Energy-Efficient Train Trajectory Optimization Approach Based on Supervised Reinforcement Learning for Urban Rail Transits. IEEE Access 2023, 11, 31508–31521. [Google Scholar] [CrossRef]
- Ning, L.; Zhou, M.; Hou, Z.; Goverde, R.M.; Wang, F.Y.; Dong, H. Deep Deterministic Policy Gradient for High-Speed Train Trajectory Optimization. IEEE Trans. Intell. Transp. Syst. 2022, 23, 11562–11574. [Google Scholar] [CrossRef]
- Gan, J.; Li, S.; Wei, C.; Deng, L.; Tang, X. Intelligent Learning Algorithm and Intelligent Transportation-Based Energy Management Strategies for Hybrid Electric Vehicles: A Review. IEEE Trans. Intell. Transp. Syst. 2023, 24, 10345–10361. [Google Scholar] [CrossRef]
- Liu, T.; Zou, Y.; Liu, D.; Sun, F. Reinforcement Learning of Adaptive Energy Management with Transition Probability for a Hybrid Electric Tracked Vehicle. IEEE Trans. Ind. Electron. 2015, 62, 7837–7846. [Google Scholar] [CrossRef]
- Zhang, Y.; Ma, R.; Zhao, D.; Huangfu, Y.; Liu, W. A Novel Energy Management Strategy Based on Dual Reward Function Q-learning for Fuel Cell Hybrid Electric Vehicle. IEEE Trans. Ind. Electron. 2022, 69, 1537–1547. [Google Scholar] [CrossRef]
- Lin, X.; Zhou, B.; Xia, Y. Online Recursive Power Management Strategy Based on the Reinforcement Learning Algorithm with Cosine Similarity and a Forgetting Factor. IEEE Trans. Ind. Electron. 2021, 68, 5013–5023. [Google Scholar] [CrossRef]
- Cong, J.; Li, B.; Guo, X.; Zhang, R. Energy Management Strategy based on Deep Q-network in the Solar-powered UAV Communications System. In Proceedings of the 2021 IEEE International Conference on Communications Workshops (ICC Workshops), Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Hu, B.; Li, J. An Adaptive Hierarchical Energy Management Strategy for Hybrid Electric Vehicles Combining Heuristic Domain Knowledge and Data-Driven Deep Reinforcement Learning. IEEE Trans. Transp. Electrif. 2022, 8, 3275–3288. [Google Scholar] [CrossRef]
- Wu, J.; He, H.; Peng, J.; Li, Y.; Li, Z. Continuous reinforcement learning of energy management with deep Q network for a power split hybrid electric bus. Appl. Energy 2018, 222, 799–811. [Google Scholar] [CrossRef]
- Han, R.; Lian, R.; He, H.; Han, X. Continuous Reinforcement Learning-Based Energy Management Strategy for Hybrid Electric-Tracked Vehicles. IEEE J. Emerg. Sel. Top. Power Electron. 2023, 11, 19–31. [Google Scholar] [CrossRef]
- Tao, F.; Zhu, L.; Fu, Z.; Si, P.; Sun, L. Frequency Decoupling-Based Energy Management Strategy for Fuel Cell/Battery/Ultracapacitor Hybrid Vehicle Using Fuzzy Control Method. IEEE Access 2020, 8, 166491–166502. [Google Scholar] [CrossRef]
- Huangfu, Y.; Yu, T.; Zhuo, S.; Shi, W.; Zhang, Z. An Optimization Energy Management Strategy Based on Dynamic Programming for Fuel Cell UAV. In Proceedings of the IECON 2021—47th Annual Conference of the IEEE Industrial Electronics Society, Toronto, ON, Canada, 13–16 October 2021; pp. 1–6. [Google Scholar]
- Lian, R.; Peng, J.; Wu, Y.; Tan, H.; Zhang, H. Rule-interposing deep reinforcement learning based energy management strategy for power-split hybrid electric vehicle. Energy 2020, 197, 117297. [Google Scholar] [CrossRef]
- Zhou, Q.; Li, J.; Shuai, B.; Williams, H.; He, Y.; Li, Z.; Xu, H.; Yan, F. Multi-step reinforcement learning for model-free predictive energy management of an electrified off-highway vehicle. Appl. Energy 2019, 255, 113755. [Google Scholar] [CrossRef]
- Zhou, Q.; Zhao, D.; Shuai, B.; Li, Y.; Williams, H.; Xu, H. Knowledge Implementation and Transfer with an Adaptive Learning Network for Real-Time Power Management of the Plug-in Hybrid Vehicle. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 5298–5308. [Google Scholar] [CrossRef]
- Noh, S.; Shim, D.; Jeon, M. Adaptive Sliding-Window Strategy for Vehicle Detection in Highway Environments. IEEE Trans. Intell. Transp. Syst. 2016, 17, 323–335. [Google Scholar] [CrossRef]
- Lhomme, W.; Letrouve, T.; Boulon, L.; Jemei, S.; Bouscayrol, A.; Chauvet, F.; Tournez, F. IEEE VTS Motor Vehicles Challenge 2019—Energy Management of a Dual-Mode Locomotive. In Proceedings of the 2018 IEEE Vehicle Power and Propulsion Conference (VPPC), Chicago, IL, USA, 27–30 August 2018; pp. 1–6. [Google Scholar] [CrossRef]
- Njoya M., S.; Tremblay, O.; Dessaint, L.A. A generic fuel cell model for the simulation of fuel cell vehicles. In Proceedings of the 2009 IEEE Vehicle Power and Propulsion Conference, Dearborn, MI, USA, 7–10 September 2009; pp. 1722–1729. [Google Scholar]
- Tremblay, O.; Dessaint, L.A. Experimental Validation of a Battery Dynamic Model for EV Applications. World Electr. Veh. J. 2009, 3, 289–298. [Google Scholar] [CrossRef]
- Ao, Y.; Laghrouche, S.; Depernet, D.; Chen, K. Proton Exchange Membrane Fuel Cell Prognosis Based on Frequency-Domain Kalman Filter. IEEE Trans. Transp. Electrif. 2021, 7, 2332–2343. [Google Scholar] [CrossRef]
- Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
Locomotive | |
---|---|
Mass | 140 t |
Number of traction drive | 4 |
Efficiency of the electric drives (considered as constant) | 85% |
Gearbox ratio | 4.14 |
Gearbox efficiency | 95% |
Diameter of a wheel | 0.92 |
Fuel Cell | |
Type of fuel cell | PEMFC |
Number of cells in series | 350 |
Number of modules in parallel | 2 |
Voltage range of a cell | 0.3–0.75 V |
Rated power of the fuel cell system | 400 kW |
Battery | |
Type of battery | LiFePO4 |
Total number of cells | 345 |
Rated voltage of a cell | 3.8 V |
Minimal voltage of a cell (charge) | 4.0 V |
Maximal voltage of a cell (discharge) | 2.8 V |
Minimal state of charge | 20% |
Maximal current of a cell | 2 C = 320 A |
Minimal current of a cell | −0.5 C = −80 A |
Algorithm | Brief Introduction | Parameter Set |
---|---|---|
Rule-based | Design power distribution according to expert experience | < 0, = ; = max; = 0; else … |
DDPG | Reinforcement learning | The loss function is shown in Equation (13) |
Frequency Decoupling | low frequency to the fuel cell and high frequency to the battery | The filtering algorithm is Fourier transform, filter frequency is shown in Equations (14) and (16) |
Algorithm | Fuel Consumption (kg) | Terminal SOC (%) | Average Efficiency of Fuel Cell (%) | Total Cost ($) | Training Time (s) |
---|---|---|---|---|---|
Rule-based | 2.78 | 62.40 | 54.78 | 56.49 | 40.12 |
DDPG | 3.84 | 63.8 | 53.38 | 21.45 | 123.56 + 33.54 |
Frequency Decoupling | 5.02 | 69.52 | 51.49 | 34.29 | 66.38 |
Proposed | 2.21 | 60.36 | 55.20 | 18.90 | 206.71 + 45.40 |
Parameter | Value |
---|---|
($/kg) | 5 |
40 | |
159 | |
105 |
Algorithm | Number of Fuel Cell Starts | Standard Deviation |
---|---|---|
Rule-based | 17 | 100.04 |
DDPG | 9 | 95.63 |
Frequency Decoupling | 13 | 88.93 |
Proposed | 6 | 86.32 |
Algorithm | Fuel Consumption (kg) | Terminal SOC (%) | Average Efficiency of Fuel Cell (%) | Total Cost ($) | Training Time (s) |
---|---|---|---|---|---|
Rule-based | 8.85 | 67.24 | 48.53 | 86.14 | 52.14 |
DDPG | 7.70 | 66.87 | 49.86 | 45.96 | 178.55 + 55.54 |
Frequency Decoupling | 12.44 | 69.73 | 44.87 | 56.91 | 105.47 |
Proposed | 5.96 | 65.53 | 51.69 | 37.70 | 252.84 + 68.40 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, H.; Kang, J.; Li, C. Energy Management Strategy Based on Reinforcement Learning and Frequency Decoupling for Fuel Cell Hybrid Powertrain. Energies 2024, 17, 1929. https://doi.org/10.3390/en17081929
Li H, Kang J, Li C. Energy Management Strategy Based on Reinforcement Learning and Frequency Decoupling for Fuel Cell Hybrid Powertrain. Energies. 2024; 17(8):1929. https://doi.org/10.3390/en17081929
Chicago/Turabian StyleLi, Hongzhe, Jinsong Kang, and Cheng Li. 2024. "Energy Management Strategy Based on Reinforcement Learning and Frequency Decoupling for Fuel Cell Hybrid Powertrain" Energies 17, no. 8: 1929. https://doi.org/10.3390/en17081929
APA StyleLi, H., Kang, J., & Li, C. (2024). Energy Management Strategy Based on Reinforcement Learning and Frequency Decoupling for Fuel Cell Hybrid Powertrain. Energies, 17(8), 1929. https://doi.org/10.3390/en17081929