Improvement of Refrigeration Efficiency by Combining Reinforcement Learning with a Coarse Model
Abstract
:1. Background and Motivation
- This is an online control strategy with data-driven control and learning simultaneously which reduces the requirement of model accuracy.
- It has the ability to get the optimal action by exploration and exploitation to achieve a better conversion efficiency which fits the current operating conditions.
- The risk on the system in the learning process, which is complex and unknown in advance is prevented mostly by a hierarchical control of the process level, which makes use of the existing knowledge of refrigeration and the loop level, which implements the tracking control.
2. Preliminaries
2.1. Aim of the Research
2.2. Principle of Compression Refrigerating System
2.2.1. Principle of the Evaporator
2.2.2. Principle of the Condenser
2.2.3. Principle of the Compressor
2.2.4. Principle of the Electronic Expansion Valve
3. Reinforcement Learning
3.1. Reinforcement Learning
3.2. Q-Algorithm
4. The Proposed Approach
4.1. Process Level
4.1.1. States Vector, Action Vector, and Cost Function
4.1.2. Exploitation and Exploration
4.1.3. Procedure of the Proposed Algorithm
- Step 1:
- Select an action randomly.
- Step 2:
- Receive immediate reward according to formula (20).
- Step 3:
- Get the new state according to the coarse model and compute value function according to formula (10).
- Step 4:
- Step 4: Renew the reward value based on current state
- Step 5:
- Adjust a new control vector with -greedy method according to (21).
- Step 6:
- Repeat step 1 to step 5 until it is convergent.
- Step 7:
- Get the optimal according to the best reward
- Step 8:
- Apply to the refrigeration system as the reference of the loop level and then get under the control .
- Step 9:
- Replace by and go to step 1.
4.2. Loop Level
5. Case Studies
5.1. Effect of Model Accuracy on Performance
5.2. Comparison with the Conventional Approach
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Consumption & Efficiency. Available online: https://www.eia.gov/consumption (accessed on 1 October 2019).
- Abdul, A.; Farrokh, J.S. Theory and applications of HVAC control systems-a review of model predictive control (MPC). Build. Environ. 2014, 72, 343–355. [Google Scholar]
- Youssef, A.; Caballero, N.; Aerts, J.M. Model-Based Monitoring of Occupant’s Thermal State for Adaptive HVAC Predictive Controlling. Processes 2019, 7, 720. [Google Scholar] [CrossRef] [Green Version]
- Wang, S.; Ma, Z. Supervisory and optimal control of building HVAC systems: A review. HVAC R Res. 2008, 14, 3–32. [Google Scholar] [CrossRef]
- Naidu, D.S.; Rieger, C.G. Advanced control strategies for heating ventilation air conditioning and refrigeration systems an overview part I hard control. HVAC R Res. 2011, 17, 2–21. [Google Scholar] [CrossRef]
- Shang, Y. Critical Stability Analysis, Optimization and Control of a Compression Refrigeration System. Ph.D. Thesis, Tianjin University, Tianjin, China, 2016. [Google Scholar]
- Sutton, R.; Barto, A. Reinforcement Learning: An Introduction; The MIT Press Cambridge: Cambridge, UK, 2005. [Google Scholar]
- Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; Van Den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484. [Google Scholar] [CrossRef]
- Vamvoudakis, K.G.; Lewis, F.L. Online actor-critic algorithm to solve the continuous-time infinite horizon optimal control problem. Automatica 2010, 46, 878–888. [Google Scholar] [CrossRef]
- Jens, K.; Andrew, B.J.; Jan, P. Reinforcement learning in robotics: A survey. Int. J. Robot. Res. 2013, 32, 1238–1274. [Google Scholar]
- Zhang, D.; Gao, Z. Reinforcement learning-based fault-tolerant control with application to flux cored wire system. Meas. Control. 2018, 51, 349–359. [Google Scholar] [CrossRef] [Green Version]
- Zhang, D.; Lin, Z.; Gao, Z. Reinforcement-learning based fault-tolerant control. In Proceedings of the 15th International Conference on Industrial Informatics (INDIN), Emden, Germany, 24–26 November 2017. [Google Scholar]
- Zhang, D.; Lin, Z.; Gao, Z. A novel fault detection with minimizing the noise-signal ratio using reinforcement learning. Sensors 2018, 18, 3087. [Google Scholar] [CrossRef] [Green Version]
- Baeksuk, C.; Jooyoung, P.; Daehie, H. Tunnel ventilation controller design using an RLS-based natural actor-critic algorithm. Int. J. Precis. Eng. Manuf. 2010, 11, 829–838. [Google Scholar]
- Fazenda, P.; Veeramachaneni, K.; Lima, P.; O’Reilly, U.M. Using reinforcement learning to optimize occupant comfort and energy usage in HVAC systems. J. Ambient Intell. Smart Environ. 2014, 6, 675–690. [Google Scholar] [CrossRef]
- Li, B.; Xia, L. A multi-grid reinforcement learning method for energy conservation and comfort of HVAC in buildings. In Proceedings of the IEEE International Conference on Automation Science and Engineering, Gothenburg, Sweden, 24–28 August 2015. [Google Scholar]
- Enda, B.; Stephen, L. Autonomous HVAC control, a reinforcement learning approach. In Proceedings of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD), Porto, Portugal, 7–11 September 2015; Volume 9286. [Google Scholar]
- Gracia Cuesta, A.D.; Fernàndez Camon, C.; Castell, A.; Mateu Piñol, C.; Cabeza, L.F. Control of a PCM ventilated facade using reinforcement learning techniques. Energy Build. (SI) 2015, 106, 234–242. [Google Scholar] [CrossRef] [Green Version]
- Yang, L.; Nagy, Z.; Goffin, P.; Schlueter, A. Reinforcement learning for optimal control of low exergy buildings. Appl. Energy 2015, 156, 577–586. [Google Scholar] [CrossRef]
- Cheng, Z.; Zhao, Q.; Wang, F.; Jiang, Y.; Xia, L.; Ding, J. Satisfaction based Q-learning for integrated lighting and blind control. Energy Build. 2016, 127, 43–55. [Google Scholar] [CrossRef]
- Wang, Y.; Velswamy, K.; Huang, B. A long-short term memory recurrent neural network based reinforcement learning controller for office heating ventilation and air conditioning systems. Processes 2017, 5, 46. [Google Scholar] [CrossRef] [Green Version]
- Chen, Y.; Norford, L.K.; Samuelson, H.W.; Malkawi, A. Optimal control of HVAC and window systems for natural ventilation through reinforcement learning. Energy Build. 2018, 169, 195–205. [Google Scholar] [CrossRef]
- Wei, T.; Wang, Y.; Zhu, Q. Deep reinforcement learning for building HVAC control. In Proceedings of the 54th ACM/EDAC/IEEE Design Automation Conference (DAC), Austin, TX, USA, 18–22 June 2017. [Google Scholar]
- Zhang, Z.; Lam, K.P. Practical implementation and evaluation of deep reinforcement learning control for a radiant heating system. In Proceedings of the 5th Conference on Systems for Built Environments, Shenzhen, China, 7–8 November 2018. [Google Scholar]
- Farahmand, A.M.; Nabi, S.; Grover, P.; Nikovski, D.N. Learning to control partial differential equations: Regularized fitted Q-iteration approach. In Proceedings of the 55th IEEE Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016. [Google Scholar]
- Wang, Y.; Velswamy, K.; Huang, B. A novel approach to feedback control with deep reinforcement learning. In Proceedings of the 10th IFAC Symposium on Advanced Control of Chemical Processes (ADCHEM), Shenyang, China, 25–27 July 2018. [Google Scholar]
- Ruelens, F.; Iacovella, S.; Claessens, B.; Belmans, R. Learning agent for a heat-pump thermostat with a set-back strategy using model-free reinforcement learning. Energies 2015, 8, 8300–8318. [Google Scholar] [CrossRef] [Green Version]
- Valladares, W.; Galindo, M.; Gutierrez, J.; Wu, W.C.; Liao, K.K.; Liao, J.C.; Lu, K.C.; Wang, C.C. Energy optimization associated with thermal comfort and indoor air control via a deep reinforcement learning algorithm. Build. Environ. 2019, 155, 105–117. [Google Scholar] [CrossRef]
- Kaelbling, L.K.; Littman, M.L.; Moore, A.W. Reinforcement learning: A survey. J. Artif. Intell. Res. 1996, 4, 237–285. [Google Scholar] [CrossRef] [Green Version]
- Ma, Z.; Yao, Y. Air Conditioning Design of Civil Buildings, 3rd ed.; Chemical Industry Press: Beijing, China, 2015. [Google Scholar]
- Tahat, M.A.; Ibrahim, G.A.; Probert, S.D. Performance instability of a refrigerator with its evaporator controlled by a thermostatic expansion-valve. Appl. Energy 2001, 70, 233–249. [Google Scholar] [CrossRef]
- Aprea, C.; Mastrullo, R.; Renno, C. Performance of thermostatic and electronic expansion valves controlling the compressor. Int. J. Energy Res. 2006, 30, 1313–1322. [Google Scholar] [CrossRef]
- Xue, H. Active Disturbance Rejection Control of Compression Refrigeration System. Master’s Thesis, Tianjin University, Tianjin, China, 2016. [Google Scholar]
- Watldns, C.J.C.H. Learning from Delayed Rewards. Ph.D. Thesis, University of Cambridge, Cambridge, UK, 1989. [Google Scholar]
- Watkins, C.J.C.H.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
- Zhang, H. Research on Internal Model Control Strategy of Compression Refrigerating System. Master’s Thesis, Tianjin University, Tianjin, China, 2013. [Google Scholar]
- Xiao, D. Theory of System Identification with Application; Tsinghua University Press: Beijing, China, 2014. [Google Scholar]
Letter | Name | Subscript | Name | Subscript | Name |
---|---|---|---|---|---|
T | Temperature, °C | k | Compressor | ei2 | Inner of superheated zone of evaporator |
M | Mass, Kg | c | Condenser | er2 | Refrigerant in superheated zone of evaporator |
ρ | Density, Kg/m3 | e | evaporator | ew2 | Tube wall of superheated zone of evaporator |
N | Power, W | val | Expansion valve | ei1 | Inner of two-phase zone of evaporator |
η | Efficiency | r | refrigerant | er1 | Refrigerant in two-phase zone of evaporator |
K | Heat transfer coefficient, W/m2·°C | w | Tube wall, water | ew1 | Tube wall of two-phase zone of evaporator |
α | Heat transfer coefficient, W/m2·°C | i | Input/inner | ci2 | Inner of superheated zone of condenser |
μ | Dynamic viscosity, Pa·s | o | Output/outer | cr2 | Refrigerant in superheated zone of condenser |
A | Area, m2 | sc | supercool | cw2 | Tube wall of superheated zone of condenser |
m | Mass flow, Kg·s | sh | superhot | ci1 | Inner of two-phase zone of condenser |
C | Specific heat capacity, KJ/Kg·°C | th | Theoretical value | cr1 | Refrigerant in two-phase zone of condenser |
ν | Specific volume, m3/Kg | g | Gas | cw1 | Tube wall of two-phase zone of condenser |
h | Specific enthalphy, KJ/Kg | l | Liquid | ci3 | Tube wall of supercool zone of condenser |
No. | Name of Devices | Specifications |
---|---|---|
1 | Compressor | Semi-closed piston compressor, rated speed 1450 rpm, 7.5 Kw/10 HP, rated discharge 38.25 m3/h, rated refrigeration capacity 30 KW |
2 | Frequency Converter for Compressor | Capacity 11 KVA, rated current 17 A, rated voltage 380 V, frequency 0–50 Hz, V/F signal 0–10 V |
3 | Condenser | Borehole heat exchangers, heat exchange 35 Kw, heat transfer area 3.4 m2, heat transfer coefficient K = 970 W/m2·°C |
4 | Evaporator | Borehole heat exchangers, heat exchange 30 kW, heat transfer area 3.02 m2, heat transfer coefficient K = 1455 W/m2·°C |
5 | Electronic Expansion Valve | Flow coefficient Kvs = 0.63 m3/h, refrigerating capacity 74 kW, caliber DN15, AC24 V, control signal 0–10 V |
6 | Cooling Water Pump | Multistage centrifugal pump: Rated power 2.2 kW, rated speed 2840 rpm, lift 32 m, flow 8.4 m3/h |
7 | Frequency Converter for Cooling Water Pump | Rated voltage 380 V, frequency 0–50 Hz, V/F signal 0–10 V |
8 | Chilled Water Pump | Single-stage centrifugal pump, rated power 1.5 kW, rated speed 2840 rpm, lift 27.4 m, flow 7.8 m3/h |
9 | Frequency Converter for Refrigerated Water Pump | Rated voltage 380 V, frequency 0-50 Hz, V/F signal 0-10 V |
10 | The Cooling Tank | Volume 1 m3 |
11 | The Chilled Tank | Volume 1 m3 |
12 | Reservoir | Volume 20 L |
C0 | C1 | C2 | C3 | C4 | C5 | C6 | C7 |
---|---|---|---|---|---|---|---|
−1672 | 1 | 5 | 1 | 456 | 0.3 | −1 | 1 |
The first coarse model | 7.9673 × 10−8 | 3.1662 × 10−5 | 0.0034666 | 1.1841 |
The second coarse model | 1.0584 × 10−7 | 3.9588 × 10−5 | 0.0043562 | 0.69583 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, D.; Gao, Z. Improvement of Refrigeration Efficiency by Combining Reinforcement Learning with a Coarse Model. Processes 2019, 7, 967. https://doi.org/10.3390/pr7120967
Zhang D, Gao Z. Improvement of Refrigeration Efficiency by Combining Reinforcement Learning with a Coarse Model. Processes. 2019; 7(12):967. https://doi.org/10.3390/pr7120967
Chicago/Turabian StyleZhang, Dapeng, and Zhiwei Gao. 2019. "Improvement of Refrigeration Efficiency by Combining Reinforcement Learning with a Coarse Model" Processes 7, no. 12: 967. https://doi.org/10.3390/pr7120967
APA StyleZhang, D., & Gao, Z. (2019). Improvement of Refrigeration Efficiency by Combining Reinforcement Learning with a Coarse Model. Processes, 7(12), 967. https://doi.org/10.3390/pr7120967