Abstract
In the context of dual-carbon goals, Park-Level Integrated Energy Systems (PIES) are pivotal for enhancing renewable energy integration and promoting clean, efficient energy use. However, the inherent non-linearity from multi-energy coupling and the high dimensionality of operational data present substantial challenges for conventional scheduling optimization methods. To overcome these obstacles, this paper introduces a novel multi-objective scheduling framework for PIES leveraging deep reinforcement learning. We innovatively formulate the scheduling task as a Markov Decision Process (MDP) and employ the Trust Region Policy Optimization (TRPO) algorithm, which is adept at handling continuous action spaces. The state and action spaces are meticulously designed according to system constraints and user demands. A comprehensive reward function is then established to concurrently pursue three objectives: minimum operating cost, minimum carbon emissions, and maximum exergy efficiency. Through comparative analyses against other AI-based algorithms, our results demonstrate that the proposed method significantly lowers operating costs and carbon footprint while enhancing overall exergy efficiency. This validates the model’s effectiveness and superiority in addressing the complex multi-objective scheduling challenges inherent in modern energy systems.