Next Article in Journal
The Influence of Variable Operating Conditions and Components on the Performance of Centrifugal Compressors in Natural Gas Storage Reservoirs
Previous Article in Journal
Green Hydrogen in Jordan: Stakeholder Perspectives on Technological, Infrastructure, and Economic Barriers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Energy Management of Electric–Hydrogen Coupled Integrated Energy System Based on Improved Proximal Policy Optimization Algorithm

1
State Grid Jiangsu Electric Power Co., Ltd., Research Institute, Nanjing 210023, China
2
State Grid Jiangsu Electric Power Co., Ltd., Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Energies 2025, 18(15), 3925; https://doi.org/10.3390/en18153925
Submission received: 12 June 2025 / Revised: 7 July 2025 / Accepted: 18 July 2025 / Published: 23 July 2025
(This article belongs to the Special Issue Advances in Hydrogen Energy and Power System)

Abstract

The electric–hydrogen coupled integrated energy system (EHCS) is a critical pathway for the low-carbon transition of energy systems. However, the inherent uncertainties of renewable energy sources present significant challenges to optimal energy management in the EHCS. To address these challenges, this paper proposes an energy management method for the EHCS based on an improved proximal policy optimization (IPPO) algorithm. This method aims to overcome the limitations of traditional heuristic algorithms, such as low solution accuracy, and the inefficiencies of mathematical programming methods. First, a mathematical model for the EHCS is established. Then, by introducing the Markov decision process (MDP), this mathematical model is transformed into a deep reinforcement learning framework. On this basis, the state space and action space of the system are defined, and a reward function is designed to guide the agent to learn to the optimal strategy, which takes into account the constraints of the system. Finally, the efficacy and economic viability of the proposed method are validated through numerical simulation.

1. Introduction

Under the background of sustained and rapid economic development and accelerated industrialization, total energy consumption has shown significant growth [1,2]. Non-renewable coal, oil, and other traditional fossil energy sources are facing a double dilemma due to their excessive use: on the one hand, their non-renewable nature has led to the depletion of reserves, and on the other hand, large-scale fossil fuel consumption results in massive greenhouse gas emissions, triggering frequent extreme weather and environmental crises like air pollution and acid rain [3].
In this context, accelerating the transformation of the energy structure, optimizing the energy consumption pattern by increasing the proportion of clean energy, and constructing a new energy system dominated by renewable energy have become an important strategic issue to break the resource and environmental constraints and promote the sustainable development of human society. The intermittent and fluctuating nature of renewable energy intensifies the complexity of the integrated energy system, which requires an energy management system to realize efficient consumption and stable regulation [4]. Depending on the model type, existing research methods can be broadly categorized into two main types: model-based optimization methods and learning-based methods [5]. Model-based optimization methods require knowledge of the state transfer mechanism of the system. The research results mainly include methods based on meta-heuristic optimization, mathematical planning methods, and methods based on control theory. The meta-heuristic optimization-based methods show unique advantages in solving high-dimensional, nonlinear, and multimodal complex optimization problems by simulating natural phenomena or intelligent behaviors of biological groups and constructing an intelligent optimization framework with a global search capability [6]. Ref. [7] developed a social group-based Gray Wolf optimizer for the economic dispatch problem, which can improve the efficiency and accuracy of the solution. However, the method has issues with search efficiency trade-offs. Therefore, mathematical programming methods already have been widely used for this problem. Ref. [8] introduced a stochastic optimization scheduling model for wind–solar–hydro hybrid systems, accounting for source and load uncertainties. The model utilizes the Vine-Copula and Monte Carlo simulations to generate correlated wind and solar power output scenarios. Ref. [9] developed a multi-objective optimization framework for the scheduling of an integrated electric–thermal energy system, which aims to constrain the scale of electric energy trading through the economic costs, so as to enhance the environmental benefits while guaranteeing the system economy. Control theory has also been applied to solve this problem in multiple research efforts [10,11]. The presented results necessitate uncertainty quantification within the system to accurately simulate the optimization problem. Furthermore, the application of solvers is required to derive the optimal scheduling policy. It is crucial to acknowledge that the selection of models and parameters significantly influences the validity of the optimization outcomes [12].
In recent years, learning-based methods have seen widespread adoption in the power systems field [13]. Such data-driven methods support the direct acquisition of models or control strategies from historical data by virtue of their model-free nature [14,15,16]. Aiming at the real-time electricity price uncertainty in microgrid energy management, Ref. [17] used deep Q-network algorithms to break through the dimensional bottleneck of Q-learning algorithms, but the need for a high-dimensional discretized action space leads to limited efficiency. In this regard, Ref. [18] develops a continuous action space deep reinforcement learning method to realize the economic dispatch of microgrids containing energy storage. Ref. [19] suffers from hyperparameter sensitivity defects, although it introduces a deep deterministic policy gradient to deal with high-dimensional continuous control [20]. Ref. [21] uses a PPO algorithm to significantly reduce the hyperparameter commissioning cost while satisfying the system N-1 security constraints, which is very important in the energy management of integrated energy systems.
In order to enhance the learning ability of the agent when interacting with the environment, one optimization method for the energy management of the EHCS based on the improved PPO algorithm (IPPO) is proposed, and the deterministic training is introduced to improve the convergence performance at the late stage of training. The so-called deterministic training in this context refers to the process whereby, during the late stages of agent training, the optimal action is selected deterministically based on a probabilistic reference aligned with a greedy policy, thereby enhancing the convergence efficiency of the algorithm. Furthermore, the algorithm’s economy efficiency and computational efficiency are validated through a comparative analysis with alternative methodologies.

2. Mathematical Description of EHCS

The energy management problem within an EHCS is formulated as a mathematical optimization model. This model seeks to optimize the dispatch of various system devices over time, subject to load demand constraints, with the objective of maximizing economic efficiency. Furthermore, with the evolution of carbon emissions trading, renewable energy quota mechanisms, and green certificate trading, the electricity–carbon–green certificate joint market presents a novel direction for electricity market development. Consequently, this paper proposes an energy management model for an EHCS, considering the integration of the electricity–carbon–green certificate market.
Take the EHCS shown in Figure 1 as an example, which contains CHP units, renewable energy generating units, hydrogen storage systems, and electric boilers.

2.1. Mathematical Modeling of Hydrogen Energy Storage System (HESS)

The HESS is mainly consisted of an electrolyzer, hydrogen storage tank, and fuel cell.
(1)
Mathematical Modeling of Electrolyzer
P H 2 , el t = η el P el t ,
where P H 2 , el t is the electrolyzer’s hydrogen-producing power. η el is the hydrogen-producing efficiency of the electrolyzer, and P el t is the power consumed by the electrolyzer.
(2)
Mathematical Modeling of Fuel Cell
P fc t = η fc , p P H 2 , fc t ,
Q fc t = η fc , q 1 η fc , p P H 2 , fc t ,
where P fc t is the hydrogen consumption power of the fuel cell, η fc , p is the electrical efficiency of the fuel cell, η fc , q is the residual heat utilization efficiency of the fuel cell, P fc t is the power to generate electricity of the fuel cell, and Q fc t is the power to produce heat of the fuel cell.
(3)
Mathematical Modeling of Hydrogen Storage Tank
E hst t = E hst t 1 + η ch , hst P H 2 , el t P H 2 , fc t / η dis , hst Δ h ,
S o h t = E hst t / M hst ,
where E hst t is the energy stored within the hydrogen storage tank, η ch , hst and η dis , hst are the hydrogen charging and discharging efficiencies of the hydrogen storage tank, respectively, M hst is the rated capacity of the hydrogen storage tank, and S o h t is the hydrogen storage state of the hydrogen storage tank.

2.2. Objective Function

The primary objective of energy management within an EHCS is to minimize operational costs. The mathematical expression is as follows:
min t = 1 T C t = min t = 1 T ( C grid ( t ) + C chp t + C eb t + C hess t + C cet t + C gct t )   ,
where C grid t is the cost of the system participating in the electricity market transaction, C chp t is the operating cost of CHP units, C eb t is the operating cost of the electric boiler, C hess t is the operating cost of the HESS, C cet t is the cost of the system participating in the carbon emission right transaction, and C gct t is the cost of the system participating in the green certificate market transaction.
(1)
Equipment Operating Costs
C eb t = κ eb P eb t Δ h ,
C hess t = κ el P el t + κ fc P fc t Δ h ,
C chp t = λ gas t P chp t η 1 + Q chp t η 2 Δ h ,
where P eb t is the consumption power of the electric boiler, and κ eb is its cost factor. P el t and P fc t are the power of the electrolyzer and fuel cell in the system, and κ el and κ fc are the unit operating costs of the electrolyzer and fuel cell, respectively. P chp t and Q chp t are the outputs of the CHP unit. λ gas t is the natural gas price, and η 1 and η 2 are the unit cost coefficients of the CHP unit.
(2)
Costs of Participating in Electricity Market Transactions
C grid t = λ e t P grid t Δ h P grid t > 0 λ e t P grid t λ disc Δ h P grid t < 0 ,
where λ e t is the electricity price, λ disc is the discount factor for selling electricity, and P grid t is the power exchanged from the electricity market.
(3)
Costs of Participating in Carbon Credit Market Transactions
The carbon emission right trading pricing model mainly includes two mechanisms: unified pricing and ladder pricing. Under the unified carbon trading pricing mechanism, the system’s participation in carbon emission right trading cost is defined as follows:
C cet t = λ cet m act t m 0 t ,
m act t = u act max 0 , P grid t + u chp P chp t ,
m 0 t = u 0 max 0 , P grid t ,
where λ cet is the carbon trading unit price, with a value of CNY 0.058 per kilogram. m act t and m 0 t are the actual carbon emissions and carbon emission quotas of the system. u 0 and u act are the carbon emission quotas and carbon emissions when the system purchases electricity from the electricity market, and u chp is the carbon emissions corresponding to the unit power of the CHP unit. The system’s participation in carbon emissions trading cost under the ladder pricing mechanism is defined as follows:
C cet t = λ cet 1 + δ m act t m 0 t 2 M < m act t m 0 t < M λ cet m act t m 0 t M < m act t m 0 t < M λ cet 1 + δ m act t m 0 t M < m act t m 0 t < 2 M ,
where δ is the growth coefficient of carbon trading unit price, with a value of 0.25. M is the length of carbon emission interval, with a value of 1000.
(4)
Costs of Participating in Green Certificate Market Transactions
C gct t = λ gct θ gct P rew t ,
where λ gct is the price of green certificates, with a value of CNY 50 per unit. θ gct is the quota parameter of green certificates, with a value of 1 unit per MWh.

2.3. Constraints

(1)
Power Balance Constraints
P rew t + P chp t + P grid t P eb t + P fc t P el t = P load t ,
Q chp t + Q eb t + Q fc t = Q load t ,
where P rew t is the power consumption of renewable energy, including the wind and photovoltaic power consumption, P load t and Q load t are the electrical and thermal loads, and Q eb t is the thermal output of the electric boiler, where Q eb t = η eb P eb t , and η eb is the efficiency parameter.
(2)
CHP Unit Operational Constraints
P chp t P chp t 1 P climb , max ,
Q chp t Q chp t 1 Q climb , max ,
where P climb , max and Q climb , max represent the maximum value of power output and thermal output climb per unit time.
In addition, as shown in Figure 2, the CHP unit needs to satisfy the polygonal constraints.
(3)
Electric Boiler Operational Constraints
P eb , min P eb t P eb , max ,
where P eb , min and P eb , max are the minimum and maximum power consumption of the electric boiler.
(4)
Energy Exchange Constraints
P grid t P grid , max ,
where P grid , max is the maximum power exchanged from the electricity market.
(5)
Renewable Energy Output Constraints
P rew t P rew , act t ,
where P rew , act t is the actual output of renewable energy in the system.
(6)
HESS Operational Constraints
P el , min P el t P el , max ,
P fc , min P fc t P fc , max ,
S o h min S o h t S o h max ,
where P el , min and P el , max are the minimum and maximum power consumption of the electrolyzer, P fc , min and P fc , max are the minimum and maximum power production of the fuel cell, and S o h min and S o h max are the minimum and maximum hydrogen storage state of the hydrogen storage tank.

3. Energy Management Strategy of EHCS Based on IPPO Algorithm

The EHCS energy management system is modeled as an agent. The states observed from the environment for the agent include the renewable energy unit output, the electric load, the thermal load, the energy price, the CHP unit output at the previous moment, and the energy storage system’s state of charge at the previous moment. Thus, the system state can be formulated as follows:
s t = P rew t , P load t , λ e t , λ gas t , S o h t 1 , P chp t 1 , Q load t , λ gas t , Q chp t 1 .
In the EHCS, the operational decision variable is the output of each energy conversion unit. To address the temporal coupling of unit outputs, this paper employs the output increment of each unit as the decision variable. Given that P eb t and P grid t are derivable from balance constraints once the CHP unit output and hydrogen storage system output are determined, the system action can be formulated as follows:
a t = Δ P chp t , Δ Q chp t , P el t , P fc t .
The reward function must encompass the system’s operational cost function to guide the agent toward the optimal policy. Given the critical need to avoid constraint violations in real-world operations, a penalty function is also integrated into the reward function, formulated as follows:
r t = a 1 C t + a 2 C ex t ,
C ex t = p en p grid t + p en p eb t ,
p en x = v x v min + v x v max v max v min ,
where C ex t is the penalty term to constrain p grid t and p eb t . a 1 represents the cost scaling factor, while a 2 denotes the penalty term scaling factor.
In this paper, an IPPO algorithm is proposed to find the optimal strategy for the energy management problem in the framework of MDP. The PPO algorithm, an off-policy method, integrates dynamic step-size and importance sampling to stabilize policy updates by clipping the surrogate objective function [22]. However, the agent’s reliance on environment interactions for training data introduces significant uncertainty. This is particularly evident in later training stages, where substantial state changes during interactions can destabilize the policy and value functions, leading to reward curve degradation and potential convergence to local optima.
Based on the PPO algorithm, deterministic training is introduced to improve the convergence performance at the late stage of training. The so-called deterministic training refers to selecting the action with the highest probability according to a certain probability deterministically, with reference to the greedy strategy at the late stage of the training of an agent, which is defined as follows:
p a = argmax a π a | s , ϑ = 1 ϵ ,
where ϵ is the greedy parameter, with a value of 0.2.
The training framework for the EHCS based on the proposed IPPO algorithm is shown in Figure 3. It is worth noting that in the later stages of training, actions are selected based on the deterministic training strategy mentioned in the preceding paragraph.

4. Case Studies

4.1. Parameter Settings

To assess the efficacy of the proposed IPPO algorithm in optimizing energy management within the EHCS, we employ the system depicted in Figure 1 for simulation and analysis. Historical data, split into training and test sets, provide the renewable energy output and system’s load profiles [23]. The system operates on a 24 h dispatch period with hourly intervals. As detailed in Table 1, the system utilizes a time-of-use pricing mechanism, with a 0.5 discount factor applied to electricity sales. Natural gas is priced at a fixed rate of CNY 0.4/kWh [15]. The hydrogen storage system commences with an initial energy level of 1000 kWh. Additional operational parameters for the system’s equipment are presented in Table 2.
The policy network and the value network both employ fully connected layers with three hidden layers (128, 64, and 32 neurons, respectively) to form the neural network architecture, using the ReLU activation function in their hidden layers. The Adam optimizer is utilized to update the network’s weights, with specific hyperparameter configurations detailed in Table 3.
The simulation experiments were based on the Python language and were run on a computer with AMD Ryzen 5 5600 G CPU and 16 G RAM, and subsequent experiments were based on this environment.

4.2. Analysis of the Training Process

The training reward curve of the IPPO and PPO algorithms is shown in Figure 4. Initially, the reward of the agent is low due to environment exploration, and then gradually increases and stabilizes with the accumulation of experience. The original PPO has fluctuating rewards in the late stage due to the random strategy, and the improved algorithm introduces a deterministic training strategy to suppress this phenomenon. The analysis demonstrates that the IPPO algorithm is capable of learning the optimal strategy with faster speed and greater stability.

4.3. Analysis of Online Operation Results

Following the training of the proposed IPPO and PPO algorithms with historical data, the saved algorithm model is deployed for energy management within the EHCS. Test data, randomly selected from the historical dataset, are utilized in the test set. The system’s load curve, along with the wind and photovoltaic output curves, are illustrated in Figure 5 and Figure 6.
The energy management optimization outcomes derived from the method proposed in this study are illustrated in Figure 7 and Figure 8. As can be seen from Figure 7, under the time-of-use pricing mechanism, the EHCS will choose to purchase electricity from the electricity market to meet the main demand of the system’s power load during the low electricity price hours (e.g., 0:00~7:00) and at the same time start the electrolyzer to store energy in the HESS (e.g., 1:00 and 5:00~6:00). During high electricity price hours, it will choose to start the fuel cell to meet the system’s power load demand instead of purchasing electricity from the electricity market, thus avoiding the purchase of electricity during peak electricity price hours, which can achieve the effect of peak shaving and valley filling. It can be noticed that at 22:00, the combined renewable energy output and CHP unit output of the system exceeds the electric load, and the system conducts a power sale. This decision of the system is mainly considered to meet the system’s renewable energy consumption, and at the same time, improve the economic efficiency of the system.
As can be seen from Figure 8, under the time-of-use pricing mechanism, the EHCS will start the fuel cell to meet part of the thermal load demand of the system during high electricity price hours (e.g., 8:00, 10:00, and 19:00–21:00), and during low electricity price hours, the system will mainly consider using the electric boiler to supplement the part of the thermal load demand that is insufficient for the removal of the CHP unit instead of choosing to start the fuel cell for the thermal supply of the system. The system’s decision is mainly based on the consideration that the electric boiler can form a complementary synergistic operation mode with the CHP unit: through the price signal of the electricity market to adjust the power of electric–thermal conversion in real time, dynamically matching the demand curve of the thermal load on the user side, and at the same time efficiently completing the dynamic regulation of electric–thermal conversion and thermal load, this synergistic mechanism can help the system to realize a higher economic efficiency.
The optimization outcomes demonstrate that the proposed energy management strategy for the EHCS based on the IPPO algorithm can dynamically adjust the output of each device within the system in response to changes in the external environment.

4.4. Analysis of System Operation Results

The following five scenarios are set up to compare the impacts of the carbon emissions trading mechanism, the green certificate trading mechanism, and different carbon emissions pricing mechanisms on the operation of the system:
Scenario 1: Participation in the electricity trading market.
Scenario 2: Participation in the electricity trading market and the green certificate trading market.
Scenario 3: Participation in the power trading market and carbon trading market and using the unified pricing mechanism.
Scenario 4: Participation in the power trading market and carbon trading market and using the ladder pricing mechanism.
Scenario 5: Participation in the power trading market, green certificate trading market, and carbon trading market, and using the ladder pricing mechanism.
To evaluate the efficacy of our proposed system energy management optimization approach, this paper has benchmarked its performance against the IPSO algorithm, the PPO algorithm, the DDPG algorithm, and the stochastic mixed integer linear programming algorithm (SMILP) [24] within a comparative experimental framework.
In the SMILP model solution, the wind power curve in the test set is categorized into 10 classes representing wind power uncertainty scenarios using a clustering algorithm [25,26]. This classification forms the basis for defining the system operation cost error e .
e = c i c SMILP c SMILP × 100 % ,
where c SMILP is the system’s cost obtained by the SMILP method, while c i is the cost obtained by other methods.
In the PSO algorithm, the particle swarm size is configured as 500, the maximum iteration count is set to 1500, and the individual learning factor and population learning factor are 1.5 and 2, respectively.
Different algorithms are employed to address each specific scenario. Table 4 shows the average operating costs of the system on a test set of 30 randomly selected days.
(1)
Comparison of Different Algorithms
(1) Analysis of system’s operation costs
As can be seen from Table 4, considering scenario 1, the system’s average operation cost obtained by the SMILP method is CNY 17,310.20. In contrast, the proposed IPPO method yields a figure of CNY 17,787.57, with a corresponding error of 2.76%. The PPO method results in an average operation cost of CNY 17,904.24, with an error of 3.43%. The DDPG method results in an average operation cost of CNY 17,943.56, with an error of 3.66%. Finally, the IPSO method produces an average operational cost of CNY 17,867.35, and an error of 3.22%. As shown in Figure 9, from the operation of the HESS, the main reason for this error is that the IPPO method proposed in this paper obtains the electrolyzer and fuel cell power time and power size, and the SMILP method power time relative to the other two methods in the low tariff time period is more consistent. The main power error occurs at time 11:00, the SMILP method carries out a hydrogen storage action, while the IPPO method does not choose to carry out the hydrogen storage action. In terms of operational cost, the IPPO method proposed in this paper demonstrates certain benefits when compared to the IPSO and PPO methods. Furthermore, this approach facilitates enhanced optimization decisions concerning the energy management challenges within the EHCS framework.
(2) Analysis of system online execution time
Using scenario 1 as an example, Table 5 displays the online execution time of various methods. As shown, the proposed IPPO algorithm achieves an online execution time of 0.41 s, matching that of the PPO and DDPG algorithm, and significantly outperforming the SMILP and IPSO methods. This efficiency stems from the offline training phase of the reinforcement learning model, which once completed, allows the trained agent to be directly deployed for real-time energy management and scheduling in the EHCS without the need for additional retraining.
(2)
Comparison of Different Scenarios
Scenario 2: Based on scenario 1, the green certificate market is analyzed. The system’s operational cost in this scenario is reduced by CNY 562.85 relative to scenario 1, primarily due to the economic advantages derived from converting wind and photovoltaic power consumption into green certificates.
Scenario 3: Based on scenario 1, a carbon trading market with a unified pricing mechanism is considered. Upon incorporating carbon emission costs, the total system operational cost increase from CNY 17,787.57 in scenario 1 to CNY 18,079.62. This escalation is attributed to the imposition of carbon emission costs. The carbon emission data for each scenario are presented in Table 6.
As can be seen from Table 6, considering the carbon trading market with a uniform pricing mechanism, the system carbon emission of scenario 3 decreases by 6.76% compared with scenario 1, which reduces the system carbon emission to a certain extent, but this also makes the system operation cost grow.
Scenario 4: Based on scenario 3, a carbon trading market with a ladder pricing mechanism is proposed to expedite the low-carbon transition of the power system. The carbon emissions of the system at each time step, derived from the IPPO algorithm, are presented in Figure 10, within the framework of scenario 1 and scenario 4.
As illustrated in Figure 10, the disparity in system carbon emissions between the two scenarios primarily manifests during periods of low electricity prices. In scenario 1, the system procures electricity from the power market to meet load demand during these periods, concurrently curtailing the CHP unit output to optimize system economics, resulting in system carbon emissions of 10,980.75 kg. Conversely, in scenario 4, the incorporation of carbon emission costs and a tiered pricing mechanism renders the economic benefits of purchasing power from the market during low-price periods insufficient to offset carbon costs. Consequently, the system increases the CHP unit output to satisfy the load demand. This strategy reduces system carbon emissions to 6064.30 kg, a 44.77% reduction compared to scenario 1. However, the system’s operational cost escalates to CNY 18,355.14, primarily due to the differential impacts of the carbon pricing mechanisms on operational expenses.
Scenario 5: Considering both the green certificate market and the carbon trading market with a ladder pricing mechanism, as illustrated in Table 5, the system’s carbon emissions in this scenario remain equivalent to those in scenario 4; however, the operational cost is reduced by CNY 562.85 relative to scenario 4. This reduction is primarily attributable to the integration of the green certificate market. Consequently, during the low-carbon transition of the power system, the incorporation of the green certificate market can mitigate, to some extent, the escalation of operational costs driven by environmental regulatory pressures.

4.5. Analysis of System Robustness

Figure 11 and Figure 12 show the impact of parameters in the proposed reward function on the agent’s accumulation of environmental experience.
For the cost discount factor a 1 , different selections influence the convergence performance of the reward function during system training. Specifically, a smaller a 1 leads to a smaller convergence value of the reward function, prompting this paper to set the cost discount factor as 0.01. Regarding the penalty factor a 2 , when a 2 = 150 , the relatively large penalty intensity increases the volatility of the system return function and reduces system stability. Conversely, when a 2 = 50 , the smaller penalty factor weakens the convergence performance of the system return function. Thus, 100 is chosen as the optimal value for the penalty discount factor.
It can be seen that the agent usually converges to relatively satisfactory results under different parameter configurations.

5. Conclusions

This paper introduces an advanced energy management optimization method for an EHCS utilizing an IPPO algorithm. Unlike conventional optimization methods, this method obviates the need for the precise prediction of system uncertainties by leveraging real-time observational data for decision-making. Compared to the PPO method, the proposed method incorporates a deterministic training strategy in the later stages of the learning process, facilitating more rapid and stable convergence toward optimal control policies. Multiple operational scenarios are simulated to evaluate the performance of the proposed method against existing methods, including IPSO, DDPG, and PPO. Results demonstrate a reduction in average system operational costs by 0.45%, 0.87%, and 0.65%, respectively. Additionally, the reinforcement learning-based methods exhibit superior online computational efficiency relative to model-based optimization methods. The findings substantiate the feasibility and efficacy of the IPPO algorithm in optimizing energy management within EHCS frameworks.
In addition, in the process of the low-carbon transition of the power system, the carbon trading market considering the ladder pricing mechanism can reduce the system carbon emissions by 44.77%. Further considerations of the green certificate market can effectively alleviate the problem of the operating cost increase caused by environmental pressure.
Future research can focus on further improving the performance of learning-based algorithms, such as considering the use of multi-agent algorithms to solve this problem, and considering the security constraints of the system, as mentioned in Ref. [27].

Author Contributions

Conceptualization, J.Z. and Z.G.; methodology, J.Z. and Z.G.; software, Z.G. and Z.C.; validation, J.Z. and Z.G.; formal analysis, J.Z. and Z.C.; investigation, J.Z. and Z.C.; resources, J.Z.; writing—original draft preparation, J.Z. and Z.G.; writing—review and editing, J.Z. and Z.C.; visualization, J.Z. and Z.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the State Grid Jiangsu Electric Power Co., Ltd. Technology Project under Grant J2024005 (Research on Planning and Operation Technology of Electro-Hydrogen Coupling System Driven by the Electric-Carbon-Green Certificate Market).

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author. The data are not publicly available due to privacy or ethical restrictions.

Conflicts of Interest

Jingbo Zhao, Zhengping Gao and Zhe Chen were employed by the State Grid Jiangsu Electric Power Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Alabi, T.M.; Aghimien, E.I.; Agbajor, F.D.; Yang, Z.; Lu, L.; Adeoye, A.R.; Gopaluni, B. A review on the integrated optimization techniques and machine learning approaches for modeling, prediction, and decision making on integrated energy systems. Renew. Energy 2022, 194, 822–849. [Google Scholar] [CrossRef]
  2. Zhao, J.; Song, Y.; Fan, H. Optimization Scheduling of Hydrogen-Integrated Energy Systems Considering Multi-Timescale Carbon Trading Mechanisms. Energies 2025, 18, 1612. [Google Scholar] [CrossRef]
  3. Yang, M.; Liu, Y. Research on multi-energy collaborative operation optimization of integrated energy system considering carbon trading and demand response. Energy 2023, 283, 129117. [Google Scholar] [CrossRef]
  4. Teng, F.; Zhang, Q.; Zou, T.; Zhu, J.; Tu, Y.; Feng, Q. Energy management strategy for seaport integrated energy system under polymorphic network. Sustainability 2022, 15, 53. [Google Scholar] [CrossRef]
  5. Liu, D.; Zang, C.; Zeng, P.; Li, W.; Wang, X.; Liu, Y.; Xu, S. Deep reinforcement learning for real-time economic energy management of microgrid system considering uncertainties. Front. Energy Res. 2023, 11, 1163053. [Google Scholar] [CrossRef]
  6. Wang, Y.; Hao, Y.; Wang, L.; Dang, X.; Jiang, L.; Zhang, Y. Multi-objective optimal dispatching for multi-energy microgrid based on improved particle swarm optimization algorithm. Electr. Meas. Instrum. 2023, 60, 29–36+59. [Google Scholar]
  7. Hosseini, S.; Beigvand, S.D.; Abdi, H.; Rastgou, A. Society-based Grey Wolf Optimizer for large scale Combined Heat and Power Economic Dispatch problem considering power losses. Appl. Soft Comput. 2022, 117, 108351. [Google Scholar] [CrossRef]
  8. Fan, Y.; Liu, W.; Zhu, F.; Wang, S.; Yue, H.; Zeng, Y.; Xu, B.; Zhong, P.-A. Short-term stochastic multi-objective optimization scheduling of wind-solar-hydro hybrid system considering source-load uncertainties. Appl. Energy 2024, 372, 123781. [Google Scholar] [CrossRef]
  9. Li, X.; Chen, Y.Z.; Li, H.W.; Liu, L.; Huang, J.Q.; Guo, P.F. Two-stage Robust Optimization of Low-Carbon Economic Dispatch for Electricity-Thermal Integrated Energy System considering Carbon Trade. Electr. Power Constr. 2024, 45, 58–69. [Google Scholar]
  10. Zhang, Y.; Fu, L.; Zhu, W.; Bao, X.; Liu, C. Robust model predictive control for optimal energy management of island microgrids with uncertainties. Energy 2018, 164, 1229–1241. [Google Scholar] [CrossRef]
  11. Zhao, Z.; Guo, J.; Luo, X.; Lai, C.S.; Yang, P.; Lai, L.L.; Li, P.; Guerrero, J.M.; Shahidehpour, M. Distributed Robust Model Predictive Control-Based Energy Management Strategy for Islanded Multi-Microgrids Considering Uncertainty. IEEE Trans. Smart Grid 2022, 13, 2107–2120. [Google Scholar] [CrossRef]
  12. Zhou, S.; Hu, Z.; Gu, W.; Jiang, M.; Chen, M.; Hong, Q.; Booth, C. Combined heat and power system intelligent economic dispatch: A deep reinforcement learning approach. Int. J. Electr. Power Energy Syst. 2020, 120, 106016. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Lin, Y.; Huang, G.; Yang, X.D.; Weng, G.Q.; Zhou, Z.Y. Review on applications of deep reinforcement learning in regulation of microgrid systems. Power Syst. Technol. 2023, 47, 2774–2788. [Google Scholar]
  14. Baldi, S.; Michailidis, I.; Ravanis, C.; Kosmatopoulos, E.B. Model-based and model-free “plug-and-play” building energy efficient control. Appl. Energy 2015, 154, 829–841. [Google Scholar] [CrossRef]
  15. Wang, X.; Zhao, Q.; Zhao, L. Energy management approach for integrated electricity-heat energy system based on deep Q-learning network. Electr. Power Constr. 2021, 42, 10–28. [Google Scholar]
  16. Feng, C.; Zhang, Y.; Wen, F.; Ye, C.; Zhang, Y.B. Energy management strategy for microgrid based on deep expected Q network algorithm. Autom. Electr. Power Syst. 2022, 46, 14–22. [Google Scholar]
  17. Ji, Y.; Wang, J.; Xu, J.; Fang, X.; Zhang, H. Real-Time Energy Management of a Microgrid Using Deep Reinforcement Learning. Energies 2019, 12, 2291. [Google Scholar] [CrossRef]
  18. Hua, H.C.; Qin, Y.C.; Hao, C.T.; Cao, J. Optimal energy management strategies for energy Internet via deep reinforcement learning approach. Appl. Energy 2019, 239, 598–609. [Google Scholar] [CrossRef]
  19. Chen, P.; Liu, M.; Chen, C.; Shang, X. A battery management strategy in microgrid for personalized customer requirements. Energy 2019, 189, 116245. [Google Scholar] [CrossRef]
  20. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2017, arXiv:1509.02971. [Google Scholar]
  21. Yang, Z.; Ren, Z.; Sun, Z.; Liu, M.; Jiang, J.; Yin, Y. Security-constrained economic dispatch of renewable energy integrated power systems based on proximal policy optimization algorithm. Power Syst. Technol. 2023, 47, 988–998. [Google Scholar]
  22. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  23. Belgium’s Electricity Transmission System Operator. Available online: https://www.elia.be/en/grid-data (accessed on 14 July 2024).
  24. Bischi, A.; Taccari, L.; Martelli, E.; Amaldi, E.; Manzolini, G.; Silva, P.; Campanari, S.; Macchi, E. A detailed MILP optimization model for combined cooling, heat and power system operation planning. Energy 2014, 76, 168–174. [Google Scholar] [CrossRef]
  25. Yin, Y.; Liu, T.; He, C. Day-ahead stochastic coordinated scheduling for thermal-hydro-wind-photovoltaic systems. Energy 2019, 187, 1552–1565. [Google Scholar] [CrossRef]
  26. Rahman, M.T.; Hasan, K.N.; Sokolowski, P. Evaluation of wind farm aggregation using probabilistic clustering algorithms for power system stability assessment. Sustain. Energy Grids Netw. 2022, 30, 100678. [Google Scholar] [CrossRef]
  27. Wei, F.X.; Yu, L.; Zhang, X.D. Fully distributed event-triggered security control for DC micro-grids subject to DoS attacks. IEEE Trans. Smart Grid 2025, 162, 929–941. [Google Scholar]
Figure 1. Framework of an EHCS.
Figure 1. Framework of an EHCS.
Energies 18 03925 g001
Figure 2. Feasible operating region of the CHP unit.
Figure 2. Feasible operating region of the CHP unit.
Energies 18 03925 g002
Figure 3. Training framework for EHCS.
Figure 3. Training framework for EHCS.
Energies 18 03925 g003
Figure 4. Training reward curve.
Figure 4. Training reward curve.
Energies 18 03925 g004
Figure 5. Schematic of load curves on the test day.
Figure 5. Schematic of load curves on the test day.
Energies 18 03925 g005
Figure 6. Schematic of renewable energy output curves on the test day.
Figure 6. Schematic of renewable energy output curves on the test day.
Energies 18 03925 g006
Figure 7. Schematic of power load optimization results for the test day.
Figure 7. Schematic of power load optimization results for the test day.
Energies 18 03925 g007
Figure 8. Schematic of thermal load optimization results for the test day.
Figure 8. Schematic of thermal load optimization results for the test day.
Energies 18 03925 g008
Figure 9. Schematic of HESS capacity change.
Figure 9. Schematic of HESS capacity change.
Energies 18 03925 g009
Figure 10. Optimization results of system carbon emissions under different scenarios.
Figure 10. Optimization results of system carbon emissions under different scenarios.
Energies 18 03925 g010
Figure 11. Average reward values for agent with different cost discount factors.
Figure 11. Average reward values for agent with different cost discount factors.
Energies 18 03925 g011
Figure 12. Average reward values for agent with different penalty discount factors.
Figure 12. Average reward values for agent with different penalty discount factors.
Energies 18 03925 g012
Table 1. Time-of-use pricing mechanism.
Table 1. Time-of-use pricing mechanism.
Time Period (h)Electricity Price (CNY/kWh)
Valley Period23:00~6:000.48
Flat Period7:00, 11:00~17:000.90
Peak Period8:00~10:00, 18:00~22:001.35
Table 2. System equipment operating parameters.
Table 2. System equipment operating parameters.
ParameterValueParameterValue
P g r i d , m a x 1000 kW η e b 0.95
P e b , m i n / P e b , m a x 0/1200 kW κ e b 0.05
P climb , max / Q climb , max 300/360 kW η 1 / η 2 0.8/0.75
P el , min / P el , max 0/200 kW M h s t 2000 kWh
S o h m i n / S o h m a x 0.2/0.9 P fc , min / P fc , max 0/200 kW
κ e l / κ f c 0.15/0.17 η c h , h s t / η d i s , h s t 0.95/0.95
η e l 0.6 η f c , p / η f c , q 0.6/0.88
u 0 / u a c t / u c h p 0.7/0.96/0.26 C (960, 400)
A (0, 0) D (200, 0)
B (960, 800) E (0, 0)
Table 3. Deep neural network training parameters.
Table 3. Deep neural network training parameters.
ParameterValue
Policy Network Learning Rate3 × 10−4
Value Network Learning Rate3 × 10−4
Reward Discount Factor0.99
Greed Parameter0.2
Clipping Parameter0.2
Replay-buffer Size10,000
Training Episode1500
Minibatch Size128
Table 4. Average operating cost of the system.
Table 4. Average operating cost of the system.
Methods\ScenariosScenario 1Scenario 2Scenario 3Scenario 4Scenario 5
IPPO17,787.5717,224.7218,079.6218,355.1417,792.29
PPO17,904.2417,341.3918,269.3518,511.3917,948.54
DDGP17,943.5617,380.7118,334.4218,586.6318,023.78
IPSO17,867.3517,304.518,187.4118,445.6217,882.77
SMILP17,310.2016,747.3517,589.2117,831.8217,268.97
Table 5. Average online execution time of the system.
Table 5. Average online execution time of the system.
MethodsOnline Execution Time(s)
IPPO0.41
PPO0.41
DDPG0.40
IPSO14.12
SMILP152.4
Table 6. Total carbon emissions of the system.
Table 6. Total carbon emissions of the system.
ScenariosScenario 1Scenario 2Scenario 3Scenario 4Scenario 5
Total carbon emissions (kg)10,980.7510,980.7510,238.216064.306064.30
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, J.; Gao, Z.; Chen, Z. Energy Management of Electric–Hydrogen Coupled Integrated Energy System Based on Improved Proximal Policy Optimization Algorithm. Energies 2025, 18, 3925. https://doi.org/10.3390/en18153925

AMA Style

Zhao J, Gao Z, Chen Z. Energy Management of Electric–Hydrogen Coupled Integrated Energy System Based on Improved Proximal Policy Optimization Algorithm. Energies. 2025; 18(15):3925. https://doi.org/10.3390/en18153925

Chicago/Turabian Style

Zhao, Jingbo, Zhengping Gao, and Zhe Chen. 2025. "Energy Management of Electric–Hydrogen Coupled Integrated Energy System Based on Improved Proximal Policy Optimization Algorithm" Energies 18, no. 15: 3925. https://doi.org/10.3390/en18153925

APA Style

Zhao, J., Gao, Z., & Chen, Z. (2025). Energy Management of Electric–Hydrogen Coupled Integrated Energy System Based on Improved Proximal Policy Optimization Algorithm. Energies, 18(15), 3925. https://doi.org/10.3390/en18153925

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop