Next Article in Journal
The Effect of Land Cover on the Nectar Collection by Honeybee Colonies in Urban and Rural Areas
Previous Article in Journal
Dynamic Obstacle Avoidance for Robotic Arms Using Deep Reinforcement Learning with Adaptive Reward Mechanisms
Previous Article in Special Issue
Learning-Based Branching Acceleration for Unit Commitment with Few Training Samples
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stable Variable Fixation for Accelerated Unit Commitment via Graph Neural Network and Linear Programming Hybrid Learning

1
Guangxi Key Laboratory of Multimedia Communication and Network Technology, School of Computer Electronics and Information, Guangxi University, Nanning 530004, China
2
School of Electrical Engineering, Guangxi University, Nanning 530004, China
3
School of Mathematics and Information Science, Guangxi University, Nanning 530004, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(8), 4498; https://doi.org/10.3390/app15084498
Submission received: 4 March 2025 / Revised: 14 April 2025 / Accepted: 17 April 2025 / Published: 18 April 2025

Abstract

:
The Unit Commitment Problem (UCP) is a critical component of power market decision-making and is typically formulated as Mixed Integer Programming (MIP). Given the complexity of solving MIPs, efficiently solving large-scale UCPs remains a significant challenge. This paper presents a hybrid Graph Neural Network (GNN)–Linear Programming (LP) framework to accelerate the solution of large-scale Unit Commitment Problems (UCPs) while maintaining the quality of solutions. By analyzing variable stability through historical branch-and-bound (B&B) trajectories, we classify MIP variables into dynamically adjustable stable and unstable groups. We adopt an MIP formulation that includes multiple types of binary variables—such as commitment, startup, and shutdown variables—and extract additional information from these auxiliary binary variables. This enriched representation provides more candidates for stable variable fixation, helping to improve variable refinement, mitigate suboptimality, and enhance computational efficiency. A bipartite GNN is trained offline to predict stable variables based on system topology and historical operational patterns. During online optimization, instance-specific root LP solutions refine these predictions, enabling adaptive variable fixation via a dual-threshold mechanism that integrates GNN confidence and LP relaxations. To mitigate suboptimality risks, we introduce temporally flexible fixation strategies—hard fixation for variables with persistent stability and soft fixation allowing limited temporal adjustments—alongside a GNN-guided branching rule to prioritize unstable variables. Numerical experiments demonstrate that jointly fixing commitment, startup, and shutdown variables yields better performance compared to fixing only commitment variables. Ablation studies further validate the importance of hard fixation and customized branching strategies, especially for large-scale systems.

1. Introduction

The Unit Commitment Problem (UCP) has garnered substantial attention from both industrial and academic communities due to its critical role in power system operations. Fundamentally, the UCP involves determining the optimal operation schedule of generating units across multiple time periods while accommodating fluctuating load demands and adhering to various operational constraints [1]. Over the years, diverse methodologies have been developed to address this complex problem, including Lagrangian relaxation [2,3], priority list methods [4], dynamic programming [5], and Mixed Integer Programming (MIP) approaches. The advancement of MIP theory, coupled with the enhanced efficiency of modern branch-and-cut solvers [6,7], has significantly increased the popularity of MIP-based solutions for UCPs. Given the NP-hard (non-deterministic polynomial-time hard) nature of UCP computations [8], substantial research efforts have been directed toward improving computational efficiency through tighter problem formulations [9] and Linear Programming (LP) relaxation [10]. In recent years, the emergence of learning approaches has opened new avenues for enhancing UCP solution processes [11,12,13], offering promising directions for future research. Notably, Graph Neural Networks (GNNs) possess a unique structural advantage due to their inherent permutation invariance [14]. This characteristic enables GNNs to generalize more effectively in UCP tasks compared to other neural network architectures [15]. Consequently, this has led to the development of numerous methods that leverage GNNs to assist MIP formulations in recent years.

1.1. Acceleration via LP Relaxation

The computational intensity associated with UCP solutions has driven the development of sophisticated LP relaxation methodologies aimed at balancing solution quality and computational efficiency. These techniques have evolved beyond traditional approaches, incorporating innovative strategies to enhance both the quality of initial solutions and the efficiency of the optimization process.
Contemporary research has demonstrated significant advancements in this domain. Recent studies [16] have introduced refined problem formulations combined with intelligent repair mechanisms for continuous relaxation variables, establishing a foundation for iterative solution enhancement. A particularly noteworthy approach [10] has emerged, focusing on strategic variable fixation through a heuristic analysis of commitment and generation variables, effectively reducing computational complexity while maintaining solution integrity. The field has further progressed with the introduction of advanced neighborhood search algorithms [17] that incorporate sophisticated relaxation-inducement techniques. Additionally, novel iterative methods [18] have been developed, combining localized branching strategies with targeted neighborhood exploration to optimize solution quality.
Reference [19] solves multiple UCPs that share the same feasible region as the original problem but have slightly different objective functions. The method determines stable variables by averaging the LP relaxation solutions and enhances computational efficiency through an adjoint model without sacrificing optimality.

1.2. Acceleration via Machine Learning

The integration of predictive analytics into power system optimization represents a paradigm shift in addressing complex operational challenges. The inherent stability of system parameters across operational instances provides a unique opportunity for knowledge transfer through machine learning techniques, enabling accelerated solution processes for recurrent optimization challenges. Research indicates that solving large-scale power systems by machine learning is a relatively accessible and straightforward challenge that can be readily addressed [20].
For constraint prediction and simplification, the authors of [21] employed support vector machines and k-nearest neighbors to identify inactive transmission constraints that can be safely omitted, using partial solutions as warm starts. The authors of [22] introduce an E-Sequence-to-Sequence data-driven framework for multi-sequence mapping, while the authors of [23] devise a machine learning classifier to predict commitment variables and subsequently restore their feasibility.

1.3. Graph Neural Networks for MIP and UCP

GNNs have attracted increasing attention in both power system optimization and MIP due to their powerful ability to model topological structures and generalize across graph-based representations.
The authors of [24] encoded power grid topologies into graphs, utilizing GNNs to simultaneously predict active transmission constraints and commitment variables. These predictions are then refined to ensure feasibility and reduce problem dimensionality.
The methods used in [25] exploit GNNs to estimate marginal probabilities of binary variables in MIP formulations, followed by constrained local searches near predicted solutions. However, such confined search spaces often compromise optimization efficiency and solution quality. The authors of [26] proposed a Graph Neural Network integrated with neural diving and branching strategies, achieving superior scalability and cost-efficiency in solving large-scale Unit Commitment Problems compared to conventional MIP solvers.

1.4. Contributions

In UCP, auxiliary variables (e.g., startup and shutdown variables) often tighten MIP formulations. While existing learning-based frameworks rarely incorporate these variables into their architectures, this may limit their ability to learn high-quality solutions efficiently.
To improve the utilization of these variables, we propose a hybrid offline–online prediction framework for three binary variables to solve the UCP. The framework is decomposed into two synergistic components, as shown in Figure 1, as an offline training phase and an online solving phase.
Offline training phase: Historical internal solution information from the B&B process of UCP instances, including variable branching patterns and LP relaxation bounds, is used to obtain the stability of each binary variable. These data are then used to train the GNN, enabling it to predict the stability of each binary variable in UCP instances.
Online solving phase: All binary variables in the UCP are first categorized into two groups: stable variables and unstable variables. This process uses the stability predicted by the GNN and the values of the LP relaxation solution for classification. The unstable variables are left to the solver for resolution, while the stable variables are further classified. The fixed approach for commitment variables is determined by auxiliary binary variables. Additionally, a learning-guided branching method is employed to accelerate the branching process.

2. Preliminaries

In this section, we introduce the fundamental formulations of the UCP, its graph-based representations, and the stability characteristics of binary variables that underpin our learning framework.

2.1. Three-Binary UCP Formulations

In this part, we introduce the classical three-binary (3-bin) formulation of the UCP from [9], i.e., commitment, startup, and shutdown variables.
1.
Objective Function:
The objective function represents the total cost of the power system, which is shown as
m i n g = 1 N t = 1 T α g u g , t + β g P g , t + S g , t
which comprises the production cost α g u g , t + β g P g , t and the startup cost S g , t .
2.
Startup cost:
The startup cost S g , t can be expressed as an MIP formulation in a 3-bin formulation, with s g , t and d g , t . The cost is determined by comparing the time since the unit was last shut down to a threshold representing the cold-start duration. It is formulated as
S g , t C h o t , g s g , t S g , t C c o l d , g s g , t τ = m a x ( t T ¯ o f f , g T c o l d , g , 1 ) t 1 d i , τ f i n i t , g , t
Since the UCP must account for the initial status of generating units, the classification of hot and cold starts in the early time periods may depend on residual states from the previous day. As a result, the standard rules for determining startup types cannot be uniformly applied across all time periods. In particular, when the following two conditions are satisfied, t T ¯ o f f , g T c o l d , g 0 and m a x ( T g , 0 , 0 ) < t T ¯ o f f , g T c o l d , g 1 + 1 , we set f i n i t , g , t = 1 , indicating that the startup at time t is considered a hot start due to the influence of the unit’s initial state. Otherwise, f i n i t , g , t = 0 , and the classification of the startup should instead be determined based on the summation term.
3.
Unit generation limits:
u g , t P ¯ g P g , t
P g , t u g , t P ¯ g
4.
Power balance constraint:
g = 1 N P g , t P D , t = 0
5.
System spinning reserve requirement:
g = 1 N u g , t P ¯ g P D , t + R t
6.
Ramp rate limits:
P g , t P g , t 1 u g , t P u p , g + P ¯ g u g , t 1 P ¯ g + s g , t P s t a r t , g P u p , g P ¯ g
P g , t 1 P g , t u g , t 1 P d o w n , g + P ¯ g u g , t P ¯ g + d g , t P s h u t , g P d o w n , g P ¯ g
7.
Minimum up/downtime constraints:
ϖ = max ( t T ¯ o n , g , 0 ) + 1 t s g , ϖ u g , t , t [ U g + 1 , , T ]
ϖ = max ( t T ¯ o f f , g , 0 ) + 1 t d g , ϖ 1 u g , t , t [ L g + 1 , , T ]
Here, ϖ is an index used to enforce the minimum up/downtime constraints. Taking Equation (9) as an example, if the unit is in the off state at time t (i.e., u g , t = 0 ), then any startup operation occurring at time ϖ would imply that the unit shuts down before completing its required minimum uptime, thereby violating the constraint.
8.
Initial status of units:
u g , t = u g , 0 , t 1 , , U g + L g
9.
State constraints:
State constraints were introduced in the 3-bin formulation to link the startup, shutdown, and state variables:
s g , t d g , t = u g , t u g , t 1

2.2. Bipartite Graph Representation for UCPs

This part describes how the UCP is transformed into a bipartite graph, a structured representation that enables GNN to learn variable–constraint interactions. Our approach adapts the MIP to the bipartite graph conversion framework proposed in [27] as shown in Figure 2.
The bipartite graph G = V , C , E represents two disjoint sets of nodes—variables and constraints—and the edges connecting them. The associated features of these nodes and edges are summarized in Table 1.
  • Variables: V = v 1 , , v n , where each variable embedding v i R 7 corresponds to a decision variable x i , partitioned as binary variables x 1 x n b (e.g., commitment, startup, and shutdown variables) and continuous variables x n b + 1 x n (e.g., power output and startup cost).
  • Constraints: C = c 1 , , c m , where each constraint embedding c j R 4 corresponds to the j -th constraint, e.g., operational rules formalized in Equations (2)–(12).
  • Edges: i , j E , which encodes the relationship where variable x i appears in constraint b j with a nonzero coefficient and has an edge embedding e i , j R .

2.3. Stable Binary Variables

A critical subset of binary variables in MIP exhibits persistent stability throughout the B&B process, where their values in solutions remain invariant from the root node to the final optimal solution. To systematically identify such variables, we propose a trajectory-based stability metric derived from the complete solution path of the B&B tree. Specifically, for each binary variable x i , we aggregate its values across the LP relaxation solution and incumbent solutions encountered during the B&B search. The stability of x i is quantified by its normalized persistence score s t a b l e i , which is defined as follows:
a v g i = 1 N + 1 k = 0 N x i k
s t a b l e i = 2 × a v g i 0.5
where N is the total number of incumbent solutions during the B&B process, and x 0 refers to the LP relaxation solution as shown in Table 2. The stability score s t a b l e i ranges from 0 to 1, with higher values indicating a greater consistency of x i .
This approach rigorously distinguishes truly stable variables from those exhibiting transient stability or noise-induced pseudo-stability. By leveraging the entire solution trajectory rather than relying on heuristic thresholds or single-node relaxations, the method ensures robustness against branching decisions and algorithmic perturbations.

3. Framework

This section presents the comprehensive framework, detailing the process from offline training of the GNN to variable fixation, initial solution setup, and ultimately obtaining the final dispatch plan. The framework is designed to enhance the efficiency and scalability of solving the UCP by leveraging data-driven insights and optimization techniques.

3.1. Dataset Collection

Real-world electricity demand data for an entire year are obtained and linearly scaled to match the capacity of the target power system. This scaled demand serves as the historical load profile. To avoid infeasible solutions caused by randomly initialized generator states (e.g., on/off statuses and continuous uptime/downtime), we sequentially solve the UCP for each day in the historical load profile. The final generator states (including continuous operating hours and power outputs) at the end of each day are preserved and used as the initial state for the next day. This ensures feasible and realistic transitions between instances.
Figure 3 shows how the dataset was collected. A subset of the load profiles is used as the test set, while the remaining profiles form the basis of the training set. For each training-day load, load i , we apply slight perturbations to generate multiple variations denoted as load 1 i , , load m i . Each perturbed load, load j i , is paired with the same initial generator state, initial i , producing a set of UCP instances M 1 i , M m i . Both the training and validation sets undergo perturbations to expand the dataset. Models trained on the training set are validated on the validation set, and the model exhibiting the best performance on the validation set is retained for testing on the test set. For the test set, instances are generated using the original training-day loads to evaluate the performance of the proposed framework in the numerical experiment section.
For each instance M j i , a bipartite graph G j i is constructed following the method described in Section 2.2, and its corresponding stability vector stable j i is calculated as described in Section 2.3. It is worth noting that it is impossible to collect all feasible solutions from the B&B search tree, so all feasible solutions explored during the solver’s process are used as a substitute.

3.2. Offline Phase

We train a bipartite GNN-based policy stable predictor π θ following previous work [27], where π denotes the policy as a function and θ represents the set of trainable parameters. The policy is designed to learn patterns from specific operational scenarios, enabling the prediction of variable stability for the UCP under unseen but similarly distributed scenarios.
D = ( G , s t a b l e ) denotes the training set. The policy is a three-layer structure shown in Figure 4 and is described as follows:
  • Input layer: This layer, consisting of a regularization layer, fully connected layers, and activation functions, inputs the embedding V 0 = v 1 0 , , v n 0 , C 0 = c 1 0 , , c m 0 , where v i 0 = v i c j 0 = c j will be embedded into a higher-dimensional latent vector V 1 C 1 .
  • Convolutional layer: The information of neighboring nodes c j l C l and v i l V l is transitioned after a multilayer convolutional layer and takes the following form:
    c j l f c ( c j l 1 , l a y e r n o r m ( i i , j E f v c ( v i l 1 , c j l 1 , e i , j ) ) )  
    v i l f v ( v i l 1 , l a y e r n o r m ( j i , j E f c v ( v i l 1 , c j l 1 , e i , j ) ) )
    where f c , f v , f v c , f c v are Multilayer Perceptrons with ReLU activation functions, and l a y e r n o r m represents the layer normalization function.
  • Output layer: This layer consists of fully connected layers and a sigmoid activation function, which maps the output to a scalar in the range [0, 1], representing the stability confidence of each variable π θ ( G ) = p 1 , , p k and p i [ 0 , 1 ] , i 1 , , k . Depending on the task setting, we can selectively retain the stability predictions of specific types of binary variables—for example, only the commitment variables, or all binary variables. The remaining variables are discarded from further consideration.
Similar to the selective retention in the output layer, the training labels are also selectively clipped. Specifically, we retain only the stability indicators corresponding to certain variables as training labels, represented as s t a b l e [ 0 , 1 ] k , while discarding the labels for the remaining variables. Given that the labels are continuous values bounded within [0, 1], the Binary Cross-Entropy (BCE) loss function is employed, which is ideally suited for such probabilistic labels. The loss function is defined as follows:
L θ = 1 D G , s t a b l e D i = 1 k [ s t a b l e i l o g p i + ( 1 s t a b l e i ) l o g 1 p i ]

3.3. Online Phase

This subsection presents the online phase of our optimization framework, which executes two interdependent routines during MILP solving: (1) machine learning-guided variable fixation to prune the search space and (2) machine learning-guided branching rules to prioritize critical variables. Together, these components construct a reduced model—a modified version of the original UCP model in which certain variable values are fixed and branching priorities are informed by learned patterns. This co-designed strategy significantly accelerates the solving process compared to the original UCP, especially under time-sensitive operational requirements.

3.3.1. Learning-Guided Fixing

The GNN is designed to output predictions exclusively for binary variables, where each value is constrained to the interval [0, 1] via a sigmoid activation function. These values represent stability confidence scores, with higher values indicating greater stability of the corresponding integer variable. To identify stable variables to be fixed, a dual-threshold criterion is applied. The set of binary variables x g , t for x u , s , d and g G ˙ , t   T ˙ is considered, where
  • u g , t s g , t , and d g , t represent commitment variables, startup variables, and shutdown variables of unit g in time t ;
  • G ˙ and T ˙ denote the set of units and time periods, respectively.
We define a stable set S s t a b l e as follows:
S s t a b l e = ( x , g , t ) x u , s , d , g G ˙ , t   T ˙ , X g , t x   τ s t a b l e r g , t x τ v r g , t x 1 τ v
where
  • X g , t x represents the stable predictor output (stability confidence score) for the ( g , t ) -th index and the variable type x ;
  • r g , t x represents the LP relaxation value for the ( g , t ) -th index and the variable type x ;
  • τ s t a b l e is the stability threshold, such that variables with X g , t x τ s t a b l e are considered stable;
  • τ v is the integrality tolerance threshold, ensuring that r g , t x is sufficiently close to a binary value.
Only variables fulfilling both conditions are provisionally classified as stable. This joint criterion ensures that stability predictions are both statistically confident and theoretically aligned with the problem’s LP relaxation properties, thereby reducing false positives caused by spurious correlations in the GNN’s latent space. The threshold parameters are tunable to balance solution quality and computational efficiency across varying UCP instances.
While fixing the generator state-transition variables s g , t (startup) and d g , t (shutdown) to 1 at specific time steps risks suboptimality due to the dynamic adjustability of power outputs, we propose a temporally flexible fixing strategy to balance solution quality and computational efficiency. Generators often exhibit flexibility in their startup/shutdown timing, and delaying or advancing commitments by several time steps may still yield feasible and near-optimal solutions. The fixation strategy can be classified into hard fix and soft fix, depending on the degree of flexibility allowed.
Rigid fixation to a single time step may exclude such solutions, while allowing a small temporal window to preserve optimality while reducing branching. For each unit g and time step t , the following fixation rules are applied:
  • Hard fix: S h a r d is denoted as the set of indices that are hard-fixed. These variables are forced into a fixed state, ensuring that the solution adheres to the rounded LP relaxation value for each time step. We define S h a r d as follows:
    S h a r d = x u , s , d } , ( x , g , t S s t a b l e j = m a x ( 0 , t ) m i n ( T , t + ) r g , j s + r g , j d = 0  
    For a given time t of unit g , if all three binary variables (commitment, startup, and shutdown) are stable, and there are no startup or shutdown actions in the neighboring time steps, then these variables will be hard-fixed to their rounded LP relaxation value. It is worth noting that a partial binary solution from the rounded LP relaxation value does not guarantee feasibility.
  • Soft fix: S s o f t is denoted as the set of indices that are soft-fixed, S s o f t = S s t a b l e S h a r d . For variables ( x , g , t ) S s o f t , we initialize the variable to the rounded LP relaxation value but allow the solver to adjust the value during the optimization process.

3.3.2. Learning-Guided Branching

To fully leverage the GNN’s stability predictions, every integer variable in the model is assigned a customized branch priority. The stability scores generated by our GNN not only guide variable fixing but also serve as a critical indicator for divide-and-conquer strategies in B&B. Variables with low stability (close to 0) exhibit inconsistent values across nodes in the B&B tree, implying that their unresolved status contributes significantly to the solution space fragmentation. Prioritizing these variables for early branching helps accelerate bound tightening, and resolving unstable variables reduces feasible region ambiguity, enabling faster pruning of suboptimal branches.
We map stability scores X g , t x [ 0 , 1 ] to integer priorities P g , t x P m i n , , P m a x via
P g , t x = P m i n + P m a x P m i n 1 X g , t x
where · is the rounding operator to the nearest integer; P m a x is the maximum priority; and P m i n is the minimum priority.
In practice, this mapping can be integrated right before invoking the solver. After constructing the MIP model, we traverse each integer variable x g , t , retrieve its predicted X g , t x from the offline-trained model, compute P g , t x according to the above formula, and set the corresponding branch priority in the solver. Once the branch priorities are updated, the solver will systematically prioritize branching on high-priority (low-stability) variables, the solver can more quickly prune large swathes of the search tree, leading to faster convergence in practice.
τ s t a b l e , τ v , are set to 0.5, 0.99, and 1, respectively, in the experiments.

4. Numerical Experiments

4.1. Case Settings

The case study is performed by Gurobi 11.0 using the Python 3.11 API on a PC with an Intel Core i7-8700 CPU running at 3.20 GHz and with 32 GB of memory. All feasible solutions produced by Gurobi in the solution process are recorded.
1.
Power system setting: The three systems are taken from [10], and their sizes are reported in Table 3. Each of these systems comprises solely thermal generating units.
2.
Demand setting: There are 365 loads from the first day of the year to the 365th day of the year, which are findings that come from [28]. In order to match the loads to the target systems, the loads for the 365 days, o r i g i n   l o a d s , are mapped to the capacity of the corresponding systems, s c a l e d   l o a d s . Specifically, for each system, the upper ( s y s u p ) and lower ( s y s l o w ) bounds of the target load mapping interval are determined based on the system’s maximum generation capacity scaled by factors s u p and s l o w :
s y s u p = maximum   generation   capacity × s u p
s y s l o w = s y s u p × s l o w
o r i g i n   l o a d s is then normalized and scaled to s c a l e d   l o a d s to fit the target system:
s c a l e d   l o a d s = o r i g i n   l o a d s m i n ( o r i g i n   l o a d s ) m a x ( o r i g i n   l o a d s ) m i n ( o r i g i n   l o a d s ) × ( s y s u p s y s l o w ) + s y s l o w
A simple rule is applied for dataset partitioning: days ending with the digit seven are assigned to the validation set, those ending with two are assigned to the testing set, and the remaining days are assigned to the training set. And each load will be perturbed to expand the dataset.
3.
Benchmark setting: The latest Gurobi solver 11.0 was used with the following settings: threads set to 1 and MIPGap set to 1 × 10−6. All other settings were kept at their default values.

4.2. Accuracy Comparison for Different Binary Variables

We compare the Mean Squared Error (MSE) of predictions for three types of binary variables ( u for commitment variables, s for startup variables, and d for shutdown variables) under three different system sizes (50 units, 100 units, and 1080 units), and the lower the better. As shown in Table 4, startup and shutdown maintain substantially lower MSE values than commitment across all system scales, indicating that GNN predicts the startup and shutdown variables with much higher accuracy than the commitment variable.
To validate the benefits of including additional variables and eliminate the influence of fixed strategies and branching strategies, we designed experiments where stable variables were simply fixed using a soft fix. Based on the selection of binary variables, the methods can be categorized into three types: only commitment variables (the approach used in most existing papers) as “M1”, only startup and shutdown variables as “M2”, or all three types of binary variables as “M3”, and we use the state-of-the-art solver Gurobi as the baseline. We use the primal gap [29] to evaluate the performance of different methods, which defined v v * v * , where v * is the best objective value obtained by any approach in the portfolio for comparison and v is the primal bound.
Figure 5 shows the primal gap as a function of runtime. In all methods applied to the three systems, a soft fix for all three variables shows the best performance. Especially in the 100-unit system, the M3 gap exhibits the fastest decline, while the other methods take more than 100 s to reach the gap achieved by M3 in just 50 s.

4.3. Ablation Study

The effectiveness of hard fix and branching priority settings is validated, as shown in Figure 6. “M4” represents the complete fixing strategy proposed in Section 3, while “M5” denotes the comprehensive online strategy introduced in the same section, which includes both the fixing strategy and a customized branching rule. By comparing M3 and M4, it can be concluded that our proposed learning-guided fixing strategy leads to a faster initial decline in the objective function compared to using only a soft fix, especially on the 100-unit and 1080-unit systems. A comparison between M4 and M5 further demonstrates the effectiveness of the branching priority strategy on the 1080-unit system, where M5 consistently achieves an earlier reduction in the gap. However, for smaller-scale systems, adjusting the branching strategy did not have a significant impact.

5. Conclusions

This work proposes a GNN-LP hybrid learning framework to accelerate UCP optimization through stability-driven variable fixation and adaptive search strategies. Key contributions include the following:
1.
Variable Stability Classification: We design a variable fixation strategy that integrates GNN-predicted stability and LP root relaxation to guide the hard and soft fixation of binary variables. Stable variables are fixed to shrink the search space, while temporal flexibility is preserved for variables with uncertain stability. This layered fixation mechanism significantly reduces suboptimality, as confirmed by ablation studies across 50-, 100-, and 1080-unit systems.
2.
Utilization of Multi-Type Binary Variables: Instead of focusing solely on commitment variables, we incorporate startup and shutdown variables into both the MIP formulation and learning framework. These auxiliary variables enhance the stability inference of commitment variables to enable safer and more accurate hard fixation, thereby reducing suboptimality and introducing additional stable candidates for soft fixation, which accelerates the solution process by further reducing the search space. Experiments show that this joint modeling approach consistently improves primal gaps and solving times over single-variable baselines, particularly in large-scale settings.
3.
Stability-Aware Branching Prioritization: We leverage predicted variable stability to guide the branching process in branch-and-bound trajectories. By assigning a higher branching priority to unstable variables, the solver is able to explore critical decisions earlier, tighten bounds faster, and accelerate convergence. This strategy achieved faster gap reduction and earlier convergence.
While the proposed branching prioritization strategy based on variable stability has demonstrated significant efficiency gains, it remains relatively simple in its current form. Additionally, the success of incorporating auxiliary binary variables into our learning and optimization framework suggests broader applicability. Future work could investigate which types or structures of multi-variable models yield the best performance, and how such strategies can be generalized to other UCP acceleration frameworks, for example, by leveraging power system topology via graph-based representations to enhance prediction accuracy and generalization.

Author Contributions

Conceptualization, L.Y. and H.Z.; methodology, L.Y.; software, P.L.; validation, L.Y., P.L. and S.C.; formal analysis, L.Y.; investigation, L.Y.; resources, H.Z.; data curation, L.Y.; writing—original draft preparation, L.Y.; writing—review and editing, P.L. and S.C.; visualization, P.L.; supervision, H.Z.; project administration, H.Z.; funding acquisition, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (72361003) and the Key Research and Development Program of Guangxi (GuiKe-AB23075199).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

Indices:
gIndex of unit
tIndex of time period
Sets:
S s t a b l e Set of stable binary variables
S h a r d Set of hard fix binary variables
S s o f t Set of soft fix binary variables
Constants:
N Total number of units
T Total number of time periods
α g , β g Coefficients of the linear production cost function of unit g
C h o t , g Hot startup cost of units g
C c o l d , g Cold startup cost of units g
T ¯ o n , g Minimum uptime of units g
T ¯ o f f , g Minimum downtime of units g
T ¯ c o l d , g Cold startup time of units g
P ¯ g Maximum power output of units g
P ¯ g Minimum power output of units g
P D , t System load demand in period t
R t Spinning reserve requirement in period t
P u p , g Ramp-up limit of unit g
P d o w n , g Ramp-down limit of unit g
P s t a r t , g Startup ramp limit of unit g
P s h u t , g Shutdown ramp limit of unit g
u g , 0 Initial commitment state of unit g
T g , 0 Number of periods unit g has been online(+) or offline(−) prior to the first period of the time span (end of period 0)
U g m a x ( 0 , m i n ( T , u g , 0 ( T ¯ o n , g T g , 0 ) )
L g m a x ( 0 , m i n ( T , u g , 0 ( T ¯ o f f , g + T g , 0 ) )
Binary variables:
u g , t Commitment status of unit g in period t , equal to 1 if unit g  is online in period t and 0 otherwise
s g , t Startup status of unit g in period t , equal to 1 if unit g starts up in period t and 0 otherwise
d g , t Shutdown status of unit g in period t , equal to 1 if unit g shuts down in period t and 0 otherwise
Continuous variables:
P g , t Power output of unit g in period t
S g , t Startup cost of unit g in period t
Binary variable states:
r LP relaxation solution of binary variables
r ¯ Rounded LP relaxation solution of binary variables
s t a b l e Stable confidence of binary variables
X GNN output of binary variables
x All binary variables include commitment, startup, and shutdown status
Thresholds:
τ s t a b l e Stability threshold
τ v Integrality threshold
Temporal flexibility threshold

References

  1. Anjos, M.F.; Conejo, A.J. Unit commitment in electric energy systems. Found. Trends Electr. Energy Syst. 2017, 1, 220–310. [Google Scholar] [CrossRef]
  2. Hua, B.; Baldick, R.; Wang, J. Representing Operational Flexibility in Generation Expansion Planning Through Convex Relaxation of Unit Commitment. IEEE Trans. Power Syst. 2017, 33, 2272–2281. [Google Scholar] [CrossRef]
  3. Ongsakul, W.; Petcharaks, N. Unit commitment by enhanced adaptive Lagrangian relaxation. IEEE Trans. Power Syst. 2004, 19, 620–628. [Google Scholar] [CrossRef]
  4. Delarue, E.; Cattrysse, D.; D’Haeseleer, W. Enhanced priority list unit commitment method for power systems with a high share of renewables. Electr. Power Syst. Res. 2013, 105, 115–123. [Google Scholar] [CrossRef]
  5. Hobbs, W.J.; Hermon, G.; Warner, S.; Shelbe, G.B. An enhanced dynamic programming approach for unit commitment. IEEE Trans. Power Syst. 1988, 3, 1201–1205. [Google Scholar] [CrossRef]
  6. Gurobi Optimization, L. Gurobi Optimizer Reference Manual. IEEE Trans. Autom. Control 2024, 28, 1–11. [Google Scholar]
  7. Ibm, L. ILOG CPLEX. Available online: http://www.ilog.com/products/cplex (accessed on 1 December 2024).
  8. Bendotti, P.; Fouilhoux, P.; Rottner, C. On the complexity of the unit commitment problem. Ann. Oper. Res. 2019, 274, 119–130. [Google Scholar] [CrossRef]
  9. Yang, L.; Zhang, C.; Jian, J.; Meng, K.; Xu, Y.; Dong, Z. A novel projected two-binary-variable formulation for unit commitment in power systems. Appl. Energy 2017, 187, 732–745. [Google Scholar] [CrossRef]
  10. Harjunkoski, I.; Giuntoli, M.; Poland, J.; Schmitt, S. Matheuristics for Speeding Up the Solution of the Unit Commitment Problem. In Proceedings of the 2021 IEEE PES Innovative Smart Grid Technologies Europe (ISGT Europe), Espoo, Finland, 18–21 October 2021; pp. 1–5. [Google Scholar]
  11. Pourahmadi, F.; Kazempour, J. Unit Commitment Predictor with a Performance Guarantee: A Support Vector Machine Classifier. IEEE Trans. Power Syst. 2024, 40, 715–727. [Google Scholar] [CrossRef]
  12. Qu, S.; Yang, Z. Optimality Guaranteed UC Acceleration via Interactive Utilization of Adjoint Model. IEEE Trans. Power Syst. 2024, 39, 5191–5203. [Google Scholar] [CrossRef]
  13. Sayed, A.R.; Zhang, X.; Wang, G.; Wang, Y.; Shaaban, M.; Shahidehpour, M. Deep Reinforcement Learning-Assisted Convex Programming for AC Unit Commitment and Its Variants. IEEE Trans. Power Syst. 2024, 39, 5561–5574. [Google Scholar] [CrossRef]
  14. Chen, Z.; Liu, J.; Wang, X.; Yin, W. On Representing Mixed-Integer Linear Programs by Graph Neural Networks. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  15. Schmitt, S.; Harjunkoski, I.; Giuntoli, M.; Poland, J.; Feng, X. Fast Solution of Unit Commitment Using Machine Learning Approaches. In Proceedings of the 2022 IEEE 7th International Energy Conference (ENERGYCON), Riga, Latvia, 9–12 May 2022; pp. 1–6. [Google Scholar]
  16. Quan, R.; Jian, J.; Mu, Y. Tighter relaxation method for unit commitment based on second-order cone programming and valid inequalities. Int. J. Electr. Power Energy Syst. 2014, 55, 82–90. [Google Scholar] [CrossRef]
  17. Ma, Z.; Zhong, H.; Xia, Q.; Kang, C.; Wang, Q.; Cao, X. A Unit Commitment Algorithm with Relaxation-Based Neighborhood Search and Improved Relaxation Inducement. IEEE Trans. Power Syst. 2020, 35, 3800–3809. [Google Scholar] [CrossRef]
  18. Huang, T.; Ferber, A.; Tian, Y.; Dilkina, B.; Steiner, B. Local branching relaxation heuristics for integer linear programs. In Proceedings of the International Conference on Integration of Constraint Programming Artificial Intelligence, and Operations Research, Nice, France, 29 May–1 June 2023; pp. 96–113. [Google Scholar]
  19. Gao, Q.; Yang, Z.; Li, W.; Yu, J. Optimality-Guaranteed Acceleration of Unit Commitment Calculation via Few-shot Solution Prediction. IEEE Trans. Power Syst. 2024, 40, 1583–1595. [Google Scholar] [CrossRef]
  20. Pineda, S.; Morales, J.M. Is learning for the unit commitment problem a low-hanging fruit? Electr. Power Syst. Res. 2022, 207, 107851. [Google Scholar] [CrossRef]
  21. Xavier, Á.S.; Qiu, F.; Ahmed, S. Learning to solve large-scale security-constrained unit commitment problems. INFORMS J. Comput. 2021, 33, 739–756. [Google Scholar] [CrossRef]
  22. Yang, N.; Yang, C.; Wu, L.; Shen, X.; Jia, J.; Li, Z.; Chen, D.; Zhu, B.; Liu, S. Intelligent data-driven decision-making method for dynamic multisequence: An E-seq2seq-based SCUC expert system. IEEE Trans. Ind. Inform. 2021, 18, 3126–3137. [Google Scholar] [CrossRef]
  23. Ramesh, A.V.; Li, X. Feasibility Layer Aided Machine Learning Approach for Day-Ahead Operations. IEEE Trans. Power Syst. 2024, 39, 1582–1593. [Google Scholar] [CrossRef]
  24. Park, S.; Chen, W.; Han, D.; Tanneau, M.; Van Hentenryck, P. Confidence-Aware Graph Neural Networks for Learning Reliability Assessment Commitments. IEEE Trans. Power Syst. 2024, 39, 3839–3850. [Google Scholar] [CrossRef]
  25. Han, Q.; Yang, L.; Chen, Q.; Zhou, X.; Zhang, D.; Wang, A.; Sun, R.; Luo, X. A GNN-Guided Predict-and-Search Framework for Mixed-Integer Linear Programming. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  26. Qin, J.; Yu, N. Solve Large-scale Unit Commitment Problems by Physics-Informed Graph Learning. arXiv 2023, arXiv:2311.15216. [Google Scholar]
  27. Gasse, M.; Chételat, D.; Ferroni, N.; Charlin, L.; Lodi, A. Exact Combinatorial Optimization with Graph Convolutional Neural Networks. arXiv 2019, arXiv:1906.01629. [Google Scholar]
  28. Gridmod. RTS-GMLC: Repository for Grid Modernization Lab Consortium (GMLC) Real-Time Simulation Project. In GitHub. 2023. Available online: https://github.com/GridMod/RTS-GMLC (accessed on 1 December 2024).
  29. Berthold, T. Primal Heuristics for Mixed Integer Programs. Master’s Thesis, Technische Universität, Berlin, Germany, 2006. [Google Scholar]
Figure 1. Overall framework of proposed acceleration method.
Figure 1. Overall framework of proposed acceleration method.
Applsci 15 04498 g001
Figure 2. Transforming a UCP formulation to a bipartite graph.
Figure 2. Transforming a UCP formulation to a bipartite graph.
Applsci 15 04498 g002
Figure 3. Dataset flowchart.
Figure 3. Dataset flowchart.
Applsci 15 04498 g003
Figure 4. Architecture of the bipartite GNN policy π θ .
Figure 4. Architecture of the bipartite GNN policy π θ .
Applsci 15 04498 g004
Figure 5. The primal gap as a function of runtime, averaged over 36 instances. From left to right, the units are 50, 100, and 1080.
Figure 5. The primal gap as a function of runtime, averaged over 36 instances. From left to right, the units are 50, 100, and 1080.
Applsci 15 04498 g005
Figure 6. The primal gap as a function of runtime, averaged over 36 instances. From left to right, the units are 50, 100, and 1080.
Figure 6. The primal gap as a function of runtime, averaged over 36 instances. From left to right, the units are 50, 100, and 1080.
Applsci 15 04498 g006
Table 2. Stable calculation.
Table 2. Stable calculation.
Index123456
x 0 0.50.30001
x 1 100101
x 2 010101
x 3 000101
x N 001001
s t a b l e 0.40.480.60.211
Table 1. Selected features for variables, constraints, and edges.
Table 1. Selected features for variables, constraints, and edges.
Feature Description
Variablesnormalized coefficient of variables in the objective function
degree of a variable node in the bipartite representation
average coefficient of the variable in all constraints
maximum value among all coefficients of the variable
minimum value among all coefficients of the variable
whether the variable is a binary variable
LP solution of the variable
Constraintsaverage of all coefficients in the constraint
degree of constraint nodes in the bipartite representation
right-hand-side value of the constraint
whether the constraint is an equation constraint
Edgescoefficient value
Table 3. Test system summary.
Table 3. Test system summary.
Case SizeVariableIntegerConstraintNonzero
506000360010,84850,678
10012,000720021,648101,349
1080129,60077,760233,3281,109,492
Table 4. Variable loss comparison.
Table 4. Variable loss comparison.
Variable Type501001080
u 0.047730.028010.06902
s 0.009510.007650.02248
d 0.009230.006460.02100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, L.; Li, P.; Chen, S.; Zheng, H. Stable Variable Fixation for Accelerated Unit Commitment via Graph Neural Network and Linear Programming Hybrid Learning. Appl. Sci. 2025, 15, 4498. https://doi.org/10.3390/app15084498

AMA Style

Yang L, Li P, Chen S, Zheng H. Stable Variable Fixation for Accelerated Unit Commitment via Graph Neural Network and Linear Programming Hybrid Learning. Applied Sciences. 2025; 15(8):4498. https://doi.org/10.3390/app15084498

Chicago/Turabian Style

Yang, Linfeng, Peilun Li, Shifei Chen, and Haiyan Zheng. 2025. "Stable Variable Fixation for Accelerated Unit Commitment via Graph Neural Network and Linear Programming Hybrid Learning" Applied Sciences 15, no. 8: 4498. https://doi.org/10.3390/app15084498

APA Style

Yang, L., Li, P., Chen, S., & Zheng, H. (2025). Stable Variable Fixation for Accelerated Unit Commitment via Graph Neural Network and Linear Programming Hybrid Learning. Applied Sciences, 15(8), 4498. https://doi.org/10.3390/app15084498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop