Next Article in Journal
Model-Based Design and Evaluation of State-of-the-Art Thermal Management Systems for Electrified Trucks
Previous Article in Journal
Frequency Regulation for High Wind Penetration Power System Based on Ocean Predator Algorithm Considering Storage Battery State
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Advanced Spatio-Temporal Graph Neural Network Framework for the Concurrent Prediction of Transient and Voltage Stability

1
State Grid Fujian Electric Power Research Institute, Fuzhou 350007, China
2
Fujian Key Laboratory of Smart Grid Protection and Operation Control, Fuzhou 350007, China
3
State Grid Fujian Electric Power Co., Ltd. Fuzhou Company, Fuzhou 350007, China
4
The School of Electrical Engineering, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Energies 2025, 18(3), 672; https://doi.org/10.3390/en18030672
Submission received: 29 December 2024 / Revised: 20 January 2025 / Accepted: 22 January 2025 / Published: 31 January 2025
(This article belongs to the Section F: Electrical Engineering)

Abstract

Power system stability prediction leveraging deep learning has gained significant attention due to the extensive deployment of phasor measurement units. However, most existing methods focus on predicting either transient or voltage stability independently. In real-world scenarios, these two types of instability often co-occur, necessitating distinct and coordinated control strategies. This paper presents a novel concurrent prediction framework for transient and voltage stability using a spatio-temporal embedding graph neural network (STEGNN). The proposed framework utilizes a graph neural network to extract topological features of the power system from adjacency matrices and temporal data graphs. In contrast, a temporal convolutional network captures the system’s dynamic behavior over time. A weighted loss function is introduced during training to enhance the model’s ability to handle instability cases. Experimental validation on the IEEE 118-bus system demonstrates the superiority of the proposed method compared to single stability prediction approaches. The STEGNN model is further evaluated for its prediction efficiency and robustness to measurement noise. Moreover, results highlight the model’s strong transfer learning capability, successfully transferring knowledge from an N-1 contingency dataset to an N-2 contingency dataset.

1. Introduction

The continuous growth in global electricity demand drives power systems to operate under increasingly high-load conditions, often nearing their stability thresholds [1]. This trend imposes substantial pressure on the reliability and resilience of modern power grids, making it imperative to address the challenges associated with maintaining stable operations. The integration of renewable energy sources (RESs), such as wind and solar power, further complicates this landscape due to their inherent intermittency and variability. While these sources offer significant environmental advantages, they introduce dynamic behaviors that, if not carefully managed, can undermine grid stability. Moreover, the widespread adoption of power electronic devices exacerbates these challenges by introducing rapid and nonlinear dynamics that may interact detrimentally with the grid [2]. Consequently, ensuring the stability of power systems under such conditions has become a paramount concern for researchers and practitioners. Accurate prediction of power system stability has become increasingly vital, as it enables operators to implement timely and effective control measures, thereby ensuring reliable electricity delivery and mitigating the risk of cascading failures [3].
Traditional model-based stability prediction methods can be categorized into time-domain simulation and energy function approaches [4]. Time-domain methods typically use low-order models to simulate system stability. For instance, C.-W. Liu and J. S. Thorp employed an implicit decoupling PQ method for stability prediction while utilizing a linear model for parallel simulation approach [5] to predict stability in the IEEE 145-bus system [6]. Time-domain simulation methods require detailed system models to solve differential or algebraic equations. However, with the continuous expansion of modern power systems and the integration of RESs, obtaining detailed models becomes impractical, and the solution process is time-consuming [2]. Energy function methods are generally based on Lyapunov’s second law. The Lyapunov exponent’s out-of-step technique [7] is used to predict system stability and energy functions [8] are employed to estimate stability margins for predicting stability in the IEEE 39-bus system. In large-scale power systems, determining an appropriate Lyapunov function is challenging, and the computation of system energy is highly time-consuming [9]. However, traditional methods have significant limitations. They require precise power system models, and inaccuracies can lead to unreliable stability analysis. Additionally, their computational time increases with system size, making them unsuitable for real-time applications in modern, large-scale power systems.
With the deployment of phasor measurement units (PMUs) and the development of wide-area monitoring systems (WAMS), machine learning (ML) has been extensively applied to power system stability prediction. ML methods are data-driven and model-free, making them well-suited for online use, where fast response times and high accuracy are required. A.Karami employed a multi-layer perceptron (MLP) to estimate the transient stability margin of power systems [10]. Convolutional neural networks (CNNs) are utilized for transient stability prediction [11]. Machine learning techniques are applied to predict voltage stability in power systems [12], while G. Wang et al. leveraged graph convolutional networks (GCN) and long short-term memory (LSTM) networks for online voltage stability prediction [13]. ML methods do not rely on the detailed modeling of power systems and possess strong generalization capabilities, making them highly promising for online diagnostics [14]. Farheen Bano et al. analyzed microgrid voltage stability using a variety of machine learning models, including random forest, gradient boosting, K-means, hierarchical clustering, and regression models [15].
Short-term power system instability includes transient and voltage instability [16]. Transient instability occurs when, after a large disturbance, the generators in the power system fail to maintain synchronous operation, causing the system to become unstable. Voltage instability occurs when a disturbance leads to a significant voltage drop, potentially resulting in voltage collapse. When a fault occurs in the system, transient and voltage instability often manifest simultaneously, or the system may transition from one instability mode to another. Different instability types require distinct control strategies [16]. For transient stability, the objective is to maintain generator synchronization, and control strategies typically focus on the generator rotor, such as excitation control or early disconnection of critical generators [2]. For short-term voltage stability, the focus is on controlling reactive power, which can be achieved by deploying reactive power compensation devices, such as capacitors, reactors, or emergency load shedding [17]. However, achieving precise classification of power system stability remains a complex challenge [18]. For instance, when the rotor angles of two generators gradually separate, the electrical distance between them increases, causing voltage drops at intermediate load nodes. This issue, caused by transient power angle instability, cannot be effectively resolved by reactive power compensation alone at the critical load node [19]. Therefore, precise classification of power system stability behaviors is crucial for successful operation.
Although current deep learning-based methods for single instability prediction perform relatively well, they focus on individual stability, such as transient or voltage stability, often overlooking the interrelated nature of instability behaviors. The coupling between phase angle and voltage instabilities is well-established, and neglecting this aspect can lead to incomplete or inaccurate control strategies. Moreover, deep learning models face the challenge of imbalanced data, as instability events are rare in power system operations. This imbalance often results in overfitting, reducing model generalizability to unseen scenarios. Current methods have not sufficiently addressed this issue, which limits their robustness in real-world applications. Furthermore, the practical applicability of these methods under more complex conditions, such as N-2 contingencies, remains underexplored. Given the significant impact of such contingencies on system stability, further research is necessary to improve the reliability and scalability of deep learning-based approaches in handling these complex scenarios. This paper proposes a co-prediction of transient and voltage stability based on a spatio-temporal embedding graph neural network (STEGNN). The main contributions of this paper are as follows:
(1) The STEGNN model proposed combines power system topology with PMU historical data to co-predicate transient and voltage stability, enabling a comprehensive analysis of power system dynamic stability for more precise operational decisions in the event of system instability.
(2) In power system operation data, instability samples constitute a small proportion. During the training process, a weighted loss function is adopted to emphasize unstable samples, effectively addressing the data imbalance issue, which enhances the model’s training stability and prevents overfitting.
(3) The proposed STEGNN model is designed for real-time applications, ensuring that it balances prediction accuracy and computational efficiency. By selecting an optimal observation window length, the model provides timely and accurate transient and voltage stability predictions, making it suitable for real-time monitoring and control in power systems.
(4) The STEGNN model is pre-trained on the N-1 contingency dataset and successfully applied to the N-2 contingency dataset, validating its effectiveness in handling various contingency scenarios and demonstrating its strong transferability.
The remainder of the paper is structured as follows: Section 2 provides a detailed exploration of the STEGNN model. Section 3 presents the experimental results from various case studies. Finally, Section 4 offers the conclusions.

2. Spatiotemporal Embedding Graph Neural Network

We propose a STEGNN framework for co-predicting the transient and voltage stability assessment. The structure of the STEGNN framework is presented below. STEGNN mainly consists of three parts: the multi-head GCN modules as topological features extractors, the temporal convolution networks (TCN) module as a temporal features extractor, and the MLP layer as the classifier. The co-prediction framework and the STEGNN model are shown in Figure 1. The structure has three main parts: the GC module, TC module, and MLP prediction layer. Each GC module has five neurons, where h i and h i (i = a to e) represent the input states of the five neurons in each GC module. The operation of the GC module is to update the output state of the focused neurons (red) using the adjacency matrix, which is only relevant to the adjacent neurons (blue) rather than others (gray). The topological features are then extracted by the FC layer and processed by the TC module, which is composed of R residual blocks with dilated factors d i to d R . In the end, the MLP layer generates the prediction probability function.

2.1. Preliminary DL Techniques

The framework of the proposed STEGNN model consists of GC and TC modules based on GCN [20] and TCN [21]. The fundamental concepts of GCN and TCN provide a foundation for understanding the STEGNN model. MLP is the classifier for the transient stability prediction.
GCN structure: The structure of the GCN layer can be represented as an undirected graph G = ( V , E , B ) , where V R N is the set of neurons, E R E is the set of neuron connections, and B R N × N is the adjacency matrix [22]. To represent the neuron connections in a GCN layer, the re-normalized adjacency matrix B ^ is often used:
B ^ = D ^ 1 / 2 B ^ D ^ 1 / 2 ,
where B ^ = B + I N is the adjacency matrix with self-loops, and I N is the identity matrix. D ^ is the diagonal degree matrix, where D ^ i , j = j B ^ i , j . The core idea of a GCN is to map a P × R input matrix to a new P × S output feature matrix for the nodes in the graph, where R and S represent the input and output feature dimensions, respectively. The layer function of GCN is defined as follows:
H = σ ( f ( B ^ , X ) ) = σ ( B ^ X W + b ) ,
where X R P × R , H R P × S , W R R × S , and b R P × S represent the input, output, network weights, and bias, respectively. σ ( · ) is the ReLU activation function.
TCN structure: The TCN model consists of sequential residual blocks, each containing a 1D fully convolutional network (FCN) [23] that uses causal and dilated convolutions and a residual connection. The causal convolution ensures that the output has the same length as the input and considers the entire history while preserving the causality of the sequence. Dilated convolutions are employed to address the depth burden introduced by causal convolutions. Dilated convolutions expand the receptive field exponentially, allowing the model to consider a larger portion of the history with a smaller depth. Specifically, the dilated convolution operation F ( · ) applied to the j-th element x j in the 1D sequence x R n , using a filter f : { 0 , , k 1 } R , is defined as follows:
F ( j ) = i = 0 k 1 f ( i ) · x j d · i ,
where k is the filter size and d is the dilation factor. The dilation factor expands the receptive field, allowing the model to capture a broader input range than standard convolutions. Residual connections stabilize deep causal dilated convolutions, enabling the layers to learn adjustments to the identity mapping. Additionally, a 1 × 1 convolution layer, denoted as Conv 1 D ( · ) , is introduced to resolve any dimensional mismatches between the input and output. Consequently, the output of a residual block is defined as follows:
o = σ Conv 1 D ( x + LF ( F ( x ) ) ) ,
where x and o are the input and output of the residual block, and LF ( · ) refers to layer normalization.
MLP structure: An MLP consists of an input layer, one or more hidden layers, and an output layer.
Input layer: The input layer accepts the input vector x = [ x 1 , x 2 , , x n ] .
Hidden layers: Each hidden layer performs a linear transformation followed by a nonlinear activation function. The output h ( l ) of the l-th hidden layer is defined as follows:
h ( l ) = σ ( W ( l ) h ( l 1 ) + b ( l ) ) ,
where W ( l ) is the weight matrix, b ( l ) is the bias vector, and σ represents the activation function (e.g., ReLU, sigmoid, and tanh).
Output layer: The output layer generates the final output vector y :
y = W ( L ) h ( L 1 ) + b ( L ) ,
where L is the total number of layers.

2.2. STEGNN Framework

The proposed STEGNN model extracts the topological spatial features from the multivariate physical PMU data input using GC modules, composed of two GCN layers and one fully connected (FC) layer with relu activations [24].
H = σ ( B ^ σ ( B ^ X W 1 + b 1 ) W 2 + b 2 ) .
Then, the TC module follows the GC modules, located after the convenience of l topological features extracted by GC modules, which were introduced in the former part. The last one is the MLP module after the TC module. MLP modules are adopted to integrate the topological and temporal features extracted by both the GC and TC modules. Finally, the softmax activation generates the prediction as a probability function.
Class-weighted loss function: Since STEGNN’s transient stability prediction gives a multi-stability probability, the class-weighted category cross-entropy (CCE) is the loss function with the L 2 regularization. Then, the loss function is defined as follows:
L o s s = i = 1 b a t c h j = 1 C α j y i j log p i j + β k = 1 M 1 2 ( w k 2 +   b k 2 ) ,
where y i j and p i j denote the label and the model output of the jth category sample, and α j represents category weight factors, j = 1 , 2 , , C . Additionally, w k and b k serve as learnable network parameters, where β is the regularization weight.
Grid-informed adjacency matrix: In spatial feature extraction within the GC modules, three distinct grid-informed adjacency matrices, B , have been conceptualized, considering a power system’s electrical and structural attributes. Initially, drawing from reference, a binary adjacency matrix with self-connections for a power system is selected (referred to as B I ). The elements of this matrix are defined by Equation (9), which accounts for the system’s topological structure while disregarding the edge weights. Subsequently, the concept of B I I is introduced, where the active power flows are integrated as weights into the adjacency matrix, as articulated in Equation (10). Lastly, the adjacency matrix B I I I is formulated by incorporating the maximum transmission capacity of transmission lines as off-diagonal elements, complemented by the active power injections for the diagonal elements, as delineated by Equation (11). This approach comprehensively represents the power system’s spatial characteristics for enhanced feature extraction.
B i j I = 1 ( i , j ) E p norm or i = j 0 ( i , j ) E p norm or ( i , j ) E p con ,
B i j I I = K i j sin ( δ i δ j ) ( i , j ) E p norm 0 ( i , j ) E p norm or ( i , j ) E p con ,
B i j I I I = K i j ( i , j ) E p norm 0 ( i , j ) E p norm or ( i , j ) E p con P i i = j ,
where E p norm denotes the ensemble of transmission lines operating under normal conditions. In contrast, E p con represents the set of transmission lines that are compromised during contingencies. The notation ( i , j ) signifies the transmission line interconnecting nodes i and j within the network. The variable K i j corresponds to the maximum permissible transmission capacity for the line connecting these two nodes. Additionally, P i and P j represent the active power injections at nodes i and j, respectively. If a transmission line ( i , j ) is an element of the set E p con , it is characterized as being in a state of malfunction during the contingency scenario. This classification is crucial for analyzing power system resilience and developing strategies to mitigate the impact of such disruptions.

2.3. Data Representation and Label Definition

The inputs to the proposed co-prediction framework are the feature matrix of transient dynamics time series data:
x i = P M U 1 i ( 0 ) P M U 2 i ( 0 ) P M U N i ( 0 ) P M U 1 i ( 1 ) P M U 2 i ( 1 ) P M U N i ( 1 ) P M U 1 i ( l 1 ) P M U 2 i ( l 1 ) P M U N i ( l 1 ) T ,
where i = 1 , 2 , . . . , F , F is the number of features, and N is the number of nodes. The jth column of x i is the post-fault time series with length l describing one of the PMU data streams P M U j i of node j, j = 1 , 2 , , N in power systems. Thus, we choose the bus relative phase, bus voltage magnitude, rotor angles, and rotor speed as the second input X = { U θ , | U | , θ , ω } , where F = 4 .
The transient stability of power systems is decided by the transient stability index (TSI) σ as [11]
σ = 180 | Δ δ i j | m a x 180 + | Δ δ i j | m a x ,
where | Δ δ i j | m a x is the maximum absolute value of the difference between the rotor angles of any two synchronous motors during the simulation process. The y T is also indicated by the sign of σ ,
y T = 0 σ > 0 1 σ 0 .
The voltage stability of power system depends on the voltage magnitude dip V dip at any bus within the evaluation period T duration [25].
y V = 0 , if V dip < 20 % within T duration 1 , otherwise ,
where T duration is typically defined as 50 cycles or 1000 ms, with 1000 ms adopted in this study.
As shown in Table 1, the label y for supervising the model training is a one-hot variable, i.e., four classes are considered according to the stability condition of transient stability label y T and voltage stability label y V . The prediction of the STEGNN model, p = { p 1 , p 2 , p 3 , p 4 } [ 0 , 1 ] 4 , is the probability that the power system will evolve to one of the four stability conditions after the softmax calculation of the last layer output of the proposed network:
p i = e z i k = 1 4 e z k ,
The power system stability co-prediction process and the detailed STEGNN model are depicted in Figure 1. The proposed framework utilizes online PMU data to derive the feature matrix X , as defined in Equation (12), and labels the stability states according to Table 1. The adjacency matrix B is constructed based on the power system topology, as defined in Equations (9)–(11). The inputs X and B are fed into the STEGNN model, where a weighted loss function is used for training, as described in Equation (8). During testing, one-hot encoding is applied to represent the predicted stability states of the power system.

3. Case Studies

The experimental results of the IEEE 118-bus power system are used to evaluate the proposed STEGNN model. We evaluate the model’s performance on the stability co-prediction and benchmark it with a single stability prediction. Additionally, we discuss the impact of grid-informed adjacency matrices and the time length of input PMU data on the STEGNN model’s performance. We also discuss the model’s robustness to noise. N-1 contingencies are considered, and the transient fault dynamics of the power system are simulated in the PSS/E program as follows:
(1)
The consumed power randomly varies between 80% and 120% of the base load levels, repeated for 50 iterations;
(2)
The load flow equation is solved to adjust the generator outputs accordingly;
(3)
Three phase-to-ground faults are applied at all 118 buses and the ends of 177 transmission lines, with random fault clearance times of 0.3, 0.4, 0.5, or 0.6 s;
(4)
The simulation is run for 5 s, and both transient stability and voltage stability labels are set.
Consequently, 59,000 samples are generated, and the number of four classes is 35381, 6389, 701 and 16,529, where the proportions of the training set, validation set, and test set are 80%, 10%, and 10%, respectively.
The confusion matrix is a valuable tool for evaluating the prediction model, defining four key values based on actual and predicted outcomes: true positive ( T P ), false positive ( F P ), true negative ( T N ), and false negative ( F N ). Here, T P or T N represents the instances where the model correctly predicts the positive or negative class. At the same time, F P or F N corresponds to instances where the model incorrectly predicts the negative or positive class. Stable (positive) and unstable (negative) are used interchangeably. To assess the performance of the STEGNN model, four metrics are employed: accuracy ( A C C ), false positive rate ( F P R ), false negative rate ( F N R ), and F-score ( F s c o r e ) where
A C C = T P + T N T P + T N + F P + F N ,
F P R = F P F P + T N ,
F N R = F N F N + T P ,
F s = ( 1 + γ 2 ) · P r e c i s i o n · R e c a l l γ 2 P r e c i s i o n + R e c a l l ,
where P r e c i s i o n = T P / ( T P + F P ) denotes the fraction of T P among the models classified as positive class, and R e c a l l = T P ( T P + F N ) denotes the fraction of T P among the total number of positive samples. While A C C , F P R , and F N R can reveal whether the predictions are good or not, F s c o r e could evaluate the prediction of the model of imbalanced samples more comprehensively, for it indicates how much more important recall is than precision or vice versa. We set γ = 1 in this paper. The STEGNN model is trained on the NVIDIA GeForce RTX 4090 platform with the Tensorflow 2.3.1 environment and the model training hyperparameters are shown in Table 2.

3.1. Concurrent Prediction Performance

ACC and loss curves during the training and validation process are shown in Figure 2a. ACC and loss denote the validation dataset’s prediction accuracy rate and class-weighted CCE. The confusion matrix for the co-prediction is given in Figure 2b.
To evaluate the effectiveness of the STEGNN model, a performance comparison is conducted between the proposed method and four existing deep learning models: MLP, CNN, TCN, and GCN. Concurrent transient stability prediction (Co-TSP) and voltage stability prediction (Co-VSP) are performed on the IEEE-118 bus power system using identical training, evaluation, and test datasets, as well as the same experimental setup. The results are presented in Table 3. As shown in Table 3, the STEGNN model outperforms the other methods in both transient stability and voltage stability prediction and the results of prediction are shown in bold.
In the IEEE 118-bus power system, the dataset of transmission line ground faults at 0.2 s (TLGF-0.2) has a size of 1770 and simultaneously has transient and voltage stability status labels. Compared to the concurrent prediction model trained with the dataset of the joint labels of transient stability and voltage stability, the single TSP and VSP prediction models are trained with the datasets with the single stability labels for the transient stability and voltage stability, respectively. In Table 4, it can be observed that the concurrent STEGNN prediction model, co-prediction, can predict the TSP and VSP simultaneously for the TLGF-0.2 dataset, while the combined prediction of the single TSP and single VSP respectively recorded 23 and 38 incorrect predictions and 54 errors in total for TSP and VSP. In Table 5 and Table 6, the co-TSP and co-VSP predictions perform better than the single TSP and single VSP. As for the joint prediction, the concurrent STEGNN model can make the perfect prediction for the TLGF-0.2 test dataset. The concurrent prediction method demonstrates superior performance in transient stability, voltage stability, and co-prediction. The results of the joint prediction in Table 4, Table 5 and Table 6 are all optimal, and are presented in bold.
Figure 3 provides a visualization of the high-dimensional features extracted from the hidden layers of the STEGNN model. The training and testing datasets samples are color-coded according to their true labels and predictions, visually assessing the model’s learning process. In the two subplots in the first row of Figure 3, the training and testing samples are colored based on their true labels, while the second row shows the same datasets colored according to the predicted output with the STEGNN model. The two subplots in the third row of Figure 3 show the prediction error. The results in Figure 3 demonstrate that the STEGNN model successfully learns the underlying features of the dataset and makes accurate predictions.

3.1.1. Effect of Time Length

The observation window length of post-fault PMU data directly impacts the real-time application of the proposed STEGNN model. A longer observation window provides more system dynamics information, which enhances prediction ( A C C ), as shown in Table 7. However, it also increases computational training time, as illustrated in Figure 4. While the training time itself does not affect real-time applications, it may become significant when scaling to larger power systems, increasing the overall computational cost.
As depicted in Figure 4, when the time length is less than 0.3 s, the A C C shows a steady increase. Beyond 0.3 s, the A C C plateaus, providing diminishing returns despite the additional observation data. To balance the demands of real-time application and predictive performance, a time length of 0.3 s is chosen and the prediction effect for time length of 0.3s is shown in bold. This selection ensures high prediction accuracy while limiting the computational cost associated with larger power systems, thereby maintaining the practical feasibility of the model.

3.1.2. Sensitivity Analysis

For the co-prediction of transient stability and voltage stability in practical power systems, the presence of noise in PMU data significantly impacts the prediction model performance. The standard deviation of noise in PMU data typically ranges from 0.0005 to 0.01, corresponding to a signal-to-noise ratio (SNR) of approximately 45 dB [26]. Table 8 presents the co-prediction performance of the IEEE 118-bus power system under various SNR levels. Even at an SNR of 50 dB (slightly lower than typical levels), the model maintains high accuracy, with the ACCs of transient stability prediction (TSP) and voltage stability prediction (VSP) decreasing by only 0.01% and 0.03%, respectively. Under high-noise conditions at an SNR of 20 dB, the prediction performance exhibits a moderate decline, with ACCs decreasing by 0.14% and 1.22%. Additionally, other evaluation metrics also show degradation as the SNR decreases for co-TSP and co-VSP. These results demonstrate that STEGNN exhibits strong robustness against noise.

3.2. Performance in N-2 Contingency

The N-2 dataset, obtained from simulating three-phase grounding at two random buses in the IEEE 118-bus system, contains 29,500 samples. We evaluated and compared the ACC of co-prediction, single prediction, and co-prediction without a weighted loss function based on the STEGNN model. Figure 5 shows that co-prediction performs best transferability for the N-2 dataset. All methods achieve high VSP accuracy, with the main differences observed in TSP. The ACC of co-prediction surpasses that of single prediction by 22.4%. It indicates a coupling relationship between transient and voltage stability and co-prediction significantly improves model transferability. Additionally, the ACC of co-prediction exceeds that of co-prediction without weighted loss by 32.3%, highlighting the importance of a weighted loss function for imbalanced datasets.

3.3. Effect of Grid-Informed Adjacency Matrices

In Figure 6, three grid-informed adjacency matrices are introduced, incorporating the power grid’s structural and electrical parameter information. The grid-informed adjacency matrices are introduced in Figure 6, where the colors of each small pixel block represent the corresponding element value of the four adjacency matrices. The corresponding prediction performances are shown in Table 9. It can be observed from Table 9 that the adjacency matrix B I I shows the worst performance. This can be explained by the very sparse adjacency matrix visualization in Figure 6, which indicates that although B I I contains information on active power flow distribution, other useful information about the power system topology and electrical properties is discarded. Under the same training conditions, B I I I demonstrates superior performance compared to B I . Therefore, the STEGNN model uses B I I I as the power grid information adjacency matrix for the co-prediction of the power system stability.
Table 9 presents a comparison of the joint prediction performance of the concurrent STEGNN model as well as the performance of co-TSP and co-VSP using three different grid-informed adjacency matrices for the IEEE-118 bus power system and the predictions of grid-informed adjacency matrice using B I I I are shown in bold. The performance is evaluated using three metrics: A C C , F P R , and F N R . The joint prediction of TSP and VSP is 99.28%. The co-TSP and co-VSP perform very well. The A C C of the co-TSP and co-VSP are 99.97% and 99.31%, respectively.

3.4. Effect of Weighted Loss

Figure 7 illustrates the impact of weighted loss on the STEGNN model during training. Figure 7a shows the ACC of co-prediction, co-TSP, and co-VSP. It can be observed that the ACC remains high whether weighted loss is applied, which, to some extent, demonstrates the superiority of the STEGNN model. The ACC is higher when the weighted loss is used. Figure 7b shows the FPR for co-TSP and co-VSP. The results indicate that without weighted loss, the FPR is significantly higher. This suggests when transient or voltage instability occurs, models without weighted loss fail to provide adequate warnings, which is hazardous and unacceptable.

4. Conclusions

In conclusion, the proposed spatiotemporal enhanced graph neural network (STEGNN) model effectively addresses the intricate challenges of concurrent transient and voltage stability prediction in power systems. The model enables comprehensive and accurate stability assessments by seamlessly integrating spatial features derived from power system topology with temporal dynamics captured from PMU data. This dual-perspective approach enhances predictive precision while capitalizing on the interdependencies between transient and voltage stability, resulting in robust and dependable performance.
The STEGNN framework employs a cohesive integration of spatial and temporal information. Spatial features are extracted via a graph convolution (GC) module, while temporal data from PMUs are processed through a temporal convolution (TC) module. This synergistic design allows for a unified and accurate analysis of both stability phenomena. The model’s capability to simultaneously predict transient and voltage stability delivers several distinct advantages:
1. Superior predictive accuracy: The STEGNN framework outperforms traditional single-focus stability prediction methods for transient stability prediction (TSP) and voltage stability prediction (VSP). By jointly analyzing transient and voltage stability, the model harnesses inherent correlations, such as those between rotor angle dynamics and voltage behavior, to produce more precise and reliable predictions.
2. Real-time prediction capability: The STEGNN model facilitates rapid stability assessments using solely PMU data, making it highly suitable for real-time and online applications. This swift response capability is critical for system operators, enabling proactive interventions and preventive measures to mitigate potential instabilities and enhance grid reliability.
3. Robustness and transferability: To address the issue of limited, unstable samples, a weighted loss function is employed to enhance the model’s generalization ability. The model exhibits exceptional resilience to noise in PMU measurements, a critical attribute for real-world implementation in operational power systems where data imperfections are unavoidable. This robustness ensures consistent performance under varying conditions. Furthermore, the STEGNN framework demonstrates strong adaptability, maintaining high accuracy when applied to more complex and dynamic system scenarios.
The STEGNN’s ability to deliver timely and reliable stability predictions establishes it as a vital tool for system operators managing increasingly complex and interconnected power grids. Its robustness to data noise and unified treatment of transient and voltage stability enhance its practical utility, ensuring reliable operation even in challenging environments. By addressing these critical aspects of power system stability, the STEGNN framework contributes significantly to the operational resilience and security of modern and future energy systems, aligning with the evolving demands of smarter, more interconnected grids.

Author Contributions

Conceptualization: C.D., L.D., W.C., J.H. and X.C.; Methodology: C.D., L.D., W.C. and X.C. Resources: J.H., J.W. and L.L.; Formal Analysis: J.W. and W.Q; Validation: L.L., W.Q. and S.L.; Data Analysis: W.Q. and S.L.; Original Draft Preparation: X.C., S.L. and W.Q.; Writing—Review and Editing: X.C., C.D., W.Q. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Science and Technology Project of State Grid Fujian Power Co., Ltd. of China (521304240018), which is named Research on an Artificial Intelligence Joint Prediction Method for Transient Stability Based on Physics Driven Feature Engineering.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

Author Lanxin Lin was employed by the company State Grid Fujian Electric Power Co., Ltd. Fuzhou Company. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Das, S.; Ketan Panigrahi, B. Prediction and control of transient stability using system integrity protection schemes. IET Gener. Transm. Distrib. 2019, 13, 1247–1254. [Google Scholar] [CrossRef]
  2. Zhu, L.; Hill, D.J.; Lu, C. Hierarchical deep learning machine for power system online transient stability prediction. IEEE Trans. Power Syst. 2019, 35, 2399–2411. [Google Scholar] [CrossRef]
  3. Hooshyar, H.; Savaghebi, M.; Vahedi, A. Synchronous generator: Past, present and future. In AFRICON 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1–7. [Google Scholar]
  4. Sobbouhi, A.R.; Vahedi, A. Transient stability prediction of power system; A review on methods, classification and considerations. Electr. Power Syst. Res. 2021, 190, 106853. [Google Scholar] [CrossRef]
  5. Zadkhast, P.; Jatskevich, J.; Vaahedi, E. A multi-decomposition approach for accelerated time-domain simulation of transient stability problems. IEEE Trans. Power Syst. 2014, 30, 2301–2311. [Google Scholar] [CrossRef]
  6. Liu, C.-W.; Thorp, J.S. New methods for computing power system dynamic response for real-time transient stability prediction. IEEE Trans. Circuits Syst. Fundam. Theory Appl. 2000, 47, 324–337. [Google Scholar]
  7. Yan, J.; Liu, C.-C.; Vaidya, U. Pmu-based monitoring of rotor angle dynamics. IEEE Trans. Power Syst. 2011, 26, 2125–2133. [Google Scholar] [CrossRef]
  8. Bhui, P.; Senroy, N. Real-time prediction and control of transient stability using transient energy function. IEEE Trans. Power Syst. 2016, 32, 923–934. [Google Scholar] [CrossRef]
  9. Farantatos, E.; Huang, R.; Cokkinides, G.J.; Meliopoulos, A. A predictive generator out-of-step protection and transient stability monitoring scheme enabled by a distributed dynamic state estimator. IEEE Trans. Power Deliv. 2015, 31, 1826–1835. [Google Scholar] [CrossRef]
  10. Karami, A. Tability margin estimation using neural networks. Int. J. Electr. Power Energy Syst. 2011, 33, 983–991. [Google Scholar] [CrossRef]
  11. Gupta, A.; Gurrala, G.; Sastry, P. An online power system stability monitoring system using convolutional neural networks. IEEE Trans. Power Syst. 2018, 34, 864–872. [Google Scholar] [CrossRef]
  12. Malbasa, V.; Zheng, C.; Chen, P.-C.; Popovic, T.; Kezunovic, M. Voltage stability prediction using active machine learning. IEEE Trans. Smart Grid 2017, 8, 3117–3124. [Google Scholar] [CrossRef]
  13. Wang, G.; Zhang, Z.; Bian, Z.; Xu, Z. A short-term voltage stability online prediction method based on graph convolutional networks and long short-term memory networks. Int. J. Electr. Power Energy Syst. 2021, 127, 106647. [Google Scholar] [CrossRef]
  14. Sarajcev, P.; Kunac, A.; Petrovic, G.; Despalatovic, M. Power system transient stability assessment using stacked autoencoder and voting ensemble. Energies 2021, 14, 3148. [Google Scholar] [CrossRef]
  15. Bano, F.; Rizwan, A.; Serbaya, S.H.; Hasan, F.; Karavas, C.-S.; Fotis, G. Integrating microgrids into engineering education: Modeling and analysis for voltage stability in modern power systems. Energies 2024, 17, 4865. [Google Scholar] [CrossRef]
  16. Kundur, P.; Paserba, J.; Ajjarapu, V.; Andersson, G.; Bose, A.; Canizares, C.; Hatziargyriou, N.; Hill, D.; Stankovic, A.; Taylor, C. Definition and classification of power system stability IEEE/CIGRE joint task force on stability terms and definitions. IEEE Trans. Power Syst. 2004, 19, 1387–1401. [Google Scholar]
  17. ALShamli, Y.; Hosseinzadeh, N.; Yousef, H.; Al-Hinai, A. A review of concepts in power system stability. In Proceedings of the 2015 IEEE 8th GCC Conference & Exhibition, Muscat, Oman, 1–4 February 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar]
  18. Abdi-Khorsand, M.; Vittal, V. Identification of critical protection functions for transient stability studies. IEEE Trans. Power Syst. 2017, 33, 2940–2948. [Google Scholar] [CrossRef]
  19. Atputharajah, A.; Saha, T.K. Power system blackouts-literature review. In Proceedings of the 2009 International Conference on Industrial and Information Systems (ICIIS), Peradeniya, Sri Lanka, 28–31 December 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 460–465. [Google Scholar]
  20. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Networks Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
  21. Tang, X.; Chen, H.; Xiang, W.; Yang, J.; Zou, M. Short-term load forecasting using channel and temporal attention based temporal convolutional network. Electr. Power Syst. Res. 2022, 205, 107761. [Google Scholar] [CrossRef]
  22. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the 5th International Conference on Learning Representations, ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  23. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  24. Nair, V.; Hinton, G.E. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  25. Shoup, D.; Paserba, J.; Taylor, C. A survey of current practices for transient voltage dip/sag criteria related to power system stability. In Proceedings of the IEEE PES Power Systems Conference and Exposition, New York, NY, USA, 10–13 October 2004; Volume 2, pp. 1140–1147. [Google Scholar]
  26. Rafique, Z.; Khalid, H.M.; Muyeen, S.; Kamwa, I. Bibliographic review on power system oscillations damping: An era of conventional grids and renewable energy integration. Int. J. Electr. Power Energy Syst. 2022, 136, 107556. [Google Scholar] [CrossRef]
Figure 1. The power system stability co-prediction process and details of the STEGNN model.
Figure 1. The power system stability co-prediction process and details of the STEGNN model.
Energies 18 00672 g001
Figure 2. Training results of the STEGNN model. Subfigure (a) shows accuracy and loss for the stability co-prediction on both the training set and validation set. Subfigure (b) shows the confusion matrix of the test set, where Class 0, Class 1, Class 2, and Class 3 represent [1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 0], and [0, 0, 0, 1] in Table 1. (a) A C C ; (b) confusion matrix.
Figure 2. Training results of the STEGNN model. Subfigure (a) shows accuracy and loss for the stability co-prediction on both the training set and validation set. Subfigure (b) shows the confusion matrix of the test set, where Class 0, Class 1, Class 2, and Class 3 represent [1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 0], and [0, 0, 0, 1] in Table 1. (a) A C C ; (b) confusion matrix.
Energies 18 00672 g002
Figure 3. Visualization of high-dimension features from the hidden layer. The first row is training and testing dataset samples colored with labels, and the second row is training and testing dataset samples colored with network outputs.
Figure 3. Visualization of high-dimension features from the hidden layer. The first row is training and testing dataset samples colored with labels, and the second row is training and testing dataset samples colored with network outputs.
Energies 18 00672 g003
Figure 4. ACC and computational training time per epoch with different time lengths for the proposed STEGNN model trained in the IEEE 118-bus power system.
Figure 4. ACC and computational training time per epoch with different time lengths for the proposed STEGNN model trained in the IEEE 118-bus power system.
Energies 18 00672 g004
Figure 5. The ACC of different methods for the N-2 contingency dataset. Co-Pre, Single-Pre, and Co-Pre without WL are abbreviations for co-prediction, TSP/VSP single prediction, and co-prediction without weight loss, respectively. TSP-VSP Co-Pre represents the co-prediction for transient and voltage stability, and the TSP–VSP single prediction represents the combined result of the single TSP and the single VSP. TSP Co-Pre represents Co-TSP, and VSP Co-Pre represents Co-VSP. TSP Single-Pre represents the single TSP and VSP Single-Pre represents the single VSP.
Figure 5. The ACC of different methods for the N-2 contingency dataset. Co-Pre, Single-Pre, and Co-Pre without WL are abbreviations for co-prediction, TSP/VSP single prediction, and co-prediction without weight loss, respectively. TSP-VSP Co-Pre represents the co-prediction for transient and voltage stability, and the TSP–VSP single prediction represents the combined result of the single TSP and the single VSP. TSP Co-Pre represents Co-TSP, and VSP Co-Pre represents Co-VSP. TSP Single-Pre represents the single TSP and VSP Single-Pre represents the single VSP.
Energies 18 00672 g005
Figure 6. Visualization of grid-informed adjacency matrices, where subfigure (a) shows B I , subfigure (b) shows B I I , and subfigure (c) shows B I I I .
Figure 6. Visualization of grid-informed adjacency matrices, where subfigure (a) shows B I , subfigure (b) shows B I I , and subfigure (c) shows B I I I .
Energies 18 00672 g006
Figure 7. Subfigure (a) shows the A C C of STEGNN with and without weighted loss. Subfigure (b) shows F N R of STEGNN with and without weighted loss, and WL is short for weighted loss. (a) A C C comparison; (b) F N R comparison.
Figure 7. Subfigure (a) shows the A C C of STEGNN with and without weighted loss. Subfigure (b) shows F N R of STEGNN with and without weighted loss, and WL is short for weighted loss. (a) A C C comparison; (b) F N R comparison.
Energies 18 00672 g007
Table 1. Definition of label y as the one-hot form.
Table 1. Definition of label y as the one-hot form.
Label y = ( y T , y V ) y T = 0 y T = 1
y V = 0 [1, 0, 0, 0][0, 1, 0, 0]
y V = 1 [0, 0, 1, 0][0, 0, 0, 1]
y T and y V are defined in Equations (14) and (15), where y T = 0 indicates transient stability and y T = 1 indicates transient instability, while y V = 0 indicates voltage stability and y V = 1 indicates voltage instability.
Table 2. The detailed seven main hyperparameters and construction of STEGNN.
Table 2. The detailed seven main hyperparameters and construction of STEGNN.
ParametersValueBlocks/LayersOutput Size
lr 0.001 input(N, F, l)
batch128GC module(N, 8, l)
β 0.0005 Flatten and Concentance( N × 8 , l)
F4TC module(1, 32)
α j  * m a x { i y i j } / i y i j fc1(16, 1)
B B I I I fc2(8, 1)
l21fc3(4)
* The calculation formula of α j in the table is generally valid, except for the case i = 1 b a t c h y i = 0 , where α j is set to 1 to prevent calculation errors.
Table 3. Comparison of STEGNN with other deep learning methods.
Table 3. Comparison of STEGNN with other deep learning methods.
Model ACC (%) FPR (%) FNR (%) F s (%)
Co-TSPMLP98.890.140.1198.82
CNN99.790.120.0399.73
TCN99.910.010.1299.84
GCN99.9600.1399.88
STEGNN99.9900.0399.99
Co-VSPMLP98.220.074.4697.66
CNN98.801.370.9498.47
TCN99.080.900.9498.83
GCN99.340.470.9299.02
STEGNN99.530.250.8199.40
Table 4. The performance comparison of the combined result of the single TSP and the VSP with the concurrent TSP and VSP for the IEEE 118-bus power system based on the TLGF-0.2 dataset.
Table 4. The performance comparison of the combined result of the single TSP and the VSP with the concurrent TSP and VSP for the IEEE 118-bus power system based on the TLGF-0.2 dataset.
MethodCorrectMissing ACC (%)
Combined (Single TSP, Single VSP)17165496.9
Co-prediction17700100
Table 5. The performance comparison of the single TSP and co-TSP for the IEEE 118-bus power system based on the TLGF-0.2 dataset.
Table 5. The performance comparison of the single TSP and co-TSP for the IEEE 118-bus power system based on the TLGF-0.2 dataset.
MethodCorrectMissing ACC (%)
Single TSP17472398.7
Co-TSP17700100
Table 6. The performance comparison of the single VSP and co-VSP for the IEEE 118-bus power system based on the TLGF-0.2 dataset.
Table 6. The performance comparison of the single VSP and co-VSP for the IEEE 118-bus power system based on the TLGF-0.2 dataset.
MethodCorrectMissing ACC (%)
Single VSP17323897.9
Co-VSP17700100
Table 7. The concurrent prediction performance with different time lengths of post-fault PMU data in the IEEE 118-bus power system.
Table 7. The concurrent prediction performance with different time lengths of post-fault PMU data in the IEEE 118-bus power system.
Time Length (s)0.050.10.150.20.250.30.350.40.450.5
Co-Prediction A C C ( % ) 99.3199.3499.3899.4599.4799.5399.5499.5599.5599.55
Co-TSP A C C ( % ) 99.8699.8999.9299.9599.9799.9999.9999.9999.9999.99
F P R ( % ) 0.020.130.110.070.0500000
F N R ( % ) 0.4000000.030.020.020.020.02
F s ( % ) 99.7799.8299.8799.9199.9499.9999.9999.9999.9999.99
Co-VSP A C C ( % ) 99.4499.4599.4599.5099.5199.5399.5499.5599.5599.55
F P R ( % ) 0.360.410.540.310.370.250.240.210.220.21
F N R ( % ) 0.870.720.560.810.670.810.650.630.610.62
F s ( % ) 99.2899.2899.2999.3699.3799.4099.4399.4599.4599.45
Table 8. Performance on co-prediction in the IEEE 118-bus power system under different SNR levels of PMU data.
Table 8. Performance on co-prediction in the IEEE 118-bus power system under different SNR levels of PMU data.
Co-PredictionSNR (dB)Metrics
ACC  (%) FNR  (%) FPR  (%) F s  (%)
Co-TSPNo99.990.03099.99
6099.990.02099.99
5099.980.05099.94
4099.950.050.0399.89
3099.910.120.0299.83
2099.850.270.2499.76
Co-VSPNo99.530.250.8199.40
6099.540.270.6999.44
5099.510.340.7199.38
4099.450.380.7799.27
3099.390.390.9199.09
2098.310.441.2598.22
Table 9. Performance comparison of grid-informed adjacency matrices.
Table 9. Performance comparison of grid-informed adjacency matrices.
Matrices B I B II B III
Co-Prediction A C C ( % ) 99.2898.9799.53
Co-TSP A C C ( % ) 99.9799.6499.99
F P R ( % ) 0.040.020
F N R ( % ) 0.031.150.03
F s ( % ) 99.9499.3999.99
Co-VSP A C C ( % ) 99.3199.3399.53
F P R ( % ) 0.320.260.25
F N R ( % ) 1.261.310.81
F s ( % ) 99.1299.1499.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Deng, C.; Dai, L.; Chao, W.; Huang, J.; Wang, J.; Lin, L.; Qin, W.; Lai, S.; Chen, X. An Advanced Spatio-Temporal Graph Neural Network Framework for the Concurrent Prediction of Transient and Voltage Stability. Energies 2025, 18, 672. https://doi.org/10.3390/en18030672

AMA Style

Deng C, Dai L, Chao W, Huang J, Wang J, Lin L, Qin W, Lai S, Chen X. An Advanced Spatio-Temporal Graph Neural Network Framework for the Concurrent Prediction of Transient and Voltage Stability. Energies. 2025; 18(3):672. https://doi.org/10.3390/en18030672

Chicago/Turabian Style

Deng, Chaoping, Liyu Dai, Wujie Chao, Junwei Huang, Jinke Wang, Lanxin Lin, Wenyu Qin, Shengquan Lai, and Xin Chen. 2025. "An Advanced Spatio-Temporal Graph Neural Network Framework for the Concurrent Prediction of Transient and Voltage Stability" Energies 18, no. 3: 672. https://doi.org/10.3390/en18030672

APA Style

Deng, C., Dai, L., Chao, W., Huang, J., Wang, J., Lin, L., Qin, W., Lai, S., & Chen, X. (2025). An Advanced Spatio-Temporal Graph Neural Network Framework for the Concurrent Prediction of Transient and Voltage Stability. Energies, 18(3), 672. https://doi.org/10.3390/en18030672

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop