Next Article in Journal
Study on Centroid Height Prediction of Non-Rigid Vehicle Based on Deep Learning Combined Model
Previous Article in Journal
Transparent EEG Analysis: Leveraging Autoencoders, Bi-LSTMs, and SHAP for Improved Neurodegenerative Diseases Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph Neural Networks for Fault Diagnosis in Photovoltaic-Integrated Distribution Networks with Weak Features

1
School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
2
Information and Communication Branch, State Grid Zhejiang Electric Power Co., Ltd., Hangzhou 310008, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5691; https://doi.org/10.3390/s25185691
Submission received: 7 August 2025 / Revised: 30 August 2025 / Accepted: 11 September 2025 / Published: 12 September 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

Effective diagnosis of distribution network faults is crucial to ensuring the reliability of power systems. However, the bidirectional power flow caused by the integration of new energy limits the effectiveness of traditional detection methods. Although data-driven approaches are not restricted by power flow direction, their performance is heavily dependent on the quantity and quality of training samples. In addition, factors such as measurement noise, variable fault impedance, and volatile photovoltaic output complicate fault information. To address this, we present a new fault diagnosis model named the dynamic, adaptive, and coupled dual-field-encoding graph neural network (DACDFE-GNN), which introduces a dynamic aggregation module to assign different weights to reduce noise interference and fully integrates information from observable nodes. On this basis, the coupled dual-field-encoding module is proposed, which encodes topological information and physical–electrical domain information as part of the initial features, thereby capturing fault features and learning the law of feature propagation. The experimental results for the IEEE 34- and IEEE 123-node feeder systems indicate that the proposed model surpasses recent fault diagnosis methods in detection performance, particularly regarding its low training sample rate.

1. Introduction

The rapid penetration of distributed energy resources and the advancement of smart metering infrastructure, while driving the intelligence of distribution networks (DNs), also pose challenges concerning system stability and protection. DN operators must thoroughly understand grid conditions, avoid faults, and promptly address any that occur to ensure power supply reliability. Therefore, it is essential to develop effective and efficient fault diagnosis methods to promptly detect, locate, and mitigate faults, ensuring the stable and reliable operation of DNs.
Traditional methods include impedance-based methods [1,2], signal injection methods [3,4], traveling wave methods [5,6,7], and matrix methods [8,9]. However, in addition to the limitations of the methods themselves, the intermittent and distributed characteristics of renewable energy also challenge these traditional coordination methods in many aspects, such as bidirectional power flow [10].
Early data-driven approaches primarily utilized multilayer perceptron (MLP) and convolutional neural network (CNN) architectures. Surin et al. [11] use five MLP networks to classify inverter output voltages and identify fault types and locations in multilevel inverter systems. Javadian et al. [12] used normalized fault currents as input to mitigate the impact of fault impedance on positioning results. Nevertheless, in practical scenarios, measurement noise, variable fault impedance, and the intermittency of new energy sources can obscure fault signatures, as illustrated in Figure 1. This figure shows the impact of adding and varying fault impedance, the photovoltaic system, and measurement noise on the data distribution, where each color signifies distinct scenarios, such as NF, LG, LL, LLG, LLLG, and LLL. Compared with (a), the other three graphs show shorter mapping distances between various faults, which means weaker fault characteristics and higher discrimination difficulty. The phenomenon shown in (b) and (c) confirms that the intermittency and randomness of new energy power generation lead to stronger dynamics and randomness in fault characteristics. In (d), the mapping distances of various faults after dimensionality reduction are the shortest. This indicates that the fault characteristics, when these three factors are taken into account, are less distinct.
Many researchers have proposed solutions to deal with these problems. For example, to address measurement noise, Aljohani et al. [13] combine the Stockwell transform (ST) with a multilayer perceptron neural network (MLP-NN), utilizing the time–frequency features of fault signals for classification. Guo et al. [14] introduced a deep learning approach based on the continuous wavelet transform (CWT) and CNNs to enable CNNs to automatically extract features and detect fault feeders. Li et al. [15] establish a relationship between the admittance matrix Y , the voltage change matrix Δ U , the current change matrix Δ C , and the unbalanced current Δ C u based on Kirchhoff’s laws. Then they further confront the variable fault impedance in the model, but they do not review a scenario with new energy. In contrast, Paul [16] and Siddique [17] propose CNN-based approaches for the new energy scenario. However, CNN-based methods require extracting temporal features from signals, necessitating the sampling of transient fault waveforms. Most existing power grid data are uploaded every five minutes, which fails to meet the algorithms’ need for high-resolution data. Moreover, as DNs expand, the above methods face more challenges beyond the lack of high-resolution data, such as the low training sample rate caused by the scarcity of fault records. In addition, the large scale of DNs causes low observation rates due to the high cost of phasor measurement unit (PMU) placement. These issues are rarely addressed in the CNN-based methods described above.
Schlichtkrull et al. [18] indicated that GCNs can be utilized to model relational data. Since then, graph neural networks (GNNs) have attracted significant attention for the fault diagnosis task of DNs because they can explicitly utilize the inherent topological information of the DN. The DN can be modeled as a homogeneous graph based on node lines, and the voltage data measured by PMUs can be used to fit GNNs. A notable study [19], by integrating voltage vectors and branch current data and leveraging GCNs’ spatial correlation capture capabilities, improved fault location accuracy in scenarios with measurement noise and variable fault impedance. Mo et al. [20] proposed a distribution network fault diagnosis method based on a depth graph attention network (GAT) to solve the problem of sparse measurement points in constant conditions. Li [21], combining the scenarios of the previous two studies, explored a fault location architecture under sparse measurement nodes to enhance fault location performance in environments with measurement noise and variable fault impedance. Based on these works, Chanda et al. [22] introduced a heterogeneous multitask learning GNN framework that simultaneously performs five tasks and independently addresses the issues of sparse measurement points and low training sample rates. Nevertheless, all the aforementioned studies were conducted in distribution network settings without new energy sources, and they ignored the analysis of the impact of new energy on fault characteristics, as shown in Figure 1d. In research on new energy fields, Ngo et al. [23] proposed a deep GAT that combines one-dimensional convolutional layers with graph attention layers, enhancing the model’s ability to process time-series and spatial data but requiring bus voltage and branch current measurements at 1 kHz. Prasad et al. [24] presented a GNN-based fault diagnosis framework (GFDF). They integrated the multihead attention mechanism with current measurements, topological information, and line parameters for fault diagnosis but did not analyze influencing factors such as low observation rates.
It is worth noting that recently a class of models known as Physics-Informed Neural Networks (PINNs) has garnered significant attention among many researchers. PINNs enable the synergistic combination of mathematical models and data [25]. Falas et al. [26] proposed a hybrid approach utilizing PINNs, leveraging graph neural network architectures to more effectively process the graph-structured data of power systems, and incorporated physical knowledge constructed from the branch current formulation, making the method adaptable to both transmission and distribution systems. Ngo et al. [27] employed domain-specific physical knowledge and graph neural networks to estimate bus voltage magnitude and phase angles under load change conditions. Unlike [19,22], these PINN-based methods are more commonly applied in the domain of distribution network state estimation and often involve the use of temporal information. Additionally, key information such as active power P and reactive power Q is crucial for constructing physics-informed models, which are not within the scope of the present study and thus are not discussed in detail.
Table 1 summarizes existing GNN methods used for fault diagnosis. Apart from the methods from [23,24], all the other methods are based on steady-state fault data recorded when the electrical quantities (voltage, current, etc.) of a power system enter a new steady-state operating point after a fault occurs. In conclusion, existing methods based on steady-state fault data have difficulties in addressing the challenge of fully considering the combined impact of sparse measurements, low training sample rates, measurement noise, variable fault impedance, and new energy fluctuations on the model. Under the co-influence of multiple perturbations, the progressive collapse of inter-class separation among fault samples markedly compromises the diagnostic accuracy of existing approaches.
Furthermore, almost all of the above models adopt the general GNN framework. The propagation of distribution network faults is closely related to network line parameters and the superior–subordinate relationship between nodes. Effective utilization of this coupling information is conducive to improving the accuracy of the fault diagnosis model. On the other hand, the above-mentioned models, such as [20], adopt a dynamic aggregation mechanism to address issues such as sparse measurement nodes, but the receptive field of local attention is limited. The computational complexity of global attention is high, and it is vulnerable to interference from redundant information and measurement noise. Furthermore, they processed the electrical quantity characteristics and topological structure separately, and the attention mechanism failed to fully integrate the electrical–topological joint features. Therefore, our work can be summarized as follows:
  • A location coding method called coupled dual-field encoding (CDFE) is proposed. A trainable location coding matrix is designed by using the transmission line characteristics and distance parameters of the connecting nodes to assist the model in understanding the law of fault propagation and learning the coupling relationship between electrical features and transmission line characteristics.
  • A dynamic adaptive graph structure-aware aggregation mechanism (DAM) is proposed. The calculation of the attention coefficient is guided by the sampled relation matrix, fully integrating the electrical–topological joint features, improving the generalization ability of the model and better capturing the rich feature information after CDFE without increasing additional computational overhead.
  • In a distribution network environment with characteristics such as sparse measurement points, a low training sample rate, measurement noise, variable fault impedance, and new energy volatility, the performance of the proposed model is superior to that of the baseline model.
The remainder of this paper is organized as follows: Section 2 first outlines the objectives and task formulation of the fault diagnosis endeavor. Then a detailed description of the proposed fault diagnosis model, DACDFE-GNN, is provided, including its architecture and key components. Section 3 sets the experimental parameters for simulation verification and discusses the experimental results. The performance of the proposed model under various conditions is analyzed and compared with existing methods. Additionally, ablation studies are conducted to substantiate the effectiveness of the modules. Finally, Section 4 offers a summary of the paper and suggestions for future research.

2. Proposed Method

2.1. Problem Formulation

Similarly to [19], this paper establishes a homogeneous graph for the distribution network. The voltage and phase V a , V b , V c , θ a , θ b , θ c R 6 at the node are used as the input data. The data for unobservable nodes and non-three-phase nodes is set to 0. Then the system measurement data sample X i R n × 6 , i 0 , 1 , 2 , , N 1 can be obtained, where n is the number of nodes observed and N is the total number of data samples.
The fault diagnosis task is divided into five subtasks, namely fault detection (FD), fault location (FL), fault classification (FC), fault impedance estimation (FIE), and fault current estimation (FCE). The first three tasks are described as classification problems, while the latter two are formulated as regression problems. It is worth noting that FC incorporates normal data, as a special type of fault sample, into the classification process, thereby avoiding the need for an additional fault detection step before performing the data classification task. Among all five tasks, FD is the most important. Only when it can detect the occurrence of faults do the subsequent location and classification tasks make sense. Secondly, FL is even more crucial in the remaining tasks because it determines the destination that the staff need to go to after a malfunction occurs. Then, the severity of the fault is closely related to the type of fault, which also makes FC quite important. Finally, FIE and FCE have the lowest priority. The importance of these tasks plays a guiding role when conducting the result analysis in Section 3.5.
In the following, we will represent the sample index with superscript k and use subscripts t = 1 , 2 , 3 , 4 , and 5, respectively, to denote the following five tasks: FD, FL, FC, FIE, and FCE. We also denote the node number index with subscript i. For a certain data sample X k R n × 6 , the corresponding label of the data sample is Y k = y 1 k , y 2 k , y 3 k , y 4 k , y 5 k .
For each task t , t = 1 , 2 , 3 , 4 , 5 , there is a separate fully connected network with its own set of parameters to achieve differentiated target outputs. Let f θ 0 represent the backbone network and h θ t denote the fully connected network for the t-th task. Then, the output of the k-th task for the t-th data sample can be represented as
y ˜ t k = h θ t f θ 0 X k
For the three classification tasks, we use cross-entropy loss L c l s , as shown in Equation (2), as the loss function. For the regression tasks, the mean-squared error L r e g , as shown in Equation (3), is used as the loss function.
L c l s = k = 1 N t r a i n y t k log exp h θ t f θ 0 X k / i = 1 m exp h θ t f θ 0 X k
where t 1 ,   2 ,   3 denotes the index of the classification task and m represents the total number of classes for that classification task. N t r a i n is used to express the number of training samples.
L r e g = k = 1 N t r a i n y t k y ˜ t k 2 / N t r a i n
where t 4 ,   5 denotes the index of the regression task.
Finally, the losses of the five subtasks are weighted and summed. The overall objective function, including the L2 regularization factor λ , can be represented as
min θ k = 1 N t r a i n t = 1 5 ω t L t θ , X t r a i n k + λ θ 2
where ω t and L t denote the weight and loss of the t-th task and θ represents the parameters in the network.

2.2. DACDFE-GNN Model

The proposed model uses multiple GNN layers as the backbone network and a multitask header to handle multiple fault diagnosis tasks simultaneously, where the subtask network includes fault detection, fault location, fault classification, fault impedance estimation, and fault current estimation.

2.2.1. Overall Architecture

The network architecture of the proposed model is shown in Figure 2. In the backbone network part, the proposed model implements message passing through the forward propagation of the DACDFE-GNN to generate graph embeddings. Only the input of the first layer of the network has CDFE. The learned graph embeddings are flattened and joined and then pooled. Finally, the feature vector is fed into five different downstream task prediction heads, corresponding to the five tasks in Section 2.1, for prediction output.
Specifically, for a system with N nodes, a low observation rate indicates the absence of a large number of node eigenvalues, which leads to the performance degradation of the GCN-based model. In order to obtain the connection relationship between the observation nodes, Dijkstra’s shortest path algorithm is used to construct the distance correlation matrix A R N × N .
For a backbone network with K layers, H k R N × n k denotes the node embeddings of the k-th layer, where k = 1 , 2 , , K . The update rule for the k-th layer is
h i k = σ h i k 1 | | h ˜ j k 1 W k , h ˜ j k 1 = Aggregate j N ˜ i m h j k 1 b i j
where | | denotes the concatenation operation, σ x represents the activation function, W k R 2 n k 1 × n k , N ˜ i m indicates the sampled set based on N i m , and b i j represents the weighted coefficients in the aggregation process, which depend on the specific aggregation function.
In contrast to existing GNN-based methods, DACDFE-GNN uses CDFE in the initial stage to enhance the understanding of fault propagation laws. The DAM is used in the process of feature propagation and aggregation to alleviate the influence of data quantity and quality. The details of the proposed method are further elaborated in the following sections: (1) CDFE Module and (2) DAM.

2.2.2. CDFE Module

Although measuring noise, variable fault impedance, and volatile photovoltaic output make it difficult to distinguish between fault characteristics for different fault locations and types, the impact of node failures in the distribution network on the remaining nodes still depends on the specific parameters of the lines. The CDFE module encodes network topology information to help the model understand the location information for the nodes in the graph and learn the law of fault propagation. By encoding the network topology with a trainable parameter matrix as part of the input features, the input features can be represented as
H 0 = A W t + p e e l e c + p e d e p W p e + X
where W t , W p e R N × n k and p e e l e c ( v i ) = d i s t ( v i ) / max v V d i s t ( v ) is the shortest path distance from the source node to other nodes, obtained by Dijkstra from the equivalent electrical distance w u v = r u v 2 + x u v 2 · d u v . The topological depth coding p e d e p is calculated from p e dep ( v i , 2 i ) = sin d ( v i ) / 10000 2 i / n k and p e dep ( v i , 2 i + 1 ) = cos d ( v i ) / 10000 2 i / n k .
In Equation (6), A represents the adjacency matrix, which captures discrete topological information. p e e l e c denotes the continuous electrical distance weighted by impedance and distance, representing physical information. Meanwhile, p e d e p signifies the topological depth, representing hierarchical information. These three elements characterize the propagation of fault signals within the power grid from three dimensions. By employing two feature mapping matrices, the structural–physical prior knowledge and electrical measurements can jointly help the model in extracting fault features.
Then, the following connections between the input layer and other layers are established through the residual network:
H k = μ H k 1 + 1 μ H 0
where μ is used to weigh the ratio of the position code to the original feature with a default value of 0.2.

2.2.3. DAM

In order to enhance the generalization ability of the model and better capture the rich feature information after CDFE, this work proposes a dynamic adaptive weight aggregation mechanism. By dynamically adjusting the weights of node aggregation, nodes focus on the key dependencies between nodes when aggregating the features of neighboring nodes, thereby helping the model to utilize the limited sample data more effectively. Its aggregation process is as follows:
Q k = H k W q k , K k = H k W k k , V k = H k W v k , E j k = B Q k j h / g : j + 1 h / g K k T
A t t j k = softmax E j k / d k / h V k
H ˜ k = Concat A t t 1 k , A t t 2 k , , A t t g k W o , H k + 1 = H k | | H ˜ k W k + 1
where W q k , W o k R n k 1 × n k , W k k , W v k R n k 1 × ( n k h g ) , l represents the l-th group, g represents the number of groups, h denotes the number of heads, and H ˜ k represents the feature representation of all nodes at layer k.
In addition, the adjacency matrix is used to guide the computation of attention. For each attention group, e i j = a ˜ i j q i k j T is used instead of e i j = q i k j T . Here, e i j and a ˜ i j denote the i-th row and j-th column of E and A ˜ , respectively, and A ˜ is the adjacency matrix obtained by subgraph sampling based on A .
It is worth noting that the values of E at the corresponding positions are set to 0. After softmax · , these values become very small but identical, which means that our model not only takes local information into account, but also retains some global aspects. In this way, the computational cost of the DAM is equivalent to that of local attention, with a computational complexity of O(z), where z represents the number of non-zero elements in the sampling matrix A ˜ . Meanwhile, the computational cost of global attention is O( N 2 ). When the sampling matrix A ˜ is sparse, the computational cost will be significantly reduced.

3. Evaluation of Proposed Model

This section shows an evaluation of the proposed model with the IEEE 123-node feeder test case for validation of the model and data generation process. Then the specific parameters and evaluation indexes of the model are given. Finally, the performance of the model is verified and compared with some baseline methods.

3.1. IEEE 123-Node Feeder System and Data Preparation

3.1.1. IEEE 123-Node Feeder System

In order to simulate the five types of short-circuit faults, including the three-phase symmetric faults LG, LL, and LLG and the three-phase asymmetric faults LLL and LLLG, we only focus on three-phase nodes, including two three-phase regulators, 150r and 160r, as indicated in Figure 3. Subsequently, we added photovoltaic generators at node 250 and node 66 to simulate a scenario with new energy access.

3.1.2. Data Generation

In the data generation section, we utilized OpenDSS as the power flow solver and interacted with it in Python 3.8 through py_dss_interface.
For the five fault detection tasks, five labels were generated for each data sample, representing whether a fault occurred, the number of nodes where the fault occurred, the type of fault, the fault impedance, and the fault current, respectively. The first three labels are discrete values corresponding to classification tasks, while the last two labels are continuous values corresponding to regression tasks.
Firstly, for the fault data, we considered variable load levels. Based on [19], we established the probability density distribution of the system’s load levels. The load levels were uniformly divided from 0.316 to 1 with 50 segments. During the generation of each data sample, sampling was performed to set the load level, thereby simulating different load levels at different moments.
Secondly, considering the photovoltaic consumption capacity of the network, we set the rated power generation capacity of the two photovoltaic systems at 500 KW. The calculation formula for photovoltaic power generation is
P out = P w × η × 1 + γ T T 0
where P out represents the actual output power, P w represents the rated power of the photovoltaic generator, and η represents the ratio of the actual radiation intensity to the standard radiation intensity, which is obtained through sampling based on the probability density distribution. γ is the temperature coefficient, usually taken as −0.5%, T 0 is the rated temperature, usually taken as 25 ∘C, and T is the actual operating temperature of the photovoltaic module.
Then, considering the practical occurrence of faults, the data generated using variable fault impedance methods is more valuable in real-world applications. When creating a set of fault impedances R f , we assumed that the fault impedance was uniformly distributed within a certain range, r U min r , max r , where r represents the sampled value and min r and max r denote the maximum and minimum values of the fault impedance under the uniform distribution, respectively. Unless otherwise specified, the fault resistance is set by default to a minimum of 0.05 Ω and a maximum of 20 Ω .
For each type of fault, we generated 8160 data samples, with each node containing 120 fault samples, covering all five types of faults. This resulted in a total of 40,800 fault data samples. For non-fault data, we only considered varying load levels and similarly generated 8160 data samples. The sequence of steps in the data generation process is illustrated in Figure 4.
  • In the first loop, iterate through all 68 three-phase nodes. For each node, perform the following two steps.
  • In the second loop, iterate through the possible states of the system, including five fault states and one normal operating state. Select the simulation parameters according to the state of the system.
  • In the third loop, first randomly sample a load level list of length 120 according to the probability density function. If it is a fault state, also sample a list of the same length of fault impedances within the specified range. Finally, iterate through the lists to perform power flow simulations.
Finally, to simulate measurement errors introduced by measuring instruments, we added noise to all 48,960 samples, including both fault and non-fault samples, to achieve a signal-to-noise ratio (SNR) of 45 dB. The noise has a mean value of zero, and the standard deviation σ n o i s e is calculated as follows:
σ n o i s e = 10 S N R 20
In the training process, due to the large fluctuation range of the fault current value, making the training process more stable, we modify the fault current label. We normalize the label y t = 5 and then take the negative logarithm, which can be expressed as
y 5 k = ln y 5 k k = 1 N t r a i n y t r a i n , 5 k
The molecule contains only data from the training set in order to avoid information leakage. To ensure stable training of the model, all data samples were normalized by subtracting the mean of the training set samples and scaling by the standard deviation of the training set.

3.2. Parameter Details

In the backbone, three DACDFE-GNN layers are used to capture common features. In subtask networks, different fully connected networks are used to extract different information. The details of the parameters for each layer and the shapes of the outputs are shown in Table 2. The required training parameters of the CDFE module depend on the number of network nodes and the dimension of the hidden layer. The computational overhead of the DAM module depends on the number of selected packets. The more packets there are, the greater the computational overhead.
This model was trained on a personal computer equipped with an Intel Core i9-13900 processor, 56 GB of memory, 48 GB of storage space, and an NVIDIA RTX 4080 SUPER graphics card. It was trained for 300 cycles using the AdamW optimizer and the early stopping method. The hyperparameters used are summarized in Table 3, and they are mainly derived from [22]. Furthermore, based on the location of the measurement node, the number of neighbors is set. This ensures that each unobserved node can be covered by an observing node within the number of neighbors of a certain observing node. For classification tasks, we report the accuracy; for regression tasks, we report the mean squared error(MSE). In particular, for fault location tasks, we also report the accuracy of 1-hop location (LAR1) and represent the accuracy as 0-hop location (LAR0).

3.3. Selection of Group Numbers

In the DAM, grouped-query attention (GQA) is used to reduce the spatial complexity of the model. We first conducted an analysis of the model’s performance across the five tasks by varying the number of groups. The configuration of other parameters is shown in Table 3.
As indicated in Table 4, the model maintains a comparable performance across different numbers of groups. Notably, when the group number is set to 2, the model achieves optimal overall performance. Consequently, the two-group configuration is selected to represent the proposed model in subsequent studies. It should be noted that in the subsequent tables, bold font is used to represent the optimal results in this group of experiments, while underscores are used to represent the suboptimal results.

3.4. Ablation Experiment

To validate the proposed module, we conducted experiments combining different configurations in four cases: with/without DAM and with/without CDFE. Additionally, we incorporated local and global attention mechanisms, leading to six models in total.
As can be seen from Table 5, in the fault location task, Mode 2 improved by approximately 15% compared to Mode 1. In all other tasks except fault detection, there were also improvements to varying degrees. This indicates that CDFE helps the GNN model better understand the law of fault propagation, thereby improving the diagnostic ability of the GNN model. Meanwhile, the comparison between Mode 3 and Mode 1 also indicates that the DAM can help the model better understand the relationships between nodes. With the attention aggregation mechanism, performance improvements are observed in all tasks, excluding fault detection. This is attributable to the dynamic aggregation mechanism’s capability to adaptively assign varying weights to different input features, allowing the model to focus more on features that are more critical to the task, thereby enhancing the model’s performance and accuracy.
Furthermore, the results of Mode 4, Mode 5 and Mode 6 show that the proposed method is superior to the local attention model but inferior to the global attention model. However, the global attention model’s enhanced performance comes from increased complexity and only shows a clear advantage in fault impedance estimation. Trading significant computational overhead for marginal performance gains is hardly a gratifying proposition. In contrast, the local attention model’s narrow focus on adjacency matrix-defined information results in information loss and slightly inferior performance compared to our method. The DAM boosts model performance by giving minimal weights to non-physically connected nodes and does not add computational costs, making it better than the local attention mechanism.
Finally, the results in Table 5 indicate that altering ω t does not lead to model failure. However, comparing mode 4 and mode 7 reveals that when the model treats the five tasks equally, the performance of the model deteriorates. Only in the FIE task does mode 7 outperform mode 4. This suggests that a rational distribution of modes can enhance the performance of the multitask model. Nevertheless, the focus of this study is not on the allocation of multitask weights. For fairness, all models in subsequent comparisons will utilize the weight parameters from [22].

3.5. Analysis of Experimental Results

To further demonstrate the advantages of our proposed model, we conducted a comparative analysis with five other baseline models under various conditions, including different scales of training samples, noise robustness, different fault impedance ranges, and changes in topology.
  • GAT [20,24]: The backbone network of this model is composed of GAT layers. In the GAT layer, a multihead local attention mechanism with eight heads is used.
  • GCN [22]: The backbone network of this model is composed of GCN layers based on the normalized Laplacian matrix.
  • GCN-Cheb [19]: The backbone network of this model is stacked by a graph convolutional neural network with Chebyshev polynomials as the kernel.
  • EATSA-GNN [28]: The backbone network of this model is composed of EATSA-GNN layers. The EATSA-GNN model uses edge awareness and a two-stage attention mechanism to enhance the classification of GNN nodes. The authors focus on the edges to capture relationship information and then apply attention to the nodes to integrate the two information sources.
  • FD-SMRIF [29]: The backbone network of this model consists of three layers of GCN networks and one layer of a normalized flow layer. The model is used for rotating mechanical samples to overcome the challenge of scarce labeled samples. Here, the input layer and output layer are modified to adapt to the fault diagnosis task of the distribution network.
The number of layers that serve as the backbone network in the baseline model is the same as in the proposed model, and the parameters of the fully connected network are consistent with those in Table 2. For all the experimental combinations, the loss in the validation set successfully converged.

3.5.1. Comparison Under Different Training Samples

The quantity of training samples significantly impacts the model performance. The experimental results in Table 6 highlight the superior performance of the proposed DACDFE-GNN model at several low label rates in comparison to other state-of-the-art models. Although all models achieve perfect accuracy for fault detection, in the more nuanced tasks of fault location and classification, the proposed DACDFE-GNN model surpasses the baseline models. At a label rate of 10%, the DACDFE-GNN shows a significant improvement in fault location accuracy. When the label rate increases to 20%, the model’s performance further improves, demonstrating a higher increase in fault location accuracy compared to the second-best model. In terms of fault classification, the DACDFE-GNN also exhibits high accuracy. All models can achieve a classification accuracy of more than 90%, but only DACDFE-GNN and GCN-Cheb can achieve a classification accuracy of more than 95% under the three verified label rates. For fault resistance and current estimation, only the performance of the EATSA-GNN is comparable to the proposed model. In most cases, the DACDFE-GNN has the lowest MSE and highest stability. This indicates that even at a low training sample rate, the proposed model is superior to other baseline models.
Furthermore, in classification tasks, GAT may underperform compared to the GCN due to its higher model complexity and sensitivity to the amount of training data, including the data sample and measurement nodes. When the number of training samples is limited, GAT may not perform as well as the GCN. However, as the number of training samples increases, GAT is able to better learn useful feature representations, thereby improving its performance. In regression tasks, the attention mechanism of GAT provides a more flexible adjustment capability for feature importance, enabling it to more effectively capture the complex relationships between nodes, thus consistently outperforming the GCN. In contrast, the DACDFE-GNN, which employs a similar propagation concept, is consistently superior to the GCN. This demonstrates that the DACDFE-GNN, when combined with DAM and CDFE, has greater advantages.
Overall, the proposed DACDFE-GNN model stands out as the most effective model for distribution network fault diagnosis across different label rates, outperforming other models by a noticeable margin and proving its superiority in handling the complexities of fault diagnosis tasks.

3.5.2. Comparison Under Different Measurement Noise Levels

Table 7 shows the performance of the DACDFE-GNN and the baseline model on five tasks under different measurement noise levels. The DACDFE-GNN demonstrated significant advantages in the two most important tasks. It is worth noting that these results use Out-of-Distribution (OOD) data, which means that the model is trained with the original data, and the data with additional noise used in the test does not exist in the training samples.
In the secondary critical task of fault location, the DACDFE-GNN achieves the highest LAR0 and LAR1 at all noise levels. At 40 dB, its LAR0 is 3.72% higher than the second-best model (GCN-Cheb), and it outperforms the worst model (GAT) by approximately 6.62%. When the noise level escalates to 30 dB, the advantage of the proposed model becomes even more pronounced, outperforming GCN-CHEB by 4.46% and the GCN by approximately 9%. For fault classification, the DACDFE-GNN achieved the second-highest accuracy rate, which was only less than 0.3% lower than the optimal GCN-CHEB.
In regression tasks, the proposed model consistently achieves the second-lowest MSE in fault impedance estimation and the lowest MSE in fault current estimation across different noise levels.
Furthermore, this can also be seen from Table 7. Models with attention aggregation mechanisms, such as the DACDFE-GNN with the DAM, GAT with local attention, and the EATSA-GNN with teacher–student attention, show greater robustness across different noise levels. This can be credited to their capacity to diminish the impact of noise through attention mechanisms.
In all three noise environments, the DACDFE-GNN shows the same performance hierarchy. By comparing the results of each model at the same noise level, the DACDFE-GNN performs better in fault location and fault current estimation and ranks second in fault classification and fault impedance estimation. Overall, the DACDFE-GNN outperforms other models across different noise levels and tasks. It shows a significant performance advantage over the second-best models and markedly outperforms the worst models, proving its effectiveness in regression tasks within distribution network fault diagnosis.

3.5.3. Comparison Under Several Topology Change Cases

To evaluate the adaptability of the proposed model to changes in the topology of distribution networks, two scenarios with altered topological connections were verified:
  • Topology 1: Open the switch between nodes 13 and 135 and close the switch between nodes 151 and 300.
  • Topology 2: Open the switch between nodes 97 and 197 and close the switch between nodes 151 and 300.
The fault electrical feature of the distribution network changes significantly with the change in node connections. This leads to a noticeable degradation in the performance of the original model. However, when data from the modified connections are incorporated into the training process, the performance of the model recovers quickly. Table 8 presents the outcomes of Topology 1 and Topology 2 following the fine-tuning of the model over 20 epochs. The results show that after the topology changes, the performance of all models can be restored with only a few epochs of fine-tuning. Among all the models, the DACDFE-GNN still maintains the optimal comprehensive performance.

3.5.4. Comparison Under Several Photovoltaic Change Cases

To observe the adaptability of the proposed model under photovoltaic changes, we set up three cases:
  • Photovoltaic 1: Keep the connection location unchanged and increase the power generation capacity to 800 kW.
  • Photovoltaic 2: Keep the power generation capacity unchanged and add a photovoltaic generator with a power generation capacity of 500 kW at node 450.
  • Photovoltaic 3: Increase the power generation capacity to 800 kW and add a photovoltaic generator with a power generation capacity of 500 kW at node 450.
Table 9 and Table 10 indicate that across these three new photovoltaic scenarios, when directly testing with out-of-distribution (OOD) data, all models show performance drops in most tasks, except for fault detection. However, the DACDFE-GNN still achieves the highest accuracy in the subcritical task of fault location, with other tasks’ accuracy slightly below the optimal model. After fine-tuning the DACDFE-GNN with new scenario data for 20 epochs, its comprehensive performance significantly surpasses the baseline model. Yet, post fine-tuning, all models experience performance declines in the fault current estimation task. This might be because during fine-tuning, the model focuses more on fault location and classification, while fault current estimation is relatively neglected.

3.5.5. Performance at Different Stages

To enhance the credibility of the proposed model, we compared its performance with other baseline models on the IEEE 34-node feeder system. Using the same parameters as in Table 3, we tested the system’s performance with training sample rates of 10% and 30%. It is worth noting that we retained the measurements of all nodes, as the focus of this study is not on the selection of observation nodes.
As can be seen from the results in Table 11, although the performance of the proposed model declined in the IEEE 34-node feeder system, compared with the baseline models, the DACDFE-GNN still retained a significant advantage in the fault location task. In the several remaining tasks, while the performance in the fault classification task is not the best, the classification accuracy remains above 95%. Furthermore, it is evident from the experimental results that GCN-ChEB has a greater advantage in small-scale graph tasks. This indicates that traditional GNN models often need to select different structures according to the scale of the graph. The proposed model performed well on both the IEEE 34- and IEEE 123-node feeder systems, which further demonstrates the effectiveness of the proposed model.

4. Discussion

In this paper, the DACDFE-GNN is designed for a new energy integration DN. This model addresses the challenge of weak fault signals that arise from measurement noise, variable fault impedance, and photovoltaic output fluctuations, especially when only limited nodes are under observation with a low training sample rate. Extensive experiments demonstrate its superiority over baseline methods in key metrics such as noise resilience, low training sample rate, topological changes, and different photovoltaic scenarios. We attribute the success of this model to CDFE and the DAM. The CDFE module guides the model to learn the propagation patterns of fault characteristics among different nodes, helping the model understand the hierarchical relationships between nodes and thereby improving the accuracy of various tasks in fault diagnosis. Then, the DAM effectively captures the rich features encoded by CDFE and enhances the adaptability of the model.
Although this experiment was only verified in the photovoltaic scenario, the model uses steady-state signals for fault diagnosis and thus can easily be extended to other new energy scenarios, such as wind power generation. Moreover, the computational complexity of the CDFE module in the proposed method is closely related to the scale of the distribution network. When the network scale is large, a hierarchical model can be sampled for integration and the network can be divided by levels. The lower-level network serves as a node of the upper-level network, and the proposed model is used to train each network.
Nevertheless, this paper is limited to simulation data. There are still some issues that need to be addressed with respect to real data, such as possible abnormal noise in real data and the scarcity of fault samples. In upcoming work, we will study the modeling of real networks, supplement the scarce fault samples through simulation, and finally use a small amount of real data to transfer the model to practical applications. Regarding the model itself, the effectiveness of the GCN model in real-world settings and the uncertainty from distributed generation’s random access remain areas for future research and improvement. Furthermore, conducting higher-precision estimation of high fault impedance in complex environments and reasonable weight design in multitask architectures will also be part of our future work.

Author Contributions

Conceptualization, Y.H., K.C. and J.Y.; validation, J.L. and G.L.; formal analysis, G.L. and Y.Y.; investigation, J.L.; data curation, G.L., S.C. and Y.X.; writing—original draft preparation, J.L.; writing—review and editing, J.L. and G.L.; visualization, J.L.; supervision, J.Y., S.C., Y.X. and Y.Y.; project administration, Y.H. and K.C.; funding acquisition, Y.H., K.C. and T.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly funded by State Grid Zhejiang Electric Power Co., Ltd., technology project (B311XT24004V).

Institutional Review Board Statement

This study does not require ethical approval.

Informed Consent Statement

This study does not involve humans.

Data Availability Statement

Data available in a publicly accessible repository. After the paper is accepted, the data generation code and model will be available on GitHub: https://github.com/dsgvsd429/GNN, accessed on 6 August 2025.

Conflicts of Interest

Authors Yuteng Huang, Ke Chen, Jiaxiang Yan, Shan Chen and Yuqing Xie were employed by the company Information and Communication Branch, State Grid Zhejiang Electric Power Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. All authors have read and agreed to the published version of the manuscript.

References

  1. Li, Z.; Wan, Y.; Wu, L.; Cheng, Y.; Weng, H. Study on wide-area protection algorithm based on composite impedance directional principle. Int. J. Electr. Power Energy Syst. 2020, 115, 105518. [Google Scholar] [CrossRef]
  2. Qi, Z.; Huang, Z.; Chen, Y. Impedance fault location technology for a distribution network based on a zero-sequence component. Power Syst. Prot. Control 2023, 51, 54–62. [Google Scholar]
  3. Chen, B.; Yu, N.; Chen, B.; Tian, C.; Chen, Y.; Chen, G. Fault location for underground cables in ungrounded MV distribution networks based on ZSC signal injection. IEEE Trans. Power Deliv. 2020, 36, 2965–2977. [Google Scholar] [CrossRef]
  4. Luo, X.; Deng, Z.; Wu, J.; Chen, K.; Jia, P. Fault nature identification and location scheme for distribution network based on active injection from IIDG. Int. J. Electr. Power Energy Syst. 2024, 156, 109706. [Google Scholar] [CrossRef]
  5. Wang, J.; Zhang, Y. Traveling wave propagation characteristic-based LCC-MMC hybrid HVDC transmission line fault location method. IEEE Trans. Power Deliv. 2021, 37, 208–218. [Google Scholar] [CrossRef]
  6. Mu, D.; Lin, S.; He, P.; Li, X. An improved method of traveling wave protection for DC lines based on the compensation of line-mode fault voltage. IEEE Trans. Power Deliv. 2022, 38, 1720–1730. [Google Scholar] [CrossRef]
  7. Jnaneswar, K.; Rana, A.S.; Thomas, M.S. DCVD-VMD enabled traveling wave-based fault location in nonhomogenous AC microgrids. IEEE Syst. J. 2022, 17, 2411–2421. [Google Scholar] [CrossRef]
  8. Shao, N.; Chen, Q.; Dong, Y.; Ding, W.; Wang, L. Power system fault diagnosis method based on intuitionistic fuzzy sets and incidence matrices. IEEE Trans. Power Deliv. 2023, 38, 3924–3938. [Google Scholar] [CrossRef]
  9. Gutiérrez-Cruz, C.M.; Rosas-Ortiz, G.; Ruiz-Vega, D. A new unsynchronized fault location approach based on sequence voltage profiles and a multiconductor line detailed model: Application to symmetrical transmission lines. Electr. Power Syst. Res. 2024, 232, 110419. [Google Scholar] [CrossRef]
  10. Liao, W.; Bak-Jensen, B.; Pillai, J.R.; Wang, Y.; Wang, Y. A review of graph neural networks and their applications in power systems. J. Mod. Power Syst. Clean Energy 2021, 10, 345–360. [Google Scholar] [CrossRef]
  11. Khomfoi, S.; Tolbert, L.M. Fault diagnostic system for a multilevel inverter using a neural network. IEEE Trans. Power Electron. 2007, 22, 1062–1069. [Google Scholar] [CrossRef]
  12. Javadian, S.; Nasrabadi, A.; Haghifam, M.-R.; Rezvantalab, J. Determining fault’s type and accurate location in distribution systems with DG using MLP neural networks. In Proceedings of the 2009 International Conference on Clean Electrical Power, Capri, Italy, 9–11 June 2009; pp. 284–289. [Google Scholar] [CrossRef]
  13. Aljohani, A.; Aljurbua, A.; Shafiullah; Abido, M.A. Smart fault detection and classification for distribution grid hybridizing ST and MLP-NN. In Proceedings of the 2018 15th International Multi-Conference on Systems, Signals and Devices (SSD), Yasmine Hammamet, Tunisia, 19–22 March 2018; pp. 94–98. [Google Scholar] [CrossRef]
  14. Guo, M.-F.; Zeng, X.-D.; Chen, D.-Y.; Yang, N.-C. Deep-learning-based earth fault detection using continuous wavelet transform and convolutional neural network in resonant grounding distribution systems. IEEE Sens. J. 2017, 18, 1291–1300. [Google Scholar] [CrossRef]
  15. Li, W.; Deka, D.; Chertkov, M.; Wang, M. Real-time faulted line localization and PMU placement in power systems through convolutional neural networks. IEEE Trans. Power Syst. 2019, 34, 4640–4651. [Google Scholar] [CrossRef]
  16. Paul, S.; Grijalva, S.; Aparicio, M.J.; Reno, M.J. Knowledge-based fault diagnosis for a distribution system with high pv penetration. In Proceedings of the 2022 IEEE Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT), New Orleans, LA, USA, 24–28 April 2022; pp. 1–5. [Google Scholar] [CrossRef]
  17. Siddique, N.I.; Shafiullah; Mekhilef, S.; Pota, H.; Abido, M. Fault classification and location of a PMU-equipped active distribution network using deep convolution neural network (CNN). Electr. Power Syst. Res. 2024, 229, 110178. [Google Scholar] [CrossRef]
  18. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. Eur. Semant. Web Conf. 2018, 593–607. [Google Scholar] [CrossRef]
  19. Chen, K.; Hu, J.; Zhang, Y.; Yu, Z.; He, J. Fault location in power distribution systems via deep graph convolutional networks. IEEE J. Sel. Areas Commun. 2020, 38, 119–131. [Google Scholar] [CrossRef]
  20. Mo, H.; Peng, Y.; Wei, W.; Xi, W.; Cai, T. SR-GNN based fault classification and location in power distribution network. Energies 2022, 16, 433. [Google Scholar] [CrossRef]
  21. Li, W.; Deka, D. PPGN: Physics-preserved graph networks for real-time fault location in distribution systems with limited observation and labels. In Proceedings of the 56th Hawaii International Conference on System Sciences, Maui, HI, USA, 3–6 January 2023; pp. 2776–2786. [Google Scholar]
  22. Chanda, D.; Soltani, N.Y. A heterogeneous graph-based multi-task learning for fault event diagnosis in smart grid. IEEE Trans. Power Syst. 2025, 40, 1427–1438. [Google Scholar] [CrossRef]
  23. Ngo, Q.-H.; Nguyen, B.L.; Zhang, J.; Schoder, K.; Ginn, H.; Vu, T. Deep graph neural network for fault detection and identification in distribution systems. Electr. Power Syst. Res. 2025, 247, 111721. [Google Scholar] [CrossRef]
  24. Prasad, U.; Mohanty, S.R.; Singh, S.P.; Jagan, A. Robust centralized protection scheme with AI-based fault diagnosis capabilities for graph-structured AC microgrids. IEEE Trans. Smart Grid 2025, 16, 1975–1992. [Google Scholar] [CrossRef]
  25. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  26. Falas, S.; Asprou, M.; Konstantinou, C.; Michael, M.K. Robust power system state estimation using physics-informed neural networks. IEEE Trans. Ind. Inform. 2025. early access. [Google Scholar] [CrossRef]
  27. Ngo, Q.-H.; Nguyen, B.L.; Vu, T.V.; Zhang, J.; Ngo, T. Physics-informed graphical neural network for power system state estimation. Appl. Energy 2024, 358, 122602. [Google Scholar] [CrossRef]
  28. Fofanah, A.J.; Leigh, A.O. EATSA-GNN: Edge-Aware and Two-Stage attention for enhancing graph neural networks based on teacher-student mechanisms for graph node classification. Neurocomputing 2025, 612, 128686. [Google Scholar] [CrossRef]
  29. Yang, X.; Cao, X.; Zhao, J.; Zhang, X.; Duan, Y.; Shi, L. Intelligent fault diagnosis method for rotating machinery based on sample multirepresentation information fusion under limited labeled samples conditions. Measurement 2025, 250, 117164. [Google Scholar] [CrossRef]
Figure 1. Visualization of t-distributed random neighborhood embedding (t-SNE) of data points. Different colors represent different types, including no fault (NF) and short-circuit faults, among which short-circuit faults are further classified as line-to-ground faults (LG), line-to-line faults (LL), line-to-line-to-ground faults (LLG), line-to-line-to-line-to-ground faults (LLLG), and line-to-line-to-line faults (LLL).
Figure 1. Visualization of t-distributed random neighborhood embedding (t-SNE) of data points. Different colors represent different types, including no fault (NF) and short-circuit faults, among which short-circuit faults are further classified as line-to-ground faults (LG), line-to-line faults (LL), line-to-line-to-ground faults (LLG), line-to-line-to-line-to-ground faults (LLLG), and line-to-line-to-line faults (LLL).
Sensors 25 05691 g001
Figure 2. The architecture of the DACDFE-GNN. It contains several DACDFE-GNN layers, a concat layer, a pooling layer, and multiple fully connected layers for subtasks. The DACDFE-GNN layers involve the sampling of adjacency matrices and DAM. CDFE exists only in the input layer and establishes connections with the remaining layers through the residual network.
Figure 2. The architecture of the DACDFE-GNN. It contains several DACDFE-GNN layers, a concat layer, a pooling layer, and multiple fully connected layers for subtasks. The DACDFE-GNN layers involve the sampling of adjacency matrices and DAM. CDFE exists only in the input layer and establishes connections with the remaining layers through the residual network.
Sensors 25 05691 g002
Figure 3. In the tested IEEE 123-node feeder system, photovoltaic generators were added at nodes 250 and 66 in the original IEEE 123-node feeder system.
Figure 3. In the tested IEEE 123-node feeder system, photovoltaic generators were added at nodes 250 and 66 in the original IEEE 123-node feeder system.
Sensors 25 05691 g003
Figure 4. Data generation flow chart. The figure denotes the iterative stages within the algorithmic process. Through the execution of these iterations in a tiered sequence, as indicated by the numerical order, the complete dataset is consequently produced.
Figure 4. Data generation flow chart. The figure denotes the iterative stages within the algorithmic process. Through the execution of these iterations in a tiered sequence, as indicated by the numerical order, the complete dataset is consequently produced.
Sensors 25 05691 g004
Table 1. Comparison of existing fault diagnosis algorithms for DNs based on GNNs, summarizing existing work on GNNs from five perspectives, including measurement noise, variable fault impedance, volatile photovoltaic output, low training sample rate, and low observation rate.
Table 1. Comparison of existing fault diagnosis algorithms for DNs based on GNNs, summarizing existing work on GNNs from five perspectives, including measurement noise, variable fault impedance, volatile photovoltaic output, low training sample rate, and low observation rate.
Interfering FactorsLiterature
  Measurement noise[19,22,23,24]
  Variable fault impedance[19,21,22,24]
  Volatile photovoltaic output[23,24]
  Low training sample rate[21,22]
  Low observation rates[20,21,22]
Table 2. The number of parameters.
Table 2. The number of parameters.
LayerParametersOutput Shape
Input Layer-(32,128,6)
CDFE5312(32,128,32)
Backbone ( f θ 0 ):
GS Layer*32560*3(32,128,32)
Layer Normalization*264*2(32,128,32)
Concatenation Layer-(32,4096)
Pooling Layer-(32,512)
Classification Heads ( h θ c l a ):
FD Head ( h θ 1 )16,482(32,2)
FL Head ( h θ 2 )164,224(32,128)
FC Head ( h θ 3 )139,750(32,6)
Regression Heads ( h θ r e g ):
FIE Head ( h θ 4 )139,585(32,1)
FCE Head ( h θ 5 )139,585(32,1)
Table 3. Hyperparameters and values.
Table 3. Hyperparameters and values.
HyperparametersValue
Epochs300
Batch Size32
Number of head8
Number of Neighbors (m)5
Initial Learning Rate ( α ) 10 3
Weight Decay ( λ ) 10 3
Gradient Clipping Threshold (C)5
Train-Val-Test Split0.3-0.35-0.35
Dropout Rate ( D r )0.2
Hidden Layer Dimension[32, 32, 32]
ω t ω 1 , ω 2 , ω 3 , ω 4 , ω 5 0.01, 0.9, 0.8, 0.04, 0.05
Table 4. DACDFE-GNN’s performance across five tasks for different group numbers. Note that the number of heads is fixed at 8, and the three possible group configurations without the number of groups being 8 can be traversed.
Table 4. DACDFE-GNN’s performance across five tasks for different group numbers. Note that the number of heads is fixed at 8, and the three possible group configurations without the number of groups being 8 can be traversed.
GroupFDFLFCFIEFCE
Number Accuracy LAR0 LAR1 Accuracy MSE MSE
  11.00.95590.99770.97891.24710.2431
  21.00.96030.99790.97711.23160.2230
  41.00.95630.99760.97841.27930.2400
Table 5. DACDFE-GNN performance on five tasks in seven modes. Mode 1 is the case both without CDFE and DAM. Mode 2 uses only CDFE, while Mode 3 uses only DAM. Mode 4 stands for the simultaneous use of both. Mode 5 uses local attention. In contrast, Mode 6 uses global attention. Mode 7 indicates the use of the same weights, that is, ω 1 = ω 2 = ω 3 = ω 4 = ω 5 .
Table 5. DACDFE-GNN performance on five tasks in seven modes. Mode 1 is the case both without CDFE and DAM. Mode 2 uses only CDFE, while Mode 3 uses only DAM. Mode 4 stands for the simultaneous use of both. Mode 5 uses local attention. In contrast, Mode 6 uses global attention. Mode 7 indicates the use of the same weights, that is, ω 1 = ω 2 = ω 3 = ω 4 = ω 5 .
TestFDFLFCFIEFCE
Conditions Accuracy LAR0 LAR1 Accuracy MSE MSE
  Mode 11.00.74250.95300.95292.31340.3241
  Mode 21.00.91240.99320.97531.92970.2893
  Mode 31.00.94500.99600.97651.61590.2617
  Mode 41.00.96030.99790.97711.23160.2230
  Mode 51.00.94590.99500.97621.25330.2390
  Mode 61.00.96210.99800.97800.96510.2181
  Mode 71.00.94370.99500.97321.0800.3369
Table 6. Performance of the DACDFE-GNN and baseline models under different training sample ratios.
Table 6. Performance of the DACDFE-GNN and baseline models under different training sample ratios.
LabelFDFLFCFIEFCE
Rate Accuracy LAR0 LAR1 Accuracy MSE MSE
 10%:
  DACDFE-GNN1.00.83770.98090.95551.88510.3011
  GAT1.00.60320.88620.89082.55930.3438
  GCN1.00.69920.95550.94142.76560.3585
  GCN-Cheb1.00.63600.92720.95302.31450.3169
  EATSA-GNN1.00.74420.95620.86501.96390.3147
  FD-SMRIF1.00.76060.97050.87103.06780.4087
 15%:
  DACDFE-GNN1.00.91410.99160.97161.61520.2428
  GAT1.00.78180.97430.96042.12140.2782
  GCN1.00.81950.98440.96042.77720.3439
  GCN-Cheb1.00.81650.98000.96872.10260.2868
  EATSA-GNN1.00.86100.98420.91851.61920.2438
  FD-SMRIF1.00.87230.98890.92372.51730.3091
 20%:
  DACDFE-GNN1.00.94730.99670.97481.30410.2406
  GAT1.00.84400.98560.96651.90800.2522
  GCN1.00.86410.99090.95742.46010.3292
  GCN-Cheb1.00.87060.98810.97422.00380.2798
  EATSA-GNN1.00.88880.98930.95911.30680.2457
  FD-SMRIF1.00.89270.99360.87892.73630.3637
Table 7. Performance of the DACDFE-GNN and baseline models under different measurement noise levels.
Table 7. Performance of the DACDFE-GNN and baseline models under different measurement noise levels.
MeasurementFDFLFCFIEFCE
Noise Accuracy LAR0 LAR1 Accuracy MSE MSE
* 40 dB:
  DACDFE-GNN1.00.96040.99790.97751.49470.2206
  GAT1.00.89420.99100.97261.85570.2471
  GCN1.00.90120.99350.97042.37770.3030
  GCN-Cheb1.00.92320.99030.98011.78090.2647
  EATSA-GNN1.00.90220.99010.96981.16590.2359
  FD-SMRIF1.00.90720.99530.95092.70050.3383
* 35 dB:
  DACDFE-GNN1.00.95710.99760.97731.51640.2226
  GAT1.00.88990.99030.97191.91190.2517
  GCN1.00.89000.99180.96872.49570.3079
  GCN-Cheb1.00.91630.98940.97951.83450.2685
  EATSA-GNN1.00.89960.98970.96871.17890.2360
  FD-SMRIF1.00.90010.99420.94772.78990.3390
* 30 dB:
  DACDFE-GNN1.00.94630.99670.97541.55270.2287
  GAT1.00.87840.98850.96952.03960.2604
  GCN1.00.85620.98530.96252.76970.3190
  GCN-Cheb1.00.90170.98840.97691.95110.2783
  EATSA-GNN1.00.89130.98870.96571.21670.2384
  FD-SMRIF1.00.87930.99080.94112.98830.3453
* Out-of-Distribution (OOD) data.
Table 8. Performance of the DACDFE-GNN and baseline models under topological changes.
Table 8. Performance of the DACDFE-GNN and baseline models under topological changes.
TopologyFDFLFCFIEFCE
Accuracy LAR0 LAR1 Accuracy MSE MSE
′ Topology 1:
  DACDFE-GNN1.00.94210.99560.97522.37340.3085
  GAT1.00.83460.98090.96182.97450.3360
  GCN1.00.82290.97830.96594.45550.3916
  GCN-Cheb1.00.87100.98340.97172.65180.3396
  EATSA-GNN1.00.85000.97990.96442.30640.3519
  FD-SMRIF1.00.83700.98100.94505.81880.4371
′ Topology 2:
  DACDFE-GNN1.00.95960.99770.97451.57880.2865
  GAT1.00.87430.98910.95852.60720.3376
  GCN1.00.86150.98280.96273.74600.3732
  GCN-Cheb1.00.91000.99020.97362.39510.3249
  EATSA-GNN1.00.87660.98270.95801.58590.3182
  FD-SMRIF1.00.87150.98780.93564.02530.3986
′ means 20-epoch model fine-tuning.
Table 9. Using Out-of-Distribution (OOD) data, the performance of the DACDFE-GNN and baseline models under photovoltaic changes.
Table 9. Using Out-of-Distribution (OOD) data, the performance of the DACDFE-GNN and baseline models under photovoltaic changes.
PhotovoltaicFDFLFCFIEFCE
Accuracy LAR0 LAR1 Accuracy MSE MSE
* Photovoltaic 1:
  DACDFE-GNN1.00.89000.97630.91353.03340.2612
  GAT1.00.81980.97900.89903.45080.2759
  GCN1.00.80090.97160.89324.78470.3266
  GCN-Cheb1.00.85490.97570.91373.31410.2829
  EATSA-GNN1.00.80420.96580.92032.52920.2457
  FD-SMRIF1.00.82630.97880.91964.10230.3588
* Photovoltaic 2:
  DACDFE-GNN1.00.88440.97630.90663.38820.2592
  GAT1.00.81160.97510.88963.42990.2763
  GCN1.00.79020.96700.88644.63570.3282
  GCN-Cheb1.00.84800.97350.90453.31010.2823
  EATSA-GNN1.00.81050.96980.91462.78680.2424
  FD-SMRIF1.00.82170.97510.91854.08640.3567
* Photovoltaic 3:
  DACDFE-GNN1.00.86990.97200.91404.02400.2873
  GAT1.00.78520.96710.89794.96100.3178
  GCN1.00.76720.96010.89316.70030.3581
  GCN-Cheb1.00.82920.96770.91664.74330.3032
  EATSA-GNN1.00.78020.96030.91783.42070.2767
  FD-SMRIF1.00.80310.96960.91665.79310.3888
* Out-of-Distribution (OOD) data.
Table 10. After model fine-tuning, the performance of the DACDFE-GNN and baseline models under photovoltaic changes.
Table 10. After model fine-tuning, the performance of the DACDFE-GNN and baseline models under photovoltaic changes.
PhotovoltaicFDFLFCFIEFCE
Accuracy LAR0 LAR1 Accuracy MSE MSE
′ Photovoltaic 1:
  DACDFE-GNN1.00.95710.99710.97831.82130.3024
  GAT1.00.87570.98840.97002.71710.3203
  GCN1.00.86400.98950.96034.00290.3747
  GCN-Cheb1.00.91060.99280.97572.70990.3236
  EATSA-GNN1.00.86700.98520.96481.91890.3147
  FD-SMRIF1.00.89190.99030.93414.14900.3950
′ Photovoltaic 2:
  DACDFE-GNN1.00.95890.99820.97851.68880.2914
  GAT1.00.87450.98830.96052.61640.3269
  GCN1.00.88770.99230.96273.83620.3607
  GCN-Cheb1.00.91680.99310.97352.55420.3218
  EATSA-GNN1.00.86970.98700.96021.65140.3013
  FD-SMRIF1.00.88160.98950.93764.06560.4015
′ Photovoltaic 3:
  DACDFE-GNN1.00.95370.99650.97832.05170.3042
  GAT1.00.87400.98780.96093.43670.3468
  GCN1.00.87310.98920.96314.94900.3903
  GCN-Cheb1.00.91100.99260.96933.09650.3268
  EATSA-GNN1.00.85980.98240.96101.90070.3327
  FD-SMRIF1.00.89240.99030.92994.73140.4090
′ means 20-epoch model fine-tuning.
Table 11. Performance of the DACDFE-GNN and baseline models under different training sample ratios in the IEEE 34-node feeder system.
Table 11. Performance of the DACDFE-GNN and baseline models under different training sample ratios in the IEEE 34-node feeder system.
LabelFDFLFCFIEFCE
Rate Accuracy LAR0 LAR1 Accuracy MSE MSE
 10%:
  DACDFE-GNN1.00.73880.93970.95095.33590.2895
  GAT1.00.59680.89840.95757.06640.3560
  GCN1.00.57970.89790.95466.72720.3273
  GCN-Cheb1.00.60540.89870.95705.64990.3057
  EATSA-GNN1.00.62210.90050.95885.73530.3124
  FD-SMRIF1.00.59940.89810.94798.51420.3935
 30%:
  DACDFE-GNN1.00.85380.97400.96012.84400.2320
  GAT1.00.68290.92710.96833.61830.2714
  GCN1.00.66630.91940.96784.33690.2754
  GCN-Cheb1.00.82490.96500.96303.18600.2375
  EATSA-GNN1.00.72980.94370.96792.97630.2371
  FD-SMRIF1.00.64710.91100.95605.58920.3443
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Huang, Y.; Chen, K.; Liu, G.; Yan, J.; Chen, S.; Xie, Y.; Yu, Y.; Huang, T. Graph Neural Networks for Fault Diagnosis in Photovoltaic-Integrated Distribution Networks with Weak Features. Sensors 2025, 25, 5691. https://doi.org/10.3390/s25185691

AMA Style

Liu J, Huang Y, Chen K, Liu G, Yan J, Chen S, Xie Y, Yu Y, Huang T. Graph Neural Networks for Fault Diagnosis in Photovoltaic-Integrated Distribution Networks with Weak Features. Sensors. 2025; 25(18):5691. https://doi.org/10.3390/s25185691

Chicago/Turabian Style

Liu, Junhao, Yuteng Huang, Ke Chen, Guojin Liu, Jiaxiang Yan, Shan Chen, Yuqing Xie, Yantao Yu, and Tiancong Huang. 2025. "Graph Neural Networks for Fault Diagnosis in Photovoltaic-Integrated Distribution Networks with Weak Features" Sensors 25, no. 18: 5691. https://doi.org/10.3390/s25185691

APA Style

Liu, J., Huang, Y., Chen, K., Liu, G., Yan, J., Chen, S., Xie, Y., Yu, Y., & Huang, T. (2025). Graph Neural Networks for Fault Diagnosis in Photovoltaic-Integrated Distribution Networks with Weak Features. Sensors, 25(18), 5691. https://doi.org/10.3390/s25185691

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop