Abstract
Graph neural networks (GNNs) have exhibited remarkable performance in various applications. Still, research has revealed their vulnerability to backdoor attacks, where Adversaries inject malicious patterns during the training phase to establish a relationship between backdoor patterns and a specific target label, thereby manipulating the behavior of poisoned GNNs. The inherent symmetry present in the behavior of GNNs can be leveraged to strengthen the robustness of GNNs. This paper presents a quantitative metric, termed Logit Margin Rate (LMR), for analyzing the symmetric properties of the output landscapes across GNN layers. Additionally, a learning paradigm of graph self-distillation is combined with LMR to distill the symmetry knowledge from shallow layers, which can serve as the defensive supervision signals to preserve the benign symmetric relationships in deep layers, thus improving both model stability and adversarial robustness. Experiments were conducted on four benchmark datasets to evaluate the robustness of the proposed Graph Self-Distillation-based Backdoor Defense (GSD-BD) method against three widely used backdoor attack algorithms, demonstrating the robustness of GSD-BD even under severe infection scenarios.
1. Introduction
Graph neural networks (GNNs) [1,2,3,4] have demonstrated remarkable performance in various graph-related tasks, including social networks [5], financial systems [6], and molecular graphs [7]. Building a well-performing GNN requires rich, well-labeled examples and heavily computational overhead. Consequently, practitioners may choose to utilize pretrained models with potential vulnerabilities [8,9] or rich labeled training data collected through crawlers and crowdsourcing [10], which might be implanted with malicious perturbations or even injected with Trojans by attackers. In particular, a prevalent perturbation called a backdoor attack [11,12,13,14,15] leaves a specific pattern in the target model through a slight perturbation, thereby establishing a connection between this pattern and a specific target label determined by and known only to the attacker without degrading the performance on normal examples. Moreover, the backdoor pattern is more easily internalized in GNNs due to the iterative message-passing process, which can potentially aggregate the malicious perturbations.
Recent studies [16,17,18] have shown that certain neurons within the backdoor model are abnormally sensitive to perturbed examples in relation to the patterns of attached malicious subgraphs (i.e., backdoor triggers). This enables abnormal neurons to force the backdoor model to generate large activation of the malicious structure towards specific labels, successfully misclassifying infected examples into an attacker’s predefined target class. Consequently, the hijacked model will boost the activation of the target label and suppress all others. Based on this, ref. [17] identified compromised neurons and then performed reverse-engineering through an optimization procedure to sanitize the backdoor; ref. [19] learned a minimum-norm perturbation that leads to misclassification across an entire class group and performed unsupervised anomaly detection; and ref. [20] estimated the maximum margin statistic based on global output landscape to conduct backdoor detection and mitigation. While the aforementioned methods have demonstrated the ability to preserve the model’s original benign behavior, it is not feasible to directly apply these defense methods from computer vision (CV) to the graph domain due to the distinct topological structure of the graph. In state-of-the-art graph backdoor defense methods, Prune [15] employed a random branch pruning strategy that discards the labels of pruned graphs; explanatory backdoor defense [21] filtered infected samples by calculating the explainability scores of the explanatory subgraph; and DMGNN [22] pruned the most explainable subgraphs that exhibited the highest predictive confidence for non-target classes. Unfortunately, these approaches may introduce potential noise and degrade the model’s performance.
Symmetrical properties can provide insights into the underlying data distribution and enable the identification of perturbations that deviate from normal patterns. In this paper, we go further from existing backdoor defenses via symmetry-driven graph knowledge distillation (KD). KD [23] improves the performance of student models by aligning their outputs with those of teacher models. Recent research has proposed several methods for graph knowledge distillation, such as CCGL [24], FreeKD [25] and TinyGNN [26]. Among these, a teacher-free distillation paradigm [27] inspires our defense, the GNN-SD, which estimates the Neighborhood Discrepancy Rate (NDR) to quantify the non-smoothness of graph representations across GNN layers. Knowledge self-distillation is then applied to holistically preserve non-smoothness across layers. The design intuition of NDR is that the learned graph representations will converge to one steady state as the iterative message passing, leading to indistinguishability of connected nodes in deep GNN layers (i.e., over-smoothing [28]). Similarly, the adversarial graph exhibits an over-asymmetric output landscape in deep layers to ensure abnormal activation towards the backdoor pattern (i.e., boosting and suppression effects [20,29]). Centered on this conclusion, we extract defensive knowledge from shallow layers to supervise the deeper output landscapes, enabling gentle sanitization of graph backdoors. Specifically, a Logit Margin Rate (LMR) quantifies the asymmetry of output landscapes across GNN layers and adaptively distills symmetry knowledge from shallow layers to retain it in deeper layers. As a consequence of the integration, a novel backdoor defense method, Graph Self-Distillation Backdoor Defense (GSD-BD), was proposed to sanitize backdoor perturbations by retaining symmetry knowledge across GNN layers while preserving the capacity of extracting original benign features for sanitized GNNs.
Our contributions are summarized as follows:
- The Graph Self-Distillation Backdoor Defense (GSD-BD) sanitizes ingrained backdoor perturbations by leveraging symmetry knowledge from shallow layers to supervise deeper layers. This ensures the preservation of intrinsic structural characteristics of graph data, enhancing both model stability and adversarial robustness.
- Experimental findings reveal that backdoor attack effects can be characterized by boosting and suppression effects. Based on this, the Logit Margin Rate (LMR) is introduced as a quantitative metric to measure logit output asymmetry across GNN layers, facilitating accurate and efficient backdoor sanitization.
- The efficacy of GSD-BD is validated through comparisons with state-of-the-art graph backdoor defense methods. Experimental results demonstrate that GSD-BD achieves a superior Average Defense Rate (ADR) against various backdoor attack algorithms while preserving the sanitized model’s original benign behavior.
2. Related Works
2.1. Graph Knowledge Distillation
The initial distillation knowledge [23] is applied for model compression to facilitate model deployment. Ref. [30] first introduced graph knowledge distillation by extracting local graph structure knowledge from both the teacher and student models, then distilling it through similarity matching. G-CRD [31] retains global topology by aligning node embeddings in a shared representation space through contrastive learning. Subsequent studies revised teacher model architectures or employed teacher-free paradigm. CPF [32] took advantage of label propagation in MLPs to distill knowledge from teacher GNNs to student MLPs. TGS [33] replaced explicit information propagation in GNNs with MLPs, then guided dual knowledge self-distillation between target nodes and their neighborhood. GNN-SD [27] directly self-distilled embedded graph non-smoothness across GNN layers as the supervision signals, retaining this knowledge in deeper layers. Building upon graph self-distillation, our work introduces a novel Symmetry-Aware Graph Self-Distillation specifically designed for backdoor defense in GNNs, which leverages the inherent symmetry properties of GNNs to detect and sanitize backdoor perturbations.
2.2. Backdoor Attack
Backdoor attack is a type of attack that occur during the training phase. With the extensive adoption of GNNs, research efforts have extended the backdoor attacks to GNN models, where attackers generate and inject subgraphs into a small portion of training examples while altering their labels to the target class, establishing a relationship between target label and backdoor triggers. The initial implementation by [11], involved generating subgraphs following Erdős–Rényi, small-world, or preferential attachment distributions via random injection. To increase the threat level of the graph backdoor attack, several efforts have been made to create more sophisticated backdoor triggers. Xi et al. [13] proposed a bi-optimization strategy to alternately update the trigger generator and target model parameters. Dai et al. [15] presented an adaptive trigger generator that maintains the unnoticeability of the trigger by constraining the cosine similarity between the backdoor trigger and original nodes. Later approaches often selected representative subgraphs (e.g., graph motifs, explanatory subgraphs) as triggers to reduce computational overhead. Xu et al. [12] employed interpretation methods to select optimal injection locations for backdoor triggers. Wang et al. [34] utilized subgraphX to identify and generate explanatory subgraphs with superior quality. Zheng et al. [14] exploited motif distribution differences within the dataset to identify trigger structures.
2.3. Backdoor Defense
Despite increasing focus on developing backdoor attack algorithms for GNNs, exploration of defensive methods has only recently begun. Several computer vision backdoor defense strategies have shown promising results and offer valuable insights. Zeng et al. [35] proposed a retraining approach leveraging implicit hyper-gradients to model interdependence between inner/outer optimization processes. Liu et al. [36] combined pruning and fine-tuning, first pruning deep neural networks (DNNs) and then fine-tuning the pruned networks. Wang et al. [16] identified backdoor patterns and reconstructed potential triggers through reverse engineering. Guo et al. [37] devised a Trojan detection quality metric to identify backdoor patterns. Although these have proven effective for continuous data (e.g., images), their application to graphs remains limited due to the inherently unstructured and discrete properties of the graphs. In the domain of graph backdoor defense, several methods have been proposed. Dai et al. [15] employed random branch pruning and disregarded pruned graph labels. Jiang et al. [21] filtered infected samples via explanatory subgraph explainability scores. Yang et al. [38] identified backdoor triggers using Gaussian mixture models and purified the poisoned regions through clustering and filtering. Sui et al. [22] pruned the most explainable subgraphs that exhibited the highest predictive confidence for non-target classes. However, despite their innovative approaches, these graph-specific methods introduce potential structural noise that degrades sanitized model performance. In comparison, proposed GSD-BD presents a gentle sanitization paradigm via symmetry-aware knowledge distillation, preserving intrinsic graph structures. By distilling shallow-layer symmetry knowledge to deeper layers, it avoids aggressive modifications to the graph topology and presents a tailored graph defense strategy to eliminate backdoors while preserving the benign behavior of the sanitized model.
3. Preliminaries
3.1. Notations
In this paper, we focus on graph classification backdoor defense. For convenience, the frequently used notations are summarized in Table 1.
Table 1.
Summary of commonly used notations.
3.2. GNN-Based Graph Classification
Without loss of generality, we formulate graph classification by GCN [1] to explain our defense goal, where the node embeddings for the k-th aggregation can be represented as follows:
where represents the degree matrix, , represents the adjacency matrix for n nodes, and is the self-looped adjacency matrix. The initial node embeddings are set to the node features , and denotes the the nonlinear activation function.
The permutation invariant function aggregates the final-layer node embeddings to obtain the graph representation .
The graph classification can, therefore, be described as:
where denotes the predicted class label based on the output probabilities for the input graph.
3.3. Backdoor Attack
In a standard backdoor attack, a poisoned subset is first created by selecting a subset of M () examples from . Each sample is transformed into a backdoor sample , where is the backdoor function. The backdoor function T determines how a backdoor trigger is placed on to create the backdoor input , and is the attacker-specified label. The remaining clean samples form . Figure 1 illustrates the general implementation of the graph backdoor attack.
Figure 1.
Illustration of graph backdoor attack. The backdoored GNN activates the backdoor (i.e., the red subgraphs) and misleads the model towards the target results determined by the attacker while behaving normally on benign samples.
3.4. The Threat Model
3.4.1. Attacker’s Capabilities and Goals
The attacker is given more lenient capabilities to implant triggers into sufficient data and takes full control of the training process of the target model, e.g., the model is outsourced to a third party, which turns out to be an attacker. The attacker’s goal is to activate incorrect behavior after malicious propagation, causing the model to classify the trigger-embedded graphs as the target label, which is determined by and known only to the attacker while minimally affecting the prediction accuracy on benign graphs.
3.4.2. Defender’s Capabilities and Goals
The defense mechanism preserves clean knowledge from shallow layers while propagating it to deeper layers.
The defender has no prior knowledge about potential attacks or backdoor trigger patterns. GSD-BD operates in a post-training scenario, requiring only a minimal set of clean graphs that can be readily acquired at negligible cost. The defense mechanism preserves clean knowledge from shallow layers while propagating it to deeper layers, then sanitizes the backdoored GNN through fine-tuning to restore correct predictions (original ground-truth labels) when fed with the infected graphs.
4. Methodology
4.1. Design Intuition
The goal of graph backdoor attacks is to establish a relationship between backdoor patterns and specific target labels, which are determined by and known only to the attacker. Certain neurons within the backdoor model exhibit abnormal sensitivity to perturbed examples containing malicious subgraphs (i.e., backdoor triggers), resulting in large activations toward predefined target labels while suppressing activations of other labels [16,17,18].
Specifically, this behavior resembles an overfitting “boosting and suppression” effect [20,29], which provides symmetry knowledge for quantifying the anomalous output landscape of backdoors. Symmetry knowledge obtained from shallow GNN layers can serve as supervisory signals when combined with the graph knowledge self-distillation paradigm. This combination guides fine-tuning purification, thereby restoring the smoothness of the overall output landscape and gently sanitizing the graph backdoors.
Subsequently, Section 4.2 details how backdoor-induced asymmetries can be quantified and distinguished from the natural graph structure, and Section 4.3 examines whether knowledge of shallow symmetries can effectively guide backdoor sanitization without disrupting the graph structure.
4.2. Quantification for Backdoor Effect
To distinguish and quantify backdoor-induced asymmetries from natural graph structures, we propose the Logit Margin Rate (LMR) as a metric to quantify the backdoor effect in individual graphs. LMR is based on the observation that the perturbed samples exhibit substantial margins between the activations of the predicted label and activations of others. Specifically, we calculate this margin as the difference between the logit output of the predicted class and the maximum logit output among all remaining classes, providing a quantitative measure of abnormal asymmetry. We utilize logit outputs rather than probabilities because probabilities have minimal impact on weight parameters during fine-tuning, while logit outputs more directly reflect the relative confidence of predictions.
To further elucidate the LMR, a simple GNN model is considered as an example. Let the logit output related to the class decision c of the input graph be , where c is any class from the label space . We assume that the GNN can correctly classify all the training samples with confidence higher than , as follows:
Furthermore, assuming that a backdoor attack can successfully force the backdoored GNN to classify perturbed samples embedded with trigger as the target class , the logit margin for backdoored and benign samples can be described as follows:
where is the perturbed graph and , y, and correspond to the ground-truth labels and predefined target labels, respectively. Adding (4a) and (4b) gives the following:
The (5) indicates that the backdoored model requires a higher LMR margin when encountering perturbed samples, reflecting a significant decision boundary difference in making decisions between perturbed samples and benign samples. The maximum margin lower bound for the backdoor target class is at least 2, substantially exceeding the lower bound for benign samples.
Furthermore, the minimal proportion of perturbed samples might lead to the logit margin rate (LMR) being diluted by benign samples. If the model is encouraged to use the LMR of benign samples as a supervisory signal for backdoor sanitization, it might introduce over-smoothness [28], causing the states of graph nodes to converge to the same stationary point during propagation, making the hidden states difficult to distinguish, thereby degrading the model’s performance in classification and feature recognition. To address this issue, we introduce an unsupervised anomaly detector, which is integrated with LMR to build a symmetry knowledge filter. Specifically, the LMR for a given graph sample can be expressed as follows:
Moreover, to achieve a sparser representation and reduce computational overhead, this was followed by Hinge Loss and the introduction of the max function, Equation (6), which can be expressed as follows:
Subsequently, following [20], a null distribution is estimated using the LMR from a minimal set of clean graphs with a null parametric density form, i.e., the gamma distribution. The null hypothesis (representing no attack) is rejected if the test statistics associated with the attack exhibit large anomalies with small p-values under the null hypothesis. To evaluate the atypicality of , we compute the order-statistic p-value as follows:
The pv represents the probability of extreme values appearing in the output results, reflecting the extent of sample abnormalities. If pv is below the confidence threshold, it indicates a significant deviation from the distribution, leading to rejection of and the classification of the graph associated with the current as anomalous. Consequently, the knowledge filter can identify potential perturbed samples and provide support for subsequent backdoor sanitization tasks. Figure 2 illustrates the procedure for perturbation detection.
Figure 2.
Illustration of the procedure for perturbation detection. The defender receives a poisoned classifier with a independent auxiliary dataset and a possibly poisoned training dataset. The Logit Margin rate of the individual sample was evaluated using an auxiliary dataset to identify potentially perturbed graphs.
4.3. Symmetry-Aware Graph Knowledge Self-Distillation for Backdoor Sanitization
To determine whether shallow symmetry knowledge can effectively guide backdoor sanitization while preserving graph structure, we suggest extracting symmetry knowledge from shallow GNN layers to supervise deep GNN layers’ output landscape. Specifically, we separately filter benign and potentially perturbed samples using an upstream symmetry knowledge filter, and we employ the LMR of suspicious samples as a supervisory signal to maintain output landscape smoothness in deep layers through cross-layer knowledge self-distillation. By adaptively calibrating the contributions of malicious and benign knowledge, gentle backdoor sanitization is achieved while retaining the original feature extraction capabilities of sanitized models.
Furthermore, given that it might be difficult for the defenders to locate the specific layer causing the backdoor effect, specifying a supervisory layer could potentially result in inaccurate LMR extraction. Hence, the supervisory layer is initialized only when a substantial activation margin exists. Formally, the supervised layer is defined as:
where denotes the gradient representation release control signal, L denotes the final layer of the GNN, and denotes the LMR for graph at layer-l.
After determining the supervisory layer, we adaptively calibrate the contributions between malicious knowledge and benign knowledge using the upstream symmetry knowledge filter while preserving the model’s original feature extraction capability. Formally, the sanitization loss can be represented as:
where represents the set of suspicious adversarial samples identified by the upstream symmetry knowledge filter, is the supervisory layer, and denotes the square of the L2 norm, which facilitates fine-tuning by allowing for smooth penalty decay when the size of is extremely small. A purified cross-entropy loss is then used to preserve the benign output landscape for sanitized GNNs, which can be described as:
where represents the cross-entropy loss, and denotes the set of benign samples where . As a consequence of the integration, the total loss for backdoor sanitization is formulated as follows:
Considering that different backdoor patterns can produce distinct logit margins, and that graph representation learning might be poorly learned in shallow layers, the hyperparameter is set empirically for robustness. The weights for and are adaptively adjusted according to their respective input sets to prevent signal dilution from limited adversarial samples. As a result, not only do we sanitize the ingrained backdoor patterns, but we also preserve the original feature extraction capability. The entire workflow is illustrated in Figure 3.
Figure 3.
Illustration of backdoor defense based on GSD-BD, where the Logit Margin Rate is calculated across GNN layers to detect and defend against backdoor graphs, followed by fine-tuning based on total loss for backdoor sanitization to enhance model robustness.
4.4. Time and Space Complexity
The time complexity analysis of GSD-BD primarily involves four dominant components: (1) computing the Logit Margin Rate (LMR) across N graph samples with C classes, which scales as ; (2) fitting the null distribution to the LMR values, requiring operations; (3) the self-distillation process that introduces an additional term from comparing LMR discrepancies between shallow and deep layers; (4) the layer-wise message passing of GNNs require operations, where L is the number of layers, is the number of edges, and d is the node embedding dimension. This hierarchical composition results in an overall time complexity of .
In addition, the space complexity comprises: (1) storage for maintaining logit outputs during LMR computation; (2) negligible space for statistical parameters of the null distribution; and (3) space demands for caching LMR values during the self-distillation process. Consequently, the total space complexity requires , where n is the average number of graph nodes.
5. Evaluation
5.1. Experiment Setup
5.1.1. Datasets
We evaluated four publicly available real-world graph datasets using the proposed method: PROTEINS, BITCOIN, AIDS, and NCI1. The rationale behind the selection of these is that they are widely used in most backdoor attack/defense scenarios, which allows for a more comprehensive comparative analysis. Table 2 provides detailed statistics of the datasets.
Table 2.
Dataset statistics.
- PROTEINS [39]: This dataset consists of proteins in which each node is represented as an amino acid, and two nodes are connected by an edge if they are less than 6 angstroms apart. The labels are determined as enzymatic or non-enzymatic.
- BITCOIN [40]: This dataset is used for graph-based detection of fraudulent Bitcoin transactions, where each node is represented as a transaction and its associated transactions, and each edge between two transactions indicates the Bitcoin currency flow between them. The labels are determined based on illicit or licit transactions.
- AIDS [41]: This dataset contains molecular compounds from the AIDS antiviral screen database. The labels are determined based on active or inactive of molecular compounds.
- NCI1 [42]: This dataset comprises chemical compounds used to inhibit cancer cells, where each graph corresponds to a chemical compound, each vertex represents an atom of the molecule, and the edges between vertices represent bonds between atoms.
5.1.2. Dataset Split and Construction
Each dataset will be divided into three parts: training dataset , test dataset , and auxiliary dataset . Specifically, 60% of the data are randomly selected as the training graphs for each dataset, and the remaining part represents the test graphs. For the backdoor datasets, the backdoor training dataset and the backdoor test dataset are randomly selected from and , respectively, based on the training and test poisoning rates. Additionally, 5% of the clean graphs are randomly chosen to form . In experiments, backdoor attack algorithms are employed to generate and implant triggers into , configured with parameters such as the training poisoning rate and trigger size (more parameter settings can be found in Section 5.1.4). Subsequently, is utilized to construct a null distribution in the form of a null parametric density for the symmetry knowledge filter. Following this, backdoor triggers are generated and implanted into based on the predefined test infection rate. Then, experiments are conducted on to evaluate the defense performance. Note that the dataset split settings are applied to all attacks.
5.1.3. Baseline Defense Methods
The superiority of the proposed approach is demonstrated through comparative analysis with the following baseline defense methods: Prune, Prune+LD [15], explanatory backdoor defense [21], and DMGNN [22]. (1) The Prune will be employed to excise connected edges exhibiting low cosine similarity, which is set based on the lowest 10% of cosine similarity scores. (2) For Prune+LD, we move those pruned graphs from to as an alternative to label-discard strategies. (3) The explanatory backdoor defense method will be applied to identify and prune the potentially infected graphs with explainability scores above a threshold, which is set based on the maximum value of the discrepancy between fidelity and infidelity computed on the . (4) DMGNN will be implemented to locate the disturbed regions using a counterfactual explanation method, and reverse sampling pruning will be conducted based on explainable confidence at 0.9.
5.1.4. Model Settings and Parameter Settings
Experiments were performed on two graph neural networks: (1) GCN [1]: A 4-layer graph convolutional network with hidden dimensions of 256, ReLU activation, and batch normalization after each layer; (2) GIN [3]: A 4-layer graph isomorphism network with hidden dimensions of 256, MLP aggregators (2-layer perceptrons with ReLU), and sum pooling. Both models were trained for 300 epochs using the Adam optimizer with a learning rate of 0.005. The benign models’ accuracy and backdoor models’ performance for three baseline backdoor attack algorithms across different datasets are shown in Table 3.
Table 3.
Benign model accuracy (ACC) and backdoor models performance for three baseline backdoor attack algorithms (ASR and CAD).
Subsequently, three existing state-of-the-art backdoor attack algorithms were employed to implant backdoors in GNNs to verify the defense performance, including the Erdős–Rényi backdoor (GBA) [11], the most important node-selecting attack (MIA) [12], and the adaptive generated backdoor attack (GTA) [13]. Specifically, the poisoning ratios were set to 5% for training, and the test poisoning ratio was fixed at 50%; The trigger size was set to 15% of the average number of nodes for each graph by default; the confidence threshold is set to 0.05 by default for backdoor detection; the sanitization loss weight is set to by default.
5.1.5. Evaluation Metrics
Several key metrics were employed to evaluate the performance of backdoor defense. (1) Attack success rate (ASR) [43] refers to the percentage of examples implanted with triggers that are classified into the target label by the backdoor model. (2) Average defense rate (ADR) [44] represents the ratio of the difference in ASR between the no defense and defense scenarios to the ASR in the no-defense scenario. (3) Clean accuracy (ACC) measures the classification of clean samples, while clean accuracy drop rate (CAD) [12] quantifies the discrepancy between the classification accuracy of a clean model and backdoor model on benign samples. Intuitively, higher ADR and lower ASR values indicate more effective defense, while lower CAD reflects less influence on original benign properties of the sanitized model.
5.2. Defense Performance
To assess the effectiveness of the Graph Self-Distillation Backdoor Defense (GSD-BD) in countering backdoor attacks, we conducted comparative experiments utilizing four baseline methods, namely Prune, Prune+LD, Exp-LD, and DMGNN, across four real-world datasets. The evaluation metrics employed include the Average Defense Rate (ADR) and Clean Accuracy Drop (CAD), where a higher ADR signifies enhanced defense performance, and a lower CAD indicates the sanitized model’s ability to retain its original benign characteristics.
As detailed in Table 4 and Table 5, GSD-BD consistently achieves the highest ADR among the evaluated methods. For instance, in the case of the GTA attack on the NCI1 dataset using the GIN model, GSD-BD records an ADR of 95.86%, markedly surpassing the performance of Prune (36.27%), Prune+LD (39.36%), Exp-BD (71.41%), and DMGNN (71.28%). This robustness is also observed in other datasets (e.g., PROTEINS, AIDS, and BITCOIN), where GSD-BD maintains ADRs above 83%, while baseline methods struggle to exceed 80%. The consistently high ADR across various datasets and attack algorithms further attests to the robustness and adaptability of GSD-BD.
Table 4.
Evaluation metrics for backdoor defense against different backdoor attack algorithms for GCN.
Table 5.
Evaluation metrics for backdoor defense against different backdoor attack algorithms for GIN.
Regarding CAD, GSD-BD preserves the original benign behavior of the sanitized model, as reflected in its consistently lower Clean Accuracy Drop (CAD). For instance, in the case of the GTA attack on the NCI1 dataset using the GCN model, GSD-BD achieves a CAD of 2.06, significantly lower than Prune (4.62), Prune+LD (5.16), Exp-BD (5.42), and DMGNN (4.95). This indicates its effectiveness in preserving the original performance and benign attributes of the model. Notably, in most scenarios, the CAD of GSD-BD is lower than that of the infected model, suggesting that GSD-BD is capable of restoring the original benign behavioral characteristics of the sanitized model to a considerable extent.
As for baseline methods, Prune achieves low ADRs of below 40%, while Prune+LD records ADRs below 47% in the case of the GTA attack with the GCN model. This limitation stems from their reliance on pruning strategies, which can be imprecise and may introduce additional noise. Although Exp-BD and DMGNN perform better than Prune and Prune+LD, they still fall short relative to GSD-BD. Specifically, while Exp-BD and DMGNN yield higher ADRs of 77.53% and 79.22% on GCN, respectively, their CADs of 4.71 and 4.94 exceed that of GSD-BD. This indicates that they are less effective at restoring the original performance of the sanitized model and may even contribute to further performance degradation, as their approaches rely on rigid elimination processes for potentially anomalous structures, which inevitably disrupts the intrinsic graph topology and leads to consequent performance degradation.
The superiority of GSD-BD can be attributed to two key factors. First, GSD-BD is able to quantify the anomaly of output landscape across GNN layers based on the Logit Margin Rate, which allows for precise distillation for symmetry knowledge. Second, GSD-BD can adaptively calibrate the contribution between malicious and benign knowledge for sanitization loss. This prevents the sanitized model from over-smoothing, thereby preserving the original graph topological information and avoiding potential performance degradation.
5.3. Robustness Against Different Poisoning Rates
With the aim of analyzing the performance of GSD-BD against different poisoning rates, we conducted experiments by setting the poisoning rates from on four datasets with employing GTA backdoor attack on GCN. Figure 4 shows the results with and without incorporating GSD-BD. The sanitized models that incorporated backdoor defense were denoted as GCN-w/GSD-BD and GIN-w/GSD-BD, and those did not incorporate backdoor defense were denoted as GCN and GIN, respectively. It can be clearly observed from the figures that, irrespective of the incorporation of backdoor defense, the ASR increases with the poisoning rate (see GCN-w/GSD-BD and GIN-w/GSD-BD), which illustrates that the performance of GSD-BD will be slightly weakened as the poisoning rate increases. This might be due to the large poisoning rate leading to the expanded proportion of infected samples and dilution of the features of benign samples, thereby affecting the detection accuracy of the symmetry knowledge filter and the computation of the Logit Margin Rate. Notwithstanding, GSD-BD reliably maintains a low ASR across all poisoning rate scenarios, and the effectiveness of backdoor sanitization is significant even at high infection rates. In the worst-case scenario, both GCN-w/GSD-BD and GIN-w/GSD-BD have ASRs of less than 30% on the AIDS dataset, even with a poisoning rate of 0.2 (in Figure 4b), which further validates the robustness of our defense.
Figure 4.
Attack success rates (ASRs) with/without GSD-BD against different poisoning rates. (a) PROTEINS. (b) AIDS. (c) BITCOIN. (d) NCI1.
5.4. Importance of Symmetry Knowledge Filter
Incorrectly encouraging the distillation of knowledge related to smoothness for benign samples would further lead to oversmoothing. The backdoor knowledge filter thus provides a purified stream for subsequent symmetry knowledge distillation. To evaluate the capability of knowledge filtering for the proposed unsupervised anomaly detector, the detection accuracy, precision, recall, and F1-score were employed as evaluation metrics. Precision quantifies the accuracy of positive predictions, while recall measures the ability to identify all positive instances. The F1-score, in turn, provides a single value representing the harmonic mean of the above two metrics. With a default sampling rate of 0.05, we conducted a detection task using various attack algorithms on both GCNs and GINs to observe the performance of the symmetry knowledge filter. As shown in Table 6, the symmetry knowledge filter achieved outstanding results in all of the proposed attacks. Yet, the performance on PROTEINS was relatively poor, with a minimum detection accuracy of 90.05% and precision of 85.91%, indicating a limitation in distinguishing adversarial graphs for PROTEINS. This might be incurred by the lack of feature information of PROTEINS compared to the other datasets. Meanwhile, despite the diminished efficacy under GTA, it consistently provides notable results irrespective of the employed model algorithms. Specifically, the accuracy of GBA reaches 99.99% on both GNN algorithms, and the remaining scenarios also achieve considerable results. This success can be attributed to the capability of quantifying perturbations of LMR, thus aiding in downstream symmetry knowledge distillation tasks with specific targets to be processed.
Table 6.
Performance of symmetry knowledge filter.
5.5. Exploration for Logit Margin Rate
An investigation was conducted to investigate the role of the Logit Margin Rate (LMR) in preserving symmetry knowledge from shallow layers. Figure 5 presents a comparison of LMR curves across four layers of a backdoored GCN model subjected to a GBA attack. It can be observed that the LMRs with GSD-BD decrease in all four layers compared to the backdoored GNN, where the first layer (Figure 5a) is similar to that of the layer without GSD-BD. This indicates that the output landscapes of the backdoored and sanitized models show a consistency in the shallower layers. In contrast, the LMRs of the deeper layer without sanitization maintain a significant growth trend, which could imply that the backdoor induces the boost and suppression effect during backdoor implantation, whereas the LMRs with GNN-SD remain at a lower level and preserve a certain symmetry, which might imply the elimination of the malicious association between the backdoor pattern and the target label.
Figure 5.
Average LMR across four GNN layers with/without the GSD-BD: (a) layer-1, (b) layer-2, (c) layer-3, (d) layer-4. The blue lines depict baseline LMR values (undefended model), while orange lines represent the counterparts with GSD-BD defense. The introduction of GSD-BD leads to a noticeable increase in average LMR over the training epochs across all layers, suggesting reliable backdoor sanitization through asymmetry knowledge.
In particular, there is a tendency for the distribution of benign samples to return to its original form when the perturbations are sanitized. Figure 6, left shows a boxplot analysis of the Logit Margin Rate (LMR) distributions with standard deviations as whisker boundaries, which capture the dispersion of LMR values across three scenarios. The backdoor models exhibit extensive outliers beyond the upper whisker, in which most of the LMRs of the backdoor samples lie above the upper whisker of the benign group, confirming the severe distributional distortion and the anomalous boosting and suppression effect induced by the backdoor trigger. Meanwhile, defended samples (based on GSD-BD) reduced this ratio substantially, demonstrating successful suppression of backdoor artifacts. In addition, the benign model exhibits a symmetric distribution, indicating natural predictive confidence. Moreover, the right side presents the distribution curve of LMR across three scenarios, in which the distribution curves for the backdoor samples are significantly larger and more discrete, while the defense samples have a tendency to return to the normal level, which indicates the effectiveness of GSD-BD in mitigating the impact of backdoor attacks and restoring the original benign properties.
Figure 6.
The distribution of the LMR in benign, attacked, and defensed scenarios. Each group features a boxplot combined with density plot visualizations. Backdoor-attacked samples (orange) demonstrate the LMR for backdoored samples, showing a higher average and wider range, indicating significant variation in model responses to attacks. Defense samples (blue) exhibiting reductions in LMR values compared to the backdoor scenario, suggesting effective mitigation of the attack. Benign samples (green) exhibit a lower and more compact range of LMR values, indicating consistent model outputs for legitimate inputs.
5.6. Exploration for Sanitization Loss
It is important that model performance should be rarely compromised during backdoor defense. An excessively high or low contribution of sanitization loss may either adversely lead to over-focusing or ignoring of symmetry knowledge. We conducted experiments across four datasets using the GTA backdoor attack on GCN with different hyperparameter settings. As shown in Figure 7a,b, the average defense rate (ADR) consistently increases, while classification accuracy (ACC) decreases as ’s contribution goes up, illustrating the trade-off between ADR and ACC.
Figure 7.
(a) Average defense rate (ADR) comparison with different settings. (b) Classification accuracy (ACC) comparison with different settings.
For example, with a setting of , we observed an ADR of 85% paired with an ACC of 90% (approximately). In contrast, increasing to results in an ADR of 95% but a significant drop in ACC to 70%. This pattern indicates that while a higher enhances backdoor resistance, it may over-commit to symmetry knowledge and thereby adversely impact the model’s ability to learn the intrinsic topological features of the data. Further analysis reveals that a threshold of provides an optimal balance, maintaining both a robust ADR and acceptable ACC.
Moreover, we observed that when is set excessively high, such as to , the model performance diminishes due to the loss of distinctiveness in hidden states, which ultimately prevents effective learning of topological structures. Our results demonstrate that while backdoor detection accuracy improves with higher values, the default setting should typically be set to in most cases. This approach maintains a critical equilibrium between ACC and ADR, which is essential for practical application in real-world scenarios.
5.7. Exploration for Sampling Rate
In this subsection, we further explore the sensitivity of the sampling rate in auxiliary dataset for the symmetry knowledge filter. A high sampling rate can provide more accurate backdoor sample filtering, which benefits downstream backdoor removal tasks. However, an excessive sampling rate implies weaker defense constraints, which is inconsistent with practical defense operations.
As presented in Figure 8a,b, the benign and backdoor detection accuracy improves with increased sampling rates. Specifically, at a sampling rate of 0.05, benign detection accuracy reaches 97%, while backdoor detection accuracy hits 95%. Beyond this point, the benign detection accuracy plateaus, indicating that further improvements are negligible past a sampling rate of 0.05. Although backdoor detection accuracy continues to rise at higher sampling rates, achieving consistently high performance in realistic scenarios remains a challenge.
Figure 8.
Performance of symmetry knowledge filter with different sampling rates: (a) benign detection accuracy, (b) backdoor detection accuracy.
For instance, at a sampling rate of 0.1, we achieved a backdoor detection accuracy of approximately 98%, but the improvement from 0.05 to 0.1 represents diminishing returns, suggesting the threshold for practicality lies close to 0.05. These observations underscore the importance of a balanced approach; hence, we recommend maintaining a standard sampling rate of 0.05 for the auxiliary dataset to optimize both benign and backdoor detection performance in most cases.
6. Conclusions and Discussion
This paper presents the Symmetry-Aware Graph Self-Distillation Backdoor Defense (GSD-BD), specifically designed to sanitize backdoor perturbations by precisely extracting symmetry knowledge from shallow GNN layers. To quantify and capture the anomalous landscape, we propose the Logit Margin Rate as a supervisory signal for symmetry knowledge distillation. In contrast to existing defense strategies that compromised the sanitized graph structure and amplified noise, GSD-BD overcomes this limitation by adaptively calibrating the contributions between malicious and benign knowledge through a unsupervised symmetric knowledge filter combined with sanitization loss, thereby ensuring effective backdoor sanitization and preserving benign model behavior.
Comparative evaluations conducted on four real-world graph datasets validate the effectiveness of the proposed method, demonstrating that GSD-BD significantly outperforms state-of-the-art methods and effectively addresses the challenge of performance loss associated with defense mechanisms. Nevertheless, the proposed method may encounter scalability challenges when applied to very large-scale graphs. Despite the fact that the upstream backdoor knowledge filter is able to reduce the computational overheads to a certain extent, in the scenario of large graphs the graph self-distillation, the computation of the Logit Margin Rate may still be computationally expensive in practice. Additionally, GSD-BD relies on retaining shallow symmetry knowledge in the deeper layers, which may limit its efficacy against more sophisticated attacks that exploit deeper layers or more complex interactions between layers.
Future work will focus on enhancing the real-time performance of the model while maintaining its robustness, particularly in applications involving real-time defense scenarios. We propose the following strategies: (1) integrating global graph knowledge distillation and (2) leveraging representative features such as explanatory subgraphs and graph motifs to reduce the computational burden. Additionally, considering that GSD-BD involves the use of original graph data, which could elevate the risk of privacy concerns, especially in sensitive contexts like social networks or medical data, the incorporation of graph federated learning could prove beneficial.
Author Contributions
Conceptualization, L.W.; methodology, H.W.; software, H.W.; validation, L.W. and X.Y.; investigation, X.Y.; data curation, H.W.; writing—original draft, H.W.; writing—review and editing, H.W.; visualization, H.W.; supervision, L.W.; project administration, L.W.; funding acquisition, L.W. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 62262004.
Data Availability Statement
AIDS, NCI1, and PROTEINS are available at https://paperswithcode.com/ (accessed on 12 June 2024). BITCOIN is available at https://github.com/zaixizhang/graphbackdoor (accessed on 12 June 2024).
Conflicts of Interest
The authors declare that this research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.
References
- Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
- Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful are Graph Neural Networks? In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
- Qiu, J.; Tang, J.; Ma, H.; Dong, Y.; Wang, K.; Tang, J. Deepinf: Social influence prediction with deep learning. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 2110–2119. [Google Scholar]
- Wang, J.; Zhang, S.; Xiao, Y.; Song, R. A Review on Graph Neural Network Methods in Financial Applications. J. Data Sci. 2022, 20, 111–134. [Google Scholar] [CrossRef]
- Kearnes, S.; McCloskey, K.; Berndl, M.; Pande, V.; Riley, P. Molecular graph convolutions: Moving beyond fingerprints. J. Comput.-Aided Mol. Des. 2016, 30, 595–608. [Google Scholar] [CrossRef] [PubMed]
- Hu, W.; Liu, B.; Gomes, J.; Zitnik, M.; Liang, P.; Pande, V.; Leskovec, J. Strategies For Pre-training Graph Neural Networks. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
- Hu, Z.; Dong, Y.; Wang, K.; Chang, K.W.; Sun, Y. Gpt-gnn: Generative pre-training of graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6–10 July 2020; pp. 1857–1867. [Google Scholar]
- Nassar, L.; Karray, F. Overview of the crowdsourcing process. Knowl. Inf. Syst. 2019, 60, 1–24. [Google Scholar] [CrossRef]
- Zhang, Z.; Jia, J.; Wang, B.; Gong, N.Z. Backdoor attacks to graph neural networks. In Proceedings of the 26th ACM Symposium on Access Control Models and Technologies, Virtual, 16–18 June 2021; pp. 15–26. [Google Scholar]
- Xu, J.; Xue, M.; Picek, S. Explainability-based backdoor attacks against graph neural networks. In Proceedings of the 3rd ACM Workshop on Wireless Security and Machine Learning, Miami, FL, USA, 15–17 May 2019; Association for Computing Machinery: New York, NY, USA, 2021; pp. 31–36. [Google Scholar]
- Xi, Z.; Pang, R.; Ji, S.; Wang, T. Graph backdoor. In Proceedings of the 30th USENIX Security Symposium (USENIX Security 21), Virtual, 11–13 August 2021; pp. 1523–1540. [Google Scholar]
- Zheng, H.; Xiong, H.; Chen, J.; Ma, H.; Huang, G. Motif-backdoor: Rethinking the backdoor attack on graph neural networks via motifs. IEEE Trans. Comput. Soc. Syst. 2023, 11, 2479–2493. [Google Scholar] [CrossRef]
- Dai, E.; Lin, M.; Zhang, X.; Wang, S. Unnoticeable backdoor attacks on graph neural networks. In Proceedings of the ACM Web Conference 2023, Austin, TX, USA, 30 April–4 May 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 2263–2273. [Google Scholar]
- Wang, B.; Yao, Y.; Shan, S.; Li, H.; Viswanath, B.; Zheng, H.; Zhao, B.Y. Neural cleanse: Identifying and mitigating backdoor attacks in neural networks. In Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 19–23 May 2019; pp. 707–723. [Google Scholar]
- Liu, Y.; Lee, W.C.; Tao, G.; Ma, S.; Aafer, Y.; Zhang, X. Abs: Scanning neural networks for back-doors by artificial brain stimulation. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, London, UK, 11–15 November 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 1265–1282. [Google Scholar]
- Li, Y.; Lyu, X.; Koren, N.; Lyu, L.; Li, B.; Ma, X. Neural Attention Distillation: Erasing Backdoor Triggers from Deep Neural Networks. In Proceedings of the International Conference on Learning Representations, Virtual, 3–7 May 2021. [Google Scholar]
- Xiang, Z.; Miller, D.J.; Kesidis, G. Detection of backdoors in trained classifiers without access to the training set. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 1177–1191. [Google Scholar] [CrossRef]
- Wang, H.; Xiang, Z.; Miller, D.J.; Kesidis, G. Mm-bd: Post-training detection of backdoor attacks with arbitrary backdoor pattern types using a maximum margin statistic. In Proceedings of the 2024 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 19–23 May 2024; pp. 1994–2012. [Google Scholar]
- Jiang, B.; Li, Z. Defending against backdoor attack on graph nerual network by explainability. arXiv 2022, arXiv:2209.02902. [Google Scholar]
- Sui, H.; Chen, B.; Zhang, J.; Zhu, C.; Wu, D.; Lu, Q.; Long, G. DMGNN: Detecting and Mitigating Backdoor Attacks in Graph Neural Networks. arXiv 2024, arXiv:2410.14105. [Google Scholar]
- Hinton, G. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
- Xu, X.; Zhou, F.; Zhang, K.; Liu, S. Ccgl: Contrastive cascade graph learning. IEEE Trans. Knowl. Data Eng. 2022, 35, 4539–4554. [Google Scholar] [CrossRef]
- Feng, K.; Li, C.; Yuan, Y.; Wang, G. Freekd: Free-direction knowledge distillation for graph neural networks. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 357–366. [Google Scholar]
- Yan, B.; Wang, C.; Guo, G.; Lou, Y. Tinygnn: Learning efficient graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6–10 July 2020; pp. 1848–1856. [Google Scholar]
- Chen, Y.; Bian, Y.; Xiao, X.; Rong, Y.; Xu, T.; Huang, J. On self-distilling graph neural network. arXiv 2020, arXiv:2011.02255. [Google Scholar]
- Li, G.; Xiong, C.; Thabet, A.; Ghanem, B. Deepergcn: All you need to train deeper gcns. arXiv 2020, arXiv:2006.07739. [Google Scholar]
- Chen, W.; Wu, B.; Wang, H. Effective backdoor defense by exploiting sensitivity of poisoned samples. Adv. Neural Inf. Process. Syst. 2022, 35, 9727–9737. [Google Scholar]
- Yang, Y.; Qiu, J.; Song, M.; Tao, D.; Wang, X. Distilling knowledge from graph convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7074–7083. [Google Scholar]
- Joshi, C.K.; Liu, F.; Xun, X.; Lin, J.; Foo, C.S. On representation knowledge distillation for graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 4656–4667. [Google Scholar] [CrossRef]
- Yang, C.; Liu, J.; Shi, C. Extract the knowledge of graph neural networks and go beyond it: An effective knowledge distillation framework. In Proceedings of the Web Conference 2021, Virtual, 19—23 April 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 1227–1237. [Google Scholar]
- Wu, L.; Lin, H.; Gao, Z.; Zhao, G.; Li, S.Z. A Teacher-Free Graph Knowledge Distillation Framework with Dual Self-Distillation. IEEE Trans. Knowl. Data Eng. 2024, 36, 4375–4385. [Google Scholar] [CrossRef]
- Wang, H.; Liu, T.; Sheng, Z.; Li, H. Explanatory subgraph attacks against Graph Neural Networks. Neural Netw. 2024, 172, 106097. [Google Scholar] [CrossRef]
- Zeng, Y.; Chen, S.; Park, W.; Mao, Z.; Jin, M.; Jia, R. Adversarial Unlearning of Backdoors via Implicit Hypergradient. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022. [Google Scholar]
- Liu, K.; Dolan-Gavitt, B.; Garg, S. Fine-pruning: Defending against backdooring attacks on deep neural networks. In International Symposium on Research in Attacks, Intrusions, and Defenses; Springer: Cham, Switzerland, 2018; pp. 273–294. [Google Scholar]
- Guo, W.; Wang, L.; Xing, X.; Du, M.; Song, D. Tabor: A highly accurate approach to inspecting and restoring trojan backdoors in ai systems. arXiv 2019, arXiv:1908.01763. [Google Scholar]
- Yang, X.; Li, G.; Tao, X.; Zhang, C.; Li, J. Black-Box Graph Backdoor Defense. In International Conference on Algorithms and Architectures for Parallel Processing; Springer: Cham, Switzerland, 2023; pp. 163–180. [Google Scholar]
- Rossi, R.; Ahmed, N. The network data repository with interactive graph analytics and visualization. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; Volume 29. [Google Scholar]
- Borgwardt, K.M.; Ong, C.S.; Schönauer, S.; Vishwanathan, S.; Smola, A.J.; Kriegel, H.P. Protein function prediction via graph kernels. Bioinformatics 2005, 21, i47–i56. [Google Scholar] [CrossRef]
- Weber, M.; Domeniconi, G.; Chen, J.; Weidele, D.K.I.; Bellei, C.; Robinson, T.; Leiserson, C. Anti-Money Laundering in Bitcoin: Experimenting with Graph Convolutional Networks for Financial Forensics. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
- Shervashidze, N.; Schweitzer, P.; Van Leeuwen, E.J.; Mehlhorn, K.; Borgwardt, K.M. Weisfeiler-lehman graph kernels. J. Mach. Learn. Res. 2011, 12, 2539–2561. [Google Scholar]
- Chen, J.; Zhang, D.; Ming, Z.; Huang, K.; Jiang, W.; Cui, C. GraphAttacker: A general multi-task graph attack framework. IEEE Trans. Netw. Sci. Eng. 2021, 9, 577–595. [Google Scholar] [CrossRef]
- Chen, J.; Lin, X.; Xiong, H.; Wu, Y.; Zheng, H.; Xuan, Q. Smoothing adversarial training for GNN. IEEE Trans. Comput. Soc. Syst. 2020, 8, 618–629. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).