Next Article in Journal
Automated Crack Detection in Micro-CT Scanning for Fiber-Reinforced Concrete Using Super-Resolution and Deep Learning
Previous Article in Journal
Diabetes Prediction and Detection System Through a Recurrent Neural Network in a Sensor Device
Previous Article in Special Issue
A Uniform Multi-Modal Feature Extraction and Adaptive Local–Global Feature Fusion Structure for RGB-X Marine Animal Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCPRES: Contrastive Deep Graph Clustering with Progressive Relaxation Weighting Strategy

1
Department of Software Engineering, College of Computer Science and Engineering, Guangxi Normal University, Guilin 541004, China
2
Guangxi Key Lab of Human–Machine Interaction and Intelligent Decision, Nanning Normal University, Nanning 530001, China
3
Guangxi Key Lab of Human–Machine Interaction and Intelligent Decision, Guangxi Academy of Science, Nanning 530007, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(21), 4206; https://doi.org/10.3390/electronics14214206
Submission received: 2 September 2025 / Revised: 12 October 2025 / Accepted: 22 October 2025 / Published: 28 October 2025
(This article belongs to the Special Issue Recent Advances in Efficient Image and Video Processing)

Abstract

Existing contrastive deep graph clustering methods typically employ fixed-threshold strategies when constructing positive and negative sample pairs, and fail to integrate both graph structure information and clustering structure information effectively. However, this fixed-threshold and binary partitioning approach is overly rigid, limiting the model’s utilization of potentially learnable samples. To address this problem, this paper proposes a contrastive deep graph clustering model with a progressive relaxation weighting strategy (DCPRES). By introducing the progressive relaxation weighting strategy (PRES), DCPRES dynamically allocates sample weights, constructing a progressive training strategy from easy to difficult samples. This effectively mitigates the impact of pseudo-label noise and enhances the quality of positive and negative sample pair construction. Building upon this, DCPRES designs two contrastive learning losses: an instance-level loss and a cluster-level loss. These respectively focus on local node information and global cluster distribution characteristics, promoting more robust representation learning and clustering performance. Extensive experiments demonstrated that DCPRES significantly outperforms existing methods on multiple public graph datasets, exhibiting a superior robustness and stability. For instance, on the CORA dataset, our model achieved a significant improvement over the static approach of CCGC, with the NMI increasing by 4.73 % , the ACC by 4.77 % , the ARI value by 7.03 % , and the F1-score by 5.89 % . It provides an efficient and stable solution for unsupervised graph clustering tasks.

1. Introduction

Graph data possess the advantage that their topological structure inherently provides natural self-supervision signals [1,2]. Consequently, this type of data holds significant application value in unsupervised learning and has greatly advanced the development of deep graph clustering (DGC) techniques. Deep graph clustering techniques aim to learn node representations and uncover latent cluster structures in an unsupervised setting. Classical methods such as ARGA [3] and DSCPS [4] typically guide clustering training based on reconstruction loss or generative modeling. However, such methods often struggle to capture the complex semantic relationships between nodes. Robustness in complex networks has also been emphasized in other domains, such as traffic network security [5], which further highlights the necessity of designing stable and resilient strategies in deep graph clustering.
Benefiting from the remarkable ability of contrastive learning [6] (CL) to learn supervision signals from unlabeled samples, contrastive deep graph clustering [1] (CDGC) has achieved an impressive performance. CDGC leverages augmented views of the graph structure and features to generate positive and negative sample pairs. By maximizing the similarity between positive pairs and minimizing the similarity with negative pairs, it guides the model to learn discriminative representations of augmented nodes, significantly enhancing the graph clustering performance. In CDGC, the selection of high-confidence positive and negative samples is crucial for improving the clustering performance. Early CDGC methods, such as MVGRL [7], primarily relied on manually designed data augmentation techniques (e.g., node feature perturbation, edge structure perturbation) to generate different views. While this approach can promote diversity in node representations, the inherent randomness of the augmentation often leads to semantic drift issues, consequently degrading the quality of the constructed positive and negative sample pairs for clustering. Yang et al. proposed CCGC [8], which utilizes high-confidence pseudo-labels to select reliable nodes for guiding the construction of positive and negative pairs. It employs dual encoders to generate different views, significantly mitigating the problems of semantic drift and pseudo-label noise. Liu et al. proposed HSAN [9], which, for the first time, introduced hard sample awareness into the CDGC framework. HSAN identifies hard positive and hard negative samples through the dual modeling of the feature and structural similarity. By dynamically weighting these hard samples during training, it enhances the model’s discriminative power on challenging samples and improves the clustering performance.
Although contrastive deep graph clustering has achieved a promising performance, we point out two drawbacks in the existing methods, as follows: (1) Existing methods overlook the role of graph structure information and clustering structure information when learning to construct negative pairs. Although CCGC [8] considers clustering structure information when selecting negative pairs and HSAN [9] considers graph structure information, no existing work has simultaneously integrated both attribute and structural information. (2) When selecting positive and negative sample pairs based on the K-means clustering results, the existing methods typically employ a fixed-threshold strategy [8,9], focusing more on samples distant from cluster centers, termed “hard samples”. However, this fixed-threshold and binary partitioning approach is overly rigid, limiting the model’s utilization of potentially learnable samples. An excessively large threshold introduces excessive noise within clusters. Conversely, an overly small threshold risks losing high-confidence samples.
To solve the mentioned problems, this paper proposes a contrastive deep graph clustering model with a progressive relaxation weighting strategy (DCPRES). Concretely, when constructing different views, we simultaneously embedded both the sample attributes and the data structure of the graph data. The embeddings from these two sources were then fused, resulting in node representations for different views that possess richer semantic information. Simultaneously, when selecting positive samples, we adopted the progressive relaxation weighting strategy (PRES). This involved introducing two small threshold parameters to divide the area around the cluster center into three confidence regions: high, medium, and low. This avoids the shortcomings of using a fixed threshold to distinguish positive and negative samples. The samples were then assigned to a confidence region based on their distance to the cluster center and weighted according to their confidence level. This ensures that the model first learns features from high-quality samples. Inspired by self-paced learning, we progressively increased the threshold parameters. This gradually introduced hard samples, guiding the model to learn their features from easy to difficult instances. This approach addresses the instability caused by semantic drift and pseudo-label noise, enhancing the model’s generalization and robustness. The main contributions of this paper are as follows:
  • We introduce a novel contrastive deep graph clustering model featuring a PRES. To improve feature representation, our model concurrently processes both sample attributes and graph structures and establishes a two-tiered contrastive learning framework at both the instance and cluster levels.
  • The progressive relaxation weight strategy implements an “easy-to-hard” learning curriculum. It guides the model to initially train on high-confidence samples to build a strong feature foundation, before gradually incorporating more challenging, lower-confidence samples. This approach enhances the model’s capacity for distinguishing between positive and negative pairs.
  • Extensive experiments on six benchmark datasets validated the superiority of our proposed model, which set new state-of-the-art results across several key quantitative metrics against competing methods.
The remainder of this paper is organized as follows. We first review related work in contrastive deep graph clustering and self-paced learning. Next, we detail the proposed DCPRES framework, including its feature encoding process, the PRES, and the dual contrastive loss functions. Subsequently, we present extensive experimental results on six benchmark datasets to demonstrate the superiority of our method, followed by ablation studies and a visual analysis to validate the effectiveness of each component. Finally, we conclude the paper with a summary of our findings.

2. Related Work

2.1. Contrastive Deep Graph Clustering

Contrastive deep clustering is a method for unsupervised representation learning whose fundamental principle is to learn highly discriminative features by attracting positive pairs and repelling negative pairs within an embedding space. This method has seen broad applications in representation learning in various domains, including computer vision [10,11,12], natural language processing [13], and graph-structured data [14,15]. It is especially effective at uncovering latent semantic structures in data where labels are absent. Central to this is the info-noise contrastive estimation (InfoNCE) loss [16], a classic paradigm in contrastive learning. It drives the learning of discriminative features by maximizing the similarity between positive pairs while minimizing their similarity to a set of negative samples. For a given anchor sample x, its corresponding positive sample x + (typically generated by data enhancement), and a set of negative samples { x i } , the InfoNCE loss is formulated as follows:
L InfoNCE = log exp ( sim ( h x , h x + ) / τ ) exp ( sim ( h x , h x + ) / τ ) + i exp ( sim ( h x , h x i ) / τ )
Here, h x , h x + , and h x i are the respective embedding vectors for the anchor x, the positive sample x + , and the negative sample { x i } . The term sim ( · , · ) represents a similarity metric (such as cosine similarity) and τ is a temperature parameter. The InfoNCE loss function considers a sample and its augmentation as a positive pair, while all other samples in the batch are treated as negative pairs. In complex graph structures, this design is prone to inaccurate positive–negative sample partitioning, thereby degrading the model performance.
To overcome this limitation, researchers have proposed several improved strategies to construct positive and negative pairs more effectively and increase the performance of contrastive learning. Prominent examples include DCHD [17], CCGC, and HSAN. These approaches typically select positive and negative pairs based on the K-means clustering results. They often focus on “hard samples”, those distant from cluster centers, and leverage confidence scores or neighborhood relationships to improve the precision of the selection process. However, these methods neglect the influence of the sample-to-center distance (i.e., the threshold) on the model. An overly large threshold introduces significant noise into the clusters, whereas a threshold that is too small results in the loss of high-confidence samples. Moreover, these approaches fail to fully leverage samples in the vicinity of cluster centers, which are high-confidence samples that could enable the model to learn essential features with greater ease.
To address this issue, we introduced a contrastive deep graph clustering model that incorporates a progressive relaxation weighting mechanism. Our approach utilizes two threshold parameters, which are relaxed as training progresses, to divide the space surrounding the cluster centers into three distinct zones: high-, medium-, and low-confidence. By assigning different weights to the samples based on their respective zones, we steer the model to prioritize learning from high-confidence samples in close proximity to the cluster centers. As training continues, these selection criteria are gradually loosened, establishing an “easy-to-hard” curriculum. This strategy significantly enhances both the robustness of the model and its overall clustering efficacy.

2.2. Semi-Supervised Clustering Based on Self-Paced Learning

Self-paced learning (SPL), initially introduced by Kumar et al. [18], was inspired by the cognitive principle that humans learn progressively, starting with simple concepts and gradually advancing to more complex ones. This approach introduces a self-pace parameter to dynamically manage the difficulty of samples during training, allowing the model to first master easier examples before tackling more challenging ones. This, in turn, enhances the robustness and generalization capacity of the model. In contrast to conventional methods that utilize the entire dataset from the outset, SPL prioritizes a sequential and phased approach to sample selection, which effectively mitigates the disruptive influence of noisy data. The efficacy and potential of SPL have been demonstrated from multiple angles in various applications, including semi-supervised image classification [19] (SPLSSE), deep subspace clustering [20] (DSC-AS), and multi-modal semi-supervised classification [21] (SSMFS). In the SPL framework, the learning pace is regulated by introducing a weight variable, v i [ 0 , 1 ] , for each sample, indicating its degree of inclusion in the training process. This joint learning over model parameters and sample weights is formulated as the following optimization problem [18]:
min w , v [ 0 , 1 ] i = 1 n v i L ( f w ( x i ) , y i ) + g ( v i , λ )
In Equation (2), L ( · ) denotes the training loss of the supervised model, while f w ( · ) represents the model parameterized by w. The term g ( v i , λ ) is a self-paced regularizer that governs the allocation of sample weights ( v i ) and steers the learning curriculum. Here, λ is the self-paced parameter that controls the quantity and difficulty of the samples selected at each stage. Initially, a small λ ensures that only a few high-confidence samples are used for training. As training advances, λ is gradually increased, exposing the model to more complex and challenging samples. In conventional self-paced learning, the value of λ imposes a hard-thresholding decision (i.e., weights of 0 or 1) on whether a sample is included in the training. In the context of unsupervised clustering, however, we argue that a soft-weighting scheme is more beneficial. Relaxing the confidence criteria allows the model to learn richer semantic representations and enhances its robustness. Therefore, inspired by self-paced learning, we conceptualized the sample-to-center distance as a self-paced parameter, λ 1 . We then relaxed this condition by introducing a second parameter, λ 2 (with λ 1 < λ 2 ), to create a soft-weighting interval. Samples with distances falling between λ 1 and λ 2 were assigned a weight between 0 and 1. Both λ 1 and λ 2 were progressively increased during training, empowering the model to more effectively select positive and negative samples.

3. Method

3.1. Overview of the Framework

We propose DCPRES, whose architectural framework is depicted in Figure 1. Initially, the model employs a feature extraction module to generate embeddings that capture both the attribute and structural characteristics of the nodes. Subsequently, our PRES module dynamically allocates weights to samples, facilitating a curriculum where the model learns progressively from high-confidence to low-confidence examples. The architecture incorporates two key loss functions: an instance-level contrastive loss (LOSS-PREi) that captures local structural details via sample-to-sample comparisons, and a cluster-level contrastive loss (LOSS-PREc) guided by our weighting strategy that refines the global cluster distribution through cluster-to-cluster comparisons. Both losses, synergistic under the progressive relaxation mechanism, jointly elevate the overall clustering performance. Consequently, the model strikes an effective balance between local similarity and global consistency, and its robustness to noisy data is significantly improved.

3.2. Attribute and Structural Feature Encoding

In this section, we model the node data as an undirected graph. Let G = ( V , E ) define an undirected graph comprising N nodes and K classes. The set of nodes is denoted by V = { v 1 , v 2 , , v N } , and the set of undirected edges M by E = { e 1 , e 2 , , e M } , which capture the connectivity between nodes. Each node is characterized by both attribute and structural information. For ease of feature extraction, we adopted a matrix representation for the graph data. The node attributes are represented by a matrix X R N × D , where each row x i corresponds to the vector of characteristics D dimensional of node v i . The structure of the graph is encoded in an adjacency matrix A R N × N , where a i j = 1 signifies an edge between nodes v i and v j , and a i j = 0 indicates no direct connection.
In line with previous work [8,9], we preprocessed the adjacency matrix with a normalization step prior to feature extraction to enhance the representational capacity of both the graph structure and node features. The procedure is as follows: first, we introduced self-loops to the adjacency matrix A by adding the identity matrix I, resulting in A ˜ = A + I . This step is crucial for ensuring that a node’s own features are preserved during the information aggregation phase. Next, we constructed a symmetrically normalized graph Laplacian matrix, L ˜ , to stabilize the information propagation and mitigate issues arising from nonuniform node degree distributions. It was formulated as follows:
L ˜ = I D ˜ 1 / 2 A ˜ D ˜ 1 / 2
Here, D ˜ = diag { d ˜ 1 , d ˜ 2 , , d ˜ N } R N × N is the degree matrix corresponding to A ˜ , with each diagonal element d ˜ i = j a ˜ i j denoting the degree of the node. This preprocessing step lays a robust foundation for the subsequent extraction of features based on the graph structure. The final smoothed node attributes are given by Equation (4), where t signifies the number of smoothing layers.
X ˜ = i = 1 t ( I L ˜ ) X = ( I L ˜ ) t X
Given the preprocessed feature matrix X ˜ and the original adjacency matrix A, we separately fed them into an attribute feature encoder (AE) and a structure feature GCN (SG). The attribute encoder A E a ( · ) maps node attributes into latent embeddings, while the structure encoder S G a ( · ) is implemented as a graph convolutional network (GCN) to capture structural dependencies. This process yields distinct representations for the node attributes and the graph structure, respectively:
Z v a = AE a ( X ˜ ) , Z v a = Z v a Z v a 2 , a { 1 , 2 } G v a = SG a ( A ) , G v a = G v a G v a 2 , a { 1 , 2 }
The GCN-based structure encoder updates node features by aggregating information from neighbors according to the normalized adjacency matrix. Its propagation rule follows the standard formulation:
H ( l + 1 ) = σ D 1 / 2 A ˜ D 1 / 2 H ( l ) W ( l )
where σ ( · ) is a non-linear activation function. After this encoding process, the attribute encoders A E 1 ( · ) and A E 2 ( · ) yield two attribute-view embeddings { Z v 1 , Z v 2 } , and the structure encoders produce two structure-view embeddings { G v 1 , G v 2 } .

3.3. Progressive Relaxation Weight Strategy

In the task of contrastive deep graph clustering, selecting high-confidence samples is paramount for improving the model stability and discriminative power. The conventional approach involves clustering the representations ( Z v 1 , Z v 2 ) from a dual-view encoder and designating the top τ samples—those closest to their cluster centroids or with the highest assignment probability—as high-confidence pairs. Traditional methods often rely on a static thresholding policy: samples satisfying a specific distance criterion are marked as positive (weight of 1), while all others are disregarded (weight of 0). This rigid, binary partitioning, however, is overly restrictive and prevents the model from leveraging other potentially valuable samples.
To address this limitation, we introduced a progressive relaxation weighting mechanism for the dynamic selection of high-confidence samples to augment the supervisory signal. Within this relaxation weighting module, a sample’s weight is determined by its feature-space distance and the clustering outcome. Specifically, we began by fusing the outputs from the two attribute encoders, Z v 1 and Z v 2 , to form a unified attribute representation, Z.
Z = 1 2 ( Z v 1 + Z v 2 )
We applied K-means clustering to the complete set of fused representations { z i } i = 1 N . This process yielded a set of cluster centroids C and a distance matrix D . To quantify the clustering confidence for each sample, we then computed the Euclidean distance between the sample and its assigned cluster centroid:
d i = z i c p 2 p = 0 K i = 0 N z i c p 2 , i ( 0 , N ) , p ( 0 , K )
Here, c p denotes the centroid of the cluster to which sample z i is assigned after K-means clustering. To improve the comparability of the distance metric and enhance the training stability, we normalized all sample distances d i , scaling them to the [ 0 , 1 ] range. A smaller value of d i signifies that the sample is closer to its cluster centroid, and thus possesses a higher clustering confidence.
For sample weight computation, we departed from traditional binary selection policies by proposing a three-stage sample-weighting mechanism based on a progressive relaxation strategy. This mechanism permits the model to initially focus exclusively on the most reliable samples, ensuring training stability. As training deepens, more challenging samples are progressively incorporated to improve the model’s discriminative power and generalization capabilities. Concurrently, by filtering out low-confidence samples, the disruptive influence of noise on the training process is significantly mitigated. Specifically, we introduced two threshold parameters, λ 1 and λ 2 (where λ 1 < λ 2 ), to define confidence regions for the samples. The parameter λ 1 functions like a conventional fixed threshold, identifying definitive positive samples. The second parameter, λ 2 , relaxes this stringent condition by allowing samples that fall within the ( λ 1 , λ 2 ) interval to be assigned a soft confidence weight. This approach expands the pool of trainable data to include potentially valuable samples, thereby enhancing the model’s feature-learning capacity. According to the functions of λ 1 and λ 2 , the dynamic formula for calculating sample weights was established as follows:
w i = 1 , d i λ 1 λ 2 d i λ 2 λ 1 , λ 1 < d i λ 2 0 , d i > λ 2
Accordingly, we classified samples into three distinct types: high-confidence ( d i λ 1 ), medium-confidence ( λ 1 < d i λ 2 ), and low-confidence ( d i > λ 2 ). At the outset of training, λ 1 and λ 2 were initialized to very small values. This guaranteed that the initial training set consisted of high-confidence samples characterized by minimal clustering distances and the most dependable feature representations. To realize an “easy-to-hard” training curriculum, we progressively increased the values of λ 1 and λ 2 as the training unfolded, thereby relaxing the conditions for sample selection. Specifically, during the t-th epoch of training, the self-paced learning parameters λ 1 and λ 2 were updated based on the following rule:
λ 1 ( t ) = λ 1 ( 0 ) + Δ λ 1 · t , λ 2 ( t ) = λ 2 ( 0 ) + Δ λ 2 · t
In Equation (10), Δ λ 1 and Δ λ 2 govern the growth increments for thresholds λ 1 and λ 2 per training epoch. The magnitudes of these increments directly dictate the pace at which the model incorporates more challenging samples into the training process. To accommodate the varying requirements of different training phases, diverse growth strategies can be employed for Δ λ 1 and Δ λ 2 , enabling a more adaptable sample-filtering procedure.
In practice, the threshold parameters λ 1 and λ 2 are initialized with small values (e.g., λ 1 0.3 , λ 2 0.6 ). This design ensures that, at the beginning of training, the model focuses only on high-confidence samples located close to the cluster centers, thereby reducing the negative influence of noisy pseudo-labels. This choice was inspired by the principle of self-paced learning, where the model first learns from “easy” and reliable samples before gradually incorporating more challenging ones.
To ensure interpretability and stability, we adopted a linear growth strategy for Δ λ 1 and Δ λ 2 , where both thresholds were steadily increased at each training epoch. For example, the growth rate of λ 1 per epoch, denoted as Δ λ 1 , is defined as ( 1.0 λ 1 ( 0 ) ) / T , where T represents the total number of training epochs. Accordingly, the value of λ 1 at the t-th epoch can be computed as λ 1 ( t ) = λ 1 ( 0 ) + Δ λ 1 · t . This formulation ensures that λ 1 linearly increases from its initial value λ 1 ( 0 ) to a maximum of 1.0 by the end of training. A similar strategy was applied to λ 2 .
In addition, the effectiveness of this design was validated through a hyperparameter sensitivity analysis (see Section 4.6) across multiple datasets. The experiments demonstrated that, when λ 1 and λ 2 are initialized within a relatively small range, the model achieves a consistently robust clustering performance.
By leveraging the aforementioned PRES for each sample { z i } in every iteration, a dedicated training weight w i is dynamically computed based on its distance metric d i relative to the thresholds λ 1 and λ 2 . The collection of these weights forms the sample weight matrix, W = { w 1 , w 2 , , w N } , which acts as a dynamic measure of sample importance. This integrated sample selection and weighting mechanism enhances the model’s robustness and generalization capabilities throughout the training process.

3.4. Progressive Relaxation Weighted Instance-Level Contrastive Loss (LOSS-PREi)

Following the K-means clustering of the fused features Z, we derived an assignment probability matrix, P R N × K . This matrix details the probability of each sample’s assignment to every cluster center. For each individual sample, we then selected the cluster center corresponding to the maximum assignment probability to serve as its pseudo-label:
q i = arg max k ( P i k )
This process generated a set of pseudo-labels, Q = { q 1 , q 2 , , q N } , which we used to construct positive and negative pairs. For any given sample Z i v 1 from the first attribute view, its positive counterpart, Z j v 2 , is a sample from the second attribute view that shares the identical pseudo-label ( q i = q j ). All samples with differing pseudo-labels are considered negative samples. An analogous protocol was employed for the structural view representations, S v 1 and S v 2 . This construction protocol ensures that nodes with coherent features across views are treated as positive pairs, thereby effectively strengthening the alignment of cross-view representations.
Building upon this, we incorporated the obtained weight W , which was automatically modulated based on its confidence score within the semantic feature space. In the initial stages of training, the model focuses exclusively on high-confidence samples. As training proceeds, more challenging medium-confidence samples are progressively incorporated, while low-confidence noisy samples are either assigned a lower weight or discarded entirely. Inspired by the InfoNCE loss [16] (Equation (1)), we designed a weighted instance contrastive loss function to encourage aggregation among positive pairs and suppress similarity among negative samples, which is defined as follows:
L ( z i v a ) = log j Pos ( z i ) w i w j · exp β sim ( Z i v a , Z j v b ) / τ I + ( 1 β ) sim ( G i v a , G j v b ) / τ I h Neg ( z i ) w i w h · exp β sim ( Z i v a , Z h v b ) / τ I + ( 1 β ) sim ( G i v a , G h v b ) / τ I
Here, Pos ( z i ) and Neg ( z i ) are the respective sets of positive and negative samples corresponding to a given sample Z i v a . The function sim ( · , · ) denotes the normalized cosine similarity. τ I is the instance-level temperature parameter, while β ( 0 , 1 ) is a learnable parameter that weighs the contributions of semantic and structural view similarities. The indices a and b denote the first and second views, respectively. This loss function, calculated for sample z i in view a, integrates information from both attribute and structural cross-view perspectives. Compared to the original InfoNCE loss, which typically relies on a single pair of positive and negative samples, our instance-level loss leverages clustering results to generate a richer set of positive and negative pairs. This allows the model to exploit more informative supervisory signals, thereby optimizing the representation learning process. Furthermore, by incorporating the PRES, each sample is assigned a confidence-based weight that fine-tunes its contribution to the final loss. This dynamic weighting mechanism reduces the influence of noisy samples and enhances the robustness of the learned representations, ultimately leading to a more stable and effective clustering performance. By incorporating confidence-based sample weighting, it steers the model to learn increasingly discriminative representations through an “easy-to-hard” curriculum. The total instance-level contrastive loss, guided by our self-paced strategy, is formulated as follows:
L PREi = 1 2 N i = 0 N L ( z i v 1 ) + L ( z i v 2 )

3.5. Progressive Relaxation Weighted Cluster-Level Contrastive Loss (LOSS-PREc)

Beyond the instance-level contrastive loss operating on individual samples, we introduced a cluster-level contrastive loss to capture higher-order semantic consistency. This loss is designed to maximize the similarity between positive cluster views while minimizing it between negative pairs. Concretely, for each of the two attribute views Z v 1 and Z v 2 , we aggregated node features into class-center representations according to their pseudo-labels. To enhance the robustness, we incorporated the sample confidence weights, W , into the formulation of the cluster centroids. The centroid for the k-th class in view v a ( a { 1 , 2 } ) , denoted c k v a , is defined as follows:
c k v a = 1 i = 1 N I ( q i = k ) · w i i = 1 N I ( q i = k ) · w i · z i v a
In this context, I ( q i = k ) serves as an indicator function, which yields a value of 1 if sample z i v a is a member of the k-th class, and 0 otherwise. To enforce consistency between centroids of the same class across different views, while simultaneously enlarging the separation between centroids of different classes, we designed the following self-paced clustering contrastive loss:
L PREc = 1 2 K k = 1 K a = 1 2 log exp ( sim ( c k v a , c k v b ) / τ c ) l = 1 K exp ( sim ( c k v a , c l v b ) / τ c ) , l k
In this context, sim ( · , · ) denotes the normalized cosine similarity and τ c is the clustering temperature parameter that modulates the sensitivity of the contrastive loss distribution. The indices a and b correspond to the first and second views, respectively. This loss function designates the centroids of the same class k from two different views, c k v a and c k v b , as a positive pair. Conversely, all other class centroids, { c l v b } l k , are treated as negative samples. This cluster-level contrastive objective optimizes the clustering consistency between the two views. Therefore, the final loss for DCPRES is a combination of the instance-level and cluster-level losses, formulated as shown below. Additionally, the complete training procedure for the model is detailed in Algorithm 1.
L = L P R E i + L P R E c
Algorithm 1 CPRES algorithm flow.
Input: 
Attribute matrix X R N × D ; Adjacency matrix A R N × N ; Epoch number I; Filtering times t; Self-paced parameters λ 1 , λ 2 ; Coefficient τ I , τ c , α , β .
Output: 
The clustering result Φ .
  1:
Obtain the filtering attribute feature X ˜ with Equation (4).
  2:
while  i < I m a x and | L i L i 1 | < ε  do
  3:
      Calculate attributes and structural embeddings Z v 1 , Z v 2 , G v 1 , G v 2 with Equation (5).
  4:
      K-means clustering using fused features Z with Equation (7) yields the result Φ .
  5:
      Obtain the sample weight W with Equation (9).
  6:
      Calculate the instance-level loss L P R E i with Equation (13).
  7:
      Calculate the clustering centers C and clustering loss L P R E c with Equations (14) and (15).
  8:
      Calculate the total loss L with Equation (16).
  9:
      Update the relaxation strategy weights λ 1 and λ 2 with Equation (10).
10:
end while
11:
return  Φ

4. Experiment

This section begins by introducing the six publicly available datasets utilized in our experiments, accompanied by a brief description of each dataset’s key characteristics. We then provide a detailed exposition of our proposed method’s implementation, covering the experimental setup, the comparative baseline models, and the evaluation metrics employed. To conduct a thorough performance assessment, we performed extensive clustering experiments on these six datasets, benchmarking our approach against several state-of-the-art methods.

4.1. Dataset

Our experimental section evaluates the performance on six common graph-structured datasets, which included standard citation networks and text-attributed graphs. The first category included CORA and CITE, two classic citation network datasets where nodes signify papers and edges denote citation links. Each node is characterized by textual features and a class label, making these datasets standard benchmarks for node classification. The second category comprised AMAP, BAT, EAT, and UAT, which are datasets built as text-attributed graphs for broad use in unsupervised or semi-supervised graph representation learning. In these datasets, nodes typically correspond to documents or text fragments, with edges established based on semantic similarity. A statistical summary of these six datasets is provided in Table 1.

4.2. Experimental Environment and Hyperparameter Settings

In our experiments, we trained the model end-to-end for 400 iterations. We employed the Adam optimizer with momentum for gradient descent-based minimization of the loss function, configured with a learning rate of 1 × 10 4 and a weight decay of 1 × 10 5 . All the experiments were executed on a system running Ubuntu 22.04.2 LTS, powered by an Intel Core i9-10980XE CPU @ 3.00 GHz and an NVIDIA GeForce RTX 4080 GPU with 16 GB of VRAM for training acceleration. The implementation was built upon the PyTorch framework (v2.4.1), utilizing CUDA 12.4 and Python 3.10. To ensure a fair comparison, all the models were trained from scratch without any pre-trained weights and under identical hyperparameter settings. Furthermore, for the sake of reproducibility, we used a fixed random seed across all experiments to maintain the reliability and comparability of our results.

4.3. Comparative Experiments

To benchmark the performance of our proposed DCPRES model, we conducted a comprehensive experimental comparison against a range of state-of-the-art deep graph clustering methods, all evaluated under identical conditions and without utilizing any pre-trained weights. Our baseline models included deep graph clustering methods such as DAEGC [22] (2019), DyFSS [23] (2024), and DeSE [24] (2025), as well as cutting-edge contrastive deep graph clustering models like GDCL [25] (2021), DCHD [17] (2024), MAGI [26] (2024), CCGC [8] (2023), and HSAN [9] (2023). To thoroughly assess the clustering performance across various aspects, we used four widely recognized evaluation metrics: the accuracy (ACC), the normalized mutual information (NMI), the adjusted Rand index (ARI), and the F1-score [4,8,27,28]. The full results of this comparative study are presented in Table 2, with the top-performing metric highlighted in bold and the runner-up underlined.
As indicated by the results in Table 2, traditional deep graph clustering models, such as DAEGC and DyFSS, show effectiveness on select datasets, but they are fundamentally limited in their ability to capture latent semantic and structural inter-node relationships. These methods generally lack explicit supervisory signals and often rely heavily on specific pre-training routines, which restricts their capacity to fully leverage the underlying graph structures. In contrast, our proposed DCPRES model integrates a contrastive learning mechanism, significantly enhancing the discriminative power of node representations. By enabling the discovery of more informative positive and negative sample pairs, DCPRES sharpens the model’s ability to detect cluster boundaries, leading to an improved clustering performance.
One of the key features of DCPRES is the progressive relaxation strategy for sample selection, which we link to SPL. Self-paced learning provides an “easy-to-hard” curriculum for the model, where it starts by focusing on high-confidence samples and gradually introduces harder, more uncertain samples as training progresses. This progressive approach is particularly effective for mitigating the impact of pseudo-label noise and stabilizing gradient updates, which is crucial for improving the clustering performance in challenging, noisy environments. Specifically, by initially training on easy-to-classify samples, DCPRES ensures a stable optimization trajectory, while progressively introducing more complex samples allows the model to learn more discriminative features as training advances. The dynamic weighting of samples based on their confidence level further refines the contrastive learning objective, making the model more robust to noise and enhancing the gradient stability.
Furthermore, we benchmarked DCPRES against leading contrastive deep graph clustering models, including GDCL, DCHD, CCGC, MAGI, and HSAN. While these models leverage contrastive learning to address the lack of supervision, they predominantly rely on static high-confidence sample selection, which can be suboptimal for both sample utilization and the representation robustness. DCPRES, in contrast, employs a dynamic, self-paced learning-based sample-scheduling strategy. This allows the model to begin by focusing on high-confidence samples to stabilize early-stage training, and then progressively include medium- and low-confidence samples, thereby adapting to harder examples. This “easy-to-hard” learning paradigm, facilitated by dynamically weighted contributions from each sample, improves both the robustness and flexibility of the contrastive loss function. The progressive relaxation strategy enables more effective use of the dataset by dynamically managing sample inclusion, mitigating early noise effects, and ultimately leading to better generalization.
For example, on the CORA dataset, DCPRES outperformed CCGC’s static approach by 4.73% for the NMI, 4.77% for the ACC, 7.03% for the ARI value, and 5.89% for the F1-score, demonstrating a superior performance and robustness in clustering. Additionally, DCPRES exhibited considerably less performance fluctuation across all metrics compared to the other models, which is reflected in the lower standard deviation. This indicates that the self-paced learning mechanism is effective at reducing the uncertainty associated with training and improving the stability of the final clustering outcomes.

4.4. Ablation Experiments

As shown in Table 3, we conducted six ablation experiments on the CORA, CITE, AMAP, and BAT datasets to evaluate the contributions of GCN-Encoder, LOSS-PREi, and LOSS-PREc. By comparing the first and third groups, as well as the second and fourth groups, we observed that removing GCN-Encoder consistently led to a decline in the performance across multiple metrics. This indicates that the structural information captured by GCN-Encoder provides essential global context that complements the final loss optimization, making it a critical component for performance improvement.
Furthermore, when either LOSS-PREi or LOSS-PREc was excluded from DCPRES, we observed a clear degradation in the results across all datasets. This demonstrates the importance of both the instance-level and cluster-level losses: they impose complementary constraints that jointly guide the representation space toward more discriminative and consistent clustering. Overall, the complete model, which integrates GCN-Encoder, LOSS-PREi, and LOSS-PREc, achieved the best results on nearly all metrics, confirming that the combination of structural encoding and dual-level loss design is both effective and necessary.
In addition, we note that the marginal contributions of each component varied across datasets. For example, removing GCN-Encoder had little effect on CORA’s NMI (+0.04), but significantly reduced the BAT NMI (−2.52). This difference reflects the varying balance of information sources: CORA node features are relatively strong, so structural encoding adds limited gain, while BAT relies more heavily on graph connectivity, making the GCN-Encoder indispensable. Similarly, the benefits of LOSS-PREi and LOSS-PREc depend on the dataset characteristics—LOSS-PREi is more effective when local neighborhoods are noisy, while LOSS-PREc contributes more when clusters are imbalanced or diverse. These results highlight that, although all components are crucial, their relative importance is dataset-dependent.
In Table 4, we compare three threshold-selection strategies—fixed, single, and DCPRES—on CORA and AMAP. Fixed applies a hard binary cut-off at 0.7; single sweeps a single threshold from 0.5 to 1.0 to derive soft weights; and DCPRES adopts a double-threshold progressive relaxation that partitions samples into high-/medium-/low-confidence bands and gradually relaxes the inclusion criteria during training. The results showed that DCPRES consistently delivered the best scores (NMI/ACC/ARI/F1), indicating that progressively widening the acceptance region with dual thresholds achieves a better trade-off between sample purity and coverage, suppresses early-stage pseudo-label noise, and incorporates medium-confidence samples at the right time, leading to more stable optimization and a stronger clustering performance.
As shown in Figure 2, this analysis shows the additional time required for calculating sample weights after incorporating the progressive relaxation mechanism. The experiment was conducted with 400 training epochs, and the final time reflects the average of 10 runs. The time to calculate the weights increased by 0.05 min, 0.06 min, and 0.22 min for the CORA, CITE, and AMAP datasets, respectively. On average, the time increased by 6.42%. The increase was relatively small on smaller datasets like CORA and CITE, but slightly higher for larger datasets like AMAP. However, this time increase is acceptable, considering the performance improvement. The efficiency gain is attributed to the use of matrix operations for the weight calculation, which is much faster than regular mathematical computations. Additionally, the memory requirements are minimal, needing only an R N × N weight matrix to store the results.

4.5. Visual Analytics

To provide a more intuitive illustration of the superiority of our proposed DCPRES model in capturing cluster structures, this section presents a visual analysis of the final node embeddings on the CORA dataset, as depicted in Figure 3. We employed the t-SNE [29] algorithm to project the high-dimensional representations into a 2D space, which allowed for a clear observation of the spatial distribution of nodes from different classes. The visualization revealed that, in comparison to other deep clustering methods, our DCPRES model carves out more distinct and semantically coherent clusters, indicating its enhanced discriminative power and ability to maintain semantic consistency.
Next, we present a visualization of the clustering results from DCPRES on the CORA dataset in Figure 4. For this, we again employed the t-SNE method to project the node embeddings into a 2D space at four different training milestones: the 100th, 200th, 300th, and 400th epochs. In these plots, low-confidence samples that have not yet been incorporated into the training are colored gray. The remaining nodes, which are actively part of the training, are colored according to their assigned cluster (seven distinct colors for seven classes). The evolution of the distributions across these stages clearly shows that DCPRES initially leverages high-confidence samples for optimization. It then progressively incorporates samples that were previously difficult to classify, ultimately encompassing the entire dataset. This “easy-to-hard” guided learning process effectively showcases the dynamic and adaptive nature of our proposed sample selection strategy in DCPRES.

4.6. Hyperparameter Analysis

To further investigate the influence of the two critical threshold parameters, λ 1 and λ 2 , within our DCPRES, we conducted a hyperparameter sensitivity analysis. This study examined the performance trends under varying sample selection intensities. These parameters govern the confidence assignment for each sample, thereby dictating whether a pseudo-labeled sample is fully trusted, partially utilized, or entirely discarded. To systematically quantify the effects of these thresholds, we designed two sets of sensitivity experiments across six datasets, including CORA.
In the first experimental set, we fixed λ 2 at 0.8 and varied λ 1 from 0.3 to 0.7 in increments of 0.1. In the second set, λ 1 was held constant at 0.5, while λ 2 ranged from 0.6 to 1.0, also in increments of 0.1. For each experiment, we assessed the performance using four standard clustering metrics (ACC, NMI, ARI, and F1) and visualized the results as bar charts to illustrate the impact of parameter changes, as detailed in Figure 5a,b.
The results from the first experimental set reveal that the model generally attained a superior performance when λ 1 was set to 0.5 or 0.6. This suggests that λ 1 significantly influences the quality of pseudo-label filtering and the initial construction of cluster centroids. On the CITE dataset, for instance, a small λ 1 (e.g., 0.3) limited the number of high-confidence samples for initialization, resulting in less representative centroids and compromising the stability and accuracy of subsequent training. As λ 1 increased, more high-confidence samples were included in the initialization, leading to an improved clustering performance. However, an excessively high λ 1 (e.g., 0.7) introduced noisy pseudo-labels, which diminished the purity of the initial centroids and adversely affected the final results.
In the second set of experiments, as λ 2 increased, the model performance followed a pattern of initial improvement followed by a decline. On the CORA dataset, for example, all four metrics peaked when λ 2 was approximately 0.8. This indicates that a moderate λ 2 value enables the model to effectively filter out highly uncertain pseudo-labels while retaining high-confidence ones, thus boosting the overall clustering performance. Conversely, as λ 2 approaches λ 1 , the model effectively degenerates to a hard-thresholding approach, using only the highest-confidence samples and discarding potentially useful information, which results in a slight performance drop. While other datasets showed similar trends, the optimal λ 2 value varied, reflecting the different data distributions and demonstrating the model’s adaptive, yet variable, sensitivity to λ 2 .
These two sets of sensitivity experiments confirm that the dual-threshold design ( λ 1 and λ 2 ) of our self-paced learning mechanism is crucial for effective pseudo-label filtering and performance enhancement in DCPRES. A well-chosen λ 1 is instrumental for constructing high-quality initial cluster centroids, while a judiciously set λ 2 improves the model’s adaptability to samples with varying confidence levels. Together, they validate the effectiveness and robustness of our self-paced sample selection strategy.

5. Conclusions

In this paper, we proposed DCPRES. By dynamically assigning weights to samples, this model enables the more effective construction of positive and negative pairs as well as more coherent cluster structures. Our framework integrates both the attribute and structural information of nodes, and employs dual contrastive learning modules—at the instance and cluster levels—guided by a self-paced strategy. This progressive learning process, moving from high-confidence to medium- and low-confidence samples, significantly improves the robustness and clustering performance. The experimental results on multiple public datasets demonstrate that DCPRES achieved superior results, exhibiting a strong stability and generalization.
Nevertheless, we acknowledge certain limitations of our work. In particular, selecting the optimal values of the hyperparameters (such as λ 1 and λ 2 ) currently requires manual tuning. While this approach is manageable for small- and medium-scale datasets, it becomes time-consuming and computationally demanding for larger datasets, which limits the efficiency and scalability of the method. To address this issue, in our future research, we plan to investigate parameter self-adjustment mechanisms. For instance, adaptive search strategies or meta-learning techniques could allow the model to dynamically determine the optimal parameter values during training, thereby reducing manual intervention and enhancing the efficiency. Moreover, we intend to explore additional optimization strategies and regularization methods to further improve the generalization and practical applicability of the model.

Author Contributions

Conceptualization, X.Q.; methodology, X.Q., L.P., and C.Y.; software, L.P.; validation, X.Q., L.P., and Z.Q.; data curation, Z.Q.; writing—original draft preparation, X.Q.; writing—review and editing, X.Q.; visualization, L.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Science and Technology of the People’s Republic of China, under grant No. STI2030-Major Projects 2021ZD0201900, and the Key Project of Science and Technology of Guangxi (Grant No. AB25069247).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets presented in this study are openly available. A comprehensive list of these datasets and access methods can be found at https://github.com/yueliu1999/Awesome-Deep-Graph-Clustering, accessed on 1 September 2025.

Acknowledgments

The authors would like to express their gratitude to all the experts who provided valuable suggestions and support for this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, Y.; Xia, J.; Zhou, S.; Yang, X.; Liang, K.; Fan, C.; Zhuang, Y.; Li, S.Z.; Liu, X.; He, K. A survey of deep graph clustering: Taxonomy, challenge, application, and open resource. arXiv 2022, arXiv:2211.12875. [Google Scholar]
  2. Ren, Y.; Pu, J.; Yang, Z.; Xu, J.; Li, G.; Pu, X.; Yu, P.S.; He, L. Deep clustering: A comprehensive survey. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 5858–5878. [Google Scholar] [CrossRef] [PubMed]
  3. Pan, S.; Hu, R.; Fung, S.; Long, G.; Jiang, J.; Zhang, C. Learning graph embedding with adversarial training methods. IEEE Trans. Cybern. 2019, 50, 2475–2487. [Google Scholar] [CrossRef] [PubMed]
  4. Qin, X.; Yuan, C.; Jiang, J.; Chen, L. Deep semi-supervised clustering based on pairwise constraints and sample similarity. Pattern Recognit. Lett. 2024, 178, 1–6. [Google Scholar] [CrossRef]
  5. Hong, S.; Yue, T.; You, Y.; Lv, Z.; Tang, X.; Hu, J.; Yin, H. A Resilience Recovery Method for Complex Traffic Network Security Based on Trend Forecasting. Int. J. Intell. Syst. 2025, 2025, 3715086. [Google Scholar] [CrossRef]
  6. Jaiswal, A.; Babu, A.R.; Zadeh, M.Z.; Banerjee, D.; Makedon, F. A survey on contrastive self-supervised learning. Technologies 2020, 9, 2. [Google Scholar] [CrossRef]
  7. Hassani, K.; Khasahmadi, A.H. Contrastive multi-view representation learning on graphs. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020. [Google Scholar]
  8. Yang, X.; Liu, Y.; Zhou, S.; Wang, S.; Tu, W.; Zheng, Q.; Liu, X.; Fang, L.; Zhu, E. Cluster-guided contrastive graph clustering network. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  9. Liu, Y.; Yang, X.; Zhou, S.; Liu, X.; Wang, Z.; Liang, K.; Tu, W.; Li, L.; Duan, J.; Chen, C. Hard sample aware network for contrastive deep graph clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  10. Li, Y.; Hu, P.; Liu, Z.; Peng, D.; Zhou, J.T.; Peng, X. Contrastive clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021. [Google Scholar]
  11. Dang, Z.; Deng, C.; Yang, X.; Wei, K.; Huang, H. Nearest neighbor matching for deep clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Event, 19–25 June 2021. [Google Scholar]
  12. Huang, D.; Deng, X.; Chen, D.H.; Wen, Z.; Sun, W.; Wang, C.D.; Lai, J.H. Deep clustering with hybrid-grained contrastive and discriminative learning. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 9472–9483. [Google Scholar] [CrossRef]
  13. Yong, Q.; Chen, C.; Zhou, X. Contrastive Learning Subspace for Text Clustering. arXiv 2024, arXiv:2408.14119. [Google Scholar] [CrossRef]
  14. Chen, M.; Wang, B.; Li, X. Deep contrastive graph learning with clustering-oriented guidance. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024. [Google Scholar]
  15. Kulatilleke, G.K.; Portmann, M.; Chandra, S.S. SCGC: Self-supervised contrastive graph clustering. Neurocomputing 2025, 611, 128629. [Google Scholar] [CrossRef]
  16. van der Oord, A.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  17. Zhang, X.; Xu, H.; Zhu, X.; Chen, Y. Deep contrastive clustering via hard positive sample debiased. Neurocomputing 2024, 570, 127147. [Google Scholar] [CrossRef]
  18. Kumar, M.; Packer, B.; Koller, D. Self-paced learning for latent variable models. Adv. Neural Inf. Process. Syst. 2010, 23, 1189–1197. [Google Scholar]
  19. Gu, N.N.; Sun, X.N.; Liu, W.; Li, L. Semi-supervised classification method based on self-paced learning and sparse self-expression. Syst. Sci. Math. 2020, 40, 191–208. [Google Scholar]
  20. Jiang, Y.Y.; Tao, C.F.; Li, P. Deep subspace clustering algorithm with data augmentation and adaptive self-paced learning. Comput. Eng. 2023, 49, 96–103. [Google Scholar]
  21. Zhang, C.; Fan, W.; Wang, B.; Chen, C.; Li, H. Self-paced semi-supervised feature selection with application to multi-modal Alzheimer’s disease classification. Inf. Fusion 2024, 107, 102345. [Google Scholar] [CrossRef]
  22. Wang, C.; Pan, S.; Hu, R.; Long, G.; Jiang, J.; Zhang, C. Attributed graph clustering: A deep attentional embedding approach. arXiv 2019, arXiv:1906.06532. [Google Scholar] [CrossRef]
  23. Zhu, P.; Wang, Q.; Wang, Y.; Li, J.; Hu, Q. Every node is different: Dynamically fusing self-supervised tasks for attributed graph clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 17184–17192. [Google Scholar]
  24. Zhang, J.; Peng, H.; Sun, L.; Wu, G.; Liu, C.; Yu, Z. Unsupervised graph clustering with deep structural entropy. In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining, V.2, Toronto, ON, Canada, 3–7 August 2025; pp. 3752–3763. [Google Scholar]
  25. Zhao, H.; Yang, X.; Wang, Z.; Yang, E.; Deng, C. Graph debiased contrastive learning with joint representation clustering. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Montreal, QC, Canada, 21–26 August 2021. [Google Scholar]
  26. Liu, Y.; Li, J.; Chen, Y.; Wu, R.; Wang, E.; Zhou, J.; Chen, L. Revisiting modularity maximization for graph clustering: A contrastive learning perspective. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Barcelona, Spain, 25–29 August 2024; pp. 1968–1979. [Google Scholar]
  27. Wan, X.; Liu, J.; Liang, W.; Liu, X.; Wen, Y.; Zhu, E. Continual multi-view clustering. In Proceedings of the 30th ACM International Conference on Multimedia, Lisbon, Portugal, 10–14 October 2022. [Google Scholar]
  28. Liu, Y.; Yang, X.; Zhou, S.; Liu, X.; Wang, S.; Liang, K.; Tu, W.; Li, L. Simple contrastive graph clustering. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 13789–13800. [Google Scholar] [CrossRef] [PubMed]
  29. Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Overall framework of the contrastive deep graph clustering model with a progressive relaxation weighting strategy (DCPRES). The diagram illustrates a dual-branch feature extraction from attributes ( X ˜ ) and structure (A) to generate embeddings and pseudo-labels. The PRES assigns dynamic weights ( w i ) using thresholds λ 1 and λ 2 to implement an “easy-to-hard” curriculum. This guides two contrastive losses: LOSS-PREi for instance-level learning and LOSS-PREc for cluster-level refinement, collectively improving the clustering performance and noise robustness.
Figure 1. Overall framework of the contrastive deep graph clustering model with a progressive relaxation weighting strategy (DCPRES). The diagram illustrates a dual-branch feature extraction from attributes ( X ˜ ) and structure (A) to generate embeddings and pseudo-labels. The PRES assigns dynamic weights ( w i ) using thresholds λ 1 and λ 2 to implement an “easy-to-hard” curriculum. This guides two contrastive losses: LOSS-PREi for instance-level learning and LOSS-PREc for cluster-level refinement, collectively improving the clustering performance and noise robustness.
Electronics 14 04206 g001
Figure 2. Soft weight calculation time consumption analysis.
Figure 2. Soft weight calculation time consumption analysis.
Electronics 14 04206 g002
Figure 3. 2D visualization of the comparison model on the CORA dataset.
Figure 3. 2D visualization of the comparison model on the CORA dataset.
Electronics 14 04206 g003
Figure 4. 2D visualization of the iterative process on the CORA dataset.
Figure 4. 2D visualization of the iterative process on the CORA dataset.
Electronics 14 04206 g004
Figure 5. Hyperparameter sensitivity analysis of the DCPRES model. (a) Performance variation across four metrics with different values of λ 1 , while λ 2 is fixed at 0.8. (b) Performance variation with different values of λ 2 , while λ 1 is fixed at 0.5. Results are shown for the six datasets.
Figure 5. Hyperparameter sensitivity analysis of the DCPRES model. (a) Performance variation across four metrics with different values of λ 1 , while λ 2 is fixed at 0.8. (b) Performance variation with different values of λ 2 , while λ 1 is fixed at 0.5. Results are shown for the six datasets.
Electronics 14 04206 g005
Table 1. Details of the graph datasets.
Table 1. Details of the graph datasets.
DatasetType#Nodes#Edges#ClassesFeature Dim
CORAGraph2708527871433
CITEGraph3327455263703
AMAPGraph7650119,0818745
BATGraph1311038481
EATGraph39959944203
UATGraph119013,5994239
Table 2. Index comparison of different graph clustering methods on six graph datasets.
Table 2. Index comparison of different graph clustering methods on six graph datasets.
DatasetMetricsDAEGCDyFSSDeSEGDCLDCHDMAGICCGCHSANDCPRES
CORANMI52.89 ± 0.6954.58 ± 0.5157.27 ± 0.3856.30 ± 0.3660.13 ± 0.9959.71 ± 0.1256.45 ± 1.0459.21 ± 1.0361.18 ± 0.61
ACC70.43 ± 0.3672.08 ± 0.3775.31 ± 0.6170.83 ± 0.4778.67 ± 0.8776.03 ± 0.2173.88 ± 1.2077.07 ± 1.5678.65 ± 0.63
ARI49.63 ± 0.4349.33 ± 0.4454.89 ± 0.4948.05 ± 0.7259.60 ± 1.6257.31 ± 0.1152.51 ± 1.8957.52 ± 0.7059.54 ± 0.76
F168.27 ± 0.5767.37 ± 0.3472.29 ± 0.3452.88 ± 0.9776.84 ± 0.7973.92 ± 0.2670.98 ± 2.7975.11 ± 1.4076.87 ± 0.62
CITENMI36.41 ± 0.8643.96 ± 0.6544.68 ± 0.3139.52 ± 0.3844.97 ± 1.2045.24 ± 0.2844.33 ± 0.7945.06 ± 0.7445.89 ± 0.40
ACC64.54 ± 1.3970.34 ± 0.9269.88 ± 0.5166.39 ± 0.6571.13 ± 1.0070.65 ± 0.3269.84 ± 0.9471.15 ± 0.8072.31 ± 0.30
ARI37.78 ± 1.2445.37 ± 0.5445.35 ± 0.4641.07 ± 0.9646.87 ± 1.0946.83 ± 0.1545.68 ± 1.8047.05 ± 1.1248.38 ± 0.54
F162.20 ± 1.3264.01 ± 0.7564.58 ± 0.4661.12 ± 0.7062.71 ± 2.0864.82 ± 0.2462.71 ± 2.0663.01 ± 1.7966.53 ± 0.62
AMAPNMI65.25 ± 0.45--37.32 ± 0.2867.08 ± 0.88-67.44 ± 0.4867.21 ± 0.3367.85 ± 0.40
ACC75.96 ± 0.23--43.75 ± 0.7877.81 ± 0.68-77.25 ± 0.4177.02 ± 0.3377.74 ± 0.44
ARI58.12 ± 0.24--21.57 ± 0.5158.71 ± 1.47-57.99 ± 0.6658.01 ± 0.4858.79 ± 0.56
F169.87 ± 0.54--38.37 ± 0.2972.37 ± 1.41-72.18 ± 0.5772.03 ± 0.4675.49 ± 1.35
BATNMI21.43 ± 0.35--31.70 ± 0.4254.30 ± 1.54-50.23 ± 2.4353.21 ± 0.9356.04 ± 0.58
ACC52.67 ± 0.00--45.42 ± 0.5478.85 ± 1.08-75.04 ± 1.7877.15 ± 0.7280.00 ± 0.46
ARI18.18 ± 0.29--19.33 ± 0.5752.75 ± 2.02-46.95 ± 3.0952.20 ± 1.1154.97 ± 0.86
F152.23 ± 0.03--39.94 ± 0.5778.80 ± 1.03-74.90 ± 1.8077.13 ± 0.7679.98 ± 0.41
EATNMI05.57 ± 0.06--13.22 ± 0.3334.87 ± 0.54-33.85 ± 0.8733.25 ± 0.4435.90 ± 0.40
ACC36.89 ± 0.15--33.46 ± 0.1857.84 ± 0.49-57.19 ± 0.6656.69 ± 0.3458.45 ± 0.29
ARI05.03 ± 0.08--04.31 ± 0.2928.03 ± 0.49-27.71 ± 0.4126.85 ± 0.5928.92 ± 0.17
F134.72 ± 0.16--25.02 ± 0.2158.16 ± 0.46-57.09 ± 0.9457.26 ± 0.2858.60 ± 0.22
UATNMI21.33 ± 0.44--25.10 ± 0.0127.33 ± 1.66-28.15 ± 1.9226.99 ± 2.1129.28 ± 1.01
ACC52.29 ± 0.49--48.70 ± 0.0656.78 ± 1.17-56.34 ± 1.1156.04 ± 0.6758.39 ± 0.83
ARI20.50 ± 0.51--21.76 ± 0.0124.85 ± 1.91-25.52 ± 2.0925.22 ± 1.9628.19 ± 1.00
F150.33 ± 0.64--45.69 ± 0.0855.93 ± 1.84-55.24 ± 1.6954.20 ± 1.8457.23 ± 0.89
Note: Bold values indicate the best performance, while underlined values indicate the second-best performance across all methods.
Table 3. Ablation studies of DCPRES on six datasets.
Table 3. Ablation studies of DCPRES on six datasets.
DatasetMetricsPREcPREiGCN + PREcGCN + PREiPREi + PREcDCPRES
CORANMI58.01 ± 0.4659.48 ± 0.5160.95 ± 0.4261.07 ± 0.4761.22 ± 0.3761.18 ± 0.61
ACC74.47 ± 0.3875.84 ± 0.4677.49 ± 0.6377.88 ± 0.7178.41 ± 0.9078.65 ± 0.63
ARI55.67 ± 0.5656.85 ± 0.6558.35 ± 0.6858.68 ± 0.6859.47 ± 0.6159.54 ± 0.76
F173.51 ± 0.2774.00 ± 0.5975.75 ± 0.7076.09 ± 0.7176.52 ± 0.9876.87 ± 0.62
CITENMI43.97 ± 0.9943.48 ± 1.0044.37 ± 0.5344.95 ± 0.3545.02 ± 0.4445.89 ± 0.40
ACC68.86 ± 0.9769.78 ± 0.7870.35 ± 0.4970.73 ± 0.5770.90 ± 0.5872.31 ± 0.30
ARI42.90 ± 1.4544.22 ± 1.1745.21 ± 0.8345.94 ± 0.5846.14 ± 0.8248.38 ± 0.54
F164.83 ± 0.6864.71 ± 0.8065.29 ± 0.6965.45 ± 0.5365.62 ± 0.3166.53 ± 0.62
AMAPNMI65.86 ± 0.3965.91 ± 0.3966.31 ± 0.3266.86 ± 0.3967.57 ± 0.4067.85 ± 0.40
ACC74.73 ± 0.4475.78 ± 0.3975.39 ± 0.1976.73 ± 0.4477.41 ± 0.5277.74 ± 0.44
ARI55.78 ± 0.5456.92 ± 0.5656.58 ± 0.2657.78 ± 0.5458.75 ± 0.5458.79 ± 0.56
F173.71 ± 1.1972.92 ± 1.0373.90 ± 1.1575.71 ± 1.1974.62 ± 1.5875.49 ± 1.35
BATNMI52.38 ± 0.5153.21 ± 0.4654.59 ± 0.7054.43 ± 0.7853.52 ± 0.5156.04 ± 0.58
ACC76.71 ± 0.6777.48 ± 0.5178.93 ± 0.6178.63 ± 0.6877.71 ± 0.4680.00 ± 0.46
ARI50.23 ± 0.8150.95 ± 0.6153.11 ± 0.9152.79 ± 1.0551.27 ± 0.6554.97 ± 0.86
F176.66 ± 0.6577.44 ± 0.4878.90 ± 0.6078.51 ± 0.7077.67 ± 0.4379.98 ± 0.41
Note: The best results are highlighted in bold.
Table 4. Threshold setting analysis on CORA and AMAP. Fixed: hard binary cutoff; single: single-threshold soft weighting; DCPRES: double-threshold progressive relaxation, splitting samples into high-/medium-/low-confidence bands.
Table 4. Threshold setting analysis on CORA and AMAP. Fixed: hard binary cutoff; single: single-threshold soft weighting; DCPRES: double-threshold progressive relaxation, splitting samples into high-/medium-/low-confidence bands.
ThresholdsCORAAMAP
NMIACCARIF1NMIACCARIF1
Fixed56.99 ± 1.1973.02 ± 2.2552.44 ± 3.2572.20 ± 2.4466.37 ± 0.4076.55 ± 0.3456.67 ± 0.4374.57 ± 1.32
Single59.85 ± 0.6075.93 ± 0.4756.20 ± 0.5873.92 ± 0.3066.86 ± 0.3777.15 ± 0.3957.82 ± 0.5375.02 ± 0.88
DCPRES61.18 ± 0.6178.65 ± 0.6359.54 ± 0.7676.87 ± 0.6267.85 ± 0.4077.74 ± 0.4458.79 ± 0.5675.49 ± 1.35
Note: The best results are highlighted in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, X.; Peng, L.; Qin, Z.; Yuan, C. DCPRES: Contrastive Deep Graph Clustering with Progressive Relaxation Weighting Strategy. Electronics 2025, 14, 4206. https://doi.org/10.3390/electronics14214206

AMA Style

Qin X, Peng L, Qin Z, Yuan C. DCPRES: Contrastive Deep Graph Clustering with Progressive Relaxation Weighting Strategy. Electronics. 2025; 14(21):4206. https://doi.org/10.3390/electronics14214206

Chicago/Turabian Style

Qin, Xiao, Lei Peng, Zhengyou Qin, and Changan Yuan. 2025. "DCPRES: Contrastive Deep Graph Clustering with Progressive Relaxation Weighting Strategy" Electronics 14, no. 21: 4206. https://doi.org/10.3390/electronics14214206

APA Style

Qin, X., Peng, L., Qin, Z., & Yuan, C. (2025). DCPRES: Contrastive Deep Graph Clustering with Progressive Relaxation Weighting Strategy. Electronics, 14(21), 4206. https://doi.org/10.3390/electronics14214206

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop