Next Article in Journal
A Review of Artificial Intelligence and Remote Sensing for Marine Oil Spill Detection, Classification, and Thickness Estimation
Previous Article in Journal
Evaluating Consistency and Accuracy of Public Tidal Flat Datasets in China’s Coastal Zone
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Global Self-Attention-Driven Graph Clustering Ensemble

by
Lingbin Zeng
,
Shixin Yao
*,
You Huang
,
Liquan Xiao
,
Yong Cheng
and
Yue Qian
National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(22), 3680; https://doi.org/10.3390/rs17223680
Submission received: 29 August 2025 / Revised: 3 November 2025 / Accepted: 5 November 2025 / Published: 10 November 2025

Highlights

What are the main findings?
  • The proposed global self-attention-driven graph clustering ensemble (GSAGCE) effectively fuses attribute and structural information through a novel global self-attention graph autoencoder that captures long-range vertex dependencies.
  • Our double-weighted graph partitioning consensus function simultaneously considers both global and local diversity within base clusterings, enhancing the overall consensus clustering performance.
What is the implication of the main finding?
  • GSAGCE addresses critical limitations in existing clustering ensemble methods for graph-structured data.
  • The self-supervised strategy we designed provides more reliable guidance for producing high-quality base clusterings, which can be extended to other domains requiring effective processing of complex graph-structured data.

Abstract

A clustering ensemble, which leverages multiple base clusterings to obtain a reliable consensus result, is a critical challenging task for Earth observation in remote sensing applications. With the development of multi-source remote sensing data, exploring the underlying graph-structured patterns has become increasingly important. However, existing clustering ensemble methods mostly employ shallow clustering in the base clustering generation stage, which fails to utilize the structural information. Moreover, the high dimensionality inherent in data further increases the difficulty of clustering. To address these problems, we propose a novel method termed Global Self-Attention-driven Graph Clustering Ensemble (GSAGCE). Specifically, GSAGCE firstly adopts basic autoencoders and global self-attention graph autoencoders (GSAGAEs) to extract node attribute information and structural information, respectively. GSAGAEs not only enhance structural information in the embedding but also have the capability to capture long-range vertex dependencies. Next, we employ a fusion strategy to adaptively fuse this dual information by considering the importance of nodes through an attention mechanism. Furthermore, we design a self-supervised strategy to adjust the clustering distribution, which integrates the attribute and structural embeddings as more reliable guidance to produce base clusterings. In the ensemble strategy, we devise a double-weighted graph partitioning consensus function that simultaneously considers both global and local diversity within the base clusterings to enhance the consensus performance. Extensive experiments on benchmark datasets demonstrate the superiority of GSAGCE compared to other state-of-the-art methods.

1. Introduction

Benefiting from the continuous advancement of unsupervised learning, clustering tasks for complex data have significantly alleviated the challenges posed by large-scale unlabeled information [1]. Among these, the clustering ensemble has emerged as a powerful paradigm in the clustering field, aiming to elegantly combine multiple weak base clusterings to achieve a more robust and superior result [2,3,4]. This approach has demonstrated remarkable success across diverse domains, including computer vision [5,6,7], bioinformatics [8,9,10], and anomaly detection [11,12]. Clustering ensemble can effectively handle and fuse this data heterogeneity through different base clusterings. However, despite its widespread adoption, two critical limitations persist in current clustering ensemble research, particularly in the context of graph-structured data.
First, existing clustering ensemble methods mainly focus on consensus function design and selective ensemble strategies, while neglecting the generation of high-quality base clusterings—a fundamental yet understudied aspect of the ensemble framework. The performance of clustering ensemble heavily depends on the diversity and quality of base clusterings, low-quality base partitions can severely degrade the final ensemble result. Unfortunately, most current approaches rely on shallow clustering algorithms (e.g., k-means, spectral clustering) for base clustering generation, which suffer from limited representation capacity and fail to capture the complex, high-dimensional structures inherent in the data [13]. As big data and remote sensing applications increasingly generate graph-structured data [14,15,16,17], these shallow methods struggle to adapt, highlighting a pressing need for more advanced deep representation learning techniques in the base clustering phase.
Second, unlike common tabular data, graph-structured data contains more abundant information, including node attribute information and topological structure information. Due to the wealth of information inherent in graph-structured data, many real-world tasks can be effectively tackled through the clustering ensemble, relying on graphs, such as social networks [18,19] and biological interaction networks [20,21]. It is worth noting that the spatial characteristics of remote sensing data are inherently consistent with the representation of graphs. Each pixel can be regarded as a graph node, and its spatial adjacency relationship constitutes the edge structure of the graph. This graph representation can effectively preserve the most crucial spatial context information. However, existing methods often inadequately exploit this structural information. Traditional shallow clustering algorithms primarily focus on node attributes while ignoring the graph topology, which encodes critical relational patterns among samples. Even recent advances in graph neural networks (GNN) [22,23]—which excel at joint feature and topology learning [24,25,26]—have limitations. State-of-the-art GNN-based clustering methods (e.g., SDCN [27], DFCN [28]) rely heavily on localized graph convolutions (GCN [29,30]) or graph autoencoders (GAE [31,32]), which capture only short-range node dependencies and fail to model global structural patterns. This design restricts their ability to uncover higher-order connectivity and long-range interactions, ultimately compromising clustering accuracy. Therefore, effectively addressing the clustering ensemble of graph-structured data is a highly challenging research topic with significant practical value.
To address the aforementioned problems, we propose a novel approach termed Global Self-Attention-driven Graph Clustering Ensemble (GSAGCE), which simultaneously fuses attribute information and structural information for consensus clustering. Specifically, GSAGCE first exploits a basic autoencoder to mine node attribute information and then constructs a global self-attention graph autoencoder (GSAGAE) to sufficiently extract structural information. Unlike the standard graph autoencoder, GSAGAE utilizes a global self-attention mechanism to enhance the capability of graph convolutional network and captures long-range vertex dependencies on a global level. Next, we employ a learning-aware feature fusion graph network, which incorporates node importance through attention mechanisms to adaptively fuse attribute and structural information. Furthermore, we design a self-supervised strategy to tune the clustering distribution, which integrates the learned representations from the autoencoder and GSAGAE as more reliable guidance to train base clusterings. Finally, in the ensemble strategy, we devise a double-weighted graph-partitioning consensus function that simultaneously considers the global diversity of base clusterings and local diversity within the base clusterings to further enhance the consensus performance.
In summary, the main contributions of our work are as follows:
  • We propose a novel global self-attention-driven graph clustering ensemble method. In the process of capturing structural information, a novel global self-attention graph autoencoder is constructed to introduce extra expressive power to the graph convolutional network. It addresses the limitations of graph convolutional network in capturing global structural information. Moreover, the self-supervised strategy we designed can guide the learning of clustering distribution to achieve clearer boundaries and higher accuracy, significantly enhancing the clustering performance after training.
  • In the ensemble strategy, a novel double-weighted graph-partitioning consensus function is devised to incorporate a global weighting uncertainty measure into a local weighting framework. Through this approach, it not only reflects the underlying relationships between clusters but also considers the differences between base clusterings, thereby enhancing the consensus clustering performance.
  • The comprehensive experiments on seven benchmark datasets have demonstrated that our method significantly outperforms comparative state-of-the-art algorithms.
The remainder of this paper is organized as follows: In Section 2, we provide a brief related work on the clustering ensemble and deep clustering algorithms. The methodology of our proposed algorithm is elaborated in Section 3. The experimental results and analysis will be discussed in Section 4. Section 5 presents the conclusion and future work of this paper.

2. Related Work

In this section, we will introduce methods and techniques related to the proposed GSAGCE, primarily focusing on two aspects: clustering ensemble and deep clustering.
Clustering ensemble is initially inspired by the idea of ensemble learning in classification [33], which involves using multiple weak learning models to construct a more robust and superior one. Based on the construction approach of the consensus function, clustering ensemble methods can be categorized into two types: median partitioning and object co-occurrence [34]. Methods of the median partitioning type attempt to find a partition in the clustering ensemble that is most similar to all other partitions. Therefore, they typically involve solving an optimization problem to achieve a consensus partition result. Methods of the object co-occurrence type, on the other hand, obtain a consensus result by voting among the objects. Specifically, they seek cluster labels associated with each object in the consensus partition.
In recent years, many classic clustering ensemble methods have been primarily focused on research around the co-occurrence type. Tao et al. [35] proposed a robust spectral ensemble clustering method to address the noise issue in the co-association matrix by learning a robust representation of the co-association matrix through low-rank constraint. Liu et al. [36] devised a spectral ensemble clustering method for large-scale data, aiming to alleviate the high time and space complexities of consensus clustering through more efficient utilization of the co-association matrix. Zhou et al. [37] developed a dense representation-based ensemble clustering algorithm by weakening the influence of outliers and theoretically verified that the final solution is the global optimal solution. Taking into account the impact of the reliability of base clusterings on the ensemble, Huang et al. [38] proposed a clustering ensemble algorithm based on ensemble-driven cluster uncertainty estimation and local weighting strategy. This algorithm is one of the most classic clustering ensemble methods in recent years. Zhou et al. [39] optimized the clustering ensemble from the perspective of improving the quality of base clustering and proposed an adaptive consensus multiple k-means method. Of course, there are many other excellent types of clustering ensemble methods used to address various constraints [40,41]. However, most of these clustering ensemble methods obtain base clusterings using shallow clustering models, with research focusing on the design of consensus functions while overlooking the importance of the mechanism for base clustering generation.
Deep clustering combines unsupervised deep representation networks into the clustering process, enabling clustering through the learning of deep embedded representations within complex data [42]. At present, researchers have explored various deep clustering methods, which are typically categorized based on the types of unsupervised deep representation networks used, such as autoencoder-based deep clustering, variational autoencoder-based deep clustering [43,44,45,46], and graph neural network-based deep clustering. The former two mainly perform clustering based on attribute features, while the latter mainly performs clustering based on graph structure features. Here, two main types of deep clustering are introduced, including autoencoder-based (AE-based) deep clustering and graph neural network-based (GNN-based) deep clustering.
Deep embedded clustering (DEC) [47], as a representative of the AE-based type, utilizes autoencoders for representation learning and enhances cluster cohesion through a KL divergence loss function. Guo et al. [48] improved DEC by incorporating a decoder network to preserve local structures in the data. They also proposed a convolutional neural network-based version specifically tailored for image data [49]. Alqahtani et al. [50] proposed a deep convolutional autoencoder embedded clustering method that is highly effective in image processing by simultaneously learning feature representations and cluster assignments. While these AE-based methods are able to capture deep intrinsic features in complex data compared to traditional shallow clustering models, they often fail to fully leverage the structural characteristics present in graph-structured data. Bo et al. [27] proposed a structural deep clustering network to address this limitation, where they integrated structural information into deep clustering by transferring representations learned by autoencoders to graph convolutional networks. This method is the first to explicitly apply structural information to deep clustering and is also the first GNN-based deep clustering method. Tu et al. [28] proposed a deep fusion clustering network by dynamically combining autoencoders and graph neural networks to leverage structural information for enhancing clustering performance. Liu et al. [51] proposed a self-supervised deep graph clustering method that addresses the issue of representation collapse during graph node encoding. They reduced information correlation in a dual manner to enhance the discriminative capability of features. These GNN-based methods, compared to AE-based methods, have the capability to incorporate local structural information into deep clustering. However, the existing graph attention mechanism learns representations through local neighbor aggregation. Although it is computationally efficient, its attention range is limited, and it is difficult to capture long-range dependencies in the graph. This is particularly crucial in clustering-ensemble tasks, as the quality of the base clusters is highly dependent on the accurate modeling of global node similarities. Therefore, we designed a global self-attention mechanism to calculate the attention weights for all node pairs, which can directly establish associations between distant nodes and avoid the information attenuation caused by multiple layers of message passing.

3. Global Self-Attention-Driven Graph Clustering Ensemble

In this section, we present the proposed model for global self-attention-driven graph clustering ensemble (GSAGCE). The overall framework of GSAGCE is depicted in Figure 1. GSAGCE can be divided into two major modules, namely the global self-attention-driven graph clustering (GSAGC) module and the double-weighted clustering ensemble (DWCE) module.
The GSAGC can be further divided into an autoencoder (AE) module, a global self-attention graph autoencoder (GSAGAE) module, a feature fusion graph network (FFGN) module, and a self-supervised strategy module. Formally, given a graph G = { V , E } with K categories of nodes, V is the node set and E is the edge set. The graph G is characterized by its attribute feature matrix X R n × d and an original adjacency matrix A R n × n , where n is the number of samples, and d is the feature dimension. The AE module captures the node attributed feature information, and the GSAGAE module extracts the topological graph structure information. The FFGN exploits a learning-aware fusion strategy, which adaptively fuses the attribute information and the structure information by considering the importance of nodes through an attention mechanism. As for the self-supervised strategy module, to generate more reliable guidance for clustering network training, we integrate the information from AE and GSAGAE to learn a more suitable clustering distribution for target distribution generation.
In the double-weighted clustering-ensemble module, we simultaneously consider the global diversity of base clusterings and the local cluster-wise diversity inside the same base clustering. Then, we design a novel double-weighted graph-partitioning consensus strategy that incorporates a global weighting uncertainty measure into a local weighting framework. After obtaining the double-weighted cluster uncertainty, the hybrid ensemble-driven cluster estimation (HECE) is proposed to further measure the reliability of the clusters. In the following subsections, we will detail the main components of our approach. Table 1 presents the summary of the main notations used in this paper.

3.1. Global Self-Attention-Driven Graph Clustering (GSAGC)

3.1.1. Attributed Feature Representation via Autoencoder

Learning an effective data representation is crucial for clustering. There are many unsupervised representation learning methods, such as variational autoencoder [52], convolutional autoencoder, and adversarial autoencoder [53]. They are the variations in the basic autoencoder. In GSAGCE, to maintain generality, we utilize a basic autoencoder (AE) to capture the low-dimensional attributed representations of raw data. We assume that there are t layers in AE, and t indicates the number of layers. Specifically, the i-th layer representation H ( i ) R n × d i in the encoder part can be obtained as follows:
H ( i ) = ϕ W e n ( i ) H ( i 1 ) + b e n ( i ) , i = { 1 , 2 , , t } ,
where W e n ( i ) and b e n ( i ) denote the weight and bias of the i-th encoder layer. ϕ is a LeakyReLU activation function (negative input slope is set as 0.2). d i is the dimension of the i-th layer network. Moreover, H ( 0 ) is the raw attribute feature matrix X , and we typically set t = 4 . The corresponding decoder part is expressed as follows:
H ^ ( i ) = ϕ W d e ( i ) H ^ ( i 1 ) + b d e ( i ) , i = { 1 , 2 , , t } ,
Here, H ^ ( i ) denotes the decoder outputs from the i-th layer, and  H ^ ( t ) = X ^ . X ^ is the reconstruction of raw data. W d e ( i ) and b d e ( i ) are the weight and bias of the i-th decoder layer. Then, we can obtain the attribute information by minimizing the reconstruction loss between X and X ^ :
L R H = X X ^ F 2 ,
where · F indicates the F-norm. By minimizing the reconstruction loss L R H , the model is able to learn to accurately reconstruct the representations of node attributes, thus enabling us to obtain the corresponding attribute information in the graph structure data.

3.1.2. Structural Information via Global Self-Attention Graph Autoencoder

The AE module is able to learn the attributed feature representation from the node feature matrix while failing to exploit the underlying structure information among graph data. To sufficiently extract the graph structural information, we propose a global self-attention graph autoencoder (GSAGAE). GSAGAE firstly transforms the l-th encoder hidden layer representation Z ( l ) ( l t ) into two feature spaces ( f , g ) to calculate the self-attention with the following equation:
β i , j = softmax j { 1 , 2 , , n } s i , j , where s i , j = Z i ( l ) W f Z j ( l ) W g T ,
β i , j indicates the attention importance of node j on node i. W f and W g are two learned weight matrices, where the first role is to reduce the computational load by dimensionality reduction, and the second is to provide additional flexibility for trainable variables. After obtaining the attention importance map, the global attention feature can be expressed with the following equation:
o i ( l ) = j = 1 n β i , j Z j ( l ) W h W r ,
where W h is for dimensionality reduction and W r projects the features back to the original size.
The global self-attention representation captures the information on a global level. However, for graph networks, local geometry is also crucial. Thus, we add the input feature map back to the global attention feature:
Z ( l + 1 ) = σ D 1 2 ( A + I ) D 1 2 Z ( l ) + γ O ( l ) W ( l ) ,
Here, D = d i a g ( D 1 , D 2 , , D n ) is the degree matrix, and D i = v j V ( a i j + I i j ) . I is the identity matrix. D 1 2 ( A + I ) D 1 2 means the normalization for the original adjacency matrix A . The reason for this normalization is that directly using the original adjacency matrix causes nodes to prefer high-degree nodes regardless of the information carried in the node. O ( l ) is the output of the global self-attention, and  W ( l ) is the feature transformation weight of l-layer. In Equation (5), o i ( l ) represents the i-th row of O ( l ) . σ is the LeakyReLU activation function (negative input slope is set as 0.2). γ > 0 is a learnable scalar with initial value 0. Introducing a learnable γ allows the network to first explore the local neighborhood and then gradually assign more weights to the global level. During model training, when the tensor parameter ‘requires_grad’ is set to True, the value of γ will be automatically updated as the gradient descends.
The core limitation of existing attention graph neural networks (such as GAT) is the “limited attention range”: they only calculate attention weights based on the local neighbors of nodes (such as K-nearest neighbors, first-order neighborhoods). Although it can capture the local structural correlations, it cannot directly model the long-range dependencies of non-neighbor nodes. The innovation of GSAGAE lies in designing a “global attention feature + local graph convolution” dynamic fusion module instead of simply using global attention. Through γ , it achieves a dynamic balance of local and global information without manually setting the neighbor range and can simultaneously cover local topology and global correlation, which is a mechanism design that existing attention GNNs have not involved.
Then, in decoder, we use the inner product of Z and Z T to reconstruct the graph topological structure A ^ :
A ^ = sigmoid Z Z ,
The reconstruction loss of GSAGAE is given by
L R A = CrossEntropyLoss ( A , A ^ ) = 1 n 2 i = 1 n j = 1 n A i j log ( A ^ i j ) + ( 1 A i j ) log ( 1 A ^ i j ) ,
By integrating Equations (3) and (8), the reconstruction loss of the GSAGC module is defined as follows:
L R = L R H + L R A ,

3.1.3. Feature Fusion Graph Network

To further enhance the capability of representation learning, we exploit a learning-aware feature fusion graph network to adaptively fuse the attributed and structural representation. Here, the attention mechanism is employed for feature fusion, aiming to address the issue of feature heterogeneity. The attribute representation H ( i ) and the structural representation Z ( i ) come from different information sources and are thus heterogeneous representations. Therefore, in the traditional fusion strategy, direct simple average accumulation will lead to the corresponding loss in heterogeneous expression fusion. Specifically, we first concatenate H ( i ) and Z ( i ) as [ H ( i ) Z ( i ) ] R n × 2 d i . Then, a weight matrix W a ( i ) R 2 d i × 2 is introduced to capture the relationship of the concatenated features. Moreover, we apply a LeakyReLU activation function (negative input slope is set as 0.2) on the product between [ H ( i ) Z ( i ) ] and W a ( i ) . Finally, through softmax function and 2 normalization, we formulate the corresponding expression as follows:
M ( i ) = 2 ( softmax   ( ( LeakyReLU   ( [ H ( i ) Z ( i ) ] W a ( i ) ) ) ) ) ,
where M ( i ) indicates the attention coefficient matrix. Thus, we can adaptively fuse the GSAGAE feature Z ( i ) and the AE feature H ( i ) as follows:
Z ˜ ( i ) = M 1 ( i ) H ( i ) + M 2 ( i ) Z ( i ) ,
M 1 ( i ) , M 2 ( i ) represent the expanded matrix of M ( i ) , and ⊙ denotes the Hadamard product. Afterwards, we obtain the final fused representation Z ˜ and directly use the softmax function to obtain the clustering probability distribution:
Z L = softmax ( Z ˜ ) , s . t . j = 1 K z i j = 1 , z i j > 0 ,
Thus, we can obtain the clustering label through Z L :
y i = arg max j z i j , j = { 1 , 2 , , K } ,

3.1.4. Self-Supervised Strategy

Due to the lack of reliable guidance for cluster probability distribution Z L when training the previous networks, we integrate the information from AE and GSAGAE to obtain suitable clustering distribution for guidance. In particular, we calculate the soft assignment distribution of H and Z using Student’s t-distribution [47]:
q h i j = 1 + h i μ j 2 1 j 1 + h i μ j 2 1 , q z i j = 1 + z i v j 2 1 j 1 + z i v j 2 1 ,
where q h i j indicates the similarity between h i and cluster center μ j , h i is the i-th row of H , and cluster center μ j is initialized by k-means. The initialization of cluster centers is carried out after the pre-training process. The pre-trained representations are then subjected to the k-means algorithm to initialize the cluster centers μ j . The reason for the existence of the Student’s t-distribution lies in that its power-law form can better model the similarity between high-dimensional data points and the cluster centers, while also alleviating the interference of outliers on the soft assignment. The interpretation of q z i j is similar to q h i j . Then, we construct the corresponding auxiliary distribution of H and Z as follows:
p h i j = q h i j 2 / i q h i j j q h i j 2 / i q h i j , p z i j = q z i j 2 / i q z i j j q z i j 2 / i q z i j ,
Thus, we can utilize auxiliary distribution p h to guide q h and q z , which help AE and GSAGAE to learn more reliable representation for clustering. Here, we minimize the KL divergence loss among them:
L K L H = K L ( p h q h ) + K L ( p h q z ) , = i j p h i j log p h i j q h i j + i j p h i j log p h i j q z i j ,
In addition, to promote a highly consistent distribution alignment to train our model, we employ auxiliary distribution p z to guide the update of q z and cluster distribution Z L by minimizing the KL divergence:
L K L Z = K L ( p z q z ) + K L ( p z Z L ) , = i j p z i j log p z i j q z i j + i j p z i j log p z i j Z i j L ,
In summary, combining the Equations (9), (16) and (17), the overall loss function of GSAGC can be written as follows:
L = L R + λ 1 L K L H + λ 2 L K L Z ,
where λ 1 > 0 and λ 2 > 0 are the trade-off hyper-parameters. For clarity, we summarize the complete training process of GSAGC in Algorithm 1.
Algorithm 1 Training process of GSAGC
Input: 
X : attribute feature matrix, A : original adjacency matrix, K: number of clusters, λ 1 , λ 2 : trade-off hyper-parameters (default 1), I t e r max : maximum iterations (default 200).
Output: 
π : clustering result
  1:
Pre-trained basic AE via Equations (1)–(3);
  2:
Pre-trained GSAGAE via Equations (4)–(8);
  3:
Initialize the k-means cluster center μ and v;
  4:
while  i t e r I t e r max   do
  5:
    Compute the attributed representations H ( i ) ;
  6:
    Reconstruct raw attribute data X ^ ;
  7:
    Compute the loss function L R H via Equation (3);
  8:
    Compute the structural representation Z ( i ) via Equation (6);
  9:
    Fuse H ( i ) and Z ( i ) via Equations (10) and (11) to obtain Z ˜ ;
10:
    Generate representations of GSAGAE via Equation (6);
11:
    Reconstruct the graph structure A ^ ;
12:
    Compute the loss function L R A via Equation (8);
13:
    Generate the cluster probability distribution Z L via Equation (12);
14:
    Compute the soft assignment distribution q h and q z via Equation (14);
15:
    Compute the corresponding auxiliary distribution p h and p z via Equation (15);
16:
    Compute the KL divergence loss L K L H and L K L Z via Equations (16) and (17);
17:
    Compute the overall loss L via Equation (18);
18:
    Backward propagate and update parameters in GSAGC;
19:
     i t e r = i t e r + 1 ;
20:
end while
21:
Obtain clustering result π via Equation (13);
22:
return  π

3.2. Double-Weighted Clustering Ensemble (DWCE)

Given the ensemble Π = { π 1 , , π M } consisting of M base clusterings, π m = { C 1 m , C 2 m , , C n m m } is the m-th base clustering. n m is the number of clusters for π m , and C i m is the i-th cluster.

3.2.1. Global Uncertainty Estimation

Considering the global diversity of base clusterings in ensemble Π , we propose a global uncertainty estimation to calculate the reliability of base clustering. Specifically, we first measure the similarity between two base clusterings π m and π n by the adjusted rand index (ARI) [54]:
Sim ( π m , π n ) = ARI ( π m , π n ) ,
Then, we can obtain the overall similarity for π m :
Sim ( π m ) = 1 M 1 π n Π , n m Sim ( π m , π n ) ,
After that, the global uncertainty estimation (GUE) in Π for π m can be defined as follows:
GUE ( π m ) = Sim ( π m ) max π n Π Sim ( π n ) ,
where the value of GUE is between [0, 1]. The higher the GUE value of a base clustering, the higher the credibility of the clustering results.

3.2.2. Local Uncertainty Measurement

Inspired by the uncertainty measurement of different clusters [38], we consider the local cluster diversity inside the same base clustering. In particular, the uncertainty of cluster C i π m ( π m Π ) is calculated as follows:
H m C i = j = 1 n m p C i , C j m log 2 p C i , C j m , p C i , C j m = C i C j m C i ,
where C j m is the j-th cluster in π m , and C i represents the number of objects. Then, the entire local uncertainty of C i in ensemble Π can be obtained via the following equation:
H Π ( C i ) = m = 1 M H m ( C i ) ,

3.2.3. Hybrid Ensemble-Driven Cluster Estimation

Based on global uncertainty estimation in Equation (21) and local uncertainty measurement in Equation (23), we can compute the double-weighted cluster uncertainty of C i in ensemble Π :
D Π ( C i ) = H Π ( C i ) GUE ( π m ) ,
When the value of GUE ( π m ) is larger, the reliability of the base clustering that generates C i is higher. Thus, the uncertainty of C i is expected to decrease. In the way, it makes the cluster-level uncertainty more accurate and further estimates the uncertainty of the cluster from the base clustering level.
After computing the double-weighted cluster uncertainty, we construct a novel hybrid ensemble-driven cluster estimation (HECE) to measure the reliability of the clusters, which is defined as follows:
HECE ( C i ) = e D Π ( C i ) θ · M ,
Here, θ > 0 adjusts the impact of uncertainty on the HECE. θ is set to 0.4 in this paper. It is easy to see that a smaller uncertainty value leads to a greater HECE value.
Finally, based on HECE, we design a double-weighted graph partitioning consensus function. The weighted graph is expressed as G = ( V , L ) , in which V = X C is the node set and L is the link set. The link weight between two nodes v i and v j is calculated as follows:
l i j = HECE v j , if v i X , v j C , and v i v j , HECE v i , if v j X , v i C , and v j v i , 0 , otherwise .
where C = C 1 , C 2 , , C N c consists of all clusters in Π . N c = n 1 + + n M denotes the total number of clusters in Π . The final consensus clustering π * is obtained by using the Tcut algorithm [39].

3.3. Method Discussions

This section concludes with discussions of the computational and space complexity of our approach. In addition, we introduce the training steps of the GSAGC part.
Computational complexity: The computational complexity of GSAGCE consists of two parts: global self-attention-driven graph clustering (GSAGC) and double-weighted clustering ensemble (DWCE). For a given data matrix X R n × d , n is the number of samples, and d is the feature dimension. t indicates the number of network layers. d i is the dimension of the i-th layer network. The graph structure G = { V , E } is divided into K categories of nodes. | E | denotes the number of edges. For the GSAGC, the complexity comes from the basic autoencoder (AE), global self-attention graph autoencoder (GSAGAE), feature fusion graph network (FFGN), and self-supervised module. The complexity of AE is O ( n i = 2 t d i 1 d i ) . For the GSAGAE, as the operation can be computed efficiently using sparse matrix computation, the complexity is only O ( | E | i = 2 t d i 1 d i + n d i = 2 t d i 1 d i + n 2 ) according to [55]. The complexity of FFGN is O ( i = 2 t d i 1 d i + K ) , and the complexity of the self-supervised module is O ( n K + n l o g n ) . For the DWCE, the complexity is O ( M n m + K n e + K ( n m ) 2 ) , where M is the number of base clusterings, n m is the number of clusters in π m , and e is the average number of links connecting to a node in the graph partitioning.
Space complexity: In the training process of GSAGC, the space complexity main composes of attribute feature matrix X , adjacency matrix A , encoder matrices H ( i ) , decoder matrices H ^ ( i ) , global attention representation matrix O ( l ) , GSAGAE representation matrix Z ( l ) , and fused representation matrix Z ˜ . Thus, the whole space cost can be analyzed as O ( n d + n 2 + 4 i = 1 t n d i + n d t ) .
GSAGC training steps: In order to learn more effective feature representations and avoid overfitting problems, we adopted pre-training and self-supervision strategies to train our GSAGC part. Specifically, we pre-train the AE and GSAGAE, respectively, and then load the pre-trained weights into the overall framework for joint training in combination with the self-supervised strategy and clustering loss.

4. Experiments and Results

In this section, to verify the effectiveness of the proposed GSAGCE algorithm, we conduct comprehensive experiments on seven benchmark datasets commonly used to compare GSAGCE against ten clustering ensemble methods and six deep clustering algorithms. Moreover, we also conduct ablation experiments, hyper-parameter sensitivity studies and visualization results to demonstrate the effectiveness of all parts in our model.
Regarding the clustering performance evaluation metric, we choose four widely used metrics, including accuracy (ACC), normalized mutual information (NMI), adjusted rand index (ARI), and macro F1. The larger the values of these metrics, the better the clustering performance. The detailed computation of these evaluation metrics is shown in the literature [56].

4.1. Datasets

We evaluate the proposed GSAGCE on seven benchmark datasets, including both graph type data and non-graph type data. Cora, ACM, Pubmed, and IMDB are typically popularly used in graph clustering literature. USPS, HAR, and REUT datasets are classically used to evaluate the performance of graph clustering algorithms on non-graph data. We summarize the brief information of datasets in Table 2. For a detailed description of datasets, please refer to the relevant literature [28,51]. For the non-graph datasets, we follow literature [27] and construct the adjacency matrix with the heat kernel method.

4.2. Experimental Settings

For all comparison baseline algorithms, for each dataset, we run 10 times and compute the average and standard deviation result for the evaluation measure. For clustering ensemble baselines, their ensemble size M is 20, and the number of clusters K is set to the true number of categories on the dataset.
For our porposed GSAGCE, the layers of the AE and GSAGAE modules are set to 4. The corresponding network dimensions of AE are 500-500-2000-20. The network dimensions of GSAGAE are the same as the AE. Before the formal training, we first pre-trained the AE module and the GSAGAE module separately. The epoch is set to 50 in the pre-training phase. The optimizer is selected as Adam, and the learning rate is dynamically adjusted using the cosine annealing strategy provided by PyTorch1.13. For the hyper-parameters λ 1 ,   λ 2 in Equation (18), we perform a parameter sensitivity experiment to select the best performing value for each data. In most cases, we can set λ 1 and λ 2 to 1 by default. The remaining parameters are automatically updated by setting the tensor parameters ‘requires_grad’ to true, thereby enabling automatic differentiation. For non-graph data, we utilize KNN to construct a 3-NN graph with the heat kernel method. To ensure fair comparison, all the comparison methods of graph neural networks are based on the same graph construction method.

4.3. Analysis of Comparison Experiment

To fully verify the superiority of the GSAGCE, we compare with ten clustering ensemble methods, including RSEC [35], SECWK [36], LWEA [38], LWGP [38], DREC [37], LRTCE [57], ECCMS [58], ACMK [39], SCCABG [40], and CEAM [41]. Moreover, we also conduct comparisons with six popular deep clustering algorithms, including AE-kmeans [59], DEC [47], IDEC [48], SDCN [27], DFCN [28], and DCRN [51]. Among them, the first three are deep clustering methods that only utilize attribute information. The remaining three are deep graph clustering methods that simultaneously employ both attribute and structural information. Here is a brief introduction to these comparative methods.
  • Clustering ensemble methods: Base clusterings employ a shallow clustering model, which can only utilize node attribute information and cannot utilize structural information.
    -
    RSEC [35] addresses the noise issue in the co-association matrix by learning a robust representation through low-rank constraint.
    -
    SECWK [36] aims to alleviate the high time and space complexities of clustering ensemble through a more efficient utilization of the co-association matrix.
    -
    LWEA and LWGP [38] consider ensemble-driven cluster uncertainty estimation and local weighting strategy in clustering ensemble.
    -
    DREC [37] uses a dense representation-based ensemble clustering algorithm by weakening the influence of outliers.
    -
    LRTCE [57] refines the co-association matrix from a global perspective through a novel constrained low-rank tensor approximation model.
    -
    ECCMS [58] utilizes a co-association matrix self-enhancement method to strengthen its quality.
    -
    ACMK [39] exploits an adaptive consensus multiple k-means method by improving the quality of base clustering.
    -
    SCCABG [40] gradually incorporates data from more reliable to less reliable sources in consensus learning using adaptive bipartite graph learning.
    -
    CEAM [41] learns an updated representation using a manifold ranking model through adaptive multiplex.
  • Deep clustering methods:
    -
    AE-kmeans [59] learns embedded vectors of raw data using a basic autoencoder and then performs the k-means algorithm on the embedding.
    -
    DEC [47] utilizes denoising autoencoders for representation learning and enhances cluster cohesion through a KL divergence loss function.
    -
    IDEC [48] enhances DEC by incorporating a decoder network to preserve local structures.
    -
    SDCN [27] integrates structural information into deep clustering for the first time by transferring representations learned by autoencoders to graph convolutional networks.
    -
    DFCN [28] enhances clustering performance by dynamically combining autoencoders and graph neural networks while leveraging both attribute and structural information.
    -
    DCRN [51] is a self-supervised deep graph clustering method that addresses the issue of representation collapse during graph node encoding.
The clustering performance results of the comparison experiment are shown in Table 3 and Table 4. For ease of presentation, we have retained four significant figures and multiplied the values of the four evaluation metrics by 100%. Table 3 shows the comparative results of clustering ensemble methods, while Table 4 presents the comparative results of deep clustering algorithms. The best clustering performance obtained for each dataset is highlighted in bold. All the results in bold are statistically superior (based on a paired t-test at the 5% significance level) to the other non-bold methods.
From Table 3, we can observe that GSAGCE significantly outperforms the comparative clustering ensemble methods. The reason is that their base clustering generation algorithm adopts a shallow clustering model, which cannot leverage the implicit information in these high-dimensional sparse data, especially for graph type data, as they overlook the graph structure information. Graph structure information is conducive to capturing the relationships between data and discovering cluster structures. The quality of base clustering plays a crucial role in the ensemble phase, low-quality base clustering will inevitably greatly reduce the performance of the ensemble. Regarding the comparative clustering ensemble baselines, we found that LWEA and LWGP achieve performance close to GSAGCE on USPS data, but significantly underperformed GSAGCE on HAR and REUT datasets. SCCABG fails to obtain reasonable clustering results on HAR data, so the evaluation metrics are empty (-). The ECCMS, ACMK, and SCCABG methods performed poorly on the four graph datasets. The reason is that they are based on adaptive strategies to explore ensemble clustering, and are not well suited for sparse graph structure data. The CEAM algorithm sometimes produces SVD errors when applied to graph type data. In terms of computational efficiency, RSEC, LRTCE, ECMMS, and ACMK exhibit exponential growth in runtime as the number of data samples and dimensions increase. SECWK is the fastest, followed by LEWA, LWGP, and GSAGCE, while the remaining methods have similar runtime durations.
From Table 4, we can observe that GSAGCE has achieved the best clustering performance compared to these deep clustering algorithms. The reason is that GSAGCE not only utilizes attribute information and structural information simultaneously but also incorporates global self-attention to enhance the capability of graph convolution network (GCN) to capture long-range vertex dependencies and feature convolutions. This capability provides additional expressive power to GCN. On the other hand, GSAGCE further enhances the capability of representation by adaptively fusing dual information in a synergistic manner. For the comparative deep clustering baselines, we found that DCRN performed poorly on three non-graph data types and encountered a CUDA out of memory (oom) issue on the Pubmed data. Even when the DCRN algorithm was imported to a larger GPU memory platform, the problem persisted.

4.4. Analysis of Ablation Experiment

To demonstrate the effectiveness of global self-attention graph autoencoder (GSAGAE), self-supervised strategy, and hybrid ensemble-driven cluster estimation (HECE) in Section 3, we conduct ablation experiments with four variants of GSAGCE in this section. “noGSA” is the variant of GSAGCE without global self-attention, which aims to verify the expressive power of GSAGAE. In noGSA, we only employ a standard graph convolution network without using global self-attention. “noKLh” and “noKLz”, respectively, denote the variant of GSAGCE that does not utilize the L K L H loss and L K L Z KL divergence loss. “noHECE” represents a variant of GSAGCE that excludes the double-weighted cluster uncertainty metrics and adopts only local uncertainty measurement. The ablation experiment results are shown in Table 5.
From Table 5, we can observe that the clustering performance of GSAGCE, considering all components together, is superior to these variant algorithms. Moreover, we found that the clustering performance of the noKLz algorithm is the worst across almost all datasets except for the ACM and REUT datasets. It indicates that in the various components of GSAGCE, utilizing p z to guide q z and cluster distribution Z L training is crucial. This is because the clustering performance is entirely dependent on the training level of the cluster distribution Z L . Only with reliable training guidance can a clustering-friendly cluster distribution be obtained. The variant with the second worst clustering performance is noKLh, and its performance gap compared to GSAGCE is quite significant. This indicates that utilizing p h to guide q h and q z training is necessary, as it can help AE and GSAGAE learn more reliable representations. noGSA has achieved very poor clustering performance on REUT data and obtained a significant performance gap to GSAGCE on other datasets. This indicates that integrating global self-attention representations into GCN is highly effective. Furthermore, we have discovered an interesting phenomenon during the training of noGSA. In formal training, noGSA often reaches its highest clustering performance only in the initial iterations of training. In subsequent training, the clustering performance deteriorates until it stabilizes at a certain value, which is usually significantly lower than the highest value. This unreasonable phenomenon would not occur in GSAGCE. Regarding the variant noHECE, its performance on all datasets is also lower than GSAGCE. This indicates that the construction of a double-weighted cluster uncertainty metric is effective, performing better than considering only local uncertainty measurement.

4.5. Analysis of Hyper-Parameters Sensitivity

In the previous ablation experiments, we found that the KL divergence loss functions, L K L H and L K L Z , play crucial roles in the entire GSAGCE component. Therefore, selecting appropriate hyper-parameters, λ 1 and λ 2 , is particularly important. In this section, we conduct sensitivity experiments on two hyper-parameters. Specifically, we take λ 1 and λ 2 through {0.001, 0.01, 0.1, 1, 10, 100, 1000} and carry out experiments to take the mean value of ARI on the Cora and ACM dataset. Selecting these two datasets as representatives for hyper-parameter sensitivity analysis, the results are shown in Figure 2.
From Figure 2, we can observe that the clustering performance on the Cora dataset is less affected by two hyper-parameters, and the performance improves as the values of λ 1 and λ 2 approach 1. On the ACM dataset, the performance is greatly influenced by two hyper-parameters; especially when λ 1 is less than λ 2 , the clustering performance is poor. Moreover, when λ 1 is less than 0.1, regardless of the value of λ 2 , the clustering performance is poor. This indicates that for the ACM dataset, when guiding the training for reliable representations simultaneously in a self-supervised strategy, the importance of guidance from p h is greater. The main reason for the difference in hyper-parameter sensitivity between the Cora dataset and the ACM dataset lies in the complexity of the data. Cora is a classic citation dataset, and compared to ACM, it has a relatively smaller number of nodes and edges, with a simpler structure. Therefore, the model’s performance on the Cora dataset is less sensitive to hyper-parameter choices, as the inherent characteristics of the data make it easier for the model to converge to a good solution. The ACM dataset has a larger number of nodes and edges, as well as a more complex structure. This implies that models may require more precise hyper-parameter tuning during training to achieve optimal performance. The optimal hyper-parameter acquisition approach for other datasets is the same as these two datasets. In most datasets, we can set λ 1 and λ 2 to 1 by default.

4.6. Analysis of Visualization

For a more intuitive display of the clustering effects in the global self-attention-driven graph clustering (GSAGC), we utilize t-SNE for the visualization of clustering on ACM and HAR data. Specifically, t-SNE visualizations are performed on the raw data matrix, the q h distribution, the q z distribution, and the cluster label distribution Z L . The visualization results are shown in Figure 3. From Figure 3, we can observe that compared to the raw data representation, the representations generated after GSAGC training are significantly better than the original data, exhibiting a more clustering-friendly nature. Therefore, this results in high-quality base clusterings generated by GSAGC, laying the foundation for the subsequent ensemble performance.

5. Conclusions

In this paper, we present a novel clustering-ensemble algorithm called the global self-attention-driven graph clustering ensemble, which introduces graph clustering to the clustering ensemble. Considering the sparse and high-dimensional characteristic of graph-structured data, as well as the inability of existing clustering-ensemble methods to fully utilize the structural information in graph data, global self-attention-driven graph clustering (GSAGC) is constructed to simultaneously fuse attribute information and structural information. Moreover, we employ dual information to guide more reliable base clustering training. To enhance the consensus performance, we propose a global and local uncertainty estimation in the ensemble phase. The sufficient experiments on benchmark datasets validate the superiority of GSAGCE. The drawback and limitation of our method lies in the fact that it cannot be applied directly when dealing with incomplete or missing graph data. When applied to large-scale graphs, direct use may fail (due to out of CUDA memory). Dimensionality reduction must be performed first before this method can be used. Meanwhile, our method is not very suitable for heterogeneous graphs. It was mainly designed for isomorphic graphs. The gap between the benchmark data performance of the method and its adaptability in practical applications mainly lies in the fact that the benchmark data are standard and complete graph data. When there are missing parts or incorrect graph node structures in real data, it will significantly affect its performance. Combining incomplete clustering or multi-view learning with the proposed method is the main direction of our future work.

Author Contributions

Conceptualization, L.Z., S.Y., and Y.Q.; methodology, L.Z., S.Y., and Y.Q.; software, S.Y.; validation, L.Z., S.Y., L.X., and Y.C.; formal analysis, Y.H. and S.Y.; investigation, S.Y. and Y.Q.; resources, L.Z., Y.H., and Y.Q.; data curation, L.Z., S.Y., and Y.C.; writing—original draft preparation, L.Z., S.Y., and Y.Q.; writing—review and editing, L.Z., S.Y., Y.Q., L.X., and Y.C.; visualization, L.X., Y.H., and Y.C.; supervision, S.Y., Y.Q., L.X., and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Publicly available datasets were analyzed in this study. Cora data can be found at https://github.com/ki-ljl/PyG-GCN (accessed on 20 January 2022). ACM, Pubmed, IMDB, USPS, HAR, and REUT data can be found at https://github.com/yueliu1999/Awesome-Deep-Graph-Clustering/tree/main/dataset (accessed on 10 January 2021).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hu, X.; Li, T.; Zhou, T.; Peng, Y. Deep Spatial-Spectral Subspace Clustering for Hyperspectral Images Based on Contrastive Learning. Remote Sens. 2021, 13, 4418. [Google Scholar] [CrossRef]
  2. Wang, W.; Wang, W.; Liu, H. Correlation-Guided Ensemble Clustering for Hyperspectral Band Selection. Remote Sens. 2022, 14, 1156. [Google Scholar] [CrossRef]
  3. Zhang, M. Weighted clustering ensemble: A review. Pattern Recognit. 2022, 124, 108428. [Google Scholar] [CrossRef]
  4. Golalipour, K.; Akbari, E.; Hamidi, S.S.; Lee, M.; Enayatifar, R. From clustering to clustering ensemble selection: A review. Eng. Appl. Artif. Intell. 2021, 104, 104388. [Google Scholar] [CrossRef]
  5. Zhai, H.; Zhang, H.; Li, P.; Zhang, L. Hyperspectral image clustering: Current achievements and future lines. IEEE Geosci. Remote Sens. Mag. 2021, 9, 35–67. [Google Scholar] [CrossRef]
  6. Zhao, Y.; Li, X. Deep spectral clustering with regularized linear embedding for hyperspectral image clustering. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5509311. [Google Scholar] [CrossRef]
  7. Guo, D.; Zhang, S.; Zhang, J.; Yang, B.; Lin, Y. Exploring Contextual Knowledge-Enhanced Speech Recognition in Air Traffic Control Communication: A Comparative Study. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 16085–16099. [Google Scholar] [CrossRef]
  8. Li, J.; Deng, W.; Dang, X.; Zhao, H. Cross-Domain Adaptation Fault Diagnosis with Maximum Classifier Discrepancy and Deep Feature Alignment Under Variable Working Conditions. IEEE Trans. Reliab. 2025, 74, 4106–4115. [Google Scholar] [CrossRef]
  9. Liu, Q.; Zhao, X.; Wang, G. A Clustering Ensemble Method for Cell Type Detection by Multiobjective Particle Optimization. IEEE/ACM Trans. Comput. Biol. Bioinform. 2023, 20, 1–14. [Google Scholar] [CrossRef]
  10. Nie, X.; Qin, D.; Zhou, X.; Duo, H.; Hao, Y.; Li, B.; Liang, G. Clustering ensemble in scRNA-seq data analysis: Methods, applications and challenges. Comput. Biol. Med. 2023, 159, 106939. [Google Scholar] [CrossRef]
  11. Yu, J.; Kang, J. Clustering ensemble-based novelty score for outlier detection. Eng. Appl. Artif. Intell. 2023, 121, 106164. [Google Scholar] [CrossRef]
  12. Ray, B.; Ghosh, S.; Ahmed, S.; Sarkar, R.; Nasipuri, M. Outlier detection using an ensemble of clustering algorithms. Multimed. Tools Appl. 2022, 81, 2681–2709. [Google Scholar] [CrossRef]
  13. Yang, Y.; Lv, H.; Chen, N. A survey on ensemble learning under the era of deep learning. Artif. Intell. Rev. 2023, 56, 5545–5589. [Google Scholar] [CrossRef]
  14. Demirbaga, U.; Aujla, G.S.; Jindal, A.; Kalyon, O. Big Data Analytics: Theory, Techniques, Platforms, and Applications; Springer Nature: Berlin/Heidelberg, Germany, 2024. [Google Scholar]
  15. Chi, M.; Plaza, A.; Benediktsson, J.A.; Sun, Z.; Shen, J.; Zhu, Y. Big data for remote sensing: Challenges and opportunities. Proc. IEEE 2016, 104, 2207–2219. [Google Scholar] [CrossRef]
  16. Chen, H.; Sun, Y.; Li, X.; Zheng, B.; Chen, T. Dual-Scale Complementary Spatial-Spectral Joint Model for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 6772–6789. [Google Scholar] [CrossRef]
  17. Deng, W.; Shen, J.; Ding, J.; Zhao, H. Robust Dual-Model Collaborative Broad Learning System for Classification Under Label Noise Environments. IEEE Internet Things J. 2025, 12, 21055–21067. [Google Scholar] [CrossRef]
  18. Chunaev, P. Community detection in node-attributed social networks: A survey. Comput. Sci. Rev. 2020, 37, 100286. [Google Scholar] [CrossRef]
  19. Kumar, S.; Mallik, A.; Khetarpal, A.; Panda, B.S. Influence maximization in social networks using graph embedding and graph neural network. Inf. Sci. 2022, 607, 1617–1636. [Google Scholar] [CrossRef]
  20. Ramos, R.H.; Ferreira, C.O.L.; Simao, A. Human protein-protein interaction networks: A topological comparison review. Heliyon 2024, 10, e27278. [Google Scholar] [CrossRef]
  21. Jha, K.; Saha, S.; Singh, H. Prediction of protein-protein interaction using graph neural networks. Sci. Rep. 2022, 12, 8360. [Google Scholar] [CrossRef]
  22. Wu, L.; Cui, P.; Pei, J.; Zhao, L.; Guo, X. Graph neural networks: Foundation, frontiers and applications. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 4840–4841. [Google Scholar]
  23. Zhao, Q.; Li, L.; Chu, Y.; Yang, Z.; Wang, Z.; Shan, W. Efficient supervised image clustering based on density division and graph neural networks. Remote Sens. 2022, 14, 3768. [Google Scholar] [CrossRef]
  24. Marfo, W.; Tosh, D.K.; Moore, S.V. Enhancing Network Anomaly Detection Using Graph Neural Networks. In Proceedings of the 22nd Mediterranean Communication and Computer Networking Conference (MedComNet), Nice, France, 11–13 June 2024; pp. 1–10. [Google Scholar]
  25. Li, M.; Micheli, A.; Wang, Y.G.; Pan, S.; Lió, P.; Gnecco, G.S.; Sanguineti, M. Guest Editorial: Deep Neural Networks for Graphs: Theory, Models, Algorithms, and Applications. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 4367–4372. [Google Scholar] [CrossRef]
  26. Villaizan-Vallelado, M.; Salvatori, M.; Carro, B.; Sanchez-Esguevillas, A.J. Graph neural network contextual embedding for deep learning on tabular data. Neural Netw. 2024, 173, 106180. [Google Scholar] [CrossRef]
  27. Bo, D.; Wang, X.; Shi, C.; Zhu, M.; Lu, E.; Cui, P. Structural deep clustering network. In Proceedings of the Web Conference, Taipei, Taiwan, 20–24 April 2020; pp. 1400–1410. [Google Scholar]
  28. Tu, W.; Zhou, S.; Liu, X.; Guo, X.; Cai, Z.; Zhu, E.; Cheng, J. Deep fusion clustering network. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 9978–9987. [Google Scholar]
  29. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph convolutional networks: A comprehensive review. Comput. Soc. Netw. 2019, 6, 11. [Google Scholar] [CrossRef]
  30. Yu, P.; Tan, Z.; Lu, G.; Bao, B.K. Multi-view graph convolutional network for multimedia recommendation. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 6576–6585. [Google Scholar]
  31. Majumdar, A. Graph structured autoencoder. Neural Netw. 2018, 106, 271–280. [Google Scholar] [CrossRef]
  32. Tu, W.; Liao, Q.; Zhou, S.; Peng, X.; Ma, C.; Liu, Z.; Liu, X.; Cai, Z.; He, K. RARE: Robust Masked Graph Autoencoder. IEEE Trans. Knowl. Data Eng. 2024, 36, 5340–5353. [Google Scholar] [CrossRef]
  33. Jurek, A.; Bi, Y.; Wu, S.; Nugent, C. A survey of commonly used ensemble-based classification techniques. Knowl. Eng. Rev. 2014, 29, 551–581. [Google Scholar] [CrossRef]
  34. Vega-Pons, S.; Ruiz-Shulcloper, J. A survey of clustering ensemble algorithms. Int. J. Pattern Recognit. Artif. Intell. 2011, 25, 337–372. [Google Scholar] [CrossRef]
  35. Tao, Z.; Liu, H.; Li, S.; Fu, Y. Robust spectral ensemble clustering. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, Indianapolis, IN, USA, 24–28 October 2016; pp. 367–376. [Google Scholar]
  36. Liu, H.; Wu, J.; Liu, T.; Tao, D.; Fu, Y. Spectral ensemble clustering via weighted k-means: Theoretical and practical evidence. IEEE Trans. Knowl. Data Eng. 2017, 29, 1129–1143. [Google Scholar] [CrossRef]
  37. Zhou, J.; Zheng, H.; Pan, L. Ensemble clustering based on dense representation. Neurocomputing 2019, 357, 66–76. [Google Scholar] [CrossRef]
  38. Huang, D.; Wang, C.D.; Lai, J.H. Locally weighted ensemble clustering. IEEE Trans. Cybern. 2017, 48, 1460–1473. [Google Scholar] [CrossRef] [PubMed]
  39. Zhou, P.; Du, L.; Li, X. Adaptive consensus clustering for multiple k-means via base results refining. IEEE Trans. Knowl. Data Eng. 2023, 35, 10251–10264. [Google Scholar] [CrossRef]
  40. Zhou, P.; Liu, X.; Du, L.; Li, X. Self-paced adaptive bipartite graph learning for consensus clustering. ACM Trans. Knowl. Discov. Data 2023, 17, 62. [Google Scholar] [CrossRef]
  41. Zhou, P.; Hu, B.; Yan, D.; Du, L. Clustering ensemble via diffusion on adaptive multiplex. IEEE Trans. Knowl. Data Eng. 2024, 36, 1463–1474. [Google Scholar] [CrossRef]
  42. Ren, Y.; Pu, J.; Yang, Z.; Xu, J.; Li, G.; Pu, X.; Philip, S.Y.; He, L. Deep clustering: A comprehensive survey. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 5858–5878. [Google Scholar] [CrossRef]
  43. Chen, Y.; Wu, W.; Ou-Yang, L.; Wang, R.; Kwong, S. GRESS: Grouping Belief-Based Deep Contrastive Subspace Clustering. IEEE Trans. Cybern. 2024, 55, 148–160. [Google Scholar] [CrossRef]
  44. Ye, X.; Zhao, J.; Zhang, L.; Guo, L. A nonparametric deep generative model for multimanifold clustering. IEEE Trans. Cybern. 2019, 49, 2664–2677. [Google Scholar] [CrossRef]
  45. Chen, R.; Tang, Y.; Tian, L.; Zhang, C.; Zhang, W. Deep convolutional self-paced clustering. Appl. Intell. 2022, 52, 4858–4872. [Google Scholar]
  46. Cai, J.; Zhang, Y.; Wang, S.; Fan, J.; Guo, W. Wasserstein Embedding Learning for Deep Clustering: A Generative Approach. IEEE Trans. Multimed. 2024, 26, 7567–7580. [Google Scholar] [CrossRef]
  47. Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 478–487. [Google Scholar]
  48. Guo, X.; Gao, L.; Liu, X.; Yin, J. Improved deep embedded clustering with local structure preservation. In Proceedings of the International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 1753–1759. [Google Scholar]
  49. Guo, X.; Liu, X.; Zhu, E.; Yin, J. Deep clustering with convolutional autoencoders. In Proceedings of the 24th International Conference on Neural Information Processing, Guangzhou, China, 14–18 November 2017; pp. 373–382. [Google Scholar]
  50. Alqahtani, A.; Xie, X.; Deng, J.; Jones, M. A deep convolutional auto-encoder with embedded clustering. In Proceedings of the 25th IEEE International Conference on Image Processing, Athens, Greece, 7–10 October 2018; pp. 4058–4062. [Google Scholar]
  51. Liu, Y.; Tu, W.; Zhou, S.; Liu, X.; Song, L.; Yang, X.; Zhu, E. Deep graph clustering via dual correlation reduction. In Proceedings of the AAAI Conference on Artificial Intelligence, Tel-Aviv, Israel, 22 February–1 March 2022; Volume 36, No. 7. pp. 7603–7611. [Google Scholar]
  52. Akkem, Y.; Biswas, S.K.; Varanasi, A. A comprehensive review of synthetic data generation in smart farming by using variational autoencoder and generative adversarial network. Eng. Appl. Artif. Intell. 2024, 131, 107881. [Google Scholar] [CrossRef]
  53. Shirazi, H.; Muramudalige, S.R.; Ray, I.; Jayasumana, A.P.; Wang, H. Adversarial autoencoder data synthesis for enhancing machine learning-based phishing detection algorithms. IEEE Trans. Serv. Comput. 2023, 16, 2411–2422. [Google Scholar] [CrossRef]
  54. Hubert, L.; Arabie, P. Comparing partitions. J. Classif. 1985, 2, 193–218. [Google Scholar] [CrossRef]
  55. Pan, S.; Hu, R.; Fung, S.F.; Long, G.; Jiang, J.; Zhang, C. Learning graph embedding with adversarial training methods. IEEE Trans. Cybern. 2020, 50, 2475–2487. [Google Scholar] [CrossRef]
  56. Oyewole, G.J.; Thopil, G.A. Data clustering: Application and trends. Artif. Intell. Rev. 2023, 56, 6439–6475. [Google Scholar] [CrossRef] [PubMed]
  57. Jia, Y.; Liu, H.; Hou, J.; Zhang, Q. Clustering ensemble meets low-rank tensor approximation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, No. 9. pp. 7970–7978. [Google Scholar]
  58. Jia, Y.; Tao, S.; Wang, R.; Wang, Y. Ensemble clustering via co-association matrix self-enhancement. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11168–11179. [Google Scholar] [CrossRef] [PubMed]
  59. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of the GSAGCE framework. GSAGCE is divided into two components, including the global self-attention-driven graph clustering (GSAGC) module and the double-weighted clustering ensemble (DWCE) module. Among them, GSAGC generates base clustering results by fully utilizing and integrating node attribute information and graph structure information, while DWCE enhances the ensemble performance by considering global and local diversity in the base clusterings.
Figure 1. Illustration of the GSAGCE framework. GSAGCE is divided into two components, including the global self-attention-driven graph clustering (GSAGC) module and the double-weighted clustering ensemble (DWCE) module. Among them, GSAGC generates base clustering results by fully utilizing and integrating node attribute information and graph structure information, while DWCE enhances the ensemble performance by considering global and local diversity in the base clusterings.
Remotesensing 17 03680 g001
Figure 2. The ARI results of hyper-parameter sensitivity in Cora and ACM.
Figure 2. The ARI results of hyper-parameter sensitivity in Cora and ACM.
Remotesensing 17 03680 g002
Figure 3. The visualization results on ACM and HAR.
Figure 3. The visualization results on ACM and HAR.
Remotesensing 17 03680 g003
Table 1. The summary of the main notations and descriptions.
Table 1. The summary of the main notations and descriptions.
NotationsDescriptions
X R n × d The attribute feature data matrix
G = { V , E } The graph structure
K The number of clusters
V , E The node set and the edge set
A R n × n The original adjacency matrix
H ( i ) R n × d i The i-th layer encoder part of autoencoder
d i The dimension of the i-th layer network
H ^ ( i ) The i-th layer decoder part of autoencoder
O ( l ) The global self-attention representation
Z ( l ) The representation of GSAGAE
A ^ The reconstruction of graph adjacency structure
M ( i ) The attention coefficient matrix of FFGN
Z ˜ The fused representation of AE and GSAGAE
Z L The clustering probability distribution
q h i j The similarity between h i and cluster center μ j
p h i j The auxiliary distribution of H
q z i j The similarity between z i and cluster center v j
p z i j The auxiliary distribution of Z
Π = { π i } i = 1 M The clustering ensemble set
MThe number of base clusterings
π m The m-th base clustering
n m The number of clusters for π m
Sim ( π m , π n ) The similarity between π m and π n
GUE ( π m ) The global uncertainty estimation of π m
H m C i The uncertainty of cluster C i
HECE ( C i ) The hybrid ensemble-driven cluster estimation of C i
Table 2. Statistics of the datasets.
Table 2. Statistics of the datasets.
Data NameSampleFeatureCategoryTypeEdges
Cora270814337graph5278
ACM302518703graph13,128
Pubmed19,7175003graph44,326
IMDB478012322graph49,005
USPS929825610image-
HAR10,2995616record-
REUT10,00020004text-
Table 3. The experimental results of clustering ensemble comparative methods on seven datasets. The best results are highlighted in bold.
Table 3. The experimental results of clustering ensemble comparative methods on seven datasets. The best results are highlighted in bold.
MethodMetricsDataset
CoraACMPubmedIMDBUSPSHARREUT
RSECACC33.62 ± 3.2366.89 ± 1.8850.65 ± 7.4862.17 ± 5.8857.17 ± 4.6367.13 ± 4.9860.63 ± 6.40
NMI12.63 ± 2.1333.36 ± 1.1715.86 ± 7.161.20 ± 0.4660.51 ± 3.4067.38 ± 5.0140.44 ± 8.29
ARI8.88 ± 1.9830.93 ± 1.1113.20 ± 8.16−3.87 ± 0.8146.39 ± 5.3657.11 ± 5.5834.01 ± 13.09
F131.43 ± 3.4167.10 ± 2.0349.16 ± 11.4844.15 ± 0.8852.81 ± 4.6964.69 ± 7.0653.57 ± 7.23
SECWKACC34.95 ± 2.2858.91 ± 7.3751.86 ± 6.6574.79 ± 1.7853.52 ± 5.7949.01 ± 5.5767.80 ± 10.58
NMI14.95 ± 2.1822.33 ± 8.5921.11 ± 11.090.34 ± 0.2954.53 ± 4.8451.25 ± 8.7743.14 ± 11.62
ARI9.25 ± 2.0019.83 ± 8.2617.98 ± 11.63−0.68 ± 1.7535.47 ± 8.9432.44 ± 10.6040.76 ± 16.86
F129.76 ± 2.5755.79 ± 10.0143.28 ± 12.3744.64 ± 1.7851.22 ± 6.3245.66 ± 6.1954.57 ± 13.69
LWEAACC35.67 ± 2.9473.36 ± 2.6552.44 ± 4.1974.03 ± 0.1874.05 ± 2.3769.37 ± 6.2967.24 ± 7.85
NMI16.34 ± 3.0438.94 ± 2.6915.49 ± 5.320.45 ± 0.0575.67 ± 1.1574.91 ± 1.8848.54 ± 5.44
ARI8.62 ± 2.6037.52 ± 3.0614.44 ± 7.41−2.08 ± 0.1467.68 ± 2.5963.92 ± 3.5045.04 ± 10.72
F126.09 ± 5.7373.55 ± 2.5949.17 ± 6.2843.99 ± 0.0871.53 ± 2.7863.23 ± 8.5156.33 ± 8.82
LWGPACC32.95 ± 3.0665.57 ± 1.9446.08 ± 4.3174.18 ± 0.4173.44 ± 5.4665.41 ± 3.3955.18 ± 7.11
NMI15.13 ± 1.9832.69 ± 0.686.92 ± 2.920.36 ± 0.1475.84 ± 1.1569.61 ± 1.8139.73 ± 10.02
ARI7.60 ± 2.5230.07 ± 1.064.51 ± 4.72−1.76 ± 0.3867.48 ± 4.3659.27 ± 2.1627.73 ± 10.58
F125.69 ± 4.0865.73 ± 2.0238.03 ± 7.6644.21 ± 0.5970.80 ± 6.4559.76 ± 4.3141.14 ± 9.09
DRECACC36.91 ± 2.1669.64 ± 1.8547.29 ± 4.1174.02 ± 0.8868.13 ± 2.8571.93 ± 3.3974.46 ± 3.59
NMI16.77 ± 2.4235.43 ± 1.608.24 ± 3.930.50 ± 0.1768.75 ± 1.2368.89 ± 2.3247.97 ± 5.56
ARI11.83 ± 2.1033.31 ± 2.085.30 ± 5.22−2.13 ± 0.5857.28 ± 1.9859.45 ± 2.8550.65 ± 9.55
F134.79 ± 2.3769.93 ± 1.8237.38 ± 6.8143.88 ± 0.1766.67 ± 3.6670.79 ± 4.8562.31 ± 4.56
LRTCEACC29.03 ± 9.6868.40 ± 7.3159.54 ± 8.3957.59 ± 4.2865.56 ± 2.1656.91 ± 4.3968.17 ± 8.12
NMI9.72 ± 8.3434.24 ± 6.6131.19 ± 7.210.22 ± 0.3062.06 ± 0.8960.20 ± 0.7148.49 ± 8.88
ARI7.05 ± 6.1731.96 ± 6.4528.05 ± 8.67−1.04 ± 1.1753.58 ± 1.7144.22 ± 2.5942.27 ± 11.66
F127.39 ± 9.5068.78 ± 7.3358.21 ± 9.0347.32 ± 1.4363.42 ± 2.2655.00 ± 4.7058.62 ± 4.65
ECCMSACC30.29 ± 0.0335.04 ± 0.1940.93 ± 2.1073.85 ± 0.4146.72 ± 11.4939.50 ± 7.8843.27 ± 6.64
NMI0.48 ± 0.050.26 ± 0.231.38 ± 1.660.48 ± 0.2860.15 ± 12.2457.47 ± 6.3417.45 ± 7.20
ARI0.03 ± 0.030.02 ± 0.020.44 ± 0.80−2.07 ± 0.6740.02 ± 14.3936.44 ± 7.623.04 ± 9.72
F16.91 ± 0.0817.61 ± 0.2522.15 ± 3.7344.16 ± 0.7228.51 ± 11.6224.92 ± 9.2225.75 ± 7.87
ACMKACC18.47 ± 0.9536.67 ± 2.6839.89 ± 1.6251.14 ± 0.9463.03 ± 3.2847.32 ± 7.3135.76 ± 6.77
NMI1.18 ± 0.290.88 ± 1.442.51 ± 0.840.08 ± 0.0559.16 ± 3.3244.44 ± 7.728.71 ± 11.16
ARI0.45 ± 0.160.87 ± 1.541.92 ± 0.840.02 ± 0.1351.10 ± 3.2230.08 ± 8.526.50 ± 7.70
F117.59 ± 0.8136.61 ± 2.6339.94 ± 1.7047.14 ± 0.9860.49 ± 3.7545.84 ± 7.4332.03 ± 4.98
SCCABGACC30.79 ± 0.7535.09 ± 0.0540.24 ± 0.4573.10 ± 0.9037.72 ± 20.66-39.16 ± 0.01
NMI1.61 ± 1.420.15 ± 0.040.94 ± 0.800.15 ± 0.1435.84 ± 34.46-14.43 ± 1.46
ARI0.15 ± 0.150.31 ± 0.050.02 ± 0.10−0.44 ± 0.7525.95 ± 25.89-−2.06 ± 0.16
F18.20 ± 3.4117.39 ± 0.0720.26 ± 1.0743.53 ± 0.2124.60 ± 20.86-21.21 ± 0.16
CEAMACC28.37 ± 2.1559.45 ± 7.2141.06 ± 2.2769.28 ± 2.0247.65 ± 14.0955.41 ± 5.5948.03 ± 10.52
NMI10.56 ± 1.5221.72 ± 7.410.51 ± 0.550.68 ± 0.0350.67 ± 13.5759.27 ± 4.3127.68 ± 13.12
ARI5.96 ± 2.1920.96 ± 7.110.29 ± 0.60−3.62 ± 0.2336.24 ± 15.6744.01 ± 6.2915.79 ± 15.21
F123.02 ± 1.1159.42 ± 7.6322.48 ± 5.5745.01 ± 0.4738.23 ± 18.6051.28 ± 6.4835.95 ± 11.89
GSAGCEACC69.08 ± 2.2191.76 ± 1.0363.57 ± 3.2175.27 ± 2.3274.38 ± 2.2382.47 ± 1.1376.63 ± 1.86
NMI51.22 ± 1.4772.00 ± 0.8623.30 ± 2.464.27 ± 1.0275.89 ± 1.3682.91 ± 1.8550.26 ± 2.18
ARI45.46 ± 1.3678.49 ± 0.7023.28 ± 3.0113.71 ± 1.5467.85 ± 2.4074.89 ± 2.9153.89 ± 2.03
F165.44 ± 2.0491.77 ± 0.9263.79 ± 2.8958.56 ± 2.1172.72 ± 1.8781.78 ± 2.5865.38 ± 2.87
Table 4. The experimental results of deep clustering comparative methods on seven datasets. The best results are highlighted in bold.
Table 4. The experimental results of deep clustering comparative methods on seven datasets. The best results are highlighted in bold.
DatasetMetricsAE-kmeansDECIDECSDCNDFCNDCRNGSAGCE
CoraACC35.44 ± 2.1245.53 ± 2.2345.20 ± 1.7446.76 ± 6.0044.67 ± 4.8554.90 ± 8.3769.08 ± 2.21
NMI14.52 ± 2.0224.26 ± 1.9425.44 ± 1.8127.17 ± 4.4533.09 ± 7.3945.55 ± 6.4151.22 ± 1.47
ARI10.69 ± 1.7219.21 ± 1.8718.82 ± 2.0821.67 ± 4.6928.70 ± 5.2433.43 ± 7.6445.46 ± 1.36
F133.43 ± 2.0944.11 ± 2.3245.00 ± 1.7839.90 ± 5.6927.43 ± 4.9248.01 ± 9.0065.44 ± 2.04
ACMACC56.03 ± 6.5959.63 ± 6.7673.57 ± 7.4687.56 ± 1.4490.77 ± 0.2590.93 ± 0.4291.76 ± 1.03
NMI19.18 ± 3.1222.84 ± 4.7836.64 ± 8.0362.24 ± 2.7769.32 ± 0.6369.66 ± 0.9572.00 ± 0.86
ARI18.61 ± 3.7921.50 ± 5.3639.94 ± 10.1567.22 ± 3.1974.77 ± 0.6275.13 ± 1.0778.49 ± 0.70
F155.97 ± 6.7159.94 ± 6.8973.38 ± 7.7287.46 ± 1.5190.71 ± 0.2690.93 ± 0.3991.77 ± 0.92
PubmedACC53.42 ± 6.3257.48 ± 4.7357.11 ± 3.3957.70 ± 4.7750.67 ± 0.36oom63.57 ± 3.21
NMI17.17 ± 4.5020.46 ± 3.3819.73 ± 2.6717.72 ± 4.596.80 ± 0.20oom23.30 ± 2.46
ARI14.26 ± 4.9417.42 ± 4.2516.26 ± 3.1315.15 ± 5.036.75 ± 0.36oom23.28 ± 3.01
F153.62 ± 6.5457.94 ± 5.1757.38 ± 3.6258.06 ± 5.5041.23 ± 0.35oom63.79 ± 2.89
IMDBACC52.67 ± 1.5552.73 ± 1.3752.28 ± 1.9853.35 ± 2.4767.94 ± 9.6571.34 ± 0.6575.27 ± 2.32
NMI0.23 ± 0.270.88 ± 0.870.98 ± 1.101.94 ± 0.992.19 ± 1.520.43 ± 0.294.27 ± 1.02
ARI0.10 ± 0.37−0.33 ± 0.330.03 ± 0.72−0.02 ± 1.431.09 ± 2.29−2.58 ± 1.1513.71 ± 1.54
F147.76 ± 1.9347.86 ± 2.7248.10 ± 3.3450.27 ± 3.4848.98 ± 5.3445.44 ± 1.0158.56 ± 2.11
USPSACC65.90 ± 3.1766.97 ± 2.8071.25 ± 3.3072.74 ± 5.0873.51 ± 0.3023.43 ± 3.7274.38 ± 2.23
NMI63.65 ± 1.9865.07 ± 1.2673.61 ± 1.7075.78 ± 2.5375.49 ± 0.1816.32 ± 7.6375.89 ± 1.36
ARI55.39 ± 2.4756.87 ± 1.7464.17 ± 2.5767.54 ± 4.1067.45 ± 0.303.21 ± 5.6267.85 ± 2.40
F163.71 ± 4.2264.99 ± 3.6069.26 ± 4.8770.43 ± 6.6872.42 ± 0.3515.14 ± 0.4572.72 ± 1.87
HARACC62.94 ± 6.1962.63 ± 3.2370.91 ± 4.7962.66 ± 5.5277.26 ± 6.4441.41 ± 2.5182.47 ± 1.13
NMI57.41 ± 3.2162.14 ± 2.9977.00 ± 4.7267.62 ± 2.4281.09 ± 4.9351.27 ± 0.6482.91 ± 1.85
ARI49.71 ± 3.3753.25 ± 3.2766.75 ± 4.5953.26 ± 3.8171.29 ± 6.0530.65 ± 1.5274.89 ± 2.91
F160.53 ± 7.3360.82 ± 3.9968.12 ± 5.2754.17 ± 7.4876.34 ± 7.9134.17 ± 2.4381.78 ± 2.58
REUTACC56.69 ± 3.9856.39 ± 2.2359.72 ± 2.8061.33 ± 6.0963.82 ± 5.2350.34 ± 4.6176.63 ± 1.86
NMI27.77 ± 4.1228.24 ± 3.4534.54 ± 2.8937.21 ± 10.3941.23 ± 4.3622.56 ± 8.5150.26 ± 2.18
ARI26.40 ± 5.0126.49 ± 3.3930.63 ± 3.3336.20 ± 11.1142.52 ± 6.7215.99 ± 6.4053.89 ± 2.03
F151.79 ± 4.2751.51 ± 2.4954.26 ± 3.7354.07 ± 9.1457.86 ± 6.1833.31 ± 3.8165.38 ± 2.87
Table 5. The ablation experiment results on seven datasets. The best results are highlighted in bold.
Table 5. The ablation experiment results on seven datasets. The best results are highlighted in bold.
DatasetMetricsnoGSAnoKLhnoKLznoHECEGSAGCE
CoraACC66.84 ± 1.7266.46 ± 1.4647.41 ± 5.7367.14 ± 1.8969.08 ± 2.21
NMI47.45 ± 1.6346.15 ± 1.4428.72 ± 4.6450.41 ± 2.0151.22 ± 1.47
ARI42.69 ± 1.2443.15 ± 2.1620.35 ± 5.7843.84 ± 1.5645.46 ± 1.36
F162.09 ± 0.5361.20 ± 1.2435.86 ± 5.2064.81 ± 2.4765.44 ± 2.04
ACMACC91.21 ± 0.8365.78 ± 1.7675.30 ± 6.3291.23 ± 0.8691.76 ± 1.03
NMI69.84 ± 2.0428.94 ± 5.0447.22 ± 10.0170.58 ± 0.6372.00 ± 0.86
ARI75.90 ± 2.0325.66 ± 3.9148.34 ± 11.0376.82 ± 1.0578.49 ± 0.70
F191.26 ± 0.8666.61 ± 1.7072.62 ± 7.7091.20 ± 1.1491.77 ± 0.92
PubmedACC61.58 ± 2.8660.17 ± 3.0552.05 ± 2.4160.32 ± 2.6863.57 ± 3.21
NMI20.00 ± 2.1212.42 ± 2.2416.42 ± 1.7920.16 ± 2.4323.30 ± 2.46
ARI18.69 ± 2.6514.96 ± 3.1417.75 ± 2.8620.16 ± 2.7923.28 ± 3.01
F161.80 ± 2.2555.13 ± 3.3948.11 ± 2.9160.58 ± 2.9363.79 ± 2.89
IMDBACC71.69 ± 0.2372.35 ± 2.3366.78 ± 6.8774.30 ± 2.4675.27 ± 2.32
NMI3.44 ± 0.233.72 ± 1.590.37 ± 2.723.94 ± 0.674.27 ± 1.02
ARI9.98 ± 6.449.55 ± 3.790.91 ± 4.8011.89 ± 1.9413.71 ± 1.54
F155.31 ± 7.2554.28 ± 4.2043.31 ± 7.2257.86 ± 2.3358.56 ± 2.11
USPSACC72.45 ± 3.5168.83 ± 6.8753.37 ± 4.3373.83 ± 2.1074.38 ± 2.23
NMI72.08 ± 6.0265.78 ± 2.7256.29 ± 4.9475.32 ± 1.1775.89 ± 1.36
ARI63.10 ± 5.5757.10 ± 4.8039.17 ± 6.6567.51 ± 2.3267.85 ± 2.40
F165.79 ± 6.6867.27 ± 7.2243.36 ± 6.1472.12 ± 2.0672.72 ± 1.87
HARACC80.32 ± 1.2778.66 ± 1.5163.06 ± 4.8581.36 ± 0.6882.47 ± 1.13
NMI80.88 ± 1.5371.06 ± 1.4961.99 ± 3.4281.76 ± 1.2082.91 ± 1.85
ARI72.36 ± 3.0663.38 ± 2.8347.41 ± 4.5872.54 ± 2.8774.89 ± 2.91
F180.42 ± 3.2277.88 ± 2.5355.16 ± 5.5780.64 ± 3.0581.78 ± 2.58
REUTACC49.85 ± 1.4739.36 ± 1.8247.54 ± 2.2575.31 ± 2.2876.63 ± 1.86
NMI14.33 ± 1.293.10 ± 3.553.66 ± 2.5648.88 ± 2.5350.26 ± 2.18
ARI7.55 ± 1.571.23 ± 2.617.37 ± 3.0052.79 ± 2.7653.89 ± 2.03
F132.65 ± 2.5126.79 ± 4.9632.02 ± 6.1464.90 ± 3.0565.38 ± 2.87
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, L.; Yao, S.; Huang, Y.; Xiao, L.; Cheng, Y.; Qian, Y. Global Self-Attention-Driven Graph Clustering Ensemble. Remote Sens. 2025, 17, 3680. https://doi.org/10.3390/rs17223680

AMA Style

Zeng L, Yao S, Huang Y, Xiao L, Cheng Y, Qian Y. Global Self-Attention-Driven Graph Clustering Ensemble. Remote Sensing. 2025; 17(22):3680. https://doi.org/10.3390/rs17223680

Chicago/Turabian Style

Zeng, Lingbin, Shixin Yao, You Huang, Liquan Xiao, Yong Cheng, and Yue Qian. 2025. "Global Self-Attention-Driven Graph Clustering Ensemble" Remote Sensing 17, no. 22: 3680. https://doi.org/10.3390/rs17223680

APA Style

Zeng, L., Yao, S., Huang, Y., Xiao, L., Cheng, Y., & Qian, Y. (2025). Global Self-Attention-Driven Graph Clustering Ensemble. Remote Sensing, 17(22), 3680. https://doi.org/10.3390/rs17223680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop