You are currently viewing a new version of our website. To view the old version click .
Remote Sensing
  • Article
  • Open Access

9 December 2025

Discriminative Anchor Learning for Hyperspectral Image Clustering

,
,
,
and
1
State Key Laboratory of Integrated Services Networks, Xidian University, Xi’an 710071, China
2
Rocket Force University of Engineering, Xi’an 710025, China
3
Cognitive Computing and Interactive Systems Laboratory, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.

Highlights

What are the main findings?
  • An anchor learning strategy that explicitly accounts for the distribution consistency between anchors and samples is proposed, and a discriminative anchor-based hyperspectral image clustering algorithm is developed accordingly.
  • Imposing low-rank and probabilistic constraints on the consensus coefficient matrix excavates the intrinsic structure of anchors and further enhances their discriminative capability.
What is the implication of the main finding?
  • The new anchor strategy provides an effective tool for anchor-based clustering algorithm.
  • The proposed algorithm significantly improves the accuracy of hyperspectral image clustering and yields high-quality clustering maps.

Abstract

Anchor-based clustering algorithms for hyperspectral image (HSI) have alleviated the computational burden and become a prominent research direction in remote sensing. The performance results of these methods are heavily influenced by the quality of the anchors. However, existing anchor-based methods directly utilize the learned anchors for clustering, ignoring the distribution consistency between anchors and samples. This leads to the degradation of anchor quality and suboptimal clustering performance. To address the problem, we propose a discriminative anchor-based hyperspectral image clustering algorithm. Specifically, by sharing the coefficient matrix among anchors and that between samples and anchors, the proposed method explicitly takes into account the distribution of samples and anchors, enabling the learned anchors to better describe the samples. We further impose low-rank and probabilistic constraints on the consensus coefficient matrix, which effectively captures latent structural information and enhances the anchors’ discriminative ability. Extensive experimental results on public HSI benchmarks demonstrate the superiority and effectiveness of the proposed method.

1. Introduction

With hundreds of spectral bands and abundant spatial information, HSI can better identify land-cover materials for real-word applications such as vegetation surveys, environmental monitoring, resource exploration and military reconnaissance [1,2]. Among these diverse applications, HSI classification remains the most fundamental task. Due to the high dimensionality of HSI, the analysis task usually requires a large number of high-quality labeled samples to avoid the Hughes phenomenon and underfitting problems. However, manual annotation directly on HSI data is inefficient and error prone, which severely restricts the application of hyperspectral remote sensing. Faced with the dilemma of label acquisition, clustering analysis that does not rely on any labeled samples has become a research hotspot.
As an unsupervised approach, HSI clustering partitions pixels into groups by exploring their intrinsic properties, ensuring that pixels within the same group have high similarity, while those between different groups show significant differences. Existing clustering approaches for HSI can be roughly classified as shallow-based methods and deep learning-based methods. Shallow-based methods mainly include centroid-based [3,4,5], density-based [6,7,8], graph-based [9] and subspace-based methods [10]. These techniques analyze the features of HSI or optimize specific objective functions to attain the clustering labels. Deep learning-based methods involve self-representation-based [11,12,13], autoencoder-based [14,15], contrastive learning-based [16,17] and transformer-based methods [18,19]. Benefiting from the strong nonlinear feature extraction, deep clustering algorithms typically achieve superior performance. However, they lack interpretability and require powerful GPU support, which limits their practical applications. Consequently, this paper focuses on shallow models for hyperspectral image clustering.
Among shallow-based methods, subspace-based clustering methods have developed rapidly and achieved success in HSI analysis. A representative method is sparse subspace clustering (SSC) [20], which learns sparse representations in a self-expression model, effectively mining the structural information of HSI. Although SSC has achieved good clustering results, it only uses the spectral features to group pixels and ignores the rich spatial information. As a result, the discriminative power of features is significantly constrained by spectral distinguishability, which limits the algorithm’s performance. To alleviate this problem, Zhang et al. [21] and Zhai et al. [22] incorporated spatial information into a sparse self-representation framework to learn the similarity matrix. Cai et al. [23] combined graph convolution and a self-representation framework to calculate the coefficient matrix, and proposed two subspace clustering models in original feature space and kernel space. These algorithms regard the entire image as a dictionary to model the low-dimensional subspaces. Nevertheless, the redundant information among the spectral features of HSIs weakens the dictionary’s expressive ability, and the construction of a coefficient matrix between all pixels usually requires quadratic computational complexity with the size of pixels, which is prohibitive for the large-scale HSI.
To tackle this problem, more efficient dictionary learning algorithms have been proposed [24,25,26]. Instead of using the entire HSI data as a dictionary, Huang et al. learned a smaller dictionary based on the random matrix and presented a large-scale SSC method [27]. Since the corresponding coefficients under the small-scale dictionary are diminished, the computational cost is accordingly lowered. To mitigate noise and spectral variability, it employs a total variation constraint on the coefficient matrix to consider the spatial dependence between adjacent pixels. In addition, Zhai et al. introduced a sketched reweighted sparse and low-rank subspace clustering, which effectively incorporated non-local spatial information through non-local mean regularization [28]. These algorithms have improved computational efficiency, but they select M anchors ( M N ) from N pixels via random sampling, which fails to guarantee the quality of anchors, resulting in unstable clustering performance. More importantly, such anchor selection strategies, inherited from machine learning, treat each pixel as an independent sample. During the selection process, they solely focus on the spectral features of pixels while ignoring the crucial spatial correlation in HSI, failing to select the satisfactory high-quality anchors. To incorporate the spectral–spatial characteristics of the pixels, superpixel-based anchor selection methods have gradually emerged [29,30,31]. Most of these methods generate homogeneous regions via superpixel segmentation and then produce representative points for these regions as anchors by weighted averaging all pixels within the same superpixel. While this strategy enhances clustering performance by incorporating spatial information, it disregards the distribution consistency between anchors and samples, resulting in the poor discriminability of anchors. As illustrated in Figure 1, the distribution consistency between anchors and samples critically affects the anchors’ representational quality. If anchor distributions deviate from pixel distributions, the anchors fail to represent samples effectively, which degrades anchor graph quality and further reduces the clustering performance.
Figure 1. Schematic illustration of anchors and pixels distributions. Taking three clusters as an example, the dots and stars represent pixels and anchors, respectively, with different colored dots denoting distinct clusters. In Figure 1a, without considering the distribution of anchors and pixels, the anchors are affected by outliers and deviate from the pixels distribution, making it difficult to accurately represent the pixels. In Figure 1b, the anchors maintain the distribution of the pixels and better represent the pixels under the constraint of distributed consistency. (a) Inconsistency. (b) Consistency.
To combat the aforementioned issue, we propose a discriminative anchor learning method for hyperspectral image clustering. Within a self-expressive framework, our model explicitly aligns the anchors distribution with the pixels distribution by sharing the same coefficient matrix, enabling the learned anchors to better represent the pixels. We impose the low-rank regularization constraint on the consensus coefficient matrix to reveal the underlying low-dimensional subspace structure. Different from the classic subspace clustering algorithms that enforce the diagonal elements of the coefficient matrix which are zero and where the row sum is one, we apply non-negative and row-sum-to-one constraints on the coefficient matrix. The non-negativity constraint ensures positive contribution between anchors, eliminating interference from negative values in similarity judgments and enhancing interpretability. The row-sum-to-one constraint normalizes the relationships between each anchor and others. Under these dual constraints, the coefficients corresponding to intra-class anchors are more concentrated, while those for inter-class anchors are sparser, which strengthens the anchors’ discriminative ability. Using the anchor graph constructed from discriminative anchors can achieve superior clustering performance.
The primary contributions of this paper are as follows:
  • We propose a discriminative anchor learning strategy to capture the intrinsic structure of anchors. By sharing the coefficient matrix, this strategy aligns the anchors distribution with the pixels distribution, facilitating the effective clustering.
  • We impose low-rank and probabilistic constraints on the consensus coefficient matrix, which not only excavates the global structures but also enhances the discriminative capability of the anchors.
  • We present an alternating optimization strategy to solve the proposed formulation. Extensive experimental results demonstrate the superiority and effectiveness of the proposed method.

3. Methodology

3.1. Problem Formulation and Objective Function

For anchor-based clustering algorithms, the selection of high-quality anchors is a pivotal challenge. Although the existing methods have achieved good performance, they directly utilize the learned anchors for clustering, overlooking the distribution consistency between anchors and samples. To address this problem, our approach integrates spectral, spatial, and distribution information to learn high-quality anchors. Then, the pixel-anchor graph constructed from discriminative anchors is used for clustering.
Superpixel Generation. Generally speaking, pixels belonging to the same class are mostly distributed within homogeneous regions. Entropy rate superpixel (ERS) segmentation [41] techniques can adaptively generate such homogeneous regions based on the spatial distribution and texture information of pixels, offering a viable approach to extract spectral–spatial structures. ERS is initially designed for RGB image segmentation, which requires converting the RGB image into a grayscale image and then segmenting the grayscale image. Given that hyperspectral remote sensing images comprise numerous spectral bands, we need to select a representative component to perform segmentation. Principal component analysis (PCA) is a common dimension reduction technique and is widely employed for data preprocessing in ERS segmentation. In detail, we adopt the max-min normalization to standardize the features to eliminate the impact of feature magnitude differences on PCA. Only the first principal component I f (PCA dimensionality = 1) is retained for executing ERS to obtain M non-overlapping superpixels as
I f = i = 1 M P i , s . t . P i P j = ,
where P i is the i-th superpixel. By mapping the segmentation results back to the original HSI, we obtain the corresponding superpixels across all spectral bands.
Each superpixel P i contains N i pixels, denoted as i = p 1 i , p 2 i , , p N i i R B × N i . The i-th anchor is computed by averaging the pixels within the same superpixel, i.e.,
a i = 1 N i j = 1 N i p j i , i = 1 , 2 , , M ,
where N = i = 1 M N i . Based on M superpixels, we derive the initialized anchor matrix A = a 1 , a 2 , , a M R B × M .
Spatial Denoising. In the acquisition of hyperspectral images, noise and outliers are inevitably introduced, which degrades image quality and affects subsequent clustering tasks. To mitigate the influence of noise and outliers, most existing algorithms augment spatial information by implementing mean filtering within small square windows. Nevertheless, such an approach is too strict to effectively extract complex spatial features. Considering that superpixel segmentation can generate homogeneous regions, conducting denoising operations within the same superpixel can result in the desired denoised pixels. Given an original pixel x ˜ i R B × 1 within X ˜ , we identify its K nearest spatial neighbors within the corresponding superpixel using the Euclidean distance and define them as Q i = { q 1 , q 2 , , q K } R B × K . To accomplish effective denoising, we denoise x ˜ i by employing the weighted sum of K nearest neighbors, and the formula is expressed as
x i = j = 1 K w j × q j , i = 1 , 2 , , N ,
where x i is the denoised pixel of x ˜ i . The weight of the spectral similarity between x ˜ i and q j is denoted by
w j = exp x ˜ i q j 2 2 / 2 t 2 h , j = 1 , 2 , , K ,
where t = 1 K j = 1 K x ˜ i q j 2 2 and h = j = 1 K exp x ˜ i q j 2 2 / 2 t 2 . By applying the above operation to all pixels, we can obtain the denoised image X = { x 1 , x 2 , , x N } R B × N . Through noise suppression, we strengthen the spatial coherence of the image while boosting the representativeness of the data.
Discriminative Anchor Learning. The initialized anchor matrix A incorporates the spatial consistency of the image, yet fails to account for the distribution consistency between anchors and pixels. Ideally, the anchors A have a consensus distribution structure with the denoised pixels X . If a pixel exhibits high affinity with certain anchors, these anchors have a high probability of belonging to the same cluster and the corresponding affinities are likely to be high as well. To achieve this, we jointly learn anchor matrix and anchor graph within a self-expression framework, allowing these two matrices to share the same coefficient representation for aligning the anchors with the pixels distribution. The loss function is
min A , S , Z X A S T F 2 + λ 1 A A Z F 2 + λ 2 S S Z F 2 + λ 3 R ( Z ) , s . t . S 0 , S · 1 = 1 . Z 0 , Z · 1 = 1 .
where S R N × M is defined as an anchor graph. R ( Z ) is the regularization term, λ 1 , λ 2 and λ 3 are trade-off parameters. Since similar pixels usually produce similar representation coefficients, dissimilar pixels yield different representation coefficients, we can regard S as the extracted features of input data X . Under the self-representation model, we learn the coefficient matrix Z corresponding to the anchor matrix A , which contains the anchor–anchor relationships. We further decompose the anchor graph S into the anchor graph S and the anchor coefficient matrix Z . Through the consistent coefficient matrix Z , we enhance the consistency between the affinity relationships among anchors and those between pixels and anchors, promoting that the anchor distribution is consistent with the pixels distribution. Building on the superpixel-averaged anchors, we iteratively update the anchor matrix through adaptive learning, fostering greater alignment between the distribution of anchors and pixels. This approach effectively mitigates the potential errors arising from initial superpixel segmentation, thereby enhancing the quality of the anchors.
To explore the global low-rank structure, we impose a nuclear norm constraint on the coefficient matrix Z , which encourages a more desirable block structure. Moreover, by constraining the coefficient matrix Z to be non-negative with row sums equal to one, the coefficient representations of intra-class anchors are more compact, while those of inter-class anchors are more dispersed. This further enhances the anchors’ discriminative capability. The final objective function is as follows:
min A , S , Z X A S T F 2 + λ 1 A A Z F 2 + λ 2 S S Z F 2 + λ 3 Z * , s . t . S 0 , S · 1 = 1 , Z 0 , Z · 1 = 1 .
where · * is a nuclear norm that represents the sum of all non-zero singular values in the matrix.

3.2. Optimization

It is challenging to directly address the model (6) when all variables are considered at once. Based on Augmented Lagrange Multiplier [42], we introduce the following auxiliary variables D , E and F . By setting E 0 , E · 1 = 1 and F 0 , F · 1 = 1 , the model (6) can be rewritten as
L ( A , S , Z , D , E , F ) = min X A S T F 2 + λ 1 A A Z F 2 + λ 2 S S Z F 2 + λ 3 D * + μ 1 2 Z D + Y 1 μ 1 F 2 + μ 2 2 S E + Y 2 μ 2 F 2 + μ 3 2 Z F + Y 3 μ 3 F 2 , s . t . E 0 , E · 1 = 1 , F 0 , F · 1 = 1 .
where Y 1 , Y 2 , and Y 3 represent Lagrange multipliers; μ 1 , μ 2 , and μ 3 are the penalty parameters. The problem (7) can be solved by an alternating optimization method.
A -subproblem. Fixing the variables S , Z , D , E and F . In this condition, the optimization of A is
min A X A S T F 2 + λ 1 A A Z F 2 .
Taking the derivative of Equation (8) w.r.t A and setting it to zero, A can be updated by
A * = X S ( S T S + λ 1 I m λ 1 ( Z + Z T ) + λ 1 Z Z T ) 1 .
S -subproblem. Fixing the variables A , Z , D , E and F . In this condition, the optimization of S is
min S X A S T F 2 + λ 2 S S Z F 2 + μ 2 2 S E + Y 2 μ 2 F 2 .
Taking the derivative of Equation (10) w.r.t S and setting it to zero, S can be updated by
S * = ( 2 X T A + μ 2 ( E Y 2 μ 2 ) ) ( 2 A T A + 2 λ 2 I m 2 λ 2 ( Z + Z T ) + 2 λ 2 Z Z T + μ 2 I m ) 1 .
Z -subproblem. Fixing the variables A , S , D , E and F . In this condition, the optimization of Z is
min Z λ 1 A A Z F 2 + λ 2 S S Z F 2 + μ 1 2 Z D + Y 1 μ 1 F 2 + μ 3 2 Z F + Y 3 μ 3 F 2 .
Taking the derivative of Equation (12) w.r.t Z and setting it to zero, Z can be updated by
Z * = ( 2 λ 1 A T A + 2 λ 2 S T S + ( μ 1 + μ 3 ) I m ) 1 ( 2 λ 1 A T A + 2 λ 2 S T S + μ 1 ( D Y 1 μ 1 ) + μ 3 ( F Y 3 μ 3 ) ) .
D -subproblem. Fixing the variables A , S , Z , E and F . In this condition, the optimization of D is
min D λ 3 D * + μ 1 2 Z D + Y 1 μ 1 F 2 = min D λ 3 μ 1 D * + 1 2 D ( Z + Y 1 μ 1 ) F 2 .
In order to solve the issue, we present the following theorem.
Theorem 1 
([43]). Suppose an arbitrary matrix Y R m × n and it is decomposed into Y = U Y D Y V Y T , τ > 0 , l = min ( m , n ) , 0 ω 1 ω 2 ω l . For the following optimization problem
arg min X 1 2 X Y F 2 + τ X * ,
the optimal solution is
X * = Υ τ Y = U Y S τ Y V Y T
where S τ Y = d i a g ( γ 1 , γ 2 , , γ l ) , which can be calculated by the soft-thresholding method [44].
According to Theorem 1, D can be updated by
D * = Υ λ 3 μ 1 ( Z + Y 1 μ 1 ) .
E -subproblem. Fixing the variables A , S , Z , D and F . In this condition, the optimization of E is
min E μ 2 2 S E + Y 2 μ 2 F 2 = min E μ 2 2 E K μ 2 F 2 , s . t . E 0 , E · 1 = 1 .
where K = μ 2 S + Y 2 . Based on [45], the closed-form solution of E is
e i * = ( K i μ 2 + θ 1 ) + ,
where ( · ) + = max ( · , 0 ) and θ stands for the Lagrangian multiplier.
F -subproblem. Fixing the variables A , S , Z , D and E . In this condition, the optimization of F is
min F μ 3 2 Z F + Y 3 μ 3 F 2 = min F μ 3 2 F T μ 3 F 2 , s . t . F 0 , F · 1 = 1 .
where T = μ 3 Z + Y 3 . Based on [45], the closed-form solution of F is
f i * = ( T i μ 3 + γ 1 ) + ,
where γ is the Lagrangian multiplier.
Furthermore, the Lagrange multipliers Y 1 , Y 2 , and Y 3 can be updated as
Y 1 = Y 1 + μ 1 ( Z D ) , Y 2 = Y 2 + μ 2 ( S E ) , Y 3 = Y 3 + μ 3 ( Z F ) .
The penalty parameters μ i are updated by
μ i = min ( μ i × 1.1 , μ max ) , i = 1 , 2 , 3 .
where μ max represents the given constant.
Ultimately, we encapsulate the pseudo code in Algorithm 1.
Algorithm 1: Discriminative anchor learning for hyperspectral image clustering.
Input: HSI matrix X R B × N , number of superpixels M, parameters λ 1 , λ 2 and λ 3 .
Output: Clustering labels.
   1:
Perform PCA and ERS to generate HSI superpixels.
   2:
Initialize A by using Equation (2). D = E = F = Y 1 = Y 2 = Y 3 = 0 , μ i = 10 3 ( i = 1 , 2 , 3 ) , μ max = 10 9 , ε = 10 6 .
   3:
while not converged do
   4:
   Update A by using Equation (9).
   5:
   Update S by using Equation (11).
   6:
   Update Z by using Equation (13).
   7:
   Update D by using Equation (17).
   8:
   Update E by using Equation (19).
   9:
   Update F by using Equation (21).
 10:
   Update Y 1 , Y 2 and Y 3 by using Equation (22).
 11:
   Update the penalty parameters μ i by using Equation (23).
 12:
   Check the convergence conditions:
  • Z D < ε & S E < ε & Z F < ε .
 13:
   Use the K-means algorithm on S to determine the clustering labels.
 14:
end while
 15:
return The final clustering results.

3.3. Complexity Analysis

Computational Complexity. The computational consumption of the proposed algorithm comprises two phases: HSI segmentation and optimization by iteratively solving Equation (6). To achieve HSI segmentation, it requires O ( B 2 N + B 3 ) for performing PCA and takes O ( N l o g ( N ) ) for conducting ERS. Subsequently, we concentrate on optimizing six variables, i.e., A , S , Z , D , E and F . To optimize the A and S subproblems, they take O ( B M N + M 2 N ) at each iteration. As for the Z and D subproblems, they need O ( M 2 N + M 3 + M 2 B ) and O ( M 3 ) at each iteration, respectively. Updating E and F take O ( N M ) , and O ( M 2 ) , respectively. Since M , B N , the primary computational complexity of the proposed algorithm in each iteration is O ( B 2 N + N l o g ( N ) + B M N + M 2 N ) .
Storage Complexity. Throughout the optimization process, the storage consumption of the proposed method mainly originates from A , S , Z , D , E and F . For these eight variables, their corresponding storage complexities are O ( B N + N M ) , O ( B N + N M ) , O ( N M + M 2 ) , O ( M 2 ) , O ( N M ) and O ( M 2 ) . Thus, the storage complexity of the proposed algorithm is O ( B N + N M + M 2 ) .

4. Experiments

4.1. Experiment Settings

Datasets. To evaluate our proposed method, we conducted extensive experiments on three commonly used HSI datasets, including Salinas, Pavia University (PaviaU), and Pavia Center (PaviaC) [29]. The detailed statistical information of these datasets is presented in Table 2.
Table 2. HSI datasets utilized in this paper.
Baselines. To demonstrate the effectiveness of our proposed method, we compare it with several representative clustering methods. FSCAG [33]: Constructing an anchor graph separately based on the original pixels and denoised pixels via mean filtering and performing fast spectral clustering to get the clustering labels. HESSC [46]: Automatically estimating the number of clusters and executing hierarchical sparse subspace clustering. NCSC [36]: Designing a superpixel pooling autoencoder to learn superpixel-level subspace representations and employing neighborhood contrastive regularization to enhance the robustness of subspace representations. S3AGC [35]: Building an anchor graph using denoised pixels and band structure information, followed by spectral clustering on similarity matrix to obtain the clustering results. BGPC [29]: Combining projection learning and bipartite graph learning with Laplacian rank constraint into a unified framework to directly obtain the clustering outcomes. SDST [19]: Designing a self-supervised double-structure transformer that contains a shared autoformer and Siamese dual-former graph module to enhance the representation performance of features for clustering. SSGCC [17]: Designing a spatial–spectral graph contrastive clustering that dynamically set the weights of sample pairs, enabling the model to prioritize handling hard samples. SAPC [39]: Learning a pixel-anchor graph and an anchor–anchor graph with Laplacian rank constraint in the projection space and designing an inference strategy to derive the clustering labels. HPCDL [40]: Imposing doubly stochastic constraints on an anchor–anchor graph and producing a better graph with strict probabilistic affinity. For a more intuitive comparison, we summarized the above algorithm from four aspects: parameter sharing, stochastic, regularization, and end-to-end, as shown in Table 3.
Table 3. Mechanism comparison table of different compared methods.
Experimental Settings.
To ensure an equitable comparison between the proposed method and the baselines, NCSC, SDST, and SSGCC are implemented on a standard Windows 10 Server with two Intel(R) Xeon(R) Platinum 8168 CPUs 2.7 GHz and 384 GB RAM. The remaining experiments are conducted using MATLAB 2020b. To quantify the clustering performance, we adopt five common metrics in HSI clustering methods, including overall clustering accuracy (OA), average clustering accuracy (AA), Kappa coefficient, normalized mutual information (NMI), and Purity. Higher values in all metrics reflect better clustering performance.
To determine the homogeneous regions, we adaptively set the number of superpixels by
M = δ N edge N ,
where N edge denotes the number of nonzero elements in the detected edge image, and δ is an adjustable parameter whose default is 2000. M is finely adjusted within the range of [100 200] to obtain a more suitable homogeneous area. To achieve optimal clustering results, we employ a grid search strategy to fine-tune the parameters included in the proposed model. Specifically, we adjust the parameters λ 1 , λ 2 and λ 3 in the range of [0.01 100]. Each experiment is repeated 10 times, and the outcomes are presented as an average.

4.2. Experimental Results and Analysis

Performance Comparison. Table 4 presents the experimental results of different clustering methods on three HSI datasets. Bold and underline are used to highlight the best and second-best results for all methods, respectively. By observing and analyzing the extensive experimental results, it can be seen that our method has achieved better clustering performance compared with most comparative algorithms. Relative to the corresponding second-best baseline, our algorithm exhibits the obvious improvement of 8.4%, 2.6%, and 7.1% in OA, AA, and Kappa on the Salinas dataset. And our approach demonstrates the apparent promotion of 8.3%, 11.9%, and 17.8% in OA, AA and Kappa over the second-best baseline on the PaviaU dataset.
Table 4. Clustering performance on the Salinas, PaviaU, and PaviaC datasets.
To be specific, HESSC yields the poorest clustering results among all the comparative algorithms. This is because it only relies on spectral information for clustering and ignores the spatial information that is crucial for HSI clustering. By incorporating spatial information, FSCAG and S3AGC achieve better clustering performance compared to HESSC. For deep learning algorithms, NCSC, SDST and SSGCC have achieved better feature representation due to their complex network design, gaining promising clustering performance. Among these methods, constrained by the pixel–superpixel relationships, the superpixel-based method NCSC exhibits inferior clustering performance. Compared with shallow methods, these deep algorithms have poorer interpretability and higher computational complexity. In contrast to the aforementioned baselines, BGPC, SAPC, and HPCDL demonstrate superior clustering performance. This is primarily because BGPC can directly obtain pixel labels in one step, whereas SAPC and HPCDL can propagate anchor labels to pixels utilizing graph-based mechanisms, effectively realizing clustering tasks. By contrast, our proposed algorithm yields the optimal clustering performance. This superiority can be attributed to the consideration of anchor distribution during the anchor selection process. When the anchor distribution aligns with that of the pixels, it helps to generate more discriminative anchors. Remarkably, even when post-processing techniques such as K-means are applied to attain clustering labels, it still maintains excellent clustering performance. To visualize the clustering effect, the clustering maps obtained by different methods are presented in Figure 2, Figure 3 and Figure 4. From these figures, it can be observed that the clustering maps produced by the proposed algorithm are close to the ground truth maps. This further substantiates the superiority of our proposed algorithm.
Figure 2. Clustering maps on the Salinas dataset. (a) GT. (b) FSCAG. (c) HESSC. (d) NCSC. (e) S3AGC. (f) BGPC. (g) SDST. (h) SSGCC. (i) SAPC. (j) HPCDL. (k) OUR.
Figure 3. Clustering maps on the PaviaU dataset. (a) GT. (b) FSCAG. (c) HESSC. (d) NCSC. (e) S3AGC. (f) BGPC. (g) SDST. (h) SSGCC. (i) SAPC. (j) HPCDL. (k) OUR.
Figure 4. Clustering maps on the PaviaC dataset. (a) GT. (b) FSCAG. (c) HESSC. (d) NCSC. (e) S3AGC. (f) BGPC. (g) SDST. (h) SSGCC. (i) SAPC. (j) HPCDL. (k) OUR.
t-SNE visualization. Figure 5 illustrates the clustering results visualized with t-distributed stochastic neighbor embedding (t-SNE) on the three HSI datasets. Each pixel is displayed using t-SNE, and distinct colors represent different cluster labels derived from our method. It can be observed that the proposed method exhibits better separability. Pixels within the same cluster are more aggregated, while those from different clusters are further dispersed, obtaining more discriminative representations.
Figure 5. Visualization of the clustering results with t-SNE on the three datasets. (a) Salinas. (b) PaviaU. (c) PaviaC.
Parameter Study.
  • Parameter M. We investigate the effect of M for our algorithm and record the clustering performance results (OA, NMI, and Kappa) as shown in Figure 6. Overall, the proposed method reaches the optimal outcomes when M is set to [120 200]. Furthermore, we can discover that these metrics curves do not monotonically increase with M. This implies that it is unnecessary to employ numerous anchors for clustering.
  • Parameters λ 1 , λ 2 and λ 3 . To facilitate observation, we analyze the effect of the parameters λ 1 and λ 2 for our method when λ 3 is fixed. The results are shown in the first row of Figure 7. It can be observed that as λ 1 and λ 2 are varied, OA exhibits significant fluctuations, indicating that these parameters need to be carefully adjusted to optimize the clustering performances. Satisfactory clustering results are typically achieved when λ 1 and λ 2 are set within the range [0.1, 10]. The second row of Figure 7 shows the variation curve corresponding to λ 3 . Obviously, the clustering results are notably impacted by these parameters, especially for the Salinas and PaviaU datasets. Because of the discrepancy in the distribution of dataset, different parameters are selected to achieve the optimal clustering results. When λ 3 is set to 1, our method achieves the optimal clustering results on the Salinas and PaviaC datasets. And when λ 3 is set to 100, our method achieves the optimal clustering results on the PaviaU dataset.
Figure 6. Clustering performance with different M on the three datasets. (a) Salinas. (b) PaviaU. (c) PaviaC.
Figure 7. Sensitivity analysis of the parameters λ 1 , λ 2 and λ 3 on the three datasets. Left: Salinas; Medium: PaviaU; Right: PaviaC.
Effectiveness of ERS Segmentation.
Table 5 shows the comparison of different superpixel segmentation methods included SLIC, LSC and ERS. The performance of ERS incorporating spatial information outperforms SLIC and LSC. The main distinction lies in the fact that ERS embraces the complete spatial context, while SLIC and LSC rely on K-nearest neighbors for segmentation. By enhancing spatial awareness and more accurately capturing spatial relationships within the data, ERS achieves improved clustering accuracy.
Table 5. Ablation study of segmentation algorithms.
Ablation Study. To demonstrate the effectiveness of discriminative anchors (DAs) and low-rank constraint (LR) in our proposed algorithm, we conduct ablation experiments on the three HSI datasets, as shown in Table 6. DA and LR respectively correspond to the third and fourth terms in the objective function. By comparing three variants, we find that explicitly learning discriminative anchors and low-rank constraint can better represent pixels, improving the quality of anchor graph and enhancing the clustering performance.
Table 6. Ablation study of the proposed method on the three datasets.
Converge Analysis. For the proposed method, we set the iteration cap to 300. When the reconstruction errors (i.e., Z D , S E and Z F ) are less than ε , the iteration stops and the model converges. To investigate the convergence of the proposed algorithm, we plot the convergence curves of reconstruction errors and clustering metrics curves (OA, NMI, and Kappa) on the three databases in Figure 8. It can be seen that our algorithm can gradually converge within 100 iterations, and the best performance after convergence is selected as the final metric. Therefore, our algorithm has good convergence and performs well in practical applications.
Figure 8. Convergence curves and clustering metrics curves on the three datasets. (a) Salinas. (b) PaviaU. (c) PaviaC.

5. Conclusions

In this paper, we proposed a discriminative anchor-based HSI clustering algorithm. It achieves consistency between the distribution of anchors and that of pixels by sharing the coefficient matrix between an anchor matrix and anchor graph, ensuring that the selected anchors can better reflect the distribution structure of pixels. By imposing low-rank and probabilistic constraints on the consensus coefficient matrix, it effectively excavates the structural information of anchors and strengthens their discriminative ability. Experimental results on three HSI databases have confirmed the superiority and effectiveness of the proposed approach.

Author Contributions

Conceptualization, Y.Y. and Q.G.; methodology, Y.Y. and Q.G.; software, Y.Y.; validation, J.Z., Y.D. and C.D.; visualization, Y.Y., J.Z. and Y.D.; supervision, Q.G., J.Z., Y.D. and C.D.; project administration, Q.G., J.Z. and Y.D.; data curation, Y.Y., J.Z. and Y.D.; writing—original draft preparation, Y.Y.; writing—review and editing, Y.Y., Q.G., J.Z. and Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China, Grant No. 62176203 and 62576263; the Natural Science Basic Research Program of Shaanxi Province, Grant No. 2025JC-QYCX-051; and the Fundamental Research Funds for the Central Universities and the Innovation Fund of Xidian University, Grant No. YJSJ25007.

Data Availability Statement

The data presented in this study are available on request corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, G.; Huang, K.; Sun, W.; Meng, X.; Mao, D.; Ge, Y. Enhanced mangrove vegetation index based on hyperspectral images for mapping mangrove. ISPRS J. Photogramm. Remote Sens. 2022, 189, 236–254. [Google Scholar] [CrossRef]
  2. Tu, B.; He, W.; Li, Q.; Peng, Y.; Plaza, A. A New Context-Aware Framework for Defending Against Adversarial Attacks in Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5505114. [Google Scholar] [CrossRef]
  3. Zhang, B.; Li, S.; Wu, C.; Gao, L.; Zhang, W.; Peng, M. A neighbourhood-constrained k-means approach to classify very high spatial resolution hyperspectral imagery. Remote Sens. Lett. 2013, 4, 161–170. [Google Scholar] [CrossRef]
  4. Ranjan, S.; Nayak, D.R.; Kumar, K.S.; Dash, R.; Majhi, B. Hyperspectral image classification: A k-means clustering based approach. In Proceedings of the 2017 4th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 6–7 January 2017; pp. 1–7. [Google Scholar]
  5. Ghaffarian, S.; Ghaffarian, S. Automatic histogram-based fuzzy C-means clustering for remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2014, 97, 46–57. [Google Scholar] [CrossRef]
  6. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef]
  7. Ester, M.; Kriegel, H.-P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the International Conference on Knowledge Discovery and Data Mining (KDD-96), Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  8. Xie, H.; Zhao, A.; Huang, S.; Han, J.; Liu, S.; Xu, X.; Luo, X.; Pan, H.; Du, Q.; Tong, X. Unsupervised Hyperspectral Remote Sensing Image Clustering Based on Adaptive Density. IEEE Geosci. Remote Sens. Lett. 2018, 15, 632–636. [Google Scholar] [CrossRef]
  9. Zhao, Y.; Yuan, Y.; Wang, Q. Fast Spectral Clustering for Unsupervised Hyperspectral Image Classification. Remote Sens. 2019, 11, 399. [Google Scholar] [CrossRef]
  10. Huang, N.; Xiao, L.; Liu, J.; Chanussot, J. Graph Convolutional Sparse Subspace Coclustering With Nonnegative Orthogonal Factorization for Large Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2023, 60, 5512016. [Google Scholar] [CrossRef]
  11. Cai, Y.; Zeng, M.; Cai, Z.; Liu, X.; Zhang, Z. Graph Regularized Residual Subspace Clustering Network for Hyperspectral Image Clustering. Inf. Sci. 2021, 578, 85–101. [Google Scholar] [CrossRef]
  12. Li, T.; Cai, Y.; Zhang, Y.; Cai, Z.; Liu, X. Deep Mutual Information Subspace Clustering Network for Hyperspectral Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6009905. [Google Scholar] [CrossRef]
  13. Lei, J.; Li, X.; Peng, B.; Fang, L.; Ling, N.; Huang, Q. Deep Spatial-Spectral Subspace Clustering for Hyperspectral Image. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 2686–2697. [Google Scholar] [CrossRef]
  14. Zhao, Y.; Li, X. Deep Spectral Clustering With Regularized Linear Embedding for Hyperspectral Image Clustering. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5509311. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Wang, Y.; Chen, X.; Jiang, X.; Zhou, Y. Spectral–Spatial Feature Extraction With Dual Graph Autoencoder for Hyperspectral Image Clustering. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 8500–8511. [Google Scholar] [CrossRef]
  16. Hu, X.; Li, T.; Zhou, T.; Peng, Y. Deep Spatial-Spectral Subspace Clustering for Hyperspectral Images Based on Contrastive Learning. Remote Sens. 2021, 13, 4418. [Google Scholar] [CrossRef]
  17. Guan, R.; Tu, W.; Li, Z.; Yu, H.; Hu, D.; Chen, Y.; Tang, C.; Yuan, Q.; Liu, X. Spatial–Spectral Graph Contrastive Clustering With Hard Sample Mining for Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5535216. [Google Scholar] [CrossRef]
  18. Liu, W.; Liu, K.; Sun, W.; Yang, G.; Ren, K.; Meng, X.; Peng, J. Self-Supervised Feature Learning Based on Spectral Masking for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4407715. [Google Scholar] [CrossRef]
  19. Luo, F.; Liu, Y.; Duan, Y.; Guo, T.; Zhang, L.; Du, B. SDST: Self-Supervised Double-Structure Transformer for Hyperspectral Images Clustering. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5511614. [Google Scholar] [CrossRef]
  20. Elhamifar, E.; Vidal, R. Sparse Subspace Clustering: Algorithm, Theory, and Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef]
  21. Zhang, H.; Zhai, H.; Zhang, L.; Li, P. Spectral-Spatial Sparse Subspace Clustering for Hyperspectral Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3672–3684. [Google Scholar] [CrossRef]
  22. Zhai, H.; Zhang, H.; Zhang, L.; Li, P.; Plaza, A. A New Sparse Subspace Clustering Algorithm for Hyperspectral Remote Sensing Imagery. IEEE Geosci. Remote Sens. Lett. 2017, 14, 43–47. [Google Scholar] [CrossRef]
  23. Cai, Y.; Zhang, Z.; Cai, Z.; Liu, X.; Jiang, X.; Yan, Q. Graph convolutional subspace clustering: A robust subspace clustering framework for hyperspectral image. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4191–4202. [Google Scholar] [CrossRef]
  24. Huang, S.; Zhang, H.; Pizurica, A. Subspace clustering for hyperspectral images via dictionary learning with adaptive regularization. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5524017. [Google Scholar] [CrossRef]
  25. Zhai, H.; Zhang, H.; Zhang, L.; Li, P. Sparsity-Based Clustering for Large Hyperspectral Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10410–10424. [Google Scholar] [CrossRef]
  26. Huang, S.; Zeng, H.; Chen, H.; Zhang, H. Spatial and Cluster Structural Prior-Guided Subspace Clustering for Hyperspectral Image. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5511115. [Google Scholar] [CrossRef]
  27. Huang, S.; Zhang, H.; Du, Q.; Pizurica, A. Sketch-Based Subspace Clustering of Hyperspectral Images. Remote Sens. 2020, 12, 775. [Google Scholar] [CrossRef]
  28. Zhai, H.; Zhang, H.; Zhang, L.; Li, P. Nonlocal Means Regularized Sketched Reweighted Sparse and Low-Rank Subspace Clustering for Large Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4164–4178. [Google Scholar] [CrossRef]
  29. Zhang, Y.; Jiang, G.; Cai, Z.; Zhou, Y. Bipartite Graph-Based Projected Clustering With Local Region Guidance for Hyperspectral Imagery. IEEE Trans. Multim. 2024, 26, 9551–9563. [Google Scholar] [CrossRef]
  30. Zhao, H.; Zhou, F.; Bruzzone, L.; Guan, R.; Yang, C. Superpixel-level Global and Local Similarity Graph-based Clustering for Large Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5519316. [Google Scholar] [CrossRef]
  31. Cui, K.; Li, R.; Polk, S.L.; Lin, Y.; Zhang, H.; Murphy, J.M.; Plemmons, R.J.; Chan, R.H. Superpixel-based and Spatially-regularized Diffusion Learning for Unsupervised Hyperspectral Image Clustering. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4405818. [Google Scholar] [CrossRef]
  32. Wang, R.; Nie, F.; Wang, Z.; He, F.; Li, X. Scalable Graph-Based Clustering With Nonnegative Relaxation for Large Hyperspectral Image. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7352–7364. [Google Scholar] [CrossRef]
  33. Wang, R.; Nie, F.; Yu, W. Fast Spectral Clustering With Anchor Graph for Large Hyperspectral Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2003–2007. [Google Scholar] [CrossRef]
  34. Wang, Q.; Miao, Y.; Chen, M.; Yuan, Y. Spatial-Spectral Clustering With Anchor Graph for Hyperspectral Image. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5542413. [Google Scholar] [CrossRef]
  35. Chen, X.; Zhang, Y.; Feng, X.; Jiang, X.; Cai, Z. Spectral-Spatial Superpixel Anchor Graph-Based Clustering for Hyperspectral Imagery. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5507405. [Google Scholar] [CrossRef]
  36. Cai, Y.; Zhang, Z.; Ghamisi, P.; Ding, Y.; Liu, X.; Cai, Z.; Gloaguen, R. Superpixel Contracted Neighborhood Contrastive Subspace Clustering Network for Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5530113. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Wang, X.; Jiang, X.; Zhang, L.; Du, B. Elastic Graph Fusion Subspace Clustering for Large Hyperspectral Image. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 6300–6312. [Google Scholar] [CrossRef]
  38. Zhang, L.; Zhang, L.; Du, B.; You, J.; Tao, D. Hyperspectral image unsupervised classification by robust manifold matrix factorization. Inf. Sci. 2019, 485, 154–169. [Google Scholar] [CrossRef]
  39. Jiang, G.; Zhang, Y.; Wang, X.; Jiang, X.; Zhang, L. Structured Anchor Learning for Large-Scale Hyperspectral Image Projected Clustering. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 2328–2340. [Google Scholar] [CrossRef]
  40. Wang, N.; Cui, Z.; Lan, Y.; Zhang, C.; Xue, Y.; Su, Y.; Li, A. Large-Scale Hyperspectral Image-Projected Clustering via Doubly Stochastic Graph Learning. Remote Sens. 2025, 17, 1526. [Google Scholar] [CrossRef]
  41. Liu, M.-Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  42. Lin, Z.; Liu, R.; Su, Z. Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation. In Proceedings of the Advances in Neural Information Processing Systems 24 (NIPS 2011), Granada, Spain, 12–15 December 2011; pp. 612–620. [Google Scholar]
  43. Gao, Q.; Zhang, P.; Xia, W.; Xie, D.; Gao, X.; Tao, D. Enhanced Tensor RPCA and Its Application. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2133–2140. [Google Scholar] [CrossRef]
  44. Cai, J.-F.; Candès, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  45. Nie, F.; Wang, X.; Jordan, M.; Huang, H. The constrained laplacian rank algorithm for graph-based clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1969–1976. [Google Scholar]
  46. Shahi, K.R.; Khodadadzadeh, M.; Tusa, L.; Ghamisi, P.; Tolosana-Delgado, R.; Gloaguen, R. Hierarchical Sparse Subspace Clustering (HESSC): An Automatic Approach for Hyperspectral Image Analysis. Remote Sens. 2020, 12, 2421. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.