Next Article in Journal
Terahertz Squint SAR Imaging Based on Decoupled Frequency Scaling Algorithm
Previous Article in Journal
Enhanced Detection of Drought Events in California’s Central Valley Basin Using Rauch–Tung–Striebel Smoothed GRACE Level-2 Data: Mechanistic Insights from Climate–Hydrology Interactions
Previous Article in Special Issue
SCSU–GDO: Superpixel Collaborative Sparse Unmixing with Graph Differential Operator for Hyperspectral Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CPMFFormer: Class-Aware Progressive Multiscale Fusion Transformer for Hyperspectral Image Classification

1
State Key Laboratory for Strength and Vibration of Mechanical Structures, Xi’an Jiaotong University, Xi’an 710049, China
2
School of Aerospace Engineering, Xi’an Jiaotong University, Xi’an 710049, China
3
School of Automation Science and Engineering, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(22), 3684; https://doi.org/10.3390/rs17223684
Submission received: 15 July 2025 / Revised: 5 November 2025 / Accepted: 6 November 2025 / Published: 10 November 2025

Highlights

What are the main findings?
  • Existing spectral weighting mechanisms mainly rely on spectral responses to measure band importance. This makes it difficult for them to effectively distinguish samples with similar spectra, but which belong to different classes. Therefore, this paper introduces class information to enable learning of class-specific spectral weights. This will help in the exploration of subtle spectral differences between ground objects.
  • Current multiscale fusion strategies generally ignore the semantic hierarchical differences between features at different scales. Directly fusing features with large-scale differences easily leads to information confusion and semantic conflict. Therefore, this paper designs a progressive multiscale fusion strategy. It achieves a smooth semantic transition by gradually fusing adjacent scale features.
What are the implications of the main findings?
  • Integrating class information into the spectral weighting mechanism provides a novel approach to handle the intra-class variability and inter-class similarity in HSI analysis. This is crucial for fine-grained classification tasks.
  • The progressive multiscale fusion strategy provides a scalable and transferable feature representation learning framework for most remote sensing image processing tasks involving multiscale spatial consistency modeling.

Abstract

Hyperspectral image (HSI) classification is a basic and significant task in remote sensing, the aim of which is to assign a class label to each pixel in an image. Recently, deep learning networks have been widely applied in HSI classification. They can extract discriminative spectral–spatial features through spectral weighting and multiscale spatial information modeling. However, existing spectral weighting mechanisms lack the ability to explore the inter-class spectral overlap caused by spectral variability. Moreover, current multiscale fusion strategies ignore semantic conflicts between features with large-scale differences. To address these problems, a class-aware progressive multiscale fusion transformer (CPMFFormer) is proposed. It first introduces class information into a spectral weighting mechanism. This helps CPMFFormer to learn class-specific spectral weights and enhance class-discriminative spectral features. Then, a center residual convolution module is constructed to extract features at different scales. It is embedded with a center feature calibration layer to achieve hierarchical enhancement of representative spatial features. Finally, a progressive multiscale fusion strategy is designed to promote effective collaboration between features at different scales. It achieves a smooth semantic transition by gradually fusing adjacent scale features. Experiments using five public HSI datasets show that CPMFFormer is rational and effective.

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) are a core data source for achieving fine observation and quantitative analysis in remote sensing. They are composed of hundreds of continuous and narrow spectral bands [1]. HSIs can provide abundant spectral–spatial information for fine-grained ground-object identification. HSIs play a crucial role in various remote sensing tasks [2]. For example, in agricultural remote sensing, by capturing subtle differences in vegetation spectra, HSIs can be used to achieve precise crop-type classification [3], dynamic monitoring of growth status [4], and early detection of pests and diseases [5]. This provides technical support for remote sensing-enabled smart agriculture. In geological remote sensing, leveraging the unique spectral absorption characteristics of minerals, HSIs can be used for lithological mapping [6], mineral resource exploration [7], and monitoring of environmental changes [8] in mining regions. This improves the efficiency and accuracy of remote sensing geological surveys. In ecological and environmental remote sensing, HSIs can be used to assess wetland ecosystem evolution [9], forest health status [10], and trace sources of soil/water pollution [11]. This provides long-term and large-scale quantitative data for remote sensing ecological monitoring. In marine remote sensing, HSIs can be used to retrieve water quality parameters [12] and provide early warnings of algal blooms [13]. This compensates for the shortcomings of traditional remote sensing in fine-scale marine monitoring. As a basic and significant task in HSI applications, HSI classification aims to assign a specific class label to each pixel [14].
In early works, non-deep learning (non-DL) models were often applied in HSI classification, including the support vector machines (SVMs) [15], random forests (RFs) [16], multinomial logistic regressions (MLRs) [17], back-propagation artificial neural networks (BP-ANNs) [18], and spectral angle mappers (SAMs) [19]. These models usually take handcrafted spectral–spatial features [20,21,22] as input. However, their limited representation capability and poor generalization performance make it difficult for non-DL methods to adapt to different scenes [23]. As a representative paradigm in artificial intelligence (AI), deep learning technology has been widely applied in HSI classification [24]. The convolutional neural networks (CNNs) and transformers are two popular deep learning methods. Through end-to-end representation learning, they can adaptively extract highly discriminative spectral–spatial features based on task requirements [25]. To effectively integrate spectral–spatial information, various AI methods have been proposed [26], including spectral methods, spatial methods, and center attention methods.
Spectral methods are mainly used to exploit subtle spectral differences between ground objects. The spectral weighting mechanism is a common spectral feature enhancement technology. It can assign an importance weight to each band. This emphasizes representative bands and suppresses redundant spectral information [27]. For example, Yang et al. [28] used global and local spectral information to enhance spectral features that contribute greatly to HSI classification. Shu et al. [29] used a spectral attention mechanism to generate a weight vector for enhancing discriminative spectral bands. Chhapariya et al. [30] used a spectral weighting mechanism to dynamically select basic features from each band. The above methods mainly measure the band contribution based on the difference in spectral responses, but their discrimination capability is still greatly restricted due to the impact of spectral variability. Spectral variability causes the spectral distributions of samples of the same ground object to diverge and the spectral responses of different ground objects to overlap with each other [31]. Existing spectral weighting mechanisms generally ignore class semantics. This makes it difficult for AI methods to accurately distinguish samples with similar spectra, but which belong to different classes.
Spatial methods are mainly used to model spatial distribution characteristics of ground objects. The multiscale CNN is a common spatial method. It uses convolution kernels with different receptive fields to extract features at different scales, which facilitates capturing diverse spatial distribution patterns of ground objects [32]. Features at different scales are complementary, and their fusion can enrich spatial semantic information [33]. For example, Ge et al. [34] first concatenated features at different scales and then used a channel attention mechanism to highlight the information-rich channel features. Similarly, Shi et al. [35] first concatenated features at different scales and subsequently adopted a spatial-channel attention mechanism to reinforce salient spatial and channel responses. Shu et al. [36] first split features into multiple branches along the channel dimension, from which they extracted features at different scales. Then, features at different scales are concatenated. The above multiscale fusion strategies directly fuse features from all scales. Although they are easy to calculate and simple in structure, these multiscale fusion strategies generally ignore the semantic hierarchical differences between features at different scales. Directly fusing features with large-scale differences easily leads to information confusion and semantic conflicts. This makes it difficult for AI systems to effectively integrate the scale information that is crucial for HSI classification.
Unlike other computer vision tasks, HSI classification usually takes a 3D patch cropped from HSI data as input and predicts the class label of the center pixel [37]. The input patch usually contains background noise and heterogeneous pixels that do not belong to the same class as the center pixel. These pixels introduce interference information [38]. The sliding sampling characteristic and parameter sharing mechanism of CNN further propagate such interference information into the convolutional features, which adversely affects the classification decision of the center pixel. Due to the limitation of the receptive field, it is difficult for the CNN to model correlations between the center pixel and its neighboring pixels [39]. Therefore, Chen et al. [40], Wang et al. [41], Jia et al. [42], Lu et al. [43], and Yu et al. [44] have proposed various center attention methods based on transformer [45]. They measure different contributions of neighboring pixels to the center pixel. This enables the classification network to learn discriminative features for accurately characterizing the center pixel. However, the above methods generally regard CNN and the center attention mechanism as two independent feature processing modules. This limits their capability to effectively suppress interference information accumulated during convolution, thereby affecting AI’s recognition accuracy and predictive performance.
In general, although AI modules such as the spectral weight learning and center attention mechanism can improve classification performance, their decision processes lack interpretability. This makes it difficult for them to intuitively explain how AI systems use spectral–spatial features for reasoning. Furthermore, in multiscale feature fusion, current AI networks tend to introduce redundant or conflicting information, which results in semantic degradation and reduces the feature representation capability.
To address the above problems, a class-aware progressive multiscale fusion transformer (CPMFFormer) is proposed. CPMFFormer first introduces class information and designs a class-aware spectral weighting module (CSWM) to enhance class-discriminative spectral features. Second, a center residual convolution module (CRCM) is constructed to extract features at different scales and suppress interference information. Then, a progressive multiscale fusion strategy is designed to gradually fuse adjacent scale features, which achieves a smooth semantic transition. Finally, the final fusion features are used to predict the class label of the center pixel. The main contributions in this paper are as follows:
(1)
CSWM introduces an aggregation operation to smooth the input patch, which reduces spectral information differences between intra-class samples. Moreover, CSWM uses a class consistency loss to learn spectral weight centers of different ground objects. This helps CSWM to obtain class-specific spectral weights.
(2)
A center feature calibration layer (CFCL) is designed to enhance the representational capability of the center pixel. Furthermore, CFCL is combined with multiscale group convolutional layers to construct CRCM, which enables continuous suppression of interference information at different spatial scales.
(3)
A cross-scale fusion bottleneck transformer (CSFBT) is built to integrate complementary information at different scales. It designs a multi-head cross-scale dual-dimensional attention (MHCSD2A) mechanism to promote semantic interaction between adjacent scale features from both spatial and channel dimensions.
To validate the superiority of CPMFFormer, this paper compares it with 15 existing networks using five public HSI datasets. Experiments show that CPMFFormer has better classification performance, stronger stability, and lower computational complexity.
The rest of this paper is organized as follows. Section 2 introduces the preliminaries. Section 3 presents CPMFFormer. Section 4 validates CPMFFormer using five public HSI datasets. Section 5 discusses CPMFFormer. Section 6 draws the conclusions.

2. Preliminaries

HSI classification is a classic task in remote sensing. Its purpose is to identify different ground objects based on spectral–spatial information. In current research, CNN and transformer are two common feature representation learning frameworks. This section provides a systematic analysis of their theoretical foundations and inherent limitations.

2.1. Question Statement

HSI classification aims to assign a class label to each pixel. A common approach is to use a 3D patch cropped from HSI data as input and predict the class label of the center pixel. Let the HSI data be X R H × W × B , where H , W , and B are the image height, image width, and the number of spectral bands, respectively. The input patch set can be represented as X 1 ,   X 2 ,   ,   X N . Here, X i R P × P × B is the input patch centered at the pixel x i , where P is the patch size. The purpose of HSI classification is to learn a mapping function f θ : X i 1,2 , , K that maps the input patch X i to the class label of the pixel x i , where θ is a set of learnable parameters and K is the number of classes.

2.2. CNN in HSI Classification

In CNN-based classification networks, spectral weighting mechanisms are often used to enhance representative spectral features. They learn an importance weight for each band based on spectral response values. Spectral variability causes samples of the same class to have different spectral responses, while samples of different classes have similar spectral characteristics. As shown in Figure 1, this section presents the average spectral response values and standard deviations of the Corn-Notill, Corn-Mintill, and Corn classes in the Indian dataset. The results indicate that the three types of ground objects have highly similar spectral curves. Moreover, the spectral distributions of samples of the three types of ground objects overlap with each other. However, current spectral weighting mechanisms generally learn band weights based on spectral responses and ignore class information. This makes it difficult for them to deal with inter-class spectral overlap and accurately distinguish samples with similar spectra, but which belong to different classes.
In HSI classification, the multiscale CNN uses convolutional kernels with different receptive fields to extract features at different scales. This helps the classification network to better model spatial information and capture diverse spatial distribution patterns of ground objects. As shown in Figure 2, this section presents the activation response heatmaps obtained from the PaviaU dataset using different convolutional kernels. The activation response represents the effective feature information obtained at a specific scale. The larger the activation response, the more effective the feature information. The red and white boxes in Figure 2a represent the Trees and Painted Metal Sheets classes, respectively. On small-scale features extracted by 3 × 3  convolutional kernel, the Trees class has  a larger activation response. Its activation response gradually decreases as  the receptive field increases. The Painted Metal Sheets class has a smaller  activation response on small-scale features extracted by 3 × 3  convolutional kernel. As the receptive  field increases, its activation response gradually increases. In particular, the  Trees and Painted Metal Sheets classes have completely different activation  responses on features extracted by 3 × 3  convolutional kernel and features  extracted by 9 × 9  convolutional kernel. In this case,  directly fusing features from all scales may cause the ground object with a higher  activation response at a specific scale to be weakened by features at other  scales. This easily leads to semantic degradation and affects classification  performance.

2.3. Transformer in HSI Classification

The limited receptive field makes it difficult for CNN to model long-range dependency. Recently, transformer based on the multi-head self-attention (MHSA) mechanism [45] has been applied in HSI classification to capture long-range context [46,47,48,49,50].
Let the input feature be F R H F × W F × C , where H F , W F , and C are the feature height, width, and channel dimension. As shown in Figure 3, the self-attention mechanism first uses three linear projection layers to generate the query matrix Q R H F × W F × D , key matrix K R H F × W F × D , and value matrix V R H F × W F × D from the input feature F , as follows:
Q = F W Q ,       K = F W K ,       V = F W V
where W Q , W K , W V R D × D are three learnable projection matrixes.
According to Figure 3, the self-attention mechanism can be formulated as follows:
A = A t t e n t i o n Q , K , V = s o f t m a x Q K T D V
MHSA first uses multiple W h Q , W h K , and W h V to generate Q h , K h , and V h . Then, Q h , K h , and V h are used to calculate the self-attention according to Equation (2) in the hth head. Finally, the self-attention calculated in all heads is concatenated, and a linear projection layer is used to obtain the output features of MHSA, as follows:
M H S A = C o n c a t A 1 , , A H W o
where H is the number of heads; W o is a learnable linear projection matrix; and A h denotes the self-attention calculated in the hth head, which can be formulated as
A h = A t t e n t i o n Q h , K h , V h
HSI classification networks usually take a square patch as input. This results in the input patch containing background noise and heterogeneous pixels that do not belong to the same class as the center pixel. The interference information introduced by these pixels affects the classification decision of the center pixel. As shown in Table 1, this section reports the number of homogeneous patches and heterogeneous patches in five public HSI datasets. The results show that the number of heterogeneous patches is significantly greater than the number of homogeneous patches.
To reduce the impact of background noise and heterogeneous pixels, various center attention mechanisms have been proposed. They measure correlations between the center pixel and its neighboring pixels. This enables the classification network to obtain discriminative spectral–spatial information from neighboring pixels that have high correlation with the center pixel. However, current center attention mechanisms usually neglect their combination with convolutional feature extraction. They take the center attention mechanism as an independent feature processing module. This may cause interference information to continue accumulate in the convolutional layer, thereby reducing the feature representation capability and overall classification performance.

3. Class-Aware Progressive Multiscale Fusion Transformer

Aiming at the above problems, a class-aware progressive multiscale fusion transformer (CPMFFormer) is proposed. It is a deep learning network based on AI reasoning. CPMFFormer enhances AI’s intelligent decision capabilities through class-aware spectral discrimination and progressive semantic fusion mechanism.

3.1. Overall Framework

As shown in Figure 4, CPMFFormer consists of the spectral weighting stage, multiscale feature extraction stage, and multiscale feature fusion and classification stages.
(1)
In the spectral weighting stage, CPMFFormer first crops a patch of size P × P × B centered at the pixel x i from HSI. Then, the patch is fed into CSWM to generate a weighted patch of size P × P × B . Finally, the weighted patch is resized to P × P × C using a 1 × 1 convolutional layer, where C is the feature channel dimension.
(2)
In the multiscale feature extraction stage, CPMFFormer uses four CRCMs with different receptive fields to extract features at different scales of size P × P × C . Moreover, CRCM introduces a center feature calibration layer (CFCL) to suppress interference information and enhance the representation capability for the center pixel.
(3)
In the multiscale feature fusion and classification stage, CPMFFormer uses the cross-scale fusion bottleneck transformer (CSFBT) to gradually fuse adjacent scale features. Moreover, a global average pooling (GAP) layer, a fully connected (FC) layer, and a S o f t m a x function are used to predict the class label of the center pixel, as follows:
y ^ i = S o f t m a x F C G A P F O U T
where F O U T is the output features of CSFBT-6 and y ^ i is the predicted class probability.

3.2. Class-Aware Spectral Weighting Module

In the first stage shown in Figure 4, to accurately distinguish samples with similar spectra but which belong to different classes, CSWM introduces class information into the spectral weighting mechanism to obtain class-specific spectral weights.
As shown in Figure 5, CSWM consists of an aggregation operation, a spectral weight learning block, and a class consistency loss function. Affected by spectral variability, pixels of the same class may exhibit significant differences across spectral bands. To mitigate this intra-class spectral fluctuation, CSWM first measures similarity relationships between bands in the spectral domain through the aggregation operation. This operation calculates the inter-band similarity matrix A R B × B using the Euclidean distance, as follows:
A = 1 S m i n S m a x S m i n S
where S R B × B is the distance matrix which can be formulated as follows:
S i j = B i B j 2 = h , w = 1 H ,   W B h w i B h w j 2
where H and W are the height and width of the HSI. B i and B j are two bands.
Next, a normalized graph Laplacian matrix L is constructed, as follows:
L = G 1 / 2 I A G 1 / 2
where I and G are the identity matrix and degree matrix, respectively.
By smoothing the input patch X i using the graph Laplacian matrix L , the aggregated patch X ~ i can be obtained. The calculation process can be formulated as follows:
X ~ i = X i L
This process can be viewed as building a fully connected graph between similar bands in the spectral domain. This allows spectral information to be propagated between semantically similar band features and helps mitigate random shifts caused by spectral fluctuations. In addition, CSWM introduces a learnable matrix W R B × B to adaptively weight the aggregated patches. This further enhances discriminative band combination. The final smoothing output in the spectral domain can be formulated as follows:
X ¯ i = X ~ i W
Through the above operations, CSWM can automatically learn the dynamic correlations between bands and effectively reduce the interference of noise bands.
To highlight key band information, CSWM uses the channel attention mechanism [51] to weight the spectral vector. Specifically, CSWM uses a global max pooling (GMP) layer and a GAP layer in parallel to generate two spatial context descriptors, as follows:
D G M P = G M P X ¯ i ,       D G A P = G A P X ¯ i
These two descriptors are fed into a shared multi-layer perceptron (MLP). The spectral weight w i R 1 × 1 × B is obtained through a nonlinear activation function, as follows:
w i = σ M L P D C M P + M L P D C A P
where σ · is the S i g m o i d function. The final weighted output can be formulated as follows:
X ^ i = X ¯ i · w i + X i
Relying only on the spectral distribution for weighting may result in band weights of different ground objects being highly similar. To enhance class-discriminative spectral features, CSWM introduces a class consistency loss, as follows:
L C = 1 M i = 1 M C y i w i 2 2
where M is the batch size; l i is the class label of the ith patch; and C y i is a learnable matrix that represents the spectral weight center of the yith class.
This constraint forces the spectral weights of samples of the same class to converge to the same center during training, thereby enhancing intra-class consistency.
By introducing spectral aggregation, attention weighting, and class constraints, CSWM improves the discriminative power of spectral features. It not only smooths spectral fluctuations and suppresses redundant bands but also achieves class-level spectral selectivity. This provides more discriminative input features for subsequent modules.

3.3. Center Residual Convolution Module

In the second stage shown in Figure 4, to extract features at different scales and reduce the impact of abnormal pixels on the classification decision of the center pixel, this section designs CRCM based on group convolution and a center attention mechanism.
As shown in Figure 6, CRCM is composed of two bottleneck residual blocks. Let the input feature size be P × P × C . The bottleneck residual block first uses a 1 × 1 convolutional layer to reduce the feature channel dimension. The reduced feature size is P × P × C R . Second, a k t × k t group convolutional layer is used to extract features at the tth scale. A batch normalization and a rectified linear unit (ReLU) function are used for feature normalization and nonlinear enhancement. Then, CRCM introduces CFCL to model the dependencies between the center pixel and its neighboring pixels. Finally, a 1 × 1 convolutional layer is used to restore the feature channel dimension.
As shown in Figure 4, to capture spatial structure information, CPMFFormer uses four CRCMs with convolution kernel sizes of 3 × 3 , 5 × 5 , 7 × 7 , and 9 × 9 , respectively, to extract features at different scales. The increase in the convolutional kernel size leads to an increase in the computational complexity. Considering that group convolution can reduce computational cost by decreasing the feature channel dimension in each group. Therefore, this paper adaptively allocates the feature channel dimension C t in each group based on the convolutional kernel size, which can be formulated as follows:
C t = C R / 2 k t 1 2
This adaptive strategy enables CRCM to achieve abundant multiscale perception while remaining lightweight. This helps improve the applicability of CPMFFormer.
The core idea of CFCL is to simulate the attention mechanism, which strengthens the key region features by calculating dependencies between the center pixel and its neighboring pixels. As shown in Figure 7, CFCL first calculates an inner product attention weight A C R P × P between each neighboring pixel and the center pixel, as follows:
A i j C = F i j · F i c j c + R i j C
where i , j is the spatial position of the neighboring pixel x n ; i c , j c is the spatial position of the center pixel x c ; A i j C is the dependency between the neighboring pixel x n and the center pixel x c ; F i j represents the feature vector of the neighboring pixel x n ; F i j represents the feature vector of the center pixel x c ; and R i j C is the relative center position encoding of the neighboring pixel x n , which can be formulated as follows:
R i j C = 1 / 1 + | | i i c | | 2 + | | j j c | | 2
Then, a center dependency matrix S C R P × P is obtained using a S o f t m a x normalization. The final output calibration features F C R P × P × C R can be expressed as follows:
F i j C = S i j C · F i j
The above process can be viewed as a weighted summation for the neighboring pixels, which enhances the feature representation for regions that are highly correlated with the center pixel. This helps to improve the distinguishability of the center pixel.
Compared with the existing center attention mechanism, this section embeds CFCL into CRCM at different spatial scales. This achieves hierarchical suppression of interference information. Among them, shallow CFCL suppresses local noise, while deep CFCL enhances structural boundaries. This module not only effectively resists interference from abnormal pixels but also enhances spatial continuity. This provides a semantically consistent feature representation for subsequent cross-scale fusion.

3.4. Cross-Scale Fusion Bottleneck Transformer

In the third stage shown in Figure 4, after completing the multiscale spatial feature extraction, CPMFFormer needs to effectively fuse features at different scales to fully explore cross-scale complementary information. Therefore, this paper builds a progressive multiscale fusion strategy. This strategy takes CSFBT as its core, which achieves a smooth semantic transition by gradually fusing adjacent scale features.
As shown in Figure 8, CSFBT takes adjacent scale features as input. It consists of a channel feature enhancement block (CFEB) and a cross-scale fusion block (CSFB). CFEB enhances valuable channel features. CSFB uses a multi-head cross-scale dual-dimensional attention (MHCSD2A) mechanism to promote semantic interaction and information redistribution across adjacent scale features from both spatial and channel dimensions.
The design idea of CFEB is to weight the channel features in adjacent scale features. This enables CFEB to highlight information-rich channel responses. For small-scale features F S R P × P × C and large-scale features F L R P × P × C , CFEB first uses two GAP layers to obtain two spatial context descriptors D G A P S , D G A P L R 1 × 1 × C , as follows:
D G A P S = G A P F S ,       D G A P L = G A P F L
Then, two MLPs are used to generate the channel attention weights w S R 1 × 1 × C and w L R 1 × 1 × C for small-scale and large-scale features, respectively, as follows:
w S = σ M L P D G A P S ,       w L = σ M L P D G A P L
where σ · denotes the S o f t m a x function.
Finally, the weight vectors w S and w L are used to reweight small-scale features and large-scale features, respectively, which can be formulated as follows:
F S = w S · F S ,       F L = w L · F L
CFEB can effectively suppress invalid channel features and make the input adjacent scale features more focused and discriminative in the channel dimension.
Next, CSFB is used to capture cross-scale dependencies from both spatial and channel dimensions. As shown in Figure 8, CSFB first uses two 1 × 1 convolutional layers to reduce the feature channel dimension. The reduced small-scale and large-scale features are denoted as F S ,   F L R P × P × C . Then, F S and F L are fed into MHCSD2A to integrate complementary information across different scales. Finally, a 1 × 1 convolutional layer is used to restore the feature channel dimension and obtain the fusion features.
Let MHCSD2A have H heads and the feature channel dimension in each head be D . As shown in Figure 9, MHCSD2A projects F S and F L into F S 1 , , F S H R P × P × D and F L 1 , , F L H R P × P × D , respectively. F S h and F L h are input into the cross-scale dual-dimensional attention (CSD2A) mechanism to enhance scale-specific information. As shown in Figure 10, CSD2A adds F S h and F L h to obtain multiscale features F M h R P × P × D . F M h contains both small-scale detail information and large-scale semantic information. CSD2A uses two 1 × 1 convolutional layers to obtain Q S R P × P × D and Q L R P × P × D from F S h and F L h , respectively. Meanwhile, CSD2A uses three 1 × 1 convolutional layers to obtain K S R P × P × D , K L R P × P × D , and V R P × P × D from F M h , which can be formulated as follows:
Q S = F S h W S Q ,       Q L = F L h W L Q
K S = F M h W S K ,       K L = F M h W L K ,       V = F M h W V
where W S Q , W L Q , W S K , W L K , W V R D × D are five learnable linear projection matrices.
On the spatial dimension, CSD2A first multiplies Q S and K S T to obtain the small-scale spatial attention matrix Z S S = Q S K S T R P 2 × P 2 . The larger the value in Z S S , the more spatially informative the position is for small-scale features. Then, Q L and K L T are multiplied to obtain the large-scale spatial attention matrix Z L S = Q L K L T R P 2 × P 2 . The larger the value in Z L S , the more spatially informative the position is for large-scale features. Finally, Z S S and Z L S are added to obtain the cross-scale spatial attention matrix Z S = Z S S + Z L S R P 2 × P 2 . To better describe spatial structures, the relative position encoding R P R P × P × D is introduced into Z S . In addition, a S o f t m a x function is used for normalization to obtain the cross-scale spatial dependency matrix G S R P 2 × P 2 , as follows:
G S = S o f t m a x Q S R P T + Q S K S T + Q L K L T D
On the channel dimension, CSD2A first multiplies Q S T and K S to obtain the small-scale channel attention matrix Z S C = Q S T K S R D × D . The larger the value in Z S C , the more channel-informative the position is for small-scale features. Second, Q L T and K L are multiplied to obtain the large-scale channel attention matrix Z L C = Q L T K L R D × D . The larger the value in Z L C , the more channel-informative the position is for large-scale features. Then, Z S C and Z L C are added to obtain the cross-scale channel attention matrix Z C = Z S C + Z L C R D × D . Meanwhile, a S o f t m a x function is used for normalization to obtain the cross-scale channel dependency matrix G c R D × D , which can be formulated as follows:
G C = S o f t m a x Q S T K S + Q L T K L D
Finally, CSD2A multiplies G S , V , and G C to obtain its output, as follows:
A h = G S V G C R P × P × D
Next, the outputs of all heads are concatenated, and a linear projection layer is used to obtain the output of MHCSD2A. MHCSD2A with H heads can be formulated as follows:
M H C S D 2 A = C o n c a t A 1 , A 2 , , A H W o
where W o R C × C represents a learnable linear projection matrix.
According to the above descriptions, MHCSD2A ensures that the fusion result can preserve small-scale detail features while integrating large-scale semantic information. This helps CPMFFormer to form a hierarchical and progressive spatial representation.
Compared with existing multiscale fusion strategies, the progressive multiscale fusion strategy not only avoids semantic conflicts between features with large-scale differences, but also maintains the continuity of semantic hierarchy across different scales. This enhances the representational power and classification robustness of CPMFFormer.

3.5. Loss Function

The loss function is used to quantify the difference between the predicted label and the actual label. By minimizing the loss function, a deep learning network can gradually adjust its parameters to make the predicted label as close to the actual label as possible. The common loss function in HSI classification is the cross-entropy loss L C E , as follows:
L C E = 1 M i = 1 M l o g p i , l i
where p i , l i is the predicted probability that the ith pixel belongs to the lith class.
The total loss L of CPMFFormer includes the class consistency loss in CSWM and the cross-entropy loss for classification, which can be formulated as follows:
L = L C E + τ L C
where τ is a balance parameter, which is set to 10 by trial and error.

3.6. Illustrative Example

To show CPMFFormer more clearly, this section takes the Indian dataset of size 145 × 145 × 200 as an example to analyze its calculation process in detail. Let the input patch size be 11 × 11 × 200 and the feature channel dimension be 128. As shown in Figure 4, a 1 × 1 convolutional layer following the weighted patch is used to adjust the feature channel dimension. It can extract features of size 11 × 11 × 128 . Moreover, this feature is used as the input of CRCM 3 × 3 . The implementation processes of different modules are shown in Table 2, Table 3 and Table 4, respectively.
For ease of description, the group convolutional layer, batch normalization, and ReLU activation function are abbreviated as GConv-BN-ReLU. The convolutional layer is abbreviated as Conv. This section only shows the implementation process of CRCM (3 × 3). The implementation processes of CRCM for other scale features extraction are similar to CRCM (3 × 3), and only the group convolutional kernel size needs to be changed according to Table 3. In Table 4, CFEB and the first Conv have two output sizes. They correspond to small-scale features and large-scale features, respectively.
Finally, a GAP layer is used to compress the output features of size 11 × 11 × 128 into a feature vector of size 1 × 1 × 128. Moreover, an FC layer and a Softmax function are used to obtain the predicted class labels. In addition, the optimization solution is performed according to the loss function until the network converges. Based on the above descriptions, the implementation process of CPMFFormer is shown in Algorithm 1.
Algorithm 1: The implementation process of CPMFFormer.
Input: HSI data, ground truth, training ratio, patch size, feature channel dimension, batch size, learning rate, the number of epochs.
Output: The predicted class labels on test samples.
1. Cropping the training, validation, and test patches from the HSI data.
2. Generating the training, validation, and test labels from the ground truth.
3. For e = 1 to epoches
4.  Using the CSWM to obtain the weighted input patch.
5.  Using a 1 × 1 convolutional layer to adjust the feature channels dimension.
6.  Using CRCM to extract the features at different scales.
7.  Using CSFBT to gradually fuse adjacent scale features.
8.  Using a GAP layer, an FC layer, and a S o f t m a x function to predict the class label.
9.  Updating the network parameters using training samples.
10.  Calculating loss function using validation samples. If loss function is smaller than of last epoch, the network parameter is saved. Conversely, it will not be saved.
11. End for
12. Predict the class labels on the test set using the best network.

4. Experiments and Analysis

4.1. Dataset’s Description

To validate the effectiveness of CPMFFormer, we use five public HSI datasets to carry out the subsequent experiments, including the Indian pines dataset, Houston dataset, WHU-Hi-HongHu dataset, Pavia University dataset, and Salinas dataset. These datasets were acquired using different sensors in multiple countries and regions.
Indian pines (Indian) dataset: This dataset was taken by AVIRIS sensor on Indian pines in northwestern Indiana, USA. It contains 145 × 145 pixels and 200 spectral bands. Its ground truth provides 10,249 labeled pixels for 16 ground objects. The types of ground objects and sample distributions in the Indian dataset are shown in Table 5.
Houston dataset: This dataset was taken by ITRES CASI-1500 sensor on the Houston university in Houston, TX, USA. It contains 349 × 1905 pixels and 144 spectral bands. Its ground truth provides 15,029 labeled pixels for 15 ground objects. The types of ground objects and sample distributions in the Houston dataset are shown in Table 6.
WHU-Hi-HongHu (HongHu) dataset: This dataset was taken by Headwall Nano-Hyperspec sensor on Honghu city, China. It contains 940 × 475 pixels and 270 spectral bands. Its ground truth provides 386,693 labeled pixels for 22 ground objects. The types of ground objects and sample distributions in the HongHu dataset are shown in Table 7.
Pavia University (PaviaU) dataset: This dataset was taken by ROSIS sensor on Pavia University in Pavia, Italy. It contains 610 × 340 pixels and 103 spectral bands. Its ground truth provides 42,776 labeled pixels for 9 ground objects. The types of ground objects and sample distributions in the PaviaU dataset are shown in Table 8.
Salinas dataset: This dataset was taken by AVIRIS sensor on Salinas Valley in California, USA. It contains 512 × 217 pixels and 204 spectral bands. Its ground truth provides 54,129 labeled pixels for 16 ground objects. The types of ground objects and sample distributions in the Salinas dataset are shown in Table 9.

4.2. Experimental Setup

Experiment Overview: To comprehensively validate the advancement and effectiveness of CPMFFormer, this section conducts the following experiments:
(1)
Parameters sensitivity analysis: This part of the experiment is used to explore the impact of important hyperparameters on network learning behavior and classification performance, which determines the appropriate parameter setting range.
(2)
Quantitative comparison: This part of the experiment is used to compare the classification performance obtained by the proposed network and existing methods on different datasets, which validates the competitiveness of CPMFFormer.
(3)
Full-scene visualization: This part of the experiment uses the complete classification results to evaluate spatial coherence and boundary retention, which intuitively reflects the expression capabilities of different networks for complex ground objects.
(4)
Network stability evaluation: This part of the experiment uses different training samples sizes to evaluate the classification performance changes in the proposed network and existing methods, which validates the reliability of CPMFFormer.
(5)
Network complexity and inference efficiency: This part of the experiment is used to measure the computational resource requirements and running speed of different networks, which evaluates the practical application prospects of CPMFFormer.
(6)
Ablation studies: This part of the experiment is mainly used to analyze the specific functions and independent contributions of each core design, which clarifies the role of different components in the feature extraction and fusion process.
(7)
Expandability analysis: This part of the experiment analyzes the value of semi-supervised learning and its improvement on classification performance, which provides a theoretical and experimental foundation for future research work.
Evaluation Metrics: In the following experiments, we use three common metrics to evaluate the classification performance of CPMFFormer, including the overall accuracy (OA), average accuracy (AA), and Kappa ( κ ) coefficient. OA is the ratio of the correctly  classified samples to all labeled samples. AA represents the average  classification accuracy across all classes. κ  measures the consistency between the predicted result  and ground truth. Furthermore, we adopt the Moran’s I (MI), boundary F1 score  (BF1), and structural similarity index measure (SSIM) to evaluate the spatial  coherence and boundary retention of the classification results of different  networks.
Implementation Details: All codes are implemented on the PyTorch platform. The basic hardware configurations are Intel Xeon Platinum 8358 CPU 2.60 GHz and NVIDIA A800 GPU 80 GB. The basic software configurations are Python 3.8, PyTorch 2.4, and CUDA v11.8. CPMFFormer is trained using an Adam optimizer. The initial learning rate is set to 0.003, the decay rate is (0.9, 0.999), and the fuzziness factor is 1e−8. The training epoch is set at 200 and the learning rate is dynamically adjusted every 15 epochs using cosine annealing. The comparison networks are set using the optimal parameters based on the recommendations in their papers and are trained to achieve the best classification performance using five public HSI datasets to the maximum extent possible. We repeat the experiment five times on each HSI dataset to eliminate the impact caused by the randomness and report the average classification performance ± standard deviation.

4.3. Parameters Sensitivity Analysis

The sensitivity analysis for the batch size: The batch size affects network training stability, convergence speed, and classification performance. A small batch size may introduce more gradient noise and lead to training instability. A large batch size may reduce the generalization performance of the classification network due to the dilution caused by the data diversity in each batch. As shown in Figure 11, to determine the optimal batch size, this paper sets the batch size to 16, 32, 48, 64, 80, and 96. According to the experimental results, when the batch size is 48, the Indian, Houston, HongHu, and PaviaU datasets reach the optimal evaluation metrics. Moreover, when the batch size is 48 or 64, the PaviaU dataset exhibits similar classification performance. In subsequent experiments, this paper uniformly sets the batch size to 48 for all HSI datasets.
The sensitivity analysis for the learning rate: The learning rate determines the update step of network parameters and plays a crucial role in the stability and convergence of network training. A small learning rate results in slow convergence and makes it difficult for the classification network to obtain optimal parameters. A large learning rate may lead to oscillation or even divergence in the training process, thereby reducing the classification performance. As shown in Figure 12, to determine the optimal learning rate, this paper sets the learning rate to 0.0003, 0.001, 0.003, 0.01, and 0.03. The experimental results show that when the learning rate is 0.003, the Indian and Houston datasets have the best classification performance. In contrast, the learning rate has a relatively small impact on the other three HSI datasets. In subsequent experiments, this paper uniformly sets the learning rate to 0.003 for all HSI datasets.
The sensitivity analysis for the number of CRCM: The number of CRCMs is related to the spatial information captured by the classification network. Fewer CRCMs make it difficult for CPMFFormer to effectively suppress interference pixels. More CRCMs may introduce redundant information and increase the computational cost. As shown in Table 10, to determine the optimal number of CRCM, this paper sets the number of CRCM to 2, 3, 4, 5, and 6. The experimental results show that when the number of CRCMs is 4, the Indian dataset achieves the best OA and AA. The Houston, HongHu, and PaviaU datasets reach the optimal evaluation metrics. The Salinas dataset has the optimal OA and κ When the  number of CRCMs is 5, the Indian dataset achieves the best κ . The Salinas  dataset achieves optimal evaluation metrics. In subsequent experiments, this  paper uniformly sets the number of CRCM to 4 for all HSI datasets.
The sensitivity analysis for the number of bottleneck residual blocks within CRCM: The number of bottleneck residual blocks within CRCM affects the efficiency of feature extraction. Fewer bottleneck residual blocks make it difficult for CRCM to extract high-level spatial semantics. More bottleneck residual blocks increase the computational cost, and cause feature redundancy. As shown in Table 11, to determine the optimal number of bottleneck residual blocks within CRCM, this paper sets the number of bottleneck residual blocks within CRCM to 1, 2, 3, 4, and 5. When the number of bottleneck residual blocks within CRCM is 2, the Indian, HongHu, PaviaU, and Salinas datasets reach the best classification performance. The Houston dataset has the optimal AA. When the number of bottleneck residual blocks within CRCM is 1, the Houston dataset achieves the best OA and κ . In subsequent experiments, this paper uniformly sets the number of bottleneck residual blocks within CRCM to 2 for all HSI datasets.
The sensitivity analysis for the patch size: The patch size P is related to the balance between the capture range of spatial information and the introduction of interference information. A small patch size may lose some representative spatial information and make it difficult for the classification network to extract discriminative spatial features. A large patch size may introduce redundant or erroneous spatial information that is not conducive to the classification decision of the center pixel. As shown in Figure 13, to determine the optimal patch size, this paper sets P to 9, 11, 13, 15, 17, 19, and 21. It can be seen from the experimental results that when P = 11 , the Indian, Houston, and PaviaU datasets reach the optimal evaluation metrics. When P = 13 , the HongHu and Salinas datasets have the best classification performance.
The sensitivity analysis for the feature channel dimension: The feature channel dimension C is related to the feature representation capability and computational efficiency. A small feature channel dimension makes it difficult for the classification network to effectively represent different ground objects. A large feature channel dimension may cause the classification network to learn redundant semantic features and increase the computational cost. As shown in Table 12, to determine the optimal feature channel dimension, this paper sets C to 32, 64, 96, 128, 160, and 192. It can be seen from the experimental results that when C = 128 , the Indian dataset has the best OA and κ . The PaviaU dataset has the best classification performance. When C = 192 , the Indian dataset reaches the optimal AA. Considering the computational cost, this paper sets the feature channel dimension to 128 for the Indian dataset. When C = 64 , the Houston, HongHu, and Salinas datasets reach the optimal evaluation metrics.

4.4. Quantitative Comparison

In this section, CPMFFormer is compared with three CNN-based classification networks (S3ARN [29], PMCN [34], and S3GAN [36]), two transformer-based classification networks (ViT [45] and SFormer [46]), and ten CNN-Transformer-based classification networks (HFormer [47], MHCFormer [35], D2S2BoT [48], SQSFormer [40], EATN [41], CFormer [42], SSACFormer [43], CPFormer [44], DSFormer [49], LGDRNet [50]).
The classification performances obtained by different networks on five HSI datasets are shown in Table 13, Table 14, Table 15, Table 16, Table 17, Table 18, Table 19, Table 20, Table 21 and Table 22. We report the class accuracy (CA), OA, AA, κ , and their standard deviation. Overall, CNN-Transformer-based classification networks outperform other networks. This is mainly because they integrate complementary characteristics of CNN and transformer. CPMFFormer achieves the optimal or suboptimal CA from most classes. Furthermore, its OA, AA, and κ are consistently optimal across different datasets.
As shown in Table 13 and Table 14, using the Indian dataset, CPMFFormer achieves the optimal CA in three classes and the suboptimal CA in two classes. It has the optimal OA, AA, and κ , which are 94.85%, 94.35%, and 94.13%, respectively. CPFormer is second only to CPMFFormer in OA and κ , with values of 94.08% and 93.26%, respectively. The differences between CPMFFormer and CPFormer in OA and κ are 0.77% and 0.87%, respectively. DSFormer is second only to CPMFFormer in AA, with a value of 90.76%. The difference between CPMFFormer and DSFormer in AA is 3.59%. CPFormer is a center attention network. However, it ignores multiscale spatial features. This makes it difficult for CPFormer to model diverse spatial distribution patterns. DSFormer can flexibly select and fuse features at different scales. However, its limitations in multiscale fusion strategy reduce the feature representation capability. Furthermore, SSACFormer is the best performing classification network in CA. It achieves the optimal CA in six classes and the suboptimal CA in one class. However, its CAs in class 7 and class 9 are poor. This is because these two classes only provide one training sample. Although SSACFormer can capture subtle spectral differences between ground objects by extracting multiscale spectral features, it ignores class information. This makes it difficult for SSACFormer to deal with inter-class spectral overlap caused by spectral variability with fewer training samples.
As shown in Table 15 and Table 16, using the Houston dataset, CPMFFormer achieves the suboptimal CA in six classes. For class 1, class 5, class 11, and class 14, the differences between the CA of CPMFFormer and the optimal CA are only 0.56%, 0.06%, 0.9%, and 0.56%, respectively. CPMFFormer has the best evaluation metrics, with OA, AA, and κ of 95.93%, 96.06%, and 95.61%, respectively. CPFormer is second only to CPMFFormer in OA and κ , with values of 94.81% and 94.39%, respectively. The differences between CPMFFormer and CPFormer in OA and κ are 1.12% and 1.22%, respectively. SSACFormer is second only to CPMFFormer in AA, with a value of 94.66%. The differences between CPMFFormer and SSACFormer in AA is 1.4%. DSFormer is the best performing classification network in CA. It achieves the optimal CA in four classes. The ground objects in the Houston dataset are more dispersed. This results in the input patch containing more interference pixels. Because DSFormer lacks a center calibration mechanism, its classification performance is inferior to that of some center attention networks, such as SSACFormer and CPFormer. Moreover, the AA of CFormer is higher than that of DSFormer. SQSFormer is also a center transformer, but its classification performance is poor. This is because it uses only one convolutional layer to learn semantic features. This makes it difficult for SQSFormer to effectively represent different ground objects.
As shown in Table 17 and Table 18, in the HongHu dataset, CPMFFormer achieves the optimal CA in seven classes and the suboptimal CA in eight classes. CPMFFormer is the best performing classification network in CA. For class 3, class 10, class 11, class 14, class 20, and class 22, the differences between the CA of CPMFFormer and the optimal CA are only 0.06%, 0.22%, 0.08%, 0.45%, 0.92%, and 0.38%, respectively. CPMFFormer has the optimal evaluation metrics, with OA, AA, and κ of 99.53%, 98.99%, and 99.41%, respectively. SSACFormer is second only to CPMFFormer in OA, AA, and κ , with values of 99.43%, 98.74%, and 99.28%, respectively. The differences between CPMFFormer and SSACFormer in OA, AA, and κ are 0.1%, 0.25%, and 0.13%, respectively. Since each ground object in the HongHu dataset has a wide distribution range, it is difficult for single-scale features to describe complex spatial relationship between pixels. Compared with other methods, multiscale classification networks perform better, such as S3ARN, PMCN, S3GAN, HFormer, D2S2BoT, CFormer, SSACFormer, and LGDRNet. MHCFormer is also a multiscale classification network, but its classification performance is poor. This is because the distributions of ground objects in the HongHu dataset are relatively continuous. MHCFormer transforms spatial semantic features into the Fourier domain to focus more on long-range spatial dependencies between high-frequency information. This makes it difficult for MHCFormer to effectively characterize different ground objects in the HongHu dataset, thereby affecting its classification performance.
As shown in Table 19 and Table 20, using the PaviaU dataset, CPMFFormer achieves suboptimal CA in four classes. Moreover, for class 2, class 4, class 5, and class 7, the differences between the CA of CPMFFormer and the optimal CA are 0.04%, 0.94%, 0.55%, and 0.15%, respectively. CPMFFormer has the best classification performance, with OA, AA, and κ of 99.50%, 99.02%, and 99.34%, respectively. S3GAN is second only to CPMFFormer in OA and κ , with values of 99.13% and 98.84%, respectively. The differences between CPMFFormer and S3GAN in OA and κ are 0.37% and 0.5%, respectively. DSFormer is second only to CPMFFormer in AA, with a value of 98.48%. The difference between CPMFFormer and DSFormer in AA is 0.54%. S3GAN is a multiscale classification network. However, it uses the concatenation operation to directly fuse features from all scales. This may lead to information confusion and semantic conflict. Therefore, the classification performance of S3GAN is inferior to that of CPMFFormer. Compared with the HongHu dataset, the PaviaU dataset has more high-frequency detail information. This enables MHCFormer to achieve a better classification performance.
As shown in Table 21 and Table 22, using the Salinas dataset, CPMFFormer achieves the optimal CA in eight classes and the suboptimal CA in two classes. For class 1, class 4, class 5, class 7, class 12, and class 16, the differences between the CA of CPMFFormer and the optimal CA are only 0.04%, 0.39%, 0.14%, 0.03%, 0.26%, and 0.3%, respectively. CPMFFormer has optimal evaluation metrics, with OA, AA, and κ of 99.84%, 99.84%, and 99.82%, respectively. S3GAN is second only to CPMFFormer in OA and κ , with values of 99.59% and 99.54%, respectively. The differences between CPMFFormer and S3GAN in OA and κ are 0.25% and 0.28%, respectively. D2S2BoT is second only to CPMFFormer in AA, with a value of 99.62%. The differences between CPMFFormer and D2S2BoT in AA is 0.22%. D2S2BoT and S3GAN are both multiscale classification networks. However, their limitations in the multiscale feature fusion strategy and their neglect of the center feature calibration mechanism make their classification performance inferior to that of CPMFFormer. Similar to the HongHu dataset, different ground objects in the Salinas dataset have a larger distribution range, and their distribution is relatively continuous. Compared with other methods, most multiscale classification networks exhibit better classification performance, such as HFormer, CFormer, SSACFormer, and DSFormer.
According to the experimental results, the classification performance obtained by different networks using the Indian and Houston datasets is significantly worse than that using the other three datasets. This performance difference mainly stems from multiple factors. On one hand, this difference is closely related to the inherent characteristics of the dataset. Specifically, in the Indian dataset, some ground objects (e.g., Corn-Notill, Corn-Mintill, and Corn) have highly similar spectral characteristics and significant intra-class spectral variability. This leads to a severe inter-class spectral overlap. As shown in Figure 1, samples in different corn-related classes share very similar spectral distributions. This makes the class separation more challenging. In the Houston dataset, most ground objects are spatially scattered. The input patches inevitably contain more background noise and heterogeneous pixels. Such interference information affects the representation of the center pixel and degrades classification stability. In contrast, ground objects in the HongHu, PaviaU, and Salinas datasets are more densely distributed and have distinct spectral differences. This makes it easier for the classification network to learn discriminative spectral–spatial features to characterize different ground objects.
On the other hand, the insufficient training samples in the Indian and Houston datasets further amplifies this performance difference. According to Table 5, Table 6, Table 7, Table 8 and Table 9, the Indian and Houston datasets only provide 212 and 308 training samples, respectively, while the HongHu, PaviaU, and Salinas datasets provide 7748, 858, and 1092 training samples, respectively. In the Indian dataset, some ground objects (e.g., Alfalfa, Corn, Grass-Pasture-Mowed, Oats, Wheat, and Stone-Steel-Towers) have relatively few training samples (only 1–5). In the Houston dataset, some ground objects (e.g., Water, Parking Lot2, and Tennis Court) have fewer than 10 training samples. Fewer training samples limits the network’s capability to comprehensively capture the high-level spectral–spatial features specific to these ground objects. This makes it difficult for the classification network to construct feature patterns that accurately distinguish ground objects with fewer training samples from other ground objects, thereby reducing classification accuracy. In contrast, the HongHu, PaviaU, and Salinas datasets have relatively more training samples for each ground objects. This allows the classification networks to learn stable and universal discriminative spectral–spatial features, thereby improving classification performance.
Nevertheless, CPMFFormer significantly outperforms other classification networks in relation to the Indian and Houston datasets. This indicates that CPMFFormer is more robust to spectral variability and spatial interference. Moreover, it can effectively reduce the impact of small sample size and imbalanced sample distributions.

4.5. Full-Scene Visualization

To further evaluate spatial coherence and boundary retention of the classification results obtained by different networks, this section conducts full-scene classification experiments. We use the classification network to predict the class label of all pixels. The full-scene classification results obtained by different networks from five HSI datasets are shown in Figure 14, Figure 15, Figure 16, Figure 17 and Figure 18. The full-scene classification performance obtained by different networks from five HSI datasets is shown in Table 23 and Table 24.
According to the experimental results, it can be seen that the full-scene classification results generated by S3ARN, HFormer, MHCFormer, and D2S2BoT are significantly worse than those of other classification networks. Although they achieve a better classification performance in Section 4.4, they perform poorly in the full-scene classification experiment. This is mainly because these models only consider the spectral–spatial features within the sample patch, while ignoring the spatial distribution differences between adjacent ground objects. Although these models can extract multiscale spatial features to better characterize complex spatial relationships and model the diverse spatial distribution patterns presented by ground objects, their limitations in the multiscale features fusion strategy may lead to information confusion and semantic conflicts. This reduces the spatial features representational capability. These networks lack the center feature calibration mechanism. This makes it difficult for them to suppress interference information and causes them to be prone to generating misclassified pixels.
In contrast, other classification networks can generate smoother full-scene classification results. Especially for the center attention networks, they perform better than other networks. This is mainly because these networks can calculate dependencies between the center pixel and its neighboring pixels. This helps them to dynamically select the valuable neighboring spectral–spatial features and effectively prevent the transmission and amplification of interference information. For the ground object boundaries, the center attention mechanism can accurately capture characteristic differences between the center pixels and their surrounding regions. This avoids class leakage and boundary blurring. These advantages enable the center attention networks to maintain stable and high-precision classification performance in full-scene classification tasks.
Compared with comparison networks, CPMFFormer can capture subtle spectral differences between ground objects. Moreover, it can hierarchically enhance discriminative spatial features at different scales. This is beneficial for reducing misclassified pixels and ensuring spatial coherence in the full-scene classification. Furthermore, CPMFFormer achieves a smooth semantic transition through the progressive multiscale fusion strategy. It can better model spatial correlations and enhance the spatial feature representational capability. As shown in Table 23 and Table 24, CPMFFormer achieved the best full-scene classification performance from most HSI datasets. Only its BF1 from the PaviaU dataset is slightly worse than that of SFormer. This is mainly because the PaviaU dataset contains a large amount of detailed information, which results in the input patch containing more interference pixels. This makes it easy for the classification network to learn incorrect spatial features. SFormer focuses more on spectral long-range dependence and has a relatively low dependence on spatial neighboring information. Its classification result is less affected by interference pixels, so it has better boundary retention performance. However, this also makes SFormer prone to generating non-smooth classification results. Therefore, SFormer has a lower spatial coherence on the PaviaU dataset.

4.6. Network Stability Evaluation

Network stability is an important indicator for evaluating classification performance. It ensures that the classification network can operate reliably and efficiently in practical applications. To validate the stability of CPMFFormer, this section sets the training sample sizes to 0.2%, 0.4%, 0.6%, 0.8%, 1%, 2%, 4%, 6%, 8%, 10%, 15%, and 20% of all label samples to analyze the performance changes in different networks. The experimental results obtained by different networks with different training sample sizes from five public HSI datasets are shown in Figure 19, Figure 20, Figure 21, Figure 22 and Figure 23.
It can be seen that under the impact of different training sample sizes, the performance of most classification networks on all HSI datasets generally shows a reasonable trend, that is, as the training sample size increases, their classification performance gradually improves. When there are sufficient training samples, most networks almost achieve optimal classification performance on five HSI datasets.
The experimental results show that compared with other classification networks, the performance curve of CPMFFormer converges faster in five HSI datasets. This demonstrates that CPMFFormer exhibits stronger stability. Furthermore, it is worth noting that at any training sample size, CPMFFormer performs better than other classification networks on all HSI datasets. Especially when there are fewer training samples (such as when the training sample size is less than 1%), the advantage of CPMFFormer in classification performance is more obvious. The above analyses indicate that the proposed classification network has better generalization performance and practicality.

4.7. Network Complexity and Inference Efficiency

Computational complexity affects training efficiency and resource consumption of the classification network. As shown in Table 25 and Table 26, this section counts the trainable parameters, computational cost, training time, and testing time of different classification networks on five HSI datasets. The trainable parameters are in millions ( M ), the computational cost is in G F L O P s , and the training and testing times are in seconds ( s ).
As shown in Table 25 and Table 26, although this paper designs multiple modules to improve the feature representation capability, the computational complexity of CPMFFormer is not the highest. This is due to our optimization of the network framework. For example, this paper introduces the bottleneck structure into CRCM and CSFBT. This reduces the feature channels’ dimension. Moreover, when extracting features at different scales using group convolution, this paper adaptively allocates the feature channels’ dimension in each group based on the convolutional kernel size. These designs reduce the computational complexity of CPMFFormer. Furthermore, it is undeniable that some classification networks have smaller computational complexity than our network. However, it is worth noting that the difference in computational complexity between CPMFFormer and these classification networks is small, but the classification performance of CPMFFormer is much better than theirs. This demonstrates that CPMFFormer achieves a good compromise between computational complexity and classification performance.

4.8. Ablation Study

To analyze the impact of some important contributions in CPMFFormer on classification performance, this section carries out the ablation study from three aspects, including the ablation study on the overall framework, the ablation study on the spectral weighting mechanism, and the ablation study on the attention mechanism.
Ablation Study on the Overall Framework: CPMFFormer constructs three main modules to enhance the feature representation capability and improve classification performance, including CSWM, CRCM, and CSFBT. To validate their effectiveness, this section removes CSWM, CRCM, and CSFBT to carry out experiments.
As shown in Figure 24, after removing each module, the classification performance of CPMFFormer decreases to varying degrees. This shows that these modules are effective in improving classification performance. Specifically, after removing CSWM, CPMFFormer focuses more on extracting multiscale spatial features and ignores spectral features. This makes it difficult for CPMFFormer to reveal subtle differences in the spectral radiation characteristics of different ground objects. Moreover, it is worth noting that after removing CRCM, CSFBT used for multiscale features’ fusion is also removed because there is no multiscale feature extraction. In this case, CPMFFormer only depends on CSWM to extract class-discriminative spectral features but cannot obtain high-level spatial semantic features to model complex spatial relationships. Therefore, the classification performance after removing CRCM is significantly worse than other experiments. After removing CSFBT, CPMFFormer mainly uses the output features of the last CRCM to predict the class probabilities. This makes CPMFFormer neglect complementary features from other scales and reduce the feature representation capability. Removing both CSWM and CSFBT causes CPMFFormer to lack spectral awareness capability and have difficulty in capturing cross-scale complementary information. The classification performance is significantly worse than that obtained after separately removing CSWM and CSFBT. This demonstrates the synergistic and promoting effects between different modules.
This paper builds a progressive multiscale fusion strategy to promote the effective collaboration between different scale features. This achieves a smooth semantic transition by gradually fusing adjacent scale features, while avoiding information confusion and semantic conflict between features with large-scale differences. As shown in Figure 4, CPMFFormer uses six CSFBTs to enhance scale-specific information. To validate their effectiveness, this section uses the progressive multiscale fusion strategy to fuse features at different scales shown in Figure 2. The activation response heatmaps obtained by different CSFBTs on the PaviaU dataset are shown in Figure 25. The experimental results show that as the fusion level increases, the progressive multiscale fusion strategy gradually integrates semantic information specific to different scales into the fusion features. As shown in Figure 25f, the fusion features obtained by CSFBT-6 not only retain the larger activation response in small-scale features to the Trees class (red box) but also retain the larger activation response in large-scale features to the Painted Metal Sheets class (white box).
To comprehensively analyze the effectiveness of the progressive multiscale fusion strategy, this paper visualizes the activation responses obtained by current multiscale fusion methods, including concatenation, element-wise addition, and attention method in the literature [35]. As shown in Figure 26, it is difficult for concatenation and element-wise additions to effectively integrate complementary information at different scales. In contrast, the attention method can enhance the activation responses to the Trees and Painted Metal Sheets classes to a certain extent. However, compared with the fusion features obtained by the progressive multiscale fusion strategy, the fusion features obtained by the attention method have smaller activation responses to different ground objects. This reduces the features representation capability and affects classification performance.
Ablation Study on the Spectral Weighting Mechanism: To obtain the class-specific spectral weights, CSWM introduces an aggregation operation and a class consistency loss into spectral weight learning (SWL). To validate its effectiveness, we remove the aggregation operation and class consistency loss in CSWM and only use SWL to learn spectral weights. As shown in Figure 27, after removing the aggregation operation and class consistency loss function, the classification performance of CPMFFormer decreases. This is mainly because spectral variability causes the spectral distribution characteristics of different ground objects to be approximately the same. SWL only relies on spectral information to learn spectral weights. This makes it difficult for CPMFFormer to distinguish samples with similar spectral characteristics, but which not belong to the same class.
As shown in Figure 1, the spectral distribution characteristics of the Corn-Notill class, Corn-Mintill class, and Corn class in the Indian dataset overlap with each other. To further validate the effectiveness of CSWM, this section visualizes the average spectral weights learned by SWL and CSWM for the Corn-Notill class, Corn-Mintill class, and Corn class, as shown in Figure 28. It can be seen from the visualization results that since SWL lacks class information, the average spectral weights learned by SWL for the Corn-Notill class, Corn-Mintill class, and Corn class are basically the same. This makes it difficult for SWL to obtain discriminative spectral features that are beneficial for characterizing different ground objects with similar spectra. In contrast, the average spectral weights learned by CSWM for the Corn-Notill class, Corn-Mintill class, and Corn class have obvious differences. This enhances class-discriminative spectral features and increases inter-class spectral differences. This enables the proposed classification network to effectively deal with inter-class spectral overlap and accurately distinguish samples with similar spectra, but which belong to different classes, thereby improving classification performance.
Ablation Study on the Attention Mechanism: This paper designs three different attention mechanisms to enhance the feature representation capability, including CFCL, CFEB, and MHCSD2A. MHCSD2A is composed of a cross-scale spatial attention (CSSA) and a cross-scale channel attention (CSCA). To validate their effectiveness, this paper removes different attention mechanisms to carry out experiments. As shown in Table 27, after removing each attention mechanism, the classification performance of CPMFFormer decreases to varying degrees. This shows that these attention mechanisms are effective in enhancing the feature representation capability and improving classification performance.

4.9. Expandability Analysis

As shown in Figure 19, Figure 20, Figure 21, Figure 22 and Figure 23, the training sample size has a significant impact on classification performance. Taking the Indian dataset as an example, when the training sample size is 0.2%, the OA, AA, and κ  of CPMFFormer from the Indian dataset are 63.50%,  68.61%, and 58.26%, respectively. When the training sample size is 20%, the OA,  AA, and κ  of CPMFFormer from the Indian dataset are 99.84%,  99.66%, and 99.82%, respectively. In other words, when there are sufficient training samples, the classification network can usually achieve optimal classification performance. However, in practical applications, the data labeling cost is relatively high. To obtain better classification performance with limited labeled samples, existing research typically introduces semi-supervised learning [52]. Semi-supervised learning is a paradigm that uses both labeled and unlabeled samples for network training. In current research, FixMatch is a representative semi-supervised learning strategy based on pseudo-labels and consistency regularization [53].
To analyze the expandability of CPMFFormer, this section introduces FixMatch into the proposed network and constructs a semi-supervised HSI classification framework that combines labeled and unlabeled samples. The experimental results are shown in Figure 29. In the semi-supervised CPMFFormer, we first use Equation (29) to calculate the supervised loss based on labeled samples. Then, weak and strong augmentations are applied to unlabeled samples to construct the consistency regularization mechanism. Meanwhile, pseudo-labels are generated by predicting the weakly enhanced samples. When the prediction confidence is higher than a preset threshold, the pseudo-label is retained. This section sets the threshold to 0.95. Furthermore, the semi-supervised loss is calculated using the corresponding strongly augmented samples and pseudo-labels according to Equation (29), thereby guiding CPMFFormer to maintain stable predictions on unlabeled data. Finally, the supervised loss and semi-supervised loss are weighted to optimize the network parameters. In this section, the weight is set to 1.
Experimental results show that the classification performance of CPMFFormer is improved across most training sample sizes after introducing FixMatch. This demonstrates that semi-supervised learning can effectively mine spectral–spatial structure information in unlabeled samples. This is beneficial for enhancing the discriminative power and generalization capability of the classification network. When labeled samples are limited, the decision boundary of the supervised network tends to be biased towards the labeled class. The consistency constraint of FixMatch provides additional manifold structure information through unlabeled samples. This guides the classification network to learn a smoother decision boundary in the feature space that better fits the data distribution, thereby improving the stability and robustness of the classification network. Meanwhile, semi-supervised learning can significantly reduce labeling costs. The pixel-level labeling of HSIs typically requires expert involvement and extensive manual review. Semi-supervised learning frameworks can fully use massive unlabeled samples. This makes the classification network more practical in real-world remote sensing scenarios.
Furthermore, we observe that when the training sample sizes are 0.6%, 0.8%, and 1%, the classification performance after introducing FixMatch is slightly lower than that of CPMFFormer. The main reason is likely the use of a fixed high confidence threshold. This may lead to the predictions of CPMFFormer for unlabeled samples being just enough to pass through a large number of pseudo-labels at these three training sample sizes, while the supervised training is insufficient to guarantee the quality of pseudo-labels. This will cause many incorrect pseudo-labels to be used in network training, thereby generating negative transfer and affecting classification performance.

5. Discussion

This section mainly discusses the usage, AI properties and interpretability limitations, adaptive range, and future optimization directions of CPMFFormer, as follows:
Usage: CPMFFormer is suitable for HSI classification tasks that require handling significant spectral variability and complex multiscale spatial structure. It is particularly valuable in the following application scenarios:
(1)
Agricultural fine-grained classification and crop monitoring: By leveraging the capability of CSWM in distinguishing corps with similar spectra but which belong to different classes, CPMFFormer can identify adjacent crop species and field boundaries across large regions of cultivated land. This provides reliable support for crop planting structure surveys, agricultural monitoring, and yield estimation.
(2)
Urban and rural land object recognition: Using CRCM and the progressive multiscale fusion strategy, CPMFFormer can identify multiple types of ground objects in urban and suburban regions with complex spectral–spatial distributions, such as roads, rooftops, green plants, water bodies, etc. It is suitable for urban planning, change detection, and land use monitoring.
(3)
Fine-grained mapping for wetlands and ecological environments: For regions where wetlands, lakes, and farmland intersect, CPMFFormer can maintains stable classification results even under strong interference caused by background noise and mixed pixel conditions. This can support fine-grained mapping for ecological protection and environmental monitoring.
(4)
Other hyperspectral application scenarios: For tasks that require high-precision classification in the presence of significant spectral drift, internal heterogeneity of ground objects, or inhomogeneous spatial texture, CPMFFormer also has strong adaptability and practical value, such as forestry resource surveys, mining region distribution identification, geological surveys, etc.
In summary, CPMFFormer is not designed solely for a single public dataset but rather addresses key challenges such as spectral variability and multiscale spatial structure integration commonly encountered in real remote sensing scenes. Therefore, its usage not only covers typical agricultural, urban, and wetland scenarios validated by public HSI datasets, but can also be extended to other ground object fine-classification tasks that require high-discriminative spectral–spatial features.
AI properties and interpretability: The class-aware spectral weighting enables each ground object to learn a differentiated spectral attention pattern. The spectral weight curves are used to visually demonstrate the decision basis of CPMFFormer. This improves its interpretability. Moreover, the center dependency modeling effectively reflects the contribution of neighboring pixels to the classification decision of the center pixel. This semantic saliency-based filtering mechanism further reduces background and inter-class interference. Combined with the progressive multiscale fusion strategy, CPMFFormer obtains a balance between high performance and low-conflict semantics. It achieves adaptive learning decisions for ground object at the AI algorithm level.
Limitations: Although CPMFFormer maintains a lead in most tasks, its performance degrades when the training sample size is small. According to the classification performance obtained by CPMFFormer with different training sample sizes, when the training sample size is only 0.2%, the OA is reduced by approximately 36% compared with when the training sample size is 20%. In the cross-domain scenarios (e.g., the Houston dataset and the WHU-Hi-HongHu dataset), OA decreases by an average of approximately 4% due to differences in sensor and spectral distribution. Moreover, although CPMFFormer balances classification performance and computational complexity through group convolution and the bottleneck structure, it still incurs significant computational and memory pressure when performing classification in large-scale high-resolution scenarios.
Adaptive range: CPMFFormer is suitable for HSI classification tasks with medium or larger training sample size, significant spectral variability and complex spatial structure. When labeled samples are scarce or cross-sensor and cross-region transfer is required, semi-supervised learning (e.g., pseudo-labeling and consistency regularization), domain adaptation (e.g., distribution alignment or feature transformation), and transfer learning can be combined to enhance generalization performance.
Future optimization directions: Future research will focus on further exploring the combination between CPMFFormer and semi-supervised learning to improve robustness and generalization performance of CPMFFormer. According to the experimental results shown in Figure 29, future research can be mainly carried out in the following aspects:
(1)
Adaptive semi-supervised framework: FixMatch-based semi-supervised learning can effectively use unlabeled samples. However, performance fluctuations at some training sample sizes (e.g., 0.6%, 0.8%, and 1%) indicate the need to introduce adaptive mechanisms. Future research will focus on the dynamic confidence threshold, class-balanced pseudo-label selection, and staged unsupervised weighting strategy. This ensures stable training of the classification network at different supervised levels.
(2)
Spectral consistency augmentation: Current data augmentation may disrupt the band continuity of HSIs. This is particularly crucial for preserving spectral features. Future research will focus on designing spectrally friendly data augmentation strategy and extending consistency regularization to the spectral dimension. This can fully use unlabeled samples while preserving physical meaning.
(3)
CFCL-based pseudo-label selection mechanism: A center-dependent pseudo-label selection method will be constructed based on CFCL. By evaluating the spectral–spatial correlation between unlabeled samples and its center pixel, only samples with high reliability and consistency are selected for training. This can effectively suppress interference from noisy pseudo-labels.
(4)
Pre-training and multi-source data fusion: Advanced semi-supervised or self-supervised pre-training methods will be further introduced. By learning a general spectral–spatial representation on large-scale unlabeled data, the initialization performance for subsequent classification tasks can be optimized.
In summary, future work will focus on building a more adaptive and physically consistent semi-supervised learning framework. This will help achieve higher classification accuracy and stronger generalization ability in real remote sensing scenarios.

6. Conclusions

To reduce the impact of spectral variability and enhance the feature representation capability, this paper proposes CPMFFormer. Since it is difficult for current spectral weighting mechanisms to effectively identify samples with similar spectra, but which belong to different classes, this paper introduces class information to enable learning of class-specific spectral weights. Unlike existing center attention mechanisms, this paper combines CFCL with multiscale convolutional layers. This achieves continuous suppression of interference information at different scales. Furthermore, current multiscale fusion strategies directly fuse features with large-scale differences. This easily leads to information confusion and semantic conflict. This paper designs a progressive multiscale fusion strategy to achieve a smooth semantic transition by gradually fusing adjacent scale features. We compared CPMFFormer with fifteen existing networks. Experiments show that CPMFFormer has better classification performance, stronger network stability, and lower computational complexity. We conducted ablation experiments, which show the effectiveness of the overall network framework and the core design in CPMFFormer.

Author Contributions

Conceptualization, M.Z., Y.Y. and D.H.; methodology, M.Z. and Y.Y.; software, M.Z., S.Z. and P.M.; validation M.Z.; writing—original draft preparation, M.Z.; writing—review and editing, Y.Y. and D.H.; funding acquisition, Y.Y. and D.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 62473304 and U22A2045).

Data Availability Statement

The Indian, PaviaU, and Salinas dataset can be obtained at https://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 5 November 2025). The Houston dataset can be obtained at https://github.com/YuxiangZhang-BIT/Data-CSHSI?tab=readme-ov-file (accessed on 5 November 2025). The HongHu dataset can be obtained at http://rsidea.whu.edu.cn/resource_WHUHi_sharing.htm (accessed on 5 November 2025). Our source code will be released at https://github.com/GitDana95/CPMFFormer (accessed on 5 November 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yue, G.; Zhang, L.; Zhou, Y.; Wang, Y.; Xue, Z. S2TNet: Spectral–spatial triplet network for few-shot hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5501705. [Google Scholar] [CrossRef]
  2. Lee, E.; Pan, L.; Li, Z.; Bhattacharyya, S.S. Efficient hyperspectral image classification using discrete cosine transform on limited-resource systems. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 24647–24661. [Google Scholar] [CrossRef]
  3. Bourriz, M.; Hajji, H.; Laamrani, A.; Elbouanani, N.; Abdelali, H.A.; Bourzeix, F.; El-Battay, A.; Amazirh, A.; Chehbouni, A. Integration of hyperspectral imaging and AI techniques for crop type mapping: Present status, trends, and challenges. Remote Sens. 2025, 17, 1574. [Google Scholar] [CrossRef]
  4. Zhang, T.; Xuan, C.; Ma, Y.; Tang, Z.; Gao, X. An efficient and precise dynamic neighbor graph network for crop mapping using unmanned aerial vehicle hyperspectral imagery. Comput. Electron. Agric. 2025, 230, 109838. [Google Scholar] [CrossRef]
  5. García-Vera, Y.E.; Polochè-Arango, A.; Mendivelso-Fajardo, C.A.; Gutiérrez-Bernal, F.J. Hyperspectral image analysis and machine learning techniques for crop disease detection and identification: A review. Sustainability 2024, 16, 6064. [Google Scholar] [CrossRef]
  6. Rahmani, N.; Sekandari, M.; Pour, A.B.; Ranjbar, H.; Carranza, E.J.M. Evaluation of support vector machine classifiers for lithological mapping using PRISMA hyperspectral remote sensing data: Sahand–Bazman magmatic arc, central Iran. Remote Sens. Appl. Soc. Environ. 2025, 37, 101449. [Google Scholar] [CrossRef]
  7. Wang, Y.; He, L.; He, Z.; Chen, J.; Luo, F. A task-oriented framework for efficient lithological mapping of imbalanced categories using hyperspectral imagery. Int. J. Appl. Earth Obs. Geoinf. 2025, 143, 104749. [Google Scholar] [CrossRef]
  8. Tolentino, V.; Ortega Lucero, A.; Koerting, F.; Savinova, E.; Hildebrand, J.C.; Micklethwaite, S. Drone-based VNIR–SWIR hyperspectral imaging for environmental monitoring of a uranium legacy mine site. Drones 2025, 9, 313. [Google Scholar] [CrossRef]
  9. Li, Z.; Liu, T.; Lu, Y.; Tian, J.; Zhang, M.; Zhou, C. Enhanced hyperspectral image classification for coastal wetlands using a hybrid CNN-Transformer approach with cross-attention mechanism. Front. Mar. Sci. 2025, 12, 1613565. [Google Scholar] [CrossRef]
  10. Ceriani, R.; Brocco, S.; Pepe, M.; Oggioni, S.; Vacchiano, G.; Motta, R.; Berretti, R.; Ascoli, D.; Garbarino, M.; Morresi, D. Hyperspectral and LiDAR space-borne data for assessing mountain forest volume and biomass. Int. J. Appl. Earth Obs. Geoinf. 2025, 141, 104614. [Google Scholar] [CrossRef]
  11. Holt, Z.K.; Khan, S.D.; Rodrigues, D.F. Hyperspectral remote sensing as an environmental plastic pollution detection approach to determine occurrence of microplastics in diverse environments. Environ. Pollut. 2025, 377, 126426. [Google Scholar] [CrossRef] [PubMed]
  12. Luque-Söllheim, A.; Martín, J.; Medina, A.; Carrasco-Acosta, M.; García-Jiménez, P.; Ferrer, N.; Bergasa, O. Hyperspectral rapid detection of bacterial content and water quality parameters in coastal bathing waters. Mar. Pollut. Bull. 2026, 222, 118705. [Google Scholar] [CrossRef] [PubMed]
  13. Arias, F.; Zambrano, M.; Galagarza, E.; Broce, K. Mapping harmful algae blooms: The potential of hyperspectral imaging technologies. Remote Sens. 2025, 17, 608. [Google Scholar] [CrossRef]
  14. Ding, S.; Ruan, X.; Yang, J.; Sun, J.; Li, S.; Hu, J. LSSMA: Lightweight spectral–spatial neural architecture with multiattention feature extraction for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 6394–6413. [Google Scholar] [CrossRef]
  15. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  16. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  17. Haut, J.; Paoletti, M.; Paz-Gallardo, A.; Plaza, J.; Plaza, A.; Vigo-Aguiar, J. Cloud implementation of logistic regression for hyperspectral image classification. In Proceedings of the 17th International Conference on Computational and Mathematical Methods in Science and Engineering (CMMSE), Cádiz, Spain, 4–8 July 2017. [Google Scholar]
  18. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. A new deep convolutional neural network for fast hyperspectral image classification. ISPRS J. Photogramm. Remote Sens. 2018, 145, 120–147. [Google Scholar] [CrossRef]
  19. Wei, L.; Ma, H.; Yin, Y.; Geng, C. Kmeans-CM algorithm with spectral angle mapper for hyperspectral image classification. IEEE Access 2023, 11, 26566–26576. [Google Scholar] [CrossRef]
  20. Beirami, B.A.; Mokhtarzade, M. Band grouping SuperPCA for feature extraction and extended morphological profile production from hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1953–1957. [Google Scholar] [CrossRef]
  21. Fang, Y.; Ye, Q.; Sun, L.; Zheng, Y.; Wu, Z. Multiattention joint convolution feature representation with lightweight transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5513814. [Google Scholar] [CrossRef]
  22. Wei, Y.-L.; Zheng, Y.-B.; Wang, R.; Li, H.-C. Quaternion convolutional neural network with EMAP representation for multisource remote-sensing data classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5508805. [Google Scholar] [CrossRef]
  23. Han, Z.; Yang, J.; Gao, L.; Zeng, Z.; Zhang, B.; Chanussot, J. Dual-branch subpixel-guided network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5521813. [Google Scholar] [CrossRef]
  24. Wei, L.; Liu, Y.; Yin, Y.; Gu, J. Improved transformer network based on multiscale grouping feature fusion for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 21305–21325. [Google Scholar] [CrossRef]
  25. Shi, C.; Yue, S.; Wang, L. A dual-branch multiscale transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5504520. [Google Scholar] [CrossRef]
  26. Bai, J.; Shi, W.; Xiao, Z.; Ali, T.A.A.; Ye, F.; Jiao, L. Achieving better category separability for hyperspectral image classification: A spatial–spectral approach. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 9621–9635. [Google Scholar] [CrossRef]
  27. Duan, Y.; Chen, C.; Fu, M.; Li, Y.; Gong, X.; Luo, F. Dimensionality reduction via multiple neighborhood-aware nonlinear collaborative analysis for hyperspectral image classification. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 9356–9370. [Google Scholar] [CrossRef]
  28. Yang, K.; Sun, H.; Zou, C.; Lu, X. Cross-attention spectral–spatial network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518714. [Google Scholar] [CrossRef]
  29. Shu, Z.; Liu, Z.; Zhou, J.; Tang, S.; Yu, Z.; Wu, X.-J. Spatial–spectral split attention residual network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 419–430. [Google Scholar] [CrossRef]
  30. Chhapariya, K.; Buddhiraju, K.M.; Kumar, A. A deep spectral–spatial residual attention network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 15393–15406. [Google Scholar] [CrossRef]
  31. Han, Z.; Yang, J.; Gao, L.; Zeng, Z.; Zhang, B.; Chanussot, J. Subpixel spectral variability network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5504014. [Google Scholar] [CrossRef]
  32. Liu, H.; Zhang, H.; Yang, R. Lithological classification by hyperspectral remote sensing images based on double-branch multi-scale dual-attention network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 14726–14741. [Google Scholar] [CrossRef]
  33. Gao, H.; Sheng, R.; Chen, Z.; Liu, H.; Xu, S.; Zhang, B. Multiscale random-shape convolution and adaptive graph convolution fusion network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5508414. [Google Scholar] [CrossRef]
  34. Ge, H.; Wang, L.; Liu, M.; Zhao, X.; Zhu, Y.; Pan, H.; Liu, Y. Pyramidal multiscale convolutional network with polarized self-attention for pixel-wise hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5516017. [Google Scholar] [CrossRef]
  35. Shi, H.; Zhang, Y.; Cao, G.; Yang, D. MHCFormer: Multiscale hierarchical conv-aided fourierformer for hyperspectral image classification. IEEE Trans. Instrum. Meas. 2024, 73, 5501115. [Google Scholar] [CrossRef]
  36. Shu, Z.; Zeng, K.; Zhou, J.; Wang, Y.; Tai, M.; Yu, Z. Spectral-spatial synergy guided attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5523316. [Google Scholar] [CrossRef]
  37. Ma, Y.; Lan, Y.; Xie, Y.; Yu, L.; Chen, C.; Wu, Y.; Dai, X. A spatial–spectral transformer for hyperspectral image classification based on global dependencies of multi-scale features. Remote Sens. 2024, 16, 404. [Google Scholar] [CrossRef]
  38. Feng, J.; Wang, Q.; Zhang, G.; Jia, X.; Yin, J. CAT: Center attention transformer with stratified spatial–spectral token for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5615415. [Google Scholar] [CrossRef]
  39. Zhang, M.; Yang, Y.; Zhang, S.; Mi, P.; Han, D. Spectral-spatial center-aware bottleneck transformer for hyperspectral image classification. Remote Sens. 2024, 16, 2152. [Google Scholar] [CrossRef]
  40. Chen, N.; Fang, L.; Xia, Y.; Xia, S.; Liu, H.; Yue, J. Spectral query spatial: Revisiting the role of center pixel in transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5402714. [Google Scholar] [CrossRef]
  41. Wang, Y.; Shu, Z.; Yu, Z. Efficient attention transformer network with self-similarity feature enhancement for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 11469–11486. [Google Scholar] [CrossRef]
  42. Jia, C.; Zhang, X.; Meng, H.; Xia, S.; Jiao, L. CenterFormer: A center spatial-spectral attention transformer network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 5523–5539. [Google Scholar] [CrossRef]
  43. Lu, Y.; Zhang, Y.; Jiang, X.; Liu, X.; Cai, Z. Dual-branch convolution-transformer network with spectral-spatial attention for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5518216. [Google Scholar] [CrossRef]
  44. Yu, C.; Zhu, Y.; Wang, Y.; Zhao, E.; Zhang, Q.; Lu, X. Concern with center-pixel labeling: Center-specific perception transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5514614. [Google Scholar] [CrossRef]
  45. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  46. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  47. Ouyang, E.; Li, B.; Hu, W.; Zhang, G.; Zhao, L.; Wu, J. When multigranularity meets spatial–spectral attention: A hybrid transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4401118. [Google Scholar] [CrossRef]
  48. Zhang, L.; Wang, Y.; Yang, L.; Chen, J.; Liu, Z.; Bian, L.; Yang, C. D2S2BoT: Dual-dimension spectral-spatial bottleneck transformer for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 2655–2669. [Google Scholar] [CrossRef]
  49. Xu, Y.; Wang, D.; Zhang, L.; Zhang, L. Dual selective fusion transformer network for hyperspectral image classification. Neural Netw. 2025, 187, 107311. [Google Scholar]
  50. Chen, Q.; Li, Z.; Yin, J.; Huang, W.; Zhan, T. Local-global feature extraction network with dynamic 3D convolution and residual attention transformer for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 9986–10001. [Google Scholar] [CrossRef]
  51. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  52. Yang, C.; Liu, Z.; Guan, R.; Zhao, H. A Semi-supervised multi-scale convolutional neural network for hyperspectral image classification with limited labeled samples. Remote Sens. 2025, 17, 3273. [Google Scholar] [CrossRef]
  53. Sohn, K.; Berthelot, D.; Carlini, N.; Zhang, Z.; Zhang, H.; Raffel, C.A.; Cubuk, E.D.; Kurakin, A.; Li, C.-L. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020. [Google Scholar]
Figure 1. Spectral curves of three types of ground objects in the Indian dataset. Each colored curve represents the average spectral response values of all pixels in the corresponding class. The shaded region around each curve represents the standard deviation of spectral response for that class (Corn-Notill: 1428 pixels; Corn-Mintill: 830 pixels; Corn class: 237 pixels).
Figure 1. Spectral curves of three types of ground objects in the Indian dataset. Each colored curve represents the average spectral response values of all pixels in the corresponding class. The shaded region around each curve represents the standard deviation of spectral response for that class (Corn-Notill: 1428 pixels; Corn-Mintill: 830 pixels; Corn class: 237 pixels).
Remotesensing 17 03684 g001
Figure 2. Activation response heatmaps obtained from the PaviaU dataset using different convolutional kernels: (a) 3 × 3 convolutional kernel; (b) 5 × 5 convolutional kernel; (c) 7 × 7 convolutional kernel; (d) 9 × 9 convolutional kernel. The red box represents the Trees class, and the white box represents the Painted Metal Sheets class.
Figure 2. Activation response heatmaps obtained from the PaviaU dataset using different convolutional kernels: (a) 3 × 3 convolutional kernel; (b) 5 × 5 convolutional kernel; (c) 7 × 7 convolutional kernel; (d) 9 × 9 convolutional kernel. The red box represents the Trees class, and the white box represents the Painted Metal Sheets class.
Remotesensing 17 03684 g002
Figure 3. The self-attention mechanism.
Figure 3. The self-attention mechanism.
Remotesensing 17 03684 g003
Figure 4. The overall framework of CPMFFormer. CPMFFormer is composed of one CSWM, four CRCMs, and six cross-scale fusion bottleneck transformers (CSFBT). CSWM, CRCM, and CSFBT are used to obtain class-specific spectral weights, features at different scales, and progressive fusion features, respectively. Following CSFBT-6, CPMFFormer uses a global average pooling (GAP) layer, a fully connected (FC) layer, and a S o f t m a x function to predict the final classification result.
Figure 4. The overall framework of CPMFFormer. CPMFFormer is composed of one CSWM, four CRCMs, and six cross-scale fusion bottleneck transformers (CSFBT). CSWM, CRCM, and CSFBT are used to obtain class-specific spectral weights, features at different scales, and progressive fusion features, respectively. Following CSFBT-6, CPMFFormer uses a global average pooling (GAP) layer, a fully connected (FC) layer, and a S o f t m a x function to predict the final classification result.
Remotesensing 17 03684 g004
Figure 5. The framework of CSWM. CSWM is composed of an aggregation operation, a spectral wight learning block, and a class consistency loss function. The aggregation operation smooths the spectral information differences between intra-class input patches. The spectral weight learning block performs adaptive spectral weighting. The class consistency loss function is used to learn the spectral weight centers of different ground objects and generate class-specific spectral weights.
Figure 5. The framework of CSWM. CSWM is composed of an aggregation operation, a spectral wight learning block, and a class consistency loss function. The aggregation operation smooths the spectral information differences between intra-class input patches. The spectral weight learning block performs adaptive spectral weighting. The class consistency loss function is used to learn the spectral weight centers of different ground objects and generate class-specific spectral weights.
Remotesensing 17 03684 g005
Figure 6. The framework of CRCM. CRCM is composed of two bottleneck residual blocks. In each bottleneck residual block, two 1 × 1 convolutional layers are used to form the bottleneck structure, and a k t × k t group convolution layer is used to extract features at the tth scale. Moreover, the embedded center feature calibration layer (CFCL) is used to model the dependencies between the center pixel and its neighboring pixels. This helps CRCM to reduce the impact of interference information introduced by abnormal pixels on the classification decision of the center pixel.
Figure 6. The framework of CRCM. CRCM is composed of two bottleneck residual blocks. In each bottleneck residual block, two 1 × 1 convolutional layers are used to form the bottleneck structure, and a k t × k t group convolution layer is used to extract features at the tth scale. Moreover, the embedded center feature calibration layer (CFCL) is used to model the dependencies between the center pixel and its neighboring pixels. This helps CRCM to reduce the impact of interference information introduced by abnormal pixels on the classification decision of the center pixel.
Remotesensing 17 03684 g006
Figure 7. The framework of CFCL. CFCL computes the correlations between the center pixel and its neighboring pixels through an inner product attention mechanism. Moreover, CFCL introduces a relative center position encoding to enhance spatial relationship modeling capabilities. This enables CFCL to highlight neighboring pixels that are relevant to the center pixel while suppressing irrelevant background information. This improves the representational capability of the center pixel.
Figure 7. The framework of CFCL. CFCL computes the correlations between the center pixel and its neighboring pixels through an inner product attention mechanism. Moreover, CFCL introduces a relative center position encoding to enhance spatial relationship modeling capabilities. This enables CFCL to highlight neighboring pixels that are relevant to the center pixel while suppressing irrelevant background information. This improves the representational capability of the center pixel.
Remotesensing 17 03684 g007
Figure 8. The framework of CSFBT. CSFBT takes adjacent scale features as input. It consists of a channel feature enhancement block (CFEB) and a cross-scale fusion block (CSFB). CFEB models inter-channel dependencies and emphasizes valuable channel features. CSFB uses a multi-head cross-scale dual-dimensional attention (MHCSD2A) mechanism to integrate complementary information across adjacent scale features from both spatial and channel dimensions.
Figure 8. The framework of CSFBT. CSFBT takes adjacent scale features as input. It consists of a channel feature enhancement block (CFEB) and a cross-scale fusion block (CSFB). CFEB models inter-channel dependencies and emphasizes valuable channel features. CSFB uses a multi-head cross-scale dual-dimensional attention (MHCSD2A) mechanism to integrate complementary information across adjacent scale features from both spatial and channel dimensions.
Remotesensing 17 03684 g008
Figure 9. The framework of MHCSD2A. MHCSD2A first maps the input features to different attention heads using five linear projection layers. Then, the cross-scale dual-dimensional attention (CSD2A) mechanism is computed in each head. Finally, the computation results of all heads are concatenated, and a linear projection layer is used to obtain the final output features.
Figure 9. The framework of MHCSD2A. MHCSD2A first maps the input features to different attention heads using five linear projection layers. Then, the cross-scale dual-dimensional attention (CSD2A) mechanism is computed in each head. Finally, the computation results of all heads are concatenated, and a linear projection layer is used to obtain the final output features.
Remotesensing 17 03684 g009
Figure 10. The framework of CSD2A. CSD2A achieves complementary semantic fusion between small-scale features and large-scale features by computing attention from both spatial and channel dimensions. This mechanism introduces the relative position encoding to better model complex spatial relationships and preserves structural details during cross-scale fusion.
Figure 10. The framework of CSD2A. CSD2A achieves complementary semantic fusion between small-scale features and large-scale features by computing attention from both spatial and channel dimensions. This mechanism introduces the relative position encoding to better model complex spatial relationships and preserves structural details during cross-scale fusion.
Remotesensing 17 03684 g010
Figure 11. Impact of batch size on classification performance: (a) OA; (b) AA; (c) κ .
Figure 11. Impact of batch size on classification performance: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g011
Figure 12. Impact of learning rate on classification performance: (a) OA; (b) AA; (c) κ .
Figure 12. Impact of learning rate on classification performance: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g012
Figure 13. Impact of patch size on classification performance: (a) OA; (b) AA; (c) κ .
Figure 13. Impact of patch size on classification performance: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g013
Figure 14. Full-Scene classification results obtained by different networks from the Indian dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Figure 14. Full-Scene classification results obtained by different networks from the Indian dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Remotesensing 17 03684 g014aRemotesensing 17 03684 g014b
Figure 15. Full-Scene classification results obtained by different networks using the Houston dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Figure 15. Full-Scene classification results obtained by different networks using the Houston dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Remotesensing 17 03684 g015aRemotesensing 17 03684 g015b
Figure 16. Full-Scene classification results obtained by different networks from the HongHu dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Figure 16. Full-Scene classification results obtained by different networks from the HongHu dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Remotesensing 17 03684 g016aRemotesensing 17 03684 g016b
Figure 17. Full-Scene classification results obtained by different networks from the PaviaU dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Figure 17. Full-Scene classification results obtained by different networks from the PaviaU dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Remotesensing 17 03684 g017aRemotesensing 17 03684 g017b
Figure 18. Full-Scene classification results obtained by different networks from the Salinas dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Figure 18. Full-Scene classification results obtained by different networks from the Salinas dataset: (a) false-color image; (b) ground truth; (c) S3ARN; (d) PMCN; (e) S3GAN; (f) ViT; (g) SFormer; (h) HFormer; (i) MHCFormer; (j) D2S2BoT; (k) SQSFormer; (l) EATN; (m) CFormer; (n) SSACFormer; (o) CPFormer; (p) DSFormer; (q) LGDRNet; (r) CPMFFormer.
Remotesensing 17 03684 g018aRemotesensing 17 03684 g018b
Figure 19. Classification performance obtained by different networks with different training sample sizes from the Indian dataset: (a) OA; (b) AA; (c) κ .
Figure 19. Classification performance obtained by different networks with different training sample sizes from the Indian dataset: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g019
Figure 20. Classification performance obtained by different networks with different training sample sizes from the Houston dataset: (a) OA; (b) AA; (c) κ .
Figure 20. Classification performance obtained by different networks with different training sample sizes from the Houston dataset: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g020
Figure 21. Classification performance obtained by different networks with different training sample sizes from the HongHu dataset: (a) OA; (b) AA; (c) κ .
Figure 21. Classification performance obtained by different networks with different training sample sizes from the HongHu dataset: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g021
Figure 22. Classification performance obtained by different networks with different training sample sizes from the PaviaU dataset: (a) OA; (b) AA; (c) κ .
Figure 22. Classification performance obtained by different networks with different training sample sizes from the PaviaU dataset: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g022
Figure 23. Classification performance obtained by different networks with different training samples sizes from the Salinas datasets: (a) OA; (b) AA; (c) κ .
Figure 23. Classification performance obtained by different networks with different training samples sizes from the Salinas datasets: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g023
Figure 24. Impact of different modules on classification performance: (a) OA; (b) AA; (c) κ .
Figure 24. Impact of different modules on classification performance: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g024
Figure 25. Activation response heatmaps obtained by different CSFBTs from the PaviaU dataset: (a) CSFBT-1; (b) CSFBT-2; (c) CSFBT-3; (d) CSFBT-4; (e) CSFBT-5; (f) CSFBT-6. The red box represents the Trees class, and the white box represents the Painted Metal Sheets class.
Figure 25. Activation response heatmaps obtained by different CSFBTs from the PaviaU dataset: (a) CSFBT-1; (b) CSFBT-2; (c) CSFBT-3; (d) CSFBT-4; (e) CSFBT-5; (f) CSFBT-6. The red box represents the Trees class, and the white box represents the Painted Metal Sheets class.
Remotesensing 17 03684 g025aRemotesensing 17 03684 g025b
Figure 26. Activation response heatmaps obtained by current multiscale fusion strategies from the PaviaU dataset: (a) concatenation; (b) element-wise addition; (c) attention method in the literature [35]. The red box represents the Trees class, and the white box represents the Painted Metal Sheets class.
Figure 26. Activation response heatmaps obtained by current multiscale fusion strategies from the PaviaU dataset: (a) concatenation; (b) element-wise addition; (c) attention method in the literature [35]. The red box represents the Trees class, and the white box represents the Painted Metal Sheets class.
Remotesensing 17 03684 g026
Figure 27. Impact of different spectral weighting mechanism on classification performance: (a) OA; (b) AA; (c) κ .
Figure 27. Impact of different spectral weighting mechanism on classification performance: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g027
Figure 28. Average spectral weights learned by CSWM and SWL: (ac): average spectral weights learned by SWL for the Corn-Notill, Corn-Mintill, and Corn classes; (df): average spectral weights learned by CSWM for the Corn-Notill, Corn-Mintill, and Corn classes.
Figure 28. Average spectral weights learned by CSWM and SWL: (ac): average spectral weights learned by SWL for the Corn-Notill, Corn-Mintill, and Corn classes; (df): average spectral weights learned by CSWM for the Corn-Notill, Corn-Mintill, and Corn classes.
Remotesensing 17 03684 g028aRemotesensing 17 03684 g028b
Figure 29. The performance changes in CPMFFormer after introducing FixMatch: (a) OA; (b) AA; (c) κ .
Figure 29. The performance changes in CPMFFormer after introducing FixMatch: (a) OA; (b) AA; (c) κ .
Remotesensing 17 03684 g029
Table 1. The number of homogeneous patches and heterogeneous patches in five public HSI datasets. The homogeneous patch indicates that all pixels in the patch belong to the same class. The heterogeneous patch means that the patch contains background noise or heterogeneous pixels.
Table 1. The number of homogeneous patches and heterogeneous patches in five public HSI datasets. The homogeneous patch indicates that all pixels in the patch belong to the same class. The heterogeneous patch means that the patch contains background noise or heterogeneous pixels.
DatasetIndianHoustonHongHuPaviaUSalinas
Homogeneous Patches14023210,74026,5619179
Heterogeneous Patches884714,997375,95327,56833,597
Table 2. The implementation process of CSWM.
Table 2. The implementation process of CSWM.
Layer NameKernel SizeOutput Size
Aggregation-11 × 11 × 200
Spectral Weight Learning-1 × 1 × 200
Element-Wise Multiplication-11 × 11 × 200
Element-Wise Addition-11 × 11× 200
Table 3. The implementation process of CRCM (3 × 3).
Table 3. The implementation process of CRCM (3 × 3).
Layer NameKernel SizeOutput Size
Conv1 × 111 × 11 × 64
GConv-BN-ReLU3 × 311 × 11 × 64
CFCL-11 × 11 × 64
Conv1 × 111 × 11 × 128
Residual Connection-11 × 11 × 128
Conv1 × 111 × 11 × 64
GConv-BN-ReLU3 × 311 × 11 × 64
CFCL-11 × 11 × 64
Conv1 × 111 × 11 × 128
Residual Connection-11 × 11 × 128
Table 4. The implementation process of CSFBT.
Table 4. The implementation process of CSFBT.
Layer NameKernel SizeOutput Size
CFEB-11 × 11 × 128
11 × 11 × 128
Conv1 × 111 × 11 × 12
11 × 11 × 12
MHCSD2A-11 × 11 × 12
Conv1 × 111 × 11 × 128
Residual Connection-11 × 11 × 128
Table 5. The types of ground objects and samples distribution in the Indian datasets.
Table 5. The types of ground objects and samples distribution in the Indian datasets.
ClassNameTrainingValidationTestTotal
1Alfalfa114446
2Corn-Notill292913701428
3Corn-Mintill1717796830
4Corn55227237
5Grass-Pasture1010463483
6Grass-Trees1515700730
7Grass-Pasture-Mowed112628
8Hay-Windrowed1010458478
9Oats111820
10Soybean-Notill2020932972
11Soybean-Mintill505023552455
12Soybean-Clean1212569593
13Wheat55195205
14Woods262612131265
15Buildings-Grass-Trees-Drives88370386
16Stone-Steel-Towers228993
Total 212212982510,249
Table 6. The types of ground objects and sample distribution in the Houston datasets.
Table 6. The types of ground objects and sample distribution in the Houston datasets.
ClassNameTrainingValidationTestTotal
1Healthy Grass262611991251
2Stressed Grass262612021254
3Synthetic Grass1414669697
4Trees252511941244
5Soil252511921242
6Water77311325
7Residential262612161268
8Commercial252511941244
9Road262612001252
10Highway252511771227
11Railway252511851235
12Parking Lot1252511831233
13Parking Lot21010449469
14Tennis Court99410428
15Running Track1414632660
Total 30830814,41315,029
Table 7. The types of ground objects and sample distribution in the HongHu datasets.
Table 7. The types of ground objects and sample distribution in the HongHu datasets.
ClassNameTrainingValidationTestTotal
1Red Roof28128113,47914,041
2Road717133703512
3Bare Soil43743720,94721,821
4Cotton32663266156,753163,285
5Cotton Firewood12512559686218
6Rape89289242,77344,557
7Chinese Cabbage48348323,13724,103
8Pakchoi828238904054
9Cabbage21721710,38510,819
10Tuber Mustard24824811,89812,394
11Brassica Parachinensis22122110,57311,015
12Brassica Chinensis18018085948954
13Small Brassica Chinensis45145121,60522,507
14Lactuca Sativa14814870607356
15Celtuce21219601002
16Film-Covered Lettuce14614669707262
17Romaine Lettuce616128883010
18Carrot656530873217
19White Radish17517583628712
20Garlic Sprout707033463486
21Broad Bean272712741328
22Tree818138784040
Total 77487748371,197386,693
Table 8. The types of ground objects and sample distribution in the PaviaU datasets.
Table 8. The types of ground objects and sample distribution in the PaviaU datasets.
ClassNameTrainingValidationTestTotal
1Asphalt13313363656631
2Meadows37337317,90318,649
3Gravel424220152099
4Trees626229403064
5Painted Metal Sheets272712911345
6Bare Soil10110148275029
7Bitumen272712761330
8Self-Blocking Bricks747435343682
9Shadows1919909947
Total 85885841,06042,776
Table 9. The types of ground objects and sample distribution in the Salinas datasets.
Table 9. The types of ground objects and sample distribution in the Salinas datasets.
ClassNameTrainingValidationTestTotal
1Broccoli Green Weeds 1414119272009
2Broccoli Green Weeds 2757535763726
3Fallow404018961976
4Fallow Rough Plow282813381394
5Fallow Smooth545425702678
6Stubble808037993959
7Celery727234353579
8Grapes Untrained22622610,81911,271
9Soil Vineyard Develop12512559536203
10Corn Senesced Green Weeds666631463278
11Lettuce Romaine 4 wk.222210241068
12Lettuce Romaine 5 wk.393918491927
13Lettuce Romaine 6 wk.1919878916
14Lettuce Romaine 7 wk.222210261070
15Vineyard Untrained14614669767268
16Vineyard Vertical Trellis373717331807
Total 1092109251,94554,129
Table 10. Impact of the number of CRCM on classification performance. The optimal results are in red.
Table 10. Impact of the number of CRCM on classification performance. The optimal results are in red.
DatasetsMetricsCRCM = 2CRCM = 3CRCM = 4CRCM = 5CRCM = 6
IndianOA (%)93.70 ± 1.0694.83 ± 0.9294.85 ± 0.6994.84 ± 0.8393.94 ± 0.78
AA (%)92.88 ± 0.9793.48 ± 1.3694.35 ± 0.6194.33 ± 0.6092.93 ± 1.21
κ (%)92.82 ± 1.2194.12 ± 1.0494.13 ± 0.8094.14 ± 0.9593.09 ± 0.89
HoustonOA (%)95.76 ± 0.6295.88 ± 0.7395.93 ± 0.3695.91 ± 0.4095.29 ± 0.52
AA (%)95.93 ± 0.8196.01 ± 0.7396.06 ± 0.2695.92 ± 0.5395.51 ± 0.64
κ (%)95.42 ± 0.6795.55 ± 0.7895.61 ± 0.3995.60 ± 0.4494.91 ± 0.56
HongHuOA (%)99.53 ± 0.0499.50 ± 0.0499.53 ± 0.0399.50 ± 0.0599.49 ± 0.05
AA (%)98.92 ± 0.2098.83 ± 0.1698.99 ± 0.1098.82 ± 0.2098.68 ± 0.06
κ (%)99.40 ± 0.0599.37 ± 0.0499.41 ± 0.0499.36 ± 0.0799.24 ± 0.06
PaviaUOA (%)99.45 ± 0.1299.45 ± 0.2299.50 ± 0.0699.48 ± 0.1399.35 ± 0.09
AA (%)98.95 ± 0.2798.88 ± 0.4599.02 ± 0.1099.01 ± 0.2098.74 ± 0.14
κ (%)99.27 ± 0.1599.27 ± 0.2999.34 ± 0.0899.31 ± 0.1899.13 ± 0.12
SalinasOA (%)99.81 ± 0.1299.77 ± 0.1399.84 ± 0.0699.84 ± 0.0699.75 ± 0.13
AA (%)99.81 ± 0.1399.79 ± 0.0899.84 ± 0.0599.86 ± 0.0299.75 ± 0.09
κ (%)99.79 ± 0.1499.74 ± 0.1499.82 ± 0.0799.82 ± 0.0799.72 ± 0.15
Table 11. Impact of the number of bottleneck residual blocks within CRCM on classification performance. The optimal results are in red.
Table 11. Impact of the number of bottleneck residual blocks within CRCM on classification performance. The optimal results are in red.
DatasetsMetricsBRB = 1BRB = 2BRB = 3BRB = 4BRB = 5
IndianOA (%)94.10 ± 0.5694.85 ± 0.6994.29 ± 0.4294.57 ± 0.8994.74 ± 0.72
AA (%)93.05 ± 1.8994.35 ± 0.6193.73 ± 0.7093.30 ± 2.4793.66 ± 0.53
κ (%)93.26 ± 0.6594.13 ± 0.8093.49 ± 0.4993.80 ± 1.0294.00 ± 0.82
HoustonOA (%)96.02 ± 0.4895.93 ± 0.3695.09 ± 1.2495.66 ± 0.5395.65 ± 0.55
AA (%)96.04 ± 0.6396.06 ± 0.2695.35 ± 1.1295.82 ± 0.5495.77 ± 0.73
κ (%)95.70 ± 0.5295.61 ± 0.3994.69 ± 1.3495.31 ± 0.5795.30 ± 0.60
HongHuOA (%)99.47 ± 0.0599.53 ± 0.0399.47 ± 0.0199.52 ± 0.0399.49 ± 0.03
AA (%)98.77 ± 0.1598.99 ± 0.1098.75 ± 0.0598.78 ± 0.1198.71 ± 0.18
κ (%)99.33 ± 0.0799.41 ± 0.0499.34 ± 0.0199.40 ± 0.0399.36 ± 0.04
PaviaUOA (%)99.39 ± 0.1099.50 ± 0.0699.38 ± 0.1599.34 ± 0.1499.43 ± 0.14
AA (%)98.77 ± 0.1999.02 ± 0.1098.80 ± 0.3898.74 ± 0.3398.82 ± 0.33
κ (%)99.19 ± 0.1399.34 ± 0.0899.18 ± 0.1999.12 ± 0.1999.25 ± 0.19
SalinasOA (%)99.83 ± 0.0999.84 ± 0.0699.79 ± 0.0699.82 ± 0.1399.79 ± 0.08
AA (%)99.83 ± 0.0599.84 ± 0.0599.79 ± 0.0799.84 ± 0.0999.81 ± 0.06
κ (%)99.81 ± 0.1099.82 ± 0.0799.76 ± 0.0799.80 ± 0.1599.77 ± 0.09
Table 12. Impact of feature channel dimension on classification performance. The optimal results are in red.
Table 12. Impact of feature channel dimension on classification performance. The optimal results are in red.
DatasetsMetricsC = 32C = 64C = 96C = 128C = 160C = 192
IndianOA (%)93.67 ± 0.4394.05 ± 1.3994.52 ± 0.9794.85 ± 0.6994.30 ± 1.0094.47 ± 0.72
AA (%)92.79 ± 1.1693.22 ± 2.3494.07 ± 1.6794.35 ± 0.6193.93 ± 1.9594.56 ± 0.68
κ (%)92.78 ± 0.5093.21 ± 1.5993.75 ± 1.1194.13 ± 0.8093.50 ± 1.1493.70 ± 0.81
HoustonOA (%)95.28 ± 1.2295.93 ± 0.3695.76 ± 0.6195.58 ± 0.8095.40 ± 0.7395.12 ± 0.94
AA (%)95.41 ± 1.5396.06 ± 0.2695.82 ± 0.6295.77 ± 0.7295.71 ± 0.7295.38 ± 0.96
κ (%)94.89 ± 1.3295.61 ± 0.3995.41 ± 0.6695.23 ± 0.8795.02 ± 0.7994.72 ± 1.02
HongHuOA (%)99.45 ± 0.0899.53 ± 0.0399.51 ± 0.0599.47 ± 0.0899.48 ± 0.0899.50 ± 0.05
AA (%)98.71 ± 0.2398.99 ± 0.1098.85 ± 0.2198.85 ± 0.1898.86 ± 0.1998.85 ± 0.18
κ (%)99.31 ± 0.1199.41 ± 0.0499.38 ± 0.0799.33 ± 0.1099.34 ± 0.1099.37 ± 0.06
PaviaUOA (%)99.44 ± 0.1299.43 ± 0.1599.47 ± 0.1099.50 ± 0.0699.39 ± 0.1299.39 ± 0.08
AA (%)98.88 ± 0.3198.89 ± 0.2698.92 ± 0.2299.02 ± 0.1098.82 ± 0.3198.87 ± 0.21
κ (%)99.25 ± 0.1699.24 ± 0.2099.29 ± 0.1399.34 ± 0.0899.19 ± 0.1699.19 ± 0.11
SalinasOA (%)99.76 ± 0.0599.84 ± 0.0699.82 ± 0.0799.83 ± 0.0899.82 ± 0.0699.76 ± 0.13
AA (%)99.75 ± 0.0599.84 ± 0.0599.82 ± 0.0699.82 ± 0.0599.79 ± 0.0799.78 ± 0.14
κ (%)99.75 ± 0.0599.82 ± 0.0799.80 ± 0.0899.81 ± 0.0999.79 ± 0.0699.73 ± 0.14
Table 13. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the Indian dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 13. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the Indian dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
152.89 ± 8.8016.00 ± 15.562.22 ± 13.76.222 ± 8.523.554 ± 2.5344.89 ± 11.942.67 ± 14.780.00 ± 12.7
288.84 ± 4.4790.82 ± 1.8593.00 ± 2.3365.75 ± 10.976.80 ± 12.891.87 ± 2.3891.39 ± 2.5685.75 ± 4.03
390.95 ± 4.7186.27 ± 4.7690.96 ± 2.6853.16 ± 10.855.33 ± 14.387.50 ± 8.9687.99 ± 7.0587.01 ± 9.05
453.53 ± 25.084.74 ± 10.184.27 ± 11.521.38 ± 10.422.76 ± 10.171.81 ± 12.776.38 ± 12.585.00 ± 13.2
589.18 ± 4.8190.86 ± 3.1190.23 ± 3.8365.12 ± 8.4675.10 ± 7.9789.77 ± 4.6691.37 ± 2.0890.36 ± 1.46
693.12 ± 2.2793.79 ± 1.4394.58 ± 3.0687.97 ± 3.1086.41 ± 4.2694.46 ± 2.8896.39 ± 2.5094.46 ± 1.99
773.33 ± 16.28.888 ± 19.987.04 ± 7.0917.78 ± 14.72.222 ± 3.3171.11 ± 18.882.22 ± 15.693.33 ± 7.59
894.49 ± 5.5999.40 ± 0.4699.79 ± 0.3097.09 ± 2.2193.59 ± 4.1599.87 ± 0.2999.96 ± 0.0999.15 ± 0.88
970.53 ± 25.41.052 ± 2.3556.58 ± 15.16.316 ± 11.42.106 ± 4.7170.53 ± 20.982.11 ± 16.161.05 ± 38.1
1078.02 ± 2.2283.68 ± 3.8382.14 ± 1.9361.32 ± 8.4370.69 ± 8.3884.45 ± 2.8583.95 ± 4.1084.51 ± 2.86
1190.54 ± 3.4993.24 ± 3.3096.30 ± 2.0076.61 ± 3.1583.21 ± 8.1296.06 ± 2.9794.80 ± 2.7993.48 ± 3.11
1266.16 ± 7.3986.68 ± 6.1889.67 ± 4.4552.98 ± 5.2552.53 ± 7.8187.37 ± 4.8690.91 ± 4.2881.89 ± 7.99
1396.10 ± 2.5396.80 ± 2.3997.38 ± 0.8586.10 ± 5.7575.10 ± 14.795.30 ± 3.6598.40 ± 1.7898.00 ± 1.87
1496.01 ± 2.8397.69 ± 1.3898.15 ± 2.0093.04 ± 2.2888.86 ± 4.6397.50 ± 1.2197.19 ± 2.9497.29 ± 2.00
1579.52 ± 14.887.67 ± 6.6088.89 ± 12.653.81 ± 16.953.97 ± 14.490.00 ± 6.7284.50 ± 5.5489.63 ± 6.60
1635.16 ± 19.882.86 ± 12.086.27 ± 9.3538.02 ± 20.732.09 ± 36.757.36 ± 23.787.91 ± 6.7394.51 ± 3.39
OA (%)86.80 ± 0.9490.64 ± 1.1592.78 ± 0.7870.73 ± 3.9174.27 ± 4.5391.62 ± 0.9891.91 ± 1.2090.65 ± 1.93
AA (%)78.02 ± 2.7475.03 ± 2.7787.34 ± 1.2355.17 ± 5.1654.64 ± 5.6783.12 ± 1.1186.76 ± 2.2088.46 ± 3.29
κ (%)84.91 ± 1.0989.31 ± 1.2991.75 ± 0.8866.28 ± 4.5870.34 ± 5.2590.42 ± 1.1190.75 ± 1.3789.33 ± 2.20
Table 14. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the Indian dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 14. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the Indian dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassSQSFormerEATNCFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
16.956 ± 3.8965.65 ± 29.980.98 ± 16.797.29 ± 2.7469.57 ± 25.295.22 ± 2.3882.61 ± 10.296.96 ± 3.64
289.10 ± 2.2089.93 ± 2.8594.28 ± 1.8890.58 ± 4.1195.73 ± 1.1591.74 ± 2.1191.88 ± 3.2794.06 ± 2.06
386.46 ± 5.6376.00 ± 4.8291.87 ± 2.3688.17 ± 2.0193.95 ± 4.6592.77 ± 3.5392.98 ± 3.5790.60 ± 5.31
477.64 ± 8.4078.06 ± 12.993.99 ± 7.2986.39 ± 14.185.23 ± 13.589.28 ± 6.3194.31 ± 8.4591.90 ± 9.81
594.16 ± 1.4691.92 ± 2.0991.77 ± 3.8189.75 ± 2.8090.17 ± 1.6091.30 ± 2.4589.86 ± 6.6590.10 ± 2.84
698.79 ± 0.5098.27 ± 1.8393.15 ± 1.1099.18 ± 0.2598.70 ± 1.4298.49 ± 1.0094.63 ± 2.1297.31 ± 1.50
720.00 ± 9.9885.71 ± 7.5848.22 ± 35.373.22 ± 48.998.22 ± 3.5791.43 ± 5.4285.72 ± 19.499.29 ± 1.60
899.92 ± 0.1299.66 ± 0.3599.27 ± 1.19100.0 ± 0.0099.69 ± 0.27100.0 ± 0.0098.95 ± 0.5799.79 ± 0.30
97.000 ± 2.7448.00 ± 23.115.00 ± 16.847.50 ± 49.926.25 ± 12.549.00 ± 14.865.00 ± 8.1688.00 ± 4.47
1081.23 ± 5.4186.52 ± 6.7185.55 ± 1.8586.86 ± 4.5290.54 ± 0.7187.99 ± 2.6190.05 ± 2.4591.93 ± 4.97
1191.22 ± 1.8887.89 ± 2.7597.02 ± 1.2995.61 ± 1.6994.16 ± 3.4895.66 ± 1.1294.69 ± 0.7896.94 ± 1.21
1288.47 ± 5.2082.19 ± 6.4488.95 ± 6.4493.30 ± 3.1888.20 ± 5.1387.62 ± 0.9187.69 ± 8.0988.40 ± 6.84
1399.51 ± 0.6099.31 ± 0.2795.24 ± 4.4699.39 ± 0.7398.66 ± 0.8399.61 ± 0.4198.41 ± 1.3499.12 ± 1.70
1497.86 ± 1.5596.33 ± 1.5799.37 ± 0.1399.78 ± 0.1799.05 ± 0.7698.62 ± 0.5698.83 ± 1.1398.59 ± 0.59
1590.98 ± 5.6183.52 ± 3.8287.50 ± 7.9096.63 ± 2.4187.83 ± 10.696.16 ± 1.3495.53 ± 3.4193.63 ± 6.31
1666.24 ± 18.891.83 ± 8.3162.63 ± 10.494.36 ± 4.4294.35 ± 4.1587.31 ± 4.3988.44 ± 10.492.90 ± 7.15
OA (%)90.22 ± 0.4888.96 ± 1.0793.39 ± 0.9393.78 ± 0.9494.08 ± 1.6094.02 ± 0.5993.69 ± 0.2294.85 ± 0.69
AA (%)74.72 ± 0.9285.05 ± 4.1482.80 ± 1.9889.87 ± 6.0288.14 ± 3.6590.76 ± 1.2690.60 ± 1.3894.35 ± 0.61
κ (%)88.84 ± 0.5587.41 ± 1.2592.46 ± 1.0692.90 ± 1.0793.26 ± 1.8193.17 ± 0.6792.82 ± 0.2694.13 ± 0.80
Table 15. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, Hformer, MHCFormer, and D2S2BoT from the Houston dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 15. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, Hformer, MHCFormer, and D2S2BoT from the Houston dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
196.78 ± 3.7596.68 ± 1.4698.50 ± 0.8495.20 ± 1.8197.66 ± 0.9497.55 ± 1.6797.89 ± 1.1596.57 ± 2.14
288.16 ± 4.4391.12 ± 2.4292.86 ± 4.6292.21 ± 3.8891.30 ± 2.6793.32 ± 4.1895.34 ± 4.9394.72 ± 3.29
398.83 ± 1.2498.10 ± 1.1699.75 ± 0.1899.97 ± 0.0798.80 ± 1.2498.83 ± 0.8798.77 ± 0.7698.48 ± 1.56
492.66 ± 1.1093.24 ± 3.8997.71 ± 2.9594.60 ± 4.6596.10 ± 1.9196.13 ± 3.5997.82 ± 2.4694.70 ± 3.10
597.25 ± 3.5499.52 ± 0.38100.0 ± 0.0099.87 ± 0.1799.79 ± 0.29100.0 ± 0.0099.95 ± 0.0499.72 ± 0.22
690.41 ± 2.4589.62 ± 3.5289.54 ± 4.1783.02 ± 7.2295.85 ± 4.7988.74 ± 2.1690.25 ± 4.8190.06 ± 4.12
783.25 ± 0.2384.30 ± 4.0993.99 ± 4.1276.57 ± 3.7385.53 ± 3.7586.78 ± 4.3793.73 ± 2.6691.30 ± 2.82
870.92 ± 4.3673.67 ± 6.2083.80 ± 6.4776.83 ± 5.7374.40 ± 7.3879.08 ± 3.9879.92 ± 5.7880.82 ± 4.51
978.71 ± 2.0879.14 ± 3.1985.86 ± 3.8172.43 ± 7.4067.91 ± 8.2780.41 ± 4.1983.56 ± 3.6284.93 ± 5.08
1093.34 ± 5.8888.84 ± 5.4698.33 ± 1.3383.53 ± 6.0487.14 ± 5.4092.28 ± 4.1295.72 ± 3.5596.41 ± 2.25
1190.74 ± 6.6691.52 ± 2.4993.91 ± 7.5687.79 ± 7.9789.59 ± 2.3195.34 ± 2.8097.50 ± 2.0795.78 ± 2.98
1296.28 ± 3.2787.45 ± 4.3489.76 ± 4.8987.18 ± 6.1979.12 ± 7.3493.26 ± 4.3892.53 ± 4.5890.60 ± 6.49
1382.46 ± 6.0183.71 ± 2.9691.21 ± 3.0766.23 ± 7.1661.22 ± 10.680.70 ± 2.8893.51 ± 0.8189.24 ± 2.05
1493.32 ± 9.4598.85 ± 0.5999.06 ± 0.8383.67 ± 11.799.67 ± 0.4999.38 ± 0.9099.71 ± 0.2098.90 ± 1.94
1599.23 ± 1.1099.72 ± 0.17100.0 ± 0.0098.36 ± 1.82100.0 ± 0.00100.0 ± 0.0099.88 ± 0.2899.91 ± 0.08
OA (%)89.67 ± 0.2189.64 ± 1.0094.05 ± 1.4286.94 ± 1.8487.74 ± 1.9891.96 ± 0.8794.04 ± 0.9293.17 ± 1.24
AA (%)90.15 ± 0.4590.37 ± 0.8594.29 ± 1.2786.50 ± 1.8688.27 ± 2.2192.12 ± 0.7094.41 ± 0.9393.48 ± 1.18
κ (%)88.83 ± 0.2288.80 ± 1.0893.57 ± 1.5485.88 ± 1.9986.74 ± 2.1491.30 ± 0.9493.56 ± 1.0092.61 ± 1.34
Table 16. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the Houston dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 16. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the Houston dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassSQSFormerEATNCFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
197.18 ± 1.0898.11 ± 1.1395.29 ± 2.9898.43 ± 0.9096.40 ± 1.6898.37 ± 0.5896.29 ± 1.3597.94 ± 0.56
295.81 ± 4.2597.92 ± 1.5890.91 ± 2.8392.85 ± 5.7293.38 ± 6.1298.64 ± 0.4597.16 ± 2.0795.49 ± 3.32
399.97 ± 0.0699.74 ± 0.1298.17 ± 0.91100.0 ± 0.00100.0 ± 0.0097.05 ± 0.8399.51 ± 0.6098.51 ± 1.30
496.61 ± 2.3498.75 ± 1.5393.11 ± 1.9997.81 ± 2.1297.35 ± 2.4497.62 ± 2.6295.79 ± 1.8597.07 ± 3.35
5100.0 ± 0.0098.87 ± 0.6099.98 ± 0.04100.0 ± 0.0099.62 ± 0.50100.0 ± 0.0098.42 ± 1.0699.94 ± 0.07
692.98 ± 3.5088.80 ± 5.6590.16 ± 5.5191.32 ± 4.3086.00 ± 13.985.35 ± 5.0886.34 ± 6.7893.41 ± 5.82
793.11 ± 2.9096.13 ± 2.3496.12 ± 1.5895.54 ± 3.8796.41 ± 1.2997.30 ± 2.4494.10 ± 2.0095.03 ± 2.70
880.61 ± 6.0689.74 ± 3.4183.02 ± 9.3582.43 ± 7.0083.42 ± 5.1587.72 ± 8.3388.17 ± 4.4089.42 ± 1.26
982.28 ± 2.5086.51 ± 5.3589.54 ± 2.5787.89 ± 5.2292.19 ± 2.3682.15 ± 6.6787.91 ± 4.3090.14 ± 3.80
1096.53 ± 3.1197.12 ± 2.9099.88 ± 0.2498.89 ± 1.3999.78 ± 0.1092.99 ± 3.0498.08 ± 1.0498.65 ± 0.92
1196.53 ± 2.1092.15 ± 3.1498.12 ± 2.0097.09 ± 1.2095.04 ± 5.7493.97 ± 4.2493.36 ± 2.0897.22 ± 1.54
1291.91 ± 3.7587.30 ± 8.3494.36 ± 5.5693.58 ± 4.4093.39 ± 3.6893.79 ± 1.6094.34 ± 3.1395.44 ± 1.26
1368.06 ± 12.056.67 ± 8.8588.01 ± 4.6084.44 ± 4.8685.87 ± 2.5190.53 ± 2.4389.98 ± 5.3393.22 ± 1.66
1498.04 ± 2.6097.71 ± 2.35100.0 ± 0.0099.91 ± 0.1398.42 ± 2.7099.67 ± 0.3997.71 ± 4.0099.44 ± 0.79
1599.55 ± 0.4899.03 ± 0.6899.96 ± 0.0899.79 ± 0.4099.89 ± 0.23100.0 ± 0.0098.91 ± 1.3899.97 ± 0.07
OA (%)93.01 ± 0.8493.54 ± 0.8094.37 ± 1.1994.71 ± 1.4394.81 ± 1.0494.49 ± 1.1094.58 ± 1.0695.93 ± 0.36
AA (%)92.61 ± 0.8292.31 ± 0.7994.44 ± 1.2394.66 ± 1.1994.48 ± 1.3494.34 ± 1.1194.40 ± 1.1296.06 ± 0.26
κ (%)92.45 ± 0.9193.01 ± 0.8793.91 ± 1.2994.28 ± 1.5494.39 ± 1.1294.04 ± 1.1994.14 ± 1.1495.61 ± 0.39
Table 17. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the HongHu dataset. The optimal and suboptimal results are in red and blue, respectively.
Table 17. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the HongHu dataset. The optimal and suboptimal results are in red and blue, respectively.
ClassS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
198.90 ± 0.5398.85 ± 0.3098.71 ± 0.8796.25 ± 0.7397.66 ± 0.4498.59 ± 0.5698.44 ± 0.5598.86 ± 0.32
291.91 ± 2.3390.32 ± 1.9591.53 ± 0.7678.43 ± 7.8088.05 ± 2.5790.32 ± 1.2793.11 ± 3.3294.24 ± 2.24
398.12 ± 0.5097.81 ± 0.3899.09 ± 0.1894.08 ± 1.6796.18 ± 0.4498.34 ± 0.4097.37 ± 0.4798.17 ± 0.27
499.84 ± 0.0399.83 ± 0.0399.95 ± 0.0299.39 ± 0.2499.67 ± 0.1099.87 ± 0.1399.63 ± 0.1399.84 ± 0.15
596.16 ± 1.5996.49 ± 0.6099.46 ± 0.4383.17 ± 2.4790.28 ± 1.9996.83 ± 1.1589.20 ± 3.8496.18 ± 1.91
699.35 ± 0.2299.60 ± 0.1999.87 ± 0.0596.96 ± 0.9899.21 ± 0.1899.74 ± 0.0899.27 ± 0.1699.53 ± 0.16
796.92 ± 0.5297.41 ± 0.2498.98 ± 0.3689.30 ± 4.1994.91 ± 1.6097.22 ± 0.6996.04 ± 0.8398.05 ± 0.32
883.31 ± 3.8896.04 ± 0.7898.88 ± 0.4954.05 ± 8.8385.55 ± 5.7194.06 ± 3.5586.17 ± 4.6892.16 ± 2.58
999.53 ± 0.2899.34 ± 0.3499.50 ± 0.1596.06 ± 1.3598.70 ± 0.6899.17 ± 0.3898.70 ± 0.4899.39 ± 0.12
1097.81 ± 0.7696.95 ± 0.8498.98 ± 0.3585.59 ± 1.8093.09 ± 0.8497.63 ± 0.7695.77 ± 0.8897.80 ± 0.89
1195.75 ± 1.7896.48 ± 0.6098.66 ± 0.4881.27 ± 3.9791.31 ± 4.0296.21 ± 1.2294.95 ± 1.6597.59 ± 0.48
1293.17 ± 2.3995.82 ± 0.7098.44 ± 0.5371.46 ± 4.0488.13 ± 1.4695.49 ± 2.3691.21 ± 1.0695.42 ± 2.13
1396.71 ± 0.5296.63 ± 1.1699.07 ± 0.3381.96 ± 3.7694.12 ± 1.2397.57 ± 0.9594.96 ± 1.6497.76 ± 0.95
1497.64 ± 0.5696.45 ± 0.7598.57 ± 0.2886.64 ± 4.9492.58 ± 1.4495.62 ± 2.0097.05 ± 0.7898.11 ± 0.72
1594.21 ± 4.3394.15 ± 2.1095.17 ± 3.6283.71 ± 2.5684.04 ± 10.996.13 ± 3.1792.64 ± 4.2894.19 ± 4.90
1698.96 ± 0.5998.05 ± 0.9199.22 ± 0.5092.88 ± 0.7196.45 ± 1.4898.51 ± 1.1598.69 ± 0.8398.11 ± 0.94
1797.55 ± 0.5897.82 ± 0.7798.84 ± 0.9282.99 ± 3.3592.81 ± 2.1498.82 ± 0.5292.82 ± 2.4798.34 ± 0.75
1892.30 ± 3.2995.77 ± 0.9696.85 ± 1.0988.43 ± 6.9093.38 ± 2.8696.45 ± 2.0294.99 ± 2.2195.91 ± 2.27
1996.54 ± 1.4196.65 ± 1.2497.36 ± 0.8690.11 ± 4.0894.49 ± 0.8997.69 ± 0.5096.69 ± 1.6197.75 ± 0.32
2096.55 ± 1.2494.83 ± 1.9497.54 ± 1.2584.91 ± 2.3991.32 ± 2.7395.66 ± 3.1494.62 ± 2.4295.56 ± 1.71
2185.37 ± 5.9695.65 ± 1.5097.73 ± 1.3556.11 ± 18.088.79 ± 2.3596.19 ± 2.5383.55 ± 3.4892.99 ± 3.73
2296.82 ± 0.4396.81 ± 1.4699.34 ± 0.4084.94 ± 6.6090.68 ± 2.6595.60 ± 2.7891.67 ± 4.7697.78 ± 1.15
OA (%)98.30 ± 0.1198.52 ± 0.0999.34 ± 0.0993.17 ± 0.7296.91 ± 0.3698.63 ± 0.2597.65 ± 0.2298.75 ± 0.10
AA (%)95.61 ± 0.4896.72 ± 0.1498.25 ± 0.0784.49 ± 1.4892.79 ± 0.9896.90 ± 0.3594.43 ± 0.7296.99 ± 0.33
κ (%)97.85 ± 0.1498.12 ± 0.1199.17 ± 0.1791.36 ± 0.9196.10 ± 0.4698.27 ± 0.3197.03 ± 0.2998.42 ± 0.13
Table 18. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the HongHu dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 18. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the HongHu dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassSQSFormerEATNCFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
198.39 ± 0.3598.13 ± 0.4699.49 ± 0.1999.42 ± 0.1599.29 ± 0.5099.21 ± 0.5899.19 ± 0.3699.45 ± 0.18
288.42 ± 1.7687.28 ± 4.9193.07 ± 2.0296.24 ± 1.9595.54 ± 0.6795.64 ± 0.7996.56 ± 1.6994.89 ± 1.21
397.04 ± 0.7396.70 ± 1.3098.85 ± 0.3999.10 ± 0.1398.79 ± 0.3698.79 ± 0.3098.96 ± 0.1699.04 ± 0.19
499.76 ± 0.1499.83 ± 0.0299.97 ± 0.0299.99 ± 0.0499.90 ± 0.0799.91 ± 0.0399.95 ± 0.0299.98 ± 0.02
594.31 ± 2.7688.68 ± 4.4999.31 ± 0.5898.95 ± 0.4399.13 ± 0.8897.30 ± 0.5497.63 ± 0.9399.57 ± 0.30
699.09 ± 0.3099.48 ± 0.1999.83 ± 0.0799.88 ± 0.1399.71 ± 0.1399.38 ± 0.0599.64 ± 0.1399.87 ± 0.04
797.16 ± 1.0795.82 ± 0.8898.88 ± 0.2099.52 ± 0.2898.45 ± 0.4898.12 ± 0.2198.73 ± 0.5099.20 ± 0.32
892.21 ± 5.1378.58 ± 6.8997.64 ± 1.9799.21 ± 0.4698.40 ± 1.3991.46 ± 1.2494.05 ± 2.4199.45 ± 0.77
999.31 ± 0.2198.38 ± 0.6199.81 ± 0.2699.70 ± 0.1099.65 ± 0.1299.64 ± 0.1499.50 ± 0.1499.79 ± 0.16
1096.22 ± 1.2896.54 ± 0.8098.98 ± 0.4198.49 ± 0.2497.89 ± 0.9398.10 ± 0.3398.42 ± 0.3498.76 ± 0.35
1194.64 ± 1.1995.26 ± 2.3698.66 ± 0.6298.57 ± 0.2498.05 ± 0.4697.49 ± 0.6598.33 ± 0.3998.58 ± 0.46
1294.37 ± 1.6493.08 ± 1.3098.53 ± 0.7299.12 ± 0.6698.47 ± 1.1395.50 ± 0.9497.99 ± 0.4299.18 ± 0.40
1394.33 ± 1.7295.04 ± 0.7999.01 ± 0.4098.63 ± 0.4198.94 ± 0.3497.12 ± 0.2998.32 ± 0.4599.13 ± 0.31
1498.11 ± 0.5496.14 ± 1.5499.18 ± 0.4598.46 ± 0.3698.17 ± 0.2597.85 ± 0.7698.99 ± 0.1398.73 ± 0.80
1587.67 ± 1.9591.50 ± 4.3495.73 ± 2.5997.80 ± 3.1595.83 ± 1.7196.51 ± 1.3494.63 ± 3.9899.38 ± 0.36
1698.04 ± 1.0497.81 ± 0.8399.38 ± 0.4299.06 ± 0.3999.37 ± 0.1798.88 ± 0.7699.61 ± 0.4099.50 ± 0.38
1797.27 ± 1.3994.65 ± 1.7599.30 ± 0.4999.30 ± 0.4899.34 ± 0.4097.86 ± 1.3399.08 ± 0.4999.48 ± 0.27
1894.38 ± 1.9097.53 ± 0.7696.85 ± 1.6499.19 ± 1.1197.49 ± 0.9497.45 ± 0.5597.68 ± 0.4897.96 ± 0.41
1995.16 ± 0.5995.48 ± 0.5698.41 ± 0.4496.38 ± 1.0097.48 ± 1.2597.34 ± 1.0997.97 ± 0.4999.17 ± 0.39
2092.72 ± 1.8593.88 ± 1 .5798.92 ± 0.5497.65 ± 0.5498.56 ± 0.8296.15 ± 0.9497.18 ± 1.0298.00 ± 1.60
2189.22 ± 9.1462.63 ± 8.6997.68 ± 1.4799.85 ± 0.5095.78 ± 6.7693.15 ± 0.7397.08 ± 0.8399.80 ± 0.20
2295.67 ± 2.1694.82 ± 1.0298.40 ± 1.2997.70 ± 0.8599.28 ± 0.5996.74 ± 2.2198.85 ± 0.3698.96 ± 0.90
OA (%)97.93 ± 0.3997.55 ± 0.1699.41 ± 0.0699.43 ± 0.0599.26 ± 0.1098.83 ± 0.0799.22 ± 0.0499.53 ± 0.03
AA (%)95.16 ± 1.0393.06 ± 0.6298.45 ± 0.2198.74 ± 0.1998.34 ± 0.4197.25 ± 0.1698.11 ± 0.1998.99 ± 0.10
κ (%)97.38 ± 0.4996.91 ± 0.2099.25 ± 0.0699.28 ± 0.0799.06 ± 0.1398.52 ± 0.0999.01 ± 0.0599.41 ± 0.04
Table 19. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the PaviaU dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 19. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the PaviaU dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
197.08 ± 1.2097.55 ± 0.3399.62 ± 0.2695.36 ± 1.5095.91 ± 1.2098.98 ± 0.4399.37 ± 0.1999.46 ± 0.38
299.85 ± 0.1599.67 ± 0.0899.98 ± 0.0299.04 ± 0.6799.73 ± 0.1199.99 ± 0.0199.83 ± 0.0799.96 ± 0.04
390.97 ± 5.1891.52 ± 2.4194.52 ± 1.8978.37 ± 1.7881.00 ± 5.2094.58 ± 1.1594.04 ± 2.8796.84 ± 1.52
498.65 ± 0.6695.27 ± 1.3297.62 ± 1.7694.66 ± 2.0695.62 ± 1.9897.39 ± 1.1598.09 ± 1.6696.68 ± 1.26
599.89 ± 0.1598.59 ± 0.96100.0 ± 0.0099.59 ± 0.7199.98 ± 0.0499.71 ± 0.4499.95 ± 0.0799.77 ± 0.39
699.52 ± 0.4698.17 ± 1.4799.92 ± 0.1495.43 ± 2.2997.50 ± 0.48100.0 ± 0.0099.59 ± 0.2399.65 ± 0.49
792.39 ± 1.7598.40 ± 0.78100.0 ± 0.0082.95 ± 10.793.95 ± 1.3999.23 ± 1.3698.65 ± 1.4299.25 ± 0.96
897.20 ± 1.0892.21 ± 3.9096.51 ± 1.8883.92 ± 6.0087.78 ± 1.5495.49 ± 2.3097.24 ± 1.4596.80 ± 0.97
999.20 ± 0.6793.81 ± 1.5197.38 ± 2.5490.09 ± 7.8595.43 ± 2.6396.96 ± 1.2999.18 ± 0.4896.72 ± 0.16
OA (%)98.39 ± 0.5197.60 ± 0.4499.13 ± 0.0594.73 ± 0.9496.37 ± 0.3098.90 ± 0.2599.05 ± 0.2399.09 ± 0.21
AA (%)97.20 ± 0.8596.13 ± 0.5898.39 ± 0.1691.05 ± 1.6494.10 ± 0.7798.04 ± 0.4398.44 ± 0.4298.35 ± 0.35
κ (%)97.86 ± 0.6896.82 ± 0.5998.84 ± 0.0793.00 ± 1.2595.17 ± 0.4098.54 ± 0.3398.74 ± 0.3098.79 ± 0.28
Table 20. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the PaviaU dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 20. Classification performances obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the PaviaU dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassSQSFormerEATNCFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
198.00 ± 1.3296.24 ± 1.3799.83 ± 0.2099.16 ± 0.3798.19 ± 1.6799.28 ± 0.4998.13 ± 0.8199.78 ± 0.16
299.25 ± 0.1899.19 ± 0.3399.92 ± 0.06100.0 ± 0.0099.90 ± 0.1499.97 ± 0.0399.89 ± 0.0599.96 ± 0.06
386.40 ± 2.4484.27 ± 4.8498.90 ± 1.1093.49 ± 4.5890.90 ± 7.2896.82 ± 1.5795.28 ± 2.3997.74 ± 1.31
496.91 ± 1.7596.27 ± 1.5492.24 ± 2.0498.33 ± 0.3398.69 ± 1.1098.15 ± 0.5395.89 ± 1.7997.75 ± 0.68
599.97 ± 0.0799.70 ± 0.5599.76 ± 0.10100.0 ± 0.0099.96 ± 0.1099.98 ± 0.0498.93 ± 0.9499.45 ± 0.56
698.75 ± 0.7297.22 ± 0.89100.0 ± 0.0099.90 ± 0.2099.85 ± 0.1499.05 ± 1.5099.44 ± 0.5999.96 ± 0.06
798.17 ± 2.0792.48 ± 1.3499.91 ± 0.1399.98 ± 0.0479.40 ± 44.498.42 ± 0.6093.46 ± 3.7599.85 ± 0.16
894.33 ± 1.4791.10 ± 1.2299.07 ± 0.6493.65 ± 2.9994.01 ± 4.6397.15 ± 1.4995.50 ± 0.5198.85 ± 0.60
999.45 ± 0.3799.45 ± 0.6788.22 ± 2.5499.89 ± 0.1197.70 ± 2.8597.50 ± 0.1296.18 ± 1.8697.89 ± 0.29
OA (%)97.77 ± 0.3996.68 ± 0.4298.98 ± 0.1598.87 ± 0.3997.91 ± 1.8999.13 ± 0.1498.36 ± 0.2699.50 ± 0.06
AA (%)96.80 ± 0.6595.10 ± 0.5997.54 ± 0.4398.27 ± 0.6895.40 ± 5.9298.48 ± 0.1596.97 ± 0.6299.02 ± 0.10
κ (%)97.05 ± 0.5295.59 ± 0.5698.64 ± 0.1998.50 ± 0.5297.22 ± 2.5398.84 ± 0.1997.82 ± 0.3499.34 ± 0.08
Table 21. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the Salinas dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 21. Classification performances obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from the Salinas dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
198.84 ± 1.2999.50 ± 0.9198.75 ± 1.3899.13 ± 0.3599.82 ± 0.1999.48 ± 0.5899.83 ± 0.2899.95 ± 0.06
299.12 ± 0.6999.87 ± 0.1299.93 ± 0.0699.81 ± 0.1599.93 ± 0.1199.92 ± 0.1499.53 ± 0.4699.92 ± 0.14
396.70 ± 2.1999.93 ± 0.0999.66 ± 0.2297.42 ± 3.9499.88 ± 0.20100.0 ± 0.0099.88 ± 0.1199.99 ± 0.02
497.83 ± 1.0697.78 ± 1.9999.51 ± 0.1597.51 ± 2.6998.26 ± 0.3499.40 ± 0.4098.84 ± 0.7499.22 ± 0.22
596.78 ± 2.2098.96 ± 0.7299.01 ± 0.4597.83 ± 1.7097.54 ± 1.0399.71 ± 0.2699.76 ± 0.2299.53 ± 0.39
699.73 ± 0.5899.87 ± 0.0599.96 ± 0.0599.95 ± 0.0499.96 ± 0.0499.86 ± 0.2299.97 ± 0.0499.96 ± 0.04
798.46 ± 1.6999.77 ± 0.1399.84 ± 0.0398.95 ± 1.0699.99 ± 0.0199.76 ± 0.1999.94 ± 0.0899.86 ± 0.13
896.39 ± 2.0797.93 ± 1.3099.58 ± 0.3593.90 ± 1.9495.92 ± 0.6299.23 ± 0.5597.80 ± 0.6199.24 ± 0.37
999.68 ± 0.1699.97 ± 0.0399.91 ± 0.1299.92 ± 0.0999.95 ± 0.0699.98 ± 0.0399.99 ± 0.01100.0 ± 0.00
1098.49 ± 0.6999.10 ± 0.4199.46 ± 0.6197.43 ± 0.6798.72 ± 0.8899.82 ± 0.2698.76 ± 0.5599.51 ± 0.46
1196.14 ± 3.6199.56 ± 0.6299.08 ± 0.3198.95 ± 0.8999.41 ± 0.7999.56 ± 0.2799.50 ± 0.2199.71 ± 0.29
1299.32 ± 0.8099.25 ± 0.5999.88 ± 0.1799.33 ± 0.4499.67 ± 0.4199.82 ± 0.24100.0 ± 0.0099.89 ± 0.10
1396.01 ± 4.8497.32 ± 0.7899.37 ± 0.3296.01 ± 3.4498.33 ± 1.2098.84 ± 0.4399.60 ± 0.3799.58 ± 0.27
1497.83 ± 1.7798.55 ± 0.7598.22 ± 2.1097.75 ± 1.3798.53 ± 0.6498.55 ± 1.3099.20 ± 0.9099.14 ± 0.78
1592.17 ± 3.6396.68 ± 1.7699.76 ± 0.0989.71 ± 0.8594.20 ± 1.9698.43 ± 1.3195.79 ± 0.8998.92 ± 0.28
1696.53 ± 5.9699.43 ± 0.5498.76 ± 0.5998.93 ± 1.3699.07 ± 0.7199.20 ± 1.3999.17 ± 1.2799.54 ± 0.55
OA (%)97.18 ± 1.3798.77 ± 0.2799.59 ± 0.0796.61 ± 0.4697.99 ± 0.1799.44 ± 0.2898.75 ± 0.1699.55 ± 0.07
AA (%)97.50 ± 1.3798.97 ± 0.2199.42 ± 0.0997.66 ± 0.2398.70 ± 0.2399.47 ± 0.2399.22 ± 0.0899.62 ± 0.09
κ (%)96.86 ± 1.5398.64 ± 0.3099.54 ± 0.0896.22 ± 0.5197.76 ± 0.1999.38 ± 0.3298.61 ± 0.1899.50 ± 0.08
Table 22. Classification performances obtained by SQSFormer, ENTA, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the Salinas dataset. The optimal results and suboptimal results are in red and blue, respectively.
Table 22. Classification performances obtained by SQSFormer, ENTA, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from the Salinas dataset. The optimal results and suboptimal results are in red and blue, respectively.
ClassSQSFormerENTACFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
199.89 ± 0.1199.97 ± 0.07100.0 ± 0.00100.0 ± 0.0098.91 ± 1.1799.43 ± 0.8599.92 ± 0.1899.96 ± 0.09
299.92 ± 0.0899.90 ± 0.21100.0 ± 0.00100.0 ± 0.0099.20 ± 1.40100.0 ± 0.0099.75 ± 0.34100.0 ± 0.00
3100.0 ± 0.0099.55 ± 0.9799.93 ± 0.11100.0 ± 0.0098.33 ± 3.45100.0 ± 0.0099.18 ± 1.67100.0 ± 0.00
499.45 ± 0.4199.86 ± 0.1299.11 ± 1.1199.83 ± 0.1299.15 ± 0.9899.70 ± 0.2998.52 ± 0.9799.47 ± 0.39
598.21 ± 1.1097.40 ± 2.9997.29 ± 3.2599.27 ± 0.2699.15 ± 0.5799.83 ± 0.1799.54 ± 0.3099.69 ± 0.26
699.94 ± 0.0699.98 ± 0.0299.49 ± 0.9899.99 ± 0.0299.90 ± 0.05100.0 ± 0.0099.65 ± 0.67100.0 ± 0.00
799.94 ± 0.0799.92 ± 0.0799.99 ± 0.03100.0 ± 0.0099.65 ± 0.4199.99 ± 0.0198.87 ± 0.9799.97 ± 0.05
894.63 ± 1.6691.98 ± 2.7398.89 ± 1.4199.05 ± 0.9196.52 ± 4.7997.99 ± 1.1797.84 ± 1.0299.62 ± 0.32
9100.0 ± 0.0099.95 ± 0.05100.0 ± 0.00100.0 ± 0.0099.98 ± 0.04100.0 ± 0.0099.81 ± 0.11100.0 ± 0.00
1098.53 ± 0.9998.55 ± 0.6198.58 ± 1.6899.74 ± 0.2499.02 ± 0.7599.02 ± 0.7998.16 ± 0.5299.89 ± 0.06
1199.42 ± 0.2898.90 ± 0.9599.74 ± 0.4899.85 ± 0.2399.19 ± 0.7699.93 ± 0.0898.71 ± 0.7499.89 ± 0.10
1299.80 ± 0.1199.99 ± 0.0299.41 ± 0.38100.0 ± 0.0099.75 ± 0.3399.98 ± 0.0399.98 ± 0.0499.74 ± 0.33
1398.34 ± 1.8799.54 ± 0.4698.23 ± 1.7098.75 ± 0.9999.43 ± 0.56100.0 ± 0.0098.86 ± 0.9999.78 ± 0.28
1494.51 ± 4.4398.99 ± 0.9899.23 ± 0.9598.73 ± 0.8397.18 ± 2.9699.68 ± 0.4799.27 ± 0.9799.83 ± 0.23
1592.28 ± 1.8987.19 ± 4.4898.25 ± 1.1198.38 ± 1.2499.59 ± 0.2997.29 ± 0.7897.14 ± 1.1799.92 ± 0.11
1698.79 ± 0.9298.88 ± 0.5899.87 ± 0.2499.98 ± 0.0399.31 ± 0.5799.93 ± 0.0999.23 ± 1.1399.68 ± 0.32
OA (%)97.44 ± 0.2796.27 ± 0.4199.17 ± 0.2499.48 ± 0.1498.80 ± 1.0399.11 ± 0.2498.74 ± 0.2499.84 ± 0.06
AA (%)98.35 ± 0.2298.16 ± 0.2499.25 ± 0.3299.60 ± 0.0799.02 ± 0.3499.55 ± 0.1099.03 ± 0.1999.84 ± 0.05
κ (%)97.15 ± 0.3095.84 ± 0.4599.08 ± 0.2699.42 ± 0.1598.66 ± 1.1499.01 ± 0.2698.59 ± 0.2799.82 ± 0.07
Table 23. Full-scene classification performance obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from five public HSI datasets. The optimal results and suboptimal results are in red and blue, respectively.
Table 23. Full-scene classification performance obtained by S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT from five public HSI datasets. The optimal results and suboptimal results are in red and blue, respectively.
DatasetsMetricsS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
IndianMI (%)36.87 ± 5.3089.33 ± 1.1879.99 ± 0.7150.83 ± 6.7862.02 ± 4.5245.74 ± 3.4767.55 ± 1.7850.63 ± 4.53
BF1 (%)63.39 ± 2.2598.26 ± 0.4897.90 ± 1.0078.23 ± 3.7383.05 ± 3.4066.68 ± 1.3485.37 ± 0.8368.45 ± 0.62
SSIM (%)53.93 ± 5.6288.92 ± 0.6388.97 ± 0.9066.35 ± 2.4772.35 ± 2.8568.99 ± 1.5173.47 ± 0.7258.72 ± 0.77
HoustonMI (%)44.51 ± 5.2191.00 ± 0.4981.41 ± 0.9765.50 ± 5.0573.69 ± 3.5149.62 ± 4.4166.18 ± 1.1360.81 ± 1.55
BF1 (%)95.86 ± 1.3496.74 ± 0.2396.98 ± 0.1296.78 ± 0.3197.04 ± 0.0594.37 ± 1.4996.81 ± 0.3695.63 ± 0.23
SSIM (%)94.40 ± 1.4398.34 ± 0.0598.56 ± 0.0498.08 ± 0.1498.20 ± 0.1097.24 ± 0.1298.05 ± 0.0796.90 ± 0.14
HongHuMI (%)61.16 ± 8.2496.97 ± 0.2294.98 ± 0.4983.96 ± 1.1390.51 ± 0.7452.68 ± 2.2166.03 ± 3.1445.44 ± 2.05
BF1 (%)26.40 ± 5.7497.32 ± 0.0895.36 ± 1.3953.36 ± 3.3677.59 ± 2.5820.60 ± 1.3826.49 ± 0.7016.32 ± 2.22
SSIM (%)26.62 ± 10.191.42 ± 0.0290.12 ± 0.7566.89 ± 2.2478.95 ± 1.3124.25 ± 9.3637.17 ± 3.7013.52 ± 0.21
PaviaUMI (%)51.99 ± 2.1890.02 ± 0.6177.37 ± 1.7963.28 ± 1.9769.59 ± 2.1053.99 ± 2.2866.56 ± 0.7352.35 ± 1.86
BF1 (%)64.96 ± 3.3884.61 ± 0.3085.21 ± 0.1685.57 ± 1.2987.03 ± 0.7872.66 ± 1.1377.30 ± 0.7566.68 ± 0.82
SSIM (%)69.03 ± 0.8684.73 ± 0.4185.56 ± 0.0881.64 ± 0.7883.40 ± 0.2071.59 ± 0.6075.47 ± 0.8969.54 ± 0.34
SalinasMI (%)51.64 ± 5.7296.05 ± 0.2894.54 ± 0.3982.53 ± 1.5489.10 ± 3.1851.58 ± 1.1671.83 ± 1.8350.82 ± 3.08
BF1 (%)29.40 ± 1.5997.11 ± 0.1496.48 ± 0.6081.43 ± 1.9989.38 ± 0.5233.26 ± 1.1449.25 ± 1.7327.54 ± 0.63
SSIM (%)55.38 ± 2.1494.19 ± 0.2293.44 ± 0.6385.22 ± 1.6388.79 ± 0.4758.38 ± 0.5465.33 ± 1.9154.65 ± 0.61
Table 24. Full-scene classification performance obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from five public HSI datasets. The optimal results and suboptimal results are in red and blue, respectively.
Table 24. Full-scene classification performance obtained by SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer from five public HSI datasets. The optimal results and suboptimal results are in red and blue, respectively.
DatasetsMetricsSQSFormerEATNCFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
IndianMI (%)78.65 ± 1.2276.89 ± 0.8472.78 ± 2.7983.54 ± 2.7580.96 ± 2.6988.53 ± 0.8387.34 ± 1.1391.78 ± 0.95
BF1 (%)96.21 ± 0.7691.00 ± 0.6693.93 ± 1.6297.19 ± 1.1597.63 ± 0.6497.96 ± 0.6998.18 ± 0.3699.17 ± 0.10
SSIM (%)86.64 ± 0.7381.74 ± 0.9582.91 ± 1.5488.54 ± 1.4987.97 ± 0.9689.31 ± 0.8787.66 ± 0.7892.73 ± 0.39
HoustonMI (%)82.25 ± 0.8473.20 ± 1.8179.67 ± 0.7583.48 ± 1.4179.01 ± 2.4785.40 ± 0.8889.22 ± 1.3091.31 ± 0.44
BF1 (%)96.77 ± 0.2397.02 ± 0.1396.99 ± 0.0696.85 ± 0.3297.04 ± 0.4196.90 ± 0.1597.09 ± 0.2097.12 ± 0.17
SSIM (%)98.44 ± 0.0698.47 ± 0.0598.46 ± 0.0498.43 ± 0.0798.44 ± 0.0898.48 ± 0.0598.44 ± 0.1198.62 ± 0.05
HongHuMI (%)92.67 ± 0.4691.18 ± 0.3792.82 ± 0.4197.19 ± 0.2696.57 ± 0.2996.24 ± 0.0896.29 ± 0.1497.66 ± 0.11
BF1 (%)87.03 ± 1.4183.09 ± 1.4288.93 ± 2.0298.24 ± 0.1897.06 ± 0.7093.93 ± 0.2395.92 ± 0.4998.53 ± 0.05
SSIM (%)86.49 ± 0.6383.99 ± 0.6384.85 ± 1.1692.13 ± 0.1691.22 ± 0.4689.67 ± 0.1290.58 ± 0.3692.54 ± 0.08
PaviaUMI (%)74.94 ± 0.6170.91 ± 0.6372.42 ± 0.6877.70 ± 2.2782.00 ± 1.8688.69 ± 0.3987.88 ± 0.6291.80 ± 0.25
BF1 (%)85.33 ± 0.5985.50 ± 0.5386.46 ± 0.6185.80 ± 0.3285.15 ± 1.2585.27 ± 0.2585.45 ± 0.4586.84 ± 0.16
SSIM (%)85.02 ± 0.2683.49 ± 0.1984.29 ± 0.3085.43 ± 0.2585.29 ± 0.9085.52 ± 0.0884.86 ± 0.4085.98 ± 0.13
SalinasMI (%)93.37 ± 0.6490.52 ± 0.8789.31 ± 1.8596.41 ± 0.2294.83 ± 0.5695.84 ± 0.2395.29 ± 0.5996.67 ± 0.11
BF1 (%)87.15 ± 1.7480.37 ± 2.0494.15 ± 0.5396.73 ± 0.9296.47 ± 1.1794.75 ± 0.9694.48 ± 1.0497.99 ± 0.20
SSIM (%)88.99 ± 0.3684.15 ± 0.7991.28 ± 0.4393.95 ± 0.693.63 ± 1.0692.57 ± 0.92.26 ± 0.4095.19 ± 0.17
Table 25. Computational complexity of S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT.
Table 25. Computational complexity of S3ARN, PMCN, S3GAN, ViT, SFormer, HFormer, MHCFormer, and D2S2BoT.
DatasetMetricsS3ARNPMCNS3GANViTSFormerHFormerMHCFormerD2S2BoT
IndianParams (M)8.11040.22880.08660.09540.34263.39180.40900.4560
FLOPs (G)1.07520.96972.60630.03810.07080.62580.05690.3742
Training Time(s)67.03130.72366.1423.8726.4059.5940.9537.07
Testing Time (s)3.317.0612.941.221.613.682.102.09
HoustonParams (M)5.26440.19360.05990.09540.22622.80420.40890.4559
FLOPs (G)0.56120.70402.44280.02750.04480.36280.05690.3742
Training Time(s)82.79137.26461.0626.5034.2176.0857.4451.93
Testing Time (s)4.417.2817.701.752.254.623.363.04
HongHuParams (M)12.32830.27300.12590.09580.54154.40710.40940.4567
FLOPs (G)1.95051.30205.78760.05140.11001.07370.05690.3742
Training Time(s)2675.177416.7031,776.6900.771006.451919.271338.531281.81
Testing Time (s)132.11360.211032.4650.3672.69125.2380.7976.82
PaviaUParams (M)3.48040.16750.04650.09500.16442.47540.40850.4552
FLOPs (G)0.29120.50942.34080.01970.02890.22390.05690.3742
Training Time(s)204.73297.331464.4270.1293.51193.33147.42140.20
Testing Time (s)12.3016.2749.144.726.0212.068.878.61
SalinasParams (M)8.33080.23130.08780.09540.35243.44070.40900.4560
FLOPs (G)1.11820.98872.61630.03890.07290.64780.05690.3742
Training Time(s)332.57735.302063.593.17147.86270.07198.95179.84
Testing Time (s)16.5135.2668.836.529.2318.4910.7616.51
Table 26. Computational complexity of SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer.
Table 26. Computational complexity of SQSFormer, EATN, CFormer, SSACFormer, CPFormer, DSFormer, LGDRNet, and CPMFFormer.
DatasetMetricsSQSFormerEATNCFormerSSACFormerCPFormerDSFormerLGDRNetCPMFFormer
IndianParams (M)0.25850.39140.61560.70600.61420.59390.35500.6688
FLOPs (G)0.08390.09490.12350.23190.25920.04212.86890.1495
Training Time(s)9.53129.3267.61182.40116.9847.1938.8833.04
Testing Time (s)0.763.016.545.5710.052.661.582.53
HoustonParams (M)0.14900.31680.61180.37990.32020.59380.35490.2065
FLOPs (G)0.04690.07690.12190.25250.13530.04213.18570.0467
Training Time(s)9.50201.3898.34262.50131.6755.9341.7042.89
Testing Time (s)0.554.057.539.6511.623.942.403.33
HongHuParams (M)0.16140.50250.62171.25271.11360.59470.35530.2855
FLOPs (G)0.05080.17010.12550.71710.46920.04212.90040.0898
Training Time(s)214.561358.341041.49239.82074.21464.41578.61557.54
Testing Time (s)15.61155.57359.00254.58450.3896.8773.89103.06
PaviaUParams (M)0.15970.26990.60760.20350.15970.59300.35460.5526
FLOPs (G)0.05080.06560.12070.11250.05080.04212.29260.1224
Training Time(s)23.75192.68114.54332.19247.22147.11133.71149.93
Testing Time (s)1.727.5314.8717.5924.369.726.629.69
SalinasParams (M)0.16060.39720.61590.73310.63890.59390.35500.2434
FLOPs (G)0.13240.09630.12360.33470.20250.04212.34630.0769
Training Time(s)93.21225.69167.73539.63339.52201.26177.06202.99
Testing Time (s)35.2613.6331.3629.2739.4013.368.5510.76
Table 27. Impact of different attention mechanisms on classification performance. The best results are in red.
Table 27. Impact of different attention mechanisms on classification performance. The best results are in red.
DatasetMetricsWithout CFCLWithout CFEBWithout CSSAWithout CSCAWithout MHCSD2ACPMFFormer
IndianOA (%)93.84 ± 0.9193.96 ± 1.0393.54 ± 1.4993.05 ± 1.0092.61 ± 2.1794.85 ± 0.69
AA (%)93.14 ± 1.3493.69 ± 2.5093.66 ± 4.3593.03 ± 2.0991.17 ± 3.1494.35 ± 0.61
κ (%)92.97 ± 0.9993.11 ± 1.1692.63 ± 1.6992.07 ± 1.1391.59 ± 2.4594.13 ± 0.80
HoustonOA (%)92.95 ± 3.1294.63 ± 1.0194.99 ± 0.5494.49 ± 1.1893.73 ± 4.1195.93 ± 0.36
AA (%)93.44 ± 2.8295.07 ± 0.8895.26 ± 0.7494.75 ± 1.1493.99 ± 3.8496.06 ± 0.26
κ (%)92.38 ± 3.3894.20 ± 1.1094.59 ± 0.5994.05 ± 1.2893.22 ± 4.4595.61 ± 0.39
HongHuOA (%)99.35 ± 0.1199.36 ± 0.0899.34 ± 0.0499.45 ± 0.0499.18 ± 0.0499.53 ± 0.03
AA (%)98.47 ± 0.6598.37 ± 0.1598.46 ± 0.1398.62 ± 0.1398.37 ± 0.1898.99 ± 0.10
κ (%)99.18 ± 0.1499.20 ± 0.1099.17 ± 0.0599.31 ± 0.0599.02 ± 0.0599.41 ± 0.04
PaviaUOA (%)99.26 ± 0.1399.31 ± 0.2199.36 ± 0.2199.27 ± 0.1799.13 ± 0.2599.50 ± 0.06
AA (%)98.64 ± 0.2598.72 ± 0.4198.78 ± 0.3698.68 ± 0.2998.20 ± 0.5699.02 ± 0.10
κ (%)99.01 ± 0.1799.08 ± 0.2899.15 ± 0.2899.04 ± 0.2298.84 ± 0.3399.34 ± 0.08
SalinasOA (%)99.75 ± 0.0799.69 ± 0.0899.68 ± 0.1099.65 ± 0.0899.37 ± 0.2099.84 ± 0.06
AA (%)99.69 ± 0.0899.65 ± 0.0999.61 ± 0.1199.57 ± 0.0999.52 ± 0.0999.84 ± 0.05
κ (%)99.72 ± 0.0999.66 ± 0.0899.65 ± 0.1199.62 ± 0.0999.19 ± 0.2299.82 ± 0.07
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, M.; Yang, Y.; Zhang, S.; Mi, P.; Han, D. CPMFFormer: Class-Aware Progressive Multiscale Fusion Transformer for Hyperspectral Image Classification. Remote Sens. 2025, 17, 3684. https://doi.org/10.3390/rs17223684

AMA Style

Zhang M, Yang Y, Zhang S, Mi P, Han D. CPMFFormer: Class-Aware Progressive Multiscale Fusion Transformer for Hyperspectral Image Classification. Remote Sensing. 2025; 17(22):3684. https://doi.org/10.3390/rs17223684

Chicago/Turabian Style

Zhang, Meng, Yi Yang, Sixian Zhang, Pengbo Mi, and Deqiang Han. 2025. "CPMFFormer: Class-Aware Progressive Multiscale Fusion Transformer for Hyperspectral Image Classification" Remote Sensing 17, no. 22: 3684. https://doi.org/10.3390/rs17223684

APA Style

Zhang, M., Yang, Y., Zhang, S., Mi, P., & Han, D. (2025). CPMFFormer: Class-Aware Progressive Multiscale Fusion Transformer for Hyperspectral Image Classification. Remote Sensing, 17(22), 3684. https://doi.org/10.3390/rs17223684

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop