Next Article in Journal
SAM2MS: An Efficient Framework for HRSI Road Extraction Powered by SAM2
Previous Article in Journal
Post-Disaster High-Frequency Ground-Based InSAR Monitoring and 3D Deformation Reconstruction of Large Landslides Using MIMO Radar
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EDTST: Efficient Dynamic Token Selection Transformer for Hyperspectral Image Classification

1
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
2
College of Science, National University of Defense Technology, Changsha 410073, China
3
College of Advanced Interdisciplinary Studies, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2025, 17(18), 3180; https://doi.org/10.3390/rs17183180
Submission received: 29 July 2025 / Revised: 9 September 2025 / Accepted: 11 September 2025 / Published: 14 September 2025

Abstract

Highlights

What are the main findings?
  • We propose EDTST, a novel and efficient Vision Transformer architecture that integrates large-kernel 3D convolution with a dynamic token selection mechanism for hyperspectral image classification.
  • EDTST achieves state-of-the-art classification accuracy with a 3% improvement in overall accuracy on the WHU-Hi-HanChuan dataset, while requiring the shortest training and inference time among recent models.
What is the implication of the main finding?
  • The model significantly enhances computational efficiency by reducing parameters and FLOPs through innovative architectural design, making it suitable for resource-constrained applications.
  • It establishes a new benchmark for balancing accuracy and efficiency in hyperspectral image analysis, providing a practical solution for real-world remote sensing tasks.

Abstract

Hyperspectral images, characterized by rich spectral information, enable precise pixel-level classification and are thus widely employed in remote sensing applications. Although convolutional neural networks (CNNs) have demonstrated effectiveness in hyperspectral image processing, their limited receptive fields constrain their capacity to capture long-range dependencies. Transformers excel at modeling long-range features for hyperspectral image classification (HSIC). Yet, they often overlook effective representation of local spectral–spatial characteristics while incurring computational redundancy from numerous classification-irrelevant tokens. To address these challenges, we propose EDTST, a state-of-the-art Vision Transformer architecture specifically designed for efficient hyperspectral image classification. The model utilizes a large-kernel 3D convolution block to extract deep spectral–spatial features. A 2D convolution block further refines these features, followed by a novel attention mechanism with dynamic token pruning that substantially reduces the computational load by focusing on the most pertinent features. The process concludes with an adaptive average pooling layer and a fully connected layer for classification. Extensive experiments on four standard hyperspectral datasets demonstrate that EDTST achieves the highest classification accuracy, with a notable 3% improvement in overall accuracy on the WHU-Hi-HanChuan dataset, while requiring the shortest training and inference time among all compared state-of-the-art models from the past three years. These results validate the efficacy of our approach in achieving superior performance with markedly improved computational efficiency.

1. Introduction

Hyperspectral images capture rich spectral information for each pixel, enabling precise classification into distinct ground object categories [1]. This capability has made them invaluable in applications such as precision agriculture, remote sensing, and material analysis. As a fundamental step in hyperspectral data processing, accurate pixel-level classification is crucial for tasks such as terrain mapping and material identification.
The performance of hyperspectral image classification hinges on the extraction of discriminative features. Conventional machine learning approaches—including random forests [2], decision trees [3], and support vector machines (SVMs) [4]—rely on handcrafted features for classification. Zhang et al. [5] pioneered edge-guided feature extraction using Extended Relative Total Variation (ERTV), where spectral dimensionality reduction precedes structure extraction guided by learned edge probability maps. This approach demonstrated significant performance gains with limited training samples. However, these methods demand domain expertise to design effective feature extraction strategies. With the rapid advancement of hyperspectral imaging technology, the increasing diversity of sensors and the exponential growth of data volume have led to greater heterogeneity across datasets. Consequently, traditional feature-based methods often face stability limitations when handling large-scale, highly variable hyperspectral imagery.
In recent years, deep learning has emerged as a powerful alternative for hyperspectral image classification. By leveraging hierarchical feature learning through diverse network architectures, deep learning models can automatically extract robust spectral–spatial features without manual engineering. This data-driven paradigm has achieved significant progress in hyperspectral image processing, demonstrating superior adaptability and performance. Notably, convolutional neural networks (CNNs) have gained widespread adoption in hyperspectral image classification due to their ability to preserve the inherent spatial structure of input data while effectively extracting discriminative spectral–spatial features. This structural preservation enables CNNs to maintain the integrity of local spectral signatures and spatial contextual information throughout the feature extraction process. For example, a multi-scale 3D CNN model [6] was introduced for hyperspectral image classification. Following this, Roy et al. [7] proposed HybridSN, an integrated model leveraging both 2D spatial and 3D spectral–spatial convolutions for hyperspectral image classification.
However, CNNs may be limited in capturing long-range dependencies and the complex spectral–spatial correlations inherent in HSI [8]. Recently, Vision Transformers (ViTs) have emerged as a powerful alternative, leveraging self-attention mechanisms to model global relationships [9]. Nevertheless, standard ViTs are computationally intensive and require large amounts of training data, making them less suitable for HSI where labeled data is limited.
Recent developments in Transformer-based models, such as the Vision Transformer (ViT) [9], have shown promising results in image processing. Several efforts have been made to apply Transformer models to HSI classification. For instance, He et al. [8] proposed a spatial–spectral transformer (SST) model which utilizes a VGGNet-like structure [10] to extract spatial features and dense transformers to model spectral relationships. Hong et al. [11] developed SpectralFormer (SF), which learns spectral features from neighboring bands using cross-layer transformer encoders. More recently, Sun et al. [12] proposed a model that integrates a CNN backbone with transformers. The CNN captures low-level spectral–spatial features, while the transformers learn contextualized spectral relationships and incorporate spatial information. Li et al. [13] proposed a lightweight self-pooling Transformer called SPFormer. SPFormer reduces model complexity by employing a one-layer self-supervised autoencoder for dimensionality reduction and introduces parameter-free modules such as Channel Shuffle for Multihead Self-Pooling with Sparse Mapping (CSSM-MHSP) and a Central Token Mixer (CTM) to enhance spectral feature mapping and pixel-wise information interaction. Liang et al. [14] introduced a transformer–CNN hybrid (FTSCN) that integrates SimAM-based attention mechanisms with hierarchical dense networks. Their dual-branch architecture leverages spatial SimAM modules for discriminative pixel weighting and squeeze-enhanced axial transformers for global spectral dependencies, achieving state-of-the-art accuracy on four datasets while reducing training time by 40–60% compared to 3D-CNN baselines. Zhang et al. [15] proposed GMLMPTF, a pyramid texture filtering (PTF) framework extracting multi-scale global structures through parameter-varied PTF operations, while local features derive from superpixel-optimized PTF outputs. Their probabilistic fusion of global and local probabilities yielded > 4% OA improvement on agricultural datasets with minimal training samples. Similarly, Zhang et al. [16] developed PFS3F to mitigate segmentation inaccuracies by probabilistically fusing superpixel-optimized spatial features with semantic-aware structural features (S2Fs) constrained by edge information. Zhang et al. [17] introduced MSTRF using exponential windowed inherent variance (eWIV) maps to guide structural-textural-aware recursive filtering. This parameter-free attention mechanism dynamically weights spatial–spectral features, enabling > 92% OA on complex urban scenes with five labeled samples/class. Earlier foundational work by Zhang et al. [18] established contour structural profiles (CSPs) using edge-aware total variation models, where KPCA-fused multi-scale CSPs preserved field boundaries with 90.78% OA on Salinas. Xue et al. introduced an Attention-Gated Tuning strategy with a Triplet Transformer (Tri-Former), enabling effective knowledge transfer across different HSI and even RGB datasets [19]. For multimodal fusion, Huang et al. presented MCFTNet, a cross-layer fusion transformer that integrates LiDAR data as an external classification token to improve feature propagation [20]. Zhu et al. proposed HMAT, a hierarchical aggregation network combining multihead axial attention and pyramid convolution to fuse hyperspectral and LiDAR features [21]. Wu et al. further developed a Spectral Spatial Window Attention Transformer that uses cross-window attention to capture long-range dependencies [22].
Despite these advances, existing approaches face several limitations: (1) high computational complexity, (2) inefficient feature extraction from both spatial and spatial–spectral domains, and (3) suboptimal performance under limited data conditions. These limitations are further summarized by the inherent trade-offs between CNN and Transformer-based models as illustrated in Table 1.
Our work addresses these challenges through a novel architecture that synergistically combines the strengths of both paradigms. This work introduces the EDTST, a novel hybrid 3D-CNN and Vision Transformer architecture specifically crafted for efficient hyperspectral image classification. The initial 3D/2D CNN layers serve as a local feature extractor, efficiently capturing discriminative spectral–spatial patterns from neighboring pixels. Subsequently, the Transformer encoder operates on this condensed feature set, modeling the long-range contextual dependencies that are crucial for disambiguating spectrally similar but semantically distinct categories. This design not only enhances model performance by leveraging both local and global information but also significantly improves computational efficiency by reducing the sequence length of tokens presented to the Transformer. The primary contributions of our research are detailed as follows:
  • To significantly reduce the parameter count and computational cost of 3D convolutions, we introduce a novel and efficient large-kernel 3D convolution block designed specifically for hyperspectral image processing. The architecture employs a spatial 7 × 7 × 7 convolution followed by two pointwise (1 × 1 × 1) convolutions. This design achieves a receptive field equivalent to that of three stacked 3 × 3 × 3 convolutions, while using only 39.8% of the parameters and 40% of the FLOPs. The module maintains strong representational capacity and is highly suitable for processing high-dimensional hyperspectral data due to its efficient parameter allocation and competitive computational efficiency.
  • To alleviate the computational redundancy associated with processing a large number of tokens in the Transformer block, we propose a dynamic token selection mechanism that significantly reduces computational complexity. By selectively retaining the 75% most informative tokens and pruning redundant ones, this strategy focuses computational resources on the most critical features, thereby maintaining or even enhancing the model’s classification accuracy while improving efficiency.
  • Extensive experimental evaluations on multiple benchmark hyperspectral datasets demonstrate that our proposed model, EDTST, consistently outperforms a range of state-of-the-art methods in classification accuracy while achieving superior computational efficiency. Notably, on the WHU-Hi-HanChuan dataset, EDTST attains a notable 3% improvement in overall accuracy compared to leading models proposed in recent years. Furthermore, EDTST requires the shortest training and inference time across all datasets tested, underscoring its practical utility for resource-constrained scenarios. These results conclusively validate the effectiveness of our architectural choices in balancing high performance with operational efficiency.
These contributions collectively enhance the applicability and effectiveness of Vision Transformers in hyperspectral image analysis, setting new benchmarks for future research in the field.
The rest of the paper is organized as follows. Section 2 introduces the related work. Section 3 provides a detailed description of the proposed framework. Section 4 presents experimental results on four hyperspectral datasets. The ablation analyses are reported in Section 5. Finally, Section 6 concludes the paper and summarizes our study.

2. Related Work

The attention mechanism in Transformer models [23] initially demonstrated remarkable success in natural language processing (NLP). Subsequent research extended this capability to computer vision, where Dosovitskiy et al. [9] pioneered Vision Transformer (ViT) for image classification. The long-range dependency modeling of ViT proves particularly advantageous for hyperspectral imagery (HSI), effectively capturing rich spectral information across hundreds of bands while overcoming the limited receptive fields of convolutional networks. This prompted further innovations: He et al. [24] developed HSI-BERT with bidirectional encoding, while He et al. [25] introduced Spa-Spe-TR, a dual-branch architecture with dedicated spectral (Spe-TR) and spatial (Spa-TR) transformer modules.
Recognizing limitations in pure transformer approaches, recent works explore hybrid CNN–transformer architectures [8,12,26]. For instance, Ouyang et al. [27] proposed HybridFormer, integrating convolutions with attention to capture global spectral–spatial dependencies. Similarly, Qi et al. [28] designed a global–local 3D convolutional transformer to model local spectral correlations within global sequences, and Zhao et al. [29] created CTFSN, fusing local–global features through additive and channel-superposition methods.
Recent advances continue to refine this hybrid paradigm. Zhao et al. proposed a Self- and Cross-Attention Enhanced Transformer (SCAET) that employs a dual-branch CNN for visible and thermal infrared HSI feature extraction, followed by interactive enhancement through attention mechanisms [30]. To improve center pixel discrimination, Jia et al. introduced CenterFormer, which employs a center spatial–spectral attention mechanism to emphasize label-relevant pixels [31], while Yu et al. developed a Center-Specific Perception Transformer (CP-Transformer) that uses label-guided attention to prioritize spectral and spatial features related to the center pixel [32].
Efficiency and deployment concerns have also been addressed through novel architectures. Hu et al. designed BinaryViT, a binary Vision Transformer that significantly reduces computational cost through adaptive binarization while maintaining competitive accuracy [33]. Wang et al. proposed an Efficient Attention Transformer Network (EATN) that incorporates self-similarity descriptors to enhance feature relevance and reduce redundancy [34].
While existing methods have made considerable progress in either improving accuracy or reducing computational overhead, few successfully balance both objectives. Many hybrid models still process a large number of tokens, leading to redundant computations, and often overlook the importance of dynamic, input-aware feature reduction. Some efficiency-oriented approaches resort to techniques such as network-wide architectural modifications, external self-similarity descriptors, or binarization. While these strategies can reduce inference cost to some extent, they often introduce additional training complexity or auxiliary computational burdens, without fundamentally addressing the core issue of token redundancy in the attention mechanism itself.
In contrast, our approach, EDTST, introduces a dynamic token selection mechanism within the transformer module to aggressively reduce sequence length without sacrificing discriminative power. Rather than relying on external aids or reduced-precision arithmetic, our method maintains full precision while adaptively focusing computation only on semantically rich regions. Furthermore, by combining 3D convolutional feature extraction with a transformer encoder, we achieve more effective local–global integration than purely attention-based or convolution-only models. This integrated design not only improves accuracy but also substantially decreases computational load during both training and inference, by directly targeting and pruning redundant tokens at the feature level.

3. Materials and Methods

The proposed EDTST architecture introduces three key innovations for efficient hyperspectral image classification: (1) a large-kernel 3D convolution block, (2) a 2D convolution block, and (3) a Transformer block. Figure 1 illustrates the overall architecture. In this section, we detail each component and their synergistic integration.

3.1. Large-Kernel 3D Convolution Block

The large-kernel 3D convolution block serves as the primary spectral–spatial feature extractor in our architecture, designed to process raw hyperspectral data cubes while preserving critical spectral information. As shown in Figure 2, this component operates directly on the input tensor X R B × 1 × D × H × W , where B denotes the batch size, D the spectral depth (number of bands), and  H × W the spatial dimensions. The block comprises three carefully designed operations.
Feature Extraction Layer: The core operation employs a multi-stage convolutional module featuring the following.
A 3D convolution with kernel size 7 × 7 × 7 and padding 3 to maintain spatial–spectral dimensions, implemented as
Y 1 = Conv 3 D 7 × 7 × 7 ( X )
Group normalization with number of groups equal to input channels, ensuring stable training dynamics:
Y 2 = γ Y 1 μ σ 2 + ϵ + β
Channel expansion via 1 × 1 × 1 convolution increasing dimensionality by an expansion ratio of r = 4 :
Y 3 = Conv 3 D 1 × 1 × 1 ( Y 2 )
Gaussian Error Linear Unit (GELU) activation for non-linear transformation:
Y 4 = 1 2 Y 3 1 + erf Y 3 2
Channel compression via 1 × 1 × 1 convolution, reducing dimensionality to target channels:
Y 5 = Conv 3 D 1 × 1 × 1 ( Y 4 )
This hierarchical design enables efficient feature extraction across spectral–spatial domains while maintaining tensor dimensions ( D , H , W ) .
Normalization Layer: The extracted features undergo 3D batch normalization:
Z = Y 5 μ B σ B 2 + ϵ γ + β
where μ B and σ B are batch-wise statistics computed across spatial–spectral dimensions. This operation mitigates internal covariate shift and accelerates convergence.
Activation Layer: Finally, GELU activation enhances representation power:
O = GELU ( Z )
producing the output tensor O R B × 8 × D × H × W .
The design prioritizes parameter efficiency through the following.
Large-kernel spatial–spectral convolution ( 7 3 ) capturing wide contextual relationships;
Channel expansion strategy ( 1 4 8 channels) enhancing representational capacity;
Minimal parameters (410 total) enabling deployment in resource-constrained environments.
As verified in Table 2, the block maintains spectral dimensionality D throughout processing, preserving critical hyperspectral information for downstream components. The output serves as enriched input to the 2D spatial processing block described in Section 3.2.

3.2. Two-Dimensional Convolution Block

Following the spectral–spatial feature extraction in Section 3.1, the 2D convolution block processes the output tensor O R B × 8 × D × H × W to extract discriminative spatial features while reducing computational complexity. As illustrated in Figure 1, this component transforms the volumetric representation into compact spatial embeddings suitable for subsequent transformer processing. As shown in Figure 3, the block employs a basic yet effective convolutional structure comprising three core operations.
Feature Transformation Layer: The primary operation utilizes a 2D spatial convolution with kernel size 3 × 3 to capture local spatial patterns:
Y = Conv 2 D 3 × 3 ( O f l a t )
where O flat R B × C in × H × W is formed by flattening the spectral dimension ( C in = 8 × D ). The convolution uses stride 1 and padding 1 to maintain spatial dimensions, formally expressed as
Y b , c o u t , i , j = k = 1 C in m = 1 1 n = 1 1 W c o u t , k , m , n O f l a t b , k , i + m , j + n + b c o u t
This operation effectively compresses spectral-channel information while preserving spatial relationships.
Normalization Layer: Batch normalization is applied to stabilize feature distributions:
Z = Y μ B σ B 2 + ϵ γ + β
where μ B , σ B are batch-wise statistics computed across spatial positions, and  γ , β R E are learnable affine parameters. This operation mitigates covariate shift during training, particularly crucial when transitioning from 3D to 2D representations.
Activation Layer: GELU activation introduces non-linear transformation while maintaining gradient continuity:
O 2 D = GELU ( Z )
producing the output tensor O 2 D R B × E × H × W .
As shown in Table 3, the block achieves linear complexity relative to input channels C in , making it scalable for high-dimensional hyperspectral data. The output O 2 D serves as spatial token embeddings for the Transformer block described in Section 3.3, preserving spatial topology while encoding rich spectral–spatial features.

3.3. Transformer Block

The Transformer block serves as the final feature refinement stage in the EDTST architecture, processing spatial embeddings O 2 D R B × E × H × W from Section 3.2 to capture long-range dependencies. As shown in Figure 4, this component consists of three sequential operations.

3.3.1. Input Transformation

Spatial features are converted to token sequences compatible with Transformer processing:
X 0 = flatten ( O 2 D ) X 0 R B × N × E
where N = H × W is the token count. Learnable positional encodings P R 1 × N × E preserve spatial relationships:
X 1 = X 0 + P

3.3.2. Attention Mechanism

The core operation of the attention mechanism can be formulated as follows. The input is first projected into queries ( Q ), keys ( K ), and values ( V ) using linear transformations, which are then segmented into H attention heads. The normalized queries and keys are used to compute a raw attention score matrix S :
S = L 2 Norm ( Q ) · L 2 Norm ( K ) · τ · 1 D h
where τ is a learnable temperature vector that scales the attention distribution per head, and  D h = C / H is the feature dimension per head.
The distinctive step in the attention mechanism is the token selection process. Instead of using a full attention matrix, the attention mechanism constructs a sparse attention matrix S ˜ by retaining only the connections to the top-k most relevant keys for each query, where k = r · N and r is a pre-defined retention ratio (e.g., r = 0.75 ). This is achieved by applying a binary mask M that identifies the indices of the top-k elements in S along the key dimension:
S ˜ = softmax S M + ( ) ( 1 M )
Here, ⊙ denotes the Hadamard product, and the mask M is defined such that M i , j = 1 if the key j is among the top-k keys for query i, and  M i , j = 0 otherwise. The softmax operation is applied to the masked scores, effectively ignoring non-top-k entries.
The final output for each head is computed as the weighted sum of the value projections using the sparsified attention weights, O h = S ˜ V h . The outputs from all heads are then concatenated and linearly projected to form the final output Z .

3.3.3. Feature Refinement Module

The Feature Refinement Module incorporates spatial reasoning and gating mechanisms to refine feature representations through a combination of convolutional processing and adaptive feature selection.
Architecture Overview
The Feature Refinement Module processes an input tensor X R B × N × C , where B is the batch size, N = 121 is the number of tokens, and  C = 64 is the feature dimension. The network consists of three main components: (1) a spatial feature reorganization block, (2) a gated feature transformation block, and (3) a projection layer.
Spatial Feature Reorganization
The Feature Refinement Module begins by reorganizing the token sequence into a spatial grid representation, leveraging the inherent spatial relationships between tokens. For an input sequence of N = 121 tokens, we reshape the tensor into an 11 × 11 spatial grid:
X spatial = Reshape ( X ) R B × C × H × W
where H = W = 11 . This spatial reorganization allows the model to capture local structural patterns that may be lost in a purely sequential representation.
The spatial tensor is then split along the channel dimension into two components:
X 1 , X 2 = Split ( X spatial )
where X 1 R B × C conv × H × W and X 2 R B × C untouched × H × W , with  C conv = C untouched = 32 .
The first component X 1 undergoes convolutional processing with a 3 × 3 kernel:
X 1 = Conv 2 D 3 × 3 ( X 1 )
This convolutional operation enhances the model’s ability to capture local spatial dependencies. The processed features are then concatenated with the untouched features:
X recon = Concat ( X 1 , X 2 ) R B × C × H × W
Finally, the spatial representation is flattened back to the original token sequence format:
X flat = Flatten ( X recon ) R B × N × C
Gated Feature Transformation
The second stage of the Feature Refinement Module employs a gating mechanism to adaptively control information flow. The input is first projected to a higher-dimensional space:
X proj = Linear C 4 C ( X flat )
The projected features are then split into two equal parts along the channel dimension:
G 1 , G 2 = Split ( X proj )
One part ( G 1 ) undergoes spatial processing through a depthwise separable convolution:
G 1 = DepthwiseConv 2 D ( G 1 )
This operation applies separate convolutional filters to each input channel, efficiently capturing spatial patterns while minimizing computational cost.
The final output is computed through a gating mechanism that combines the spatially processed features with the original projected features:
X gate = G 1 G 2
where ⊙ denotes element-wise multiplication.
Output Projection
The gated output is projected back to the original feature dimension:
Z = Linear 2 C C ( X gate )

3.3.4. Architectural Advantages

The proposed design provides:
Adaptive Computation: A novel token selection mechanism (with a pruning ratio of k = 0.75) dynamically reduces the sequence length during attention calculation, significantly lowering the quadratic complexity based on input content.
Hybrid Representation Learning: The model synergistically combines the global contextualization of multihead self-attention (H = 8) with the local feature extraction prowess of 3 × 3 convolutions, effectively capturing both long-range dependencies and fine-grained local patterns.
Parallel Feature Processing: The Feature Refinement Module (FRM) employs a split–process–merge strategy, applying spatial convolutions to a subset of channels and channel-wise projections to the remainder, enabling simultaneous and efficient local and global feature refinement within a single module.
As verified in Table 4, the block achieves significant efficiency gains. The output X L R B × N × E undergoes global average pooling before the classification layer, forming the final hyperspectral representation that integrates spectral–spatial features with global contextual relationships.

4. Results

This section consists of datasets, backbone models, and result analysis. To evaluate the model performance, we adopted five metrics: overall accuracy (OA), average accuracy (AA), Kappa coefficient (Kappa), Matthews correlation coefficient (MCC), geometric mean (G-Mean), training time, and testing time. OA measures the ratio of correctly classified pixels over the total pixels. AA is the average accuracy of each class, and Kappa measures the consistency between the result and the truth. MCC comprehensively considers true positives, true negatives, false positives, and false negatives, providing a balanced measure especially useful in scenarios of class imbalance. G-Mean, the geometric mean of recall rates of all classes, reflects the model’s balanced performance across different categories and is particularly suitable for evaluating classification effectiveness on imbalanced datasets. The training time is the time required for model training, and the testing time is the time required to classify the test set.
The metrics are computed as follows. Let n be the total number of samples, C the number of classes, and  M the C × C confusion matrix, where M i j represents the number of samples of class i predicted as class j. We define the following:
  • Row sum: R i = j = 1 C M i j (actual samples in class i).
  • Column sum: C j = i = 1 C M i j (predicted samples in class j).
  • Overall Accuracy (OA) is the proportion of correctly classified samples:
    OA = i = 1 C M i i n × 100 %
  • Average Accuracy (AA) is the mean of class-specific producer’s accuracies:
    AA = 1 C i = 1 C M i i R i × 100 %
  • Kappa Coefficient measures agreement between predictions and ground truth:
    κ = p a p e 1 p e × 100 %
    where p a = OA is the observed agreement, and  p e is the expected agreement by chance:
    p e = 1 n 2 i = 1 C ( R i × C i )
  • Matthews Correlation Coefficient (MCC) for multi-class classification is calculated using the standard formula:
    MCC = n × k = 1 C M k k k = 1 C ( R k × C k ) ( n 2 k = 1 C R k 2 ) × ( n 2 k = 1 C C k 2 ) × 100 %
    This formulation corresponds to the multi-class implementation in scikit-learn’s matthews_corrcoef function.
  • Geometric Mean (G-Mean) is the geometric mean of class-wise recall:
    G - Mean = i = 1 C M i i R i 1 / C × 100 %
    In implementation, we apply smoothing max ( M i i R i , 10 7 ) to avoid zero values.
  • Training Time and Testing Time are recorded in seconds during model training and inference.

4.1. Data Description

In this study, we utilized four distinct hyperspectral datasets to evaluate our image classification methods: QUH-Tangdaowan [35], WHU-Hi-HanChuan [36], Indian Pines, Salinas. Each dataset was selected for its unique properties and relevance to specific aspects of hyperspectral imaging and classification.
The QUH-Tangdaowan dataset, acquired over Tangdao Bay National Wetland Park, Qingdao, China, on 18 May 2021, features ultra-high spatial resolution (0.15 m) imagery captured at a 300 m flight altitude. Comprising 1740 × 860 pixels across 176 spectral bands (400–1000 nm), this dataset encompasses 18 challenging land cover classes, including diverse vegetation (Coniferous pine, Buxus sinica, Populus, Ulmus pumila L, Ligustrum vicaryi, Photinia serrulata, Bulrush, Spiraea, and Grassland), artificial surfaces (rubber track, asphalt, boardwalk, flagging, and gravel road), and natural features (sandy, rocky shallows, bare soil, and seawater).
WHU-Hi-HanChuan, acquired on 17 June 2016, documents a diverse peri-urban landscape near Hanchuan, also in Hubei Province. It includes agricultural land interspersed with small settlements and natural vegetation, presenting a complex mosaic of land uses. The dataset features 274 spectral bands and spans an area of 1217 × 303 pixels, with a very high spatial resolution of about 0.109 m, enabling detailed analyses of both crop types and urban infrastructure.
The Indian Pines dataset, sourced from the AVIRIS sensor over northwestern Indiana, USA, is particularly noted for its coverage of mixed agricultural and forest land. The original data comprises 224 spectral bands. After removing bands covering the water absorption regions ([104–108], [150–163], and 220), the dataset used in this study consists of 200 spectral bands in the range of 0.4–2.5 μm and 145 × 145 pixels. This dataset captures a variety of perennial and seasonal vegetation as well as man-made features, offering a broad spectrum for algorithm testing and development in rural settings.
The Salinas dataset, also captured by the AVIRIS sensor, focuses on the Salinas Valley, California, a region known for its intensive agriculture, particularly vegetables and vineyards. The original imagery has 224 spectral bands. The water absorption bands ([108–112], [154–167], and 224) are discarded, resulting in a processed dataset of 204 spectral bands. It features a high spatial resolution of 3.7 m per pixel, covering an area of 512 lines by 217 samples. The Salinas dataset is ideal for analyzing agricultural health and productivity due to its detailed ground truth data and extensive spectral coverage.
Each dataset contributes uniquely to our study, providing a rich basis for exploring and refining hyperspectral image classification methods across different landscapes and land use types. The diverse range of environments from urban to agricultural and natural settings enhances the robustness and applicability of our findings in the hyperspectral image classification domain.

4.2. Experiment Settings

4.2.1. Implementation Details

To ensure the statistical reliability of our experimental results, we conducted 5 independent runs with different random seeds for both model initialization and training sample selection. In total, 25 samples per class were selected for training in the QUH-Tangdaowan, WHU-Hi-HanChuan and Salinas datasets, while 10 samples per class were used for the Indian Pines dataset due to its smaller size. Table 5 provides detailed information about the numbers of training and test samples for each dataset. The performance metrics (overall accuracy (OA), average accuracy (AA), Kappa coefficient, Matthews correlation coefficient (MCC), geometric mean (G-Mean), training time, and testing time) are reported as mean ± standard deviation across these 5 runs to demonstrate the robustness and stability of our proposed method. The confusion matrices for the five individual runs and the average confusion matrix of our proposed method are provided in Appendix A.
For data preprocessing, we applied Principal Component Analysis (PCA) to reduce the spectral dimensionality to 40 components while preserving the essential spectral information. The selection of Principal Component Analysis (PCA) for dimensionality reduction was motivated by a comparative assessment of its advantages over alternative techniques in the context of hyperspectral data. Unlike supervised methods such as Linear Discriminant Analysis (LDA), which are prone to overfitting under the small-sample-size scenario typical of HSI classification, PCA operates unsupervisedly, avoiding this pitfall. Moreover, while non-linear methods like autoencoders or t-SNE can model complex structures, they are often computationally intensive and prioritize visualization over classification clarity. PCA, in contrast, provides a computationally efficient solution that directly addresses the core issue of band multicollinearity by deriving orthogonal components. This transformation not only reduces dimensionality but also preserves the dominant, discriminative spectral variances essential for a classifier, establishing PCA as the optimal choice for this study. The input patch size was set to 11 × 11 × 40 for all experiments. All experiments were conducted on a server equipped with a NVIDIA RTX 6000 GPU and 128 GB RAM. The network was trained using the AdamW optimizer with an initial learning rate of 1 × 10−4 for 100 epochs, using a mini-batch size of 64. As a preprocessing step for model training, the input image patches were normalized to the range [0, 1].

4.2.2. Baseline Methods

For comprehensive evaluation, we compared our proposed EDTST with several representative methods proposed in recent three years, including the following:
  • Three CNN-based models: 3D ASPP multi-scale feature fusion network (3A-MFFN) [37], spatial–spectral ConvNeXt (SS-ConvNeXt) [38], and dual-branch convolutional transformer network (DCTN) [39].
  • Four state-of-the-art transformer-based models: global–local multigranularity transformer (GLMGT) [40], multi-scale hierarchical conv-aided fourierformer (MHCFormer) [41], 3D-convolution guided spectral–spatial transformer (3D-ConvSST) [42], and dual selective fusion transformer (DSFormer) [43].
For fair comparison, all baseline methods were implemented using their original configurations as reported in their respective papers.

4.3. Classification Results

4.3.1. Classification Results of the QUH-Tangdaowan Dataset

As shown in Table 6 and Figure 5, the proposed EDTST achieves superior performance on the QUH-Tangdaowan dataset, attaining the highest OA of 85.75 % ± 1.62 % , alongside leading results in AA ( 91.75 % ± 0.58 % ), κ ( 84.07 ± 1.76 ), MCC ( 84.38 % ± 1.67 % ), and G-Mean ( 91.26 % ± 0.69 % ). The strong performance across all metrics, particularly MCC and G-Mean, indicates both excellent per-class balance and overall classification capability. Notably, EDTST delivers the best accuracy in several challenging classes, including C5 ( 98.16 % ± 1.16 % ), C7 ( 83.25 % ± 3.25 % ), C10 ( 99.92 % ± 0.05 % ), C13 ( 100.00 % ± 0.00 % ), C14 ( 100.00 % ± 0.00 % ), and C16 ( 70.56 % ± 4.99 % ). While certain methods excel in specific classes (e.g., GLMGT in C1 and C8, and MHCFormer in C3 and C4), EDTST maintains highly competitive performance across all categories. A significant advantage of our approach is its computational efficiency. EDTST requires only 12.15 ± 0.02 seconds for training and 59.92 ± 0.13 seconds for testing, substantially outperforming all competitors in both training and inference speed.

4.3.2. Classification Results of the WHU-Hi-HanChuan Dataset

Table 7 and Figure 6 confirm the strong generalization of EDTST on the WHU-Hi-HanChuan dataset. Our method achieves the highest OA ( 89.78 % ± 0.53 % ), AA ( 88.86 % ± 0.53 % ), κ ( 88.10 ± 0.61 ), MCC ( 88.16 % ± 0.61 % ), and G-Mean ( 88.33 % ± 0.64 % ), outperforming the second-best method by a considerable margin. The superior G-Mean score demonstrates the effectiveness of EDTST in handling class imbalance, while the high MCC value reflects a strong correlation between predicted and actual classifications. EDTST obtains the best accuracy in seven classes, most notably in C2 ( 82.62 % ± 5.78 % ), C8 ( 77.81 % ± 5.80 % ), C9 ( 81.88 % ± 7.37 % ), and C13 ( 72.68 % ± 5.72 % )—categories characterized by complex spatial–spectral characteristics. The computational efficiency is again remarkable, with EDTST completing training in 10.71 ± 0.01 seconds and testing in 27.81 ± 0.07 seconds, representing a 1.48 × and 1.09 × speedup over the fastest competitor, respectively.

4.3.3. Classification Results of the Indian Pines Dataset

The results on the Indian Pines dataset (Table 8) highlight a particularly interesting phenomenon: the AA ( 89.93 % ± 0.41 % ) significantly exceeds the OA ( 82.06 % ± 0.39 % ) across all methods, including our proposed EDTST. This discrepancy arises from the highly imbalanced class distribution inherent to the Indian Pines dataset. Several classes (e.g., C1, C7, C9, C13, and C16) contain very few samples (ranging from 10 to 195 test samples) but achieve near-perfect classification accuracy (often 100%). These high accuracies disproportionately elevate the AA, which is the arithmetic mean of per-class accuracies. In contrast, the OA is weighted by sample count and is heavily influenced by the performance on majority classes (e.g., C2, C3, C10, C11, C12), where accuracy is notably lower due to higher spectral confusion and within-class variability.
Despite these challenges, EDTST achieves the highest OA, AA, κ ( 79.80 ± 0.42 ), MCC ( 80.07 % ± 0.36 % ), and G-Mean ( 89.11 % ± 0.33 % ). The strong G-Mean performance is particularly noteworthy, as it demonstrates our method’s ability to maintain balanced sensitivity across both minority and majority classes. EDTST shows particularly strong gains in the majority classes: C2 ( 76.78 % ± 5.34 % ), C10 ( 82.91 % ± 4.80 % ), C11 ( 69.31 % ± 4.40 % ), and C12 ( 73.55 % ± 3.64 % ). EDTST also maintains the fastest computational time, requiring only 4.24 ± 0.07 seconds for training and 1.05 ± 0.01 seconds for testing. The classification maps in Figure 7 further validate the effectiveness of EDTST, showing fewer misclassified pixels in agriculturally dominant regions.

4.3.4. Classification Results of the Salinas Dataset

Results on the Salinas dataset (Table 9) demonstrate the robustness of EDTST in a controlled agricultural setting. Our method achieves the highest OA ( 97.36 % ± 0.82 % ), AA ( 98.93 % ± 0.27 % ), κ ( 97.07 ± 0.91 ), MCC ( 97.09 % ± 0.89 % ), and G-Mean ( 98.89 % ± 0.29 % ), with particularly notable performance in the more challenging classes: C8 ( 91.05 % ± 3.39 % ) and C15 ( 95.60 % ± 2.53 % ). The near-perfect G-Mean score ( 98.89 % ) indicates exceptional balance in class-wise sensitivity and specificity. The Salinas scene, with its extensive homogeneous regions, is generally easier to classify as reflected in the high accuracies across all methods. Nevertheless, EDTST consistently outperforms its counterparts, achieving perfect ( 100 % ) accuracy in five of the sixteen classes. Computationally, EDTST remains the most efficient, with training and testing times of 10.79 ± 0.01 and 5.80 ± 0.01 seconds, respectively. The classification maps in Figure 8 are visually similar across all advanced methods.

4.3.5. Overall Performance

The consistent superiority of EDTST across all four datasets and all evaluation metrics underscores its effectiveness in capturing discriminative spatial–spectral features for HSI classification. The integration of efficient depthwise transformers and spectral attention mechanisms allows the model to achieve state-of-the-art accuracy while significantly reducing the computational overhead. The observed phenomenon in the Indian Pines dataset, where AA exceeds OA, is a classic characteristic of imbalanced data and highlights the importance of evaluating multiple metrics—AA ensures good performance across all classes, while OA reflects the effectiveness of the dataset as a whole. The strong performance in MCC and G-Mean across all datasets demonstrates the robustness of EDTST in handling both balanced and imbalanced classification scenarios. The minimal visual differences in classification maps among the top-performing methods suggest that quantitative metrics provide a more discriminative evaluation than qualitative inspection alone. EDTST strikes an optimal balance between accuracy and efficiency, making it particularly suitable for practical applications requiring rapid and reliable analysis.

5. Discussion

In this section, we conduct a comprehensive ablation study to evaluate the contributions of key components in the proposed method, including the large-kernel 3D convolution block, the Transformer block, the patch size selection, and the use of PCA for dimensionality reduction. Performance is assessed using five widely recognized metrics: overall accuracy (OA), average accuracy (AA), Kappa coefficient ( κ ), Matthews correlation coefficient (MCC), and G-Mean.

5.1. Effect of Large-Kernel 3D Convolution Block and Transformer Block

The ablation results presented in Table 10 provide several important insights into the role of each component in our architecture:
The combined use of large-kernel 3D convolution block and Transformer block consistently yields the best performance across all datasets, indicating that these two components capture complementary types of features. The large-kernel 3D convolution block excels at extracting local spectral–spatial characteristics, while the Transformer block effectively models long-range dependencies within the data. This synergistic effect is particularly evident in the WHU-Hi-HanChuan dataset, where the full model achieves an OA of 89.78%, outperforming both individual components.
Notably, the large-kernel 3D convolution block demonstrates fundamental importance for hyperspectral data processing. Across three of the four datasets (WHU-Hi-HanChuan, Indian Pines, and Salinas), models incorporating 3D convolution show better performance than those using only the Transformer block. For instance, on the Salinas dataset, the large-kernel 3D convolution block alone achieves an OA of 97.22%, significantly higher than the Transformer-only configuration (95.76%). This performance gap highlights the indispensable role of explicitly modeling spectral–spatial relationships, which provides a stronger foundational feature representation for the transformer to refine.
For challenging scenarios with class imbalance such as Indian Pines, the combined model shows notable improvements in AA (89.93%) and G-Mean (89.11%), indicating better handling of minority classes.
Additionally, the full model exhibits enhanced stability during training as evidenced by reduced standard deviations across multiple metrics, particularly on the Indian Pines dataset. The consistent improvements in κ and MCC values further confirm that our integrated approach produces more reliable classifications across diverse hyperspectral scenes.

5.2. Influence of Patch Size

We systematically evaluate the impact of patch size on classification performance, with the results summarized in Table 11. A clear trend emerges across all datasets: larger patch sizes generally improve classification accuracy up to a certain point, beyond which gains diminish or become negligible.
For all four datasets, a patch size of 11 provides the optimal balance between classification performance and computational efficiency. On the QUH-Tangdaowan dataset, increasing the patch size from 5 to 11 improves OA by nearly 10 percentage points (76.46% to 85.75%), with corresponding improvements across all other metrics. Similar trends are observed for WHU-Hi-HanChuan and Indian Pines, where patch size 11 delivers the best performance (89.78% and 82.06% OA, respectively).
Notably, larger patch sizes not only improve accuracy but also enhance result stability. For Indian Pines, using a patch size of 11 reduces the standard deviation of OA to just ±0.39%, compared to ±3.16% for size 9. This increased stability is consistent across datasets and suggests that larger receptive fields help the model learn more robust features.
Based on these experiments, we establish patch size 11 as the optimal configuration for our method, providing excellent performance across diverse hyperspectral scenes while maintaining reasonable computational requirements.

5.3. Influence of PCA Dimensionality Reduction

The results of our PCA dimensionality reduction analysis are presented in Table 12. These experiments reveal that appropriate dimensionality reduction significantly enhances model performance compared to using raw spectral bands, with optimal results achieved at 40 principal components across all datasets.
Using PCA preprocessing consistently improves classification accuracy over using the full spectral dimension. For the WHU-Hi-HanChuan dataset, employing 40 PCA components yields an OA of 89.78%, compared to 84.08% without PCA—an improvement of more than 5 percentage points. Similar substantial gains are observed for QUH-Tangdaowan (85.75% vs. 71.95%) and Salinas (97.36% vs. 93.06%). This demonstrates that judicious dimensionality reduction removes redundant spectral information while preserving discriminative features.
Beyond improving accuracy, PCA preprocessing also enhances training stability as evidenced by reduced standard deviations across multiple runs. For Indian Pines, using 40 components reduces the OA standard deviation to just ±0.39%, compared to ±2.06% without PCA. This stabilizing effect is consistent across datasets and suggests that PCA helps the model learn more robust representations by eliminating noise and redundancy in the spectral domain.
These results confirm that PCA is an essential preprocessing step for our method, providing both computational efficiency through dimensionality reduction and improved classification performance through noise removal and feature enhancement.

5.4. Overall Discussion

Our comprehensive ablation study demonstrates that each component of the proposed method contributes significantly to its overall performance. The large-kernel 3D convolution block provides essential local spectral–spatial feature extraction capabilities, while the Transformer block effectively captures long-range contextual relationships. The combination of these components yields synergistic effects that outperform either individual approach.
The selection of an appropriate patch size and PCA dimensions further optimizes performance. A patch size of 11 provides sufficient spatial context for accurate classification without excessive computational burden. Similarly, reducing spectral dimensionality to 40 principal components preserves discriminative information while improving training stability.
These findings highlight the importance of considering both architectural innovations and appropriate preprocessing techniques when designing hyperspectral image classification systems. Our method strikes an effective balance between model complexity and performance, delivering state-of-the-art results across diverse hyperspectral datasets.

6. Conclusions

This paper presents EDTST, a novel Vision Transformer architecture that effectively addresses the challenges of hyperspectral image classification through several key innovations. By integrating 3D convolutions for initial spectral–spatial feature extraction, our model demonstrates superior feature learning capabilities while maintaining computational efficiency. The proposed progressive token pruning strategy in our attention mechanism has proven particularly effective, reducing 25% of tokens while preserving classification performance. This adaptive approach not only decreases computational complexity but also helps the model focus on the most informative spectral–spatial features.
Despite these advantages, the proposed method has certain limitations. First, the fixed token pruning ratio of 25% may not be optimal for all datasets, as the redundancy of spectral–spatial information can vary significantly across different scenes and sensors. Second, the pruning strategy relies solely on the self-attention weights to determine token importance, which may not always accurately reflect the semantic or physical relevance of tokens. A more nuanced criterion that incorporates domain knowledge about spectral characteristics could further improve the pruning effectiveness.
For future work, we plan to extend the EDTST architecture in several directions. We will investigate adaptive pruning strategies that automatically determine the optimal pruning ratio based on the redundancy degree of the input data. Additionally, we aim to develop more advanced token selection mechanisms, such as incorporating learnable gates or leveraging physical spectral properties to identify and retain the most informative tokens. We also envision extending EDTST to support multi-scale feature extraction and exploring self-supervised pre-training strategies to better leverage unlabeled hyperspectral data. Finally, we will study the interpretability of the model and the relationship between pruned tokens and physical spectral characteristics to provide deeper insights for hyperspectral image analysis.

Author Contributions

Conceptualization, T.Z. and X.H.; methodology, T.Z. and X.H.; software, X.H. and Z.Z.; validation, Z.Z., J.Z., L.Z., Y.T. and Y.P.; formal analysis, T.Z. and X.H.; investigation, T.Z., X.H., Z.Z., J.Z., L.Z. and Y.T.; resources, T.Z., X.H. and Z.Z.; data curation, T.Z., X.H., Z.Z., J.Z., L.Z., Y.T. and Y.P.; writing—original draft preparation, T.Z. and X.H.; writing—review and editing, T.Z. and X.H.; visualization, X.H. and Z.Z.; supervision, T.Z.; project administration, T.Z.; funding acquisition, T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Independent Innovation Science Foundation (24-ZZCX-BC-01).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

To provide a comprehensive analysis of the classification performance and the consistency across different runs, this appendix presents the detailed confusion matrices for each dataset. For each of the four datasets—QUH-Tangdaowan, WHU-Hi-HanChuan, Indian Pines, and Salinas—we display the confusion matrices from all five independent training runs (subfigures a–e), followed by the average confusion matrix (subfigure f) calculated across these runs. The average matrix offers a consolidated view of the model’s overall performance and its stability, highlighting any consistent patterns of misclassification between classes. These visuals complement the aggregated performance metrics reported in the main paper by illustrating the per-class precision and recall in detail.
Figure A1. Confusion matrices of the QUH-Tangdaowan dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Figure A1. Confusion matrices of the QUH-Tangdaowan dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Remotesensing 17 03180 g0a1
Figure A2. Confusion matrices of the WHU-Hi-HanChuan dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Figure A2. Confusion matrices of the WHU-Hi-HanChuan dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Remotesensing 17 03180 g0a2
Figure A3. Confusion matrices of the Indian Pines dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Figure A3. Confusion matrices of the Indian Pines dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Remotesensing 17 03180 g0a3
Figure A4. Confusion matrices of the Salinas dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Figure A4. Confusion matrices of the Salinas dataset. (ae) The confusion matrices from all five independent training runs. (f) The average confusion matrix.
Remotesensing 17 03180 g0a4

References

  1. Teke, M.; Deveci, H.S.; Haliloğlu, O.; Gürbüz, S.Z.; Sakarya, U. A short survey of hyperspectral remote sensing applications in agriculture. In Proceedings of the 2013 6th International Conference on Recent Advances in Space Technologies (RAST), Istanbul, Turkey, 12–14 June 2013; pp. 171–176. [Google Scholar] [CrossRef]
  2. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  3. Delalieux, S.; Somers, B.; Haest, B.; Spanhove, T.; Borre, J.V.; Mücher, C. Heathland conservation status mapping through integration of hyperspectral mixture analysis and decision tree classifiers. Remote Sens. Environ. 2012, 126, 222–231. [Google Scholar] [CrossRef]
  4. Gualtieri, J.; Chettri, S. Support vector machines for classification of hyperspectral data. In Proceedings of the IGARSS 2000, IEEE 2000 International Geoscience and Remote Sensing Symposium, Taking the Pulse of the Planet: The Role of Remote Sensing in Managing the Environment, Proceedings (Cat. No. 00CH37120), Honolulu, HI, USA, 24–28 July 2000; Volume 2, pp. 813–815. [Google Scholar]
  5. Zhang, Y.; Duan, P.; Kang, X.; Mao, J. Edge Guided Structure Extraction for Hyperspectral Image Classification. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 5993–5996. [Google Scholar] [CrossRef]
  6. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  7. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  8. He, X.; Chen, Y.; Lin, Z. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  9. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  10. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  11. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4509416. [Google Scholar] [CrossRef]
  12. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  13. Li, Z.; Xue, Z.; Xu, Q.; Zhang, L.; Zhu, T.; Zhang, M. SPFormer: Self-Pooling Transformer for Few-Shot Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5502019. [Google Scholar] [CrossRef]
  14. Liang, L.; Zhang, Y.; Zhang, S.; Li, J.; Plaza, A.; Kang, X. Fast Hyperspectral Image Classification Combining Transformers and SimAM-Based CNNs. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5522219. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Liang, L.; Mao, J.; Wang, Y.; Jia, L. From Global to Local: A Dual-Branch Structural Feature Extraction Method for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 1778–1791. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Duan, P.; Liang, L.; Kang, X.; Li, J.; Plaza, A. PFS3F: Probabilistic Fusion of Superpixel-wise and Semantic-aware Structural Features for Hyperspectral Image Classification. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 8723–8737. [Google Scholar] [CrossRef]
  17. Zhang, Y.; Liang, L.; Li, J.; Plaza, A.; Kang, X.; Mao, J.; Wang, Y. Structural and Textural-Aware Feature Extraction for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5502305. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Duan, P.; Mao, J.; Kang, X.; Fang, L.; Ghamisi, P. Contour Structural Profiles: An Edge-Aware Feature Extractor for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5545914. [Google Scholar] [CrossRef]
  19. Xue, X.; Zhang, H.; Jing, H.; Tao, L.; Bai, Z.; Li, Y. Bridging Sensor Gaps via Attention-Gated Tuning for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 10075–10094. [Google Scholar] [CrossRef]
  20. Huang, W.; Wu, T.; Zhang, X.; Li, L.; Lv, M.; Jia, Z.; Zhao, X.; Ma, H.; Vivone, G. MCFTNet: Multimodal Cross-Layer Fusion Transformer Network for Hyperspectral and LiDAR Data Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 12803–12818. [Google Scholar] [CrossRef]
  21. Zhu, F.; Shi, C.; Shi, K.; Wang, L. Joint Classification of Hyperspectral and LiDAR Data Using Hierarchical Multimodal Feature Aggregation-Based Multihead Axial Attention Transformer. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5503817. [Google Scholar] [CrossRef]
  22. Wu, X.; Arshad, T.; Peng, B. Spectral Spatial Window Attention Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5519413. [Google Scholar] [CrossRef]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Red Hook, NY, USA, 4–9 September 2017; pp. 6000–6010. [Google Scholar]
  24. He, J.; Zhao, L.; Yang, H.; Zhang, M.; Li, W. HSI-BERT: Hyperspectral Image Classification Using the Bidirectional Encoder Representation From Transformers. IEEE Trans. Geosci. Remote Sens. 2020, 58, 165–178. [Google Scholar] [CrossRef]
  25. He, X.; Chen, Y.; Li, Q. Two-Branch Pure Transformer for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6015005. [Google Scholar] [CrossRef]
  26. Yang, H.; Yu, H.; Zheng, K.; Hu, J.; Tao, T.; Zhang, Q. Hyperspectral Image Classification Based on Interactive Transformer and CNN With Multilevel Feature Fusion Network. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5507905. [Google Scholar] [CrossRef]
  27. Ouyang, E.; Li, B.; Hu, W.; Zhang, G.; Zhao, L.; Wu, J. When Multigranularity Meets Spatial–Spectral Attention: A Hybrid Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4401118. [Google Scholar] [CrossRef]
  28. Qi, W.; Huang, C.; Wang, Y.; Zhang, X.; Sun, W.; Zhang, L. Global–Local 3-D Convolutional Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5510820. [Google Scholar] [CrossRef]
  29. Zhao, F.; Li, S.; Zhang, J.; Liu, H. Convolution Transformer Fusion Splicing Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5501005. [Google Scholar] [CrossRef]
  30. Zhao, E.; Su, Y.; Qu, N.; Wang, Y.; Gao, C.; Zeng, J. Self- and Cross-Attention Enhanced Transformer for Visible and Thermal Infrared Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 13408–13422. [Google Scholar] [CrossRef]
  31. Jia, C.; Zhang, X.; Meng, H.; Xia, S.; Jiao, L. CenterFormer: A Center Spatial–Spectral Attention Transformer Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 5523–5539. [Google Scholar] [CrossRef]
  32. Yu, C.; Zhu, Y.; Wang, Y.; Zhao, E.; Zhang, Q.; Lu, X. Concern With Center-Pixel Labeling: Center-Specific Perception Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5514614. [Google Scholar] [CrossRef]
  33. Hu, X.; Liu, T.; Guo, Z.; Tang, Y.; Peng, Y.; Zhou, T. BinaryViT: Binary Vision Transformer for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 20469–20486. [Google Scholar] [CrossRef]
  34. Wang, Y.; Shu, Z.; Yu, Z. Efficient Attention Transformer Network With Self-Similarity Feature Enhancement for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 11469–11486. [Google Scholar] [CrossRef]
  35. Fu, H.; Sun, G.; Zhang, L.; Zhang, A.; Ren, J.; Jia, X.; Li, F. Three-dimensional singular spectrum analysis for precise land cover classification from UAV-borne hyperspectral benchmark datasets. ISPRS J. Photogramm. Remote Sens. 2023, 203, 115–134. [Google Scholar] [CrossRef]
  36. Zhong, Y.; Hu, X.; Luo, C.; Wang, X.; Zhao, J.; Zhang, L. WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2) benchmark datasets and classifier for precise crop identification based on deep convolutional neural network with CRF. Remote Sens. Environ. 2020, 250, 112012. [Google Scholar] [CrossRef]
  37. Zhu, T.; Liu, Q.; Zhang, L. 3D atrous spatial pyramid pooling based multi-scale feature fusion network for hyperspectral image classification. In Proceedings of the International Conference on Remote Sensing, Mapping, and Geographic Systems (RSMG 2023), SPIE, Kaifeng, China, 7–9 July 2023; Volume 12815, pp. 225–231. [Google Scholar]
  38. Zhu, Y.; Yuan, K.; Zhong, W.; Xu, L. Spatial–Spectral ConvNeXt for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 5453–5463. [Google Scholar] [CrossRef]
  39. Zhou, Y.; Huang, X.; Yang, X.; Peng, J.; Ban, Y. DCTN: Dual-Branch Convolutional Transformer Network With Efficient Interactive Self-Attention for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5508616. [Google Scholar] [CrossRef]
  40. Meng, Z.; Yan, Q.; Zhao, F.; Chen, G.; Hua, W.; Liang, M. Global-Local MultiGranularity Transformer for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 112–131. [Google Scholar] [CrossRef]
  41. Shi, H.; Zhang, Y.; Cao, G.; Yang, D. MHCFormer: Multiscale Hierarchical Conv-Aided Fourierformer for Hyperspectral Image Classification. IEEE Trans. Instrum. Meas. 2024, 73, 5501115. [Google Scholar] [CrossRef]
  42. Varahagiri, S.; Sinha, A.; Dubey, S.R.; Singh, S.K. 3D-Convolution Guided Spectral-Spatial Transformer for Hyperspectral Image Classification. In Proceedings of the 2024 IEEE Conference on Artificial Intelligence (CAI), Singapore, 25–27 June 2024; pp. 8–14. [Google Scholar]
  43. Xu, Y.; Wang, D.; Zhang, L.; Zhang, L. Dual selective fusion transformer network for hyperspectral image classification. Neural Netw. 2025, 187, 107311. [Google Scholar] [CrossRef]
Figure 1. The proposed EDTST architecture for hyperspectral image analysis.
Figure 1. The proposed EDTST architecture for hyperspectral image analysis.
Remotesensing 17 03180 g001
Figure 2. The large-kernel 3D convlution block in the proposed EDTST model.
Figure 2. The large-kernel 3D convlution block in the proposed EDTST model.
Remotesensing 17 03180 g002
Figure 3. Architecture of the 2D convolution block.
Figure 3. Architecture of the 2D convolution block.
Remotesensing 17 03180 g003
Figure 4. The Transformer block in the proposed EDTST model.
Figure 4. The Transformer block in the proposed EDTST model.
Remotesensing 17 03180 g004
Figure 5. Classification maps of the QUH-Tangdaowan dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (proposed).
Figure 5. Classification maps of the QUH-Tangdaowan dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (proposed).
Remotesensing 17 03180 g005
Figure 6. Classification maps of the WHU-Hi-HanChuan dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (Proposed).
Figure 6. Classification maps of the WHU-Hi-HanChuan dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (Proposed).
Remotesensing 17 03180 g006
Figure 7. Classification maps of the Indian Pines dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (Proposed).
Figure 7. Classification maps of the Indian Pines dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (Proposed).
Remotesensing 17 03180 g007
Figure 8. Classification maps of the Salinas dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (proposed).
Figure 8. Classification maps of the Salinas dataset. (a) Ground-truth map. (b) 3A-MFFN. (c) SS-ConvNeXt. (d) DCTN. (e) GLMGT. (f) MHCFormer. (g) 3D-ConvSST. (h) DSFormer. (i) EDTST (proposed).
Remotesensing 17 03180 g008
Table 1. Comparison of CNN and transformer architectures for HSI classification.
Table 1. Comparison of CNN and transformer architectures for HSI classification.
AspectCNN-Based ModelsTransformer-Based Models
Local Feature Extraction✓Strong; inductive bias favors local patterns× Weak; requires large data to learn locality
Global Dependency Modeling× Limited; requires deep stacks or large kernels✓Strong; innate self-attention mechanism
Computational Efficiency✓High; linear in image size× Quadratic in token sequence length
Data Efficiency✓Moderate; benefits from convolutional priors× Low; requires pre-training or large datasets
Table 2. Parameter distribution in large-kernel 3D convolution block ( B   =   64 ,   D   =   40 ,   H   =   11 ,   W   =   11 ), showing that our convolutional kernel design reduces computational cost and parameter count compared to the traditional 3 × 3 × 3 combination.
Table 2. Parameter distribution in large-kernel 3D convolution block ( B   =   64 ,   D   =   40 ,   H   =   11 ,   W   =   11 ), showing that our convolutional kernel design reduces computational cost and parameter count compared to the traditional 3 × 3 × 3 combination.
Layer TypeOutput ShapeKernelParams
Input [ 64 , 1 , 40 , 11 , 11 ] --
Conv3D [ 64 , 1 , 40 , 11 , 11 ] 3 × 3 × 3 7 × 7 × 7 28 344
GroupNorm [ 64 , 1 , 40 , 11 , 11 ] -2
Conv3D [ 64 , 4 , 40 , 11 , 11 ] 3 × 3 × 3 1 × 1 × 1 112 8
GELU [ 64 , 4 , 40 , 11 , 11 ] -0
Conv3D [ 64 , 8 , 40 , 11 , 11 ] 3 × 3 × 3 1 × 1 × 1 872 40
BatchNorm3d [ 64 , 8 , 40 , 11 , 11 ] -16
GELU [ 64 , 8 , 40 , 11 , 11 ] -0
Total Parameters1030→ 410
100%→ 39.8%
FLOPs4,990,040 →1,989,240
100% → 40%
Table 3. Computational characteristics of 2D convolution block.
Table 3. Computational characteristics of 2D convolution block.
OperationFLOPsParameters
Conv 2 D 3 × 3 9 H W C in E 9 C in E + E
BatchNorm2d 2 E H W 2 E
GELU 3 E H W 0
Total 𝒪 ( H W C in E ) 𝒪 ( C in E )
Table 4. Computational complexity comparison ( N = 121 , E = 64 ).
Table 4. Computational complexity comparison ( N = 121 , E = 64 ).
ComponentFLOPs
Standard Self-Attention 2 N 2 E + 4 E 2 N
Proposed Attention 1.75 N 2 E + 4 E 2 N
Total Saving6%
Table 5. Training and testing set distribution for QUH-Tangdaowan, WHU-Hi-HanChuan, Indian Pines, and Salinas datasets.
Table 5. Training and testing set distribution for QUH-Tangdaowan, WHU-Hi-HanChuan, Indian Pines, and Salinas datasets.
QUH-TangdaowanWHU-Hi-HanChuan
NO.ClassTrainTestClassTrainTest
C1Rubber track2525,824Strawberry2544,710
C2Flaggingv2555,528Cowpea2522,728
C3Sandy2534,012Soybean2510,262
C4Asphalt2560,665Sorghum255328
C5Boardwalk251837Water spinach251175
C6Rocky shallows2537,100Watermelon254508
C7Grassland2514,102Greens255878
C8Bulrush2564,062Trees2517,953
C9Gravel road2530,670Grass259444
C10Ligustrum vicaryi251758Red roof2510,491
C11Coniferous pine2521,211Gray roof2516,886
C12Spiraea25724Plastic253654
C13Bare soil251661Bare soil259091
C14Buxus sinica25861Road2518,535
C15Photinia serrulata2513,995Bright object251111
C16Populus25140,879Water2575,376
C17Ulmus pumila L259777
C18Seawater2542,250
 Total450556,916Total400257,530
Indian PinesSalinas
NO.ClassTrainTestClassTrainTest
C1Alfalfa1036Brocoli_green_weeds_1251984
C2Corn-notill101418Brocoli_green_weeds_2253701
C3Corn-mintill10820Fallow251951
C4Corn10227Fallow_rough_plow251369
C5Grass-pasture10473Fallow_smooth252653
C6Grass-trees10720Stubble253934
C7Grass-pasture-mowed1018Celery253554
C8Hay-windrowed10468Grapes_untrained2511,246
C9Oats1010Soil_vinyard_develop256178
C10Soybean-notill10962Corn_senesced_green_weeds253253
C11Soybean-mintill102445Lettuce_romaine_4wk251043
C12Soybean-clean10583Lettuce_romaine_5wk251902
C13Wheat10195Lettuce_romaine_6wk25891
C14Woods101255Lettuce_romaine_7wk251045
C15Buildings-Grass-Trees-Drives10376Vinyard_untrained257243
C16Stone-Steel-Towers1083Vinyard_vertical_trellis251782
 Total16010,089Total40054,429
Table 6. The classification results of the QUH-Tangdaowan dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Table 6. The classification results of the QUH-Tangdaowan dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Class
(Test Samples)
3A-MFFN
[37]
SS-ConvNeXt
[38]
DCTN
[39]
GLMGT
[40]
MHCFormer
[41]
3D-ConvSST
[42]
DSFormer
[43]
EDTST
(Proposed)
C1 (25,824)97.31 ± 0.9099.18 ± 0.3199.54 ± 0.4399.58 ± 0.4099.39 ± 0.3699.49 ± 0.3899.33 ± 0.2398.82 ± 0.46
C2 (55,528)63.81 ± 3.2477.21 ± 5.4479.87 ± 13.3410.38 ± 9.7079.15 ± 5.2465.16 ± 5.8479.32 ± 6.4079.33 ± 7.58
C3 (34,012)75.48 ± 4.2487.07 ± 11.9289.23 ± 9.8447.36 ± 22.2394.31 ± 2.8076.62 ± 17.1292.67 ± 2.5891.33 ± 4.45
C4 (60,665)81.11 ± 7.6786.65 ± 5.1887.47 ± 7.8976.69 ± 10.4095.48 ± 3.2688.41 ± 5.5089.08 ± 2.4691.67 ± 4.23
C5 (1837)95.32 ± 1.0895.95 ± 0.5890.23 ± 9.5846.67 ± 35.4797.69 ± 1.7997.04 ± 2.7996.83 ± 1.2398.16 ± 1.16
C6 (37,100)68.31 ± 4.8764.84 ± 6.3675.62 ± 2.7971.48 ± 9.5677.36 ± 1.4666.92 ± 8.6781.39 ± 3.7080.89 ± 4.52
C7 (14,102)75.55 ± 1.9068.29 ± 5.1082.26 ± 8.1516.16 ± 13.3079.44 ± 4.5066.00 ± 12.1880.92 ± 5.8683.25 ± 3.25
C8 (64,062)94.75 ± 3.0797.96 ± 1.3997.99 ± 1.5198.96 ± 0.3098.09 ± 1.7596.65 ± 1.2498.35 ± 1.0698.65 ± 1.26
C9 (30,670)87.51 ± 4.6782.81 ± 14.4994.29 ± 2.7495.28 ± 4.4995.01 ± 3.6569.93 ± 17.0393.31 ± 4.3394.47 ± 3.77
C10 (1758)88.07 ± 6.2997.69 ± 2.1598.69 ± 1.1256.05 ± 33.2798.03 ± 2.7687.97 ± 7.9599.53 ± 0.4099.92 ± 0.05
C11 (21,211)58.56 ± 10.8155.31 ± 12.6055.50 ± 12.2064.88 ± 32.4787.44 ± 2.5552.07 ± 23.4781.30 ± 3.4585.20 ± 3.29
C12 (724)95.83 ± 1.5099.48 ± 0.4597.93 ± 1.6396.74 ± 2.4399.70 ± 0.4396.49 ± 6.7498.98 ± 1.5299.64 ± 0.35
C13 (1661)97.86 ± 1.02100.00 ± 0.0099.98 ± 0.05100.00 ± 0.00100.00 ± 0.0099.88 ± 0.2499.98 ± 0.05100.00 ± 0.00
C14 (861)91.50 ± 5.6897.58 ± 1.5799.86 ± 0.2396.35 ± 2.2299.91 ± 0.1495.35 ± 3.6399.44 ± 0.70100.00 ± 0.00
C15 (13,995)77.92 ± 1.5287.77 ± 1.2782.25 ± 3.8889.90 ± 3.9887.30 ± 2.2786.56 ± 1.1790.27 ± 1.5789.85 ± 2.52
C16 (140,879)28.51 ± 8.7749.21 ± 11.6646.82 ± 20.9311.92 ± 23.5255.43 ± 6.1445.34 ± 24.7168.21 ± 1.9070.56 ± 4.99
C17 (9777)68.46 ± 5.4385.41 ± 4.0258.77 ± 30.8473.47 ± 20.1587.75 ± 2.9184.14 ± 4.9091.25 ± 1.9991.24 ± 3.10
C18 (42,250)96.69 ± 1.2299.18 ± 0.4099.84 ± 0.2599.94 ± 0.0899.75 ± 0.4399.90 ± 0.1699.55 ± 0.7698.54 ± 1.97
OA (%)67.59 ± 1.3575.92 ± 1.5576.96 ± 4.9356.22 ± 4.9782.21 ± 1.9372.37 ± 5.2384.79 ± 0.7185.75 ± 1.62
AA (%)80.14 ± 0.6085.09 ± 1.4385.34 ± 2.3969.55 ± 3.8390.62 ± 0.5381.89 ± 2.3891.09 ± 0.4191.75 ± 0.58
κ × 100 64.76 ± 1.3273.42 ± 1.5074.66 ± 5.0653.10 ± 4.7180.32 ± 2.0569.64 ± 5.1783.01 ± 0.7784.07 ± 1.76
MCC (%)66.18 ± 1.1374.18 ± 1.1475.77 ± 4.2956.04 ± 3.7581.03 ± 1.8670.75 ± 4.1183.32 ± 0.7384.38 ± 1.67
G-Mean (%)77.09 ± 1.0082.97 ± 1.3781.90 ± 3.7929.55 ± 15.4289.70 ± 0.6977.14 ± 5.3090.56 ± 0.4791.26 ± 0.69
Training Time (s)29.65 ± 0.0246.03 ± 0.06137.77 ± 0.3614.20 ± 0.0318.03 ± 0.0718.89 ± 0.0424.03 ± 0.0212.15 ± 0.02
Testing Time (s)106.87 ± 0.03181.95 ± 0.37471.99 ± 0.4460.71 ± 0.8864.40 ± 0.4277.70 ± 0.40162.58 ± 0.1859.92 ± 0.13
Table 7. The classification results of the WHU-Hi-HanChuan dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Table 7. The classification results of the WHU-Hi-HanChuan dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Class
(Test Samples)
3A-MFFN
[37]
SS-ConvNeXt
[38]
DCTN
[39]
GLMGT
[40]
MHCFormer
[41]
3D-ConvSST
[42]
DSFormer
[43]
EDTST
(Proposed)
C1 (44,710)73.74 ± 5.9591.03 ± 2.6985.59 ± 12.2085.48 ± 5.9385.52 ± 7.9687.07 ± 7.8785.79 ± 3.2488.83 ± 3.58
C2 (22,728)48.09 ± 6.3658.56 ± 12.0077.80 ± 5.2046.31 ± 14.4277.65 ± 2.6779.39 ± 6.3574.42 ± 2.2082.62 ± 5.78
C3 (10,262)65.93 ± 5.2783.17 ± 3.1089.95 ± 5.3688.63 ± 5.0991.43 ± 4.8886.05 ± 8.3173.56 ± 6.8988.73 ± 3.86
C4 (5328)88.19 ± 2.9497.48 ± 0.9598.45 ± 1.3798.96 ± 0.6698.16 ± 0.7495.77 ± 3.5197.58 ± 1.0997.83 ± 1.75
C5 (1175)88.82 ± 2.7699.83 ± 0.1699.76 ± 0.48100.00 ± 0.00100.00 ± 0.0099.93 ± 0.0885.29 ± 4.4799.85 ± 0.20
C6 (4508)39.05 ± 2.2371.85 ± 5.0570.25 ± 9.9923.10 ± 11.0168.46 ± 6.5578.39 ± 5.0760.01 ± 5.5178.17 ± 8.36
C7 (5878)89.00 ± 2.5591.42 ± 2.6686.43 ± 18.2596.61 ± 1.9093.56 ± 1.7090.79 ± 5.7688.49 ± 3.7495.07 ± 0.97
C8 (17,953)53.90 ± 5.8557.38 ± 9.2765.56 ± 2.3027.83 ± 19.9762.99 ± 5.2669.87 ± 6.3060.41 ± 5.4377.81 ± 5.80
C9 (9444)48.24 ± 10.3067.99 ± 8.8773.81 ± 8.554.67 ± 3.5570.09 ± 2.6677.52 ± 13.8971.10 ± 7.1581.88 ± 7.37
C10 (10,491)84.17 ± 7.5794.04 ± 3.4193.25 ± 6.2197.39 ± 1.9896.45 ± 1.5583.85 ± 8.8296.20 ± 0.5496.31 ± 2.56
C11 (16,886)82.18 ± 11.0788.82 ± 5.1083.61 ± 8.1777.22 ± 29.2493.79 ± 1.5481.23 ± 10.3187.79 ± 4.2495.07 ± 1.66
C12 (3654)62.47 ± 7.4168.68 ± 10.5781.66 ± 11.8956.23 ± 24.1788.32 ± 3.2885.99 ± 20.8959.14 ± 1.2794.71 ± 3.82
C13 (9091)58.20 ± 3.9559.50 ± 3.0069.01 ± 5.8129.66 ± 10.5070.41 ± 6.9359.78 ± 8.7052.66 ± 6.7472.68 ± 5.72
C14 (18,535)70.60 ± 4.7860.12 ± 8.9783.39 ± 2.4058.19 ± 15.6078.74 ± 7.1777.33 ± 8.4066.68 ± 3.7781.99 ± 3.01
C15 (1111)76.71 ± 3.4991.54 ± 3.7094.01 ± 3.2996.15 ± 1.0096.31 ± 1.8595.59 ± 2.8990.98 ± 3.0392.58 ± 5.10
C16 (75,376)91.30 ± 6.1797.82 ± 1.7198.49 ± 0.5994.00 ± 3.0897.44 ± 1.7594.74 ± 3.0695.69 ± 2.4897.63 ± 2.37
OA(%)74.15 ± 1.8482.81 ± 2.0786.52 ± 3.8672.79 ± 0.9486.68 ± 1.4685.06 ± 2.3482.39 ± 1.1789.78 ± 0.53
AA(%)70.04 ± 1.4479.95 ± 2.1284.44 ± 3.5867.53 ± 2.5385.58 ± 0.7383.96 ± 2.0777.86 ± 0.8588.86 ± 0.53
κ × 100 70.22 ± 2.0280.02 ± 2.3584.25 ± 4.5368.61 ± 1.0684.52 ± 1.6582.68 ± 2.6379.55 ± 1.3288.10 ± 0.61
MCC (%)70.50 ± 1.9280.16 ± 2.3384.41 ± 4.4269.28 ± 1.0884.63 ± 1.6282.84 ± 2.5879.64 ± 1.2988.16 ± 0.61
G-Mean(%)67.63 ± 1.9478.05 ± 2.5383.39 ± 3.8149.81 ± 4.0684.56 ± 0.8382.76 ± 2.2576.29 ± 0.8688.33 ± 0.64
Training Time (s)42.17 ± 0.0642.36 ± 0.07163.91 ± 0.4615.34 ± 0.1115.81 ± 0.0324.31 ± 0.0221.15 ± 0.0110.71 ± 0.01
Testing Time (s)76.07 ± 0.1492.03 ± 0.11303.26 ± 0.9236.43 ± 0.4530.24 ± 0.0749.97 ± 0.1775.53 ± 0.0627.81 ± 0.07
Table 8. The classification results of the Indian Pines dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Table 8. The classification results of the Indian Pines dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Class
(Test Samples)
3A-MFFN
[37]
SS-ConvNeXt
[38]
DCTN
[39]
GLMGT
[40]
MHCFormer
[41]
3D-ConvSST
[42]
DSFormer
[43]
EDTST
(Proposed)
C1 (36)89.44 ± 7.7499.44 ± 1.11100.00 ± 0.0071.67 ± 39.3099.44 ± 1.11100.00 ± 0.0097.78 ± 2.72100.00 ± 0.00
C2 (1418)30.47 ± 7.3347.46 ± 6.3260.48 ± 4.0324.33 ± 19.9067.45 ± 13.3969.48 ± 4.8450.06 ± 5.5176.78 ± 5.34
C3 (820)33.90 ± 6.6543.71 ± 7.8868.76 ± 6.8011.85 ± 12.2169.15 ± 4.5882.76 ± 2.4167.17 ± 7.4176.37 ± 7.20
C4 (227)40.18 ± 7.8893.57 ± 4.5091.37 ± 5.1992.33 ± 5.4491.72 ± 6.4596.56 ± 2.3989.78 ± 4.4696.12 ± 4.06
C5 (473)71.54 ± 7.3174.46 ± 7.0174.63 ± 5.4445.54 ± 25.7579.49 ± 8.5479.37 ± 8.4278.60 ± 8.6679.41 ± 7.89
C6 (720)79.44 ± 4.7595.22 ± 2.2195.17 ± 0.7797.75 ± 0.5997.58 ± 1.5898.08 ± 1.0698.56 ± 1.2695.94 ± 1.31
C7 (18)98.89 ± 2.22100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
C8 (468)89.36 ± 4.3797.31 ± 3.2499.74 ± 0.4190.47 ± 18.6499.57 ± 0.8599.83 ± 0.3493.33 ± 6.8699.87 ± 0.26
C9 (10)100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0088.00 ± 24.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
C10 (962)56.53 ± 10.7468.92 ± 11.8675.82 ± 8.9046.30 ± 34.8182.56 ± 5.2482.06 ± 4.3175.70 ± 6.0382.91 ± 4.80
C11 (2445)50.82 ± 9.5063.98 ± 7.2966.04 ± 9.4464.33 ± 14.1365.73 ± 7.1867.43 ± 7.3357.59 ± 7.4669.31 ± 4.40
C12 (583)38.39 ± 8.9057.50 ± 4.2758.83 ± 5.9443.33 ± 16.6061.78 ± 4.8867.96 ± 3.8748.40 ± 8.3373.55 ± 3.64
C13 (195)97.74 ± 1.82100.00 ± 0.0099.90 ± 0.21100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.59 ± 0.82100.00 ± 0.00
C14 (1255)72.24 ± 6.0893.34 ± 5.4493.40 ± 1.6195.41 ± 3.1389.85 ± 4.1394.17 ± 3.1491.54 ± 4.0295.54 ± 3.26
C15 (376)50.69 ± 2.8687.07 ± 4.5992.98 ± 1.0080.27 ± 6.9287.87 ± 8.4294.41 ± 3.0989.10 ± 5.2193.03 ± 4.01
C16 (83)97.83 ± 2.6899.76 ± 0.48100.00 ± 0.00100.00 ± 0.0098.80 ± 1.08100.00 ± 0.0099.28 ± 1.45100.00 ± 0.00
OA (%)55.20 ± 3.6170.77 ± 3.2076.17 ± 2.5260.41 ± 4.3877.67 ± 2.8580.73 ± 2.9971.61 ± 2.7782.06 ± 0.39
AA (%)68.59 ± 1.3582.61 ± 1.5686.07 ± 0.7571.97 ± 2.9486.94 ± 1.3889.51 ± 1.5583.53 ± 1.4789.93 ± 0.41
κ × 100 49.65 ± 3.7567.08 ± 3.4873.21 ± 2.7055.33 ± 4.5774.87 ± 3.1278.33 ± 3.2968.18 ± 2.9579.80 ± 0.42
MCC (%)50.01 ± 3.5367.40 ± 3.3773.58 ± 2.4156.60 ± 4.3775.28 ± 2.8778.64 ± 3.1268.64 ± 2.6380.07 ± 0.36
G-Mean (%)62.99 ± 2.5079.59 ± 2.2184.41 ± 0.9740.55 ± 16.8585.52 ± 1.8388.51 ± 1.8781.04 ± 1.9289.11 ± 0.33
Training Time (s)11.39 ± 0.1816.47 ± 0.1253.40 ± 0.154.79 ± 0.026.67 ± 0.0310.94 ± 0.038.81 ± 0.024.24 ± 0.07
Testing Time (s)2.04 ± 0.033.22 ± 0.039.19 ± 0.041.05 ± 0.011.17 ± 0.012.71 ± 0.012.96 ± 0.001.05 ± 0.01
Table 9. The classification results of the Salinas dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Table 9. The classification results of the Salinas dataset. The number in parentheses after each class label is the number of test samples. The reported metrics are overall accuracy (OA, %), average accuracy (AA, %), Cohen’s Kappa coefficient ( κ × 100 ), Matthew’s correlation coefficient (MCC, %), G-Mean (%), training time (seconds), and testing time (seconds). The best results are highlighted in bold.
Class
(Test Samples)
3A-MFFN
[37]
SS-ConvNeXt
[38]
DCTN
[39]
GLMGT
[40]
MHCFormer
[41]
3D-ConvSST
[42]
DSFormer
[43]
EDTST
(Proposed)
C1 (1984)98.40 ± 1.01100.00 ± 0.0099.39 ± 0.9899.99 ± 0.0299.99 ± 0.0299.96 ± 0.0899.79 ± 0.37100.00 ± 0.00
C2 (3701)99.49 ± 0.41100.00 ± 0.00100.00 ± 0.0098.04 ± 2.2299.97 ± 0.0696.40 ± 7.14100.00 ± 0.00100.00 ± 0.00
C3 (1951)97.58 ± 1.3699.84 ± 0.3399.77 ± 0.3490.28 ± 14.42100.00 ± 0.0092.33 ± 10.1499.95 ± 0.06100.00 ± 0.00
C4 (1369)99.24 ± 0.5699.94 ± 0.1299.27 ± 1.2899.52 ± 0.4699.55 ± 0.5699.63 ± 0.3499.33 ± 0.3899.49 ± 0.55
C5 (2653)98.30 ± 0.3998.90 ± 0.5196.88 ± 3.3096.87 ± 1.3999.29 ± 0.4899.13 ± 0.6999.19 ± 0.3498.90 ± 0.95
C6 (3934)99.92 ± 0.14100.00 ± 0.00100.00 ± 0.0099.98 ± 0.0499.97 ± 0.06100.00 ± 0.00100.00 ± 0.0099.95 ± 0.09
C7 (3554)98.93 ± 0.4799.98 ± 0.0399.88 ± 0.2299.97 ± 0.0499.98 ± 0.0399.64 ± 0.3699.99 ± 0.0199.97 ± 0.03
C8 (11,246)71.78 ± 2.8177.71 ± 6.0475.82 ± 5.0663.10 ± 38.2886.73 ± 5.2679.66 ± 13.1485.27 ± 3.0991.05 ± 3.39
C9 (6178)97.08 ± 1.37100.00 ± 0.0197.12 ± 5.7699.67 ± 0.65100.00 ± 0.0099.97 ± 0.0499.99 ± 0.01100.00 ± 0.00
C10 (3253)91.42 ± 1.8395.96 ± 0.5794.85 ± 3.2094.33 ± 1.1597.72 ± 0.9196.71 ± 1.4498.40 ± 0.8298.83 ± 0.73
C11 (1043)95.59 ± 2.5999.64 ± 0.1599.85 ± 0.1796.16 ± 6.6499.96 ± 0.0899.60 ± 0.4699.90 ± 0.11100.00 ± 0.00
C12 (1902)99.86 ± 0.2299.99 ± 0.0299.85 ± 0.1896.48 ± 1.77100.00 ± 0.0099.82 ± 0.0999.27 ± 0.7299.79 ± 0.33
C13 (891)99.96 ± 0.09100.00 ± 0.0099.73 ± 0.34100.00 ± 0.0099.87 ± 0.1899.78 ± 0.3599.66 ± 0.5199.71 ± 0.38
C14 (1045)97.76 ± 0.6099.48 ± 0.3099.73 ± 0.1699.83 ± 0.1499.62 ± 0.2699.22 ± 0.8799.46 ± 0.3099.79 ± 0.18
C15 (7243)70.83 ± 9.6288.84 ± 9.3775.06 ± 4.7446.33 ± 43.5689.12 ± 6.4083.95 ± 12.5883.39 ± 6.1595.60 ± 2.53
C16 (1782)93.11 ± 3.1798.93 ± 0.5199.84 ± 0.2498.99 ± 0.8399.01 ± 0.5099.62 ± 0.6899.55 ± 0.2699.74 ± 0.22
OA (%)88.58 ± 1.5093.47 ± 0.3590.70 ± 1.0783.77 ± 3.4595.52 ± 0.5592.73 ± 2.1194.45 ± 0.2897.36 ± 0.82
AA (%)94.33 ± 0.6897.45 ± 0.2696.07 ± 0.8492.47 ± 1.2298.17 ± 0.1496.59 ± 0.9997.70 ± 0.2398.93 ± 0.27
κ × 100 87.31 ± 1.6792.75 ± 0.3989.66 ± 1.1981.96 ± 3.6595.02 ± 0.6191.92 ± 2.3293.83 ± 0.3197.07 ± 0.91
MCC (%)87.36 ± 1.7192.90 ± 0.3989.72 ± 1.1983.71 ± 2.6795.07 ± 0.5792.13 ± 2.0793.85 ± 0.3397.09 ± 0.89
G-Mean (%)93.76 ± 0.9197.20 ± 0.3095.65 ± 0.9082.86 ± 8.4098.06 ± 0.1696.17 ± 1.2897.53 ± 0.2798.89 ± 0.29
Training Time (s)31.19 ± 0.0440.86 ± 0.07132.83 ± 0.2513.10 ± 0.0415.68 ± 0.0418.80 ± 0.0821.27 ± 0.0610.79 ± 0.01
Testing Time (s)11.70 ± 0.0217.81 ± 0.3750.20 ± 0.166.11 ± 0.016.21 ± 0.038.18 ± 0.0315.74 ± 0.025.80 ± 0.01
Table 10. Analysis of the impact of large-kernel 3D convolution block and Transformer block in the proposed network architecture. The best results are highlighted in bold.
Table 10. Analysis of the impact of large-kernel 3D convolution block and Transformer block in the proposed network architecture. The best results are highlighted in bold.
Dataset3D ConvTransformerOA(%)AA(%) κ × 100 MCC (%)G-Mean (%)
QUH-Tangdaowan 85.28 ± 0.4691.39 ± 0.7383.56 ± 0.5283.89 ± 0.5490.83 ± 0.73
84.46 ± 1.8091.52 ± 0.6482.71 ± 1.9383.15 ± 1.7990.88 ± 0.80
85.75 ± 1.6291.75 ± 0.5884.07 ± 1.7684.38 ± 1.6791.26 ± 0.69
WHU-Hi-HanChuan 88.32 ± 1.4386.98 ± 0.6886.41 ± 1.6386.48 ± 1.6186.32 ± 0.78
89.22 ± 0.9988.11 ± 0.5487.48 ± 1.1287.55 ± 1.0987.44 ± 0.62
89.78 ± 0.5388.86 ± 0.5388.10 ± 0.6188.16 ± 0.6188.33 ± 0.64
Indian Pines 79.56 ± 2.4089.02 ± 1.1577.03 ± 2.6377.38 ± 2.4687.97 ± 1.41
79.58 ± 2.3588.86 ± 1.2277.06 ± 2.5777.43 ± 2.3787.65 ± 1.57
82.06 ± 0.3989.93 ± 0.4179.80 ± 0.4280.07 ± 0.3689.11 ± 0.33
Salinas 95.76 ± 0.3098.26 ± 0.1595.28 ± 0.3395.33 ± 0.3198.16 ± 0.18
97.22 ± 0.5898.79 ± 0.2396.90 ± 0.6596.92 ± 0.6498.75 ± 0.25
97.36 ± 0.8298.93 ± 0.2797.07 ± 0.9197.09 ± 0.8998.89 ± 0.29
Table 11. The classification results of different patch sizes over four datasets. The best results are highlighted in bold.
Table 11. The classification results of different patch sizes over four datasets. The best results are highlighted in bold.
DatasetPatch SizeOA(%)AA(%) κ × 100 MCC(%)G-Mean (%)
QUH-Tangdaowan576.46 ± 2.4386.51 ± 0.8874.03 ± 2.5574.80 ± 2.3285.14 ± 1.19
781.12 ± 1.4989.24 ± 0.4879.03 ± 1.6079.57 ± 1.4788.34 ± 0.65
983.93 ± 2.2390.80 ± 0.6582.09 ± 2.3982.47 ± 2.2190.13 ± 0.76
1185.75 ± 1.6291.75 ± 0.5884.07 ± 1.7684.38 ± 1.6791.26 ± 0.69
WHU-Hi-HanChuan580.50 ± 0.8777.43 ± 1.7177.43 ± 1.0077.62 ± 0.9875.56 ± 2.30
786.24 ± 1.3584.26 ± 0.9284.01 ± 1.5484.10 ± 1.5183.37 ± 1.06
988.49 ± 1.3486.86 ± 1.0386.60 ± 1.5386.67 ± 1.5086.18 ± 1.24
1189.78 ± 0.5388.86 ± 0.5388.10 ± 0.6188.16 ± 0.6188.33 ± 0.64
Indian Pines570.98 ± 2.9782.13 ± 1.8467.35 ± 3.2367.70 ± 2.9779.89 ± 2.42
776.84 ± 2.4786.60 ± 1.5373.88 ± 2.8174.11 ± 2.7985.21 ± 1.80
979.72 ± 3.1688.68 ± 1.6177.18 ± 3.5277.48 ± 3.3987.61 ± 1.99
1182.06 ± 0.3989.93 ± 0.4179.80 ± 0.4280.07 ± 0.3689.11 ± 0.33
Salinas593.99 ± 0.3097.42 ± 0.2293.31 ± 0.3393.37 ± 0.3297.24 ± 0.27
795.33 ± 0.4998.14 ± 0.1994.81 ± 0.5494.87 ± 0.5298.04 ± 0.21
996.42 ± 0.5098.59 ± 0.1796.02 ± 0.5696.07 ± 0.5498.52 ± 0.18
1197.36 ± 0.8298.93 ± 0.2797.07 ± 0.9197.09 ± 0.8998.89 ± 0.29
Table 12. The classification performance with different PCA dimensions. The best results are highlighted in bold.
Table 12. The classification performance with different PCA dimensions. The best results are highlighted in bold.
DatasetPCA DimsOA (%)AA (%) κ × 100 MCC (%)G-Mean (%)
QUH-Tangdaowan2085.37 ± 1.9491.43 ± 0.6683.66 ± 2.1083.97 ± 1.9890.88 ± 0.79
3083.85 ± 3.7591.28 ± 0.8382.06 ± 3.9682.56 ± 3.5490.47 ± 1.30
4085.75 ± 1.6291.75 ± 0.5884.07 ± 1.7684.38 ± 1.6791.26 ± 0.69
None71.95 ± 4.7083.21 ± 1.2269.25 ± 4.7770.43 ± 4.1180.49 ± 2.21
WHU-Hi-HanChuan2087.87 ± 1.5286.86 ± 1.0185.90 ± 1.7385.98 ± 1.7086.20 ± 1.21
3089.39 ± 1.0487.80 ± 0.7187.64 ± 1.2087.70 ± 1.1887.06 ± 0.76
4089.78 ± 0.5388.86 ± 0.5388.10 ± 0.6188.16 ± 0.6188.33 ± 0.64
None84.08 ± 4.6082.62 ± 2.8681.53 ± 5.1781.70 ± 5.0781.37 ± 3.31
Indian Pines2076.62 ± 2.3886.98 ± 0.7173.79 ± 2.5674.23 ± 2.3785.55 ± 0.94
3079.31 ± 1.5788.85 ± 0.6876.76 ± 1.7177.04 ± 1.6487.78 ± 0.89
4082.06 ± 0.3989.93 ± 0.4179.80 ± 0.4280.07 ± 0.3689.11 ± 0.33
None77.67 ± 2.0687.64 ± 0.9374.93 ± 2.2275.38 ± 2.0086.26 ± 1.12
Salinas2097.28 ± 0.9598.83 ± 0.3196.97 ± 1.0597.01 ± 1.0198.79 ± 0.33
3097.02 ± 1.2198.54 ± 0.5996.68 ± 1.3496.74 ± 1.2898.49 ± 0.63
4097.36 ± 0.8298.93 ± 0.2797.07 ± 0.9197.09 ± 0.8998.89 ± 0.29
None93.06 ± 1.7696.89 ± 0.6592.30 ± 1.9392.47 ± 1.7096.61 ± 0.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, X.; Zhang, Z.; Zhai, J.; Zhang, L.; Tang, Y.; Peng, Y.; Zhou, T. EDTST: Efficient Dynamic Token Selection Transformer for Hyperspectral Image Classification. Remote Sens. 2025, 17, 3180. https://doi.org/10.3390/rs17183180

AMA Style

Hu X, Zhang Z, Zhai J, Zhang L, Tang Y, Peng Y, Zhou T. EDTST: Efficient Dynamic Token Selection Transformer for Hyperspectral Image Classification. Remote Sensing. 2025; 17(18):3180. https://doi.org/10.3390/rs17183180

Chicago/Turabian Style

Hu, Xiang, Zhiwen Zhang, Jianghe Zhai, Longlong Zhang, Yuxiang Tang, Yuanxi Peng, and Tong Zhou. 2025. "EDTST: Efficient Dynamic Token Selection Transformer for Hyperspectral Image Classification" Remote Sensing 17, no. 18: 3180. https://doi.org/10.3390/rs17183180

APA Style

Hu, X., Zhang, Z., Zhai, J., Zhang, L., Tang, Y., Peng, Y., & Zhou, T. (2025). EDTST: Efficient Dynamic Token Selection Transformer for Hyperspectral Image Classification. Remote Sensing, 17(18), 3180. https://doi.org/10.3390/rs17183180

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop