Next Article in Journal
A Robust Tracking Method for Aerial Extended Targets with Space-Based Wideband Radar
Previous Article in Journal
CAU2DNet: A Dual-Branch Deep Learning Network and a Dataset for Slum Recognition with Multi-Source Remote Sensing Data
Previous Article in Special Issue
HG-Mamba: A Hybrid Geometry-Aware Bidirectional Mamba Network for Hyperspectral Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HyperTransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Hyperspectral Image Classification

1
School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
2
School of Information Engineering, East China Jiaotong University, Nanchang 330013, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(14), 2361; https://doi.org/10.3390/rs17142361
Submission received: 3 June 2025 / Revised: 30 June 2025 / Accepted: 3 July 2025 / Published: 9 July 2025
(This article belongs to the Special Issue AI-Driven Hyperspectral Remote Sensing of Atmosphere and Land)

Abstract

Recent advances in hyperspectral image (HSI) classification have demonstrated the effectiveness of hybrid architectures that integrate convolutional neural networks (CNNs) and Transformers, leveraging CNNs for local feature extraction and Transformers for global dependency modeling. However, existing fusion approaches face three critical challenges: (1) insufficient synergy between spectral and spatial feature learning due to rigid coupling mechanisms; (2) high computational complexity resulting from redundant attention calculations; and (3) limited adaptability to spectral redundancy and noise in small-sample scenarios. To address these limitations, we propose HyperTransXNet, a novel CNN-Transformer hybrid architecture that incorporates adaptive spectral-spatial fusion. Specifically, the proposed HyperTransXNet comprises three key modules: (1) a Hybrid Spatial-Spectral Module (HSSM) that captures the refined local spectral-spatial features and models global spectral correlations by combining depth-wise dynamic convolution with frequency-domain attention; (2) a Mixture-of-Experts Routing (MoE-R) module that adaptively fuses multi-scale features by dynamically selecting optimal experts via Top-K sparse weights; and (3) a Spatial-Spectral Tokens Enhancer (SSTE) module that ensures causality-preserving interactions between spectral bands and spatial contexts. Extensive experiments on the Indian Pines, Houston 2013, and WHU-Hi-LongKou datasets demonstrate the superiority of HyperTransXNet.

1. Introduction

Hyperspectral images (HSIs), characterized by their contiguous spectral bands that span the electromagnetic spectrum, offer a distinctive three-dimensional spatial-spectral representation of surface materials. This dense spectral sampling enables the precise discrimination of land cover types, rendering HSIs indispensable in applications such as environmental monitoring [1], precision agriculture, and urban planning [2]. Nevertheless, the high dimensionality of HSIs introduces inherent challenges, including spectral redundancy, spatial heterogeneity, and the curse of dimensionality, all of which complicate feature extraction and classification. Traditional machine learning methods, such as support vector machines (SVMs) and random forests, often struggle with these complexities due to their dependence on handcrafted features and their limited capacity to model nonlinear spectral-spatial relationships.
Recent advances in deep learning have transformed hyperspectral imaging (HSI) classification by automating feature extraction [3,4]. Early studies primarily employed two-dimensional convolutional neural networks (2D-CNNs) to capture spatial patterns by sliding fixed kernels across hyperspectral image (HSI) patches. For example, Paoletti et al. [5] proposed a 2D-CNN with shortcut connections. Although such architectures are proficient at extracting local textures and edges, they inherently disregard the rich spectral correlations present across hundreds of contiguous bands. To address this limitation, three-dimensional convolutional neural networks (3D-CNNs) emerged as a transformative approach by explicitly modeling spectral-spatial cubes via volumetric convolutions. A notable instance is HybridSN [6], which hierarchically integrates 3D convolutions for joint spectral-spatial encoding with 2D convolutions for spatial refinement. This hybrid architecture significantly enhances classification accuracy by leveraging inter-band dependencies. However, 3D-CNNs are challenged by prohibitive computational costs and parameter redundancy, especially in scenarios with limited labeled data.
Although 2D- and 3D-CNNs have laid a robust foundation for automated feature learning, they are subject to two main limitations. First, there is a spatial-spectral decoupling issue: 2D-CNNs prioritize spatial details at the expense of spectral coherence, while 3D-CNNs encounter challenges in balancing computational efficiency with spectral fidelity. Second, both architectures inadequately capture global dependencies; their localized receptive fields hinder the modeling of long-range contextual relationships, which can lead to misclassifications in heterogeneous or fragmented regions. These constraints underscore the necessity for novel architectures that seamlessly integrate spectral-spatial dynamics while maintaining computational efficiency—a gap that subsequent advancements in attention mechanisms and hybrid frameworks seek to address.
The emergence of Vision Transformers (ViTs) has introduced a paradigm shift in hyperspectral image (HSI) classification by leveraging self-attention mechanisms to capture global spectral-spatial dependencies. In contrast to convolutional neural networks (CNNs), which are limited to localized convolutions, ViTs represent HSIs as sequences of tokens, thereby enabling the modeling of long-range interactions. For example, the Spectral-Spatial Feature Tokenization Transformer (SSFTT) [7] pioneers a hierarchical framework in which preliminary spectral-spatial features are extracted using shallow CNN layers, tokenized, and subsequently processed by Transformer blocks to capture global contextual relationships. This hybrid design harmonizes the preservation of local details with the attainment of global coherence, achieving state-of-the-art accuracy on datasets such as Indian Pines. Despite these advancements, Transformer-based methods are hindered by two primary limitations: (1) quadratic complexity, whereby self-attention scales quadratically with sequence length, making it computationally prohibitive for high-resolution HSIs; and (2) spatial discontinuity, as fixed-size patch tokenization disrupts the inherent spatial continuity of HSIs, thereby generating artifacts in fine-grained classification maps, particularly for small or irregular targets.
To mitigate these issues, Mamba-based architectures have emerged as lightweight alternatives. By combining the computational efficiency of convolutions with selective state transitions, Mamba models achieve linear computational scaling while preserving global sequence modeling capabilities. For instance, SpectralMamba [8] introduces a dynamic spectral masking mechanism and a bidirectional scanning strategy that adaptively groups redundant bands and propagates contextual dependencies across the spectral dimension. Similarly, DualMamba [9] presents a dual-path architecture incorporating parallel spatial and spectral SSM blocks, which simultaneously model local spatial details and global spectral trends. However, Mamba-based methods are still constrained by their implicit feature fusion mechanisms; for example, SpectralMamba relies on post-hoc concatenation of spectral and spatial features and lacks dynamic interaction modules capable of suppressing noise or enhancing discriminative bands.
Despite recent advancements, several critical challenges remain unresolved.
  • Current fusion methodologies exhibit insufficient spectral-spatial synergy; they tend to integrate local and global features in a rigid manner, thereby failing to adaptively balance their respective contributions.
  • Computational inefficiencies persist, as redundant attention calculations and suboptimal parameterizations impede real-time deployment.
  • Conventional architectures demonstrate a pronounced sensitivity to spectral noise, struggling to suppress redundant bands and enhance discriminative features, particularly under conditions of significant noise or limited sample sizes.
These limitations present critical challenges for hyperspectral image classification, particularly in effectively leveraging high-dimensional spectral features and modeling spatial-spectral relationships. To address these challenges, we propose HyperTransXNet, a novel CNN-Transformer hybrid architecture that integrates spectral and spatial dynamics through three meticulously designed modules. First, the Hybrid Spatial-Spectral Module (HSSM) combines depth-wise dynamic convolutions with frequency-domain attention to simultaneously refine local details and global spectral correlations. Second, the Mixture-of-Experts Routing (MoE-R) module utilizes a dynamic gating mechanism to select optimal multi-scale features via Top-K sparse activation. Third, the Spatial-Spectral Tokens Enhancer (SSTE) module bolsters robustness by preserving causality-aware interactions among tokens, aligning spectral and spatial features with 1 × 1 convolutions, and further refining them through MoE-R, thereby effectively suppressing noise while maintaining structural coherence. The main contributions of this paper are as follows:
  • We introduce HyperTransXNet, a novel hybrid framework that integrates CNN-driven local feature extraction, Transformer-based global dependency modeling, and Mamba-inspired efficiency through dynamic spectral-spatial fusion, thereby addressing the limitations of isolated architectural paradigms.
  • We propose a Hybrid Spatial-Spectral Module (HSSM) that integrates depth-wise dynamic convolutions with frequency-domain attention. This design facilitates the simultaneous enhancement of local textures and global spectral correlations without incurring additional computational overhead.
  • We employ the Top-K sparse Mixture-of-Experts routing mechanism, marking its inaugural application in hyperspectral classification, which achieves parameter-efficient multi-scale feature aggregation via context-aware expert selection.
  • Extensive experiments conducted on three public HSI datasets demonstrate that the proposed HyperTransXNet outperforms state-of-the-art methods based on CNNs and Transformers.
The remainder of this paper is organized as follows. Section 2 reviews the literature on image classification methods based on convolutional neural networks (CNNs), Transformer architectures, and studies on Mixture-of-Experts (MoE) routing methods. Section 4 details the proposed approach and its constituent components. Section 5 introduces several benchmark high-resolution hyperspectral image (HSI) datasets, describes the experimental settings, and presents the results along with a thorough analysis. Finally, Section 6 concludes the paper.

2. Related Works

Hyperspectral image (HSI) classification methodologies have advanced through three distinct research paradigms: convolutional neural networks (CNNs)-based, Transformers-based, and Mamba-based approaches. Each paradigm targets specific facets of spectral-spatial feature learning while simultaneously addressing the intrinsic limitations associated with the individual methodologies.

2.1. CNN-Based Hyperspectral Image Classification

With the rise of deep learning, CNN-based hyperspectral image (HSI) classification methods have also received extensive research attention [3,10,11,12,13,14,15,16,17,18,19,20]. For example, Paoletti et al. [5] proposed a 2D-CNN model for HSI classification, introducing shortcut connections to allow low-level features to directly pass to subsequent layers, thereby enhancing feature expression capabilities. However, 2D-CNNs inherently neglect spectral correlations across hundreds of contiguous bands, limiting their discriminative power for materials with subtle spectral variations (e.g., vegetation subtypes).
To address this limitation, 3D convolutional neural networks (3D-CNNs) have been developed to explicitly model spectral-spatial cubes using volumetric convolutions. For example, Roy et al. [6] introduced a hybrid 3D-2D CNN model (HybridSN) that hierarchically combines 3D convolutions for joint spectral-spatial encoding with 2D convolutions for spatial refinement, achieving a 12% improvement in classification accuracy over conventional 2D-CNNs on the Indian Pines dataset. Similarly, Mei et al. [21] proposed an approach that integrates attention mechanisms with bidirectional long short-term memory (BiLSTM) networks within a CNN framework. Nonetheless, despite the superior local feature extraction capabilities of CNNs, they exhibit notable limitations in representing global features and capturing long-range dependencies in HSI data. Moreover, 3D-CNNs confront two critical challenges: (1) fixed receptive fields limit adaptability to varying spectral curve complexities, and (2) quadratic growth in the number of parameters increases the risk of overfitting, particularly in small-sample scenarios.

2.2. Transformers-Based Hyperspectral Image Classification

To address the limitations of traditional CNNs, some researchers have introduced Transformer to solve the problem of HSI classification [22,23,24,25,26,27,28,29]. For example, Vision Transformers (ViTs) [30] have redefined hyperspectral image (HSI) classification by leveraging self-attention mechanisms to model global spectral-spatial dependencies, thereby overcoming the localized receptive field limitations inherent in convolutional neural networks (CNNs). In contrast to traditional CNNs, ViTs conceptualize HSIs as sequences of tokens, which facilitates interactions between distant pixels and spectral bands. A seminal example of this approach is the Spectral-Spatial Feature Tokenization Transformer (SSFTT) [7], which initially extracts shallow spatial-spectral features using hierarchical 3D-CNN layers and subsequently tokenizes these features into semantic vectors for processing by Transformer blocks. This hybrid architecture achieves an overall accuracy (OA) by effectively integrating local texture details with global spectral trends. Another innovative architecture, MorphFormer [31], integrates morphological convolution operations with spectral-spatial attention mechanisms to enhance structural feature learning. By applying learnable morphological filters (e.g., erosion and dilation) to HSI tokens, MorphFormer captures geometric primitives such as edges and shapes, which are critical for identifying irregular targets like roads and parking lots.
Recent advancements further explore specialized attention mechanisms for HSI. For example, HiT [32] embeds 3D convolutions within Transformer blocks to propagate local spectral context. Despite their success, Transformer-based methods face two fundamental challenges: (1) Quadratic Computational Complexity: self-attention scales quadratically with the number of tokens, rendering them inefficient for high-resolution HSIs. (2) Spatial discontinuity: fixed-size patch tokenization disrupts the inherent spatial continuity of HSIs, leading to misclassification at patch boundaries. To mitigate these limitations, recent works such as Swin-HSI [33] employ shifted window attention to reduce computational costs. However, these adaptations often sacrifice global context modeling for efficiency, underscoring the need for architectures that balance computational tractability with comprehensive spectral-spatial reasoning—a goal central to our proposed HyperTransXNet.

2.3. Mamba-Based Hyperspectral Image Classification

Recent advancements in state space models (SSMs) have driven the development of Mamba-based architectures, which integrate the computational efficiency of convolutional operations with the extensive sequence modeling capabilities of Transformers. By employing selective state transitions and linear-time scanning mechanisms, Mamba models offer enhanced scalability for high-dimensional hyperspectral data while ensuring minimal computational overhead [8,34,35].
For example, SpectralMamba [8] represents a pioneering contribution by introducing a dynamic spectral masking mechanism that adaptively clusters redundant spectral bands. Additionally, it employs bidirectional spectral scanning, processing spectral sequences in both forward and backward directions, to effectively model long-range dependencies across bands. Building on this concept, DualMamba [9] proposes a dual-path architecture featuring parallel spatial and spectral SSM blocks. In this framework, the spatial branch applies a 2D-SSM to capture local textures, whereas the spectral branch utilizes a 1D-SSM to model inter-band correlations, resulting in an overall accuracy of 98.74% on the WHU-Hi-LongKou dataset through disentangled yet complementary feature learning. Moreover, emerging variants such as HyperMamba [36] further enhance the applicability of the Mamba paradigm by integrating 3D spectral-spatial SSM blocks; HyperMamba processes hyperspectral image (HSI) cubes as continuous three-dimensional sequences and dynamically adjusts state transitions based on spectral gradients.
Despite significant advancements in hyperspectral image (HSI) classification, three fundamental limitations persist across current paradigms:
  • Ineffective joint spectral-spatial representation learning: CNN-based methods are constrained by fixed receptive fields that fail to adapt to the varying complexities of spectral curves. In contrast, Mamba architectures suffer from fragmented local context modeling due to their unidirectional scanning. Although Transformer-based approaches effectively model global dependencies, they incur quadratic computational costs, thereby limiting their applicability to large-scale HSIs.
  • Suboptimal feature fusion strategies: Contemporary methods predominantly rely on static fusion techniques to integrate spectral and spatial features. These heuristic approaches lack the adaptability needed to suppress noisy bands or to emphasize discriminative spectral regions, resulting in performance degradation in complex scenes, such as those encountered in urban-rural transitions.
In this paper, we propose HyperTransXNet, a unified architecture that redefines spectral-spatial fusion through three key modules: (1) The Hybrid Spatial-Spectral Module (HSSM) integrates depth-wise dynamic convolutions, with kernel parameters adaptively generated from the global context and frequency domain attention to jointly refine local textures and global spectral trends. This approach effectively eliminates the spatial-spectral decoupling commonly observed in conventional CNNs and Mamba models. (2) Mixture-of-Experts Routing (MoE-R) utilizes Top-K sparse gating to dynamically select optimal experts for feature aggregation. (3) By integrating lightweight dynamic operations with frequency-domain regularization, HyperTransXNet significantly reduces parameter redundancy.

2.4. Different from TransXNet

There are three main differences between TransXNet [37] and our proposed HyperTransXNet. The first difference lies in the feature extraction method. TransXNet utilizes deep convolution and global attention modules to extract local features and global representations, respectively. In contrast, HyperTransXNet employs two different modules: the Spatial-Spectral Local Block (SSLB) and the Spatial-Spectral Global Block (SSGB). SSLB achieves local feature extraction through adaptively generated convolution kernels, while SSGB combines frequency domain attention and multi-head attention mechanisms to capture global representations. The second difference is in the application of the methods. TransXNet is primarily applied in general computer vision fields, whereas our proposed HyperTransXNet focuses on hyperspectral image classification. The third difference is in the feature fusion mechanism. TransXNet mainly uses an improved multi-scale feedforward network for feature fusion, while HyperTransXNet introduces a Spatial-Spectral Tag Enhancer (SSTE) to merge local and global features. In SSTE, we introduce a Mixture-of-Experts routing mechanism (MoE-R) that selectively activates suitable experts through a Top-K sparse gating strategy and constructs a selection matrix, ultimately merging the outputs of both to achieve feature aggregation.

3. MoE

The Mixture-of-Experts (MoE) model is an advanced neural network architecture designed to enhance overall performance by integrating predictions from multiple expert sub-models. Due to its sophistication, this architecture has been widely applied in various fields, including natural language processing [38,39,40,41,42]. The fundamental concept underlying the MoE model involves partitioning the input data among specialized sub-models, each proficient in handling specific tasks and subsequently aggregating their outputs to generate a final prediction. This allocation mechanism is typically dynamic, adjusting based on the characteristics of the input data to ensure that each expert focuses on the data type or task for which it is best suited, thereby achieving efficient and accurate predictions. As shown in Figure 1.The MoE model consists of two main components: the Gating Network and the experts. The Gating Network dynamically determines which expert sub-model or models should process a given input by analyzing its features and computing the corresponding weights or importance for each expert, subsequently allocating the input based on these weights. The experts comprise multiple independent neural networks, each dedicated to processing a specific subset of the input data or a particular task. Ultimately, the MoE model consolidates the prediction results from multiple experts, leveraging the unique strengths of each to improve overall performance. The formulation of the MoE model is expressed as follows:   
g i ( x ) = softmax W g · x E i ( x ) = f expert i ( x ; θ i ) output = i = 1 N g i ( x ) · E i ( x )
Here, W g is the weight matrix of the gated network, N is the number of experts, θ i is the parameter of the i t h expert, and  f expert i can be any neural network (such as fully connected layers, Transformers, etc.). According to the formula, the gated network calculates the weights of each expert for the input sample x and generates a probability distribution g i ( x ) . The  E i ( x ) in the experts network corresponds to the output of the i-th expert.

4. Proposed Methodology

4.1. Overview

Figure 2 illustrates the proposed HyperTransXNet architecture, a hierarchical hybrid model that integrates convolutional neural networks and Transformers to harmonize spectral-spatial dynamics through adaptive feature fusion. The architecture comprises three primary components. First, the Stem module performs lightweight spectral-spatial preprocessing for initial feature extraction. Second, the HyperTransXNet Blocks consist of cascaded units that integrate Hybrid Spatial-Spectral Modules (HSSMs) and a Spatial-Spectral Tokens Enhancer (SSTE) to enable joint local and global refinement. Third, the Classification Head dynamically aggregates features and generates predictions through adaptive pooling and fully connected layers.

4.2. Stem

The Stem module performs lightweight spectral-spatial preprocessing to reduce dimensionality while preserving discriminative features. It combines 1 × 1 convolutional layers for spectral compression and 3 × 3 convolutions for spatial detail extraction. The Stem can be expressed as follows:
Stem ( X ) = Conv 3 × 3 ( ReLU ( Conv 1 × 1 ( X ) ) )
Here, X R B × C × H × W represents the input feature map.

4.3. HyperTransXNet Block

To enhance the local feature extraction capability of the Transformer model, numerous studies have endeavored to integrate convolution and attention mechanisms. However, these methods rely on fixed-sized convolution kernels, rendering them less adaptable to the significant variations in spectral curves across different land cover categories. To address this challenge, we propose a novel HyperTransXNet Block that dynamically fuses global and local information to precisely enhance the frequency bands critical to classification tasks. As illustrated in Figure 2, the input feature map X R B × C × H × W is initially processed by the Stem module. Thereafter, a branch replication operation generates two identical feature maps, which are respectively forwarded to the Spatial-Spectral Local Block (SSLB) and the Spatial Spectral Global Block (SSGB), resulting in the feature maps Y S S L B and Y S S G B . Ultimately, the Spatial Spectral Tokens Enhancer fuses the local features and global representations efficiently to produce the output feature map Y R B × C × H × W . Overall, the proposed HyperTransXNet Block can be expressed as follows:
X 1 , X 2 = Copy ( Stem ( X ) ) Y S S L B , Y S S G B = SSLB ( X 1 ) , SSGB ( X 2 ) Y = SSTE ( Y S S L B , Y S S G B )

Hybrid Spatial-Spectral Module (HSSM)

Traditional hyperspectral image classification methods often depend excessively on a single algorithm or approach, limiting their ability to fully harness the complementary nature of spectral and spatial features. To overcome this limitation, we propose a novel Hybrid Spatial-Spectral Module (HSSM) that integrates multiple sub-modules designed to capitalize on the strengths of diverse architectural approaches, thereby significantly enhancing classification accuracy. Specifically, the HSSM comprises two complementary sub-modules: the Spatial-Spectral Local Block (SSLB), which efficiently extracts local spectral-spatial features via depth-wise separable convolution, and the Spatial-Spectral Global Block (SSGB), which captures global spectral-spatial representations using multi-head attention mechanisms. This modular design not only enables the synergistic optimization of local and global features but also fully exploits the underlying relationships between spectral and spatial information, resulting in an improved feature representation for hyperspectral image classification.
Spatial-Spectral Local Block(SSLB): The SSLB efficiently extracts local information, utilizing depth-wise separable convolution with adaptively generated kernels. Its core structure comprises three components: a dynamic weight generation network, an input unfolding layer, and the application of the dynamic weights (see Figure 3 and Algorithm 1). Given an input feature X 1 R B × C × H × W , the spatial dimensions are first compressed to ( B , C , 1 , 1 ) via an adaptive average pooling (AAP) layer within the dynamic weight generation network. Subsequently, two consecutive 1 × 1 convolutions interleaved with GELU activations are employed to generate the dynamic convolution kernel parameters. The first 1 × 1 convolution maintains channel consistency, while the second increases the number of 1 × 1 output channels by a factor of K 2 . The input feature is then padded using reflection padding (with a padding width of P = K / 2 ) to preserve edge information. Following this, a sliding window unfolding operation with a stride of 1 is performed, transforming the padded input into a neighborhood vector matrix. Finally, the generated dynamic convolution kernel parameters are applied through channel-wise point-wise multiplication with the unfolded input to produce the final output. The SSLB can be formally described as follows:
W dynamic = Reshape ( Conv 1 × 1 C ( C · K 2 ) ( GELU ( Conv 1 × 1 C C ( AAP ( X 1 ) ) ) ) ) X pad = Pad reflect ( X 1 , P ) , P = K / 2 U = Unfold ( X pad ) R B × ( C · K 2 ) × ( H · W ) X unfold = Reshape X R B × C × K 2 × H × W Y S S L B = k = 1 K 2 X unfold · W dynamic
where W dynamic is the kernel dynamically generated from global context via adaptive average pooling, and  U is the unfolded neighborhood matrix. Pad reflect is utilized for reflection padding operations, while Reshape is employed to ensure that the input is expanded for subsequent channel-wise multiplication, all while maintaining consistent dimensions. X pad is a neighborhood vector matrix. Unfold is used to expand and align the input, so as to facilitate the subsequent element-wise multiplication operation across channels. Y S S L B represents the local features extracted from SSLB.
Algorithm 1 Generate dynamic convolution kernels
Input: 
Input feature map X 1 R B × C × H × W
Output: 
Dynamic weights W dynamic R B × C × K × K
  1:
assert  C = = dim             # Channel dimension consistency check.
  2:
weights AdaptiveAvgPool 2 d ( X 1 )        # Global descriptor: B × C × 1 × 1
  3:
weights Conv 1 × 1 C C ( weights )         # Channel-wise feature transformation
  4:
weights GELU ( weights )                 # Nonlinear activation
  5:
weights Conv 1 × 1 C ( C · K 2 ) ( weights )           # Generate kernel parameters: B × ( C × K 2 ) × 1 × 1
  6:
W dynamic reshape ( weights , ( B , C , K , K ) )          # Reshape to B × C × K × K
Spatial-Spectral Global Block (SSGB): The SSGB integrates frequency-domain attention with multi-head attention (MHA) to model global dependencies. It enhances both the spectral discriminability and the spatial contextual representation of hyperspectral data by integrating frequency-domain correction with global spatial modeling. As illustrated in Figure 3, the SSGB consists of two primary components: spectral frequency-domain correction and global spatial attention. The core concept of the SSGB is to utilize the Fourier transform to learn inter-channel spectral feature correlations in the frequency domain, followed by the establishment of long-range spatial dependencies through multi-head self-attention. This design effectively captures the spectral characteristics and spatial relationships within hyperspectral data, thereby improving overall classification performance.
Spectral Frequency Domain Correction: Traditional convolution operations are constrained by their limited local receptive field, impeding their capacity to model global correlations among spectral channels. To address this limitation, spectral frequency-domain correction is implemented via frequency-domain transformations. Initially, the input feature map X 2 R B × C × H × W is reformatted into a sequential representation to facilitate spectral frequency analysis. Specifically, the spatial dimensions ( B , C , H , W ) are flattened and merged into a shape of ( B · H · W , C ) . Subsequently, a fully connected layer is applied to perform a nonlinear mapping of the spectral channels, followed by a real Fourier transform to project the features into the frequency domain. After frequency-domain correction, an inverse Fourier transform is applied to restore the time-domain signal. Finally, the reconstructed features are normalized using Layer Normalization (LayerNorm), and the original tensor structure is recovered. This process can be expressed by the following formula:
X freq = F ( Linear ( rearrange ( X 2 ) ) ) X spec = F 1 ( X freq ) X spec = rearrange ( LayerNorm ( X spec ) ) R B × C × H × W
where F ( · ) denotes the real Fourier transform along the channel dimension, and rearrange is referred to as “dimension reorganization.”
Global spatial attention: The inherent locality of traditional convolutional kernels constrains the model’s ability to capture long-range spatial structures. In contrast, the multi-head attention (MHA) mechanism explicitly establishes global contextual relationships by computing the similarity between any two spatial locations, rendering it particularly effective for detecting sparse yet semantically critical targets in hyperspectral images. In our approach, spectrally corrected features, denoted as X spec , are input into the MHA module to model long-range spatial dependencies. Initially, the feature map is reformulated into a sequence format. For the i t h attention head (among N heads), three independent linear projection matrices are employed to transform the features into query, key, and value vectors, respectively. Subsequently, the spatial attention weight matrix is computed via a scaled dot-product mechanism and normalized using a Softmax function. The outputs from all attention heads are then concatenated and subjected to a final linear transformation to restore the original dimensionality. An inverse transformation follows to reconstruct the spatial feature map, denoted as Y S S G B R B × C × H × W . This process is mathematically formalized as follows:
Q = K = V = rearrange ( X spec ) Q i = Q W i Q K i = K W i K V i = V W i V Attention i = Softmax Q i K i T d k V i R ( H · W ) × B × d k Y S S G B = rearrange ( MHA ( Attention 1 , , Attention N ) )
where W i Q , W i K , W i V R C × d k ( d k = C / N is the subspace dimension) represent three independent linear projection matrices. Y S S G B is the global feature captured by SSGB.
Many existing approaches have predominantly concentrated on single-path models or simple feature combinations. In contrast, HSSM is the first to unify dynamic convolution for local adaptability and frequency-domain attention for global spectral coherence, eliminating handcrafted fusion heuristics. Consequently, employing the HSSM significantly enhances classification accuracy in hyperspectral image classification compared to traditional single-branch or single-algorithm methods.

4.4. Spatial-Spectral Tokens Enhancer (SSTE)

In the process of HSI classification, it is typically necessary to fuse response information from different spectral bands or spatial contexts. However, this process becomes complex due to variations in noise levels, sensor characteristics, and environmental conditions.
To address this challenge, we propose a Spatial-Spectral Tokens Enhancer (SSTE). As shown in Figure 4. The SSTE module first achieves channel alignment through 1 × 1 convolution, ensuring dimensional consistency for feature fusion. It then enhances nonlinear expressiveness through Batch Normalization (BN) and ReLU activation. The output features undergo multi-scale feature fusion via MoE-R, and finally, residual connections effectively preserve the original distribution of the features. The SSTE can be expressed as follows:   
Y i n t = Conv 1 × 1 ( BN ( Y S S G B + Y S S L B ) ) Y = X + MoE - R ( Y i n t )
Here, Y i n t is obtained through the initial fusion of local features and global features.
MoE-R dynamically fuses multi-scale features via Top-K sparse gating, as illustrated in Figure 4. The module comprises two components: parallel multi-scale expert networks and a lightweight routing network. By employing a routing weight mechanism, it directs weighted operations within the experts to achieve hierarchical feature fusion. This design not only enhances the model’s capacity to capture complex patterns in hyperspectral data but also offers an effective solution to the challenge of multi-scale feature fusion in hyperspectral image classification.
Multi-scale Feature Extraction by Expert Networks: Each expert network employs a cascaded structure consisting of a 3 × 3 depth-wise separable convolution followed by a 1 × 1 point convolution, designed to capture spectral-spatial features at multiple scales. Given an input feature map Y e R B × C × H × W , the  3 × 3 depth-wise separable convolution extracts locally detailed features along the spatial dimension, concentrating on high-frequency texture information. Subsequently, the GELU activation function introduces nonlinearity to enhance the model’s expressive capacity. In the subsequent stage, the 1 × 1 point convolution facilitates cross-channel interaction by projecting the local features into a global context space, thereby integrating low-frequency background information. This architecture endows each expert with the simultaneous capabilities of localized, fine-grained perception and global context modeling, forming the fundamental unit of the multi-scale feature pyramid. The four expert networks collectively produce differentiated features that ultimately constitute the multi-scale feature space F ( e ) , emphasizing responses across diverse frequency bands and spatial patterns. The expert network can be expressed by the following formula:
F local ( e ) = GELU DWConv 3 × 3 ( Y e ) F global ( e ) = Conv 1 × 1 ( F local ( e ) ) F ( e ) = F global ( 1 ) , , F global ( 4 ) R B × 4 × C × H × W .
Here, DWConv 3 × 3 represents a 3 × 3 depth-wise separable convolution, and  Conv 1 × 1 represents a 1 × 1 point convolution. The value of e ranges from 1 to 4, corresponding to four experts. Y e represents the input feature map of the expert network.
Dynamic routing and weight assignment: The routing network functions as the module’s control center, tasked with parsing the global statistical characteristics of the input features and dynamically generating expert weights to facilitate the adaptive fusion of multi-scale features. Initially, the network compresses spatial dimensions using adaptive average pooling (AAP), thereby preserving the global spectral distribution information inherent in the input features. Subsequently, a fully connected layer maps the resulting descriptive vector to the expert weight space, and the weights are then normalized into a probability distribution using the Softmax function. As shown in Algorithm 2, A Top-K sparse activation strategy (default K = 1 ) is employed, which retains only the experts with the highest weights for further computation while silencing the remaining expert paths. After normalization, a selection matrix is constructed and fused with the outputs of the selected experts to produce the final output. The routing network can be formally expressed by the following formula:
w b = Softmax GELU ( Linear ( AAP ( Y int ) ) ) R B × E G b , e = exp ( w b , k ) k T b exp ( w b , e ) e T b 0 else Y = e = 1 E G b , e F ( e )
In this work, we set the total number of experts as E = 4 . We define T b = TopK ( w b , K ) to represent the indices of the activated experts for the b t h sample. The operator ⊙ denotes element-wise multiplication and addition, while F ( e ) represents the output of the e t h expert.
Algorithm 2 Top-K Gating Mechanism for Mixture-of-Experts
Require: 
Routing weights w b R B × E , Input feature map X R B × C × H × W , Top-K value K, Expert functions { Expert 1 , , Expert E }
Ensure: 
Fusion output Y R B × C × H × W
  1:
( w top , idx top ) TopK ( w b , k = K , dim = 1 )           # Select top-K experts
  2:
w top w top / ( k = 1 K w top [ : , k ] + ϵ )              # Renormalize weights
  3:
gates Zeros ( B , E )                    # Initialize gate matrix
  4:
gates Scatter ( gates , 1 , idx top , w top )        # Assign weights to selected experts
  5:
expert _ outputs Stack ( [ Expert e ( X ) e 1 . . E ] )      # Compute expert outputs: B × E × C × H × W
  6:
Y e = 1 E gates [ : , e ] expert _ outputs [ : , e ]
Overall, the input features are first broadcast to all the expert networks. Each expert independently performs the operations of depth-wise separable convolution and point-wise convolution to generate a multi-dimensional feature stack tensor. At the same time, the routing network concurrently computes the gated weights and selects the key expert indices. Finally, by performing tensor multiplication and addition with the selected matrix constructed by the routing weights, local features and global features are fused. It is noted that MoE-R is the first Top-K sparse gating mechanism in HSI classification, enabling parameter-efficient aggregation of discriminative spectral-spatial cues.

5. Experiment

5.1. Datasets and Setting

5.1.1. Datasets

Our experiments utilize three benchmark HSI datasets: Indian Pines, Houston 2013, and WHU-Hi-LongKou.
  • The Indian Pines Scene dataset was collected by the AVIRIS sensor in 1992 in the northwestern part of Indiana, USA. It contains hyperspectral image data with a spatial resolution of 145 × 145 and 220 spectral bands. During preprocessing, 20 noisy bands were removed, leaving 200 valid bands for analysis. The dataset includes 16 land-cover classes, such as Alfalfa, Corn, and Woods, and is widely used in hyperspectral image classification tasks. In the experiment, 10% of the samples were randomly selected for training, and the remaining 90% were used for testing to evaluate the model’s performance.
  • Houston 2013 dataset: This dataset encompasses the geographical information of Houston, Texas, USA, and its surrounding areas, acquired using the ITRES CASI-1500 imaging spectrometer. It features a spatial resolution of 349 × 1905 pixels and covers 144 spectral bands. The dataset is based on a high-quality, cloud-free image provided by the Geo-Science and Remote Sensing Society (GRRSS). It contains 15 distinct land-use categories, such as highways, roads, and vegetation. In our experimental setup, we randomly selected 10% of the dataset for model training, with the remaining 90% reserved for validation and testing.
  • The WHU-HI-Longkou (WHL) dataset was collected on 17 July 2018, in Longkou Town, Hubei Province, China, using a DJI M600 Pro drone equipped with a Headwall Nano Hyperspectral Imaging Sensor. The sensor has an 8 mm focal length, and the drone flew at an altitude of 500 m, capturing images with a resolution of 550 × 400 pixels and covering 270 spectral bands, with wavelengths ranging from 400 to 1000 nm. The dataset includes nine land cover types with a total of 204,542 labeled samples, providing rich hyperspectral information for classification method evaluation. In this study, 1% of the labeled samples were used for training, and the remaining 99% for testing, simulating classification tasks with limited labeled data to assess the model’s generalization ability in small sample learning.

5.1.2. Evaluation Metrics

The classification performance was evaluated using three commonly employed metrics: overall accuracy (OA), average accuracy (AA), and the Kappa coefficient ( κ ).

5.1.3. Comparison Methods

Comparison methods: We choose seven state-of-the-art methods to compare with HyperTransXNet, including CNNs (e.g., 2D-CNN [43], 3D-CNN [3], and Hybridsn [6]), Transformers (such as HiT [32], MorphFormer [31], and SSFTT [7]), and Mamba (Mambahsi [44]).

5.1.4. Setting

During training, we randomly extracted 100 patches of size 15 × 15 as input data. We set the total number of iterations to 100 and employed all comparison methods alongside HyperTransXNet using the PyTorch (version 2.6.0) framework. HyperTransXNet was optimized using the Adam optimizer with a learning rate of 1 × 10 3 and a batch size of 100.

5.2. Results and Analysis

We conducted experiments on three widely used HSI datasets and evaluated their performance using two metrics. The results, presented in Table 1, Table 2 and Table 3 (with the best values highlighted in bold), reveal that the proposed HyperTransXNet achieves superior classification accuracy across all datasets. This performance is primarily attributed to two key innovations. First, HyperTransXNet incorporates a Hybrid Spatial-Spectral Module (HSSM) that employs parallel local dynamic convolutions alongside global frequency-domain attention mechanisms, thereby capturing both fine-grained spatial details and broader spectral correlations. Second, the architecture integrates a Mixture-of-Experts Routing mechanism (MoE-R) that adaptively fuses multi-scale features through a learnable routing process. Notably, HyperTransXNet delivers satisfactory classification outcomes on the unbalanced WHU-HI-Longkou dataset and demonstrates enhanced performance across nearly all categories.
An additional observation is that Transformer architectures augmented with local feature extraction components (e.g., SSFTT, HiT, and MorphFormer) demonstrably outperform conventional Transformer models (e.g., ViT and DeepViT). This advantage arises because, although traditional Transformers effectively capture global representations from patch embeddings, they often fall short in emphasizing localized features. In contrast, while straightforward feature fusion strategies may introduce noise and redundancy, the integration of local features with global representations occurs in HyperTransXNet. For example, on the Indian Pines dataset, HyperTransXNet achieved a classification accuracy of 94.74%, surpassing the 94.03% of MorphFormer by 0.71%, thereby underscoring the efficacy of combining local and global feature representations.
Furthermore, convolutional neural networks (CNNs), as classical deep learning models, excel in image processing tasks due to their proficiency in extracting local features. However, their focus on local connectivity limits their ability to capture long-range dependencies and global contextual relationships, as evidenced by our experimental comparisons. Specifically, on the Houston 2013 dataset, HyperTransXNet achieved an accuracy of 98.46%, outperforming traditional 2D-CNN and 3D-CNN models. This notable improvement emphasizes the critical importance of integrating both local and global features and highlights the potential for further architectural enhancements in CNN frameworks.
In addition, the recently introduced Mamba series models have emerged as promising deep learning architectures with significant advantages in hyperspectral image (HSI) classification. These models have achieved breakthroughs in multi-scale feature extraction by innovatively integrating the Selective State Space Model (SSM) with a dynamic scanning mechanism. Their architectural design combines the strengths of convolutional neural networks (CNNs) in local feature extraction with the robust global context modeling capabilities of Transformer architectures. However, a fundamental limitation of these models is the absence of an explicit feature fusion module, necessitating external components for effective feature integration in tasks that demand complex feature collaboration. Analysis of experimental results reveals that the proposed HyperTransXNet, an advanced architecture that effectively fuses both local and global features, outperforms the Mamba series models. For example, on the Houston 2013 dataset, HyperTransXNet achieves a 1.66% higher overall accuracy compared to Mambahsi. This outcome underscores the importance of incorporating a fusion module and further demonstrates the potential of the Mamba series models in the realm of HSI classification.
Figure 5, Figure 6 and Figure 7 present a comparative performance analysis of the proposed HyperTransXNet method across three widely adopted benchmark datasets in remote sensing image analysis: the Indian Pines scene, Houston 2013, and WHU-HI-Longkou datasets. The visual outcomes clearly demonstrate that HyperTransXNet achieves exceptionally high classification accuracy across all categories. Moreover, the classification maps produced by the model exhibit strong concordance with the ground truth, indicating that HyperTransXNet effectively integrates both local features and global representations. This superior performance can be primarily attributed to the design of HyperTransXNet, which is specifically optimized for the fusion of heterogeneous features, thereby significantly enhancing classification performance. A detailed examination of the comparative charts reveals that most alternative methods yield classification maps with substantial noise levels, a finding that corresponds with the quantitative results summarized in previous tables. This observation suggests that conventional convolutional neural networks (CNNs), frequently employed in computer vision tasks, often encounter difficulties in capturing global representations. Similarly, recent transformer-based models, despite their rising prominence, also face challenges in effectively merging local and global features.
These results underscore the necessity of developing methodologies that can adeptly balance the integration of local and global features in remote sensing image classification tasks, thereby enhancing overall performance. In summary, through its innovative architectural design, HyperTransXNet not only overcomes the limitations of traditional feature fusion methods but also offers a more efficient and accurate solution for remote sensing image classification.
To comprehensively evaluate the feature representation capabilities of the proposed HyperTransXNet method, we conducted a t-SNE visualization comparative analysis against six benchmark methods: CNN-based approaches (2D-CNN and HybridSN), visual Transformer series models (SSFTT and MorphFormer), and Mamba series models (Mambahsi). Figure 8 displays the t-SNE visualization results of each method on the Indian Pines dataset, a widely used resource in remote sensing image analysis due to its rich hyperspectral information and diverse land cover types. The visualization clearly indicates that HyperTransXNet outperforms the other methods in terms of clustering performance within the feature space, with its t-SNE map exhibiting more compact intra-class clusters and more distinct inter-class boundaries. This superior performance primarily arises from the model’s unique architectural design, where the integration of a global frequency domain attention mechanism with local dynamic convolution effectively captures both global dependencies and local contextual information in hyperspectral images. Such capabilities are essential for accurately identifying land cover categories characterized by complex spectral features. In contrast, traditional convolutional neural networks, constrained by limited long-range dependency modeling, demonstrate more dispersed cluster distributions and substantial inter-class interference. Transformer-based models, despite improvements in processing global information, still reveal limitations in capturing fine-grained local details, as evidenced by less cohesive intra-class aggregation and noticeable misclassification. Mamba series models, which typically employ unidirectional or bidirectional global scanning without targeted modeling of local spatial-spectral cubes, struggle to capture subtle local features, resulting in inadequate intra-class cohesion and pronounced misclassification. This visualization analysis not only corroborates the outstanding performance of HyperTransXNet in hyperspectral classification tasks but also provides intuitive evidence of its advantages, thereby reinforcing its effectiveness in feature representation and establishing a robust theoretical foundation for its broad application in remote sensing image analysis.

5.3. Comparison of Computational Complexity

This study presents a computational complexity analysis of various comparison methods alongside the HyperTransXNet network, all assessed using the Houston 2013 dataset. As detailed in Table 4, HyperTransXNet exhibits considerable advantages while maintaining high classification accuracy. We can observe that its overall accuracy (OA) reaches 98.46% and its Kappa coefficient is 98.33%, outperforming current mainstream Transformer architectures. Notably, HyperTransXNet achieves the highest OA performance with only 1.08 MB of parameters.
The computational complexity analysis further indicates that the CvT network requires 9.04 G floating-point operations (FLOPs) and a training time of 2758.06 s (approximately 46 min), with its high computational density and parameter count (17.77 MB) potentially leading to training efficiency bottlenecks. In contrast, HyperTransXNet completes training in just 201.39 s, utilizing only 0.24 FLOPs and 1.08 of parameters, while its testing time is limited to 0.13 s. At the same time, its memory consumption is only 707.06 MB, which is the lowest among all comparison models. This optimized balance between accuracy and efficiency renders the approach highly applicable to hyperspectral classification scenarios that demand real-time processing.

5.4. Ablation Studies

Ablation study of the input size: Table 5 presents an investigation into the impact of input size on the classification performance across three datasets. By varying the input size from 11 × 11 to 19 × 19 , the resulting classification accuracies, as detailed in Table 5, demonstrate that performance varies significantly with different input dimensions. Notably, smaller input sizes yield higher overall accuracy (OA) and Kappa coefficients, likely because they facilitate the extraction of more precise features, thereby enhancing the identification of land cover types. Conversely, larger input sizes may introduce extraneous noise. These findings further substantiate the robustness of the proposed HyperTransXNet method. Consequently, to comprehensively evaluate HyperTransXNet, the input size was optimally set to 15 × 15 .
Ablation study of different modules: Table 6 presents an ablation study assessing the impacts of the SSLB and SSGB configurations on the Houston 2013 dataset, with the 2D-CNN serving as the baseline method. As shown in Table 6, the proposed models (HyperTransXNet_SSLB and HyperTransXNet_SSGB) outperform the conventional 2D-CNN in terms of overall accuracy (OA) and the Kappa coefficient. Specifically, HyperTransXNet_SSLB improves OA by 0.73% and the Kappa coefficient by 0.79%, while HyperTransXNet_SSGB achieves improvements of 0.86% in OA and 0.94% in the Kappa coefficient relative to the 2D-CNN. This enhancement is attributed to the SSLB module’s ability to effectively capture local contextual features and the SSGB module’s strength in modeling global dependency relationships. These findings underscore that integrating optimal local contextual features with global dependencies significantly enhances classification performance.
Ablation study of SSTE: In order to verify the superiority of the proposed SSTE module for feature fusion, Table 7 presents the ablation study results on the Houston 2013 dataset. The experimental results indicate that the model enhanced with SSTE: achieves superior performance in both overall accuracy (OA) and the Kappa coefficient. Notably, the HyperTransXNet_SSLB_SSGB_SSTE model shows improvements of 0.30% in OA and 0.32% in κ compared to the HyperTransXNet_SSLB_SSGB. These findings not only validate the effectiveness of SSTE in integrating corresponding local features with the global representation but also underscore the merits of incorporating this module.
Ablation study of SSTE and MoE-R: In Table 8, we conducted an ablation study on the SSTE and MoE-R modules using the Houston 2013 dataset. The results indicate that HyperTransXNet_SSLB_SSGB_SSTE significantly outperforms HyperTransXNet_SSLB_SSGB_MoE-R in terms of OA (4.73%) and κ (5.11%). This superior performance can be attributed to the ability of SSTE to achieve channel alignment through 1 × 1 convolution, which ensures consistent dimensions during feature fusion by MoE-R. Additionally, SSTE effectively preserves the original distribution of features through residual connections. The experimental results confirm that SSTE and MoE-R exhibit a complementary relationship rather than redundancy.
Ablation study of the dynamic kernel size in the HSSM: By default, we set the dynamic kernel size (K) in HSSM to 3. Since variations in kernel size can significantly impact the performance of the module, we conducted experiments to assess the effects of different kernel sizes on overall accuracy (OA), specifically using sizes of 1, 3, and 5. As shown in Figure 9, K = 3 achieved the highest classification accuracy (OA = 98.46%) on the Houston 2013 dataset. In contrast, K = 1 and K = 5 resulted in performance declines of 0.48% (vs. 97.98%) and 0.43% (vs. 98.03%), respectively, with the model’s OA fluctuation remaining below 0.5%. According to the experimental results, the robustness of the model under varying kernel sizes is fully demonstrated.
Ablation study of the expert count in the MoE-R: We set the expert count (E) in the Mixture-of-Experts Routing mechanism (MoE-R) to 4 by default. To validate this choice, we evaluated the impact of different expert counts; specifically, E { 2 , 4 , 8 } on parameter efficiency and overall accuracy (OA) using the Houston 2013 dataset. As shown in Table 9, increasing E from 2 to 4 resulted in an improvement in OA by 0.40% with a Marginal Parameter Efficiency (MPE) of 4.44. Conversely, increasing E from 4 to 8 led to a decrease in OA by 0.50%, accompanied by a negative MPE of −2.78. These results demonstrate that E = 4 optimally balances MPE (4.44) and OA (98.46%), confirming the scientific rationale behind our default selection and indicating model robustness, as evidenced by an OA fluctuation of ≤0.50%, even with suboptimal E values.
Ablation study of training sample ratio: This study investigates the impact of training sample ratio on the classification performance of three hyperspectral benchmark datasets. To evaluate the model’s performance across varying levels of data availability, we selected a small sample size of 5%, representing an extreme scenario in which resources are severely limited in real-world applications. Additionally, we included sample sizes ranging from 20% to 50% to identify the saturation point of model performance. According to Figure 10, there is a significant positive correlation between overall classification accuracy (OA) and the number of training samples. For instance, in the Indian Pines dataset, the OA increased from 94.74% to 95.84% as the number of training samples gradually increased. This trend indicates that the model’s ability to capture a consistent data distribution improves with the expansion of the training scale. Similarly, in the Houston 2013 dataset, the OA rose from 98.46% to 99.41%, further confirming that sufficient training data enhances the model’s cross-class generalization ability. These findings suggest that increasing the number of training samples not only mitigates the impact of intra-class feature heterogeneity but also improves the model’s capacity to delineate inter-class decision boundaries. Overall, this study provides valuable insights for hyperspectral classification tasks, indicating that expanding the training dataset significantly enhances the robustness and reliability of the classification system by strengthening the model’s structured representation of the feature space.

6. Conclusions

In this study, we leverage the advantages of CNNs and vision Transformers to propose a novel hybrid network, HyperTransXNet, for hyperspectral image (HSI) classification. Unlike previous approaches, we introduce a HSSM that synergistically combines global frequency-domain attention mechanisms with local dynamic convolutions, enabling the effective capture of both global dependencies and local contextual information in hyperspectral imagery. To adaptively fuse the corresponding local features and global representations, we further propose an SSTE mechanism for robust information integration. Extensive experiments demonstrate that our proposed HyperTransXNet outperforms state-of-the-art CNNs and vision Transformers. In future work, we aim to explore additional fusion modules to further enhance the learning capabilities of Transformers.

Author Contributions

Software, S.X.; Validation, L.L.; Writing—original draft, Z.L.; Writing—review & editing, X.D. and X.Y.; Funding acquisition, X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (NSFC) Fund under Grant 62301174 and 62462031, and Guangzhou basic and applied basic research topics under Grant 2024A04J2081 and Grant 2025A04J3375. This work was also supported by the Natural Science Foundation of Jiangxi Province under Grant 20242BAB26023.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kang, X.; Wang, Z.; Duan, P.; Wei, X. The potential of hyperspectral image classification for oil spill mapping. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar]
  2. Chen, F.; Wang, K.; Van de Voorde, T.; Tang, T.F. Mapping urban land cover from high spatial resolution hyperspectral data: An approach based on simultaneously unmixing similar pixels with jointly sparse spectral mixture analysis. Remote Sens. Environ. 2017, 196, 324–342. [Google Scholar]
  3. Yang, X.; Zhang, X.; Ye, Y.; Lau, R.Y.; Lu, S.; Li, X.; Huang, X. Synergistic 2D/3D convolutional neural network for hyperspectral image classification. Remote Sens. 2020, 12, 2033. [Google Scholar]
  4. Shen, J.; Zhang, D.; Dong, G.; Sun, D.; Liang, X.; Su, M. Classification of hyperspectral images based on fused 3D inception and 3D-2D hybrid convolution. Signal Image Video Process. 2024, 18, 3031–3041. [Google Scholar]
  5. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep dense convolutional neural network for hyperspectral image classification. Remote Sens. 2018, 10, 1454. [Google Scholar]
  6. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar]
  7. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar]
  8. Yao, J.; Hong, D.; Li, C.; Chanussot, J. Spectralmamba: Efficient mamba for hyperspectral image classification. arXiv 2024, arXiv:2404.08489. [Google Scholar]
  9. Sheng, J.; Zhou, J.; Wang, J.; Ye, P.; Fan, J. Dualmamba: A lightweight spectral-spatial mamba-convolution network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 63, 5501415. [Google Scholar]
  10. Li, Y.; Zhang, H.; Shen, Q. Spectral-spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar]
  11. Xu, H.; Yao, W.; Cheng, L.; Li, B. Multiple spectral resolution 3D convolutional neural network for hyperspectral image classification. Remote Sens. 2021, 13, 1248. [Google Scholar]
  12. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral image classification using a hybrid 3D-2D convolutional neural networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar]
  13. Zhang, H.; Li, Y.; Zhang, Y.; Shen, Q. Spectral-spatial classification of hyperspectral imagery using a dual-channel convolutional neural network. Remote Sens. Lett. 2017, 8, 438–447. [Google Scholar]
  14. Haque, M.R.; Mishu, S.Z.; Palash Uddin, M.; Al Mamun, M. A lightweight 3D-2D convolutional neural network for spectral-spatial classification of hyperspectral images. J. Intell. Fuzzy Syst. 2022, 43, 1241–1258. [Google Scholar]
  15. Luo, Y.; Zou, J.; Yao, C.; Zhao, X.; Li, T.; Bai, G. HSI-CNN: A novel convolution neural network for hyperspectral image. In Proceedings of the 2018 International Conference on Audio, Language and Image Processing (ICALIP), Shanghai, China, 16–17 July 2018; pp. 464–469. [Google Scholar]
  16. Liang, L.; Zhang, S.; Li, J.; Plaza, A.; Cui, Z. Multi-scale spectral-spatial attention network for hyperspectral image classification combining 2D octave and 3D convolutional neural networks. Remote Sens. 2023, 15, 1758. [Google Scholar]
  17. Diakite, A.; Jiangsheng, G.; Xiaping, F. Hyperspectral image classification using 3D 2D CNN. IET Image Process. 2021, 15, 1083–1092. [Google Scholar]
  18. Cao, J.; Li, X. A 3D 2D convolutional neural network model for hyperspectral image classification. arXiv 2021, arXiv:2111.10293. [Google Scholar]
  19. Liu, H.; Li, W.; Xia, X.G.; Zhang, M.; Gao, C.Z.; Tao, R. Central Attention Network for Hyperspectral Imagery Classification. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 8989–9003. [Google Scholar] [CrossRef]
  20. Liu, H.; Li, W.; Xia, X.G.; Zhang, M.; Tao, R. Multiarea Target Attention for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5524916. [Google Scholar] [CrossRef]
  21. Mei, S.; Li, X.; Liu, X.; Cai, H.; Du, Q. Hyperspectral image classification using attention-based bidirectional long short-term memory network. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5509612. [Google Scholar]
  22. Ayas, S.; Tunc-Gormus, E. SpectralSWIN: A spectral-swin transformer network for hyperspectral image classification. Int. J. Remote Sens. 2022, 43, 4025–4044. [Google Scholar]
  23. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. QTN: Quaternion transformer network for hyperspectral image classification. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 7370–7384. [Google Scholar]
  24. Qiao, X.; Huang, W. A dual frequency transformer network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 10344–10358. [Google Scholar]
  25. Li, Y.; Yang, X.; Tang, D.; Zhou, Z. RDTN: Residual Densely Transformer Network for hyperspectral image classification. Expert Syst. Appl. 2024, 250, 123939. [Google Scholar]
  26. Yang, X.; Cao, W.; Tang, D.; Zhou, Y.; Lu, Y. ACTN: Adaptive Coupling Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5503115. [Google Scholar]
  27. Gong, Z.; Zhou, X.; Yao, W. MultiScale spectral–spatial convolutional transformer for hyperspectral image classification. IET Image Process. 2024, 18, 4328–4340. [Google Scholar]
  28. Luo, Y.; Tang, D.; Yang, X.; Li, Y. Spectral-spatial attention transformer network for hyperspectral image classification. In Proceedings of the The International Conference Optoelectronic Information and Optical Engineering (OIOE2024), Wuhan, China; 2024; Volume 13513, pp. 833–838. [Google Scholar]
  29. Liu, H.; Li, W.; Xia, X.G.; Zhang, M.; Guo, Z.; Song, L. SegHSI: Semantic Segmentation of Hyperspectral Images with Limited Labeled Pixels. IEEE Trans. Image Process. 2024, 33, 6469–6482. [Google Scholar] [CrossRef]
  30. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  31. Roy, S.K.; Deria, A.; Shah, C.; Haut, J.M.; Du, Q.; Plaza, A. Spectral–spatial morphological attention transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5503615. [Google Scholar]
  32. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral image transformer classification networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528715. [Google Scholar]
  33. Liu, B.; Liu, Y.; Zhang, W.; Tian, Y.; Kong, W. Spectral Swin Transformer Network for Hyperspectral Image Classification. Remote Sens. 2023, 15, 3721. [Google Scholar] [CrossRef]
  34. Wang, C.; Huang, J.; Lv, M.; Du, H.; Wu, Y.; Qin, R. A local enhanced mamba network for hyperspectral image classification. Int. J. Appl. Earth Obs. Geoinf. 2024, 133, 104092. [Google Scholar]
  35. He, Y.; Tu, B.; Liu, B.; Li, J.; Plaza, A. 3DSS-Mamba: 3D-spectral-spatial mamba for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5534216. [Google Scholar]
  36. Liu, Q.; Yue, J.; Fang, Y.; Xia, S.; Fang, L. HyperMamba: A Spectral-Spatial Adaptive Mamba for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5536514. [Google Scholar]
  37. Lou, M.; Zhang, S.; Zhou, H.Y.; Yang, S.; Wu, C.; Yu, Y. TransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Visual Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 11534–11547. [Google Scholar] [CrossRef]
  38. Fedus, W.; Zoph, B.; Shazeer, N. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. J. Mach. Learn. Res. 2022, 23, 1–39. [Google Scholar]
  39. Li, J.; Su, Q.; Yang, Y.; Jiang, Y.; Wang, C.; Xu, H. Adaptive gating in mixture-of-experts based language models. arXiv 2023, arXiv:2310.07188. [Google Scholar]
  40. Wu, J.; Hou, M. Enhancing diversity for logical table-to-text generation with mixture of experts. Expert Syst. 2024, 41, e13533. [Google Scholar]
  41. Shazeer, N.; Mirhoseini, A.; Maziarz, K.; Davis, A.; Le, Q.; Hinton, G.; Dean, J. Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv 2017, arXiv:1701.06538. [Google Scholar]
  42. Lin, B.; Tang, Z.; Ye, Y.; Cui, J.; Zhu, B.; Jin, P.; Huang, J.; Zhang, J.; Pang, Y.; Ning, M.; et al. Moe-llava: Mixture of experts for large vision-language models. arXiv 2024, arXiv:2401.15947. [Google Scholar]
  43. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.; Zhang, X.; Huang, X. Hyperspectral image classification with deep learning models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar]
  44. Li, Y.; Luo, Y.; Zhang, L.; Wang, Z.; Du, B. MambaHSI: Spatial-spectral mamba for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5524216. [Google Scholar]
Figure 1. The basic framework of MoE.
Figure 1. The basic framework of MoE.
Remotesensing 17 02361 g001
Figure 2. The overall architecture of the proposed HyperTransXNet.
Figure 2. The overall architecture of the proposed HyperTransXNet.
Remotesensing 17 02361 g002
Figure 3. Workflow of the proposed HyperTransXNet Block.
Figure 3. Workflow of the proposed HyperTransXNet Block.
Remotesensing 17 02361 g003
Figure 4. Our MoE-R performs multi-scale token aggregations.
Figure 4. Our MoE-R performs multi-scale token aggregations.
Remotesensing 17 02361 g004
Figure 5. The classification maps obtained using different methods on the Houston 2013 scene dataset (with 10% training samples).
Figure 5. The classification maps obtained using different methods on the Houston 2013 scene dataset (with 10% training samples).
Remotesensing 17 02361 g005
Figure 6. The classification maps obtained using different methods on the Indian Pines Scene dataset (with 10% training samples).
Figure 6. The classification maps obtained using different methods on the Indian Pines Scene dataset (with 10% training samples).
Remotesensing 17 02361 g006
Figure 7. The classification maps obtained using different methods on the WHL dataset (with 1% training samples).
Figure 7. The classification maps obtained using different methods on the WHL dataset (with 1% training samples).
Remotesensing 17 02361 g007
Figure 8. The T-SNE results obtained using different methods on the Indian Pines dataset (with 10% training samples).
Figure 8. The T-SNE results obtained using different methods on the Indian Pines dataset (with 10% training samples).
Remotesensing 17 02361 g008
Figure 9. Impact of different kernel sizes for the OA on the Houston 2013 dataset.
Figure 9. Impact of different kernel sizes for the OA on the Houston 2013 dataset.
Remotesensing 17 02361 g009
Figure 10. Impact of training sample ratio on classification performance across datasets.
Figure 10. Impact of training sample ratio on classification performance across datasets.
Remotesensing 17 02361 g010
Table 1. Comparison with the state-of-the-art CNNs and Transformers on the Indian Pines Scene dataset (10% training samples).
Table 1. Comparison with the state-of-the-art CNNs and Transformers on the Indian Pines Scene dataset (10% training samples).
Class2D-CNN3D-CNNHybridsnHiTSSFTTMorphformerMambaHSIHyperTransXNet
Alfalfa92.82 ± 4.8069.95 ± 17.9218.85 ± 29.7916.34 ± 15.7089.65 ± 6.9182.13 ± 22.4076.83 ± 22.1589.51 ± 14.39
Corn-notill93.81 ± 1.9788.61 ± 1.1784.84 ± 11.2290.48 ± 1.2294.11 ± 1.0893.38 ± 2.1490.97 ± 3.5094.23 ± 1.72
Corn-mintill92.19 ± 1.7786.74 ± 1.9075.93 ± 18.4093.13 ± 3.6390.13 ± 2.6691.28 ± 3.6690.46 ± 5.0596.22 ± 2.30
Corn97.94 ± 1.5093.92 ± 2.4480.93 ± 17.0191.17 ± 5.1294.90 ± 3.4695.26 ± 4.0490.38 ± 6.2493.90 ± 5.63
Grass-pasture93.09 ± 3.3293.44 ± 0.7073.56 ± 16.4381.89 ± 10.5993.08 ± 2.5294.72 ± 1.5595.26 ± 2.1495.61 ± 2.37
Grass-trees95.65 ± 2.9794.82 ± 0.6775.90 ± 15.7595.11 ± 0.9895.98 ± 1.2995.47 ± 1.3894.70 ± 2.3797.26 ± 1.22
Grass-pasture-mowed7.94 ± 18.290.00 ± 0.000.00 ± 0.000.00 ± 0.0054.69 ± 35.8471.56 ± 18.0144.40 ± 32.5228.40 ± 30.54
Hay-windrowed99.69 ± 0.5598.87 ± 1.1087.72 ± 13.0999.86 ± 0.2198.77 ± 1.3599.81 ± 0.3999.72 ± 0.46100.00 ± 0.00
Oats73.30 ± 29.090.00 ± 0.002.45 ± 5.090.00 ± 0.0054.34 ± 32.0421.28 ± 30.1457.22 ± 25.9550.56 ± 36.47
Soybean-notill87.78 ± 1.6083.25 ± 1.2278.05 ± 9.8777.61 ± 2.0787.11 ± 1.7788.80 ± 3.5884.25 ± 5.7483.09 ± 2.03
Soybean-mintill96.26 ± 1.2494.38 ± 0.5191.41 ± 4.1696.09 ± 1.7496.78 ± 0.8296.27 ± 0.5996.12 ± 3.5498.40 ± 0.54
Soybean-clean91.80 ± 2.2189.11 ± 1.7178.53 ± 12.7691.18 ± 3.4689.52 ± 3.3587.66 ± 5.0892.96 ± 5.9993.16 ± 4.09
Wheat98.12 ± 1.3286.71 ± 7.8154.68 ± 33.4391.89 ± 2.9695.00 ± 3.6294.35 ± 4.8895.14 ± 2.9090.76 ± 5.41
Woods98.28 ± 2.4297.81 ± 0.5894.10 ± 5.2299.45 ± 0.3898.67 ± 0.6698.87 ± 0.3199.16 ± 0.4399.49 ± 0.39
Buildings-Grass-Trees-Drives97.82 ± 1.4693.48 ± 2.2573.19 ± 17.0987.90 ± 11.8396.08 ± 2.7696.06 ± 1.4890.86 ± 4.1296.31 ± 3.60
Stone-Steel-Towers52.74 ± 21.3946.87 ± 17.6241.04 ± 32.9116.79 ± 20.5539.20 ± 32.3025.11 ± 33.422.98 ± 18.6945.36 ± 17.65
OA (%)94.48 ± 1.4191.48 ± 0.5283.10 ± 10.1990.85 ± 1.1394.09 ± 0.9794.03 ± 0.9092.72 ± 1.1594.74 ± 0.39
AA (%)83.81 ± 3.3873.92 ± 2.1262.87 ± 12.0270.56 ± 2.8484.54 ± 3.9182.75 ± 2.8582.59 ± 3.3584.52 ± 2.76
κ (%)93.69 ± 1.6190.25 ± 0.5980.70 ± 11.5289.53 ± 1.3193.25 ± 1.1093.18 ± 1.0391.69 ± 1.3093.99 ± 0.44
Table 2. Comparison with the state-of-the-art CNNs and Transformers on the Houston 2013 dataset (10% training samples).
Table 2. Comparison with the state-of-the-art CNNs and Transformers on the Houston 2013 dataset (10% training samples).
Class2D-CNN3D-CNNHybridSNHiTSSFTTMorphformerMambaHSIHyperTransXNet
Healthy Grass95.21 ± 1.6691.06 ± 3.4490.76 ± 1.8890.02 ± 3.4096.42 ± 1.1685.97 ± 8.7596.94 ± 1.8098.41 ± 1.19
Stressed grass95.51 ± 1.7288.85 ± 6.4686.26 ± 8.2795.63 ± 0.8397.16 ± 1.4288.97 ± 5.5894.21 ± 2.6997.25 ± 2.50
Synthetic GrassTrees99.24 ± 0.4597.36 ± 3.0695.85 ± 3.6797.48 ± 0.6699.30 ± 0.5891.37 ± 13.8497.26 ± 1.0697.62 ± 0.73
Trees94.72 ± 2.1888.28 ± 4.5279.71 ± 7.4687.64 ± 4.9697.68 ± 0.9090.96 ± 4.9594.49 ± 2.3196.98 ± 1.97
Soil99.81 ± 0.1795.72 ± 3.7196.41 ± 3.8399.90 ± 0.1499.43 ± 0.6397.96 ± 2.1799.63 ± 0.3899.97 ± 0.04
Water94.05 ± 0.9786.16 ± 1.9291.69 ± 3.1274.69 ± 3.1493.61 ± 3.0790.85 ± 4.2590.86 ± 6.6992.57 ± 3.57
Residential96.82 ± 1.0283.00 ± 6.9878.74 ± 17.1273.54 ± 8.9897.52 ± 0.8085.78 ± 24.7395.88 ± 1.5298.89 ± 0.68
Commercial97.62 ± 1.5189.55 ± 1.6193.01 ± 2.6587.67 ± 4.7597.88 ± 1.2490.76 ± 9.5696.46 ± 1.3896.06 ± 1.43
Road95.42 ± 1.5185.03 ± 3.3873.95 ± 14.3681.45 ± 3.1596.70 ± 1.5687.54 ± 7.9996.16 ± 1.5999.00 ± 0.90
Highway99.09 ± 1.5791.67 ± 6.5491.92 ± 7.9296.51 ± 2.2599.78 ± 0.3893.89 ± 9.7699.94 ± 0.19100.00 ± 0.00
Railway99.58 ± 0.4681.57 ± 5.8785.29 ± 7.9293.82 ± 3.7899.76 ± 0.4491.69 ± 16.2697.00 ± 1.9299.62 ± 0.80
Parking Lot 197.88 ± 1.8394.30 ± 1.8195.79 ± 2.7193.77 ± 3.0998.70 ± 1.2488.63 ± 13.8997.32 ± 1.7199.22 ± 0.84
Parking Lot 297.55 ± 2.4481.89 ± 6.2486.51 ± 8.0890.28 ± 3.6098.54 ± 1.2392.25 ± 6.6697.27 ± 1.7998.13 ± 2.36
Tennise Court99.98 ± 0.0798.60 ± 0.9792.98 ± 6.1699.82 ± 0.5599.67 ± 0.4799.04 ± 1.4699.77 ± 0.50100.00 ± 0.00
Running Track97.87 ± 1.7194.77 ± 5.4191.48 ± 5.6099.21 ± 1.3798.95 ± 0.7193.32 ± 7.4997.17 ± 4.63100.00 ± 0.00
OA (%)97.32 ± 0.4889.54 ± 3.2188.19 ± 4.7590.67 ± 1.0198.15 ± 0.5390.98 ± 7.7196.80 ± 0.5598.46 ± 0.29
AA (%)97.03 ± 0.3789.74 ± 2.8788.97 ± 4.0990.76 ± 0.9297.81 ± 0.5990.99 ± 7.4696.69 ± 0.6998.25 ± 0.22
κ ( % ) 97.10 ± 0.5188.70 ± 3.4787.24 ± 5.1389.91 ± 1.0998.00 ± 0.5890.24 ± 8.3696.54 ± 0.5998.33 ± 0.31
Table 3. Comparison with the state-of-the-art CNNs and Transformers on the WHL dataset (1% training samples).
Table 3. Comparison with the state-of-the-art CNNs and Transformers on the WHL dataset (1% training samples).
Class2D-CNN3D-CNNHybridSNHiTSSFTTMorphformerMambaHSIHyperTransXNet
Corn99.87 ± 0.0399.34 ± 0.4099.34 ± 0.5899.79 ± 0.0899.88 ± 0.0499.89 ± 0.0699.73 ± 0.2499.84 ± 0.11
Cotton99.72 ± 0.0996.30 ± 1.2497.55 ± 3.1897.26 ± 2.0599.69 ± 0.1199.78 ± 0.1399.78 ± 0.2599.19 ± 0.75
Sesame94.97 ± 1.2755.88 ± 29.7981.52 ± 14.5494.55 ± 3.2498.93 ± 0.5499.44 ± 0.4295.44 ± 1.7098.82 ± 0.77
Broad-leaf soybean99.12 ± 0.0896.24 ± 0.8998.22 ± 0.6499.77 ± 0.1099.63 ± 0.0599.65 ± 0.1699.61 ± 0.2299.83 ± 0.13
Narrow-leaf soybean95.42 ± 0.7187.80 ± 2.4481.06 ± 10.1988.16 ± 3.4798.30 ± 0.4897.60 ± 1.5095.25 ± 3.2496.38 ± 1.97
Rice98.57 ± 0.1998.11 ± 0.6697.28 ± 0.9199.36 ± 0.1998.96 ± 0.4499.05 ± 0.2398.33 ± 0.7298.54 ± 0.49
Water99.74 ± 0.0499.51 ± 0.2297.11 ± 0.5399.99 ± 0.0199.56 ± 0.1299.49 ± 0.1599.86 ± 0.1199.85 ± 0.10
Roads and houses85.56 ± 0.7882.84 ± 2.2278.60 ± 7.0984.38 ± 3.8789.32 ± 2.8291.03 ± 1.8287.57 ± 3.6087.32 ± 2.85
Mixed weed78.58 ± 2.0479.05 ± 2.9937.10 ± 12.8670.44 ± 3.6584.48 ± 3.0983.76 ± 1.7986.64 ± 5.8988.04 ± 4.99
OA(%)98.35 ± 0.0996.60 ± 0.6395.77 ± 0.8098.12 ± 0.1498.86 ± 0.2098.74 ± 0.1698.74 ± 0.2298.91 ± 0.16
AA (%)93.24 ± 0.4684.58 ± 3.9582.29 ± 3.5992.63 ± 0.8095.72 ± 0.7295.03 ± 0.6795.80 ± 0.9496.42 ± 0.69
κ ( % ) 97.82 ± 0.1395.49 ± 0.8594.38 ± 1.0797.52 ± 0.1998.50 ± 0.2698.34 ± 0.2198.35 ± 0.2998.57 ± 0.21
Table 4. Comparison of computational complexity.
Table 4. Comparison of computational complexity.
MethodsFLOPs (G)Param (MB)Inference Speed
(sample/s)
Peak Memory (MB)Training Time (s)Testing Time (s)OA (%)AA (%) κ (%)
ViT2.7152.22565.30715.06727.893.4390.27 ± 1.6990.45 ± 1.4289.48 ± 1.83
DeepViT2.7152.22352.53881.061497.976.8291.58 ± 1.5091.38 ± 1.4390.89 ± 1.62
CvT9.0417.77149.528494.942758.0611.6593.52 ± 1.8893.75 ± 1.9592.99 ± 2.04
HiT1.8116.94103.918314.00112.046.795.20 ± 0.3394.80 ± 0.3294.81 ± 0.35
SSFTT0.9787.04148.298198.00389.621.5998.15 ± 0.5397.82 ± 0.5998.00 ± 0.58
MorphFormer0.7487.0411.448200.00874.994.1690.98 ± 7.7190.99 ± 7.4690.24 ± 8.36
MambaHSI0.0060.10440.888202.00170.710.1096.80 ± 0.5596.69 ± 0.6996.54 ± 0.59
HyperTransXNet0.241.08278.63707.06201.390.1398.46 ± 0.2998.25 ± 0.2298.33 ± 0.31
Table 5. Study of the input size.
Table 5. Study of the input size.
SizesIndianPinesWHU-HI-LongkouHouston2013
OA Kapp OA Kapp OA Kapp
11 × 11 97.43 ± 0.3497.07 ± 0.3999.25 ± 0.0899.01 ± 0.1198.78 ± 0.2798.69 ± 0.29
13 × 13 95.82 ± 0.5195.23 ± 0.5899.14 ± 0.1898.87 ± 0.2398.50 ± 0.3098.38 ± 0.33
15 × 15 94.74 ± 0.3993.99 ± 0.4498.91 ± 0.1698.57 ± 0.2198.46 ± 0.2998.33 ± 0.31
17 × 17 92.88 ± 0.6391.86 ± 0.7198.49 ± 0.2998.01 ± 0.3898.14 ± 0.3597.99 ± 0.38
19 × 19 91.12 ± 1.5389.85 ± 1.7498.27 ± 0.1397.72 ± 0.1797.58 ± 0.2397.38 ± 0.25
Table 6. Study of different modules.
Table 6. Study of different modules.
MethodsSSLBSSGBOA (%) κ (%)
2D-CNN××97.32 ± 0.4897.10 ± 0.51
HyperTransXNet×98.05 ± 0.43 (↑0.73%)97.89 ± 0.46 (↑0.79%)
HyperTransXNet×98.18 ± 0.25 (↑0.86%)98.04 ± 0.27 (↑0.94%)
Table 7. Study of SSTE.
Table 7. Study of SSTE.
MethodsSSTEOA (%) κ (%)
HyperTransXNet_SSLB_SSGB×98.16 ± 0.3298.01 ± 0.34
HyperTransXNet_SSLB_SSGB98.46 ± 0.29 (↑0.30%)98.33 ± 0.31 (↑0.32%)
Table 8. Study of SSTE and MoE-R.
Table 8. Study of SSTE and MoE-R.
MethodsSSTEMoE-ROA (%) κ (%)
HyperTransXNet_SSLB_SSGB×93.73 ± 1.5393.22 ± 1.65
HyperTransXNet_SSLB_SSGB×98.46 ± 0.29 (↑4.73%)98.33 ± 0.31 (↑5.11%)
Table 9. Impact of different expert counts for the OA and parameter efficiency.
Table 9. Impact of different expert counts for the OA and parameter efficiency.
Number of Experts (E)FLOPs (G)Param (MB)Parameter Efficiency
(Acc/Param)
Marginal Parameter Efficiency
( Δ Acc/ Δ Param)
OA (%)AA (%) κ (%)
20.210.9999.05-98.06 ± 0.4497.63 ± 0.3997.90 ± 0.48
40.241.0891.174.4498.46 ± 0.2998.25 ± 0.2298.33 ± 0.31
80.301.2677.75−2.7897.96 ± 0.4997.51 ± 0.5497.80 ± 0.53
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dai, X.; Li, Z.; Li, L.; Xue, S.; Huang, X.; Yang, X. HyperTransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Hyperspectral Image Classification. Remote Sens. 2025, 17, 2361. https://doi.org/10.3390/rs17142361

AMA Style

Dai X, Li Z, Li L, Xue S, Huang X, Yang X. HyperTransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Hyperspectral Image Classification. Remote Sensing. 2025; 17(14):2361. https://doi.org/10.3390/rs17142361

Chicago/Turabian Style

Dai, Xin, Zexi Li, Lin Li, Shuihua Xue, Xiaohui Huang, and Xiaofei Yang. 2025. "HyperTransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Hyperspectral Image Classification" Remote Sensing 17, no. 14: 2361. https://doi.org/10.3390/rs17142361

APA Style

Dai, X., Li, Z., Li, L., Xue, S., Huang, X., & Yang, X. (2025). HyperTransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Hyperspectral Image Classification. Remote Sensing, 17(14), 2361. https://doi.org/10.3390/rs17142361

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop