Next Article in Journal
Multi-Level Intertemporal Attention-Guided Network for Change Detection in Remote Sensing Images
Next Article in Special Issue
HyperTransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Hyperspectral Image Classification
Previous Article in Journal
Enhanced Detection of Permafrost Deformation with Machine Learning and Interferometric SAR Along the Qinghai–Tibet Engineering Corridor
Previous Article in Special Issue
PAMFPN: Position-Aware Multi-Kernel Feature Pyramid Network with Adaptive Sparse Attention for Robust Object Detection in Remote Sensing Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HG-Mamba: A Hybrid Geometry-Aware Bidirectional Mamba Network for Hyperspectral Image Classification

1
School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
2
School of Information Engineering, East China Jiaotong University, Nanchang 330044, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(13), 2234; https://doi.org/10.3390/rs17132234
Submission received: 24 May 2025 / Revised: 26 June 2025 / Accepted: 27 June 2025 / Published: 29 June 2025
(This article belongs to the Special Issue AI-Driven Hyperspectral Remote Sensing of Atmosphere and Land)

Abstract

Deep learning has demonstrated significant success in hyperspectral image (HSI) classification by effectively leveraging spatial–spectral feature learning. However, current approaches encounter three challenges: (1) high spectral redundancy and the presence of noisy bands, which impair the extraction of discriminative features; (2) limited spatial receptive fields inherent in convolutional operations; and (3) unidirectional context modeling that inadequately captures bidirectional dependencies in non-causal HSI data. To address these challenges, this paper proposes HG-Mamba, a novel hybrid geometry-aware bidirectional Mamba network for HSI classification. The proposed HG-Mamba synergistically integrates convolutional operations, geometry-aware filtering, and bidirectional state-space models (SSMs) to achieve robust spectral–spatial representation learning. The proposed framework comprises two stages. The first stage, termed spectral compression and discrimination enhancement, employs multi-scale spectral convolutions alongside a spectral bidirectional Mamba (SeBM) module to suppress redundant bands while modeling long-range spectral dependencies. The second stage, designated spatial structure perception and context modeling, incorporates a Gaussian Distance Decay (GDD) mechanism to adaptively reweight spatial neighbors based on geometric distances, coupled with a spatial bidirectional Mamba (SaBM) module for comprehensive global context modeling. The GDD mechanism facilitates boundary-aware feature extraction by prioritizing spatially proximate pixels, while the bidirectional SSMs mitigate unidirectional bias through parallel forward–backward state transitions. Extensiveexperiments on the Indian Pines, Houston2013, and WHU-Hi-LongKou datasets demonstrate the superior performance of HG-Mamba, achieving overall accuracies of 94.91%, 98.41%, and 98.67%, respectively.

Graphical Abstract

1. Introduction

Hyperspectral imaging (HSI) acquires hundreds of narrow spectral bands for each pixel, thereby enabling the precise identification of materials based on their distinct spectral signatures. This innovative technology has proven indispensable in remote sensing and has driven significant advancements in agricultural monitoring [1], mineral exploration [2], urban development [3], and environmental conservation [4]. The primary objective of hyperspectral image classification is to assign semantic labels to individual pixels, a task that critically depends on robust spectral–spatial feature extraction. However, the inherent challenges of HSI—including spectral redundancy, spatial heterogeneity, and the scarcity of labeled data—necessitate the development of sophisticated methodologies that effectively balance discriminative power with computational efficiency.
Deep learning has achieved remarkable success in HSI classification, particularly through Convolutional Neural Networks (CNNs). Early 2D-CNN approaches [5,6,7,8,9] processed spatial patches while treating spectral bands independently, thereby limiting the potential for joint spectral–spatial learning. To address this limitation, 3D-CNNs employing volumetric convolutions were introduced to concurrently model spectral and spatial dimensions. For example, Yang et al. [7] proposed a hybrid 2D-3D CNN to enhance feature fusion. However, these 3D-CNN-based methods [10,11] require extensive training data and remain computationally intensive. Additionally, although dilated convolutions expand receptive fields, they can introduce grid artifacts, and depthwise separable convolutions reduce parameter counts at the expense of feature richness. It is noted that CNNs treat all spatial neighbors uniformly, thereby disregarding the influence of geometric distance. This oversight becomes particularly significant in boundary regions, where distant pixels contribute minimally to the semantic context.
Transformers address the limited global context inherent in CNNs by incorporating self-attention mechanisms [12,13]. For example, SpectralFormer [12] groups spectral embeddings and fuses cross-layer features, while SSFTT [14] combines 3D-2D convolutions with transformer encoders to facilitate multi-scale learning. Nevertheless, these advancements come at the cost of quadratic computational complexity relative to sequence length [15], making transformers impractical for high-resolution hyperspectral images. Furthermore, MorphFormer [13] integrates morphological convolutions into the attention layers. However, this approach incurs prohibitive memory costs for large-scale datasets.
Recent state-space models (SSMs), particularly Mamba [16], have attracted attention in HSI classification due to their linear-time sequence modeling capabilities. Building on this framework, MambaHSI [17] pioneered a joint spatial–spectral modeling approach utilizing spatial and spectral Mamba blocks, while SpectralMamba [18] enhanced computational efficiency through dynamic masked convolutions. However, these methods exhibit three critical limitations: (1) unidirectional scanning that fails to capture the bidirectional dependencies inherent in non-causal HSI data; (2) static spatial weighting that does not adapt to variations in geometric distance; and (3) insufficient spectral compression, which can allow redundant bands to obscure discriminative features.
Based on the preceding discussion, although deep learning methods have demonstrated considerable success in HSI classification, current modeling strategies continue to exhibit three critical limitations:
1.
HSpectral redundancy: high inter-band correlation compounded by noise not only elevates computational costs but also impairs the discriminability of features.
2.
Spatial insensitivity: fixed convolutional kernels prove inadequate for modeling boundary regions, where spatial importance diminishes with increased distance.
3.
Unidirectional bias: Mamba’s causal state transitions truncate the contextual dependencies that are essential in bidirectional hyperspectral imaging data.
To address these challenges, we introduce HG-Mamba, a novel hybrid framework for HSI classification that leverages bidirectional state-space modeling and geometry-aware filtering. This framework employs a synergistic two-stage architecture to fundamentally enhance spectral–spatial feature learning for robust classification. In the spectral compression stage, redundant bands are suppressed using pointwise convolution and multi-scale spectral convolutions, while the bidirectional Spectral Mamba (SeBM) module captures both local and global spectral dependencies through parallel forward–backward state transitions, thereby overcoming the limitations of unidirectional state-space models (SSMs). In the subsequent spatial modeling stage, the framework incorporates a geometry-aware Gaussian Distance Decay (GDD) mechanism to adaptively reweight spatial neighbors based on geometric proximity, improving boundary fidelity. Additionally, the Spatial Bidirectional Mamba (SaBM) module models global context with linear complexity and is refined through Conv-Mamba residual blocks, which synergistically integrate convolutional locality with SSM-based globality. This work bridges the critical gap between local geometric sensitivity and global dependency learning, establishing a new paradigm for efficient and accurate HSI analysis.
The major contributions of this paper are summarized as follows:
1.
We propose the first framework that unifies convolutional operations, geometry-aware filtering, and bidirectional state-space models (SSMs) into a hierarchical architecture. This design enables joint extraction of local spectral details and global contextual dependencies while reducing computational complexity to linear time.
2.
We develop a sequential two-stage architecture, which comprises a bidirectional Spectral Mamba (SeBM) module for enhancing spectral discriminability and a bidirectional Spatial Mamba (SaBM) module for resolving spatial heterogeneity and capturing long-range dependencies.
3.
We also introduce geometry-aware Gaussian Distance Decay (GDD) to dynamically extract spatial feature, which is a novel mechanism that adaptively reweights spatial neighbors based on Euclidean distances.

2. Related Work

2.1. Convolution Neural Network-Based Methods for HSI Classification

Convolutional Neural Networks (CNNs) have been foundational in HSI classification [19,20,21,22], leveraging their hierarchical feature learning capabilities. Early 2D-CNN architectures processed spatial patches independently across spectral bands, but their separation of spatial and spectral dimensions limited joint feature learning. For example, Yang et al. [6] proposed a 2D-CNN for HSI classification, constructed by stacking three convolutional layers. Sharma et al. [23] designed a simple 2D-CNN-based method to identify hyperspectral scenes. However, such 2D-CNN approaches separately process spatial and spectral information, which limits their ability to capture joint spatial–spectral features. Three-dimensional CNNs [7,24] addressed this by applying volumetric convolutions to spectral–spatial cubes, such as the hybrid 2D/3D CNN by Yang et al. [7], which improved feature fusion but required substantial training data and suffered from high computational costs. Dilated convolutions [25,26] expanded receptive fields without increasing parameters, yet introduced grid artifacts, while static convolutional weights ignored spatial distance variations, leading to boundary blurring in heterogeneous regions [10,11]. For example, SSBLS [27] and SANet [28] incorporated distance-weighted mechanisms, but their static spatial weights lacked end-to-end adaptability, limiting performance in complex scenes. Moreover, to reduce the reliance on large annotated datasets, recent studies have explored lightweight architectures and semi-supervised learning strategies. For example, Xcep-Dense [29] introduces a lightweight model based on an extreme inception design, enabling efficient hyperspectral image classification under limited supervision.
While CNNs excel at local feature extraction, their inability to model global dependencies and long-range spatial–spectral interactions necessitated the adoption of alternative architectures.

2.2. Transformer-Based Methods for HSI Classification

Transformers have revolutionized HSI classification by enabling the modeling of global context through self-attention mechanisms [30,31]. Hong et al. [12] groups spectral embeddings and fuses cross-layer features to enhance spectral representations, while Roy et al. [13] integrates morphological convolutions into attention layers to capture structural details. In addition, Sun et al. [14] and Zhou et al. [32] combine convolutional and Transformer branches to facilitate multi-scale feature learning. However, these models suffer from quadratic computational complexity [15], rendering them impractical for large-scale HSI. For instance, processing high-resolution scenes, such as the Houston2013 image, incurs significant memory and computational overhead, thereby limiting scalability in real-world applications.
Despite their global modeling prowess, Transformers’ efficiency challenges inspired the exploration of linear-time sequence models, such as Mamba, for HSI classification.

2.3. Mamba-Based Methods for HSI Classification

Recently, state-space models (SSMs), particularly Mamba [16], have emerged as efficient alternatives with linear computational complexity. For example, Yao et al. [18] proposed the SpectralMamba model, which integrates dynamic masked convolution with state-space modeling to improve both classification performance and computational efficiency. Subsequently, Li et al. [17] introduced MambaHSI, the first model to achieve joint spatial–spectral modeling at the image level. Pan et al. [33] developed MambaLG, which sequentially integrates local and global spatial features (SpaM) and performs short- and long-range spectral dynamic perception (SpeM). A gated attention unit is also introduced to enhance global contextual modeling while preserving fine spatial details. Ahmad et al. [34] presented MorpMamba, which combines morphological operations with the Mamba architecture, achieving improved classification performance and higher parameter efficiency.
To address the challenges of spectral redundancy, spatial distance insensitivity, and unidirectional contextual truncation, we propose HG-Mamba, a hybrid bidirectional Mamba network that synergizes convolutional, geometric, and state-space modeling. HG-Mamba employs a two-stage hybrid architecture. The first stage (spectral compression and discriminative enhancement) uses pointwise convolution, multi-scale spectral convolutions, and a bidirectional Spectral Mamba (SeBM) module to reduce spectral redundancy, extract multi-scale spectral features, and model cross-band dependencies. The second stage (spatial structure perception and context modeling) incorporates a Gaussian Distance Decay (GDD) module for geometry-aware spatial weighting and a bidirectional Spatial Mamba (SaBM) module to capture long-range spatial dependencies, with residual blocks fusing local and global features.

3. Preliminaries

State-space models (SSMs) [35] serve as fundamental dynamic systems for sequence modeling, rooted in continuous-time linear systems. For an input sequence x ( t ) R M , the hidden state h ( t ) R N , and output y ( t ) R O . The general form can be written as:
h ( t ) = A h ( t ) + B x ( t )
y ( t ) = C h ( t ) + D x ( t )
where A R N × N is the state transition matrix, and B R N × M , C R O × N , and D R O × M represent the mappings from input to state, state to output, and input to output, respectively.
To facilitate training within deep learning frameworks, Equation (1) is commonly discretized into a discrete-time system with time step Δ using zero-order hold (ZOH). By reparameterizing matrices A and B , we obtain:
A ¯ = exp ( Δ A ) B ¯ = ( Δ A ) 1 ( A ¯ I ) ( Δ B ) ( Δ A ) 1 ( Δ A ) ( Δ B )
= Δ B            
Mamba enhances traditional SSMs by parameterizing B, C, and  Δ  as input-dependent functions, introducing adaptivity through a selective SSM (S6) framework. This design allows dynamic adjustment of internal mappings based on input characteristics, improving the model’s capacity to capture context-varying dependencies in sequences. By integrating input-conditioned gates and projections, Mamba achieves linear-time complexity while maintaining expressive power for long-range dependencies.
After discretization, the SSM takes the following recursive form:
h k = A ¯ h k 1 + B ¯ x k
y k = C h k + D x k
To further enhance the modeling capability of global context in visual sequences, Vision Mamba [36] introduces a Bidirectional Mamba Block, which models the input sequence from both forward and backward directions in parallel, as illustrated in Figure 1.
Given an input sequence T in R B × L × D , a linear projection is first applied to divide it into the state modeling branch and the residual branch:
[ x , z ] = Linear ( T in ) ,
where x is used for state modeling and z is reserved for residual connection.
Then, x is simultaneously fed into forward and backward modeling paths. The forward path uses causal convolution followed by a forward state-space model, while the backward path reverses the input, applies anti-causal modeling, and flips the output back:
y f = SSM f ( Conv 1 D f ( x ) ) ,
y b = Flip ( SSM b ( Conv 1 D b ( Flip ( x ) ) ) ) ,
The outputs from both directions are then fused as:
y bi = y f + y b ,
Then the result is projected back to the original dimension and added with residual input to yield the final output:
T out = W out y bi + T in .
This bidirectional architecture retains linear computational complexity while improving the model’s ability to capture non-causal context dependencies, thus enhancing performance in various vision tasks.

4. Proposed Approach

4.1. Overall Architecture

HG-Mamba is a hybrid bidirectional Mamba network developed to address challenges such as spectral redundancy, spatial heterogeneity, and unidirectional modeling in HSI classification. The framework employs a two-stage sequential pipeline (as illustrated in Figure 2). It integrates convolutional operations, geometry-aware filtering, and bidirectional state-space models (SSMs) to facilitate robust spectral–spatial feature learning. In Stage 1, termed spectral compression and discrimination enhancement, redundant spectral bands are suppressed while discriminative features are enhanced through multi-scale spectral convolutions and bidirectional SSMs. Stage 2, named spatial structure perception and context modeling, focuses on capturing geometry-aware spatial dependencies and long-range contextual relationships via a Gaussian Distance Decay (GDD) module and bidirectional spatial Mamba blocks.

4.2. Spectral Compression and Discrimination Enhancement

The first stage of HG-Mamba aims to reduce spectral redundancy and enhance discriminative spectral features that are relevant to the classification task. Given an input hyperspectral image patch X R B × C × P × P , where C denotes the number of spectral bands and P × P is the spatial patch size, a point-wise 1 × 1 convolution is first applied to compress the spectral dimension into a lower-dimensional embedding space of size d:
X 0 = ϕ ( BN ( Conv 1 × 1 ( X ) ) )
where X 0 R B × d × P × P , ϕ denotes the GELU activation function, and BN ( · ) represents the batch normalization operation.
To enhance the extraction of localized spectral details, the feature map X 0 is reshaped to R B × H × W × d and passed through two parallel spectral convolution branches. These branches apply 3D convolutions with kernel sizes of 3 × 1 × 1 and 7 × 1 × 1 , respectively, to capture contextual spectral dependencies at different receptive fields:
S 3 = ϕ ( LN ( Conv 3 × 1 × 1 ( X 0 ) ) )
S 7 = ϕ ( LN ( Conv 7 × 1 × 1 ( X 0 ) ) )
where ϕ denotes the GELU activation function, and LN ( · ) represents the layer normalization operation, as illustrated in Figure 3.
The outputs of both branches are fused with the original feature map by element-wise addition to obtain the enhanced spectral representation:
S = S 3 + S 7 + X 0
Next, global average pooling is performed along the spatial dimensions to obtain a compressed spectral representation S g R B × d × 1 × 1 . Then, the spectral dimension is partitioned into multiple overlapping sub-bands, where each sub-band has a width of s (i.e., spec_num) and a sliding stride of m (i.e., move_num). This results in a set of spectral sub-sequences { S ( i ) } i = 1 n , with each S ( i ) R B × s . The above parameter configuration follows the sliding sub-band partitioning strategy proposed by Pan et al. [33]. Specifically, we set the sub-band width s = 12 and the sliding stride m = 6 , corresponding to a coverage rate of c = 0.5 . A smaller s captures fine-grained details but may lose contextual information, while a larger s increases redundancy. A moderate window length balances local detail and global context. Similarly, a moderate coverage rate provides sufficient overlap between sub-bands to ensure robustness without incurring excessive computational cost. Based on the embedding dimension d, the number of sub-bands n is computed as:
n = d s m + 1
To model inter-sub-band contextual dependencies, we introduce the Spectral Bidirectional Mamba (SeBM) module. Given the sequence of spectral sub-bands { S ( i ) } i = 1 n generated earlier, SeBM performs bidirectional state-space modeling to capture global correlations among different sub-bands, thereby enhancing cross-band information integration. The entire process can be described as:
Z s = M SeBM ( { S ( i ) } ) R B × n × s
Next, we flatten Z s into R B × ( n · s ) , and feed it into a fully connected layer with LayerNorm and GELU activation to obtain a compact global spectral descriptor:
z = ϕ ( LN ( FC ( Z s ) ) ) R B × d
This descriptor is then reshaped to R B × d × 1 × 1 , and used to modulate the original spectral features X 0 through element-wise multiplication, forming a spectral attention mechanism:
X 1 = X 0 z
Such modulation enhances the response of spectrally informative components while suppressing noisy or redundant ones, thereby improving feature discriminability. The modulated features X 1 are then forwarded into the spatial modeling stage as input.

4.3. Spatial Structure Perception and Context Modeling

This stage is dedicated to modeling spatial dependencies to enhance the extraction of local textures and global contextual semantics. This stage takes the spectrally modulated feature map from Stage 1, X 1 R B × d × P × P , as input.

4.3.1. Multi-Scale Spatial Structure Extraction

To effectively capture multi-scale spatial structural information, we design a multi-branch convolutional feature extraction module to simultaneously model local texture details at different spatial scales. As illustrated in Figure 4, the input feature map X 1 R B × d × P × P is first passed through three parallel convolutional branches with varying receptive fields (i.e., kernel sizes of 1 × 1 , 3 × 3 , and 5 × 5 ). The small kernel helps extract fine-grained edge and texture features, while the larger kernels enhance regional semantics. These three branches produce intermediate features F 1 , F 3 , and F 5 R B × d × P × P , respectively. These are then concatenated with the original feature X 1 along the channel dimension to construct a multi-scale fused representation F cat R B × 4 d × P × P , with the concatenation operation defined as:
F cat = Concat ( X 1 , F 1 , F 3 , F 5 )

4.3.2. Gaussian Distance Decay (GDD)

To incorporate geometry-awareness into spatial modeling, we design a Gaussian distance-weighted spatial filtering mechanism, referred to as Gaussian Distance Decay (GDD). The GDD module is embedded within the fusion operation to enhance spatial interactions with distance sensitivity.
This module employs a learnable geometry-aware convolution kernel to assign distance-based weights to different spatial locations, thereby emphasizing structurally salient regions while suppressing weakly correlated areas. The core idea is as follows. The importance of a pixel in spatial structures is not solely determined by its content features. It is also influenced by its geometric distance from the central position. To this end, we use a learnable Gaussian kernel to reweight spatial responses based on relative distances, enabling a geometry-sensitive filtering operation.
Specifically, GDD first constructs a 2D Euclidean distance grid G ( i , j ) of size k × k , where each element represents the distance from pixel ( i , j ) to the kernel center:
G ( i , j ) = ( i c ) 2 + ( j c ) 2
where c = k 1 2 denotes the coordinate of the kernel center. Based on this distance map, GDD constructs a learnable Gaussian weighting kernel as follows:
W ( i , j ) = 1 Z exp G ( i , j ) 2 2 σ 2
Here, σ is a learnable smoothing parameter that controls the decay rate of spatial responses with respect to distance. Z is a normalization factor that ensures the sum of the kernel weights equals 1. To enable group-wise depthwise convolution, the Gaussian kernel is expanded to shape [ C , 1 , k , k ] , so that each channel uses an independent geometric filter.
After constructing the Gaussian kernel, we apply it in a depthwise convolutional manner to reweight spatial features. Let F cat R B × 4 d × P × P denote the concatenated multi-scale features obtained from the previous convolutional branches. These features are first compressed back to the original embedding dimension d via a 1 × 1 convolution:
F proj = Conv 1 × 1 ( F cat ) R B × d × P × P
Then, the geometry-aware GDD filtering is applied to F proj as a channel-wise depthwise convolution:
F gdd = GDD ( F proj ) R B × d × P × P
Finally, the filtered feature F gdd is normalized and activated to obtain the fused spatial representation:
F fused = ϕ ( BN ( F gdd ) ) R B × d × P × P
Here, ϕ ( · ) denotes the GELU activation function and BN ( · ) is batch normalization. This fusion process integrates multi-scale spatial features and enhances geometry-aware representation for subsequent contextual modeling.

4.3.3. Hierarchical Refinement with Spatial Bidirectional Mamba

To further capture long-range spatial dependencies, we flatten the fused features F fused along spatial dimensions into a sequence T R B × ( P · P ) × d , which is then fed into an SaBM module to enable bidirectional sequence modeling:
T = M SaBM ( T ) R B × ( P · P ) × d
The output sequence is reshaped back to its original spatial format:
F ctx = Reshape ( T ) R B × d × P × P
To further refine the hierarchical spatial representations, the context-enhanced feature F ctx is passed through a stack of L Conv-Mamba residual blocks. In each block, local features are extracted using a convolutional layer followed by GDD, batch normalization, and GELU activation. These locally refined features are then passed into a SaBM module to capture long-range spatial dependencies. A residual connection integrates the Mamba-enhanced output with the input, forming the final representation X 2 after N residual refinements.
The final output feature map X 2 R B × d × P × P is globally pooled and fed into a linear classifier to produce the final predictions. This process is formulated as:
y ^ = FC ( GAP ( X 2 ) ) R B × N c
where GAP ( · ) denotes the global average pooling operation, FC ( · ) is a fully connected layer, N c is the number of target classes, and y ^ represents the predicted class logits.

5. Experiments

5.1. Datasets and Setting

Datasets. Three HSI datasets are utilized in our experiments, including the Indian Pines scene, Houston2013, and WHU-Hi-LongKou (WHL) datasets.
  • Indian Pines Scene: This HSI dataset was acquired in 1992 by the Airborne Visible Imaging Spectrometer (AVIRIS) instrument over a mixed agricultural and forest area in Northwestern Indiana, USA. The original imagery comprised 220 spectral bands. After preprocessing that involved the removal of 20 noisy bands, 200 spectral bands were retained for analysis. The spatial dimensions of the image are 145 × 145 pixels. The scene includes 16 distinct land cover classes, encompassing a variety of agricultural and natural surface types. The complete set of classes is as follows: Alfalfa, Corn-notill, Corn-mintill, Corn, Grass-pasture, Grass-trees, Grass-pasture-mowed, Hay-windrowed, Oats, Soybean-notill, Soybean-mintill, Soybean-clean, Wheat, Woods, Buildings-Grass-Trees-Drives, and Stone-Steel-Towers.
    For the purpose of training and testing, a subset constituting 10% of the total labeled samples was allocated for model training, with the remaining 90% designated for performance evaluation.
  • Houston2013 Dataset: This dataset captures an urban scene covering the University of Houston and its vicinities in Texas, USA. The data was collected using the ITRES CASI-1500 sensor. The imagery provides spatial dimensions of 349 × 1905 pixels and contains 144 spectral band. Provided as a cloud-free image by the Geo-science and Remote Sensing Society (GRSS), it serves as a standard benchmark. The dataset is composed of a total of 15 labeled land cover classes. The complete set of classes includes: Grass-healthy, Grass-stressed, Grass-synthetic, Tree, Soil, Water, Residential, Commercial, Road, Highway, Railway, Parking Lot 1, Parking Lot 2, Tennis Court, and Running Track.
    In our experimental setup, 10% of the available samples were selected to form the training set, while the remaining samples constituted the test set.
  • WHU-Hi-LongKou (WHL) Dataset: Acquired on 17 July 2018, this dataset covers Longkou Town, Hubei province, China. Data collection was performed via a UAV platform (DJI Matrice 600 Pro) equipped with a Headwall Nano-Hyperspec imaging sensor featuring an 8 mm focal length. The UAV flew at an altitude of 500 m, resulting in imagery with a spatial resolution of 550 × 400 pixels. The spectral range of the sensor spans from 400 to 1000 nm, capturing data across 270 bands. The dataset comprises 204,542 labeled samples distributed among nine distinct land cover classes. The complete set of classes includes: Corn, Cotton, Sesame, Broad-leaf soybean, Narrow-leaf soybean, Rice, Water, Roads and houses, and Mixed weed.
    In this study, a small fraction, specifically 1%, of the labeled samples was utilized for training, with the predominant portion (99%) reserved for testing.
Evaluation Metrics. To quantify the performance of the classification methods, three standard metrics were employed: Overall Accuracy (OA), Average Accuracy (AA), and the Kappa Coefficient ( κ ).
Comparison Methods. We conducted comprehensive comparisons with state-of-the-art methods across different architectural paradigms: CNN-based (e.g., 2D-CNN [6], 3D-CNN [7], HybridSN [37]), Transformer-based (e.g., ViT [38], DeepViT [39], CvT [40], HiT [41], MorphFormer [13], SSTMNet [42], DCTN [32], and SSFTT [14]), and SSM-based (MambaHSI [17], S 2 Mamba [43]).
Setting. For the training process, 100 patches with spatial dimensions of 15 × 15 pixels were randomly cropped and utilized as input samples. The training procedure was executed for 100 iterations. All comparison methods and the proposed HG-Mamba model were implemented within the PyTorch 2.1.1 framework. The HG-Mamba model’s parameters were updated using the Adam optimizer. Furthermore, the learning rate was configured to 1 × 10−3, and the batch size was set at 100.

5.2. Results and Analysis

In this section, we provide a comprehensive analysis of the experimental results obtained from the Indian Pines, Houston2013, and WHU-Hi-LongKou (WHL) benchmark datasets, as presented in Table 1, Table 2 and Table 3. Our analysis focuses on evaluating the overall performance of the proposed HG-Mamba against state-of-the-art comparison methods from CNN-based, Transformer-based, and SSM-based categories, as well as examining per-class accuracies and model robustness.
As shown in Table 1, Table 2 and Table 3, the proposed HG-Mamba consistently achieves the highest Overall Accuracy (OA), Average Accuracy (AA), and Kappa Coefficient ( κ ) across all three benchmark datasets. On the Indian Pines dataset (10% training samples, Table 1), HG-Mamba achieves an OA of 94.91%, an AA of 90.83%, and a κ of 94.18%, outperforming all comparison methods in terms of overall and average accuracy. Similarly, for the Houston2013 dataset (Table 2), HG-Mamba leads with an OA of 98.41%, AA of 98.07%, and κ of 98.28%. The performance gains are particularly significant on the challenging WHL dataset, which uses only 1% of training samples (Table 3). Here, HG-Mamba demonstrates superior capability in learning from extremely limited data, achieving an OA of 98.67%, AA of 95.61%, and κ of 98.25%. These results strongly validate the effectiveness of the proposed HG-Mamba framework.
Comparing HG-Mamba with methods from different architectural paradigms reveals the advantages of our hybrid design. CNN-based methods (2D-CNN, 3D-CNN, HybridSN) generally show lower overall accuracies compared to Transformer and SSM-based methods, highlighting their limitation in capturing global context. Transformer-based methods, leveraging self-attention for long-range dependencies, generally perform well, with recent architectures like DCTN and SSFTT serving as strong baselines. However, HG-Mamba still surpasses these methods, suggesting that our combination of convolutional local feature extraction with efficient SSM-based global modeling is more effective for HSI classification. The SSM-based MambaHSI, while efficient, is consistently outperformed by HG-Mamba, indicating that the bidirectional modeling and geometry-aware components in our architecture contribute significantly beyond a basic spatial–spectral Mamba model.
Furthermore, we analyze the per-class classification accuracies to assess the model’s performance on individual land cover categories, particularly those with limited training samples. As seen in Table 1, classes like ‘Grass-pasture-mowed’ and ‘Oats’ have very few labeled samples, leading to significantly lower accuracies and high standard deviations for many methods (some even achieving 0% accuracy). HG-Mamba demonstrates remarkably better performance on these challenging classes (e.g., 83.60% for ’Grass-pasture-mowed’, 90.00% for ’Oats’) compared to most baselines, indicating its robustness and ability to handle class imbalance and learn discriminative features even from scarce examples. This robustness is also reflected in the consistently lower standard deviations of OA, AA, and κ for HG-Mamba across all datasets, suggesting higher stability in performance compared to several comparison methods.
The superior performance of HG-Mamba, corroborated by the visually clean and highly faithful classification maps (Figure 5, Figure 6 and Figure 7), can be attributed to the synergistic integration of its key components, which effectively combine local details and global representations. Specifically, the spectral compression and Spectral Bidirectional Mamba efficiently handle high spectral dimensionality and capture long-range spectral dependencies, enhancing discriminability. The multi-scale spatial convolutions extract hierarchical local features, while the GDD module accurately models spatial heterogeneity and boundary regions. Furthermore, the Spatial Bidirectional Mamba and Conv-Mamba residual blocks effectively capture global spatial context and refine spatial representations with linear complexity, overcoming limitations of unidirectional SSMs and computationally expensive attention mechanisms. In contrast, the classification maps produced by most comparison methods exhibit significant noise levels (Figure 5, Figure 6 and Figure 7), a finding consistent with the quantitative results (Table 1, Table 2 and Table 3). This can be explained by the fact that CNNs primarily focus on local features and overlook global representations, while Transformer models, despite their ability to model global features, typically suffer from high computational complexity. Although SSM-based models (like MambaHSI) offer an efficient alternative for long-range dependencies, they may lack proficiency in handling complex spatial details and boundaries. These limitations highlight that simply relying on a single architecture or limited fusion mechanisms may be insufficient for generating high-quality classification maps. Therefore, the ability to effectively and efficiently balance the extraction of local details and global contextual modeling is crucial, which is precisely the advantage of HG-Mamba’s hybrid architecture, clearly demonstrated in both its state-of-the-art performance and the smoother, more ground-truth-faithful classification maps it generates.
We also conducted a t-SNE visualization comparative analysis of the proposed HG-Mamba and five comparison methods, including CNN-based (e.g., 2D-CNN and 3D-CNN), Transformer-based (e.g., SSFTT and MorphFormer), and SSM-based (e.g., MambaHSI). The visualization results of different methods on the Indian Pines dataset are shown in Figure 8. According to Figure 8, we find that HG-Mamba can reduce inter-class confusion and enhance intra-class clustering, which is mainly due to HG-Mamba’s powerful feature learning ability; its hybrid convolutional–state-space model (Conv-SSM) structure effectively integrating local details and global representations, capturing more discriminative spatial–spectral features.

5.3. Comparison of Computational Complexity

Table 4 presents a comparison of the computational complexity (FLOPs and parameters), runtime (training and testing time), and classification performance of the proposed HG-Mamba and five representative comparison methods on the Indian Pines dataset.
As shown in Table 4, HG-Mamba achieves state-of-the-art classification accuracy in terms of OA, AA, and κ . Importantly, compared to Transformer-based methods such as Deep ViT (52.75 MB, 13.69 G) and HiT (27.22 MB, 2.20 G), HG-Mamba has significantly lower parameters (1.30 MB) and FLOPs (0.57 G), which is consistent with the theoretical efficiency advantages of SSMs. Compared to CNN-based methods (2D-CNN, 3D-CNN), HG-Mamba has competitive or slightly lower parameters, although its FLOPs are in a similar range to 2D-CNN but lower than 3D-CNN. While 2D-CNN and 3D-CNN offer faster training and testing times, HG-Mamba provides higher accuracy, especially in terms of AA. MambaHSI stands out with its extremely low FLOPs (0.006 G) and parameters (0.02 MB), demonstrating the significant advantages of SSMs in terms of model size and theoretical computation. However, HG-Mamba achieves significantly higher classification accuracy than MambaHSI, indicating that its hybrid architecture and specific components make important contributions to improving performance, despite the computational cost being slightly higher than MambaHSI itself. The training and testing times for HG-Mamba are moderate, faster than HiT but slower than CNNs and Deep ViT. This suggests a certain trade-off between speed and accuracy. Overall, the results show that HG-Mamba achieves an excellent balance between computational efficiency and classification performance, realizing state-of-the-art accuracy with a favorable complexity profile (especially in terms of model size and theoretical operations).

6. Ablation Studies

6.1. Patch Size Sensitivity Analysis

Figure 9 illustrates the sensitivity of HG-Mamba’s classification performance (OA, AA, κ ) to varying patch sizes ( 9 × 9 to 17 × 17 ) across three datasets. The optimal patch size is dataset-dependent, reflecting their distinct spatial characteristics. For Indian Pines (Figure 9, left), smaller patches yield better results, with performance decreasing as size increases. In contrast, Houston2013 (Figure 9, middle) and WHL (Figure 9, right) show performance peaking at moderate patch sizes (around 13 and 11, respectively). Across all datasets, performance consistently degrades with excessively large patches, suggesting that overly wide spatial context introduces noise. This highlights the importance of empirically selecting the appropriate patch size for optimal performance based on dataset characteristics.

6.2. Training Sample Ratio Sensitivity Analysis

Figure 10 illustrates the classification performance (OA, AA, and κ ) of HG-Mamba as the training sample ratio increases. As expected, performance generally improves with increased training data, but the rate of improvement and saturation point vary by dataset. For Indian Pines (Figure 10, left), OA and κ show improvement, but AA exhibits a counter-intuitive trend, decreasing as the training ratio increases. This behavior in AA may be attributed to the severe class imbalance in the Indian Pines dataset; while the model learns better on more abundant classes with increased data, its performance on certain difficult or extremely sparse classes might not improve proportionally or could even slightly decrease relative to its overall learning focus, thus affecting the average per-class accuracy.
For Houston2013 (Figure 10, middle), all metrics show an initial sharp upward trend, followed by relatively quick saturation. On the challenging WHL dataset (Figure 10, right), the proposed model also shows some improvement, which highlights the model’s data efficiency.

6.3. Ablation Study of Different Modules

To analyze the contribution of each proposed module to the overall performance of HG-Mamba, we conducted an ablation study. Table 5 presents the classification performance (OA, AA, and κ ) of the model when different modules or stages are removed from the full architecture on the Indian Pines dataset.
As shown in Table 5, removing any module or stage results in a degradation of classification performance, indicating that all components contribute to HG-Mamba’s effectiveness. Removing Stage 2 (spatial structure perception and context modeling) leads to the most significant performance drop across all metrics (OA: 1.93%↓, AA: 9.37%↓, κ : 2.20%↓), highlighting the critical importance of the spatial processing stage. Removing Stage 1 (spectral compression and discrimination enhancement) also causes a substantial decrease (OA: 0.99%↓, AA: 4.96%↓, κ : 1.13%↓), underscoring the necessity of effective spectral processing.
Analyzing individual modules, removing the GDD module results in a notable drop in performance, particularly in AA (1.93%↓), confirming its role in capturing geometry-aware spatial features and enhancing per-class accuracy. The removal of spectral compression also leads to a considerable decrease (OA: 0.90%↓, AA: 4.65%↓), emphasizing its importance in reducing redundancy and improving feature discriminability. Removing the spectral bidirectional Mamba causes a drop (OA: 0.38%↓, AA: 3.58%↓), demonstrating the contribution of bidirectional spectral dependency modeling. The Spatial Bidirectional Mamba module, while contributing to global spatial context, shows the smallest impact on OA (0.07%↓) when removed, although its removal still affects AA (2.94%↓) and κ (0.08%↓).
These results quantitatively validate the functional significance of each component within the HG-Mamba framework.

6.4. Analysis of Network Depth

The performance of deep learning models is often sensitive to network depth. Figure 11 illustrates the classification performance (OA, AA, and κ ) of HG-Mamba with varying numbers of stacked layers (Depth) on the Indian Pines, Houston2013, and WHL datasets.
As depicted in Figure 11, the optimal network depth varies across the datasets, reflecting their distinct characteristics. For the Indian Pines dataset (Figure 11, left), OA and κ generally improve with increasing depth, peaking at Depth 4 before slightly declining. AA shows a more volatile trend with larger standard deviations, reaching its highest mean value at Depth 6. The performance suggests that increasing depth up to a certain point enhances feature learning for this dataset, but excessive depth or specific architectural interactions at deeper levels might introduce instability or reduce gains, particularly affecting per-class accuracy as seen in the drop at Depth 5.
On the Houston2013 dataset (Figure 11, middle), all performance metrics consistently improve as the network depth increases from 1 to 6. The absence of performance degradation within this range suggests that for this urban dataset, deeper models are more effective at capturing the complex spatial–spectral patterns required for classification, and performance may continue to improve or saturate at even greater depths.
For the challenging WHL dataset (Figure 11, right), OA, AA, and κ generally improve with increasing depth, reaching peak performance around Depth 5, followed by a slight decrease at Depth 6. This indicates that a moderate depth is optimal for this dataset, providing sufficient capacity to learn from limited training data while avoiding potential overfitting or diminished returns from overly deep architectures.
Overall, the analysis reveals that while increasing network depth is generally beneficial across datasets, the optimal depth is dataset-dependent.

6.5. Limitations and Future Directions

Although HG-Mamba achieves impressive performance in hyperspectral image classification, two aspects of its current design warrant further improvement. First, the spectral sub-band partitioning strategy is manually defined with fixed hyperparameters, lacking adaptability to varying data distributions and task requirements. This rigidity may degrade performance when generalizing across scenes. Future work could incorporate learnable or data-driven partitioning schemes that dynamically optimize sub-band grouping according to spectral characteristics, thereby enhancing model flexibility and generalization. Second, the proposed Gaussian Distance Decay (GDD) module introduces spatial priors via a Gaussian kernel and effectively models local spatial continuity, yet it relies on a fixed Euclidean distance and applies identical decay weights across all channels, limiting its capacity to capture semantic boundaries. Subsequent research may explore more expressive spatial modeling strategies such as channel aware Gaussian kernels, anisotropic distance metrics, or attention-based spatial weighting to better encode complex spatial semantics in heterogeneous regions.

7. Conclusions

In this paper, we address the challenges of spectral redundancy, spatial distance insensitivity, and unidirectional contextual truncation in HSI classification by introducing HG-Mamba, a hybrid geometry-aware bidirectional Mamba network. The proposed framework employs a two-stage architecture that integrates convolutional operations, bidirectional state-space models (SSMs), and geometry-aware filtering. In Stage 1 (spectral compression and discrimination enhancement), multi-scale spectral convolutions and a Spectral Bidirectional Mamba (SeBM) module suppress redundant bands while modeling long-range inter-band dependencies, thereby enhancing spectral discriminability. In Stage 2 (spatial structure perception and context modeling), a Gaussian Distance Decay (GDD) module adaptively reweights spatial neighbors based on geometric proximity to mitigate boundary heterogeneity, and a Spatial Bidirectional Mamba (SaBM) module captures global spatial dependencies with linear computational complexity. Residual blocks further refine the features by fusing local convolutional representations with global SSM-based information. Extensive experiments on three benchmark datasets demonstrate that the proposed HG-Mamba outperforms CNNs, Transformers, and previous Mamba variants (e.g., MambaHSI). Future research will investigate extending the framework to real-time applications and multi-modal remote sensing data, as well as exploring its scalability to ultra-high-resolution HSI.

Author Contributions

Software, S.X.; Data curation, L.L. and S.X.; Writing—original draft, J.Y.; Writing—review & editing, X.Y. and H.T.; Visualization, L.L. and S.X.; Funding acquisition, H.S. and X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (NSFC) Fund under Grant 62301174 and 62462031, and Guangzhou basic and applied basic research topics under Grant 2024A04J2081 and Grant 2025A04J3375. This work was also supported by the Natural Science Foundation of Jiangxi Province under Grant 20242BAB26023.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bhadra, S.; Sagan, V.; Sarkar, S.; Braud, M.; Mockler, T.C.; Eveland, A.L. PROSAIL-Net: A transfer learning-based dual stream neural network to estimate leaf chlorophyll and leaf angle of crops from UAV hyperspectral images. ISPRS J. Photogramm. Remote Sens. 2024, 210, 1–24. [Google Scholar] [CrossRef]
  2. Hong, D.; Yokoya, N.; Chanussot, J.; Zhu, X.X. An augmented linear mixing model to address spectral variability for hyperspectral unmixing. IEEE Trans. Image Process. 2018, 28, 1923–1938. [Google Scholar] [CrossRef] [PubMed]
  3. Guo, J.; Hong, D.; Zhu, X.X. High-resolution satellite images reveal the prevalent positive indirect impact of urbanization on urban tree canopy coverage in South America. Landsc. Urban Plan. 2024, 247, 105076. [Google Scholar] [CrossRef]
  4. Ali, M.A.; Lyu, X.; Ersan, M.S.; Xiao, F. Critical evaluation of hyperspectral imaging technology for detection and quantification of microplastics in soil. J. Hazard. Mater. 2024, 476, 135041. [Google Scholar] [CrossRef] [PubMed]
  5. He, X.; Chen, Y. Optimized input for CNN-based hyperspectral image classification using spatial transformer network. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1884–1888. [Google Scholar] [CrossRef]
  6. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.; Zhang, X.; Huang, X. Hyperspectral image classification with deep learning models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  7. Yang, X.; Zhang, X.; Ye, Y.; Lau, R.Y.; Lu, S.; Li, X.; Huang, X. Synergistic 2D/3D convolutional neural network for hyperspectral image classification. Remote Sens. 2020, 12, 2033. [Google Scholar] [CrossRef]
  8. Mou, L.; Zhu, X.X. Learning to pay attention on spectral domain: A spectral attention module-based convolutional network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 110–122. [Google Scholar] [CrossRef]
  9. Sun, W.; Du, Q. Hyperspectral band selection: A review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 118–139. [Google Scholar] [CrossRef]
  10. Tu, C.; Liu, W.; Jiang, W.; Zhao, L. Hyperspectral image classification based on residual dense and dilated convolution. Infrared Phys. Technol. 2023, 131, 104706. [Google Scholar] [CrossRef]
  11. Ye, Z.; Wang, J.; Bai, L. Multi-Scale Spatial-Spectral Feature Extraction Based on Dilated Convolution for Hyperspectral Image Classification. In Proceedings of the 2023 6th International Conference on Image and Graphics Processing, Chongqing, China, 6–8 January 2023; pp. 97–103. [Google Scholar]
  12. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  13. Roy, S.K.; Deria, A.; Shah, C.; Haut, J.M.; Du, Q.; Plaza, A. Spectral–spatial morphological attention transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  14. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  15. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  16. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  17. Li, Y.; Luo, Y.; Zhang, L.; Wang, Z.; Du, B. MambaHSI: Spatial-spectral mamba for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5524216. [Google Scholar] [CrossRef]
  18. Yao, J.; Hong, D.; Li, C.; Chanussot, J. Spectralmamba: Efficient mamba for hyperspectral image classification. arXiv 2024, arXiv:2404.08489. [Google Scholar]
  19. Lee, H.; Kwon, H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef]
  20. Xu, Y.; Du, B.; Zhang, L. Self-attention context network: Addressing the threat of adversarial attacks for hyperspectral image classification. IEEE Trans. Image Process. 2021, 30, 8671–8685. [Google Scholar] [CrossRef]
  21. Wang, Z.; Chen, B.; Lu, R.; Zhang, H.; Liu, H.; Varshney, P.K. FusionNet: An unsupervised convolutional variational network for hyperspectral and multispectral image fusion. IEEE Trans. Image Process. 2020, 29, 7565–7577. [Google Scholar] [CrossRef]
  22. Zhao, C.; Zhu, W.; Feng, S. Superpixel guided deformable convolution network for hyperspectral image classification. IEEE Trans. Image Process. 2022, 31, 3838–3851. [Google Scholar] [CrossRef] [PubMed]
  23. Sharma, V.; Diba, A.; Tuytelaars, T.; Van Gool, L. Hyperspectral CNN for Image Classification & Band Selection, with Application to Face Recognition; Technical Report KUL/ESAT/PSI/1604; KU Leuven; ESAT: Leuven, Belgium, 2016. [Google Scholar]
  24. Ahmad, M.; Khan, A.M.; Mazzara, M.; Distefano, S.; Ali, M.; Sarfraz, M.S. A fast and compact 3-D CNN for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar] [CrossRef]
  25. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  26. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  27. Zhao, G.; Wang, X.; Kong, Y.; Cheng, Y. Spectral-spatial joint classification of hyperspectral image based on broad learning system. Remote Sens. 2021, 13, 583. [Google Scholar] [CrossRef]
  28. Wei, Y.; Zhou, Y. Spatial-aware network for hyperspectral image classification. Remote Sens. 2021, 13, 3232. [Google Scholar] [CrossRef]
  29. Ranjan, P.; Girdhar, A. Xcep-Dense: A novel lightweight extreme inception model for hyperspectral image classification. Int. J. Remote Sens. 2022, 43, 5204–5230. [Google Scholar] [CrossRef]
  30. Tu, B.; Liao, X.; Li, Q.; Peng, Y.; Plaza, A. Local semantic feature aggregation-based transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  31. Yang, X.; Cao, W.; Tang, D.; Zhou, Y.; Lu, Y. ACTN: Adaptive Coupling Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5503115. [Google Scholar] [CrossRef]
  32. Zhou, Y.; Huang, X.; Yang, X.; Peng, J.; Ban, Y. DCTN: Dual-branch convolutional transformer network with efficient interactive self-attention for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  33. Pan, Z.; Li, C.; Plaza, A.; Chanussot, J.; Hong, D. Hyperspectral Image Classification with Mamba. IEEE Trans. Geosci. Remote Sens. 2025, 63, 1–14. [Google Scholar] [CrossRef]
  34. Ahmad, M.; Butt, M.H.F.; Khan, A.M.; Mazzara, M.; Distefano, S.; Usama, M.; Roy, S.K.; Chanussot, J.; Hong, D. Spatial–spectral morphological mamba for hyperspectral image classification. Neurocomputing 2025, 636, 129995. [Google Scholar] [CrossRef]
  35. Kalman, R.E. A new approach to linear filtering and prediction problems. J. Basic Eng. Mar. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  36. Zhu, L.; Liao, B.; Zhang, Q.; Wang, X.; Liu, W.; Wang, X. Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model. arXiv 2024, arXiv:2401.09417. [Google Scholar]
  37. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  38. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  39. Zhou, D.; Kang, B.; Jin, X.; Yang, L.; Lian, X.; Jiang, Z.; Hou, Q.; Feng, J. Deepvit: Towards deeper vision transformer. arXiv 2021, arXiv:2103.11886. [Google Scholar]
  40. Wu, H.; Xiao, B.; Codella, N.; Liu, M.; Dai, X.; Yuan, L.; Zhang, L. Cvt: Introducing convolutions to vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 22–31. [Google Scholar]
  41. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral image transformer classification networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  42. Huang, X.; Zhou, Y.; Yang, X.; Zhu, X.; Wang, K. Ss-tmnet: Spatial–spectral transformer network with multi-scale convolution for hyperspectral image classification. Remote Sens. 2023, 15, 1206. [Google Scholar] [CrossRef]
  43. Wang, G.; Zhang, X.; Peng, Z.; Zhang, T.; Jiao, L. S2mamba: A spatial-spectral state space model for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5511413. [Google Scholar] [CrossRef]
Figure 1. Bidirectional Mamba block architecture with forward and backward SSM.
Figure 1. Bidirectional Mamba block architecture with forward and backward SSM.
Remotesensing 17 02234 g001
Figure 2. Overview of the proposed HG-Mamba framework. Here, c denotes the number of original spectral bands, p is the spatial patch size, d is the spectral embedding dimension, s is the sub-band width, m is the sub-band sliding stride, and N is the number of stacked Conv-Mamba residual blocks in the spatial stage.
Figure 2. Overview of the proposed HG-Mamba framework. Here, c denotes the number of original spectral bands, p is the spatial patch size, d is the spectral embedding dimension, s is the sub-band width, m is the sub-band sliding stride, and N is the number of stacked Conv-Mamba residual blocks in the spatial stage.
Remotesensing 17 02234 g002
Figure 3. Visualization of spectral compression and multi-scale spectral convolution in Stage 1.
Figure 3. Visualization of spectral compression and multi-scale spectral convolution in Stage 1.
Remotesensing 17 02234 g003
Figure 4. Architecture of the multi-branch spatial feature extraction module, where C denotes the concatenation of features from different branches along the channel dimension.
Figure 4. Architecture of the multi-branch spatial feature extraction module, where C denotes the concatenation of features from different branches along the channel dimension.
Remotesensing 17 02234 g004
Figure 5. Classification maps obtained by different methods on the Indian Pines Scene dataset (with 10% training samples).
Figure 5. Classification maps obtained by different methods on the Indian Pines Scene dataset (with 10% training samples).
Remotesensing 17 02234 g005
Figure 6. Classification maps obtained by different methods on the Houston2013 dataset (with 10% training samples).
Figure 6. Classification maps obtained by different methods on the Houston2013 dataset (with 10% training samples).
Remotesensing 17 02234 g006aRemotesensing 17 02234 g006b
Figure 7. Classification maps obtained by different methods on the WHL dataset (with 1% training samples).
Figure 7. Classification maps obtained by different methods on the WHL dataset (with 1% training samples).
Remotesensing 17 02234 g007
Figure 8. The t-SNE results obtained by different methods on the Indian Pines dataset (with 10% training samples).
Figure 8. The t-SNE results obtained by different methods on the Indian Pines dataset (with 10% training samples).
Remotesensing 17 02234 g008
Figure 9. Impact of patch size on classification performance across datasets.
Figure 9. Impact of patch size on classification performance across datasets.
Remotesensing 17 02234 g009
Figure 10. Impact of training sample ratio on classification performance across datasets.
Figure 10. Impact of training sample ratio on classification performance across datasets.
Remotesensing 17 02234 g010
Figure 11. Impact of network depth on classification performance across datasets.
Figure 11. Impact of network depth on classification performance across datasets.
Remotesensing 17 02234 g011
Table 1. Comparisons with the state-of-the-art models on the Indian Pines Scene dataset (10% training samples).
Table 1. Comparisons with the state-of-the-art models on the Indian Pines Scene dataset (10% training samples).
Class2D-CNN3D-CNNHybridsnViTDeep ViTCvTHiTSSFTTMorphFormerSS_TMNetDCTNMambaHSIS2MambaHG-Mamba
Alfalfa92.82 ± 4.8069.95 ± 17.9218.85 ± 29.7980.76 ± 5.7481.65 ± 10.8448.74 ± 15.9714.63 ± 19.6889.65 ± 6.9182.13 ± 22.4087.48 ± 8.1572.29 ± 11.0454.15 ± 14.9110.73 ± 11.9194.15 ± 4.39
Corn-notill93.81 ± 1.9788.61 ± 1.1784.84 ± 11.2294.29 ± 0.7794.14 ± 1.0691.40 ± 2.2490.74 ± 0.8994.11 ± 1.0893.38 ± 2.1488.56 ± 2.3495.70 ± 1.5789.07 ± 2.2284.01 ± 3.2992.61 ± 3.28
Corn-mintill92.19 ± 1.7786.74 ± 1.9075.93 ± 18.4093.63 ± 0.5892.86 ± 1.3085.23 ± 4.8986.88 ± 1.9090.13 ± 2.6691.28 ± 3.6676.50 ± 3.2388.93 ± 1.4089.10 ± 6.2390.19 ± 4.6497.32 ± 1.85
Corn97.94 ± 1.5093.92 ± 2.4480.93 ± 17.0199.62 ± 0.2899.22 ± 1.2292.32 ± 4.4079.81 ± 11.1394.90 ± 3.4695.26 ± 4.0482.19 ± 3.9292.77 ± 2.7791.60 ± 5.3193.43 ± 6.3297.70 ± 2.94
Grass-pasture93.09 ± 3.3293.44 ± 0.7073.56 ± 16.4392.00 ± 1.1189.45 ± 2.9188.40 ± 2.4083.91 ± 9.2493.08 ± 2.5294.72 ± 1.5581.71 ± 3.4992.04 ± 1.2887.38 ± 7.6775.82 ± 20.8495.54 ± 2.30
Grass-trees95.65 ± 2.9794.82 ± 0.6775.90 ± 15.7590.79 ± 1.4190.74 ± 2.8593.92 ± 1.0294.06 ± 0.6395.98 ± 1.2995.47 ± 1.3897.76 ± 0.9298.64 ± 0.6893.21 ± 2.3091.54 ± 2.3695.95 ± 5.36
Grass-pasture-mowed7.94 ± 18.290.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.008.08 ± 22.040.00 ± 0.0054.69 ± 35.8471.56 ± 18.0172.09 ± 16.3073.76 ± 9.760.00 ± 0.000.00 ± 0.0083.60 ± 28.65
Hay-windrowed99.69 ± 0.5598.87 ± 1.1087.72 ± 13.0999.77 ± 0.2299.79 ± 0.1799.31 ± 0.81100.00 ± 0.0098.77 ± 1.3599.81 ± 0.3994.39 ± 0.4797.75 ± 0.9599.79 ± 0.63100.00 ± 0.00100.00 ± 0.00
Oats73.30 ± 29.090.00 ± 0.002.45 ± 5.094.29 ± 10.0925.28 ± 27.218.49 ± 15.630.00 ± 0.0054.34 ± 32.0421.28 ± 30.1468.66 ± 17.2676.70 ± 15.730.00 ± 0.000.00 ± 0.0090.00 ± 13.56
Soybean-notill87.78 ± 1.6083.25 ± 1.2278.05 ± 9.8789.90 ± 0.3688.77 ± 1.2284.38 ± 1.4184.11 ± 0.2787.11 ± 1.7788.80 ± 3.5887.19 ± 1.9893.57 ± 1.2180.54 ± 3.2081.28 ± 2.0082.15 ± 3.53
Soybean-mintill96.26 ± 1.2494.38 ± 0.5191.41 ± 4.1696.55 ± 0.1296.65 ± 0.5594.64 ± 0.6397.06 ± 0.0696.78 ± 0.8296.27 ± 0.5990.70 ± 1.6395.46 ± 0.5397.61 ± 1.3397.59 ± 1.2299.03 ± 0.40
Soybean-clean91.80 ± 2.2189.11 ± 1.7178.53 ± 12.7692.96 ± 1.3493.57 ± 1.3086.16 ± 5.2491.20 ± 0.4289.52 ± 3.3587.66 ± 5.0881.85 ± 3.9794.68 ± 1.3993.03 ± 4.3894.51 ± 2.0194.91 ± 1.81
Wheat98.12 ± 1.3286.71 ± 7.8154.68 ± 33.4396.73 ± 1.4397.08 ± 1.6489.64 ± 4.40100.00 ± 0.0095.00 ± 3.6294.35 ± 4.8897.18 ± 3.0299.89 ± 0.1894.38 ± 7.5586.11 ± 6.0192.38 ± 6.74
Woods98.28 ± 2.4297.81 ± 0.5894.10 ± 5.2298.27 ± 0.3897.75 ± 0.7598.39 ± 0.5299.56 ± 0.1798.67 ± 0.6698.87 ± 0.3196.21 ± 0.8998.49 ± 0.4099.58 ± 0.4499.52 ± 0.2598.88 ± 1.28
Buildings-Grass-Trees-Drives97.82 ± 1.4693.48 ± 2.2573.19 ± 17.0998.00 ± 0.6098.25 ± 1.1291.41 ± 3.6086.17 ± 1.4796.08 ± 2.7696.06 ± 1.4863.92 ± 3.7877.76 ± 1.6086.48 ± 9.2787.67 ± 5.3395.19 ± 2.69
Stone-Steel-Towers52.74 ± 21.3946.87 ± 17.6241.04 ± 32.9153.30 ± 17.4963.78 ± 21.7767.35 ± 23.508.33 ± 22.8239.20 ± 32.3025.11 ± 33.4787.73 ± 3.1596.80 ± 1.9829.05 ± 25.480.00 ± 0.0043.93 ± 23.09
OA (%)94.48 ± 1.4191.48 ± 0.5283.10 ± 10.1994.50 ± 0.3094.28 ± 0.4391.41 ± 0.9391.02 ± 0.9594.09 ± 0.9794.03 ± 0.9084.67 ± 1.2592.85 ± 0.4191.50 ± 0.6789.84 ± 1.2694.91 ± 0.51
AA (%)83.81 ± 3.3873.92 ± 2.1262.87 ± 12.0278.40 ± 1.1879.90 ± 2.1177.31 ± 2.4169.78 ± 4.5284.54 ± 3.9182.75 ± 2.8585.56 ± 4.9186.55 ± 3.5574.06 ± 1.5068.27 ± 1.8390.83 ± 2.84
κ (%)93.69 ± 1.6190.25 ± 0.5980.70 ± 11.5293.71 ± 0.3593.46 ± 0.4990.20 ± 1.0689.72 ± 1.0793.25 ± 1.1093.18 ± 1.0382.66 ± 1.4191.87 ± 0.4790.28 ± 0.7788.36 ± 1.4594.18 ± 0.59
Table 2. Comparisons with the state-of-the-art models on the Houston2013 dataset (10% training samples).
Table 2. Comparisons with the state-of-the-art models on the Houston2013 dataset (10% training samples).
Class2D-CNN3D-CNNHybridSNViTDeep ViTCvTHiTSSFTTMorphformerSS_TMNetDCTNMambaHSIS2MambaHG-Mamba
Healthy Grass95.21 ± 1.6691.06 ± 3.4490.76 ± 1.8887.43 ± 3.2591.33 ± 2.1787.30 ± 10.7394.49 ± 0.5296.42 ± 1.1685.97 ± 8.7597.60 ± 0.6498.86 ± 0.5094.08 ± 2.2993.14 ± 1.4896.47 ± 0.85
Stressed grass95.51 ± 1.7288.85 ± 6.4686.26 ± 8.2780.16 ± 5.7683.59 ± 2.7190.21 ± 4.4091.22 ± 2.1097.16 ± 1.4288.97 ± 5.5898.44 ± 0.5699.33 ± 0.2297.40 ± 1.2997.50 ± 0.9497.96 ± 1.47
Synthetic GrassTrees99.24 ± 0.4597.36 ± 3.0695.85 ± 3.6797.95 ± 0.8598.64 ± 0.5697.70 ± 2.8098.97 ± 0.2999.30 ± 0.5891.37 ± 13.8499.50 ± 0.2399.74 ± 0.2994.95 ± 2.0595.27 ± 2.0798.24 ± 1.13
Trees94.72 ± 2.1888.28 ± 4.5279.71 ± 7.4681.31 ± 6.6583.87 ± 4.6289.45 ± 3.2789.26 ± 1.2097.68 ± 0.9090.96 ± 4.9597.26 ± 0.9699.21 ± 0.4197.17 ± 0.8792.68 ± 3.4799.10 ± 0.87
Soil99.81 ± 0.1795.72 ± 3.7196.41 ± 3.8397.80 ± 1.2598.62 ± 1.5199.17 ± 0.6799.53 ± 0.2499.43 ± 0.6397.96 ± 2.1798.19 ± 0.3398.65 ± 0.1299.50 ± 0.5699.87 ± 0.1399.86 ± 0.43
Water94.05 ± 0.9786.16 ± 1.9291.69 ± 3.1286.50 ± 1.7987.48 ± 2.3693.58 ± 2.2386.87 ± 0.8293.61 ± 3.0790.85 ± 4.2593.67 ± 2.3398.75 ± 1.0784.38 ± 4.0480.75 ± 4.7789.04 ± 5.59
Residential96.82 ± 1.0283.00 ± 6.9878.74 ± 17.1287.38 ± 1.9987.28 ± 2.6892.51 ± 3.8291.87 ± 0.8397.52 ± 0.8085.78 ± 24.7394.54 ± 1.0398.20 ± 0.4597.00 ± 1.6988.33 ± 5.1298.35 ± 0.76
Commercial97.62 ± 1.5189.55 ± 1.6193.01 ± 2.6590.55 ± 3.3494.05 ± 1.9594.16 ± 3.5596.41 ± 0.6497.88 ± 1.2490.76 ± 9.5695.74 ± 1.3598.22 ± 0.6993.78 ± 1.6495.32 ± 1.9897.69 ± 1.29
Road95.42 ± 1.5185.03 ± 3.3873.95 ± 14.3686.62 ± 1.5787.98 ± 1.7887.78 ± 2.9389.78 ± 0.9996.70 ± 1.5687.54 ± 7.9994.29 ± 1.3297.66 ± 0.5692.94 ± 1.7488.00 ± 3.0698.97 ± 0.81
Highway99.09 ± 1.5791.67 ± 6.5491.92 ± 7.9296.43 ± 3.2295.01 ± 3.4497.50 ± 1.5697.93 ± 0.5199.78 ± 0.3893.89 ± 9.7696.91 ± 0.8198.89 ± 0.4999.27 ± 1.0399.97 ± 0.0899.98 ± 0.05
Railway99.58 ± 0.4681.57 ± 5.8785.29 ± 7.9293.86 ± 2.5193.32 ± 4.1294.73 ± 3.2299.37 ± 0.4599.76 ± 0.4491.69 ± 16.2694.94 ± 0.7298.51 ± 0.3398.79 ± 1.1697.41 ± 1.5099.39 ± 1.05
Parking Lot 197.88 ± 1.8394.30 ± 1.8195.79 ± 2.7191.42 ± 4.9193.49 ± 3.9695.83 ± 3.4798.68 ± 0.2198.70 ± 1.2488.63 ± 13.8996.50 ± 1.0098.96 ± 0.2295.81 ± 1.7597.49 ± 0.9297.32 ± 1.59
Parking Lot 297.55 ± 2.4481.89 ± 6.2486.51 ± 8.0891.84 ± 1.5589.01 ± 4.0395.80 ± 2.8795.93 ± 1.5398.54 ± 1.2392.25 ± 6.6693.42 ± 1.6197.68 ± 1.0195.85 ± 2.4496.23 ± 2.1698.86 ± 1.41
Tennise Court99.98 ± 0.0798.60 ± 0.9792.98 ± 6.1699.20 ± 0.8598.43 ± 2.6697.96 ± 3.2899.99 ± 0.0499.67 ± 0.4799.04 ± 1.4699.88 ± 0.19100.00 ± 0.0099.97 ± 0.08100.00 ± 0.00100.00 ± 0.00
Running Track97.87 ± 1.7194.77 ± 5.4191.48 ± 5.6096.93 ± 1.8998.14 ± 0.7296.65 ± 3.1298.39 ± 0.3398.95 ± 0.7193.32 ± 7.4998.98 ± 0.5899.24 ± 0.8899.21 ± 1.04100.00 ± 0.0099.83 ± 0.40
OA (%)97.32 ± 0.4889.54 ± 3.2188.19 ± 4.7590.27 ± 1.6991.58 ± 1.5093.52 ± 1.8895.20 ± 0.3398.15 ± 0.5390.98 ± 7.7196.22 ± 0.3598.31 ± 0.1696.33 ± 0.3894.98 ± 0.5598.41 ± 0.26
AA (%)97.03 ± 0.3789.74 ± 2.8788.97 ± 4.0990.45 ± 1.4291.38 ± 1.4393.74 ± 1.9594.80 ± 0.3297.81 ± 0.5990.99 ± 7.4695.15 ± 0.4597.32 ± 0.3696.01 ± 0.4494.80 ± 0.5998.07 ± 0.34
κ (%)97.10 ± 0.5188.70 ± 3.4787.24 ± 5.1389.48 ± 1.8390.89 ± 1.6292.99 ± 2.0494.81 ± 0.3598.00 ± 0.5890.24 ± 8.3695.92 ± 0.3898.17 ± 0.1796.03 ± 0.4194.57 ± 0.6098.28 ± 0.28
Table 3. Comparisons with the state-of-the-art models on the WHL dataset (1% training samples).
Table 3. Comparisons with the state-of-the-art models on the WHL dataset (1% training samples).
Class2D-CNN3D-CNNHybridSNViTDeep ViTCvTHiTSSFTTMorphformerSS_TMNetDCTNMambaHSIS2MambaHG-Mamba
Corn99.87 ± 0.0399.34 ± 0.4099.34 ± 0.5899.33 ± 0.1399.66 ± 0.0899.83 ± 0.0799.77 ± 0.0499.71 ± 0.2399.70 ± 0.2899.86 ± 0.1199.99 ± 0.0199.43 ± 0.2599.73 ± 0.1999.87 ± 0.12
Cotton99.72 ± 0.0996.30 ± 1.2497.55 ± 3.1883.45 ± 0.8996.41 ± 1.6099.39 ± 0.2397.74 ± 0.6999.69 ± 0.2598.18 ± 1.7498.55 ± 1.5099.98 ± 0.0198.12 ± 1.9099.56 ± 0.3998.86 ± 0.91
Sesame94.97 ± 1.2755.88 ± 29.7981.52 ± 14.5451.31 ± 21.7088.79 ± 4.2997.91 ± 0.7191.44 ± 1.5092.35 ± 5.0394.05 ± 5.5695.42 ± 4.0098.80 ± 0.6869.48 ± 24.8897.66 ± 1.6698.90 ± 1.34
Broad-leaf soybean99.12 ± 0.0896.24 ± 0.8998.22 ± 0.6496.51 ± 0.6798.52 ± 0.2299.49 ± 0.0998.71 ± 0.2199.69 ± 0.1699.51 ± 0.4199.79 ± 0.1099.95 ± 0.0299.73 ± 0.0999.89 ± 0.0999.86 ± 0.09
Narrow-leaf soybean95.42 ± 0.7187.80 ± 2.4481.06 ± 10.1942.38 ± 7.0889.80 ± 2.3197.10 ± 1.0095.35 ± 1.4191.00 ± 4.0988.15 ± 8.3291.07 ± 4.0894.13 ± 1.4789.56 ± 3.8294.16 ± 1.4696.56 ± 1.61
Rice98.57 ± 0.1998.11 ± 0.6697.28 ± 0.9198.54 ± 0.4098.87 ± 0.1698.80 ± 0.1199.26 ± 0.0599.06 ± 0.6698.77 ± 1.5999.23 ± 0.2599.18 ± 0.1898.13 ± 0.5299.24 ± 0.3199.09 ± 0.42
Water99.74 ± 0.0499.51 ± 0.2297.11 ± 0.5399.28 ± 0.1199.39 ± 0.1899.70 ± 0.1299.45 ± 0.0499.95 ± 0.0599.97 ± 0.0199.98 ± 0.01100.00 ± 0.0099.99 ± 0.0199.99 ± 0.0099.81 ± 0.38
Roads and houses85.56 ± 0.7882.84 ± 2.2278.60 ± 7.0980.96 ± 2.0483.05 ± 1.4284.75 ± 0.9687.96 ± 0.7485.45 ± 5.9189.66 ± 4.2288.66 ± 3.0480.26 ± 4.5389.38 ± 3.5285.76 ± 5.0880.02 ± 6.18
Mixed weed78.58 ± 2.0479.05 ± 2.9937.10 ± 12.8677.37 ± 2.0174.30 ± 2.8979.17 ± 2.8480.65 ± 1.0184.71 ± 9.5277.91 ± 7.9678.67 ± 8.0877.96 ± 4.8774.92 ± 5.5076.94 ± 3.0287.52 ± 6.65
OA (%)98.35 ± 0.0996.60 ± 0.6395.77 ± 0.8095.28 ± 0.4897.55 ± 0.1798.50 ± 0.1298.18 ± 0.1198.58 ± 0.3798.39 ± 0.3898.61 ± 0.2698.54 ± 0.1497.96 ± 0.4498.61 ± 0.1598.67 ± 0.29
AA (%)93.24 ± 0.4684.58 ± 3.9582.29 ± 3.5977.94 ± 3.0890.01 ± 1.0494.34 ± 0.5892.43 ± 0.5394.62 ± 1.2493.99 ± 1.7694.58 ± 1.0494.47 ± 0.5790.97 ± 3.2294.77 ± 0.5195.61 ± 0.99
κ (%)97.82 ± 0.1395.49 ± 0.8594.38 ± 1.0793.74 ± 0.6596.76 ± 0.2398.03 ± 0.1697.60 ± 0.1598.13 ± 0.4997.88 ± 0.5098.17 ± 0.3598.08 ± 0.1897.30 ± 0.5998.17 ± 0.1998.25 ± 0.38
Table 4. Comparison of computational complexity and performance.
Table 4. Comparison of computational complexity and performance.
MethodFLOPs (G)Param (MB)Training Time (s)Testing Time (s)OA (%)AA (%) κ (%)
2D-CNN0.771.7130.131.2894.48 ± 1.4193.81 ± 3.3893.69 ± 1.61
3D-CNN1.351.4581.482.5691.48 ± 0.5273.92 ± 2.1290.25 ± 0.50
Deep ViT13.6952.75142.503.6994.28 ± 0.4379.90 ± 2.1193.46 ± 0.49
HiT2.2027.22534.298.0691.02 ± 0.9569.78 ± 4.5289.72 ± 1.07
MambaHSI0.0060.02174.556.9891.50 ± 0.6774.06 ± 1.5090.28 ± 0.77
HG-Mamba0.571.30248.208.8894.91 ± 0.5190.83 ± 2.8494.18 ± 0.59
Table 5. Study of module contribution via ablation analysis.
Table 5. Study of module contribution via ablation analysis.
Module RemovedOA (%)AA (%) κ (%)
Val ± Std Δ Val ± Std Δ Val ± Std Δ
Stage 193.92 ± 0.48(0.99↓)85.87 ± 3.42(4.96↓)93.05 ± 0.55(1.13↓)
Stage 292.98 ± 0.54(1.93↓)81.46 ± 2.83(9.37↓)91.98 ± 0.63(2.20↓)
Spectral Bidirectional Mamba94.53 ± 0.43(0.38↓)87.25 ± 2.38(3.58↓)93.75 ± 0.49(0.43↓)
Spatial Bidirectional Mamba94.84 ± 0.61(0.07↓)87.89 ± 2.85(2.94↓)94.10 ± 0.69(0.08↓)
Only Unidirectional Mamba94.19 ± 0.80(0.72↓)86.46 ± 2.68(4.37↓)93.36 ± 0.91(0.82↓)
Spectral Compression94.01 ± 0.47(0.90↓)86.18 ± 3.98(4.65↓)93.16 ± 0.54(1.02↓)
GDD94.45 ± 0.73(0.46↓)88.90 ± 2.48(1.93↓)93.66 ± 0.84(0.52↓)
Full Model (None Removed)94.91 ± 0.5190.83 ± 2.8494.18 ± 0.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, X.; Yang, J.; Li, L.; Xue, S.; Shi, H.; Tang, H.; Huang, X. HG-Mamba: A Hybrid Geometry-Aware Bidirectional Mamba Network for Hyperspectral Image Classification. Remote Sens. 2025, 17, 2234. https://doi.org/10.3390/rs17132234

AMA Style

Yang X, Yang J, Li L, Xue S, Shi H, Tang H, Huang X. HG-Mamba: A Hybrid Geometry-Aware Bidirectional Mamba Network for Hyperspectral Image Classification. Remote Sensing. 2025; 17(13):2234. https://doi.org/10.3390/rs17132234

Chicago/Turabian Style

Yang, Xiaofei, Jiafeng Yang, Lin Li, Suihua Xue, Haotian Shi, Haojin Tang, and Xiaohui Huang. 2025. "HG-Mamba: A Hybrid Geometry-Aware Bidirectional Mamba Network for Hyperspectral Image Classification" Remote Sensing 17, no. 13: 2234. https://doi.org/10.3390/rs17132234

APA Style

Yang, X., Yang, J., Li, L., Xue, S., Shi, H., Tang, H., & Huang, X. (2025). HG-Mamba: A Hybrid Geometry-Aware Bidirectional Mamba Network for Hyperspectral Image Classification. Remote Sensing, 17(13), 2234. https://doi.org/10.3390/rs17132234

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop