Next Article in Journal
Synergizing Multi-Temporal Remote Sensing and Systemic Resilience for Rainstorm–Flood Risk Zoning in the Northern Qinling Foothills: A Geospatial Modeling Approach
Next Article in Special Issue
A Lightweight Remote-Sensing Image-Change Detection Algorithm Based on Asymmetric Convolution and Attention Coupling
Previous Article in Journal
Comprehensive Assessment of Ocean Surface Current Retrievals Using SAR Doppler Shift and Drifting Buoy Observations
Previous Article in Special Issue
Multimodal Prompt-Guided Bidirectional Fusion for Referring Remote Sensing Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HyperSMamba: A Lightweight Mamba for Efficient Hyperspectral Image Classification

1
College of Computer Science and Technology, Xinjiang University, Urumqi 830017, China
2
Xinjiang Multimodal Intelligent Processing and Information Security Engineering Technology Research Center, Urumqi 830017, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(12), 2008; https://doi.org/10.3390/rs17122008
Submission received: 18 April 2025 / Revised: 3 June 2025 / Accepted: 8 June 2025 / Published: 11 June 2025

Abstract

Deep learning has recently achieved remarkable progress in hyperspectral image (HSI) classification. Among these advancements, the Transformer-based models have gained considerable attention due to their ability to establish long-range dependencies. However, the quadratic computational complexity of the self-attention mechanism limits its application in hyperspectral image classification (HSIC). Recently, the Mamba architecture has shown outstanding performance in 1D sequence modeling tasks owing to its lightweight linear sequence operations and efficient parallel scanning capabilities. Nevertheless, its application in HSI classification still faces challenges. Most existing Mamba-based approaches adopt various selective scanning strategies for HSI serialization, ensuring the adjacency of scanning sequences to enhance spatial continuity. However, these methods lead to substantially increased computational overhead. To overcome these challenges, this study proposes the Hyperspectral Spatial Mamba (HyperSMamba) model for HSIC, aiming to reduce computational complexity while improving classification performance. The suggested framework consists of the following key components: (1) a Multi-Scale Spatial Mamba (MS-Mamba) encoder, which refines the state-space model (SSM) computation by incorporating a Multi-Scale State Fusion Module (MSFM) after the state transition equations of original SSMs. This module aggregates adjacent state representations to reinforce spatial dependencies among local features; (2) our proposed Adaptive Fusion Attention Module (AFAttention) to dynamically fuse bidirectional Mamba outputs for optimizing feature representation. Experiments were performed on three HSI datasets, and the findings demonstrate that HyperSMamba attains overall accuracy of 94.86%, 97.72%, and 97.38% on the Indian Pines, Pavia University, and Salinas datasets, while maintaining low computational complexity. These results confirm the model’s effectiveness and potential for practical application in HSIC tasks.

1. Introduction

Hyperspectral images (HSIs) are image data collected across dozens to hundreds of bands, covering wavelengths from the visible to infrared spectrum [1,2]. Unlike conventional RGB images, HSIs provide rich spatial information and higher spectral resolution, offering more detailed information about land cover [3]. Therefore, HSIs have attracted increasing attention in remote sensing research, and have been extensively applied in both military and civilian domains, including precision agriculture, defense reconnaissance, and environmental monitoring [4].
Hyperspectral image classification (HSIC) techniques aim to assign a precise land cover category label to each pixel. As the quality of hyperspectral image data continues to improve and its applications expand, how to achieve high accuracy and computational efficiency of HSIC has become one of the hotspots of current research [5].
In recent years, deep learning has made groundbreaking advances across various domains. Unlike traditional machine learning algorithms, deep learning can automatically extract high-level representations without relying on manual feature engineering. This allows classification models to better express the intrinsic characteristics of datasets. Owing to their ability to autonomously extract discriminative features, deep learning techniques have been increasingly applied to HSIC. Representative models include Auto-Encoders (AEs) [6], Convolutional Neural Networks (CNNs) [7], Graph Convolutional Networks (GCNs) [8], Long Short-Term Memory networks (LSTM) [9], Recurrent Neural Networks (RNNs) [10], and Transformers [11].
Convolutional Neural Networks (CNNs) are one of the most widely utilized architectures for feature extraction. In addition to delivering strong performance, they have found widespread use in edge-aware networks. Numerous studies have aimed to enhance the efficiency of CNN-based modeling, including the use of grouped convolutions [12] and depthwise separable convolutions [13] (consisting of depthwise convolutions and pointwise convolutions). Moreover, to fully exploit the high dimensionality of HSI, several specialized CNN architectures have been developed. The foundational 3D-CNN framework introduced by Li et al. [14] accomplishes joint spatial–spectral feature learning. Zhou et al. [15] addressed the spectral discontinuity issue between bands through the proposed Multi-scale Fusion Spectral Attention Mechanism. Xu et al. [16] developed an optimization method to automatically extract local spectral features. Roy et al. [17] constructed a Hybrid CNN Network (HybridSN) that better captures deep spatial–spectral joint features. CNNs model neighborhood-level relationships and generally achieve more satisfactory classification results. However, they rely on fixed-size convolutional kernels to extract local area information, which limits their ability to consider global context [18].
Compared with CNNs, the Transformer architecture is capable of modeling long-range dependencies between any pair of pixels, enabling a more comprehensive understanding of global spatial–spectral relationships. Therefore, Transformer-based models have attracted considerable attention in HSIC tasks. To enhance spectral representation learning, Hong et al. [19] proposed the Spectral Transformer (SpectralFormer). Following this, a variety of Transformer-driven models have emerged, such as the Convolutional Transformer Network (CTN) [20], Spatial–Spectral Transformer Network (SST) [21], and Neighborhood Enhancement Hybrid Transformer Network (NEHT) [22]. Moreover, recent efforts have been directed toward joint modeling of spectral–spatial and semantic representations, as demonstrated by the Spectral–Spatial Feature Tokenization Transformer (SSFTT) [23], Hyperspectral Image Transformer (HiT) [24], and Hierarchical Unified Spectral–Spatial Aggregated Transformer (HUSST) [25]. However, despite their powerful modeling capabilities, Transformer architectures suffer from high computational complexity.
The Mamba architecture draws inspiration from state-space models in control theory, achieving linear scaling of complexity with sequence length [26]. Mamba utilizes parameterized matrices and selective mechanisms, combined with hardware-aware parallel optimization strategies, to dynamically retain or discard information according to inter-element correlations in sequences. This design efficiently captures long-range dependencies while maintaining linear time complexity [27].
Researchers are actively exploring the application potential of Mamba-based models in visual tasks. Zhu et al. [28] introduced Vision Mamba (ViM), a model based on a bidirectional scanning mechanism designed for general-purpose visual tasks. Meanwhile, Liu et al. [29] proposed a Visual State-Space Model (VMamba), utilizing a cross-scanning mechanism to convert 2D images into 1D sequences. In addition, EfficientVMamba [30] introduces the Efficient 2D Scanning (ES2D) technique, which uses a dilated selective scanning mechanism to eliminate redundant tokens while maintaining a broad receptive field. Huang et al. [31] argued that global image scanning may fail to effectively model local spatial dependencies. To address this issue, they proposed LocalMamba, which partitions the input into multiple local windows through localized scanning to better capture fine-grained spatial features.
In the field of HSIC, Mamba-based research has made significant progress. Yao et al. [32] proposed SpectralMamba by mining and analyzing the spectral information using Mamba. Yang et al. [33] incorporated CNN layers after the Mamba block to jointly capture global spectral features and local spatial patterns. Chen et al. [34] enhanced the model’s expressiveness through a dynamic multi-path activation mechanism. Zhuang et al. [35] developed the Frequency Group Embedding Module to capture high-frequency and low-frequency features, and then used bidirectional Mamba to obtain long-range spatial relationships. Furthermore, numerous models, such as Mamba in Mamba (MiM) [36] and Spectral-Spatial Mamba (SSMamba) [37], have been proposed for spectral–spatial joint modeling in HSIC. By employing spatial cross-scanning and spectral bidirectional scanning techniques, these models enhance the alignment between spectral and spatial features.
Unlike sequential data such as text and audio, HSIs contain rich spatial and spectral information. Mamba selectively propagates or filters information along the sequence, meaning that the modeling process is influenced only by past and current inputs within the sequence, making it difficult to fully model the global relationships in the 2D space. Existing visual Mamba models [31,32,33,35,36,37] often adopt directional scanning strategies that reshape 2D visual inputs into multiple 1D sequences from various orientations. These approaches partially alleviate spatial structure misalignment and information loss. However, these methods face challenges regarding model effectiveness and efficiency. Directional scanning can distort the original spatial relationships between pixels, compromising the spatial context [38]. Moreover, the use of multiple scanning directions significantly increases the computational cost [39]. Therefore, there is a pressing need for more efficient strategies to optimize Vision Mamba (ViM) architectures for hyperspectral data, enabling the full exploitation of complex spatial–spectral representations while meeting the demand for accurate and efficient classification.
To address the aforementioned issues and achieve efficient joint spatial–spectral features extraction, we propose a novel model, Hyperspectral Spatial Mamba (HyperSMamba). This architecture inherits the Vision Mamba (ViM) model’s efficient capability to model long-range dependencies, while being structurally optimized and modularly designed to address the specific characteristics of HSI classification tasks. We extract long-range dependencies through the Multi-Scale Spatial Mamba (MS-Mamba) module, and then design the Adaptive Fusion Attention (AFAttention) module to dynamically integrate bidirectional features to enhance feature representation.
The main contributions of this paper are as follows:
1.
The proposed Vision Mamba-based HSIC framework, HyperSMamba, significantly improves the extraction of long-range spatial–spectral features while maintaining high computational efficiency.
2.
FusionSSM improves the SSM computation process by incorporating the innovative Multi-Scale Feature Fusion Module (MSFM) after the state transition equation, which facilitates the flow of spatial context information and alleviates the limitations imposed by causal relationships.
3.
The Adaptive Fusion Attention Module (AFAttention) is proposed to interact and fuse multi-source spatial–spectral features, allowing the model to autonomously focus on key feature regions.

2. Methods

2.1. Overview

In order to achieve collaborative deep mining of spectral features in HSI, this study develops the end-to-end framework HyperSMamba, as illustrated in Figure 1.
We first apply principal component analysis (PCA) [40] to obtain a reduced representation I R H × W × C , where H, W, and C represent the height, width, and the number of retained principal components. Then, we adopt the Follow Patch method proposed by Yang et al. [41] to divide the reduced HSI into localized 3D patches X R P × P × C , where P denotes the patch size. Unlike traditional padding-based or Vision Transformer (ViT)-style patches, this approach ensures that the target pixel is positioned as close to the center of the patch as possible, thus retaining more effective neighborhood information. To further exploit neighborhood dependencies, we apply horizontal forward and backward scanning to each patch, generating two directional sequences x ( t ) R P 2 × C .These sequences are then independently fed into the Multi-Scale Spatial Mamba (MS-Mamba) module for bidirectional spatial feature extraction. In MS-Mamba, the state variables computed from the state transition equations of the original SSM are reshaped into a visual format. The MSFM module is then employed to autonomously select meaningful neighboring state information to update each state variable, thereby enhancing spatial interaction. To further enhance feature fusion capability, we propose the AFAttention module, which integrates features extracted from forward and backward sequences. This module consists of spatial attention, channel attention, and pixel-wise attention mechanisms. We assign a unique importance weight to each pixel to refine forward and backward features, effectively reducing redundant information while preserving critical features unchanged. Finally, the improved features are supplied to the fully connected layer after being residually merged with the original and linearly modified features.

2.2. State-Space Models

The structured State-Space Sequence Model originates from the Kalman filter framework and excels at catching long-range relationships while demonstrating efficient parallel processing capabilities [42]. Through a learnable hidden state h ( t ) R N , the model converts a 1D input sequence x ( t ) into an output sequence y ( t ) R L , where t indicates the time step, N and L denote the dimensions of the latent space and the sequence.
h ˜ ( t ) = A h ( t ) + B x ( t ) ,
y ( t ) = C h ˜ ( t ) + D x ( t ) .
The state transition equation describes the temporal dynamics of the hidden state, governed by its current value and the external input. Specifically, A R N × N is the state transition matrix, representing the transition relationship from the current state to the next state. B R N × L denotes the input projection matrix, capturing the influence of external inputs on the hidden state. h ˜ ( t ) = d h ( t ) d t represents the time derivative of h ( t ) . The output equation generates y ( t ) by projecting the updated hidden state into the output space via the observation matrix C R L × N , while D acts as a residual connection, often omitted in practical implementations.
To adapt this continuous system for discrete sequence data, the Zero-Order Hold (ZOH) rule is applied to discretize the continuous parameters. The ZOH rule approximates inputs as constant within each sampling interval, ensuring consistency between the continuous model and its discrete-time implementation.
A ¯ = exp ( Δ A ) ,
B ¯ = ( Δ A ) 1 exp ( Δ A ) I · Δ B .
Specifically, a time scaling parameter Δ R L is introduced. Here, exp ( · ) denotes the matrix exponential function, which propagates the system dynamics over a time interval Δ , and I represents the identity matrix. The matrices A ¯ R N × N and B ¯ R N × L are the discretized forms of A and B. After discretization, the state-space model equations are shown below:
h ( t ) = A ¯ h ( t 1 ) + B ¯ x ( t ) ,
y ( t ) = C h ( t ) + D x ( t ) .
However, Gu et al. [43] pointed out that the matrices A, B, and C in traditional state-space models typically remain fixed throughout the entire sequence, thereby exhibiting Linear Time Invariance (LTI). This property limits the model’s expressiveness when dealing with complex, dynamically changing sequence patterns. To overcome this limitation, Mamba introduces a selective scanning mechanism, allowing parameters to change with input, improving the model’s capacity in dynamic contexts. Additionally, Mamba adopts a hardware-efficient parallel implementation to resolve the inference latency bottlenecks observed in S4, marking a notable advancement in the evolution of state-space models.

2.3. Multi-Scale Spatial Mamba

Since the state transition equation of Mamba is influenced solely by the past and current input states within the sequence, the model is unable to leverage information from unscanned data. This limitation restricts its receptive field and impairs its ability to capture global contextual information, potentially leading to incomplete or biased feature representations.
To establish spatial dependencies with unscanned adjacent features, we introduce Multi-Scale Spatial Mamba (MS-Mamba). In this module, FusionSSM incorporates the Multi-Scale State Fusion Module (MSFM) into the original Mamba structure. It performs self-learning weighting of adjacent state variables to aggregate information from neighboring states, thereby enhancing the representation of spatial relational features.
As depicted in Figure 1, each sequence is processed through a 1D convolutional (1D-CNN) layer and SiLU activation function. Subsequently, the FusionSSM module captures long-range dependencies while simultaneously incorporating latent local spatial information. The overall process of FusionSSM can be described by three equations: the state transition equation, the MSFM, and the observation equation.
h ( t ) = A ¯ h ( t 1 ) + B ¯ x ( t ) ,
h ( t ) = MSFM h ( t ) ,
y ( t ) = C h ( t ) + D x ( t ) .
Specifically, the current state h ( t ) is first computed using the state transition equation of the original SSM and then reshaped into a visualizable format. To enable each state to perceive its neighboring states, we employ the MSFM module to perform Multi-Scale State Fusion on the processed state variables. This procedure integrates local dependency information into the new state h ( t ) , resulting in a more contextually informed representation. Then, the new state variables h ( t ) are reverted to the original format and fed to the observation equation to compute the final result. Finally, the resulting output from the observation equation is multiplied by a Gaussian decay mask to optimize the feature distribution, yielding the forward features X for R P × P × C and backward features X back R P × P × C .

2.4. Multi-Scale State Fusion Module

Hyperspectral images present local spatial heterogeneity, meaning that the same land cover class may exhibit different spectral responses at different spatial locations. Traditional CNNs, owing to the restricted local receptive field, often struggle to fully extract effective features [44]. Moreover, convolutional features extracted at different scales possess varying degrees of importance, and simple fusion strategies fail to fully exploit their complementary characteristics. Therefore, it is essential to design a model that can effectively extract spatial features at multiple scales and adaptively fuse them based on their contextual relevance.
We introduce the Multi-Scale State Fusion Module (MSFM), designed to dynamically select spatial features at various scales through multi-scale convolutions, as shown in Figure 2.
The MSFM module adopts a parallel three-branch convolution architecture. For input X i n , the first branch uses standard depthwise convolution (DWConv) to capture fine-grained local features. The second and third branches utilize depthwise dilated convolutions (DDConv) with different kernel sizes to approximate the receptive field of large kernel convolutions, identifying regions with similar local features but more distant distributions. Then, a richer feature representation X concat is created by concatenating the outputs of three branches along the channel dimension.
X concat = Concat ( DWConv 3 × 3 ( X in ) , DDConv 5 × 5 ( X in ) , DDConv 7 × 7 ( X in ) ) .
Here, DDConv k × k ( · ) and DWConv k × k ( · ) denote dilated depthwise convolution and standard depthwise convolution using a k × k convolutional kernel. The operator Concat ( · ) performs channel-wise concatenation.
Next, we use the average pooling layer and two convolutional layers to calculate the attention weights of the feature map, guiding the model to learn the interactions between channels at different scales during the fusion process. Finally, these attention weights are applied to the convolutional inputs of the corresponding scales to achieve weighting of the important channels and help the model automatically focus on more representative features.
[ W 1 ; W 2 ; W 3 ] = Conv 1 × 1 Conv 1 × 1 AvgPool ( X concat ) ,
h ( t ) = W 1 · DWConv 3 × 3 ( X in ) + W 2 · DDConv 3 × 3 ( X in ) + W 3 · DDConv 5 × 5 ( X in ) .
Here, AvgPool ( · ) represents the average pooling, Conv 1 × 1 ( · ) is a point-wise convolutional layer used for channel compression and attention extraction. [ W 1 ; W 2 ; W 3 ] represents the learned attention weights corresponding to three convolutional branches. The final output h ( t ) is the weighted sum of the three convolutional paths.

2.5. Adaptive Fusion Attention Module

To more effectively integrate key information from the forward and backward features while suppressing redundant and noisy components, we designed the Adaptive Fusion Attention Module (AFAttention), as illustrated in Figure 3. The module dynamically evaluates the relative importance of forward and backward features, achieving intelligent control of the bidirectional feature flow. AFAttention is composed of spatial attention, channel attention, and pixel attention.
We perform element-wise addition of the forward features X for R P × P × C and backward features X back R P × P × C to obtain X sum R P × P × C . Subsequently, X sum is fed into the spatial attention and channel attention, which generate spatial attention weights W s R P × P × 1 and channel attention weights W c R 1 × 1 × C , respectively.
The spatial attention (Figure 3b) applies max pooling and average pooling along the channel dimension. These pooled features are concatenated and processed by a 7 × 7 convolutional layer, capturing important features from local regions and enhancing the model’s spatial discriminative ability.
The channel attention (Figure 3d) compresses spatial dimensions using global average pooling and generates channel weights through a series of nonlinear transformations. This mechanism helps enhance the critical spectral band information in hyperspectral data while suppressing redundant channels. Subsequently, W s and W c are fused together through simple addition to generate the spatial–spectral importance matrix W sc R P × P × C .
W s = Conv 7 × 7 AvgPool ( X sum ) ; MaxPool ( X sum ) ,
W c = W 1 W 0 ( AvgPool ( X sum ) ) ,
W s c = W s + W c .
Here, Conv 7 × 7 represents a convolution operation with a 7 × 7 kernel, and MaxPool refers to channel-wise max pooling. W 0 R C × C / r , W 1 R C / r × C are nonlinear transformation coefficients.
Finally, we use pixel attention (Figure 3c) to obtain the final pixel attention matrix W and apply X s u m as guidance to adaptively adjust each channel of W s c . Specifically, each channel of W s c and X s u m is rearranged in an alternating manner through Channel Shuffle [45] and fused using a 7 × 7 group convolution that allows for full interaction of information from different sources. Then, the attention weights are normalized using the sigmoid activation to obtain the final weights W.
W = σ GConv 7 × 7 ( CS ( [ X sum , W sc ] ) ) .
where GConv 7 × 7 ( · ) denotes a grouped convolution operation with a 7 × 7 kernel. CS ( · ) is a channel rearrangement operation. σ ( · ) represents the sigmoid activation function.
Due to the different importance of forward and backward features, they are separately element-wise multiplied with the pixel attention matrix W and then combined using the weighted average method with the self-learned parameter α .
y output = α · ( X for W ) + ( 1 α ) · ( X back W )
where ⊙ denotes element-wise dot multiplication, y output is the final result of the model.

3. Results

3.1. Datasets Description

To systematically evaluate the model, we conduct comprehensive evaluations on three commonly used HSI datasets. The sample distribution is detailed in Table 1, Table 2 and Table 3.
1.
Indian Pines (IP): The IP dataset was collected from the agricultural region of Indiana, USA, and contains 145 × 145 pixels. 200 valid bands were retained, covering wavelengths from 400 nm to 2500 nm. The dataset covers 16 different land cover types.
2.
Pavia University (UP): The UP dataset was collected from the University of Pavia, Italy. It consists of 610 × 340 pixels and a total of 103 spectral bands. The dataset covers nine different land cover types.
3.
Salinas (SA): This dataset was collected from the Salinas region of California, USA. The HSI size is 512 × 217 pixels, covering 16 land cover types. After preprocessing, 204 valid bands were retained.

3.2. Experimental Setup

1.
Implementation Details: All models are implemented using PyTorch 2.5 and trained on an NVIDIA Tesla T4 GPU (16 GB VRAM). The batch size for all models is set to 128, and the number of epochs is set to 400. We use the cross-entropy loss function to update parameters. The AdamW optimizer is used with an initial learning rate of 5 × 10 4 . In the proposed HyperSMamba model, all feature dimensions are normalized to 50. The initial patch sizes are set to 9 × 9 for the Indian Pines (IP) dataset and 11 × 11 for both the Pavia University (UP) and Salinas (SA) datasets. To ensure reproducibility of the results, the random seed was fixed, and each experiment was repeated 10 times.
2.
Evaluation Metrics: We adopt three standard evaluation metrics, including average accuracy (AA), overall accuracy (OA), and Kappa coefficient ( κ ) .

3.3. Comparative Experimentation

3.3.1. Contrast Experiment Results

In order to evaluate the effectiveness of the proposed method, we conducted a comparative analysis against a series of baselines. These models include 2D-CNN [46], HybridSN [17], SpectralFormer [19], HiT [24], HUSST [25], SSFTT [23], and SSM-based ViM [28], MiM [36], SSMamba [37], and FAHM [35]. All models were tested under the optimal experimental conditions. Details can be found in the original paper.
Table 4, Table 5 and Table 6 show the average classification metrics and standard deviations of all comparison models after 10 experiments on three datasets. To further illustrate the comparative performance and stability of the proposed method, we include a detailed bar chart with error bars in Appendix A, Figure A1.
As a hybrid 2D–3D convolution-based method, HybridSN extracts joint local spectral–spatial features, thus outperforming certain Transformer-based methods. For example, on the IP and SA datasets, the classification accuracy of HybridSN is higher than SpectralFormer, which uses a Transformer only in the spectral dimension. SSFTT also utilizes 3D and 2D convolutions to extract information before the Transformer establishes long-range dependencies. This integrated spatial–spectral feature ensures that the Transformer receives inputs enriched with sufficient local semantic information. SSFTT achieves better performance than HiT, which embeds convolutions within the Transformer. The results illustrate that capturing local spatial features before establishing long-distance dependencies improves classification accuracy.
From the results, ViM does not perform well in hyperspectral image classification tasks. In contrast, MiM leverages the cross-sequence scanning mechanism to capture spatial information from four directions, leading to more focused and discriminative feature representations. Thus, MiM outperforms ViM in classification accuracy. SSMamba further extends the receptive field of spatial Mamba to large-scale 27 × 27 token groups, while cross-band feature calibration is achieved through spectral Mamba, resulting in superior classification performance compared with MiM. In comparison, FAHM extracts both high-frequency and low-frequency features and improves the model’s capacity to capture contextual dependencies across spectral bands via bidirectional spectral analysis.
The proposed HyperSMamba model, built on ViM, does not enhance the ability to capture spatial neighborhood information by introducing cross-scanning sequences from additional directions. However, it still achieves excellent performance across all three datasets. Notably, on UP and SA datasets, the OA improves by 3.42% and 2.72%, compared with SSMamba, which confirms the proposed model’s superiority in HSIC tasks. Further analysis of the classification accuracy for each category reveals that HyperSMamba performs well on dominant categories and edge pixels. Although the classification accuracy for partly minor categories is sub-optimal, it remains stable and does not lead to serious misclassification due to data imbalance, and the overall performance is balanced. Additionally, as shown in Appendix A, Figure A1, the standard deviation of HyperSMamba’s classification metrics on all datasets is very small, which indicates that the algorithm’s classification performance is more balanced on all datasets, and it is able to maintain high robustness under different data distributions.

3.3.2. Visual Analysis of Classification Results

To facilitate an intuitive comparison of the performance across various classification methods, this study visualizes and analyzes the classification results. Figure 4, Figure 5 and Figure 6 show the classification visualization maps for the IP, UP, and SA datasets.
From the classification maps, it is evident that convolution-based models are significantly limited by local receptive fields, which hinders their ability to capture global spatial context. This limitation results in reduced classification accuracy at category boundaries and disrupts spatial consistency. The Transformer-based models possess global modeling capabilities, which make the category edge transition smoother and provide a clearer distinction between flat land cover areas and boundaries. However, the under-utilization of spatial information in small-scale areas leads to misclassification of fine-grained categories and loss of texture information, which affects overall classification performance.
Compared with other methods, HyperSMamba effectively extracts both local and global spatial features. Across multiple datasets, it demonstrates superior texture discrimination and boundary recognition ability. The classification maps produced by HyperSMamba exhibit smoother visual transitions, sharper inter-class boundaries, and significantly reduced misclassification errors.

3.4. Ablation Study

3.4.1. Ablation for the HyperSMamba Architecture

To assess the contribution of the MSFM and AFAttention modules to the performance of the proposed HyperSMamba model, we conducted ablation experiments. The results are summarized in Table 7.
The inclusion of MSFM significantly improves the evaluation metrics across all three datasets. This finding indicates that this component can partially compensate for the loss of spatial structural information caused by flattening the input into 1D sequences. By integrating local spatial context into the latent state representation, MSFM improves spatial feature representation.
The AFAttention module adaptively adjusts the importance of spatial and channel features, optimizing the complementary relationship between bidirectional states and enhancing both feature fusion quality and classification accuracy. The accuracy of all datasets improved after adding the attention fusion module, further confirming the effectiveness of the AFAttention module.
When both MSFM and AFAttention are incorporated, the model benefits from enhanced global context modeling, as well as multi-scale spatial feature extraction. The ablation results clearly demonstrate that the full model significantly outperforms its counterparts without either component, confirming the synergistic effect of the two modules on classification performance.

3.4.2. Ablation for Multi-Scale State Fusion Module Design

To further analyze the effectiveness of each branch in the proposed Multi-Scale State Fusion Module (MSFM) module, we conduct an ablation study by evaluating different combinations of three convolutional operations: 3 × 3 depthwise convolution (DWConv), 3 × 3 dilated depthwise convolution (DDConv), and 5 × 5 DDConv. The experimental results on the Indian Pines (IP), Pavia University (UP), and Salinas (SA) datasets are summarized in Table 8.
Across all three datasets, combining any two branches results in noticeable performance improvements. On the IP dataset, the combination of 3 × 3 DWConv and 5 × 5 DDConv achieves the most substantial improvement among the two-branch configurations, with an overall accuracy (OA) of 94.01%. This enhancement can be attributed to the presence of heterogeneous and fine-scale objects in the IP scene. Specifically, the large receptive field introduced by the 5 × 5 DDConv effectively captures broad spatial context, while the 3 × 3 DWConv preserves detailed edge and boundary information. On the UP and SA datasets, the combination of 3 × 3 DWConv and 3 × 3 DDConv produces the most significant performance improvements among two-branch combinations. This improvement is mainly attributed to the fact that both datasets exhibit prominent textures and structured spatial patterns.
Moreover, activating all three branches simultaneously consistently produces the best classification performance across all datasets. This indicates that the complementary nature of multi-scale convolutions effectively facilitates the extraction of both fine-grained local features and broader contextual information.

4. Discussion

4.1. Discussion of the Computational Costs

To assess the computational efficiency of different methods, we calculated the number of parameters and Floating Point Operations (FLOPs) for each model on the IP dataset, and summarized the results in Table 9. In order to further visualize the differences in resource consumption, a combined bar and line chart is provided in Appendix A Figure A2. Tables and charts indicate significant differences across methods in terms of computational efficiency and parameter scale.
HybridSN employs local convolution computations, but its computational complexity and parameter count increase rapidly with larger patch sizes, limiting its applicability in high-resolution scenarios. While both Transformer-based and Mamba-based models can establish long-range dependencies, they exhibit notable differences in computational complexity and parameter scale. The Mamba-based model has superior computing efficiency. For instance, ViM achieves the lowest number of parameters and computational complexity, requiring only 79.87k parameters and 12.02M FLOPs. Similarly, the proposed HyperSMamba maintains strong classification performance with only 16.93M FLOPs, which is more than 90% lower than the computational costs of HiT and HybridSN.
Focusing on Mamba-based architectures, the proposed HyperSMamba model achieves the highest classification accuracy among its peers. It is noteworthy that HyperSMamba exhibits a substantial reduction in both computational complexity and parameter count compared with SS-Mamba, MiM, and FAHM. On the Indian Pines dataset, although it introduces a slight increase in computational cost and parameter usage relative to ViM, it achieves a 9.22% improvement in overall accuracy (OA), demonstrating an excellent trade-off between efficiency and classification performance. This favorable balance highlights HyperSMamba’s strong potential for practical and industrial deployment.

4.2. Learned Feature Visualizations by T-SNE

The high-dimensional features extracted by the model are mapped into a 2D space using t-Distributed Stochastic Neighbor Embedding (T-SNE) to intuitively assess the model’s feature extraction capability. The visualization outcomes of each comparative experiment on the UP dataset are shown in Figure 7. In the T-SNE plots, each point represents a pixel from the Pavia University test set, colored by its ground-truth class. The horizontal and vertical axes denote the first and second components of the t-SNE embedding, both normalized to the range [ 0 , 1 ] . Although these axes have no physical units, they reflect the relative positioning of feature representations in the projected space.
Specifically, the features of 2D-CNN and ViM are relatively scattered, and the class boundaries are unclear, reflecting the insufficient feature extraction capability of these models, which fail to adequately distinguish the feature representations of different classes in the high-dimensional feature space. Our method produces more compact and well-separated clusters, with clearer boundaries between categories after dimensionality reduction, indicating that the model has learned more discriminative features in the high-dimensional space. Furthermore, the proposed model exhibits smoother transitions between similar categories, reducing inter-class overlap by capturing subtle feature differences more precisely.

4.3. Parameters Analyzed

We explore the impact of two key hyperparameters, patch size and the number of bands after PCA processing, on the classification accuracy of HSI.

4.3.1. Impact of the Patch Size

In HSIC tasks, patch size directly influences the ability to extract spatial information. We calculated the classification accuracy for each dataset under different follow-up patch sizes.
As illustrated in Figure 8a, appropriately increasing patch size allows for the capture of more spatial contextual information while retaining local details. However, when the patch size becomes too large, the model may incorporate excessive background information, leading to information overlap between classes and consequently reducing classification accuracy, while significantly increasing computational cost. The IP dataset attained the maximum classification accuracy with a patch size of 9 × 9 , while the SA and UP datasets reached optimal performance with a patch size of 11 × 11 .

4.3.2. Impact of the Number of Bands After PCA

Principal component analysis (PCA) is used to retain key information while decreasing computational complexity. We tested the effect of different dimensions after PCA on classification performance, as shown in Figure 8b.
In the IP dataset, spectral differences between classes are subtle. Lower PCA dimensions result in significant information loss, making it difficult to effectively distinguish fine-grained classes and leading to significantly lower classification accuracy. As the number of retained components increases to 50, the OA rises sharply, primarily due to the enhanced discriminative capability resulting from the preservation of more critical spectral features. With further increases in dimensionality, the OA exhibits a further upward trend, indicating that high-dimensional representations are more beneficial for classifying the IP dataset. However, this also leads to a significant increase in model parameters, thereby introducing additional computational overhead. In contrast, the UP and SA datasets have higher spatial resolution and fewer spectral channels. When the PCA dimension is set to 10, there is considerable information loss, resulting in lower classification accuracy. However, OA reaches the maximum value when the PCA dimension is increased to 50. When the PCA dimension exceeds 50, excessive noise and redundant features are retained, which reduces classification accuracy and increases computational complexity.

4.4. Discussion of the Strengths and Limitations

The key to HSIC lies in the effective utilization of spectral and spatial features. Insufficient spatial information may lead to misclassification of land cover types that share similar spectral characteristics but differ in spatial distribution. Conversely, excessive spatial information without adequately leveraging spectral data may cause the model to overfit to local spatial patterns while neglecting the discriminative potential of spectral signatures. To improve the model’s capacity for establishing long-range dependencies in spectral and spatial dimensions while decreasing the number of parameters and computational expenses, we employed ViM as the base architecture. However, as shown in the comparative experiments in Table 4, Table 5 and Table 6, the inherent causal properties of SSM limit its ability to extract spatial features. Therefore, we designed a lightweight MSFM to extract features at multiple scales and used this module to integrate spatial state information, thereby enhancing the representation of local spatial features. The comparative results and classification maps clearly illustrate the benefits of HyperSMamba. In contrast to CNN-based models, HyperSMamba establishes long-range dependencies and improves the recognition accuracy of boundary pixels between classes. Although Transformer-based models (such as SpectralFormer and HiT) possess global modeling capabilities, their computational complexity is high, and they underutilize spatial information in small-scale regions, leading to misclassification of fine-grained categories. In contrast, HyperSMamba achieves higher classification accuracy while reducing computational overhead, particularly in areas where class boundaries are unclear and spectral similarity is high, exhibiting greater classification stability.
Despite the excellent performance of HyperSMamba in hyperspectral image classification tasks, the method still exhibits several potential limitations that warrant further investigation and improvement. One major issue is the class imbalance commonly observed in hyperspectral datasets, where certain classes contain relatively few and sparsely distributed samples. This often leads the model to focus on optimizing the classification accuracy for dominant classes, while overlooking underrepresented ones. Another limitation lies in the conversion of 2D spatial images into sequential inputs for the state-space mechanism (SSM). The current implementation employs simple horizontal or vertical scanning strategies, which may fail to capture more complex and semantically meaningful spatial dependencies.
To address these limitations, future work will focus on improving spectral feature extraction mechanisms, enabling the model to adaptively emphasize challenging categories during training. We also performed preliminary experiments to model spectral sequences directly using SSM, but the results were not promising. This finding suggests that introducing an effective feature extraction module before the SSM for refining and condensing spectral features could significantly enhance the model’s performance. Regarding spatial sequence construction, we plan to explore more expressive strategies, such as graph-based or tree-based transformations, to better preserve spatial structure and improve the representational capacity of the SSM module.

5. Conclusions

To further explore the application of Vision Mamba in hyperspectral imaging, we propose a hyperspectral classification framework, HyperSMamba. The model introduces a Multi-Scale State Fusion Module (MSFM) to enhance the recognition capability of features across multiple scales and integrate the spatial relationships of the state matrix in FusionSSM. Furthermore, the AFAttention module is employed to enhance and fuse spectral–spatial features, enabling the extraction of more discriminative information. Ablation studies demonstrate that the proposed MSFM and AFAttention are effective. The model achieves overall accuracies of 94.86%, 97.72%, and 97.38% on the IP, UP, and SA datasets, respectively. Both visual and quantitative analyses indicate that the model exhibits strong robustness and effectiveness in hyperspectral image classification tasks, fully leveraging features to improve classification accuracy. Although the model achieves excellent performance, it still faces several challenges, such as overcoming the limitations of causal properties of the state-space model, and further optimizing the feature representation capability of the Mamba architecture while maintaining high computational efficiency. These are key directions for our future research.

Author Contributions

Conceptualization, M.S.; methodology, M.S.; software, M.S.; validation, L.W., S.J. and S.C.; formal analysis, L.W., S.J. and S.C.; data curation, M.S.; writing—original draft preparation, M.S.; writing—review and editing, L.W., S.J., S.C. and L.T.; visualization, M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China (Grant No. 2022ZD0115802).

Data Availability Statement

The Indian Pines, Pavia University, and Salinas datasets are available at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 10 June 2025). The core code will be released at https://github.com/Mengyuan-Sun/HyperSMamba (accessed on 10 June 2025).

Conflicts of Interest

There are no conflicts of interest.

Appendix A

Appendix A.1. Visualization of Comparative Experimental Results

To provide a more intuitive comparison of performance and stability across different methods, we present a bar chart with error bars illustrating the variance in model outputs. The horizontal axis denotes the evaluated models, while the vertical axis represents the evaluation metric, expressed as a percentage (%). As shown in Figure A1, the proposed HyperSMamba model not only achieves higher evaluation values but also exhibits smaller error margins. These results demonstrate that HyperSMamba outperforms the compared methods in both accuracy and stability, validating its robustness and reliability in practical scenarios.
Figure A1. Bar plot comparison of OA (%), AA (%), and scaled κ ( × 100 ) for different classification methods on three datasets.
Figure A1. Bar plot comparison of OA (%), AA (%), and scaled κ ( × 100 ) for different classification methods on three datasets.
Remotesensing 17 02008 g0a1

Appendix A.2. Visualization of Model Complexity

Figure A2 presents a comparison of FLOPs and parameter counts (both plotted on a logarithmic scale) for different models evaluated on the Indian Pines (IP) dataset. The bar charts display the absolute values of FLOPs (M) and parameters (K), while the dashed lines highlight the overall trends across the models. This figure clearly demonstrates that HyperSMamba achieves both low computational complexity and low parameter count, significantly outperforming most baseline models in terms of efficiency.
Figure A2. A combined bar and line plot showing the number of parameters (K) and computational complexity (MFLOPs) of different comparison models on the IP dataset.
Figure A2. A combined bar and line plot showing the number of parameters (K) and computational complexity (MFLOPs) of different comparison models on the IP dataset.
Remotesensing 17 02008 g0a2

References

  1. Li, D.; Wu, J.; Zhao, J.; Xu, H.; Bian, L. SpectraTrack: Megapixel, Hundred-fps, and Thousand-channel Hyperspectral Imaging. Nat. Commun. 2024, 15, 9459. [Google Scholar] [CrossRef] [PubMed]
  2. Jia, J.W.; Wang, Y.H.; Chen, J.Y.; Guo, R.L.; Shu, R.M.; Wang, J.Q. Status and Application of Advanced Airborne Hyperspectral Imaging Technology: A Review. Infrared Phys. Technol. 2020, 104, 103115. [Google Scholar] [CrossRef]
  3. Bhargava, A.; Sachdeva, A.; Sharma, K.; Alsharif, M.H.; Uthansakul, P.; Uthansakul, M. Hyperspectral imaging and its applications: A review. Heliyon 2024, 10, 33208. [Google Scholar] [CrossRef]
  4. Ahmad, M.; Shabbir, S.; Roy, S.K.; Hong, D.; Wu, X.; Yao, J. Hyperspectral Image Classification—Traditional to Deep Models: A Survey for Future Prospects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 968–999. [Google Scholar] [CrossRef]
  5. Zhang, W.; Zhao, L. The Track, Hotspot and Frontier of International Hyperspectral Remote Sensing Research 2009–2019—A Bibliometric Analysis Based on SCI Database. Measurement 2022, 187, 110229. [Google Scholar] [CrossRef]
  6. Ma, X.; Wang, H.; Geng, J. Spectral–spatial Classification of Hyperspectral Image Based on Deep Auto-encoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  7. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  8. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Cai, W.; Yu, C.; Yang, N.; Cai, W. Multi-feature Fusion: Graph Neural Network and CNN Combining for Hyperspectral Image Classification. Neurocomputing 2022, 501, 246–257. [Google Scholar] [CrossRef]
  9. Song, T.; Wang, Y.; Gao, C.; Chen, H.; Li, J. MSLAN: A Two-branch Multi-directional Spectral–Spatial LSTM Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528814. [Google Scholar] [CrossRef]
  10. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  11. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral Image Classification Using Group-Aware Hierarchical Transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  13. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  14. Li, Y.; Zhang, H.; Shen, Q. Spectral–Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  15. Zhou, J.; Zeng, S.; Xiao, Z.; Zhou, J.; Li, H.; Kang, Z. An Enhanced Spectral Fusion 3D CNN Model for Hyperspectral Image Classification. Remote Sens. 2022, 14, 5334. [Google Scholar] [CrossRef]
  16. Xu, Z.; Su, C.; Wang, S.; Zhang, X. Local and Global Spectral Features for Hyperspectral Image Classification. Remote Sens. 2023, 15, 1803. [Google Scholar] [CrossRef]
  17. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3D-2D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  18. Chen, J.; Wang, X.J.; Guo, Z.C.; Zhang, X.Y.; Sun, J. Dynamic Region-aware Convolution. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar] [CrossRef]
  19. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  20. Zhao, Z.; Hu, D.; Wang, H.; Yu, X. Convolutional Transformer Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  21. He, X.; Chen, Y.; Lin, Z. Spatial-spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  22. Ma, C.; Jiang, J.; Li, H.; Mei, X.; Bai, C. Hyperspectral Image Classification Via Spectral Pooling and Hybrid Transformer. Remote Sens. 2022, 14, 4732. [Google Scholar] [CrossRef]
  23. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  24. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral Image Transformer Classification Networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  25. Zhou, W.; Kamata, S.I.; Luo, Z.; Chen, X. Hierarchical Unified Spectral-Spatial Aggregated Transformer for Hyperspectral Image Classification. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022. [Google Scholar] [CrossRef]
  26. Gu, A.; Goel, K.; Ré, C. Efficiently Modeling Long Sequences with Structured State Spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar] [CrossRef]
  27. Gu, A.; Dao, T. Mamba: Linear-time Sequence Modeling with Selective State Spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar] [CrossRef]
  28. Zhu, L.; Liao, B.; Zhang, Q.; Wang, X.; Liu, W.; Wang, X. Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model. arXiv 2024, arXiv:2401.09417. [Google Scholar] [CrossRef]
  29. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Jiao, J.; Liu, Y. Vmamba: Visual State Space Model. arXiv 2024, arXiv:2401.10166. [Google Scholar] [CrossRef]
  30. Pei, X.; Huang, T.; Xu, C. Efficientvmamba: Atrous Selective Scan for Lightweight Visual Mamba. arXiv 2024, arXiv:2403.09977. [Google Scholar] [CrossRef]
  31. Huang, T.; Pei, X.; You, S.; Wang, F.; Qian, C.; Xu, C. Localmamba: Visual State Space Model with Windowed Selective Scan. arXiv 2024, arXiv:2403.09338. [Google Scholar] [CrossRef]
  32. Yao, J.; Hong, D.; Li, C.; Chanussot, J. Spectralmamba: Efficient Mamba for Hyperspectral Image Classification. arXiv 2024, arXiv:2404.08489. [Google Scholar] [CrossRef]
  33. Yang, J.X.; Zhou, J.; Wang, J.; Tian, H.; Liew, A.W.C. Hsimamba: Hyperspectral Imaging Efficient Feature Learning with Bidirectional State Space for Classification. arXiv 2024, arXiv:2404.00272. [Google Scholar] [CrossRef]
  34. Chen, K.; Chen, B.; Liu, C.; Li, W.; Zou, Z.; Shi, Z. Rsmamba: Remote Sensing Image Classification with State Space Model. arXiv 2024, arXiv:2403.19654. [Google Scholar] [CrossRef]
  35. Zhuang, P.; Zhang, X.; Wang, H.; Zhang, T.; Liu, L.; Li, J. FAHM: Frequency-Aware Hierarchical Mamba for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 6299. [Google Scholar] [CrossRef]
  36. Zhou, W.; Kamata, S.I.; Wang, H.; Wong, M.S.; Hou, H.C. Mamba-in-Mamba: Centralized Mamba-Cross-Scan in Tokenized Mamba Model for Hyperspectral Image Classification. Neurocomputing 2025, 613, 128751. [Google Scholar] [CrossRef]
  37. Huang, L.; Chen, Y.; He, X. Spectral-Spatial Mamba for Hyperspectral Image Classification. Remote Sens. 2024, 16, 2449. [Google Scholar] [CrossRef]
  38. Liu, X.; Zhang, C.; Zhang, L. Vision Mamba: A Comprehensive Survey and Taxonomy. arXiv 2024, arXiv:2405.04404. [Google Scholar] [CrossRef]
  39. Xu, R.; Yang, S.; Wang, Y.; Cai, Y.; Du, B.; Chen, H. Visual Mamba: A Survey and New Outlooks. arXiv 2024, arXiv:2404.18861. [Google Scholar] [CrossRef]
  40. Ma, P.; Ren, J.; Zhao, H.; Sun, G.; Murray, P.; Zheng, J. Multiscale 2-D Singular Spectrum Analysis and Principal Component Analysis for Spatial-Spectral Noise-robust Feature Extraction and Classification of Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1233–1245. [Google Scholar] [CrossRef]
  41. Yang, A.; Li, M.; Ding, Y.; Hong, D.; Lv, Y.; He, Y. GTFN: GCN and Transformer Fusion Network with Spatial-Spectral Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  42. Dao, T.; Gu, A. Transformers Are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality. arXiv 2024, arXiv:2405.21060. [Google Scholar] [CrossRef]
  43. Gu, A.; Johnson, I.; Goel, K.; Saab, K.; Dao, T.; Rudra, A. Combining Recurrent, Convolutional, and Continuous-Time Models with Linear State Space Layers. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), San Diego, CA, USA, 6–14 December 2021. [Google Scholar]
  44. Piqueras, S.; Burger, J.; Tauler, R.; de Juan, A. Relevant Aspects of Quantification and Sample Heterogeneity in Hyperspectral Image Resolution. Chemom. Intell. Lab. Syst. 2012, 117, 169–182. [Google Scholar] [CrossRef]
  45. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar] [CrossRef]
  46. Zhu, X.; Cheng, D.; Zhang, Z.; Lin, S.; Dai, J. An Empirical Study of Spatial Attention Mechanisms in Deep Networks. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of HyperSMamba.
Figure 1. The overall architecture of HyperSMamba.
Remotesensing 17 02008 g001
Figure 2. Illustration of the Multi-Scale State Fusion module.
Figure 2. Illustration of the Multi-Scale State Fusion module.
Remotesensing 17 02008 g002
Figure 3. (a) The structure of the Adaptive Fusion Attention Module, which comprises the following three components: (b) spatial attention; (c) pixel attention; (d) channel attention.
Figure 3. (a) The structure of the Adaptive Fusion Attention Module, which comprises the following three components: (b) spatial attention; (c) pixel attention; (d) channel attention.
Remotesensing 17 02008 g003
Figure 4. Land cover maps of all comparative experiments on the Indian Pines dataset: (a) ground truth, (b) 2D CNN, (c) HybridSN, (d) SpectralFormer, (e) HiT, (f) HUSST, (g) SSFTT, (h) ViM, (i) MiM, (j) SS-Mamba, (k) FAHM, (l) HyperSMamba.
Figure 4. Land cover maps of all comparative experiments on the Indian Pines dataset: (a) ground truth, (b) 2D CNN, (c) HybridSN, (d) SpectralFormer, (e) HiT, (f) HUSST, (g) SSFTT, (h) ViM, (i) MiM, (j) SS-Mamba, (k) FAHM, (l) HyperSMamba.
Remotesensing 17 02008 g004
Figure 5. Land cover maps of all comparative experiments on the Pavia University dataset: (a) ground truth, (b) 2D CNN, (c) HybridSN, (d) SpectralFormer, (e) HiT, (f) HUSST, (g) SSFTT, (h) ViM, (i) MiM, (j) SS-Mamba, (k) FAHM, (l) HyperSMamba.
Figure 5. Land cover maps of all comparative experiments on the Pavia University dataset: (a) ground truth, (b) 2D CNN, (c) HybridSN, (d) SpectralFormer, (e) HiT, (f) HUSST, (g) SSFTT, (h) ViM, (i) MiM, (j) SS-Mamba, (k) FAHM, (l) HyperSMamba.
Remotesensing 17 02008 g005
Figure 6. Land cover maps of all comparative experiments on the Salinas dataset: (a) ground truth, (b) 2D CNN, (c) HybridSN, (d) SpectralFormer, (e) HiT, (f) HUSST, (g) SSFTT, (h) ViM, (i) MiM, (j) SS-Mamba, (k) FAHM, (l) HyperSMamba.
Figure 6. Land cover maps of all comparative experiments on the Salinas dataset: (a) ground truth, (b) 2D CNN, (c) HybridSN, (d) SpectralFormer, (e) HiT, (f) HUSST, (g) SSFTT, (h) ViM, (i) MiM, (j) SS-Mamba, (k) FAHM, (l) HyperSMamba.
Remotesensing 17 02008 g006
Figure 7. Visualization of T-SNE results on the Pavia University dataset. (a) 2D CNN, (b) HybridSN, (c) SpectralFormer, (d) HiT, (e) HUSST, (f) SSFTT, (g) ViM, (h) MiM, (i) SS-Mamba, (j) FAHM, (k) HyperSMamba.
Figure 7. Visualization of T-SNE results on the Pavia University dataset. (a) 2D CNN, (b) HybridSN, (c) SpectralFormer, (d) HiT, (e) HUSST, (f) SSFTT, (g) ViM, (h) MiM, (i) SS-Mamba, (j) FAHM, (k) HyperSMamba.
Remotesensing 17 02008 g007
Figure 8. Impacts of the different parameters on overall accuracy (OA) for three datasets: (a) Impact of the patch size. (b) Impact of the number of retained bands after PCA.
Figure 8. Impacts of the different parameters on overall accuracy (OA) for three datasets: (a) Impact of the patch size. (b) Impact of the number of retained bands after PCA.
Remotesensing 17 02008 g008
Table 1. Description of land cover types and number of train–test samples in the Indian Pines dataset.
Table 1. Description of land cover types and number of train–test samples in the Indian Pines dataset.
Class IDCategoryTrainingTestingTotal
1Alfalfa153146
2Corn-notill4013881428
3Corn-mintill40790830
4Corn40197237
5Grass-pasture40443483
6Grass-trees40690730
7Grass-pasture-mowed151328
8Hay-windrowed40438478
9Oats15520
10Soybeans-notill40932972
11Soybean-mintill4024152455
12Soybean-clean40553593
13Wheat40165205
14Woods4012251265
15Build-grass-trees-drivers40346386
16Stones-steel-towers405393
Total565968410,249
Table 2. Description of land cover types and number of train–test samples in the Pavia University dataset.
Table 2. Description of land cover types and number of train–test samples in the Pavia University dataset.
Class IDCategoryTrainingTestingTotal
1Asphalt3066016631
2Meadows3018,61918,649
3Gravel3020692099
4Trees3030343064
5Metal Sheets3013151345
6Bare-soil3049995029
7Bitumen3013001300
8Bricks3036523682
9Shadows30917947
Total27042,50642,776
Table 3. Description of land cover types and number of train–test samples in the Salinas dataset.
Table 3. Description of land cover types and number of train–test samples in the Salinas dataset.
Class IDCategoryTrainingTestingTotal
1Brocoli_green_weeds_13019792009
2Brocoli_green_weeds_23036963726
3Fallow3019461976
4Fallow_rough_plow3013641394
5Fallow_smooth3026482678
6Stubble3039293959
7Celery3035493579
8Grapes_untrained3011,24111,271
9Soil_vinyard_develop3061736203
10SCorn_senesced_green_weeds3032483278
11Lettuce_romaine_4wk3010381068
12Lettuce_romaine_5wk3018971927
13Lettuce_romaine_6wk30886916
14Lettuce_romaine_7wk3010401070
15Vinyard_untrained3072387268
16Vinyard_vertical_trellis3017771807
Total48053,64954,129
Table 4. Quantitative performance of different classification methods in terms of OA (%), AA (%), scaled κ ( × 100 ), and per-class accuracy (%) on the Indian Pines dataset. The best results are highlighted in bold.
Table 4. Quantitative performance of different classification methods in terms of OA (%), AA (%), scaled κ ( × 100 ), and per-class accuracy (%) on the Indian Pines dataset. The best results are highlighted in bold.
Class
No.
2D
CNN
Hybrid
SN
Spectral
Former
HiTHUSSTSSFTTViMMiMSS-
Mamba
FAHMHyper-
SMamba
186.21 ± 3.3492.83 ± 6.3097.29 ± 3.2785.05 ± 2.3580.88 ± 1.8789.13 ± 9.8289.79 ± 5.0595.65 ± 7.4999.38 ± 1.8895.90 ± 4.0899.10 ± 1.24
271.86 ± 2.1386.76 ± 1.8879.15 ± 2.7481.89 ± 1.4282.16 ± 0.9488.94 ± 2.1777.44 ± 1.3083.40 ± 8.8383.65 ± 5.7388.78 ± 2.0089.99 ± 2.04
369.52 ± 2.5390.27 ± 2.9385.73 ± 4.0185.59 ± 0.9287.55 ± 1.6588.36 ± 2.9282.19 ± 2.4978.80 ± 6.9492.10 ± 3.6095.49 ± 1.2292.88 ± 1.31
464.92 ± 6.2188.17 ± 5.1680.41 ± 4.9486.42 ± 1.7392.52 ± 0.9292.72 ± 5.4484.04 ± 2.9977.64 ± 10.2699.23 ± 2.1693.10 ± 4.1694.54 ± 2.29
580.43 ± 3.7591.10 ± 2.5293.02 ± 3.3889.74 ± 1.1295.37 ± 0.9792.28 ± 1.9388.53 ± 3.6286.13 ± 3.2694.06 ± 1.9197.00 ± 1.8997.81 ± 1.12
693.23 ± 0.9895.20 ± 3.6798.63 ± 0.6295.02 ± 0.7397.75 ± 0.3695.30 ± 3.0395.54 ± 1.0596.30 ± 0.7898.89 ± 0.4998.88 ± 0.7199.42 ± 0.26
781.67 ± 5.7790.80 ± 9.6388.22 ± 9.2589.12 ± 5.1095.44 ± 4.2775.86 ± 2.6493.43 ± 4.6382.14 ± 7.14100.00 ± 0.0091.48 ± 8.3695.32 ± 2.68
895.63 ± 0.3899.24 ± 0.8699.31 ± 0.7796.87 ± 0.4999.39 ± 0.5998.45 ± 1.1496.59 ± 0.6198.95 ± 0.63100.00 ± 0.0099.75 ± 0.34100.00 ± 0.00
970.18 ± 12.2472.90 ± 19.6273.92 ± 14.0760.29 ± 9.0477.14 ± 8.1151.29 ± 32.9174.60 ± 14.1170.00 ± 0.15100.00 ± 0.0093.22 ± 1.9994.87 ± 4.74
1066.41 ± 3.1780.88 ± 4.2878.85 ± 6.7078.00 ± 1.0179.93 ± 1.3783.51 ± 1.4074.11 ± 2.0782.10 ± 5.2591.15 ± 4.7694.03 ± 3.0890.58 ± 0.78
1176.38 ± 2.0886.49 ± 2.4480.25 ± 5.6288.99 ± 1.2184.74 ± 1.3591.68 ± 1.7686.07 ± 1.0684.93 ± 5.3789.81 ± 5.0991.33 ± 2.4591.72 ± 2.83
1268.28 ± 5.0787.42 ± 3.3772.95 ± 4.4679.06 ± 2.2483.53 ± 2.3491.24 ± 4.2179.84 ± 1.9684.49 ± 4.8994.07 ± 4.2093.92 ± 4.2694.34 ± 0.99
1395.03 ± 1.9597.08 ± 2.3598.40 ± 2.2396.70 ± 0.8098.80 ± 0.5896.40 ± 2.5894.81 ± 1.3095.61 ± 3.16100.00 ± 0.0099.64 ± 0.5597.71 ± 1.58
1496.14 ± 1.0597.37 ± 1.0397.81 ± 0.6497.22 ± 0.3598.82 ± 0.2897.30 ± 0.9496.40 ± 0.5496.60 ± 1.4298.86 ± 0.6498.74 ± 0.9799.35 ± 0.35
1585.14 ± 7.1489.30 ± 2.7088.70 ± 4.1090.77 ± 1.7795.60 ± 1.1992.26 ± 4.8086.84 ± 3.1791.71 ± 2.6097.33 ± 2.4698.13 ± 1.4597.82 ± 1.57
1680.49 ± 3.1388.42 ± 10.3386.80 ± 8.8363.85 ± 2.6089.34 ± 2.1181.54 ± 4.5870.19 ± 5.6684.95 ± 7.9899.05 ± 1.0590.92 ± 2.3695.75 ± 3.68
OA (%)79.06 ± 1.8489.50 ± 1.4885.69 ± 3.1487.89 ± 1.4788.79 ± 0.6991.38 ± 1.2585.64 ± 0.7187.20 ± 1.9992.94 ± 1.9193.06 ± 1.2394.86 ± 0.67
AA (%)80.84 ± 2.7394.70 ± 0.7293.18 ± 1.4692.96 ± 1.3094.83 ± 0.2794.31 ± 1.0087.30 ± 1.4186.83 ± 2.8496.37 ± 0.8896.71 ± 0.4697.60 ± 0.20
κ × 100 76.13 ± 1.9688.02 ± 1.6783.79 ± 3.4586.21 ± 1.5287.21 ± 0.7790.17 ± 1.4183.67 ± 0.7485.41 ± 0.7992.00 ± 2.1693.07 ± 1.3994.13 ± 0.76
Table 5. Quantitative performance of different classification methods in terms of OA (%), AA (%), scaled κ ( × 100 ), and per-class accuracy (%) on the Pavia University dataset. The best results are highlighted in bold.
Table 5. Quantitative performance of different classification methods in terms of OA (%), AA (%), scaled κ ( × 100 ), and per-class accuracy (%) on the Pavia University dataset. The best results are highlighted in bold.
Class
No.
2D
CNN
Hybrid
SN
Spectral
Former
HiTHUSSTSSFTTViMMiMSS-
Mamba
FAHMHyper-
SMamba
177.86 ± 1.4875.40 ± 6.3888.55 ± 2.5778.53 ± 5.7184.75 ± 1.5189.42 ± 7.1575.39 ± 2.3185.99 ± 2.0494.14 ± 3.4795.64 ± 1.3696.77 ± 0.58
288.67 ± 0.8994.62 ± 3.7094.80 ± 1.5594.07 ± 2.7897.64 ± 0.3696.95 ± 2.1694.78 ± 0.6487.17 ± 1.5990.31 ± 7.3498.95 ± 0.3999.66 ± 0.10
380.59 ± 6.1876.78 ± 6.0672.20 ± 4.5075.63 ± 7.4387.60 ± 1.1285.69 ± 6.3573.61 ± 2.7383.46 ± 7.3593.67 ± 11.9895.44 ± 2.0193.09 ± 1.67
479.70 ± 1.3383.52 ± 4.8093.69 ± 1.8078.94 ± 5.0477.36 ± 0.6989.45 ± 2.1181.08 ± 1.3285.67 ± 5.5599.09 ± 0.3895.19 ± 1.0895.15 ± 0.67
588.97 ± 1.8999.33 ± 0.6899.85 ± 0.0697.88 ± 0.4596.12 ± 0.3298.59 ± 1.3498.57 ± 0.2299.33 ± 0.64100.00 ± 0.0098.16 ± 1.0599.34 ± 0.10
685.29 ± 1.5791.26 ± 5.4684.87 ± 4.4684.28 ± 5.9395.28 ± 1.1790.60 ± 6.1784.87 ± 1.7193.96 ± 4.3799.43 ± 0.2398.28 ± 1.3299.58 ± 0.16
783.54 ± 1.8287.51 ± 9.6289.94 ± 2.7193.23 ± 3.9198.35 ± 0.4694.84 ± 3.6190.32 ± 3.5195.64 ± 3.4499.96 ± 0.1298.40 ± 0.8299.02 ± 0.78
861.33 ± 4.1268.54 ± 7.0973.30 ± 3.8486.90 ± 4.3579.49 ± 1.0479.40 ± 7.2766.93 ± 2.5184.68 ± 2.4798.15 ± 3.0995.64 ± 2.2492.23 ± 1.42
973.72 ± 1.9068.72 ± 12.1093.76 ± 2.9365.30 ± 12.5580.47 ± 1.8185.17 ± 9.7170.72 ± 3.7986.38 ± 9.7699.99 ± 0.0394.24 ± 1.7896.20 ± 1.06
OA (%)83.72 ± 0.6986.59 ± 4.2089.45 ± 1.9685.50 ± 3.7791.49 ± 0.4392.12 ± 4.0483.80 ± 0.9187.91 ± 1.8194.30 ± 3.3995.63 ± 0.6397.72 ± 0.31
AA (%)79.96 ± 2.0987.01 ± 3.4789.87 ± 1.1984.86 ± 2.2790.04 ± 0.2691.18 ± 3.3283.63 ± 2.0189.15 ± 1.1697.25 ± 1.8094.94 ± 0.5996.98 ± 0.47
κ × 100 77.11 ± 0.9182.44 ± 5.3786.11 ± 2.5181.07 ± 4.6788.75 ± 0.5589.56 ± 5.1380.20 ± 1.2083.98 ± 2.1692.62 ± 4.2694.21 ± 0.8396.97 ± 0.41
Table 6. Quantitative performance of different classification methods in terms of OA (%), AA (%), scaled κ ( × 100 ), and per-class accuracy (%) on the Salinas dataset. The best results are highlighted in bold.
Table 6. Quantitative performance of different classification methods in terms of OA (%), AA (%), scaled κ ( × 100 ), and per-class accuracy (%) on the Salinas dataset. The best results are highlighted in bold.
Class
No.
2D
CNN
Hybrid
SN
Spectral
Former
HiTHUSSTSSFTTViMMiMSS-
Mamba
FAHMHyper-
SMamba
199.51 ± 0.8799.98 ± 0.0294.19 ± 1.78100.00 ± 0.00100.00 ± 0.0099.68 ± 0.5696.84 ± 1.1995.81 ± 0.12100.00 ± 0.00100.00 ± 0.0099.99 ± 0.01
299.60 ± 0.5399.93 ± 0.0698.91 ± 0.0899.34 ± 0.4999.56 ± 0.2699.65 ± 0.3498.12 ± 1.0199.95 ± 0.05100.00 ± 0.0099.97 ± 0.0499.99 ± 0.01
398.59 ± 0.9699.75 ± 0.2492.02 ± 0.7499.82 ± 0.1699.99 ± 0.0199.29 ± 1.1799.22 ± 0.0199.85 ± 0.0299.86 ± 0.40100.00 ± 0.00100.00 ± 0.00
498.67 ± 0.3398.61 ± 1.3197.16 ± 1.2597.49 ± 0.1498.12 ± 0.4295.79 ± 0.9597.46 ± 0.6596.48 ± 1.6999.73 ± 0.3198.92 ± 1.0198.19 ± 0.89
597.37 ± 0.8298.78 ± 0.8189.23 ± 1.2598.62 ± 0.2298.48 ± 0.2697.06 ± 2.1697.97 ± 0.5798.24 ± 1.0197.87 ± 2.0199.23 ± 0.3399.42 ± 0.26
699.95 ± 0.0599.68 ± 0.2499.67 ± 0.5599.43 ± 0.2199.72 ± 0.1398.13 ± 1.0199.05 ± 0.3098.38 ± 0.0999.97 ± 0.0399.52 ± 0.4899.92 ± 0.06
799.59 ± 0.3298.74 ± 0.7197.80 ± 0.2099.41 ± 0.2399.47 ± 0.1498.37 ± 1.0498.98 ± 0.3398.66 ± 0.0799.83 ± 0.1199.44 ± 0.5299.93 ± 0.07
886.41 ± 1.5490.29 ± 4.5573.06 ± 2.4783.58 ± 3.6186.43 ± 1.5089.16 ± 2.8380.38 ± 4.7088.66 ± 1.2782.19 ± 17.6593.37 ± 4.9194.42 ± 2.14
999.84 ± 0.1799.15 ± 0.6098.76 ± 0.1299.96 ± 0.0299.97 ± 0.0298.99 ± 1.1197.96 ± 1.0499.58 ± 0.0699.99 ± 0.01100.00 ± 0.0099.83 ± 0.07
1097.63 ± 1.0696.27 ± 1.7994.40 ± 1.6599.49 ± 0.1298.53 ± 0.6496.44 ± 1.6096.85 ± 1.4497.53 ± 0.9999.32 ± 0.6499.73 ± 0.1998.20 ± 1.09
1198.65 ± 0.7696.36 ± 1.5198.73 ± 0.5098.87 ± 0.0298.72 ± 0.4598.49 ± 1.7698.70 ± 0.1597.00 ± 1.7199.74 ± 0.2199.95 ± 0.0598.60 ± 1.64
1299.44 ± 0.5098.67 ± 1.3898.86 ± 0.4898.46 ± 0.1798.90 ± 0.3597.37 ± 2.5897.45 ± 0.2597.09 ± 0.5399.89 ± 0.3399.94 ± 0.0399.83 ± 0.15
1398.81 ± 0.5985.54 ± 5.5697.86 ± 0.6397.13 ± 0.5498.39 ± 0.5794.84 ± 3.3196.10 ± 0.6896.94 ± 1.0299.86 ± 0.1899.81 ± 0.1399.47 ± 0.42
1498.43 ± 0.5697.40 ± 4.0498.57 ± 0.8797.76 ± 0.4799.14 ± 0.1693.05 ± 5.4396.35 ± 0.6296.54 ± 1.5899.26 ± 0.9599.31 ± 0.3297.78 ± 1.74
1579.16 ± 2.3863.28 ± 3.2578.46 ± 4.8075.76 ± 4.1378.58 ± 4.9184.85 ± 2.4872.26 ± 3.1786.03 ± 1.9389.56 ± 18.5890.08 ± 5.4191.49 ± 3.31
1699.43 ± 0.1599.49 ± 0.3596.48 ± 0.3499.11 ± 0.7699.26 ± 0.3498.82 ± 1.2297.19 ± 1.0197.34 ± 0.0699.80 ± 0.2899.87 ± 0.1999.91 ± 0.16
OA (%)88.50 ± 0.4692.07 ± 1.0689.13 ± 1.5192.84 ± 0.9493.84 ± 0.9494.74 ± 1.2590.85 ± 1.0894.65 ± 0.8894.66 ± 3.3696.03 ± 0.9597.38 ± 0.89
AA (%)92.57 ± 2.9195.12 ± 1.2193.76 ± 0.6596.53 ± 1.5797.11 ± 0.4597.18 ± 0.8595.07 ± 2.2996.52 ± 3.6997.94 ± 1.2498.64 ± 0.4598.78 ± 0.37
κ × 100 87.21 ± 0.5291.18 ± 1.1887.90 ± 1.6992.02 ± 1.0493.14 ± 0.2794.15 ± 1.3889.81 ± 1.2194.04 ± 0.4394.07 ± 3.7296.69 ± 0.4297.08 ± 0.10
Table 7. Ablation study on the impact of different modules in HyperSMamba on OA (%), AA (%), and scaled κ ( × 100 ) across three datasets. The best results are highlighted in bold.
Table 7. Ablation study on the impact of different modules in HyperSMamba on OA (%), AA (%), and scaled κ ( × 100 ) across three datasets. The best results are highlighted in bold.
MSFMAFAttentionIndian PinesPavia UniversitySalinas
OA (%)AA (%) κ × 100 OA (%)AA (%) κ × 100 OA (%)AA (%) κ × 100
××86.4988.4985.2185.9786.1382.6894.0297.4094.08
×90.6494.2389.5893.9292.9492.9196.4498.5496.04
×92.8696.6191.6296.6795.6795.5996.8298.6896.46
94.8697.6094.1397.7296.9896.9797.3898.7897.08
Table 8. Ablation study of convolutional components in Multi-Scale State Fusion Module (MSFM) on OA (%) across three datasets. The best results are highlighted in bold.
Table 8. Ablation study of convolutional components in Multi-Scale State Fusion Module (MSFM) on OA (%) across three datasets. The best results are highlighted in bold.
3 × 3
DWConv
3 × 3
DDConv
5 × 5
DDConv
Indian PinesPavia UniversitySalinas
×××90.6493.9296.44
×92.2196.1096.77
×94.0196.2796.71
×92.8596.7397.09
94.8697.7297.38
Table 9. The number of parameters (K) and computational complexity (MFLOPs) of different comparison experiments on the IP dataset.
Table 9. The number of parameters (K) and computational complexity (MFLOPs) of different comparison experiments on the IP dataset.
Methods2D CNNHybrid-SNSpectral-FormerHiTSSFTTHUSSTViMMiMSS-MambaFAHMHyper-Mamba
FLOPs (M)68.53512.9420.77587.9259.2514.9612.02139.430.8944.1216.93
Parameters (K)537.875490.43407.9640,670.19875.221310.7179.87217.36360.62784.74110.88
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, M.; Wang, L.; Jiang, S.; Cheng, S.; Tang, L. HyperSMamba: A Lightweight Mamba for Efficient Hyperspectral Image Classification. Remote Sens. 2025, 17, 2008. https://doi.org/10.3390/rs17122008

AMA Style

Sun M, Wang L, Jiang S, Cheng S, Tang L. HyperSMamba: A Lightweight Mamba for Efficient Hyperspectral Image Classification. Remote Sensing. 2025; 17(12):2008. https://doi.org/10.3390/rs17122008

Chicago/Turabian Style

Sun, Mengyuan, Liejun Wang, Shaochen Jiang, Shuli Cheng, and Lihan Tang. 2025. "HyperSMamba: A Lightweight Mamba for Efficient Hyperspectral Image Classification" Remote Sensing 17, no. 12: 2008. https://doi.org/10.3390/rs17122008

APA Style

Sun, M., Wang, L., Jiang, S., Cheng, S., & Tang, L. (2025). HyperSMamba: A Lightweight Mamba for Efficient Hyperspectral Image Classification. Remote Sensing, 17(12), 2008. https://doi.org/10.3390/rs17122008

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop