Next Article in Journal
GDEIM-SF: A Lightweight UAV Detection Framework Coupling Dehazing and Low-Light Enhancement
Next Article in Special Issue
QHAWAY: An Instance Segmentation and Monocular Distance Estimation ADAS for Vulnerable Road Users in Informal Andean Urban Corridors
Previous Article in Journal
PSO-Based Dynamic RSU Role Assignment Framework for Scalable and Reliable Content Delivery in VANETs
Previous Article in Special Issue
Benchmark Dataset and Deep Model for Monocular Camera Calibration from Single Highway Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HL-Mamba: A High–Low Frequency Interaction Mamba Network for Hyperspectral Image Classification

1
School of Land and Resources Engineering, Kunming University of Science and Technology, Kunming 650093, China
2
Application Engineering Research Center of Spatial Information Surveying and Mapping Technology in Plateau and Mountainous Areas Set by Universities in Yunnan Province, Kunming 650093, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(5), 1556; https://doi.org/10.3390/s26051556
Submission received: 17 January 2026 / Revised: 26 February 2026 / Accepted: 26 February 2026 / Published: 2 March 2026

Abstract

Deep-learning-based methods have achieved remarkable success in hyperspectral image (HSI) classification tasks due to their promising ability. However, the high dimensionality and spectral–spatial correlations of HSIs usually lead to information redundancy and feature entanglement, limiting the classification performance. To address these issues, we propose a novel high–low frequency interaction Mamba network, called HL-Mamba, which achieves effective decoupling and interaction between global structures and edge details of HSIs in the frequency domain, thereby improving spectral–spatial representation for HSI classification. Specifically, a high–low frequency decomposition Mamba module is designed to decompose the HSI into low-frequency structural and high-frequency edge detail components, which allows the model to learn global structures and fine-grained details, enhancing classification performance. By employing two parallel Mamba branches to model long-range dependencies across different frequency components, the network achieves efficient global modeling while mitigating information redundancy. Furthermore, a cross-frequency interaction module is designed to establish complementary information flow between high- and low-frequency features through a dynamic attention mechanism. In this way, low-frequency structural features guide the aggregation of high-frequency details, whereas high-frequency textures refine global structural representations, yielding more discriminative spectral–spatial features for HSI classification. In addition, a frequency alignment loss is designed to enhance the consistency and complementarity between high- and low-frequency features, further improving classification performance. Extensive experiments on four public benchmark datasets (i.e., Indian Pines, Pavia University, WHU-Hi-HanChuan, and Houston datasets) demonstrate that the proposed HL-Mamba significantly outperforms eight comparison methods, achieving an overall accuracy of 94.07%, 93.82%, 95.28%, and 87.32%, respectively. Ablation studies further verify the effectiveness of core component within the network.

1. Introduction

Hyperspectral imaging technology has emerged as one of the most significant research topics in remote sensing, as it provides rich spectral and spatial information through hundreds of contiguous spectral bands [1,2]. The remarkable potential of hyperspectral images (HSIs) plays a pivotal role in diverse applications, such as environmental monitoring [3], agricultural management [4], mineral exploration [5], and urban planning [6]. HSI classification serves as the most critical step in many of these applications, aiming to assign a distinct land cover category to each pixel [7].
Over the past decade, numerous methods have been proposed to enhance the performance of HSI classification [8]. Early methods mainly relied on traditional machine learning techniques such as support vector machines [9], k-nearest neighbors [10], and random forests [11], which primarily focused on spectral information. However, these methods often ignored spatial correlations, leading to spectral confusion and classification errors in complex scenes. To mitigate this limitation, spatial–spectral feature extraction methods are introduced, incorporating local spatial context via morphological profiles, Markov random fields [12], or principal component analysis (PCA)-based dimensionality reduction [13]. Nevertheless, these hand-crafted spatial–spectral feature extraction strategies still suffer from limited capability to adaptively capture the complex nonlinear relationships embedded in HSI data.
With the rapid advancements in deep learning, numerous deep-learning-based models have been proposed to improve HSI classification by jointly learning hierarchical spectral–spatial representations, which have shown significant superiority over traditional hand-crafted feature methods [14,15]. Among them, convolutional neural networks (CNNs) have gained significant attention due to their strong ability to capture local spatial–spectral dependencies of HSIs [16]. For instance, 2D CNNs and 3D CNNs are widely employed to learn hierarchical feature representations from pixel values and spectral reflectance sequences [17]. Yang et al. proposed an enhanced multiscale feature fusion network, which extracts multiscale features from parallel multipath structures across three stages to boost classification performance [18]. In addition to CNNs, recurrent neural networks and long short-term memory networks have been applied to HSI data to model sequential dependencies along the spectral dimension, fully leveraging the intrinsic correlation between contiguous spectral bands [19,20,21]. More recently, Transformer-based models, known for their powerful self-attention mechanisms, have also been adapted for HSI classification, achieving promising results by capturing long-range dependencies between spectral bands and spatial regions [22,23]. For example, Yang et al. proposed a center-to-surrounding interactive learning network via integrating Transformer architecture for more effective spectral–spatial feature learning [24]. However, the self-attention mechanism in standard Transformers incurs a quadratic computational complexity with the number of pixels, resulting in substantial computational redundancy and limiting its efficiency in processing HSI data.
To alleviate the above computational redundancy issue, subsequent studies have explored a series of efficient model designs, including CNN variants with depthwise separable convolutions [25], compact RNN architectures with gated mechanisms, and improved Transformer models with sparse attention or linear attention [26]. These methods reduce computational overhead for HSI classification tasks, enabling efficient processing of high-dimensional HSI data [27]. Meanwhile, state space models (SSMs) represented by Mamba have also been introduced into HSI classification recently, due to their linear computational complexity and efficient long-range dependency modeling capability. For instance, Wang et al. proposed MambaHSI+, a framework that integrates bidirectional state-space modeling with spectral trajectory learning for HSI classification [28]. Liu et al. proposed a dual-classification-head self-training network that performs class-wise feature alignment across domains, achieving strong cross-domain classification performance [29]. Despite their success in efficiency and accuracy, these deep-learning-based methods still face nonnegligible limitations. They rely on complex network architectures and large-scale training data, and more importantly, they inherently entangle global structural features with fine-grained edge details in the feature learning process, leading to redundant and suboptimal feature representations, and making it difficult to simultaneously capture both global structural information and local fine-grained textures.
To address the feature entanglement between global structure and local details, frequency-based methods have recently emerged as a promising paradigm for HSI classification, by explicitly decomposing HSI data into different frequency components corresponding to structural and detail information, respectively [30]. Early attempts at frequency-based analysis for HSI classification mainly adopted Fourier transform and wavelet transform to decompose HSI into multiresolution frequency components, enabling multiscale feature extraction [31]. Wavelet-based methods have shown unique advantages in handling the high dimensionality of HSIs, as they support the decomposition of spectral and spatial features at different scales [32]. However, these methods typically rely on hand-crafted frequency decomposition and feature extraction strategies, which may limit their generalization and performance when dealing with complex and diverse HSI datasets. To overcome this limitation, recent studies have combined deep learning techniques with frequency-domain representations, integrating frequency decomposition into end-to-end networks [33]. These frequency-based deep learning models take the decomposed frequency components as part of the input to CNNs, enabling the model to learn discriminative frequency-domain representations in an end-to-end manner [34], and have achieved significant improvements in classification accuracy [35]. For example, Yang et al. proposed the ITER method for weakly supervised HSI classification, which employs a high-frequency-aware self-attention in a high-enhanced transformer to refine detailed feature representation for pixel-level prediction [36]. However, a critical limitation of existing frequency-based methods is the lack of effective interaction between different frequency components, failing to fully exploit the complementary properties of high- and low-frequency features.
To address aforementioned core challenges of computational redundancy, feature entanglement between global structure and local details, and insufficient cross-frequency interaction in existing methods, we propose a high–low frequency interaction Mamba network (HL-Mamba) framework, tailored for HSI classification. This framework innovatively integrates frequency-domain feature decoupling with the efficient long-range modeling capability of Mamba to tackle the core pain points in HSI classification. First, a high–low frequency decomposition Mamba module is designed to decompose HSI into low-frequency structural and high-frequency detail components. The low-frequency branch efficiently captures long-range dependencies in global structures, while the high-frequency branch focuses on modeling fine-grained textures and edges. This frequency-wise separation manner enables the network to learn complementary features without mutual interference between global structural and edge details. Furthermore, a dynamic cross-frequency interaction module is proposed to establish bidirectional information flow. Thus, low-frequency structural features guide high-frequency detail aggregation, while high-frequency textures refine global structural representations, yielding more discriminative spectral–spatial features for HSI classification. To enhance feature consistency, and complementarity, a frequency alignment loss is designed, enhancing the classification performance. The main contributions of this article can be summarized as follows:
1.
A high–low frequency interaction Mamba network is proposed, which innovatively integrates frequency-domain feature decoupling with the efficient long-range modeling capability of Mamba to tackle the core issues of feature entanglement and computational redundancy in HSI classification.
2.
A dynamic cross-frequency interaction module is designed, where low-frequency structural features guide aggregation of discriminative high-frequency details and high-frequency textures refine global structural representations, yielding more discriminative spectral–spatial features for HSI classification.
3.
A frequency alignment loss is designed to constrain the consistency and complementarity of distribution between the high- and low-frequency components, further enhancing the discriminability of classification.

2. Materials and Methods

2.1. Dataset Description

Four benchmark datasets are selected to ensure experimental diversity across different sensors. The detailed number of training and testing samples is summarized in Table 1.
1.
Indian Pines dataset [37]: Captured over northwestern Indiana, USA, this dataset has a spatial resolution of 20 m and 220 spectral bands. The spatial size of image is 145 × 145 pixels, containing 16 land-cover classes dominated by agricultural crops (e.g., corn, soybean), forests, and low-density residential areas. The strong spectral similarity between corn, soybean, and grass species poses a significant challenge for HSI classification.
2.
Pavia University dataset [38]: Acquired by the ROSIS sensor over Pavia, Italy, this dataset features a high spatial resolution of 1.3 m and 103 spectral bands. A spatial size of 610 × 340 pixels and 9 urban surface classes (e.g., asphalt, meadows, bare soil, self-blocking bricks), it is ideal for evaluating the fine-grained spatial discrimination capability of models due to its rich spatial details.
3.
Houston dataset [39]: Released by the IEEE GRSS Data Fusion Contest 2013, this dataset includes 349 × 1905 pixels and 144 spectral bands, covering 15 complex urban and suburban classes (e.g., roads, parking lots, vegetation, buildings). High intra-class variability and mixed pixels caused by shadows and structural occlusions make it a challenging benchmark for urban HSI classification.
4.
WHU-Hi-HanChuan dataset [40]: A large-scale dataset captured by the Wuhan University airborne hyperspectral sensor over Hanchuan, China. It covers 1100 × 5100 pixels with 270 spectral bands and a fine spatial resolution of 1 m. The 16 land-cover categories include agricultural crops (strawberry, soybean, sorghum), vegetation, water bodies, and built-up areas. The vast scene size and high spectral diversity enable a rigorous test of the model’s generalization to real-world scenarios.

2.2. Data Processing Methods

To ensure fair comparison with HSI classification methods, we adopt the widely accepted training–test split protocols consistent with mainstream works for the four datasets, randomly selecting a fixed proportion of labeled samples per land cover category as the training set and using remaining labeled samples as the independent test set, with the random seed fixed throughout the splitting process to guarantee result reproducibility and avoid data leakage. For data preprocessing, we first perform per-band min–max normalization on raw HSI data to map reflectance values to the range of [0, 1]. We then apply PCA with the transformation matrix fitted only on the training set. To construct standardized model inputs, we adopt a center-pixel neighborhood cropping strategy with a uniform spatial window size of K = 7 for all datasets to balance local spatial context capture and computational efficiency, and use padding for edge pixels to supplement missing neighborhood information and ensure valid input patches for all pixels without introducing extra noise.

3. Proposed Method

In this section, the details of the HL-Mamba are presented. The core idea is to separate HSI into low-frequency and high-frequency components, enabling complementary feature learning via the Mamba network. A cross-frequency interaction module is also designed for feature interaction and learning.

3.1. Mamba: State Space Model for Sequence Modeling

Mamba, a novel state space mode proposed for sequence modeling, has recently shown remarkable superiority in processing long-range sequential data due to its linear computational complexity ( O ( N ) ) and efficient state propagation mechanism [28]. Different from traditional transformer-based architectures relying on self-attention, Mamba captures temporal and spatial dependencies by modeling the evolution of hidden states through a continuous-time SSM, which is particularly suitable for HSI classification tasks where spectral–spatial features exhibit sequential characteristics along spectral bands.
The core of Mamba is the discrete-time state space update, defined as:
h t = A h t 1 + B x t , y t = C h t + D x t
where h t R d denotes the hidden state at position t, x t R d is the input feature, y t R d represents the output, and A R d × d , B R d × d , C R d × d , D R d are learnable parameter matrices. To adapt to HSI’s multidirectional feature propagation, the diagonalized version of A (denoted as A ¯ ) is adopted to accelerate computation, leading to the simplified inference formula:
h t = diag ( exp ( A ¯ ) ) h t 1 + B x t
where exp ( · ) denotes element-wise exponential operation and diag ( · ) converts a vector to a diagonal matrix. This design enables Mamba to efficiently model the long spectral sequences of HSI pixels while maintaining low computational cost, making it a promising backbone for HSI classification.

3.2. Overall Framework

HSI classification faces two core challenges, i.e., spectral–spatial feature entanglement and computational redundancy in long-range dependency modeling. Traditional HSI classification methods mainly rely on convolutions or self-attention mechanisms for feature learning [17,23]. However, convolutions are limited in capturing long-range dependences, while self-attention suffers from quadratic computational complexity. In addition, these methods treat the spectral–spatial features of HSIs as spatial domain, ignoring the inherent frequency properties of HSIs. Different frequency components preserve sharp edges and global structural features that are crucial for distinguishing similar classes. To address the above issues, an HL-Mamba is proposed (Figure 1). It integrates frequency-domain decomposition and long-range modeling of Mamba to explicitly process high- and low-frequency features. This reduces feature entanglement and computational redundancy, while boosting the discriminability of spectral–spatial representations for HSI classification.
Given an input HSI cube X R H × W × B , where H, W, and B denote the spatial height, width, and number of spectral bands, respectively, we first apply principal component analysis (PCA) [41] to reduce spectral redundancy and preserve the most informative bands:
X = PCA ( X ) , X R H × W × B
where B B represents the reduced spectral dimension.
For pixel-wise HSI classification, 3D patches centered at each pixel are extracted as input samples:
P i R K × K × B
where K denotes the spatial window size and P i is the local spatial–spectral cube of the i-th sample.
Each patch is processed by the proposed HL-Mamba, whose core components include a high–low frequency decomposition Mamba module, a cross-frequency interaction module, and a classification head. The overall forward process is defined as:
y i = f cls f CFI ( f LF ( P i ) , f HF ( P i ) )
where f LF ( · ) and f HF ( · ) denote the low- and high-frequency feature extraction branches, f CFI ( · ) represents the cross-frequency interaction module, and f cls ( · ) is the final classification head output y i R C with C classes corresponding to the number of land-cover types.
This framework achieves frequency decomposition, long-range feature modeling, and adaptive interaction, ensuring that the model can simultaneously capture fine-grained edge features and global structural information, thus achieving high-precision HSI classification.

3.3. High–Low Frequency Decomposition Mamba Module

HSI contains rich but redundant spectral–spatial information, and directly feeding deep networks leads to entangled feature learning. Moreover, the complex spatial–spectral correlation in HSI requires the model to have strong long-range dependency modeling capabilities, but traditional methods (convolutions or self-attention) are difficult to balance modeling effectiveness and computational redundancy. To address these problems, a high–low frequency decomposition Mamba module is designed (Figure 2), which first explicitly disentangles HSI to global structural representations and fine-grained texture representations, and then employs two parallel Mamba branches to model long-range dependencies under different frequency components.
Given a patch P i , we perform a 2D discrete Fourier transform (DFT) along the spatial dimensions to convert the spatial domain information into the frequency domain:
F = F ( P i ) , F C K × K × B
where F ( · ) is the DFT operator. In the frequency domain, each spatial frequency component corresponds to different levels of spatial variation and structural smoothness. Low-frequency components are distributed in the central region, corresponding to smooth global structures and semantic information, while high-frequency components are distributed in the peripheral region, corresponding to sharp edges and local texture details.
To accurately separate high- and low-frequency components, we define two complementary frequency masks based on the distance from the frequency domain center:
M LF ( u , v ) = 1 , u 2 + v 2 r c 0 , otherwise , M HF = 1 M LF
where r c denotes the cutoff radius determining the low–high frequency boundary, which is a learnable hyperparameter adjusted according to different HSI datasets.
The high- and low-frequency components in the frequency domain are reconstructed back to the spatial domain via inverse DFT to facilitate subsequent feature extraction and modeling:
P LF = F 1 ( F M LF ) , P HF = F 1 ( F M HF )
where ⊙ is element-wise multiplication. P LF and P HF , respectively, represent the low-frequency structural component and high-frequency detail component of the original patch.
To fully capture the distinct characteristics of different frequency components, each frequency component is fed into an independent Mamba network for long-range dependency modeling and feature enhancement:
Z LF = f Mamba - LF ( P LF ) , Z HF = f Mamba - HF ( P HF )
where Z LF R K × K × D and Z HF R K × K × D are the low-frequency structural feature map and high-frequency detail feature map, respectively. D is the channel dimension after Mamba network.
The Mamba architecture, based on selective state-space modeling, efficiently captures long-range dependencies with linear complexity. Compared with Transformer-based models, Mamba reduces computational costs, allowing better feature representation for HSI classification.

3.4. Cross-Frequency Interaction Module

Although low-frequency and high-frequency branches capture complementary information, processing them independently may lead to insufficient representation. To address this, a cross-frequency interaction module is designed that enables bidirectional information exchange and adaptive feature fusion (Figure 3).
We first summarize global information from each branch using global average pooling (GAP) and project it through linear layers:
α LF HF = σ W 1 [ GAP ( Z LF ) ] , α HF LF = σ W 2 [ GAP ( Z HF ) ]
where σ ( · ) denotes the sigmoid activation function, W 1 R D × D and W 2 R D × D are learnable projection matrices, and GAP ( · ) represents the global average pooling operation. α LF HF measures the contribution of low-frequency structural features to high-frequency detail feature enhancement, and α HF LF measures the contribution of high-frequency texture features to low-frequency structural feature refinement.
Based on the generated dynamic attention weights, adaptive feature enhancement on each frequency branch is performed to achieve bidirectional information interaction:
Z ˜ HF = Z HF + α LF HF Z LF , Z ˜ LF = Z LF + α HF LF Z HF
After bidirectional enhancement, the two frequency feature maps are concatenated and project them through a fusion layer to generate the final integrated spectral–spatial feature map:
Z fuse = ϕ [ Z ˜ LF | | Z ˜ HF ] W f
where | | denotes concatenation, ϕ ( · ) represents a GELU activation, and W f is a learnable fusion matrix. The GELU activation is mathematically defined as:
ϕ ( x ) = x · Φ ( x ) = 0.5 x 1 + erf x 2
where ϕ ( x ) is the cumulative distribution function of the standard normal distribution, and erf denotes the Gaussian error function. Unlike traditional piecewise activation functions (e.g., ReLU), GELU introduces soft stochasticity by weighting input values based on their probability of being active, which effectively alleviates overfitting and improves the generalization ability of deep learning models fo feature learning.
Finally, the integrated feature map is input into the classification head to obtain the final classification prediction. The classification head consists of a global average pooling layer, a fully connected layer, and a softmax activation function:
y i = Softmax ( W c · Pool ( Z fuse ) + b c )
where Pool ( · ) is the global average pooling operation and W c R D × C and b c R C are the weight matrix and bias vector of the fully connected layer, respectively. To avoid noise addition during cross-frequency interaction, we adopt the dynamic attention weights generated by global average pooling which can adaptively weight the contribution of different frequency features, reducing the impact of noise. In addition, the GELU activation function in the fusion layer has a slight regularization effect, suppressing the propagation of noise components.

3.5. Frequency Alignment Loss

Although the cross-frequency interaction module promotes information exchange between high- and low-frequency features, during the training process, the two frequency branches may still converge to disjoint feature subspaces due to differences in the characteristics of the structural and texture features. To address this issue, a frequency alignment loss function is designed to explicitly constrain the distribution consistency and feature complementarity between high- and low-frequency features.
The total loss function of the HL-Mamba is a weighted combination of the cross-entropy loss ( L cls ) and the frequency alignment loss ( L align ). Among them, the cross-entropy loss is used to supervise the classification performance of model, ensuring that the integrated features can accurately distinguish different land-cover classes:
L cls = 1 N i = 1 N c = 1 C y i , c log ( y ^ i , c )
where N is the number of training samples and y i , c and y ^ i , c denote the ground truth label and predicted probability of the i-th sample for the c-th class, respectively.
The frequency alignment loss is designed to constrain the distribution consistency of high- and low-frequency features. We first normalize the high- and low-frequency feature maps to eliminate the impact of different feature scales, then calculate the L2 distance between the normalized features, and finally take the average of all samples as the alignment loss:
L align = 1 N i = 1 N Norm ( Z LF ( i ) ) Norm ( Z HF ( i ) ) 2 2
where Norm ( · ) denotes the normalization operation and Z LF ( i ) and Z HF ( i ) are the low-frequency and high-frequency feature maps of the i-th sample, respectively. Minimizing this loss can force the high- and low-frequency features to be distributed in similar subspaces, ensuring semantic consistency, while the differences in features are retained to maintain complementarity.
The total loss function of the network is defined as:
L total = L cls + λ L align
where λ is a hyperparameter controlling the balance between the classification accuracy and the inter-frequency feature coherence. Through experiments, we set λ = 0.2 to ensure that the frequency alignment loss can effectively promote feature consistency without compromising the classification performance.

4. Results

4.1. Experimental Setup

All experiments are implemented using PyTorch 2.2 and are conducted on a NVIDIA RTX 4090 GPU. The Adam optimizer is used for model training with an initial learning rate of 1 × 10 3 and a batch size of 64. During data preprocessing, PCA is used to reduce the spectral dimension to 30 to preserve key spectral information while reducing redundancy. The network depth is configured as three Mamba modules, with a total of six Mamba modules for feature extraction in the dual branches. The loss weight λ in the total loss function is adjusted through experiments, and the value was set to 0.2. The training curves of the Indian Pines dataset are shown in Figure 4. All models are trained for a maximum of 100 epochs, and the model with the highest validation accuracy is selected for testing.

4.2. Evaluation Metrics

Three metrics are used to quantitatively evaluate the classification performance, i.e., overall accuracy (OA), average accuracy (AA), and Kappa coefficient. OA measures the global classification performance as the ratio of correctly classified pixels to the total number of pixels, which is mathematically defined as:
OA = i = 1 C n i i i = 1 C j = 1 C n i j × 100 %
where C denotes the total number of land cover categories, n i i represents the number of pixels correctly classified into class i, and n i j is the number of pixels belonging to class i but predicted as class j. AA addresses category imbalance by calculating the arithmetic mean of class-wise accuracies, reflecting the model’s ability to classify each category evenly. Its formulation is given by:
AA = 1 C i = 1 C n i i j = 1 C n i j × 100 %
The Kappa coefficient is a robust metric that evaluates the agreement between predicted results and ground truth while eliminating the impact of random classification, providing a more reliable assessment beyond mere chance. It is calculated as:
Kappa = OA P e 1 P e
where P e = 1 N 2 i = 1 C j = 1 C n i j j = 1 C n j i is the expected accuracy due to random chance, and N = i = 1 C j = 1 C n i j is the total number of test pixels.

4.3. Performance Comparison

To further validate the performance of HL-Mamba, we compare it with eight HSI classification methods, including CNN-based methods (CDCNN [42], FDSSC [43], M3D-CNN [44], DGCNet [45], LGCNet [46]) and spectral–spatial feature fusion methods (DBDA [47], DBSSAN [48], SSFTT [22]). The quantitative results on the four datasets are summarized in Table 2, Table 3, Table 4 and Table 5, and the qualitative classification maps are shown in Figure 5, Figure 6, Figure 7 and Figure 8. The optimal values for OA, AA, and Kappa are indicated in bold. The second-best values are indicated with double underlines, and the third-best values are indicated with a single underline.

4.3.1. Quantitative Results and Analysis

1.
Indian Pines Dataset: As shown in Table 2, HL-Mamba achieves the highest OA (94.07%), AA (90.40%), and Kappa (93.25%) among all compared methods. Compared with the second-best method DBDA (OA = 93.44%), HL-Mamba improves OA by 0.63%. The key reason is that the high–low frequency decomposition effectively separates the global structure and edge details of crops, and the cross-frequency interaction module enhances the discriminability of spectrally similar classes (e.g., Alfalfa and Oats), which is difficult for comparison methods to achieve. Notably, HL-Mamba shows significant advantages in class-level accuracy for challenging classes with high spectral similarity. For example, the accuracy of Alfalfa (a minority class) reaches 53.78%, which is 2.45% higher than DBDA (51.33%) and far superior to CDCNN (0.89%) and FDSSC (24.00%). For Oats, HL-Mamba achieves 79.47% accuracy, outperforming M3DCNN (78.42%) and DBDA (73.16%). This indicates that the frequency-aware feature learning of HL-Mamba effectively enhances the discrimination of spectrally similar classes.
2.
Pavia University Dataset: Table 3 presents the quantitative results of HL-Mamba and comparison methods on the Pavia University dataset. This dataset is characterized by high resolution and large samples for most classes, which demands models to effectively capture fine-grained spatial details in processing large-scale data. HL-Mamba achieves the highest AA of 90.66% and Kappa of 91.79%. In terms of class-level accuracy, HL-Mamba shows remarkable performance in classes with similar spatial textures and spectral properties. For example, the Asphalt and Bitumen classes, which are easily misclassified due to their similar gray-scale textures in spatial images, HL-Mamba achieves accuracies of 90.82% and 95.18%, respectively. For the Shadows class, a typical low-frequency structural class with large intra-class variability, HL-Mamba achieves an accuracy of 82.72%. This advantage comes from the low-frequency branch’s efficient modeling of global structural context, which accurately identifies shadow regions. For the high-resolution urban scene of Pavia University, the key advantage lies in the HLFDMM’s ability to capture fine-grained spatial details (via high-frequency branch) and global structural consistency (via low-frequency branch). For example, the Asphalt class (easily confused with Bitumen due to similar gray textures) achieves 90.82% accuracy, and the Shadows class (with large intra-class variability) reaches 82.72% accuracy, far higher than M3DCNN’s 63.55%. This is because the low-frequency branch models the global context of shadow regions, while the cross-frequency interaction module refines edge details of urban structures, effectively mitigating misclassification caused by spatial similarity.
3.
Houston Dataset: The Houston dataset is a challenging urban hyperspectral dataset with complex mixed pixels, shadows, and occlusions, requiring models to extract spectral–spatial feature better. As shown in Table 4, HL-Mamba achieves the highest OA of 87.32%, AA of 88.18%, and Kappa of 86.29%. The dataset’s challenges (mixed pixels, shadows, occlusions) are addressed by the synergy of CFIM and FAL: for the Commercial class (affected by high-rise building shadows), HL-Mamba’s accuracy (63.68%) is 3.45% higher than DBDA (60.23%); for the Parking Lot 2 class (small-scale and easily mixed with roads), accuracy reaches 86.90%, outperforming the second-best FDSSC (80.80%). The FAL enhances feature invariance to illumination variations, while the CFIM fuses structural (low-frequency) and texture (high-frequency) features, making the model robust to urban interference factors. For classes affected by shadows and occlusions, such as Commercial and Railway, HL-Mamba shows significant advantages. The Commercial class, which is often misclassified due to the shadow of high-rise buildings, achieves an accuracy of 63.38% with HL-Mamba, 3.45% higher than DBDA (92.48%). This is because the frequency alignment loss enhances the invariance of features to illumination variations, making the model less sensitive to shadow-induced spectral distortions.
4.
WHU-Hi-HanChuan Dataset: As a large-scale agricultural and suburban hyperspectral dataset, WHU-Hi-HanChuan features a vast scene size and high spectral diversity, testing the model’s performance. Table 5 presents the quantitative results, where HL-Mamba achieves the highest OA of 95.28%, AA of 89.81%, and Kappa of 94.47%, outperforming the second-ranked M3DCNN (OA = 94.26%, Kappa = 993.28%). For crop classes with overlapping spectral bands (e.g., Soybean and Sorghum), HL-Mamba achieves 92.70% and 97.42% accuracy, respectively, benefiting from HLFDMM’s effective decomposition of spectral–spatial features. For the Water-spinach class (a minority crop with sparse samples), accuracy reaches 85.78%, 15.31% higher than DBDA (70.75%), as the cross-frequency interaction module aggregates discriminative details from high-frequency features, compensating for the lack of training samples. For agricultural crop classes with similar spectral characteristics, such as Soybean and Sorghum, HL-Mamba achieves accuracies of 97.42% and 85.78%, respectively. This demonstrates that the frequency-aware feature learning effectively extracts discriminative spectral features from overlapping spectral bands of different crops.

4.3.2. Visualization Results and Analysis

Figure 5, Figure 6, Figure 7 and Figure 8 illustrate classification maps generated by the proposed method and baselines. Visually, the proposed method yields smoother boundaries and fewer misclassified pixels, especially along class edges and shadowed regions. The high-frequency branch captures subtle texture transitions, while the low-frequency branch preserves large-area semantic consistency. The resulting maps exhibit clear object details.
1.
Indian Pines Dataset: Qualitatively, Figure 5 visualizes the classification maps of HL-Mamba and comparison methods on the Indian Pines dataset, which is dominated by agricultural land cover with similar crop types. The classification map of the HL-Mamba exhibits sharp edge details across the entire scene. In contrast, CDCNN and FDSSC suffer from severe salt-and-pepper noise in minority class regions, and their classification results show blurred boundaries between Oats and surrounding grassland classes. For the Grass-pasture-mowed class, a sparse category prone to misclassification, HL-Mamba clearly demarcates its distribution with minimal pixel confusion. This visual superiority stems from the CFIM and frequency-aware feature learning of HL-Mamba. The high-frequency branch captures fine texture differences between similar crops, while the low-frequency branch anchors the global structural distribution of sparse classes.
2.
Pavia University: Qualitatively, the classification map of HL-Mamba (Figure 6) shows clear edge details and minimal misclassification noise compared to other methods. For instance, in the buildings, HL-Mamba accurately distinguishes the classes with no obvious mixed pixels, while CDCNN and FDSSC exhibit slight misclassification in these edge regions. Additionally, HL-Mamba effectively suppresses the “salt-and-pepper” noise in the classification map of the Gravel class, which is attributed to the CFIM integration of global and local features, enhancing the spatial consistency of classification results.
3.
Houston Dataset: Figure 7 presents the qualitative classification results on the Houston dataset, a complex urban scene with mixed pixels. The classification map of HL-Mamba accurately delineates the boundaries of distinct urban functional areas, with no obvious mixed pixels in the transition zones. For shadow-affected regions, HL-Mamba maintains consistent classification accuracy without spectral distortion-induced misclassification, while DBDA and M3DCNN show scattered misclassified pixels in these shadowed areas. Notably, HL-Mamba precisely identifies small-scale high-texture objects embedded in complex urban landscapes, with clear and complete details of their shapes. Additionally, for the Healthy grass and Stressed grass classes with high intra-class variability, HL-Mamba’s classification map shows smooth spatial transitions between the two grass types, reflecting the effective fusion of global structure and local details via the CFIM.
4.
WHU-Hi-HanChuan Dataset: Qualitatively, Figure 8 shows that HL-Mamba’s classification map accurately captures the large-scale distribution of agricultural crops and urban areas, with no obvious misclassification in the transition zones between Strawberry fields and Soybean fields. Compared to CDCNN and FDSSC, HL-Mamba’s classification map has better spatial continuity, especially in the Water class, where the boundaries of water bodies are clearly delineated without being affected by surrounding vegetation shadows. This confirms that HL-Mamba is well-suited for large-scale HSI classification tasks.

4.3.3. Computational Complexity

To evaluate the efficiency of HL-Mamba, we analyzed its model parameters, computational complexity (FLOPs), and test time on the Indian Pines dataset and compared them with eight methods under the same hardware platform. As shown in Table 6, HL-Mamba achieves an excellent balance between classification performance and model efficiency. The visualization results of the number of parameters and the of OA for each method in the Indian Pines dataset are shown in Figure 9. It has a compact parameter count of 0.2957 M, a low FLOP count of 0.0354 G, and a fast average test time of 0.2934 s, which demonstrates its lightweight and high-efficiency characteristics for HSI classification. Compared with other methods, HL-Mamba shows obvious advantages in reducing computational overhead and accelerating inference. HL-Mamba significantly reduces both model parameters and inference latency, which benefits from the linear-complexity Mamba backbone, frequency decomposition-based redundant computation reduction, and a lightweight cross-frequency interaction mechanism. Although slightly higher in parameters and FLOPs than some extremely lightweight methods (e.g., DBDA with 0.0389 M parameters, M3DCNN with 0.0095 G FLOPs), HL-Mamba outperforms these methods by a large margin in classification accuracy (OA) on the Indian Pines dataset, which verifies its trade-off of performance and efficiency.

5. Discussion

To validate the effectiveness of key components in HL-Mamba and the influence of critical hyperparameters, we carried out ablation experiments on the Indian Pines dataset. The Indian Pines dataset was selected due to its high class diversity and spectral similarity challenges, which can effectively highlight the performance of model. Two sets of ablation experiments were designed. The first was a sensitivity analysis of hyperparameter λ and the number of Mamba modules. The second was module ablation. These experiments aimed to validate the contributions of the high–low frequency decomposition Mamba module (HLFDMM), CFIM, and frequency alignment loss (FAL).

5.1. Sensitivity Analysis of Hyperparameter

To investigate the sensitivity of the model’s classification performance to the number of Mamba modules, experiments were conducted with the number of Mamba modules set to 2, 3, 4, and 5, respectively. The results are presented in Table 7. It can be observed that the classification metrics exhibit a trend of first increasing and then decreasing with the increase in the number of Mamba modules. When the number of Mamba modules is 3, the model achieves the optimal performance, with OA reaching 94.07%, AA 90.40%, and Kappa 93.25%. This indicates that an appropriate increase in the number of Mamba modules can enhance the model’s ability, thereby improving HSI classification accuracy. However, when the number of modules exceeds 3, all metrics show a decline. This phenomenon may be attributed to the excessive complexity of the model caused by redundant Mamba modules, leading to overfitting and a reduction in feature representation efficiency. Therefore, the optimal number of Mamba modules for the proposed model is determined to be 3.
The loss weight λ balances the classification loss L cls and the frequency alignment loss L align . To determine the optimal λ , experiments are conducted with λ varying from 0 to 0.6 at intervals of 0.1. The results are summarized in Table 8. When λ = 0 (i.e., no frequency alignment loss), the OA is 92.86%, AA is 88.57%, and Kappa is 91.85%. With the increase of λ from 0 to 0.2, all metrics gradually improve, reaching the highest performance (OA = 94.07%, AA = 90.40%, Kappa = 93.25%) at λ = 0.2 . This indicates that the frequency alignment loss effectively enhances the consistency and complementarity between high- and low-frequency features, thereby improving HSI classification performance. However, when λ exceeds 0.2, the metrics begin to decrease. For example, when λ = 0.6 , OA drops to 92.53%, which is lower than the result at λ = 0.2 . This is because a large λ over-constrains the feature distribution of the two frequency branches, leading to the loss of complementary details between high- and low-frequency features. Thus, setting λ = 0.2 achieves the optimal balance between classification accuracy and inter-frequency feature.

5.2. Module Ablation

To validate the contributions of the high–low frequency decomposition Mamba module (HLFDMM), CFIM, and frequency alignment loss (FAL), module ablation experiments were conducted, as shown in Table 9. The baseline model without any of the three modules achieves an OA of 89.25%, AA of 84.11%, and Kappa of 87.98%. Specifically, enabling only HLFDMM yields the most significant standalone performance improvement (OA = 92.86%), confirming its core role in decomposing HSIs into discriminative high–low frequency features to capture global structures and fine-grained details. In contrast, enabling only CFIM improves OA to 90.18% by facilitating cross-frequency feature interaction, while enabling only FAL raises OA to 89.76% by aligning feature distributions across frequency branches; both exhibit limited standalone gains because CFIM merely establishes interaction without relying on effective frequency decomposition and FAL lacks the support of decoupled high–low frequency features. Furthermore, among dual-module combinations, the HLFDMM + CFIM configuration performs best (OA = 93.69%), as CFIM can fully exploit the discriminative value of HLFDMM-decomposed features to enhance inter-frequency complementarity, while the HLFDMM + FAL combination also delivers solid results (OA = 93.24%) by integrating decomposition and alignment. Notably, activating all three modules achieves the optimal performance (OA = 94.07%, AA = 90.40%, Kappa = 93.25%), demonstrating that only when combined with HLFDMM, whose frequency decomposition lays the foundation for discriminative feature learning, can CFIM fully leverage the complementary advantages of different frequency features and FAL effectively enhance the consistency of decoupled features, ultimately achieving synergistic performance improvement in HSI classification.

5.3. Limitations

Despite the promising performance of the proposed HL-Mamba model demonstrated by the above experiments, it still has limitations. The frequency decomposition strategy adopted in HLFDMM is fixed, and it lacks adaptability to different types of ground objects in HSIs, which may restrict the model’s ability to capture task-specific frequency features.

6. Conclusions

This study makes three key contributions that integrate frequency-domain decomposition and Mamba to effectively address feature entanglement and computational redundancy, designing a CFIM to achieve bidirectional information complementation between high- and low-frequency features, and designing an FAL to enhance feature consistency. Specifically, by decomposing HSI into complementary low-frequency structural components and high-frequency edge detail components, and employing parallel Mamba branches to model different frequency component, the proposed method mitigates the problems of feature entanglement and information redundancy. Furthermore, through the bidirectional information flow between different frequency features established by the cross-frequency interaction module, and the consistency enhancement of frequency-domain features by the frequency alignment loss, HL-Mamba is able to learn more discriminative spectral–spatial representations and significantly enhance the HSI classification capability. Extensive experiments conducted on four public benchmark datasets demonstrate that the proposed HL-Mamba consistently achieves state-of-the-art classification performance while maintaining high computational efficiency, benefiting from the linear-complexity Mamba backbone and lightweight cross-frequency interaction module. Ablation studies and visualization analyses further confirm the effectiveness of each core component of the overall framework. Future research will focus on developing adaptive frequency separation schemes, exploring multi-scale frequency fusion, and extending the model to large-scale or real-time remote sensing applications.

Author Contributions

Conceptualization, Y.T., S.G. and X.Y.; methodology, Y.T., S.G. and X.Y.; software, Y.T.; validation, Y.T. and S.G.; formal analysis, S.G.; investigation, Y.T.; resources, Y.T.; data curation, Y.T.; writing—original draft preparation, Y.T.; writing—review and editing, S.G. and X.Y.; visualization, Y.T.; supervision, S.G.; project administration, S.G.; funding acquisition, S.G. and X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grant number 62266026 and supported by the analysis and testing foundation of Kunming University of Science and Technology, number 2022P20191101006.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data and code are publicly available at https://github.com/tengye663/HL-Mamba, accessed on 25 February 2026.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HSIHyperspectral image
CNNConvolutional neural network
PCAPrincipal component analysis
RNNRecurrent neural network
HL-MambaHigh–low frequency interaction Mamba network
DFTDiscrete Fourier transform

References

  1. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral image transformer classification networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528715. [Google Scholar] [CrossRef]
  2. Ahmad, M.; Ghous, U.; Usama, M.; Mazzara, M. WaveFormer: Spectral–spatial wavelet transformer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5502405. [Google Scholar] [CrossRef]
  3. Rajabi, R.; Zehtabian, A.; Singh, K.D.; Tabatabaeenejad, A.; Ghamisi, P.; Homayouni, S. Hyperspectral imaging in environmental monitoring and analysis. Front. Environ. Sci. 2024, 11, 1353447. [Google Scholar] [CrossRef]
  4. Avola, G.; Matese, A.; Riggi, E. An overview of the special issue on “precision agriculture using hyperspectral images”. Remote Sens. 2023, 15, 1917. [Google Scholar] [CrossRef]
  5. Booysen, R.; Lorenz, S.; Thiele, S.T.; Fuchsloch, W.C.; Marais, T.; Nex, P.A.M.; Gloaguen, R. Accurate hyperspectral imaging of mineralised outcrops: An example from lithium-bearing pegmatites at Uis, Namibia. Remote Sens. Environ. 2022, 269, 112790. [Google Scholar] [CrossRef]
  6. Yuan, J.; Wang, S.; Wu, C.; Xu, Y. Fine-grained classification of urban functional zones and landscape pattern analysis using hyperspectral satellite imagery: A case study of Wuhan. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3972–3991. [Google Scholar] [CrossRef]
  7. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral image classification using group-aware hierarchical transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5539014. [Google Scholar] [CrossRef]
  8. He, L.; Li, J.; Liu, C.; Li, S. Recent advances on spectral–spatial hyperspectral image classification: An overview and new guidelines. IEEE Trans. Geosci. Remote Sens. 2017, 56, 1579–1597. [Google Scholar] [CrossRef]
  9. Harikiran, J.J.H. Hyperspectral image classification using support vector machines. IAES Int. J. Artif. Intell. 2020, 9, 684. [Google Scholar] [CrossRef]
  10. Florimbi, G.; Fabelo, H.; Torti, E.; Lazcano, R.; Madroñal, D.; Ortega, S.; Salvador, R.; Leporati, F.; Danese, G.; Báez-Quevedo, A. Accelerating the K-nearest neighbors filtering algorithm to optimize the real-time classification of human brain tumor in hyperspectral images. Sensors 2018, 18, 2314. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Cao, G.; Li, X.; Wang, B.; Fu, P. Active semi-supervised random forest for hyperspectral image classification. Remote Sens. 2019, 11, 2974. [Google Scholar] [CrossRef]
  12. Imani, M.; Ghassemian, H. Edge patch image-based morphological profiles for classification of Multispectral and hyperspectral data. IET Image Process. 2017, 11, 164–172. [Google Scholar] [CrossRef]
  13. Uddin, M.P.; Mamun, M.A.; Afjal, M.I.; Hossain, M.A. Information-theoretic feature selection with segmentation-based folded principal component analysis (PCA) for hyperspectral image classification. Int. J. Remote Sens. 2021, 42, 286–321. [Google Scholar] [CrossRef]
  14. Liu, Y.; Hu, J.; Kang, X.; Luo, J.; Fan, S. Interactformer: Interactive transformer and CNN for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531715. [Google Scholar] [CrossRef]
  15. Liu, Y.; He, Q.; Li, J.; Liu, X.; Fiorio, P.R.; Nakai, É.S.; Yang, B. Towards resolution-arbitrary remote sensing change detection with Spatial-frequency dual domain learning. ISPRS J. Photogramm. Remote Sens. 2026, 231, 137–150. [Google Scholar] [CrossRef]
  16. Slavkovikj, V.; Verstockt, S.; De Neve, W.; Van Hoecke, S.; Van de Walle, R. Hyperspectral image classification with convolutional neural networks. In Proceedings of the 23rd ACM international conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 1159–1162. [Google Scholar]
  17. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral image classification using a hybrid 3D-2D convolutional neural networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar] [CrossRef]
  18. Yang, J.; Wu, C.; Du, B.; Zhang, L. Enhanced multiscale feature fusion network for HSI classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10328–10347. [Google Scholar] [CrossRef]
  19. Zhou, W.; Kamata, S.-I.; Luo, Z.; Wang, H. Multiscanning strategy-based recurrent neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5521018. [Google Scholar] [CrossRef]
  20. Qi, W.; Zhang, X.; Wang, N.; Zhang, M.; Cen, Y. A spectral-spatial cascaded 3D convolutional neural network with a convolutional long short-term memory network for hyperspectral image classification. Remote Sens. 2019, 11, 2363. [Google Scholar] [CrossRef]
  21. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  22. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  23. Peng, Y.; Zhang, Y.; Tu, B.; Li, Q.; Li, W. Spatial–spectral transformer with cross-attention for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5537415. [Google Scholar] [CrossRef]
  24. Yang, J.; Du, B.; Zhang, L. From center to surrounding: An interactive learning framework for hyperspectral image classification. ISPRS J. Photogramm. Remote Sens. 2023, 197, 145–166. [Google Scholar] [CrossRef]
  25. Dang, L.; Pang, P.; Lee, J. Depth-wise separable convolution neural network with residual connection for hyperspectral image classification. Remote Sens. 2020, 12, 3408. [Google Scholar] [CrossRef]
  26. Tan, Y.; Li, M.; Yuan, L.; Shi, C.; Luo, Y.; Wen, G. Hyperspectral image classification with embedded linear vision transformer. Earth Sci. Inform. 2025, 18, 69. [Google Scholar] [CrossRef]
  27. Ma, X.; Wang, W.; Li, W.; Wang, J.; Ren, G.; Ren, P.; Liu, B. An ultralightweight hybrid CNN based on redundancy removal for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5506212. [Google Scholar] [CrossRef]
  28. Wang, Y.; Liu, L.; Xiao, J.; Yu, D.; Tao, Y.; Zhang, W. MambaHSI+: Multidirectional State Propagation for Efficient Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 4411414. [Google Scholar] [CrossRef]
  29. Liu, R.; Liang, J.; Yang, J.; Hu, M.; He, J.; Zhu, P.; Zhang, L. DHSNet: Dual Classification Head Self-Training Network for Cross-Scene Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5534515. [Google Scholar] [CrossRef]
  30. Qiao, X.; Huang, W. A dual frequency transformer network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 10344–10358. [Google Scholar] [CrossRef]
  31. Zhuang, P.; Zhang, X.; Wang, H.; Zhang, T.; Liu, L.; Li, J. FAHM: Frequency-aware hierarchical mamba for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 6299–6313. [Google Scholar] [CrossRef]
  32. Anand, R.; Veni, S.; Aravinth, J. Robust classification technique for hyperspectral images based on 3D-discrete wavelet transform. Remote Sens. 2021, 13, 1255. [Google Scholar] [CrossRef]
  33. Wang, K.; Gu, X.F.; Yu, T.; Meng, Q.Y.; Zhao, L.M.; Feng, L. Classification of hyperspectral remote sensing images using frequency spectrum similarity. Sci. China Technol. Sci. 2013, 56, 980–988. [Google Scholar] [CrossRef]
  34. Gong, J.; Li, F.; Wang, J.; Yang, Z.; Ding, X. A split-frequency filter network for hyperspectral image classification. Remote Sens. 2023, 15, 3900. [Google Scholar] [CrossRef]
  35. Hong, D.; Wu, X.; Ghamisi, P.; Chanussot, J.; Yokoya, N.; Zhu, X.X. Invariant attribute profiles: A spatial-frequency joint feature extractor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3791–3808. [Google Scholar] [CrossRef]
  36. Yang, J.; Du, B.; Wang, D.; Zhang, L. ITER: Image-to-pixel representation for weakly supervised HSI classification. IEEE Trans. Image Process. 2024, 33, 257–272. [Google Scholar] [CrossRef]
  37. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  38. Liu, X.; Hu, Q.; Cai, Y.; Cai, Z. Extreme learning machine-based ensemble transfer learning for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3892–3902. [Google Scholar] [CrossRef]
  39. Chen, H.; Miao, F.; Chen, Y.; Xiong, Y.; Chen, T. A hyperspectral image classification method using multifeature vectors and optimized KELM. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2781–2795. [Google Scholar] [CrossRef]
  40. Hu, X.; Zhong, Y.; Luo, C.; Wang, X. WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2) benchmark datasets for hyperspectral image classification. arXiv 2020, arXiv:2012.13920. [Google Scholar]
  41. Maćkiewicz, A.; Ratajczak, W. Principal components analysis (PCA). Comput. Geosci. 1993, 19, 303–342. [Google Scholar] [CrossRef]
  42. Lee, H.; Kwon, H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef]
  43. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A fast dense spectral–spatial convolution network framework for hyperspectral images classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef]
  44. Hu, W.; Shi, W.; Lan, C.; Li, Y.; He, L. Hyperspectral Image Classification with Multi-Path 3D-CNN and Coordinated Hierarchical Attention. Remote Sens. 2025, 17, 4035. [Google Scholar] [CrossRef]
  45. Li, G.; Ye, M. DGCNet: An efficient 3d-densenet based on dynamic group convolution for hyperspectral remote sensing image classification. Spectrosc. Lett. 2025, 1–14. [Google Scholar] [CrossRef]
  46. Li, G.; Ye, M. Spatial-spectral hyperspectral classification based on learnable 3D group convolution. Spectrosc. Lett. 2025, 58, 704–716. [Google Scholar] [CrossRef]
  47. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  48. Zhao, J.; Wang, J.; Ruan, C.; Dong, Y.; Huang, L. Dual-branch spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5504718. [Google Scholar] [CrossRef]
Figure 1. Architecture of the proposed HL-Mamba. The input HSI X is first processed by PCA to obtain X . Then, high–low frequency decomposition is performed to get P L F and P H F . After Mamba modeling, Z L F and Z H F are generated and input into the cross-frequency interaction module to obtain Z f u s e . Finally, the classification result is output through the classification head. The loss function includes cross-entropy loss and frequency alignment loss, where λ is the balance weight.
Figure 1. Architecture of the proposed HL-Mamba. The input HSI X is first processed by PCA to obtain X . Then, high–low frequency decomposition is performed to get P L F and P H F . After Mamba modeling, Z L F and Z H F are generated and input into the cross-frequency interaction module to obtain Z f u s e . Finally, the classification result is output through the classification head. The loss function includes cross-entropy loss and frequency alignment loss, where λ is the balance weight.
Sensors 26 01556 g001
Figure 2. Architecture of the proposed high–low frequency decomposition Mamba module. An HSI patch P i is transformed to frequency domain by 2D DFT, split into low-frequency and high-frequency components via learnable masks with adaptive r c , reconstructed to spatial domain by inverse DFT, and processed with two Mamba branches to output enhanced Z LF and Z HF for cross-frequency interaction.
Figure 2. Architecture of the proposed high–low frequency decomposition Mamba module. An HSI patch P i is transformed to frequency domain by 2D DFT, split into low-frequency and high-frequency components via learnable masks with adaptive r c , reconstructed to spatial domain by inverse DFT, and processed with two Mamba branches to output enhanced Z LF and Z HF for cross-frequency interaction.
Sensors 26 01556 g002
Figure 3. Architecture of the proposed cross-frequency interaction module. This module takes the enhanced low-frequency structural feature Z LF and high-frequency detail feature Z HF as inputs, first generates bidirectional dynamic attention weights, then implements adaptive bidirectional feature enhancement to realize mutual guidance between global structure and local details, and finally obtains the discriminative fused spectral–spatial feature Z fuse .
Figure 3. Architecture of the proposed cross-frequency interaction module. This module takes the enhanced low-frequency structural feature Z LF and high-frequency detail feature Z HF as inputs, first generates bidirectional dynamic attention weights, then implements adaptive bidirectional feature enhancement to realize mutual guidance between global structure and local details, and finally obtains the discriminative fused spectral–spatial feature Z fuse .
Sensors 26 01556 g003
Figure 4. Training and evaluation curves on the Indian Pines dataset, which include training loss, validation loss, training accuracy, and test accuracy, intuitively reflecting the convergence process, fitting status, and generalization performance of the proposed HL-Mamba.
Figure 4. Training and evaluation curves on the Indian Pines dataset, which include training loss, validation loss, training accuracy, and test accuracy, intuitively reflecting the convergence process, fitting status, and generalization performance of the proposed HL-Mamba.
Sensors 26 01556 g004
Figure 5. Classification maps on the Indian Pines dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Figure 5. Classification maps on the Indian Pines dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Sensors 26 01556 g005
Figure 6. Classification maps on the Pavia University dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Figure 6. Classification maps on the Pavia University dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Sensors 26 01556 g006
Figure 7. Classification maps on the Houston dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Figure 7. Classification maps on the Houston dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Sensors 26 01556 g007aSensors 26 01556 g007b
Figure 8. Classification maps on the WHU-Hi-HanChuan dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Figure 8. Classification maps on the WHU-Hi-HanChuan dataset, containing results of CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method, showing the better classification performance of the proposed method.
Sensors 26 01556 g008
Figure 9. The visualization results of the number of parameters and the of OA for each method in the Indian Pines dataset, containing CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method.
Figure 9. The visualization results of the number of parameters and the of OA for each method in the Indian Pines dataset, containing CDCNN, FDSSC, DBDA, SSFTT, DBSSAN, DGCNet, LGCNet, M3DCNN, and the proposed method.
Sensors 26 01556 g009
Table 1. Detailed category information of four datasets.
Table 1. Detailed category information of four datasets.
Indian Pines DatasetHouston Dataset
ClassClass NameTrainTestClassClass NameTrainTest
1Alfalfa1451Healthy grass131238
2Corn-notill4313852Stressed grass131241
3Corn-mintill258053Synthetic grass7690
4Corn72304Trees121232
5Grass-pasture144695Soil121230
6Grass-trees227086Water3322
7Grass-pasture-mowed1277Residential131255
8Hay-windrowed144648Commercial121232
9Oats1199Road131239
10Soybean-notill2994310Highway121215
11Soybean-mintill73238211Railway121223
12Soybean-clean1857512Parking Lot 1121221
13Wheat619913Parking Lot 25464
14Woods38122714Tennis Court4424
15Buildings-Grass-Trees-Drives1237415Running Track7653
16Stone-Steel-Towers390-Total15014,879
-Total3079942
Pavia University DatasetWHU-Hi-HanChuan Dataset
ClassClass NameTrainTestClassClass NameTrainTest
1Asphalt3365981Strawberry22444,511
2Meadows9318,5562Cowpea114
3Gravel1020893Soybean5110,236
4Trees1530494Sorghum275326
5Painted metal sheets713385Water-spinach61194
6Bare Soil2550046Watermelon234510
7Bitumen713237Greens295874
8Self-Blocking Bricks1836648Trees9017,888
9Shadows59429Grass479422
-Total21342,56310Red-roof5210,464
11Gray-roof8416,827
12Plastic183661
13Bare-soil469070
14Road9318,467
15Bright-object61130
16Water37775,024
-Total1287256,243
Table 2. Classification results on the Indian Pines dataset.
Table 2. Classification results on the Indian Pines dataset.
ClassCDCNNFDSSCDBDASSFTTDBSSANDGCNetLGCNetM3DCNNOurs
Alfalfa0.8924.0051.3338.6715.5620.0020.6742.2253.78
Corn-notill78.0686.8292.3581.8673.6073.6972.5892.0992.53
Corn-mintill88.0584.1791.1386.4161.1366.4563.6590.9392.37
Corn79.7080.5287.7877.2657.7452.8355.3589.8790.70
Grass-pasture90.3087.7491.6283.5467.0174.0372.8187.8090.00
Grass-trees91.2791.8198.8194.0494.6596.4394.8397.1696.84
Grass-pasture-mowed12.5956.6790.3750.3720.7427.4118.5290.7493.33
Hay-windrowed97.4197.3399.2097.7698.4398.3697.5698.9799.46
Oats0.0032.1173.1667.3711.5825.2617.3778.4279.47
Soybean-notill79.9587.2590.3680.8469.0971.6367.3089.8589.70
Soybean-mintill81.3891.0595.2790.3482.6783.7283.7495.2695.93
Soybean-clean84.2379.9782.1278.0051.1054.3552.0587.0190.38
Wheat89.6595.0898.1489.4090.9094.0789.5097.7498.04
Woods98.2197.3998.3292.6797.0797.4695.8396.5098.04
Buildings-Grass-Trees-Drives84.1790.1191.3988.5675.9174.4969.7392.7894.92
Stone-Steel-Towers78.0087.3397.0084.0084.7882.6780.6787.4490.89
OA (%)84.9389.13 93.44 _ _ 86.9777.9279.4977.8993.0694.07
AA (%)70.8779.33 89.27 _ _ 80.0765.7568.3065.7688.4290.40
Kappa (%)83.0087.62 92.52 _ _ 85.1574.7376.5274.6492.1193.25
Table 3. Classification results on the Pavia University dataset.
Table 3. Classification results on the Pavia University dataset.
ClassCDCNNFDSSCDBDASSFTTDBSSANDGCNetLGCNetM3DCNNOurs
Asphalt96.3088.1787.4590.0981.5284.6985.8589.8590.82
Meadows95.3699.1599.5299.2698.7299.2989.3099.4199.28
Gravel73.6277.4582.1669.5148.8664.1357.5877.1476.09
Trees80.6584.4589.9986.7670.4374.3670.3784.6885.89
Painted metal sheets99.5799.3299.7898.8399.1999.5199.1399.7899.79
Bare Soil90.7696.4794.0196.3886.9285.1985.1397.0997.78
Bitumen86.0296.5397.9982.0059.0665.3155.5992.8095.18
Self-Blocking Bricks81.9990.4481.5678.0064.6465.2860.7085.5683.02
Shadows77.3183.3080.6172.5259.8471.6064.1763.5582.70
OA (%)91.1393.8493.4692.1785.1887.2781.6193.33 93.82 _ _
AA (%)86.84 90.59 _ _ 90.3485.9374.3578.8274.2087.7690.66
Kappa (%)88.5791.06 91.32 _ _ 89.5979.9482.8076.5391.1491.79
Table 4. Classification results on the Houston dataset.
Table 4. Classification results on the Houston dataset.
ClassCDCNNFDSSCDBDASSFTTDBSSANDGCNetLGCNetM3DCNNOurs
Healthy grass81.3988.3490.3586.9586.7989.9989.3990.7490.55
Stressed grass87.5188.7692.3989.1986.4590.6189.3389.4693.22
Synthetic grass98.5598.3298.1398.8696.7095.1396.0998.3398.94
Trees89.3388.9891.7885.9883.6883.1286.5789.1690.32
Soil98.0797.3299.5995.8397.2098.6797.7999.1199.62
Water79.1978.7682.1475.9066.9676.3472.4271.9381.34
Residential92.6886.6984.3285.2268.1375.7878.2180.9684.72
Commercial61.0065.8660.2360.6262.4967.8267.6662.8363.68
Road81.8279.4574.6174.4765.8563.5469.2373.6179.01
Highway89.5390.2590.6292.8384.0585.4586.5391.6585.99
Railway95.0893.0792.3191.1081.2985.4183.7089.9089.22
Parking Lot 174.8376.5281.8882.2474.2880.4880.0983.7682.83
Parking Lot 273.4780.8086.5180.2881.0882.5082.5686.5586.90
Tennis Court99.6099.72100.0095.6486.9694.7693.9299.0399.34
Running Track99.8699.4097.6497.9297.0099.2599.6599.5196.98
OA (%)86.3286.83 87.24 _ _ 85.7080.6483.6884.2786.5087.32
AA (%)86.7987.48 88.17 _ _ 86.2081.2684.5984.8887.1088.18
Kappa (%)85.2185.77 86.20 _ _ 84.5479.0682.3682.9985.4186.29
Table 5. Classification results on the WHU-Hi-HanChuan dataset.
Table 5. Classification results on the WHU-Hi-HanChuan dataset.
ClassCDCNNFDSSCDBDASSFTTDBSSANDGCNetLGCNetM3DCNNOurs
Strawberry87.1998.1897.2098.5897.4297.2097.0297.8398.40
Cowpea86.2894.6592.0092.7786.0689.6190.8893.1294.01
Soybean82.1988.2488.9293.0988.3288.6088.3389.9292.70
Sorghum94.4496.7397.0595.8495.3197.1096.2996.0397.42
Water-spinach58.7386.3870.7578.0250.6463.2063.8970.4785.78
Watermelon60.7874.4469.2074.6757.7161.2852.0971.6576.06
Greens86.6092.7590.6793.0282.8189.1585.9291.6692.61
Trees86.6689.9587.9587.2883.2883.4483.1088.4990.79
Grass83.3987.7187.5186.6971.6185.7277.8289.0190.12
Red-roof96.3296.1195.7896.2593.7593.5193.6597.4097.80
Gray-roof93.4496.5195.8297.1794.2395.2893.3996.2697.30
Plastic79.5386.5882.6685.1446.7855.8348.0782.8688.01
Bare-soil72.1473.4576.8673.8268.8070.7266.6277.8377.76
Road92.8093.6890.0589.6186.1189.9388.7791.8693.73
Bright-object68.9268.2766.4957.4240.7362.2953.6960.0664.64
Water99.0499.5699.6399.7699.4499.6899.6699.6899.75
OA (%)90.21 94.60 _ _ 93.6294.1890.2592.0791.0794.2695.28
AA (%)83.03 88.95 _ _ 86.7887.4477.6982.6679.9587.1389.81
Kappa (%)88.63 93.67 _ _ 92.5393.1888.5690.7089.5393.2894.47
Table 6. Model parameters and computational complexity comparison.
Table 6. Model parameters and computational complexity comparison.
MetricCDCNNFDSSCDBDASSFTTDBSSANDGCNetLGCNetM3DCNNOurs
Parameters (M)0.30220.18390.03890.14850.07213.36640.37820.44120.2957
FLOPs (G)0.05060.49640.05560.01140.13980.03440.01930.00950.0354
Test Time (s)0.39792.08120.99610.57961.463123.591029.62714.28030.2934
Table 7. Classification performance under different numbers of Mamba modules on the Indian Pines dataset.
Table 7. Classification performance under different numbers of Mamba modules on the Indian Pines dataset.
Number of Mamba ModulesOA (%)AA (%)Kappa (%)
292.1587.3290.98
3 (Ours)94.0790.4093.25
493.5889.7692.71
592.8988.6191.88
Table 8. Classification performance under different loss weight λ on the Indian Pines dataset.
Table 8. Classification performance under different loss weight λ on the Indian Pines dataset.
λ OA (%)AA (%)Kappa (%)
0.092.8688.5791.85
0.193.5289.6392.58
0.2 (Ours)94.0790.4093.25
0.393.8489.9592.99
0.493.2189.1292.27
0.592.9388.7691.93
0.692.5388.2491.48
Table 9. Classification results of different module combinations on the Indian Pines dataset.
Table 9. Classification results of different module combinations on the Indian Pines dataset.
HLFDMMCFIMFALOA (%)AA (%)Kappa (%)
×××89.2584.1187.98
××92.8688.5791.85
××90.1885.6289.03
××89.7684.8388.59
×93.6989.9292.88
×93.2489.3192.37
×90.8786.4589.81
94.0790.4093.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Teng, Y.; Gan, S.; Yuan, X. HL-Mamba: A High–Low Frequency Interaction Mamba Network for Hyperspectral Image Classification. Sensors 2026, 26, 1556. https://doi.org/10.3390/s26051556

AMA Style

Teng Y, Gan S, Yuan X. HL-Mamba: A High–Low Frequency Interaction Mamba Network for Hyperspectral Image Classification. Sensors. 2026; 26(5):1556. https://doi.org/10.3390/s26051556

Chicago/Turabian Style

Teng, Yehong, Shu Gan, and Xiping Yuan. 2026. "HL-Mamba: A High–Low Frequency Interaction Mamba Network for Hyperspectral Image Classification" Sensors 26, no. 5: 1556. https://doi.org/10.3390/s26051556

APA Style

Teng, Y., Gan, S., & Yuan, X. (2026). HL-Mamba: A High–Low Frequency Interaction Mamba Network for Hyperspectral Image Classification. Sensors, 26(5), 1556. https://doi.org/10.3390/s26051556

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop