Next Article in Journal
Reconstructing a Century of Urban Growth Through Deep Learning-Based Colorization and Segmentation of Historical Aerial and Satellite Imagery: Les Sables-d’Olonne, France (1920–2024)
Previous Article in Journal
Using Satellite-Based NDVI to Monitor Subtle Changes in Native Grassland Condition Across Multiple Years
Previous Article in Special Issue
Unsupervised Hyperspectral Image Denoising via Spectral Learning Preference of Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

DualMambaFormer: A Parallel Hybrid Transformer–Mamba Network for Hyperspectral Image Classification

1
China Three Gorges Construction Engineering Corporation, Chengdu 610041, China
2
Changjiang River Scientific Research Institute, Changjiang Water Resources Committee, Wuhan 430010, China
3
School of Geosciences, Yangtze University, Wuhan 430100, China
4
Hubei Engineering Research Center of Unconventional Petroleum Geology and Engineering, Wuhan 430100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2026, 18(10), 1516; https://doi.org/10.3390/rs18101516
Submission received: 25 March 2026 / Revised: 3 May 2026 / Accepted: 9 May 2026 / Published: 11 May 2026

Abstract

Hyperspectral image classification (HSIC) plays a crucial role in fine-grained Earth observation tasks. However, balancing efficient long-range dependency modeling with the extraction of fine-grained local features remains a significant challenge, primarily due to the inherent high-dimensional spectral redundancy and complex spatial variability of hyperspectral data. Existing modeling paradigms exhibit distinct limitations: Convolutional Neural Networks (CNNs) are constrained by localized receptive fields, while Vision Transformers (ViTs), despite their global receptive capabilities, incur prohibitive quadratic computational complexity. Meanwhile, the emerging Mamba architecture has demonstrated remarkable effectiveness in sequence modeling with linear complexity, but it often lacks sufficient sensitivity to local textures when directly applied to non-causal 2D images. To address these limitations, this paper proposes a novel parallel hybrid architecture termed DualMambaFormer. Deviating from the traditional serial stacking paradigm, the proposed network utilizes a dual-stream design to achieve the complementary fusion of global static attention and dynamic sequence reasoning. Specifically, the model first employs an SS-ResNet for spectral dimensionality reduction and local feature embedding. Subsequently, the architecture bifurcates into a parallel encoding stage: one branch leverages Multi-Head Self-Attention (MHSA) to capture global spatial correlations, while the other introduces a Local Enhanced Mamba (LEM) branch. By integrating State Space Models (SSM) with depthwise separable convolutions, the LEM branch simultaneously captures long-range causal dependencies and local spatial context. Finally, a dual class token fusion strategy is designed to integrate heterogeneous representations at the decision level. Extensive experiments on four benchmark datasets—Indian Pines, Pavia University, Salinas, and WHU-HongHu—show that DualMambaFormer achieves OA values of 96.56%, 98.95%, 97.60%, and 96.09%, respectively, with consistently high AA and Kappa coefficients. These results demonstrate the effectiveness, robustness, and generalization capability of the proposed method for hyperspectral image classification. Compared with the second-best competing methods, DualMambaFormer improves OA by 5.55, 2.30, 1.68, and 4.30 percentage points on the Pavia University, Indian Pines, Salinas, and WHU-HongHu datasets, respectively.
Keywords: hyperspectral image classification; Mamba; transformer; parallel hybrid architecture; state space models hyperspectral image classification; Mamba; transformer; parallel hybrid architecture; state space models

Share and Cite

MDPI and ACS Style

Yu, J.; Li, J.; Sun, G.; Lu, J.; Cheng, X.; Zhou, R.; Sun, W.; Gao, X. DualMambaFormer: A Parallel Hybrid Transformer–Mamba Network for Hyperspectral Image Classification. Remote Sens. 2026, 18, 1516. https://doi.org/10.3390/rs18101516

AMA Style

Yu J, Li J, Sun G, Lu J, Cheng X, Zhou R, Sun W, Gao X. DualMambaFormer: A Parallel Hybrid Transformer–Mamba Network for Hyperspectral Image Classification. Remote Sensing. 2026; 18(10):1516. https://doi.org/10.3390/rs18101516

Chicago/Turabian Style

Yu, Jiang, Jingwei Li, Gan Sun, Jingying Lu, Xuejun Cheng, Ruimeng Zhou, Wei Sun, and Xianjun Gao. 2026. "DualMambaFormer: A Parallel Hybrid Transformer–Mamba Network for Hyperspectral Image Classification" Remote Sensing 18, no. 10: 1516. https://doi.org/10.3390/rs18101516

APA Style

Yu, J., Li, J., Sun, G., Lu, J., Cheng, X., Zhou, R., Sun, W., & Gao, X. (2026). DualMambaFormer: A Parallel Hybrid Transformer–Mamba Network for Hyperspectral Image Classification. Remote Sensing, 18(10), 1516. https://doi.org/10.3390/rs18101516

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop