Next Article in Journal
Landslide Susceptibility Assessment in Tongren County, Qinghai Province, Using Machine Learning and Multi–Source Data Integration: A Comparative Analysis of Models
Previous Article in Journal
S3R-GS: Saliency-Guided Gaussian Splatting for Arbitrary-Scale Spacecraft Image Super-Resolution
Previous Article in Special Issue
Semi-Supervised Remote Sensing Image Semantic Segmentation Based on Multi-Scale Consistency and Cross-Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Multi-Scale Transformer-Based Neural Architecture Search for Hyperspectral Image Classification

1
Heilongjiang Province Key Laboratory of Laser Spectroscopy Technology and Application, Harbin University of Science and Technology, Harbin 150080, China
2
School of Integrated Circuit, Shenzhen Polytechnic University, Shenzhen 518115, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2026, 18(10), 1586; https://doi.org/10.3390/rs18101586
Submission received: 9 April 2026 / Revised: 12 May 2026 / Accepted: 14 May 2026 / Published: 15 May 2026

Abstract

Hyperspectral image classification (HSIC) is a crucial task for remote sensing applications, requiring accurate pixel-level labeling while effectively capturing both spectral and spatial information. Traditional convolutional neural network architectures often struggle to balance local texture detail and global contextual consistency, and existing neural architecture search (NAS) methods rarely incorporate attention mechanisms, limiting their performance. To address these challenges, this study proposes a multi-scale Transformer-based NAS framework (TR-NAS) for fine-grained hyperspectral image classification. The framework combines local cube sampling, shallow and deep multi-scale convolutions, and a searchable Transformer module that adaptively selects global, local window, and multi-scale attention operators. Lightweight enhanced convolution operators, including dual-gated (DG-Conv) and mixed depthwise (MixConv) convolutions, are incorporated to improve spectral discrimination and scale robustness. Extensive experiments on the PU and Hanchuan datasets demonstrate that TR-NAS achieves superior classification accuracy, stability, and boundary consistency compared to traditional methods and existing NAS architectures, showing improved robustness to spectral similarity and spatial heterogeneity in complex remote sensing scenes.
Keywords: hyperspectral image classification; neural architecture search; multi-scale transformer hyperspectral image classification; neural architecture search; multi-scale transformer

Share and Cite

MDPI and ACS Style

Wang, A.; Liu, X.; Chen, H. Multi-Scale Transformer-Based Neural Architecture Search for Hyperspectral Image Classification. Remote Sens. 2026, 18, 1586. https://doi.org/10.3390/rs18101586

AMA Style

Wang A, Liu X, Chen H. Multi-Scale Transformer-Based Neural Architecture Search for Hyperspectral Image Classification. Remote Sensing. 2026; 18(10):1586. https://doi.org/10.3390/rs18101586

Chicago/Turabian Style

Wang, Aili, Xinyu Liu, and Haisong Chen. 2026. "Multi-Scale Transformer-Based Neural Architecture Search for Hyperspectral Image Classification" Remote Sensing 18, no. 10: 1586. https://doi.org/10.3390/rs18101586

APA Style

Wang, A., Liu, X., & Chen, H. (2026). Multi-Scale Transformer-Based Neural Architecture Search for Hyperspectral Image Classification. Remote Sensing, 18(10), 1586. https://doi.org/10.3390/rs18101586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop