WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion
Abstract
1. Introduction
- We explore a novel feature learning paradigm for the hierarchical DWT structure. By proposing the -FEM with a specialized alignment loss, we demonstrate that robust spatial–spectral representations can be learned directly from decoupled wavelet sub-bands, bypassing the inverse wavelet transform.
- We investigate the interaction mechanism between varying frequency layers. To this end, we design the SCA-FFM, which reveals that leveraging high-frequency details to guide the refinement of low-frequency features via cross-attention effectively mitigates the information loss in deep networks.
- We present a high-efficiency framework that achieves a superior trade-off between accuracy and speed. Comprehensive experiments on three benchmark datasets validate that our method reduces computational overhead, offering a promising solution for time-critical remote sensing applications.
2. Background and Related Works
2.1. Wavelet Compression Method
2.2. DCT Compressed-Domain Computing
2.3. DWT Compressed-Domain Computing
3. Preliminaries: Wavelet Transform Based Data Structure
4. Methods
4.1. Overview
- (1)
- Data preprocessing: The HSI classification task is pixel-level, requiring prediction results for each pixel. Since the resolution of DWT domain data cannot be spatially aligned with the original image, all sub-bands are first upsampled to the spatial size of the original data using bilinear interpolation. Then, to reduce input dimension and improve computational speed, Principal Component Analysis is applied to reduce the spectral dimension of each sub-band to . To expand the contextual spatial information of input pixels, patches are extracted with each pixel as the center and r as the radius, using the patch instead of a single pixel as the input.
- (2)
- Shallow feature extraction: Since LL, HL, LH, and HH represent features in different directions, each branch needs to be processed separately. As shown in Figure 5, we designed -FEM to extract shallow features from the four sub-bands. This module has four branches with the same network structure. Each branch extracts spectral features through 3D convolution and then spatial features through 2D convolution, obtaining shallow features from the four sub-bands. To ensure the alignment of features across different branches, we employ SFA loss. This loss functions by using the spatial layout as an intermediary, compelling the channels in each sub-band feature to exhibit specific spatial activation patterns.
- (3)
- Deep feature fusion: represents low-frequency features that retain most information and are of highest importance for classification, while other high-frequency features play auxiliary roles. To fully fuse the information from low-frequency and high-frequency features, we propose SCA-FFM. This module uses the features of HL, LH, and HH as queries, and the features of LL as keys and values to compute attention, obtaining high-level fused features.
- (4)
- Classification: Following the traditional vision transformer-based classification process, we pass the fused features into the transformer encoder, obtain the classification token after a series of nonlinear transformations, and finally obtain the classification result through a linear layer.
4.2. Model Architecture
4.2.1. Multi-Branch Multi-Scale Spatial–Spectral Feature Extraction Module
4.2.2. Sub-Band Cross-Attention Feature Fusion Module
4.3. Training Strategy
4.3.1. Sub-Band Random Masking Strategy
4.3.2. Loss Function
5. Experimental Setup
5.1. Datasets
- Indian Pines: It is acquired by NASA’s airborne visible/infrared imaging spectrometer sensor over northwestern Indiana, contains pixel imagery with 200 spectral bands (400–2500 nm) after preprocessing. The scene was captured in June, predominantly features farmland and woodlands with 16 land-cover categories. Ground truth annotations are available through Purdue University’s MultiSpec platform.
- Pavia University: It comprises a HSI captured by reflective optics system imaging spectrometer sensor over Pavia, Italy. It consists of pixels with 225 spectral bands (430–860 nm), each with a spatial resolution of 1.3 m. The image is divided into 9 classes: asphalt, meadows, gravel, trees, painted metal sheets, bare soil, bitumen, self-blocking bricks, and shadows.
- Kennedy Space Center: It is acquired by NASA’s sensor over Kennedy Space Center, Florida, on 23 March 1996, with 18 m spatial resolution from 20 km altitude, featuring 224 spectral bands (10 nm width, 400–2500 nm center wavelengths) and 176 retained after excluding water absorption and low signal-to-noise ratio bands. The dataset encompasses 13 land cover classes, developed to reflect discernible functional types, with classification challenges due to similar spectral signatures among certain vegetation.
5.2. Implementation Details
5.3. Evaluation Metrics
5.4. Baselines
6. Results and Analysis
6.1. Ablation Studies
6.1.1. Layers in Wavelet Transform
6.1.2. Sub-Band Feature Alignment
6.1.3. Number of Cross-Attention Blocks
6.1.4. Effectiveness of Different Sub-Band Random Masking Strategies
6.2. Comparative Analysis of Classification Accuracy
6.3. Efficiency Analysis
7. Discussion
7.1. Trade-Off Between Accuracy and Efficiency
7.2. Limitations and Future Work
8. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| HSI | Hyperspectral Image |
| DCT | Discrete Cosine Transform |
| DWT | Discrete Wavelet Transform |
| EBCOT | Embedded Block Coding with Optimized Truncation |
| JPEG | Joint Photographic Experts Group |
| JPEG2000 | Joint Photographic Experts Group 2000 |
| LL | low-frequency sub-band |
| HL, LH, HH | high-frequency sub-bands |
| MOE | Mixture of Experts |
| -FEM | Multi-branch Multi-scale Spatial–Spectral Feature Extraction Module |
| SCA-FFM | Sub-band Cross-attention Feature Fusion |
| SFA | Sub-band Feature Alignment |
| KA | Kappa coefficient |
| OA | Overall Accuracy |
| AA | Average Accuracy |
| SE | Squeeze-and-Excitation |
| RF | Random Forest |
| SVM | Support Vector Machine |
| L1, L2, L3 | 1st, 2nd, and 3rd Wavelet Decomposition Levels |
| L3+2 | Combination of L3 and L2 |
| L3+2+1 | Combination of L3, L2 and L1 |
| FULL | full decompression |
References
- UCLouvain; OpenJPEG Contributors. OpenJPEG: JPEG 2000 Reference Software. Open-Source Implementation Officially Recognized by ISO/IEC and ITU-T in July 2015. Available online: https://www.openjpeg.org/ (accessed on 15 January 2025).
- Chen, Z.; Li, Q.; Xia, Y. Performance analysis of AVS2 for remote sensing image compression. In Proceedings of the 2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), Chengdu, China, 14–18 July 2014; pp. 1–4. [Google Scholar]
- Yu, X.; Zhao, J.; Zhu, T.; Lan, Q.; Gao, L.; Fan, L. Analysis of JPEG2000 compression quality of optical satellite images. In Proceedings of the 2022 2nd Asia-Pacific Conference on Communications Technology and Computer Science (ACCTCS), Shenyang, China, 25–27 February 2022; pp. 500–503. [Google Scholar]
- Gueguen, L.; Sergeev, A.; Kadlec, B.; Liu, R.; Yosinski, J. Faster neural networks straight from jpeg. Adv. Neural Inf. Process. Syst. 2018, 31, 3933–3944. [Google Scholar]
- Park, J.; Johnson, J. Rgb no more: Minimally-decoded jpeg vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22334–22346. [Google Scholar]
- Li, X.; Zhang, Y.; Yuan, J.; Lu, H.; Zhu, Y. Discrete cosin transformer: Image modeling from frequency domain. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 5468–5478. [Google Scholar]
- Cheng, K.O.; Law, N.F.; Siu, W.C. Fast extraction of wavelet-based features from JPEG images for joint retrieval with JPEG2000 images. Pattern Recognit. 2010, 43, 3314–3323. [Google Scholar] [CrossRef]
- Byju, A.P.; Sumbul, G.; Demir, B.; Bruzzone, L. Remote-sensing image scene classification with deep neural networks in JPEG 2000 compressed domain. IEEE Trans. Geosci. Remote Sens. 2020, 59, 3458–3472. [Google Scholar] [CrossRef]
- Byju, A.P.; Demir, B.; Bruzzone, L. A progressive content-based image retrieval in JPEG 2000 compressed remote sensing archives. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5739–5751. [Google Scholar] [CrossRef]
- Chebil, F.; Miled, M.H.; Islam, A.; Willner, K. Compressed domain editing of jpeg2000 images. IEEE Trans. Consum. Electron. 2005, 51, 710–717. [Google Scholar] [CrossRef]
- Zhao, F.; Liu, G.; Ren, F. Adaptive blind watermarking for JPEG2000 compression domain. In Proceedings of the 2009 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 October 2009; pp. 1–5. [Google Scholar]
- Bisen, T.; Javed, M.; Nagabhushan, P.; Watanabe, O. Segmentation-less extraction of text and non-text regions from jpeg 2000 compressed document images through partial and intelligent decompression. IEEE Access 2023, 11, 20673–20687. [Google Scholar] [CrossRef]
- Li, Z.; Li, B.; Eliceiri, K.W.; Narayanan, V. Computationally efficient adaptive decompression for whole slide image processing. Biomed. Opt. Express 2023, 14, 667–686. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Y.; Cai, R.; Chen, T.; Zhang, G.; Zhang, H.; Chen, P.Y.; Chang, S.; Wang, Z.; Liu, S. Robust Mixture-of-Expert Training for Convolutional Neural Networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 90–101. [Google Scholar]
- Dai, D.; Deng, C.; Zhao, C.; Xu, R.X.; Gao, H.; Chen, D.; Li, J.; Zeng, W.; Yu, X.; Wu, Y.; et al. DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models. arXiv 2024. [Google Scholar] [CrossRef]
- Liu, Y.; Cai, Y.; Jia, Q.; Qiu, B.; Wang, W.; Pu, N. Novel Class Discovery for Ultra-Fine-Grained Visual Categorization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 17679–17688. [Google Scholar]
- Yuan, Z.; Ding, X.; Xia, X.; He, Y.; Fang, H.; Yang, B.; Fu, W. Structure-Aware Progressive Multi-Modal Fusion Network for RGB-T Crack Segmentation. J. Imaging 2025, 11, 384. [Google Scholar] [CrossRef]
- Chen, C.; Liu, X.; Zhou, M.; Li, Z.; Du, Z.; Lin, Y. Lightweight and Real-Time Driver Fatigue Detection Based on MG-YOLOv8 with Facial Multi-Feature Fusion. J. Imaging 2025, 11, 385. [Google Scholar] [CrossRef]
- Shen, J.; Chen, Y.; Liu, Y.; Zuo, X.; Fan, H.; Yang, W. ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection. Pattern Recognit. 2024, 145, 109913. [Google Scholar] [CrossRef]
- Wei, X.; Zhang, T.; Li, Y.; Zhang, Y.; Wu, F. Multi-modality cross attention network for image and sentence matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 10941–10950. [Google Scholar]
- Li, M.; Liu, Y.; Xue, G.; Huang, Y.; Yang, G. Exploring the relationship between center and neighborhoods: Central vector oriented self-similarity network for hyperspectral image classification. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1979–1993. [Google Scholar] [CrossRef]
- Jiang, M.; Su, Y.; Gao, L.; Plaza, A.; Zhao, X.L.; Sun, X.; Liu, G. GraphGST: Graph Generative Structure-Aware Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5534815. [Google Scholar] [CrossRef]
- Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef]
- Rigatti, S.J. Random forest. J. Insur. Med. 2017, 47, 31–39. [Google Scholar] [CrossRef]
- Lee, H.; Kwon, H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef]
- Zhu, M.; Jiao, L.; Liu, F.; Yang, S.; Wang, J. Residual spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 449–462. [Google Scholar] [CrossRef]
- Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.S. Spectral–spatial transformer network for hyperspectral image classification: A factorized architecture search framework. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5514715. [Google Scholar] [CrossRef]
- Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3232–3245. [Google Scholar] [CrossRef]
- Zhang, X.; Sun, G.; Jia, X.; Wu, L.; Zhang, A.; Ren, J.; Fu, H.; Yao, Y. Spectral–Spatial Self-Attention Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5512115. [Google Scholar] [CrossRef]
- Hang, R.; Li, Z.; Liu, Q.; Ghamisi, P.; Bhattacharyya, S.S. Hyperspectral image classification with attention-aided CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 59, 2281–2293. [Google Scholar] [CrossRef]
- Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-based adaptive spectral–spatial kernel ResNet for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7831–7843. [Google Scholar] [CrossRef]
- He, Y.; Tu, B.; Jiang, P.; Liu, B.; Li, J.; Plaza, A. IGroupSS-Mamba: Interval group spatial-spectral mamba for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5538817. [Google Scholar] [CrossRef]






| Method | Year | Domain | Task | Strengths | Limitations |
|---|---|---|---|---|---|
| DCT-based Methods | |||||
| Gueguen et al. [4] | 2018 | DCT | Natural Image Classification | First efficient CNN on JPEG; Significant speedup | Requires heavy CNN modification; Block artifacts |
| Jeongsoo Park et al. [5] | 2023 | DCT | Natural Image Classification | ViT structure fits block-features; Fast training | Limited to block-based DCT; No spectral modeling |
| DCFormer [6] | 2023 | DCT | Natural Image Cls/Det/Seg | Learnable frequency selection; Effective trade-off | Hard to generalize to hierarchical DWT structures |
| DWT-based Methods | |||||
| Chebil et al. [10] | 2005 | DWT | Natural Image Editing | Efficient for brightness/contrast adjustment | Signal processing only; No semantic feature learning |
| Zhao et al. [11] | 2009 | DWT | Watermarking | Effective copyright protection in bitstream | Specific to security tasks; No semantic analysis |
| Akshara et al. [8] | 2020 | DWT | RGB Remote Sensing Classification | First DL on compressed RGB remote sensing images | Limited to RGB; Simple architecture lacks fusion |
| Li et al. [13] | 2023 | DWT | Whole Slide Image Classification | Efficient pyramid structure for medical slides | Domain specific (Medical); Specific magnification logic |
| Bisen et al. [12] | 2023 | DWT | Region Extraction | Segmentation-less text extraction | Specific to document layout; Not for general classification |
| Ours | - | DWT | HSI Classification | Preserves spectral-spatial info; Cross-attention fusion | - |
| No. | Indian Pines | Pavia University | Kennedy Space Center | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Class | Train. | Val. | Test. | Class | Train. | Val. | Test. | Class | Train. | Val. | Test. | |
| 1 | Alfalfa | 4 | 1 | 41 | Asphalt | 332 | 33 | 6266 | Scrub | 76 | 8 | 677 |
| 2 | Corn-notill | 143 | 14 | 1271 | Meadows | 932 | 93 | 17,624 | Willow-swamp | 24 | 3 | 216 |
| 3 | Corn-mintill | 83 | 8 | 739 | Gravel | 104 | 11 | 1984 | Cp-hammock | 25 | 3 | 228 |
| 4 | Corn | 24 | 2 | 211 | Trees | 154 | 15 | 2895 | Cp/O-hammock | 25 | 3 | 224 |
| 5 | Grass-pasture | 48 | 5 | 430 | Painted-m-s | 67 | 7 | 1271 | Slash-pine | 16 | 2 | 143 |
| 6 | Grass-trees | 73 | 7 | 650 | Bare-soil | 252 | 25 | 4752 | Oak/B-hammock | 23 | 2 | 204 |
| 7 | Grass-pasture-m | 3 | 1 | 24 | Bitumen | 66 | 7 | 1257 | HW-swamp | 11 | 1 | 93 |
| 8 | Hay-windrowed | 48 | 5 | 425 | Self-block-b | 184 | 18 | 3480 | Graminoid-marsh | 43 | 4 | 384 |
| 9 | Oats | 2 | 1 | 17 | Shadows | 47 | 5 | 895 | Spartine-marsh | 52 | 5 | 463 |
| 10 | Soybean-notill | 97 | 10 | 865 | Cattail-marsh | 40 | 4 | 360 | ||||
| 11 | Soybean-mintill | 245 | 25 | 2185 | Salt-marsh | 42 | 4 | 373 | ||||
| 12 | Soybean-clean | 59 | 6 | 528 | Mud-flats | 50 | 5 | 448 | ||||
| 13 | Wheat | 21 | 2 | 182 | Water | 93 | 9 | 825 | ||||
| 14 | Woods | 126 | 13 | 1126 | ||||||||
| 15 | Buildings-g-t | 39 | 4 | 343 | ||||||||
| 16 | Stone-steel-t | 9 | 1 | 83 | ||||||||
| Total | 1024 | 105 | 9120 | 2138 | 214 | 40,424 | 520 | 53 | 4638 | |||
| Dataset | L3 | L2 | L1 | L3+2 | L3+2+1 | Full Decompression | |
|---|---|---|---|---|---|---|---|
| Inverse Wavelet Transform | / | L3 → L2 | L3 → L2 → L1 | L3 → L2 | L3 → L2 → L1 | L3 → L2 → L1 → FULL | |
| Indian Pines | KA (%) | 96.61 | 97.27 | 98.27 | 98.63 | 98.68 | / |
| OA (%) | 96.78 | 97.43 | 98.48 | 98.80 | 98.84 | / | |
| AA (%) | 97.03 | 96.28 | 98.98 | 98.08 | 99.02 | / | |
| Decode Time (ms) | / | 5.9 | 14.8 | 5.9 | 14.8 | 33.9 | |
| Pavia University | KA (%) | 94.39 | 98.16 | 99.17 | 99.66 | 99.63 | / |
| OA (%) | 95.75 | 98.23 | 99.25 | 99.74 | 99.72 | / | |
| AA (%) | 94.12 | 97.90 | 98.80 | 99.39 | 99.31 | / | |
| Decode Time (ms) | / | 10.4 | 37.8 | 10.4 | 37.8 | 111.7 | |
| Kennedy Space Center | KA (%) | 96.59 | 97.19 | 99.15 | 99.19 | 99.46 | / |
| OA (%) | 96.70 | 97.29 | 99.22 | 99.25 | 99.50 | / | |
| AA (%) | 96.92 | 97.36 | 99.34 | 99.39 | 99.51 | / | |
| Decode Time (ms) | / | 41.1 | 173.9 | 41.1 | 173.9 | 1305.0 |
| Metric | 1 | ||||
|---|---|---|---|---|---|
| KA (%) | 98.46 | 98.89 | 99.39 | 98.89 | 98.65 |
| OA (%) | 98.84 | 99.16 | 99.54 | 99.16 | 98.98 |
| AA (%) | 97.73 | 98.27 | 99.07 | 98.18 | 97.92 |
| Metric | L3 | L2 | L1 | L3+2 | L3+2+1 | |
|---|---|---|---|---|---|---|
| KA (%) | 98.67 | 99.25 | 99.30 | 99.13 | 99.50 | |
| Without SFA | OA (%) | 98.99 | 99.44 | 99.47 | 99.34 | 99.63 |
| AA (%) | 97.99 | 98.84 | 99.06 | 98.64 | 99.24 | |
| KA (%) | 98.70 | 99.32 | 99.41 | 99.39 | 99.52 | |
| With SFA | OA (%) | 99.02 | 99.49 | 99.56 | 99.54 | 99.64 |
| AA (%) | 98.07 | 99.03 | 99.27 | 99.07 | 99.19 |
| Metric | 1 | 2 | 3 | 4 | 5 | 6 | |
|---|---|---|---|---|---|---|---|
| KA (%) | 97.79 | 98.14 | 98.58 | 98.64 | 98.73 | 98.49 | |
| L3+2 | OA (%) | 97.94 | 98.31 | 98.74 | 98.79 | 98.88 | 98.68 |
| AA (%) | 97.52 | 97.59 | 98.14 | 97.66 | 98.61 | 98.58 | |
| KA (%) | 98.44 | 98.68 | 98.97 | 98.76 | 99.01 | 98.63 | |
| L3+2+1 | OA (%) | 98.63 | 98.84 | 99.10 | 98.92 | 99.13 | 98.80 |
| AA (%) | 98.24 | 99.02 | 98.68 | 99.07 | 98.38 | 98.55 |
| Metric | No Mask | All | All | No LL | No LL | |
|---|---|---|---|---|---|---|
| =0.1 | =0.01 | Others = 0.1 | Others = 0.01 | |||
| KA (%) | 98.63 | 98.73 | 98.50 | 98.43 | 98.40 | |
| L3+L2 | OA (%) | 98.80 | 98.88 | 98.69 | 98.62 | 98.60 |
| AA (%) | 98.08 | 98.56 | 97.91 | 97.53 | 98.33 | |
| KA (%) | 98.68 | 98.80 | 98.26 | 98.73 | 98.59 | |
| L3+2+1 | OA (%) | 98.84 | 98.95 | 98.47 | 98.88 | 98.76 |
| AA (%) | 99.02 | 99.09 | 98.62 | 99.08 | 98.39 |
| Class | SVM (2004) | RF (2005) | ContextNet (2017) | RSSAN (2020) | SSTN (2022) | SSAN (2020) | SSSAN (2022) | SSAtt (2021) | A2S2KResNet (2021) | CVSSN (2023) | IGroupSS-Mamba (2024) | GraphGST (2024) | Ours |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Alfalfa | 46.95 ± 13.99 | 71.04 ± 22.75 | 95.86 ± 6.64 | 93.68 ± 11.46 | 84.29 ± 22.19 | 97.08 ± 4.52 | 89.73 ± 9.99 | 97.10 ± 6.48 | 95.38 ± 5.56 | 94.63 ± 5.24 | 97.56 ± 1.99 | 99.21 ± 1.12 | 99.15 ± 1.21 |
| Corn-notill | 75.08 ± 1.92 | 68.77 ± 1.79 | 79.48 ± 3.54 | 79.81 ± 5.57 | 91.15 ± 4.90 | 86.33 ± 2.47 | 93.09 ± 2.85 | 94.17 ± 1.54 | 98.08 ± 0.94 | 97.55 ± 0.76 | 93.07 ± 3.00 | 95.64 ± 1.37 | 95.25 ± 0.56 |
| Corn-mintill | 76.99 ± 4.31 | 75.93 ± 2.89 | 82.01 ± 5.14 | 85.31 ± 4.82 | 97.78 ± 1.23 | 85.71 ± 3.27 | 95.21 ± 1.92 | 96.92 ± 1.18 | 98.80 ± 0.90 | 99.05 ± 0.86 | 99.78 ± 0.32 | 96.16 ± 0.33 | 99.72 ± 0.31 |
| Corn | 63.25 ± 4.34 | 57.31 ± 5.18 | 85.20 ± 4.65 | 79.49 ± 14.25 | 96.07 ± 3.75 | 88.87 ± 6.27 | 92.11 ± 3.79 | 97.71 ± 1.73 | 96.19 ± 3.55 | 98.57 ± 1.60 | 98.59 ± 1.15 | 93.61 ± 0.96 | 100.00 ± 0.00 |
| Grass-pasture | 89.68 ± 3.00 | 85.67 ± 3.20 | 94.15 ± 3.13 | 92.09 ± 4.68 | 93.02 ± 6.48 | 93.97 ± 3.57 | 97.50 ± 1.80 | 97.26 ± 2.07 | 98.09 ± 1.74 | 98.10 ± 1.45 | 98.39 ± 0.99 | 95.78 ± 0.39 | 99.43 ± 0.23 |
| Grass-trees | 89.16 ± 1.55 | 81.36 ± 1.45 | 95.88 ± 1.78 | 94.38 ± 2.22 | 98.31 ± 1.25 | 96.08 ± 2.48 | 98.34 ± 2.38 | 98.64 ± 1.84 | 99.48 ± 0.88 | 99.62 ± 0.34 | 98.22 ± 1.76 | 99.95 ± 0.07 | 99.68 ± 0.13 |
| Grass-pasture-m | 85.52 ± 8.07 | 32.95 ± 42.12 | 90.79 ± 11.24 | 85.43 ± 12.16 | 83.60 ± 13.08 | 90.59 ± 11.43 | 93.40 ± 7.44 | 97.65 ± 4.71 | 92.96 ± 10.23 | 94.33 ± 8.69 | 100.00 ± 0.00 | 96.15 ± 0.00 | 95.83 ± 3.40 |
| Hay-windrowed | 91.34 ± 1.49 | 85.23 ± 1.72 | 95.70 ± 2.79 | 94.73 ± 2.57 | 98.77 ± 2.50 | 97.00 ± 2.78 | 98.08 ± 1.33 | 98.45 ± 1.13 | 99.00 ± 1.23 | 99.93 ± 0.21 | 100.00 ± 0.00 | 100.00 ± 0.00 | 99.92 ± 0.12 |
| Oats | 65.94 ± 13.55 | 53.33 ± 44.44 | 91.59 ± 10.96 | 92.24 ± 10.68 | 73.47 ± 18.20 | 86.58 ± 14.01 | 85.73 ± 10.84 | 95.96 ± 5.62 | 97.09 ± 3.83 | 94.97 ± 10.08 | 81.48 ± 18.33 | 75.93 ± 6.93 | 92.16 ± 5.54 |
| Soybean-notill | 74.10 ± 2.06 | 72.77 ± 4.40 | 87.61 ± 3.93 | 86.52 ± 4.28 | 92.75 ± 3.58 | 91.10 ± 3.54 | 95.56 ± 2.26 | 94.84 ± 2.34 | 96.77 ± 1.90 | 97.53 ± 1.05 | 97.52 ± 1.50 | 95.93 ± 0.75 | 97.66 ± 0.55 |
| Soybean-mintill | 75.63 ± 1.35 | 69.60 ± 1.47 | 88.42 ± 2.11 | 85.57 ± 4.28 | 95.27 ± 2.12 | 90.54 ± 1.69 | 96.18 ± 1.22 | 95.26 ± 1.57 | 98.29 ± 0.87 | 98.31 ± 0.70 | 98.99 ± 0.35 | 93.82 ± 0.70 | 99.11 ± 0.16 |
| Soybean-clean | 78.06 ± 4.72 | 62.62 ± 5.73 | 75.12 ± 6.35 | 73.23 ± 8.45 | 90.77 ± 6.49 | 84.54 ± 4.02 | 89.46 ± 3.36 | 92.40 ± 2.98 | 96.28 ± 2.00 | 97.34 ± 1.11 | 92.63 ± 5.26 | 90.45 ± 1.00 | 98.48 ± 0.52 |
| Wheat | 91.63 ± 5.35 | 85.38 ± 4.80 | 97.28 ± 2.93 | 93.58 ± 5.95 | 97.62 ± 2.54 | 96.76 ± 3.48 | 98.58 ± 2.68 | 99.89 ± 0.22 | 99.37 ± 1.28 | 99.63 ± 0.79 | 98.74 ± 0.67 | 99.10 ± 0.25 | 99.04 ± 1.35 |
| Woods | 91.48 ± 1.23 | 89.70 ± 0.98 | 95.68 ± 2.35 | 95.33 ± 1.30 | 97.16 ± 1.69 | 96.02 ± 1.34 | 97.98 ± 1.08 | 97.65 ± 0.87 | 98.89 ± 0.71 | 98.68 ± 0.64 | 99.94 ± 0.08 | 99.50 ± 0.04 | 99.87 ± 0.09 |
| Buildings-g-t | 78.47 ± 5.12 | 63.49 ± 5.85 | 89.88 ± 3.05 | 86.58 ± 3.82 | 91.85 ± 3.99 | 90.27 ± 4.52 | 93.80 ± 2.23 | 94.67 ± 1.46 | 94.62 ± 2.19 | 95.81 ± 2.07 | 96.54 ± 0.71 | 98.95 ± 0.14 | 98.78 ± 0.66 |
| Stone-steel-t | 97.44 ± 2.67 | 97.12 ± 2.56 | 90.53 ± 8.17 | 90.54 ± 9.41 | 89.30 ± 9.42 | 92.84 ± 6.14 | 95.22 ± 5.59 | 95.28 ± 6.12 | 96.45 ± 3.11 | 95.95 ± 3.28 | 96.03 ± 1.48 | 99.21 ± 0.56 | 97.47 ± 0.00 |
| KA | 77.42 ± 0.59 | 71.28 ± 0.92 | 86.32 ± 1.88 | 84.97 ± 3.98 | 93.68 ± 2.70 | 89.59 ± 1.13 | 94.87 ± 0.69 | 95.34 ± 0.60 | 97.70 ± 0.44 | 97.95 ± 0.24 | 97.30 ± 1.30 | 95.59 ± 0.36 | 98.41 ± 0.19 |
| OA | 80.33 ± 0.50 | 75.12 ± 0.82 | 88.04 ± 1.64 | 86.88 ± 3.45 | 94.46 ± 2.37 | 90.89 ± 0.98 | 95.51 ± 0.61 | 95.92 ± 0.53 | 97.98 ± 0.39 | 98.20 ± 0.21 | 97.63 ± 1.14 | 96.12 ± 0.31 | 98.60 ± 0.17 |
| AA | 79.42 ± 1.69 | 72.02 ± 3.91 | 89.70 ± 1.97 | 88.03 ± 3.34 | 91.95 ± 3.54 | 91.52 ± 1.48 | 94.37 ± 1.23 | 96.49 ± 0.56 | 97.23 ± 1.26 | 97.50 ± 1.07 | 96.72 ± 1.97 | 95.59 ± 0.66 | 98.22 ± 0.74 |
| Class | SVM (2004) | RF (2005) | ContextNet (2017) | RSSAN (2020) | SSTN (2022) | SSAN (2020) | SSSAN (2022) | SSAtt (2021) | A2S2KResNet (2021) | CVSSN (2023) | IGroupSS-Mamba (2024) | GraphGST (2024) | Ours |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Asphalt | 91.23 ± 0.85 | 89.85 ± 1.34 | 97.15 ± 0.58 | 96.51 ± 0.62 | 97.66 ± 1.04 | 97.71 ± 0.51 | 99.05 ± 0.45 | 97.76 ± 0.58 | 99.48 ± 0.16 | 99.63 ± 0.19 | 98.87 ± 1.22 | 98.65 ± 0.18 | 99.94 ± 0.04 |
| Meadows | 92.64 ± 0.35 | 87.99 ± 0.42 | 99.15 ± 0.26 | 99.55 ± 0.29 | 99.49 ± 0.52 | 99.66 ± 0.17 | 99.67 ± 0.16 | 99.76 ± 0.12 | 99.91 ± 0.07 | 99.94 ± 0.04 | 99.79 ± 0.10 | 99.89 ± 0.07 | 99.99 ± 0.00 |
| Gravel | 82.30 ± 1.37 | 73.81 ± 2.27 | 92.70 ± 2.41 | 93.36 ± 2.21 | 96.84 ± 2.66 | 95.57 ± 1.40 | 97.79 ± 1.05 | 97.61 ± 1.13 | 99.25 ± 0.70 | 99.43 ± 0.54 | 97.48 ± 1.88 | 91.29 ± 1.13 | 99.95 ± 0.04 |
| Trees | 96.42 ± 0.68 | 93.32 ± 1.00 | 99.12 ± 0.70 | 99.23 ± 0.39 | 97.18 ± 1.75 | 99.46 ± 0.39 | 99.54 ± 0.28 | 99.29 ± 0.51 | 99.64 ± 0.29 | 99.33 ± 0.43 | 99.03 ± 0.29 | 96.11 ± 1.57 | 98.55 ± 0.18 |
| Painted-m-s | 98.83 ± 0.82 | 97.16 ± 1.39 | 99.76 ± 0.51 | 98.76 ± 0.81 | 99.24 ± 0.73 | 99.70 ± 0.36 | 99.70 ± 0.55 | 99.69 ± 0.28 | 99.60 ± 0.38 | 99.82 ± 0.15 | 99.61 ± 0.17 | 100.00 ± 0.00 | 99.12 ± 0.37 |
| Bare-soil | 94.06 ± 0.70 | 87.46 ± 1.43 | 98.67 ± 0.32 | 98.58 ± 0.51 | 99.33 ± 1.80 | 98.94 ± 0.62 | 99.37 ± 0.36 | 99.03 ± 0.72 | 99.74 ± 0.24 | 99.70 ± 0.15 | 99.36 ± 0.91 | 99.78 ± 0.02 | 100.00 ± 0.00 |
| Bitumen | 85.47 ± 3.02 | 81.71 ± 2.91 | 96.15 ± 1.50 | 96.04 ± 2.02 | 98.97 ± 1.30 | 97.31 ± 1.91 | 99.46 ± 0.57 | 98.31 ± 1.38 | 99.63 ± 0.54 | 99.65 ± 0.31 | 100.00 ± 0.00 | 96.88 ± 0.40 | 99.83 ± 0.13 |
| Self-block-b | 80.59 ± 2.41 | 77.62 ± 2.11 | 92.92 ± 1.43 | 92.27 ± 2.20 | 96.38 ± 2.10 | 94.79 ± 1.14 | 97.12 ± 0.84 | 95.67 ± 0.97 | 98.50 ± 0.54 | 99.13 ± 0.34 | 99.17 ± 0.39 | 94.86 ± 1.20 | 99.51 ± 0.14 |
| Shadows | 99.94 ± 0.07 | 99.88 ± 0.09 | 99.43 ± 0.88 | 98.17 ± 1.12 | 96.37 ± 1.48 | 99.30 ± 0.52 | 99.57 ± 0.67 | 99.93 ± 0.11 | 99.81 ± 0.22 | 99.51 ± 0.57 | 97.15 ± 1.39 | 98.28 ± 0.31 | 98.08 ± 0.87 |
| KA | 88.43 ± 0.42 | 82.93 ± 0.31 | 97.15 ± 0.45 | 97.13 ± 0.62 | 98.02 ± 0.67 | 98.08 ± 0.35 | 98.95 ± 0.22 | 98.44 ± 0.32 | 99.51 ± 0.12 | 99.60 ± 0.08 | 99.09 ± 0.58 | 97.92 ± 0.09 | 99.69 ± 0.07 |
| OA | 91.38 ± 0.31 | 87.38 ± 0.23 | 97.85 ± 0.34 | 97.84 ± 0.46 | 98.51 ± 0.50 | 98.55 ± 0.26 | 99.21 ± 0.17 | 98.82 ± 0.24 | 99.63 ± 0.09 | 99.70 ± 0.06 | 99.32 ± 0.44 | 98.44 ± 0.07 | 99.77 ± 0.05 |
| AA | 91.28 ± 0.42 | 87.65 ± 0.51 | 97.23 ± 0.54 | 96.94 ± 0.68 | 97.94 ± 0.57 | 98.05 ± 0.38 | 99.03 ± 0.20 | 98.56 ± 0.31 | 99.51 ± 0.15 | 99.57 ± 0.10 | 98.94 ± 0.58 | 97.30 ± 0.20 | 99.63 ± 0.27 |
| Class | SVM (2004) | RF (2005) | ContextNet (2017) | RSSAN (2020) | SSTN (2022) | SSAN (2020) | SSSAN (2022) | SSAtt (2021) | A2S2KResNet (2021) | CVSSN (2023) | IGroupSS-Mamba (2024) | GraphGST (2024) | Ours |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Scrub | 91.16 ± 1.33 | 91.00 ± 1.38 | 98.22 ± 1.60 | 98.09 ± 2.35 | 99.37 ± 1.48 | 98.86 ± 1.03 | 99.21 ± 1.12 | 99.32 ± 0.87 | 99.48 ± 0.58 | 99.76 ± 0.42 | 98.15 ± 1.85 | 100.00 ± 0.00 | 99.85 ± 0.21 |
| Willow-swamp | 89.32 ± 3.72 | 77.79 ± 5.87 | 81.89 ± 10.01 | 93.38 ± 3.88 | 95.75 ± 3.19 | 91.51 ± 7.16 | 96.06 ± 2.37 | 95.80 ± 2.41 | 96.88 ± 2.35 | 94.27 ± 12.80 | 78.54 ± 0.37 | 88.89 ± 6.87 | 95.17 ± 4.01 |
| Cp-hammock | 69.71 ± 5.57 | 87.72 ± 4.81 | 77.13 ± 4.58 | 79.34 ± 6.55 | 92.39 ± 6.34 | 77.02 ± 5.09 | 90.09 ± 5.79 | 89.41 ± 4.05 | 87.87 ± 4.52 | 97.61 ± 1.83 | 89.28 ± 2.57 | 94.53 ± 1.02 | 99.85 ± 0.21 |
| Cp/O-hammock | 49.07 ± 3.26 | 59.60 ± 4.32 | 70.78 ± 6.43 | 72.83 ± 7.56 | 80.74 ± 7.31 | 73.63 ± 2.28 | 82.99 ± 7.81 | 84.46 ± 6.04 | 86.32 ± 4.13 | 90.47 ± 6.49 | 73.57 ± 7.18 | 44.58 ± 8.13 | 84.23 ± 2.21 |
| Slash-pine | 70.71 ± 10.19 | 70.98 ± 6.46 | 65.16 ± 5.78 | 71.08 ± 6.17 | 82.08 ± 8.37 | 73.06 ± 7.51 | 80.52 ± 8.11 | 84.07 ± 6.02 | 87.01 ± 5.59 | 85.10 ± 28.64 | 85.75 ± 0.65 | 0.22 ± 0.31 | 99.77 ± 0.33 |
| Oak/B-hammock | 71.89 ± 7.47 | 61.87 ± 7.26 | 85.64 ± 5.36 | 83.83 ± 8.51 | 90.58 ± 5.14 | 85.72 ± 4.37 | 91.59 ± 6.40 | 90.27 ± 5.75 | 93.61 ± 5.41 | 90.81 ± 14.63 | 89.16 ± 5.02 | 78.75 ± 19.68 | 98.51 ± 1.08 |
| HW-swamp | 75.95 ± 4.12 | 73.51 ± 3.32 | 86.99 ± 6.69 | 90.92 ± 7.31 | 91.02 ± 5.74 | 85.01 ± 8.88 | 89.58 ± 9.94 | 95.23 ± 6.37 | 95.22 ± 5.03 | 95.92 ± 10.61 | 100.00 ± 0.00 | 86.33 ± 7.93 | 97.10 ± 1.36 |
| Graminoid-marsh | 88.36 ± 3.78 | 76.33 ± 4.62 | 93.18 ± 2.47 | 95.85 ± 2.51 | 96.51 ± 2.30 | 95.25 ± 4.01 | 97.61 ± 2.35 | 97.88 ± 1.45 | 97.50 ± 2.20 | 99.61 ± 0.60 | 94.24 ± 2.27 | 88.13 ± 2.09 | 95.78 ± 2.05 |
| Spartine-marsh | 89.17 ± 2.63 | 84.66 ± 2.57 | 96.03 ± 2.05 | 97.73 ± 1.67 | 98.62 ± 1.18 | 97.10 ± 2.71 | 99.20 ± 0.87 | 99.13 ± 1.35 | 99.15 ± 0.81 | 99.33 ± 1.16 | 93.80 ± 3.04 | 100.00 ± 0.00 | 99.85 ± 0.21 |
| Cattail-marsh | 97.22 ± 2.94 | 89.19 ± 6.49 | 94.57 ± 2.89 | 96.69 ± 2.06 | 98.69 ± 1.78 | 96.55 ± 2.30 | 99.39 ± 1.32 | 99.86 ± 0.23 | 99.83 ± 0.28 | 99.97 ± 0.09 | 79.49 ± 5.63 | 98.61 ± 0.75 | 99.81 ± 0.26 |
| Salt-marsh | 96.51 ± 1.19 | 97.59 ± 1.94 | 99.45 ± 0.79 | 99.57 ± 0.58 | 98.82 ± 2.19 | 99.56 ± 0.66 | 99.95 ± 0.16 | 99.74 ± 0.63 | 99.45 ± 0.85 | 99.95 ± 0.11 | 99.73 ± 0.38 | 95.91 ± 0.31 | 99.73 ± 0.00 |
| Mud-flats | 93.37 ± 2.35 | 92.11 ± 2.71 | 94.97 ± 3.54 | 94.45 ± 2.97 | 98.61 ± 1.69 | 96.90 ± 2.06 | 99.60 ± 0.46 | 98.94 ± 0.88 | 98.38 ± 0.75 | 99.58 ± 0.61 | 97.72 ± 1.33 | 100.00 ± 0.00 | 99.77 ± 0.18 |
| Water | 99.80 ± 0.24 | 98.41 ± 0.63 | 99.18 ± 1.10 | 99.06 ± 0.98 | 99.38 ± 1.05 | 99.57 ± 0.62 | 99.81 ± 0.38 | 99.85 ± 0.36 | 99.95 ± 0.11 | 100.00 ± 0.00 | 100.00 ± 0.00 | 100.00 ± 0.00 | 100.00 ± 0.00 |
| KA | 86.87 ± 0.46 | 84.85 ± 1.10 | 91.52 ± 1.49 | 93.04 ± 1.33 | 95.84 ± 0.99 | 93.24 ± 1.52 | 96.32 ± 0.92 | 96.61 ± 0.55 | 96.85 ± 0.50 | 97.64 ± 2.88 | 92.29 ± 0.96 | 89.76 ± 1.74 | 98.25 ± 0.33 |
| OA | 88.22 ± 0.42 | 86.41 ± 0.99 | 92.39 ± 1.34 | 93.75 ± 1.19 | 96.26 ± 0.89 | 93.93 ± 1.36 | 96.70 ± 0.83 | 96.95 ± 0.50 | 97.17 ± 0.45 | 97.88 ± 2.58 | 93.08 ± 0.86 | 90.82 ± 1.56 | 98.43 ± 0.29 |
| AA | 83.25 ± 1.25 | 81.60 ± 1.55 | 87.94 ± 2.17 | 90.22 ± 2.06 | 94.04 ± 1.14 | 89.98 ± 2.33 | 94.28 ± 1.70 | 94.92 ± 1.06 | 95.43 ± 0.87 | 96.34 ± 4.96 | 90.72 ± 0.80 | 82.77 ± 2.15 | 97.65 ± 0.48 |
| Dataset | Metric | IGroupSS-Mamba | GraphGST | Ours |
|---|---|---|---|---|
| (2024) | (2024) | (L3+2+1) | ||
| Indian Pines () | Decode time | 33.9 | 33.9 | 14.8 |
| Inference time | 1475.7 | 156.4 | 433.3 | |
| Total time | 1509.6 | 190.3 | 448.1 | |
| Pavia University () | Decode time | 111.7 | 111.7 | 37.8 |
| Inference time | 7330.1 | 475.3 | 1384.7 | |
| Total time | 7441.8 | 587.0 | 1422.5 | |
| Kennedy Space Center () | Decode time | 1305.0 | 1305.0 | 173.9 |
| Inference time | 1888.5 | 78.6 | 342.7 | |
| Total time | 3193.5 | 1383.6 | 516.6 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, X.; Sun, B. WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion. J. Imaging 2025, 11, 441. https://doi.org/10.3390/jimaging11120441
Li X, Sun B. WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion. Journal of Imaging. 2025; 11(12):441. https://doi.org/10.3390/jimaging11120441
Chicago/Turabian StyleLi, Xin, and Baile Sun. 2025. "WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion" Journal of Imaging 11, no. 12: 441. https://doi.org/10.3390/jimaging11120441
APA StyleLi, X., & Sun, B. (2025). WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion. Journal of Imaging, 11(12), 441. https://doi.org/10.3390/jimaging11120441

