Next Article in Journal
TENet: A Texture-Enhanced Network for Intertidal Sediment and Habitat Classification in Multiband PolSAR Images
Next Article in Special Issue
A Super-Resolution Reconstruction Model for Remote Sensing Image Based on Generative Adversarial Networks
Previous Article in Journal
Infrared–Visible Image Fusion through Feature-Based Decomposition and Domain Normalization
Previous Article in Special Issue
FusionHeightNet: A Multi-Level Cross-Fusion Method from Multi-Source Remote Sensing Images for Urban Building Height Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

S3L: Spectrum Transformer for Self-Supervised Learning in Hyperspectral Image Classification

1
State Key Laboratory of Dynamic Measurement Technology, School of Instrument and Electronics, North University of China, Taiyuan 030051, China
2
Department of Transportation Information Engineering, Henan College of Transportation, Zhengzhou 451460, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(6), 970; https://doi.org/10.3390/rs16060970
Submission received: 24 January 2024 / Revised: 4 March 2024 / Accepted: 8 March 2024 / Published: 10 March 2024

Abstract

:
In the realm of Earth observation and remote sensing data analysis, the advancement of hyperspectral imaging (HSI) classification technology is of paramount importance. Nevertheless, the intricate nature of hyperspectral data, coupled with the scarcity of labeled data, presents significant challenges in this domain. To mitigate these issues, we introduce a self-supervised learning algorithm predicated on a spectral transformer for HSI classification under conditions of limited labeled data, with the objective of enhancing the efficacy of HSI classification. The S3L algorithm operates in two distinct phases: pretraining and fine-tuning. During the pretraining phase, the algorithm learns the spatial representation of HSI from unlabeled data, utilizing a masking mechanism and a spectral transformer, thereby augmenting the sequence dependence of spectral features. Subsequently, in the fine-tuning phase, labeled data is employed to refine the pretrained weights, thereby improving the precision of HSI classification. Within the comprehensive encoder–decoder framework, we propose a novel spectral transformer module specifically engineered to synergize spatial feature extraction with spectral domain analysis. This innovative module adeptly navigates the complex interplay among various spectral bands, capturing both global and sequential spectral dependencies. Uniquely, it incorporates a gated recurrent unit (GRU) layer within the encoder to enhance its ability to process spectral sequences. Our experimental evaluations across several public datasets reveal that our proposed method, distinguished by its spectral transformer, achieves superior classification performance, particularly in scenarios with limited labeled samples, outperforming existing state-of-the-art approaches.

1. Introduction

Hyperspectral imaging (HSI), a pioneering remote sensing technology, has proven its distinct benefits across various domains in recent years. This technology surpasses traditional imaging by offering richer, more detailed data for precise object identification and analysis. It achieves this by capturing continuous spectral information ranging from visible light to near-infrared bands. The evolution of HSI has not only revolutionized traditional Earth observation and environmental monitoring but also demonstrated its exceptional application value in urban planning, disaster management, and agriculture [1,2,3,4]. However, the intricate and high-dimensional nature of hyperspectral data poses significant challenges to effective data analysis and processing. This complexity necessitates ongoing research to discover more efficient data processing and analysis methods.
In the initial stages, traditional methods such as support vector machines (SVMs) [5,6], random forests (RFs) [7], k-NN [8], and PCA [9] were commonly employed for hyperspectral image classification. Among these, the SVM model, despite its compactness, struggled with identifying an appropriate nonlinear kernel function. RF, by integrating multiple decision trees, exhibited superior generalization capabilities but demanded substantial computational resources. While k-NN offered flexibility by adjusting the k value to suit different problems, it often underperformed with category-imbalanced datasets. PCA, based on linear assumptions, found it challenging to handle nonlinear data structures. Furthermore, these traditional methods typically relied on manually designed features and lacked the capacity to extract deep features, rendering them often ineffective for hyperspectral image classification tasks.
In recent years, the swift advancement of deep learning technology has significantly propelled hyperspectral image classification technology. The goal is to harness the capabilities of sophisticated neural network architectures for more robust and efficient classification. Early deep learning methods, such as stacked autoencoders (SAEs) and deep belief networks (DBNs) [10], set the groundwork by emphasizing feature extraction through fully connected layers. However, these methods’ reliance on fully connected layers led to a large number of network parameters and necessitated diverse training data.
To mitigate these issues, the convolutional neural network (CNN) [11] was introduced into HSI classification. Qing et al. [12] proposed a multiscale residual convolutional neural network model, MRA-NET, for hyperspectral image classification, focusing on efficient channel attention network fusion. Bhatti et al. [13] introduced the local similarity projection Gabor filter (LSPGF) algorithm for hyperspectral image classification, combining dimensionality reduction via local similarity projection and 2D Gabor filtering with CNN-based feature extraction. Several researchers have employed 3D CNNs for hyperspectral image classification tasks. Yue et al. [14] proposed an HSI classification method based on adaptive spatial pyramid constraint (ASPC), which leverages the global spatial neighborhood information of labeled samples to enhance the model’s generalization ability in scenarios with limited training data. Zhu et al. [15] integrated global convolutional long short-term memory and a global joint attention mechanism to address the challenge of insufficient and imbalanced sample data, introducing the SSDGL framework for HSI classification. Despite their success, these methods still grapple with the challenge that CNNs are local and translation-invariant, potentially failing to capture long-range dependencies in hyperspectral data. The self-attention mechanism, a key feature of the transformer architecture, allows the model to dynamically weigh and integrate information from the entire input spectrum [16]. The transformer model’s self-attention mechanism enables it to capture long-range dependencies in the data, and transformers are permutation-invariant, making them more suitable for processing unordered feature sets [17]. Sun et al. [18] proposed a spectral–spatial feature tag transformer (SSFTT) method for HSI classification, which captures spectral–spatial features and high-level semantic features, outperforming several state-of-the-art methods. Yang et al. [19] proposed a hyperspectral image transformer (HiT) classification network that embeds convolution operations into the transformer structure to capture subtle spectral differences and convey local spatial context information.
Despite substantial advancements in HSI classification using deep learning, these methods continue to face challenges. End-to-end supervised learning often necessitates a large number of labeled samples to optimize deep models. To mitigate these issues, strategies such as lightweight modeling [20], active learning [21], and self-supervised pretraining modeling (SSL) have been introduced. These methods aim to enhance generalization capabilities and utilize unlabeled raw data, which are more readily available than labeled samples. In this paper, we introduce a spectral transformer-based self-supervised learning algorithm (S3L) for HSI classification when labeled data are scarce. The goal is to optimize the performance of hyperspectral image classification. This method comprises two stages: pretraining and fine-tuning. In the pretraining stage, a mask mechanism is employed to learn the spatial representation of HSI, and the spectral features are modeled through the spectral transformer module. In the fine-tuning stage, labeled data are used to optimize pretraining weights and enhance classification accuracy. Experimental results on multiple public datasets demonstrate that the proposed method achieves state-of-the-art performance.
The main contributions are summarized as follows:
1. We propose a unique spectral transformer structure specifically designed to capture and model the complex relationships between different spectral bands in hyperspectral images. This structure enhances the model’s sensitivity to spectral information, allowing it to more accurately capture subtle changes in HSI data.
2. We adopt a pretraining and fine-tuning strategy, using a mask mechanism and a spectral transformer to learn the spatial representation of HSI from unlabeled data. This approach enhances the sequence dependence of spectral features and learns robust spatial representation of the hyperspectral image by introducing a mask mechanism.
3. Experimental results on multiple public datasets demonstrate that the proposed S3L outperforms other methods, achieving state-of-the-art performance.
In the subsequent sections, we delve into the specifics of the proposed S3L. Section 2 reviews previous work. Section 3 provides a detailed introduction to the S3L. In Section 4, we carry out comparative and ablation experiments and analyze the results to further validate the effectiveness of the proposed S3L. Finally, Section 5 offers a comprehensive summary of the entire text.

2. Related Work

2.1. Hyperspectral Image Classification

HSI classification serves as a foundational pillar for applications of HSI, finding extensive utility across various domains including agriculture, forestry, urban planning, military defense, and environmental surveillance. The realm of hyperspectral image classification algorithms bifurcates into two primary categories: conventional methodologies and those predicated on deep learning techniques.
(1) Traditional methods: HSI classification has predominantly concentrated on extracting spectral features. Tang et al. [22] introduced two manifold-based sparse representation algorithms aimed at addressing the instability issues of 1-based sparse representation. By integrating local structure and smoothness properties into the sparse representation, they significantly enhanced classification performance. Gu et al. [23] developed a novel framework, multistructural element nonlinear multikernel learning (MultiSE-NMKL), which merges spectral and spatial features. This approach generates extended morphological profiles (EMPs) using multistructural elements to encapsulate spatial–spectral information. Furthermore, it employs nonlinear multikernel learning (NMKL) to derive optimal combination kernels from predefined linear basis kernels, thereby boosting classification accuracy. Samaniego et al. [24] outlined a general strategy to identify the Euclidean metric in a low-dimensional space that minimizes the variance for a specific class label. This method primarily addresses challenges in remote sensing, pattern recognition, and statistics by improving classification accuracy through ensemble prediction, tackling the issue of nonlinear relationships in object classification. Ren et al. [25] proposed an innovative classification technique that enhances HSI classification by integrating random forests with label constraints. This method leverages spectral and spatial information through principal component analysis and morphological profiling. By incorporating spatial continuity label constraints into a random forest classifier, it achieves superior accuracy over traditional random forest and support vector machine (SVM) methods.
Banki et al. [26] introduced a wavelet-based kernel function for SVMs, termed the wavelet kernel, to advance HSI classification. This kernel demonstrates its effectiveness in comparison with traditional Gaussian and polynomial kernels in remote sensing applications. Hsu et al. [27] presented a classification method that utilizes a learning dictionary to amalgamate spectral and spatial information into a joint sparse representation, addressing the challenge of dimensionality reduction in HSI analysis. Melgani et al. [6] conducted a thorough evaluation of SVMs for hyperspectral remote sensing image classification. They compared the SVMs’ efficiency and effectiveness against traditional feature reduction techniques and other nonparametric classifiers. Additionally, they explored various multiclass strategies to overcome the challenges of applying binary SVMs to multiclass problems in hyperspectral data analysis. Zhang et al. [7] introduced a cascaded random forest (CRF) method that enhances the random forest algorithm by integrating a hierarchical random subspace method for feature selection with boosting. This combination, along with a novel out-of-bag error method for updating sample weights, aims to improve classification performance without risking overfitting. Heras et al. [28] advocated for the use of extreme learning machines (ELMs) in classifying hyperspectral images for land cover classification. They introduced two ELM-based techniques to integrate spectral and spatial information, which not only improved classification accuracy and efficiency but also reduced execution time compared to similar technologies. Li et al. [29] proposed the robust sparse representation-based classification (RSRC) method and its extension, the joint RSRC (JRSRC), to overcome the limitations of traditional SRC methods. By effectively handling outliers in HSI classification, these methods demonstrated improved performance over existing techniques, including orthogonal matching pursuit and other popular classifiers.
(2) Deep Learning: Deep learning-based algorithms for hyperspectral image classification can be broadly categorized into three main approaches: convolutional neural network (CNN)-based methods (encompassing both 2D and 3D architectures), attention mechanism-based methods, and those leveraging transformer architectures. Lu et al. [30] proposed an evolutionary block-based CNN (EB-CNN), which employs a genetic algorithm to automatically determine the optimal CNN architecture for hyperspectral image classification. Zhou et al. [31] introduced multiscale convolutional gradient features (MCGFs), refined through pseudo-Siamese networks, to capture finer common features and overcome the limitations of traditional handcrafted features and deep learning methods requiring extensive training. Cai et al. [32] introduced the transformer network (TRN) into the field of automatic modulation classification, highlighting its ability to fuse global information of sample sequences to improve classification accuracy, especially in low signal-to-noise ratios. TRN outperforms other deep models and traditional methods in terms of classification accuracy. Hong et al. [33] developed a general multimodal deep learning (MDL) framework that combines data fusion and network architecture strategies to extend from pixel-level classification to spatial information modeling using CNNs. Wu et al. [34] proposed a deep learning-based framework, CCR-Net, which utilizes convolutional neural networks and a novel cross-channel reconstruction module to classify multimodal remote sensing data. Yu et al. [35] proposed an image-based global learning framework of a dual-channel convolutional network (DCCN) that optimizes the utilization of global and multiscale information for HSI classification. Qing et al. [36] introduced a 3D self-attention multiscale feature fusion network (3DSA-MFN) for HSI classification, incorporating 3D multihead self-attention to capture interactive features over long distances and effectively fuse spatial and spectral features. Zhong et al. [37] introduced a spectral space transform network (SSTN), with spatial attention and spectral correlation modules, and a factorized architecture search (FAS) framework for hyperspectral image classification. Peng et al. [38] used a dual-branch structure to capture fine-grained spatial information and extract spectral features, proposing a CASST method for hyperspectral image classification, which achieved better accuracy than the existing transformer classification model. Qing et al. [39] utilized spectral attention and self-attention mechanisms to extract spectral–spatial features and proposed an end-to-end transformer model, SAT Net, for hyperspectral image classification.

2.2. Self-Supervised Learning

Deep learning methodologies predominantly rely on data-driven approaches, necessitating extensive efforts in data collection and annotation—a process that is both time-consuming and resource-intensive. Self-supervised learning [40] emerges as a potent solution, capable of extracting visual features from vast quantities of unlabeled images or videos, thereby significantly reducing the dependency on labeled datasets for deep learning applications. Within the realm of self-supervision theory, Teng et al. [41] introduced an innovative approach that leverages data from downstream tasks to refine unlabeled data, enhancing self-supervised learning. This method primarily addresses the challenges arising when the conditional independence (CI) condition is not met, which otherwise leads to a marked increase in sample complexity for downstream tasks. Bansal et al. [42] established a novel upper bound to gauge the generalization gap of classifiers, while Huang et al. [43] developed a technique to mathematically evaluate data augmentation through the introduction of a specific metric. This metric provides an upper limit on the error rate for downstream classification tasks, offering a theoretical framework to comprehend the impact of contrastive learning on the generalization capabilities within self-supervised learning.
In the domain of image representation learning, Pandey et al. [44] unveiled a cutting-edge semisupervised method for 2D medical image segmentation. This method employs contrastive learning (CL) on image patches to capture local feature representations, coupled with a novel consistency regularization (CR) strategy. This combination effectively addresses the issue of confirmation bias and fosters improved feature-space clustering. Park et al. [45] introduced an innovative model, RUC, designed to tackle the challenges of misprediction and overconfidence frequently encountered in unsupervised image clustering techniques. RUC’s innovation lies in its use of pseudolabels from existing image clustering models as a noisy dataset, which may include misclassified samples. Through a retraining process, RUC aims to rectify misaligned knowledge, thus alleviating the overconfidence issue in predictions. In the realm of video representation learning, Jenni et al. [46] introduced an innovative self-supervised approach for learning video representations attuned to changes in motion dynamics. This method trains a neural network to differentiate between video sequences and their temporally transformed counterparts, eliminating the need for manual labeling. Han et al. [47] developed a vision-centric self-supervised technique for video representation learning that enhances information–noise contrastive estimation (InfoNCE) training. This enhancement is achieved through two key innovations: the inclusion of semantic class positive samples to boost performance and the introduction of a novel self-supervised cotraining strategy. This strategy leverages the complementary information from different visual perspectives (RGB flow and optical flow) to generate positive samples from one view based on another. In the domain of 3D feature learning, Xie et al. [48] put forward an unsupervised pretraining methodology for 3D point cloud comprehension, diverging from the traditional focus on advanced scene understanding tasks. This method employs unsupervised pretraining on a vast dataset of 3D scenes, utilizing a unified architecture, source datasets, and contrastive losses to achieve performance that outstrips recent state-of-the-art results in segmentation and detection tasks across six diverse indoor and outdoor, real and synthetic datasets. The notable enhancement in performance underscores the adaptability of the learned representations across different domains.

3. Methodology

In order to achieve efficient classification performance with limited HSI annotation data, we propose a self-supervised learning algorithm, S3L (spectral–spatial self-learning), which deeply processes spectral features. The design of S3L is inspired by current state-of-the-art techniques in self-supervised learning, particularly their effectiveness in processing high-dimensional feature spaces. The overall architecture of this algorithm is shown in Figure 1. At the core of S3L is its unique spectral transformer architecture, specifically designed to parse and learn from the complexity and high dimensionality of HSI data. The introduction of the spectral transformer structure aims to enhance the model’s sensitivity to spectral information and capture subtle changes in HSI data more accurately. The S3L algorithm is divided into two main stages: pretraining and fine-tuning. In the pretraining stage, we adopt an innovative approach that does not rely on labeled data but introduces a mask mechanism to learn a robust spatial representation of hyperspectral images. The goal of this stage is to enable the model to autonomously learn and understand the intrinsic structure and characteristics of HSI data without explicit label guidance. Concurrently, the introduction of the spectral transformer enables the model to conduct in-depth modeling of spectral features in HSI data, capturing richer and more detailed information. To further enhance the model’s sequential dependence on spectral features, we innovatively couple the gated recurrent unit (GRU) module with the transformer’s encoder module. This coupling not only improves the model’s ability to process time series data but also enhances its ability to understand and represent spectral features. In this way, S3L can handle spatiotemporal dynamic changes in HSI data more effectively, providing more accurate classification performance.
Throughout the pretraining process, we continually optimize and adjust the model’s parameters through a comprehensive optimization strategy, encompassing both classification and reconstruction functions. This strategy ensures that the model can glean as much effective information as possible during the pretraining phase, laying a solid foundation for the subsequent fine-tuning phase. Upon entering the fine-tuning phase, we introduce limited labeled data to further refine and optimize the model weights obtained during the pretraining phase. The goal of this stage is to enable the model to adapt more accurately to specific classification tasks and further enhance the accuracy of hyperspectral image classification. Through the robust feature representation obtained in the pretraining stage coupled with the fine adjustments in the fine-tuning stage, S3L can significantly improve the overall performance of classification tasks.
In the following section, we first delve into the details of the spectral transformer module in S3L, explaining its design concept and working mechanism. We then explore the pretraining and fine-tuning process of S3L and how these two stages interact with and complement each other to advance the hyperspectral image classification task.

3.1. Spectrum Transformer

Masked image modeling in self-supervised learning methods is a transformative technology, particularly crucial in HSI classification. Existing masking methods primarily focus on modeling spatial features, often overlooking the importance of spectral features. To address this, we propose a unique spectral transformer structure specifically designed to capture and model the complex relationships between different spectral bands in hyperspectral images, as depicted in Figure 2.
Specifically, for the input hyperspectral feature matrix, we first normalize it to eliminate scale differences between different bands. Next, we perform carefully designed partitioning and splicing operations on the spectral features. This not only preserves the integrity of the spatial features but also significantly enhances the model’s ability to capture spectral sequential feature dependencies when processing input shifts by introducing changes in spectral dimensions. We further linearly map the normalized feature vectors to recalibrate the spectral dimensions, a step that critically influences the model’s ability to understand and process spectral data.
By further dividing, we obtain two components, v and g a t e , which play a central role in the model. Additionally, the feature vector generates q u e r y and k e y through another linear mapping, two components crucial in the transformer architecture. We further divide them, resulting in different linear components (such as q 1 , q 2 , k 1 , and k 2 ), and through multiple position shifts, we enable the model to more deeply understand and model the complex interactions between spectral features. For k 2 , we adopt an innovative random masking strategy. By setting the value of the mask part to zero, the model can ignore certain noncritical parts and focus on more important information. We also introduce rotational position encoding to enhance the model’s understanding of sequential relationships in hyperspectral sequence data. Next, the input data are padded to fit the needs of the transformer model and facilitate parallel processing of spectral features. This step enables the model to efficiently capture different granularity levels in the spectral data. Simultaneously, the mask is also padded accordingly to ensure its consistency with the input data length. The core of the transformer model lies in its attention mechanism. We obtain the final similarity matrix by calculating the similarity between q u e r y and k e y and adjusting it using the relative position deviation. These attention weights are then applied to the feature variables V, producing the output matrix a t t 2 . For k 1 and q 1 , we adopt a similar approach to obtain the feature output a t t 1 . By weighting information at different locations, the model is able to obtain a more accurate and richer data representation. Finally, by weighted combination of a t t 1 and a t t 2 , we obtain the final feature output. This approach enables the model to effectively handle complex relationships between adjacent pixels and spectral features, further enhancing the performance of S3L. The entire implementation process of the spectral transformer demonstrates in detail how to effectively improve the accuracy of hyperspectral image classification through carefully designed structures and strategies. The detailed execution process is shown in Algorithm 1.
Algorithm 1 Spectrum transformer
1:
Input: Hyperspectral feature matrix x
2:
Output: Processed feature matrix
3:
Normalize x to obtain n o r m e d _ x
4:
if s e l f . s h i f t _ t o k e n s is true then
5:
   Split and concatenate n o r m e d _ x to introduce spectral dimension variations
6:
end if
7:
Apply linear mapping to n o r m e d _ x , split into v and g a t e
8:
Generate query and key from n o r m e d _ x , split into q 1 , q 2 , k 1 , and k 2
9:
if mask exists then
10:
   Apply masking to k 2 , setting masked values to 0
11:
   Apply rotary position encoding to q 1 , q 2 , k 1 , k 2 to enhance sequence understanding
12:
end if
13:
Pad input data for suitable reorganization and chunking
14:
Apply the same padding to the mask to match the input data length
15:
Compute similarity s i m between queries and keys with relative position bias
16:
Apply attention mechanism to v to produce output matrices a t t 1 and a t t 2
17:
Weight and combine a t t 1 and a t t 2 to obtain the final feature output
18:
Enhance performance by processing relationships between adjacent pixels and spectral features
19:
return Final feature output

3.2. Self-Supervised Pretraining

During the pretraining phase, our goal is to extract robust spatial and spectral features from hyperspectral images using a self-supervised learning approach, which does not require labeled data. We achieve this by implementing a masking mechanism that randomly occludes parts of the input hyperspectral images, creating an information bottleneck. This compels the model to predict the occluded region’s content based solely on the surrounding unoccluded context, thereby enhancing the model’s understanding of spatial and spectral dependencies in the data. The entire procedure is delineated in Algorithm 2.
Drawing inspiration from the work of Sun et al. [49], a variety of dimensionality reduction techniques were explored, including linear discriminant analysis (LDA), independent component analysis (ICA), local linear embedding (LLE), and principal component analysis (PCA). LDA is designed to enhance the separation between classes, yet it may falter with hyperspectral data where class distinctions are subtle. ICA aims to identify independent components, a challenge in hyperspectral imagery due to the often-correlated spectral signatures of materials. LLE prioritizes preserving the local structure of data, potentially overlooking broader trends. Conversely, PCA focuses on capturing components with significant variance, effectively mitigating noise by discarding low-variance elements. This attribute of PCA not only addresses the noise issue prevalent in hyperspectral imagery but also adeptly manages the sparsity associated with high-dimensional data spaces. To effectively handle noise and redundant information in hyperspectral images, we employ the principal component analysis (PCA) algorithm for data preprocessing. PCA linearly transforms the original hyperspectral data I R H × W × C into a new coordinate system, the basis of which is the data’s principal component direction. The PCA-processed data are denoted as I p R H × W × L , where L signifies the number of spectral channels postdimensionality reduction. We then use a sliding window of size s × s to divide I p into N small blocks of size s × s × L . Each small block is represented as P R s × s × L , where N = ( H s + 1 ) ( W s + 1 ) .
Algorithm 2 Pretraining process for hyperspectral image classification
1:
Initialize model parameters θ .
2:
for each training epoch e p o c h = 1 , 2 , , E  do
3:
   for each batch ( X , Y ) in the training set do
4:
     Generate a masked version of the input X m a s k e d .
5:
     Transfer X m a s k e d to the computation device (e.g., GPU).
6:
     Compute the reconstruction loss L r e c .
7:
     if labelled data is available then
8:
        Compute the classification loss L c l s .
9:
     end if
10:
     Calculate the total loss L = α L c l s + β L r e c .
11:
     Update the model parameters θ to minimize L.
12:
   end for
13:
   Update the learning rate.
14:
end for
15:
Save the model state.
For these small patches P, we apply random mask operations and add positional encoding. The processed data are then passed through the encoder and decoder, which are composed of spectral transformers in sequence. In the encoder’s penultimate layer, we incorporate a GRU network structure to capture the sequential dependence of spectral features while maintaining the capture of global dependencies. This is necessary as our proposed spectral transformer performs multiple splits and calibrations when handling spectral dimensions, which may overlook the spectral continuity between adjacent pixels.
To optimize the entire pretraining process, we use two loss functions: classification loss (Cls loss) and reconstruction loss (Rec loss). The classification loss aims to preserve the model’s discriminative ability during pretraining, even when faced with incomplete spectral information. This is achieved by making accurate class predictions on unoccluded parts of the image, thereby enhancing the model’s ability to infer global information from partial data. The classification loss is expressed as
Cls = i = 1 N y i log ( p i )
where y i is the true class label and p i is the class probability predicted by the model.
The reconstruction loss is designed to aid the model in learning the complex structure of hyperspectral images, deeply exploring spatial–spectral features by reconstructing the mask area, and identifying subtle spectral differences of different ground objects. This enables comprehensive characterization of hyperspectral images, capturing more spatial and spectral details. The reconstruction loss is expressed as mean square error (MSE):
Rec = 1 N i = 1 N ( I i I ^ i ) 2
where I i is the pixel value of the original image and I ^ i is the pixel value of the reconstructed image.

3.3. Fine-Tuning

Following the initial deep learning and feature extraction in the pretraining phase, the model undergoes further refinement to enhance its performance in specific classification tasks. Fine-tuning primarily involves adjusting the pretrained weights using labeled data to align the learned features more accurately with the dataset’s specific categories. The network structure during the fine-tuning phase largely mirrors the pretrained model. The key distinction is the exclusive use of the encoder, omitting the decoder. The encoder employs the spectrum transformer to amalgamate and process spectral data, ensuring that the spatial and spectral features align with the labeled data. This method preserves the original spatial and spectral representation during refinement, enhancing its consistency with the labeled data.
A classification head is incorporated during the fine-tuning process, positioned at the model’s end. This component transforms the rich spectral–spatial features into category probability distributions. It is optimized to map learned representations to specific categories effectively, thereby minimizing classification errors. The classification head also strikes a balance between the generalized representation acquired during pretraining and the detailed information necessary for precise classification, ensuring the model’s final output accurately predicts hyperspectral image class labels. It is important to note that a lower learning rate is maintained throughout the fine-tuning process. This strategy preserves the general features acquired during pretraining while facilitating necessary model adjustments for new classification tasks. Consequently, the fine-tuning process can optimize the model and make necessary adjustments while retaining its robust feature extraction capabilities, thereby enhancing the model’s performance in specific tasks. In summary, the fine-tuning phase is a refinement and enhancement of the pretraining phase. It fine-tunes the learned features and adjusts the network structure to better suit specific classification tasks. This process not only boosts the model’s accuracy but also ensures its comprehensive understanding and effective processing of hyperspectral images. Through fine-tuning, our model exhibits superior performance across a range of complex classification scenarios.

4. Datasets and Experimental Setting

This section commences with an introduction to three publicly accessible datasets, followed by a quantitative and qualitative comparative analysis with other sophisticated methods. Subsequently, we perform a series of ablation experiments to evaluate and highlight the efficacy of each component within the S3L framework.

4.1. Datasets

Our research utilizes four renowned hyperspectral image datasets, Indian Pines (IN), University of Pavia (UP), Salinas (SA), and Houston 2013, to assess the performance of our proposed method. These datasets, widely recognized in the hyperspectral image processing domain, offer hyperspectral images with diverse characteristics, facilitating a thorough evaluation of our method’s effectiveness and generalizability. Table 1 summarizes the basic information of the four datasets. Table 2 and Table 3 illustrate the training and test samples.
Indian Pines: The Indian Pines dataset, captured by the Airborne Visible/Infrared Imaging Spectroradiometer (AVIRIS) in 1992, represents a 145 × 145 pixel region in northwest Indiana’s Indian Pines area. It comprises 21,025 samples, initially with 224 spectral bands in the 0.4 to 2.5 µm range. However, due to water absorption and noise, 24 bands were discarded, leaving 200 usable bands. The spatial resolution is 20 m. The dataset, known for its diverse class representation, includes 16 different classes with a total of 10,249 labeled samples, primarily consisting of agricultural land and natural perennial vegetation.
Salinas: The Salinas dataset, collected by the AVIRIS sensor in California’s Salinas Valley, is characterized by a high spatial resolution of 3.7 m per pixel. The original dataset includes 224 bands with a spectral range from 400 to 2500 nm. After excluding 20 bands due to water absorption, 204 bands were retained for analysis. The dataset, covering an area of 512 × 217 pixels or a total of 111,104 samples, includes 16 different categories with 54,129 labeled samples, featuring diverse landscapes such as vegetable plots, bare soil, and vineyards.
University of Pavia: The University of Pavia dataset, also known as PaviaU, was collected by the ROSIS-03 sensor in the urban area of Pavia, Northern Italy. It boasts a high geometric resolution of 1.3 m, consisting of 610 × 340 pixels, with a total of 207,400 samples. Initially, 115 frequency bands were included, 12 of which were discarded due to noise, leaving 103 frequency bands for our experiments. The dataset features urban landscapes with nine categories and 42,776 labeled samples, including various urban surfaces such as asphalt, bricks, grass, and trees.
Houston 2013: The Houston 2013 dataset, captured using CASI-1500 sensors, encompasses the University of Houston campus and adjacent regions. It comprises 349 by 1905 pixels, featuring a comprehensive collection of 144 spectral bands. These bands span a spectral range from 380 nm to 1050 nm, with each pixel representing a spatial resolution of 2.5 m. The dataset is categorized into 15 distinct classes, encompassing various natural and human-made surfaces such as grass, trees, soil, and water, among others. It includes a total of 15,029 labeled pixels, facilitating detailed analysis and classification tasks.

4.2. Implementation Details

We employed the PyTorch 1.13.1 library to construct an experimental framework for hyperspectral image classification. The system used for this experiment was equipped with an NVIDIA GPU 4090 (Santa Clara, CA, USA), boasting 24 GB of memory, and operated within a Python 3.8 environment. In the experiment, we configured 20 training samples per category, set the window sizes to 27, and established a batch size of 512. We ran the model for 300 epochs with a learning rate of 1 × 10 3 . Regarding the mask settings, we maintained a mask ratio of 0.8 and an MLP ratio of 2.0, and we assigned 128 to the number of hidden channels. The dimensions of both the model’s encoder and decoder were set to 128. However, they differed in complexity: the encoder consisted of two layers, while the decoder was more intricate with six layers. Each layer contained eight‘ attention heads. We set the temperature to 1.0 and the hierarchical loss rate to 0.005. The model was trained using the Adam optimizer. To minimize the effects of random sampling, we conducted the experiment multiple times, each time with different initial training samples.

4.3. Comparative Analysis

To assess the efficacy of our proposed method, we juxtapose it with eight distinct methods: HybridSN [50], 3DAES [51], SSFTT [18], FDSSC [52], DCFSL [53], CLB [54], and DBDA [55]. HybridSN [50] is a spectral–spatial 3D-CNN supplemented with a spatial 2D-CNN. This method jointly extracts spatial–spectral features from a multitude of spectral bands and further learns a spatial representation at a more abstract level. 3DAES [51] is a semisupervised Siamese network that incorporates an autoencoder module and a Siamese network. This network explores information in large volumes of unlabeled data and rectifies it with a limited set of labeled samples. SSFTT [18] captures spectral–spatial features and high-level semantic features using a spectral–spatial feature extraction module and a transformer encoder module. FDSSC [52] is an end-to-end fast dense spectral–spatial convolution framework. It employs different convolution kernel sizes to extract spectral and spatial features, respectively, and utilizes a densely connected structure for deep learning of features. DCFSL [53] employs a conditional adversarial domain adaptation strategy to address the few-shot learning and domain adaptation problems within a unified framework. CLB [54] is an unsupervised framework that leverages contrastive learning methods and transformer models for hyperspectral image classification. DBDA [55] is a dual-branch dual-attention mechanism network that captures a vast number of spectral and spatial features contained in HSI. It uses channel attention blocks and spatial attention blocks to refine and optimize the extracted feature maps.

4.3.1. Quantitative Analysis

Table 4 displays the quantitative experimental results of various methods on the IN dataset. The proposed method consistently outperforms other techniques across various indicators. Specifically, S3L achieves an OA of 93.45%, which is 1.8% higher than the closest competitor, DBDA, and significantly surpasses traditional methods such as HybridSN. In the AA domain, the proposed method again leads with 92.2%, slightly higher than SSFTT’s 91.68%. This underscores the balanced performance of the proposed method across different categories. SSFTT can achieve higher classification performance with limited labeled samples. For instance, the OA of the SSFTT method is improved by 9.58% over the HybridSN method and 3.83% over the 3DAES method. The addition of methods like CLB and DBDA, which utilize contrastive learning and attention mechanisms, respectively, are beneficial for handling classification boundary issues. As shown in the per-class accuracy results, our proposed method can handle boundary information better, achieving 100% accuracy for Classes 6, 9, and 14.
Table 5 showcases the experimental results on the SA dataset. The OA, AA, and Kappa values of the proposed S3L are 94.87%, 94.31%, and 92.46, respectively, which surpass the corresponding values of other methods. 3DAES and SSFTT also achieved competitive results, with OA values of 92.13% and 93.01%, AA values of 92.47% and 94.26%,and Kappa values of 92.68 and 92.34, respectively. The proposed S3L obtains additional semantic information by processing spectral and spatial data, effectively capturing and utilizing spectral and spatial features in hyperspectral images.
Table 6 presents the quantitative experimental results of various methods on the PU dataset. The proposed method, S3L, achieved an overall accuracy (OA) of 94.64% and an average accuracy (AA) of 92.78%, demonstrating its consistent performance across all categories and its effective integration of spectral and spatial features. The Kappa coefficient is 93.54, the second highest, indicating a high agreement between the predicted and actual class labels. Among other methods, 3DAES and SSFTT performed commendably, with overall accuracy rates of 91.56% and 90.65% respectively. FDSSC, known for its rapid and accurate feature learning, excelled in Class 1 and Class 6 but fell short in other categories. This discrepancy may be attributed to the method’s emphasis on speed, potentially compromising its accuracy in complex scenes. SSFTT and CLB also demonstrated notable performance, particularly in Class 2 and Class 3 for SSFTT, and Class 2 and Class 4 for CLB, indicating their effective feature extraction capabilities.
Table 7 presents a comparative analysis of the performance metrics for various methodologies applied to the Houston 2013 dataset. The method proposed in this study attains an overall accuracy (OA) of 86.50%, an average accuracy (AA) of 85.25%, and a Kappa coefficient of 85.85. Concurrently, the SSFTT and FDSSC methods demonstrate robust performance, with SSFTT excelling in Class 4 with a 90.98% accuracy rate, and FDSSC achieving the highest classification accuracy in Class 5 at 91.57%. Nonetheless, when juxtaposed with the outcomes from other datasets, a marginal decline in the performance of all evaluated methods on the Houston 2013 dataset is observed. This decrement can likely be attributed to the larger size and the more complex spatial and spectral characteristics of the Houston 2013 dataset compared to the others.

4.3.2. Qualitative Analysis

Figure 3 presents the visual results of the hyperspectral image classification task on the Indian Pines (IN) dataset. The classification results of different methods (HybridSN, 3DAES, SSFTT, FDSSC, DCFSL, CLB, and DBDA) and the ground truth are displayed in sections a–i. It is evident that the proposed method excels in preserving plot boundaries and minimizing classification errors, particularly in identifying complex surfaces such as Building–Grass–Trees–Drives and Stone–Steel–Towers. The classification results of the proposed method are more coherent and the color blocks are more compact, indicating higher classification accuracy and spatial continuity. For different types of crops, such as corn (Corn-notill, Corn-mintill, Corn) and soybeans (Soybean-notill, Soybean-mintill, Soybean-clean) with varying farming methods, the proposed method appears to distinguish more accurately between different farming statuses. This is reflected in the classification results in the figure, with clear color distinctions and fewer misclassifications. When analyzing large areas of single color such as wheat fields (Wheat) and fallow fields (Fallow, Fallow-rough plow, Fallow-smooth), the proposed method effectively reduces noise and misclassification areas, demonstrating smoother and more consistent classification results.
Figure 4 displays the visualization results of the hyperspectral image classification task on the Salinas Area (SA) dataset. The proposed method proves effective in identifying small plots and complex boundaries, such as vineyards (Grapes), lettuce (Lettuce-4wk to Lettuce-7wk), and Vineyard-untrained, accurately depicting the edges of plots. Compared to other methods, oversmoothing is reduced, thereby better preserving the spatial characteristics of the original features. Additionally, the results of the proposed method are more distinct in color distinction, reducing confusion and misclassification, especially between categories with similar colors, such as lettuce at different stages. Throughout the visualization outcomes, objects categorized as Grapes and Fallow-plow exhibit a higher propensity for misclassification.
Figure 5 illustrates the visualization results of the hyperspectral image classification task on the PU dataset. When classifying urban structures such as buildings, roads, and lanes, the proposed method produces more refined and coherent results, better preserving the structural features of the buildings, such as metal surfaces and brick distinctions. Furthermore, S3L effectively reduces noise, displaying a more uniform and consistent classification. Throughout the visualization outcomes, objects categorized as Meadows and Bare Soil exhibit a higher propensity for misclassification.
Figure 6 presents the comparative visualization outcomes for various methodologies applied to the Houston 2013 dataset, with inaccurately classified regions delineated by yellow boxes. The analysis reveals that HybridSN, DCFSL, and SSFTT exhibit a higher incidence of misclassifications, particularly with indistinct demarcations between Tennis Court and Running Track areas. Both CLB and the newly introduced S3L method encounter some classification inaccuracies within the Healthy Grass categories. Conversely, FDSSC and the novel S3L method demonstrate superior qualitative visualization outcomes. Notably, S3L stands out for its minimal misclassifications, distinct boundary delineations, and adeptness at accurately classifying small-area features as well as features possessing similar spectral signatures. Throughout the visualization outcomes, objects categorized as Stressed Grass and Water exhibit a higher propensity for misclassification.
Figure 7 presents the visualization curves of the proposed S3L method during training and validation on three hyperspectral datasets: IN, UP, and SA. For the IN dataset, the initial training loss of S3L is high, reflecting the uncertainty in the initial model. As the number of epochs increases, it eventually stabilizes around 0.2, indicating effective model convergence, with the validation accuracy stabilizing at approximately 91%. The training accuracy of S3L on the UP dataset approaches 100%, the validation accuracy nears 94%, and the final training loss is close to 0.1. The performance on the SA dataset mirrors that of the IN and UP datasets, with a similar pattern observed: the training loss significantly decreases and levels off, with the training set accuracy nearing 100% and the validation set accuracy approaching 93%.

4.4. Ablation Experiment

To validate the effectiveness of each proposed module, we designed several sets of ablation experiments. In Experiment 1, the spectral transformer module was replaced with the standard transformer module (Exp 1) to evaluate the performance of the standard transformer in processing hyperspectral data. In Experiment 2, the GRU module (Exp 2) was removed to analyze its contribution to the order dependence of captured spectral features. No pretraining was performed in Experiment 3 (Exp 3). The experimental results are displayed in Table 8.
When replacing the spectral transformer module with the classic transformer module, performance deteriorates on all datasets. OA on the UP dataset dropped to 91.2%, and Kappa dropped to 0.905; on the IN dataset, OA dropped to 90.3%, and Kappa dropped to 0.895; and on the SA dataset, OA dropped to 88.5%, and Kappa dropped to 0.880. This demonstrates the critical role of the specialized design of the spectral transformer in processing hyperspectral data for the overall performance of the model. Classic transformers, while effective in various tasks, may not be as proficient as spectral transformers in capturing complex spectral dependencies in hyperspectral data.
In the experiment of removing the GRU module, although the performance decreased, the impact was less than that of Experiment 1. The OA on the UP dataset is 92.7% and the Kappa is 0.920; the IN dataset has an OA of 91.8% and a Kappa of 0.910; and the SA dataset has an OA of 89.9% and a Kappa of 0.895. This result illustrates the important role of the GRU module in the model, especially in capturing the order dependence of spectral features. However, the model can still maintain good performance even without the GRU module, indicating that other components (such as the spectral transformer) also play a significant role in the model.
The experimental results without pretraining performed the worst on all indicators. On the UP dataset, OA dropped to 87.3%, and Kappa dropped to 0.865; on the IN dataset, OA dropped to 86.5%, and Kappa dropped to 0.850; and on the SA dataset, OA dropped to 84.7%, and Kappa dropped to 0.835. This significant performance drop underscores the importance of the pretraining phase. Pretraining allows the model to learn robust feature representations without labeled data, providing a solid foundation for the fine-tuning phase.

5. Conclusions

This paper introduces a self-supervised learning algorithm, S3L, based on the spectral transformer for HSI classification. The S3L algorithm operates in two stages: pretraining and fine-tuning. During the pretraining stage, a mask mechanism is employed to learn the spatial representation of HSI, and spectral features are modeled through the spectral transformer module. In the fine-tuning stage, labeled data are utilized to optimize pretraining weights and enhance classification accuracy. Additionally, a GRU layer is integrated into the algorithm to strengthen the sequence dependence of spectral features. Experimental results on multiple datasets demonstrate that the S3L algorithm performs commendably when labeled samples are limited and is competitive with current advanced methods. Future work will explore the application of S3L in a broader range of remote sensing data classification tasks and further optimize the algorithm structure.
While the proposed method has demonstrated remarkable performance, it is not without its limitations. Compared to convolutional neural networks (CNNs), the self-supervised learning (S3L) approach, which leverages the transformer architecture, encounters certain deployment challenges. Primarily, the intricate self-attention mechanism inherent to transformers demands substantial memory during inference, posing significant demands on hardware resources. Additionally, the scarcity of hardware platforms capable of accelerating transformer models represents another critical issue that needs addressing. Moving forward, our research will focus on developing a lightweight transformer architecture and exploring hardware acceleration capabilities for transformer-based models. Furthermore, the training, fine-tuning, and testing processes have been confined to a single dataset. Future efforts will aim to extend these processes across diverse datasets, enhancing the method’s adaptability and robustness.

Author Contributions

Methodology, H.G. and W.L.; Validation, H.G.; Writing—original draft, H.G.; Writing—review & editing, H.G. and W.L.; Funding acquisition, W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Innovative Research Group Project of the National Science Foundation of China under Grant No. 51821003, in part by the National Science Foundation of Shanxi Province under Grant No. 201701D121065, and in part by the Fundamental Research Program of Shanxi Province under Grant No. 20210302124329.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ghamisi, P.; Plaza, J.; Chen, Y.; Li, J.; Plaza, A.J. Advanced spectral classifiers for hyperspectral images: A review. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–32. [Google Scholar] [CrossRef]
  2. Ghamisi, P.; Maggiori, E.; Li, S.; Souza, R.; Tarablaka, Y.; Moser, G.; De Giorgi, A.; Fang, L.; Chen, Y.; Chi, M.; et al. New frontiers in spectral-spatial hyperspectral image classification: The latest advances based on mathematical morphology, Markov random fields, segmentation, sparse representation, and deep learning. IEEE Geosci. Remote Sens. Mag. 2018, 6, 10–43. [Google Scholar] [CrossRef]
  3. Datta, D.; Mallick, P.K.; Bhoi, A.K.; Ijaz, M.F.; Shafi, J.; Choi, J. Hyperspectral image classification: Potentials, challenges, and future directions. Comput. Intell. Neurosci. 2022, 2022, 3854635. [Google Scholar] [CrossRef] [PubMed]
  4. Wijata, A.M.; Foulon, M.F.; Bobichon, Y.; Vitulli, R.; Celesti, M.; Camarero, R.; Di Cosimo, G.; Gascon, F.; Longépé, N.; Nieke, J.; et al. Taking Artificial Intelligence Into Space Through Objective Selection of Hyperspectral Earth Observation Applications: To bring the “brain” close to the “eyes” of satellite missions. IEEE Geosci. Remote Sens. Mag. 2023, 11, 10–39. [Google Scholar] [CrossRef]
  5. Kuo, B.C.; Ho, H.H.; Li, C.H.; Hung, C.C.; Taur, J.S. A kernel-based feature selection method for SVM with RBF kernel for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 317–326. [Google Scholar] [CrossRef]
  6. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Cao, G.; Li, X.; Wang, B. Cascaded random forest for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1082–1094. [Google Scholar] [CrossRef]
  8. Yang, J.M.; Yu, P.T.; Kuo, B.C. A nonparametric feature extraction and its application to nearest neighbor classification for hyperspectral image data. IEEE Trans. Geosci. Remote Sens. 2009, 48, 1279–1293. [Google Scholar] [CrossRef]
  9. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J.A. PCA-based edge-preserving features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  10. Zhong, P.; Gong, Z.; Li, S.; Schönlieb, C.B. Learning to diversify deep belief networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3516–3530. [Google Scholar] [CrossRef]
  11. Wu, Y.; Hu, X.; Zhang, Y.; Gong, M.; Ma, W.; Miao, Q. SACF-Net: Skip-Attention Based Correspondence Filtering Network for Point Cloud Registration. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3585–3595. [Google Scholar] [CrossRef]
  12. Qing, Y.; Liu, W. Hyperspectral image classification based on multi-scale residual network with attention mechanism. Remote Sens. 2021, 13, 335. [Google Scholar] [CrossRef]
  13. Bhatti, U.A.; Yu, Z.; Chanussot, J.; Zeeshan, Z.; Yuan, L.; Luo, W.; Nawaz, S.A.; Bhatti, M.A.; Ain, Q.U.; Mehmood, A. Local similarity-based spatial–spectral fusion hyperspectral image classification with deep CNN and Gabor filtering. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5514215. [Google Scholar] [CrossRef]
  14. Yue, J.; Zhu, D.; Fang, L.; Ghamisi, P.; Wang, Y. Adaptive spatial pyramid constraint for hyperspectral image classification with limited training samples. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5512914. [Google Scholar] [CrossRef]
  15. Zhu, Q.; Deng, W.; Zheng, Z.; Zhong, Y.; Guan, Q.; Lin, W.; Zhang, L.; Li, D. A spectral-spatial-dependent global learning framework for insufficient and imbalanced hyperspectral image classification. IEEE Trans. Cybern. 2021, 52, 11709–11723. [Google Scholar] [CrossRef]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  17. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  18. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  19. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral image transformer classification networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528715. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Tao, Q.; Yin, Y. A Lightweight Man-Overboard Detection and Tracking Model Using Aerial Images for Maritime Search and Rescue. Remote Sens. 2024, 16, 165. [Google Scholar] [CrossRef]
  21. Wang, J.; Wang, X.; Guan, J.; Zhang, L.; Zhang, F.; Chang, T. STPF-Net: Short-Term Precipitation Forecast Based on a Recurrent Neural Network. Remote Sens. 2024, 16, 52. [Google Scholar] [CrossRef]
  22. Tang, Y.Y.; Yuan, H.; Li, L. Manifold-Based Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7606–7618. [Google Scholar] [CrossRef]
  23. Gu, Y.; Liu, T.; Jia, X.; Benediktsson, J.A.; Chanussot, J. Nonlinear Multiple Kernel Learning with Multiple-Structure-Element Extended Morphological Profiles for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3235–3247. [Google Scholar] [CrossRef]
  24. Samaniego, L.; Bardossy, A.; Schulz, K. Supervised Classification of Remotely Sensed Imagery Using a Modified k-NN Technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  25. Ren, Y.; Zhang, Y.; Wei, W.; Li, L. A spectral-spatial hyperspectral data classification approach using random forest with label constraints. In Proceedings of the 2014 IEEE Workshop on Electronics, Computer and Applications, Ottawa, ON, Canada, 8–9 May 2014; pp. 344–347. [Google Scholar] [CrossRef]
  26. Banki, M.H.; Shirazi, A.A.B. New kernel function for hyperspectral image classification. In Proceedings of the 2010 the 2nd International Conference on Computer and Automation Engineering (ICCAE), Singapore, 26–28 February 2010; Volume 1, pp. 780–783. [Google Scholar] [CrossRef]
  27. Hsu, P.H.; Cheng, Y.Y. Hyperspectral Image Classification via Joint Sparse Representation. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 2997–3000. [Google Scholar] [CrossRef]
  28. Dora, B.; Heras, F.A.; Quesada-Barriuso, P. Exploring ELM-based spatial–spectral classification of hyperspectral images. Int. J. Remote Sens. 2014, 35, 401–423. [Google Scholar] [CrossRef]
  29. Li, C.; Ma, Y.; Mei, X.; Liu, C.; Ma, J. Hyperspectral Image Classification With Robust Sparse Representation. IEEE Geosci. Remote Sens. Lett. 2016, 13, 641–645. [Google Scholar] [CrossRef]
  30. Lu, Z.; Liang, S.; Yang, Q.; Du, B. Evolving block-based convolutional neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5525921. [Google Scholar] [CrossRef]
  31. Zhou, L.; Ye, Y.; Tang, T.; Nan, K.; Qin, Y. Robust matching for SAR and optical images using multiscale convolutional gradient features. IEEE Geosci. Remote Sens. Lett. 2021, 19, 4017605. [Google Scholar] [CrossRef]
  32. Cai, J.; Gan, F.; Cao, X.; Liu, W. Signal modulation classification based on the transformer network. IEEE Trans. Cogn. Commun. Netw. 2022, 8, 1348–1357. [Google Scholar] [CrossRef]
  33. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More diverse means better: Multimodal deep learning meets remote-sensing imagery classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4340–4354. [Google Scholar] [CrossRef]
  34. Wu, X.; Hong, D.; Chanussot, J. Convolutional neural networks for multimodal remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5517010. [Google Scholar] [CrossRef]
  35. Yu, H.; Zhang, H.; Liu, Y.; Zheng, K.; Xu, Z.; Xiao, C. Dual-channel convolution network with image-based global learning framework for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 6005705. [Google Scholar] [CrossRef]
  36. Qing, Y.; Huang, Q.; Feng, L.; Qi, Y.; Liu, W. Multiscale Feature Fusion Network Incorporating 3D Self-Attention for Hyperspectral Image Classification. Remote Sens. 2022, 14, 742. [Google Scholar] [CrossRef]
  37. Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.S. Spectral–spatial transformer network for hyperspectral image classification: A factorized architecture search framework. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5514715. [Google Scholar] [CrossRef]
  38. Peng, Y.; Zhang, Y.; Tu, B.; Li, Q.; Li, W. Spatial–spectral transformer with cross-attention for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5537415. [Google Scholar] [CrossRef]
  39. Qing, Y.; Liu, W.; Feng, L.; Gao, W. Improved transformer net for hyperspectral image classification. Remote Sens. 2021, 13, 2216. [Google Scholar] [CrossRef]
  40. Wu, Y.; Liu, J.; Gong, M.; Gong, P.; Fan, X.; Qin, A.K.; Miao, Q.; Ma, W. Self-Supervised Intra-Modal and Cross-Modal Contrastive Learning for Point Cloud Understanding. IEEE Trans. Multimed. 2024, 26, 1626–1638. [Google Scholar] [CrossRef]
  41. Teng, J.; Huang, W.; He, H. Can pretext-based self-supervised learning be boosted by downstream data? A theoretical analysis. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Virtual, 28–30 March 2022; pp. 4198–4216. [Google Scholar]
  42. Bansal, Y.; Kaplun, G.; Barak, B. For self-supervised learning, rationality implies generalization, provably. arXiv 2020, arXiv:2010.08508. [Google Scholar]
  43. Huang, W.; Yi, M.; Zhao, X.; Jiang, Z. Towards the generalization of contrastive self-supervised learning. arXiv 2021, arXiv:2111.00743. [Google Scholar]
  44. Pandey, P.; Pai, A.; Bhatt, N.; Das, P.; Makharia, G.; Prathosh, A.P. Contrastive semi-supervised learning for 2D medical image segmentation. arXiv 2021, arXiv:2106.06801. [Google Scholar]
  45. Park, S.; Han, S.; Kim, S.; Kim, D.; Park, S.; Hong, S.; Cha, M. Improving Unsupervised Image Clustering With Robust Learning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12273–12282. [Google Scholar] [CrossRef]
  46. Jenni, S.; Meishvili, G.; Favaro, P. Video representation learning by recognizing temporal transformations. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 425–442. [Google Scholar]
  47. Han, T.; Xie, W.; Zisserman, A. Self-supervised co-training for video representation learning. Adv. Neural Inf. Process. Syst. 2020, 33, 5679–5690. [Google Scholar]
  48. Xie, S.; Gu, J.; Guo, D.; Qi, C.R.; Guibas, L.; Litany, O. Pointcontrast: Unsupervised pre-training for 3D point cloud understanding. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part III 16; Springer: Cham, Switzerland, 2020; pp. 574–591. [Google Scholar]
  49. Sun, W.; Du, Q. Hyperspectral Band Selection: A Review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 118–139. [Google Scholar] [CrossRef]
  50. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  51. Jia, S.; Jiang, S.; Lin, Z.; Xu, M.; Sun, W.; Huang, Q.; Zhu, J.; Jia, X. A semisupervised Siamese network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5516417. [Google Scholar] [CrossRef]
  52. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A fast dense spectral–spatial convolution network framework for hyperspectral images classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef]
  53. Li, Z.; Liu, M.; Chen, Y.; Xu, Y.; Li, W.; Du, Q. Deep Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5501618. [Google Scholar] [CrossRef]
  54. Hu, X.; Li, T.; Zhou, T.; Liu, Y.; Peng, Y. Contrastive learning based on transformer for hyperspectral image classification. Appl. Sci. 2021, 11, 8670. [Google Scholar] [CrossRef]
  55. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of Hyperspectral Image Based on Double-Branch Dual-Attention Mechanism Network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
Figure 1. Architecture of the S3L algorithm for enhanced hyperspectral image classification with limited annotations. The figure illustrates the two-phase self-supervised learning approach, incorporating a spectral transformer and GRU for robust feature extraction during pretraining, followed by fine-tuning with labeled data to refine classification accuracy.
Figure 1. Architecture of the S3L algorithm for enhanced hyperspectral image classification with limited annotations. The figure illustrates the two-phase self-supervised learning approach, incorporating a spectral transformer and GRU for robust feature extraction during pretraining, followed by fine-tuning with labeled data to refine classification accuracy.
Remotesensing 16 00970 g001
Figure 2. Spectral transformer structure for enhanced hyperspectral image classification: a focus on spectral feature modeling and sequential dependency capturing in self-supervised learning.
Figure 2. Spectral transformer structure for enhanced hyperspectral image classification: a focus on spectral feature modeling and sequential dependency capturing in self-supervised learning.
Remotesensing 16 00970 g002
Figure 3. Comparative analysis of hyperspectral image classification on the Indian Pines dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Figure 3. Comparative analysis of hyperspectral image classification on the Indian Pines dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Remotesensing 16 00970 g003
Figure 4. Enhanced boundary discrimination in hyperspectral image classification on the Salinas Area dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Figure 4. Enhanced boundary discrimination in hyperspectral image classification on the Salinas Area dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Remotesensing 16 00970 g004
Figure 5. Urban structure classification precision in hyperspectral imagery on the PU dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Figure 5. Urban structure classification precision in hyperspectral imagery on the PU dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Remotesensing 16 00970 g005
Figure 6. Urban structure classification precision in hyperspectral imagery on the Houston 2013 dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Figure 6. Urban structure classification precision in hyperspectral imagery on the Houston 2013 dataset: a study of (a) original image, (b) HybridSN, (c) DCFSL, (d) 3DAES, (e) SSFTT, (f) FDSSC, (g) CLB, (h) DBDA, (i) proposed method, and (j) ground truth.
Remotesensing 16 00970 g006
Figure 7. Visualization of S3L method performance on hyperspectral datasets: (a) IN, (b) UP and (c) SA: training and validation metrics analysis.
Figure 7. Visualization of S3L method performance on hyperspectral datasets: (a) IN, (b) UP and (c) SA: training and validation metrics analysis.
Remotesensing 16 00970 g007
Table 1. Summary of hyperspectral image datasets.
Table 1. Summary of hyperspectral image datasets.
INUPSAHouston
Region or countryIndiana, USAPavia, ItalyCalifornia, USATexas, USA
Sensor typeAVIRISROSISAVIRISITRES CASI
Number of bands200103204144
Wavelength ( μ m)0.4–2.50.43–0.860.4–2.50.38–1.05
Spatial resolution20 m1.3 m3.7 m2.5 m
Image size 145 × 145 610 × 340 512 × 217 349 × 1905
Number of labeled samples10,24942,77654,12915,429
Number of classes1691615
Table 2. The number of training and testing samples across the University of Pavia, Indian Pines, and Salinas datasets).
Table 2. The number of training and testing samples across the University of Pavia, Indian Pines, and Salinas datasets).
Indian PinesSalinasUniversity of Pavia
ClassLand Cover TypeTrainingTestingLand Cover TypeTrainingTestingLand Cover TypeTrainingTesting
1Alfalfa2046Weeds-1202009Asphalt206631
2Corn-notill201428Weeds-2203726Meadows2018,649
3Corn-mintill20830Fallow201976Gravel202099
4Corn20237Fallow-plow201394Trees203064
5Grass-past20483Fallow-sm202678Metal sheets201345
6Grass-trees20730Stubble203959Bare Soil205029
7Grass-pas-m2028Celery203579Bitumen201330
8Hay-windr20478Grapes2011,271Bricks203682
9Oats2020Soil206203Shadows20947
10Soybean-n20972Corn203278
11Soybean-m202455Lettuce-4wk201068
12Soybean-c20593Lettuce-5wk201927
13Wheat20205Lettuce-6wk20916
14Woods201265Lettuce-7wk201070
15Buildings20386Vineyard-unt207268
16Stone2093Vineyard-t201807
  Total  320  10,249  Total  320  54,129  Total  180  42,776
Table 3. The number of labeled training samples and testing samples of the Houston 2013 dataset.
Table 3. The number of labeled training samples and testing samples of the Houston 2013 dataset.
Land Cover TypeTrainingTesting
1Healthy Grass201251
2Stressed Grass201254
3Synthetic Grass20697
4Trees201244
5Soil201242
6Water20325
7Residential201268
8Commercial201244
9Road201252
10Highway201227
11Railway201235
12Parking Lot 1201233
13Parking Lot 220469
14Tennis Court20428
15Running Track20660
Total30015,029
Table 4. Comparative analysis of quantitative classification outcomes using various algorithms on the IN dataset. The best performing results are shown in bold.
Table 4. Comparative analysis of quantitative classification outcomes using various algorithms on the IN dataset. The best performing results are shown in bold.
Class No.HybridSNDCFSL3DAESSSFTTFDSSCCLBDBDAProposed
191.87 ± 3.295.26 ± 2.395.50 ± 0.9100.00 ± 0.091.38 ± 5.894.59 ± 2.788.95 ± 6.188.47 ± 5.5
258.38 ± 5.458.86 ± 7.688.79 ± 1.487.54 ± 8.391.32 ± 6.788.59 ± 7.293.34 ± 2.194.82 ± 0.7
382.77 ± 6.366.92 ± 9.170.78 ± 4.886.22 ± 7.489.60 ± 8.677.12 ± 3.596.97 ± 2.295.27 ± 0.8
477.38 ± 8.784.61 ± 5.377.82 ± 6.298.94 ± 0.681.78 ± 7.985.33 ± 4.433.22 ± 9.484.13 ± 8.1
594.41 ± 4.577.91 ± 8.885.31 ± 3.692.08 ± 5.198.20 ± 0.946.25 ± 8.995.15 ± 1.896.70 ± 3.3
670.61 ± 7.187.45 ± 4.154.11 ± 9.597.11 ± 2.096.22 ± 3.165.89 ± 9.391.86 ± 1.1100.00 ± 0.0
793.89 ± 2.697.83 ± 0.6100.00 ± 0.0100.00 ± 0.098.75 ± 0.8100.00 ± 0.098.68 ± 0.694.40 ± 5.6
880.55 ± 6.489.50 ± 3.761.06 ± 8.699.07 ± 0.697.58 ± 0.534.30 ± 8.272.00 ± 6.993.16 ± 4.3
998.27 ± 1.297.02 ± 0.796.49 ± 1.2100.00 ± 0.098.70 ± 1.389.82 ± 7.8100.00 ± 0.0100.00 ± 0.0
1091.11 ± 0.969.07 ± 9.685.88 ± 2.888.70 ± 7.077.37 ± 8.484.31 ± 5.990.28 ± 3.485.74 ± 6.5
1197.16 ± 2.565.86 ± 8.098.90 ± 1.184.86 ± 9.890.90 ± 7.588.18 ± 6.670.50 ± 7.788.69 ± 4.7
1276.39 ± 8.958.28 ± 9.487.11 ± 2.276.18 ± 8.685.33 ± 5.189.32 ± 3.997.04 ± 1.596.77 ± 0.8
1388.73 ± 7.296.93 ± 0.599.49 ± 0.598.90 ± 1.192.46 ± 4.889.60 ± 5.798.60 ± 1.492.53 ± 6.3
1496.72 ± 2.387.36 ± 9.597.84 ± 1.697.81 ± 1.296.64 ± 1.3100.00 ± 0.0100.00 ± 0.0100.00 ± 0.0
1574.90 ± 9.868.37 ± 8.799.20 ± 0.894.51 ± 2.984.92 ± 7.492.65 ± 4.686.73 ± 8.294.47 ± 3.1
1678.03 ± 7.396.83 ± 2.296.30 ± 3.7100.00 ± 0.082.19 ± 9.173.80 ± 8.489.76 ± 9.988.73 ± 5.4
OA (%)83.17 ± 2.181.99 ± 1.887.78 ± 2.991.14 ± 2.591.24 ± 2.288.43 ± 1.991.65 ± 1.793.45 ± 1.5
AA (%)82.13 ± 2.880.73 ± 3.489.20 ± 1.991.68 ± 2.585.56 ± 2.785.38 ± 2.187.09 ± 2.992.20 ± 1.4
Kappa × 10077.10 ± 2.579.36 ± 2.291.62 ± 1.889.24 ± 2.687.90 ± 2.388.67 ± 1.791.16 ± 1.292.78 ± 2.0
Table 5. Comparativeanalysis of quantitative classification outcomes using various algorithms on the SA dataset. The best performing results are shown in bold.
Table 5. Comparativeanalysis of quantitative classification outcomes using various algorithms on the SA dataset. The best performing results are shown in bold.
Class No.HybridSNDCFSL3DAESSSFTTFDSSCCLBDBDAProposed
186.16 ± 5.797.37 ± 0.696.99 ± 0.394.64 ± 2.694.25 ± 5.292.34 ± 7.893.42 ± 6.994.17 ± 5.8
288.67 ± 6.896.19 ± 0.299.10 ± 0.897.84 ± 1.795.55 ± 3.393.45 ± 7.194.26 ± 5.897.68 ± 2.2
391.49 ± 7.989.84 ± 8.092.07 ± 4.896.38 ± 3.294.69 ± 7.393.68 ± 8.894.56 ± 5.196.31 ± 3.6
490.72 ± 8.099.47 ± 0.491.60 ± 3.495.05 ± 8.793.59 ± 5.692.36 ± 6.292.60 ± 7.497.26 ± 2.7
588.45 ± 5.391.46 ± 7.290.67 ± 8.789.90 ± 7.692.12 ± 5.791.96 ± 5.490.55 ± 6.792.97 ± 7.0
693.24 ± 7.498.69 ± 0.499.20 ± 0.896.70 ± 3.388.43 ± 7.687.43 ± 8.388.45 ± 5.791.53 ± 8.4
789.79 ± 6.297.53 ± 1.199.87 ± 0.198.40 ± 1.695.39 ± 3.692.20 ± 7.589.37 ± 8.691.46 ± 8.5
893.56 ± 8.788.72 ± 9.383.40 ± 6.189.67 ± 7.290.15 ± 8.492.77 ± 7.391.25 ± 6.593.34 ± 6.6
992.91 ± 7.794.54 ± 0.498.58 ± 1.093.01 ± 6.392.20 ± 5.892.87 ± 9.694.56 ± 8.297.68 ± 2.3
1095.18 ± 6.689.60 ± 7.091.87 ± 8.092.13 ± 9.191.81 ± 8.889.66 ± 7.991.29 ± 6.496.18 ± 3.7
1192.01 ± 5.596.70 ± 3.393.55 ± 0.494.20 ± 5.894.83 ± 6.194.67 ± 7.490.64 ± 5.998.47 ± 1.5
1294.20 ± 6.797.62 ± 0.197.08 ± 0.989.00 ± 5.890.87 ± 6.486.28 ± 7.687.99 ± 8.288.49 ± 5.4
1395.55 ± 7.899.16 ± 0.798.34 ± 0.291.26 ± 6.583.87 ± 7.988.45 ± 8.488.38 ± 5.689.36 ± 7.1
1491.39 ± 8.696.06 ± 1.196.05 ± 2.789.02 ± 8.686.44 ± 5.385.10 ± 6.788.08 ± 7.592.72 ± 7.2
1589.95 ± 9.877.59 ± 8.089.31 ± 7.192.98 ± 6.291.70 ± 7.389.98 ± 5.490.47 ± 7.793.87 ± 6.1
1691.72 ± 5.691.56 ± 6.888.49 ± 9.086.84 ± 7.290.53 ± 8.490.97 ± 6.989.28 ± 5.592.55 ± 7.4
OA (%)92.50 ± 1.792.70 ± 1.592.13 ± 1.893.01 ± 1.692.83 ± 1.992.27 ± 3.392.17 ± 2.794.87 ± 3.1
AA (%)93.75 ± 3.492.87 ± 2.892.47 ± 2.594.26 ± 2.893.51 ± 3.493.87 ± 3.492.20 ± 3.894.31 ± 3.2
Kappa × 10092.61 ± 1.992.31 ± 3.392.68 ± 2.792.34 ± 2.592.03 ± 2.092.49 ± 2.692.10 ± 2.892.46 ± 2.4
Table 6. Comparative analysis of quantitative classification outcomes using various algorithms on the PU dataset. The best performing results are shown in bold.
Table 6. Comparative analysis of quantitative classification outcomes using various algorithms on the PU dataset. The best performing results are shown in bold.
Class No.HybridSNDCFSL3DAESSSFTTFDSSCCLBDBDAProposed
185.26 ± 5.486.68 ± 7.198.99 ± 6.388.58 ± 8.297.56 ± 1.890.42 ± 7.496.83 ± 3.998.76 ± 1.2
298.13 ± 0.291.57 ± 8.898.51 ± 1.796.79 ± 5.191.53 ± 6.696.41 ± 0.796.55 ± 3.295.46 ± 3.5
362.08 ± 6.174.35 ± 5.284.53 ± 8.496.60 ± 7.590.23 ± 6.877.28 ± 5.951.34 ± 8.187.75 ± 4.4
473.44 ± 7.392.59 ± 8.694.16 ± 5.487.37 ± 6.285.60 ± 7.897.09 ± 2.990.46 ± 5.691.37 ± 2.3
593.63 ± 8.798.58 ± 0.599.96 ± 0.399.62 ± 6.488.06 ± 7.696.14 ± 1.098.48 ± 1.792.13 ± 2.5
698.97 ± 6.578.59 ± 5.777.67 ± 8.294.94 ± 7.994.01 ± 3.471.26 ± 6.393.36 ± 7.497.10 ± 2.9
790.58 ± 7.284.73 ± 8.581.26 ± 6.692.63 ± 5.586.93 ± 7.784.30 ± 8.343.93 ± 6.096.47 ± 2.5
880.70 ± 8.977.26 ± 7.391.92 ± 6.787.66 ± 5.983.95 ± 8.194.32 ± 7.693.29 ± 6.486.25 ± 6.0
960.47 ± 5.697.46 ± 1.397.62 ± 0.194.04 ± 7.086.88 ± 6.995.65 ± 5.297.73 ± 4.089.27 ± 5.7
OA (%)85.97 ± 1.985.65 ± 1.591.56 ± 1.890.65 ± 1.789.13 ± 1.790.45 ± 1.492.55 ± 1.694.64 ± 1.5
AA (%)84.13 ± 2.884.75 ± 3.292.77 ± 2.693.44 ± 1.590.86 ± 2.090.42 ± 1.983.10 ± 1.892.78 ± 1.7
Kappa × 10088.39 ± 1.284.94 ± 1.294.06 ± 0.991.77 ± 1.593.13 ± 2.089.94 ± 1.391.30 ± 1.893.54 ± 1.6
Table 7. Estimated comparative analysis of quantitative classification outcomes using various algorithms on the Houston 2013 dataset. The best performing results are shown in bold.
Table 7. Estimated comparative analysis of quantitative classification outcomes using various algorithms on the Houston 2013 dataset. The best performing results are shown in bold.
Class No.HybridSNDCFSL3DAESSSFTTFDSSCCLBDBDAProposed
148.23 ± 7.288.12 ± 8.164.89 ± 5.691.78 ± 0.582.67 ± 6.437.54 ± 7.974.32 ± 8.581.23 ± 5.1
284.68 ± 6.863.89 ± 9.482.56 ± 7.378.42 ± 8.285.29 ± 5.585.77 ± 6.187.13 ± 9.187.59 ± 0.8
375.47 ± 8.767.83 ± 7.588.24 ± 6.980.57 ± 5.880.86 ± 9.872.74 ± 8.387.72 ± 7.188.19 ± 0.9
469.28 ± 9.574.52 ± 8.471.97 ± 7.290.98 ± 0.773.64 ± 6.576.08 ± 5.427.83 ± 8.977.05 ± 7.6
585.14 ± 7.469.79 ± 5.979.76 ± 8.687.07 ± 6.791.57 ± 0.687.58 ± 8.887.73 ± 9.090.09 ± 6.2
687.39 ± 8.087.84 ± 7.790.67 ± 9.490.92 ± 5.390.89 ± 7.891.16 ± 6.692.07 ± 5.793.94 ± 0.5
765.79 ± 6.392.49 ± 8.953.92 ± 7.093.02 ± 0.660.72 ± 5.593.02 ± 0.775.18 ± 6.987.53 ± 8.7
891.33 ± 9.282.18 ± 6.093.02 ± 0.892.36 ± 7.692.95 ± 5.891.76 ± 9.793.02 ± 0.993.02 ± 0.7
971.87 ± 8.192.27 ± 7.948.68 ± 6.793.02 ± 0.826.01 ± 9.083.02 ± 5.285.24 ± 7.386.12 ± 8.2
1086.72 ± 0.566.38 ± 7.480.72 ± 6.482.60 ± 8.072.21 ± 9.377.61 ± 5.983.84 ± 6.878.68 ± 7.2
1188.90 ± 8.671.71 ± 5.489.74 ± 7.877.68 ± 6.185.16 ± 9.081.80 ± 8.492.47 ± 0.881.65 ± 7.3
1269.02 ± 6.869.68 ± 9.179.12 ± 8.270.36 ± 7.480.90 ± 5.681.16 ± 9.989.03 ± 8.789.71 ± 0.6
1380.14 ± 7.692.31 ± 8.392.45 ± 0.591.82 ± 6.985.57 ± 5.783.20 ± 8.591.42 ± 7.985.49 ± 6.4
1490.70 ± 8.481.36 ± 7.191.28 ± 6.291.71 ± 9.891.53 ± 5.993.02 ± 0.693.02 ± 0.793.02 ± 0.8
1565.77 ± 9.676.63 ± 8.891.33 ± 0.689.87 ± 6.376.19 ± 7.586.48 ± 8.180.17 ± 5.488.33 ± 9.9
OA (%)74.32 ± 1.573.75 ± 4.880.21 ± 1.983.05 ± 1.782.68 ± 1.683.52 ± 1.885.24 ± 1.986.50 ± 1.2
AA (%)75.24 ± 1.477.92 ± 1.681.17 ± 2.584.40 ± 2.877.03 ± 1.779.39 ± 1.582.58 ± 1.985.25 ± 1.3
Kappa × 10066.89 ± 1.266.71 ± 1.582.42 ± 1.681.79 ± 1.481.33 ± 1.882.27 ± 1.786.41 ± 1.985.85 ± 1.6
Table 8. Comparison of ablation experiment results (UP, IN, SA datasets). The best performing results are shown in bold.
Table 8. Comparison of ablation experiment results (UP, IN, SA datasets). The best performing results are shown in bold.
ExperimentalUPINSA
OA (%)KappaOA (%)KappaOA (%)Kappa
Full94.640.935493.50.92591.80.910
Exp 191.20.90590.30.89588.50.880
Exp 292.70.92091.80.91089.90.895
Exp 387.30.86586.50.85084.70.835
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, H.; Liu, W. S3L: Spectrum Transformer for Self-Supervised Learning in Hyperspectral Image Classification. Remote Sens. 2024, 16, 970. https://doi.org/10.3390/rs16060970

AMA Style

Guo H, Liu W. S3L: Spectrum Transformer for Self-Supervised Learning in Hyperspectral Image Classification. Remote Sensing. 2024; 16(6):970. https://doi.org/10.3390/rs16060970

Chicago/Turabian Style

Guo, Hufeng, and Wenyi Liu. 2024. "S3L: Spectrum Transformer for Self-Supervised Learning in Hyperspectral Image Classification" Remote Sensing 16, no. 6: 970. https://doi.org/10.3390/rs16060970

APA Style

Guo, H., & Liu, W. (2024). S3L: Spectrum Transformer for Self-Supervised Learning in Hyperspectral Image Classification. Remote Sensing, 16(6), 970. https://doi.org/10.3390/rs16060970

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop