Next Article in Journal
Privacy-Preserving Fine-Grained Redaction with Policy Fuzzy Matching in Blockchain-Based Mobile Crowdsensing
Previous Article in Journal
A Fast Gradient Iterative Affine Motion Estimation Algorithm Based on Edge Detection for Versatile Video Coding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Few-Shot Hyperspectral Image Classification Based on Convolutional Residuals and SAM Siamese Networks

1
School of Information, Yunnan University, Kunming 650504, China
2
Kunming Enersun Technology Co., Ltd., Kunming 650504, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(16), 3415; https://doi.org/10.3390/electronics12163415
Submission received: 30 June 2023 / Revised: 6 August 2023 / Accepted: 9 August 2023 / Published: 11 August 2023
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)

Abstract

:
With the development of few-shot learning, significant progress has been achieved in hyperspectral image classification using related networks, leading to improved classification outcomes. However, practical few-shot hyperspectral image classification encounters challenges such as network overfitting and insufficient feature extraction during the model training process. To address these issues, we propose a model called CRSSNet (Convolutional Residuals and SAM Siamese Networks) for few-shot hyperspectral image classification. In this model, we deepen the network depth and employ the convolutional residual technique to enhance the feature extraction capabilities and alleviate the problem of network gradient degradation. Additionally, we introduce the Spatial Attention Mechanism (SAM) to effectively leverage spatial information features in hyperspectral images. Lastly, metric learning is employed by comparing the distance between two output feature vectors to determine the label category. Experimental results demonstrate that our method achieves superior classification performance compared to other methods.

1. Introduction

Hyperspectral images (HSI) are obtained and compiled using hyperspectral sensors or imaging spectrometers, which contain numerous continuous bands that offer abundant spectral and spatial information. As a result, hyperspectral images find valuable applications in mineral resources [1], agricultural production [2], environmental monitoring [3,4], and astronomy [5]. Previous studies have employed various methods for hyperspectral image classification, as depicted in Figure 1a, including Support Vector Machines (SVM) [6], Random Forests (RF) [7], sparse representation-based [8] and K Nearest Neighbors (KNN) [9]. However, these classification approaches are not well-suited for multi-classification problems and fail to effectively leverage the high-dimensional hyperspectral data. Consequently, they do not fully exploit the abundant spectral and spatial information available in hyperspectral images, leading to unsatisfactory classification results.
With the growing significance of deep learning in computer vision, deep-learning models exhibit greater network depth and enhanced data-mining capabilities compared to traditional network structures [10]. In the realm of deep learning, various structural models have been applied to hyperspectral image classification, as illustrated in Figure 1b, including Convolutional Neural Networks (CNN) [11], Stacked Autoencoder Networks (SAN) [12], Deep-Belief Networks (DBN) [13], and Recurrent Neural Networks (RNN) [14]. Among them, the Convolutional Neural Network (CNN) has emerged as the primary method for hyperspectral image classification. Yang et al. [15] proposed a multilevel spectral–spatial transformation network (MSTNet) for hyperspectral image classification (HSIC). The network utilizes transformer encoders to learn feature representations and decoders to integrate multi-level features, resulting in accurate classification results. VGGNet [16], introduced by Simonyan K et al., utilizes small convolutional kernels (3 × 3) and small pooling kernels (3 × 3), allowing for deeper network models while managing computational growth. ResNet [17], also devised by He K M et al., enhances overall network performance through residual learning using convolutional layers and increased network depth. These advancements aim to improve the effectiveness of hyperspectral image classification. Zhu et al. [18]. introduced a short-range and long-range graph convolution (SLGConv) based on the graph convolutional neural network. They utilized a three-layer SLGConv to construct the short- and long-range graph convolution network (SLGCN) for extracting both global and local spatial–spectral information to improve hyperspectral image classification. However, Liao et al. [19] proposed a spectral–spatial fusion transformer network (S2FTNet) for HSI classification. S2FTNet leverages the transformer framework to create a spatial transformer module (SpaFormer) and a spectrum converter module (SpeFormer) to capture long-distance dependencies between image space and spectrum, enhancing the classification performance.
Although deep learning demonstrates excellent performance in hyperspectral image classification, its effectiveness heavily relies on a substantial number of labeled samples. Unfortunately, the process of labeling hyperspectral image samples is labor-intensive and costly, resulting in the limited availability of labeled data [20].
The scarcity of labeled samples poses challenges such as overfitting the deep-learning model and reduced accuracy. Consequently, the primary focus of future hyperspectral image classification lies in achieving satisfactory results with limited samples. Finding ways to optimize deep-learning models under such circumstances is a key research direction in the field.
In recent years, few-shot learning has gained popularity and is commonly applied to few-shot target classification tasks. Two prominent approaches in few-shot learning are metric learning and meta learning, as depicted in Figure 1c. Metric-learning methods focus on learning feature embeddings or similarity measures to enable classification based on discriminative similarities between query set labels and support set-labeled examples. These methods often employ episodic training mechanisms to acquire transferable knowledge. Meanwhile, the meta-learning method aims to learn an optimal initialization or optimizer that can generalize the model to unseen tasks. For instance, recurrent neural networks are employed to train cross-task meta-learners, enhancing knowledge transfers across different tasks and improving generalization performance. Both approaches share the common objective of learning a model capable of classifying query set labels with limited support set labels.
This paper combines various common classification methods for hyperspectral imagery with approaches used to address the issue of limited samples in traditional image processing. The main structure of this article is outlined as follows:
  • Related work: This section primarily focuses on presenting the relevant theories of metric learning and identifying the deficiencies in existing partial networks, subsequently leading to the introduction of the main innovations of our proposed network.
  • Materials and methods: We provide a concise overview of the designed network architecture, followed by a more detailed exposition of the concepts related to the spectral module, spatial module, and loss function.
  • Experimental results: This section encompasses the introduction of three commonly used hyperspectral image datasets, the description of the relevant experimental procedures, and a comprehensive analysis and discussion of the experimental results.
  • Conclusions: This section summarizes the main findings of the study, reviews the key points of the paper, and identifies the current limitations of the research, followed by proposing suggestions for future work.

2. Related Work

2.1. Few-Shot Learning Based on Metric Learning

Metric learning, also known as similarity learning, is a type of the few-shot learning approach based on migration learning [21]. It involves calculating the distance between the sample to be classified and known samples, analyzing the distances to measure their similarity, and finding the nearest matching sample for classification [22].
The Siamese network is a neural network model based on metric learning. In this model, pairs of samples are fed into the dual-channel Siamese network for feature extraction. For the input sample pair belonging to the same class, it is labeled as a positive sample (Label 1), while input sample pairs from different classes are labeled as negative samples (Label 0). The ratio of positive and negative samples has an impact on the network’s feature extraction classification accuracy. To ensure a balanced effect, this paper maintains a 1:1 ratio between positive and negative samples. The input data is then transformed into a target space using the Siamese network. Subsequently, a similarity calculation is performed using the distance function. This involves computing the distance between the output feature vectors of the two samples being tested and the feature vectors of the known samples in the feature vector space. The known labeled sample with the smallest distance to the test sample is selected as its classification class, thus completing the classification process.
The utilization of the dual-channel Siamese network, which employs metric learning in the few-shot learning method, leads to enhanced classification performance on samples. In contrast to the traditional single-channel network as depicted in Figure 2, which solely focuses on simple inter-class sample classification, it lacks the ability to effectively discern relationships and distinguish between different samples. However, the dual-channel Siamese network addresses this limitation by establishing stronger connections within the same categories while expanding the separation between different categories. This is accomplished by leveraging input sample pairs that facilitate differentiation of various categories, consequently leading to improved classification outcomes.

2.2. Related Network Models

There are three main types of few-shot networks currently applied to hyperspectral image classification [23]: prototypical networks, relation networks, and Siamese networks.
Tang et al. [24]. proposed a novel multiscale spatial spectroscopic PN (MSSPN). The network is a multi-scale spatial spectral feature extraction algorithm based on a trapezoidal structure, effectively achieving fusion of spatial spectral features at different scales. It also designs a multi-scale spatial spectral prototype representation based on the trapezoidal structure extraction algorithm theory, which offers improved scalability and better classification effectiveness. Zhang [25] proposed a “Global Prototypical Network” for hyperspectral image classification using a prototype network. Although the current prototype network shows effectiveness in classifying hyperspectral images, its generalization ability to new samples is limited. When encountering unseen samples, the classification performance trend decreases.
Sung F et al. proposed the relation network [26], in which the network initially processes randomly selected samples from the query set and sample set through the embedding layer to process and obtain the feature map. These feature maps are then concatenated and passed through the relation network to calculate the relation score, determining whether the samples belong to the same category. Gao et al. [27] employed a meta-learning strategy to design a feature-learning module and a relation-learning module, aiming to enhance the effectiveness of hyperspectral classification. However, the method based on the relational network relies on the selection and quantity of sample data. An insufficient sample size, or poor sample quality, can lead to the underutilization of hyperspectral image sample information in small sample scenarios, impacting feature extraction capabilities and resulting in decreased classification accuracy.
Wang et al. [28] propose a Siamese CNN with a soft-loss function that adapts based on coupling substitution data enhancement. This approach effectively addresses the HSI classification challenge with limited training samples. Huang et al. [29] utilized a bidirectional Siamese network to separately fuse spectral space features, thereby enhancing the network’s capability for feature extraction in hyperspectral images. However, the Siamese network is susceptible to overfitting or underfitting when dealing with small samples. In the context of hyperspectral image classification, the high dimensionality of the data and the limited number of samples can lead to good performance on the training set but poor performance on the test set.
To address the aforementioned challenges, this paper presents a few-shot hyperspectral image classification based on convolutional residuals and SAM Siamese networks. The key contributions are as follows:
  • We propose a novel dual-channel Siamese network architecture for spatial–spectral feature extraction in hyperspectral images. This model leverages convolutional neural networks of different dimensions, employing one-dimensional convolution for spectral feature extraction and two-dimensional convolution for spatial feature extraction.
  • We introduce a feature residual-extraction module specifically tailored for hyperspectral image classification. This module not only reduces the network’s 0complexity, but it also leverages multi-layer features to effectively extract spectral–spatial information.
  • In order to enhance the utilization of spatial information, we incorporate a Spatial Attention Mechanism (SAM) into the spatial feature-extraction module. Additionally, to mitigate overfitting caused by network deepening and improve the model’s generalization ability, we adopt a superior label-smoothing cross-entropy loss function [30].

3. Materials and Methods

The network architecture of this study is depicted in Figure 3. Initially, the hyperspectral image data samples are divided into a training set and test set. Each data sample is further segmented into data blocks of varying sizes, denoted as W1 × H1 for larger data blocks and W2 × H2 for smaller data blocks. During the training process, the data blocks of all samples are randomly selected and paired to create positive and negative sample pairs. These pairs are then input into the feature extraction network of the dual-channel Siamese network to extract deeper spectral and spatial information of hyperspectral data. The network model is continuously optimized through metric learning, which involves extracting feature vectors from hyperspectral image data. This enables the network to learn the classification ability for different classes. The parameters of the feature extraction module are adjusted via back-propagation of the loss function, facilitating the fine-tuning of the network.
During the testing process, the sample data in the test set is divided into a support set and a query set. The query set consists of the data to be detected, while the support set consists of the known class data. These two sets form a sample pair, where each pair is composed of a data block from the query set and a data block from the support set. The sample pairs are then input into the trained model. The trained model outputs feature vectors for each sample pair. The distance between the feature vector of the data in the query set and the feature vector of the data in the support set is calculated. The sample pair with the highest similarity in terms of feature vectors, specifically between the feature vectors of the support set and the query set, is selected. The predicted class for the data in the query set is determined based on this selected sample pair. This process completes the classification of hyperspectral images.

3.1. Model Method

3.1.1. Spectral Feature Residual Extraction Module

In the scenario of limited samples, the deep-learning network often encounters the issue of vanishing gradients as the number of network layers increases. To address this problem, the convolution residual method is incorporated into the spectral extraction module, which helps mitigate gradient degradation during the deepening of network. A lightweight implementation of the residual method is adopted. Each lightweight residual extractor, as depicted in Figure 4, comprises convolutional layers, Batch Normalization (BN) layers, and ReLU layers, applied in sequence.
As illustrated in Figure 4, the initial data input to the lightweight residual extractor is denoted as x . Each convolutional layer is subsequently followed by a Batch Norm layer [31], facilitating network convergence and improving generalization ability. This data processing results in an output following a normal distribution with a mean value of 0 and a variance of 1. Furthermore, a Rectified Linear Unit (ReLU) activation function is employed to reduce network computational complexity and address the issue of gradient vanishing in deep network.
Y ( x ) = G ( x ) + F ( x )
The spectral residual-extraction module is depicted in Figure 5. In the first lightweight residual extractor, the convolution has a kernel size of 3 × 1 × 1 and 16 channels. For the second lightweight residual extractor, the convolution layer has a kernel size of 3 × 1 × 1 and 32 channels. The third lightweight residual extractor has a convolution layer with a kernel size of 3 × 1 × 1 and 64 channels. Additionally, the output of the first ReLU activation function F x is added to the output of the third ReLU activation function, G x , to obtain the overall output Y x . The computation process is shown in Equation (1). The output Y x is passed to the next lightweight residual extractor. The feature map obtained after extraction by three lightweight residual extractors is denoted as Z R ( W × H × 1 × 64 ) .
Due to the limited number of samples in the few-shot scenario, it is important to control the number of parameters in the network model. Therefore, in the spectral feature-extraction part of the network, the special size of the three-dimensional convolution is set to 1, which effectively reduces the model parameters by 2/3.
The implementation of one-dimensional convolution in the network involves setting the spatial size of the three-dimensional convolution to 1, as shown in Equation (2). The value v i j x y z of the one-dimensional convolution represents the computation of the neuron at position ( x , y , z ) in the j -th feature map of the i -th layer and is provided by the following equation:
v i j x y z = f ( b i j + m p = 1 1 q = 1 1 r = 0 R i 1 w i j m p q r v ( i 1 ) m x y ( z + r ) )
where m represents the index of the feature map connected to the j -th feature map in the ( i 1 ) -th layer, R i is the size of the convolution kernel along with the spectrum dimension; p denotes the length of the spatial convolution kernel, q denotes the width of the spatial convolution kernel, and p and q are both set to 1; w i j m p q r is the value connected to the position ( p , q , r ) in the m -th feature map; b i j is the bias of the j -th feature map in the i -th layer; the function f ( ) is the ReLU activation function.

3.1.2. Spatial Feature Extraction Module

In scenarios of limited samples, effectively utilizing spatial information can significantly enhance hyperspectral image-classification capabilities. Based on the structural characteristics of HSI, this paper proposes a novel spatial feature extractor, depicted in Figure 6. The spatial feature extractor primarily consists of a spatial residual feature extractor and a spatial attention mechanism.

Spatial Residual Feature Extractor

The spatial feature-extraction component is illustrated in Figure 6. The spatial feature-extraction module effectively captures the spatial information of the feature map obtained from generating the spectral spatial-connection layer. This is achieved by employing a combination of a two-dimensional convolution residual block and a spatial attention mechanism. The resulting feature map generated by this module contains discriminative information related to different classes of hyperspectral images. In this process, the feature map Z R ( W × H × 1 × 64 ) is transformed into Z R ( W × H × K ) through the utilization of this module, where K represents the dimension of the feature map.
Using a similar approach to one-dimensional convolution, the spatial feature extraction component sets the spatial size of the three-dimensional convolution to 3 × 3, forming a two-dimensional convolution. The v i j x y of the two-dimensional convolution represents the calculated value of the neuron at the position ( x , y ) in the j th feature map of the i th layer, as shown in Equation (3):
v i j x y = f ( b i j + m p = 0 p i 1 q = 0 q i 1 w i j m p q v ( i 1 ) m ( x + p ) ( y + q ) )
where m represents the index of the feature map connected to the j -th feature map in the ( i 1 ) -th layer; p i represents the length of the spatial convolution kernel in the i -th layer, and q i represents the width of the spatial convolution kernel in the i -th layer; w i j m p q represents the value connected to the position ( p , q ) of the m -th feature map; b i j is the bias of the j -th feature map in the i -th layer; and the function f ( ) represents the ReLU activation function.

Spatial Attention Mechanisms

The feature map Z R ( W × H × K ) , obtained from the two-dimensional convolutional layer, serves as the input feature map of the spatial attention mechanism, as depicted in Figure 7 below. Initially, the input feature map is fed into the spatial attention mechanism, where maximum pooling and average pooling operations are performed to capture different information. This result in two feature maps of Z R ( W × H × 1 ) , each of size W × H × 1 , where every pixel in the generated image incorporates features from all channels at that position.
Then, the two generated feature maps Z , both of size Z R ( W × H × 1 ) , are then concatenated along the channel dimension. Subsequently, a 7 × 7 convolutional layer is applied to transform the feature map into a single-channel representation Z of size Z R ( W × H × 1 ) . The Sigmoid activation function is utilized to map the pixel values in the feature map to the probability space of 0 to 1, capturing the more prominent feature information in the image. This mapping generates spatial attention weight coefficients M s . Finally, the attention weights are multiplied element-wise, channel-by-channel, with the input feature map of the module, denoted as “ M s × input features”, resulting in a new feature Z of size Z R ( W × H × K ) . The attention mechanism is expressed by Equation (4).
M s ( F ) = σ f 7 × 7 A v g P o o l ( F ) , M a x P o o l ( F )
where M s represents the attention weight coefficient, F represents the input feature, σ represents the Sigmoid activation function, f 7 × 7 represents the 7 × 7 convolution kernel, and AvgPool and MaxPool represent average pooling and maximum pooling respectively.

3.2. Loss Function

In the training process of the dual-channel Siamese network, the loss function is calculated using a weighted-comparison loss function and a label-smoothing cross-entropy loss function for each training iteration. The total loss function is expressed as Equation (5):
L = L c t r + L s c e
where L c t r is the weighted contrast loss function, and L s c e is the label-smoothing cross-entropy loss function.
The dual-channel Siamese network outputs two feature maps of different sizes, Z 1 H 1 × H 1 × K and Z 2 H 2 × H 2 × K , and extracts their center vectors, L 1 1 × K and L 2 1 × K , respectively, where H 1 and H 2 represent the sizes of two feature maps. The cosine distance w between two input sample pairs can be calculated using Equation (6), and the weighted comparison loss function is defined by Equation (7):
w = L 1 × L 2 L 1 L 2
L c t r ( y , d p o s , d n e g ) = y × ( 1 w ) × d p o s + ( 1 y ) × w × max ( m d n e g , 0 )
Equation (7) represents the weighted comparison loss function, where ( 1 w ) is multiplied by the positive sample y pair, and w is multiplied by the negative sample pair; d p o s denotes the distance between the feature vectors of the same class samples in a sample pair, while d n e g represents the distance between the feature vectors of different class samples in a sample pair. The variable m represents the maximum distance between different samples.
The dual-channel Siamese network receives different information, resulting in a diversification of the output feature vector. To optimize the network parameters effectively, noise is introduced simultaneously to prevent excessive confidence in the correct label and enhance the network’s generalization ability. Hence, the label-smoothing cross-entropy loss function is introduced, defined by Equation (10) as follows:
p k = e x T w k l = 1 L e x T w l
y k = y k ( 1 α ) + α K
L s c e = k = 1 K y k log ( p k )
In Equation (8), p k represents the probability of each category of data, w k corresponds to the weights associated with the k -th category sample pairs, and x is the vector of activations from the second-to-last layer of the network. T denotes the transpose of x .
In Equations (9) and (10), y k represents the label vector, α is the label-smoothing factor, and K represents the total number of categories being classified.

4. Experimental Results

4.1. Experimental Datasets

To evaluate the efficacy of the proposed few-shot hyperspectral image classification method, three publicly available hyperspectral image datasets were chosen for experimental validation: the University of Pavia, Indian Pines, and Salinas [32]. The University of Pavia and Salinas datasets represent agricultural farmland hyperspectral image datasets, while the University of Pavia dataset pertains to an urban hyperspectral image dataset. Table 1 provides detailed information regarding these three public datasets, including the pseudo-color image and the ground-truth image. Please refer to Figure 8 for visual representations of the datasets.

4.2. Experiment Setup

The experimental hardware computer configuration consists of Intel® Core™i7-6700 K CPU @4.00 GHz RAM 32 GB and NVIDIA TITAN X (Pascal) 12 GB graphics card. The training and testing are conducted using the Pytorch neural network framework.
In the network, the Patch Size for hyperspectral input is set to 13 × 13 and 7 × 7 for two channels [23]. During training, five labeled samples are randomly selected for each class for model training, while the remaining samples are used for testing. The model is optimized by the Adam optimizer, with a learning rate of 0.001. This setting effectively trains the network, accelerates its convergence, and reduces the training time. To mitigate the impact of random results on the classification performance, all experimental data in this study are averaged over 10 experimental results obtained with randomly selected samples.
In this paper, the quantitative evaluation indicators for assessing the performance of hyperspectral image classification methods are as follow: overall accuracy (OA), average accuracy (AA), and the Kappa coefficient. The overall accuracy represents the ratio of correctly classified pixels to the total number of tested pixels in the hyperspectral data. Its calculation equation for overall accuracy is as follows:
O A i = C i i N
In the Equation, N represents the total sample size, and C represents the confusion matrix of size n × n .
The average classification accuracy represents the average accuracy of each category’s classification. Its calculation equation is as follows:
A A = i = 1 n O A i N
In the Equation, N represents the total number of samples, and n represents the number of hyperspectral image categories.
The Kappa coefficient is used as an indicator to measure the agreement between the classification results of the hyperspectral dataset and the actual effect. Typically, the Kappa coefficient ranges from 0 and 1. A Kappa value of 1 indicates complete agreement, Kappa ≥ 0.75 indicates satisfactory agreement, and Kappa < 0.4 indicates a less-than-ideal agreement. The calculation equation for the Kappa coefficient is as follows:
K a p p a = N i = 1 n C i i i = 1 n C i i × N i N 2 i = 1 n C i i × N i
In the Equation, N represents the total number of samples, n represents the number of hyperspectral image categories, N i represents the number of samples of the i -th class, and C represents the confusion matrix of size n × n .
To evaluate the impact of label factors in the label-smoothing loss function on OA, AA, and Kappa coefficients, this study conducts experiments on three datasets by varying the label factor from 0.1 to 0.9. The experimental results are depicted in Figure 9. It is observed that when the label factor is set to 0.2, the OA value is generally higher than other parameter settings. Therefore, for subsequent ablation experiments and comparative experiments, a label factor of 0.2 is chosen.

4.3. Ablation Experiments

To verify the improvement in overall accuracy of the residual feature extractor and the spatial attention mechanism in the proposed CRSSNet method, it is compared with the enhanced network CRSSNet (no SAM) that does not utilize the spatial attention mechanism, as well as the unimproved original network S3net. Two sets of ablation experiments are conducted on three datasets to validate the effectiveness of the residual feature extractor and spatial attention mechanism in the CRSSNet method. The flat average OA, AA, and Kappa data from 10 experiments of the two ablation models are presented in Table 2, Table 3 and Table 4, respectively.
From Table 2, Table 3 and Table 4, it can be observed that there is a significant performance improvement in OA when using the residual volume feature extractor in the network. The sample cases make full use of their hyperspectral data for training. Additionally, after incorporating the spatial-attention mechanism, the network’s OA also slightly increases, indicating that the spatial feature extraction part, using SAM after the spatial residual feature extractor extracts, further utilizes spatial information and enhances the network’s ability to utilize spatial information, resulting in higher accuracy. Furthermore, from the results of the Kappa coefficient, it can be observed that CRSSNet also exhibits a significant improvement in terms of consistency compared to CRSSNet (no SAM) and S3Net.
From the ablation experimental-feature classification diagram in Figure 10, it can be observed that the output features are transformed into two-dimensional space using PCA. The Siamese network reduces the intra-class distance and increases the inter-class distance, making the network highly separable. However, compared to S3Net, CRSSNet further improves the separability of the feature maps due to enhanced feature-extraction capabilities, leading to better classification accuracy.

4.4. Performance on Cross-Scene HSI Classification

To validate the effectiveness of the proposed method in this paper, classic and advanced hyperspectral image classification methods are selected, including 3DCNN [33], SSRN [34], DFSL + SVM [35], Gia-CFSL [36], DCFSL [37], and S3Net, among other methods. In the aforementioned method experiment, a random selection of five labeled samples per class is performed for training. Comparative analysis is conducted to verify the effectiveness of the proposed method.
Under the setting of a label factor of 0.2, the label-smoothing loss function, the experimental results of the proposed method in this paper and the above-mentioned hyperspectral image classification methods on the Indian Pines, Pavia University, and Salinas datasets are presented in Table 5, Table 6 and Table 7, respectively.
According to the results of Indian Pines, Pavia University, and Salinas datasets in Table 5, Table 6, and Table 7, respectively, this method achieves the highest overall classification accuracy of 78.03%, 88.00%, and 96.20%. Compared with Siamese network S3Net in transfer learning, the overall classification accuracy of the proposed algorithm increased by 2.04%, 3.6%, and 2.17%, the average classification accuracy increased by 0.83%, 2.33% and 1.51%, and the Kappa coefficient increased by 0.0225, 0.0423, and 0.0235, respectively. Compared with cross-domain method DCFSL in transfer learning, the overall classification accuracy increased by 11.22%, 4.35%, and 6.86%, the average classification accuracy increased by 8.83%, 6.72%, and 3.59%, and the Kappa coefficient increased by 0.1269, 0.0564, and 0.0756, respectively. This method improves the network structure by using a residual feature extractor and introducing a spatial-attention mechanism. These enhancements enable the feature-extraction network to have better feature extraction capabilities for both spectral and spatial information in hyperspectral data. Additionally, the inclusion of the label smooth cross-entropy loss function helps reduce network overfitting. Compared to alternative approaches, the classification performance of hyperspectral images can be notably enhanced, especially in scenarios with limited sample sizes. Experimental results from three distinct datasets demonstrate the superior consistency of the proposed method.
As depicted in the visualization results presented in Figure 11, Figure 12 and Figure 13, CRSSNet exhibits superior classification performance compared to other methods. In the Indian Pines dataset shown in Figure 11, it is evident that different classification methods exhibit certain classification errors and fail to achieve satisfactory results with limited samples. However, when compared to other methods, the proposed method achieves the best classification performance in the categories of Corn and Buildings Grass Tress Drives, among others. In the classification results of the Pavia University dataset in Figure 12, compared to other methods, CRSSNet shows lower misclassifications in the categories of Sheets and Bare soil, resulting in a superior classification effect map. In the Salinas dataset, shown in Figure 13, the classification effect map generated by CRSSNet closely resembles the real image, with only a few misclassifications observed in individual categories such as Grapes untrained. However, in categories like Brocoli Green Weeds_2 and Fallow, among others, CRSSNet demonstratess the best classification performance.

5. Conclusions

To address the challenges of insufficient spatial and spectral information extraction, as well as network overfitting in small-sample hyperspectral image classification, a method based on convolutional residuals and SAM Siamese networks is proposed. This method incorporates a feature residual-extraction module and a spatial attention mechanism to enhance feature information extraction. Furthermore, a more effective loss function is employed, which introduces noise and explores label factors to improve the classification capability of hyperspectral images with limited samples. Experimental results demonstrate the effectiveness of the proposed method on publicly available hyperspectral datasets, including Indian Pines, Pavia University, and Salinas. When compared to other methods, the proposed approach achieves superior classification performance in metrics such as OA, AA, and Kappa. These results highlight the method’s enhanced classification performance and improved generalization ability. However, there are still some challenges that need to be addressed. As observed from Figure 11, Figure 12 and Figure 13, the classification accuracy is relatively low when dealing with similar categories.
Currently, there are still challenges in solving the classification of hyperspectral images with limited samples, especially regarding the inherent high-dimensional nature of such images. Although our proposed network incorporates PCA for dimensionality reduction, the issue of high dimensionality persists due to the oversampling of the spectral dimensions in hyperspectral images. Recently, some graph-based methods for dimensionality reduction of hyperspectral images have been proposed. For instance, Luo et al. [38] introduced an Enhanced Hybrid Graph Discriminative Learning (EHGDL) based on hypergraphs, while Zhang et al. [39] presented Multilayer Graph Spectral Analysis of Hyperspectral Images using Multilayer Graph Signal Processing (M–GSP), among others, which have shown promising results. In future research, it would be beneficial to explore and build upon these graph-based methods to achieve better dimensionality reduction performance in hyperspectral image classification tasks.
Moreover, despite the need for a limited number of labeled samples in hyperspectral image classification, the reliance on manual labeling for training remains crucial. Therefore, future research should explore the potential of conducting hyperspectral image classification without the dependence on labeled samples. Such developments hold significant promise in contributing to the field and its practical applications.

Author Contributions

Conceptualization, M.X.; methodology, L.Y. and K.X.; software, M.X.; writing—original draft preparation, M.X.; writing—review and editing, H.Z. and G.Y.; supervision, Y.R. and Z.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Major Science and Technology Project in Yunnan Province, grant number 202202AD080004. The APC was funded by Yunnan Province Science and Technology Department.

Data Availability Statement

The datasets presented in this work are openly available on the website. Available online: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 6 August 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mao, Y.X.; Zhao, H.Q.; Feng, S.Q.; Xu, H.X.; He, T.; Song, L.J. Research on hyperspectral remote sensing openpit minerals identification method based on spectral matching. Nat. Resour. Informatiz. 2022, 130, 28–32+15. [Google Scholar]
  2. Zhao, Y.Y. Research on Nondestructive Detection Methods of Crop Seed Quality Based on Hyperspectral Imaging Technique; Zhejiang University: Hangzhou, China, 2021. [Google Scholar]
  3. Fang, Y.; Hu, Z.; Xu, L.; Wong, A.; Clausi, D.A. Estimation of Iron Concentration in Soil of a Mining Area from Uav-Based Hyperspectral Imagery. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  4. Zhang, Y. Inversion Study of Heavy Metals in Soils of Potentially Polluted Sites Based on UAV Hyperspectral Data and Machine Learning Algorithms. In Proceedings of the 2021 11th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 14–16 January 2021; pp. 1–5. [Google Scholar]
  5. Guilloteau, C.; Oberlin, T.; Berné, O.; Dobigeon, N. Informed Spatial Regularizations for Fast Fusion of Astronomical Images. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 2221–2225. [Google Scholar]
  6. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  7. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in hyperspectral image and signal processing a comprehensive overview of the state of the art. IEEE Geosci. Remote Sens. 2017, 5, 37–78. [Google Scholar] [CrossRef] [Green Version]
  8. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  9. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X.; Guo, L. KNN-based representation of super pixels for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  10. Zhang, H.K.; Li, Y.; Jiang, Y.N. Deep Learning for Hyperspectral Imagery Classification: The State of the Art and Prospects. Acta Autom. Sin. 2018, 44, 961–977. [Google Scholar]
  11. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 13–16 December 2012; Curran Associates: Red Hook, NY, USA; pp. 1097–1105. [Google Scholar]
  12. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  13. Hinton, G.E.; Osindero, S.; The, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  14. Zaremba, W.; Sutskever, I.; Vinyals, O. Recurrent neural network regularization. arXiv 2014, arXiv:1409.2329. [Google Scholar]
  15. Yang, H.; Yu, H.; Hong, D.; Xu, Z.; Wang, Y.; Song, M. Hyperspectral Image Classification Based on Multi-Level Spectral-Spatial Transformer Network. In Proceedings of the 2022 12th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Rome, Italy, 13–16 September 2022; pp. 1–4. [Google Scholar]
  16. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2017, arXiv:1409.1556v6. [Google Scholar]
  17. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  18. Zhu, W.; Zhao, C.; Qin, B.; Feng, S. Short and Long Range Graph Convolution Network for Hyperspectral Image Classification. In Proceedings of the IGARSS 2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 3564–3567. [Google Scholar]
  19. Liao, D.; Shi, C.; Wang, L. A Spectral–Spatial Fusion Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  20. Cui, B.G.; Ma, X.D.; Xie, X.Y. Hyperspectral image de-noising and classification with small training samples. J. Remote Sens. 2017, 21, 728–738. [Google Scholar] [CrossRef]
  21. Zhao, K.L.; Jin, X.L.; Wang, Y.Z. Survey on few-shot learning. Ruan Jian Xue Bao/J. Softw. 2021, 32, 349–369. [Google Scholar]
  22. Aurélien, B.; Amaury, H.; Marc, S. A survey on metric learning for feature vectors andstructured data. arXiv 2013, arXiv:1306.6709. [Google Scholar]
  23. Xue, Z.H.; Zhou, Y.Y.; Du, P.J. S3Net: Spectral–Spatial Siamese Network for Few-Shot Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
  24. Tang, H.; Huang, Z.; Li, Y.; Zhang, L.; Xie, W. A Multiscale Spatial–Spectral Prototypical Network for Hyperspectral Image Few-Shot Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  25. Zhang, C.; Yue, J.; Qin, Q. Global Prototypical Network for Few-Shot Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4748–4759. [Google Scholar] [CrossRef]
  26. Sung, F.; Yang, Y.; Zhang, L. Learning to compare: Relation network for few-shot learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
  27. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep Relation Network for Hyperspectral Image Few-Shot Classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef] [Green Version]
  28. Wang, W.; Chen, Y.; He, X.; Li, Z. Soft Augmentation-Based Siamese CNN for Hyperspectral Image Classification With Limited Training Samples. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  29. Huang, L.; Chen, Y. Dual-Path Siamese CNN for Hyperspectral Image Classification With Limited Training Samples. IEEE Geosci. Remote Sens. Lett. 2021, 18, 518–522. [Google Scholar] [CrossRef]
  30. Müller, R.; Kornblith, S.; Hinton, G. When Does Label Smoothing Help? arXiv 2019, arXiv:1906.02629. [Google Scholar]
  31. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  32. Hyperspectral Remote Sensing Scenes. Available online: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 31 July 2023).
  33. Li, Y.; Zhang, H.; Shen, Q. Spectral–Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef] [Green Version]
  34. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral-spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  35. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep few-shot learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  36. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph Information Aggregation Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef] [PubMed]
  37. Li, Z.; Liu, M.; Chen, Y.; Xu, Y.; Li, W.; Du, Q. Deep Cross Domain Few-Shot Learning For Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  38. Luo, F.; Zhang, L.; Du, B.; Zhang, L. Dimensionality reduction with enhanced hybrid-graph discriminant learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5336–5353. [Google Scholar] [CrossRef]
  39. Zhang, S.; Deng, Q.; Ding, Z. Multilayer graph spectral analysis for hyperspectral images. EURASIP J. Adv. Signal Process. 2022, 2022, 92. [Google Scholar] [CrossRef]
Figure 1. The main classification methods of hyperspectral images: (a) Hyperspectral image classification based on machine learning; (b) Hyperspectral image classification based on deep learning; (c) Hyperspectral image classification based on few-shot learning.
Figure 1. The main classification methods of hyperspectral images: (a) Hyperspectral image classification based on machine learning; (b) Hyperspectral image classification based on deep learning; (c) Hyperspectral image classification based on few-shot learning.
Electronics 12 03415 g001
Figure 2. Difference between metric learning-based single-channel neural network and dual-channel Siamese network.
Figure 2. Difference between metric learning-based single-channel neural network and dual-channel Siamese network.
Electronics 12 03415 g002
Figure 3. CRSSNet network structure.
Figure 3. CRSSNet network structure.
Electronics 12 03415 g003
Figure 4. Lightweight residual extractor.
Figure 4. Lightweight residual extractor.
Electronics 12 03415 g004
Figure 5. Spectral residual-extraction module.
Figure 5. Spectral residual-extraction module.
Electronics 12 03415 g005
Figure 6. Spatial feature extraction module.
Figure 6. Spatial feature extraction module.
Electronics 12 03415 g006
Figure 7. Spatial Attention Mechanisms.
Figure 7. Spatial Attention Mechanisms.
Electronics 12 03415 g007
Figure 8. Indian Pines datasets. (a) False-color composite image; (b) Ground-truth map; University of Pavia datasets. (c) False-color composite image; (d) Ground-truth map; Salinas datasets. (e) False-color composite image; (f) Ground-truth map.
Figure 8. Indian Pines datasets. (a) False-color composite image; (b) Ground-truth map; University of Pavia datasets. (c) False-color composite image; (d) Ground-truth map; Salinas datasets. (e) False-color composite image; (f) Ground-truth map.
Electronics 12 03415 g008
Figure 9. Impact of labeling factors on evaluation indicators under three datasets. (a) OA; (b) AA; (c) Kappa.
Figure 9. Impact of labeling factors on evaluation indicators under three datasets. (a) OA; (b) AA; (c) Kappa.
Electronics 12 03415 g009
Figure 10. Feature separability in ablation experiments. Indian Pines dataset: (a) S3Net, (b) CRSSNet (no SAM), and (c) CRSSNet. Salinas dataset: (d) S3Net, (e) CRSSNet (no SAM), and (f) CRSSNet; Pavia University dataset: (g) S3Net, (h) CRSSNet (no SAM), and (i) CRSSNet.
Figure 10. Feature separability in ablation experiments. Indian Pines dataset: (a) S3Net, (b) CRSSNet (no SAM), and (c) CRSSNet. Salinas dataset: (d) S3Net, (e) CRSSNet (no SAM), and (f) CRSSNet; Pavia University dataset: (g) S3Net, (h) CRSSNet (no SAM), and (i) CRSSNet.
Electronics 12 03415 g010
Figure 11. Results of classification for Indian Pines dataset.
Figure 11. Results of classification for Indian Pines dataset.
Electronics 12 03415 g011
Figure 12. Results of classification for Pavia University dataset.
Figure 12. Results of classification for Pavia University dataset.
Electronics 12 03415 g012
Figure 13. Results of classification for Salinas dataset.
Figure 13. Results of classification for Salinas dataset.
Electronics 12 03415 g013
Table 1. Details of the three hyperspectral image datasets.
Table 1. Details of the three hyperspectral image datasets.
Indian PinesPavia UniversitySalinas
Pixel resolution145 × 145610 × 310512 × 217
Spectral range (nm)400–2500430–860400–2500
Number of bands200103204
Spatial resolution (m)201.33.7
SensorAVIRISROSISAVIRIS
Number of categories16916
Total number of pixels21,025207,400111,104
Table 2. Classification accuracy of Pavia University dataset ablation experiments.
Table 2. Classification accuracy of Pavia University dataset ablation experiments.
NumberCategoryS3NetCRSSNet
(No SAM)
CRSSNet
1Asphalt83.7277.4680.05
2Meadows82.2987.0189.48
3Gravel79.6193.5188.79
4Tress91.2992.1491.73
5Sheets99.66100.00100.00
6Bare soil81.3189.6484.17
7Bitumen99.6599.9599.99
8Bricks84.7371.6786.29
9Shadow91.2394.9893.94
OA%84.4086.1988.00
AA%88.1689.5990.49
Kappa%80.1182.1984.34
Table 3. Classification accuracy of Indian Pines dataset ablation experiments.
Table 3. Classification accuracy of Indian Pines dataset ablation experiments.
NumberCategoryS3NetCRSSNet (No SAM)CRSSNet
1Alfalfa100.00100.00100.00
2Corn Notill59.4476.8775.32
3Corn Mintill62.7261.7962.39
4Corn92.7299.3599.66
5Grass Pasture81.2189.1889.41
6Grass Tree92.4691.5690.97
7Grass Pasture Mowed100.00100.00100.00
8Hay Windrowed99.4998.0597.93
9Oats100.00100.00100.00
10Soybean Notill73.2160.7258.13
11Soybean Mintill65.9863.7668.99
12Soybean Clean67.4063.2760.82
13Wheat97.3589.8590.30
14Woods95.0299.0399.01
15Buildings Grass Tress Drives90.2191.9995.33
16Stone Steel Towers97.1698.6499.20
OA%75.9977.2278.03
AA%85.8986.5086.72
Kappa%73.0874.4775.33
Table 4. Classification accuracy of Salinas dataset ablation experiments.
Table 4. Classification accuracy of Salinas dataset ablation experiments.
NumberCategoryS3NetCRSSNet (No SAM)CRSSNet
1Brocoli–Green–Weeds–199.04100.00100.00
2Brocoli–Green–Weeds–299.1399.9999.97
3Fallow98.80100.00100.00
4Fallow–Rough–Plow99.4799.8999.83
5Fallow–Smooth95.8996.3595.57
6Stubble98.5099.1199.55
7Celery99.7999.94100.00
8Grapes–Untrained82.6388.1688.91
9Soil–Vinyard–Develop99.8999.7299.71
10Corn–Senesced–Green–Weeds92.4496.7396.49
11Lettuce–Romain–4wk99.5099.2999.45
12Lettuce–Romain–5wk93.6498.1797.97
13Lettuce–Romain–6wk94.0091.6893.12
14Lettuce–Romain–7wk93.0598.4298.21
15Vinyard–Untrained93.4792.5594.24
16Vinyard–Vertical–Trellis98.8296.8198.06
OA%94.0395.6996.20
AA%96.1297.3097.63
Kappa × 100%93.3895.2195.73
Table 5. Classification accuracy of hyperspectral remote-sensing image dataset of Indian Pines.
Table 5. Classification accuracy of hyperspectral remote-sensing image dataset of Indian Pines.
Number3DCNNSSRNDFSL + SVMGia–CFSLDCFSLS3NetOurs
195.1218.3896.7595.1295.37100100
237.7064.7936.3847.3643.2659.4475.32
319.7727.6538.3437.9457.9562.7262.39
432.5126.9777.1678.4580.6092.7299.66
588.4580.7673.9272.8072.9181.2189.41
673.6586.8786.2573.3887.9692.4690.97
781.8232.2497.1010099.57100100
853.3510081.8291.5486.2699.4997.93
910057.6975.5610099.33100100
1041.3559.6952.2266.9162.4473.2158.13
1166.7170.8759.9667.0262.7565.9868.99
1237.4045.0036.5627.3848.7267.4060.82
1385.7188.2998.0096.5099.3597.3590.30
1462.5797.1884.6391.5985.4095.0299.01
1556.4236.6474.1063.7866.6990.2195.33
1690.3660.9810098.8697.6197.1699.20
OA%54.7661.3661.6965.7566.8175.9978.03
AA%63.9359.7573.0575.5477.8985.8986.72
Kappa × 100%48.7256.9156.7861.2462.6473.0875.33
Table 6. Classification accuracy of hyperspectral remote-sensing image dataset of Pavia University.
Table 6. Classification accuracy of hyperspectral remote-sensing image dataset of Pavia University.
Number3DCNNSSRNDFSL + SVMGia-CFSLDCFSLS3NetOurs
159.8291.8473.4379.9482.2083.7280.05
263.0595.1389.2585.8887.7482.2989.48
368.9155.2348.0947.6667.4679.6188.79
477.3178.0284.7296.0893.1691.2991.73
590.7798.3499.6599.9399.4999.66100
663.4053.5667.8165.5577.3281.3184.17
787.6460.0764.4877.2881.1899.6599.99
857.2785.3467.3762.9366.7384.7386.29
995.5798.0892.9299.7998.6691.2393.94
OA%65.7476.2679.6379.9383.6584.4088.00
AA%73.7279.5176.4179.4583.7788.1690.49
Kappa × 100%57.3770.5673.0573.8178.7080.1184.34
Table 7. Classification accuracy of hyperspectral remote-sensing image dataset of Salinas.
Table 7. Classification accuracy of hyperspectral remote-sensing image dataset of Salinas.
Number3DCNNSSRNDFSL + SVMGia-CFSLDCFSLS3NetOurs
195.2997.5573.9297.8599.4099.04100
297.2098.9796.8599.7699.7699.1399.97
391.4592.4796.2899.5491.9698.80100
497.3196.5099.1195.3999.5599.4799.83
591.2494.2080.7291.5892.7095.8995.57
698.8099.2891.6399.8099.5298.5099.55
799.6999.9897.7398.7798.8899.79100
866.4086.9082.3377.3674.5782.6388.91
996.2599.6494.4499.1499.5999.8999.71
1070.7292.0180.9666.9796.4292.4496.49
1193.1595.8693.3890.5996.6199.5099.45
1299.6599.1597.9499.9599.9393.6497.97
1392.6383.2495.7999.3499.3094.0093.12
1493.5695.1598.8798.4098.8593.0598.21
1568.0255.9791.1370.4575.3893.4794.24
1681.4198.9190.5792.9792.2298.8298.06
OA%84.2086.3986.9588.0089.3494.0396.20
AA%89.5693.2490.0892.3694.0496.1297.63
Kappa × 100%82.4684.9585.5186.6488.1793.3895.73
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xia, M.; Yuan, G.; Yang, L.; Xia, K.; Ren, Y.; Shi, Z.; Zhou, H. Few-Shot Hyperspectral Image Classification Based on Convolutional Residuals and SAM Siamese Networks. Electronics 2023, 12, 3415. https://doi.org/10.3390/electronics12163415

AMA Style

Xia M, Yuan G, Yang L, Xia K, Ren Y, Shi Z, Zhou H. Few-Shot Hyperspectral Image Classification Based on Convolutional Residuals and SAM Siamese Networks. Electronics. 2023; 12(16):3415. https://doi.org/10.3390/electronics12163415

Chicago/Turabian Style

Xia, Mengen, Guowu Yuan, Lingyu Yang, Kunming Xia, Ying Ren, Zhiliang Shi, and Hao Zhou. 2023. "Few-Shot Hyperspectral Image Classification Based on Convolutional Residuals and SAM Siamese Networks" Electronics 12, no. 16: 3415. https://doi.org/10.3390/electronics12163415

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop