Next Article in Journal
Multidimensional Assessment of Lake Water Ecosystem Services Using Remote Sensing
Previous Article in Journal
Deriving Aerodynamic Roughness Length at Ultra-High Resolution in Agricultural Areas Using UAV-Borne LiDAR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Boosting Few-Shot Hyperspectral Image Classification Using Pseudo-Label Learning

1
School of Computer Science and Technology, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
2
Shaanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi’an 710121, China
3
Xi’an Key Laboratory of Big Data and Intelligent Computing, Xi’an 710121, China
4
Shaanxi Key Lab of Speech & Image Information Processing (SAIIP), School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710129, China
5
National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, Xi’an 710129, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(17), 3539; https://doi.org/10.3390/rs13173539
Submission received: 11 August 2021 / Revised: 1 September 2021 / Accepted: 3 September 2021 / Published: 6 September 2021

Abstract

:
Deep neural networks have underpinned much of the recent progress in the field of hyperspectral image (HSI) classification owing to their powerful ability to learn discriminative features. However, training a deep neural network often requires the availability of a large number of labeled samples to mitigate over-fitting, and these labeled samples are not always available in practical applications. To adapt the deep neural network-based HSI classification approach to cases in which only a very limited number of labeled samples (i.e., few or even only one labeled sample) are provided, we propose a novel few-shot deep learning framework for HSI classification. In order to mitigate over-fitting, the framework borrows supervision from an auxiliary set of unlabeled samples with soft pseudo-labels to assist the training of the feature extractor on few labeled samples. By considering each labeled sample as a reference agent, the soft pseudo-label is assigned by computing the distances between the unlabeled sample and all agents. To demonstrate the effectiveness of the proposed method, we evaluate it on three benchmark HSI classification datasets. The results indicate that our method achieves better performance relative to existing competitors in few-shot and one-shot settings.

1. Introduction

Hyperspectral images (HSIs) collect the continuous reflectance of the imaging scene under a wide spectral range, which is often represented as a three-dimensional (3D) data cube and consists of hundreds of narrow spectral bands with a high spectral resolution [1]. Because they contain abundant spectral information, HSIs have been widely employed in many remote sensing-related applications [1,2,3,4,5,6], including environmental monitoring, target localization, military surveillance, and terrain classification, etc.
Among the various HSI-related applications, one of the most essential tasks is HSI classification, which aims at assigning a predefined class label to each pixel. Although many conventional machine learning-based methods, e.g., support vector machine (SVM) and k-nearest neighbors ( k NN) [7] etc., have been proposed for HSI classification, the inherently shallow structures of their feature extraction models prevent these methods from learning discriminative features, as well as from appropriately generalizing to challenging cases. As a result, performance may be further improved. Other mathematics based methods are also used in the HSI classification tasks with effective results [8,9].
Unlike conventional machine learning-based methods, deep neural networks can cast feature representation and classifier construction to train an end-to-end hierarchical network with labeled samples. Due to their deep structures, deep neural networks have the powerful ability to learn discriminative features for classification [10,11,12,13,14,15]. To date, extensive deep neural network-based methods for HSI classification [16,17,18,19,20,21] have been proposed and have achieved promising classification performance. For example, Liang et al. [16] propose extracting the deep feature representation of the samples using a stacked denoising auto encoder (SDAE), which is learned without supervision, to then predict the label of samples with a trained back-propagation network under supervision. In [18], Lakhal et al. introduce a deep recurrent neural network that incorporates high-level feature descriptors to tackle the challenging HSI classification problem. After witnessing the success of deep convolution neural networks (DCNNs) in various computer vision tasks [22,23], an increasing number of deep HSI classification methods have adopted the DCNN structure as their backbone. For example, Chen et al. [22] employed several 3D convolutional and pooling layers for deep feature extraction. Moreover, several other strategies to mitigate over-fitting, such as l 2 regularization and dropout, have been investigated. In [23], Gong et al. propose integrating the DCNN with multi-scale convolution with determinantal point process-based diversity-promoting deep metrics to obtain discriminative features for HSI classification. Recent progress has shown that due to the gradient vanishing problem, it is difficult to train a deep neural network as the network depth increases. To address this problem, He et al. [24] proposed a novel residual learning structure that can successfully mitigate the gradient vanishing problem. Inspired by this, many recent HSI classification methods have also begun to take advantage of residual learning for training deeper neural networks. For example, Zhang et al. [25] employed a 3D densely connected convolutional neural network (3D-DenseNet) to learn the spectral-spatial features of HSIs for classification. With its densely connected structure, the deeper network can be easier to train. Zhong et al. [26] develop an end-to-end spectral-spatial residual network (SSRN) that takes raw 3D cubes as input data, then employs residual blocks to connect every other 3D convolutional layer through identity mapping for HSI classification purposes.
Compared with conventional machine learning-based HSI classification methods, deep neural network-based methods often achieve much better performance. In order to succeed, these approaches require the network to be trained with sufficient labeled samples. In the HSI classification task, labeled samples are obtained by manually annotating some samples in the given HSI data set with a set of predefined class labels. In practice, since manually annotating an HSI pixel is generally laborious, time-consuming, and can only be accomplished by experts, the labeled samples obtained are always insufficient or even deficient. In such cases, the trained networks are prone to be overfitting (i.e., fail to appropriately generalize to the unseen test samples). To alleviate this problem, some transfer learning-based methods [27,28,29] have been developed. These methods aim to train on an HSI dataset (i.e., source domain) and classify the pixels of same classes on another HSI dataset (i.e., target domain).
However, this kind of method requires a large number of labeled samples to be available in the source domain, which is still unfeasible for lots of applications. Thus, some methods [30,31,32] propose utilizing data augmentation to increase the number of labeled training samples on the given dataset (i.e., target domain), without relying on another, similar dataset (i.e., source domain). For example, in [30], Li et al. proposed to substantially increase the number of labeled training samples using a novel pixel-pair strategy, which combines two pixels (to form a pair of pixels) and then classifies the pixel-pair using a trained DCNN. Although methods of this kind can generate more labeled samples from the existing labeled ones, they rely heavily on those samples. Once the number of the labeled samples has been heavily limited (i.e., providing only one labeled sample per class), the labeled samples generated will still be insufficient to overcome the overfitting problem. Other methods emphasize the need to find a suitable way to use the unlabeled samples, since there are plenty of unlabeled samples in real applications. These methods are also denoted as semi-supervised methods [33,34,35,36,37,38,39]. Such methods have been explored in the context of conventional machine learning-based approaches. For example, Yang et al. [39] proposed a new spatio-spectral Laplacian support vector machine (SS-LapSVM) method for semi-supervised HSI classification, which utilizes a clustering assumption on spectral vectors and the neighborhood spatial constraints of the given HSI to formulate a manifold regularizer and a spatial regularizer separately. These regularizers are then utilized to regularize the classification results of the conventional SVM-based methods. Meanwhile, the semi-supervised methods have been applied in deep learning-based methods. The main idea of Zhu et al. [39] is based on an auxiliary classifier generative adversarial network (ACGAN) [40], which uses a generator that regards the label and noise information as input to create fake data in order to train the discriminator jointly with the labeled data. Liu et al. [38] proposed a multi-task framework based on a ladder network [41] that combines an encoder-decoder for utilizing the unlabeled data. Although the above methods can to some extent mitigate the problem resulting from insufficient labeled samples, few of them have attempted to deal with a more challenging deep learning-based HSI classification problem in which only a few or even only one labeled sample per class is provided for training. For simplicity, we refer to this problem as few-shot or one-shot HSI classification.
To adapt the deep neural network-based HSI classification method to cases with few or even only one labeled sample, we propose a novel few-shot deep learning framework for HSI classification, which aims to appropriately mitigate the overfitting problem resulting from a limitation in the number of labeled samples. To this end, we first revisit the HSI utilized for classification. More specifically, in HSI, there are a large number of unlabeled samples in addition to a few labeled ones. These unlabeled samples have underlying relations with the labeled samples (e.g., intra-class similarity or inter-class dissimilarity). Inspired by this observation, we propose to annotate these unlabeled samples with pseudo-labels by using their relation to the labeled samples as an auxiliary dataset, then incorporate this dataset into a joint learning framework in order to assist in training the network on a few labeled samples. In more detail, we first develop a new two-heads DCNN architecture in which the labeled and unlabeled samples share the same feature extractor sub-network, but two different classifier sub-networks (i.e., heads) to achieve classification on the few provided labeled samples and the auxiliary unlabeled dataset separately. To this end, by considering each labeled sample as a reference agent, we assign each unlabeled sample a soft pseudo-label by computing the distances between the unlabeled sample and all agents, which is a totally different approach from that adopted by the existing semi-supervised HSI classification method. Through joint learning on both labeled and unlabeled samples, the proposed framework automatically borrows the weak supervision from the unlabeled samples in order to enhance the supervision from the labeled samples for feature learning purposes, and thus appropriately mitigates overfitting. It is notable that the proposed framework provides a general learning framework, through the use of which any kind of deep neural network can be exploited as the feature extractor. With extensive experiments on three HSI classification benchmark datasets, the proposed method exhibits obvious superiority over existing competitors in few-shot and one-shot settings.
The main contributions of this study can be summarized as follows:
(1) We propose a novel few-shot deep learning framework for HSI classification, which borrows the supervision from the unlabeled samples to assist in training the deep neural network on a few labeled samples with a joint learning strategy.
(2) A soft pseudo-label learning method is proposed to assign soft pseudo-labels to unlabeled samples for joint learning.
(3) The proposed framework provides a general learning framework, with which any kind of deep neural network can be exploited as the feature extractor sub-network.
(4) The proposed method demonstrates state-of-the-art few-shot and one-shot HSI classification performance on three benchmark HSI classification datasets.

2. The Proposed Methods

2.1. The Proposed Few-Shot Deep Learning Framework

In this study, we denote the labeled training samples and unlabeled samples as T = x 1 , y 1 , x 2 , y 2 , , x n , y n and U = z 1 , z 2 , , z m , respectively. x i d × 1 denotes the i -th d -dimensional labeled sample, y i is the class label of x i and z j d × 1 is the j -th unlabeled sample.
In essence, the HSI classification method aims to learn a mapping function between the training sample x and its label y , then exploits the learned function to predict the label of the unlabeled sample z . Since the deep neural network has a powerful non-linearity representation ability, it is suitable for building such a mapping function. Thus, large numbers of deep learning-based methods have been proposed for HSI classification, among which DCNNs and residual networks have attracted significant attention.
An important aspect underpinning the success of deep learning-based methods is the training of a deep neural network containing large numbers of parameters with sufficient training samples. However, due to the high cost of labeling in real applications, only a few labeled samples (i.e., n is very small) or even only one labeled sample per class can be provided for network training, which results in the deep network overfitting, i.e., fails to appropriately generalize to the unseen test samples. It is therefore necessary to adapt the deep learning-based method to a few labeled training samples in real applications.
Although collecting labeled samples is difficult, a large number of unlabeled samples are always available (i.e., m can be very large). More importantly, these unlabeled samples exhibit underlying relations to these labeled ones, such as intra-class similarity or inter-class dissimilarity, which inspires us to capture the underlying relations and exploit these relations for few-shot HSI classification. In more detail, we propose to exploit pseudo-labels to reflect such underlying relations, then annotate the unlabeled samples with those pseudo-labels, from which an auxiliary dataset containing pseudo-labels S = z 1 , y 1 , z 2 , y 2 , , z m , y m can be obtained (the details of this pseudo-label learning can be seen in Section 2.2). Based on the generated auxiliary dataset, we then borrow the supervision from the auxiliary unlabeled dataset to assist the feature in learning on a few labeled samples, which allows us to appropriately mitigate the overfitting problem.
In line with this concept, we propose a two-heads few-shot deep learning framework, illustrated in Figure 1. Two different classifier sub-networks, C and C separately achieve classification on T and S , but share a common feature extractor sub-network F θ , θ denotes the parameters in the feature extractor sub-network. For the classification branch on T , the predicted label can be obtained as follows:
y ^ k = C F θ x k ,
where y ^ k denotes the predicted label for sample x k . In addition, for the classification branch on S , the relationship between the pseudo-label and the sample z k is formulated as follows:
y ^ k = C F θ z k ,
in which y ^ k is the predicted label for the unlabeled sample z k .
We regard four samples as one batch just for convenience, though it is irrelevant to the real settings in experiments. Thanks to the shared feature extractor sub-network, both labeled and unlabeled samples are integrated into one framework, to which more samples are introduced in order to regularize the original few-shot supervised branch. Through joint learning the proposed network via a back-propagation algorithm, the supervision from the auxiliary unlabeled dataset is exploited to appropriately mitigate the overfitting problem resulting from the existence of few labeled samples, i.e., the auxiliary dataset acts as a regularizer that trains the feature extractor. The details of training the proposed model can be seen in Section 2.4.
It is notable that the proposed framework is a general learning framework, through the use of which any kind of deep neural network can be exploited as the feature extractor sub-network. In this paper, we use DCNNs and residual networks, two state-of-the-art network structures, to construct the extraction sub-network. The details are outlined in Section 2.3.

2.2. Soft Pseudo-Label Learning

The proposed framework is trained in a supervised manner using both the labeled and unlabeled samples with pseudo-labels. To achieve this, we require a suitable strategy that assigns a pseudo-label to each unlabeled sample. Some semi-supervised methods conduct clustering (e.g., k -means algorithm) on the unlabeled samples to obtain the cluster labels [42], and the cluster labels are then exploited as the pseudo-labels. The k -means method can only assign each unlabeled sample a specific unique cluster label that is equivalent to the hard label. For this reason, we use the resulting cluster labels obtained via k -means clustering as the hard pseudo-labels. However, this simple strategy has a number of limitations. First, it ignores the relationship between the labeled samples. Second, the hard label cannot adequately represent samples in which the distance to one cluster center is closer than the distances to the other cluster centers. Thus, the classification performance of these methods using hard pseudo-labels is still limited, especially in the few-shot or one-shot setting.
In contrast to hard labels, soft labels have high entropy, and can thus convey more valuable information during the classification process. In addition, soft labels can lead to continuous gradients during network training, which is beneficial for convergence and mitigates the problem of the algorithm becoming trapped into bad local minima [43]. Inspired by this, we propose a soft pseudo-label learning method that operates by exploiting the underlying relation between the labeled and unlabeled sample. In particular, we conclude that specific underlying relations exist such as intra-class similarity and inter-class dissimilarity between the unlabeled sample and the labeled sample. Samples (labeled or not) that belong to the same class show high similarity, while samples (labeled or not) that belong to different classes exhibit obvious difference, and vice versa. Therefore, by considering each labeled sample as a reference agent, we can assign each unlabeled sample a soft pseudo-label by computing the distances between the unlabeled sample and all agents.
In more detail, taking an unlabeled sample z p as an example, we generate its soft pseudo-label y p L as follows. Firstly, each labeled sample is adopted as an agent, i.e.,
a g e n t i j = x i j ,
in which x i j denotes the j -th training sample from i -th class. We then compute the distances between the unlabeled sample z p and all agents as follows:
d i s t i j p = s i m a g e n t i j , z p ,
Here, d i s t i j p is the distance between the agent a g e n t i j and the unlabeled sample z p , which is calculated via the similarity function s i m , . For simplicity, we utilize the Euclidean distance as the function s i m , , the features used in the function are the original pixel of the HSI. From the distances between z p and all agents in class i , we select the one with the smallest distance as the reference agent in the i -th class as follows:
a g e ^ n t i p = arg min a g e n t i j d i s t i j p ,
Then, the i -th element of pseudo-label y p can thus be obtained via
y p [ i ] = s o f t m a x i 1 s i m a g e ^ n t i p , z p ,
By applying the same operations on all classes, the soft pseudo-label of the unlabeled sample z p can be finally represented as:
y p = y p [ 1 ] , y p [ 2 ] , , y p [ L ] ,
It should be noted that all labeled samples of each class are exploited as agents in the proposed method. Although we can average the labeled samples to create a representative agent, this averaging strategy is likely to impair the underlying relation to some extent. For example, some samples are more similar to a specific agent than the average agent, and thus degrade the performance of the proposed method. The experimental results can be seen from the ablation study in Section 3.4.

2.3. The Architecture of the Adopted Feature Extractor Sub-Network and Classifier Sub-Network

Figure 1 illustrates the pipeline of the proposed two-heads few-shot deep learning framework, which includes one shared-feature extractor sub-network, one few-shot classifier sub-network for labeled data, and one auxiliary classifier sub-network for unlabeled data with soft pseudo-labels. Because both labeled and unlabeled data are utilized for training, the proposed method can be seen as a semi-supervised deep learning method. If we utilized a feature extractor sub-network and a few-shot classifier sub-network only within the proposed framework, the proposed framework would degenerate into a standardly used supervised deep learning method. Thus, by comparing the proposed method with its supervised counterpart (i.e., utilizing the feature extractor sub-network and few-shot classifier sub-network only), the advantage of introducing unlabeled data for HSI classification can be observed. For this reason, we refer to the supervised counterpart as the backbone. It is notable that the proposed network is a two-head network structure, while its supervised counterpart is a one-head network structure (with a few-shot classifier sub-network, but without an auxiliary classifier sub-network).
The proposed framework is a general learning framework, which enables any kind of deep neural network to be exploited as the feature extractor sub-network. In various applications, to demonstrate the adaptability of the proposed framework to different network structures, two network structures that are widely used for HSI classification (i.e., 3D-CNN and SSRN) are adopted as the feature extraction sub-network. It is notable that the structure of the classifier sub-network (i.e., the heads) is fixed because it is rather simple (i.e., the classifier sub-network is implemented via a fully connected layer).
(1) 3D-CNN feature extractor sub-network: It has been demonstrated that a 3D-CNN structure that utilizes both spatial and spectral information can result in better HSI classification performance when compared with a method which utilizes only the spectrum information. This is because HSI data is inherently a 3D data cube that contains both spatial and spectral information, which makes 3D-CNN a natural structure choice. Thus, 3D-CNN is adopted as the first backbone network for feature extraction. By following the settings of existing 3D-CNN structures for HSI classification, we constructed a 3D-CNN network for feature extraction, as illustrated in Figure 2. The network consists of four convolutional layers, each of which is equipped with 3D convolutional kernels for feature extraction. The detailed parameter settings (such as convolution kernels and stride) are also included in Figure 2. The network contains 3 convolutional layers and a fully connected layer. In the first convolutional layer, the kernel size is 5 × 1 × 8 and 64 feature maps; in the second convolutional layer, the kernel size is 1 × 5 × 3 and 128 feature maps; in the third convolutional layer, the kernel size is 1 × 1 × 3 and 256 feature maps, s is the step size. The finally fully connected layer includes 512 nodes.
(2) SSRN feature extractor sub-network: Although DCNN has a strong feature extraction ability, due to the gradient vanishing problem it becomes difficult to train a deep neural network as the network depth increases. To overcome the gradient vanishing problem due to increasing network depth, He et al. propose a novel residual learning structure that can successfully mitigate the gradient vanishing problem [24]. The key to residual learning is to introduce a shortcut connection into the original network structure, with which the gradient vanishing problem can be well mitigated. Inspired by this, many recent HSI classification methods have also begun to take advantage of residual learning to train a deeper neural network. For this purpose, SSRN, a typical residual learning structure-based network, is exploited as the feature extractor sub-network within the proposed network. More specifically, we constructed a spectral-spatial residual network to capture both the spectral and spatial information. In this network, the spectral and spatial information is extracted consecutively based on the spectral and spatial residual blocks. For the spectral residual learning block, the convolutional layers, together with two shortcut connections, are utilized to extract the spectral features. For the spatial residual learning block, the convolutional layers, together with one shortcut connection, are exploited to extract spatial information. The details of the modeled residual network can be seen in Figure 3, which also includes the architecture and the parameter settings.
(3) Classifier sub-network: In general, the classifier sub-network adopts a much simpler architecture when compared with the feature extractor sub-network above. For the supervised deep learning-based classification task, it often consists of a fully connected layer. In this study, we use two fully connected layers to form the classifier sub-networks on T and S independently.

2.4. Loss Function

The constructed deep neural network contains a number of unknown parameters. To train and determine those parameters using a back-propagation algorithm, an appropriate loss needs to be defined beforehand.
Since the proposed network is a two-branch network, which includes two separate classification tasks on T and S , for the classification branch on T , we adopt the cross-entropy loss, defined as follows:
L f e w = k = 1 n j = 1 L 1 [ j = y k ] ln y ^ k ,
Here, y ^ k is an L -dimensional vector calculated via Equation (1), which predicts the label of x k .
It is notable that this achieves a classification task with very limited labeled samples. However, a deep neural network with a large number of parameters will confront a heavily ill-posed problem when provided with few labeled training samples (e.g., five samples per class) for an HSI classification task. This is primarily because a large number of parameters cannot be fully and directly trained with a small number of labeled training samples; i.e., the parameters obtained by optimizing Equation (8) are unstable.
To resolve this issue, an effective method in the machine learning field is to enforce a suitable regularization item into Equation (8), from which the solution space can be effectively reduced to find a stable parameter. Inspired by this, we propose utilizing the unlabeled data with soft labels (i.e., the classification task on the auxiliary dataset S ) as the regularization item, as follows:
L a u x = k = 1 m j = 1 L 1 [ j = y k ] ln y ^ k ,
where y ^ k is an L -dimensional vector calculated by Equation (2).
With these two loss functions defined above, we can formulate the overall loss function of the proposed two-branch network as follows:
L = L f e w + λ L a u x ,
where λ is a predefined weighting factor. It is notable that the shared feature extractor module can guarantee these two branches are trained jointly. Then, by minimizing the loss L , the parameters can be trained with the back-propagation algorithm.

3. Experiments

To demonstrate the effectiveness of the proposed method in terms of few-shot HSI classification, we conducted experiments on three benchmark HSI datasets. In the below, we will first introduce the adopted benchmark HSI datasets and the experimental settings. We will next explain the undertaken appropriate and adequate comparison experiments between the proposed method and some state-of-the-art HSI classification methods in various cases, followed by some ablation studies.

3.1. Datasets

In this study, three benchmark HSI datasets were adopted to verify the efficacy of the proposed method including the Pavia University (PaviaU) dataset, the Salinas dataset, and the Indian Pines dataset [44].
PaviaU dataset: The PaviaU dataset was acquired by a sensor during a flight campaign over the university of Pavia, Italy, and includes 9 classes and 610 × 340 pixels, as shown in Table 1. In addition, the false color image and the ground truth map are shown in Figure 4. For the PaviaU dataset, each pixel has 103 spectral reflectance bands ranging from 430 to 860 nm, and the spatial resolution is 1.3 m. In these experiments, all bands are utilized for HSI classification.
Salinas dataset: The Salinas dataset was acquired using the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) from California, which includes 16 classes and 512 × 217 pixels (shown in Table 1). The spatial resolution is 3.7 m. The false color image and the ground truth map are shown in Figure 5. For this dataset, each pixel has 224 bands ranging from 400 to 2450 nm. After the water-absorption bands were removed, 204 bands were retained for the experiments.
Indian Pines dataset: The Indian Pines dataset was collected by AVIRIS from the agricultural Indiana Pines test site in northwestern Indiana. It contains 16 land cover classes and images of 145 × 145 pixels. The detailed class name, the numbers of labeled pixels and the class map can be seen in Table 1 and Figure 6. The spatial resolution of HSI in this dataset is about 20 m. Each pixel has 220 spectral reflectance bands ranging from 400 to 2450 nm. All 220 bands were utilized for HSI classification in this experiment.
These three datasets have a public characteristic, which are all unbalanced datasets.

3.2. Comparison Methods and Measures

To verify the performance of the proposed method, we compared it with Support Vector Machine (SVM), Spatio-Spectral Laplacian Support Vector Machine (SS-LapSVM) [37], 3D-GAN [39], SS-CNN [38], 3D-CNN, SSRN [26], and 3D Densely Connected Convolutional Network (3D-Densenet) [25]-based HSI classification methods. Among these competing methods, the SVM-based method is a conventional machine learning-based method that specializes in dealing with small sample problems. SS-LapSVM is a traditional semi-supervised method that utilizes both labeled and unlabeled samples. 3D-GAN and SS-CNN are deep learning-based methods for semi-supervised HSI classification. 3D-CNN and SSRN are two representative deep learning-based methods, which utilize the labeled samples only for classification and also serve as the backbone for feature extraction. Based on these two backbone feature extractor networks, we obtained two different versions of the proposed method. For the sake of simplicity, we denote the proposed method based on 3D-CNN and SSRN as 3 D - CN N o u r s and SSR N o u r s , respectively. To facilitate further comparison, variants of the proposed two-branch networks were specifically designed so that only the soft pseudo-labels are replaced by the hard pseudo-labels generated via k-means clustering [42]. We denote these two variant methods as 3 D - CN N k - m e a n s and SSR N k - m e a n s . For the adopted k-means clustering method, the number of clusters is set equal to the categories within the HSI dataset (e.g., we set the number of clusters to 16, since there are 16 categories in the Indian Pines dataset). To further testify the effectiveness of proposed soft pseudo-labels, we generate the hard label via the proposed soft pseudo-label learning method, in which one sample can only belong to one category. The proposed method with hard pseudo-label is denoted as 3 D - CN N h a r d l a b e l and SSR N h a r d l a b e l . In addition, 3D-Densenet, a newly proposed deep learning-based method that achieves excellent HSI classification performance when provided with plenty of labeled samples was also introduced as a competing method.
As for the evaluation measurement, three conventional evaluation metrics including overall accuracy (OA), average accuracy (AA) and the Kappa coefficient k were adopted to evaluate the HSI classification performance of different methods.

3.3. Experimental Results

In this subsection, we focus primarily on comparing the proposed method with other competing methods in order to demonstrate its effectiveness in the field of HSI classification under few-shot and one-shot settings. Since the problem we are addressing involves obtaining an appropriate deep learning-based method when heavily limited labeled samples are provided for training, we conducted experiments with three and five labeled samples per class and experiments with one labeled sample per class. In more detail, the experiments with three and five labeled samples per class correspond to the few-shot HSI classification problem, while the experiments with one labeled sample are for a more challenging one-shot HSI classification task.
We repeated the same HSI classification process for each number of labeled samples. Taking three labeled samples as an example, we randomly chose three labeled samples per class as the training set, then treated the rest of the samples as the test set in the classification experiments. To eliminate the influence of randomness on classification accuracy when selecting training samples, the final classification accuracy for each method was calculated by averaging the classification results independently obtained from ten random training and testing rounds. The average classification results for all methods on different datasets are summarized and reported in Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10. To more accurately assess the performance of the different methods, the classification maps of these methods on three datasets are also visualized in Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14 and Figure 15, where (a)–(m) represent the results from different classification methods. From these listed results, we obtained the following conclusions.
(1) Comparison with conventional machine learning-based method: Although the effectiveness of the deep learning-based methods has been demonstrated when a large number of labeled samples are provided for training, it can be seen from the results in Table 2 that the performance of the deep learning-based methods cannot be guaranteed when only a few labeled samples are provided. In most cases, the performance of 3D-Densenet and SSR N k - m e a n s are even inferior to that of SVM. This is because deep structured network-based methods are prone to overfitting when few labeled training samples are provided. Nevertheless, the proposed deep learning methods obtain better performance in the majority of cases than SVM. For example, the proposed method improves OA over SVM by about 5.52% when five labeled samples in PaviaU dataset are provided, which demonstrates that the proposed networks can better adapt to the few-shot HSI classification problem.
(2) Comparison with semi-supervised method: In all methods, SS-LapSVM, 3D-GAN, SS-CNN, the proposed methods (i.e., 3 D - CN N o u r s and SSR N o u r s ), two variants of our methods (including 3 D - CN N k - m e a n s and SSR N k - m e a n s ), and the proposed method with hard pseudo-label (including 3 D - CN N h a r d l a b e l and SSR N h a r d l a b e l ) are semi-supervised methods. From the results, we can observe that the proposed methods have obvious advantages over the competing semi-supervised methods in terms of HSI classification when only a very limited number of training samples are provided for training. It is notable that the traditional semi-supervised method SS-LapSVM achieves almost the worst performance out of all comparison methods. Compared with deep learning-based semi-supervised methods such as 3D-GAN and SS-CNN, the proposed method (i.e., 3 D - CN N o u r s ) shows promising results with a higher accuracy. Considering all these methods are semi-supervised methods, we can conclude that the proposed method can better utilize the unlabeled data for better HSI classification results. In addition, by comparing the performance of the proposed methods with 3 D - CN N k - m e a n s and SSR N k - m e a n s as well as 3 D - CN N h a r d l a b e l and SSR N h a r d l a b e l , we can see that the soft pseudo-labels also contribute to the improvement of HSI classification.
(3) Comparison with deep learning-based method: 3D-Densenet, 3D-CNN, SSRN, the proposed methods ( 3 D - CN N o u r s and SSR N o u r s ) and the two variants of the proposed methods ( 3 D - CN N k - m e a n s and SSR N k - m e a n s ) are all deep learning-based methods. From the results, we can see that the proposed methods achieve better classification performance than other competing deep learning-based methods. This is because the competing methods, such as 3D-Densenet, are more likely to be overfitting when heavily few-labeled samples are available for HSI classification. By contrast, with the regularization from the auxiliary dataset with soft pseudo-labels, the overfitting problem of the proposed method can be effectively reduced, with which the classification accuracy is improved.
(4) Applicability to different kinds of networks: The proposed framework is a general semi-supervised framework, with which the unlabeled samples could be effectively used to mitigate overfitting and improve the HSI classification accuracy. To demonstrate the applicability of the proposed framework to different kinds of networks, we used a 3D-CNN and SSRN as the backbones (i.e., only utilizing the labeled samples for training). By comparing the results between 3 D - CN N o u r s (i.e., the proposed method) and 3D-CNN, we can find that the proposed method based on the 3 D - CN N o u r s structure could obtain better performance on three datasets with different number of training samples. For example, when provided five labeled training samples on the PaviaU dataset, the average OA of 3 D - CN N o u r s is improved by 3.89% over 3D-CNN. Similar, the experimental results from SSRN and SSR N o u r s also reflect that the SSR N o u r s obtains better classification results. In summary, from the results of both 3D-CNN and SSRN, we can conclude that the proposed framework is applicable to different kinds of networks for few-shot learning.
(5) Applicability to one-shot HSI classification: The proposed methods do not only improve the few-shot HSI classification with heavily limited training samples (i.e., three and five labeled samples), but can also improve the one-shot HSI classification when only one training sample per class is provided. Although the performance from the one-shot classification cannot be compared with that of the few-shot classification for all methods, it can be observed in Table 8 that the proposed methods still achieve better performance relative to the other comparison methods. For example, the proposed method outperforms SVM by 8.68% in terms of OA on the Indian Pines dataset. In addition, it can be observed that when only one labeled sample is provided, the SS-LapSVM method even fails to classify the test samples in these benchmark datasets, which indicates the difficulty of utilizing the unlabeled samples with only one labeled training sample. Nevertheless, the proposed method effectively utilizes the unlabeled samples and thus yields better classification results. We can therefore conclude that the proposed method is also effective for one-shot classification.

3.4. Ablation Study

(1) Effectiveness of the reference agent: In this paper, a group of agent references were adopted to assign soft pseudo-labels, among which all labeled samples of each class were exploited as agents. In addition to this strategy, we also average all labeled samples to generate a representative agent for each class, then utilized the representative agent to obtain the soft pseudo-label, which is referred to as Average in this study. To demonstrate the superiority of the proposed agent learning method, we conducted comparison experiments between Ours and Average. Table 11 reports the comparison results, in which 3D-CNN was adopted as the backbone and five labeled samples were provided for training. It can thus be seen that the performance of Ours achieves promising results with a higher accuracy than that of Average, as shown in Table 11. For example, the OA of Ours outperforms Average by 1.45%. This is because the underlying relation is impaired to some extent when averaging all labeled samples of each class, whereas the method we adopt preserves these relations.
(2) Sensitivity of λ : We used the loss function defined in Equation (10) to train and determine the parameters contained in the proposed two-branch deep neural network, where λ was used to balance the loss function from each branch. Figure 16 illustrates the performance of OA, AA and the Kappa coefficient κ as it varies with different λ , in which the OA, AA and the Kappa coefficient k are average results from 10 rounds of experiments with the 3D-CNNours feature extractor, trained with three labeled samples. From the results, we can determine that the classification performance of the proposed method first increases and then decreases as λ increases from 0 to 2. When λ is around 0.5, the proposed method can obtain its best performance.
(3) The impact of feature extraction: According to the proposed framework, the features were automatically extracted from the data since the proposed method is an end-to-end deep learning-based method. To verify that the proposed network may also be utilized with other feature extraction methods (i.e., we input the extracted feature rather than the original data into the constructed network), we utilized a commonly used feature extraction method termed principal component analysis (PCA) for experiments. More specifically, we extract the PCA features from the original spectrum for each pixel, then input the PCA feature instead of the original spectrum into the constructed two-head network. The results are summarized in the following Table 12. It can be seen that the proposed method still achieves better performance when input with PCA features.
(4) Effectiveness of the two-head network: To demonstrate the effectiveness of the constructed two-head network, we constructed a one-head network structure which includes only the shared feature extraction module and one classification module. Both labeled data (with true-class label) and un-labeled data (with pseudo-label) were trained with the same network structure. In this case, the only difference between the proposed method and this competing method is that the proposed method utilizes a two-branch network structure while the competing method utilizes a one-branch network. Results from the experiments conducted on the PaviaU dataset are summarized in Table 13. It can be clearly seen that the constructed two-head network obtains better performance compared with the one-head structure-based method in most cases, to which we can attribute the improvement of the proposed method over one-head 3D-CNN to the two-branch strategy.
(5) Effect of the proposed pseudo-label generation method: To further verify the effectiveness of the adopted soft pseudo-label generation method based on reference agent, we further conducted a comparison method, in which only the soft pseudo-labels were generated based on the k-means method. Specifically, we first run k -means to get the cluster centers, and then utilized the distances between the unlabeled data and the cluster centers to compute the soft-labels for the unlabeled data. The relative experiments are summarized in Table 14. It can be clearly seen that the proposed pseudo-label generation method based on a reference agent has better performance compared with the pseudo-label generation method based on k -means.

4. Conclusions

The training process of deep neural networks usually depends on the availability of a large number of training labeled samples. However, owing to the heavy labeling cost for HSI data, the number of labeled samples available in practical applications is heavily limited, resulting in the deep neural network-based HSI classification method overfitting and achieving a poor generalization ability. To address this issue, this paper has proposed a novel few-shot learning method for HSI classification, which borrows the supervision from the auxiliary dataset with soft pseudo-labels in order to train the feature extractor on the few-shot dataset. By employing joint training on both a few-shot dataset and auxiliary dataset with soft pseudo-labels, the proposed method can achieve better classification results on both few-shot HSI classification and one-shot HSI classification tasks. Experimental results demonstrate the superiority of the proposed approach on three hyperspectral benchmark datasets, compared with other competing methods.

Author Contributions

All of the authors made significant contributions to this work. C.D., W.W. and L.Z. devised the approach and analyzed the data; L.Z., Y.Z. and W.W. helped design the remote sensing experiments and provided advice for the preparation and revision of the work; and Y.L., Y.W. and M.Z. performed the experiments. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundations of China (Grant no.61901369 and Grant no.62071387), the Foundation of National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology (Grant No.20200203) and the National Key Research and Development Project of China (No. 2020AAA0104603).

Data Availability Statement

Not applicable.

Acknowledgments

The paper is funded by the key discipline special fund construction project of Shaanxi Province. We acknowledge the author for the HSI datasets.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in Hyperspectral Image and Signal Processing: A Comprehensive Overview of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef] [Green Version]
  2. Wei, W.; Zhang, L.; Jiao, Y.; Tian, C.; Wang, C.; Zhang, Y. Intracluster Structured Low-Rank Matrix Analysis Method for Hyperspectral Denoising. IEEE Trans. Geosci. Remote Sens. 2018, 57, 866–880. [Google Scholar] [CrossRef]
  3. He, L.; Li, J.; Liu, C.; Li, S. Recent Advances on Spectral-Spatial Hyperspectral Image Classification: An Overview and New guidelines. IEEE Trans. Geosci. Remote Sens. 2017, 56, 1579–1597. [Google Scholar] [CrossRef]
  4. Guerra, R.; Barrios, Y.; Díaz, M.; Santos, L.; López, S.; Sarmiento, R. A new algorithm for the on-board compression of hyperspectral images. Remote Sens. 2018, 10, 428. [Google Scholar] [CrossRef] [Green Version]
  5. Zhang, L.; Wei, W.; Bai, C.; Gao, Y.; Zhang, Y. Exploiting Clustering Manifold Structure for Hyperspectral Imagery Super-Resolution. IEEE Trans. Image Process. 2018, 27, 5969–5982. [Google Scholar] [CrossRef] [PubMed]
  6. Weiss, M.; Jacob, F.; Duveiller, G. Remote sensing for agricultural applications: A meta-review. Remote Sens. Environ. 2020, 236, 111402. [Google Scholar] [CrossRef]
  7. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X.; Guo, L. KNN-Based Representation of Superpixels for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  8. Andriyanov, N.; Dementiev, V.; Kondratiev, D. Tracking of Objects in Video Sequences. Intell. Decis. Technol. 2021, 253–262. [Google Scholar]
  9. Andriyanov, N.A.; Vasil’ev, K.K.; Dement’ev, V.E. Investigation of Filtering and Objects Detection Algorithms for A Multizone Image Sequence. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, 4, 7–10. [Google Scholar] [CrossRef] [Green Version]
  10. Hu, H.; Gu, J.; Zhang, Z.; Dai, J.; Wei, Y. Relation Networks for Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3588–3597. [Google Scholar]
  11. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  12. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 833–851. [Google Scholar]
  13. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding Convolution for Semantic Segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  14. Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context Encoding for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7151–7160. [Google Scholar]
  15. Yalniz, I.Z.; Jégou, H.; Chen, K.; Paluri, M.; Mahajan, D. Billion-scale semi-supervised learning for image classification. arXiv 2019, arXiv:1905.00546. [Google Scholar]
  16. Liang, P.; Shi, W.; Zhang, X. Remote Sensing Image Classification Based on Stacked Denoising Autoencoder. Remote Sens. 2017, 10, 16. [Google Scholar] [CrossRef] [Green Version]
  17. Paul, S.; Kumar, D.N. Spectral-spatial classification of hyperspectral data with mutual information based segmented stacked autoencoder approach. ISPRS J. Photogramm. Remote Sens. 2018, 138, 265–280. [Google Scholar] [CrossRef]
  18. Lakhal, M.I.; Çevikalp, H.; Escalera, S.; Ofli, F. Recurrent neural networks for remote sensing image classification. IET Comput. Vis. 2018, 12, 1040–1045. [Google Scholar] [CrossRef] [Green Version]
  19. Liu, B.; Yu, X.; Zhang, P.; Yu, A.; Fu, Q.; Wei, X. Supervised Deep Feature Extraction for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 56, 1909–1921. [Google Scholar] [CrossRef]
  20. Jiao, L.; Liang, M.; Chen, H.; Yang, S.; Liu, H.; Cao, X. Deep Fully Convolutional Network-Based Spatial Distribution Prediction for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5585–5599. [Google Scholar] [CrossRef]
  21. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Li, J. Hyperspectral Image Classification Using Random Occlusion Data Augmentation. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1751–1755. [Google Scholar] [CrossRef]
  22. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  23. Gong, Z.; Zhong, P.; Yu, Y.; Hu, W.; Li, S.A. Cnn with multiscale convolution and diversified metric for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3599–3618. [Google Scholar] [CrossRef]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  25. Zhang, C.; Li, G.; Du, S.; Tan, W.; Gao, F. Three-dimensional densely connected convolutional network for hyperspectral remote sensing image classification. J. Appl. Remote Sens. 2019, 13, 016519. [Google Scholar] [CrossRef]
  26. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  27. Qu, Y.; Baghbaderani, R.K.; Qi, H. Few-Shot Hyperspectral Image Classification Through Multitask Transfer Learning. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  28. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  29. Masarczyk, W.; Glomb, P.; Grabowski, B.; Ostaszewski, M. Effective transfer learning for hyperspectral image classification with deep convolutional neural networks. arXiv 2019, arXiv:1909.05507. [Google Scholar]
  30. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2016, 55, 844–853. [Google Scholar] [CrossRef]
  31. Wei, W.; Zhang, J.; Zhang, L.; Tian, C.; Zhang, Y. Deep Cube-Pair Network for Hyperspectral Imagery Classification. Remote Sens. 2018, 10, 783. [Google Scholar] [CrossRef] [Green Version]
  32. Li, W.; Chen, C.; Zhang, M.; Li, H.; Du, Q. Data Augmentation for Hyperspectral Image Classification with Deep CNN. IEEE Geosci. Remote Sens. Lett. 2018, 16, 593–597. [Google Scholar] [CrossRef]
  33. Kipf, T.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2017, arXiv:1609.02907. [Google Scholar]
  34. Sakai, T.; Plessis MC du Niu, G.; Sugiyama, M. Semi-supervised classification based on classification from positive and unlabeled data. In Proceedings of the ICML’17 Proceedings of the 34th International Conference on Machine Learning-Volume 70, Sydney, NSW, Australia, 6–11 August 2017; pp. 2998–3006. [Google Scholar]
  35. Liao, R.; Brockschmidt, M.; Tarlow, D.; Gaunt, A.; Urtasun, R.; Zemel, R. Graph Partition Neural Networks for Semi-Supervised Classification. arXiv 2018, arXiv:1803.06272. [Google Scholar]
  36. Tan, K.; Hu, J.; Li, J.; Du, P. A novel semi-supervised hyperspectral image classification approach based on spatial neighborhood information and classifier combination. ISPRS J. Photogramm. Remote Sens. 2015, 105, 19–29. [Google Scholar] [CrossRef]
  37. Yang, L.; Yang, S.; Jin, P.; Zhang, R. Semi-Supervised Hyperspectral Image Classification Using Spatio-Spectral Laplacian Support Vector Machine. IEEE Geosci. Remote Sens. Lett. 2013, 11, 651–655. [Google Scholar] [CrossRef]
  38. Liu, B.; Yu, X.; Zhang, P.; Tan, X.; Yu, A.; Xue, Z. A semi-supervised convolutional neural network for hyperspectral image classification. Remote Sens. Lett. 2017, 8, 839–848. [Google Scholar] [CrossRef]
  39. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  40. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier GANs. In Proceedings of the ICML’17 Proceedings of the 34th International Conference on Machine Learning-Volume 70, Sydney, NSW, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  41. Rasmus, A.; Valpola, H.; Honkala, M.; Berglund, M.; Raiko, T. Semi-supervised learning with Ladder networks. In Proceedings of the NIPS’15 the 28th International Conference on Neural Information Processing Systems-Volume 2, Montreal, QC, Canada, 7–12 December 2015; Volume 28, pp. 3546–3554. [Google Scholar]
  42. Zhang, J.; Wei, W.; Zhang, L.; Zhang, Y. Improving Hyperspectral Image Classification with Unsupervised Knowledge Learning. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 June–2 August 2019; pp. 2722–2725. [Google Scholar]
  43. Hinton, G.E.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  44. Aviris, N. Indianas indian pines 1992 data, s.e.t. 2012. [Google Scholar]
Figure 1. Pipeline of the proposed two-branch few-shot deep learning framework for HSI classification.
Figure 1. Pipeline of the proposed two-branch few-shot deep learning framework for HSI classification.
Remotesensing 13 03539 g001
Figure 2. Illustration of the adopted 3D-CNN architecture for the feature extractor sub-network, together with parameter settings (such as dimension of feature (f), convolution kernels (k) and stride (s)) for all convolutional layers and feature maps.
Figure 2. Illustration of the adopted 3D-CNN architecture for the feature extractor sub-network, together with parameter settings (such as dimension of feature (f), convolution kernels (k) and stride (s)) for all convolutional layers and feature maps.
Remotesensing 13 03539 g002
Figure 3. Illustrations of the adopted SSRN architecture for the feature extractor sub-network, together with parameter settings (such as dimension of feature (f), convolution kernels (k) and stride (s)) for all layers and feature maps. (a) Spectral residual learning block, (b) spatial residual learning block. The two residual learning blocks contain shortcut connections between the two convolutional layers.
Figure 3. Illustrations of the adopted SSRN architecture for the feature extractor sub-network, together with parameter settings (such as dimension of feature (f), convolution kernels (k) and stride (s)) for all layers and feature maps. (a) Spectral residual learning block, (b) spatial residual learning block. The two residual learning blocks contain shortcut connections between the two convolutional layers.
Remotesensing 13 03539 g003
Figure 4. The false color images and ground truth map of the PaviaU dataset. The original image size is 610 × 340 pixels.
Figure 4. The false color images and ground truth map of the PaviaU dataset. The original image size is 610 × 340 pixels.
Remotesensing 13 03539 g004
Figure 5. The false color image and ground truth map of the Salinas dataset. The original image size is 512 × 217 pixels.
Figure 5. The false color image and ground truth map of the Salinas dataset. The original image size is 512 × 217 pixels.
Remotesensing 13 03539 g005
Figure 6. The false color image and ground truth map of the Indian Pines dataset. The original image size is 145 × 145 pixels.
Figure 6. The false color image and ground truth map of the Indian Pines dataset. The original image size is 145 × 145 pixels.
Remotesensing 13 03539 g006
Figure 7. The classification maps of all methods with three samples on PaviaU dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 7. The classification maps of all methods with three samples on PaviaU dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g007
Figure 8. The classification maps of all methods with five samples on PaviaU dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 8. The classification maps of all methods with five samples on PaviaU dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g008aRemotesensing 13 03539 g008b
Figure 9. The classification maps of all methods with three samples on Salinas dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 9. The classification maps of all methods with three samples on Salinas dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g009aRemotesensing 13 03539 g009b
Figure 10. The classification maps of all methods with five samples on Salinas dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 10. The classification maps of all methods with five samples on Salinas dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g010aRemotesensing 13 03539 g010b
Figure 11. The classification maps of all methods with three samples on Indian Pines dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 11. The classification maps of all methods with three samples on Indian Pines dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g011
Figure 12. The classification maps of all methods with five samples on Indian Pines dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 12. The classification maps of all methods with five samples on Indian Pines dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g012
Figure 13. The classification maps of all methods with one sample on PaviaU dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 13. The classification maps of all methods with one sample on PaviaU dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g013
Figure 14. The classification maps of all methods with one sample on Salinas dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s -means method, soft label with the proposed method. The experiments were conducted with five training samples per class on the Indian Pines and PaviaU dataset.
Figure 14. The classification maps of all methods with one sample on Salinas dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s -means method, soft label with the proposed method. The experiments were conducted with five training samples per class on the Indian Pines and PaviaU dataset.
Remotesensing 13 03539 g014
Figure 15. The classification maps of all methods with one sample on Indian Pines dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Figure 15. The classification maps of all methods with one sample on Indian Pines dataset. (a) SVM; (b) SS-LapSVM; (c) 3D-DENSEnet; (d) 3D-Gan; (e) SS-CNN; (f) 3D-CNN; (g) 3 D - CN N k - m e a n s (h) 3 D - CN N o u r s ; (i) 3 D - CN N h a r d l a b e l ; (j) SSRN; (k) SSR N k - m e a n s ; (l) SSR N o u r s ; (m) SSR N h a r d l a b e l .
Remotesensing 13 03539 g015
Figure 16. Illustration of the sensitivity of λ . It reflects the classification accuracy of OA, AA and k with different values of λ . (a) OA; (b) AA; (c) k .
Figure 16. Illustration of the sensitivity of λ . It reflects the classification accuracy of OA, AA and k with different values of λ . (a) OA; (b) AA; (c) k .
Remotesensing 13 03539 g016
Table 1. The name and number of samples per class on the Indian Pines dataset, PaviaU dataset, and Salinas dataset respectively.
Table 1. The name and number of samples per class on the Indian Pines dataset, PaviaU dataset, and Salinas dataset respectively.
No.PaviaUSalinasIndian Pines
Class NameNumberClass NameNumberClass NameNumber
1Asphalt6631Weeds 12009Alfafa46
2Meadows18,649Weeds 23726Corn-notill1428
3Gravel2099Fallow1976Corn-mintill830
4Trees3064Fallow rough piow1394Corn237
5Sheets1345Fallow smooth2678Grass-pasture483
6Bare Soil5029Stubble3959Grass-tress730
7Bitumen1330Celery3579Grass-pasture-mowed28
8Bricks3682Grapes11,271Hay-windrowed478
9Shadows947Soil6203Oats20
10 Corn3278Soybean-notill972
11 Lecture 4 wk1068Soybean-mintill2455
12 Lecture 5 wk1927Soybean-clean593
13 Lecture 6 wk916Wheat205
14 Lecture 7 wk1070Woods1265
15 Vinyard untrained7268Buildings-Grass386
16 Vinyard trellis1807Stone-Steel-Towers93
sum42,77654,12910,249
Table 2. Classification accuracy (%) for all methods with three training samples on the PaviaU dataset. Best results are in bold. The numbers in parentheses are standard variances (%).
Table 2. Classification accuracy (%) for all methods with three training samples on the PaviaU dataset. Best results are in bold. The numbers in parentheses are standard variances (%).
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
162.1066.6925.3975.5483.2759.0457.2070.3050.6760.8466.2062.2565.64
253.2427.7468.6047.8169.6064.6676.2066.1467.5264.7251.3860.9762.28
356.7544.2427.0449.7424.9255.9550.0348.1454.4954.5352.6551.2146.33
475.3573.7663.7767.0921.7473.3750.6373.3069.5550.9077.4871.4272.72
597.0885.9097.5286.3267.2499.0296.6198.5398.8294.3998.8895.2997.28
649.4161.3329.4558.2032.9241.4129.6548.4837.8148.8347.2149.4648.11
788.7160.4511.9457.5321.5579.1369.6575.2085.6568.5982.1684.4886.53
849.6056.6421.8557.1049.8562.5951.3257.1755.5933.7059.3761.7661.27
999.8085.9892.0478.5134.0396.7992.4494.1096.6387.8299.8898.0198.63
OA59.15(3.80)48.46(8.65)50.56(6.25)46.65(3.08)39.92(6.38)63.31(5.77)63.32(2.61)65.48(3.18)62.08(3.30)59.65(6.24)59.32(4.68)63.09(4.60)63.67(4.56)
AA70.23(2.37)62.53(8.93)48.62(4.17)22.14(9.10)45.01(6.00)70.22(4.02)63.75(2.35)70.15(3.73)68.52(2.16)62.70(6.97)70.58(1.80)70.76(3.31)70.98(2.98)
k 49.69(4.07)39.57(8.63)33.15(7.37)12.40(9.76)31.85(5.67)53.74(6.27)51.95(2.78)56.28(3.60)51.76(3.88)48.82(7.25)50.09(4.31)53.74(5.38)54.35(5.42)
Table 3. Classification accuracy (%) for all methods with five training samples on the PaviaU dataset.
Table 3. Classification accuracy (%) for all methods with five training samples on the PaviaU dataset.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
164.3967.1418.8865.7684.9765.1763.5374.7464.0167.4071.1272.8971.13
250.3140.5444.6048.6084.6156.9969.5065.3762.6166.1657.3563.3863.75
362.5654.8916.1353.7228.0961.6534.8961.1755.1941.5155.6756.0253.85
488.4782.5242.8654.2344.3586.3378.6185.7380.0785.9082.8592.0882.42
598.6098.8856.1079.7697.3899.0599.3799.8199.0496.8099.9099.2498.60
663.8869.4614.7652.2340.3747.4435.3850.9552.6349.8252.3655.2352.93
785.0086.2419.2766.9827.6781.5562.1776.0682.4573.6080.4485.5786.29
867.7553.6917.4168.9862.3771.1468.0755.7260.1560.4469.4864.6666.14
999.8299.9156.5776.7951.2897.8896.0092.2697.6896.1799.8798.7899.13
OA62.68(5.81)57.48(8.50)55.13(10.07)46.89(3.47)56.61(6.79)63.66(5.26)64.69(2.27)67.55(3.75)64.86(3.61)65.98(3.86)64.67(5.47)68.20(3.79)67.25(4.31)
AA75.64(2.38)72.59(7.98)53.07(4.42)23.43(11.06)57.90(3.86)74.13(3.17)67.50(2.20)73.53(3.26)72.65(1.88)70.85(2.87)74.34(1.67)76.09(1.91)74.92(1.88)
k 54.46(6.09)48.94(9.59)40.73(11.18)13.23(12.22)48.20(6.45)54.92(5.57)54.71(1.93)59.00(4.12)55.86(3.67)56.92(3.89)56.14(5.55)60.09(3.89)58.78(4.12)
Table 4. Classification accuracy (%) for all methods with three training samples on the Salinas dataset.
Table 4. Classification accuracy (%) for all methods with three training samples on the Salinas dataset.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
195.0395.3065.8158.8620.1694.6494.5993.9594.8276.0995.5794.4095.15
287.2479.6844.3262.3470.5895.9496.2097.2596.5066.1388.2593.3291.97
369.9566.3940.1635.8847.5257.4851.6966.1163.1545.6665,5172.8165.16
499.0597.3575.9536.3363.2398.3193.2198.8298.9284.3699.2698.3498.64
590.0287.7660.5572.2479.2587.5284.1388.4087.8277.8490.0687.7485.14
683.7996.5493.1581.7699.5198.4898.2798.6798.4994.4496.9698.0098.29
795.5385.3262.2484.3190.7495.5593.2893.8792.6076.4684.2290.7391.29
871.3152.8841.5150.14665.3359.0364.1969.3565.4958.0455.1857.2563.73
995.5387.2962.6635.8378.9295.6793.5894.8695.5379.5794.4792.9694.48
1071.3161.5739.4241.4859.7465.2965.4673.0473.5246.8955.0662.7056.59
1183.5563.5453.8348.2231.9985.5486.8784.2082.5168.7980.5380.5280.06
1294.7988.4162.4368.2867.8483.8679.7894.5298.7962.0896.6994.1997.81
1398.0895.0869.9337.8699.0894.8491.7693.7397.2783.5992.1295.5495.96
1488.1576.5164.8937.4181.3188.3083.2592.5992.4866.6388.9393.7291.24
1564.0554.4950.8444.3043.0266.1962.4264.8966.8238.7459.1466.1859.56
1676.2972.3651.6073.7774.5370.9966.6076.4772.6948.2053.9765.2862.71
OA78.33(2.25)72.99(3.79)55.35(7.81)31.05(3.42)63.06(4.75)79.08(3.26)78.35(3.26)82.38(2.49)81.72(3.12)64.18(7.62)75.51(2.97)78.47(3.14)78.32(3.19)
AA84.70(1.36)78.78(4.23)58.70(8.06)37.01(4.05)66.96(3.27)83.60(2.18)81.58(2.07)86.29(1.18)86.09(2.13)67.10(10.84)81.04(2.60)83.98(2.08)82.99(2.59)
k 75.93(2.41)70.13(4.09)50.90(8.42)24.58(3.25)59.02(4.76)76.82(3.57)75.99(3.57)80.42(2.71)79.75(3.41)60.44(8.49)72.92(3.21)76.18(3.39)75.95(3.52)
Table 5. Classification accuracy (%) for all methods with five training samples on the Salinas dataset.
Table 5. Classification accuracy (%) for all methods with five training samples on the Salinas dataset.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
194.3297.5781.9875.2946.5393.8693.7894.4294.7990.3894.6196.1995.11
294.2491.1754.5850.2886.7197.4697.3397.7396.3187.6994.5295.8496.86
386.7686.0341.4043.9677.7485.2976.3189.9789.7077.2284.5690.8086.06
499.3798.3986.2045.1175.9798.7898.1598.9598.9694.7999.4599.1699.30
589.5690.6768.1068.8289.4385.7583.3885.0285.3283.9492.2187.6886.55
698.6296.5096.1188.6799.7897.4697.1398.3998.1296.4298.0891.3598.12
795.1294.6987.5482.9692.8991.5090.9393.8291.7487.9290.5390.4991.48
857.5554.5647.8655.6566.2366.2768.0765.2466.8960.5061.1764.5265.68
994.2494.8881.2335.6793.0294.9692.2994.8794.6194.9593.3493.7693.34
1072.8367.1337.1743.7687.6974.8171.7476.8175.9760.1963.0970.0669.04
1190.4090.6654.0469.9263.2889.1788.6588.4390.3885.6486.1191.1389.67
1298.2998.1488.4665.9476.6094.3696.5596.7495.7785.4893.6294.8495.72
1397.5197.4788.2534.9395.5793.0691.3993.1993.6894.9790.9694.5294.68
1489.2888.2781.8237.2895.5191.4886.6990.3992.7279.7586.3994.3793.90
1566.8760.0948.1743.2846.2664.6064.0167.1365.9158.9057.2763.9865.43
1679.9777.8859.3976.2897.4374.3068.5476.3777.6162.9261.3979.7675.18
OA81.23(1.64)79.02(4.32)64.04(6.07)41.99(5.17)72.51(3.82)82.13(2.91)81.15(2.08)82.90(2.28)82.85(1.77)76.72(2.45)78.55(1.26)81.78(2.84)81.88(2.56)
AA87.75(1.20)86.51(1.82)68.92(6.18)49.58(5.75)80.88(3.52)87.07(1.67)85.31(1.80)87.98(1.76)88.03(1.46)81.37(3.68)84.08(1.87)87.78(1.68)87.26(2.22)
k 79.08(1.78)76.76(4.74)60.30(6.60)36.81(5.27)69.42(4.19)80.20(3.13)79.12(2.26)81.04(2.48)80.98(1.94)74.25(2.75)76.23(1.38)79.81(3.11)79.90(2.84)
Table 6. Classification accuracy (%) for all methods with three training samples on the Indian Pines dataset.
Table 6. Classification accuracy (%) for all methods with three training samples on the Indian Pines dataset.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
176.3065.8751.7420.5636.5248.6047.3344.6558.8431.8663.0265.5868.60
226.8816.5122.8445.7838.8431.2233.7831.9431.6014.1330.6727.3530.26
325.9823.8018.3564.7128.2736.3739.6735.1135.9020.0223.7625.7423.35
435.7029.9623.8834.1421.2350.5146.5349.9651.1128.6342.2736.7539.91
560.7711.5722.1345.7641.7856.3147.8046.8556.696.5436.8146.8150.81
657.6723.9577.7842.2270.7263.1861.8061.8861.2919.5658.0155.7162.68
789.6492.5089.645.9515.6976.0077.7883.6083.2054.0071.2074.4080.40
850.7566.8679.3980.1487.9880.7277.5382.3279.4382.5983.2876.7476.21
982.5072.0094.5013.5514.6294.1291.5893.5391.1884.7190.0099.4197.65
1044.7331.1639.8440.8525.3165.7564.2663.0260.9433.9545.4649.8546.35
1141.1015.0646.6168.0350.1747.4844.2348.2549.9656.1344.7149.8344.84
1224.6216.4618.6744.9421.1126.6827.5726.1227.8613.9220.3722.9523.86
1382.9383.3262.9867.7849.6686.5382.7084.8086.7847.0887.5777.5286.53
1463.2173.7386.4880.2875.5367.3976.6174.8567.3588.9571.0971.1770.47
1522.8211.5027.8037.0550.4936.6838.2340.2638.5915.3822.4836.4524.28
1688.7189.3560.7526.6277.0199.0098.7099.5698.8983.1187.5689.5686.11
OA43.60(3.84)30.50(5.48)45.68(2.81)39.82(3.28)43.33(4.28)51.43(2.68)51.59(1.77)51.88(2.38)51.62(2.73)40.98(2.48)45.90(2.86)47.66(2.68)46.76(3.29)
AA54.64(2.20)45.22(3.20)51.45(1.77)58.75(2.37)44.06(1.68)60.41(2.08)59.76(21.80)60.42(2.35)61.23(1.93)42.54(5.84)54.89(2.00)56.61(3.55)57.02(2.86)
k 36.86(4.17)23.74(5.35)38.66(2.74)34.41(2.97)36.69(4.10)45.80(3.05)46.11(1.98)46.21(2.78)45.99(2.94)32.49(2.72)39.37(3.16)41.24(2.80)40.29(3.32)
Table 7. Classification accuracy (%) for all methods with five training samples on the Indian Pines dataset.
Table 7. Classification accuracy (%) for all methods with five training samples on the Indian Pines dataset.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
186.0974.5762.2119.9944.1362.4362.2863.6166.3454.1574.3986.1073.90
230.5129.2734.8754.9046.6338.8339.8037.4235.3819.9624.3132.0329.01
339.2724.7719.5875.4133.3244.2538.8640.3541.1420.0124.7334.7029.15
456.2962.3625.1953.9823.0561.6455.1360.5259.7446.4761.9050.6546.38
564.7662.4435.7159.4557.7761.0354.8757.2657.4915.7543.0357.5960.59
676.0571.4874.9349.8678.9873.3069.7274.0474.2247.2365.9072.1974.36
789.6492.8693.937.3322.7891.7491.7493.4894.7881.3089.5786.9694.35
863.4973.3180.9690.4093.8581.1280.0680.5979.2672.9086.2674.9984.33
978.0074.0098.0017.0915.7790.0088.6791.3392.0083.3387.3399.3393.33
1038.7440.7531.4162.0333.4758.3165.9461.2157.6120.9442.2144.4142.37
1138.5515.2746.1865.9556.3247.5748.4947.1551.5954.1244.8545.7849.69
1230.4623.8122.4163.2925.0837.3936.9940.0235.949.1027.4428.4929.76
1390.3492.0065.2764.7860.3990.3088.6092.6092.8066.6582.2089.5587.10
1464.8568.1383.1582.5181.2272.5475.8677.0573.8287.7872.1476.3374.19
1532.3839.0721.5350.2463.7841.2941.9446.9050.3427.5927.6641.4434.70
1688.4991.5165.3842.0988.3398.7597.9599.8998.9891.8290.6894.4390.91
OA47.83(2.64)41.47(5.96)46.85(4.31)49.52(6.57)51.73(3.12)55.23(4.34)55.50(3.03)55.70(3.72)55.65(3.89)43.07(3.55)46.98(3.68)50.97(3.50)50.92(3.07)
AA60.50(2.01)58.48(2.56)53.80(3.40)64.89(7.05)51.54(1.98)65.67(2.23)64.83(3.33)66.45(1.74)66.34(2.09)49.94(6.16)59.05(2.62)63.44(2.43)62.13(2.48)
k 41.53(2.71)35.24(5.70)39.81(4.57)44.15(6.72)45.28(2.75)50.01(4.77)50.24(3.40)50.48(4.05)50.35(4.18)35.00(3.58)40.49(3.66)44.98(3.67)44.78(3.05)
Table 8. Classification accuracy (%) for all methods with one training samples on the PaviaU dataset.
Table 8. Classification accuracy (%) for all methods with one training samples on the PaviaU dataset.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
155.397.1344.0560.1664.3759.9659.0854.8135.8248.0020.2153.0958.69
239.2516.9161.7644.7866.3750.1065.77664.2747.0933.4632.7349.8940.35
324.699.2523.6485.2520.1118.2316.9121.3225.8822.6115.7929.5930.07
488.5655.2354.3890.0018.6370.2560.6058.4574.7888.7193.5986.7084.25
579.435.3070.5694.8041.0483.9794.3493.3598.4288.7699.4095.2696.09
642.5215.7722.1364.4038.4832.8414.0726.3940.0433.9142.1334.1643.01
779.8822.779.1882.7314.5729.6717.3152.7275.0565.7363.3352.7772.02
868.8216.1854.6255.7339.2835.5133.2434.3754.3151.4482.8354.8156.37
999.858.9565.1971.4219.3875.4072.2366.5696.8583.7999.8497.7397.84
OA51.38(11.07)18.71(6.59)50.53(4.38)43.07(4.92)29.76(9.36)49.21(6.79)52.62(7.20)53.86(7.98)49.66(9.27)44.59(9.98)48.92(8.30)53.17(8.62)51.53(9.64)
AA64.27(4.77)22.80(5.27)45.06(4.26)14.70(3.97)35.41(4.30)50.66(15.80)48.17(8.10)52.47(7.93)60.91(3.51)57.38(6.28)64.43(4.18)61.55(4.74)64.30(2.93)
k 42.21(10.75)9.85(4.64)31.16(3.87)4.04(4.83)22.07(6.59)35.65(10.66)38.39(5.40)39.97(9.21)39.05(7.89)34.99(8.37)39.91(7.85)42.89(7.82)42.08(8.84)
Table 9. Classification accuracy (%) for all methods with one training sample on the Salinas dataset. The symbol ‘-’ means the method failed to classify.
Table 9. Classification accuracy (%) for all methods with one training sample on the Salinas dataset. The symbol ‘-’ means the method failed to classify.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
197.16-46.5548.0323.1094.9394.9193.9595.3258.5695.0496.2296.61
264.30-45.0924.9111.0768.3971.9084.9978.8821.3680.4370.4472.38
354.28-17.0433.7928.6437.2238.3357.4153.5018.9247.3150.9554.04
498.63-54.4519.0033.4996.4095.8197.4297.8245.1899.4397.7098.52
589.30-45.9345.1551.2789.1668.5587.0385.7532.8491.0084.9686.42
696.09-81.9658.8097.4397.8897.4398.3898.4644.3995.9397.5697.70
790.96-48.2065.0649.3188.8677.6582.5993.8856.0684.4693.7492.56
850.50-54.6144.9754.0946.9554.3952.9454.0656.7758.3147.0544.74
992.37-60.2443.3459.4294.9286.5292.6093.6278.2092.3392.4192.75
1033.80-47.6641.8459.7932.1228.5034.5832.7010.4625.0931.0829.45
1174.19-19.7035.4912.9983.0577.4582.5183.0232.1778.8978.7982.07
1280.79-27.1247.319.3573.5259.2483.1084.7913.1287.9981.9783.10
1395.82-36.9121.43100.0089.3375.7591.8495.2866.6192.4893.3094.56
1478.85-39.6727.766.6083.0158.5484.8988.8420.4885.9390.9185.91
1544.36-33.8130.7419.6565.5353.4664.5957.3828.5937.0657.1559.88
1641.06-38.6957.6818.3447.8128.2552.0850.9014.3638.4443.5941.94
OA68.03(5.01)-42.42(7.08)25.06(2.92)37.57(7.75)69.92(5.90)65.15(6.30)72.86(3.99)72.48(4.12)42.44(6.78)68.91(5.47)69.58(4.31)69.66(4.35)
AA73.90(3.01)-43.60(7.24)27.51(6.33)39.28(4.84)74.32(4.80)66.67(6.77)77.56(2.99)77.76(2.48)37.38(12.73)74.38(3.59)75.49(2.52)75.79(2.61)
k 64.58(5.45)-36.95(6.96)16.43(5.22)32.53(6.85)32.53(6.85)61.48(6.84)69.99(4.33)69.99(4.42)36.00(7.74)65.54(5.98)66.40(4.59)66.50(4.62)
Table 10. Classification accuracy (%) for all methods with one training sample on the Indian Pines dataset. The symbol ‘-’ means the method is failed to classify.
Table 10. Classification accuracy (%) for all methods with one training sample on the Indian Pines dataset. The symbol ‘-’ means the method is failed to classify.
SVMDenseNetCNNResNet
SVMSS-LapSVMDenseNet3D-GanSS-CNN3D-CNN 3 D - C N N k - m e a n s 3 D - C N N o u r s 3 D - C N N h a r d l a b e l SSRN S S R N k - m e a n s S S R N o u r s S S R N h a r d l a b e l
173.04-17.5110.7718.6139.5629.1150.8946.4412.4443.7860.6757.11
224.09-7.3244.1130.0833.9733.3133.3431.5216.5720.7628.8428.09
324.23-28.5244.6113.9422.0121.3121.6420.495.6317.4216.4115.42
417.34-21.9036.8316.8135.0828.7728.7321.4010.9324.3619.4919.58
545.24-32.0250.2027.5740.2733.0137.4539.3210.3534.9836.8940.95
637.16-50.0148.6258.9646.0542.9547.8251.3914.6435.6743.7347.08
788.57-10.7113.2615.4361.1142.2263.3355.1930.7459.2674.4470.00
841.99-53.0058.6680.3171.9163.7566.2151.5149.9859.9644.5354.57
962.00-12.577.9911.0967.8945.2677.8975.7935.2657.3783.6883.68
1023.91-20.8637.2925.3536.0833.7936.3128.048.9120.5527.7723.25
1129.88-47.0151.5645.3235.3436.9239.6141.4765.7732.2631.5537.68
1219.09-34.1956.9021.7413.4314.8015.0514.972.3514.4918.7515.84
1370.34-42.1031.5032.9247.1153.5849.9554.417.9461.7243.2839.66
1439.57-55.3277.1165.5353.3158.7360.5357.1881.8066.0058.5056.06
1516.22-45.5137.6728.3322.2922.3925.1221.6610.8816.5724.0017.48
1667.31-42.6126.7554.6294.3586.2092.8389.4638.9158.5974.3562.72
OA31.24(2.67)-32.92(3.87)25.66(2.87)36.77(5.11)38.21(4.35)37.80(3.61)39.92(4.34)38.19(4.29)34.86(3.10)33.36(3.53)34.29(3.84)35.13(3.77)
AA42.50(4.74)-34.45(4.08)38.23(7.47)34.65(3.78)44.98(3.99)40.38(4.24)46.67(3.05)43.76(2.53)25.20(5.03)39.98(3.27)42.93(3.87)41.82(2.33)
k 24.15(2.40)-24.96(3.56)18.23(3.80)28.51(4.51)30.99(4.59)30.45(3.44)32.67(4.22)30.67(4.00)23.72(3.62)25.82(2.99)27.17(3.64)27.46(3.58)
Table 11. Classification accuracy of 3D-CNN with five training samples via average and ours on the PaviaU dataset respectively.
Table 11. Classification accuracy of 3D-CNN with five training samples via average and ours on the PaviaU dataset respectively.
MethodAverageOurs
Criterion
OA66.1067.55
AA73.3273.53
k 57.2459.00
Table 12. Classification accuracy (%) with five training samples of the Salinas dataset after using PCA to reserve 100 important bands.
Table 12. Classification accuracy (%) with five training samples of the Salinas dataset after using PCA to reserve 100 important bands.
Method3D-CNN 3 D - C N N o u r s
Criterion
OA23.9830.34
AA26.6434.19
k 17.8524.74
Table 13. Classification accuracy (%) of the proposed two-head network and the comparison one-head network. The experiment is conducted with one, three and five training samples per class on the PaviaU dataset.
Table 13. Classification accuracy (%) of the proposed two-head network and the comparison one-head network. The experiment is conducted with one, three and five training samples per class on the PaviaU dataset.
3D-CNNSSRN
1 Sample per Class3 Samples per Class5 Samples per Class1 Sample per Class3 Samples per Class5 Samples per Class
One-HeadTwo-HeadOne-HeadTwo-HeadOne-HeadTwo-HeadOne-HeadTwo-HeadOne-HeadTwo-HeadOne-HeadTwo-Head
132.2154.8147.8970.3063.5874.7426.5153.0943.6362.2563.2972.89
251.9664.2769.2366.1466.0465.3759.8449.8981.2460.9778.3563.38
322.6321.3249.7448.1452.1061.1714.4629.5926.3351.2142.9056.02
474.8658.4558.0273.3064.9885.7347.0686.7035.8871.4255.2492.08
584.1893.3591.8598.5393.7899.8154.4995.2656.6995.2980.1299.24
633.9826.3937.2148.4844.0850.9534.1034.1625.3649.4629.9755.23
764.0252.7283.7675.2080.1976.0647.6952.7756.7184.4871.9685.57
851.8034.3751.2057.1757.9455.7242.2654.8142.2161.7653.7264.66
985.1266.5695.2394.1095.7692.2666.2797.7365.7898.0196.3198.78
OA49.09(10.05)53.86(7.98)60.58(4.06)65.48(3.18)63.58(3.76)67.55(3.75)46.59(14.93)53.17(8.62)57.67(9.57)63.09(4.60)65.07(7.39)68.20(3.79)
AA55.64(15.24)52.47(7.93)64.90(2.66)70.15(3.73)68.72(3.54)73.53(3.26)43.63(21.70)61.55(4.74)48.20(24.35)70.76(3.31)63.54(14.03)76.09(1.91)
k 35.52(14.66)39.97(9.21)49.59(4.74)56.28(3.60)53.58(3.77)59.00(4.12)28.63(20.30)42.89(7.82)36.91(24.35)53.74(5.38)51.83(15.96)60.09(3.89)
Table 14. Classification accuracy (%) with hard label obtained from -means method, soft label with the proposed method. The experiments were conducted with five training samples per class on the Indian Pines and PaviaU dataset.
Table 14. Classification accuracy (%) with hard label obtained from -means method, soft label with the proposed method. The experiments were conducted with five training samples per class on the Indian Pines and PaviaU dataset.
Indian PinesPaviaU
3D-CNNSSRN3D-CNNSSRN
k - m e a n s h k - m e a n s s O u r s s k - m e a n s h k - m e a n s s O u r s s k - m e a n s h k - m e a n s s O u r s s k - m e a n s h k - m e a n s s O u r s s
OA55.5055.3755.7046.9841.6350.9764.6966.2167.5564.6767.5068.20
AA64.8365.9766.4559.0548.5263.4467.5071.0673.5374.3474.4576.09
k 50.2450.0750.4840.4934.4344.9854.7156.5059.0056.1459.2460.09
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ding, C.; Li, Y.; Wen, Y.; Zheng, M.; Zhang, L.; Wei, W.; Zhang, Y. Boosting Few-Shot Hyperspectral Image Classification Using Pseudo-Label Learning. Remote Sens. 2021, 13, 3539. https://doi.org/10.3390/rs13173539

AMA Style

Ding C, Li Y, Wen Y, Zheng M, Zhang L, Wei W, Zhang Y. Boosting Few-Shot Hyperspectral Image Classification Using Pseudo-Label Learning. Remote Sensing. 2021; 13(17):3539. https://doi.org/10.3390/rs13173539

Chicago/Turabian Style

Ding, Chen, Yu Li, Yue Wen, Mengmeng Zheng, Lei Zhang, Wei Wei, and Yanning Zhang. 2021. "Boosting Few-Shot Hyperspectral Image Classification Using Pseudo-Label Learning" Remote Sensing 13, no. 17: 3539. https://doi.org/10.3390/rs13173539

APA Style

Ding, C., Li, Y., Wen, Y., Zheng, M., Zhang, L., Wei, W., & Zhang, Y. (2021). Boosting Few-Shot Hyperspectral Image Classification Using Pseudo-Label Learning. Remote Sensing, 13(17), 3539. https://doi.org/10.3390/rs13173539

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop