Next Article in Journal
Improving Estimates and Change Detection of Forest Above-Ground Biomass Using Statistical Methods
Previous Article in Journal
A Convolutional Neural Network for Large-Scale Greenhouse Extraction from Satellite Images Considering Spatial Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AC-WGAN-GP: Generating Labeled Samples for Improving Hyperspectral Image Classification with Small-Samples

1
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
2
School of Mathematics and Statistics, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(19), 4910; https://doi.org/10.3390/rs14194910
Submission received: 23 August 2022 / Revised: 18 September 2022 / Accepted: 27 September 2022 / Published: 1 October 2022
(This article belongs to the Section AI Remote Sensing)

Abstract

:
The lack of labeled samples severely restricts the classification performance of deep learning on hyperspectral image classification. To solve this problem, Generative Adversarial Networks (GAN) are usually used for data augmentation. However, GAN have several problems with this task, such as the poor quality of the generated samples and an unstable training process. Thereby, knowing how to construct a GAN to generate high-quality hyperspectral training samples is meaningful for the small-sample classification task of hyperspectral data. In this paper, an Auxiliary Classifier based Wasserstein GAN with Gradient Penalty (AC-WGAN-GP) was proposed. The framework includes AC-WGAN-GP, an online generation mechanism, and a sample selection algorithm. The proposed method has the following distinctive advantages. First, the input of the generator is guided by prior knowledge and a separate classifier is introduced to the architecture of AC-WGAN-GP to produce reliable labels. Second, an online generation mechanism ensures the diversity of generated samples. Third, generated samples that are similar to real data are selected. Experiments on three public hyperspectral datasets show that the generated samples follow the same distribution as the real samples and have enough diversity, which effectively expands the training set. Compared to other competitive methods, the proposed framework achieved better classification accuracy with a small number of labeled samples.

1. Introduction

Hyperspectral imaging sensors can capture more spectral data than just visible light. The data takes the form of continuous spectral features and can be used for the accurate identification of a variety of surface materials on planet earth [1]. With the development of hyperspectral sensors, the spatial and spectral resolution of the collected hyperspectral images (HSI) are increasing higher and higher [2]. Therefore, hyperspectral remote sensing finds important applications in many fields [3,4], including mining [5], astronomy [6], agriculture [7], environmental science [8,9], wasteland fire tracking, and biological threat detection [10]. HSI classification technology is an important content of hyperspectral remote sensing of earth observation technology. Its specific task is the classification of the objects represented by each pixel in HSI.
Early HSI classification mostly used traditional machine learning and statistical methods, such as K Nearest Neighbor (KNN) [11], support vector machine (SVM) [12], distance classifier [13], and naive Bayes classifier [14]. These methods rely on manually designed features. However, the characteristics of hyperspectral images such as high dimensionality and much spatial information make it difficult for a single traditional classification model to achieve good results. Some methods of dimension reduction are used, such as principal component analysis (PCA) [15], linear discriminant analysis [16], and band selection methods [17]. A space-spectral fusion method based on conditional random fields is proposed in [18]. In another method, the original spectral features and extracted spatial features through the Gabor filter bank form a space-spectral fusion feature [19]. The above methods use original hyperspectral samples for training. Those conventional models require a lot of prior knowledge and expert experience. Although manual features based on prior knowledge show good performance for some datasets, they are not related to the data and task themselves. So, their generalization ability is limited [20].
Recently, classifiers based on deep learning are widely used in HSI classification. Deep learning can extract deep features automatically, which is more convenient and effective. For example, the stack autoencoder network (SAE) [21] and the convolutional neural network (CNN) [22] are used to extract spectral and spatial features. 1D-CNN only focuses on hyperspectral vectors, but good classification accuracy has been obtained [23]. 3D-CNN performs convolution operations on three-dimensional HSI patch samples, considering both the spatial and spectral dimensions [24]. Pu et al. proposed a spatial and spectral convolutional neural network to extract spatial-spectral features [25]. Ding et al. combined spectral information and spatial coordinates to generate probability maps to fuse spectral and spatial information [26]. Very recently, transformer shows great potential in the field of computer vision [27]. L. Sun, G. Zhao, Y. Zheng, and Z. Wu used CNN to obtain HSI feature maps, then the serialized feature map is fed into the transformer module [28]. The deep features are extracted adaptively according to different training sets. Thus, classifiers based on deep features are generally better than conventional ones [20]. However, they always require a great number of labeled samples to optimize the parameters, labelling samples are expensive in the field of hyperspectral image processing. Thus, most hyperspectral classification is performed with small sample sizes.
Hyperspectral image classification with small sample sizes involves semi-supervised learning, self-supervised learning, and sample augmentation methods. Both semi-supervised learning and self-supervised learning try to mine information from abundant unlabelled samples. DAE-GCN [29] propose a deep autoencoder model to extract relevant features from the HSI and constructs a spectral-spatial graph to train a semi-supervised graph convolutional network. L. Zhao, W. Luo, Q. Liao, S. Chen, and J [30] introduced a module to generate HSI sample pairs and used the available samples for training a self-supervised learning model based on a Siamese network. Then labeled samples are used to fine-tune the model. Li T. et al. [31] proposed a dual-branch residual neural network. A self-supervised learning pre-training method with the idea of recovering intermediate unlabelled pixel information through artificially divided image cube samples is designed. Then pre-trained weights and a few labeled samples are used for the training classifier. However, the semi-supervised and self-supervised learning methods are based on existing enough unlabeled samples. In fact, for some hyperspectral classifiers, unlabeled samples are also insufficient. So, in order to increase the number of samples, many data augmentation methods are proposed. Early sample augmentation methods [32] generate new samples by rotating, adding noise, and linearly combining existing original samples, assuming that these newly generated samples and original samples share the same label. Wang C. et al. established a data mixture model to augment the labeled training set quadratically [33]. The paradigm of conventional data augmentation methods is relatively fixed. And it cannot guarantee that the generated samples conform to the correct distribution and provide useful information.
Nowadays, many scholars use generative adversarial networks to learn the implicit distribution function of the base samples, and then produce new samples with the same distribution by random sampling. Traditional GAN networks [34] only generate samples of the same class, but cannot generate samples of multiple classes, resulting in that the generated samples cannot improve the classification accuracy effectively. To generate labeled samples, Odena et al. proposed the Auxiliary Classifier GAN (AC-GAN) [35]. As an approach to solving the multi-classification problem, its discriminator is used to output corresponding label probabilities, and each generated sample has a corresponding class label. In practice, both the truthfulness of the data and the likelihood of correct classification are taken into account by the objective function of the discriminator. The classification branch is added to alleviate the HSI classification problem using limited training samples. Based on these ideas, Y. Zhan et al. designed a GAN network with a one-dimensional spectral structure [36], trained a one-dimensional GAN network with unlabeled samples, and then converted the trained discriminator into a classification network. Chen et al. designed a 1D and 3D GAN network for HSI classification, combining the generated samples with the real training samples into a new class [37], which was fine-tuned in the discriminator to improve the final classification performance. Multiscale conditional adversarial networks [38] use multiple scales and stages to achieve a coarse-to-fine fashion. This method can achieve high-quality data augmentation with a small number of training samples. J. Feng et al. combined a self-supervised classifier and GAN [39]. The pretext cluster task was designed by leveraging abundant unlabeled samples, then transferring the learned cluster representation from the cluster task. HyperViTGAN [40] is proposed to deal with the class imbalance problem of HSI data. It also involves an external semi-supervised classifier to share the task of the discriminator in GAN. In response to gradient disappearance and mode collapse, Liang et al. proposed an average minimization loss constrained by unlabeled data for HSI [41]. Gulrajani et al. introduced WGAN-GP into the network to make training smoother and more efficient [42]. We find that when GAN is used to generate hyperspectral data, inputs are often noise or noise-label, lacking guidance and constraint from prior knowledge. It leads to unstable and low-quality generated results on multi-category tasks. Besides, although the generator may fit the distribution of real data well, the network that has been trained with fixed parameters cannot generate data that satisfies the diversity under the same distribution. Additionally, some of the generated samples may be far from the real samples and not all the generated samples are helpful to the hyperspectral image classification.
In this article, we propose a new generative network named AC-WGAN-GP based on AC-GAN and WGAN-GP. The proposed framework utilizes guidance from prior knowledge and improves label reliability by a separate classifier. The online generation mechanism improves the diversity of generated sample sets. A selection algorithm based on KNN is presented to choose more reliable samples for training. The proposed framework can offer high-quality labeled samples with diversity and veracity to expand the training set. The contributions of this work can be summarized as follows.
  • We construct a new generative network named AC-WGAN-GP to generate labeled samples of different classes. We also design the new input of the generator including PCA features and category information are used to guide the process of generating ad noise to maintain the diversity of samples. Considering the task of generating multi-category labeled samples, we add a separate classifier to strengthen the difference between samples of different categories.
  • The online generation mechanism is studied profoundly. Instead of generating samples after the network has converged, the online generation mechanism makes AC-WGAN-GP periodically keep the generated samples during the training process, thereby significantly improving the diversity of the generated sample set.
  • A lightweight sample selection method is designed to efficiently select samples that are similar to real ones from the generated sample set. The function of the proposed algorithm also includes smoothing the label to reduce the error of using cross-entropy loss. Finally, the augmented training set is constructed by the generated samples and original real samples.
We organize the rest of this article as follows. Section 2 is used to review a series of GAN. The detailed introduction of the proposed method is presented in Section 3. Section 4 evaluates the proposed method and comparison with competing methods in this paper, Finally, Section 5 concludes the paper.

2. Related Work

In this section, we have a review of GAN, WGAN, and WGAN-GP and analyze their advantages and disadvantages.

2.1. GAN

The core idea of GAN originates from the celebrated Nash equilibrium of game theory. In GAN, the two players are set as one generator and one discriminator. While the generator strives to learn the distribution of real data, the discriminator aims to correctly determine whether the input data comes from real data or from the generated fake data. Each of these two players seeks to win the competitive game by constantly optimizing itself and improving its generation or discrimination ability. This is a learning optimization process and the goal is to attain Nash equilibrium between the two competing players.
For generator G , the random noise variable z (characterized by the distribution p ( z ) ) serves as input. The sample generated by G conforming as far as possible to the distribution p ( x ) of real data is characterized by G ( z ) . Meanwhile, the input sample to the discriminator D is labeled 1 if it is a real sample x , and it is labeled 0 if it is generated by G ( z ) . We reiterate that D aims to achieve the discrimination of the data sources, while G seeks to ensure consistency between the performance D ( G ( z ) ) of the sample G ( z ) on D and the performance D ( x ) of the real data x on D . The generator and discriminator are competing with each other and the iterative optimization process leads to a continuous improvement of the performance for each of them. Finally, the game approaches a state that the discrimination ability cannot be further improved. At this state the generator G might be safely assumed that it attains its goal of learning the distribution of the real data. From the above process, the task undertaken by the GAN is to address a minimax problem. The objective function of the GAN can be described as the minmax formulation:
min G   max D V ( θ D , θ G ) = E x p ( x ) [ log D ( x ) ] + E z p ( z ) [ log ( 1 D ( G ( z ) ) ) ]
where E is the mathematical expectation, x p ( x ) is the random variables that fit the probability distribution of real samples, z p ( z ) is the random variables that fit the probability distribution of generated samples, θ D , θ G are two parts of loss function from generator and discriminator. The GAN is trained via an alternating optimization method. We start training by fixing the generator G and optimizing the discriminator D . Then, we alternate by fixing the discriminator D and optimizing the generator G . The global optimal solution is reached when and only when p ( g ) = p ( x ) . The parameters of D are generally updated k times while those of G are updated only once when training.

2.2. WGAN and WGAN-GP

However, the problems undermining the utility of the traditional GAN include training difficulties, incapability of the optimization function of the generator, and the lack of diversity of the generated samples. Moreover, the GAN suffers the shortcoming of gradient disappearance when training is based on gradient descent. The Jensen–Shannao divergence measures the overlap between two distributions. The shortcoming arises when the common part between the real and generated sample distributions is diminished, and the Jensen–Shannon divergence of the objective function used by conventional GAN becomes a constant, thereby producing a discontinuity of the optimization objective. With an eye on handling the deficiency of a disappearing training gradient, Arjovsky et al. were the first to introduce the Wasserstein GAN (WGAN) [43] model, together with the associated concept of the Wasserstein distance proposed by Rubner, Y., Tomasi, C. and Guibas, L.J. [44] A mathematical transformation is used to convert the Wasserstein distance into a solvable form, and this distance can be approximated through adding a parameter’s numerically limited in discriminator network. Subject to the approximate optimal discriminator obtained, optimizing the generator to decrease the Wasserstein distance can effectively shorten the distance between the generated distribution p ( g ) and the distribution p ( x ) of real data. Mathematically, the objective function of WGAN is given by:
min G   max D E x p ( x ) [ D ( x ) ] + E x ˜ p ( g ) [ D ( x ˜ ) ]
where x ˜ is a random variable that fit the distribution of generated samples. Here D is the set of 1-Lipschitz functions, which needs to satisfy Lipchitz continuity so as to use the Wasserstein distance. Lipschitz continuity is a restriction on continuous functions, which requires that the derivative of the functions must be less than a constant K .
To satisfy this condition, I. Gulrajani [42] et al. limits the weight of D to a range, thereby introducing WGAN-GP as an improved version of WGAN. A new Lipchitz continuous restriction technique named gradient punishment is proposed to handle both problems of gradient disappearance and gradient explosion. Instead of directly constraining the value of the gradient, the authors added a regularization term. Detailed formulas and explanations are shown in the next section.

3. The Proposed Method

Figure 1 shows the overall framework for the HSI classification with a small sample size based on the AC-WGAN-GP, which is composed of four parts: the preprocessing based on Gaussian smoothing, the AC-WGAN-GP network, online sample generation, and sample selection algorithm based on KNN.

3.1. Smoothing-Based Preprocessing

The spectral vectors of neighboring pixels are assumed to be related because they are likely to be part of an image of a semantically homogeneous component.
This paper uses gaussian filter to exploit the neighboring information because gaussian filter has rotation invariance. Gaussian kernel has fixed parameters. Therefore, each pixel will perform a weighted calculation with the neighboring pixels through the Gaussian kernel. Gaussian filter weighted the sum of the pixels according to the spatial distance between the neighboring pixels and the central pixels to obtain smoothed HSI patches. As the data complexity increases, the structure of networks used for learning also needs to be designed to be more complex. The smoothed HSI patches contain some spatial information while discarding harmful information. We think that the generator and discriminator can be designed with a simpler structure because of the smoothed input patches. The Gaussian smoothing process can be considered a spatial feature extraction.
After the normalization of the original hyperspectral image, a patch is taken as the input of gaussian filtering for each pixel x m n , which can be expressed as x m n R M × N × H (where M is the width, N is the height, and H is the number of bands), and the smoothed image is:
x m n s m o o t h = i j x i j exp ( ( i , j ) ( m , n ) 2 / 2 σ 2 ) i j exp ( ( i , j ) ( m , n ) 2 / 2 σ 2 )
where x i j is the pixel in x R M × N × H , ( i , j ) is the spatial coordinate of the pixel, ( m , n ) is the spatial coordinate of the central pixel x m n . In practice, we restrict the sums to a distance of 3 σ from ( i , j ) , since pixels far from that have a negligible contribution. 3 σ is the window size and σ is adjusted when experimenting and selected from collection [1, 1.67, 2.33, 3, 3.67, 4.33, 5].

3.2. AC-WGAN-GP

In the proposed method, the AC-WGAN-GP is constructed based on part of the theories of AC-GAN and WGAN-GP. We design an input of the generator so that the generator can be guided from the manual feature and different labels. So, the difficulty of generating high quality samples in different categories decreases. We also design a new structure by adding an auxiliary classifier to form a separate classifier C . Accordingly, the new optimization function is also designed. Thereby, the output fake labels and the generated samples are in parallel. Finally, fake labels and fake samples can be generated simultaneously from a limited number of real training samples thus expanding the training set. The architecture of AC-WGAN-GP is shown in Figure 2.
A network used to generate multi-category samples in different distribution is difficult. Besides adding label information to the input of generator, we believe that the manual features also can be used as prior information. So, the input of generator G introduces the principal components of real sample pixel X r e a l as additional constraints to guide G . The vector of noise extends a number of dimensions to accommodate the principal components extracted from X r e a l and one-hot labels. With the above input, the generator can be guided to generate high-quality samples in a targeted manner. Furtherly, we employ the Principal component analysis (PCA) to reduce the dimensionality of the HSIs. Indian pines dataset has 200 spectral bands, Salinas dataset has 224 spectral bands, and KSC dataset has 176 spectral bands. We select the first 30 principal components (PC), expressed as X P C . So, the input of generator G includes three parts: Gaussian noise variable z ; one-hot coded class information c ; and 30-dimensional PC x P C of single pixel. On the basis of traditional Gaussian noise, prior knowledge of various labels and samples are added. The generated sample can be expressed as X f a k e = G ( z , c , x p c ) .
In our AC-WGAN-GP, an independent classifier C is used to output fake one-hot labels independently. Refs. [39,40] show the methods of combining an independent classifier with GAN. The task of the generator in AC-WGAN-GP is generating samples of different categories, which is difficult to control. So, we have to add the category-correct constraints to the objective function. To reinforce the function of generating samples from different categories, we separate a classifier C that uses part of the loss function of the generator. Besides, it is difficult for the discriminator to discriminate between the real and fake samples while outputting the category of the generated samples. The function of C includes two points. Firstly, C is responsible for generating the labels of fake samples efficiently. Secondly, when optimizing AC-WGAN-GP, the loss function of C could ensure the generated samples belong to the corresponding input category information, thereby increasing the gap between samples of different categories. A classifier based on CNN is suitable. The input of the classifier C is the same as the input of the discriminator. The classifier is trained by real samples and their labels and then predicts the labels of fake samples. The cross-entropy loss of C is formally unified with the loss functions of the generator and discriminator. Therefore, we consider AC-WGAN-GP as a whole network during training.
The input of discriminator D is composed of the real single pixel X r e a l , the generated fake sample X f a k e labeled as c, and the output of D is a probability distribution P ( S | X ) . The function of the discriminator is to judge the real or fake samples. The purpose of designing D and C is that each part of the AC-WGAN-GP is assigned a specific task.
In the above framework, referring to Deep Convolution Generative Adversarial Network (DCGAN) [45], G adopts the form of a fractionally-stride convolutional neural network, while D and C adopt the form of a standard CNN. The Batch norm layer is used in generator G and discriminator D to normalize features, which improves the training speed and makes the training more stable. The Leak-ReLU activation function, not the ReLU one, is used in the discriminator to prevent gradient sparsity, and no activation function is used at the last layer of the discriminator. The ReLU activation function is employed in the generator, while the tanh activation function is utilized solely in the output layer. Classifier C consists of a convolutional layer and a full-connection one, and finally connects a Softmax activation function to output the probability of classification. In the training process, according to the idea of fixing one part to train another part, G , D and C fix the parameters of two networks, optimize the remaining network parameters, and iterate alternately in this way to complete the whole training process. The specific network parameter design is shown in Section 4. The input to the generator is our designed vector, a noise-label-principal component vector. We can select the output of the generator to the parts we need, such as fake labels or fake pixel vectors.
After designing the architecture of the AC-WGAN-GP, we propose a suitable optimization function. The multi-classification problem can be realized by the Auxiliary Classifier GAN (AC-GAN), which was introduced by Odena et al. [35]. We treat its loss function as a prototype. The objective function contains the likelihood of data source L S and the likelihood of category L C . Updating the parameters of D is achieved by maximizing L S + L C , while updating the parameters of G is attained by maximizing L C L S .
L s = E [ log P ( S = r e a l | X r e a l ) ] + E [ log P ( S = f a k e | X f a k e ) ]
L c = E [ log P ( C = c | X r e a l ) ] + E [ log P ( C = c | X f a k e ) ]
The structure of L s , L c is similar and both use cross entropy loss. The two parts of L s are responsible for judging the real samples and fake samples respectively. S represents the judgement result of the discriminator. Discrimination can also be viewed as a binary classification problem. And the two parts of L c classify real and fake samples. C defines as labels the classifier output, and c represents the ground truth labels.
Based on above equations, the discriminator optimizes L D in the training process is obtained. We define x ^ as sampling uniformly along straight lines between pairs of points sampled from the data distribution p ( x ) and the generator distribution p ( g ) in sample space. x ^ p ( x ^ ) represents the random variable that fits the distribution of p ( x ^ ) . x ˜ is the random variable that fits the distribution of fake samples G ( z , c , x p c ) . is the gradient. The first and second part is responsible for judging whether the sample is real or fake. The third part is the gradient penalty. L D is showed in the following formula:
L D = E x ˜ p ( g ) [ D ( x ˜ ) ] E x p ( x ) [ D ( x ) ] + λ E x ^ p ( x ^ ) [ ( x ^ D ( x ^ ) 2 1 ) 2 ]
As a regular choice and part of the objective function of the generator, the class cross entropy loss is the objective functions of the classifier. As shown in the following formula:
L c = E x p ( x ) [ log P ( C = c | X r e a l ) ]
where C is defined as the labels of the network forward output, and c is the ground truth labels of training samples.
The objective function of the generator consists of two parts. One part comes from discriminator D , which ensures that the discriminator does not recognize the sample produced by the generator. The other part comes from classifier C , which ensures that the samples generated by the generator belong to the corresponding class c to the greatest extent. The second part reinforces the connection between category information and generated samples. Therefore, the sum of these two parts constitutes the objective function for the generator G is defined as:
L G = E x ˜ p ( g ) [ D ( x ˜ ) ] E [ log p ( C = c | X f a k e ) ]

3.3. Online Sample Generation

Generally, after GAN is trained, the parameters of each network are fixed, then the required input data generator is sent to the generator to generate false samples. Although the trained network can fit the distribution of training samples of different classes and input gaussian noise Z is also random, the parameters and other inputs of GAN are fixed, so the diversity of fake data generated by AC-WGAN-GP is poor. Therefore, we find it is appropriate to design a GAN-friendly mechanism named online sample generation.
Firstly, AC-WGAN-GP is trained by the above training methods described in Section 3.2, and then we observe the loss value of the network. After the network reaches a certain section of convergence, the online generation strategy starts. Although the network is not well-trained, and the loss curve is still trending downward, we can start to gather some of the faked samples and labels at one certain epoch. Then AC-WGAN-GP will continue to train to optimize network parameters. After a certain number of training epochs, AC-WGAN-GP will generate samples online again. Finally, the samples generated online in each time are collected. In order to ensure the diversity of samples and the balance between different classes, we can keep the abundant online generated samples which might include all categories of hyperspectral samples if the computer memory allows. In this way, we can also alleviate the pressure to tune the variables of the online generation mechanism. As shown in Figure 3, online sample generation can take advantage of plenty of models to generate samples with higher diversity. And the abundant generated fake samples and labels will be filtrated in the next part.

3.4. Sample Selection Algorithm Based on KNN

The online sample generation guarantees the diversity of generated samples, but due to GAN’s unique properties, it is not guaranteed that all generated samples have good quality and be similar to real samples. Therefore, for the samples ( X f a k e , C f a k e ) generated by GAN, an extra selection method is needed to select samples that are close to real samples thus can promote classification accuracy. The algorithm includes two steps: samples selection and label smoothing.
To ensure the total number of parameters and operation time of the entire framework are appropriate, a brief and effective method is demanded. Firstly, the total set of original real data is divided into a training set ( X t r a i n , C t r a i n ) and a temporary set X t . We use the clustering algorithm to divide the temporary set into N clusters. And then the central samples of each cluster and a part of the temporary set samples around central samples X t ( X t X t ) are randomly taken to form a mini set. The clustering algorithm we employ are K-MEANS clustering algorithm, DBSCAN clustering algorithm, EM algorithm, and Mean shift clustering algorithm. For each algorithm, we repeat the above operations, and finally take the union of the mini sets to form X u . In the generated sample set, KNN is used to select the closest sample to each test sample x u ( x u X u ). Finally, the selected generated samples constitute the ultimate generated sample set. The generated fake samples selected by this algorithm can be representative and similar with the real samples.
The samples in ( X f a k e , C f a k e ) originate from the overall generated distribution. Each sample is characterized by its own label, but, nevertheless, it might mix information from other classes. In deep learning, when calculating cross-entropy with one-hot coded labels, only the loss of the correct class (the class with the label coded as 1) is considered, but the loss of the wrong class (the class with the label coded as 0) is not considered. It may lead to overfitting, gradient explosion or gradient disappearance [46]. In order to distinguish the labels of generated samples and labels of real samples and reduce the negative impact of labelling errors when training, we use label smoothing regularization (LSR) [47] to consider the distribution of wrong classes. We conduct the LSR processing on the labels of fake samples set. C f a k e represents the generated one-hot label, ε [ 0 , 1 ] , K represents the number of classes. The smoothing process is shown as:
C f a k e = C f a k e ( 1 ε ) + ε K
After the above operations, smooth labels improve the generalization ability of the network.
After implementing the algorithm, we obtain the final fake hyperspectral samples and their corresponding labels to augment limited training datasets. At the same time, the quality and diversity of the generated samples are also guaranteed to a certain extent. Algorithm 1 summarizes selection of generated samples and label smoothing.
Algorithm 1. Samples selection and label smoothing
Input: Generated sample set G, dataset D, the number of categories N, clustering algorithm list L (K-MEANS, DBSCAN, EM and Mean shift), the hyperparameter represents the volume of random set M, the hyperparameter for KNN k
Output: selected samples and smoothing labels ( X f a k e , C f a k e )
Step1:Randomly divided D into training set A and temporary set B
Step2:Union the clustering set:
          For i = 0 , , 3 , do
s e t B i = L [ i ] ( B )
S e t   c e n t r a l   C = Θ i B i , l e n ( Θ i ) = N
S e t   r a n d o m   R = Φ i B i , l e n ( Φ i ) = M
Step3: Union   the   set :   B = Θ 1   Θ 2   Θ 3   Θ 4   Φ 1   Φ 2   Φ 3   Φ 4 ,   l e n ( B ) = n ,
Step4: On   the   G = { ( x 1 , c 1 ) , , ( x q , c q ) } ,
                G = K N N ( B ,   G ,   k )
          For c   i n   G , do
                 c = L S R ( c )
Step5: Output   G = { ( x 1 , c 1 ) , , ( x p , c p ) } ,

4. Experiments and Result Analysis

In this experiment, firstly we introduce basic information such as datasets, training samples, and evaluation metrics. Then detailed experimental settings are given. The third section analyzes the quality of pixels generated using proposed method. In next section, we study the effect of real and generated pixel ratios on the classification results and perform the ablation experiments of the final sample selection module. Finally, we compare AC-GAN-GP and CNN classifier with classifiers based on traditional methods, convolution and GAN, demonstrating the effectiveness of our proposed method on task of hyperspectral classification with small samples. We implemented AC-WGAN-GP with the tensorflow framework on a PC server with two NVIDIA GTX1080TI GPU and 22 GB memory. The average time for training AC-WGAN-GP, generating and classifying is 210 min and 15 s.

4.1. Hyperspectral Datasets

In this experiment, three popular datasets of hyperspectral images are used as experimental data named the Indian Pines dataset, the Salinas dataset, and the Kennedy Space Centre (KSC) dataset.
Indian Pines: The dataset of the Indian Pines comprises a hyperspectral image of agricultural and forest areas in India, which is collected by remote sensing equipment utilizing an airborne visible/infrared imaging spectrometer (AVIRIS). Note that the image in the dataset comprises 145 × 145 pixels, and its spatial resolution is 20 m/pixel. The Indian Pines dataset consists of 220 spectral bands, 20 of which severely water-absorbing bands are removed, and we conduct experiments on the remaining 200 spectral bands. For this dataset, we consider 16 classes to be of interest, excluding background pixels. Figure 4a shows the three-band false color image and the ground reference map of the Indian Pines image.
Salinas: Salinas data was also taken by the aforementioned AVIRIS imaging spectrometer, and it is an image of the Salinas valley, a prominent valley notable for being a highly productive region in California, the United States of America (USA) The image is with a spatial resolution of 3.7 m, and it originally had 224 bands, out of which we used only 204 bands for classification. We exclude 20 bands that could not be reflected by water, namely the 108th–112th, 154th–167th bands and the 224th band. The image is characterized by a size of 512 × 217 pixels, which are divided into 16 classes. Figure 4b shows the three-band false color image and ground reference map of the Salinas image.
KSC: The NASA’s AVIRIS equipment collected data over the Kennedy Space Centre (KSC) in the south-eastern state of Florida, United States of America almost a quarter of a century ago. This spectroradiometer equipment obtained 224 bands each characterized by a width of 10 nm and having a median wavelength ranging from 400 nm to 2500 nm. The image consists of 512 × 217 pixels and possesses a spatial resolution of 18 m/pixel. Only 176 bands are used for further analysis after all water absorbing bands and low SNR bands are eliminated. For classification purposes, we define 13 classes, which represent a variety of land cover types in the pertinent environment for the given site. Figure 4c shows the three-band false color image and the ground reference map of the KSC image.
For each of these three datasets, we divide the initial dataset into two subsets named a training set and a testing one. The training set is composed of 200 samples, and is extracted by random sampling. In other words, the proportion of training samples on the three datasets are 2.0%, 1.0%, 3.8%, which satisfies the requirement of few-shot experiments. The remaining samples are taken as the testing set. Table 1, Table 2 and Table 3 show the legend of each category. Table 4, Table 5 and Table 6 present the number of training samples selected from each dataset and the total number of samples.

4.2. Experimental Setting

The training/testing sample used in the experiment is a single pixel. Each pixel can be used as feature to train the AC-WGAN-GP and the CNN classifiers, as it corresponds to a unique label. The number of classes in the dataset is represented by n. All HSIs data are normalized between −1 and 1 at the beginning of the experiment. We utilize randomly-selected training and testing sets to repeat the experiment 10 times, and subsequently report the average obtained accuracy. For quantitative evaluation of the experimental results, we utilize the popular metrics of the overall accuracy (OA), the average accuracy (AA) and the kappa correlation coefficient ( κ ). The definitions of all OA, AA, and Kappa are shown as follows:
(1) OA: OA assesses the proportion of correctly identified samples to all the samples.
O A = i = 1 C ˉ h i i N
where N is the total number of labeled samples, h i i represents the number of class i samples divided into class i, and C ˉ is the total number of categories.
(2) AA: AA represents the mean of the percentage of the correctly identified samples.
A A = 1 C ˉ i = 1 C ˉ h i i N i
where C ˉ is the total number of categories, h i i represents the number of samples of category i divided into category i, and N i represents the number of samples of category i.
(3) Kappa: Kappa coefficient denotes the interrater reliability for categorical variables.
K a p p a = N i C ˉ h i i i C ˉ ( h i + h + i ) N 2 i C ˉ ( h i + h + i )
where h i + and h + i , respectively, represent the total number of samples of category i true category and the number of samples predicted to be category i.
The size of the window in Gaussian smoothing kernel is 11 and the sigma for the Gaussian smoothing kernel is 0.1. The network structure and parameters for generators, discriminators, and classifiers in AC-WGAN-GP are described in detail in Table 7. In the table Deconv represents the fractionally-strided convolutional neural network layer, conv represents the convolutional neural network layer, Fn represents the full connection layer, BN represents the Batchnorm layer, stride represents the step size of convolution, and Padding represents the way of filling. The input of generator G is a vector composed of noise, one-hot label and 30-dimensional principal components. The sample generated by G is a single pixel with the size of H × 1 × 1 . The input of D is ( X f a k e , C f a k e ) and ( X r e a l , C r e a l ) , and the output is a scalar, which represents whether the input sample is a real sample or a false sample. C has the same input as D , and the output is the probability that pixels belong to each category. The size of the mini-batch is set to 64 for training AC-WGAN-GP and CNN, the learning rate ranges from 0.1 to 0.0001, and the value of the label smoothing parameter is 0.1.

4.3. Analysis of Generated Samples

Before using the generated samples for classification, we have to check the distribution of the generated samples. We check whether the distributions of generated samples and real samples are consistent. Because the distribution of each class is different, each class has its own distribution. When generating samples, our method generates labeled samples, so the generated samples belonging to the same category should possess the same distribution as that of the real samples. In order to verify whether the generated samples and real samples of the same class have the same distribution, we extract the principal components of the generated samples and real samples through PCA first, and then select the two first principal components. As shown in Figure 5, Figure 6 and Figure 7, the red triangle represents the generated sample, the blue point represents the real test sample, and the green point represents the real training sample. The figure shows that, for most categories, the AC-WGAN-GP network can generate samples that have a similar distribution to original samples in the feature space. However, for the classes with too few real training samples, the generated sample distribution and the real sample distribution show some differences. Such as Grass-pasture-mowed class in the Indian Pines dataset, the number and distribution of real samples and fake samples are different. In addition, it can be seen from the figure that AC-WGAN-GP tends to imitate simple distributions better, but sometimes complex distributions cannot be well fitted. There are two reasons for this difference. One is that too few training samples are sent to AC-WGAN-GP network, the convergence is not complete; the other is that the randomly selected training samples are not evenly distributed in the class, which cannot represent the overall distribution of this class. For these two reasons, GAN does not perform well in generating a small number of class samples. This is also a common problem of GAN. Intuitively, the sample distribution generated by AC-WGAN-GP is correct.
In the visualization experiment, three classes of generated fake samples and real samples were selected from the three data sets for visual display, as shown in Figure 8, Figure 9 and Figure 10. The solid red line represents the real sample and the dotted black line represents the generated sample. The figure indicates that the generated sample is very similar to the real sample of the corresponding class, but not completely consistent, meet the demands that generated samples need to be consistent with the real sample distribution and have a certain diversity. The generator can learn the different characteristics of each class and generate different samples according to the class. We obtain a model that fits multiple class distributions and fills in the lack of diversity in the sample space using limited samples by utilizing AC-WGAN-GP.
In addition to analyzing the distribution difference according to the visual results obtained by the category, the experiments in this subsection also give numerical results on whether the real samples and the generated samples are consistent in the global distribution. A 1-Nearest Neighbor classifier (1-NN classifier) is used for evaluating whether the two distributions are consistent. Given two sets of samples, the real samples are labeled as positive samples and denoted as S r , and the generated samples are labeled as negative samples and denoted as S f . Train 1 nearest neighbor classifier by S r and S f , and then use them as test samples to obtain new labels, calculate the classification accuracy of the 1-NN classifier, and express it as the Transfer(T) accuracy. When the number of samples is very large, if two distributions are consistent and not completely replicated, the T accuracy of the 1-NN classifier should be 50%. Because when the generated samples are only the results of simple replication of real samples, the T accuracy rate is 0%. When the two distributions do not match at all, the T accuracy rate is 100%. Table 8 calculates the T accuracy of the three HSIs datasets, and the table also shows the average spectral distance between the generated samples and the real samples. We select the number of samples shown in the first row of the table for calculation. The average T accuracies of the three datasets are 57.11%, 69.70% and 63.91%, respectively, which are relatively close to the ideal generative distribution effect. When the average T accuracy rate is closer to 50% and the average spectral distance is smaller, the effect of adding the generated samples to the real training samples for data enhancement is more obvious.

4.4. Generated Sample and Real Sample Mixed Ratio Analysis

The ratio of the real samples and generated samples in training sets may affect the result of classification. So, we conduct an experiment to investigate the effect of injecting different proportions of fake samples on classification accuracy. We selected several representative and common mixing proportions 1:0, 4:1, 2:1, 1:1, 1:2, 1:4. Table 9 lists the experimental results, which indicate that both the Indian pines dataset and the KSC dataset achieve the best classification accuracy when the ratio between the real samples and the fake samples is 1:1, and the Salinas dataset achieves the best precision at 2:1. The accuracy of the three datasets with fake samples is improved compared with that without fake samples.
This experiment indirectly proves that the distribution of samples generated by AC-WGAN-GP is correct and makes up for the lack of diversity of small training samples in the sample space. The method has a positive impact on the classification results. Among them, Indian Pines and KSC classification accuracy are the best when the ratio of real and false samples is 1:1, and classification accuracy on Salinas is the best when the ratio of real and false samples is 2:1. It may indicate that the sample quality generated by the model on datasets Indian Pines and KSC is slightly better. From Table 9, Table 10, Table 11, Table 12 and Table 13, the bold number in the tables represents the best result in the comparison.

4.5. Effectiveness Analysis of Sample Selection Algorithm

We make a comparison of the classification accuracy of KNN selecting fake samples and randomly selected fake samples to verify the function of our selection algorithm. Table 10 presents the experimental results and indicates that the accuracy of the KNN sample selection method on the three datasets is higher than that of random sample selection method. The OA, AA and KAPPA index using KNN selection on Indian Pines dataset are respectively higher than the random selection method 0.2%, 1.06%, 0.21%. On Salinas and KSC datasets, the number is 0.4%, 0.43%, 0.44% and 0.48%, 0.44%, 0.64%. It shows that the KNN sample selection method is effective and necessary. The selection algorithm we proposed selects the generated samples that are similar to the real samples and has a positive effect on classification.

4.6. Classification Result

As a most important way of assessing the performance, we compare it with several other competing algorithms including SVM, CNN, and GAN on small-size training sets (See Section 4.2 for specific quantities). Recognizing the well-known advantages of support vector machines (SVM), this paper introduces some HSIs classifiers based on SVM for comparison, namely 3D-RBF-SVM and EMP-SVM [12]. 3D-RBF-SVM input is an image block, the SVM classifiers are using the radial basis function kernel. As a typical deep learning model, 1D-CNN [23] and 3D-CNN [24] also have good classification performance, and are used to compare. Furthermore, 3D-Aug-GAN [37] also uses a GAN network to augment the training set and improves classification accuracy, which is used for comparison with our AC-WGAN-GP. Meanwhile, 1D-S-SVM and 1D-S-CNN are also used for comparison. Where S stands for gaussian smoothing, 1D means that the input is a single pixel, and the structure of CNN is the same as that of classifier C. These improved classifiers will verify the effect of smoothing operation and whether AC-WGAN-GP augments the training set thus improving the accuracy of classification. All of the above methods use the same training samples in a small size. Besides, we add a HSI classifiers also using samples with small size named AML [48] for comparison.
The qualitative evaluation of various methods is shown in Table 11, Table 12 and Table 13. Table 14 shows the comparison of the recent classifier AML and AC-WGAN-GP. The visual classification results are shown in Figure 11, Figure 12 and Figure 13. Based on the above experimental results, some observations and discussions can be focused on.
First of all, it can be seen from the visual classification map in Figure 11, Figure 12 and Figure 13. Gaussian smoothing of the data with the 1D-CNN classifier, and our proposed AC-WGAN-GP model. It can be seen that the third and fourth column results are significantly better than the second column. And significant differences between the third and fourth sets of prediction plots are marked by white boxes. In Indian Pines, the classification result of the 1D-S-CNN algorithm demonstrates that compared with the method only using CNN, the noise inside the class is small, but the class boundary produces more error points, which shows that the use of Gaussian smoothing can effectively smooth the spectral samples inside the class, but it is easier to confuse the spatial features at the boundary, and obtain the wrong class boundary. But after using AC-WGAN-GP to augment data, the class boundary in the white box is significantly improved. In Salinas, algorithms confuse some of the categories 8 and 15 samples in this data set, making more misclassification points. The reason for this problem is that the spectral characteristics of the two types are relatively similar, and the classification difficulty comes higher. It is difficult for many algorithms to completely distinguish these two categories. Compared with the other two algorithms, the proposed algorithm has a better classification effect on these two categories. And except for some noise points and misclassification points at the boundary of the 10th and 5th categories, the noise points of other categories in this dataset are less and the boundary positioning is clearer. Compared with 1D-CNN, the noise points of the proposed algorithm are greatly reduced, and compared with 1D-S-CNN, there are fewer areas of error points. In KSC, the pixels are relatively discrete. Figure 13e,f shows that the main misclassification point of the algorithm is that the samples of the fifth class close to the fourth class are wrongly classified into the fourth class. The reason for this error is that the spectrum has the problem of the same spectrum including different objects and the same object scattering in a different spectrum, and the number of training samples is few. The classification results of our proposed method still have obvious advantages in the same region. It can be seen that the positive effect of AC-WGAN-GP data augmentation on classification.
From the Tables, it can be seen intuitively that our proposed method outperforms other methods on most of the three metrics OA, AA, Kappa, and most of the category classification accuracy.
Firstly, we observe that the method (i.e., 3D-RBF-SVM, EMP-SVM, 1D-CNN) uses the single pixel or image block of the original data as input, and their classification accuracy is lower than that of other methods (i.e., SVM, 1D-S-SVM, 1D-S-CNN, AC-WGAN-GP) that adopt gaussian smoothing processing. For example, as can be seen from Table 11, the OA of EMP-SVM is 18.97%, 21.94%, and 23.17% lower than 1D-S-SVM, 1D-S-CNN, and AC-WGAN-GP, respectively. Similar properties can be found in Table 12 and Table 13. The above phenomenon shows that gaussian smoothing can improve the classification accuracy, because gaussian smoothing not only simply filters hyperspectral pixels, but also adds neighboring information. At the same time, gaussian smoothing also makes the learning task easier.
Secondly, the classification accuracy of AC-WGAN-GP is higher than that of 1D-S- CNN with only original training samples. As shown in the experimental results of KSC dataset, OA, AA, and Kappa of AC-WGAN-GP are all higher than those of 1D-S-CNN, which are 1.36%, 2.03%, and 1.51% higher, respectively. The same results can be obtained in the Salinas dataset. On the Indian Pines dataset, OA and Kappa were significantly improved, while AA decreased slightly. Compared with 3D-CNN, the proposed method still has an advantage. The experimental results of the Indian Pines dataset, OA, AA, and Kappa of AC-WGAN-GP are all higher than those of 3D-CNN, which are 6.04%, 20.18% and 7.34% higher, respectively. In addition to the third category, we also outperform 3D-CNN in specific categories. For example categories 12, 13, 16, and 3D-CNN are far below the average due to poor training samples to learn and the proposed AC-WGAN-GP does not have this problem.
Thirdly, by comparing the 3D-Aug–GAN with our proposed AC-WGAN-GP method, it is observed from Table 12 that the OA, AA and Kappa of AC-WGAN-GP are 19.65%, 17.02% and 0.25% higher than 3D-Aug–GAN, respectively. In specific categories, our method is still leading. We can observe the same results from Indian pines and KSC data. The above analysis verifies that the proposed method has higher classification accuracy than that of 3D-Aug-GAN. We have made a little attribution on a GAN-based hyperspectral data augmentation and classifier compared with the above old models.
Fourthly, AML is a method that combines LSTM and attention and aims at HSI classification for small training size. On different proportion of training sets, we compare the AML with our AC-WGAN-GP. The result in Table 14 shows that the performance of our method is ahead of AML slightly.
Finally, we notice that high accuracy can be achieved using only 1D-CNN and the smoothing module on the KSC dataset. Only categories 3, 4 and 5 do not achieve 100.00 accuracy. Even in this case, the method using AC-WGAN-GP framework is still 1.36%, 2.03%, 1.51% higher on OA, AA, and Kappa, respectively. Specially, in Figure 14, we list distribution of generated samples from three categories. These samples have low quality and lead to low accuracy. In Alfalfa, we can see the distribution of real samples is discrete and irregular. So it is difficult for AC-WGAN-GP to learn a better distribution. The generated samples have obvious wrong. In Stone-Steel-Towers, the situation is similar. In OAK/Broadleaf, the generated samples conform to the distribution of the real samples to some extent, but the network has not learned the right sparse and density of real distribution. From the above instances, we find that some categories have complex and uneven distribution. This tests the performance of GAN. For AC-WGAN-GP proposed by us, the result is not ideal.

5. Conclusions

This paper discusses the feasibility of using labeled fake samples as a method of data augmentation in hyperspectral image classification with small sample sizes. Gaussian smoothing makes use of neighboring information and makes the learning task of the network simpler. The proposed AC-WGAN-GP model extends the traditional GAN framework based on the WGAN-GP and AC-GAN. Then, a new sampling method is introduced to generate labeled samples based online generation mechanism. Lastly, a sample selection method is designed. The LSR processing of fake labelled samples can make the use of fake samples more reasonable. The KNN method is used to select samples similar to the test samples from the generated samples and add them to the original training set. The combined training set is used to train the CNN and complete the hyperspectral images classification task with small samples. We verified the results of the proposed method on three popular well-known datasets by selecting few training samples with labels. When we only use 2.0% samples in Indian Pines set, 0.37% samples in Salinas set, 3.8% samples in KSC set, the classification accuracy obtained by our model is 92.51%, 96.86%, and 98.92%, respectively. Compared with the models based on 3DCNN, the proposed model in this paper increases by 6.04%, 8.71%, 3.29%, respectively. Compared with models based on 3DGAN, the proposed model in this paper increases by 1.41%, 3.19%, 0.8%, respectively. However, the complex distribution of original small samples is sometimes difficult for our framework to learn. In our future work, we will try to design more effective GAN to generate better samples or add other frameworks such as in [49].

Author Contributions

C.S. wrote the whole article, adjusted it and adjusted the structure of the article and checked out some mistakes. X.Z. directed the completion of the algorithm and put forward key suggestions. X.C. reviewed the whole paper carefully and gave some key suggestions on the format and writing style of the paper. J.Z. checked out some grammatical errors in the paper. H.M. reviewed the article and made some suggestions. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 61877066, Aero-Science Fund under Grant 20175181013, Science and technology plan project of Xi’an under Grant 21RGZN0010.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ghamisi, P.; Maggiori, E.; Li, S.; Souza, R.; Tarablaka, Y.; Moser, G.; De Giorgi, A.; Fang, L.; Chen, Y.; Chi, M.; et al. New Frontiers in Spectral-Spatial Hyperspectral Image Classification: The Latest Advances Based on Mathematical Morphology, Markov Random Fields, Segmentation, Sparse Representation, and Deep Learning. IEEE Geo-Sci. Remote Sens. Mag. 2018, 6, 10–43. [Google Scholar] [CrossRef]
  2. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  3. Landgrebe, D. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  4. Nasrabadi, N.M. Hyperspectral Target Detection: An Overview of Current and Future Challenges. IEEE Signal Process. Mag. 2013, 31, 34–44. [Google Scholar] [CrossRef]
  5. van der Meer, F. Analysis of spectral absorption features in hyperspectral imagery. Int. J. Appl. Earth Obs. Geoinf. 2004, 5, 55–68. [Google Scholar] [CrossRef]
  6. Hege, E.K.; O’Connell, D.; Johnson, W.; Basty, S.; Dereniak, E.L. Hyperspectral imaging for astronomy and space surveillance. In Imaging Spectrometry IX; SPIE: Bellingham, WA, USA, 2004; pp. 380–391. [Google Scholar]
  7. Lacar, F.M.; Lewis, M.M.; Grierson, I.T. Use of hyperspectral imagery for mapping grape varieties in the Barossa Valley, South Australia. In Proceedings of the IGARSS 2001, Scanning the Present and Resolving the Future, Proceedings of the IEEE 2001 International Geoscience and Remote Sensing Symposium (Cat. No.01CH37217), Sydney, NSW, Australia, 9–13 July 2001; Volume 6, pp. 2875–2877. [Google Scholar]
  8. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  9. Malthus, T.; Mumby, P.J. Remote sensing of the coastal zone: An overview and priorities for future research. Int. J. Remote Sens. 2003, 24, 2805–2815. [Google Scholar] [CrossRef]
  10. Plaza, A.; Du, Q.; Chang, Y.-L.; King, R.L. High Performance Computing for Hyperspectral Remote Sensing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2011, 4, 528–544. [Google Scholar] [CrossRef]
  11. Liu, Q.; Liu, C. A Novel Locally Linear KNN Method With Applications to Visual Recognition. IEEE Trans. Neural Networks Learn. Syst. 2016, 28, 2010–2021. [Google Scholar] [CrossRef]
  12. Fung, G.M.; Mangasarian, O.L. Multicategory Proximal Support Vector Machine Classifiers. Mach. Learn. 2005, 59, 77–97. [Google Scholar] [CrossRef] [Green Version]
  13. Pu, H.; Chen, Z.; Wang, B.; Jiang, G.M. A Novel Spatial–Spectral Similarity Measure for Dimensionality Reduction and Classification of Hyper-spectral Imagery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7008–7022. [Google Scholar]
  14. Touil, M.; Boudebza, I.E.; Daamouche, A. Classification of hyperspectral data using grey model. In Proceedings of the 2015 4th International Conference on Electrical Engineering (ICEE), Boumerdes, Algeria, 13–15 December 2015; pp. 1–5. [Google Scholar]
  15. Wang, F.; Zhang, R.; Wu, Q. Hyperspectral image classification based on PCA network. In Proceedings of the 2016 8th Workshop on Hyper-Spectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016; pp. 1–4. [Google Scholar]
  16. Cihan, M.; Ceylan, M. Comparison of Linear Discriminant Analysis, Support Vector Machines and Naive Bayes Methods in the Classification of Neonatal Hyperspectral Signatures. In Proceedings of the 2021 29th Signal Processing and Communications Applications Conference (SIU), Istanbul, Turkey, 9–11 June 2021; pp. 1–4. [Google Scholar] [CrossRef]
  17. Su, P.; Liu, D.; Li, X.; Liu, Z. A Saliency-Based Band Selection Approach for Hyperspectral Imagery Inspired by Scale Selection. IEEE Geosci. Remote Sens. Lett. 2018, 15, 572–576. [Google Scholar] [CrossRef]
  18. Wei, L.; Yu, M.; Zhong, Y.; Yuan, Z.; Huang, C. Conditional random field hyperspectral image classification method based on space-spectral fusion. Chin. J. Surv. Mapp. 2020, 49, 343–354. [Google Scholar]
  19. Qian, X. Research on Hyperspectral Image Classification Combining Spatial Information and Spectral Information; Harbin Engineering University: Harbin, China, 2014. [Google Scholar]
  20. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  21. Özdemir, A.O.B.; Gedik, B.E.; Çetin, C.Y.Y. Hyperspectral classification using stacked autoencoders with deep learning. In Proceedings of the 2014 6th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lausanne, Switzerland, 24–27 June 2014. [Google Scholar]
  22. Pooja, K.; Nidamanuri, R.R.; Mishra, D. Multi-Scale Dilated Residual Convolutional Neural Network for Hyperspectral Image Classification. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  23. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H.-C. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sensors 2015, 2015, 258619. [Google Scholar] [CrossRef] [Green Version]
  24. Li, Y.; Zhang, H.; Shen, Q. Spectral–Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef] [Green Version]
  25. Pu, C.; Huang, H.; Li, Z. Spatial-Spectral Combination Convolutional Neural Network for Hyperspectral Image Classification. In Proceedings of the IGARSS 2020–2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2037–2040. [Google Scholar]
  26. Ding, Y.; Chong, Y.; Pan, S.; Wang, Y.; Nie, C. Spatial-Spectral Unified Adaptive Probability Graph Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–15. [Google Scholar] [CrossRef]
  27. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  28. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classi-fication. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar]
  29. Hanachi, R.; Sellami, A.; Farah, I.R.; Mura, M.D. Semi-supervised Classification of Hyperspectral Image through Deep Encoder-Decoder and Graph Neural Networks. In Proceedings of the 2021 International Congress of Advanced Technology and Engineering (ICOTEN), Taiz, Yemen, 4–5 July 2021; pp. 1–8. [Google Scholar] [CrossRef]
  30. Zhao, L.; Luo, W.; Liao, Q.; Chen, S.; Wu, J. Hyperspectral Image Classification with Contrastive Self-Supervised Learning Under Limited Labeled Samples. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  31. Li, T.; Zhang, X.; Zhang, S.; Wang, L. Self-Supervised Learning with a Dual-Branch ResNet for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  32. Qin, K.; Ge, F.; Zhao, Y.; Zhu, L.; Li, M.; Shi, C.; Li, D.; Zhou, X. Hapke Data Augmentation for Deep Learning-Based Hyperspectral Data Analysis with Limited Samples. IEEE Geosci. Remote Sens. Lett. 2020, 18, 886–890. [Google Scholar] [CrossRef]
  33. Wang, C.; Zhang, L.; Wei, W.; Zhang, Y. Hyperspectral Image Classification with Data Augmentation and Classifier Fusion. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1420–1424. [Google Scholar] [CrossRef]
  34. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  35. Odena, A.; Olah, C.; Shlens, J. Conditional Image Synthesis With Auxiliary Classifier GANs. arXiv 2016, arXiv:1610.09585. [Google Scholar]
  36. Zhan, Y.; Hu, D.; Wang, Y.; Yu, X. Semisupervised Hyperspectral Image Classification Based on Generative Adversarial Networks. IEEE Geosci. Remote Sens. Lett. 2017, 15, 212–216. [Google Scholar] [CrossRef]
  37. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  38. Li, Y.; Lyu, X.; Frery, A.C.; Ren, P. Oil Spill Detection with Multiscale Conditional Adversarial Networks with Small-Data Training. Remote Sens. 2021, 13, 2378. [Google Scholar] [CrossRef]
  39. Feng, J.; Zhao, N.; Shang, R.; Zhang, X.; Jiao, L. Self-Supervised Divide-and-Conquer Generative Adversarial Network for Classification of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  40. He, Z.; Xia, K.; Ghamisi, P.; Hu, Y.; Fan, S.; Zu, B. HyperViTGAN: Semisupervised Generative Adversarial Network With Transformer for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 6053–6068. [Google Scholar] [CrossRef]
  41. Liang, H.; Bao, W.; Lei, B.; Zhang, J.; Qu, K. Adaptive Neighborhood Strategy Based Generative Adversarial Network for Hyperspectral Image Classification. In Proceedings of the IGARSS 2020–2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 862–865. [Google Scholar] [CrossRef]
  42. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved Training of Wasserstein GANs. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  43. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. [Google Scholar]
  44. Rubner, Y.; Tomasi, C.; Guibas, L.J. The Earth Mover’s Distance as a Metric for Image Retrieval. Int. J. Comput. Vis. 2000, 40, 99–121. [Google Scholar] [CrossRef]
  45. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  46. Xu, Q.; Huang, G.; Yuan, Y.; Guo, C.; Sun, Y.; Wu, F.; Weinberger, K. An empirical study on evaluation metrics of generative adversarial networks. arXiv 2018, arXiv:1806.07755. [Google Scholar]
  47. Sun, W.; Zhang, L.; Du, B. Feature extraction using near-isometric linear embeddings for hyperspectral imagery classification. In Proceedings of the 2016 8th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016. [Google Scholar]
  48. Wang, Z.; Zou, C.; Cai, W. Small Sample Classification of Hyperspectral Remote Sensing Images Based on Sequential Joint Deeping Learning Model. IEEE Access 2020, 8, 71353–71363. [Google Scholar] [CrossRef]
  49. Hou, S.; Shi, H.; Cao, X.; Zhang, X.; Jiao, L. Hyperspectral Imagery Classification Based on Contrastive Learning. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–13. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed method for HSIs classification.
Figure 1. The framework of the proposed method for HSIs classification.
Remotesensing 14 04910 g001
Figure 2. The architecture of the proposed AC-WGAN-GP.
Figure 2. The architecture of the proposed AC-WGAN-GP.
Remotesensing 14 04910 g002
Figure 3. The proposed online sample generation model.
Figure 3. The proposed online sample generation model.
Remotesensing 14 04910 g003
Figure 4. Ground truths (first row) and False-color composites (second row) of experimental HSI datasets. Each color represents one kind of object. (a) Indian Pines; (b) Salinas; (c) KSC.
Figure 4. Ground truths (first row) and False-color composites (second row) of experimental HSI datasets. Each color represents one kind of object. (a) Indian Pines; (b) Salinas; (c) KSC.
Remotesensing 14 04910 g004aRemotesensing 14 04910 g004b
Figure 5. The distribution of real samples and generated samples in different classes on Indian Pines dataset. (a) Corn-notill class; (b) Grass-trees class; (c) Grass-pasture class; (d) Grass-pasture-mowed class.
Figure 5. The distribution of real samples and generated samples in different classes on Indian Pines dataset. (a) Corn-notill class; (b) Grass-trees class; (c) Grass-pasture class; (d) Grass-pasture-mowed class.
Remotesensing 14 04910 g005
Figure 6. The distribution of real samples and generated samples in different classes on Salinas dataset. (a) Stubble class; (b) Grapes_untrained class; (c) Vinyard_untrained class; (d) Fallow_rough_plow class.
Figure 6. The distribution of real samples and generated samples in different classes on Salinas dataset. (a) Stubble class; (b) Grapes_untrained class; (c) Vinyard_untrained class; (d) Fallow_rough_plow class.
Remotesensing 14 04910 g006
Figure 7. The distribution of real samples and generated samples in different classes on KSC dataset. (a) Spartina marsh class; (b) Willow swamp class; (c) Water class; (d) Hardwood class.
Figure 7. The distribution of real samples and generated samples in different classes on KSC dataset. (a) Spartina marsh class; (b) Willow swamp class; (c) Water class; (d) Hardwood class.
Remotesensing 14 04910 g007
Figure 8. The real data and generated fake data with same labels in different classes on Indian Pines dataset. (a) Alfalfa class; (b) Woods class; (c) Soybean-clean class.
Figure 8. The real data and generated fake data with same labels in different classes on Indian Pines dataset. (a) Alfalfa class; (b) Woods class; (c) Soybean-clean class.
Remotesensing 14 04910 g008
Figure 9. The real data and generated fake data with same labels in different classes on Salinas dataset. (a) Brocoli_green_weeds_1 class; (b) Stubble class; (c) Fallow_rough_plow class.
Figure 9. The real data and generated fake data with same labels in different classes on Salinas dataset. (a) Brocoli_green_weeds_1 class; (b) Stubble class; (c) Fallow_rough_plow class.
Remotesensing 14 04910 g009
Figure 10. The real data and generated fake data with same labels in different classes on KSC dataset. (a)Scrub class; (b) Spartina marsh class; (c) Water class.
Figure 10. The real data and generated fake data with same labels in different classes on KSC dataset. (a)Scrub class; (b) Spartina marsh class; (c) Water class.
Remotesensing 14 04910 g010
Figure 11. The visual Indian Pines classification map. Mark clearly dominant areas with boxes. (a) Ground Truth; (b) 1D-CNN; (c) 1D-S-CNN; (d) Proposed method.
Figure 11. The visual Indian Pines classification map. Mark clearly dominant areas with boxes. (a) Ground Truth; (b) 1D-CNN; (c) 1D-S-CNN; (d) Proposed method.
Remotesensing 14 04910 g011
Figure 12. The visual Salinas classification map. Mark clearly dominant areas with boxes. (a) Ground Truth; (b) 1D-CNN; (c) 1D-S-CNN; (d) Proposed method.
Figure 12. The visual Salinas classification map. Mark clearly dominant areas with boxes. (a) Ground Truth; (b) 1D-CNN; (c) 1D-S-CNN; (d) Proposed method.
Remotesensing 14 04910 g012
Figure 13. The partially enlarged visual KSC classification map. Mark clearly dominant areas with boxes. (a) Ground Truth; (b) 1D-CNN; (c) 1D-S-CNN; (d) Proposed method; (e) Enlarged 1D-S-CNN; (f) Enlarged proposed method.
Figure 13. The partially enlarged visual KSC classification map. Mark clearly dominant areas with boxes. (a) Ground Truth; (b) 1D-CNN; (c) 1D-S-CNN; (d) Proposed method; (e) Enlarged 1D-S-CNN; (f) Enlarged proposed method.
Remotesensing 14 04910 g013aRemotesensing 14 04910 g013b
Figure 14. Some generated samples that lead to the degradation of classification accuracy (a) Indian Pines Alfalfa; (b) Indian Pines Stone-Steel-Towers; (c) KSC Oak/Broadleaf.
Figure 14. Some generated samples that lead to the degradation of classification accuracy (a) Indian Pines Alfalfa; (b) Indian Pines Stone-Steel-Towers; (c) KSC Oak/Broadleaf.
Remotesensing 14 04910 g014
Table 1. Land cover types and total number of samples in the Indian Pines dataset.
Table 1. Land cover types and total number of samples in the Indian Pines dataset.
No.ColorNameNumberNo.ColorNameNumber
1Remotesensing 14 04910 i001Alfalfa469Remotesensing 14 04910 i002Oats20
2Remotesensing 14 04910 i003Corn-notill142810Remotesensing 14 04910 i004Soybean-notill972
3Remotesensing 14 04910 i005Corn-mintill83011Remotesensing 14 04910 i006Soybean-mintill2455
4Remotesensing 14 04910 i007Corn23712Remotesensing 14 04910 i008Soybean-clean593
5Remotesensing 14 04910 i009Grass-pasture48313Remotesensing 14 04910 i010Wheat205
6Remotesensing 14 04910 i011Grass-trees73014Remotesensing 14 04910 i012Woods1265
7Remotesensing 14 04910 i013Grass-pasture-mowed2815Remotesensing 14 04910 i014Buildings-Grass-Trees386
8Remotesensing 14 04910 i015Hay-windrowed47816Remotesensing 14 04910 i016Stone- Steel-Towers93
Total Numbers10,249
Table 2. Land cover types and total number of samples in the Salinas dataset.
Table 2. Land cover types and total number of samples in the Salinas dataset.
No.ColorNameNumberNo.ColorNameNumber
1Remotesensing 14 04910 i001Brocoli_green_weeds_120099Remotesensing 14 04910 i002Soil_vinyard_develop6203
2Remotesensing 14 04910 i003Brocoli_green_weeds_2372610Remotesensing 14 04910 i004Corn_senesced_green_weeds3278
3Remotesensing 14 04910 i005Fallow197611Remotesensing 14 04910 i006Lettuce_romaine_4wk1068
4Remotesensing 14 04910 i007Fallow_rough_pow139412Remotesensing 14 04910 i008Lettuce_romaine_5wk1927
5Remotesensing 14 04910 i009Fallow_smooth267813Remotesensing 14 04910 i010Lettuce_romaine_6wk916
6Remotesensing 14 04910 i011Stubble395914Remotesensing 14 04910 i012Lettuce_romaine_7wk1070
7Remotesensing 14 04910 i013Celery357915Remotesensing 14 04910 i014Vinyard_untrained7268
8Remotesensing 14 04910 i015Grapes_untrained11,27116Remotesensing 14 04910 i016Vinyard_vertical_trellis1807
Total Numbers54,129
Table 3. Land cover types and total number of samples in the KSC dataset.
Table 3. Land cover types and total number of samples in the KSC dataset.
No.ColorNameNumberNo.ColorNameNumber
1Remotesensing 14 04910 i001Scrub7618Remotesensing 14 04910 i015Graminoid marsh431
2Remotesensing 14 04910 i003Willow swamp2439Remotesensing 14 04910 i002Spartina marsh520
3Remotesensing 14 04910 i005CP hammock25610Remotesensing 14 04910 i004Cattail marsh404
4Remotesensing 14 04910 i007Slash pine25211Remotesensing 14 04910 i006Salt marsh419
5Remotesensing 14 04910 i009Oak/Broadleaf16112Remotesensing 14 04910 i008Mud flats503
6Remotesensing 14 04910 i011Hardwood22913Remotesensing 14 04910 i010Water927
7Remotesensing 14 04910 i013Swamp105
Total Numbers5211
Table 4. Land cover classes, number of training samples and total samples in Indian Pines.
Table 4. Land cover classes, number of training samples and total samples in Indian Pines.
Classes NameTrain NumbersAll Numbers
Alfalfa346
Corn-notill291428
Corn-min17830
Corn5237
Grass-pasture10483
Grass-trees15730
Grass-pasture mowed328
Hay-windrowed10478
Oats320
Soybean-notill19972
Soybean-mintill492455
Soybean-clean12593
Wheat4205
Woods251265
Buildings-Grass-Trees8386
Stone-Steel-Towers393
Total21510,249
Table 5. Land cover classes, number of training samples and total samples in Salinas.
Table 5. Land cover classes, number of training samples and total samples in Salinas.
Classes NameTrain NumbersAll Numbers
Brocoli_green_weeds_1202009
Brocoli_green_weeds_2373726
Fallow201976
Fallow_rough_plow141394
Fallow_smooth272678
Stubble403959
Celery363579
Grapes_untrained11311,271
Soil_vinyard_develop626203
Corn_senesced_green_weeds333278
Lettuce_romaine_4wk111068
Lettuce_romaine_4wk191927
Lettuce_romaine_4wk9916
Lettuce_romaine_4wk111070
Vinyard_untrained737268
Vinyard_vertical_trellis181807
Total54354,129
Table 6. Land cover classes, number of training samples, and total samples in KSC.
Table 6. Land cover classes, number of training samples, and total samples in KSC.
Classes NameTrain NumbersAll Numbers
Scrub29761
Willow swamp9243
CP hammock10256
Slash pine10252
Oak/Broadleaf6161
Hardwood11229
Swamp4105
Graminoid marsh16431
Spartina marsh20520
Cattail marsh15404
Salt marsh16419
Mud flats19503
Water35927
Total2005211
Table 7. The architecture of the AC-WGAN-GP.
Table 7. The architecture of the AC-WGAN-GP.
NetNumberInput SizeLayerBnStridePaddingActivation FunctionOutput Size
G1100+30+classnumberFn
(1/16H×512)
Reshape
Yes--ReLU1/16H×1×512
21/16H×1×512Deconv1d
(3×1×512×256)
Yes2×1SAMEReLU1/8H×1×256
31/8H×1×256Deconv1d
(3×1×256×128)
Yes2×1SAMEReLU1/4H×1×128
41/4H×1×128Deconv1d
(3×1×128×64)
Yes2×1SAMEReLU1/2H×1×64
51/2H×1×64Deconv1d
(3×1×64×1)
No2×1SAMETanh1/16H×1×1
D1H×1×1Conv1d
(3×1×1×64)
No2×1SAMELeakyReLU1/2H×1×64
21/2H×1×64Conv1d
(3×1×64×128)
Yes2×1SAMELeakyReLU1/4H×1×128
31/4H×1×128Conv1d
(3×1×128×256)
Yes2×1SAMELeakyReLU1/8H×1×256
41/8H×1×256Conv1d
(3×1×256×512)
Yes2×1SAMELeakyReLU1/16H×1×512
51/16H×1×512Flatten
Fn(32H×1)
No---1
C2H×1×1Conv
(15×1×1×64)
Flatten
No15×1SAMETanh1/15H×1×64
31/15H×1×64Fn
(64/15H×C’)
No--SoftmaxC’
Table 8. 1-NN accuracy (%) and average spectral distance of generated and real samples.
Table 8. 1-NN accuracy (%) and average spectral distance of generated and real samples.
DatasetIndian PinesSalinasKSC
| S r | = | S f | 10,24910,8525211
1-NN accuracy (Real) (%)58.2265.0374.24
1-NN accuracy (Fake) (%)56.0062.8065.16
1-NN accuracy (Average) (%)57.1163.9169.70
Average Spectral Distance0.50280.34210.1135
Table 9. Effect of different mixing proportions in training sets.
Table 9. Effect of different mixing proportions in training sets.
Real:Fake1:04:12:11:11:21:4
DatasetIndian Pines
OA (%)91.2891.8892.4492.5191.8792.43
AA (%)92.1588.8788.9090.5989.1590.33
Kappa×10090.8890.7491.3991.4690.7491.39
DatasetSalinas
OA (%)96.0696.8396.9796.8696.7896.73
AA (%)97.4398.2698.4797.9398.2997.79
Kappa×10095.6196.4896.6396.5096.4196.37
DatasetKSC
OA (%)97.5698.7198.8498.9298.8698.91
AA (%)95.4397.3897.6697.4697.6997.43
Kappa×10097.2998.5798.7298.8098.7498.74
Table 10. Effect of fake sample selection method.
Table 10. Effect of fake sample selection method.
MethodKNN SelectionRandom Selection
Indian PinesSalinasKSCIndian PinesSalinasKSC
OA (%)92.5196.9798.9292.3196.5798.44
AA (%)90.5998.4797.4689.5398.0497.02
Kappa×10091.4696.6398.8091.2596.1998.26
Table 11. The Classification Accuracy of various methods for Indian Pines. Bold values indicate the best result for a row.
Table 11. The Classification Accuracy of various methods for Indian Pines. Bold values indicate the best result for a row.
Indian Pines
Method3D-RBF-
SVM
EMP-SVM1D-CNN1D-S-SVM3D-Aug-GAN1D-S-CNN3D-CNNAC-WGAN-GP
OA (%)58.0169.3467.1888.3191.1091.2886.4792.51
AA (%)50.5652.6355.1881.0783.7692.1570.4190.59
Kappa×10052.0764.5162.3786.6989.9590.0884.1291.46
177.3515.948.7039.1331.2484.7814.7071.74
218.5239.1657.1494.0482.0594.0186.3494.26
354.6670.7528.6771.8177.1979.4089.4980.12
432.3054.7425.7497.0591.3196.2042.0093.67
59.7368.3776.6082.1988.5184.6885.9185.71
685.5396.2191.3797.6793.5198.3692.7598.90
710.27100.007.1475.0041.76100.0012.00100.00
878.5085.0098.3394.1498.3299.37100.00100.00
912.3215.785.0045.0021.96100.0010.0090.00
1062.2875.9264.3085.1875.1785.0878.7290.12
1166.8681.2372.5088.3191.3790.2695.5292.83
1228.2531.8556.6686.3486.4590.7389.4792.75
1399.1898.2393.6779.0252.2098.0580.0099.02
1485.9890.8590.5197.0794.6499.7684.5599.92
1513.8294.2826.9475.9195.2676.4269.5479.79
1687.7895.2479.5789.2589.3896.7789.3480.64
Table 12. The Classification Accuracy of various methods for Salinas. Bold values indicate the best result for a row.
Table 12. The Classification Accuracy of various methods for Salinas. Bold values indicate the best result for a row.
Salinas
Method3D-RBF-
SVM
EMP-SVM1D-CNN1D-S-SVM3D-Aug-GAN1D-S-CNN3D-CNNAC-WGAN-GP
OA (%)83.0985.9085.2895.7693.6796.0688.1596.86
AA ($)85.4682.5389.2996.9690.8997.4377.7697.93
Kappa×10081.0784.0283.6195.2892.5595.6186.0596.50
194.1577.9798.3697.0199.1497.3160.1799.90
298.5799.7598.5598.3694.23100.0095.04100.00
390.5650.4085.4390.6476.7498.7984.69100.00
498.9398.7296.2798.21100.0096.7797.6397.20
595.2397.4481.6396.2390.1499.1499.9598.69
699.2599.9499.8599.9799.70100.0099.9699.19
798.8299.8899.53100.0099.9098.8887.5099.89
878.5098.5077.2791.9490.5792.8389.6591.15
994.1199.3398.6999.55100.00100.0099.38100.00
1085.6593.9981.6690.5499.0387.5898.2095.91
1190.6382.3078.4698.6997.6998.2292.2099.53
1299.48100.0093.4199.9099.56100.0034.5499.84
1320.0899.1297.7199.7879.6099.8919.2090.83
1466.2997.6491.6899.1678.8899.6390.3899.91
1559.1413.8055.7492.0671.3390.3691.5995.03
1666.9693.7994.5299.3490.1299.5018.4999.83
Table 13. The Classification Accuracy of various methods for KSC. Bold values indicate the best result for a row.
Table 13. The Classification Accuracy of various methods for KSC. Bold values indicate the best result for a row.
KSC
Method3D-RBF-
SVM
EMP-SVM1D-CNN1D-S-SVM3D-Aug-GAN1D-S-CNN3D-CNNAC-WGAN-GP
OA (%)76.1390.5988.0496.5498.1297.5695.6398.92
AA ($)64.6085.8381.3594.4694.7695.4389.6597.46
Kappa×10073.5289.4186.6796.1598.0597.2994.9598.80
188.6887.8296.32100.0098.71100.0091.71100.00
269.7780.1488.07100.0081.28100.0089.73100.00
370.7387.6485.1677.7398.5776.9592.1697.66
457.3286.6451.1988.8987.6389.6886.94100.00
542.8577.0249.6973.9198.8973.9194.7969.56
632.2489.6654.5999.56100.00100.0090.92100.00
710.0083.3773.33100.0098.20100.0091.57100.00
842.2391.7885.8589.0975.05100.0096.2299.77
984.0097.1294.23100.0099.32100.0099.53100.00
1081.8897.0699.00100.00100.00100.0099.81100.00
1196.7599.6491.8998.81100.00100.0099.79100.00
1286.3299.2489.07100.0097.71100.0097.69100.00
13100.00100.0099.14100.00100.00100.00100.00100.00
Table 14. Compare the AML and AC-WGAN-GP with different sizes of training sets on Indian Pines.
Table 14. Compare the AML and AC-WGAN-GP with different sizes of training sets on Indian Pines.
DatasetIndian PinesSalinas
Training set proportion5%10%1%5%
methodsAML
OA (%)77.0482.9591.6394.54
AA (%)77.7283.2994.5897.15
Kappa×10074.4681.8993.1595.52
DatasetAC-WGAN-GP
OA (%)92.4493.5896.8697.04
AA (%)92.7694.2397.9397.88
Kappa×10091.6193.0196.5096.95
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sun, C.; Zhang, X.; Meng, H.; Cao, X.; Zhang, J. AC-WGAN-GP: Generating Labeled Samples for Improving Hyperspectral Image Classification with Small-Samples. Remote Sens. 2022, 14, 4910. https://doi.org/10.3390/rs14194910

AMA Style

Sun C, Zhang X, Meng H, Cao X, Zhang J. AC-WGAN-GP: Generating Labeled Samples for Improving Hyperspectral Image Classification with Small-Samples. Remote Sensing. 2022; 14(19):4910. https://doi.org/10.3390/rs14194910

Chicago/Turabian Style

Sun, Caihao, Xiaohua Zhang, Hongyun Meng, Xianghai Cao, and Jinhua Zhang. 2022. "AC-WGAN-GP: Generating Labeled Samples for Improving Hyperspectral Image Classification with Small-Samples" Remote Sensing 14, no. 19: 4910. https://doi.org/10.3390/rs14194910

APA Style

Sun, C., Zhang, X., Meng, H., Cao, X., & Zhang, J. (2022). AC-WGAN-GP: Generating Labeled Samples for Improving Hyperspectral Image Classification with Small-Samples. Remote Sensing, 14(19), 4910. https://doi.org/10.3390/rs14194910

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop