Next Article in Journal
Fine-Grained Urban Air Quality Mapping from Sparse Mobile Air Pollution Measurements and Dense Traffic Density
Next Article in Special Issue
A Hybrid Privacy-Preserving Deep Learning Approach for Object Classification in Very High-Resolution Satellite Images
Previous Article in Journal
A Swin Transformer-Based Encoding Booster Integrated in U-Shaped Network for Building Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarial Representation Learning for Hyperspectral Image Classification with Small-Sized Labeled Set

1
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
2
School of Mathematics and Statistics, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(11), 2612; https://doi.org/10.3390/rs14112612
Submission received: 30 April 2022 / Revised: 26 May 2022 / Accepted: 26 May 2022 / Published: 29 May 2022

Abstract

:
Hyperspectral image (HSI) classification is one of the main research contents of hyperspectral technology. Existing HSI classification algorithms that are based on deep learning use a large number of labeled samples to train models to ensure excellent classification effects, but when the labeled samples are insufficient, the deep learning model is prone to overfitting. In practice, there are a large number of unlabeled samples that have not been effectively utilized, so it is meaningful to study a semi-supervised method. In this paper, an adversarial representation learning that is based on a generative adversarial networks (ARL-GAN) method is proposed to solve the small samples problem in hyperspectral image classification by applying GAN to the representation learning domain in a semi-supervised manner. The proposed method has the following distinctive advantages. First, we build a hyperspectral image block generator whose input is the feature vector that is extracted from the encoder and use the encoder as a feature extractor to extract more discriminant information. Second, the distance of the class probability output by the discriminator is used to measure the error between the generated image block and the real image instead of the root mean square error (MSE), so that the encoder can extract more useful information for classification. Third, GAN and conditional entropy are used to improve the utilization of unlabeled data and solve the small sample problem in hyperspectral image classification. Experiments on three public datasets show that the method achieved better classification accuracy with a small number of labeled samples compared to other state-of-the-art methods.

1. Introduction

Hyperspectral remote sensing technology uses a large number of spectral bands to image surface objects, these spectral bands are composed of ten to hundreds non-overlapping narrow bands and it combines spatial information with its own unique spectral information to acquire hyperspectral images on a pixel-by-pixel basis. An important research branch of hyperspectral remote sensing technology is HSI classification technology and it is the task of assigning categories to each pixel in a hyperspectral image. At present, this technology is widely used in many fields [1] such as modern agriculture [2], aviation [3], mineral exploration [4], astronomy [5], biomedicine [6], national defense, and homeland security monitoring [7].
For hyperspectral image classification techniques, the early studies only used the spectral information in hyperspectral datasets but ignored the spatial information that was contained in hyperspectral datasets, such as k-nearest neighbor (KNN) classifier [8], support vector machine (SVM) [9], distance classifiers [10], and extreme learning machines [11], etc. As the spectrum of pixel is affected by the surrounding pixel and noise, the algorithm using spectral information alone cannot obtain good classification performance. In subsequent studies, many researchers have introduced spatial information and formed the classification method of HSI that is based on spatial-spectral combination. For example, the spatial-spectral composite kernel function method that is based on SVM [12]. This method combines hidden Markov random field segmentation with SVM [13]. The non-local weighted joint sparse representation classification method [14], and the joint sparse representation classification method that is based on shape adaptation [15]. Although the above methods have achieved some improvement in classification performance, their features are designed manually.
Deep learning is able to extract features from data automatically compared to ordinary machine learning that requires manual feature design, so many researchers have proposed HSI classification algorithms that are based on deep learning. For example, a stacked autoencoder [16] is used to extract hyperspectral spatial-spectral joint features [17]. Deep belief network (DBN) [18] is used to extract both spectral and spatial features of hyperspectral images and then cascade them to form spatial-spectral features for classification [19]. Convolutional neural network (CNN) [20] has features such as local perception, weight sharing, and multiple convolutional kernels, and thus has received attention from many researchers and many algorithms have been proposed [21]. For example, 1D-CNN [22] and 2D-CNN [23] are used to extract both spectral and spatial features, respectively. 1D-CNN and 2D-CNN are also combined to extract joint spatial-spectral features [24]. 3D-CNN is used to extract features directly from image blocks [25]. Pixel-pair convolutional neural network approach (PPF-CNN) [26] extends the training dataset by reorganizing and rescaling the training samples. However, the labeled samples of HSI are difficult to obtain and the training of deep learning requires a huge number of labeled samples data.
For solving the small sample problem, some scholars have proposed the use of transfer learning (TL) to classify hyperspectral images. For example, Lin et al. proposed an unsupervised classification method that was based on sample migration by transforming the source domain samples and target domain samples into the feature subspace [27], finding the transformation matrix between them and obtaining the initial classification results. Persello et al. used a feature migration strategy to reduce the difference between the source and target domain data by projecting the source domain data into the regenerative Hilbert space and calculate the distance between the conditional distributions of the source and target domains [28]. Some scholars have proposed some semi-supervised approaches, for example, Sun Z. proposed a semi-supervised support vector machine algorithm [29], which uses both labeled and unlabeled data to find a classification surface with maximum class spacing. Aydemir M. S. et al. proposed a graph-based HSI classification algorithm [30], which assumes that similar input data should share similar output labels, the vertices of the graph are labeled or unlabeled training samples, and the weights of the edges are the similarity between examples, followed by specifying the objective function to be optimized and using the smoothness as a regular term to find the optimal parameters. However, the distribution of unlabeled data that is used by the TL methods does not match the distribution of real samples and manual domain adaptation operations cannot completely solve this problem.
For the above problems, scholars have applied generative adversarial network (GAN) [31] to hyperspectral image classification. The method, that is based on GAN, does not need to provide the sample distribution in advance and the samples distribution can be learned automatically by the neural network. At the same time, the methods that are based on GAN are also able to use a huge number of unlabeled samples. 1DGAN and 3DGAN are used as spectral classification and spatial image block classification [32], respectively, by combining a conditional generative adversarial network (CGAN) [33] with an auxiliary classifier generative adversarial network (AC-GAN) [34], with category information as a conditional input. AC-GAN modifies its discriminator to a multi-category discriminator by viewing the generated samples as a new class and training the discriminator simultaneously with the real labeled samples. HSGAN [35] uses GAN to extract spectral features with discriminators in an unsupervised manner and train classifiers in a supervised manner, and finally the test samples labels are generated by voting on their neighboring image element classes. Zhang M and Sun Q et al. used Wasserstein GANs (WGANs) [36] to extract features of hyperspectral image blocks by discriminators in an unsupervised manner, and then trained the classifier with a small number of samples. The above GAN-based methods utilize only spectral information or spatial information. Jie Feng [37] et al. combined the two and proposed a novel multi-class null-spectrum GAN method, which has two conditional generators to generate spectral samples and spatial image block samples of corresponding classes, and the discriminator takes both spectral and spatial image block samples as inputs, extracts their respective features, and then cascades them, and finally classifies them, and the generated samples are discriminated in the discriminator as not belonging to any class. These common GAN-based semi-supervised classification algorithms use discriminators to extract features for classification. Although the adversarial training of discriminators can enhance the ability of discriminators to extract features to a certain extent, the discriminators still need to force the distribution of the generated samples and real samples to be separated when the distributions are close to each other in the later stages of training, which deteriorates the classification ability of the discriminators. In general, the existing hyperspectral image classification algorithms that are based on GAN are not conducive to extracting discriminant information in the later training period because it uses discriminator as the feature extractor for classification. Although some methods expand the data by generating pseudo samples, they do not effectively use a large number of unlabeled samples to solve the problem of small samples.
In this paper, an adversarial representation learning that is based on generative adversarial networks (ARL-GAN) is proposed to extend GAN to the domain of representation learning. We used PCA to compress the raw data and added an encoder in front of the generator to combine it with the generator as a stack autoencoder. The discriminator is modified to be a multi-category discriminator and it is used instead of MSE to measure the difference between the generated image and the real image, so as to guide feature extraction and image generation through the high-level semantic information. In addition, conditional entropy is added to the objective function to increase the use of unlabeled samples in the network. This algorithm uses GAN to improve the representation capability of the encoder, and finally uses the features that are extracted by the encoder for classification, which is not affected by the above factors in the later stages of training, while training the GAN with a huge number of unlabeled samples. This makes the output features of the feature extractor are more conducive to classification, and also takes full advantage of the large amount of unlabeled data, which greatly improves the hyperspectral image classification performance under the small sample problem. The main contributions of this article are as follows.
  • We construct a hyperspectral image block generator that is based on PCA, whose input is the feature vector that is extracted by the encoder rather than the noise, and the encoder is used as the feature extractor of classification.
  • The distance of the class vector output by the discriminator replaces MSE in stacked autoencoder (SAE) to measure the error between the generated image block and the real image block, so that the features that are extracted by the encoder have more useful information for classification.
  • GAN and conditional entropy are used to improve the utilization of unlabeled data when training discriminators, solving the problem of small samples in HSI classification.
The rest of this paper is organized as following. Section 2 describes the methodology of this paper in detail. Section 3 verifies the advancement of the proposed model by comparing it with other excellent HSI classification models. Finally, Section 4 summarizes this article.

2. Proposed Method

This section is structured as follows. First, we introduce the structure of the ARL-GAN algorithm that is proposed in this paper and its corresponding parameter settings. Second, we investigate the use of spectral regularization for ARL-GAN to stabilize the training process. Third, we present the use of conditional entropy in detail. Finally, we describe the encoder E, the classifier C, the generator G, and the discriminator D of ARL-GAN.

2.1. ARL-GAN

The structure diagram of the ARL-GAN algorithm that is proposed in this paper is shown in Figure 1. ARL-GAN includes five parts. In the first part, PCA is used to extract the principal components in the spectral dimension of hyperspectral images so as to reduce redundancy and reduce the difficulty of image generation. In the second part, we extract the features of real hyperspectral image blocks using the encoder E based on a two-dimensional convolution neural network. In the third part, we generate (reconstruct) hyperspectral image blocks of the same class using the feature that is extracted by E as input using the generator G based on two-dimensional transposed convolution neural network. In the fourth part, we measure the generation errors of real hyperspectral image blocks and the generated hyperspectral image blocks by using a discriminator D, which is designed as a two-dimensional convolution neural network. The errors are calculated by the class similarity of the two image blocks. In the fifth part, we add a fully connected layer and the Softmax activation function after encoder E for classification.
In this paper, x is used to represent the samples of HSI blocks with M × N × H size, M and N are used to represent the length and width of the blocks, H is used to indicate the number of channels of the HSI after PCA operation, and y is used to represent one_hot label corresponding to the real sample x . The real labeled samples are represented as ( x L , y L ) , and the real unlabeled samples are represented as x U . In the encoding phase, x is used as input to the encoder E, and the encoded output is characterized by f . In the generation phase, the feature f that is extracted by the encoder is used as input to the generator G, and the hyperspectral image blocks that are generated by the generator are represented as x G ( f ) . The discriminator takes the resulting sample x G ( f ) , the real labeled sample ( x L , y L ) , and the real untagged sample x U as input. The real labeled samples ( x L , y L ) are used to learn the correct image block–category mapping function based on the discriminator. The unlabeled sample x U uses conditional entropy (CE) [38] to enhance discriminator classification, that is, to force the discriminator to output a class probability vector with an advantage class for the unlabeled sample. Conversely, the discriminator needs to output the class probability vectors with no dominant classes for x G ( f ) to distinguish between the real samples and the generated samples. For generators, the dominant classes output by the discriminator are required to be consistent between the generated samples and the corresponding real samples. Finally, the feature that is extracted by the encoder is classified by a fully connected layer and Softmax score.

2.2. Spectral Regularization for Stabilizing ARL-GAN Training Process

GAN is prone to pattern collapse, gradient disappearance, and gradient explosion during the training process, leading to an extremely unstable training process. Therefore, many studies have been conducted to solve this problem. For example, WGAN and WGAN-GP solve this problem by modifying the loss function by replacing the Jensen-Shannon (JS) Divergence JS in the classical GAN loss function with the Wasserstein distance and implement the required Lipschitz continuity condition at that distance in two different ways. In addition to solving this problem by modifying the loss function, some researchers have proposed to solve this problem by modifying the network structure of generative adversarial networks. For example, the deep convolution generative adversarial network (DCGAN) [39], which mitigates the problem of network training instability through a well-designed network structure. In this paper, spectral normalization (SN) [40] is used to make the discriminator satisfy the Lipschitz continuity condition to solve the above problem.
SN applies regularization in the form of spectral norms over layer parameters, so that the discriminator D has Lipschitz continuity conditions. Compared with the stable GAN approach of WGAN-GP, SN has a better effect in GAN [41]. The method that is presented in this paper requires the generation of HSI blocks, which is more difficult than the generation of hyperspectral vectors. Therefore, a spectral regularization method is used to stabilize the GAN.
In GAN, if the discriminator D is K-Lipschitz continuous, any x 1 and x 2 in image space can be represented as Equation (1), where · 2 represents the L 2 norm, K is called the Lipschitz constant of D ( x ) , and the minimum of K is called D L i p . The spectral norm definition of the matrix is shown in the Formula (2), and s u p represents the upper bound.
D ( x 1 ) D ( x 2 ) x 1 x 2 2 K , x 1 , x 2
A 2 = s u p A x 2 x 2 = σ ( A )
To analyze whether the discriminator D satisfies the Lipschitz continuity condition after SN is used, the discriminator function can be expressed as x l = a l ( W l x l 1 + b l ) , if D is a fully connected network with a non-linear activation function, each layer is recorded as θ   = { W l , b l } l = 1 L , and the parameter of D is recorded as D θ ( x 0 ) = x L . Within a small neighborhood of the value, the discriminator can be considered as a linear function, D θ ( x ) = W θ , x x + b θ , x . Formula (3) can be derived from the property of matrix norm σ ( A B ) σ ( A ) σ ( B ) .
The activation function uses ReLU, and a l ( x l 1 ) can be thought of as D i a g θ , x l x l 1 . D i a g θ , x l represents a diagonal matrix, with 1 at the non-negative corresponding position of x l 1 and 0 at the rest, so that there is σ ( D i a g θ , x l ) = D i a g θ , x l 2 1 , from which the Formulas (4) and (5) can be derived.
D θ ( x + δ ) D θ ( x ) 2 δ 2 = W θ , x δ 2 δ 2 σ ( W θ , x ) = s u p δ 0 W θ , x δ 2 δ 2
W θ , x = l = 1 L D i a g θ , x l W θ , x l
σ ( W θ , x ) l = 1 L σ ( D i a g θ , x l ) σ ( W θ , x l ) l = 1 L σ ( W l )
If SN is used for each network layer of discriminator D, the Formulas (6)–(8) can be derived, and then discriminator D satisfies the 1-Lipschitz continuous condition.
W ^ S N l = W l σ ( W l )
σ ( W ^ S N l ) = σ ( W l ) σ ( W l ) = 1
D θ ( x + δ ) D θ ( x ) 2 δ 2 σ ( W θ , x ) 1
SN is essentially a layer-by-layer singular value decomposition (SVD) process by dividing the parameter matrix of each layer by the spectral norm of the parameter matrix, but the process of solving SVD is time-consuming, so it is approximated by power iteration. The entire algorithm flow is shown in Algorithm 1.
Algorithm 1. Spectral Normalization (SN)
Step1: For   l = 1 , , L network layers, the random initialization vector u ˜ l with Gaussian distribution is used; The number of power iterations is N.
Step2:  Calculate the values of v ˜ l and   u ˜ l :
For   l = 1 , , L
For   n 1 = 1 , , N 1
v ˜ l = ( W l ) T u ˜ l / ( W l ) T u ˜ 2
u ˜ l = W l v ˜ l / W l v ˜ l 2
Step3:  Calculate spectral norm normalization parameters according to v ˜ l and   u ˜ l :   W ¯ SN l ( W l ) = W l / σ ( W l ) , where σ ( W l ) = u ˜ l T W l v ˜ l .
Step4: On the training dataset D M = { ( x i 1 , y i 1 ) , , ( x i k , y i k ) } of mini-batch, update the
 parameter W l with the learning rate α :
W l = W l - α W l ( W ¯ SN l ( W l ) , D M )

2.3. Conditional Entropy

The cost of labeled samples in HSI datasets is high and the available labeled samples are limited, so the introduction of unlabeled samples into the deep learning algorithm needs to be considered to enhance the classification ability of the model. Classical GAN is an unsupervised training method, but the generated samples have a poor enhancement effect on the classifier, so both CGAN and ACGAN introduce category information as additional information to constrain GAN. The algorithm in this paper also introduces category information by changing the discriminator to a multi-classification discriminator. For unlabeled samples without category information, the algorithm in this paper exploits it by adding conditional entropy to the objective function.
To enhance the ability of the generator to generate samples, the discriminator needs to have a strong ability to discriminate between true and false for unlabeled data. However, the binary classification objective function of the classical GAN does not force a dominant class, and only needs to satisfy Equation (9) to obtain the decision boundary of the correct data source, while the class probability P D ( y c | x ) of the real unlabeled samples may be evenly distributed. In order to ensure the discriminator has strong ability to distinguish true from false, a conditional entropy is added to the algorithm in this paper. At the same time, in practice, the real labeled samples do not know the specific label, but they should belong to a class, so conditional entropy is added as an a priori condition in the classifier to enhance the performance of the classifier. The mathematical expression of conditional entropy is shown in Equation (10). When the probability of one class is 1 and the predicted probability of other classes is 0, the conditional entropy obtains the maximum value and when the predicted probability of each class is equal, the conditional entropy obtains the minimum value.
c = 1 C ¯ P D ( y c | x ) > P D ( y C ¯ + 1 | x )
C E = E x U c = 1 C ¯ p C ( y c | x ) log p C ( y c | x )

2.4. Encoder E and Classifier C of ARL-GAN

The role of the encoder E is to map the input hyperspectral image block x L U from the input space to the feature space f . The encoder maps the hyperspectral image block with original size of M × N × H to a feature vector with size of 1 × 1 × F by four-step convolution. ReLU is used as the nonlinear activation function for each layer. The process of extracting the features by the encoder can be represented as Equation (11).
f = E ( x L U )
The general semi-supervised classification algorithm that is based on GAN uses a discriminator to extract features for classification. Although the adversarial training of the discriminator can enhance the ability of the discriminator to extract features to a certain extent, the discriminator still needs to force the distribution of the generated samples and the real samples to be separated when the distribution is close to each other in the later stage of training, which affects the classification ability of the discriminator. In contrast, the algorithm in this paper uses GAN to improve the representation capability of the encoder, and ultimately uses the features that are extracted by the encoder for classification, which is not affected by the above factors in the later stages of training. Compared with the semi-supervised SAE algorithm without the antagonistic training, the feature that is extracted by the encoder in this paper is more suitable for classification tasks. The classifier adds a full connection layer with weight normalization (WN) [42] to the encoder E. The input image block’s class probability distribution p C ( y c | x ) is obtained by passing the output of the full connection layer through the Softmax, as shown in Formula (12). The objective function of classifier C can be expressed as Formula (13).
p C ( y c | x ) = softmax ( W n _ F c ( f ) )
max L C = E ( x L , y L ) log p C ( y c | x , y ) + λ E x U c = 1 C ¯ p C ( y c | x ) log p C ( y c | x )
In (13) the first term represents the cross-entropy that is calculated from the labeled samples, the second term represents the conditional entropy that is calculated from the unlabeled samples, and λ is the weight of the conditional entropy. As whether the predicted result of the classifier is correct or not, the conditional entropy will promote the network to further enhance the current prediction. If the classifier is immature and has poor recognition ability, it will lead to the degradation of classification. Therefore, when the number of iterations is low in the early stages of training, λ is small. When the classifier is fully trained with the increase of training steps, the accuracy of the prediction results increases, and the weight λ should also increase with the increase of steps.

2.5. Generator G and Discriminator D of ARL-GAN

The generator consists of four transposed convolution layers, which take the feature that is extracted by the encoder with a size of 1 × 1 × F as input to generate (rebuild) an image block with a size of M × N × H , using ReLU as the activation function and Tanh as the activation function in the output layer to match the range of the true hyperspectral data. The discriminator is designed by ten layers of convolution layers that are normalized by spectral norms and full connection layer normalized by weights. Because the discriminator is not only used to correctly classify the real labeled samples, but also to determine which is the generated sample and which is the real sample, the task is heavier than the classifier, so more network layers and more network parameters are used. The discriminator uses LeakyReLU as the activation function except for the final fully connected layer, which uses the Softmax function. The arrows represent the direction of data flow.
Generator G is used to generate hyperspectral image blocks using the features that are extracted by the encoder E as input. The stacked autoencoder uses the root mean square error to calculate the reconstruction error so that the input image block and the reconstructed image block are identical at the pixel level. The encoder of SAE can be considered as a non-linear dimension reduction algorithm. Although the redundancy of data is reduced, the ability of the encoder to extract features is improved to a certain extent by. However, the features that are extracted by this encoder are not completely suitable for classification tasks. In this algorithm, MSE is replaced by a discriminator, which takes the generated image block and the real image block as input, and then constrains of the generator according to the advanced semantic information of category, making the generated image block consistent with the corresponding real image block category, thus making the features and categories extracted by the encoder highly relevant. The extracted features contain the main information of the class and remove the remaining noise information. The encoder and generator are trained with a discriminator to further enhance the ability of encoder and generator.
The generator’s target function is shown in Equation (14), where p D ( y c | x ) represents the predictive probability of the discriminator, G ( f ) represents the generated data, p D ( y c | x L U ) denotes the class probability of the discriminator for the sample output, and p D ( y c | x G ( f ) ) denotes the class probability of the discriminator for the generated sample output. This objective function can constrain the generation process so that the generated image block and the real image block belongs to the same class, rather than making them identical at the pixel level. The generators and encoders are a complete generation network when updating parameters, so the parameters of encoders E and G are updated during reverse propagation.
min L G , E = E x ( 1 C ¯ c = 1 C ¯ ( p D ( y c | x L U ) p D ( y c | x G ( f ) ) ) 2 )
The function of the discriminator is mainly to distinguish the generated image blocks from the real data blocks. However, unlike the classical GAN which uses binary classification to discriminate, the discriminator D in the proposed method is a multi-class discriminator. The objective function of D is shown in Equation (15). The objective function of the discriminator contains three terms, the first term is the cross-entropy for the true labeled samples; the second term is the conditional entropy for the real labeled samples, which gives the discriminator an advantage class for the class probability of the output of the real labeled samples; and the third term is the objective function for generating image blocks. Unlike other semi-supervised GAN algorithms, the generated samples are considered as a new class [43,44]. The label of the sample that is generated by this algorithm is y G = [ 1 / C ¯ , 1 / C ¯ ] where C ¯ is the number of classes, which is also contrary to the true unlabeled sample, which does not have any dominant classes and outputs a uniform class probability distribution.
max L D = E ( x L , y L ) log p D ( y c | x , y ) + λ E x U c = 1 C ¯ p D ( y c | x ) log p D ( y c | x ) + E x G ( f ) c = 1 C ¯ y G ( c ) log ( p D ( y c | x ) )
The optimization process of encoder E, classifier C, generator G, and discriminator D in the whole network uses an alternating optimization strategy. When updating the parameters, G and E are considered as a whole generation network. When C and G(E) network parameters are fixed, D’s network parameters are optimized. Conversely, when D is frozen, the parameters of the remaining networks are updated. All the networks are optimized using the Adam optimization algorithm. Through this end-to-end network structure and alternate optimization, the test samples are finally classified by classifier C.

3. Experiments and Result Analysis

In this section, three well-known HSI datasets are described, then we introduce the implementation details of the algorithm in this paper and compare it with other excellent methods. The effectiveness of adversarial learning and conditional entropy are analyzed experimentally and a sensitivity analysis with the number of labeled samples is performed. Finally, a comparison with existing GAN-based hyperspectral image classification algorithms is performed to evaluate the HSI image classification performance of the proposed algorithm.

3.1. Hyperspectral Datasets

In this experiment, three widely-used datasets were used and they are Indian Pines dataset, Pavia University dataset (PaviaU), and Salinas. Figure 2a–c shows the false-color image and ground truth of Indian Pines, Pavia University, and Salinas dataset, respectively.
The Indian Pines dataset was gathered by the AVIRIS sensor in the northeast of Indiana. It is 145 × 145 in size and has a spatial resolution of 20 m per pixel. A total of 20 bands were removed due to water absorption, leaving 200 bands for our experimental analysis, with a spectral range of 400 to 2500 nm. This dataset contains 16 categories and background information.
The PaviaU dataset was gathered in northern Italy by the Reflective Optics System Imaging Spectrometer. Its spatial size is 610 × 340 and the resolution is 1.3 m per pixel. It has 103 bands, which range from 430 to 860 nm.
The Salinas dataset was gathered by the AVIRIS sensor over the Salinas Valley. The dataset is of size 512 × 217 × 224, and it has a spatial resolution of 3.7 m per pixel with 16 land-cover classes.
Real samples were divided into a training set and a test set by random sampling. A total of 5% samples were selected in Indian Pines, 1% samples were selected from the Salina dataset, and 3% samples were selected from the PaviaU dataset. The values of all HSIs datasets are normalized between −1 and 1, and then the spectral dimension is reduced to 10 dimensions by the PCA algorithm. The size of the input image block is 32 × 32 × 10 in Indian Pines and Salinas, and the size of the input image block is 8 × 8 × 10 in PaviaU, because PaviaU is a collection of hyperspectral image data for urban areas, while Indian Pines and Salinas are a hyperspectral dataset for agricultural areas, where the distribution of land objects is more complex and homogeneous. Therefore, the size of the input image block that was selected by this algorithm on the PaviaU dataset is small. In the training, the number of training samples per small batch is set to 32; the learning rate is set by simulated annealing algorithm, the learning rate range is 0.0 to 0.002; the conditional entropy weight λ uses step parameters, the range is 0.5 to 1, and the λ increases by 0.05; and the total number of training steps is 1000 for each 100 step increase in training steps. The experiments in this paper were run 10 times independently, and each time the training set and test set were randomly divided, and the average and standard deviation of the experimental results were calculated. All the experiments in this paper are run on an Nvidia 1080ti graphics card. Moreover, Table 1, Table 2 and Table 3 listed the land cover type and the total number of samples in the three HSI datasets.
In order to quantitatively evaluate the experimental results, three common evaluation criteria are used: OA, AA, and Kappa coefficients. The definitions of all evaluation criteria are shown as follows:
(1) OA: OA assesses the proportion of correctly identified samples to all the samples.
OA = i = 1 C ¯ h i i N
where N is the total number of labeled samples, h i i represents the number of class i samples divided into class i, and C ¯ is the total number of categories.
(2) AA: AA represents the mean of the percentage of the correctly identified samples.
AA = 1 C ¯ i = 1 C ¯ h i i N i
where C ¯ is the total number of categories, h i i represents the number of samples of category i divided into category i, and N i represents the number of samples of category i.
(3) Kappa: Kappa coefficient denotes the interrater reliability for categorical variables.
Kappa = N i C ¯ h i i i C ¯ ( h i + · h + i ) N 2 i C ¯ ( h i + · h + i )
where h i + and h + i , respectively, represent the total number of samples of category i true category and the number of samples predicted to be category i.

3.2. Visual Analysis of Generated Image Blocks

In order to display the generated image blocks conveniently, the original dataset is first reduced to 3-dimensional using PCA, and the last output channel of the generated network is changed from 10 to 3 to generate three-channel image blocks as shown in Figure 3.
The real image blocks that are drawn with pseudo color images and the corresponding generated image blocks are also provided. By comparing the real image block and the generated image block, we can see that they are not exactly the same. The generated image block and the real image block are very similar in the central pixel part, but the surrounding pixels are not exactly the same because the generation model of the algorithm in this paper only makes the two categories consistent, not on the pixels, so this phenomenon meets the theoretical requirements of the algorithm.

3.3. Numerical Results and Visualization of HSI Classification

To evaluate the performance of this classification algorithm, this section compares it with other algorithms. In this paper, six representative hyperspectral image classification algorithms are selected, which are RBF-SVM [9], SAE [17], DBN [45], CNN [46], PPFCNN [26], and 3DCNN [47]. RBF-SVM is a classical SVM method that is based on radial basis function and the other methods are based on deep learning algorithm. The SVM in the RBF-SVM algorithm extends it from a two-class model to a multi-class model through a one-to-many strategy. The window size and other parameter settings in SAE, DBN, CNN, PPFCNN, and 3DCNN are the optimal experimental settings that are directly used in the original papers. The spatial dimensions of PPFCNN sample pairs are also set according to the original paper. The window size in a 3DCNN is determined by searching for the optimal size within a range of 5 to 33.
The classification results mainly include two parts, one are the numerical classification results, the other is the visual classification results. These two classification results on the Indian Pines, PaviaU, and Salinas datasets are shown, respectively.
Indian Pines: The numerical classification results for the Indian Pines dataset are shown in Table 4. The first three lines show AA, OA, and Kappa coefficients that were calculated from all samples. The last 16 rows record the classification accuracy of the each classes. Under the three evaluation criteria of AA, OA, and Kappa coefficients, the proposed algorithm performs better than the other six algorithms. Compared with RBF-SVM, the classification accuracy of the other five classification algorithms based deep learning has been greatly improved, which shows that the deep learning algorithm can extract distinguished features more effectively. PPF-CNN is more effective than CNN in this dataset because the number of training samples is increased and the classification accuracy of the model is improved by using pairs of samples. Compared with SAE, this algorithm increases the OA, AA, and Kappa coefficients by 16.4%, 25.6%, and 18.7%, respectively. This shows that the features that are extracted by the encoder in this algorithm are more suitable for classification. Three-dimensional convolution can effectively use the spatial-spectral information to extract the combined features of spatial spectrum and improve the classification accuracy. Compared with three-dimensional CNN, the algorithm in this paper improves the OA, AA, and Kappa coefficients by 5.5%, 5.6%, and 6.1%, respectively. The optimal classification results for each class are represented by bold numbers in the table. There are 13 classes (16 classes in total) of the algorithm that have obtained the optimal classification results. At the same time, it can be found that RBF-SVM and PF-CNN have zero classification accuracy for the seventh and ninth classes because the number of samples is much less than other classes. The accuracy of the algorithms in this paper is 82.9 ± 16.2 and 92.0 ± 16.0, respectively. At the same time, the proposed method achieves the optimal value on AA. Therefore, the algorithm in this paper can effectively handle the extreme sample imbalance problem on the Indian Pines dataset and achieve good results.
Figure 4 shows the visual classification results of the different algorithms on the Indian Pines dataset. First, we can see that the RBF-SVM, DBN, and other spectral classification algorithms only use spectrum to classify. There are a lot of noise points in the classification results, while the 3DCNN and the proposed method both use spatial information and have relatively few noise points. The classification map of the 3DCNN contains a large number of error points at the boundary. The algorithm in this paper has fewer error points than that of 3DCNN for those pixels that are close to the boundary. It is concluded that the algorithm that is presented in this paper can reduce the noise points and maintain the classification boundary better than other algorithms.
PaviaU: The classification results of the PaviaU dataset are shown in Table 5. Under the three evaluation criteria of AA, OA, and Kappa coefficient, the classification effect of the algorithm that is proposed in this paper is better than the other six algorithms. Consistent with the conclusion for the Indian dataset, the classification effect of the six classification algorithms that are based on deep learning are better than the traditional algorithm that is based on SVM. The classification effect of 3DCNN on PaviaU is lower than that of PPF-CNN, which is different from the conclusion on Indian Pines dataset. It can be seen that these two algorithms are more sensitive to datasets. The algorithm in this paper has achieved the best classification effect on these two datasets as it shows it is less sensitive to datasets than other algorithms and is easy to be transplanted to different datasets for classification. Compared with the SAE algorithm on this dataset, the OA, AA, and Kappa coefficients of the proposed ARL-GAN have increased by 7.4%, 10.3%, and 9.8%, respectively. Since the classification accuracy of all algorithms on PaviaU is higher than that for the Indian Pines dataset, the growth rate on this dataset is lower than that for the Indian Pines dataset, However, it can also be verified that the features that were extracted by the encoder through the algorithm in this paper are more suitable for classification. Among the six comparison algorithms, PPF-CNN achieved the best classification results on the dataset. Compared with PPF-CNN, the OA, AA, and Kappa coefficients of this algorithm were improved by 2.9%, 4.5%, and 3.7%, respectively. At the same time, our algorithm achieves the best classification accuracy on all classes. Figure 5 shows the visual classification results of different algorithms on the PaviaU dataset, from which we can obtain the similar conclusion for the Indian Pines dataset. The visual classification results of the algorithm in this paper have less noise points, and shows better performance for those boundary samples.
Salinas: The classification results of the Salinas dataset are shown in Table 6. Firstly, from the three evaluation criteria of AA, OA, and Kappa coefficients, the classification effect of the algorithm that is proposed in this paper is better than the other six algorithms. Consistent with the conclusions from the the first two datasets, the classification effect of the six classification algorithms that are based on deep learning is better than the traditional methods based on SVM. On Salinas, the classification effect of 3DCNN is higher than that of PPF-CNN, which is consistent with the conclusion for the Indian Pines dataset. Therefore, based on the results of the three datasets, the classification result of 3DCNN is better than that of PPF-CNN as a whole thanks to its comprehensive utilization of spatial-spectral information by means of three-dimensional convolution. Compared with the SAE algorithm, the algorithm in this paper increases the OA, AA, and Kappa coefficients by 8.4%, 5.0%, and 9.3%, respectively. For the three datasets, the algorithm in this paper has a greater improvement than the SAE algorithm in the three evaluation index values, so it further proves that the features that are extracted by the encoder of this paper algorithm are more suitable for classification. Among the six algorithms, 3DCNN achieves the best classification results on the dataset. Compared with 3DCNN, the proposed method improves the OA, AA, and Kappa coefficients by 2.0%, 1.9%, and 2.0%, respectively. The algorithm in this paper achieves the optimal classification accuracy in all classes, with 100% accuracy in 10 classes and nearly 100% accuracy in the other 6 classes.
Figure 6 is the visual classification result of different algorithms for the Salinas dataset. First, it can be seen that RBF-SVM, DBN, SAE, CNN, and PPF-CNN classify many of the eighth-class samples into fifteen categories, and many of the fifteenth-class samples into eighth categories. This is due to the problems of “synonyms spectrum” and “foreign body with the spectrum” in hyperspectral images. Compared with the five algorithms, 3DCNN has fewer points of error. This is because it uses the joint information of the spatial spectrum to reduce the impact of this problem. Compared with these algorithms, the algorithm in this paper can well distinguish between Class 8 and Class 15 samples, extract more distinguishing features, and achieve better classification results.
The above experimental results show that compared with 3DCNN, the algorithm in this paper generally has fewer misclassification points at the boundary, and, compared with PPF-CNN, the algorithm in this paper has fewer misclassification points inside the region. The reason for this is that 3DCNN uses 3D convolutional networks, and the network model needs to be trained with more labeled samples in order to obtain excellent classification performance while PPF-CNN extends the training dataset by reorganizing the existing training samples in pairs and relabeling them but ignores the use of unlabeled samples. However, due to the addition of Conditional Entropy and GAN, the algorithm in this paper makes full use of labeled samples and a large amount of unlabeled sample data to enhance the feature extraction ability of the encoder.

3.4. Effectiveness Analysis of Adversarial Learning and Conditional Entropy

In order to verify the effectiveness of the conditional entropy loss function and adversarial learning that is proposed by the algorithm in this paper, the classification results of CNN, CNN-CE, and the algorithm in this paper are compared for the three datasets. The CNN classifiers in CNN and CNN-CE adopt the same network structure and consistent parameters as the classifiers in this paper. CNN-CE represents the algorithm model of semi supervised training with conditional entropy. Table 7 lists the classification results of the three algorithms on OA, AA, and Kappa coefficients. As shown in the table, compared with CNN and CNN-CE, the proposed method improved the OA by 0.42%, 0.89%, and 0.84%, respectively. It can be seen that semi-supervised training through conditional entropy can effectively improve the classification effect.
As shown in the table, compared with CNN-CE, the algorithms in this paper have improved by 0.79%, 0.05%, and 0.01% respectively on the three datasets under the OA evaluation standard. The improvement rate of confrontation learning on the Indian dataset is relatively large, and the improvement is not large on the other two datasets because the training samples of the Indian Pines dataset are more limited, and the dataset is more noisy and more difficult to be classified so there is more room for improvement. The other two datasets have sufficient samples and are much easier to be classified, so there is less room for improvement. Through the overall analysis of the three datasets, the conditional entropy loss function and confrontation learning can improve the classification effect to a certain extent. The reason is that conditional entropy can constrain the classifier, train the network with a large number of unlabeled samples, and enhance the feature extraction ability of the network for hyperspectral data. Adversarial learning can improve both the generative performance of the generator and the discriminant ability of the discriminator, as well as the final classification performance.

3.5. Number Sensitivity Analysis of Labeled Samples

The method that is based on deep learning is highly dependent on labeled training samples, and the number of labeled training samples will greatly affect the classification performance. Therefore, it is necessary to study the sensitivity of the classification algorithm to the number of labeled training samples. Figure 7, Figure 8 and Figure 9 show the overall classification accuracy (OA) of seven algorithms on three datasets: Indian Pines, PaviaU, and Salinas with different numbers of training samples; the horizontal coordinate is the number of labeled samples. In the experiment, the number of labeled training samples on the Indian Pines dataset ranged from 1% to 9% with an interval of 2%. The number of labeled training samples in the PaviaU dataset ranged from 1% to 5% with an interval of 1%. The number of labeled training samples in the Salinas dataset ranged from 1% to 3%, with an interval of 0.5%. For the Indian Pines dataset, the classification accuracy of the seven algorithms decreases significantly with the decrease of the number of training samples. However, compared with the other six algorithms, the algorithm in this paper has the lowest reduction rate, with 85% classification accuracy even when only 1% of the training samples are available. The other six algorithms have less than 60% classification accuracy except 3DCNN. The algorithm that is presented in this paper can effectively handle small sample problems on the Indian Pines dataset. Among the six comparison algorithms, PPF-CNN, and 3DCNN achieve better classification results because they use the combination of sample pair and spatial-spectral to effectively reduce the impact of data volume on the classification results. The conclusion for the Pavia University dataset is basically the same as that for the Indian Pines dataset. The classification accuracy of the algorithms in this paper was more than 99% on the Salinas dataset, with no significant decrease. At the same time, the algorithm in this paper obtains optimal classification results on three datasets.
Therefore, the algorithm in this paper is a better choice when there are a limited number of labeled samples. This result is attributed to the full utilization of the large amount of unlabeled data by the algorithm in this paper and the excellent design of the classifier, which can obtain a good classification accuracy even when there are very few labeled samples.

3.6. Comparison with Existing Hyperspectral Image Classification Algorithms Based GAN

In this section, three generative adversarial networks that are based HSI classification methods are also introduced for comparison. The comparison results are shown in Table 8.
3DGAN [18] uses image blocks as training samples and uses antagonistic learning to improve the discriminator’s classification ability. This paper also uses image blocks as training samples, but uses the antagonistic network to perform representation learning and uses the features that are extracted by the encoder to classify. The algorithm in this paper improves the OA index of Indian Pines, PaviaU, and Salinas datasets by 31.5%, 1.63%, and 1.07%, respectively, compared with 3DGAN. HSGAN [22] uses one-dimensional spectral samples as training samples, uses GAN to extract spectral features with discriminators in an unsupervised manner, trains classifiers in a supervised manner, and finally votes for the categories of the test samples from their neighboring cells. Compared with the HSGAN algorithm in this paper, OA improves by 24.15%, 14.13%, and 11.67% on three datasets, respectively. Using image blocks as training samples can greatly improve the classification accuracy of the algorithm. MSGAN uses two generators and a discriminator to form a variety of spatial spectrum generation adversarial networks hyperspectral image classification. The two generators generate spectral samples and spatial image block samples respectively. The discriminator fuses the spatial spectrum features to improve the discriminator’s classification performance in an end-to-end manner. MSGAN uses spatial spectrum information and is more accurate than 3DGAN and HSGAN. Although the algorithm in this paper also utilizes spatial and partial spectral information, the classification effect is better than MSGAN, and the OA is improved by 2.65%, 0.63%, and 0.87% for the three datasets, respectively, which shows that the algorithm in this paper can improve the classification accuracy more effectively by using the GAN approach. The reason is that the algorithm adds the encoder as the final classifier and modifies the discriminator to guide the feature extraction of the encoder, which is more suitable for the classification task than the previous algorithms that are based on GAN that uses the discriminator as the final classifier. In addition, with the addition of conditional entropy, the algorithm makes full use of the information of unlabeled samples, so it is more suitable for improving the problem of small samples.

4. Conclusions

In this paper, for the problem of small samples in hyperspectral image classification tasks, an HSI classification algorithm that is based on adversarial representation learning is proposed, which introduces unlabeled samples and uses conditional entropy and GAN for semi-supervised learning. The algorithm also extending GAN to the field of representation learning, and uses adversarial learning to improve the ability of the encoder to extract features that are more suitable for the classification task. In the experimental part of this paper, the effectiveness of the algorithm is verified by experiments. For the three datasets, the classification accuracy that is obtained by the algorithm is 98.3%, 99.8%, and 99.9%, respectively. Compared with the optimal results of other algorithms, the algorithm in this paper increases by 5.5%, 2.9%, and 2.0%, respectively. Conditional entropy is used as constraint to enhance the performance of the classifier. However, conditional entropy also increases the error prediction of classifier. The proposed algorithm reduces the adverse effects of conditional entropy by manually-designed step weights. In our future work, we will try to select appropriate weights adaptively and adopt other regularization constraints. In the process of training, the model training speed of this method is relatively slow.

Author Contributions

S.Z. wrote the whole article and adjusted it. X.Z. directed the completion of the algorithm and put forward key suggestions. T.L. adjusted the structure of the article and checked out some mistakes. H.M. reviewed the article and made some suggestions. X.C. reviewed the whole paper carefully and gave some key suggestions on the format and writing style of the paper. L.W. checked out some grammatical errors in the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61877066, Aero-Science Fund under Grant 20175181013 and Science and technology plan project of Xi’an under Grant 2021JH-05-0052.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nasrabadi, N.M. Hyperspectral target detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2013, 31, 34–44. [Google Scholar] [CrossRef]
  2. Lacar, F.M.; Lewis, M.M.; Grierson, I.T. Use of hyperspectral imagery for mapping grape varieties in the Barossa Valley, South Australia. In Proceedings of the Geoscience and Remote Sensing Symposium (IGARSS), Sydney, NSW, Australia, 9–13 July 2001. [Google Scholar]
  3. Brown, A.J.; Walter, M.R.; Cudahy, T.J. Hyperspectral imaging spectroscopy of a Mars analogue environment at the North Pole Dome, Pilbara Craton, Western Australia. J. Geol. Soc. Aust. 2005, 52, 353–364. [Google Scholar] [CrossRef]
  4. Meer, F.V.D. Analysis of spectral absorption features in hyperspectral imagery. Int. J. Appl. Earth Obs. Geoinf. 2004, 5, 55–68. [Google Scholar]
  5. Hege, E.K.; O‘Connell, D.; Johnson, W.; Basty, S.; Dereniak, E.L. Hyperspectral imaging for astronomy and space surveillance. In Proceedings of the Optical Science and Technology, San Diego, CA, USA, 3–8 August 2003; pp. 380–391. [Google Scholar]
  6. Carrasco, O.; Gomez, R.B.; Chainani, A.; Roper, W.E. Hyperspectral imaging applied to medical diagnoses and food safety. In Proceedings of the SPIE Aerosense Conference (Aerospace and Defense Sensing and Control Technology Symposium), Orlando, FL, USA, 21–25 April 2003; pp. 215–221. [Google Scholar]
  7. Yuen, P.W.; Richardson, M. An introduction to hyperspectral imaging and its application for security, surveillance and target acquisition. Imaging Sci. J. 2010, 58, 241–253. [Google Scholar] [CrossRef]
  8. Samaniego, L.; Bárdossy, A.; Schulz, K. Supervised classification of remotely sensed imagery using a modified k-NN technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  9. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  10. Kumar, S.; Ghosh, J.; Crawford, M.M. Best-bases feature extraction algorithms for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2001, 39, 1368–1379. [Google Scholar] [CrossRef] [Green Version]
  11. Li, W.; Chen, C.; Su, H.; Du, Q. Local binary patterns and extreme learning machine for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3681–3693. [Google Scholar] [CrossRef]
  12. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Marí, J.; Vila-Francés, J.; Calpe-Maravilla, J. Composite kernels for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2006, 3, 93–97. [Google Scholar] [CrossRef]
  13. Ghamisi, P.; Benediktsson, J.A.; Ulfarsson, M.O. Spectral–spatial classification of hyperspectral images based on hidden Markov random fields. IEEE Trans. Geosci. Remote Sens. 2013, 52, 2565–2574. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, H.; Li, J.; Huang, Y.; Zhang, L. A nonlocal weighted joint sparse representation classification method for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 2056–2065. [Google Scholar] [CrossRef]
  15. Fu, W.; Li, S.; Fang, L.; Kang, X.; Benediktsson, J.A. Hyperspectral image classification via shape-adaptive joint sparse representation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 9, 556–567. [Google Scholar] [CrossRef]
  16. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  18. Hinton, G.E.; Osindero, S.; Teh, Y.-W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  19. Le, J.H.; Yazdanpanah, A.P.; Regentova, E.E.; Muthukumar, V. A deep belief network for classifying remotely-sensed hyperspectral data. In Proceedings of the International Symposium on Visual Computing, Las Vegas, NV, USA, 14–16 December 2015; pp. 682–692. [Google Scholar]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems 2012, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  21. Mustaqeem; Kwon, S. 1D-CNN: Speech emotion recognition system using a stacked network with dilated CNN features. CMC Comput. Mater. Contin. 2021, 67, 4039–4059. [Google Scholar] [CrossRef]
  22. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef] [Green Version]
  23. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  24. Yang, J.; Zhao, Y.; Chan, J.C.-W.; Yi, C. Hyperspectral image classification using two-channel deep convolutional neural network. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5079–5082. [Google Scholar]
  25. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef] [Green Version]
  26. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2016, 55, 844–853. [Google Scholar] [CrossRef]
  27. Lin, J.; He, C.; Wang, Z.J.; Li, S. Structure preserving transfer learning for unsupervised hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1656–1660. [Google Scholar] [CrossRef]
  28. Persello, C.; Bruzzone, L. Kernel-based domain-invariant feature selection in hyperspectral images for transfer learning. IEEE Trans. Geosci. Remote Sens. 2015, 54, 2615–2626. [Google Scholar] [CrossRef]
  29. Sun, Z.; Wang, C.; Li, D.; Li, J. Semisupervised classification for hyperspectral imagery with transductive multiple-kernel learning. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1991–1995. [Google Scholar]
  30. Aydemir, M.S.; Bilgin, G. Semisupervised hyperspectral image classification using small sample sizes. IEEE Geosci. Remote Sens. Lett. 2017, 14, 621–625. [Google Scholar] [CrossRef]
  31. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Annual Conference on Neural Information Processing Systems 2014, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  32. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative adversarial networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  33. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  34. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. In Proceedings of the International Conference on Machine Learning, Sydney, NSW, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  35. Zhan, Y.; Wu, K.; Liu, W.; Qin, J.; Yang, Z.; Medjadba, Y.; Wang, G.; Yu, X. Semi-supervised classification of hyperspectral data based on generative adversarial networks and neighborhood majority voting. In Proceedings of the IGARSS IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 23–27 June 2018; pp. 5756–5759. [Google Scholar]
  36. Arjocvsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. [Google Scholar]
  37. Feng, J.; Yu, H.; Wang, L.; Cao, X.; Zhang, X.; Jiao, L. Classification of hyperspectral images based on multiclass spatial–spectral generative adversarial networks. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5329–5343. [Google Scholar] [CrossRef]
  38. Springenberg, J.T. Unsupervised and semi-supervised learning with categorical generative adversarial networks. arXiv 2015, arXiv:1511.06390. [Google Scholar]
  39. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  40. Yoshida, Y.; Miyato, T. Spectral norm regularization for improving the generalizability of deep learning. arXiv 2017, arXiv:1705.10941. [Google Scholar]
  41. Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y. Spectral normalization for generative adversarial networks. arXiv 2018, arXiv:1802.05957. [Google Scholar]
  42. Salimans, T.; Kingma, D.P. Weight normalization: A simple reparameterization to accelerate training of deep neural networks. In Proceedings of the Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  43. Yin, J.; Li, W.; Han, B. Hyperspectral image classification based on generative adversarial network with dropblock. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 405–409. [Google Scholar]
  44. Zhao, W.; Chen, X.; Bo, Y.; Chen, J. Semisupervised hyperspectral image classification with cluster-based conditional generative adversarial net. IEEE Geosci. Remote Sens. Lett. 2019, 17, 539–543. [Google Scholar] [CrossRef]
  45. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  46. Yu, S.; Jia, S.; Xu, C. Convolutional neural networks for hyperspectral image classification. Neurocomputing 2017, 219, 88–98. [Google Scholar] [CrossRef]
  47. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Structure diagram of the ARL-GAN algorithm.
Figure 1. Structure diagram of the ARL-GAN algorithm.
Remotesensing 14 02612 g001
Figure 2. False-color composites (first row) and ground truths (second row) of experimental HSI datasets. Each color represents one kind of object. (a) Indian Pines; (b) PaviaU; (c) Salinas.
Figure 2. False-color composites (first row) and ground truths (second row) of experimental HSI datasets. Each color represents one kind of object. (a) Indian Pines; (b) PaviaU; (c) Salinas.
Remotesensing 14 02612 g002
Figure 3. Visualization of the real image blocks (a) and the corresponding resulting image block (b).
Figure 3. Visualization of the real image blocks (a) and the corresponding resulting image block (b).
Remotesensing 14 02612 g003
Figure 4. Visualization of the classification results on the Indian Pines dataset. (a) Ground Truth. (b) RBF-SVM. (c) DBN. (d) SAE. (e) CNN. (f) 3DCNN. (g) PPF-CNN. (h) ARL-GAN.
Figure 4. Visualization of the classification results on the Indian Pines dataset. (a) Ground Truth. (b) RBF-SVM. (c) DBN. (d) SAE. (e) CNN. (f) 3DCNN. (g) PPF-CNN. (h) ARL-GAN.
Remotesensing 14 02612 g004
Figure 5. Visualization of the classification results on the PaviaU dataset. (a) Ground Truth. (b) RBF-SVM. (c) DBN. (d) SAE. (e) CNN. (f) 3DCNN. (g) PPF-CNN. (h) ARL-GAN.
Figure 5. Visualization of the classification results on the PaviaU dataset. (a) Ground Truth. (b) RBF-SVM. (c) DBN. (d) SAE. (e) CNN. (f) 3DCNN. (g) PPF-CNN. (h) ARL-GAN.
Remotesensing 14 02612 g005
Figure 6. Visualization of the classification results for the Salinas dataset. (a) Ground Truth. (b) RBF-SVM. (c) DBN. (d) SAE. (e) CNN. (f) 3DCNN. (g) PPF-CNN. (h) ARL-GAN.
Figure 6. Visualization of the classification results for the Salinas dataset. (a) Ground Truth. (b) RBF-SVM. (c) DBN. (d) SAE. (e) CNN. (f) 3DCNN. (g) PPF-CNN. (h) ARL-GAN.
Remotesensing 14 02612 g006
Figure 7. Accuracy of seven classification algorithms for the Indian Pines dataset under different numbers of training samples.
Figure 7. Accuracy of seven classification algorithms for the Indian Pines dataset under different numbers of training samples.
Remotesensing 14 02612 g007
Figure 8. Accuracy of seven classification algorithms for the PaviaU dataset under different numbers of training samples.
Figure 8. Accuracy of seven classification algorithms for the PaviaU dataset under different numbers of training samples.
Remotesensing 14 02612 g008
Figure 9. Accuracy of seven classification algorithms for the Salinas dataset under different numbers of training samples.
Figure 9. Accuracy of seven classification algorithms for the Salinas dataset under different numbers of training samples.
Remotesensing 14 02612 g009
Table 1. Land cover types and total number of samples in the Indian Pines dataset.
Table 1. Land cover types and total number of samples in the Indian Pines dataset.
No.ColorNameNumberNo.ColorNameNumber
1 Remotesensing 14 02612 i001Alfalfa469 Remotesensing 14 02612 i002Oats20
2 Remotesensing 14 02612 i003Corn-notill142810 Remotesensing 14 02612 i004Soybean-notill972
3 Remotesensing 14 02612 i005Corn-mintill83011 Remotesensing 14 02612 i006Soybean-mintill2455
4 Remotesensing 14 02612 i007Corn23712 Remotesensing 14 02612 i008Soybean-clean593
5 Remotesensing 14 02612 i009Grass-pasture48313 Remotesensing 14 02612 i010Wheat205
6 Remotesensing 14 02612 i011Grass-trees73014 Remotesensing 14 02612 i012Woods1265
7 Remotesensing 14 02612 i013Grass-pasture-mowed2815 Remotesensing 14 02612 i014Buildings-Grass-Trees386
8 Remotesensing 14 02612 i015Hay-windrowed47816 Remotesensing 14 02612 i016Stone-Steel-Towers93
Total Numbers10,249
Table 2. Land cover types and total number of samples in the PaviaU dataset.
Table 2. Land cover types and total number of samples in the PaviaU dataset.
No.ColorNameNumberNo.ColorNameNumber
1 Remotesensing 14 02612 i017Asphalt66316 Remotesensing 14 02612 i018Bare Soil5029
2 Remotesensing 14 02612 i019Meadows18,6497 Remotesensing 14 02612 i020Bitumen1330
3 Remotesensing 14 02612 i021Gravel20998 Remotesensing 14 02612 i022Self-Blocking Bricks3682
4 Remotesensing 14 02612 i023Trees30649 Remotesensing 14 02612 i024Shadows947
5 Remotesensing 14 02612 i025Painted metal sheets1345Total Numbers42,776
Table 3. Land cover types and total number of samples in the Salinas dataset.
Table 3. Land cover types and total number of samples in the Salinas dataset.
No.ColorNameNumberNo.ColorNameNumber
1 Remotesensing 14 02612 i026Brocoli_green_weeds_119779 Remotesensing 14 02612 i027Soil_vinyard_develop6197
2 Remotesensing 14 02612 i028Brocoli_green_weeds_2372610 Remotesensing 14 02612 i029Corn_senesced_green_weeds3249
3 Remotesensing 14 02612 i030Fallow197611 Remotesensing 14 02612 i031Lettuce_romaine_4wk1058
4 Remotesensing 14 02612 i032Fallow_rough_pow139412 Remotesensing 14 02612 i033Lettuce_romaine_5wk1908
5 Remotesensing 14 02612 i034Fallow_ smooth267813 Remotesensing 14 02612 i035Lettuce_romaine_6wk909
6 Remotesensing 14 02612 i036Stubble395914 Remotesensing 14 02612 i037Lettuce_romaine_7wk1061
7 Remotesensing 14 02612 i038Celery357915 Remotesensing 14 02612 i039Vinyard_untrained7164
8 Remotesensing 14 02612 i040Grapes_untrained11,21316 Remotesensing 14 02612 i041Vinyard_vertical_trellis1737
Total Numbers53,785
Table 4. Classification comparison result of ARL-GAN with different algorithms on Indian Pines dataset.
Table 4. Classification comparison result of ARL-GAN with different algorithms on Indian Pines dataset.
MethodRBF-SVMDBNSAECNN3DCNNPPF-CNNARL-GAN
OA (%)77.8 ± 0.880.6 ± 0.181.9 ± 0.185.4 ± 0.892.8 ± 0.887.9 ± 0.898.3 ± 0.4
AA (%)61.3 ± 1.468.3 ± 1.769.4 ± 1.979.4 ± 1.689.4 ± 1.476.5 ± 0.695.0 ± 2.1
Kappa74.5 ± 1.077.8 ± 1.379.3 ± 1.184.3 ± 2.991.9 ± 0.986.3 ± 0.998.0 ± 0.4
16.1 ± 11.213.6 ± 5.610.0 ± 6.478.4 ± 10.250.4 ± 8.450.4 ± 8.480.4 ± 26.2
272.9 ± 3.679.8 ± 2.979.7 ± 2.375.4 ± 2.492.7 ± 3.589.2 ± 2.199.0 ± 1.0
358.0 ± 3.670.5 ± 2.274.9 ± 4.882.8 ± 3.387.2 ± 10.477.1 ± 2.796.7 ± 1.8
439.0 ± 15.071.3 ± 6.662.8 ± 8.389.2 ± 3.583.4 ± 8.387.7 ± 3.797.5 ± 4.9
587.0 ± 4.580.1 ± 4.184.2 ± 3.369.0 ± 4.684.0 ± 5.792.7 ± 1.096.3 ± 1.1
692.4 ± 2.094.2 ± 2.494.3 ± 1.792.8 ± 2.593.4 ± 2.593.1 ± 1.999.2 ± 1.0
70 ± 028.1 ± 22.624.4 ± 18.851.1 ± 12.397.2 ± 4.80 ± 082.9 ± 16.2
898.1 ± 1.498.5 ± 1.598.8 ± 0.497.1 ± 1.697.4 ± 2.899.6 ± 0.3100.0 ± 0.0
90 ± 09.5 ± 2.411.1 ± 10.141.6 ± 9.977.0 ± 11.10 ± 092.0 ± 16.0
1065.8 ± 3.773.2 ± 4.773.6 ± 3.881.0 ± 2.693.3 ± 5.085.6 ± 2.896.2 ± 3.1
1185.3 ± 2.982.7 ± 2.283.4 ± 2.087.2 ± 1.594.9 ± 2.783.8 ± 1.699.3 ± 0.7
1269.6 ± 6.562.0 ± 5.870.4 ± 8.084.4 ± 2.389.8 ± 4.390.4 ± 3.197.3 ± 3.2
1392.3 ± 4.189.7 ± 10.694.2 ± 4.383.1 ± 4.292.8 ± 5.997.8 ± 0.996.9 ± 3.8
1496.6 ± 1.094.4 ± 1.694.2 ± 1.598.2 ± 0.898.3 ± 1.395.5 ± 1.1100.0 ± 0.0
1541.7 ± 7.064.2 ± 6.566.1 ± 5.684.7 ± 4.577.8 ± 13.478.0 ± 2.499.1 ± 1.7
1675.2 ± 9.080.5 ± 13.287.6 ± 8.176.0 ± 8.188.4 ± 5.397.3 ± 1.388.0 ± 3.0
The bold value indicates the optimal value.
Table 5. Classification comparison results of ARL-GAN with different algorithms on the PaviaU dataset.
Table 5. Classification comparison results of ARL-GAN with different algorithms on the PaviaU dataset.
MethodRBF-SVMDBNSAECNN3DCNNPPF-CNNARL-GAN
OA (%)90.3 ± 0.691.9 ± 0.392.4 ± 0.393.0 ± 0.695.2 ± 0.796.9 ± 0.299.8 ± 0.1
AA (%)86.6 ± 0.988.5 ± 0.889.3 ± 0.788.6 ± 0.891.2 ± 1.195.1 ± 0.299.6 ± 0.2
Kappa87.1 ± 0.889.2 ± 0.489.9 ± 0.490.7 ± 0.793.8 ± 0.996.0 ± 0.299.7 ± 0.1
190.7 ± 1.191.6 ± 0.892.3 ± 1.193.1 ± 1.495.5 ± 1.298.0 ± 0.1100.0 ± 0.0
296.8 ± 0.797.4 ± 0.497.6 ± 0.397.6 ± 0.999.4 ± 0.399.2 ± 0.2100.0 ± 0.0
360.2 ± 5.469.7 ± 6.072.1 ± 3.577.9 ± 4.592.6 ± 5.484.9 ± 1.897.9 ± 1.4
490.8 ± 2.091.2 ± 1.490.9 ± 1.486.4 ± 3.675.2 ± 4.995.8 ± 0.899.0 ± 0.6
598.8 ± 0.498.6 ± 0.698.7 ± 0.498.5 ± 1.495.4 ± 4.399.8 ± 0.1100.0 ± 0.0
679.5 ± 4.985.6 ± 2.286.9 ± 1.991.0 ± 2.899.4 ± 0.696.4 ± 0.3100.0 ± 0.0
774.3 ± 5.174.8 ± 4.878.1 ± 4.981.2 ± 2.991.5 ± 3.489.2 ± 0.8100.0 ± 0.0
888.8 ± 2.288.2 ± 1.387.8 ± 1.492.5 ± 2.294.8 ± 1.493.7 ± 1.2100.0 ± 0.0
999.8 ± 0.199.6 ± 0.199.5 ± 0.379.0 ± 4.177.4 ± 2.898.5 ± 0.7100.0 ± 0.0
The bold value indicates the optimal value.
Table 6. Classification comparison result of ARL-GAN with different algorithms for the PaviaU dataset.
Table 6. Classification comparison result of ARL-GAN with different algorithms for the PaviaU dataset.
MethodRBF-SVMDBNSAECNN3DCNNPPF-CNNARL-GAN
OA (%)89.3 ± 0.790.8 ± 0.691.5 ± 0.192.3 ± 1.297.9 ± 0.492.8 ± 0.499.97 ± 0.02
AA (%)92.5 ± 0.694.5 ± 0.894.9 ± 0.294.2 ± 0.998.0 ± 0.695.5 ± 0.799.97 ± 0.01
Kappa88.0 ± 0.889.7 ± 0.790.6 ± 0.491.4 ± 1.497.9 ± 0.591.9 ± 0.499.96 ± 0.02
197.4 ± 1.598.5 ± 0.897.9 ± 0.593.3 ± 8.799.4 ± 0.698.5 ± 0.5100.0 ± 0.0
299.7 ± 0.298.9 ± 0.299.1 ± 0.597.4 ± 1.299.4 ± 0.699.7 ± 0.2100.0 ± 0.0
393.7 ± 1.597.5 ± 0.195.3 ± 0.686.4 ± 4.198.8 ± 1.899.8 ± 0.1100.0 ± 0.0
497.8 ± 1.399.0 ± 0.399.5 ± 0.698.2 ± 1.898.8 ± 1.999.7 ± 0.2100.0 ± 0.0
597.5 ± 1.197.5 ± 0.298.5 ± 0.498.1 ± 1.099.0 ± 0.796.8 ± 0.2100.0 ± 0.1
699.5 ± 0.399.3 ± 0.199.9 ± 0.199.9 ± 0.299.8 ± 0.399.8 ± 0.3100.0 ± 0.0
799.3 ± 0.299.0 ± 0.399.2 ± 0.199.0 ± 0.997.9 ± 2.599.5 ± 0.299.9 ± 0.1
888.9 ± 2.983.0 ± 1.482.7 ± 0.788.4 ± 2.896.7 ± 2.289.9 ± 0.9100.0 ± 0.0
999.2 ± 0.399.0 ± 0.199.2 ± 0.195.1 ± 0.799.8 ± 0.399.8 ± 0.2100.0 ± 0.0
1088.8 ± 1.992.8 ± 0.188.9 ± 0.993.6 ± 2.498.2 ± 2.188.3 ± 2.799.9 ± 0.1
1187.5 ± 4.691.7 ± 0.193.6 ± 7.097.6 ± 1.097.3 ± 3.893.4 ± 2.9100.0 ± 0.0
1298.0 ± 2.399.0 ± 0.196.8 ± 1.098.9 ± 1.198.5 ± 2.199.7 ± 0.799.9 ± 0.1
1398.0 ± 0.899.3 ± 0.299.2 ± 0.794.7 ± 2.497.4 ± 5.498.6 ± 0.7100.0 ± 0.0
1489.6 ± 2.692.0 ± 7.494.8 ± 0.292.5 ± 3.998.9 ± 1.092.3 ± 1.799.9 ± 0.2
1553.9 ± 7.669.5 ± 0.275.9 ± 2.480.1 ± 5.396.5 ± 1.572.9 ± 2.599.8 ± 0.1
1690.8 ± 5.096.1 ± 2.596.1 ± 1.993.6 ± 2.694.7 ± 5.195.7 ± 2.6100.0 ± 0.0
The bold value indicates the optimal value.
Table 7. Classification result of CNN, CNN-CE, and ARL-GAN.
Table 7. Classification result of CNN, CNN-CE, and ARL-GAN.
DatasetMethodsCNNCNN-CEARL-GAN
Indian Pines (Train: 5%)OA97.04 ± 0.4297.46 ± 0.4398.25 ± 0.36
AA96.17 ± 0.9393.63 ± 0.4695.03 ± 2.07
Kappa96.62 ± 0.4797.10 ± 0.4998.01 ± 0.42
Pavia University (Train: 3%)OA98.89 ± 0.0899.78 ± 0.0699.83 ± 0.07
AA98.21 ± 0.2399.55 ± 0.1299.66 ± 0.18
Kappa98.53 ± 0.1499.71 ± 0.1199.77 ± 0.09
Salinas
(Train: 1%)
OA99.13 ± 0.1399.96 ± 0.0299.97 ± 0.02
AA98.27 ± 0.2399.98 ± 0.0199.97 ± 0.01
Kappa99.03 ± 0.0999.97 ± 0.0299.96 ± 0.02
The bold value indicates the optimal value.
Table 8. Classification results of 3DGAN, HSGAN, MSGAN, and ARL-GAN.
Table 8. Classification results of 3DGAN, HSGAN, MSGAN, and ARL-GAN.
DatasetMethodsHSGAN3DGANMSGANARL-GAN
Indian Pines (5%)OA74.1495.1395.6198.25
AA65.6384.8191.4495.03
Kappa70.4494.8995.0298.01
PaviaU (3%)OA85.7198.2299.1799.83
AA81.4287.1798.4299.66
Kappa84.2894.9398.8999.77
Salinas (1%)OA88.2998.8899.1199.97
AA92.3192.6199.2399.97
Kappa87.0397.9299.0399.96
The bold value indicates the optimal value.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, S.; Zhang, X.; Li, T.; Meng, H.; Cao, X.; Wang, L. Adversarial Representation Learning for Hyperspectral Image Classification with Small-Sized Labeled Set. Remote Sens. 2022, 14, 2612. https://doi.org/10.3390/rs14112612

AMA Style

Zhang S, Zhang X, Li T, Meng H, Cao X, Wang L. Adversarial Representation Learning for Hyperspectral Image Classification with Small-Sized Labeled Set. Remote Sensing. 2022; 14(11):2612. https://doi.org/10.3390/rs14112612

Chicago/Turabian Style

Zhang, Shuhan, Xiaohua Zhang, Tianrui Li, Hongyun Meng, Xianghai Cao, and Li Wang. 2022. "Adversarial Representation Learning for Hyperspectral Image Classification with Small-Sized Labeled Set" Remote Sensing 14, no. 11: 2612. https://doi.org/10.3390/rs14112612

APA Style

Zhang, S., Zhang, X., Li, T., Meng, H., Cao, X., & Wang, L. (2022). Adversarial Representation Learning for Hyperspectral Image Classification with Small-Sized Labeled Set. Remote Sensing, 14(11), 2612. https://doi.org/10.3390/rs14112612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop