Next Article in Journal
Geohash-Based Rapid Query Method of Regional Transactions in Blockchain for Internet of Vehicles
Next Article in Special Issue
An Efficient End-to-End Multitask Network Architecture for Defect Inspection
Previous Article in Journal
Sensory Interaction and Balancing Ability Evaluation of the Elderly Using a Simplified Force Plate System
Previous Article in Special Issue
SA-FEM: Combined Feature Selection and Feature Fusion for Students’ Performance Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New SCAE-MT Classification Model for Hyperspectral Remote Sensing Images

1
School of Computer Science, China West Normal University, Nanchong 637002, China
2
School of Electronic Information and Automation, Civil Aviation University of China, Tianjin 300300, China
3
Traction Power State Key Laboratory, Southwest Jiaotong University, Chengdu 610031, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(22), 8881; https://doi.org/10.3390/s22228881
Submission received: 11 October 2022 / Revised: 15 November 2022 / Accepted: 15 November 2022 / Published: 17 November 2022
(This article belongs to the Special Issue Recent Trends and Advances in Fault Detection and Diagnostics)

Abstract

:
Hyperspectral remote sensing images (HRSI) have the characteristics of foreign objects with the same spectrum. As it is difficult to label samples manually, the hyperspectral remote sensing images are understood to be typical “small sample” datasets. Deep neural networks can effectively extract the deep features from the HRSI, but the classification accuracy mainly depends on the training label samples. Therefore, the stacked convolutional autoencoder network and transfer learning strategy are employed in order to design a new stacked convolutional autoencoder network model transfer (SCAE-MT) for the purposes of classifying the HRSI in this paper. In the proposed classification method, the stacked convolutional au-to-encoding network is employed in order to effectively extract the deep features from the HRSI. Then, the transfer learning strategy is applied to design a stacked convolutional autoencoder network model transfer under the small and limited training samples. The SCAE-MT model is used to propose a new HRSI classification method in order to solve the small samples of the HRSI. In this study, in order to prove the effectiveness of the proposed classification method, two HRSI datasets were chosen. In order to verify the effectiveness of the methods, the overall classification accuracy (OA) of the convolutional self-coding network classification method (CAE), the stack convolutional self-coding network classification method (SCAE), and the SCAE-MT method under 5%, 10%, and 15% training sets are calculated. When compared with the CAE and SCAE models in 5%, 10%, and 15% training datasets, the overall accuracy (OA) of the SCAE-MT method was improved by 2.71%, 3.33%, and 3.07% (on average), respectively. The SCAE-MT method is, thus, clearly superior to the other methods and also shows a good classification performance.

1. Introduction

Hyperspectral remote sensing images (HRSI) have the characteristics of “homospectral foreign objects” and “homospectral foreign objects”. In order to accurately identify and classify the ground objects, the deep features of the HRSI need to be extracted. These deep features can fully represent the original ground objects, such as spectral features, spatial features, and local correlation features, in addition to the shallow and middle-level features such as the shape, contour, and texture of the images [1,2,3,4]. In recent years, an autoencoder network has obtained good results in the use of the deep feature extraction of HRSI via the decoding and coding of images, feature mapping, and image reconstruction in multi-layer networks [5,6,7]. Zhang et al. [8] presented an unsupervised feature learning method using a RAE (recursive autoencoders) network. However, the recursive structure will occupy a lot of space and has low efficiency. If the recursive structure is too deep, it will cause stack overflow and program crash. Shi and Pun [9] proposed a recursive neural network that was based on a stack autoencoder in order to learn high-order multiscale spectral spatial features. However, this was also troubled by efficiency and easy “overfitting”. Mei et al. [10] proposed a spectral and spatial feature extraction method based on a three-dimensional convolutional autoencoder (3D-CAE) in order to extract the deep features of the HRSI. However, in the 3D operation of the deep neural network, the training process is complex, i.e., the “small sample” and efficiency problems are still the bottleneck of the classification task. Zhao et al. [11] proposed a simplified three-dimensional convolutional self-encoder (S3D-CAE) for the purposes of extracting the spatial features of deep spectrum from the dual temporal images without prior information. Although the spectral redundancy is reduced and the data processing speed is improved, it is easy for “salt noise” to appear when the classification task is completed. Zhang et al. [12] designed a bi-directional self-encoder, which reconstructed hyperspectral data and lidar data together, as well as sent the fused data to a dual-branch CNN for final classification, which improved the classification accuracy. However, the model training still requires one to consider many parameters.
The autoencoder network can effectively extract the deep features of the HRSI through the layer-by-layer training of the deep neural network structure [13,14,15]. However, when the model is trained, the scale of involved parameters in the training is large, the computational complexity is high, and the efficiency is low. In the full connected network, it is also easy to ignore local dependencies in the domain space. On the other hand, it is difficult to manually label samples of the HRSI data, which are examples of typical “small sample” data. Insufficient training samples and labeled samples will further lead to difficult and unstable training parameters of the classification model; further, the insufficient space-spectral feature space will also affect classification accuracy [16,17,18,19]. Dai et al. [20] combined a residual convolutional neural network (RESNET) with an ensemble learning (EL) strategy in order to establish discriminant image representation to complete the classification task. Consequently, they obtained good classification results. However, the classification task was completed via supervised learning, which has a requirement to be based on a certain number of labeled samples. Yan et al. [21] introduced the semi-supervised learning into the generative adversarial network (GAN) and introduced the aliasing data enhancement method in the classification model, which increased the available marker data for classification and improved the classification accuracy. However, the model is prone to “over-fitting”. Miao et al. [22] proposed a semi-supervised representation-consistent Siamese network (SS-RCSN) in order to reduce the difference between labeled and unlabeled data to a certain extent, which, thus, prevents the model from “overfitting”. However, the intra-class diversity and inter-class similarity make it easy to generate “salt and pepper noise”, which affects the accuracy of classification. In addition, some other methods are also proposed to deal with images and other problems [23,24,25,26,27,28,29,30,31,32,33,34].
On basis of the idea of semi-supervised learning, the insufficient labeled sample data can be solved to a certain extent [35,36,37,38,39,40,41,42]. However, for the HRSI data—due to the characteristics of the spatial structure, scale size, spectral confusion, correction error, etc.—the real labels of the feature data and the learned pseudo labels from the training data are easy to mislabel and are inconsistent. Further, if this part of the unlabeled sample data is forcibly added to the training dataset of labeled samples, it will easily affect the classification accuracy. He et al. [43] presented a HRSI classification method based on heterogeneous transfer learning. However, the “fine-tuning” process of the structure of two heterogeneous datasets during transfer is complicated. Chen et al. [44] proposed a semi-supervised dual-dictionary nonnegative matrix factorization (SSDDNMF) heterogeneous transfer learning algorithm that projects source domain features and target domain features. In the dimensional subspace, the difference between feature spaces is eliminated, but its key features are easily lost when the features are shared. Additionally, the determination of size difference has a greater impact on the classification accuracy. Qu et al. [45] reconciled the gap between the source domain dataset and the target domain dataset by projecting the HRSI data of the source domain and the target domain into the shared abundance space, as based on their physical characteristics. Yang et al. [46] designed a hierarchical depth neural network in order to realize the shallow feature transfer and depth feature classification, as well as to reduce the demand for source domain data.
To summarize, the autoencoder network is an unsupervised learning model. When the hyperspectral remote sensing image features are learned, it can extract the spatial spectrum features, but it will lose the spatial information of the original images in the training process. On the other hand, the deep network model needs a large number of labeled samples in order to participate in the training, while the hyperspectral remote sensing image data is an example of typical “small sample” data and the lack of training samples is easy to make the model produce an “over fitting” result. The transfer strategy is introduced in order to solve the training problem of “small sample” data to a certain extent. However, the different acquisition conditions of the HRSI usually leads to spectral changes, while the training and test data are usually located in different domains, and the data distribution changes not only with time and space, but also with different situations. It is sensitive to aim of reconciling the gap, but its classification robustness is poor. In order to solve these problems, a stacked convolutional autoencoder network model is employed in order to extract the deep features from the HRSI. As the different training datasets have different distribution spaces during model training, and the learning model cannot use the same data distribution to treat new training data, a transfer learning strategy is introduced in order to modify the learning model to be more robust. The model parameters, structure, and other knowledge transfers are trained; further, the learning model is continuously fine-tuned during the transfer process, such that the trained model can be adjusted before the new data set is employed in order to solve the model sensitivity and poor classification robustness under “small samples” readings. It is, therefore, easy to obtain the unsatisfactory classification accuracy caused by “over-fitting”. Moreover, the classification effectiveness and accuracy of the HRSI are improved.
The innovations and main work of this paper are described as follows.
  • A new stacked convolutional autoencoder network model transfer (SCAE-MT) is developed;
  • The stacked convolutional auto-encoding network is used to effectively extract the deep features of the HRSI;
  • The transfer learning strategy is employed in order to develop a SCAE network model transfer under small and limited training samples;
  • The SCAE-MT is used to propose a new HRSI classification method in order to solve the small samples that can be found in the HRSI.

2. Image Classification with SCAE-MT

2.1. The Idea of Image Classification

In order to accurately identify and classify the ground objects, when the massive high-dimensional hyperspectral remote sensing image data is used, it is necessary to extract image features that can fully represent the original ground objects [47,48,49,50]. Therefore, in this paper, the stack convolution auto-coding network is employed in order to extract the deep features of hyperspectral remote sensing images. All convolution cores, weights, and offsets in the deep neural network are trained through a deep iteration process. Further, in the process of training, the optimal convolution kernel parameter weight w and bias b are obtained by repeatedly alternating a forward propagation and backward error return. Clearly, when the stack convolutional auto-coding network is to be iteratively trained, sufficient marker samples are required in order to participate in the training of the deep neural network. The marker samples involved in the training are greater, the accuracy of acquiring the deep features is higher, and the recognition accuracy of ground objects are also higher. However, the characteristics of the HRSI data determine that it is difficult to label samples [51,52,53,54]. Therefore, if the HRSI dataset is simply trained by a deep neural network—which is based on a stack convolution self-encoder—due to the fact that the involved label samples in training are limited, then it is easy to produce the phenomenon of “over fitting”.
In order to solve this problem, a transfer learning strategy was introduced when the stack convolution self-encoder model was trained. As the HRSI datasets usually have different distributions of their source domain datasets and target domain datasets, the model-based transfer learning is adopted. Firstly, the source domain datasets are pre-trained, and the source domain datasets and target domain datasets share some general prior knowledge on the model. The HRSI data sets from different scenes collected by the same sensor are common to the image itself in terms of shape, corner, and other characteristics. Though the source domain data set is pre-trained, the transfer knowledge and code for it to model parameters will participate in model fine-tuning; in addition, it will help complete the classification task of target domain datasets via the adjusting of previous prior knowledge and pre-trained models.

2.2. Processes of Implementation

For the purposes of the HRSI classification, the HRSI datasets from different scenes collected by the same sensor are common to the image itself, such as the shape and corner. In this paper, the transfer learning strategy was used to transfer these common features from the source dataset to the target dataset. Additionally, the transferred knowledge is applied to the model prior to knowledge acquisition, model parameter training, and model architecture of the target dataset. For the top layer of the target scene dataset, the deep neural network is used to extract the deep features of the images and classify the target scene data [55,56,57]. The specific implementation steps are described as follows.
  • Step 1: Deep feature extraction
The HRSI data is composed of two-dimensional spatial information and one-dimensional spectral information. It is a three-dimensional data. Compared with ordinary remote sensing images, it contains more abundant spectral information and spatial information. For the HRSI data, the invariance and limited robustness of it shallow features are difficult to cope with in regard to the high intra-class variability and low inter-class variability. The auto-encoding network cascades multiple layers of features and extracts higher-level features hierarchically. For the HRSI data, it can effectively combine its spectral features and spatial features, which is beneficial in order to improve the classification accuracy of ground and object. However, during the training of the self-encoding network model, because the HRSI data is usually represented as a vector, the full connected layer will lose spatial information. With the increase in the number of neurons in the hidden layer, the parameter size of the weight matrix will also increase sharply. Many parameters will lead to difficult and inefficient training; further, the model is prone to “over-fitting”. Therefore, the convolutional layers with local connection and weight sharing characteristics are used to replace the convolutional autoencoder network with a full connected layer of simple autoencoder networks for the purposes of deep feature extraction of the HRSI [58,59,60]. When the convolutional self-encoding network is trained—due to the fact that the neurons in each layer are arranged in three dimensions of width, height, and depth—the neurons in the layer are only connected to a small area in the previous layer, instead of it taking a full connection. The involved parameters in training are relatively small, which reduces the computational complexity and solves the problem of model “over-fitting” to a certain extent. On the other hand, the local connection and weight sharing of the convolutional layer and the three-dimensional arrangement of the feature vectors enable the convolutional auto-encoding neural network to better preserve the local spatial information of the data, as well as capture a specific type of input data through a convolution kernel. The local features of the HRSI can be extracted by using multiple different convolutions in order to extract multi-level features.
In order to strengthen the ability of the convolutional autoencoder network to better extract the deep features of the HRSI, the classification accuracy requires further improvement. The convolutional autoencoder networks in a stack manner are organized to construct an unsupervised deep neural network for the purposes of extracting the deep features of the HRSI through stacked convolutional autoencoder networks [61,62,63,64,65,66]. The stacked convolutional autoencoder network utilizes the characteristics of local connections and the weight sharing of convolutional layers; further, it is trained layer-by-layer from low layers to high layers. While the computational complexity of the training model is reduced, the features of each layer are fused and reconstructed in order to generate fusion deep-level feature representation for spectral and spatial features of the HRSI. The stacked convolutional self-encoding network model is shown in Figure 1.
Suppose the stacked convolutional auto-encoding network is composed of n convolutional auto-encoding networks and that the input feature is X = { x i | ( i = 1 ,   2 , m ) } . Then the features of the output of the l ( l = 1 ,   2 , n ) convolutional layer are expressed as follows:
z l = σ ( x i w l k × k + b l )
Y = σ ( d = 1 l x i d w d k × k + b d l )
Among these, z l is the feature map of the l layer, Y is the superposition of all feature maps, w l k × k is the size of k × k the convolution kernel, b l is the bias, and ⊗ represents the convolution operation. As the image has a local correlation, the input feature vector can extract the features that have a local correlation through the convolution operation with the convolution kernel. After the convolution operation, the pooling layer is used for the purposes of pooling in order to reduce the amount of training parameters and to also prevent over-fitting. In this paper, mean pooling is used; further, σ ( · ) is a nonlinear differentiable activation function, and the ReLU function with fast convergence speed and simple calculation is employed.
In order to reduce the number of parameters and computational complexity, as well as to improve the training efficiency, the high-dimensional features are converted into low-dimensional features through operating a series of convolution pooling; then, the merged smaller images are unsampled into the original image dimension. The low-dimensional feature vectors are used to reconstruct the images, that is, the low-dimensional features are mapped to high-dimensional features. After each convolution layer passes through the activation function, the feature map z l is obtained, and the transposition of the corresponding convolution kernel is used to perform the convolution operation by feature reconstruction. The specific expression is described as follows:
Y = σ ( i = 1 l z l w l k × k ˜ + b l )
Among this, the w l k × k transpose matrix is of w l k × k ˜ , and the reverse w l k × k mapping of low-dimensional features to high-dimensional features is realized by transposing.
The parameters in the stacked convolutional autoencoder network are the weights and biases in the convolutional layer, and the backpropagation (BP) algorithm is used in order to update the parameters. The updating process of the parameters via the use of the BP algorithm is divided into forward propagation and back propagation. According to the initialization weight w and bias b given by the input sample, the square error is used as the loss function, which is then used to calculate the error value δ between the output value y and the actual value x.
δ = 1 2 i = 1 l ( y i x i ) 2
The above is a forward calculation process. If the error δ is within the threshold range at this time, it stops updating the weights w and bias b. On the contrary, if the error δ value is not within the threshold range, then the error is returned for back propagation. During the process of backpropagation, the gradient descent method is used to solve the weight w and bias b, such that the minimum error reaches the threshold range. In the stacked convolutional auto-encoding process, the gradient is backpropagated through the error value δ of each layer. Therefore, during backpropagation, the rule is employed in order to calculate the gradient of the convolution layer parameters. Further, the loss function is used to calculate the gradient of the l layer. Partial derivatives for weights w l and bias b l are described as follows.
δ w l = δ z l x l 1
δ b l = δ z l x l 1
In the deep feature extraction of the HRSI, the model is pretrained by using stacked convolutional autoencoder networks. Further, in the pre-training process, the first-order features are obtained from the bottom layer; that is, the first layer is used as the input of the second layer. When compared with the first-order features of the first layer, the second-order features including the first-order feature pattern of the next layer will be learned in the second layer. By the use of analogy and training layer-by-layer, the trestle convolution self-coding network will eventually learn the higher-level features of the HRSI data. Compared with the convolutional autoencoder network, the stack-type convolutional autoencoder network is stacked in a stacking manner in the training phase, such that the spatial correlation features of the underlying features are retained in the upper-layer feature map. The local correlation of features within the domain is well preserved. This solves the problem that a single convolutional autoencoder network easily loses, i.e., the spatial correlation features of the domain.
  • Step 2: Classification based on SCAE-MT
The stacked convolutional autoencoder network is used as the training and base model for transfer learning. The pre-training is performed on the source-domain HRSI dataset. At this time, the parameter weights w i and biases b i of the stacked convolutional auto-encoding network are randomly generated during initialization. After the pre-training of the source domain dataset is completed, the model parameters obtained by the pre-training network are transferred by the transfer strategy of sharing model structure and parameters. The weights w i and biases b i are considered as initialization parameters in regard to the target domain training dataset model for the purposes of training the target domain dataset. In regard to the ground feature classification, in order to improve the dataset due to limited sample labels, the model parameters of the weights and biases for the target domain dataset are obtained by fine-tuning within the process of the training. Additionally, the transfer strategy is used to complete the feature vector construction of the “small sample” dataset, as well as to reduce the impact of insufficient training samples on the classification performance. Moreover, the classification framework is transferred based on the stacked convolutional autoencoder model in Figure 2.
In the transfer process, as based on the stacked convolutional auto-encoding network model, the average value of the softmax activation function of all source samples of the category class n and the class n weak label t are calculated. Further, that which can be extracted then its average value t ( n ) is used here. The cross-entropy loss of the source domain dataset and the target domain dataset can be defined as follows:
ζ = 1 n i ζ i = 1 n i c = 1 n x i c log t i c n
When the trained model on the source domain data set is used as prior knowledge in order to train the model of the target domain dataset, whether the target domain sample belongs to a category c i or is calculated by defining an excitation function, then the excitation function is defined as follows:
f = ( y i , c i ) = t ( n ) · θ ¯ ( y i , c i ) = t ( 0 ) · θ 0 ( x , y ) + z = 1 c o v n t ( c i , z ) · θ ( c i , z )
where θ ( · ) is the feature mapping function, and covn is the number of layers.
The steps of the fine-tuning are described as follows:
Step 1. Train a stack convolutional auto-coding model on the source domain dataset as the source model.
Step 2. The target domain dataset copies the model structures and parameters on the source domain model except in regard to the output layer. Then the stack convolutional auto-coding network model of the target domain dataset is established.
Step 3. Add an output layer whose output size is the number of target domain dataset categories to the target domain model, as well as randomly initialize the model parameters of this layer.
Step 4. The target model is trained on the target domain dataset. At this time, the structure and weight parameters of all previous layers (i.e., those that do not participate in training, which is to say do not back propagate) are frozen. Then, a smaller learning rate and the gradient descent method are used in order to retrain the output layer.

3. Experiment Results and Analysis

In order to verify the effectiveness of the proposed classification method based on the SCAE-MT, the Indian Pines dataset and the Salinas dataset are selected here. They are the two commonly hyperspectral remote sensing image datasets by the same sensor in different areas and they are used as the experimental datasets.

3.1. Experimental Datasets

Both the Indian Pines and the Salinas datasets were acquired by the airborne visual/infrared imaging spectrometer AVIRIS, the basic information of which is shown in Table 1. The Indian Pines dataset is an image of an Indian pine tree in Indiana, USA; further, the size of 145 × 145 is then intercepted and marked as a dataset for the HRSI classification, which is one of the most commonly used datasets. The data contains a total of 21,025 pixels—excluding a large number of background pixels—and the remaining pixels containing ground objects are a total of 10,249, which includes 16 types of ground objects and 200 valid bands. The statistics of these ground object categories are shown in Table 2. For the Salinas dataset, it originates from the Salinas Valley in California, USA. Compared with the Indian Pines dataset, the Salinas dataset is larger, has more samples, and has a higher spatial resolution of 3.7 m. The image size is 512 × 217, as such it contains 111,104 pixels in total, of which 56,975 are background pixels. Additionally, there are 54,129 pixels that can be used for the purposes of classification, including 16 types of ground objects and 204 valid bands. The category statistics of this are shown in Table 2.

3.2. Experiment Environment and Parameter Settings

The experiment environment was configured as follows: thinkstationp920 was the workstation, the CPU@2 *Xeon silver 4210, memory (RAM) was 128G, the graphics card was an NVIDIA Titan RTX Titan (video memory 24G), and the programming language used was Matlab R2018a.
During the experiment, the kernel size of the convolution self-coding network was the ownership value matrix of the full connection layer that was initialized according to a Gaussian random distribution; further, the standard deviation was set to 0.05. Additionally, the mean value was set to 0. The bias initialization was set to 0, the mean pooling method was adopted, and the activation function was the ReLu function. The learning rate in the standard random gradient descent method was fixed to 0.0001, the momentum was set to 0.9, and the batch size (bottom, 2012) was set to 128. The maximum number of iterations was 80.

3.3. Experimental Results and Analysis

In order to verify the learning ability of the trestle convolutional self-coding network, an experimental comparison of the sensitivity of the number of stacked convolutional self-coding networks was executed on the classification performance in the network structure. The 40% and 80% of the Indian Pines dataset were randomly selected as training samples according to prior experiences and the literature; the rest was the test data set. The training data set was used to train the model, and the test data set was used to verify the effectiveness of the model, which were trained under the model with the number of convolutional self-coding networks of 1, 2, and 3, respectively. Additionally, the overall accuracy (OA) of the classification was calculated. The overall accuracy, i.e., the OA values of classification under different convolutional self-coding networks and also under different training samples are shown in Table 3. Table 3 shows the overall classification accuracy of the Indian Pines dataset by randomly selecting 40% and 80% of the samples as the training set, then training and calculating when the number of convolutional self-encoders is 1, 2, 3. The comparative histogram of the sensitivity of the number of convolutional self-coding networks to classification performance in the trestle convolutional self-coding model is shown in Figure 3. In order to clearly compare the impact of different number of self-encoders on the classification accuracy, the column chart is drawn to show that in the stack type convolutional self-coding network, with the increase in the number of convolutional auto-encoders, the overall classification accuracy will increase. A reference is provided for subsequent experiments to help determine the number of convolutional auto- encoders.
From Table 3 and Figure 3, the overall accuracy (OA) of classification is the lowest when there is only one convolutional self-coding network in the trestle convolutional self-coding network that is under 40% or 80% in the training sample sets. When the number of convolutional self-coding networks in the trestle convolutional self-coding model structure increases, the overall accuracy (OA) of classification will also increase. However, it can be seen that when the number of convolutional self-coding networks in the model result is three, then the overall accuracy of classification is improved by 0.17% and 0.14%, respectively, when under 40% and 80% in regard to the training sample sets by comparing with the case of only two networks. The increasing number of convolutional self-coding networks has little impact on the classification accuracy. Considering the running cost of adding one convolutional self-coding network, as well as in balancing the time-space efficiency and classification accuracy (OA), two convolutional self-coding networks are selected in order to participate in the training of the trestle convolutional self-coding model.
In order to verify the effectiveness of the transfer learning strategy in the proposed classification method—assuming that the Indian Pines dataset has sufficient training samples, which is considered as the source domain dataset—and the Salinas data set acquired by the same sensor in different regions is the target data set, then the following steps should be taken. Firstly, the effectiveness of the model transfer strategy is verified. The 80% of the samples on the Indian pines data set is selected for the purposes of pre-training on the trestle convolutional self-coding network model. Further, the Indian pines data set with sufficient samples was selected as the source domain dataset. Then, the “small sample” data on the Salinas data set are randomly selected as the training set, and the 5% on the Salinas data set, in turn, are randomly selected. The 10% and 15% of the samples were used as training sample sets; further, experiments were conducted on the influence of a non-transfer learning strategy and the introduction of a transfer learning strategy on the overall classification accuracy of “small sample” data sets. Under the no transfer learning strategy, the “small sample” training sets of 5%, 10%, and 15% on the Salinas data set were trained, respectively. In addition, the overall classification accuracy (OA) was obtained after the random initialization parameters were trained. After the transfer learning strategy was used, the model-based transfer learning strategy was introduced when the target domain dataset Salinas 5%, 10% and 15% of the “small sample” training sets are trained. The parameters of all layers in the structure of the pre-training model are used as the initial parameters for training under different numbers of small samples in the Salinas dataset of the target domain, and the gradient descent method was used to fine tune them to obtain the overall classification accuracy (OA) under each proportion of the training sample set. The overall classification accuracy (OA) of the “small sample” training set without a transfer strategy, as well as after the introduction of a transfer strategy on the Salinas dataset, are both shown in Table 4. The comparative histogram of the overall classification accuracy (OA) of the three ratios without a transfer strategy and after the introduction of a transfer strategy are shown in Figure 4.
From Table 4 and Figure 4, after the transfer strategy is introduced, the overall accuracy of the target domain in regard to the Salinas dataset, under the small-sample training set of 5%, 10%, and 15%, is improved 2.53%, 6.18%, and 3.11%, respectively. This is achieved by comparing with the without the introduction of the transfer strategy method, which indicates that the model-based transfer strategy proposed in this paper has a certain effect on the improvement of the classification accuracy of the “small sample” dataset, which, in turn, verifies the effectiveness of the transfer strategy.
In addition, in order to verify the classification performance of the classification method based on the SCAE-MT on the “small sample” dataset. The classification method based on the convolutional autoencoder network (CAE); the classification method based on stacking convolutional autoencoder network (SCAE); and the classification method based on optimization of the stacking convolutional autoencoder network model transfer (SCAE-MT) is calculated. The overall classification accuracy (OA) of the method at 5%, 10%, and 15% of the training set, as well as the overall accuracy (OA) of each method under the three proportions of the training samples are shown in Table 5. In regard to the Salinas dataset, the classification effect of each method under a 15% proportion of training samples is shown in Figure 5. Figure 5a–d are the original classification effects map, the classification effects map of the CAE method, the classification effects map of the SCAE method, and the classification effect map of the SCAE-MT method, respectively.
As can be seen in Table 5 and Figure 5, the SCAE-MT has an average improvement of 2.71%, 3.33%, and 3.07% in terms of overall accuracy (OA) when compared with the CAE and the SCAE methods under the training datasets of 5%, 10%, and 15%, respectively. The SCAE-MT method is significantly better than its comparative methods, which show a good classification performance. The transfer learning strategy can improve the classification accuracy of the HRSI. From the experimental results of the CAE and SCAE methods, it can be seen that the overall accuracy (OA) of the SCAE is improved by 2.43%, 1.7%, and 1.29% under the training data sets of 5%, 10%, and 15%, respectively. In regard to the deep neural network structure that stacks the convolutional autoencoder (CAE) network in a stacking manner, it can be seen that it exerts its advantages of a layer-by-layer feature extraction, indicating that each layer of the SCAE method cannot simply serve as a convolutional layer during training. The self-encoding (CAE) network is trained separately, and the small-size feature space following the upper-layer feature mapping participates in the training of the next layer, which results in it performing better in regard to classification. This, therefore, shows it has a higher classification accuracy than the CAE method. A better overall accuracy (OA) for the SCAE and CAE methods under a 5% training set can been to be 0.79% higher than that under 10% training sample set, which is 1.14% higher than that under 15% training sample set. Therefore, the experiment results show that the SCAE-MT model has a certain effect on the improvement of the classification performance of “small sample” data.

4. Conclusions and Prospect

In this paper, a new stacked convolutional autoencoder network model transfer (SCAE-MT) is developed in order to realize a HRSI classification method that can better solve the problems of insufficient image feature extraction, can produce large scale model training parameters, and can also more easily produce the “overfitting” and “small sample” features of HRSI data. The comparative experiments show that the overall classification accuracy of the target data set after the introduced transfer strategy is 3.94% higher than that without the transfer strategy when compared with those under a different proportion of training sample sets. The overall classification accuracy (OA) of the target domain data under the 5%, 10%, and 15% training sets is improved by 1.79% on average by organizing two convolutional self-encoders in a trestle manner and through the use of simple convolutional self-encoders. It can be seen that the trestle convolutional self-encoder used in the study can effectively extract the features of the HRSI. As a machine learning method that uses and adjusts the knowledge and model of the source domain in order to help solve the problem of the target domain, transfer learning can solve the problem of “small samples” in the target domain due to the lack of training data.
However, the proposed classification method is based on the premise that the labeled samples of the source domain dataset are sufficient. In future works, we will further study how to improve the classification accuracy and feature recognition ability of the HRSI by the utilization of a transfer learning strategy.

Author Contributions

Conceptualization, Y.C. and Q.W.; methodology, H.C.; software, Q.W.; validation, T.C. and Q.W.; investigation, T.C.; resources, H.Z.; data curation, Y.C.; writing—original draft preparation, H.C.; writing—review and editing, T.C.; visualization, Q.W.; supervision, H.Z.; project administration, H.C.; funding acquisition, H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by: the Natural Science Foundation of Sichuan Province under Grant 2022NSFSC0536; the Research Foundation for Civil Aviation University of China under Grant 2020KYQD123 and Grant 3122022PT02; and the Open Project Program of the Traction Power State Key Laboratory of Southwest Jiaotong University (TPL2203).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chang, C.I.; Kuo, Y.M.; Chen, S.; Liang, C.C.; Ma, K.Y.; Hu, P.F. Self-mutual information-based band selection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5979–5997. [Google Scholar] [CrossRef]
  2. Song, M.; Shang, X.; Chang, C.I. 3-D receiver operating characteristic analysis for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8093–8115. [Google Scholar] [CrossRef]
  3. Shang, X.; Song, M.; Chang, C.I. An iterative random training sample selection approach to constrained energy minimization for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 1625–1629. [Google Scholar] [CrossRef]
  4. Yu, H.; Zhang, H.; Liu, Y. Dual-channel convolution network with image-based global learning framework for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  5. Chen, S.; Zhong, S.; Xue, B.; Li, X.; Zhao, L.; Chang, C.I. Iterative scale-invariant feature transform for remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3244–3265. [Google Scholar] [CrossRef]
  6. Huang, C.; Zhou, X.; Ran, X.J.; Liu, Y.; Deng, W.Q.; Deng, W. Co-evolutionary competitive swarm optimizer with three-phase for large-scale complex optimization problem. Inf. Sci. 2022. [Google Scholar] [CrossRef]
  7. Sun, H.; Ren, J.; Zhao, H.; Yuen, P.; Tschannerl, J. Novel Gumbel-Softmax Trick Enabled Concrete Autoencoder with Entropy Constraints for Unsupervised Hyperspectral Band Selection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5506413. [Google Scholar] [CrossRef]
  8. Zhang, X.; Liang, Y.; Li, C.; Huyan, N.; Jiao, L.; Zhou, H. Recursive autoencoders-based unsupervised feature learning for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1928–1932. [Google Scholar] [CrossRef] [Green Version]
  9. Shi, C.; Pun, C. Multiscale Superpixel-based hyperspectral image classification using recurrent neural networks with stacked autoencoder. IEEE Trans. Multimed. 2020, 22, 487–501. [Google Scholar] [CrossRef]
  10. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised Spatial—Spectral Feature Learning by 3D Convolutional Autoencoder for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  11. Zhao, C.; Cheng, H.; Feng, S. A Spectral—Spatial change detection method based on simplified 3D convolutional autoencoder for multitemporal hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar]
  12. Zhang, M.; Li, W.; Tao, R.; Li, H.; Du, Q. Information Fusion for Classification of Hyperspectral and LiDAR Data Using IP-CNN. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5506812. [Google Scholar] [CrossRef]
  13. Dou, Z.; Gao, K.; Zhang, X.; Wang, H.; Han, L. Band Selection of hyperspectral images using attention-based autoencoders. IEEE Geosci. Remote Sens. Lett. 2021, 18, 147–151. [Google Scholar] [CrossRef]
  14. Ghasrodashti, E.K.; Sharma, N. Hyperspectral image classification using an extended Auto-Encoder method. Signal Process. Image Commun. 2020, 92, 116111. [Google Scholar] [CrossRef]
  15. Bansal, V.; Buckchash, H.; Raman, B. Discriminative Auto-Encoding for classification and representation learning problems. IEEE Signal Process. Lett. 2021, 28, 987–991. [Google Scholar] [CrossRef]
  16. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Li, J.; Plaza, A. Active Learning with Convolutional Neural Networks for Hyperspectral Image Classification Using a New Bayesian Approach. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6440–6461. [Google Scholar] [CrossRef]
  17. Rao, M.; Tang, P.; Zhang, Z. Spatial—Spectral Relation Network for Hyperspectral Image Classification with Limited Training Samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 5086–5100. [Google Scholar] [CrossRef]
  18. Prasad, S.; Labate, D.; Cui, M.; Zhang, Y. Morphologically Decoupled Structured Sparsity for Rotation-Invariant Hyperspectral Image Analysis. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4355–4366. [Google Scholar] [CrossRef]
  19. Gao, K.; Guo, W.; Yu, X.; Liu, B.; Yu, A.; Wei, X. Deep Induction Network for Small Samples Classification of Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3462–3477. [Google Scholar] [CrossRef]
  20. Dai, X.; Wu, X.; Wang, B.; Zhang, L. Semisupervised scene classification for remote sensing images: A method based on convolutional neural networks and ensemble learning. IEEE Geosci. Remote Sens. Lett. 2019, 16, 869–873. [Google Scholar] [CrossRef]
  21. Yan, P.; He, F.; Yang, Y.; Hu, F. Semi-supervised representation learning for remote sensing image classification based on generative adversarial networks. IEEE Access 2020, 8, 54135–54144. [Google Scholar] [CrossRef]
  22. Miao, W.; Geng, J.; Jiang, W. Semi-Supervised Remote-Sensing Image Scene Classification Using Representation Consistency Siamese Network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5616614. [Google Scholar] [CrossRef]
  23. Deng, W.; Zhang, L.; Zhou, X.; Zhou, Y.; Sun, Y.; Zhu, W.; Chen, H.; Deng, W.; Chen, H.; Zhao, H. Multi-strategy particle swarm and ant colony hybrid optimization for airport taxiway planning problem. Inf. Sci. 2022, 612, 576–593. [Google Scholar] [CrossRef]
  24. Lv, Z.; Huang, H.; Li, X.; Zhao, M.; Benediktsson, J.A.; Sun, W.; Falco, N. Land cover change detection with heterogeneous remote sensing images: Review, Progress and Perspective. Proc. IEEE 2022. [Google Scholar] [CrossRef]
  25. Wei, Y.Y.; Zhou, Y.Q.; Luo, Q.F.; Deng, W. Optimal reactive power dispatch using an improved slime Mould algorithm. Energy Rep. 2021, 7, 8742–8759. [Google Scholar] [CrossRef]
  26. Zhang, H.; Liu, Y.; Li, X.; Feng, R.; Gong, Y.; Jiang, Y.; Guan, X.; Li, S. Combing remote sensing information entropy and machine learning for ecological environment assessment of Hefei-Nanjing-Hangzhou region, China. J. Environ. Manag. 2022, 325, 116553. [Google Scholar] [CrossRef]
  27. Cao, H.; Shao, H.; Zhong, X.; Deng, Q.; Yang, X.; Xuan, J. Unsupervised domain-share CNN for machine fault transfer diagnosis from steady speeds to time-varying speeds. J. Manuf. Syst. 2022, 62, 186–198. [Google Scholar] [CrossRef]
  28. Song, Y.; Cai, X.; Zhou, X.; Zhang, B.; Chen, H.; Li, Y.; Deng, W.; Deng, W. Dynamic hybrid mechanism-based differential evolution algorithm and its application. Expert Syst. Appl. 2022, 213, 118834. [Google Scholar] [CrossRef]
  29. Zhao, H.; Liu, J.; Chen, H.; Chen, J.; Li, Y.; Xu, J.; Deng, W. Intelligent Diagnosis Using Continuous Wavelet Transform and Gauss Convolutional Deep Belief Network. IEEE Trans. Reliab. 2022. early access. [Google Scholar] [CrossRef]
  30. Zhou, X.; Ma, H.; Gu, J.; Chen, H.; Deng, W. Parameter adaptation-based ant colony optimization with dynamic hybrid mechanism. Eng. Appl. Artif. Intell. 2022, 114, 105139. [Google Scholar] [CrossRef]
  31. Chen, H.; Miao, F.; Chen, Y.; Xiong, Y.; Chen, T. A Hyperspectral Image Classification Method Using Multifeature Vectors and Optimized KELM. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2781–2795. [Google Scholar] [CrossRef]
  32. Deng, W.; Ni, H.; Liu, Y.; Chen, H.; Zhao, H. An adaptive differential evolution algorithm based on belief space and generalized opposition-based learning for resource allocation. Appl. Soft Comput. 2022, 127, 109419. [Google Scholar] [CrossRef]
  33. Yao, R.; Guo, C.; Deng, W.; Zhao, H. A novel mathematical morphology spectrum entropy based on scale-adaptive techniques. ISA Trans. 2022, 126, 691–702. [Google Scholar] [CrossRef] [PubMed]
  34. Feng, R.; Li, X.; Bai, J.; Ye, Y. MID: A Novel Mountainous Remote Sensing Imagery Registration Dataset Assessed by a Coarse-to-Fine Unsupervised Cascading Network. Remote Sens. 2022, 14, 4178. [Google Scholar] [CrossRef]
  35. He, Z.Y.; Shao, H.D.; Wang, P.; Janet, L.; Cheng, J.S.; Yang, Y. Deep transfer multi-wavelet auto-encoder for intelligent fault diagnosis of gearbox with few target training samples. Knowl.-Based Syst. 2020, 191, 105313. [Google Scholar] [CrossRef]
  36. Li, W.; Zhong, X.; Shao, H.; Cai, B.; Yang, X. Multi-mode data augmentation and fault diagnosis of rotating machinery using modified ACGAN designed with new framework. Adv. Eng. Inform. 2022, 52, 101552. [Google Scholar] [CrossRef]
  37. Ren, Z.; Han, X.; Yu, X.; Skjetne, R.; Leira, B.J.; Sævik, S.; Zhu, M. Data-driven simultaneous identification of the 6DOF dynamic model and wave load for a ship in waves. Mech. Syst. Signal Process. 2023, 184, 109422. [Google Scholar] [CrossRef]
  38. Zhao, W.; Chen, X.; Bo, Y.; Chen, J. Semisupervised Hyperspectral Image Classification with Cluster-Based Conditional Generative Adversarial Net. IEEE Geosci. Remote Sens. Lett. 2019, 17, 539–543. [Google Scholar] [CrossRef]
  39. Zheng, C.; Wang, N.; Cui, J. Hyperspectral image classification with small training sample size using superpixel-guided training sample enlargement. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7307–7316. [Google Scholar] [CrossRef]
  40. Yu, C.; Zhou, S.; Song, M.; Chang, C.I. Semisupervised hyperspectral band selection based on dual-constrained low-rank representation. IEEE Geosci. Remote. Sens. 2022, 19, 5503005. [Google Scholar] [CrossRef]
  41. Zhang, Z.; Huang, W.G.; Liao, Y.; Song, Z.; Shi, J.; Jiang, X.; Shen, C.; Zhu, Z. Bearing fault diagnosis via generalized logarithm sparse regularization. Mech. Syst. Signal Process. 2022, 167, 108576. [Google Scholar] [CrossRef]
  42. Sha, A.; Wang, B.; Wu, X.; Zhang, L. Semisupervised Classification for Hyperspectral Images Using Graph Attention Networks. IEEE Geosci. Remote Sens. Lett. 2020, 18, 157–161. [Google Scholar] [CrossRef]
  43. He, X.; Chen, Y.; Ghamisi, P. Heterogeneous transfer learning for hyperspectral image classification based on convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3246–3263. [Google Scholar] [CrossRef]
  44. Chen, H.; Ye, M.; Lei, L.; Lu, H.; Qian, Y. Semisupervised dual-dictionary learning for heterogeneous transfer learning on cross-scene hyperspectral images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3164–3178. [Google Scholar] [CrossRef]
  45. Qu, Y.; Baghbaderani, R.K.; Li, W.; Gao, L.; Zhang, Y.; Qi, H. Physically constrained transfer learning through shared abundance space for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021. early access. [Google Scholar] [CrossRef]
  46. Yang, B.; Hu, S.; Guo, Q.; Hong, D. Multisource Domain Transfer Learning Based on Spectral Projections for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3730–3739. [Google Scholar] [CrossRef]
  47. Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G. A Particle Swarm Optimization-Based Flexible Convolutional Autoencoder for Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 2295–2309. [Google Scholar] [CrossRef] [Green Version]
  48. Sun, Z.; Li, J.; Liu, P.; Cao, W.; Yu, T.; Gu, X. SAR Image Classification Using Greedy Hierarchical Learning with Unsupervised Stacked CAEs. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5721–5739. [Google Scholar] [CrossRef]
  49. Xing, F.; Xie, Y.; Su, H.; Liu, F.; Yang, L. Deep Learning in Microscopy Image Analysis: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4550–4568. [Google Scholar] [CrossRef]
  50. Zhu, Y.; Zhang, S.; Li, X.; Zhao, H.; Zhu, L.; Chen, S. Ground Target Recognition Using Carrier-Free UWB Radar Sensor with a Semi-Supervised Stacked Convolutional Denoising Autoencoder. IEEE Sens. J. 2021, 21, 20685–20693. [Google Scholar] [CrossRef]
  51. Li, X.; Zhang, L.; Du, B.; Zhang, L.; Shi, Q. Iterative Reweighting Heterogeneous Transfer Learning Framework for Supervised Remote Sensing Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2022–2035. [Google Scholar] [CrossRef]
  52. Deng, C.; Xue, Y.; Liu, X.; Li, C.; Tao, D. Active Transfer Learning Network: A Unified Deep Joint Spectral–Spatial Feature Learning Model for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1741–1754. [Google Scholar] [CrossRef] [Green Version]
  53. He, X.; Chen, Y. Transferring CNN Ensemble for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 876–880. [Google Scholar] [CrossRef]
  54. Yan, H.; Zhang, E.; Wang, J.; Leng, C.; Peng, J. MTFFN: Multimodal Transfer Feature Fusion Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6008005. [Google Scholar] [CrossRef]
  55. Liu, S.; Cao, Y.; Wang, Y.; Peng, J.; Mathiopoulos, P.T.; Li, Y. DFL-LC: Deep Feature Learning with Label Consistencies for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 3669–3681. [Google Scholar] [CrossRef]
  56. Aydemir, M.S.; Bilgin, G. Semisupervised Hyperspectral Image Classification Using Deep Features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3615–3622. [Google Scholar] [CrossRef]
  57. Wei, W.; Xu, S.; Zhang, L.; Zhang, J.; Zhang, Y. Boosting Hyperspectral Image Classification with Unsupervised Feature Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502315. [Google Scholar] [CrossRef]
  58. Li, J.; Huang, X.; Gamba, P.; Bioucas-Dias, J.M.; Zhang, L.; Benediktsson, J.A.; Plaza, A. Multiple feature learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1592–1606. [Google Scholar] [CrossRef] [Green Version]
  59. Li, J.; Bruzzone, L.; Liu, S. Deep feature representation for hyperspectral image classification. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4951–4954. [Google Scholar]
  60. Miao, J.; Wang, B.; Wu, X.; Zhang, L.; Hu, B.; Zhang, J.Q. Deep feature extraction based on Siamese network and auto-encoder for hyperspectral image classification. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 8 July–2 August 2019; pp. 397–400. [Google Scholar]
  61. Du, B.; Xiong, W.; Wu, J.; Zhang, L.; Zhang, L.; Tao, D. Stacked convolutional denoising auto-encoders for feature representation. IEEE Trans. Cybern. 2016, 47, 1017–1027. [Google Scholar] [CrossRef]
  62. Yu, Y.; Hao, Z.; Li, G.; Liu, Y.; Yang, R.; Liu, H. Optimal search mapping among sensors in heterogeneous smart homes. Math. Biosci. Eng. 2023, 20, 1960–1980. [Google Scholar] [CrossRef]
  63. Xu, G.; Dong, W.; Xing, J.; Lei, W.; Liu, J.; Gong, L.; Feng, M.; Zheng, X.; Liu, S. Delay-CJ: A novel cryptojacking covert attack method based on delayed strategy and its detection. Digit. Commun. Netw. 2022. [Google Scholar] [CrossRef]
  64. Chen, H.Y.; Fang, M.; Xu, S. Hyperspectral remote sensing image classification with CNN based on quantum genetic-optimized sparse representation. IEEE Access 2020, 8, 99900–99909. [Google Scholar] [CrossRef]
  65. Xu, G.; Bai, H.; Xing, J.; Luo, T.; Xiong, N.N. SG-PBFT: A secure and highly efficient distributed blockchain PBFT consensus algorithm for intelligent Internet of vehicles. J. Parallel Distrib. Comput. 2022, 164, 1–11. [Google Scholar] [CrossRef]
  66. Yu, C.; Liu, C.; Yu, H.; Song, M.; Chang, C.I. Unsupervised domain adaptation with dense-based compaction for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12287–12299. [Google Scholar] [CrossRef]
Figure 1. The stacked convolutional autoencoder network model.
Figure 1. The stacked convolutional autoencoder network model.
Sensors 22 08881 g001
Figure 2. The classification framework based on the stacked convolutional autoencoder.
Figure 2. The classification framework based on the stacked convolutional autoencoder.
Sensors 22 08881 g002
Figure 3. The classification effect of different convolutional self-coding networks.
Figure 3. The classification effect of different convolutional self-coding networks.
Sensors 22 08881 g003
Figure 4. The impact of the transfer strategy on the classification accuracy.
Figure 4. The impact of the transfer strategy on the classification accuracy.
Sensors 22 08881 g004
Figure 5. (a) The original images; (b) the CAE classification effect; (c) the SCAE classification; (d) the SCAE-MT classification effect; and (e) the 16 class labels.
Figure 5. (a) The original images; (b) the CAE classification effect; (c) the SCAE classification; (d) the SCAE-MT classification effect; and (e) the 16 class labels.
Sensors 22 08881 g005
Table 1. Basic information of the Indian Pines and the Salinas datasets.
Table 1. Basic information of the Indian Pines and the Salinas datasets.
DataIndian PinesSalinas
Collection locationIndiana, USACalifornia, U.S.
Collection equipmentAVIRISAVIRIS
Spectral coverage (um)0.4∼2.50.4∼2.5
Data size (pixel)145 × 145512 × 217
Spatial resolution (m)203.7
Number of bands220224
Number of bands after denoising200204
Sample size10,24954,129
Number of categories1616
Table 2. The 16 types of ground features in Indian Pines and Salinas datasets.
Table 2. The 16 types of ground features in Indian Pines and Salinas datasets.
Indian PinesSalinas
CategoryClass NameNumber of Samples Class NameNumber of Samples
1Alfalfa46Brocoli_green_weeds_12009
2Corn-notill1428Brocoli_green_weeds_223726
3Corn-min830Fallow1976
4Corn237Fallow_rough_plow1394
5Grass-pasture483Fallow_smooth2678
6Grass-trees730Stubble3959
7Grass-pastue-mowed28Celery3579
8Hay-windrowed478Grapes_untrained11,271
9Oats20Soil_vinyard_develop6203
10Soybean-notill972Corn_senesced_green_weec3278
11Soybean-min2455Lettuce_romainc_4wk1068
12Soybean-clean593Lettuce_romainc_5wk1927
13Wheat205Lettuce_romainc_6wk916
14Woods1265Lettuce_romainc_7wk1070
15Bldg-Grass-Tree-Drives386Vinyard_untraincd7268
16Stone-Steel-Towers93Vinyard_vertical_trellis1870
Table 3. Effects of different convolutional self-encoders on classification accuracy.
Table 3. Effects of different convolutional self-encoders on classification accuracy.
Training Set (40%)Training Set (80%)
Number123123
OA (%)90.6795.6595.8293.3297.0397.17
Table 4. Overall classification accuracy under different proportions of training sample sets.
Table 4. Overall classification accuracy under different proportions of training sample sets.
IndexMethodSalinas Dataset
5%10%15%
No transfer strategy80.5882.7487.13
OA (%)Introduction of transfer strategy83.1188.9290.24
Table 5. The overall classification accuracy under different proportions of sample sets.
Table 5. The overall classification accuracy under different proportions of sample sets.
MethodCAE (OA%)SCAE (OA%)SCAE-MT (OA%)
5%79.1981.6283.11
10%84.7786.4188.92
15%86.5387.8290.24
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, H.; Chen, Y.; Wang, Q.; Chen, T.; Zhao, H. A New SCAE-MT Classification Model for Hyperspectral Remote Sensing Images. Sensors 2022, 22, 8881. https://doi.org/10.3390/s22228881

AMA Style

Chen H, Chen Y, Wang Q, Chen T, Zhao H. A New SCAE-MT Classification Model for Hyperspectral Remote Sensing Images. Sensors. 2022; 22(22):8881. https://doi.org/10.3390/s22228881

Chicago/Turabian Style

Chen, Huayue, Ye Chen, Qiuyue Wang, Tao Chen, and Huimin Zhao. 2022. "A New SCAE-MT Classification Model for Hyperspectral Remote Sensing Images" Sensors 22, no. 22: 8881. https://doi.org/10.3390/s22228881

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop