Next Article in Journal
Solid Precipitation and Visibility Measurements at the Centre for Atmospheric Research Experiments in Southern Ontario and Bratt’s Lake in Southern Saskatchewan
Previous Article in Journal
Quantifying the Loss of Coral from a Bleaching Event Using Underwater Photogrammetry and AI-Assisted Image Segmentation
Previous Article in Special Issue
Rethinking Representation Learning-Based Hyperspectral Target Detection: A Hierarchical Representation Residual Feature-Based Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Random Shuffling Data for Hyperspectral Image Classification with Siamese and Knowledge Distillation Network

1
The School of Information and Electromechanical Engineering, Jiangxi Science and Technology Normal University, Nanchang 330013, China
2
Key Laboratory of System Control and Information Processing, Ministry of Education, Shanghai 200240, China
3
Shanghai Key Laboratory of Intelligent Sensing and Recognition, Shanghai Jiao Tong University, Shanghai 200240, China
4
School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 639798, Singapore
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(16), 4078; https://doi.org/10.3390/rs15164078
Submission received: 26 July 2023 / Revised: 13 August 2023 / Accepted: 17 August 2023 / Published: 18 August 2023

Abstract

:
Hyperspectral images (HSIs) are characterized by hundreds of spectral bands. The goal of HSI is to associate the pixel with a corresponding category label by analyzing subtle differences in the spectrum. Due to their excellent local context modeling capabilities, Convolutional Neural Network (CNN)-based methods are often adopted to complete the classification task. To verify whether the patch-data-based CNN methods depend on the homogeneity of patch data during the training process in HSI classification, we designed a random shuffling strategy to disrupt the data homogeneity of the patch data, which is randomly assigning the pixels from the original dataset to other positions to form a new dataset. Based on this random shuffling strategy, we propose a sub-branch to extract features on the reconstructed dataset and fuse the loss rates (RFL). The loss rate calculated by RFL in the new patch data is cross combined with the loss value calculated by another sub-branch in the original patch data. Moreover, we construct a new hyperspectral classification network based on the Siamese and Knowledge Distillation Network (SKDN) that can improve the classification accuracy on randomly shuffled data. In addition, RFL is introduced into the original model for hyperspectral classification tasks in the original dataset. The experimental results show that the improved model is also better than the original model, which indicates that RFL is effective and feasible. Experiments on four real-world datasets show that, as the proportion of randomly shuffling data increases, the latest patch-data-based CNN methods cannot extract more abundant local contextual information for HSI classification, while the proposed sub-branch RFL can alleviate this problem and improve the network’s recognition ability.

1. Introduction

Compared to traditional two-dimensional RGB images, hyperspectral images (HSIs) have many continuous spectral bands. Therefore, they also contain more information that allows us to better perform tasks, such as object detection, change detection, and geological exploration. Land cover classification is one of the most important issues within these fields, as it aims to apply specific semantic labels to the pixels of the entire HSI on the basis of the unique spatial-spectral features of the HSI. In recent years, there has been increasing interest in Deep Learning (DL) methods for solving the problem of HSI classification. Depending on the shape of the training data, we can divide these methods into the following two categories: the single-pixel-based methods and the patch-data-based methods.
Existing DL models often ignore the inherent relationship between pixels in patch data because they are designed for Euclidean data. In recent years, Graph Convolutional Networks (GCNs) have received increasing attention as a representative of the single-pixel-based methods due to their ability to perform convolutions on arbitrarily structured graphs. By encoding HSI into a graph, the correlation between adjacent land cover can be explicitly exploited, and the spatial context structure of HSI can be better modeled by GCNs [1]. Shahraki et al. [2] proposed a model combining 1-D Convolutional Neural Networks and GCNs for HSI classification. However, due to the large number of pixels in HSIs, using each pixel as a node of a graph has been shown to incur huge computational costs and limit its applicability. To solve this issue, Hong et al. [3] developed a new model named mini-GCN, which can be used to train large GCNs in a mini-batch process. Liu et al. [1] proposed a heterogeneous model called CEGCN, in which CNN and GCN sub-networks generate complementary feature information at pixel and super-pixel levels. To explore Non-Euclidean structures and reduce high computational costs, Bai et al. [4] developed a graph attention model with an adaptive graph structure mining approach (GAT-AGSM). Liu et al. [5] designed a multi-level network based on U-Net operating on a superpixel structured graph, named MSSGU, to learn multi-level features on multi-level graphs and overcome the limitation of specific superpixel segmentation in modeling.
Generally, the methods mentioned above mainly belong to single-pixel-based methods. Recently, researchers still pay more attention to the patch-data-based methods, which can extract more discriminative information and deeper features in model training. In particular, Convolutional Neural Network (CNN) is very popular among patch-data-based methods. Since the feature information contained in a single pixel is limited, CNN cannot extract local context information from the data. To solve this problem, researchers changed the shape of the input data. Specifically, the pixel is placed in the center of the patch data. The pixels around the classified pixel are also included in the patch data. This allows the model to extract more local contextual information from the patch data to enhance the classification accuracy in the central pixel.
In the early days of 2016, researchers mainly focused on HSI classification based on 1D-CNNs and 2D-CNNs [6]. Chen et al. [7] proposed a model that uses Stacked Autoencoders (SAE) for high-level information extraction and 1D-CNN for low-level information extraction. In addition to methods based on 1D-CNNs, 2D-CNN-based methods are often introduced into the HSI network to extract spatial feature information. Shen et al. [8] proposed a model called ENL-FCN to incorporate the long-range context information, which is constructed by a deep fully convolutional network and an efficient non-local module. Since residual blocks have been extensively adopted in the HSI network, Zhong et al. [9] developed a model to improve the accuracy of HIS classification, which is named the Spectral-Spatial Residual Network (SSRN), and greatly enhanced the feature utilization rate by using front-layer feature information to complement back-layer features. Furthermore, Paoletti et al. [10] proposed a model named DPRN that also adopted the residual module. Inspired by the 3D convolution, 3D convolution blocks have been widely applied to HSI networks, which utilize spectral-spatial information to achieve more satisfactory classification results. In [11], Zhong et al. developed 3D Deep-Residual Networks (3D-ResNets) to reduce the influence of model size. To mine deeper spectral-spatial feature information from HSIs, Zhang et al. [12] also used 3D-CNNs to construct a 3D-DenseNet model.
In recent years, combining 2D and 3D convolutions and optimizing the structure of CNN-based models have gradually become a hotspot. Combining the advantage of 2D-CNN and 3D-CNN, some researchers designed a dual-branche structure to enhance the classification accuracy of HSI. For instance, Zheng et al. [13] proposed a mixed CNN with covariance pooling, named MCNN-CP, where covariance pooling is used to mine second-order information from the spectral-spatial feature. Roy et al. [14] developed a model that is a 3D convolution block combined with a 2D convolution block, which is called a hybrid spectral CNN (HybridSN). By introducing an efficient residual structure, the network parameters can be optimized, and a lightweight design can reduce the computational complexity of the model. Wu et al. [15] designed a re-parametrized network, abbreviated as RepSSRN, which reparametrizes the Spectral-Spatial Residual Network (SSRN). Inspired by the Dense Convolutional Network, Wang et al. [16] proposed a fast dense spectral-spatial convolution (FDSSC) algorithm. In another work [17], a network named CMR-CNN adopted the 3D residual blocks followed by the 2D residual blocks together to capture the spatial-spectral feature information of the HSI. To reduce the network parameters and computational complexity, Meng et al. [18] proposed a lightweight spectral-spatial convolution HSI classification module (LS2CM), and Li et al. [19] designed a lightweight network architecture (LiteDenseNet). The computational complexity and network parameters in their work are much lower than counter-intuitive deep learning methods. Although these methods mentioned above are effective in improving the classification performance of HSI, it is difficult to overcome the issue caused by the limited training samples and increasing network layers. So, much work has been done to introduce the attention mechanism and transformer models into the HSI network. Sun et al. [20] constructed a model called the spectral-spatial attention network (SSAN) to acquire important spectral-spatial information in the attention regions of patch data. Li et al. [21] proposed a model called DBDA to capture a variety of spectral-spatial information, which is a dual-branch network with two attention mechanisms. He et al. [22] designed a model called the Spatial-Spectral Transformer (SST), which extracted spatial features via the VGGNet network and established the relationship between adjacent spectra by using the dense transformer blocks. Similarly, Sun et al. [23] designed a network that adopts a Gaussian weighted feature tokenizer to capture high-level semantic features, which is called SSFTT. By improving the traditional transformer model, Hong et al. [24] designed a novel method that is widely used in HSI classification tasks, which is called SpectralFormer (SF). Although these networks are excellent at capturing spectral signatures, these models cannot capture the local contextual information of patch data well, and they make insufficient use of the spatial features in HSI.
Besides the above limitations, the patch-data-based methods also tend to rely on the local neighborhood information of the patch data in the training process, and sometimes, overfitting is caused by improper setting of the training proportion. To verify whether the patch-data-based CNN methods depend on the homogeneity of patch data during the training process and evaluate the ability of the HSI method to extract spatial location information, we rethink the HSI classification process from the data perspective in the patch-data-based method and design a novel strategy to reconstruct the original dataset. Based on this strategy, we also propose a new model based on the Siamese and Knowledge Distillation Network (SKDN) to complete the classification task. The most important contributions of this work can be summarized in the following way.
  • We use the proposed strategy of randomly shuffling data to explore the influence of patch data homogeneity features in HSI classification networks. Specifically, this strategy involves randomly assigning the pixels in the original dataset to other locations to construct a new dataset. Therefore, the new patch data after the random strategy contains richer information about species categories;
  • We propose a sub-branch to extract features from the reconstructed dataset and fuse the loss values (RFL), which uses a designed loss function in RFL to compute and fuse loss values from two sub-branches. The novelty of this loss function is that the loss rate computed by RFL in the new data is cross combined with the loss rate calculated in the original data from another sub-branch to further optimize the network. Thus, the proposed network can not only enhance the recognition ability of the model, but also increase the classification accuracy on randomly shuffled data, which is based on the Siamese and Knowledge Distillation Network (SKDN) and is constructed from two sub-branches;
  • We also introduce the proposed sub-branch RFL into the original network to further explore the effectiveness of the RFL, and we let the original model and its improved model achieve HSI classification on the original dataset. The experiments show that the classification performance of the improved model is better than that of the original model, so it can also prove that the proposed sub-branch is effective and feasible;
  • Experiments conducted on several typical datasets show that, as the proportion of randomly shuffled data increases, the latest patch-data-based CNN methods are unable to extract more abundant local contextual information for HSI classification, while the proposed sub-branch can alleviate this problem.
The rest of this paper consists of the following sections: Section 2 presents the methodology, Section 3 presents and analyzes the experiments, Section 4 discusses the usefulness of the proposed methods, and Section 5 is the work’s conclusion.

2. Methodology

More details of the developed random shuffling strategy and the new network SKDN constructed by the proposed sub-branch will be introduced and discussed carefully in this section.

2.1. Randomly Shuffling the Pixels

It is well known that most pixels in the patch data belong to the same category in the CNN-based networks in HSI classification. CNNs have shown that they can capture spectral-spatial feature information and local contextual information in training patch data. They largely rely on the data homogeneity of the patch data in the model’s training, leading to overfitting if the value of the training ratio is not chosen correctly. Here, we designed a random shuffling strategy to disrupt the data homogeneity of the patch data, which is randomly assigning the pixels from the original dataset to other positions to form a new dataset. The random shuffling of the pixel scheme can be seen in Figure 1. It consists of reconstructing the dataset by randomly shuffling the position of pixels. The false color image of the rebuilt dataset based on Indian Pines with various random shuffling ratios is displayed in Figure 2. We found that as the proportion of random shuffling increased, so did the category differences in each pixel cube. This random shuffling strategy can increase the diversity of extracted features from the new dataset. Specifically, for each new neighboring cube that contains many types of categories, CNNs can learn more category information about different ground objects and extract more local context information during feature extraction. We also present hereafter the advantage of the proposed random shuffling strategy over the traditional training dataset in HSI classification. In general, the 3D-conv operations are adopted to acquire more spectral-spatial features in HSI, while the 2D-con operations pay more attention to acquiring more spatial features. If most of the training samples in the neighboring cube belong to the same category, the model can easily lead to ignoring some category information about other categories, resulting in a large reliance on the data homogeneity of the patch data in the training process. By randomly rearranging the pixel positions, the patch data of HSI can be reconstructed. This not only serves to test the robustness of the network, but also helps the CNN-based model extract more local context information and category information in the patch data. This strategy also increases the training difficulty of the CNN-based network in the training process. Therefore, in comparison to the traditional method of training patch data, the proposed strategy enables us to rethink the task of HSI classification from the perspective of patch data with feature extraction.

2.2. Hyperspectral Image Classification Network SKDN

By giving the traditional Siamese network a classification structure, it can have the ability to classify and additionally calculates a cross-entropy loss for each individual sample. Inspired by this method, we construct our network for HSI classification, with the difference that we only compute the cross-entropy loss in our structure. The Knowledge Distillation (KD) network is designed to transfer the extracted knowledge from a larger model into a smaller network for knowledge preserving and computationally inexpensive deep models [25]. Based on this theory, we design two branches to construct our model, where one branch extracts features from the original dataset as a teacher network, and the other branch extracts features from the reconstructed dataset as a student network. The difference is that we design a loss function to merge the loss rates of the two branches. Although the neighborhood of the central pixel in the reconstructed dataset contains a variety of species, it also makes it more difficult for the model to classifying central pixels of patch data. Therefore, we propose a network SKDN based on the input form of the Siamese network and the distillation mode of the KD network to improve the classification ability. The structure of the network SKDN is illustrated in Figure 3. It consists of two sub-branches, which are used to extract deep features and local context information in the patch data from the original dataset and the reconstructed dataset, respectively. The weights are not shared between the backbones used to extract features in the two branches. Finally, the proposed new loss function is used to fuse the loss rates of the two branches in order to improve the classification effect of the model.
Figure 1. The use of random shuffling strategy to reconstruct the HSI dataset. H, W, and B, respectively, represent the length of HSI, the width of HSI, and the spectral number of HSI. It is randomly assigning the pixels from the original dataset to other positions to form a new dataset. We can that see the False Color Image of the new dataset is different from the False Color Image of the original dataset. The Planar Effect Picture shows the strategy applied to a sample 3 × 3 image. Note that the position of the spectrum corresponding to the shuffled pixel also changes as its pixel position changes.
Figure 1. The use of random shuffling strategy to reconstruct the HSI dataset. H, W, and B, respectively, represent the length of HSI, the width of HSI, and the spectral number of HSI. It is randomly assigning the pixels from the original dataset to other positions to form a new dataset. We can that see the False Color Image of the new dataset is different from the False Color Image of the original dataset. The Planar Effect Picture shows the strategy applied to a sample 3 × 3 image. Note that the position of the spectrum corresponding to the shuffled pixel also changes as its pixel position changes.
Remotesensing 15 04078 g001
Figure 2. The False Color Image of the reconstructed dataset based on Indian Pines with different random shuffling ratios.
Figure 2. The False Color Image of the reconstructed dataset based on Indian Pines with different random shuffling ratios.
Remotesensing 15 04078 g002
Figure 3. The architecture of the developed network SKDN, which consists of two sub-branches. Sub-branch 1 extracts the information and computes the loss value as the knowledge of the original dataset. Indeed, the sub-branch RFL extracts the feature and computes the loss value of the reconstructed dataset, then fuses the knowledge in the new loss function Loss.
Figure 3. The architecture of the developed network SKDN, which consists of two sub-branches. Sub-branch 1 extracts the information and computes the loss value as the knowledge of the original dataset. Indeed, the sub-branch RFL extracts the feature and computes the loss value of the reconstructed dataset, then fuses the knowledge in the new loss function Loss.
Remotesensing 15 04078 g003
Differentiated input strategy. Based on the random shuffling strategy, we also design a different input strategy that feeds patch data from different datasets into the dual branches of our network SKDN, providing more local context information to the model, while most existing methods use the original dataset. To be specific, we are feeding two sets as the input patch data for two branches in SKDN: the data input of one branch from the original patch data, and the other one from the reconstructed patch data. So, the input data in the network SKDN are different from the traditional Siamese networks, which often feed with different size input data from the same dataset for its model training. Since the neighborhood of the central pixel in the reconstructed dataset contains a variety of species, these patch data have richer species categories than the original, which means that they can provide more local context information to the model’s training to strengthening the model’s recognition ability. We also transfer the knowledge in the feature extraction backbones to improve the adaptability of the two sub-branches with different input data.
Weighted Cross-Entropy Loss. The sub-branch RFL of network SKDN includes two loss functions. The computational process of the loss function is given by
L = α × L ori + β × L rec L ori y ^ n , y n = L rec y ^ n , y n = 1 N n = 0 N 1 y n log y ^ n
We use two weighted parameters ( α and β ) to fuse the two loss values learned from the dual branch to balance the differential information due to different inputs. Here, we also set the initial values of α and β to 1 and 0.5, respectively. y n is the label of the input patch-data, and y ^ n is the predicted value of the network output result. The L rec represents the loss value calculated on the RFL, where the computational process is performed on the reconstructed dataset. The L ori represents the loss value calculated on sub-branch 1, where the computational process is performed on the original dataset. Note that the input patch data that are sent to the two sub-branches are different, so the two labels y n in L ori and L rec are also different. Furthermore, we still adopt a common Adam optimizer to optimize the network with L.

3. Experiments

To test the feasibility of the strategy, four real-world datasets from HSI were selected for the experiment in this paper. They are Indian Pines, Pavia University, Salinas, and Kennedy Space Center. We can see more details of these public datasets in Table 1.
Indian Pines (IP) is taken by the AVIRIS sensors. The size of this HSI dataset is 145 × 145, including 200 spectral bands and 24 noisy bands that cannot be reflected by water. It has 10,249 pixels that can be divided into 16 land categories for HSI classification. The False Color image and Ground Truth map are shown in Figure 4.
Pavia University (PU) is taken by the ROSIS sensors. The size of this image data is 615 × 345, which includes 9 categories. This PU dataset only leaves 103 bands and removes 12 bands. The False Color image and Ground Truth map are shown in Figure 5.
Salinas (SA) is taken by the AVIRIS sensors. It remains 204 bands after removing the noisy bands in HSI. The dataset size of SA is 512 × 217, which holds 54,129 pixels that can be divided into 16 land categories with true labels. The False Color image and Ground Truth map are shown in Figure 6.
Kennedy Space Center (KSC) is taken by the AVIRIS sensors. The size of KSC data is 512 × 614, which contains 13 land categories with true labels in total. This dataset only leaves 176 bands and removes 48 bands that cannot be reflected by water. The False Color image and Ground Truth map are shown in Figure 7.
In addition to reducing the spectral bands of the PU by PCA to 90, the spectral bands of the KSC by PCA to 120, and the spectral bands of the IP and SA dataset by PCA to 110, we also set some other parameters. During the training process, we set the learning rate, training epochs, and patch size to 0.001, 200, and 13, respectively. Note that due to the reconstruction of the HSI dataset during training, we still chose the usual training ratio in these four datasets, which is set to 0.05, 0.1, 0.005, and 0.2, respectively. In the reconstructed datasets experiments, we additionally select some training samples from the original dataset to compute the knowledge loss that is used in the SKDN training. In the original datasets experiments, we additionally choose some training samples from the reconstructed dataset to compute the knowledge loss. In addition, three well-known numerical indicators, namely overall accuracy (OA), average accuracy (AA), and kappa coefficient (kappa), are used to evaluate the performance of different classification methods. OA is the ratio of the number of correctly classified samples to the total number of test samples, AA is the average accuracy across the accuracy of all classes, and Kappa is an available measure of agreement between ground truth and classification maps. All experiments were performed using four patch-data-based CNN methods in HSI classification, which are DPRN [10], CEGCN [1], SSTFF [23] and MRViT [26]. All results of these experiments in four real-world datasets are shown in Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11 and Figure 8, Figure 9, Figure 10 and Figure 11.

3.1. The Experimental Results of the SKDN on Reconstructed Data Based Four Real-World Datasets

Since our random shuffling strategy disrupts the data homogeneity of the original patch data, we introduce the designed sub-branch into all original networks to form it improved network SKDN, and we complete these experiments on four reconstructed datasets. Note that we additionally chose some training samples from the original datasets for the improved model (SKDN)’s training, where the proportions of these original datasets in the model’s training are the same as the ratio in reconstructed datasets. In these experiments, we set the initial values of α and β to 0.5 in the weighted cross-entropy loss function when the random shuffling ratio was 0%. Table 2, Table 3 and Table 4 report the experimental results of the networks on these four datasets, and the visual results of these networks can be seen in Figure 8 through Figure 11.
(1) The analysis of the experiments on the reconstructed dataset based on Indian Pines: the second row in Table 2 lists the quantitative results of the different methods, and Figure 8 shows the prediction maps corresponding to these methods for the reconstructed IP dataset. Compared to the network DPRN, the SKDN(DPRN) model improves the OA by 11.45% and achieves a better visual result (Figure 8c) when the random ratio is 20%. It is 12.88% higher than DPRN when the random mixture ratio is 50%. This shows that the result of the improved model is better than that of the original model, which proves the effectiveness of RFL in the network. Obviously, SKDN(SSTFF) also has a better classification performance compared to SSTFF. The value of OA is 5.11% higher than that of SSTFF when the random shuffling ratio is 20%. It is 2.85% higher than that of SSTFF when the ratio of the random mixture is 50%. From Figure 8f, it is easy to see that the improved model can extract more discriminative local context information and category information in the patch data. The corresponding values of the SKDN(MRViT) model in Table 2 also show that the improved network performs better than MRViT. The value of OA is 5.5% higher than MRViT when the random mixture proportion is 20%. It is 7.09% higher than MRViT when the proportion of the random mixture is 50%. This also proves that the designed sub-branch RFL is more effective in the reconstructed dataset.
(2) The analysis of the experiments on the reconstructed Pavia University dataset: the third row in Table 2 illustrates the quantitative results of the different methods, and Figure 8 shows the prediction maps corresponding to these methods on for reconstructed PU dataset. It is clear that our improved-method SKDN(DPRN) performs better than the original method DPRN. The value of the OA is 5.37% higher than that of DPRN when the random mixture proportion is 20%. It is 7.36% higher than that of DPRN when the proportion of random mixture is 50%. By analyzing these visual maps, it is easy to find that SKDN(DPRN) in Figure 9c has a lower prediction error range than DPRN in Figure 9b. In contrast to SSTFF, the improved SKDN(SSTFF) model in Figure 9e also has fewer misclassifications. This can be confirmed by the improved OA value (0.55% and 1.19%) in Table 2. Figure 9g also shows that the improved SKDN(MRViT) method has better performance, which is also confirmed by its OA value (85.19% and 66.79%) in Table 2. This is direct evidence that the sub-branching used to construct the SKDN is effective.
(3) The analysis of the experiments on the reconstructed Salinas dataset: the fourth row in Table 2 shows the classification results of the different methods. The corresponding prediction graphs for the reconstructed SA dataset are shown in Figure 10. Compared with the first network, the improved-network SKDN(DPRN) improves the OA by 17.9% when the random shuffling ratio is 20% and achieves a better classification result with fewer misclassified regions in the second row in Figure 10c. It is 11.7% higher than DPRN when the proportion of random shuffling is 50%. These results show that RFL is useful for the model. It is clear that the experimental results of SKDN(SSTFF) are superior to those of SSTFF, which can also be seen in Figure 10e and is confirmed by the better OA value (5.49% and 12.91%) in Table 2. In Figure 10f,g, the experimental results of SKDN(MRViT) are superior to those of MRViT with fewer training samples. The SKDN(MRViT) model improves OA by 4.01% when the random ratio is 20%. It is 10.72% higher than MRViT when the proportion of random shuffling is 50%. These results can prove that RFL can extract more discriminative local context information in the new dataset.
(4) The analysis of the experiments on the reconstructed dataset Kennedy Space Center: the fifth row in Table 2 shows the quantitative results of different methods, and Figure 11 is the prediction maps corresponding to these methods on the reconstructed KSC dataset. It is noticeable that the OA value in the experimental results of the improved model is better than that of the original model. In the following, the results of these methods are only briefly analyzed. Compared with the network DPRN, the model SKDN(DPRN) improves OA by 25.68% and achieves a better result, as shown in Figure 11c, where the proportion of random shuffling is 20%. It is 28.67% higher than the DPRN when the proportion of random mixture is 50%. Moreover, the classification capability of SKDN(SSTFF) is better than that of SSTFF, which is also evidenced by the better OA value (5.8% and 4.45%). Compared to the MRViT, SKDN(MRViT) also achieves better classification performance (82.38% and 50.91%) in Table 2. SKDN(MRViT) has less misclassification than MRViT in its domain, which can be seen in Figure 11g. All the results of these experiments proved that our model SKDN can indeed extract more local context information and category information from random shuffling patch data for HSI classification. Our proposed sub-branch RFL, which fuses the loss rates of two sub-branches in its loss function, can effectively improve the classification accuracy in random shuffling data. All results of these experiments have proven that our model SKDN can really extract more local context information and category information from random shuffling patch data for HSI classification. Our proposed sub-branch RFL, which fuses the loss rates of two sub-branches in its loss function, can effectively improve the classification accuracy in random shuffling data.

3.2. The Experimental Results of the SKDN on Four Real-World Datasets

To further validate the usefulness of the proposed sub-branch RFL, we introduce RFL into the original network to form its improved network SKDN and let the improved model and the original model achieve the HSI classification task on the four original datasets. Note that we additionally chose some training samples from the reconstructed datasets for the improved model (SKDN) training, where the proportions of these reconstructed datasets in model training are the same as the ratio in the original datasets. Table 5 lists the quantitative results of the original methods and the improved methods on four original datasets. We also set the initial values of α and β to 0.5 in the weighted cross-entropy loss function when the random shuffling ratio was 0%. Hereinafter, we just briefly analyze the results of these methods on four original datasets.
The second row in Table 5 lists the experimental results of the original networks and the improved models for the IP dataset. The classification result of SKDN(DPRN) is 2.43% higher than the result of DPRN for OA when the proportion of random mixture is 20%; the value of the OA also improves by 7.76% when the random shuffling ratio is 50%. Compared with the network SSTFF, the improved model SKDN(SSTFF) has a better classification performance, which can be verified by its improved OA value (11.35% and 41.6%) in Table 5. We also find that the improved network SKDN(MRViT) (97.24% and 96.48%) performs better than MRViT. This proves that the proposed RFL is effective in the original dataset.
The third row in Table 5 shows the experimental results of the original networks and the improved models for the PU dataset. Compared with the original method DPRN, the improved-method SKDN(DPRN) improves OA by 4.15% when the random shuffling ratio is 20%. It is 4.12% higher than DPRN when the proportion of random shuffling is 50%. We also can see that the improved-method SKDN(SSTFF) also performs better than SSTFF, which can also be verified by its OA value (99.61% and 98.61%). Compared to the MRViT, SKDN(MRViT) also achieves a better classification performance, which can be verified by its improved OA value (2.82% and 4.78%). These results demonstrate that the RFL is useful in our model.
The fourth row in Table 5 shows the experimental results of the original networks and the improved networks for the SA dataset. Compared with DPRN, the improved method SKDN(DPRN) improves OA by 2.77% when the random shuffling ratio is 20%. It is 4.94% higher than DPRN when the proportion of random shuffling is 50%. It is shown that the improved-method SKDN(SSTFF) also performs better than SSTFF, which can be verified by its improved OA value (6.35% and 35.6%). Compared to the MRViT, SKDN(MRViT) also achieves a better classification performance, which can also be seen in its OA value (96.24% and 94.07%) in Table 5. These results also prove that the RFL can capture the discriminative feature in the original dataset.
The fifth row in Table 5 lists the experimental results of the original networks and the improved networks for the KSC dataset. Compared with DPRN, the OA of SKDN(DPRN) improves by 4.31% and 5.86% when the random shuffling ratios are 20% and 50%, respectively. Moreover, the OA of SKDN(SSTFF) is 31.08% higher than that of SSTFF when the proportion of random shuffling is 20%. It is 39.59% higher than SSTFF when the random shuffling ratio is 50%. Compared to the MRViT, FSKDN(MRViT) also achieves a better classification performance, which can be verified by its improved OA value (2.74% and 8.6%). All results of these experiments show that the classification capacity of the improved model SKDN is superior to that of the original model, which in turn proves that the proposed RFL sub-branch is effective and feasible.

3.3. The Experimental Results of Single Pixel-Based Methods on Reconstructed Data Based on Four Real-World Datasets

We set the value of the patch-size parameter to 1 in the patch-data-based CNN methods in the HSI classification so that they become the single-pixel-based methods in this experiment. Table 8, Table 9, Table 10 and Table 11 show the results of each network with different random shuffling ratios and training ratios in the HSI classification. We see that the accuracy of these models improves slightly when the training fraction is increased and the random mixture ratio is kept constant. However, when the same training fraction is kept and the random mixture ratio is increased, the classification results of the CNN-based method and the Transformer-based method, such as the DPRN and the SSTFF, vary only within a certain value interval, and the OA value does not change significantly, but the classification results of the GCN-based network show a downward trend. Using the experiments of the reconstructed dataset based on IP as an analysis case and setting the proportion of each method to 5% in the model’s training, we find that the results of the DPRN are in the range of 64% to 69% when the proportion of random shuffling increases, the results of the SSTFF are in the range of 62% to 65%, the results of the MRViT are in the range of 65% to 68%, and the results of the SKDN(MRViT) are in the range of 72% to 75%. However, the results of the CEGCN decreased from 97% to 19%. These experiments show that no matter how the proportion of random shuffling increases, the classification accuracy results of the single pixel-based models for HSI classification do not change significantly. This may indicate that the single pixel-based methods do not use the local context relationship between pixels for the classification of HSIs, which also proves that the strategy of reconstructing the dataset has no major effect on the experimental classification results of the single pixel-based networks for the classification of HSIs.

4. Discussion

Several conclusions can be drawn regarding the role of the random shuffle strategy and the sub-branch RFL. First, the classification results of these experiments on the reconstructed datasets and original datasets prove that our model SKDN can extract more local context information and category information from random shuffling patch data for HSI classification. Furthermore, these experiments reveal that the patch-data-based CNN methods really depend on the data homogeneity of patch-data in the training process. This can also be verified by the classification results of these original models that present a downward trend as the proportion of random shuffling increases. What is more, the RFL is introduced into the original model to construct the SKDN, and the new loss function in the SKDN can fuse loss values from two sub-branches that enhance the recognition ability of the original method. Although the combination of the random shuffling strategy and the RFL cannot achieve a significant effect in the single pixel-based method’s training process, it also enriches the related research on the single pixel-based methods in HSI classification. The work in this paper provides certain reference values for subsequent research into the patch-data-based CNN methods.

5. Conclusions

To resolve the issue of whether the patch-data-based CNN methods really depend on the data homogeneity of the patch-data in HSI, we propose a strategy of randomly shuffling pixel data to validate the influence of the characteristics of the patch-data in HSI classification networks. Specifically, it is to randomly assign the pixels in the original dataset to other locations. To ensure that the classification networks of HSI can learn more local contextual information and category information regarding the patch data, we also propose a sub-branch to fuse the loss rates. In detail, the loss rate calculated by this sub-branch in the new patch data is cross combined with the loss rate calculated by another branch in the original patch data to construct a new hyperspectral classification network named SKDN. Extensive experiments on these four datasets indicate that as the proportion of randomly shuffled data increases, the patch-data-based CNN methods are no longer able to extract more discriminative local context information for classifying pixels in HSI. However, the network SKDN constructed by our proposed sub-branch can effectively address this problem, which can improve the classification accuracy of randomly shuffled data. In addition, we also introduce the proposed sub-branch into the original network and let the improved model and the original model achieve HSI classification on the original dataset to explore the effectiveness of the sub-branch RFL. The results of all experiments show that the classification performance of the improved model is better than that of the original model, so they also prove that the proposed sub-branch is effective and feasible.

Author Contributions

Conceptualization, Z.Y. and Y.C.; methodology, Z.Y. and Y.C.; validation, Y.C.; formal analysis, Z.Y., Y.C., T.Z. and J.J.; writing—original draft preparation, Z.Y., Y.C., T.Z. and J.J.; funding acquisition, Z.Y., X.Z. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62261026, 62262026, 62003251, and 62201343), the Education Department Foundation of Jiangxi Province (No. GJJ2201357 and GJJ211111), the Jiangxi Natural Science Foundation (No. 20232ACB212006 and 20232BAB202020), the Key Laboratory of System Control and Information Processing, Ministry of Education (Scip202106), and the Shanghai Key Laboratory of Navigation and Location Based Services (No. SKLNLBS2023001).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of the data; in the writing of the manuscript; nor in the decision to publish the results.

References

  1. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. CNN-Enhanced Graph Convolutional Network With Pixel and Superpixel-Level Feature Fusion for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8657–8671. [Google Scholar] [CrossRef]
  2. Shahraki, F.F.; Prasad, S. Graph Convolutional Neural Networks for Hyperspectral Data Classification. In Proceedings of the IEEE Global Conference on Signal and Information Processing, Anaheim, CA, USA, 26–28 November 2018. [Google Scholar]
  3. Hong, D.; Gao, L.; Wu, X.; Yao, J.; Zhang, B. Revisiting Graph Convolutional Networks with Mini-Batch Sampling for Hyperspectral Image Classification 2021. In Proceedings of the 2021 11th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 March 2021; pp. 1–5. [Google Scholar] [CrossRef]
  4. Bai, J.; Shi, W.; Xiao, Z.; Regan, A.C.; Ali, T.A.A.; Zhu, Y.; Zhang, R.; Jiao, L. Hyperspectral Image Classification Based on Superpixel Feature Subdivision and Adaptive Graph Structure. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5524415. [Google Scholar] [CrossRef]
  5. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. Multilevel Superpixel Structured Graph U-Nets for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5516115. [Google Scholar] [CrossRef]
  6. Xue, X.; Zhang, H.; Fang, B.; Bai, Z.; Li, Y. Grafting Transformer on Automatically Designed Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531116. [Google Scholar] [CrossRef]
  7. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  8. Shen, Y.; Zhu, S.; Chen, C.; Du, Q.; Xiao, L.; Chen, J.; Pan, D. Efficient Deep Learning of Nonlocal Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6029–6043. [Google Scholar] [CrossRef]
  9. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral-Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  10. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.J.; Pla, F. Deep Pyramidal Residual Networks for Spectral-Spatial Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 740–754. [Google Scholar] [CrossRef]
  11. Zhong, Z.; Li, J.; Ma, L.; Jiang, H.; Zhao, H. Deep residual networks for hyperspectral image classification. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 1824–1827. [Google Scholar] [CrossRef]
  12. Zhang, C.; Li, G.; Du, S.; Tan, W.; Gao, F. Three-dimensional densely connected convolutional network for hyperspectral remote sensing image classification. J. Appl. Remote Sens. 2019, 13, 1. [Google Scholar] [CrossRef]
  13. Zheng, J.; Feng, Y.; Bai, C.; Zhang, J. Hyperspectral Image Classification Using Mixed Convolutions and Covariance Pooling. IEEE Trans. Geosci. Remote Sens. 2021, 59, 522–534. [Google Scholar] [CrossRef]
  14. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D-2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  15. Wu, Y.; Zhou, T.; Hu, X.; Shi, L.; Yang, W. RepSSRN: The Structural Reparameterization Applied to SSRN for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  16. Wenju, W.; Shuguang, D.; Zhongmin, J.; Liujie, S. A Fast Dense Spectral-Spatial Convolution Network Framework for Hyperspectral Images Classification. Remote Sens. 2018, 10, 1068. [Google Scholar]
  17. Yang, Z.; Xi, Z.; Zhang, T.; Guo, W.; Zhang, Z.; Li, H.C. CMR-CNN: Cross-Mixing Residual Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8974–8989. [Google Scholar] [CrossRef]
  18. Meng, Z.; Jiao, L.; Liang, M.; Zhao, F. A Lightweight Spectral-Spatial Convolution Module for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  19. Rui, L.; Chenxi, D. LiteDenseNet: A Lightweight Network for Hyperspectral Image Classification. arXiv 2020, arXiv:eess.IV/2004.08112. [Google Scholar]
  20. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral-Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3232–3245. [Google Scholar] [CrossRef]
  21. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of Hyperspectral Image Based on Double-Branch Dual-Attention Mechanism Network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  22. He, X.; Chen, Y.; Lin, Z. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  23. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral-Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  24. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification With Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5518615. [Google Scholar] [CrossRef]
  25. Zhao, W.; Peng, R.; Wang, Q.; Cheng, C.; Emery, W.J.; Zhang, L. Life-Long Learning With Continual Spectral-Spatial Feature Distillation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5544214. [Google Scholar] [CrossRef]
  26. Cao, Y.; Wang, Y.; Yin, Z.; Yang, Z. Mixed Residual Convolutions with Vision Transformer in Hyperspectral Image Classification. In Proceedings of the 2022 IEEE 22nd International Conference on Communication Technology (ICCT), Nanjing, China, 11–14 November 2022; pp. 1595–1599. [Google Scholar] [CrossRef]
Figure 4. Indian Pines (IP) dataset. (a) False Color image. (b) Ground Truth map.
Figure 4. Indian Pines (IP) dataset. (a) False Color image. (b) Ground Truth map.
Remotesensing 15 04078 g004
Figure 5. Pavia University (PU) dataset. (a) False Color image. (b) Ground Truth map.
Figure 5. Pavia University (PU) dataset. (a) False Color image. (b) Ground Truth map.
Remotesensing 15 04078 g005
Figure 6. Salinas (SA) dataset. (a) False Color image. (b) Ground Truth map.
Figure 6. Salinas (SA) dataset. (a) False Color image. (b) Ground Truth map.
Remotesensing 15 04078 g006
Figure 7. Kennedy Space Center (KSC) dataset. (a) False Color image. (b) Ground Truth map.
Figure 7. Kennedy Space Center (KSC) dataset. (a) False Color image. (b) Ground Truth map.
Remotesensing 15 04078 g007
Figure 8. The first row is the prediction map of Indian Pines (IP) with a random shuffling ratio of 0, the second row is the experimental results of IP with a random shuffling proportion of 0.2, and the third row is the experimental results of IP with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Figure 8. The first row is the prediction map of Indian Pines (IP) with a random shuffling ratio of 0, the second row is the experimental results of IP with a random shuffling proportion of 0.2, and the third row is the experimental results of IP with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Remotesensing 15 04078 g008
Figure 9. The first row is the prediction map of Pavia University (PU) with a random shuffling ratio of 0, the second row is the experimental results of PU with a random shuffling proportion of 0.2, and the third row is the experimental results of PU with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Figure 9. The first row is the prediction map of Pavia University (PU) with a random shuffling ratio of 0, the second row is the experimental results of PU with a random shuffling proportion of 0.2, and the third row is the experimental results of PU with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Remotesensing 15 04078 g009
Figure 10. The first row is the prediction map of Salinas (SA) with a random shuffling ratio of 0, the second row is the experimental results of SA with a random shuffling proportion of 0.2, and the third row is the experimental results of SA with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Figure 10. The first row is the prediction map of Salinas (SA) with a random shuffling ratio of 0, the second row is the experimental results of SA with a random shuffling proportion of 0.2, and the third row is the experimental results of SA with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Remotesensing 15 04078 g010
Figure 11. The first row is the prediction map of KSC with a random shuffling ratio of 0, the second row is the experimental results of KSC with a random shuffling proportion of 0.2, and the third row is the experimental results of KSC with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Figure 11. The first row is the prediction map of KSC with a random shuffling ratio of 0, the second row is the experimental results of KSC with a random shuffling proportion of 0.2, and the third row is the experimental results of KSC with a random shuffling ratio of 0.5. (a) Ground Truth map; (b) DPRN; (c) SKDN(DPRN); (d) SSFTT; (e) SKDN(SSTFF); (f) MRViT; (g) SKDN(MRViT).
Remotesensing 15 04078 g011
Table 1. Numbers of samples and land cover classes in four public datasets.
Table 1. Numbers of samples and land cover classes in four public datasets.
Indian PinesPavia UniversitySalinasKennedy Space Center
IDClass NameSamplesIDClass NameSamplesIDClass NameSamplesIDClass NameSamples
1Alfalfa461Asphalt66311Brocoli_green_weeds_120091Scurb761
2Corn-notill14282Meadows18,6492Brocoli_green_weeds_223,7262Willow swamp243
3Corn-mintill8303Gravel20993Fallow19763CP hammock256
4Corn2374Trees30644Fallow_rough_plow13944Slash pine252
5Grass-pasture4835Painted metal sheets13455Fallow_smooth26785Oak/Broadleaf161
6Grass-trees7306Bare Soil50296Stubble39596Hardwood229
7Grass-pasture-mowed287Bitumen13307Celery35797Swamp105
8Hay-windrowed4788Self-Blocking Bricks36828Grapes_untrained11,2718Graminiod marsh431
9Oats209Shadows9479Soil_vinyard_develop62039Spartina marsh520
10Soybean-notill97210Background164,62410Corn_senesced_green_weeds327810Catiail marsh404
11Soybean-mintill2455 11Lettuce_romaine_4wk106811Salt marsh419
12Soybean-clean593 12Lettuce_romaine_5wk192712Mud flats503
13Wheat205 13Lettuce_romaine_6wk91613Water927
14Woods1265 14Lettuce_romaine_7wk107014Background56,975
15Buildings-Grass-Trees-Drives386 15Vinyard_untrained7268
16Stone-Steel-Towers93 16Vinyard_vertical_trellis1807
17Background10,776 17Background56,975
Total Samples21,025 Total Samples207,400 Total Samples111,104 Total Samples314,368
Table 2. Classification results’ OA from the improved methods and the original methods on four reconstructed datasets.
Table 2. Classification results’ OA from the improved methods and the original methods on four reconstructed datasets.
DatasetRandom RatioTraining RatioDPRNSKDN(DPRN)SSTFFSKDN(SSTFF)MRViTSKDN(MRViT)
0%5%95.1795.3296.3996.4698.0898.12
IP20%5%65.476.8568.9474.0573.7279.22
50%5%32.6145.4937.6140.4643.3450.43
0%10%95.0299.9199.1299.7799.0899.51
PU20%10%81.8887.2580.2380.7881.685.19
50%10%64.5171.8753.5954.7862.1166.79
0%0.50%95.5895.9296.3996.996.2796.73
SA20%0.50%55.7373.6370.8976.3874.578.51
50%0.50%24.9636.6632.5745.4840.451.12
0%20%97.0197.8298.0898.198.1798.21
KSC20%20%57.7383.4170.3776.1781.7582.38
50%20%26.6955.3643.0547.550.2650.91
Table 3. Classification results’ AA from the improved methods and the original methods on four reconstructed datasets.
Table 3. Classification results’ AA from the improved methods and the original methods on four reconstructed datasets.
DatasetRandom RatioTraining RatioDPRNSKDN(DPRN)SSTFFSKDN(SSTFF)MRViTSKDN(MRViT)
0%5%92.0093.0192.6893.5693.3995.26
IP20%5%54.4967.2661.9870.2164.5574.24
50%5%17.6331.8530.4131.9832.1144.69
0%10%90.395.6698.4498.7298.6599.03
PU20%10%73.2184.1676.2276.8477.0284.52
50%10%47.0961.2741.845.8347.7860.17
0%0.50%95.3295.4496.1296.6396.0996.72
SA20%0.50%54.2174.1271.6474.6572.9275.91
50%0.50%23.2433.8128.7442.0737.2748.36
0%20%96.597.2397.6598.0397.3797.85
KSC20%20%48.6181.7669.369.6579.4380.05
50%20%18.2851.738.3643.9345.4350.37
Table 4. Classification results’ kappa from the improved methods and the original methods on four reconstructed datasets.
Table 4. Classification results’ kappa from the improved methods and the original methods on four reconstructed datasets.
DatasetRandom RatioTraining RatioDPRNSKDN(DPRN)SSTFFSKDN(SSTFF)MRViTSKDN(MRViT)
0%5%95.8295.9195.8896.2297.3797.81
IP20%5%60.4473.6064.4470.4169.4274.81
50%5%21.7236.8728.6030.0134.8843.01
0%10%93.4795.6199.5399.6098.4298.91
PU20%10%75.3483.0573.7774.5175.5177.87
50%10%50.2261.9336.7638.5448.5358.02
0%0.50%95.3195.4296.2396.4396.3695.82
SA20%0.50%50.3470.7567.8073.7071.5175.45
50%0.50%17.4729.9724.3738.6832.5143.92
0%20%96.5797.0197.5397.9797.6298.27
KSC20%20%53.1181.4767.0267.8374.6678.27
50%20%20.3950.1835.9241.4644.4546.29
Table 5. The experimental results’ OA from the improved models and the original models on four real-world datasets.
Table 5. The experimental results’ OA from the improved models and the original models on four real-world datasets.
DatasetRandom RatioTraining RatioDPRNSKDN(DPRN)SSTFFSKDN(SSTFF)MRViTSKDN(MRViT)
0%5%95.1795.3296.3996.4698.0898.12
IP20%5%92.6195.0483.394.6591.7497.24
50%5%74.4582.2152.4994.0976.7696.48
0%10%95.0299.9199.1299.7799.0899.51
PU20%10%95.2799.4295.5199.6196.1698.98
50%10%92.2596.3785.6698.6192.3797.15
0%0.50%95.5895.9296.3996.996.2796.73
SA20%0.50%90.1292.8989.6696.0191.1296.24
50%0.50%84.0789.0160.3395.9377.1594.07
0%20%97.0197.8298.0898.198.1798.21
KSC20%20%95.1399.4434.7165.7991.4994.23
50%20%71.2877.1417.6657.2577.0185.61
Table 6. The experimental results’ AA from the improved models and the original models on four real-world datasets.
Table 6. The experimental results’ AA from the improved models and the original models on four real-world datasets.
DatasetRandom RatioTraining RatioDPRNSKDN(DPRN)SSTFFSKDN(SSTFF)MRViTSKDN(MRViT)
0%5%92.8193.0492.8693.7396.4897.66
IP20%5%88.6289.1584.1988.3992.6494.97
50%5%69.4476.8483.6584.7980.7385.89
0%10%98.7999.5399.0699.2498.6799.44
PU20%10%98.2699.3194.6299.0996.198.59
50%10%94.6295.0487.4497.8593.3496.18
0%0.50%92.1294.7995.6996.5595.5196.79
SA20%0.50%90.9392.8489.3396.3990.7597.21
50%0.50%83.7188.7960.1294.5172.4695.68
0%20%95.0196.5792.1893.5992.0492.96
KSC20%20%93.2296.1657.3459.1690.7690.91
50%20%73.7876.341.4545.2690.1891.45
Table 7. The experimental results’ kappa from the improved models and the original models on four real-world datasets.
Table 7. The experimental results’ kappa from the improved models and the original models on four real-world datasets.
DatasetRandom RatioTraining RatioDPRNSKDN(DPRN)SSTFFSKDN(SSTFF)MRViTSKDN(MRViT)
0%5%93.6894.1094.0994.7196.3397.21
IP20%5%93.1794.3593.1994.6594.4496.85
50%5%74.2979.9692.2393.2693.8995.97
0%10%96.7398.7797.0199.5397.7499.03
PU20%10%97.1698.2396.7899.4896.8398.65
50%10%94.7495.1589.9598.1590.4196.21
0%0.50%93.2795.1194.2296.2494.3396.08
SA20%0.50%92.592.0993.1895.5692.0995.81
50%0.50%85.9387.7770.5795.0190.9393.41
0%20%96.0797.3585.6689.7395.7697.73
KSC20%20%84.5197.2460.1461.5589.1893.57
50%20%70.3774.4150.4951.5783.6287.94
Table 8. The experimental results from different single-pixel-based models on Indian Pines.
Table 8. The experimental results from different single-pixel-based models on Indian Pines.
MethodsDPRNCEGCNSSTFFMRViTSKDN(MRViT)
Random ShuffleTraining RatioOAOAOAOAOA
0%5%68.8497.6162.5565.1974.51
20%5%64.4973.8163.6167.8672.86
50%5%67.8034.2464.9466.3674.83
100%5%68.7519.2464.9365.3773.17
20%20%87.7475.680.8783.6286.31
50%20%87.4347.4779.9683.5285.79
100%20%87.1419.4579.8484.6886.47
20%50%90.5574.5690.6893.1495.97
50%50%90.6748.490.3592.4594.75
100%50%90.1824.3187.0192.3194.39
Table 9. The experimental results from different single-pixel-based models on Pavia University.
Table 9. The experimental results from different single-pixel-based models on Pavia University.
MethodsDPRNCEGCNSSTFFMRViTSKDN(MRViT)
Random ShuffleTraining RatioOAOAOAOAOA
0%10%94.1296.7390.6684.0684.79
20%10%93.8283.591.8284.7688.77
50%10%92.0164.2992.2285.6288.14
100%10%93.250.8991.3985.9787.39
20%30%95.083.7895.7192.193.66
50%30%94.7265.1695.3792.3493.17
100%30%94.555.3395.0592.0694.09
20%50%95.8581.9896.5796.2896.75
50%50%94.3860.9596.996.3195.73
100%50%94.1656.4596.6496.1896.71
Table 10. The experimental results from different single-pixel-based models on Salinas.
Table 10. The experimental results from different single-pixel-based models on Salinas.
MethodsDPRNCEGCNSSTFFMRViTSKDN(MRViT)
Random ShuffleTraining RatioOAOAOAOAOA
0%0.50%84.7799.1385.8476.3177.87
20%0.50%85.9380.1386.5178.7682.79
50%0.50%84.8653.0584.5275.5883.52
100%0.50%84.0718.9884.5676.1682.71
50%5%88.9578.5892.7690.4792.77
20%5%89.1287.7292.7490.9993.18
100%5%88.4464.7192.8190.8692.27
20%20%93.7391.9495.7494.2995.57
50%20%95.1289.9695.0394.1495.01
100%20%94.7983.1295.5694.1594.35
20%50%95.1692.5597.1497.2397.59
50%50%94.7190.6597.5997.4597.74
100%50%94.8488.6597.6997.2297.48
Table 11. The experimental results from different single-pixel-based models on Kennedy Space Center.
Table 11. The experimental results from different single-pixel-based models on Kennedy Space Center.
MethodsDPRNCEGCNSSTFFMRViTSKDN(MRViT)
Random ShuffleTraining RatioOAOAOAOAOA
0%20%93.1395.8632.7188.7689.59
20%20%90.5277.8811.7779.2581.88
50%20%85.9652.9915.575.8880.41
100%20%87.7913.9316.6880.279.74
20%40%91.4378.8615.3687.490.82
50%40%90.2954.7415.282.2290.38
100%40%90.1418.248.7584.5790.43
20%60%93.6678.658.6493.9894.01
50%60%92.8856.8914.8993.5593.71
100%60%92.1720.9415.2392.8293.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Cao, Y.; Zhou, X.; Liu, J.; Zhang, T.; Ji, J. Random Shuffling Data for Hyperspectral Image Classification with Siamese and Knowledge Distillation Network. Remote Sens. 2023, 15, 4078. https://doi.org/10.3390/rs15164078

AMA Style

Yang Z, Cao Y, Zhou X, Liu J, Zhang T, Ji J. Random Shuffling Data for Hyperspectral Image Classification with Siamese and Knowledge Distillation Network. Remote Sensing. 2023; 15(16):4078. https://doi.org/10.3390/rs15164078

Chicago/Turabian Style

Yang, Zhen, Ying Cao, Xin Zhou, Junya Liu, Tao Zhang, and Jinsheng Ji. 2023. "Random Shuffling Data for Hyperspectral Image Classification with Siamese and Knowledge Distillation Network" Remote Sensing 15, no. 16: 4078. https://doi.org/10.3390/rs15164078

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop