You are currently viewing a new version of our website. To view the old version click .
Remote Sensing
  • Article
  • Open Access

25 March 2018

Semi-Supervised Classification of Hyperspectral Images Based on Extended Label Propagation and Rolling Guidance Filtering

,
,
,
and
1
The College of Computer Science and Engineering, Shandong University of Science and Technology, Qingdao 266590, Shandong Province, China
2
The School of Information and Control Engineering, Qingdao University of Technology, No.777 Jialingjiang Road, Huangdao District, Qingdao 266520, Shandong Province, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Hyperspectral Imaging and Applications

Abstract

Semi-supervised classification methods result in higher performance for hyperspectral images, because they can utilize the relationship between unlabeled samples and labeled samples to obtain pseudo-labeled samples. However, how generating an effective training sample set is a major challenge for semi-supervised methods, In this paper, we propose a novel semi-supervised classification method based on extended label propagation (ELP) and a rolling guidance filter (RGF) called ELP-RGF, in which ELP is a new two-step process to make full use of unlabeled samples. The first step is to implement the graph-based label propagation algorithm to propagate the label information from labeled samples to the neighboring unlabeled samples. This is then followed by the second step, which uses superpixel propagation to assign the same labels to all pixels within the superpixels that are generated by the image segmentation method, so that some labels wrongly labeled by the above step can be modified. As a result, so obtained pseudo-labeled samples could be used to improve the performance of the classifier. Subsequently, an effective feature extraction method, i.e., RGF is further used to remove the noise and the small texture structures to optimize the features of the initial hyperspectral image. Finally, these produced initial labeled samples and high-confidence pseudo-labeled samples are used as a training set for support vector machine (SVM). The experimental results show that the proposed method can produce better classification performance for three widely-used real hyperspectral datasets, particularly when the number of training samples is relatively small.

1. Introduction

Hyperspectral images have been widely used for many applications, such as classification [1], spectral unmixing [2], target detection [3], environmental monitoring [4] and anomaly detection [5]. Among these applications, classification is one of the most crucial branches. There are more than 100 spectral bands that provide detailed information to discriminate the object in a hyperspectral image [6]. However, the high dimensions of hyperspectral images require a more complicated model, while such a complicated model also requires more training samples to support it. Thus, the imbalance between the number of training samples and the high dimensions may cause the well-known “Hughes” phenomenon [7]. The existence of the “Hughes” phenomenon poses restrictions on performance improvement for the hyperspectral image classification.
In hyperspectral image classification, traditional spectral-based classification methods are widely used, such as support vector machines (SVM) [8], the back-propagation neural network (BP) [9], random forest (RF) [10] and the 1D deep convolutional neural network (1D CNN) [11]. However, all of these methods are sensitive to the quality and number of training samples; thus, the classification performance is limited when a small amount of training samples is provided. In order to further improve the classification performance, the rich spatial-contextual information is used in pixel-wise classification methods [12,13,14,15]. For instance, Pan et al. [16] introduced the hierarchical guidance filtering to extract the different spatial contextual information at different filter scales in hyperspectral images. In [17], a new network that utilizes the spectral and spatial information simultaneously was proposed to achieve more accurate classification results.
During the past few decades, the semi-supervised learning methods have shown excellent performance in hyperspectral image classification [18,19]. One goal of the semi-supervised learning method is to select the most useful unlabeled samples and to determine the label information of these new selected samples. Generally, semi-supervised learning can be classified into the generative model [20], the co-training model [21], the graph-based method [18,19,20,21,22], etc. All of those methods are based on an assumption that similar samples have the same labels. Hence, graph-based semi-supervised methods have attracted increasing attention in hyperspectral image classification [18,22]. For example, in [23], Wang et al. proposed a novel graph-based semi-supervised learning approach based on a linear neighborhood model to propagate the labels from the labeled samples to the whole dataset using these linear neighborhoods with sufficient smoothness. In [18], the wealth of unlabeled samples is exploited through a graph-based methodology to handle the special characteristics of hyperspectral images. The label propagation algorithm (LP) [24,25] is a widely-used method in graph-based semi-supervised learning [26,27], as in [28]; unlabeled data information is effectively exploited by combing the Gaussian random field model and harmonic function. Wang et al. [24] proposed an approach based on spatial-spectral label propagation for the semi-supervised classification method, in which labels were propagated from labeled samples to unlabeled samples with the spatial-spectral graph to update the training set. However, there are three main difficulties of the aforementioned graph-based semi-supervised classification methods: (1) how to significantly generate the pseudo-labeled samples with a high quality; (2) how to expand the propagation scope of the samples as much as possible; (3) how to modify the labels that wrongly propagate to other classes.
Recently, the superpixel technique [29] has been an effective way to introduce the spatial information for hyperspectral image classification [16,30,31]. Each superpixel is a homogeneous region, whose size and shape are adaptive. The commonly-used superpixel segmentation methods include the SLIC method [32], normalized cut method [33], regional growth method [34], etc. Moreover, superpixel-based classification methods [35,36] have shown a good robustness in the result of hyperspectral image classification. Motivated by the idea of a superpixel, we design a novel superpixel-based label propagation framework, extended label propagation (ELP), which uses a two-step propagation process to significantly extend the number of pseudo-labeled samples. In ELP, the spatial-spectral weighted graph is first constructed with the labeled samples and unlabeled samples from the spatial neighbors of the labeled samples to propagate the class labels to unlabeled samples. Second, the multi-scale segmentation algorithm [37] is used to generate superpixels, and then, superpixel propagation is introduced to assign the same label to all pixels within a superpixel. Finally, a threshold is defined; when the confidence of pseudo-labeled samples is higher than the defined threshold, they will be selected to enrich the training sample set. Note that the second step of the ELP method, i.e., extended label propagation with superpixel segmentation, is the innovation of the proposed method, because it can generate a large number of high-confidence pseudo-labeled samples.
In this paper, the motivations include three aspects. First, we would like to extend the number of high-confidence pseudo-labeled samples based on a two-step propagation process. Second, rolling guidance filtering is used to optimize the feature of the initial hyperspectral image. In the optimized image, the noise and small texture are removed, while the strong structure of the image is preserved, enhancing the discrimination within and between classes. Third, we want to modify the labels that wrongly propagate by the label propagation algorithm. The proposed ELP-RGF can effectively improve the classification performance with less training samples. The contributions of the proposed method consist of:
(1) We propose a novel extended label propagation component that is based on the label propagation algorithm. The second step of ELP, that is superpixel propagation, is the most innovative of the proposed method, because it not only expands the scope of the label propagation, but also generates a large number of high-confidence pseudo-labeled samples. Therefore, it has a good performance for hyperspectral image classification.
(2) In the step of superpixel propagation, the labels of pixels within the superpixel are obtained by a majority vote with the labeled samples belonging to that superpixel. Therefore, some pseudo-labeled samples with wrong labels that are obtained by the first step of the ELP method can be modified. Furthermore, we can show that the variation of ELP-RGF is much more stable compared to the result in [38] and [24].
(3) Optimized image features with the rolling guidance filter (RGF) [39] can eliminate the noise of the initial image. The filtered image is treated as an input to the SVM method to help improve the result of the final classification.
The remainder of this paper is organized as follows. The related work is described in Section 2. The proposed method is introduced in Section 3. The discussion is provided in Section 5. Finally, conclusions are given in Section 6.

3. Proposed Method

Figure 2 shows the schematic diagram of the proposed semi-supervised classification method based on extended label propagation and rolling guidance filtering for the hyperspectral image, which consists of the following steps: First, the extended label propagation method is used to obtain an effective set of pseudo-labeled samples. This step is a two-step process. The first step is that the neighboring unlabeled samples from initial labeled samples are assigned labels by using the graph-based spatial-spectral label propagation method. The second step is that all pixels within the superpixel to which the labeled samples belong are assigned the same labels to achieve further label propagation. Then, pseudo-labeled samples with confidence less than the constant threshold will not be added into the training sample set. Then, rolling guidance filtering is used to optimize the feature of the original image, and the filtered result is used to extend the feature vector that is an input to the SVM. Finally, the initial labeled samples and pseudo-labeled samples are merged in training by SVM.
Figure 2. Schematic of the proposed semi-supervised classification method of hyperspectral images based on extended label propagation and rolling guidance filtering.
The proposed semi-supervised classification method based on extended label propagation and rolling guidance filtering (ELP-RGF) method can be shown by Algorithm 1:
Algorithm 1: proposed ELP-RGF method
Input: the dataset X, the initial labeled training sample set V L , the weight μ , the width of spatial neighborhood system d, the segmentation scale S, the unlabeled samples V U
1. Superpixels segmentation:
      Obtain Y = ( Y 1 , , Y M ) , where Y i is the i-th superpixel, based on the multi-scale segmentation algorithm for X.
2. Extended label propagation method:
      Obtain the pseudo-labeled training sample set V ˜ n e w .
      (1) Label propagation:
               Selection of the unlabeled training set V U from the neighbors of the labeled samples.
               Construction of the weighted graph G and weighted matrix W i j by Equations (4)–(6).
               Calculation of the probability matrix P by Equations (7)–(10).
               Prediction of the labels of V U by Equation (11) and generation of the pseudo-labeled sample set V n e w .
      (2) Superpixel propagation:
               Observation of the labels of labeled samples belonging to superpixel Y i , and then, the majority vote method is used to assign the labels for all pixels within Y i .
3. Rolling guidance filtering:
      Extraction of the spectral features of initial image X, and the filtered image I ˜ is obtained by Equations (10)–(12).
4. SVM classification:
       V L and V ˜ n e w are merged as the final training sample set, and then, train SVM to obtain the prediction of labels of the testing set. The input feature vector to the SVM is the filtered image by the rolling guidance filtering.
Note that we perform the SVM to obtain the final classification result, because it has a good performance for the non-linear problem [41]. The goal of SVM is to find an optimal decision hyperplane that can maximize the distance between the two nearest samples on the two sides of the plane for classification. In this paper, the “one against rest” strategy [42] is adopted to achieve the multi-classification.

4. Experiment

In this section, the experimental results are performed on three real hyperspectral datasets to evaluate the performance of the proposed ELP-RGF method.

4.1. Datasets Description

In our experiments, three hyperspectral image datasets including the Indian Pines image, the University of Pavia image and the Kennedy Space Center image are utilized to evaluate the performance of the ELP-RGF.
(1) Indian Pines dataset: This image was acquired by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor, which captured an Indian Pines unlabeled agricultural site of northwestern Indiana and contains 220 × 145 × 145 bands. Twenty water absorption bands (Nos. 104–108, 150–163 and 220) were removed before hyperspectral image classification. The spatial resolution of the Indian Pines image is 20 m per pixel, and the spectral coverage ranges from 0.4–2.5 μ m. Figure 3 shows a color composite and the corresponding ground-truth data of the Indian Pines image.
Figure 3. Indian Pines dataset. (a) False-color composite; (b,c) Ground-truth data.
(2) University of Pavia dataset: This image capturing the University of Pavia, Italy, was recorded by the Reflective Optics System Imaging Spectrometer (ROSIS). This image contains 115 bands and a size 610 × 340 with a spatial resolution of 1.3 m per pixel and a spectral coverage ranging from 0.43–0.86 μ m. Using a standard preprocessing approach before hyperspectral image classification, 12 noisy channels were removed. Nine classes of interest are considered for this image. Figure 4 shows the color composite and the corresponding ground-truth data of the University of Pavia image.
Figure 4. University of Pavia image. (a) False-color composite; (b,c) Ground-truth data.
(3) Kennedy Space Center dataset: The Kennedy Space Center (KSC) image was captured by the National Aeronautics and Space Administration (NASA) Airborne Visible/Infrared Imaging Spectrometer instrument at a spatial resolution of 18 m per pixel. The KSC image contains 224 bands with a spatial size of 512 × 614, and the water absorption and low signal-to-noise ratio (SNR) bands were discarded before the classification. Figure 5 shows the KSC image and the corresponding ground-truth data.
Figure 5. (a,b) Ground truth data of the Kennedy Space Center images.

4.2. Parameter Analysis of the Proposed Method

In the experiments, the original images were segmented by the multi-scale segmentation method (MSS) [37]. In this section, we fix the shape parameters v = 0.1 and the smoothness parameter to u 1 = 0.5 in MSS. For the proposed method, there are three hyperparameters that have to be adjusted, namely weight parameter u, segmentation scale S and the width of spatial neighborhood d. The three hyperparameters were selected using the cross-validation strategy. Figure 6 shows the classification results obtained by the ELP-RGF method with different weight parameters μ and segmentation scales S. From Figure 6, we can see that the result of Figure 6a–b is visually more satisfactory than that of Figure 6c–d. If the process of label propagation entirely relied on the spatial graph, that is, u= 1 is applied, the result of ELP-RGF is poor, as Figure 6d shows. Therefore, μ = 0.001 to μ = 0.01 is considered the most optimal weight parameter range. We can see that the classification result of Figure 6f–g is better than Figure 6e,h, especially for the landscape of “Soybeans-min till“ and “Hay-windrowed“. In addition, Figure 7 shows the OA curves of ELP-RGF in the different u and dto illustrate the spatial weight parameter playing an important role in the process of label propagation. Furthermore, Figure 8 shows that the classification accuracies and computing time of the proposed method are significantly affected by S. When the two factors of the classification accuracy and the computing time are taken into full consideration (see Figure 8) and observing the selected parameters obtained by cross-validation, we can know that the optimal parameter range is 4–6.
Figure 6. The analysis for the hyperparameters μ and S for the Indian Pines image. In the first row, S is fixed as five. (ad) respectively showthe classification results obtained by the extended label propagation (ELP)-RGF method with (a) u = 0.001, (b) u = 0.01, (c) u = 0.1 and (d) u = 1. In the second row, μ is fixed as 0.01; (eh) respectively showthe classificationmaps obtained by ELP-RGFmethodwith (e) S = 1, (f) S = 3, (g) S = 5 and (h) S = 9.
Figure 7. Influence of μ and d on the Kappa coefficient of ELP-RGF for the Indian Pines dataset.
Figure 8. The effect of the different segmentation size on OA accuracy and computing times for three datasets.

4.3. Comparison with Other Classification Methods

In this section, the proposed ELP-RGF method is compared with several hyperspectral image classification methods, i.e., the typical SVM method, more advanced extended random walkers (ERW) [43] and semi-supervised methods (the Laplacian support vector machine (LapSVM) [38] and spectral-spatial label propagation (SSLP-SVM) [24]). In addition, a post-processing-based edge-preserving filtering (EPF) [44] and the rolling guidance filtering method (RGF) [39] are also used as the comparison methods. The parameter settings for the EPF, ERW and SSLP-SVM methods are given in the corresponding papers. The evaluation indexes in Table 1, Table 2, Table 3 and Table 4 are given in the form of the mean ± standard deviation.
Table 1. Comparison of classification accuracies (in percentage) provided by different methods using different training samples per class (Indian Pines image). EPF, edge-preserving filtering; ERW, extended random walkers; LapSVM, Laplacian SVM; SSLP, spectral-spatial label propagation.
Table 2. Overall accuracy of the various methods for the University of Pavia image (average of 10 runs with thestandard deviation; the bold values indicate the greatest accuracy among the methods in each case).
Table 3. Individual class accuracies, OA, AA and Kappa coefficient (in percentage) for the Kennedy Space Center images.
Table 4. Overall accuracy of the various methods for the Kennedy Space Center image (average of 10 runs with thestandard deviation; the bold values indicate the greatest accuracy among the method in each case).
For the Indian Pines dataset, Table 1 shows the OA, AA and Kappa coefficient of different methods with the 5/10/15 training numbers per class (represented as s). From Table 1 we can see that the OA accuracy and Kappa coefficient of the proposed ELP-RGF method are better than other methods when the number of training samples is relatively small. In particular, when the s = 5, the OA accuracy of the proposed ELP-RGF method increases 14.29% and 36.63% compared to that of the SSLP-SVM method and the LapSVM method. The Kappa coefficient of ELP-RGF method is 13.52% higher than SSLP-SVM when s = 10, which fully shows the superiority of the two-step method proposed in this paper. We can see that the performance of the proposed ELP-RGF is always superior to that of the ERW method. As the number of training samples increases, the accuracy of the increase rate has decreased, however, there is still a large gap compared with other methods. Figure 9 shows the classification maps obtained by different methods. It can be seen that the classification map of the proposed method has less noise, and the boundary region in the classification map is also much clearer.
Figure 9. Classification maps of different methods for the Indian Pines image. (a) Reference image; (b) SVM method; (c) EPF method; (d) RGF method; (e) ERW method; (f) LapSVM method; (g) SSLP-SVM method and (h) ELP-RGF method.
For the University of Pavia image, we randomly selected 5, 10 and 15 samples from each class as the training samples. Table 2 shows the OA, AA and Kappa coefficient of the different methods with different s. According to Table 2, the proposed ELP-RGF, SSLP-SVM, LapSVM and ERW can produce greater classification accuracy than the SVM at the same s. However, the degrees of the improvement of ERW, SSLP-SVM and LapSVM are smaller compared with ELP-RGF. For example, the OA of ERW and SSLP-SVM increased by 1.16% and 10.24%. The experimental result indicated that the proposed ELP-RGF outperforms the compared methods. The results show that the OA of the proposed method is 96.02%, which is 10.25% higher than that of the SSLP-SVM and 0.91% higher than that of the ERW when s = 15. The OA accuracy and kappa coefficients of the ELP-RGF method are always the highest, which demonstrates that the ELP-RGF is the most accurate classifier among these methods. We can see that compared with SVM, SSLP-SVM and ERW, the OA accuracy and Kappa coefficients of the ELP method are more competitive. Figure 10 shows the classification maps of different methods when s = 15. The figure shows the effectiveness of the proposed method. The proposed method presents more accurate classification results for the class of MetalSheetsand Gravel, and its classification result is better than those of other methods.
Figure 10. Classification maps of different methods for the University of Pavia image (a) Reference image; (b) SVM method; (c) EPF method; (d) RGF method; (e) ERWmethod; (f) LapSVM method; (g) SSLP-SVM method and (h) ELP-RGF method.
For the Kennedy Space Center dataset, we evaluated the classification accuracies of different methods using 39 training samples collected from each class. Table 3 shows the OAs, AAs, Kappa and individual classification accuracies obtained for the various methods. From Table 3, it is demonstrated that the OA, AA and Kappa accuracy of the proposed method are the highest in all comparative methods. Most of individual accuracies are significantly higher than other methods. For the class of Willowswamp, the accuracies of the proposed method and SSLP-SVM are 99.75% and 77.38%; thus, the accuracy gain is 22.37%. For the class of Oak/Broadleaf, the proposed method can produce 59.21% and 57.11% OA improvements compared with SSLP-SVM and SVM. Table 4 provides the OA accuracies of the various methods. Observing the values in Table 4, we can see that the classification accuracy is proportional to s. Moreover, the performance of the proposed method is not only higher than the other semi-supervised classification methods, but also can improve more than that of ERW. Figure 11 shows that the proposed ELP-RGF method can achieve better classification performance and produce little noise compared with other methods.
Figure 11. Classification maps of different methods for the Kennedy Space Center image. (a) Reference image; (b) SVM method; (c) EPF method; (d) RGF method; (e) ERW method; (f) LapSVM method; (g) SSLP-SVM method; (h) ELP-RGF method.
Table 5 lists the number of samples generated by the two semi-supervised methods (i.e., SSLP-SVM and ELP-RGF) under three different datasets and the correct rate of these new labeled samples. We can see that the total number of labeled samples generated by the ELP-RGF method is almost 7–22-times more than that generated by the SSLP-SVM method for three datasets. Although the correct rate of the SSLP-SVM method is slightly higher, the ELP-RGF method is also competitive. More importantly, the proposed ELP-RGF can produce more labeled samples.
Table 5. Correct rate of samples generated by the two semi-supervised methods of three datasets.
Table 6 illustrates the effect of the superpixel by comparing with the RGF method, the combination of label propagation and RGF and the combination of superpixel propagation and RGF. The table shows that the superpixel propagation plays a major role in the proposed method. For example, for the Indian Pines image, the OA accuracy of the proposed ELP-RGF is 79.13%, while the accuracy obtained by SP-RGF and RGF is 75.62% and 56.14%, respectively. For the Kennedy Space Center image, the OA accuracy of the SP-RGF method is 4.64% higher than that of LP-RGF. As Table 6 shows, the accuracy of the SL-RGF is more than that of the LP-RGF method when RGF is used in those methods. While LP-RGF is higher than SP-RGF, the gap is small. Thus, the process of superpixel propagation is very useful to help improve the classification result.
Table 6. Overall accuracy of the various combined methods involved in the proposed method for three datasets.

5. Discussion

In this paper, the proposed ELP-RGF method is used to increase the number of training samples and optimize the features of the initial hyperspectral image. Previous label propagation-based works, such as SSLP-SVM, only increased a small number of training samples, which are neighboring the labeled samples, and the computational expense is large. If the scope of propagation is beyond neighbors, the computing time will rapidly increase. Furthermore, in the process of label propagation, some wrongly-labeled samples may be introduced to train the model, resulting in misclassification. In our ELP-RGF method, a two-step label prorogation process called ELP is proposed, which first utilized the spatial-spectral label propagation to propagate the label information from labeled samples to the neighboring unlabeled samples. Then, superpixel propagation is used to expand the scope of propagation to the entire superpixel to increase the huge number of training samples, and it is less time consuming compared to the propagation beyond neighbors. Compared with other semi-supervised classification methods, ELP has two obvious advantages: on the one hand, it can generate a large number of pseudo-labeled samples for model training; on the other hand, it can ensure the ‘effectiveness’ of the increased pseudo-labeled samples; here, ‘effectiveness’ means that almost all of the labels of the pseudo-labeled samples are correct, which was shown in Table 6. Moreover, as shown in Figure 12, the wrongly-labeled samples in the first step of the ELP method can be modified by the superpixel propagation. Thus, the proposed ELP-RGF method shows a better classification performance than other comparative methods. However, the greatest limitation of the proposed method is that the classification result is over-reliant on the segmentation scale. As shown in Figure 8, the difference in classification results with different segmentation scales is larger.
Figure 12. The process of modifying the wrongly-labeled samples. (a) The five wrongly-labeled pseudo-labeled samples are provided; (b) shows that the first and second labels of pseudo-labeled have been modified by superpixel propagation; (c) shows the real labels of the provided wrongly-labeled pseudo-labeled samples.

6. Conclusions

In this paper, a novel semi-supervised classification method of hyperspectral images based on extended label propagation and rolling guidance filtering is proposed. The first advantage of this method is that the number of pseudo-labeled training samples is significantly increased. The second advantage is that the diversity of training samples is improved to enhance the generalization of the proposed method. The third advantage is that the spatial information is fully considered using graphs and superpixels. The experimental results on three different hyperspectral datasets demonstrate that the proposed ELP-RGF method offers an excellent performance in terms of both visual quality and quantitative evaluation indexes. In particular, when the number of training samples is relatively small, the improvement is more obvious.

Acknowledgments

This work was co-supported by the National Natural Science Foundation of China (NSFC) (41406200, 61701272) and Shandong Province Natural Science Foundation of China (ZR2014DQ030, ZR2017PF004).

Author Contributions

Binge Cui conceived of the idea of this paper. Xiaoyun Xie designed the experiments and drafted the paper. Binge Cui, Siyuan Hao, Jiandi Cui and Yan Lu revised the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhong, Z.; Fan, B.; Duan, J.; Wang, L.; Ding, K.; Xiang, S. Discriminant Tensor Spectral–Spatial Feature Extraction for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2017, 12, 1028–1032. [Google Scholar] [CrossRef]
  2. Xu, X.; Shi, Z. Multi-objective based spectral unmixing for hyperspectral images. ISPRS J. Photogramm. Remote Sens. 2017, 124, 54–69. [Google Scholar]
  3. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. Sparse Transfer Manifold Embedding for Hyperspectral Target Detection. IEEE Geosci. Remote Sens. Lett. 2013, 52, 1030–1043. [Google Scholar]
  4. Pan, B.; Shi, Z.; An, Z.; Jiang, Z.; Ma, Y. Sparse Transfer Manifold Embedding for Hyperspectral Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 99, 1–13. [Google Scholar]
  5. Kang, X.; Zhang, X.; Li, S.; Li, K.; Li, J.; Benediktsson, J.A. Hyperspectral Anomaly Detection with Attribute and Edge-Preserving Filters. IEEE Trans. Geosci. Remote Sens. 2017, 99, 1–12. [Google Scholar] [CrossRef]
  6. Chang, C.I. Hyperspectral Imaging: Techniques for Spectral Detection and Classification; Plenum Publishing Co.: New York, NY, USA, 2003. [Google Scholar]
  7. Hughes, G. On the mean accuracy of statistical pattern recognizers. Inf. Theory IEEE Trans. 1968, 14, 55–63. [Google Scholar] [CrossRef]
  8. Gotsis, P.K.; Chamis, C.C.; Minnetyan, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar]
  9. Benediktsson, J.A.; Swain, P.H.; Ersoy, O.K. Neural Network Approaches versus Statistical Methods in Classification of Multisource Remote Sensing Data. Geosci. Remote Sens. Symp. 1989, 28, 540–552. [Google Scholar] [CrossRef]
  10. Xia, J.; Du, P.; He, X.; Chanussot, J. Hyperspectral Remote Sensing Image Classification Based on Rotation Forest. IEEE Geosci. Remote Sens. Lett. 2013, 11, 239–243. [Google Scholar] [CrossRef]
  11. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  12. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE. 2013, 101, 652–675. [Google Scholar] [CrossRef]
  13. Pan, B.; Shi, Z.; Xu, X. R-VCANet: A New Deep-Learning-Based Hyperspectral Image Classification Method. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 99, 1–12. [Google Scholar] [CrossRef]
  14. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J.A. PCA-Based Edge-Preserving Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 99, 1–12. [Google Scholar] [CrossRef]
  15. Pan, B.; Shi, Z.; Xu, X. MugNet: Deep learning for hyperspectral image classification using limited samples. ISPRS J. Photogramm. Remote Sens. 2017. [Google Scholar] [CrossRef]
  16. Pan, B.; Shi, Z.; Xu, X. Hierarchical Guidance Filtering-Based Ensemble Classification for Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2017, 99, 1–13. [Google Scholar] [CrossRef]
  17. Pan, B.; Shi, Z.; Zhang, N.; Xie, S. Hyperspectral Image Classification Based on Nonlinear Spectral–Spatial Network. IEEE Geosci. Remote Sens. Lett. 2016, 99, 1–5. [Google Scholar] [CrossRef]
  18. Camps-Valls, G.; Marsheva, T.V.B.; Zhou, D. Semi-Supervised Graph-Based Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3044–3054. [Google Scholar] [CrossRef]
  19. Shao, Y.; Sang, N.; Gao, C. Probabilistic Class Structure Regularized Sparse Representation Graph for Semi-Supervised Hyperspectral Image Classification. Pat. Recognit. 2017, 63, 102–114. [Google Scholar] [CrossRef]
  20. Krishnapuram, B.; Carin, L.; Figueiredo, M.A.T.; Member, S. Sparse multinomial logistic regression: Fast algorithms and generalization bounds. IEEE Trans. Pat. Anal. Mach. Intell. 2005, 27, 957–968. [Google Scholar] [CrossRef] [PubMed]
  21. Ando, R.K.; Zhang, T. Two-view feature generation model for semi-supervised learning. In Proceedings of the International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007. [Google Scholar]
  22. Meng, J.; Jung, C. Semi-Supervised Bi-Dictionary Learning for Image Classification With Smooth Representation-Based Label Propagation. IEEE Transa. Multimedia. 2016, 18, 458–473. [Google Scholar]
  23. Wang, F.; Zhang, C. Label Propagation through Linear Neighborhoods. IEEE Trans. Knowl. Data Eng. 2008, 20, 55–67. [Google Scholar] [CrossRef]
  24. Wang, L.; Hao, S.; Wang, Q.; Wang, Y. Semi-supervised classification for hyperspectral imagery based on spatial-spectral Label Propagation. Isprs J. Photogramm. Remote Sens. 2014, 97, 123–137. [Google Scholar] [CrossRef]
  25. Zhu, X.; Ghahramani, Z.; Mit, T.J. Semi-Supervised Learning with Graphs. In Proceedings of the International Joint Conference on Natural Language Processing, Carnegie Mellon University, Pittsburgh, PA, USA, January 2005; pp. 2465–2472. [Google Scholar]
  26. Cheng, H.; Liu, Z. Sparsity induced similarity measure for label propagation. In Proceedings of the IEEE International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009. [Google Scholar]
  27. Karasuyama, M.; Mamitsuka, H. Multiple Graph Label Propagation by Sparse Integration. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 1999–2012. [Google Scholar] [CrossRef] [PubMed]
  28. Zhu, X.; Ghahramani, Z. Semi-Supervised Learning Using Gaussian Fields and Harmonic Functions. In Proceedings of the Twentieth International Conference on International Conference on Machine Learning, Washington, DC, USA, 21–24 August 2003. [Google Scholar]
  29. Moore, A.P.; Prince, S.J.D.; Warrell, J.; Mohammed, U. Superpixel lattices. In Proceedings of the Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008. [Google Scholar]
  30. Wang, L.; Zhang, J.; Liu, P.; Choo, K.K.R.; Huang, F. Spectral-spatial multi-feature-based deep learning for hyperspectral remote sensing image classification. Soft Comput. A Fusion Found. Methodol. Appl. 2017, 21, 213–221. [Google Scholar] [CrossRef]
  31. Zhang, S.; Li, S.; Fu, W.; Fang, L. Multiscale Superpixel-Based Sparse Representation for Hyperspectral Image Classification. Remote Sens. 2017, 9, 139. [Google Scholar] [CrossRef]
  32. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC Superpixels; EPFL: Lausanne, Switzerland, 2010. [Google Scholar]
  33. De Carvalho, M.A.G.; da Costa, A.L.; Ferreira, A.C.B.; Junior, R.M.C. Image Segmentation Using Watershed and Normalized Cut. In Proceedings of the SIBIGRAPI, Gramado, Brazil, 30 August–3 September 2009. [Google Scholar]
  34. Ugarriza, L.G.; Saber, E.; Vantaram, S.R.; Amuso, V.; Shaw, M.; Bhaskar, R. Automatic image segmentation by dynamic region growth and multiresolution merging. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2009, 18, 2275–2288. [Google Scholar] [CrossRef] [PubMed]
  35. Jin, X.; Gu, Y. Superpixel-Based Intrinsic Image Decomposition of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2017, 99, 1–11. [Google Scholar]
  36. Jia, S.; Deng, B.; Zhu, J.; Jia, X.; Li, Q. Superpixel-Based Multitask Learning Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 99, 1–14. [Google Scholar] [CrossRef]
  37. Zhang, D.; Yang, Y.; Song, K. Research on a Multi-Scale Segmentation Algorithm Based on High Resolution Satellite Remote Sensing Image. International Conference on Intelligent Control and Computer Application. 2016.
  38. Belkin, M.; Niyogi, P.; Sindhwani, V. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Mach. Learn. Res. 2006, 7, 2399–2434. [Google Scholar]
  39. Zhang, Q.; Shen, X.; Xu, L. Rolling Guidance Filter. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  40. Rohban, M.H.; Rabiee, H.R. Supervised neighborhood graph construction for semi-supervised classification. Pat. Recognit. 2012, 4, 1363–1372. [Google Scholar] [CrossRef]
  41. Vapnik, V.N. The Nature of Statistical Learning Theory. IEEE Trans. Neural Netw. 1997, 8, 1564. [Google Scholar]
  42. Wang, L.; Zhang, Y.; Gu, Y. The research of simplification of structure of multiclass classifier of support vector machine. Image Graph. 2005, 5, 571–572. [Google Scholar]
  43. Kang, X.; Li, S.; Fang, L.; Li, M.; Benediktsson, J.A. Extended Random Walker-Based Classification of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2014, 53, 144–153. [Google Scholar] [CrossRef]
  44. Kang, X.; Li, S.; Benediktsson, J.A. Spectral–Spatial Hyperspectral Image Classification With Edge-Preserving Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.