Next Article in Journal
Flood Hazard Assessment Supported by Reduced Cost Aerial Precision Photogrammetry
Next Article in Special Issue
Hyperspectral Image Restoration under Complex Multi-Band Noises
Previous Article in Journal
Unsupervised Feature Selection Based on Ultrametricity and Sparse Training Data: A Case Study for the Classification of High-Dimensional Hyperspectral Data
Previous Article in Special Issue
Urban Change Detection Based on Dempster–Shafer Theory for Multitemporal Very High-Resolution Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Image Classification Based on Two-Stage Subspace Projection

1
School of Computer Science, China University of Geosciences, Wuhan 430074, China
2
School of Computer, Wuhan University, Wuhan 430072, China
3
Department of Computing, The Hong Kong Polytechnic University, Kowloon, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2018, 10(10), 1565; https://doi.org/10.3390/rs10101565
Submission received: 31 August 2018 / Revised: 27 September 2018 / Accepted: 28 September 2018 / Published: 30 September 2018

Abstract

:
Hyperspectral image (HSI) classification is a widely used application to provide important information of land covers. Each pixel of an HSI has hundreds of spectral bands, which are often considered as features. However, some features are highly correlated and nonlinear. To address these problems, we propose a new discrimination analysis framework for HSI classification based on the Two-stage Subspace Projection (TwoSP) in this paper. First, the proposed framework projects the original feature data into a higher-dimensional feature subspace by exploiting the kernel principal component analysis (KPCA). Then, a novel discrimination-information based locality preserving projection (DLPP) method is applied to the preceding KPCA feature data. Finally, an optimal low-dimensional feature space is constructed for the subsequent HSI classification. The main contributions of the proposed TwoSP method are twofold: (1) the discrimination information is utilized to minimize the within-class distance in a small neighborhood, and (2) the subspace found by TwoSP separates the samples more than they would be if DLPP was directly applied to the original HSI data. Experimental results on two real-world HSI datasets demonstrate the effectiveness of the proposed TwoSP method in terms of classification accuracy.

1. Introduction

Due to rapid development, hyperspectral images (HSIs) play a very significant role in various hyperspectral remote sensing applications, e.g., military [1], astronomy [2], and classification [3,4,5]. Among these mentioned tasks, HSI classification is a fundamental yet important application to provide primary information for the subsequent tasks, which is the main focus of this paper.
The goal of HSI classification is to distinguish the land-cover types of each pixel, which often has hundreds of spectral bands [6]. Although the high-dimensional features may provide the advantages for more accurate classification, the Hughes phenomenon [7] still exists in the classification process. An effective way is to perform dimensionality reduction of these features before HSI classification.
The existing dimensionality reduction methods can be divided into two categories: feature selection [8,9] and feature extraction [10,11]. The design of feature selection is to select some valuable features from the original HSI data. By contrast, the focus of feature extraction is to project the original high-dimensional feature data into an optimal low-dimensional subspace, which is able to construct valuable features in the projective transformation. Consequently, many feature extraction methods have been presented [12,13,14,15,16,17,18]. Several popular feature extraction methods are principal component analysis (PCA) [19], independent component analysis (ICA) [20], linear discriminant analysis (LDA) [21], and locality preserving projection (LPP) [22]. In general, PCA is a popular global dimensionality reduction method, while LPP is an effective local dimensionality reduction method. Both global and local structures are important for projecting the original high-dimensional data into a low-dimensional subspace while preserving the valuable information. LDA constructs a linear transformation by minimizing the within-class scatter and maximizing the between-class distance. Many extended versions of LDA methods have been presented to improve the classification performance. For instance, a regularized version of LDA (RLDA) [23] was proposed by exploiting a regularized with-class scatter matrix. Wang et al. [24] proposed an HSI classification method by constructing a scatter matrix from a small neighborhood. Since the global structure of HSI data may be inconsistent with the local structure, the classification accuracy of the LDA-based methods is low for the HSI data [25]. To deal with the nonlinear problem involved in the original feature space, the kernel technique has been widely used, which is able to project the data from the original feature space into a kernel-induced space. The corresponding kernel-based versions have kernel principal component analysis (KPCA) [26], kernel independent component analysis (KICA) [27], and kernel discriminant analysis (KDA) [28]. To alleviate the nonlinear and inconsistent problems, Li et al. [29] proposed an unsupervised subspace projection method for single image super-resolution. However, it is difficult to obtain better classification performance because the discrimination information is underutilized.
The classifiers of the HSI data fall into two categories: generative and discriminative [30]. The generative classifiers are to learn the joint probability densities with the feature data and the label information and then compute the posterior probabilities via naive Bayesian [31] or Gaussian mixture model [32]. Although the generative model exploits the feature data exhaustively, it still lacks the discriminative information. The discriminative classifiers are able to find the optimal decision boundaries among different classes, including neighbor neighbors (NN) [33], logistic regression [34], support vector machine (SVM) [35], and random forest (RF) [36]. Compared to the generative classification methods, the discriminative model is able to distinguish between classes. Hence, we use the discriminative model for the corresponding HSI classification.
Due to high-dimensionality of the HSI data, the relationship between the features is often nonliear. Directly applying the linear transformation method for the high-dimensional feature data may lead to over-fitting problem during the training process and may provide low accuracy in the classification process. Kernel-based techniques [37] are designed to deal with the nonlinear problem. We chose the most popular kernel-based feature representation method, i.e., KPCA, as the first-stage subspace projection. However, the original feature data will be projected onto a higher-dimensional subspace to acquire the approximate linear relationship. In order to practically address the dimensionality reduction problem, we propose a new discrimination-information based locality preserving projection (DLPP) method by computing the kernel distances in a k-nearest neighborhood for the foregoing KPCA projected data when the training samples are belong to the same class. Overall, the proposed two-stage subspace projection framework first applies KPCA to the original high-dimensional HSI data and then exploits the proposed DLPP method for the preceding KPCA feature data, which is able to preserve both global and local structures.
In this work, we propose a new dimensionality reduction method for supervised HSI classification, termed as Two-stage Subspace Projection (TwoSP). In order to exploit the global and local structures of the original HSI data, the proposed TwoSP framework first projects the data onto the KPCA space to preserve the global structure and to alleviate the nonlinear problem. Furthermore, the discrimination information in a k-nearest neighborhood for the within-class samples is used to learn the DLPP transformation matrix in the training process. The final subspace found by TwoSP substantially separates the testing samples from different classes. In summary, the main contributions of this paper can be summarized as follows:
(1)
The discrimination information is utilized to compute the samples’ kernel distances from a k-nearest neighborhood in the subspace, so the local structure of the original HSI data can be captured adaptively.
(2)
The proposed TwoSP framework is an effective way for supervised HSI classification by combining the existing KPCA and the proposed DLPP methods, which not only extracts the nonlinear feature, but also exploits the discrimination information to preserve both global and local structures of the original HSI data. In an optimal low-dimensional subspace, the classification boundary can be found for the HSI data.
The remainder of this paper is organized as follows. In Section 2, we briefly introduce the traditional KPCA method. In Section 3, we describe the proposed DLPP method in detail. Section 4 summaries the proposed TwoSP framework for the HSI data. Section 5 evaluates the TwoSP on two real-world HSI databases compared with several previous works. Finally, we provide a conclusion of this work in Section 6.

2. Kernel Principal Component Analysis

In this section, we briefly introduce the kernel principal component analysis (KPCA) method [26,29]. Let X = [ x 1 , x 2 , , x n ] R d × n be the original feature data, where d is the data dimensionality and n is the number of the input samples. We then consider the nonlinear problem in a feature subspace F induced by a mapping function ϕ : R d F . The projected feature data becomes linearly related in F , which is also named as reproducing kernel Hilbert space.
First, the mapping data, ϕ ( x i ) should be transformed by zero-mean normalization, i.e., subtracting the mean vector u = 1 / n i = 1 n ϕ ( x i ) . Similar to principal component analysis (PCA) [19], the corresponding covariance matrix C for KPCA transformation can be defined as
C = 1 n i = 1 n ( ϕ ( x i ) u ) ( ϕ ( x i ) u ) T .
We seek the optimal projection vector v that maximizes the covariance matrix after projection, i.e., solving the eigenvalue problem, that is
Cv = λ v
where λ and v are the eigenvalue and the corresponding eigenvector, respectively. The eigenvector v can be expanded as
v = i = 1 n α i ( ϕ ( x i ) u )
where α i is the i-th weighted coefficient and all the n weights are grouped into α = [ α 1 , α 2 , , α n ] T . Substituting Equations (1) and (3) into Equation (2), the eigenvalue problem can be reduced to the following equation (for the details, see [29]):
K ˜ α = λ α
where K ˜ is the centered kernel matrix with size of n × n . The relationship between K ˜ and the kernel matrix K of the original feature data can be defined as K ˜ = GKG , where G = I n 1 / n 1 n , I n is an identity matrix with size of n × n , and 1 n is a matrix all for 1 with size of n × n . Gaussian radial basis is the most popular kernel function, which is used in this paper, defined as K i , j = e x p ( x i x j 2 2 / σ ) , where σ is the kernel parameter [38].
The optimal KPCA projection matrix W for problem (4) is formed by the r eigenvectors of the centered kernel matrix K ˜ with respect to the r largest eigenvalues. Therefore, the original feature data can be converted onto the KPCA subspace:
X r = W T K ˜ .
Therefore, the original features are mapped from the d-dimensional space R d to a r-dimensional subspace R r , where the optimal value of r may be larger than d. Although the original features may be projected onto a higher-dimensional feature space, the linear relationship and global structure can be captured, which is important for the subsequent dimensionality reduction.

3. Discrimination-Information-Based Locality Preserving Projection

In this section, the proposed discrimination-information based locality preserving projection (DLPP) method is introduced in detail.
After applying the first-stage KPCA subspace projection, we can obtain the mapping data X r = [ x r 1 , x r 2 , , x r n ] R r × n , which preserves the global structure of the original high-dimensional feature data. However, in the real-world HSI classification, the local structure may be inconsistent with the global structure. Therefore, the extraction of the local structure should be taken into consideration in the process of dimensionality reduction.
Denote another nonlinear mapping function as ψ : R r F . Denote the value of kernel matrix be K i , j r = ψ ( x r i ) T ψ ( x r j ) . As mentioned, we also chose the Gaussian kernel function in the computation of the kernel distances, i.e., K i , j r = e x p ( x r i x r j 2 2 / ρ ) , where ρ is the corresponding kernel parameter. Then, we can compute the kernel distances among the KPCA mapping data X r as follows:
D i , j ψ = ψ ( x r i ) ψ ( x r j ) 2 = ψ ( x r i ) T ψ ( x r i ) + ψ ( x r j ) T ψ ( x r j ) 2 ψ ( x r i ) T ψ ( x r j ) = K i , i r + K j , j r 2 K i , j r = 2 2 K i , j r K i , i r = 1   and   K j , j r = 1 .
To further preserve the local structures, an adjacency matrix S is designed to measure the similarity relationship between feature vectors x r i and x r j that are from the same class, i.e.,
S i , j = 1 D i , j ψ , if x r i N ( x r j ) or x r j N ( x r i ) and ( x r i ) = ( x r j ) 0 , otherwise
where N ( x r i ) and N ( x r j ) indicate the k-nearest neighbors of x r i and x r j , respectively. ( · ) is the function to obtain the class information of the input feature vector. That is to say, when the two feature vectors x r i and x r j have the same class information and exist in each other’s neighborhood, the value of the adjacency matrix S is computed by the corresponding kernel distances; otherwise, zero.
The proposed DLPP is to adjust the adjacency during the dimensionality reduction. Denote  P be the projection matrix that embeds the first-stage KPCA mapping data X r into an optimal low-dimensional subspace transformed by X m = P T X r , which yields the following formula:
min P i , j = 1 n P T ( x r i x r j ) 2 2 S i , j .
Problem (8) can be further reduced to
min P tr 1 2 i , j = 1 n P T x r i x r j x r i x r j T P S i , j = min P tr 1 2 i , j = 1 n 2 P T x r i S i , j x r i T P 2 P T x r i S i , j x r i T P T = min P tr i = 1 n P T x r i j = 1 n S i , j x r i T P i , j = 1 n P T x r i S i , j x r j T P = min P tr P T X r Z s X r T P P T X r S X r T P = min P tr P T X r L s X r T P
where L s = Z s S and Z s = diag ( S 1 ˜ ) , 1 ˜ is a vector all for 1 with size of n × 1 , tr ( · ) and diag are the trace function and the diagonal function. To obtain an optimal solution of problem (9), a normalized scale constraint is imposed as
P T X r Z s X r T P = I m
where I m is a m × m identity matrix. Therefore, the optimal DLPP projection matrix P * can be computed as
P * = a r g min P tr P T X r L s X r T P s . t . P T X r Z s X r T P = I m
which can be solved analytically through generalized eigenvalue decomposition between X r L s X r T and X r Z s X r T . P * is then formed by the m eigenvectors corresponding to the m smallest eigenvalues.
When the optimal DLPP projection matrix P * is obtained, the second-stage projected feature data can be computed by
X m = P * T X r .
After the two-stage subspace projection, the dimensionality of the final projected feature data is  m , which is significantly lower than the original dimensionality, i.e., m d .

4. The Proposed Framework

In this section, we will show the proposed TwoSP method can be applied to the HSI data. According to the class information in an ascending sort order, the input HSI feature data X = [ x 1 , x 2 , , x n ] is changed into X = [ X 1 , X 2 , , X c ] , where X i = [ x n i 1 + 1 , x n i 1 + 2 , , x n i ] , i = 1 , 2 , , c , n = n 1 + n 2 + + n c , and n 0 = 0 (i.e., excluding the samples with the class information of 0). These samples are then further partitioned as the training samples and the test ones. Like the dataset partition method in [39], a small portion of the training set is enough for a good classification performance. For each subset X i ( i = 1 , 2 , , c ), we randomly select 5 % samples to compose the training subset and all the remaining samples are used as the test subset. Therefore, we can denote the training dataset as X s = [ X s , 1 , X s , 2 , , X s , c ] , where X s , i = [ x s , i 1 , x s , i 2 , , x s , i n s , i ] , and n s , i = n i * 5 % , and the test subset is defined as X t = [ X t , 1 , X t , 2 , , X t , c ] , where X t , i = [ x t , i 1 , x t , i 2 , , x t , i n t , i ] , and n t , i = n i n s , i . Moreover, the number of training and test samples are denoted as n s = n s , 1 + n s , 2 + + n s , c and n t = n t , 1 + n t , 2 + + n t , c .
For simplicity, the training dataset and the corresponding class information are marked as X s = [ x s 1 , x s 2 , , x s n s ] and Y s = [ y s 1 , y s 2 , , y s n s ] , where y s i [ 1 , 2 , , c ] . On the other hand, the test dataset is labeled as X t = [ x t 1 , x t 2 , , x t n t ] . After the HSI classification, the estimated class information of the test samples is denoted as Y t = [ y t 1 , y t 2 , , y t n t ] , where y t i also belongs to [ 1 , 2 , , c ] . The details of the whole framework is described in Algorithm 1.
Algorithm 1: The Proposed Framework for HSI classification.
Input: 
Training dataset X s , training class information set Y s , test dataset X t , the number of neighbors in DLPP k , first-stage dimensionality r, and second-stage dimensionality m.
Output: 
Estimate the test class information set Y t .
First-Stage Subspace Projection:
  • merge X s and X t into the whole dataset X = [ X s , X t ] = [ x 1 , , x n s , x n s + 1 , , x n s + n t ] ;
  • compute the original kernel matrix K , where K i , j = e x p ( x i x j 2 2 / σ ) and σ = 3 i , j = 1 n x i x j 2 / n 2 2 ; n is equal to n s + n t ;
  • compute the centered kernel matrix K ˜ = GKG , where G = I n 1 / n 1 n ;
  • solve the eigenvalue problem as Equation (4);
  • select the r eigenvectors of K ˜ corresponding to the r largest eigenvalues, to construct the first-stage subspace projection matrix W ;
  • according to Equation (5), obtain the KPCA feature data X r ;
Second-Stage Subspace Projection:
7.
extract the projected training samples X s , r = [ x s , r 1 , x s , r 2 , , x s , r n s ] , and compute the corresponding kernel matrix K r , where K i , j r = e x p ( x s , r i x s , r j 2 2 / ρ ) and ρ = 3 i , j = 1 n s x s , r i x s , r j 2 / n s 2 2 ;
8.
compute the adjacency matrix S in the corresponding k-nearest neigbhors with the same class using Equation (7), where y s i = ( x s , r i ) ;
9.
solve problem (11), where Z s = diag ( S 1 ˜ ) and L s = Z s S ;
10.
choose the m eigenvectors corresponding to the m smallest eigenvalues, to construct the second-stage subspace projection matrix P * ;
11.
for the training samples, the DLPP feature data X s , m is obtained using Equation (12), i.e., X s , m = P * T X s , r .;
Classification:
for i = 1 , 2 , , n t do
12.
also according to Equation (12), the final projected feature vector for each test sample can be computed as x t , m i = P * T x t , r i ;
13.
using the NN classifier, find the nearest training sample in the optimal low-dimensional feature space, i.e., j * = a r g min j x t , m i x s , m j 2 2 , where j = 1 , , n s ;
14.
obtain the corresponding class information, i.e., y t i = ( x s , m j * ) ; and then set i = i + 1 ;
end
Result: Obtain the class estimation for all the test samples, i.e., Y t .

5. Experimental Results

To validate the effectiveness of the proposed algorithm for hyperspectral image classification, experiments are conducted to compare with several existing dimensionality reduction methods including principal component analysis (PCA) [19], independent component analysis (ICA) [20], linear discriminant analysis (LDA) [21], kernel discriminant analysis (KDA) [28], kernel principal component analysis (KPCA) [26], the proposed discrimination-information based locality preserving projection (DLPP), and the proposed two-stage subspace projection (TwoSP). In addition, we also compare the classification results of the raw spectral features (RAW).

5.1. Experimental Setting

In this paper, we conduct the experiments on two real-world HSI datasets, i.e., Indian Pines and Kennedy Space Center (KSC) datasets [40], to demonstrate the effectiveness of the proposed TwoSP method compared with the existing dimensionality reduction algorithms.
The Indian Pines dataset in the corrected version consists of 145 × 145 pixels and 200 spectral bands, which was gathered by an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor over Northwestern Indiana. By removing the background with the class information of 0, 10,249 pixels are annotated from 16 classes. On the other hand, the KSC dataset was acquired by an AVIRIS sensor over the Kennedy Space Center, Florida. The size of this HSI is 512 × 614. Each pixel has 176 spectral bands with the annotated 13 classes. By removing the background pixels, 5211 valuable pixels remain. Table 1 shows various land-cover types and the selected sizes of the training and test subsets for the two aforementioned HSI datasets.
For classification, we use the nearest neighbor (NN), support vector machine (SVM), and random forest (RF) classifiers to obtain the estimated classes of the test samples. For SVM, we use the “libsvm” toolbox in a Matlab version with a linear kernel [41]. After that, we select three widely used classification measurements, i.e., average accuracy (AA), overall accuracy (OA), and kappa coefficient (KC) between the estimates and the ground-truths of all the classes, to objectively assess the performance of HSI classification. All the experiments we performed on a personal computer with Intel Xeon CPU E5-2643 v3, 3.40 GHz, 64 GB memory, and 64-bit Windows 7 using Matlab R2017b.

5.2. Performance on Hyperspectral Image Datasets

To alleviate the random error caused by the randomly selecting 5 % samples as the training set and all the remaining samples as the test set, we repeated the experimental results five times and reported the average accuracy for each class, average AAs for all the classes, average OAs and average KCs for all the test samples with their corresponding standard deviations (STDs) in this paper. The quantitative classification accuracy of these methods are given in Table 2 and Table 3 for the Indian Pines dataset and KSC dataset, respectively. Each method uses its best reduced dimensionality shown in brackets.
From the two tables, we can see that TwoSP outperforms all the competitors in terms of AA, OA, and KC. Especially, the classification performance of TwoSP is better than directly applying the proposed DLPP method to the original HSI feature data. PCA neglects the nonlinear relationship from the original high-dimensional feature data, although it achieves the dimensionality reduction. Due to the nonlinear problem, which may lead to a one-to-many relationship among the feature vector, ICA is unable to capture the global and local structures. LDA preserves the local manifold structure by exploiting the discrimination information of the training samples. However, the global structure is lost in the dimensionality reduction. KDA considers both the global and local data relationship in the discriminant analysis. For a small number of training samples, the classification accuracy is often low. For instance, “Class No. 9” in Table 2, the average OA value (%) for individual class No. 9 of KDA, is 6.3, while that of our TwoSP reaches up to 30.5 in spite of having only one training sample. To achieve dimensionality reduction, KDA introduces the discrimination information of all the within-class and between-class samples. However, the within-class samples that are far away from the referred sample may have a negative effect on the final classification accuracy. Therefore, the discrimination information in a small neighborhood should be taken into consideration in the dimensionality reduction process, which is proposed in this paper. KPCA largely alleviates the nonlinear problem from the HSI data, but it cannot preserve the local data relationship with the valuable discrimination information. The proposed DLPP method directly applying to the original feature data is better than the traditional linear feature transformation approaches, i.e., PCA, ICA, and LDA. The proposed TwoSP method investigates the local structure of the HSI data adaptively, and preserves the global structure in the first-stage subspace projection. Therefore, TwoSP can achieve the best classification performance on a large proportion of the occasions shown in the quantitative results.
Furthermore, the classification maps of the aforementioned methods on Indian Pines and KSC datasets are visualized in Figure 1 and Figure 2. RAW, PCA, ICA, and KPCA have low classification performance because they cannot use the discrimination information in the subspace projection process. Although LDA and KDA use the discrimination information, they introduce all the within-class and between-class samples in the discriminant analysis, which is difficult to preserve the local manifold structure. From the subfigures (f) and (g) in Figure 1 and Figure 2, a clear classification map is shown for several classes. TwoSP, which enforces the discrimination information in a small neighborhood, shows better visualization quality than the others, since it not only preserves the global structure in the first-stage KPCA subspace projection, but also learns the local data relationship in the second-stage DLPP subspace projection. It also demonstrates that the utilization of discrimination information within a small neighborhood can largely improve the classification performance in the corresponding nonlinear feature space.
To further validate the effectiveness of the proposed method, we conducted an additional experiment by the effective cross validation method, i.e., the McNemar test [42]. Table 4 shows the results of the McNemar test for the proposed method and the baselines, where the methods on the vertical direction are the test methods, while those on the horizontal direction are the reference methods. When the value is greater than zero, it illustrates that the classification performance of the test method is better than that of the reference method; otherwise, the reference method has more advantages. Generally, the threshold of significance in the McNemar test is set to 0.05 . Furthermore, when the absolute value is larger than 1.96 , it indicts that the two methods have obvious differences. From Table 4, we can see that the TwoSP has the highest classification performance when compared with the existing approaches. In addition, the simplified version of our TwoSP, i.e., DLPP, also has good performance to some extent.

5.3. Discussion on Computational Cost

In this subsection, we only discuss the computational cost of the proposed method compared with state-of-the-art approaches. Considering the main steps in Algorithm 1, the proposed HSI classification method takes account of three parts: KPCA subspace projection, DLPP subspace projection, and classification.
Let n s and n t be the number of training and test samples, respectively. n = n s + n t . The computational complexity of first-stage subspace projection matrix, i.e., W , is O ( n 3 ) . Here, we can obtain the KPCA feature data X r with the dimensionality of r. k is denoted as the number of neighbors in the computation of adjacency matrix. The computational complexity of the second-stage subspace projection process is then O ( r k n s + r 3 ) . After the two-stage subspace projection framework, the original HSI data is projected onto an optimal low-dimensional feature space with the dimensionality of m. For classification using NN classifier, each test sample is compared with all the training samples to find the nearest neighbor. Therefore, the computational complexity of classification is O ( m n t n s ) .
Table 5 shows the computational time (in terms of seconds) of the proposed method and the baselines, where T1 represents the computation of projection for different methods, and T2, T3, and T4 are the classification time obtained by NN, RF, and SVM classifiers, respectively. Since the first-stage projection of TwoSP needs to compute the kernel matrix (i.e., K ) with all the training and test samples, T1 of TwoSP is larger than that of other methods. The T1s of KPCA and TwoSP are similar because they both involve in the computation of a large kernel matrix. KDA uses the discriminant information of a small portion of training samples, so the computation of its kernel matrix is small. RAW directly puts the original HSI data into the classification process, so its T1 is null. Among the T2, T3, and T4 of different methods, we can see that the RF classifier needs more time to classify all test samples. Considering classification performance and computational time simultaneously, we chose the NN classifier in the experiments.

5.4. Performance of Reduced Dimensionality

Each method is performed with different reduced dimensionality. To demonstrate the optimal reduced dimensionality for the corresponding method, the classification accuracy in terms of overall accuracy is obtained by the NN classifier when the dimensionality varies within { 2 , 3 , , 30 } on Indian Pines and KSC, respectively.
Figure 3 shows the curves of OA versus the reduced dimensionality on two different HSI datasets. The quantitative results are obtained with a random dataset partition. The proposed TwoSP method achieves the highest OA constantly. Especially, the OA value obtained by TwoSP exceeds that of the other methods to a large extent when the reduced dimensionality is more than 7. In Figure 3, the classification performance becomes stable when the dimensionality increases to a certain value. In most cases, the performance with low-dimensional projected data is better than the original high-dimensional data, which also validates that the dimensionality reduction does improve the classification accuracy.

5.5. Analysis of Classifier

To evaluate the classification performance of each dimensionality reduction method with three different classifiers, i.e., NN, RF, and SVM, we randomly selected five different dataset partitions and then computed the average OA values. After the projected feature data was obtained by one of these dimensionality reduction methods, the classes of the test set were discriminated by the NN, RF, and SVM classifiers, respectively. Figure 4 shows the classification results vs. different classifiers on two HSI datasets.
In Figure 4a,b, the proposed TwoSP method presents the best classification performance for different classifiers when compared with the other dimensionality reduction methods, which also demonstrates that TwoSP has better robustness in different classifiers. For most cases on the Indian Pines dataset, the NN and RF classifiers generates better OAs than the SVM classifier. On the other hand, the RF classifier achieves better results on the KSC dataset, while the results with NN are superior to RF for LDA and DLPP. It also illustrates the classification accuracy of different classifiers on different datasets may have relatively large differences. To unify the utilization of the classification model, we applied the NN classifier to evaluate the classification performance of each class in Table 2 and Table 3.

5.6. Analysis of Parameters

In our proposed TwoSP method, there are three parameters set first, as shown in the input of Algorithm 1. Therefore, the first- and second-stage subspace projection dimensionality r and m and the number of neighbors k are studied experimentally. We randomly choose one of the training and test set partition and mainly take the Indian Pines dataset for instance. The objective values of the parameters are changed during the analysis process.
Figure 5a shows the overall accuracy results obtained by simultaneously varying the dimensionalities r and m. Since the first-stage subspace projection is designed to preserve the global structure of the original HSI data, the dimensionality of the second-stage subspace projection m should be smaller than r. From this subfigure, we can see that TwoSP is robust to r and m in a wide range. When r and m increase to a certain value, the classification performance in terms of OA is the best. Therefore, we select r = 45 and m = 20 for the Indian Pines dataset in the experiments. Figure 5b shows the different OA values vs. different number of neighbors in the second-stage subspace projection. TwoSP achieves the highest OA value when the number of neighbors k = 200 , which is chosen in this paper. Note that the neighbors are selected by comparing the kernel distances first. Only the within-class samples are used to compute the value of adjacency matrix, and zero otherwise. Therefore, in practice a small number of training samples are used to learn the local structure of the first-stage projected data.

6. Conclusions

In this paper, we proposed the TwoSP method on the basis of the preservation of global and local structures to learn the optimal low-dimensional feature space for HSI classification. TwoSP first applies the traditional KPCA method to address the nonlinear problem which often exists in the HSI data. However, the dimensionality of the first-stage subspace projection is not small enough. It needs to apply the second-stage subspace projection to the preceding projected features. TwoSP exploits the a priori knowledge of the training samples to construct an adjacency matrix for the within-class samples, which can enhance the discrimination information of projected features.
Compared with the state of the art, TwoSP is better able to learn the data manifold relationship in the desired feature space, which creates various valuable features for HSI classification. In addition, TwoSP strongly retains the local smoothness within a small neighborhood. Through the experiments on two real-world HSI datasets, i.e., Indian Pines and KSC, TwoSP provides better classification performance than the existing dimensionality reduction methods, which also validates the effectiveness of the proposed method.
Our future work will focus on how to extend the proposed method to train the optimal transformation matrix for each test sample quickly. It is desirable to improve the classification performance and increase the computation efficiency of the dimensionality reduction.

Author Contributions

All authors designed the study and discussed the basic structure of the manuscript. X.L. carried out the experiments and finished the first version of this manuscript. L.Z. provided the framework of the experiments. J.Y. reviewed and edited this paper.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grants 61601416, 61711530239 and 41701417, the Fundamental Research Funds for the Central Universities, China University of Geosciences, Wuhan, under Grant CUG170612, the Hong Kong Scholars Program under Grant XJ2017030, and the Hong Kong Polytheistic University Research Fund.

Conflicts of Interest

All the authors declare no conflict of interest.

References

  1. Stein, D.W.J.; Beaven, S.G.; Hoff, L.E.; Winter, E.M.; Schaum, A.P.; Stocker, A.D. Anomaly detection from hyperspectral imagery. IEEE Signal Process. Mag. 2002, 19, 58–69. [Google Scholar] [CrossRef] [Green Version]
  2. Gevaert, C.M.; Suomalainen, J.; Tang, J.; Kooistra, L. Generation of Spectral-Temporal Response Surfaces by Combining Multispectral Satellite and Hyperspectral UAV Imagery for Precision Agriculture Applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3140–3146. [Google Scholar] [CrossRef]
  3. Luo, F.; Du, B.; Zhang, L.; Zhang, L.; Tao, D. Feature Learning Using Spatial-Spectral Hypergraph Discriminant Analysis for Hyperspectral Image. IEEE Trans. Cybern. 2018, PP, 1–14. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, W.; Ma, X.; Zhou, Y.; Tao, D.; Cheng, J. p-Laplacian Regularization for Scene Recognition. IEEE Trans. Cybern. 2018, 1–14. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene Classification with Recurrent Attention of VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2018, PP, 1–13. [Google Scholar] [CrossRef]
  6. Ma, X.; Liu, W.; Li, S.; Tao, D.; Zhou, Y. Hypergraph p-Laplacian Regularization for Remotely Sensed Image Recognition. IEEE Trans. Geosci. Remote Sens. 2018, 1–11. [Google Scholar] [CrossRef]
  7. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef]
  8. Mitra, P.; Murthy, C.A.; Pal, S.K. Unsupervised feature selection using feature similarity. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 301–312. [Google Scholar] [CrossRef] [Green Version]
  9. Wang, Q.; Lin, J.; Yuan, Y. Salient Band Selection for Hyperspectral Image Classification via Manifold Ranking. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1279–1289. [Google Scholar] [CrossRef] [PubMed]
  10. Liu, W.; Yang, X.; Tao, D.; Cheng, J.; Tang, Y. Multiview dimension reduction via Hessian multiset canonical correlations. Inf. Fusion 2018, 41, 119–128. [Google Scholar] [CrossRef]
  11. Wang, Q.; He, X.; Li, X. Locality and Structure Regularized Low Rank Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, PP, 1–13. [Google Scholar] [CrossRef]
  12. Zhang, L.; Zhang, L.; Tao, D.; Huang, X.; Du, B. Hyperspectral Remote Sensing Image Subpixel Target Detection Based on Supervised Metric Learning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4955–4965. [Google Scholar] [CrossRef]
  13. Luo, Y.; Wen, Y.; Tao, D.; Gui, J.; Xu, C. Large Margin Multi-Modal Multi-Task Feature Extraction for Image Classification. IEEE Trans. Image Process. 2016, 25, 414–427. [Google Scholar]
  14. Zhang, L.; Zhang, Q.; Du, B.; Huang, X.; Tang, Y.Y.; Tao, D. Simultaneous Spectral-Spatial Feature Selection and Extraction for Hyperspectral Images. IEEE Trans. Cybern. 2018, 48, 16–28. [Google Scholar] [CrossRef] [PubMed]
  15. Tao, D.; Tang, X.; Li, X. Which Components Are Important For Interactive Image Searching? IEEE Trans. Circuits Syst. Video Technol. 2008, 18, 3–11. [Google Scholar]
  16. Sun, W.; Tian, L.; Xu, Y.; Du, B.; Du, Q. A Randomized Subspace Learning Based Anomaly Detector for Hyperspectral Imagery. Remote Sens. 2018, 10, 417. [Google Scholar] [CrossRef]
  17. Luo, F.; Huang, H.; Duan, Y.; Liu, J.; Liao, Y. Local Geometric Structure Feature for Dimensionality Reduction of Hyperspectral Imagery. Remote Sens. 2017, 9, 790. [Google Scholar] [CrossRef]
  18. Li, X.; He, H.; Wang, R.; Tao, D. Single Image Superresolution via Directional Group Sparsity and Directional Features. IEEE Trans. Image Process. 2015, 24, 2874–2888. [Google Scholar] [CrossRef] [PubMed]
  19. Prasad, S.; Bruce, L.M. Limitations of Principal Components Analysis for Hyperspectral Target Recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  20. Wang, J.; Chang, C.I. Independent component analysis-based dimensionality reduction with applications in hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1586–1600. [Google Scholar] [CrossRef] [Green Version]
  21. Wang, H.; Lu, X.; Hu, Z.; Zheng, W. Fisher Discriminant Analysis With L1-Norm. IEEE Trans. Cybern. 2014, 44, 828–842. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, Z.; He, B. Locality perserving projections algorithm for hyperspectral image dimensionality reduction. In Proceedings of the 2011 19th International Conference on Geoinformatics, Shanghai, China, 24–26 June 2011; pp. 1–4. [Google Scholar]
  23. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of Hyperspectral Images With Regularized Linear Discriminant Analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, Q.; Meng, Z.; Li, X. Locality Adaptive Discriminant Analysis for Spectral-Spatial Classification of Hyperspectral Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2077–2081. [Google Scholar] [CrossRef]
  25. Dong, Y.; Du, B.; Zhang, L.; Zhang, L. Dimensionality Reduction and Classification of Hyperspectral Images Using Ensemble Discriminative Local Metric Learning. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2509–2524. [Google Scholar] [CrossRef]
  26. Fauvel, M.; Chanussot, J.; Benediktsson, J.A. Kernel Principal Component Analysis for the Classification of Hyperspectral Remote Sensing Data over Urban Areas. EURASIP J. Adv. Signal Process. 2009, 2009, 783194. [Google Scholar] [CrossRef]
  27. Mei, F.; Zhao, C.; Wang, L.; Huo, H. Anomaly Detection in Hyperspectral Imagery Based on Kernel ICA Feature Extraction. In Proceedings of the 2008 Second International Symposium on Intelligent Information Technology Application, Shanghai, China, 20–22 December 2008; Volume 1, pp. 869–873. [Google Scholar]
  28. Baudat, G.; Anouar, F. Generalized Discriminant Analysis Using a Kernel Approach. Neural Comput. 2000, 12, 2385–2404. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Li, X.; He, H.; Yin, Z.; Chen, F.; Cheng, J. Single image super-resolution via subspace projection and neighbor embedding. Neurocomputing 2014, 139, 310–320. [Google Scholar] [CrossRef]
  30. Feng, J.; Liu, L.; Cao, X.; Jiao, L.; Sun, T.; Zhang, X. Marginal Stacked Autoencoder With Adaptively-Spatial Regularization for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3297–3311. [Google Scholar] [CrossRef]
  31. Borges, J.S.; Bioucas-Dias, J.M.; Marcal, A.R.S. Bayesian Hyperspectral Image Segmentation With Discriminative Class Learning. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2151–2164. [Google Scholar] [CrossRef] [Green Version]
  32. Fauvel, M.; Dechesne, C.; Zullo, A.; Ferraty, F. Fast Forward Feature Selection of Hyperspectral Images for Classification With Gaussian Mixture Models. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2824–2831. [Google Scholar] [CrossRef]
  33. Cariou, C.; Chehdi, K. A new k-nearest neighbor density-based clustering method and its application to hyperspectral images. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 6161–6164. [Google Scholar]
  34. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised Hyperspectral Image Segmentation Using Multinomial Logistic Regression With Active Learning. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4085–4098. [Google Scholar] [CrossRef]
  35. Xu, Y. Maximum Margin of Twin Spheres Support Vector Machine for Imbalanced Data Classification. IEEE Trans. Cybern. 2017, 47, 1540–1550. [Google Scholar] [CrossRef] [PubMed]
  36. Pullanagari, R.R.; Kereszturi, G.; Yule, I. Integrating Airborne Hyperspectral, Topographic, and Soil Data for Estimating Pasture Quality Using Recursive Feature Elimination with Random Forest Regression. Remote Sens. 2018, 10, 1117. [Google Scholar] [CrossRef]
  37. De Bie, T.; Cristianini, N. Kernel Methods for Exploratory Pattern Analysis: A Demonstration on Text Data. In Structural, Syntactic, and Statistical Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2004; pp. 16–29. [Google Scholar]
  38. Li, X.; He, H.; Yin, Z.; Chen, F.; Cheng, J. KPLS-based image super-resolution using clustering and weighted boosting. Neurocomputing 2015, 149, 940–948. [Google Scholar] [CrossRef]
  39. Chen, M.; Wang, Q.; Li, X. Discriminant Analysis with Graph Learning for Hyperspectral Image Classification. Remote Sens. 2018, 10, 836. [Google Scholar] [CrossRef]
  40. Hyperspectral Remote Sensing Scenes. Available online: http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 31 July 2018).
  41. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. Available online: http://www.csie.ntu.edu.tw/~cjlin/libsvm (accessed on 15 July 2018). [CrossRef]
  42. Eyden, R.J.V.; Wit, P.W.C.D.; Arron, J.C. Predicting company failure-a comparison between neural networks and established statistical techniques by applying the McNemar test. In Proceedings of the Conference on Computational Intelligence for Financial Engineering (CIFEr), New York, NY, USA, 9–11 April 1995; pp. 91–96. [Google Scholar]
Figure 1. Classification maps for the Indian Pines dataset with different dimensionality reduction algorithms using NN classifier. (a) The ground truth of the classes. (b) Randomly selected training samples. (c) Result obtained by using RAW features. (d) PCA. (e) ICA. (f) LDA. (h) KDA. (g) KPCA. (i) the proposed DLPP. (j) The proposed TwoSP.
Figure 1. Classification maps for the Indian Pines dataset with different dimensionality reduction algorithms using NN classifier. (a) The ground truth of the classes. (b) Randomly selected training samples. (c) Result obtained by using RAW features. (d) PCA. (e) ICA. (f) LDA. (h) KDA. (g) KPCA. (i) the proposed DLPP. (j) The proposed TwoSP.
Remotesensing 10 01565 g001
Figure 2. Classification maps (cropped to a size of 290 × 310) for the KSC dataset with different dimensionality reduction algorithms using NN classifier. (a) The ground truth of the classes. (b) Randomly selected training samples. (c) Result obtained by using RAW features. (d) PCA. (e) ICA. (f) LDA. (h) KDA. (g) KPCA. (i) The proposed DLPP. (j) The proposed TwoSP.
Figure 2. Classification maps (cropped to a size of 290 × 310) for the KSC dataset with different dimensionality reduction algorithms using NN classifier. (a) The ground truth of the classes. (b) Randomly selected training samples. (c) Result obtained by using RAW features. (d) PCA. (e) ICA. (f) LDA. (h) KDA. (g) KPCA. (i) The proposed DLPP. (j) The proposed TwoSP.
Remotesensing 10 01565 g002
Figure 3. Overall Accuracy (OA) obtained by the NN classifier versus the reduced dimensionality of various different methods on two real-world HSI datasets, i.e., (a) Indian Pines and (b) KSC.
Figure 3. Overall Accuracy (OA) obtained by the NN classifier versus the reduced dimensionality of various different methods on two real-world HSI datasets, i.e., (a) Indian Pines and (b) KSC.
Remotesensing 10 01565 g003
Figure 4. Classification Results in terms of Overall Accuracy (OA) of each dimensionality reduction method with different classifiers on (a) Indian Pines and (b) KSC.
Figure 4. Classification Results in terms of Overall Accuracy (OA) of each dimensionality reduction method with different classifiers on (a) Indian Pines and (b) KSC.
Remotesensing 10 01565 g004
Figure 5. Parameter analysis of (a) OAs with simultaneously varying the first-stage subspace projection dimensionality r and the second-stage subspace projection dimensionality m on Indian Pines, and (b) OAs with varying the number of neighbors k in the computation of adjacency matrix.
Figure 5. Parameter analysis of (a) OAs with simultaneously varying the first-stage subspace projection dimensionality r and the second-stage subspace projection dimensionality m on Indian Pines, and (b) OAs with varying the number of neighbors k in the computation of adjacency matrix.
Remotesensing 10 01565 g005
Table 1. Land-cover classes with the number of training and test samples for the Indian Pines and KSC.
Table 1. Land-cover classes with the number of training and test samples for the Indian Pines and KSC.
Indian PinesKSC
Class No.Land CoverTrainingTestClass No.Land CoverTrainingTest
1Alfalfa3431Scurb39722
2Corn-notill7213562Willow-swamp13230
3Corn-mintill427883Cabbage-palm-hammock13243
4Corn122254Cabbage-palm/oak-hammock13239
5Grass-pasture254585Slash-pine9152
6Grass-tree376936Oak/broadleaf-hammock12217
7Grass-pasture-mowed2267Hardwood-swamp699
8Hay-windrowed244548Graminoid-marsh22409
9Oats1199Spartina-marsh26494
10Soybeans-notill4992310Cattail-marsh21383
11Soybeans-mintill123233211Salt-marsh21398
12Soybeans-clean3056312Mud-flats26477
13Wheat1119413Water47880
14Woods641201
15Bldg-grass-tree-drives20366
16Stone-steel-towers588
Total 5209729Total 2684943
Table 2. Performance (%) of different methods (with the best reduced dimensionality in brackets) on the Indian Pines dataset using the NN classifier.
Table 2. Performance (%) of different methods (with the best reduced dimensionality in brackets) on the Indian Pines dataset using the NN classifier.
Class No.RAW (200)PCA (17)ICA (17)LDA (11)KDA (16)KPCA (40)DLPP (14)TwoSP (20)
140.0 ± 12.642.3 ± 20.641.4 ± 22.242.8 ± 17.845.1 ± 11.243.3 ± 23.549.3 ± 15.731.6 ± 12.2
248.1 ± 1.249.3 ± 2.038.9 ± 4.157.0 ± 2.168.1 ± 3.639.5 ± 3.359.8 ± 3.167.0 ± 1.7
344.2 ± 2.444.0 ± 2.329.0 ± 3.242.6 ± 6.659.4 ± 3.429.9 ± 4.448.6 ± 5.456.1 ± 3.2
430.2 ± 6.328.2 ± 6.929.4 ± 5.627.9 ± 3.239.3 ± 5.029.9 ± 5.831.8 ± 5.340.1 ± 4.8
576.3 ± 5.175.9 ± 3.644.4 ± 3.283.6 ± 5.582.6 ± 2.947.8 ± 4.183.4 ± 3.384.1 ± 4.4
692.2 ± 1.992.4 ± 2.278.0 ± 4.591.2 ± 2.192.0 ± 2.578.8 ± 4.190.1 ± 2.592.3 ± 1.7
780.0 ± 10.079.2 ± 10.455.4 ± 15.576.2 ± 14.773.1 ± 12.559.2 ± 15.085.4 ± 9.282.3 ± 6.4
894.5 ± 2.394.9 ± 2.587.5 ± 2.996.0 ± 1.484.7 ± 1.887.1 ± 1.093.7 ± 2.889.3 ± 5.8
914.7 ± 10.814.7 ± 10.81.1 ± 2.412.6 ± 9.66.3 ± 6.92.1 ± 4.724.2 ± 13.230.5 ± 9.4
1061.3 ± 6.562.2 ± 6.750.3 ± 6.146.8 ± 3.861.2 ± 3.051.3 ± 5.157.4 ± 2.969.1 ± 1.4
1167.4 ± 3.467.4 ± 2.156.4 ± 2.464.8 ± 1.775.9 ± 2.857.6 ± 2.363.1 ± 1.174.8 ± 2.0
1235.0 ± 2.634.6 ± 3.729.5 ± 1.946.8 ± 3.959.3 ± 8.029.6 ± 1.948.4 ± 4.956.0 ± 6.3
1393.3 ± 1.593.2 ± 1.777.9 ± 6.792.6 ± 3.797.4 ± 1.078.6 ± 6.693.7 ± 5.096.4 ± 1.3
1490.3 ± 2.989.3 ± 3.881.1 ± 3.093.4 ± 1.694.2 ± 1.181.6 ± 2.894.0 ± 2.095.2 ± 2.1
1527.2 ± 2.027.3 ± 2.320.3 ± 3.244.6 ± 8.144.2 ± 3.320.9 ± 3.644.5 ± 7.746.7 ± 6.1
1686.4 ± 3.086.4 ± 3.189.1 ± 4.979.8 ± 8.376.6 ± 9.980.9 ± 5.682.1 ± 6.086.4 ± 3.2
AA61.3 ± 1.461.3 ± 2.050.6 ± 1.962.4 ± 2.766.2 ± 1.551.1 ± 2.165.6 ± 0.868.6 ± 0.9
OA64.8 ± 1.064.9 ± 0.853.6 ± 1.065.8 ± 1.773.4 ± 0.954.4 ± 0.767.5 ± 0.973.9 ± 1.2
KC59.7 ± 1.159.8 ± 0.847.1 ± 1.160.8 ± 2.161.0 ± 18.448.0 ± 0.762.9 ± 1.070.1 ± 1.4
Table 3. Performance (%) of different methods (with the best reduced dimensionality in brackets) on the KSC dataset using the NN classifier.
Table 3. Performance (%) of different methods (with the best reduced dimensionality in brackets) on the KSC dataset using the NN classifier.
Class No.RAW (176)PCA (23)ICA (8)LDA (10)KDA (10)KPCA (87)DLPP (41)TwoSP (22)
187.7 ± 3.287.6 ± 3.260.9 ± 3.786.1 ± 4.266.9 ± 3.465.9 ± 5.782.4 ± 1.890.6 ± 2.3
275.7 ± 13.375.7 ± 13.250.9 ± 13.486.4 ± 1.958.2 ± 14.349.7 ± 12.787.8 ± 2.477.4 ± 2.0
366.7 ± 3.766.8 ± 3.826.2 ± 3.153.4 ± 7.127.7 ± 5.128.0 ± 4.655.3 ± 7.777.4 ± 3.5
450.0 ± 3.349.9 ± 3.024.4 ± 2.539.9 ± 4.526.0 ± 4.823.3 ± 3.739.3 ± 5.164.0 ± 4.2
545.8 ± 13.145.4 ± 12.929.3 ± 8.750.4 ± 4.733.6 ± 12.727.9 ± 11.251.3 ± 1.973.0 ± 1.8
634.2 ± 5.234.5 ± 5.123.8 ± 0.450.6 ± 8.222.8 ± 3.523.7 ± 5.552.9 ± 8.859.0 ± 7.5
764.0 ± 11.263.0 ± 10.323.0 ± 6.453.7 ± 13.023.6 ± 11.124.7 ± 7.447.3 ± 9.278.8 ± 9.5
869.4 ± 6.869.0 ± 7.442.4 ± 6.878.4 ± 4.446.0 ± 7.344.7 ± 7.777.7 ± 3.883.1 ± 3.5
988.5 ± 4.088.5 ± 4.067.4 ± 6.382.3 ± 3.267.7 ± 7.868.2 ± 5.780.7 ± 2.393.5 ± 2.6
1081.1 ± 2.481.2 ± 2.550.2 ± 8.894.8 ± 0.861.9 ± 5.554.6 ± 6.091.6 ± 1.682.5 ± 1.7
1192.8 ± 1.692.8 ± 1.688.3 ± 1.787.6 ± 2.694.4 ± 1.688.2 ± 3.687.5 ± 3.185.2 ± 3.0
1278.3 ± 4.778.2 ± 4.662.9 ± 9.189.1 ± 2.668.1 ± 6.164.1 ± 5.388.3 ± 4.180.9 ± 4.2
1398.4 ± 0.998.4 ± 0.998.0 ± 1.099.6 ± 0.498.1 ± 0.798.2 ± 0.998.5 ± 0.998.5 ± 0.9
AA71.7 ± 1.871.6 ± 1.849.8 ± 1.773.3 ± 1.153.5 ± 1.950.9 ± 1.772.4 ± 1.380.3 ± 0.9
OA79.6 ± 0.679.5 ± 0.660.9 ± 1.681.4 ± 0.664.6 ± 0.962.4 ± 1.080.3 ± 0.685.0 ± 0.6
KC77.3 ± 0.677.2 ± 0.656.5 ± 1.879.3 ± 0.760.6 ± 1.058.1 ± 1.178.1 ± 0.783.3 ± 0.8
Table 4. McNemar test of different methods on the Indian Pines and KSC datasets.
Table 4. McNemar test of different methods on the Indian Pines and KSC datasets.
MethodsIndian PinesKSC
PCAICALDAKDAKPCADLPPTwoSPPCAICALDAKDAKPCADLPPTwoSP
RAW−0.320.8−1.9−1.320.2−4.9−18.70.926.6−3.024.326.9−1.3−9.4
PCA-21.5−1.7−1.220.9−4.7−18.1-26.5−3.124.226.8−1.2−9.5
ICA--−19.3−19.1−1.5−22.1−33.7--−24.5−4.7−1.0−21.7−25.4
LDA---0.718.6−3.6−15.5---21.924.35.5−7.1
KDA----18.5−3.9−17.4----4.2−18.8−23.1
KPCA-----−21.4−33.2-----−21.5−25.1
DLPP------−12.8------−10.2
Table 5. Computational time (s) of different methods on Indian Pines and KSC datasets.
Table 5. Computational time (s) of different methods on Indian Pines and KSC datasets.
MethodsIndian PinesKSC
T1T2T3T4T1T2T3T4
RAW-1.6735.482.84-0.4220.632.88
PCA0.680.7021.480.920.280.1912.810.67
ICA11.690.6621.363.195.410.178.440.52
LDA0.200.6217.640.630.080.187.580.31
KDA1.050.6919.420.800.350.167.640.30
KPCA423.630.8326.031.6089.230.3116.700.66
DLPP0.610.6718.480.530.170.2210.200.41
TwoSP428.250.6821.280.8389.380.1912.000.41

Share and Cite

MDPI and ACS Style

Li, X.; Zhang, L.; You, J. Hyperspectral Image Classification Based on Two-Stage Subspace Projection. Remote Sens. 2018, 10, 1565. https://doi.org/10.3390/rs10101565

AMA Style

Li X, Zhang L, You J. Hyperspectral Image Classification Based on Two-Stage Subspace Projection. Remote Sensing. 2018; 10(10):1565. https://doi.org/10.3390/rs10101565

Chicago/Turabian Style

Li, Xiaoyan, Lefei Zhang, and Jane You. 2018. "Hyperspectral Image Classification Based on Two-Stage Subspace Projection" Remote Sensing 10, no. 10: 1565. https://doi.org/10.3390/rs10101565

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop