Next Article in Journal
Recent Large Scale Environmental Changes in the Mediterranean Sea and Their Potential Impacts on Posidonia Oceanica
Next Article in Special Issue
Multi-Feature Manifold Discriminant Analysis for Hyperspectral Image Classification
Previous Article in Journal
Farmland Extraction from High Spatial Resolution Remote Sensing Images Based on Stratified Scale Pre-Estimation
Previous Article in Special Issue
Unsupervised Feature Selection Based on Ultrametricity and Sparse Training Data: A Case Study for the Classification of High-Dimensional Hyperspectral Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Locally Weighted Discriminant Analysis for Hyperspectral Image Classification

1
School of Computer Science, China University of Geosciences, Wuhan 430074, China
2
School of Computer, Wuhan University, Wuhan 430072, China
3
Department of Computing, The Hong Kong Polytechnic University, Kowloon, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(2), 109; https://doi.org/10.3390/rs11020109
Submission received: 26 November 2018 / Revised: 27 December 2018 / Accepted: 3 January 2019 / Published: 9 January 2019
(This article belongs to the Special Issue Dimensionality Reduction for Hyperspectral Imagery Analysis)

Abstract

:
A hyperspectral image (HSI) contains a great number of spectral bands for each pixel, which will limit the conventional image classification methods to distinguish land-cover types of each pixel. Dimensionality reduction is an effective way to improve the performance of classification. Linear discriminant analysis (LDA) is a popular dimensionality reduction method for HSI classification, which assumes all the samples obey the same distribution. However, different samples may have different contributions in the computation of scatter matrices. To address the problem of feature redundancy, a new supervised HSI classification method based on locally weighted discriminant analysis (LWDA) is presented. The proposed LWDA method constructs a weighted discriminant scatter matrix model and an optimal projection matrix model for each training sample, which is on the basis of discriminant information and spatial-spectral information. For each test sample, LWDA searches its nearest training sample with spatial information and then uses the corresponding projection matrix to project the test sample and all the training samples into a low-dimensional feature space. LWDA can effectively preserve the spatial-spectral local structures of the original HSI data and improve the discriminating power of the projected data for the final classification. Experimental results on two real-world HSI datasets show the effectiveness of the proposed LWDA method compared with some state-of-the-art algorithms. Especially when the data partition factor is small, i.e., 0.05, the overall accuracy obtained by LWDA increases by about 20 % for Indian Pines and 17 % for Kennedy Space Center (KSC) in comparison with the results obtained when directly using the original high-dimensional data.

1. Introduction

A hyperspectral image (HSI) is captured by an image spectrometer with hundreds of spectral bands for each image pixel, which often plays an important role in the fields of urban planning, precision agriculture, and land-cover classification [1,2,3,4,5]. Generally, the spectral bands of each pixel are considered to be the features with high dimensionality. The high dimensionality of the original HSI data significantly leads to feature redundancy problem and increases the computational complexity [6,7]. To overcome these drawbacks, it is critical to perform dimensionality reduction, which is designed to project the original high-dimensional data into a low-dimensional feature subspace while preserving some desirable information.
The existing dimensionality reduction approaches can be classified into two categories: feature selection [8,9,10,11] and feature extraction. The focus of this paper is feature extraction, which is designed to construct a low-dimensional embedding subspace and then create meaningful information by the projection of the original high-dimensional data. Then, the existing traditional classification methods (e.g., support vector machine classifier) can be directly applied to the projected data. Therefore, the HSI classification of low-dimensional data is conducive to avoiding feature redundancy and the Hughes phenomenon [12] and to reducing the computational complexity. Consequently, lots of feature extraction approaches have been presented [6,13,14,15,16,17,18,19,20]. Popular feature extraction methods include principal component analysis (PCA) [21], linear discriminant analysis (LDA) [22], locality preserving projection (LPP) [23], and modified locality preserving projection (MLPP) [24]. Compared with PCA and LPP, LDA can learn a linear transformation by simultaneously minimizing the intraclass distances and maximizing the interclass discrepancy. However, when directly applying LDA to process HSI data, it still faces several problems [25]: (1) when the dimensionality of data exceeds the size of training samples, LDA suffers from an ill-posed problem; (2) when the reduced dimensionality is less than the number of classes, LDA has an over-reducing problem; (3) LDA neglects the spatial information in the discriminant analysis; (4) LDA assumes that all the samples obey the Gaussian distribution, which is difficult for constructing the local classification boundary.
Recently, there have been many variants of LDA that try to improve the classification performance using some constraints, such as regularized local discriminant embedding (RLDE) [26], local geometric structure Fisher analysis (LGSFA) [18], and locality adaptive discriminant analysis (LADA) [27]. RLDE employs a regularized discriminant model to preserve the local structure of the HSI data. LGSFA retains the local structure among the within-class and between-class samples during the analysis process. LADA constructs a scatter matrix for each pixel with its small neighborhood, which is considered a regularization term. The above approaches can alleviate the ill-posed and over-reducing problems of the original LDA method. However, they only represent the local structure relationship of the HSI data as one-to-one. Moreover, the preservation of local structure still remains an open issue. Some new methods have been developed on the basis of graph learning. Ly et al. [28] used graph learning to construct scatter matrices, and then they conducted a discriminant analysis. Li et al. [29] proposed a two-stage framework to learn the data graph in the low-dimensional feature subspace. However, the above methods only consider the spectral information of the HSI data, which cannot accurately determine the local classification boundary.
To effectively exploit the spectral information and spatial information of the high-dimensional HSI data, a locally weighted discriminant analysis (LWDA)-based dimensionality reduction method is proposed for HSI classification in this paper. In order to apply the spatial information to the projection process, the proposed method learns the data structures adaptively during the transformation of subspace projection. Furthermore, to guarantee the spatial consistency of land cover, samples within a small neighborhood in the embedding space should be similar, which is considered a regularized constraint term during the optimization. The main contributions of this paper can be summarized as follows:
(1)
A weighted scatter matrix model is proposed by exploiting the label information and spectral information of the samples, which is able to reduce the effect of the image difference of the HSI data.
(2)
The proposed method considers the spatial consistency and the similarity relationship among the samples in a small spatial neighborhood, which is able to describe the local structure of the samples.
(3)
An optimization function is constructed on the basis of the spatial-spectral information and label information, which is able to preserve the within-class characteristics and suppress the between-class properties in the embedding feature subspace.
The remainder of this paper is organized as follows. Section 2 briefly introduces some related works, including the original LDA and MFA approaches. Section 3 provides our proposed method in detail. In Section 4.2, experimental results are presented to demonstrate the effectiveness of the proposed method compared with several state-of-the-art dimensionality reduction algorithms. Finally, a conclusion of this work is provided in Section 5.

2. Related Works

Let X = [ x 1 , x 2 , , x n ] R d × n be the original HSI data, where d is the number of spectral bands for each image pixel, i.e., the data dimensionality of the HSI data, and n represents the number of the image pixels considered as samples. The label information of the ith pixel is denoted as ( x i ) , which belongs to { 1 , 2 , , c } , and c is the number of classes. The goal of dimensionality reduction is to construct a projection matrix P R d × m , where m is the reduced dimensionality of the projected data. For the linear mapping function, the projected data is indicted as Y = P T X . Generally, the value of m is considerably smaller than d.

2.1. Linear Discriminant Analysis

Linear discriminant analysis (LDA) is a supervised method and able to compact the within-class samples and separate the between-class samples. It defines a between-class scatter matrix S b and a within-class scatter matrix S w as follows:
S b = k = 1 c n k u k u ¯ u k u ¯ T ,
S w = k = 1 c i = 1 n k x k i u k x k i u k T ,
where n k is the number of the kth class, and x k i is the ith sample from the kth class. u k is the mean of the kth class, computed by u k = i = 1 n k x k i / n k . Similarly, u ¯ is the mean of all the samples, i.e., u ¯ = k = 1 c i = 1 n k x k i / k = 1 c n k . In addition, T represents the transpose operation.
With the above definitions, LDA tries to learn the linear transformation matrix P by maximizing the ratio of the between-class scatter and the within-class scatter. The projection matrix can be obtained by the following optimization function [22]:
max P t r P T S b P P T S w P ,
where t r ( · ) represents the trace operator. The optimal projection matrix P can be obtained by analytically solving the generalized eigenvalue decomposition and then choosing the m eigenvectors that correspond to the m largest eigenvalues. Then, the m-dimensional projected data can be computed by Y = P T X .
Equations (1) and (2) reveal that the between-class scatter matrix is easily reflected by the subtraction of the total mean. Moreover, it is unable to capture the local manifold structure of the HSI data. Due to the two drawbacks, it is difficult for LDA to achieve satisfactory performance in real-world HSI applications.

2.2. Marginal Fisher Analysis

Marginal Fisher analysis (MFA) is a supervised graph learning method, which constructs an inherent graph and a penalty graph [18]. The inherent graph tries to obtain certain geometrical information of the input dataset, while the penalty graph reveals the unwanted properties of the inputs. MFA designs two weight matrices. Let W = { w i j } i , j = 1 n and W p = { w i j p } i , j = 1 n be the similarity matrix and the penalty matrix. w i j represents the similarity relationship between the two data points x i and x j , which are from the same class. On the other hand, w i j p describes the similarity characteristic between x i and x j that are from different classes. The mathematical descriptions of w i j and w i j p are defined as follows:
w i , j = 1 , i f x i N 1 ( x j ) o r x j N 1 ( x i ) a n d ( x i ) = ( x j ) , 0 , o t h e r w i s e .
w i , j p = 1 , i f x i N 2 ( x j ) o r x j N 2 ( x i ) a n d ( x i ) ( x j ) , 0 , o t h e r w i s e .
where N 1 ( x i ) and N 1 ( x j ) represent the k 1 nearest neighbors of data points x i and x j that are from the same class, and N 2 ( x i ) and N 2 ( x j ) represent the k 2 nearest neighbors of data points x i and x j that are from different classes, respectively.
With the definition of the two weight matrices, the optimization function of MFA is designed to obtain a projection matrix P as follows:
min P i = 1 n j = 1 n P T ( x i x j ) 2 2 w i j i = 1 n j = 1 n P T ( x i x j ) 2 2 w i j p ,
min P t r P T X L X T P P T X L p X T P ,
where L and L p are the Laplacian matrices, which are defined as L = D W , D = d i a g ( { j = 1 n w i j } i = 1 n ) , and L p = D p W p , D p = d i a g ( { j = 1 n w i j p } i = 1 n ) . d i a g ( · ) represents the matrix diagonal element extraction operation.
Equation (7) can be solved analytically through the generalized eigenvalue decomposition of X L X T and X L p X T . Then, the optimal projection matrix P is formed by the m eigenvectors corresponding to the m smallest eigenvalues.
MFA tries to enhance the compactness of the data points from the same class and to improve the separability of the data points from different classes in the embedding feature subspace. However, the similarity relationship between two data points in a small neighborhood is simplified as 1, which will limit it to learn a certain local manifold structure of the HSI data.

3. Proposed Method

To take advantage of the discriminant information and the spatial-spectral information of the input HSI data, a new supervised dimensionality reduction method, named locally weighted discriminant analysis (LWDA), is presented for HSI classification. LWDA constructs a weighted scatter matrix model on the basis of the within-class and between-class scatter matrices of the traditional LDA method. The weighted scatter matrix model defines a weighted within-class scatter matrix and a weighted between-class scatter matrix to improve the discriminating power. Furthermore, LWDA preserves the spatial consistency among the samples in a small spatial neighborhood. To construct the optimal low-dimensional feature subspace, the proposed method obtains the corresponding projection matrix by compacting the nature of the weighted within-class scatter matrix and the spatial consistency, and suppressing the property of the weighted between-class scatter matrix.
The flowchart of the proposed method is shown in Figure 1, where the high-dimensional HSI data are projected onto a two-dimensional subspace for visualization. Taking the classification process of the Indian Pines dataset as an example, the steps of the proposed algorithm can be summarized as follows: (1) on the basis of the training samples and the corresponding training labels, the weighted within- and between-class scatter matrices are computed; (2) with the help of the training and test samples, the spatial consistency matrix for each training sample can be computed; (3) with the foregoing weighted scatter matrices and spatial consistency matrix, the optimal projection matrix corresponding to each training sample is obtained; (4) for each test sample, the spatially closest training sample’s projection matrix can be obtained, which is used to construct the embedding features; (5) the class estimation of each test sample is obtained by a certain classifier with the training labels and the embedding features.

3.1. Weighted Scatter Matrix Model

LDA assumes all the samples possess the same contribution, i.e., the Gaussian distribution. In LDA, the within-class scatter matrix only considers the data variances of the within-class samples, while the between-class scatter matrix just considers the data variance between the mean of each individual class and the total mean. However, different within-class samples should have different contribution rates in the within-class scatter matrix. Moreover, the properties of any two different individual class means may be different in the between-class scatter matrix. To better represent the similarity characteristic of the within-class samples and different individual class means, the proposed method constructs two weighted scatter matrices, i.e., the weighted within-class scatter matrix and the weighted between-class scatter matrix, which are defined as follows:
S ˜ w = k = 1 c i = 1 n k j = 1 n k x k i u k g i , j k x k j u k T ,
S ˜ b = i = 1 c j = 1 c n i u i u j h i , j u i u j T ,
where g i , j k is the similarity weight between the samples x k i and x k j , and h i , j is the similarity weight between the one-class means u i and u j . The similarity weights are represented as
g i , j k = e x p x k i x k j 2 2 ρ k i 2 + ε ,
h i , j = e x p u i u j 2 2 σ i 2 + ε ,
where ρ k i = j = 1 n k x k i x k j / n k , σ i = j = 1 c u i u j / c , and ε is a small value for avoiding zero in the denominator.
Similar to LDA, the optimization function is designed to improve the aggregation of the within-class samples and enhance the diversity of the between-class samples in a low-dimensional feature subspace. So, the optimal projection matrix can be obtained by the following formula:
min P t r P T S ˜ w P P T S ˜ b P .
Supposing the minimum value of the above function is α , the optimal P should make the value of t r P T S ˜ w P α t r P T S ˜ b P close to 0. Thus, Equation (12) is equivalent to
min P t r P T S ˜ w P α t r P T S ˜ b P .

3.2. Spatial Consistency Matrix

For real-world HSI data, the data points within a small spatial region are often highly correlated and are classified as the same class [25]. Hence, spatial consistency is essential for an accurate classification. Given a data point x i R d × 1 ( i = 1 , , n ) , the spatial surroundings are found within a search region with a size of r × r , where r must be odd. Therefore, the r 2 1 neighbors are selected for each sample, which are denoted as Z i = z i 1 , z i 2 , , z i r 2 1 and z i j X ( j = 1 , , r 2 1 ) . For different samples x i and x j , the subsets Z i and Z j may partially overlap. In a desired feature subspace, these neighbors are encouraged to be close to each other. The problem of spatial consistency can be defined as
min P j , k = 1 r 2 1 P T z i j z i k 2 2 , x i , i = 1 , 2 , , n .
The spatial consistency matrix is defined as
S ˜ z = j , k = 1 r 2 1 z i j z i k z i j z i k T .
Then, Equation (14) can be further reduced to
min P t r P T S ˜ z P .

3.3. Optimization Algorithm

Integrating Equations (13) and (16) together, the objective function of the proposed LWDA method is summarized as
min P i t r P i T S ˜ w P i α t r P i T S ˜ b P i + β t r P i T S ˜ z P i ,
min P i t r P i T S ˜ w α S ˜ b + β S ˜ z P i ,
where α and β are parameters, and P i is the desired projection matrix for the sample x i ( i = 1 , 2 , , n ) . With the proposed objective function, the spatial consistency between the data points is captured, and the local data relationship is also investigated during the discriminant analysis. The optimal P i for Equation (18) can be obtained by the m ( m d ) eigenvectors of the critical matrix S ˜ w α S ˜ b + β S ˜ z corresponding to the m smallest eigenvalues.
With a certain dataset partition factor τ ( 0 < τ < 1 ) , the input HSI dataset can be divided into the training subset X s and the test subset X t . That is to say, for the kth class, the number of the randomly selected samples for the training subset is n s , k = n k τ , while the number of the samples chosen for the test subset is n t , k = n k n s , k , where n k is the total number of samples belonging to the kth class. For simplicity, the training and test subsets are denoted as X s = x s 1 , x s 2 , , x s n s and X t = x t 1 , x t 2 , , x t n t . The label information of the training subset is marked as Y s = y s 1 , y s 2 , , y s n s , where y s i [ 1 , 2 , , c ] , i = 1 , 2 , , n s . The details of the whole framework are described in Algorithm 1.
Algorithm 1: Locally weighted discriminant analysis (LWDA).
  Input: Training dataset X s , training class information set Y s , test dataset X t , parameters α and β , dimensionality of desired projection matrix m.
  Output: Estimate the test class information set Y t .
  Training:
      1. Compute the weighted within-class scatter matrix S ˜ w according to Equation (8);
      2. Compute the weighted between-class scatter matrix S ˜ b according to Equation (9);
  for i = 1 , 2 , , n s do
          3. Construct the neighbor set Z i and then compute the spatial consistency matrix S ˜ z according to Equation (15);
          4. Compute the total matrix S ˜ = S ˜ w α S ˜ b + β S ˜ z ;
          5. Obtain the optimal projection matrix P i by choosing the m eigenvectors of S ˜ corresponding to the m smallest eigenvalues; then, set i = i + 1 ;
  end
  Testing:
  for i = 1 , 2 , , n t do
          6. For each test sample x t i , find the spatially nearest training sample x s i and the corresponding optimal projection matrix, denoted as P ˜ i = P i , where i = 1 , , n s ;
          7. Low-dimensional embedding features are computed as X s , m = P ˜ i T X s , and x t , m i = P ˜ i T x t i .
          8. Using the nearest-neighbor classifier, find the serial number of the nearest training sample in the low-dimensional feature subspace, computed as j = a r g min j x t , m i x s , m j 2 2 , where j = 1 , , n s ;
          9. Obtain the corresponding class information, i.e., y t i = y s j ; then, set i = i + 1 .
  end (

4. Experimental Results

4.1. Experimental Setting

In the experiments, two real-world hyperspectral image datasets were employed, i.e., Indian Pines and Kennedy Space Center (KSC) datasets [30]. The Indian Pines dataset contains 10,249 data points from 16 classes. Each data point (pixel) has 200 spectral bands. The KSC dataset annotates 5211 valid pixels (excluding the background pixels with the class information of 0) from 13 classes. Each pixel has 176 spectral bands. The two HSI datasets were both captured by an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor. Table 1 shows various land-cover types and the corresponding number of samples for the two aforementioned HSI datasets.
To investigate the classification performance, each dataset was randomly divided into the training and test samples with a data partition factor τ . For instance, the total number of samples for the “Corn-notill” land-cover type is 1428, shown in Table 1. Setting τ to 0.05 , the number of samples chosen for training is 1428 × 0.05 = 72 , while the remaining 1356 samples were used for testing.
For a quantitative comparison, seven dimensionality reduction algorithms were taken as competitors, including the original raw spectral feature (RAW), principal component analysis (PCA) [21], linear discriminant analysis (LDA) [22], discrimination-information-based locality preserving projection (DLPP) [24], marginal Fisher analysis (MFA) and local geometric structure Fisher analysis (LGSFA) [18], two-stage subspace projection (TwoSP) [29], as well as discriminant analysis with graph learning (DAGL) [25]. In LWDA, the value of parameter α is set to 10 3 .
According to different dimensionality reduction algorithms, the projection matrix can be obtained to achieve the low-dimensional embedding features of training and test samples. After that, a certain classifier, e.g., nearest neighbor (NN) and support vector machine (SVM) [31], is exploited to discriminate the land-cover types of the test samples with the help of the class information of training samples. In this study, three widely used classification measurements, i.e., average classification accuracy of all the classes (AA), overall classification accuracy (OA), and kappa coefficient (KC), were used to evaluate the objective results of each method. To alleviate the random error caused by the partition of training and test samples, each experiment was repeated five times in each condition; reported are the average AAs, the average OAs, the average KCs, and their standard deviations (STDs). All the experiments were performed on a personal computer with Intel Xeon CPU E5-2643 v3, 3.40 GHz, 64 GB memory, and 64-bit Windows 7 using Matlab R2017b.

4.2. Performance on Hyperspectral Image Datasets

The quantitative results of the proposed method and the baselines are given in Table 2 and Table 3 for the Indian Pines dataset and KSC dataset, respectively. The two tables show the average classification accuracy of each class, the AAs, OAs, and KCs, as well as their STDs, which were obtained by repeating each experiment five times. All the values are represented in terms of percentage.
Table 2 and Table 3 demonstrate that the proposed LWDA method achieves better classification results in most classes compared with other methods, and it outperforms all the competitors in terms of AAs, OAs, and KCs. PCA neglects the nonlinear relationship from the original high-dimensional feature data, although it achieves dimensionality reduction. LDA and DLPP preserve the local manifold structure by exploiting the discrimination information of the training samples. However, they neglect the global structure in the dimensionality reduction. Since MFA simply considers the similarity relationship between two samples in a small neighborhood as one, it is difficult to learn an accurate local manifold structure. LGSFA can retain the local structure among the within- and between-class samples during the discriminant analysis. TwoSP preserves the global structure in the first-stage subspace projection, and it investigates the local structure of the HSI data adaptively. DAGL combines the spatial neighborhood information and data graph in the discriminant analysis process. However, the data graph is constructed in the original high-dimensional space, which still introduces data noise into the final projection process. The proposed LWDA method exploits the spectral and spatial information of the HSI data, and it then enhances the spatial consistency during the discriminant analysis. Therefore, LWDA produces the best classification performance on the two HSI datasets.
Moreover, the classification maps of the aforementioned methods on the Indian Pines and KSC datasets are shown in Figure 2 and Figure 4. For better visualization, the local magnifications with a certain magnification factor are displayed in Figure 3 and Figure 5. From Figure 2, Figure 3, Figure 4 and Figure 5, the proposed LWDA method generates smoother classification maps and poses more homogeneous areas. LWDA enforces the spectral-spatial information during the discriminant analysis. It not only preserves the global structure but also the local manifold structure with the help of the proposed weighted scatter matrix model and the construction of spatial consistency.
Furthermore, a McNemar test [32,33] was conducted by pairwise comparison to validate the effectiveness of different methods. In the McNemar test, the threshold of significance is set to 0.05 . The results of the McNemar test are shown in Table 4, where the methods in the horizontal direction are denoted as the test methods while those in the vertical direction are marked as the reference methods. When the value is smaller than zero, it indicates that the classification performance of the test method is better than that of the reference method; otherwise, the reference method outperforms the test method. Moreover, when the absolute value is larger than 1.96 , the two methods have obvious differences. Compared with all the competitors, the absolute values in the LWDA column are larger than 10, which demonstrates that LWDA has a distinct advantage.

4.3. Analysis of Computational Cost

T1 is denoted as the running time of the construction of projection matrix. T2 and T3 are defined as the classification times obtained by the NN and SVM classifiers in the testing process.
Table 5 shows T1, T2, and T3 in terms of seconds of the different methods. The low dimensionality of the embedding features can reduce the running time in the classification process. To obtain the optimal projection matrix, the proposed LWDA method estimates the class information for each test sample. In addition, LWDA takes the most running time to extract the spectral-spatial information in the discriminant analysis for each training sample. So, the computation of dimensionality reduction in LWDA is larger than the others, excluding TwoSP because it involves a large kernel matrix computation. Compared with T3, the running time using the NN classifier is smaller, which also illustrates that the NN classifier for the classification process has a better advantage. LWDA with the SVM classifier needs to construct an SVM model for each test sample, so T3 of LWDA is much larger than the baselines. Therefore, the NN classifier was applied in the experiments.

4.4. Analysis of Reduced Dimensionality

The optimal reduced dimensionality of each method is discussed in this section. Figure 6 shows the curves of OAs varying with different dimensionalities of the projection matrix, from 2 to 50, for the Indian Pines and KSC datasets.
Figure 6 demonstrates that the proposed LWDA method achieves the highest OAs constantly. In particular, LWDA exceeds the baselines to a large extent when the dimensionality is less than 5. Furthermore, the classification performance of all the methods becomes stable or decreases when the dimensionality increases to a certain value, which also indicates that a low-dimensional feature subspace is sufficient for preserving the valuable information of HSI data.

4.5. Analysis of Classifier

To evaluate the classification performance of each method with two different classifiers, i.e., NN and SVM, the experiments were repeated five times. For SVM, the LibSVM Toolbox in a MATLAB version was applied with a radial basis function (RBF) kernel [34]. Once the projected features were obtained by each method, the NN and SVM classifiers were applied for the classification process, respectively. Figure 7 shows the classification results obtained by different methods with NN and SVM classifiers on two HSI datasets.
Figure 7 illustrates that the proposed LWDA method with the NN classifier presents the best classification performance compared with the other dimensionality reduction methods. For most cases on the two dataset, the results with the NN classifier are superior to those with SVM. To unify the classifier in the classification process, the NN classifier was used in all the experiments.

4.6. Analysis of Parameters

The proposed LWDA method has two trade-off parameters r and β . The value of r affects the number of neighbors in the spatial space, while the value of β balances the contribution between the weighted scatter matrix model and the spatial consistency matrix. r was tuned with the set { 3 , 5 , 7 , 9 , 11 , 13 , 15 , 17 , 19 , 21 , 23 , 25 , 27 } , and β was varied with the set { 0 , 0.001 , 0.005 , 0.01 , 0.02 , 0.04 , 0.05 , 0.06 , 0.08 , 0.1 , 0.5 , 1 , 5 , 10 , 50 , 100 } . Figure 8 shows the average OAs with respect to the values of parameters r and β .
According to Figure 8, when the value of β increases, the OAs display a subtle change with a fixed r. That is because the spatial consistency matrix generates a similar contribution for the Indian Pines and KSC datasets. An increased r leads to introducing more between-class samples in the construction of a spatial consistency matrix. For the Indian Pines dataset, a peak value is generated in the curved surface map when the value of r reaches 11, and then the OAs begin to slowly decrease when the value of r continues to increase. If r is a large value, the contribution of the preservation of the local manifold structure will be reduced. Similarly, a peak value of the curved surface is obtained when r increases to 25 for the KSC dataset. Therefore, the parameters r and β were set to 11 and 0.05 for the Indian Pines dataset and 25 and 0.04 for the KSC dataset in the experiments.
The proposed method can be divided into two versions: online and offline. The online version of LWDA needs to construct the optimal projection matrix for each test sample, which leads to a large computational cost. To reduce the computational time, the experiments in this paper used the offline version. Figure 9 shows the histograms of the spatial distance between the input test sample and its nearest training sample. The two histograms for the Indian Pines and KSC datasets illustrate that the spatial distance is mainly distributed in the range of [ 1 , 5 ] . The samples in a small neighborhood should be close to each other in the desired feature subspace. Furthermore, the experimental results shown in the Section 4.2 demonstrate that the offline version of the proposed method achieves better classification performance than the existing approaches.

4.7. Analysis of Data Partition Factor τ

The data partition factor τ affects the number of selected samples in the training process. The influence of different values of τ was investigated, and results are presented in this section. The value of τ was tuned with the set { 0.05 , 0.06 , 0.07 , 0.08 , 0.09 , 0.1 } . Table 6 and Table 7 show the classification results in terms of average OAs, average KCs, and their STDs with the Indian Pines and KSC datasets.
In Table 6 and Table 7, the OAs and KCs improve with the increase in the data partition factor for all the methods on the two HSI datasets. It implies that a large number of training samples contain more valuable information in the feature representation. DLPP achieves better classification results than PCA and LDA, because DLPP applies the discrimination information to preserve the local structure. TwoSP shows better OAs and KCs than MFA and LGSFA in most conditions, since TwoSP effectively alleviates the nonlinear problem in HSI data and simultaneously preserves the global and local structures. In all experiments, the proposed LWDA method achieves the best classification results under different values of τ , especially with a small τ value, which indicates that a small number of training samples is enough for a good performance. Moreover, LWDA applies the spectral-spatial information to the discriminant analysis process, which largely enhances the discriminating power of low-dimensional embedding features.

5. Conclusions

In this paper, a new supervised dimensionality reduction method, named LWDA, is proposed on the basis of the spectral-spatial information of HSI data. During the discriminant analysis, LWDA uses the proposed weighted scatter matrix model and computes the spatial consistency matrix for each data sample, which can adaptively learn local manifold structures of the original HSI data. In addition, LWDA preserves the within-class properties and suppresses the between-class characteristics in an optimal low-dimensional feature subspace.
Through the experiments on two real-world HSI datasets, i.e., Indian Pines and KSC, LWDA achieves better classification performance than the existing dimensionality reduction approaches. In particular, a small data portion of the training set is enough for a satisfactory classification performance. The overall accuracy obtained by LWDA increases by at least 17 % in comparison with RAW when the data partition factor is 0.05. In addition, the McNemar test demonstrates that LWDA has statistical significance when compared with the baselines. For LWDA, the absolute value of the McNemar test is at least 10 > 1.96 . LWDA learns similarity relationships of the within-class samples and the means of different classes, as well as creates more available information in the subsequent classification. Hence, LWDA achieves the qualitative and quantitative results in the experiments.
Our future work will focus on how to extend the online version of the proposed method to quickly represent the spectral-spatial information and improve the computational efficiency.

Author Contributions

All the authors designed the study and discussed the basic structure of the manuscript. X.L. carried out the experiments and finished the first version of this manuscript. L.Z. provided the framework of the experiments. J.Y. reviewed and edited this paper.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grants 61601416, 61771349, 61711530239 and 41701417, the Fundamental Research Funds for the Central Universities, China University of Geosciences, Wuhan, under Grant CUG170612, the Hong Kong Scholars Program under Grant XJ2017030, and the Hong Kong Polytheistic University Research Fund.

Conflicts of Interest

All the authors declare no conflict of interest.

References

  1. Gevaert, C.M.; Suomalainen, J.; Tang, J.; Kooistra, L. Generation of Spectral-Temporal Response Surfaces by Combining Multispectral Satellite and Hyperspectral UAV Imagery for Precision Agriculture Applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3140–3146. [Google Scholar] [CrossRef]
  2. Luo, F.; Du, B.; Zhang, L.; Zhang, L.; Tao, D. Feature Learning Using Spatial-Spectral Hypergraph Discriminant Analysis for Hyperspectral Image. IEEE Trans. Cybern. 2018, 1–14. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, X.; Sun, Y.; Zhang, J.; Wu, P.; Jiao, L. Hyperspectral Unmixing via Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1–5. [Google Scholar] [CrossRef]
  4. Liu, W.; Ma, X.; Zhou, Y.; Tao, D.; Cheng, J. p-Laplacian Regularization for Scene Recognition. IEEE Trans. Cybern. 2018, 1–14. [Google Scholar] [CrossRef]
  5. Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene Classification with Recurrent Attention of VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2018, 1–13. [Google Scholar] [CrossRef]
  6. Liu, Y.; Zhong, Y.; Qin, Q. Scene Classification Based on Multiscale Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 7109–7121. [Google Scholar] [CrossRef]
  7. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A Superpixelwise PCA Approach for Unsupervised Feature Extraction of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef]
  8. Wang, Q.; Zhang, F.; Li, X. Optimal Clustering Framework for Hyperspectral Band Selection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5910–5922. [Google Scholar] [CrossRef]
  9. Liu, W.; Yang, X.; Tao, D.; Cheng, J.; Tang, Y. Multiview dimension reduction via Hessian multiset canonical correlations. Inf. Fusion 2018, 41, 119–128. [Google Scholar] [CrossRef]
  10. Wang, Q.; Qin, Z.; Nie, F.; Li, X. Spectral Embedded Adaptive Neighbors Clustering. IEEE Trans. Neural Netw. Learn. Syst. 2018, 1–7. [Google Scholar] [CrossRef]
  11. Wang, Q.; Chen, M.; Nie, F.; Li, X. Detecting Coherent Groups in Crowd Scenes by Multiview Clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 1. [Google Scholar] [CrossRef] [PubMed]
  12. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef]
  13. Zhang, L.; Zhang, L.; Tao, D.; Huang, X.; Du, B. Hyperspectral Remote Sensing Image Subpixel Target Detection Based on Supervised Metric Learning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4955–4965. [Google Scholar] [CrossRef]
  14. Luo, Y.; Wen, Y.; Tao, D.; Gui, J.; Xu, C. Large Margin Multi-Modal Multi-Task Feature Extraction for Image Classification. IEEE Trans. Image Process. 2016, 25, 414–427. [Google Scholar]
  15. Zhang, L.; Zhang, Q.; Du, B.; Huang, X.; Tang, Y.Y.; Tao, D. Simultaneous Spectral-Spatial Feature Selection and Extraction for Hyperspectral Images. IEEE Trans. Cybern. 2018, 48, 16–28. [Google Scholar] [CrossRef] [PubMed]
  16. Tao, D.; Tang, X.; Li, X. Which Components Are Important For Interactive Image Searching? IEEE Trans. Circuits Syst. Video Technol. 2008, 18, 3–11. [Google Scholar]
  17. Sun, W.; Tian, L.; Xu, Y.; Du, B.; Du, Q. A Randomized Subspace Learning Based Anomaly Detector for Hyperspectral Imagery. Remote Sens. 2018, 10, 417. [Google Scholar] [CrossRef]
  18. Luo, F.; Huang, H.; Duan, Y.; Liu, J.; Liao, Y. Local Geometric Structure Feature for Dimensionality Reduction of Hyperspectral Imagery. Remote Sens. 2017, 9, 790. [Google Scholar] [CrossRef]
  19. Li, X.; He, H.; Wang, R.; Tao, D. Single Image Superresolution via Directional Group Sparsity and Directional Features. IEEE Trans. Image Process. 2015, 24, 2874–2888. [Google Scholar] [CrossRef]
  20. Su, H.; Zhao, B.; Du, Q.; Du, P. Kernel Equation Collaborative Representation With Local Correlation Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 1–12. [Google Scholar] [CrossRef]
  21. Prasad, S.; Bruce, L.M. Limitations of Principal Components Analysis for Hyperspectral Target Recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  22. Wang, H.; Lu, X.; Hu, Z.; Zheng, W. Fisher Discriminant Analysis With L1-Norm. IEEE Trans. Cybern. 2014, 44, 828–842. [Google Scholar] [CrossRef] [PubMed]
  23. Wang, Z.; He, B. Locality perserving projections algorithm for hyperspectral image dimensionality reduction. In Proceedings of the 2011 19th International Conference on Geoinformatics, Shanghai, China, 24–26 June 2011; pp. 1–4. [Google Scholar]
  24. Li, X.; He, H.; Yin, Z.; Chen, F.; Cheng, J. Single image super-resolution via subspace projection and neighbor embedding. Neurocomputing 2014, 139, 310–320. [Google Scholar] [CrossRef]
  25. Chen, M.; Wang, Q.; Li, X. Discriminant Analysis with Graph Learning for Hyperspectral Image Classification. Remote Sens. 2018, 10, 836. [Google Scholar] [CrossRef]
  26. Zhou, Y.; Peng, J.; Chen, C.L.P. Dimension Reduction Using Spatial and Spectral Regularized Local Discriminant Embedding for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1082–1095. [Google Scholar] [CrossRef]
  27. Wang, Q.; Meng, Z.; Li, X. Locality Adaptive Discriminant Analysis for Spectral-Spatial Classification of Hyperspectral Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2077–2081. [Google Scholar] [CrossRef]
  28. Ly, N.H.; Du, Q.; Fowler, J.E. Collaborative Graph-Based Discriminant Analysis for Hyperspectral Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2688–2696. [Google Scholar] [CrossRef]
  29. Li, X.; Zhang, L.; You, J. Hyperspectral Image Classification Based on Two-stage Subspace Projection. Remote Sens. 2018, 10, 1565. [Google Scholar] [CrossRef]
  30. Hyperspectral Remote Sensing Scenes. Available online: http://www.ehu.eus/ccwintco/index.php?title= Hyperspectral_Remote_Sensing_Scenes (accessed on 31 July 2014).
  31. Dong, A.; Chung, F.; Deng, Z.; Wang, S. Semi-Supervised SVM with Extended Hidden Features. IEEE Trans. Cybern. 2016, 46, 2924–2937. [Google Scholar] [CrossRef] [PubMed]
  32. Eyden, R.J.V.; Wit, P.W.C.D.; Arron, J.C. Predicting company failure-a comparison between neural networks and established statistical techniques by applying the McNemar test. In Proceedings of the 1995 Conference on Computational Intelligence for Financial Engineering (CIFEr), New York, NY, USA, 9–11 April 1995; pp. 91–96. [Google Scholar]
  33. Foody, G.M. Latent Class Modeling for Site- and Non-Site-Specific Classification Accuracy Assessment Without Ground Data. IEEE Trans. Geosci. Remote Sens. 2012, 50, 2827–2838. [Google Scholar] [CrossRef]
  34. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed LWDA method. With a partition factor τ , the training and test sample set (shown in two-dimensional space), as well as the training label set, are obtained. Then, the dataset is used to construct the weighted scatter matrix model, spatial consistency matrix, and optimal projection matrix for each training sample. The next step is to find the spatially nearest training sample’s projection matrix, which is applied to construct the embedding features of training samples and input test sample. Once all the predicted test labels are obtained, the classification map (including the training labels), is generated by exploiting a fixed classifier.
Figure 1. Flowchart of the proposed LWDA method. With a partition factor τ , the training and test sample set (shown in two-dimensional space), as well as the training label set, are obtained. Then, the dataset is used to construct the weighted scatter matrix model, spatial consistency matrix, and optimal projection matrix for each training sample. The next step is to find the spatially nearest training sample’s projection matrix, which is applied to construct the embedding features of training samples and input test sample. Once all the predicted test labels are obtained, the classification map (including the training labels), is generated by exploiting a fixed classifier.
Remotesensing 11 00109 g001
Figure 2. Classification maps of different dimensionality reduction methods with NN classifier on the Indian Pines dataset ( τ = 0.05 ). (a) Ground truth; (b) original raw spectral feature (RAW); (c) principal component analysis (PCA); (d) linear discriminant analysis (LDA); (e) discrimination- information-based locality preserving projection (DLPP); (f) marginal Fisher analysis (MFA); (g) local geometric structure Fisher analysis (LGSFA); (h) two-stage subspace projection (TwoSP); (i) discriminant analysis with graph learning (DAGL); (j) proposed locally weighted discriminant analysis (LWDA).
Figure 2. Classification maps of different dimensionality reduction methods with NN classifier on the Indian Pines dataset ( τ = 0.05 ). (a) Ground truth; (b) original raw spectral feature (RAW); (c) principal component analysis (PCA); (d) linear discriminant analysis (LDA); (e) discrimination- information-based locality preserving projection (DLPP); (f) marginal Fisher analysis (MFA); (g) local geometric structure Fisher analysis (LGSFA); (h) two-stage subspace projection (TwoSP); (i) discriminant analysis with graph learning (DAGL); (j) proposed locally weighted discriminant analysis (LWDA).
Remotesensing 11 00109 g002
Figure 3. Local magnification (with a magnification factor of 3) of the output classification maps shown in Figure 2. (a) Selected local region; (b) magnified local region; (c) PCA; (d) LDA; (e) DLPP; (f) MFA; (g) LGSFA; (h) TwoSP; (i) DAGL; (j) proposed LWDA.
Figure 3. Local magnification (with a magnification factor of 3) of the output classification maps shown in Figure 2. (a) Selected local region; (b) magnified local region; (c) PCA; (d) LDA; (e) DLPP; (f) MFA; (g) LGSFA; (h) TwoSP; (i) DAGL; (j) proposed LWDA.
Remotesensing 11 00109 g003
Figure 4. Classification maps of different dimensionality reduction methods with NN classifier on the KSC dataset ( τ = 0.05 ). (a) Ground truth; (b) PCA; (c) LDA; (d) DLPP; (e) MFA; (f) LGSFA; (g) TwoSP; (h) DAGL; (i) proposed LWDA.
Figure 4. Classification maps of different dimensionality reduction methods with NN classifier on the KSC dataset ( τ = 0.05 ). (a) Ground truth; (b) PCA; (c) LDA; (d) DLPP; (e) MFA; (f) LGSFA; (g) TwoSP; (h) DAGL; (i) proposed LWDA.
Remotesensing 11 00109 g004
Figure 5. Local magnification (with a magnification factor of 4) of the output classification maps shown in Figure 4. (a) Selected local region; (b) magnified local region; (c) LDA; (d) DLPP; (e) MFA; (f) LGSFA; (g) TwoSP; (h) DAGL; (i) proposed LWDA.
Figure 5. Local magnification (with a magnification factor of 4) of the output classification maps shown in Figure 4. (a) Selected local region; (b) magnified local region; (c) LDA; (d) DLPP; (e) MFA; (f) LGSFA; (g) TwoSP; (h) DAGL; (i) proposed LWDA.
Remotesensing 11 00109 g005
Figure 6. Overall classification accuracy (OA) versus the reduced dimensionality of different methods with NN classifier on the (a) Indian Pines and (b) KSC datasets using τ = 0.05 .
Figure 6. Overall classification accuracy (OA) versus the reduced dimensionality of different methods with NN classifier on the (a) Indian Pines and (b) KSC datasets using τ = 0.05 .
Remotesensing 11 00109 g006
Figure 7. OAs obtained by different methods with the NN and support vector machine (SVM) classifiers on the (a) Indian Pines and (b) KSC datasets using τ = 0.05 .
Figure 7. OAs obtained by different methods with the NN and support vector machine (SVM) classifiers on the (a) Indian Pines and (b) KSC datasets using τ = 0.05 .
Remotesensing 11 00109 g007
Figure 8. OAs versus the value of parameters r and β in the proposed LWDA method with NN classifier on the (a) Indian Pines and (b) KSC datasets using τ = 0.05 .
Figure 8. OAs versus the value of parameters r and β in the proposed LWDA method with NN classifier on the (a) Indian Pines and (b) KSC datasets using τ = 0.05 .
Remotesensing 11 00109 g008
Figure 9. Histogram of spatial distance for the (a) Indian Pines and (b) KSC datasets.
Figure 9. Histogram of spatial distance for the (a) Indian Pines and (b) KSC datasets.
Remotesensing 11 00109 g009
Table 1. Number of total, training, and test samples, with a partition factor τ = 0.05 of each land-cover class for the Indian Pines and Kennedy Space Center (KSC) datasets.
Table 1. Number of total, training, and test samples, with a partition factor τ = 0.05 of each land-cover class for the Indian Pines and Kennedy Space Center (KSC) datasets.
Indian PinesKSC
Class No.Land CoverSamplesTrainingTestClass No.Land CoverSamplesTrainingTest
1Alfalfa463431Scurb76139722
2Corn-notill14287213562Willow-swamp24313230
3Corn-mintill830427883Cabbage-palm-hammock25613243
4Corn237122254Cabbage-palm/oak-hammock25213239
5Grass-pasture483254585Slash-pine1619152
6Grass-tree730376936Oak/broadleaf-hammock22912217
7Grass-pasture-mowed282267Hardwood-swamp105699
8Hay-windrowed478244548Graminoid-marsh43122409
9Oats201199Spartina-marsh52026494
10Soybeans-notill9724992310Cattail-marsh40421383
11Soybeans-mintill2455123233211Salt-marsh41921398
12Soybeans-clean5933056312Mud-flats50326477
13Wheat2051119413Water92747880
14Woods1265641201
15Bldg-grass-tree-drives38620366
16Stone-steel-towers93588
Table 2. Classification results (%) of each class ( τ = 0.05 ) with nearest-neighbor (NN) classifier on the Indian Pines dataset.
Table 2. Classification results (%) of each class ( τ = 0.05 ) with nearest-neighbor (NN) classifier on the Indian Pines dataset.
Class No.RAWPCALDADLPPMFALGSFATwoSPDAGLLWDA
140.0 ± 12.642.3 ± 20.642.8 ± 17.849.3 ± 15.750.2 ± 11.732.1 ± 12.831.6 ± 12.288.4 ± 10.590.7 ± 8.4
248.1 ± 1.249.3 ± 2.057.0 ± 2.159.8 ± 3.159.1 ± 2.661.3 ± 2.467.0 ± 1.775.2 ± 2.681.2 ± 2.8
344.2 ± 2.444.0 ± 2.342.6 ± 6.648.6 ± 5.444.8 ± 3.549.4 ± 3.756.1 ± 3.273.2 ± 5.977.2 ± 7.8
430.2 ± 6.328.2 ± 6.927.9 ± 3.231.8 ± 5.329.6 ± 4.129.3 ± 4.940.1 ± 4.875.1 ± 6.276.6 ± 10.2
576.3 ± 5.175.9 ± 3.683.6 ± 5.583.4 ± 3.383.6 ± 3.284.5 ± 2.684.1 ± 4.484.5 ± 2.386.7 ± 1.2
692.2 ± 1.992.4 ± 2.291.2 ± 2.190.1 ± 2.592.9 ± 0.991.2 ± 2.392.3 ± 1.784.6 ± 2.287.0 ± 2.6
780.0 ± 10.079.2 ± 10.476.2 ± 14.785.4 ± 9.285.4 ± 5.053.8 ± 28.482.3 ± 6.43.8 ± 15.924.4 ± 17.8
894.5 ± 2.394.9 ± 2.596.0 ± 1.493.7 ± 2.893.9 ± 3.296.7 ± 1.789.3 ± 5.898.5 ± 1.399.0 ± 0.6
914.7 ± 10.814.7 ± 10.812.6 ± 9.624.2 ± 13.220.0 ± 11.417.9 ± 11.530.5 ± 9.41.5 ± 9.85.3 ± 9.1
1061.3 ± 6.562.2 ± 6.746.8 ± 3.857.4 ± 2.954.7 ± 6.352.9 ± 3.169.1 ± 1.465.0 ± 8.174.8 ± 7.7
1167.4 ± 3.467.4 ± 2.164.8 ± 1.763.1 ± 1.171.1 ± 4.969.2 ± 1.874.8 ± 2.085.9 ± 1.788.4 ± 1.4
1235.0 ± 2.634.6 ± 3.746.8 ± 3.948.4 ± 4.943.9 ± 4.849.8 ± 4.656.0 ± 6.376.2 ± 5.576.0 ± 6.8
1393.3 ± 1.593.2 ± 1.792.6 ± 3.793.7 ± 5.095.2 ± 3.597.3 ± 1.396.4 ± 1.393.3 ± 2.081.3 ± 11.9
1490.3 ± 2.989.3 ± 3.893.4 ± 1.694.0 ± 2.094.0 ± 1.995.0 ± 1.295.2 ± 2.196.2 ± 1.397.7 ± 0.7
1527.2 ± 2.027.3 ± 2.344.6 ± 8.144.5 ± 7.743.3 ± 6.948.4 ± 8.046.7 ± 6.186.9 ± 7.392.7 ± 1.1
1686.4 ± 3.086.4 ± 3.179.8 ± 8.382.1 ± 6.081.1 ± 4.484.3 ± 4.986.4 ± 3.245.5 ± 25.159.8 ± 22.0
AA61.3 ± 1.461.3 ± 2.062.4 ± 2.765.6 ± 0.865.2 ± 1.063.3 ± 2.968.6 ± 0.970.9 ± 2.974.9 ± 1.8
OA64.8 ± 1.064.9 ± 0.865.8 ± 1.767.5 ± 0.968.6 ± 1.469.3 ± 1.173.9 ± 1.281.7 ± 1.285.1 ± 1.0
KC59.7 ± 1.159.8 ± 0.860.8 ± 2.162.9 ± 1.063.9 ± 1.564.7 ± 1.370.1 ± 1.479.2 ± 1.583.0 ± 1.1
Table 3. Classification results (%) of each class ( τ = 0.05 ) with NN classifier on the KSC dataset.
Table 3. Classification results (%) of each class ( τ = 0.05 ) with NN classifier on the KSC dataset.
Class No.RAWPCALDADLPPMFALGSFATwoSPDAGLLWDA
187.7 ± 3.287.6 ± 3.286.1 ± 4.282.4 ± 1.891.3 ± 2.085.7 ± 2.390.6 ± 2.394.3 ± 1.198.6 ± 1.3
275.7 ± 13.375.7 ± 13.286.4 ± 1.987.8 ± 2.488.7 ± 2.890.0 ± 3.177.4 ± 2.085.2 ± 5.884.0 ± 12.7
366.7 ± 3.766.8 ± 3.853.4 ± 7.155.3 ± 7.760.2 ± 6.158.4 ± 4.777.4 ± 3.578.2 ± 4.293.2 ± 5.1
450.0 ± 3.349.9 ± 3.039.9 ± 4.539.3 ± 5.145.0 ± 9.547.0 ± 4.664.0 ± 4.276.6 ± 5.185.1 ± 7.0
545.8 ± 13.145.4 ± 12.950.4 ± 4.751.3 ± 1.948.2 ± 4.246.2 ± 5.073.0 ± 1.888.8 ± 3.288.9 ± 8.2
634.2 ± 5.234.5 ± 5.150.6 ± 8.252.9 ± 8.834.6 ± 7.143.3 ± 8.759.0 ± 7.575.1 ± 8.4100.0 ± 0.0
764.0 ± 11.263.0 ± 10.353.7 ± 13.047.3 ± 9.252.7 ± 11.544.2 ± 15.278.8 ± 9.598.0 ± 2.9100.0 ± 0.0
869.4 ± 6.869.0 ± 7.478.4 ± 4.477.7 ± 3.881.6 ± 6.282.7 ± 5.183.1 ± 3.591.7 ± 5.893.3 ± 6.4
988.5 ± 4.088.5 ± 4.082.3 ± 3.280.7 ± 2.387.7 ± 5.184.3 ± 4.493.5 ± 2.697.8 ± 3.299.1 ± 1.9
1081.1 ± 2.481.2 ± 2.594.8 ± 0.891.6 ± 1.694.2 ± 2.594.9 ± 1.682.5 ± 1.796.9 ± 0.5100.0 ± 0.0
1192.8 ± 1.692.8 ± 1.687.6 ± 2.687.5 ± 3.188.7 ± 4.684.9 ± 5.585.2 ± 3.096.2 ± 1.399.6 ± 0.5
1278.3 ± 4.778.2 ± 4.689.1 ± 2.688.3 ± 4.182.2 ± 3.983.7 ± 2.780.9 ± 4.292.7 ± 2.798.0 ± 1.5
1398.4 ± 0.998.4 ± 0.999.6 ± 0.498.5 ± 0.998.2 ± 0.598.8 ± 0.798.5 ± 0.998.1 ± 0.2100.0 ± 0.0
AA71.7 ± 1.871.6 ± 1.873.3 ± 1.172.4 ± 1.373.3 ± 1.072.6 ± 1.680.3 ± 0.990.0 ± 0.795.4 ± 0.8
OA79.6 ± 0.679.5 ± 0.681.4 ± 0.680.3 ± 0.682.0 ± 0.981.2 ± 1.585.0 ± 0.692.3 ± 0.896.8 ± 0.7
KC77.3 ± 0.677.2 ± 0.679.3 ± 0.778.1 ± 0.780.0 ± 1.079.0 ± 1.783.3 ± 0.891.4 ± 0.796.4 ± 0.8
Table 4. McNemar test of different methods on the Indian Pines and KSC datasets.
Table 4. McNemar test of different methods on the Indian Pines and KSC datasets.
MethodsIndian PinesKSC
PCALDADLPPMFALGSFATwoSPDAGLLWDAPCALDADLPPMFALGSFATwoSPDAGLLWDA
RAW−0.3−1.9−4.9−7.6−8.5−18.9−28.0−34.91.3−4.5−2.7−5.5−3.9−6.4−21.5−27.6
PCA-−1.7−4.7−7.3−8.2−18.4−28.6−34.8-−4.7−2.9−5.7−4.1−6.5−21.5−27.7
LDA--−3.6−5.6−7.5−15.9−31.1−33.4--2.6−1.00.9−5.3−19.3−24.8
DLPP---−2.4−4.1−13.3−26.1−31.0---−3.4−1.7−6.4−20.2−25.7
MFA----−1.5−11.7−24.0−29.6----1.8−5.1−19.8−24.2
LGSFA-----−10.0−23.7−28.7-----−7.4−21.3−25.0
TwoSP------−15.5−21.4------−12.5−21.0
DAGL-------−17.7-------−10.9
Table 5. Average computational time (unit: second) of different methods with NN classifier on the Indian Pines and KSC datasets using τ = 0.05 .
Table 5. Average computational time (unit: second) of different methods with NN classifier on the Indian Pines and KSC datasets using τ = 0.05 .
MethodsRAWPCALDADLPPMFALGSFATwoSPDAGLLWDA
Time
Indian PinesT1-0.70.20.61.31.3428.3588.1229.3
T21.70.70.60.70.60.70.738.136.9
T3112.822.220.121.723.325.827.97145.348,347.3
KSCT1-0.30.10.20.50.489.4218.7157.5
T20.40.20.20.20.20.20.216.014.0
T328.47.06.27.510.18.48.33598.619,316.0
Table 6. Classification results (%) with different values of τ on the Indian Pines dataset with NN classifier. Each method has two rows, where the first row is the OA ± standard deviation (STD) and the second row is the kappa coefficient (KC) ± STD.
Table 6. Classification results (%) with different values of τ on the Indian Pines dataset with NN classifier. Each method has two rows, where the first row is the OA ± standard deviation (STD) and the second row is the kappa coefficient (KC) ± STD.
Methods τ = 0.05 τ = 0.06 τ = 0.07 τ = 0.08 τ = 0.09 τ = 0.1
RAW64.8 ± 1.065.5 ± 0.766.1 ± 0.467.0 ± 0.567.3 ± 0.567.7 ± 0.4
59.7 ± 1.160.6 ± 0.761.3 ± 0.562.3 ± 0.662.7 ± 0.663.1 ± 0.5
PCA64.9 ± 0.865.6 ± 0.566.2 ± 0.367.2 ± 0.567.5 ± 0.367.7 ± 0.4
59.8 ± 0.860.7 ± 0.561.5 ± 0.362.5 ± 0.662.9 ± 0.463.1 ± 0.5
LDA65.8 ± 1.767.9 ± 0.869.7 ± 1.271.0 ± 0.671.8 ± 0.972.6 ± 1.2
60.8 ± 2.163.1 ± 0.965.3 ± 1.466.9 ± 0.767.7 ± 1.068.7 ± 1.3
DLPP67.5 ± 0.970.3 ± 0.572.8 ± 0.873.9 ± 0.575.5 ± 0.676.7 ± 1.1
62.9 ± 1.066.0 ± 0.668.9 ± 0.970.2 ± 0.672.0 ± 0.773.3 ± 1.3
MFA68.6 ± 1.471.6 ± 1.673.5 ± 1.874.8 ± 1.576.6 ± 1.177.7 ± 1.2
63.9 ± 1.567.3 ± 1.869.6 ± 2.171.1 ± 1.773.2 ± 1.374.4 ± 1.4
LGSFA69.3 ± 1.171.6 ± 1.273.4 ± 1.075.0 ± 1.076.6 ± 0.677.1 ± 1.0
64.7 ± 1.367.4 ± 1.469.5 ± 1.271.3 ± 1.273.2 ± 0.873.8 ± 1.2
TwoSP73.9 ± 1.275.3 ± 0.876.3 ± 1.276.9 ± 0.677.7 ± 1.178.6 ± 0.9
70.1 ± 1.471.8 ± 0.972.9 ± 1.473.6 ± 0.774.5 ± 1.275.6 ± 1.1
DAGL81.7 ± 1.283.5 ± 1.384.1 ± 1.085.2 ± 0.985.7 ± 1.086.5 ± 1.2
79.2 ± 1.581.1 ± 1.582.8 ± 1.383.9 ± 1.384.5 ± 0.985.1 ± 0.8
LWDA85.1 ± 1.087.0 ± 1.687.7 ± 1.188.5 ± 1.088.9 ± 1.289.3 ± 0.7
83.0 ± 1.185.2 ± 1.986.0 ± 1.286.9 ± 1.287.3 ± 1.387.8 ± 0.7
Table 7. Classification results (%) with different values of τ on the KSC dataset with NN classifier. Each method has two rows, where the first row is the OA ± STD and the second row is the KC ± STD.
Table 7. Classification results (%) with different values of τ on the KSC dataset with NN classifier. Each method has two rows, where the first row is the OA ± STD and the second row is the KC ± STD.
Methods τ = 0.05 τ = 0.06 τ = 0.07 τ = 0.08 τ = 0.09 τ = 0.1
RAW79.7 ± 0.679.8 ± 0.980.3 ± 0.481.1 ± 0.482.0 ± 0.482.2 ± 0.4
77.2 ± 0.677.2 ± 1.078.1 ± 0.478.9 ± 0.479.9 ± 0.580.2 ± 0.5
PCA79.5 ± 0.679.4 ± 0.780.2 ± 0.380.9 ± 0.481.9 ± 0.482.0 ± 0.5
77.2 ± 0.677.0 ± 0.878.0 ± 0.378.8 ± 0.479.8 ± 0.580.0 ± 0.6
LDA81.4 ± 0.684.3 ± 1.285.8 ± 1.186.7 ± 1.087.8 ± 0.788.3 ± 1.0
79.3 ± 0.782.5 ± 1.384.1 ± 1.285.2 ± 1.186.4 ± 0.886.9 ± 1.1
DLPP80.3 ± 0.684.1 ± 1.085.6 ± 0.786.8 ± 1.187.7 ± 0.988.0 ± 1.0
78.1 ± 0.782.3 ± 1.183.9 ± 0.785.3 ± 1.286.3 ± 1.086.6 ± 1.1
MFA82.0 ± 0.984.1 ± 0.585.3 ± 0.585.9 ± 0.686.3 ± 0.886.9 ± 0.5
80.0 ± 1.082.3 ± 0.583.6 ± 0.684.3 ± 0.784.7 ± 0.985.4 ± 0.6
LGSFA81.2 ± 1.584.4 ± 1.186.5 ± 0.987.6 ± 1.288.7 ± 1.189.5 ± 0.7
79.0 ± 1.782.7 ± 1.285.0 ± 0.986.2 ± 1.387.4 ± 1.388.3 ± 0.8
TwoSP84.5 ± 0.685.3 ± 1.286.8 ± 1.088.0 ± 1.089.0 ± 1.090.1 ± 1.0
83.3 ± 0.884.2 ± 1.185.4 ± 1.086.4 ± 1.087.8 ± 1.188.9 ± 1.1
DAGL92.3 ± 0.792.7 ± 1.393.3 ± 1.193.7 ± 0.993.9 ± 0.694.0 ± 0.5
91.4 ± 0.992.0 ± 1.292.9 ± 1.393.4 ± 1.093.6 ± 0.793.8 ± 0.6
LWDA96.8 ± 0.797.2 ± 1.197.9 ± 1.398.3 ± 0.898.4 ± 0.798.5 ± 0.3
96.4 ± 0.896.9 ± 1.297.7 ± 1.498.1 ± 0.998.2 ± 0.898.4 ± 0.4

Share and Cite

MDPI and ACS Style

Li, X.; Zhang, L.; You, J. Locally Weighted Discriminant Analysis for Hyperspectral Image Classification. Remote Sens. 2019, 11, 109. https://doi.org/10.3390/rs11020109

AMA Style

Li X, Zhang L, You J. Locally Weighted Discriminant Analysis for Hyperspectral Image Classification. Remote Sensing. 2019; 11(2):109. https://doi.org/10.3390/rs11020109

Chicago/Turabian Style

Li, Xiaoyan, Lefei Zhang, and Jane You. 2019. "Locally Weighted Discriminant Analysis for Hyperspectral Image Classification" Remote Sensing 11, no. 2: 109. https://doi.org/10.3390/rs11020109

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop