Next Article in Journal
Synthetic Aperture Radar Image Segmentation with Reaction Diffusion Level Set Evolution Equation in an Active Contour Model
Previous Article in Journal
Seasonal and Decadal Groundwater Changes in African Sedimentary Aquifers Estimated Using GRACE Products and LSMs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Joint Sparse Models for Hyperspectral Image Classification Based on a Novel Neighbour Selection Strategy

1
School of Civil and Environmental Engineering, University of New South Wales, Sydney, NSW 2052, Australia
2
School of Engineering and Information Technology, University of New South Wales, Canberra, ACT 2612, Australia
*
Author to whom correspondence should be addressed.
Remote Sens. 2018, 10(6), 905; https://doi.org/10.3390/rs10060905
Submission received: 1 May 2018 / Revised: 30 May 2018 / Accepted: 5 June 2018 / Published: 8 June 2018
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Joint sparse representation has been widely used for hyperspectral image classification in recent years, however, the equal weight assigned to each neighbouring pixel is less realistic, especially for the edge areas, and one fixed scale is not appropriate for the entire image extent. To overcome these problems, we propose an adaptive local neighbour selection strategy suitable for hyperspectral image classification. We also introduce a multi-level joint sparse model based on the proposed adaptive local neighbour selection strategy. This method can generate multiple joint sparse matrices on different levels based on the selected parameters, and the multi-level joint sparse optimization can be performed efficiently by a simultaneous orthogonal matching pursuit algorithm. Tests on three benchmark datasets show that the proposed method is superior to the conventional sparsity representation methods and the popular support vector machines.

Graphical Abstract

1. Introduction

In recent years, remote sensing images have played an important role in many areas, such as surveillance, land-use classification, forest disturbance, and urban planning [1]. How to exploit the information of remotely-sensed images has been a popular research problem for decades. Hyperspectral images (HSI) have attracted a significant amount of attention due to its high spectral resolution and wide spectral range which make it possible to analyse and distinguish various objects with higher accuracy [2].
One of the most important applications of HSI is supervised classification which assigns a specific class to each pixel based on the spectral information [3]. Various techniques have been employed for this task, such as support vector machines (SVM) [4,5,6], random forest (RF) [7], multinomial logistic regression (MLR) [8], and neural networks (NN) [9]. Among these techniques, SVM has shown its effectiveness for HSI classification, especially when dealing with the Hughes phenomenon of very high-dimensional data [4]. Dimensionality reduction methods were developed to deal with the high dimensionality of HSI data and obtained some promising results [6,10,11]. Although these methods have provided some reasonable solutions to the problem, spatial context has not been fully utilized in those conventional classifiers. Without spatial information being involved, the classification map may produce a more noisy appearance and a lower accuracy [12]. During the past decade, many attempts have been made to integrate the spatial context in the classification tasks. Some methods have focused on feature extraction, such as extended morphological profiles [13,14] and attribute profiles [15,16], which improved the classification results by taking the morphological properties into account. Kernel-based methods, such as the one in [17], use a composite kernel to incorporate spectral and spatial properties, and then images are classified by SVM. In addition, multiple feature learning approaches [18,19], which combine different features learned by various methods, are proven to enhance the classification results. Markov random fields (MRF) [20,21] focus on developing classifiers that can preserve the spatial context by formulating a minimization function of spatial and spectral energy, and the result changes when different weights are assigned to spectral and spatial energy terms. All these techniques are able to incorporate spatial and spectral information.
In the last few years, sparse representation (SR) has been a promising tool in solving many image processing problems, such as denoising [22], fusion [23], and image compression [24]. In [25], SR was used to detect the boundary points and outliers from images, which is proven to be very useful for image processing applications. SR assumes that a natural signal can be linearly expressed by a few coefficients from a so-called dictionary [26]. Now SR has been extended to the classification of HSI based on the assumption that the pixels of a class usually lie in a low-dimensional subspace, despite their high-dimensional characteristics [27]. This enables a test pixel with an unknown label to be linearly represented by a few elements, and then the label can be determined after the coefficient vectors are recovered from a training dictionary. Many studies [28,29,30] have provided a promising result for HSI classification based on SR. Moreover, SR can also learn the probability for a decision fusion classifier [31]. It should be noted that the reconstruction work mechanism makes it very efficient, and can easily accommodate new training data by reconstructing the class-specific dictionary (e.g., directly adding the new training data to the dictionary matrix which corresponding to the label class of the training data) without retraining the whole dataset, which is required for other classifiers, such as SVM and MLR. Another advantage of SR compared to other conventional binary classifiers is that it can label the pixels from multiple classes directly. Many studies have been attempted to explore the merits of SR. Cui and Prasad [32] proposed a class-dependent sparse representation classifier which utilized the k-nearest method to select atoms for SR, and they considered the label information by computing the Euclidean distances between the test sample and k-nearest pixels in a class-specific manner. The class-dependent SR can also be extended to a kernelized variant by using a kernel function. Similar work has been done in [33]. In the report of [34], the authors proposed a multi-layer spatial-spectral sparse representation specifically for hyperspectral image classification. The class residual is computed in the first sparse representation layer, and then the sparse coefficients are determined to be updated or not in the next layer based on a class-dependent residual distribution. The multi-layer strategy is implemented sequentially and the selected atoms’ indices are determined by the classes’ rankings within the minimal residuals. The experimental results have demonstrated its superior results when compared to the traditional sparse coding methods.
In order to further exploit the spatial information, a joint sparse model (JSM) has been proposed [28]. For HSI, neighbouring pixels tend to have similar contextual properties and are highly correlated with each other [35]. JSM assumes that pixels constrained by a region scale share a common sparsity pattern, and these pixels can be linearly represented by a few common atoms which are associated with different sparsity coefficients. Based on this assumption, the test pixel can be replaced with its surrounding neighbours in the JSM model to seek a more reliable representation. Recently, JSM has achieved a better performance when compared to pixel-wise SR methods [36]. However, JSM is sensitive to the selected region scale because near-edge areas require a small region scale and smooth areas need a large region scale. Some experiments have shown that, if an oversized area is selected for a specific test pixel, the accuracy tends to decrease [37]. If the scale is too small, then insufficient contextual properties are included; hence, it is difficult to choose an optimal region scale for JSM.
On the other hand, for a given specific area, distinct structures and characteristics, as well as some irrelevant information, will exhibit; however, some pixels with different spectral structures of the test pixel also exist in this region. If a strategy aims to find the most similar pixels to the test pixel and reject the dissimilar neighbouring pixels, information of correlated spatial context should be more representative for classification. Hence, we propose an adaptive neighbour selection strategy which computes the weights based on distances between pixels, with the labels of training data as a priori information. The structural similarity between the central pixel and its neighbours can be exploited in a more sensible way by considering the different contribution of each spectral band. Based on this, a novel joint sparse model-based classification approach, namely ‘adaptive weighted joint sparse model’ (AJSM) is proposed in this paper. Moreover, we propose a novel classification method with a name ‘multi-level joint sparse representation model’ (MLSR), in order to take advantage of the correlations among neighbouring pixels in a region. The procedures of MLSR are summarized as: (1) Local matrices are obtained by the proposed adaptive neighbour selection strategy. Different thresholds of distances can result in different local matrices corresponding to different levels; therefore (2) different joint sparse representations of the test pixel from different levels can be constructed. Since pixels with similar distances can be simultaneously sparsely represented by the features in the same subspace, and pixels from multiple levels may share different sparsity patterns, MLSR is designed to learn the dictionary for each joint sparse model separately; and (3) a simultaneous orthogonal matching pursuit (SOMP) algorithm is employed to learn the multi-level classification task.
The weight matrix for AJSM and MLSR is constructed by the ratio of the between-class and within-class distances with the consideration of a priori label information. This alleviates the negative impact when we classify the mixed pixels and similar pixels. In addition, the proposed MLSR performs on one region scale with different levels, and the sparse coding procedures at different levels are independent of each other. To sum up the main advantage of the proposed multi-level method, various parameter values can generate multiple sparse models to represent the different inner contextual structures among pixels, thereby improving the HSI classification accuracy.
The remainder of this paper is organized as follows: Section 2 reviews the sparsity representation and joint sparse models briefly. Section 3 describes the proposed MLSR method in detail for HSI classification. Experimental results on three benchmark datasets are presented in Section 4. Finally, conclusions and future work are provided in Section 5.

2. Classification of HSI via SR and JSM

2.1. Sparsity Representation Classification Model

For the sparsity representation classification (SRC) model, assume that there are N   training pixels belonging to C classes, and   x is a L dimensional pixel. Let D be the dictionary learnt by training samples, and x can be linearly represented by the combination of D :
x = [ D 1 , D 2 D C ] [ r 1 r 2 r C ] = D r
where D c L × N c ( N 1 + N c + + N C = N ) is the sub-dictionary for the c-th class, r c N c × 1 is the sparse coefficients corresponding to D c . In an ideal situation, if x belongs to the c-th class, then r j = 0 , j = 1 C , j c . Given the dictionary D , coefficient vectors can be recovered by solving the optimization problem:
r ^ = arg min r r 0 s u b j e c t t o D r = x
Considering empirical error tolerance σ , Equation (2) can be relaxed with the following inequality:
r ^ = arg min r r 0 s u b j e c t t o D r x 2 σ
Equation (3) can also be replaced by a sparse objective function:
r ^ = arg min r x D r 2 s u b j e c t t o r 0 P
where P is a predefined sparsity parameter corresponding to the number of zero entries in r . This nondeterministic polynomial-time hard (NP-hard) problem can be optimized by greedy pursuit algorithms. Orthogonal Matching Pursuit (OMP) [38] is a typical algorithm that solves this NP-hard problem, in which the residual is always orthogonal to the span of the already selected atoms, and r is updated by the residual in each iteration. This problem can also be relaxed to a basis pursuit problem by replacing the 0 norm with other form of regularization as follows:
r ^ = arg min r x D r 2 + λ r q
where λ is a regularization parameter, and the norm is l 1 and l 2 when q = 1 and q = 2 respectively. Normally, 1 norm is more effective in solving the convex optimization problems than 0 norm is, and 2 norm can avoid the overfitting issue. The detailed procedure to solve this convex problem can be found in [39].
The label of x can be directly determined by the recovered sparse coefficients and reconstruction error. Let e represent the residual error between the test sample and the reconstruction term by sparse representation:
e c = x D c r ^ c 2 c = 1 , 2 C
where r c represents the residual computed by dictionary and an optimal sparse coefficient for the c-th class. Then the class label of test sample x can be obtained according to the minimum residual:
C l a s s ( x ) = arg min c = 1 , 2 C e c ( x )

2.2. Joint Sparsity Model

Since spatial information has been considered very important for HSI classification tasks, it is essential to embed spatial contextual information into the SR model as well. A joint sparsity model (JSM) [28] was proposed to exploit the correlation between neighbouring pixels and the centre pixel. Given a patch of W × W pixels where W is a square number, let X = [ x 1 , x 2 x W ] be the joint signal matrix consisting of all the neighbouring pixels in this patch. In other words, the test pixel is located at the centre of the selected region, the remaining pixels in X are its neighbours. According to the report [28], X can be expressed as:
X = [ x 1 , x 2 x W ] = [ D r 1 , D r 2 D r W ] = D [ r 1 , r 2 r W ] = D R
where R = [ r 1 , r 2 r W ] N × W is the sparsity matrix, and the selected atoms in dictionary D are determined by the nonzero coefficients in R . Therefore, the common sparsity pattern for pixels can be recognized by enforcing the indices of nonzero atoms in the sparsity coefficient matrix.
Given the dictionary D , the matrix R can be optimized by solving the following object function:
R = arg min X D R F s u b j e c t t o R r o w , 0 P
where F is the Frobenius norm, and R r o w , 0 denotes the number of nonzero rows of R . Equation (9) is also an NP-hard problem. Simultaneous OMP (SOMP) [36] is a generalized OMP algorithm which can be used to efficiently solve this problem.
The label of x can be directly determined by the recovered sparse coefficients and reconstruction error. Let e represent the residual error between the test sample and the reconstruction term by sparse representation:
e c = X D c R c 2 c = 1 , 2 C
where R c represents the reconstruction residual corresponding to the c-th class. Then the class label of test sample x can be obtained according to the minimum residual:
C l a s s ( x ) = arg min c = 1 , 2 C e c ( x )
JSM can achieve a better classification result by incorporating the contextual information of neighbouring pixels when compared with a pixel-based SRC. However, different areas need different region scales, and there may exist some less-correlated pixels in one local patch due to the spectrally heterogeneous features in HSI scenes even though neighbouring pixels tend to have similar spectral signatures. Another situation that should be considered according to Li [40] is that the general dictionary constructed by the whole set of training samples may include outliers.

3. Adaptive Weight Joint Sparse Model (AJSM) and Multi-Level Sparse Representation Model (MLSR)

We introduce an adaptive weight joint sparse model (AJSM) and a multi-level joint sparse representation model (MLSR) for HSI classification in this paper. Multiple local signal matrices are constructed using different parameters to realize the similarity learning in MLSR. In fact, AJSM is a simple form of MLSR. The proposed AJSM is expected to improve the classification accuracy in these areas by not taking all the neighbouring pixels to construct the joint sparse matrix. Additionally, MLSR improves the classification results by selecting the neighbour pixels from various levels using the proposed adaptive neighbour selection strategy.
To better understand the procedure of the proposed method, a flowchart is shown in Figure 1 where each component of the method is explained in detail in the following sections.

3.1. Adaptive Local Signal Matrix

In order to select reasonable neighbours to construct the joint matrix, the weighted Euclidean distances between the test pixel and its neighbours are used. We first select a region with a window size W × W , which is centred at the test pixel x i . Different weights are given to each spectral band according to their contribution to the whole spectral characteristics. The weighting strategy is described as follows:
A < x i , x j > = l = 1 L w l ( x i l x j l ) 2 w l = exp ( α I l ) l = 1 L exp ( α I l ) I l = i = 1 c I n ( y i = c ) ( x ¯ c l x ¯ l ) 2 i = 1 c I n ( y i = c ) ( x i l x ¯ c l ) 2
where A < x i , x j > is the weight distance between pixels x i and x j , w l is the weight for the l-th feature, and w l is determined by training samples from different classes. α is a positive parameter that controls the influence of a class-specific distance I l . If α = 0 , the distance between two pixels decreases to the equal weight Euclidean distance. If α is large enough, the change will be reflected on I . I n ( ) denotes an indicator function which takes between-class and within-class distances into account. x ¯ c l is the average of the c-th class of the l-th feature, and x ¯ l represents the average of all training samples of the l-th feature; y i represents the label of pixel x i .
Pixels with a predefined distance value can be selected as similar neighbours according to this method. In other words, this adaptive neighbour selection strategy can identify the samples with similar characteristics to form a group. The superiority of this weight strategy over other weighting schemes is that it considers the spectral similarities on a pixel level, and the discriminative information among different groups which can be obtained from training samples.

3.2. Adaptive Weight Joint Sparse Model

The goal of Equation (12) is to find the optimal samples to reconstruct the central pixel. Once the appropriate weights are assigned to each spectral band, the weight distances between the test pixel and its neighbouring pixels can be evaluated. Based on the top N-nearest strategy, N nearest neighbouring pixels can be chosen as the adaptive weight joint sparse matrix to relax the joint sparse model (Equation (9)). Here we define S N as the weight matrix chosen from the original joint sparse matrix X = [ x 1 , x 2 x W ] . In other words, N nearest pixels are selected from the W pixels based on the previous adaptive weight scheme. The adaptive weight joint sparse model can be expressed as:
R = arg min S N D R F s u b j e c t t o R r o w , 0 P
The label of central pixel can be identified by minimizing the class residual:
C l a s s ( x ) = arg min c = 1 , 2 C S N D c R c F
The procedure of AJSM is summarized below in Algorithm 1.
Algorithm 1. The implementation of AJSM
Input: training datasets belong to the c-th class: X c , region scale: W , top number of nearest neighbours: N , test datasets X T .
Initialization: initialize dictionary D with training samples, and normalize the columns of D to have unit 2 norm.
1. Compute the w l for each spectral band according to Equation (12);
2. For each test pixel in x i X T :
 Construct the weight matrix S N according to Equation (12) and normalize the columns of S N to have a unit 2 norm;
 Calculate the sparse coefficient matrix R and dictionary D from Equation (13) using SOMP;
 Determine the class label y t for each test pixel x i X T by Equation (14).
Output: 2-dimensional classification map.
It has been identified that neighbouring pixels consist of different types of materials in the heterogeneous areas in HSI. JSM cannot perform well on such areas due to its definition of neighbouring pixels, which tend to have similar labels. The proposed AJSM is expected to improve the classification accuracy in these areas by not taking all the neighbouring pixels to construct the joint sparse matrix.

3.3. Multi-Level Weighted Joint Sparse Model

The neighbour pixels selected from a fixed scale using single level criteria as seen in JSM and AJSM may not contain the complementary and accurate information, and the neighbour pixels selected from different criteria levels can help represent the data wholly. Herewith we propose a multi-level weighted joint sparse model to fully integrate the neighbour information, as well as to avoid the outliers dominating in the sparse coding. For a test pixel, its neighbour pixels are selected by the proposed adaptive neighbour selection strategy with different distance threshold level values. Then the multiple joint sparse matrices are constructed by the corresponding neighbour pixels with different distance threshold level values. The details of this method are described as follows.
Assume that S i , k is the k-th joint sparse matrix constructed for pixel x i . Here we define S i , k using a weight matrix i.e., S i , k = [ ϖ < x i , x i , 1 > x i , 1 , ϖ < x i , x i , 2 > x i , 2 , , ϖ < x i , x i , j > x i , j , , ϖ < x i , x i , W > x i , W ] where ϖ < x i , x i , j > is a function that determines if pixel x i , j can be preserved to reconstruct x i , x i , j is the j-th sample in the given region which is restricted by the scale W × W . In (12), A < x i , x i , j > is a monotonously increasing function of the weighted distances. Although there are many ways to define ϖ < x i , x i , j > , we define it as a piecewise constant to simplify the selection of different joint sparse matrices as follows:
ϖ < x i , x i , j > = { 1 , A < x i , x i , j > ε 0 , A < x i , x i , j > > ε
where ε is a threshold controlling the value of the corresponding element in S i , k . According to (12) and (15), when a pixel in S i , k has the corresponding weighted distance with the test pixel x i : A < x i , x i , j > > ε it will not be selected in the joint sparse model. Otherwise, if A < x i , x i , j > ε , the corresponding term will be selected to reconstruct the test pixel. In other words, S i , k is constructed by the terms that have the weighted distances less than ε between itself and the test pixel x i .
By using the proposed scheme, we can generate different patches with various values of ε :
  • ε = 0 : This is an independent set. In this situation, only the central pixel itself is selected. This means that the joint sparse model becomes a pixel-wise sparse representation model.
  • ε 1 : Because A < x i , x i , j > 1 in this situation, all the neighbours of the test pixel in the given area are selected.
  • 0 < ε < 1 : The sparsity representation of x i , j is satisfied. A smaller number of pixels are selected for the reconstruction of x i .
As described above, for each test pixel x i , when different parameters of { ε 1 , , ε k , , ε K } are applied, K different patches can be generated to represent this pixel with the inner contextual information involved. Our next task is to construct the multi-level joint sparse representation model for the test pixel.

3.4. Multi-Level Joint Sparse Representation

The SRC has been successfully used for HSI classification, herein we extend it to a multi-level version for the classification task. After K different patches are constructed for each pixel, the patches for the test pixel can be arranged as a feature matrix: S i = [ S i , 1 , , S i , k , , S i , K ] ( k = 1 , 2 , , K ), where S i , k be the k-th joint sparsity matrix constructed for the test pixel x i .
In this paper, let D = { D 1 , , D k , D K } be a set of dictionary which can be learnt from all the training data for K patches, and D k is the dictionary learnt for the k-th level. Each dictionary D k is composed of all the sub-dictionaries for each labelled class as D k = [ D 1 k , , D c k , , D C k ] , where D c k denotes the sub-dictionary of the c-th labelled class.
The sparse representation of the test pixel x i with its k-th patch can be described as:
min Q k S i , k D k Q k F
where Q k is the sparse representation coefficients for the specific patch S i , k . Equation (16) expresses how to sparsely represent each of the K patches when the sparse coefficient vector is given. Considering all the K patches, Equation (16) can be rewritten as:
min Q k = 1 K ( S i , k D k Q k F )
where Q = [ Q 1 , , Q k , Q K ] is composed of K columns of the coefficient vectors. Each column of the matrix is the sparse representation coefficients corresponding to a dictionary over a specific patch.
Since the pixels belonging to the same class should have the dictionary in the same subspace spanned by the training samples, the class-specific level joint representation optimization problem can be written as:
D ^ , Q ^ = arg min D , Q k = 1 K ( S i , k D k Q k F )
This problem can be decomposed into K sub-problems. In this paper, the SOMP is used to solve the optimization function (Equation (18)) and it can efficiently solve this problem in several iterations. Algorithm 2 introduces the implementation of the proposed framework.
After the sparsity coefficients are obtained, for a given test pixel x i , it would be assigned to the class which gives the smallest reconstruction residual:
y i = arg min c 1 , 2 , , C E c ( x i )
where E c ( x i ) is the reconstruction residual of x i , as:
E c ( x i ) = k = 1 K S i , k D c k Q c k 2 2
where D c k is the dictionary for the c-th class over the k-th patch, and Q c k denotes the sparse coefficient matrix corresponding to D c k .
Algorithm 2. The implementation of the proposed algorithm.
Input: training datasets belonging to the c-th class: X c , region scale: W , number of levels: K , distance threshold controlling parameter: ε , test datasets X T
Initialization: initialize dictionary D c = X c , and normalize the columns of dictionary to have unit 2
1. Compute w l ( l = 1 , 2 , , L ) according to Equation (12) using the training data sets X c and the corresponding labels
2. For each test pixel x i X T
 Compute adaptive weight distances between the test pixel and all the pixels in the selected neighbour region to construct A < x i , x j > based on Equation (12)
3. Compute S i , k based on Equation (15).
4. For k = 1 : K
 Compute Q k for each level for each class using SOMP.
5. Compute the class label y i for test pixel based on Equations (19) and (20).
Output: 2-dimensional classification map.

4. Experimental Results and Analysis

4.1. Data Description

To validate the proposed methods, three benchmark datasets are used in the experiments.
1. Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) dataset: Indian Pines, which was obtained by an AVIRIS sensor over the site in Northwest Indiana, United States of America (USA). This imagery has 16 labelled classes. The dataset has 220 spectral bands ranging from 0.2 to 2.4 µ m wavelength, and each channel has 145   ×   145 pixels with a spatial resolution of 20 m. Twenty water absorption bands (no. 104–108, 150–163, and 220) are removed in the experiments. This data is widely used due to the presence of mixed pixels in available classes and the unbalanced number of training samples per class.
2. Reflective Optics System Imaging Spectrometer (ROSIS) data set: University of Pavia, Italy. This image was acquired during a fight campaign over Pavia in Northern Italy. It has nine labelled ground truth classes with 610   ×   340 pixels. This dataset contains urban features, as well as vegetation and soil features. Each pixel has a 1.3 m spatial resolution. With water absorption bands removed, 103 bands are used in the experiments.
3. AVIRIS data set: Salinas. The image was also acquired by an AVIRIS sensor over Salinas Valley, CA, USA. The image is of 512   ×   217 size with 224 spectral bands. In the experiments, 20 water absorption bands (no. 108–112, 154–167, and 224) are removed. Salinas has a 3.7 m resolution per pixel and 16 different classes. It includes vegetables, bare soils, and vineyard fields. Due to the spectral similarity of most classes, this dataset has been frequently used as a benchmark for HSI classification.
The ground truths of three datasets, as well as the false colour composite images are illustrated in Figure 2.

4.2. Description of Comparative Classifiers and Parameters Setting

In this paper, the proposed AJSM and MLSR are compared with several benchmark classifiers: pixel-wise SVM (referred to as SVM), EMP with SVM (referred to as EMP), pixel-wise SRC (referred to as SRC), and JSM with a greedy pursuit algorithm [28]. Pixel-wise SVM and pixel-wise SRC classify the images with only spectral information, while JSM, AJSM, and MLSR are sparse representation-based classifiers with spatial information utilized.
During the experiments, the range of parameters is empirically determined and the optimal values are determined by cross-validation. The parameters for pixel-wise SVM are set as the default ones in [4] and implemented using the SVM library with Gaussian kernels [41]. Parameters for EMP and pixel-wise SRC are set up by following the instructions in [14] and [28], respectively. The selected regions for JSM, AJSM, and MLSR are set as 3   ×   3, 5   ×   5, 7   ×   7, 9   ×   9, 11   ×   11, 13   ×   13 and 15   ×   15, and the best result is described in this paper. For AJSM, the number of pixels selected in the given region is set as: 7, 20, 40, 50, 50, 50, and 50 for the abovementioned scales, respectively. For the proposed MLSR, the number of threshold parameter ε is set as seven, and the threshold values are:   { 0.1 ,   0.2 ,   0.3 ,   0.4 ,   0.5 ,   0.7 ,   1 } . The predefined sparsity level is set as 3 for each dataset.
Quantitative analysis metrics, overall accuracy (OA), average accuracy (AA), and Kappa coefficient are adopted to validate the proposed method. All the experiments in this paper are repeatedly implemented ten times and the mean accuracy is presented.

4.3. Experimental Results

The first experiment was performed on the Indian Pines image. We randomly selected 10% of the samples from each class as training data and the remaining as a test dataset. The optimal parameters in this experiment are set as: α = 0.2, W = 13 × 13 . The numbers of training and test data for each class are described in Table 1. Classification results are listed in Table 2, and the classification maps are shown in Figure 3. One can observe that the classification maps obtained by pixel-wise SVM and pixel-wise SRC have a more noisy appearance than other classifiers, which confirms that the contextual information is important for hyperspectral image classification. Considering the spatial information, JSM gives a smoother result; however, it still fails to classify some near-edge areas. EMP, AJSM, and the proposed MLSR deliver better results, and MLSR shows the highest classification accuracy. From Figure 3, one can see that MLSR further provides a smoother classification result and preserves more useful information for HSI.
The proposed AJSM improves the classification capability of JSM by exploring the different contributions of the neighbouring pixels in the selected region. This confirms the effectiveness of the adaptive weight matrix scheme. However, one can see that AJSM produces a relatively lower accuracy for oats, which has limited training samples. The improvement of MLSR-based classification of alfalfa and oats, which have been considered as small classes, indicates that the proposed method can perform well on classes with fewer training samples. In addition, the adaptive local matrix imposes the local constraint on the sparsity, which would improve the performance. As can be observed from the classification maps, our proposed method has a better capability to identify the near-edge areas and it benefits from the selection of most similar pixels to reconstruct the test pixel. The accuracies for MLSR are very high, which indicates that JSM can be significantly improved by multiple feature extraction approaches.
The second experiment is conducted on the Pavia University image, and Table 3 shows the class information. We randomly selected 250 samples as the training data, and the rest as test data. The optimal parameters in this experiment are set as: α = 0.2, W = 15 × 15 . Classification results and maps are illustrated in Table 4 and Figure 4, respectively. It is obvious that the multi-level information can, indeed, improve the results of classification of the Pavia University image compared to other SRC based methods and the popular SVMs. The improvement of MLSR compared to JSM suggests that the local adaptive matrix can preserve the most useful information and reduce the redundant information. The result is consistent with the previous experiment on the Indian Pines image where the edge pixels are predicted more precisely.
The third experiment is conducted on the Salinas imagery. For each class, 1.5% samples are selected as the training data, and the remaining as the test dataset. The optimal parameters in this experiment are set as: α = 0.2, W = 15 × 15 . The class information and classification results are given in Table 5 and Table 6, respectively. The results are also visualized in classification maps as shown in Figure 5. One can observe that the proposed MLSR yields the best accuracy for most of the classes, especially for classes 15 and 16. Furthermore, the proposed MLSR identified the edge areas best.

4.4. Effects of Different Kinds of Parameters

This section focuses on the effects of the parameters setting on the classification performance. We first varied the value of positive parameter α that controls the influence of the ratio of the between-class and within-class distances, and the value was varied from 0 to 1 at 0.2 intervals. The experiments were conducted with AJSM on three datasets and the window sizes were fixed as the corresponding optimal values. In Figure 6, the overall accuracies for three datasets fluctuate in a small range, and the best performances were obtained when α was set as 0.2 for all three datasets though the trends for them were different. As α only controls the influence of each feature band, it is reasonable to apply the same value for MLSR in the experiments.
The effect of region scales for JSM, AJSM, and MLSR has also been analysed in the experiments. In order to simply show the trends, the numbers of training and test datasets are selected to be the same as in the previous experiments. OA is shown in Figure 7. For JSM, AJSM, and MLSR, the region scales ranging from 3   ×   3 to 29   ×   29 at 2   ×   2 intervals. As shown in Figure 7, the best OA is achieved for JSM when the scale is set as 7   ×   7, 11   ×   11, and 15   ×   15 for Indian Pines, Pavia University, and Salinas, respectively. If the scale increases, the accuracy decreases dramatically. In most situations, AJSM performs better than JSM because the most useful information is preserved and the redundant information is rejected by the selection strategy. The accuracy for MLSR becomes stable when a larger region is selected. More specifically, the proposed MLSR performs better than other joint sparsity-based models in most regions. This result actually benefits from its mechanism of discarding outliers in the specific area, which provides a more reliable dictionary.
Another consideration is the number of patches that should be tested, i.e., is having more patches better? To evaluate this, the adaptive framework is used to generate more patches. Especially, with ε set to { 0.1 ,   0.2 ,   0.3 ,     0.4 ,   0.5 ,   0.6 ,   0.7 ,   0.8 ,   0.9 ,   1 } , we can define 11 patches. In each experiment, we randomly selected a patch subset with the number of K { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 } from these 11 patches and evaluated the performance of the method on three datasets. For each value of K , the experiment procedure is repeated 10 times with different subset selection. Figure 8 shows the average OA result of the 10 iterations. As K increases, the performance of the framework also increases when K 7 ; however, it slightly decreases when K 8 . This trend shows that a certain number of patches are necessary for the improvement of the performance of the proposed method. However, too many patches can also result in a slight decrease in performance. In the experiment, we fixed five values {0.1, 0.2, 0.3, 0.4, 0.5}, and the last two values are determined from the remaining five values {0.6, 0.7, 0.8, 0.9, 1} by cross-validation.
We also conducted the experiments to evaluate the impact of the number of training samples per class for pixel-wise SVM, pixel-wise SRC, EMP with SVM, single-scale JSM, and the proposed MLSR. AJSM is not considered in this experiment as they exhibit a similar trend with JSM. Training samples are randomly chosen, and the rest are used as test samples. For the Indian Pines dataset, the number of training data ranges from 5% to 40% of the whole pixel counts at 5% intervals; for the Pavia University dataset, the number of training samples per class ranges from 150 to 500 at 50 intervals; For the Salinas dataset, the number of training samples per class ranges from 50 to 400 at 50 intervals. Figure 9 illustrates the classification results (OA) for these three datasets. As can be observed, less than 5% of the samples are needed for each class to obtain an OA over 90% for the Indian Pines datasets using the proposed MLSR. This is very promising because it is often difficult to collect a large training datasets in practice. For the Pavia University dataset, only 150 training samples are needed to obtain an OA of 95%. In fact, this accuracy is 3% higher than that by JSM and 4.5% higher than that by EMP with SVM. This is due to the fact that the local information included by the proposed MLSR outperforms the others. The same trend can be concluded for the Salinas dataset. In addition, the proposed MLSR produces very high accuracy and shows robustness with an increase of the number of training samples, and it can be observed that MLSR performs very well when training samples are limited.

5. Conclusions and Future Research Lines

In this paper, we have introduced two novel sparse representation-based hyperspectral classification methods. These proposed methods employ an adaptive weight matrix scheme as the neighbour selection strategy for the joint sparse matrix construction. The adaptive weight joint sparse model outperforms the traditional joint sparse models, however, it is designed for simple cases rather than complicated situations where the number of labelled training samples is not sufficient. This was overcome by introducing the second model, i.e., the multi-level joint sparse model that can solve the complex classification problem in a more effective way. The multi-level joint sparse model consists of two main parts: adaptive locality patches and a multi-level joint sparse representation model. This model is introduced to fully explore the spatial context within a given region for the test pixel. The proposed methods locally smooth the classification maps and preserve the relevant information for most labelled classes. Compared with other spatial-spectral methods and sparse representation-based approaches, the proposed methods can provide a better performance on real hyperspectral scenes. This is consistent with the observation from the classification maps. Moreover, the experiments on the impact of the number of training samples also indicate that the proposed multi-level sparse approach leads to a more reliable result when only a limited number of training samples are available.

Author Contributions

Q.G. formulated and directed the methodology. S.L. and X.J. supervised the data processing. Q.G. prepared the manuscript and interpreted the results supported by S.L. and X.J. All authors contributed to the methodology validation, results analysis, and reviewed the manuscript.

Acknowledgments

The authors would like to thank D. Landgrebe from Purdue University for providing the free downloads of the hyperspectral AVIRIS dataset, Paolo Gamba from the Telecommunications and Remote Sensing Laboratory for providing the Pavia University dataset, the California Institute of Technology for providing the Salinas dataset, and the Associate Editor and anonymous reviewers for their careful reading and helpful comments which significantly helped in improving this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Campbell, J.B.; Wynne, R.H. Introduction to Remote Sensing; Guilford Press: New York, NY, USA, 2011. [Google Scholar]
  2. Landgrebe, D.A. Signal Theory Methods in Multispectral Remote Sensing; John Wiley & Sons: Hoboken, NJ, USA, 2005; Volume 29. [Google Scholar]
  3. Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Multiple spectral–spatial classification approach for hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4122–4132. [Google Scholar] [CrossRef]
  4. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  5. Pal, M.; Foody, G.M. Feature selection for classification of hyperspectral data by SVM. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2297–2307. [Google Scholar] [CrossRef]
  6. Sun, W.; Liu, C.; Xu, Y.; Tian, L.; Li, W. A Band-Weighted Support Vector Machine Method for Hyperspectral Imagery Classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1710–1714. [Google Scholar] [CrossRef]
  7. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef] [Green Version]
  8. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised hyperspectral image classification using soft sparse multinomial logistic regression. IEEE Geosci. Remote Sens. Lett. 2013, 10, 318–322. [Google Scholar]
  9. Kuching, S. The performance of maximum likelihood, spectral angle mapper, neural network and decision tree classifiers in hyperspectral image analysis. J. Comput. Sci. 2007, 3, 419–423. [Google Scholar]
  10. Sun, W.; Halevy, A.; Benedetto, J.J.; Czaja, W.; Li, W.; Liu, C.; Shi, B.; Wang, R. Nonlinear dimensionality reduction via the ENH-LTSA method for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 375–388. [Google Scholar] [CrossRef]
  11. Wang, Q.; Zhang, F.; Li, X. Optimal Clustering Framework for Hyperspectral Band Selection. IEEE Trans. Geosci. Remote Sens. 2018, 1–13. [Google Scholar] [CrossRef]
  12. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in spectral-spatial classification of hyperspectral images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef]
  13. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  14. Fauvel, M.; Benediktsson, J.A.; Chanussot, J.; Sveinsson, J.R. Spectral and spatial classification of hyperspectral data using SVMs and morphological profiles. IEEE Trans. Geosci. Remote Sens. 2008, 46, 3804–3814. [Google Scholar] [CrossRef]
  15. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of hyperspectral images by using extended morphological attribute profiles and independent component analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef]
  16. Song, B.; Li, J.; Dalla Mura, M.; Li, P.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A.; Chanussot, J. Remotely sensed image classification using sparse representations of morphological attribute profiles. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5122–5136. [Google Scholar] [CrossRef]
  17. Li, J.; Marpu, P.R.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A. Generalized composite kernel framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  18. Yue, J.; Zhao, W.; Mao, S.; Liu, H. Spectral–spatial classification of hyperspectral images using deep convolutional neural networks. Remote Sens. Lett. 2015, 6, 468–477. [Google Scholar] [CrossRef]
  19. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On combining multiple features for hyperspectral remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 879–893. [Google Scholar] [CrossRef]
  20. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  21. Zhang, B.; Li, S.; Jia, X.; Gao, L.; Peng, M. Adaptive Markov random field approach for classification of hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2011, 8, 973–977. [Google Scholar] [CrossRef]
  22. Elad, M.; Aharon, M. Image denoising via learned dictionaries and sparse representation. In Proceedings of the IEEE Conference Computer Visison and Pattern Recognition (CVPR), New York, NY, USA, 17–22 June 2006; pp. 895–900. [Google Scholar]
  23. Wu, Y.; Blasch, E.; Chen, G.; Bai, L.; Ling, H. Multiple source data fusion via sparse representation for robust visual tracking. In Proceedings of the 14th International Conference Information Fusion (FUSION), Chicago, IL, USA, 5–8 July 2011; pp. 1–8. [Google Scholar]
  24. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed]
  25. Li, X.; Lv, J.; Yi, Z. An efficient representation-based method for boundary point and outlier detection. IEEE Trans. Neural Netw. Learn. Syst. 2016, 29, 51–61. [Google Scholar] [CrossRef] [PubMed]
  26. Huang, K.; Aviyente, S. Sparse representation for signal classification. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 3–6 December 2007; MIT Press: Boston, MA, USA, 2007; pp. 609–616. [Google Scholar]
  27. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification via multiscale adaptive sparse representation. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7738–7749. [Google Scholar] [CrossRef]
  28. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  29. Liu, J.; Wu, Z.; Wei, Z.; Xiao, L.; Sun, L. Spatial-spectral kernel sparse representation for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 2462–2471. [Google Scholar] [CrossRef]
  30. Dong, W.; Fu, F.; Shi, G.; Cao, X.; Wu, J.; Li, G.; Li, X. Hyperspectral image super-resolution via non-negative structured sparse representation. IEEE Trans. Image Process. 2016, 25, 2337–2352. [Google Scholar] [CrossRef] [PubMed]
  31. Li, J.; Zhang, H.; Zhang, L. Supervised segmentation of very high resolution images by the use of extended morphological attribute profiles and a sparse transform. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1409–1413. [Google Scholar]
  32. Cui, M.; Prasad, S. Class-dependent sparse representation classifier for robust hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2683–2695. [Google Scholar] [CrossRef]
  33. Zou, J.; Li, W.; Du, Q. Sparse representation-based nearest neighbor classifiers for hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2418–2422. [Google Scholar]
  34. Bian, X.; Chen, C.; Xu, Y.; Du, Q. Robust Hyperspectral Image Classification by Multi-Layer Spatial-Spectral Sparse Representations. Remote Sens. 2016, 8, 985. [Google Scholar] [CrossRef]
  35. Iordache, M.-D.; Bioucas-Dias, J.M.; Plaza, A. Collaborative sparse regression for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2014, 52, 341–354. [Google Scholar] [CrossRef]
  36. Tropp, J.A.; Gilbert, A.C.; Strauss, M.J. Algorithms for simultaneous sparse approximation. Part I: Greedy pursuit. Signal Process. 2006, 86, 572–588. [Google Scholar] [CrossRef]
  37. Roscher, R.; Waske, B. Shapelet-based sparse representation for landcover classification of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1623–1634. [Google Scholar] [CrossRef]
  38. Mallat, S.G.; Zhang, Z. Matching pursuits with time-frequency dictionaries. IEEE Trans. Signal Process. 1993, 41, 3397–3415. [Google Scholar] [CrossRef] [Green Version]
  39. Yang, J.; Yu, K.; Huang, T. Supervised translation-invariant sparse coding. In Proceedings of the IEEE Conference Computer Visison and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 3517–3524. [Google Scholar]
  40. Zhang, H.; Li, J.; Huang, Y.; Zhang, L. A nonlocal weighted joint sparse representation classification method for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2056–2065. [Google Scholar] [CrossRef]
  41. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Marí, J.; Vila-Francés, J.; Calpe-Maravilla, J. Composite kernels for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2006, 3, 93–97. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed AJSM and MLSR methods.
Figure 1. Flowchart of the proposed AJSM and MLSR methods.
Remotesensing 10 00905 g001
Figure 2. False color composite images with spectral bands 50, 27 and 17 (left) and ground truth (middle) with legends (right) for the three test datasets: (a) Indian Pines; (b) University of Pavia; and (c) Salinas.
Figure 2. False color composite images with spectral bands 50, 27 and 17 (left) and ground truth (middle) with legends (right) for the three test datasets: (a) Indian Pines; (b) University of Pavia; and (c) Salinas.
Remotesensing 10 00905 g002aRemotesensing 10 00905 g002b
Figure 3. Classification maps of Indian Pines: (a) SVM; (b) EMP; (c) SRC; (d) JSM; (e) AJSM; and (f) MLSR.
Figure 3. Classification maps of Indian Pines: (a) SVM; (b) EMP; (c) SRC; (d) JSM; (e) AJSM; and (f) MLSR.
Remotesensing 10 00905 g003
Figure 4. Classification maps of the University of Pavia: (a) SVM; (b) EMP; (c) SRC; (d) JSM; (e) AJSM; and (f) MLSR.
Figure 4. Classification maps of the University of Pavia: (a) SVM; (b) EMP; (c) SRC; (d) JSM; (e) AJSM; and (f) MLSR.
Remotesensing 10 00905 g004
Figure 5. Classification maps of the Salinas scene: (a) SVM; (b) EMP; (c) SRC; (d) JSM; (e) AJSM; and (f) MLSR.
Figure 5. Classification maps of the Salinas scene: (a) SVM; (b) EMP; (c) SRC; (d) JSM; (e) AJSM; and (f) MLSR.
Remotesensing 10 00905 g005
Figure 6. The effect of controlling parameter α on classification results for three data sets.
Figure 6. The effect of controlling parameter α on classification results for three data sets.
Remotesensing 10 00905 g006
Figure 7. The effects of region scales on JSM, AJSM, and MLSR: (a) Indian Pines (b) Pavia University and (c) the Salinas scene.
Figure 7. The effects of region scales on JSM, AJSM, and MLSR: (a) Indian Pines (b) Pavia University and (c) the Salinas scene.
Remotesensing 10 00905 g007
Figure 8. The effect of number of patches of MLSR on three datasets.
Figure 8. The effect of number of patches of MLSR on three datasets.
Remotesensing 10 00905 g008
Figure 9. The effect of numbers of training data on five different methods: (a) Indian Pines; (b) University of Pavia; and (c) the Salinas scene.
Figure 9. The effect of numbers of training data on five different methods: (a) Indian Pines; (b) University of Pavia; and (c) the Salinas scene.
Remotesensing 10 00905 g009
Table 1. Class information for the Indian Pines dataset.
Table 1. Class information for the Indian Pines dataset.
Class Class NameTraining Test
1Alfalfa541
2Corn-no till1431285
3Corn-min till83747
4Corn24213
5Grass/trees49434
6Grass/pasture73657
7Grass/pasture-mowed325
8Hay-windrowed48430
9Oats218
10Soybeans-no till97875
11Soybeans-min till2462209
12Soybeans-clean till60533
13Wheat21184
14Woods1271138
15Buildings-grass-trees39347
16Stone-steel towers984
Total10299220
Table 2. Classification accuracies (%) for the Indian Pines image. The best results are shown in boldface.
Table 2. Classification accuracies (%) for the Indian Pines image. The best results are shown in boldface.
ClassSVMEMPSRCJSMAJSMMLSR
142.4070.4932.4874.0784.0992.60
275.0691.5573.3194.9792.1694.63
359.9185.6358.1291.8295.1699.88
450.9879.4947.5387.1593.6796.15
586.9795.8382.0496.6396.6793.36
693.8498.1989.3398.8898.6397.99
789.6696.3039.6894.1599.98100.00
899.57100.0093.2299.7999.1699.80
966.6792.8632.7378.5750.0082.00
1062.4985.9660.9691.1693.3097.83
1183.3194.3982.6195.6192.7197.41
1272.1788.9670.0092.0595.4391.37
1390.0498.0778.7499.5096.1099.53
1496.9398.7794.8399.0397.78100.00
1552.8283.5749.9089.8895.8498.95
1682.6190.8260.4094.5798.9190.53
OA75.4190.7765.8292.5294.7497.08
AA75.3490.6865.3792.7392.4895.75
Kappa73.7191.2069.9094.2594.0296.79
Table 3. Class information for the University of Pavia image.
Table 3. Class information for the University of Pavia image.
Class No.Class NameTraining Test
1Asphalt2506381
2Meadows25018,399
3Gravel2501849
4Trees2502814
5Meta sheets2501095
6Bare soil2504779
7Bitumen2501080
8Bricks2503432
9Shadows250697
Total225040,526
Table 4. Classification accuracies (%) for the University of Pavia image. The best results are shown in boldface.
Table 4. Classification accuracies (%) for the University of Pavia image. The best results are shown in boldface.
ClassSVMEMPSRCJSMAJSMMLSR
177.4284.3873.4686.8499.4196.32
296.3497.8195.3598.0792.5999.38
384.0391.4479.5391.4787.0999.90
472.4181.5868.0783.2398.5097.68
599.9299.9399.9299.1199.11100.00
682.6788.8178.5290.1990.16100.00
795.0796.6593.6497.2795.1999.92
892.3995.9789.7496.4684.8799.65
999.8999.5497.6586.54100.0095.04
OA88.9092.9086.2192.1393.3098.85
AA88.9292.9586.4793.7394.1098.65
Kappa84.3890.2280.7691.5191.2198.47
Table 5. Class Information for the Salinas image.
Table 5. Class Information for the Salinas image.
Class No.Class NameTraining Test
1Weeds_1301979
2Weeds_2563670
3Fallow301946
4Fallow plow211373
5Fallow smooth402638
6Stubble603899
7Celery543525
8Grapes16911,102
9Soil936110
10Corn493229
11Lettuce 4 week161052
12Lettuce 5 week291898
13Lettuce 6 week14902
14Lettuce 7 week161054
15Vineyard untrained1107158
16Vineyard trellis271780
Total81453,315
Table 6. Classification accuracies (%) for the Salinas image. The best results are shown in boldface.
Table 6. Classification accuracies (%) for the Salinas image. The best results are shown in boldface.
ClassSVMEMPSRCJSMAJSMMLSR
198.6299.5098.6799.3199.7597.83
299.6599.7699.6599.7099.5897.97
395.4497.9595.6397.0298.73100.00
497.2598.4397.3997.9498.9298.85
597.7398.3097.7698.1898.6299.51
6100.0099.90100.0099.9299.7799.92
798.1199.4298.1699.1199.8699.72
878.8790.7479.9186.7980.1889.77
999.5699.7799.6199.7698.9099.48
1090.6096.0391.1394.1293.7298.96
1189.0195.7589.7993.0799.44100.00
1296.0598.4596.4097.5599.07100.00
1394.7796.6794.8795.6297.6099.13
1487.1194.2587.5691.4496.1798.13
1559.0979.5060.6271.8782.4497.41
1698.2899.1798.3999.1698.7898.56
OA87.6494.3188.2091.9892.5897.25
AA92.5196.4792.8595.0496.3398.77
Kappa86.2993.6886.9191.0992.0096.94

Share and Cite

MDPI and ACS Style

Gao, Q.; Lim, S.; Jia, X. Improved Joint Sparse Models for Hyperspectral Image Classification Based on a Novel Neighbour Selection Strategy. Remote Sens. 2018, 10, 905. https://doi.org/10.3390/rs10060905

AMA Style

Gao Q, Lim S, Jia X. Improved Joint Sparse Models for Hyperspectral Image Classification Based on a Novel Neighbour Selection Strategy. Remote Sensing. 2018; 10(6):905. https://doi.org/10.3390/rs10060905

Chicago/Turabian Style

Gao, Qishuo, Samsung Lim, and Xiuping Jia. 2018. "Improved Joint Sparse Models for Hyperspectral Image Classification Based on a Novel Neighbour Selection Strategy" Remote Sensing 10, no. 6: 905. https://doi.org/10.3390/rs10060905

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop