Next Article in Journal
Lateral Offset Quality Rating along Low Slip Rate Faults: Application to the Alhama de Murcia Fault (SE Iberian Peninsula)
Previous Article in Journal
Hyperspectral Differentiation of Phytoplankton Taxonomic Groups: A Comparison between Using Remote Sensing Reflectance and Absorption Spectra
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Weighted-Fusion-Based Representation Classifiers for Hyperspectral Imagery

1
College of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100029, China
2
Department of Electrical and Computer Engineering, Mississippi State University, Starkville, MS 39762, USA
3
Intelligent Fusion Technology, Germantown, MD 20876, USA
*
Authors to whom correspondence should be addressed.
Remote Sens. 2015, 7(11), 14806-14826; https://doi.org/10.3390/rs71114806
Submission received: 17 June 2015 / Revised: 9 September 2015 / Accepted: 30 October 2015 / Published: 6 November 2015

Abstract

:
Spatial texture features have been demonstrated to be very useful for the recently-proposed representation-based classifiers, such as the sparse representation-based classifier (SRC) and nearest regularized subspace (NRS). In this work, a weighted residual-fusion-based strategy with multiple features is proposed for these classifiers. Multiple features include local binary patterns (LBP), Gabor features, and the original spectral signatures. In the proposed classification framework, representation residuals for a testing pixel from using each type of features are weighted to generate the final representation residual, and then the label of the testing pixel is determined according to the class yielding the minimum final residual. The motivation of this work is that different features represent pixels from different perspectives and their fusion in the residual domain can enhance the discriminative ability. Experimental results of several real hyperspectral image datasets demonstrate that the proposed residual-based fusion outperforms the original NRS, SRC, support vector machine (SVM) with LBP, and SVM with Gabor features, even in small-sample-size (SSS) situations.

Graphical Abstract

1. Introduction

Containing hundreds of spectral narrow bands, hyperspectral imagery has a very high spectral resolution that provides potential for more accurate object classification. Numerous classification algorithms [1,2] using hyperspectral data have been developed for a variety of applications, such as mineral exploration, environmental monitoring, etc. With the advance of sensor technology, hyperspectral images with high spatial resolution are becoming more available.
Recently, representation-based classification, which does not assume any data density distribution, has attracted great interest. In our previous work, nearest regularized subspace (NRS) [3] was proposed for hyperspectral image classification, where each testing pixel was represented by a linear combination of all available labeled samples per class and the class label was the one whose training samples provide the lowest representation residual. Moreover, a distance-weighted Tikhonov regularization was employed to impose adaptive penalty on representation coefficients of different training samples in the way that samples more similar to the testing pixel could be assigned with larger coefficients. The NRS has a closed-form solution, and has been demonstrated to be more efficient than the state-of-the-art classifiers, such as support vector machine (SVM) [2].
Sparse representation-based classification (SRC) [4] was originally introduced in computer vision [5,6], and first applied to hyperspectral image classification in [7]. Tang et al. [8] proposed a classifier based on the SRC, which outperformed some popular classification methods, such as the SVM, subspace projection (SP), orthogonal matching pursuit (OMP), and so on. Similar to the NRS, the SRC determines the class label of a testing pixel by solving a problem of sparse regression between training samples and a testing pixel. However, the solution of sparse coefficients may not be unique or stable when the atoms in a dictionary contain high mutual coherence [9,10]. In order to alleviate such problems, improvement strategies have been applied on the SRC with structured priors, including joint sparsity (JS) [11], low-rank (LR) Lasso [12], collaborative hierarchical Lasso (CHiLasso) [13], and so on.
With the improvement of spatial resolution, spatial information has been utilized for hyperspectral image classification [14,15,16,17,18,19]. For example, composite kernels (CK) for both spectral and spatial features were employed by a SVM classifier, referred to as SVM-CK [14]. Recently, spatial features, such as morphological attribute profile (MAP) [15,20,21], 2-D Gabor features [16,22], 3-D Gabor features [17], gray level co-occurrence matrix (GLCM) [18], local binary patterns (LBP) [19], and shape features [23], were investigated, and a number of spatial features were combined for the representation-based classifiers. In [16], the benefits of using spatial features extracted from a Gabor filter for the NRS were demonstrated. In [24], weighted joint collaborative representation (WJCR) classification was investigated. For face recognition, Lee et al. [25] introduced a robust method that applied the Gabor-edge components histogram on the SRC. A joint sparse representation model (JSRM) was proposed in [26,27,28]. Multiscale adaptive sparse representation was developed to reduce the sensitivity to region size in the JSRM [29].
Along with algorithm development for hyperspectral image classification, feature-level and decision-level fusion have been investigated [23,30,31]. Waske et al. [30] analyzed the robustness of multiple classifier systems based on the SVM and random feature selection, and it was demonstrated that the combination of the selected features for multiple classifiers could truly improve classification accuracy. In [31], an adaptive affinity propagation algorithm was proposed for dimension reduction for classification improvement. In [32,33], novel frameworks by combining supervised and unsupervised learning methods (e.g., SVM and fuzzy-c-means clustering algorithm) were developed. Decision fusion could also improve the performance of target and anomaly detection [34]. In [35], a wavelet-based NRS classifier for noise-robust hyperspectral image classification based on redundant discrete wavelet transformation was introduced. In [36], a dynamic classifier selection approach was presented, where both spatial and spectral information were used to determine the label of a testing pixel once the remaining neighborhood meets the threshold.
In this paper, a weighted-residual-fusion-based strategy with multiple features is proposed for representation-based classifiers. Multiple features include LBP features, Gabor features, and the original spectral signatures. In the proposed classification framework, each type of feature is used for the NRS or SRC, generating multiple representation residuals, and then all these residuals are added together with different weights [37] and the label of the testing pixel is determined according to the class yielding the minimum weighted sum of the residuals. The LBP and Gabor feature are employed due to the facts that the LBP has an excellent ability to describe local image texture features and a Gabor filter is able to extract global features. It is expected that different types of spatial features reflect the characteristics of a pixel from different perspectives and their fusion in the residual domain is able to enhance class separability even in small-sample-size (SSS) situations.

2. Proposed Classification Framework

2.1. Gabor-Filter

A Gabor filter [38] is an orientation-dependent band-pass filter with orientation sensitive but rotation invariant characteristics. Here, Gabor features are the magnitudes of each Gabor-filtered image, which is equivalent to signal power in the corresponding filter pass band. In a 2-D (a, b) coordinate system, a Gabor filter, including a real and imaginary term, is represented as,
G δ , θ , ϕ , σ σ ( a , b ) = exp ( a 2 + γ 2 b 2 2 σ 2 ) exp ( j ( 2 π a δ + ϕ ) )
where:
a = a cos θ + b sin θ
b = a sin θ + b cos θ
In Equation (1), δ is the wavelength of the sinusoidal factor, θ is the orientation separation angle (e.g., π 8 π 4 π 2 etc.) of Gabor kernels, φ represents the phase offset, σ is the standard derivation of the Gaussian envelope, and γ is the spatial aspects ratio specifying the ellipticity of the support of the Gabor function. ϕ = 0 and ϕ = π 2 return the real and imaginary parts of the Gabor filter, respectively. Parameter σ is determined by δ and spatial frequency bandwidth bw as,
σ = δ π ln 2 2 2 b w + 1 2 b w 1

2.2. LBP

Recently, the original LBP [39] and its variants have been used to extract rotation invariant features for classification. An entire image is convoluted by a small mask, e.g., a 3 × 3 window; for each central pixel in the window, its gray value is compared with its surrounding 8 pixels; a binary label is then obtained, which can be either “0” or “1” (0 means a smaller gray value and 1 means a larger one). For a general case, let m represent the number of surrounding neighbors and let these pixels be denoted as { g i } i = 0 m 1 . For the center pixel gc, its LBP code is calculated as,
L B P m , r ( g c ) = i = 0 m 1 U ( g i g c ) 2 i
where the function U(.) is defined as
U ( g ) = { 1 , g i g c 0 , g i < g c
As illustrated in Figure 1, for each pixel, a binary number is obtained by concatenating all these binary values in a clockwise direction, which starts from one of its top-middle neighbors. It is easy to observe that such an LBP operator actually represents the texture orientation and smoothness information in a local region. For a 3 × 3 mask, the eight labels compose the center pixel’s binary numbers that can be further converted into decimal numbers as an LBP code. Then, an occurrence histogram, as a nonparametric statistical estimate, is calculated; during the process, a binning procedure is required to ensure that the extracted histogram features have the same dimensions. Furthermore, to deal with textures at different scales, the LBP operator may use windows of different sizes [39]. The original rotation invariant LBP is achieved by circularly rotating each bit pattern to the minimum value. In our work, a variant of LBP called “uniform-LBP” is employed, where the histogram has a separate bin for every uniform pattern, while all non-uniform patterns are assigned to a single bin. Here, a pattern is called uniform if the binary pattern contains at most two circular 0–1 and 1–0 transitions. Selecting uniform patterns appropriately can reduce the length of the feature vector, thereby improving classification performance [19,40,41].
Figure 1. An example of LBP binary thresholding: (a) a 3 × 3 window; and (b) binary labels of the surrounding eight neighbors (in the clockwise direction).
Figure 1. An example of LBP binary thresholding: (a) a 3 × 3 window; and (b) binary labels of the surrounding eight neighbors (in the clockwise direction).
Remotesensing 07 14806 g001

2.3. NRS

Consider a dataset with training samples X = { x i } i = 1 n in a d-dimensionality space with class labels being denoted as Q i { 1 , 2 , , C } , where C is the number of classes, n is the total number of training samples. Let the number of training samples in the l-th class be represented as nl. In the NRS, yl represents the class-specific approximation in the l-th class of a test pixel y, which is calculated via a linear combination of available training samples in the l-th class Xl, with the weight factor αl as,
y l = X l α l
The optimal αl can be solved by an 2−norm regularization described as,
α l = arg min α l y X l α l 2 2 + λ Γ l , y α l 2 2
where λ is a regularization parameter , and the biasing Tikhonov matrix defined as,
Γ l , y = [ y x l , 1 2 0 0 y x l , n l 2 ]
which can reflect the data locality structure in the calculated weight coefficients; that is, the labeled samples that are most dissimilar to the testing pixel will provide a much smaller contribution to the linear representation. A closed-form solution of the weight vector αl in Equation (8) can be directly calculated as,
α l = ( X l T X l + λ 2 Γ l , y T Γ l , y ) 1 X l T y
After the weight vector is obtained, the class label of the testing pixel is determined according to the class that minimizes the Euclidean distance between yl and y, i.e.,
r l ( y ) = y l y 2 = X l α l y 2
and class
( y ) = arg min l = 1 , , C r l ( y )

2.4. SRC

The basic idea of SRC is to represent the testing pixel y as a sparse linear combination of a training sample dictionary X containing samples from all the C classes. Then, the sparse representation coefficients α = [ α 1 , α i , , α c ] can be estimated by solving the following l1 minimization problem:
α ^ = arg min α y X α 2 2 + λ α 1
In our work, the l1-norm minimization is solved using the l1_ls package (SRC with the l1-norm minimization is implemented using the l1_ls package [42]. Then, similarly, the label of the testing pixel is determined by computing the representation residual per class.

2.5. Residual-Fusion-Based-Collaborative Representation

Figure 2 illustrates the flowchart of the proposed weighted-residual-fusion-based classifier (NRC or SRC) that merges three representation residuals from using three types of features (i.e., LBP features, Gabor features, and the original spectral features). During the feature exaction process, band-selection-based dimensionality reduction is conducted with the linear prediction error (LPE) criterion [43,44], which is an unsupervised band selection method to find a small set of distinctive and informative bands. Based on the concept that a band yielding the maximum reconstruction error is considered as the most dissimilar band, LPE begins with the best two-band combination, and then augments this two-band combination to three, four, and so on, until a desired number of bands is selected. For each selected band, the LBP operator and the Gabor filter are used to extract spatial features as described previously. As shown in Equation (11), each NRS and SRC classifier (before assigning class membership) produces a certain representation residual. The residuals from using different types of features are combined to generate a weighted sum as
r l ( y ) = w 1 r l S ( y ) + w 2 r l L ( y ) + w 3 r l G ( y ) s.t.     w 1 + w 2 + w 3 = 1 ,   w 1 0     w 2 0 , w 3 0
where r l S ( y ) , r l L ( y ) and r l G ( y ) are the residuals when using the l-th class’ labeled samples with the original spectral features (the resulting classifiers are denoted as Spec-NRS for NRS or Spec-SRC for SRC), the LBP features (denoted as LBP-NRS or LBP-SRC), and with Gabor features (denoted as Gabor-NRS or Gabor-SRC), respectively. The class label of the testing pixel y is determined by the one producing the minimum weighted sum of residual. Here, w1, w2, and w3 are the weights of residuals when using the spectral signatures, LBP features, and Gabor features, respectively. A larger weight means the corresponding feature is more important in decision making, which is obviously data and class dependent. Note that these weights are imposed with non-negativity and sum-to-one constraints.
Figure 2. Flowchart of the proposed residual-fusion-based strategy with multiple features and representation-based classifiers.
Figure 2. Flowchart of the proposed residual-fusion-based strategy with multiple features and representation-based classifiers.
Remotesensing 07 14806 g002
In the proposed weighted-residual-fusion-based classifier, both LBP and Gabor features are extracted to present spatial information. Note that Gabor features are produced by the average magnitude response for each Gabor filtered image reflecting the global signal power, while the LBP-coded image is an expression of detailed local spatial features, such as edges, corners, and knots. Thus, Gabor filtering is a supplement to the local LBP that lacks the consideration of distant pixel interactions. It is expected that different features represent a testing pixel from different perspectives, and fusion of these features can enhance discriminative power.

3. Experimental Section

In this section, the performance of the proposed weighted-residual-fusion-based representation classifiers, denoted as RF-NRS and RF-SRC, on two real hyperspectral datasets is investigated. The classification results are compared with other algorithms, such as the original NRS or SRC using spectral signatures only, LBP-NRS or LBP-SRC using LBP features only, Gabor-NRS or Gabor-SRC using Gabor features only, and SVM using Gabor-filtering and LBP features. All experimental data are downloaded from a public website [45].
The first experimental data were collected by the Reflective Optics System Imaging Spectrometer (ROSIS) sensor [46]. The image scene, with a spatial size of 610 × 340 pixels, covers the city of Pavia, Italy. The data set has 91 spectral bands after water-band removal. It has a spectral coverage from 0.43–0.86 um and a spatial resolution of 1.3 m. There are nine classes from the ground truth map.
The second data were collected by the AVIRIS sensor over an area in Salinas Valley, California with a spatial resolution of 3.7 m. The image comprises 512 × 217 pixels with 204 bands (after 20 water absorption bands are removed). It mainly contains vegetables, bar soils, and vineyard fields. There are 16 classes from the ground truth map. With the same sensor, another scene was collected over northwest Indiana’s Indian Pine test site in 1992. The scene consists of 145 × 145 pixels with a spatial resolution of 20 m. There are also 16 different land-cover classes.
Table 1. Labeled samples for the University of Pavia dataset.
Table 1. Labeled samples for the University of Pavia dataset.
ClassNumber of Samples
NO.Name
1Asphalt6631
2Meadows18,649
3Gravel2099
4Trees3064
5Painted metal sheets1345
6Bare soil5029
7Bitumen1330
8Self-blocking bricks3682
9Shadows947
Total42,776
The information of classes and their labeled samples for both the Salinas and University of Pavia datasets are given in Table 1 and Table 2. All experiments are carried out using MATLAB on an Intel i7 quad core 3.40-GHz machine with 10 GB of RAM.
Table 2. Labeled samples for the Salinas dataset.
Table 2. Labeled samples for the Salinas dataset.
ClassNumber of Samples
NO.Name
1Brocoli_green_weeds_12009
2Brocoli_green_weeds_23726
3Fallow1976
4Fallow_rough_plow1394
5Fallow_smooth2678
6Stubble3959
7Celery3579
8Grapes_untrained11,271
9Soil_vinyard_develop6203
10Corn_senesced_green_weeds3278
11Lettuce_romaine_4wk1068
12Lettuce_romaine_5wk1927
13Lettuce_romaine_6wk916
14Lettuce_romaine_7wk1070
15Vinyard_untrained7268
16Vinyard_vertical_trellis1807
Total50,550

3.1. Parameter Tuning

Several significant parameters for the proposed classification framework are tuned with the leave-one-out cross validation (LOOCV) strategy based on available training samples. For the parameters of Gabor filtering, all the eight orientations [ 0 π 8 π 4 3 π 8 π 2 5 π 8 3 π 4 7 π 8 ] shown in Figure 3 are used in this work according to [47], and the parameter bw in Equation (4) is set to five and 10 bands for the Gabor filter according to our empirical study shown in Figure 4. Note that selecting more bands may not significantly improve classification accuracy, but definitely increases computational cost due to the resulting higher feature dimensionality. Thus, only three bands are selected for the LBP operator and 10 bands for the Gabor filter. Figure 5 further illustrates the effect of patch sizes on the LBP. It can be seen that classification accuracy tends to be the maximum with 21 × 21 patch size for the University of Pavia data.
To estimate an optimal λ, Figure 6 shows accuracy at different values of λ for NRS and SRC classifiers. The number of training samples per class is 30 in the Salinas dataset, and 60 in the University of Pavia data. In different datasets, the optimal values of λ are different as shown in Figure 6. In the Salinas data, the optimal parameter is 0.5 for NRS and 0.1 for SRC; in the University of Pavia data, it is 1 for NRS and 0.1 for SRC.
Figure 3. (a) Band 95 of the University dataset. (b)–(i) Eight Gabor feature images corresponding to eight orientations of the band.
Figure 3. (a) Band 95 of the University dataset. (b)–(i) Eight Gabor feature images corresponding to eight orientations of the band.
Remotesensing 07 14806 g003
Another three important parameters are w1, w2 and w3. Table 3, Table 4, Table 5 and Table 6 demonstrate classification accuracy versus the values of w1 and w2. It is interesting to observe that in University of Pavia dataset, the Gabor features provide a better performance, and the LBP features need higher weights in Salinas dataset explaining the high classification accuracy of LBP-NRS in Figure 6a,b. As shown in Table 3, Table 4, Table 5 and Table 6, the best values of w1, w2 and w3 are: 0.2, 0.3, 0.5; 0.6, 0.1, 0.4; 0.2, 0.7, 0.1; 0.6(0.7), 0.1, 0.3 (0.2), respectively. The weight of each residual indicates the importance of the corresponding feature. For example, in Table 3, as Gabor features perform well in the University of Pavia dataset, its weight is 0.5, much higher than the other two; in Table 4 for the Salinas dataset, where the spectral signatures actually provide the highest classification accuracy, the corresponding weight is 0.6, larger than others.
Figure 4. Classification accuracy versus the number of selected bands for: (a) University of Pavia dataset (60 training samples per class); (b) Salinas dataset (30 training samples per class).
Figure 4. Classification accuracy versus the number of selected bands for: (a) University of Pavia dataset (60 training samples per class); (b) Salinas dataset (30 training samples per class).
Remotesensing 07 14806 g004
Figure 5. Impact of patch size on LBP generation using the University of Pavia dataset.
Figure 5. Impact of patch size on LBP generation using the University of Pavia dataset.
Remotesensing 07 14806 g005

3.2. Classification Performance

The effectiveness of the proposed RF-NRS classifier is evaluated by the comparison with the original NRS [3], Gabor-NRS [16], and LBP-NRS; similarly, RF-SRC is compared with the original SRC [6], Gabor-SRC, and LBP-SRC. In the University of Pavia dataset, the number of training samples per class is varied from 30–100. In the dataset of Salinas Valley, the number of training samples per class is from 10–30. All the samples are selected randomly to avoid bias.
Figure 6. Classification accuracy (%) on two datasets using different parameters for λ: (a) and (b) Salinas dataset with 30 training samples per class, and (c) and (d) University of Pavia dataset with 60 training samples per class.
Figure 6. Classification accuracy (%) on two datasets using different parameters for λ: (a) and (b) Salinas dataset with 30 training samples per class, and (c) and (d) University of Pavia dataset with 60 training samples per class.
Remotesensing 07 14806 g006
Table 3. Classification accuracy (%) versus the values of w1 and w2 for RF-NRS in University of Pavia dataset with 60 training samples per class.
Table 3. Classification accuracy (%) versus the values of w1 and w2 for RF-NRS in University of Pavia dataset with 60 training samples per class.
w100.10.20.30.40.50.60.70.80.91
w2
00.82570.96370.97960.98330.98350.98230.97950.97550.96670.94640.9067
0.10.93120.98130.98740.98840.98780.98550.98160.97610.95950.9248-
0.20.95960.98720.99150.99120.98990.98740.98310.97240.9394--
0.30.96970.9890.99230.99220.99080.98730.9810.9536---
0.40.9730.98880.99190.99210.99040.98560.9666----
0.50.97250.98690.99080.99110.98810.9753-----
0.60.97010.98390.98780.98770.9794------
0.70.96670.97970.98290.9784-------
0.80.96250.97360.973--------
0.90.95520.9626---------
10.9428----------
Table 4. Classification accuracy (%) versus the values of w1 and w2 for RF-NRS in Salinas dataset with 30 training samples per class.
Table 4. Classification accuracy (%) versus the values of w1 and w2 for RF-NRS in Salinas dataset with 30 training samples per class.
w100.10.20.30.40.50.60.70.80.91
w2
00.87590.96520.97950.98390.98520.98530.98560.98520.98370.97870.9667
0.10.90060.9690.98080.98480.98560.98580.98590.98490.98170.9755-
0.20.91740.9690.97950.98320.98460.98550.98480.98270.9785--
0.30.93010.9690.97750.98120.98310.98360.9820.9792---
0.40.93740.96850.97590.97940.9810.98110.9792----
0.50.94070.9670.97390.97730.97850.9779-----
0.60.9410.96520.97180.97490.9756------
0.70.94010.96330.96990.9716-------
0.80.93820.96090.967--------
0.90.93650.9575---------
10.9332----------
Table 5. Classification accuracy (%) versus the values of w1 and w2 for RF-SRC in University of Pavia dataset with 60 training samples per class.
Table 5. Classification accuracy (%) versus the values of w1 and w2 for RF-SRC in University of Pavia dataset with 60 training samples per class.
w100.10.20.30.40.50.60.70.80.91
w2
00.73590.82970.88950.92690.94950.9620.96760.96280.94910.92780.901
0.10.75810.84860.90720.9410.96070.97020.9710.96050.940.9146-
0.20.77730.86780.9230.95370.96940.97570.97080.95370.9246--
0.30.79510.88680.93830.96450.97650.97710.9650.9387---
0.40.81170.90450.95230.9740.98030.97220.9531----
0.50.82860.92380.9650.98110.97710.9612-----
0.60.84570.9420.97530.9810.9667------
0.70.8640.95870.98140.9712-------
0.80.88540.97010.9745--------
0.90.90680.9678---------
10.9215----------
To validate the effectiveness of the proposed RF-NRS and RF-SRC, the original NRS, SRC, LBP-NRS, LBP-SRC, Gabor-NRS, and Gabor-SRC are compared. Figure 7 illustrates the classification accuracy (%) of these methods versus the number of training samples per class for the two experimental datasets. It is obvious that the accuracy of all the classifiers increases with more training samples, and the classifiers using both spectral and spatial information are much better than the one with solely spectral information. For instance, in Figure 7d, the classification accuracy of the original SRC is at least 23% lower than that of RF-SRC with 60 training samples per class. Furthermore, the proposed RF-based classifiers are consistently much better than that of LBP-NRS, Gabor-NRS, LBP-SRC, and Gabor-SRC. Taking the Salinas Valley as an example, the accuracy difference between RF-NRS and LBP-NRS is approximately 2%. When the number of training samples per class is 30 for the University of Pavia dataset, the accuracy of RF-NRS is 6% higher than the others, indicating the proposed method is more robust in SSS situations. For the Salinas dataset, Figure 7 shows that the classification accuracy of RF-SRC is 1% higher than that of LBP-SRC. For the University of Pavia dataset, RF-SRC obviously outperforms others, with an approximate 4% improvement.
Table 6. Classification accuracy (%) versus the values of w1 and w2 for RF-SRC in the Salinas dataset with 30 training samples per class.
Table 6. Classification accuracy (%) versus the values of w1 and w2 for RF-SRC in the Salinas dataset with 30 training samples per class.
w100.10.20.30.40.50.60.70.80.91
w2
00.88530.93870.9580.96710.97240.97530.97660.9770.97660.97370.9622
0.10.8880.94170.96030.96910.97360.97610.97710.97710.97560.9702-
0.20.89010.94440.96190.96980.9740.9760.97630.97530.9717--
0.30.89280.9470.96350.97080.97430.97510.9750.9717---
0.40.89560.95020.96490.97090.97340.97410.9716----
0.50.89870.95390.96650.97170.97260.9711-----
0.60.9020.95750.96790.97090.97------
0.70.90660.96050.96850.969-------
0.80.9110.96220.9664--------
0.90.91610.9597---------
10.9144----------
Figure 7. Classification accuracy (%) versus different numbers of training samples per class for two experimental datasets: (a) and (b) Salinas Valley dataset; (c) and (d) University of Pavia dataset.
Figure 7. Classification accuracy (%) versus different numbers of training samples per class for two experimental datasets: (a) and (b) Salinas Valley dataset; (c) and (d) University of Pavia dataset.
Remotesensing 07 14806 g007
The classification accuracy for each class and overall accuracy (OA) are listed in Table 7 and Table 8 for the two datasets. For the Salinas dataset, 30 samples are randomly selected from each class for training and the rest for testing, while for the University of Pavia dataset, the number of training samples is 60 per class. As can be seen in Table 7 and Table 8, using spatial features enhances classification accuracy. For example, in Table 7, accuracy is increased by approximately 8% with the integration of the LBP feature, and Table 8 shows the Gabor feature brings about 9% improvement. Figure 8 and Figure 9 are the thematic maps of these hyperspectral datasets. Clearly, the classification maps of the proposed residual-fusion methods are less noisy and more accurate, which are consistent with the results listed in Table 7 and Table 8.
Table 7. Classification accuracy (%) for the Salinas dataset.
Table 7. Classification accuracy (%) for the Salinas dataset.
ClassSamplesLBP-NRSGabor-NRSNRSRF-NRSLBP-SRCGabor-SRCSRCRF-SRCLBP-SVMGabor-SVM
TrainTest
130197910098.9599.1510099.899.0599.3510010099.24
230369697.0596.7899.699.4696.5194.6692.3899.4994.5396.19
330194610099.8598.8910010092.7690.0810099.6999.49
430136499.2898.1398.4910098.2898.0699.7199.7195.5397.73
530264895.2997.7697.3598.1796.1298.3296.8397.6195.9699.06
630392997.1799.9799.4210097.1799.7799.7799.9597.2399.72
730354997.1599.8699.4410096.4599.1999.599.9797.2199.86
8301124193.4378.1269.9694.3792.579.579.991.5393.3477.36
930617399.2198.0596.9499.6198.3197.4599.5899.7198.6498.67
1030324899.3998.2394.8799.9799.3396.7791.7699.6699.3897.14
1130103810098.2296.6399.7299.8197.9499.2599.6310098.17
1230189796.4210010010095.910098.1399.9597100
133088692.997.1699.6799.3494.2195.7497.1697.699.2398.87
1430104097.4898.8892.0699.9195.3398.0493.1898.696.8396.83
1530723894.8584.373.3996.3194.779.5560.6194.792.0584.84
1630177710099.6798.1710099.7298.7898.2899.1799.7299.04
OA--96.6792.4588.8798.1496.2291.4488.5397.296.0192.29
Table 8. Classification accuracy (%) for the University of Pavia dataset.
Table 8. Classification accuracy (%) for the University of Pavia dataset.
ClassSamplesLBP-NRSGabor-NRSNRSRF-NRSLBP-SRCGabor-SRCSRCRF-SRCLBP-SVMGabor-SVM
TrainTest
160657183.1991.3180.7799.5383.1286.0876.3498.9179.9780.5
2601858992.4394.0184.7498.6691.5993.0566.6894.9889.9290.55
360203994.3396.986.7199.7195.0994.2474.1899.2990.7390.34
460300473.8996.8798.499.4572.3296.6194.3999.3581.9594.49
560128595.3210099.9310094.2810099.5599.9394.05100
660496997.9796.982.0499.5696.8294.5983.1899.5498.4995.32
760127097.8994.8993.6899.9298.3594.1472.9398.6597.8994.45
860362294.6291.0976.7599.7695.0686.2360.9798.3293.0687.78
96088783.4295.6799.6899.4783.5394.1910098.179.8295.09
OA--90.6794.2885.2899.290.192.1573.9597.2889.2890.08
Figure 8. Thematic maps for the University of Pavia dataset with 60 training samples per class; (a) Pseudo-color Image; (b) Ground truth Map; (c) NRS; (d) SRC; (e) NRS + Gabor; (f) NRS + LBP; (g) SRC + Gabor; (h) SRC + LBP; (i) RF-NRS; (j) RF-SRC.
Figure 8. Thematic maps for the University of Pavia dataset with 60 training samples per class; (a) Pseudo-color Image; (b) Ground truth Map; (c) NRS; (d) SRC; (e) NRS + Gabor; (f) NRS + LBP; (g) SRC + Gabor; (h) SRC + LBP; (i) RF-NRS; (j) RF-SRC.
Remotesensing 07 14806 g008
Figure 9. Thematic maps for the Salinas dataset with 30 training samples per class; (a) Pseudo-color Image; (b) Ground truth Map; (c) NRS; (d) SRC; (e) NRS + Gabor; (f) NRS + LBP; (g) SRC + Gabor (h) SRC + LBP; (i) RF-NRS; (j) RF-SRC.
Figure 9. Thematic maps for the Salinas dataset with 30 training samples per class; (a) Pseudo-color Image; (b) Ground truth Map; (c) NRS; (d) SRC; (e) NRS + Gabor; (f) NRS + LBP; (g) SRC + Gabor (h) SRC + LBP; (i) RF-NRS; (j) RF-SRC.
Remotesensing 07 14806 g009
Classification results from the aforementioned classifiers using the Indian Pines data are shown in Table 9 with the same number of training and testing samples as in [48]. It is apparent that the proposed RF-NRS and RF-SRC still provide superior performance, which further affirms that classification accuracy can be greatly improved by fusing two complementary spatial features (i.e., Gabor features and LBP features).
In Figure 10, the classification results of the proposed methods are compared with SVM using spatial features, i.e., LBP-SVM [49] and Gabor-SVM [50]. As can be seen, the classification accuracy of RF-NRS is more than 1% higher than that of RF-SRC. It is due to the calculation of weight coefficients is sparse in SRC, which is prone to misclassify a testing sample when the number of training samples is limited or the quality of training samples is poor. The proposed RF-NRS and RF-SRC have higher classification accuracy (at least 4%) than Gabor-SVM and LBP-SVM especially in the University of Pavia dataset. Moreover, with the increase of the number of training samples, the proposed RF-NRS and RF-SRC outperform Gabor-SVM and LBP-SVM. It also illustrates that the residual-based fusion methods provide even more solid and robust performance in SSS situations (e.g., 30 training samples per class).
Table 9. Classification accuracy (%) for the Indian Pine dataset.
Table 9. Classification accuracy (%) for the Indian Pine dataset.
ClassSamplesLBP-NRSGabor-NRSRF-NRSLBP-SRCGabor-SRCRF-SRCSVM-LBPSVM-Gabor
TrainTest
1153910096.390.7410096.310010094.87
250138487.9580.0692.6888.2171.699186.6377.96
35078497.7290.6598.5696.8881.7798.4498.4788.61
45018410098.7299.1598.7295.399.1599.46100
55044797.1896.5898.9997.1893.9697.7997.9994.85
65069797.5996.9297.9996.9293.0498.896.797.56
71511100100100100100100100100
850439100100100100100100100100
9155100100100100100100100100
105091886.8885.1290.3988.0282.5488.3386.9375.49
1150241889.6782.4694.0489.2678.0888.5386.3573.9
125057491.2193.1697.7291.5394.395.9390.690.43
135016299.0698.1110098.5898.5899.5310099.38
1450124496.9993.669997.7695.2999.6997.5196.78
155033099.2196.3210099.4795.5310099.3998.18
165045100100100100100100100100
OA--93.1689.1296.0493.2385.7394.3192.0885.43
Figure 10. Classification accuracy (%) versus the number of training samples per class for: (a) University of Pavia dataset; (b) Salinas dataset.
Figure 10. Classification accuracy (%) versus the number of training samples per class for: (a) University of Pavia dataset; (b) Salinas dataset.
Remotesensing 07 14806 g010

4. Conclusions

In this paper, a weighted-residual-fusion-based classification framework was proposed. The overall classification utilized multiple features, including LBP features, Gabor features, and the original spectral features. A representation-based classifier, such as NRS or SRC, was applied to each type of feature, and fusion was achieved by a weighted combination of their residuals. It was found that the resulting classifiers, i.e., RF-NRS and RF-SRC, were more discriminative than the original spectral classifiers and the classifiers with the Gabor feature and the LBP feature only. Experimental results from several real hyperspectral images demonstrated that the proposed residual-fusion-based classification methods consistently outperformed the traditional classifiers and other state-of-the-art classification algorithms (e.g., Gabor-SVM and LBP-SVM) when the number of training samples per class was varied. Specifically, the proposed RF-NRS achieved about 6% classification accuracy improvement over the traditional Gabor-NRS and Gabor-SVM, and about 2% over the LBP-NRS and LBP-SVM for the Salinas dataset when 30 training samples per class were chosen. RF-NRS outperformed Gabor-NRS by about 5%, LBP-NRS by about 9%, LBP-SVM by about 10%, and Gabor-SVM by about 9% for the University of Pavia data when 60 training samples per class were chosen. In addition, the proposed RF-NRS and RF-SRC tended to be more robust in SSS situations.
In future work, the spectral-local-global feature fusion strategy will be exploited with more state-of-the-art features, and automatic estimation of optimal/suboptimal weighting parameters will be investigated.

Acknowledgments

This work was supported by the National Natural Science Foundation of China under Grants No. NSFC-61571033, 61302164, and partly by the Research Fund for Basic Researches in Central Universities under Grant No. YS1404.

Author Contributions

All authors conceived and designed the study. Bing Peng and Wei Li carried out the experiments. All authors discussed the basic structure of the manuscript, and Bing Peng finished the first draft. Wei Li, Xiaoming Xie, Qian Du, and Kui Liu reviewed and edited the draft.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Landgrebe, D.A. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  2. Li, W.; Prasad, S.; Fowler, J.E. Decision fusion in kernel-induced spaces for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3399–3411. [Google Scholar] [CrossRef]
  3. Li, W.; Tramel, E.W.; Prasad, S.; Fowler, J.E. Nearest regularized subspace for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 477–489. [Google Scholar] [CrossRef]
  4. Wright, J.; Ma, Y.; Mairal, J.; Sapiro, G.; Huang, T.S.; Yan, S. Sparse representation for computer vision and pattern recognition. Proc. IEEE. 2010, 98, 1031–1044. [Google Scholar] [CrossRef]
  5. Wright, J.; Yang, A.; Ganesh, A.; Sastry, S.; Ma, Y. Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 210–227. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, H.; Nasrabadi, N.; Zhang, Y.; Huang, T. Joint Dynamic Sparse Representation for Multiview Face Recognition. Pattern Recognit. 2012, 45, 1290–1298. [Google Scholar] [CrossRef]
  7. Chen, Y.; Nasrabadi, M.; Tran, T. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  8. Tang, Y.; Li, X.; Liu, Y.; Wang, J.; Liu, C.; Liu, S. Hyperspectral image classification using sparse representation-based classifier. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Quebec City, QC, Canada, 13–18 July 2014; pp. 3450–3453.
  9. Iordache, M.; Bioucas-Dias, J.; Plaza, A. Sparse unmixing of hyperspectral data. IEEE Geosci. Remote Sens. 2011, 49, 2014–2039. [Google Scholar] [CrossRef]
  10. Sun, X.; Qu, Q.; Nasrabadi, N.M.; Tran, T.D. Structured priors for sparse-representation-based hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1235–1239. [Google Scholar] [CrossRef]
  11. Berg, E.; Friedlander, M. Joint-sparse recovery from multiple measurements. IEEE Trans. Inf. Theor. 2010, 56, 2516–2527. [Google Scholar] [CrossRef]
  12. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 171–184. [Google Scholar] [CrossRef] [PubMed]
  13. Sprechmann, P.; Ramirez, I.; Sapiro, G.; Eldar, Y. C-HiLasso: A collaborative hierarchical sparse modeling framework. IEEE Trans. Signal Process. 2011, 59, 4183–4198. [Google Scholar] [CrossRef]
  14. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Marí, J.; Vila-Francés, J.; Calpe-Maravilla, J. Composite kernels for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2006, 3, 93–97. [Google Scholar] [CrossRef]
  15. Song, B.; Li, J.; Dalla Mura, M.; Li, P.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A.; Chanussot, J. Remotely sensed image classification using sparse representations of morphological attribute profiles. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5122–5136. [Google Scholar] [CrossRef]
  16. Li, W.; Du, Q. Gabor-filtering-based nearest regularized subspace for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 1012–1022. [Google Scholar] [CrossRef]
  17. Jia, S.; Jia, L.; Li, Q. Gabor feature-based collaborative representation for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2015, 52, 1118–1129. [Google Scholar]
  18. Tsai, F.; Tsai, J.S. Feature extraction of hyperspectral image cubes using three dimensional gray level co-occurrence. IEEE Trans. Geosci. Remote Sens. 2013, 5, 3504–3514. [Google Scholar] [CrossRef]
  19. Li, W.; Chen, C.; Su, H.; Du, Q. Local binary patterns and extreme learning machine for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3681–3693. [Google Scholar] [CrossRef]
  20. Quesada-Barriuso, P.; Arguello, F.; Heras, D.B. Spectral-Spatial Classification of Hyperspectral Images Using Wavelets and Extended Morphological Profiles. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 1177–1185. [Google Scholar] [CrossRef]
  21. Ghamisi, P.; Benediktsson, J.A.; Gavallaro, G.; Plaza, A. Automatic Framework for Spectral-Spatial Classification Based on Supervised Feature Extraction and Morphological Attribute Profiles. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 2147–2160. [Google Scholar] [CrossRef]
  22. Rajadell, O.; Garcia-Sevilla, P.; Pla, F. Spectral-Spatial Pixel Characterization Using Gabor Filters for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2013, 10, 860–864. [Google Scholar] [CrossRef]
  23. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On combining multiple features for hyperspectral remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 879–893. [Google Scholar] [CrossRef]
  24. Xiong, M.; Ran, Q.; Li, W.; Zou, J.; Du, Q. Hyperspectral image classification using weighted joint collaborative representation. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1209–1213. [Google Scholar] [CrossRef]
  25. Lee, H.; Chung, Y.; Yoo, J.H.; Won, C. Face recognition based on sparse representation classifier with gabor-edge components histogram. In Proceedings of the 8th International Conference on Signal Image Technology and Internet Based Systems (SITIS), Naples, Italy, 25–29 November 2012; pp. 105–109.
  26. Li, W.; Du, Q. Joint within-class collaborative representation for hyperspectral image classification. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 2200–2208. [Google Scholar] [CrossRef]
  27. Tropp, J.A.; Gilbert, A.C.; Strauss, M.J. Algorithms for simultaneous sparse approximation. Part I: Greedy pursuit. IEEE Signal Process. 2006, 86, 572–588. [Google Scholar] [CrossRef]
  28. Wang, J.; Jiao, L.; Liu, H.; Yang, S.; Liu, F. Hyperspectral image classification by spatial-spectral derivative-aided kernel joint sparse representation. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 2485–2500. [Google Scholar] [CrossRef]
  29. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral-spatial hyperspectral image classification via multiscale adaptive sparse representation. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7738–7749. [Google Scholar] [CrossRef]
  30. Waske, B.; Linden, S.; Benediktsson, J.A.; Rabe, A.; Hostert, P. Sensitivity of support vector machines to random feature selection in classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2880–2889. [Google Scholar] [CrossRef]
  31. Su, H.; Sheng, Y.; Du, P.; Liu, K. Adaptive affinity propagation with spectral angle mapper for semi-supervised hyperspectral band selection. Appl. Opt. 2012, 51, 2656–2663. [Google Scholar] [CrossRef] [PubMed]
  32. Alajlan, N.; Bazi, Y.; Melgani, F.; Yager, R.R. Fusion of supervised and unsupervised learning for improved classification of hyperspectral images. Inf. Sci. 2012, 217, 39–55. [Google Scholar] [CrossRef]
  33. Yang, H.; Du, Q.; Ma, B. Decision fusion on supervised and unsupervised classifiers for hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2010, 7, 875–879. [Google Scholar] [CrossRef]
  34. Li, W.; Du, Q. Decision fusion for dual-window based hyperspectral anomaly detection. J. Appl. Remote Sens. 2015, 9. [Google Scholar] [CrossRef]
  35. Li, W.; Liu, K.; Su, H. Wavelet-based nearest-regularized subspace for noise-robust hyperspectral image classification. J. Appl. Remote Sens. 2014, 8, 083665. [Google Scholar] [CrossRef]
  36. Su, H.; Yong, B.; Du, P.; Liu, H.; Chen, C.; Liu, K. Dynamic classifier selection using spectral-spatial information for hyperspectral image classification. J. Appl. Remote Sens. 2014, 8, 085095. [Google Scholar] [CrossRef]
  37. Chi, Y.; Porikli, F. Classification and boosting with multiple collaborative representations. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1519–1531. [Google Scholar] [CrossRef] [PubMed]
  38. Daugman, J.G. Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters. J. Optim. Soc. Am. 1985, 2, 1160–1169. [Google Scholar] [CrossRef]
  39. Ojala, T.; Pietikainen, M.; Maenpaa, T.T. Multi resolution gray scale and rotation invariant texture classification with local binary pattern. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  40. Ahonen, T.; Hadid, A.; Pietikainen, M. Face description with local binary patterns: Application to face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 2037–2041. [Google Scholar] [CrossRef] [PubMed]
  41. Ren, J.; Jiang, X.; Yuan, J. Noise-resistant local binary pattern with an embedded error-correction mechanism. IEEE Trans. Image Process. 2013, 22, 4049–4060. [Google Scholar] [CrossRef] [PubMed]
  42. The l1_ls Package. Available online: https://stanford.edu/~boyd/l1_ls/ (accessed on 17 June 2015).
  43. Du, Q.; Yang, H. Similarity-based unsupervised band selection for hyperspectral image analysis. IEEE Geosci. Remote Sens. Lett. 2008, 5, 564–568. [Google Scholar] [CrossRef]
  44. Yang, H.; Du, Q.; Chen, G. Unsupervised hyperspectral band selection using graphics processing units. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2011, 4, 660–668. [Google Scholar] [CrossRef]
  45. Hyprespectral Remote Sensing Scenes. Available online: http://www.ehu.eus/ccwintco/index.php?title=Hyprespectral_Remote_Sensing_Scenes (accessed on 17 June 2015).
  46. Gamba, P. A collection of data for urban area characterization. In Proceedings of the IEEE International Geoscience and Remote Sens. Symposium, Anchorage, AK, USA, 20–24 September 2004; pp. 69–72.
  47. Chen, C.; Li, W.; Su, H.; Liu, K. Spectral-spatial classification of hyperspectral image based on kernel extreme learning machine. Remote Sens. 2014, 6, 5795–5814. [Google Scholar] [CrossRef]
  48. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in spectral-spatial classification of hyperspectral images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef]
  49. Tian, Q.; Tian, G. Development and testing of a LBP-SVM based teeth visibility recognizer. In Proceedings of the IEEE International Conference on Computational Intelligence for Measurement Systems and Applications (CIMSA), Tianjin, China, 2–4 July 2012; pp. 114–119.
  50. Huang, D.; Hu, W.; Chang, S. Vision-based hand gesture recognition using PCA + Gabor Filters and SVM. In Proceedings of the 5th International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP’09), Kyoto, Japan, 12–24 September 2009; pp. 1–4.

Share and Cite

MDPI and ACS Style

Peng, B.; Li, W.; Xie, X.; Du, Q.; Liu, K. Weighted-Fusion-Based Representation Classifiers for Hyperspectral Imagery. Remote Sens. 2015, 7, 14806-14826. https://doi.org/10.3390/rs71114806

AMA Style

Peng B, Li W, Xie X, Du Q, Liu K. Weighted-Fusion-Based Representation Classifiers for Hyperspectral Imagery. Remote Sensing. 2015; 7(11):14806-14826. https://doi.org/10.3390/rs71114806

Chicago/Turabian Style

Peng, Bing, Wei Li, Xiaoming Xie, Qian Du, and Kui Liu. 2015. "Weighted-Fusion-Based Representation Classifiers for Hyperspectral Imagery" Remote Sensing 7, no. 11: 14806-14826. https://doi.org/10.3390/rs71114806

Article Metrics

Back to TopTop