Next Article in Journal
ML-Based Analysis of Particle Distributions in High-Intensity Laser Experiments: Role of Binning Strategy
Next Article in Special Issue
A New Hybrid Possibilistic-Probabilistic Decision-Making Scheme for Classification
Previous Article in Journal
Structural Statistical Quantifiers and Thermal Features of Quantum Systems
Previous Article in Special Issue
Modified Gerchberg–Saxton (G-S) Algorithm and Its Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hyperspectral Image Classification Approach Based on Feature Fusion and Multi-Layered Gradient Boosting Decision Trees

1
State Key Laboratory of Geo-Information Engineering, Xi’an 710054, China
2
College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
3
College of Communication Engineering, Xidian University, Xi’an 710071, China
4
Business College, Central South University of Forestry and Technology, Changsha 410004, China
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(1), 20; https://doi.org/10.3390/e23010020
Submission received: 24 November 2020 / Revised: 22 December 2020 / Accepted: 23 December 2020 / Published: 25 December 2020

Abstract

:
At present, many Deep Neural Network (DNN) methods have been widely used for hyperspectral image classification. Promising classification results have been obtained by utilizing such models. However, due to the complexity and depth of the model, increasing the number of model parameters may lead to an overfitting of the model, especially when training data are insufficient. As the performance of the model mainly depends on sufficient data and a large network with reasonably optimized hyperparameters, using DNNs for classification requires better hardware conditions and sufficient training time. This paper proposes a feature fusion and multi-layered gradient boosting decision tree model (FF-DT) for hyperspectral image classification. First, we fuse extended morphology profiles (EMPs), linear multi-scale spatial characteristics, and nonlinear multi-scale spatial characteristics as final features to extract both special and spectral features. Furthermore, a multi-layered gradient boosting decision tree model is constructed for classification. We conduct experiments based on three datasets, which in this paper are referred to as the Pavia University, Indiana Pines, and Salinas datasets. It is shown that the proposed FF-DT achieves better performance in classification accuracy, training conditions, and time consumption than other current classical hyperspectral image classification methods.

1. Introduction

Hyperspectral image (HSI) processing is one of the core research areas in the field of HSI remote sensing [1,2,3,4,5], which is an important means of earth observation. HSIs have many spectral bands and large amounts of data, many of which are redundant. Designing an algorithm and obtaining efficient performance is key in this field. There are two important points in HSI classification: proper feature representation and an efficient classification model.
Extended shape contours are constructed by mathematical operations of structural elements to form shapes of various sizes [6,7], and morphological attribute profiles obtained through more complex morphological operations [8,9,10] are representative in spatial information. In [11], a spline wavelet, which has the characteristic of spatial translation invariance, was used to extract spatial spectrum features. In order to improve the performance, multiple extension methods of SVM applied to the HSI field have been proposed [12]. Convolutional neural networks use local receptive fields to extract spatial information in images and local sharing mechanisms to reduce network training parameters. Convolutional neural networks are now being applied to HSI classification tasks with greater frequency [13,14,15,16]. In [17], the authors used a CNN in the study of depth representation methods based on spectral features, and the classification effect was better than the traditional SVM algorithm. In [18], the authors utilized CNN and obtained a higher accuracy in classification capabilities by learning pixel-pair features. In [19], the authors proposed a multi-region CNN classification method for HSIs.
However, in order to obtain better accuracy by using DNN models, much effort is needed to adjust the parameters of deep model. Furthermore, when the data change, the deep network also needs to adjust the network structure. Beyond that, there are problems caused by a vanishing gradient and an exploding gradient when using DNN for classification: overfitting and underfitting. Though a large amount of HSI data has been obtained due to advanced HSI remote sensing technology, there are only a few labeled HSI data. Without sufficient training data, it is difficult to obtain a satisfactory result for a DNN model with many parameters, due to intensive computing conditions. An efficient classification network is needed that can be applied to small devices, such as drones and remote sensing platforms.
Additionally, in HSI classification tasks, preprocessing-based methods and hybrid methods often use extracted features and classifiers to implement classification tasks. Therefore, the quality of the extracted features and the selected classifier obviously affects classification results. Feature extraction is used to transform original features into features with actual physical meaning, such as texture features, geometric features, and edge features in the image. Choosing appropriate features can reduce redundant information in the data and mine the potential deep-level information of the data, which will greatly help the subsequent classification.
Summarizing, the HSI classification problem is facing several technical challenges:
(i)
How can the right features be chosen where multiple features can be fused?
(ii)
For classification, how can model training be effective with a few parameters and low computational complexity?
(iii)
How can a satisfactory classification model be obtained in a short time under limited hardware conditions?
To solve the above-mentioned problems, we strive to propose a new strategy via leveraging feature fusion and a multi-level gradient boosting decision tree (mGBDT) method [20]. The main contributions are outlined as follows.
  • We extract extended morphology profiles, linear multi-scale spatial characteristics, and nonlinear multi-scale spatial characteristics as final features. The original data of the HSI is a three-dimensional image, and the spatial dependence complementary to the spectral information behavior is naturally another information source. The introduction of spatial information improves the possibility of pixel-by-pixel classification.
  • We utilize a decision tree-based model, namely, mGBDT, which has fewer parameters and is easier to train. Compared with deep learning model, the proposed model is easy for theoretical analysis and practical training, and only requires simple hardware conditions to perform model training.
In this paper, feature fusion and a mGDBT model (FF-DT) for HSI classification is proposed. In Section 2, we briefly introduce works relevant to the proposed method. In Section 3, the structure of our method is presented. In Section 4, the experiment design is detailed. Section 5 shows the results. Finally, a summary of the paper is given.

2. Related Work

2.1. Principal Component Analysis

In [21], principal component analysis of a data matrix was proposed to extract the dominant patterns in a matrix in terms of a complementary set of score and loading plots, alleviating computing pressure and accelerating computing speed.
In [22], a folded PCA (F-PCA), in which both global and local structures were taken into account, preserved all useful properties of PCA. The work simplified the analysis of the high dimensional nature of a hyperspectral image.
With the improvement of sensor technology, the dimensions of hyperspectral images are generally in the tens to hundreds of dimensions. The amount of data is large and redundant. If it is trained, it takes a long time and requires good hardware. Thus, our purpose is to reduce the dimensionality of n-dimensional data to k-dimensional data, that is, to find a k-dimensional plane to represent this dataset. In order to increase the training rate, a certain amount of data needs to be reduced. In this paper, we use principal component analysis (PCA) to reduce the spectral dimension of the hyperspectral image.

2.2. Extended Morphological Features

In [23], a semisupervised classification algorithm for hyperspectral remote sensing images based on spectral and spatial information was proposed. The spatial information was extracted by building extended morphological profiles (EMPs) based on principle components of the hyperspectral image.
In [24], morphological profiles were used to fuse spectral and spatial information to produce improved classification results. Opening and closing morphological transforms were used in order to isolate bright (opening) and dark (closing) structures in images, where bright/dark means brighter/darker than the surrounding features in the images.
The morphological feature is a nonlinear method. The basic operations included are opening operation, corrosion, expansion, and closing operation. Given a set A, the structural element B corrodes A, and the output result is the set of the origin position of B when B is scanned by A and B.
A B = z B z A
The expansion is to use the structural element B to scan from the beginning to touch the set A to find the set of the origin of B, so at least one element of A and B overlaps [25].
A B = z [ B ^ ) z A
The opening and closing operations are defined by the sequence of corrosion and expansion. Corrosion before expansion is an opening operation, which can eliminate smaller bright spots and retain the characteristics of a larger bright area [26]. Expansion after corrosion is a closing operation, which can bridge narrower discontinuities and remove smaller dark spots in the image.
M P ( n ) ( I ) = φ r ( n ) ( I ) , , φ r ( 1 ) ( I ) , I , , γ r ( n ) ( I )
where n is the number of open and closed operations, I is the image to be processed, φ r ( I ) is the closing operation, and γ r ( I ) is the opening operation. Because hyperspectral images have a large number of spectral dimensions, in order to utilize these spectral dimensions, extended morphological feature extraction methods are used first by reducing the spectral dimension of the hyperspectral image, then by extracting m principal components, and then extracting the morphological features of each component [27].
E M P m ( n ) ( I ) = M P 1 ( n ) ( I ) , M P m ( n ) ( I )

3. Methodologies

Because the traditional method does not combine spatial information of hyperspectral imagery, we propose a hyperspectral image classification method based on feature fusion and an mGBDT. The method first uses PCA to reduce the original hyperspectral data, and then extracts its extensible morphological features (EMP), linear multi-scale features, and nonlinear multi-scale features from the reduced-dimensional data. The fusion of these three features is used as the input of the deep forest classifier. The experimental results prove that the hyperspectral classification method based on feature fusion and deep forest is efficient and has excellent performance, even on small samples.

3.1. Linear Multi-Scale Spatial Characteristics

The scale space aims to obtain images at different scales. In [28], the linear and nonlinear multiscale spatial characteristics are used in the proposed algorithm. In [29], mathematical forms are extracted to prove that the Gaussian kernel is the only transformation kernel that realizes scale transformation in different ways. The Gaussian filter is an optimal filter for establishing a linear scale space. For a 2D image I(x, y), the image filtered by a Gaussian filter can be represented as follows,
L ( x , y , δ ) = G ( x , y , δ ) I ( x , y )
where * represents the convolution operator. ( x , y ) represents the position of the pixels in the image, and G ( x , y , δ ) represents a Gaussian function whose formula can be expressed as
G ( x , y , δ ) = 1 2 π δ 2 e x 0 m / 2 2 + y 0 n / 2 2 2 δ 2
where x 0 , y 0 is the coordinate of the center point. δ is a scale parameter, which is a continuous transformation value that determines the degree of smoothness of the transformed image. The larger δ is, the better the smoothness becomes, and the overall structure of the image can be captured. The smaller δ is, the weaker the smoothness becomes, and the details of the image can be retained. The algorithm here is based on Gaussian-scale space. Therefore, when scale transformation is performed, an operator is added so that the edge information can be preserved while extracting spatial features. The new scale transformation formula is
L z ( x , y , δ ) = e 1 1 + x i ( G ( x , y , d ) I ( x , y ) )
where x i represents the Sobel operator, which can be expressed as
x i = 1 4 h X i 2 + v X i 2 + r X i 2 + i X i 2
where h X i , v X i , r X i , and l X i denote the first-order horizontal, vertical, and two diagonal gradients of a pixel, respectively.

3.2. Nonlinear Multi-Scale Spatial Features

In order to obtain translation invariance, a non-subsampled contourlet (NSCT) is considered as a nonlinear multiscale spatial feature. Compared with two-dimensional discrete wavelet transform, contour wave transform is better at representing two-dimensional signals and can significantly improve the performance of image denoising, image texture features, and shape features extraction. Through the pyramid directional filter bank (pdfb), the image is decomposed into directional sub-bands of different scales by a contourlet. Pdfb consists of two parts: a Laplacian pyramid (LP) and a directional filter bank.
Figure 1 shows the decomposition steps of NSCT. The basis of a NSCT is a non-subsampled pyramid (NSP) and non-subsampled directional filter banks (NSDFBs). There are two main steps in a NSCT: First, the input image is decomposed into high-pass and low-pass parts by an NSP tower, and second the high-frequency sub-band is then decomposed into multiple directional sub-bands by an NSDFB.

3.3. mGBDT

In this paper, after data processing and feature extraction of the hyperspectral images, the extracted features are used as input to the mGBDT model to train it. The mGBDT model uses the GBDT tree structure as the basic module, the network structure is simple, and the number of participants is small, so the training is convenient. For the adjustment of parameters, for different data, the deep network needs to adjust the network structure. The mGBDT model, with the same parameter settings, can meet the training needs of different data. The design of the model is ideal for theoretical analysis and practical training. Only simple hardware conditions are required to train the model. In order to improve classification accuracy, users must adjust a deep model. As the model is a multi-layered tree structure, backpropagation cannot be used during training to modify the parameters, like traditional deep neural networks. We know that, in a multilayer network, there is a forward propagation, F, between layers, i.e.,
F t O i 1 = O i
The parameters of forward propagation are optimized to minimize the loss of the output of the last layer and groundtruth. Generally, it is effective to use backpropagation directly in different deep neural network models, but for multi-layer networks with a tree structure, it is necessary to define a parameter iteration mechanism. The mGBDT uses a pseudo-label idea to iterate through the parameters layer by layer. For the forward propagation of each layer, we set a reverse mapping,
G i t F i t 1 O i 1 O i 1 ,
compute the loss and G,
G ^ t = argmin E x G i F i O i 1 + ε O i 1 + ε ,
and introduce some noise to increase model robustness. The purpose of the iterative update of F and G is to obtain the error between the ideal output and the actual output of the current layer. First, it is necessary to obtain the ideal output of each layer. This seems to be difficult, but a pseudo-label can be set to approximate the real output. The top-level pseudo-labels can be calculated using ground truth, i.e.,
Z m t = O M L O M , y O M
Inverse mapping is used to update Z from back to front, and F can then be updated toward the residual:
L F M t 1 O M 1 , Z M t F M t 1 O M 1

3.4. Hsi Classification Based on Feature Fusion and mGBDT

The selection of features is a key step in the hyperspectral classification task. This paper uses three features: EMP, linear multiscale features, and nonlinear multiscale features. EMP effectively extracts the morphological features of each layer of the hyperspectral image and retains the geometric information of the image well. Gaussian filtering and Sobel operators are used for linear multi-scale spatial features to sufficiently suppress image noise while retaining image edge features. Nonlinear multi-scale features use the NSCT feature, which has good multi-directionality and multi-scale properties, and have translation invariance, which can effectively extract the contour features of the image. The combination of linear and nonlinear features is effective. The edge and contour information of the image is preserved. The combination of the three features guarantees the adequacy, efficiency, and accuracy of the proposed features and provides favorable input features for subsequent classification models.
In this paper, the proposed FF-DT is based on feature fusion and mGBDT. Figure 2 shows the architecture of the proposed method. Let I R W H C represent a hyperspectral image, and let X n R 1 C , n [ 1 , N ] represent an HSI pixel. First, because the hyperspectral image has a high spectral dimension, we use the PCA method to reduce the redundancy, retaining the first principal components to contain most of the information I p c R W H C . The spectral-spatial features of the input data are then extracted by EMP, linear multi-scale spatial characteristics, and nonlinear multi-scale spatial characteristics. At last, each pixel from the hyperspectral images is transformed into a one-dimensional feature vector, I R W H C .
On the basis of feature fusion, we introduce a multi-level gradient boosted decision tree (mGBDT). The mGBDT is a model based on a decision tree whose overall amount of parameters is small. Compared with many current deep models with wider and deeper network parameters, network parameter training is easier and faster, and it can obtain better results in a shorter time when the CPU and other hardware conditions are poor. In addition, in order to fully train a deep model with a large number of parameters, a sufficient amount of training data is required to improve model performance. However, the mGBDT can obtain better results with a small number of training samples. In this paper, we define the framework: input-l1-l2-output, which is simple but useful. More detailed parameter analysis will be discussed in the experiment section.

4. Experiment Designs

4.1. Datasets

In the experimental part, we use multiple datasets to verify the validity of the model. The images in the Pavia University dataset cover the city of Pavia in Italy. This dataset was captured by a reflective optics system imaging spectrometer (ROSIS) over the city. Each image is 610 * 340 pixels. The number of spectral bands is 103, with a spectral coverage ranging from 0.43 to 0.86 m and a spatial resolution of 1.3 m per pixel. There are nine different classes: asphalt, meadows, gravel, trees, painted metal sheets, bare soil, bitumen, self-blocking bricks, and shadows. Different colors denote different types.
The Indian Pines dataset depicts Indian pine forests in Indiana, Northwestern United States. They were captured by an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor over Northwestern Indiana and contain 220 spectral bands with a spectral coverage ranging from 0.4 to 2.5 m. When there were 20 spectral bands that could not be reflected by water, they were removed. The remaining 200 spectral bands were chosen as the experimental objects, which include 16 classes.
The images in the Salinas dataset were captured by an AVIRIS sensor over the city of Salinas Valley in the state of California, USA. These images contain 224 bands with 16 classes, including vegetables, vineyard fields, and bare soil.

4.2. Parameter Analysis

The proposed method is based on extended morphological features and an mGBDT. First, at the data level, data preprocessing is required. For the redundant spectral dimensions of the original input data, PCA was used to extract principal components and reduce the data dimensions. The first five principal components in the spectral dimension were retained because they contain more than 99 percent of the information. EMP was then used to extract the morphological features of the hyperspectral data, with four openings and closings (range from 2 to 8, with a step size of 2). Each pixel was a 1-d vector with 90 dimensional features, and the model was trained and predicted using the mGBDT.
Furthermore, the maximum depth of each tree used by the forward network and inverse mapping in the model was 7, the learning rate was 0.2, one-hot coding training and prediction data was used, and the inverse mapping layer used Gaussian noise with a mean value of 0 and a standard deviation value of 0.3. In order to avoid training the inverse mapping at the top level, the top-level classification layer was set to a linear function with cross entropy loss, the learning rate was set at 0.03, and other layers used GBDT for inverse mapping.

5. Results

The proposed model was compared with SVM-RBF [30], SVM-EMP [31], CNN [32], 3D-CNN [33], FuNet-C [34], and MDGCN [35] to explore whether the proposed model was effective. SVM-RBF used the spectral features of the hyperspectral data as input for training. SVM-EMP used the EMP features as input for training. The CNN model used only spectral features and used a one-dimensional convolution kernel to extract them. 3D-CNN introduced the spatial dimension. The image blocks were input into the CNN network, and the results were output. FuNet-C is a semisupervised network based on graph neural network. MDGCN is applicable to the irregular image regions represented by graph topological information. This experiment used three hyperspectral datasets: the Pavia University, Indian Pines, and Salinas datasets.

5.1. Classification Results of the Pavia University Dataset

The Pavia University data were divided into a training set and a test set at a ratio of 5% and 95%. Each model used the same dataset to train and test.
Table 1 and Table 2 and Figure 3 show the classification performances and classification visualizations of the Pavia University dataset for various algorithms. Figure 3 shows the results of seven different classification algorithms. Different colors represent different types of experimental objects. The many different colors between RBF-SVM (a) and ground truth (h) in the center of these two pictures indicate the poor classification performance of RBF-SVM (a). The EMP-SVM (b) outperforms the RBF-SVM (a), but there is a gap between EMP-SVM (b) and ground truth (h). The 2D-CNN (c) and 3D-CNN (d) obtain better classification results compared to the former two algorithms because of the complex computing models and hyperparameters. The 3D-CNN outperforms the 2D-CNN. The FuNet-C (e) and MDGCN (f), that are based on the graph, do not obtain a better performance than the mGBDT. Our proposed model FF-DT (g) performs the best in all the models, as there are significant similarities between the results of FF-DT (g) and ground truth (h). Furthermore, it relies on a small complex network and minimal computing conditions, making the HSI classification more practical.
Table 1 and Table 2 show that, for the dataset of Pavia University, the overall accuracy (OA) of SVM using EMP is greatly improved compared with that of SVM-PCA using spectral features. The CNN method effectively utilized the features of the hyperspectral images. Even the spectral dimension features are relatively good. The 3D-CNN extract spectral-spacial features of hyperspectral images were also improved. However, due to the structural limitations of CNNs, if a high precision is required, a large number of training data and a wider and deeper network are needed. The FuNet-C model is based on a graph requiring a high computation complexity; however, its classification accuracy on this dataset is lower than our proposed model. The same as MDGCN, that is based on graph neural network. The MDGCN performs worse than the mGBDT. The mGBDT network using EMP features can effectively and quickly train the network. In this experiment, the proposed model is 11% higher in OA than the r-SVM. Furthermore, it is 8.7% higher in KAPPA than the E-SVM. It can be seen that in most of types, FF-DT performs the best in all the algorithms. Even compared to the CNN based model, it obtains a great result in both training speed and classification accuracy in most of the class. It is shown in Table 2 that our proposed model obtains the best classification result with or without considering the weight of each component, illustrating the superiority of FF-DT.

5.2. Classification Results on the Indian Pines Dataset

The images in the Indian pines dataset have a small space size but contain 16 types of features. The data were divided into a training set and a test set at a ratio of 10% and 90%.
The classification performance and visualization of the Indian Pine dataset by this experiment and multiple comparison algorithms are shown in Table 3 and Table 4 and Figure 4.
Figure 4 shows results of five different classification algorithms on the Indian Pines dataset. There are many noisy points in the RBF-SVM (a) and EMP-SVM (b), which indicates their poor classification performance. The 2D-CNN (c) and 3D CNN (d) outperform the former two algorithms, because their inner structures are complex. The FuNet-C (e) and MDGCN (f) obtain a better performance than the former models, for the complex network architecture based on graph. Our proposed model, FF-DT (g), is the most similar to the ground truth (h).
Table 3 and Table 4 show that the SVM classification and CNN effect are not particularly good if only spectral features are used. As the Indian Pines dataset is small, there are too many categories. However, extended morphological features can effectively represent the information contained in the hyperspectral images, and the performance of SVM is effectively improved. 3D-CNN used spectral-spacial information, and the classification effect was better. The FuNet-C and MDGCN perform well in this dataset; however, their classification accuracy are lower than our proposed mGBDT. With the importing of mGBDT algorithm. The fusion of the three features fully embodies the spatial and spectral features of the hyperspectral images and the spectral information of the datasets. Moreover, the mGBDT model makes a great contribution to the classification results. From the Table 3, the FF-DT model is 27% higher in AA than the r-SVM. In OA and KAPPA, it outperforms all the baselines. From Table 4, it is shown that the FF-DT obtains a best classification precision in 5% Indian Pines dataset. In weighted average value, FF-DT obtains the best performance in all the evaluation metrics, illustrating the importance of weighting each component in classification process.

5.3. Classification Results on the Salinas Dataset

The Salinas dataset was divided: 5% for training data and 95% for test data. Table 5 and Table 6 and Figure 4 show the classification results and classification visualizations of the different methods for the Salinas dataset.
Figure 5 shows the results of five different classification algorithms on the Salinas dataset. It can be seen that the SVM using EMP (b) features outperforms the RBF-SVM (a). Many different color points are shown in the RBF-SVM (a) and EMP-SVM (b), indicating that these two algorithms do not classify well. The 2D-CNN (c) and 3D CNN (d) outperform the former two algorithms, because they have imported spatial information and possess large amounts of hyperparameters. The FuNet-C model (e) and MDGCN model (f) conducted on the dataset show a good classification result, based on the graph neural network. Our proposed model, FF-DT (g), appears to have the least points compared to the other algorithms, representing its superiority in HSI classification.
Table 5 and Table 6 show that, due to the characteristics of the Salinas hyperspectral imagery, even if a support vector machine is used to train the spectral dimension, it can achieve better results. Similarly, a SVM using EMP features will have better results. Because of the integration of spatial dimensions, the 3D-CNN leads to an improved prediction accuracy than the 2D-CNN network. Furthermore, FuNet-C and MDGCN show much improvement in classification accuracy. However, the mGBDT network using feature fusion is still the best on the whole. From the Table 5, it can be seen that the FF-DT model is 21% higher in KAPPA than the r-SVM. In OA and AA, FF-DT obtains the best performance. From Table 6, we found that the FF-DT performs best in 5% Indian Pines dataset, especially when the weight of each component is considered. It also can be seen that the EMP-SVM performs great in all evaluation metrics, illustrating the superiority of extended morphological features.

5.4. The Effect of Multi-Feature Fusion

In order to show the effect of multi-feature fusion, we added a comparison experiment of single feature and multi-feature fusion, shown in Table 7. The experiment showed that multi-feature fusion has an improved classification effect. The feature vector dimensions extracted by EMP, Linear MSSC, Nonlinear MSSC, and Feature Fusion are 30, 18, 42, and 90, respectively. For the Pavia University dataset in Table 1, Feature Fusion was 9.1% higher than the single EMP feature, 4.6% higher than the single Linear MSSC, and 0.2% higher than the single Nonlinear MSSC. For the Salinas dataset in Table 1, Feature Fusion was 0.4% higher than the single EMP feature, 15.2% higher than the single Linear MSSC, and 0.8% higher than the single Nonlinear MSSC. The Indian Pines dataset had similar results compared to the Salinas dataset.
In the experiments on the hyperspectral data, we used the same model structure, input-32-32-output, for the three different datasets with different numbers and classifications. The classification results show that the mGBDT model is more robust. In terms of training time, the mGBDT model was trained on a personal PC. The time required was also close to the time required to run a deep network on a GPU device.
Our proposed method consists of feature fusion and an effective classifier mGBDT. In order to obtain better classification accuracy, we need to spend a lot of effort to adjust the parameters of deep model, and when data changes, the deep network also needs to adjust the network structure. The mGBDT is a new model of tree structure proposed in 2018. Compared with the deep learning model, the design of the model is easy for theoretical analysis and practical training, and only requires simple hardware conditions to perform model training. Considering the advantages of mGBDT, we first introduce mGBDT to hyperspectral classification. The mGBDT model, with the same parameter settings, can meet the training of different data (We use the same parameter settings, in the experiment of hyperspectral data. The model structure, intput-32-32-output, can obtain better classification results, showing that the mGBDT model has better robustness.), and mGBDT also has other advantages. To illustrate the advantages of mGBDT in terms of computational consumption, we have added speed comparison experiments of 3D-CNN and the proposed FF-DT shown as Figure 6. Graphics processing often requires a higher-cost GPU, while mGDBT requires only a normal CPU to achieve faster convergence speed. (The selected data set is the commonly used hyperspectral data set Pavia, 3D-CNN uses GPU: GeFore GTX 1080Ti; the proposed FF-DT uses CPU: Intel i7). Compared with 3D-CNN, shorter training time and faster convergence is much more competitive in some certain scenarios. Therefore, this model can be effectively applied to some lightweight remote sensing devices.
In conclusion, the FF-DT method is proposed to solve the problem that the traditional method cannot effectively use the spatial information of hyperspectral data. The extracted EMP features greatly preserve the geometric information of the image, while the linear and nonlinear multi-scale spatial features well retain the edge and contour information of the image. After the three features are fused, the target image’s effective information will be well retained, which could be beneficial for subsequent classification tasks. On the classifier, we choose the mGBDT with strong representation ability and adaptive adjustment ability of structure complexity to improve the classification performance. From the comparative analysis with three comparative experiments, it can be seen that the algorithm in this chapter excels in the classification task of hyperspectral images. Even for Indian Pines with small homogeneous regions, the algorithm in this chapter can greatly retain its edge information, making the classification process more efficient and accurate. From the experimental classification accuracy table and classification result graph, the efficiency of the algorithm in this chapter can be seen.

6. Conclusions

Models based on deep convolutional neural networks are widely used in hyperspectral image processing fields, such as image classification, but deep networks require sufficient training data, appropriate adjustments of many parameters, and high hardware conditions. It is difficult to obtain good results for small sample data because of the disadvantages of back-propagation algorithms. To overcome these shortcomings, the proposed method in this paper introduces feature fusion and an mGBDT model to better extract and classify the spectral and spatial features of hyperspectral images. First, principal component analysis was used to reduce the dimensions of the data. Extended morphology profiles (EMPs), linear multi-scale spatial characteristics, and nonlinear multi-scale spatial characteristics were used to extract spectral features. Finally, a mGBDT model was constructed for classification. The model based on the decision tree requires fewer hyperparameters and simple training conditions. This paper proposes an algorithm that can achieve relatively good classification accuracy with limited hardware conditions and short iterations. In the three hyperspectral image classification experiments, the proposed model combined the excellent performance of the tree and the use of EMP features to effectively extract the spatial and spectral information of hyperspectral images and can thus achieve better performance than other current classical models.

Author Contributions

Formal analysis, H.W.; Investigation, S.L.; Resources, Z.X.; Software, F.Z.; Writing—original draft, S.X.; Validation, S.L.; Writing—review and editing, W.C.; Supervision, Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (grant number:no.19CGL043).

Institutional Review Board Statement

Ethical review and approval were waived for this study, due to REASON that our work involves no human or animals.

Informed Consent Statement

Patient consent was waived due to REASON that our work involves no human or animals.

Data Availability Statement

Data available in a publicly accessible repository The data presented in this study are openly available in https://dx.doi.org/10.21227/eqk7-wa46.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, S.; Xu, P.; Song, R.; Li, P.; Ma, H. Development of High Performance Quantum Image Algorithm on Constrained Least Squares Filtering Computation. Entropy 2020, 22, 1207. [Google Scholar] [CrossRef] [PubMed]
  2. Vukotić, V.; Chappelier, V.; Furon, T. Are Classification Deep Neural Networks Good for Blind Image Watermarking? Entropy 2020, 22, 198. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Mohamed, H.G.; ElKamchouchi, D.H.; Moussa, K.H. A novel color image encryption algorithm based on hyperchaotic maps and mitochondrial DNA sequences. Entropy 2020, 22, 158. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Gao, H.; Yao, D.; Wang, M.; Li, C.; Liu, H.; Hua, Z.; Wang, J. A Hyperspectral Image Classification Method Based on Multi-Discriminator Generative Adversarial Networks. Sensors 2019, 19, 3269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Chaudhary, S.; Ninsawat, S.; Nakamura, T. Non-Destructive Trace Detection of Explosives Using Pushbroom Scanning Hyperspectral Imaging System. Sensors 2019, 19, 97. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Plaza, A.; Martinez, P.; Perez, R.; Plaza, J. A new approach to mixed pixel classification of hyperspectral imagery based on extended morphological profiles. Pattern Recognit. 2004, 37, 1097–1116. [Google Scholar] [CrossRef]
  7. Bai, J.; Yuan, A.; Xiao, Z.; Zhou, H.; Wang, D.; Jiang, H.; Jiao, L. Class Incremental Learning With Few-Shots Based on Linear Programming for Hyperspectral Image Classification. IEEE Trans. Cybernet. 2020, 1–12. [Google Scholar] [CrossRef]
  8. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of Hyperspectral Images by Using Extended Morphological Attribute Profiles and Independent Component Analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef] [Green Version]
  9. Sánchez-Sánchez, M.; Conde, C.; Gómez-Ayllón, B.; Ortega-Delcampo, D.; Tsitiridis, A.; Palacios-Alonso, D.; Cabello, E. Convolutional Neural Network Approach for Multispectral Facial Presentation Attack Detection in Automated Border Control Systems. Entropy 2020, 22, 1296. [Google Scholar] [CrossRef]
  10. Xia, J.; Dalla Mura, M.; Chanussot, J.; Du, P.; He, X. Random Subspace Ensembles for Hyperspectral Image Classification with Extended Morphological Attribute Profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4768–4786. [Google Scholar] [CrossRef]
  11. He, L.; Li, Y.; Li, X.; Wu, W. Spectral–Spatial Classification of Hyperspectral Images via Spatial Translation-Invariant Wavelet-Based Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2696–2712. [Google Scholar] [CrossRef]
  12. Cui, Y.; Zeng, Z. Remote Sensing Image Classification Based on the HSI Transformation and Fuzzy Support Vector Machine. In Proceedings of the 2009 International Conference on Future Computer and Communication, Kuala Lumpar, Malaysia, 3–5 April 2009; pp. 632–635. [Google Scholar]
  13. Feng, F.; Wang, S.; Wang, C.; Zhang, J. Learning Deep Hierarchical Spatial-Spectral Features for Hyperspectral Image Classification Based on Residual 3D-2D CNN. Sensors 2019, 19, 5276. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Zhu, J.; Hu, J.; Jia, S.; Jia, X.; Li, Q. Multiple 3-D Feature Fusion Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1873–1886. [Google Scholar] [CrossRef]
  15. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  16. Wu, C.; Wang, M.; Gao, L.; Song, W.; Tian, T.; Choo, K.R. Convolutional Neural Network with Expert Knowledge for Hyperspectral Remote Sensing Imagery Classification. KSII Trans. Internet Inf. Syst. 2019, 13, 3917–3941. [Google Scholar] [CrossRef]
  17. Hu, W.; Huang, Y.; Li, W.; Zhang, F.; Li, H. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015, 258619:1–258619:12. [Google Scholar] [CrossRef] [Green Version]
  18. Roodposhti, M.S.; Aryal, J.; Lucieer, A.; Bryan, B.A. Uncertainty Assessment of Hyperspectral Image Classification: Deep Learning vs. Random Forest. Entropy 2019, 21, 78. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Zhang, M.; Li, W.; Du, Q. Diverse Region-Based CNN for Hyperspectral Image Classification. IEEE Trans. Image Process. 2018, 27, 2623–2634. [Google Scholar] [CrossRef]
  20. Feng, J.; Yu, Y.; Zhou, Z. Multi-Layered Gradient Boosting Decision Trees. In Advances in Neural Information Processing Systems 31: Proceedings of the Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, QC, Canada, 3–8 December 2018; Bengio, S., Wallach, H.M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; 2018; pp. 3555–3565. Available online: https://papers.nips.cc/paper/2018/file/39027dfad5138c9ca0c474d71db915c3-Paper.pdf (accessed on 24 December 2020).
  21. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  22. Deepa, P.; Thilagavathi, K. Feature extraction of hyperspectral image using principal component analysis and folded-principal component analysis. In Proceedings of the 2015 2nd International Conference on Electronics and Communication Systems (ICECS), Coimbatore, India, 26–27 February 2015; pp. 656–660. [Google Scholar] [CrossRef]
  23. Wang, J.; Zhang, G.; Cao, M.; Jiang, N. Semi-supervised classification of hyperspectral image based on spectral and extended morphological profiles. In Proceedings of the 2016 8th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016; pp. 1–4. [Google Scholar] [CrossRef]
  24. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  25. Plaza, A.; Martínez, P.; Plaza, J.; Pérez, R. Dimensionality reduction and classification of hyperspectral image data using sequences of extended morphological transformations. IEEE Trans. Geosci. Remote Sens. 2005, 43, 466–479. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, F.; Bai, J.; Zhang, J.; Xiao, Z.; Pei, C. An Optimized Training Method for GAN-Based Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 1–5. [Google Scholar] [CrossRef]
  27. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear versus nonlinear PCA for the classification of hyperspectral data based on the extended morphological profiles. IEEE Geosci. Remote Sens. Lett. 2011, 9, 447–451. [Google Scholar] [CrossRef] [Green Version]
  28. Manzo, M. Attributed Relational SIFT-Based Regions Graph: Concepts and Applications. Mach. Learn. Knowl. Extr. 2020, 2, 233–255. [Google Scholar] [CrossRef]
  29. Lindeberg, T. Scale-Space Theory in Computer Vision; The Springer International Series in Engineering and Computer Science; Springer: Boston, MA, USA, 1994; Volume 256. [Google Scholar] [CrossRef] [Green Version]
  30. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  31. Fauvel, M.; Chanussot, J.; Benediktsson, J.A. A spatial-spectral kernel-based approach for the classification of remote-sensing images. Pattern Recognit. 2012, 45, 381–392. [Google Scholar] [CrossRef]
  32. Jia, P.; Zhang, M.; Yu, W.; Shen, F.; Shen, Y. Convolutional neural network based classification for hyperspectral data. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5075–5078. [Google Scholar] [CrossRef]
  33. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  34. Mou, L.; Lu, X.; Li, X.; Zhu, X.X. Nonlocal Graph Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 54, 8246–8257. [Google Scholar] [CrossRef]
  35. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multiscale dynamic graph convolutional network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3162–3177. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Illustration of the non-subsampled contourlet (NSCT).
Figure 1. Illustration of the non-subsampled contourlet (NSCT).
Entropy 23 00020 g001
Figure 2. The overall framework of the proposed model.
Figure 2. The overall framework of the proposed model.
Entropy 23 00020 g002
Figure 3. Classification results map of various algorithms on the Pavia University dataset: (a) RBF-SVM, (b) EMP-SVM, (c) 2D-CNN, (d) 3D-CNN, (e) FuNet-C, (f) MDGCN, (g) FF-DT, and (h) ground truth.
Figure 3. Classification results map of various algorithms on the Pavia University dataset: (a) RBF-SVM, (b) EMP-SVM, (c) 2D-CNN, (d) 3D-CNN, (e) FuNet-C, (f) MDGCN, (g) FF-DT, and (h) ground truth.
Entropy 23 00020 g003
Figure 4. Classification results map of various algorithms on the Indian Pines dataset: (a) RBF-SVM, (b) EMP-SVM, (c) 2D-CNN, (d) 3D-CNN, (e) FuNet-C, (f) MDGCN, (g) FF-DT, and (h) ground truth.
Figure 4. Classification results map of various algorithms on the Indian Pines dataset: (a) RBF-SVM, (b) EMP-SVM, (c) 2D-CNN, (d) 3D-CNN, (e) FuNet-C, (f) MDGCN, (g) FF-DT, and (h) ground truth.
Entropy 23 00020 g004
Figure 5. Classification results map of various algorithms on the Salinas dataset: (a) RBF-SVM, (b) EMP-SVM, (c) 2D-CNN, (d) 3D-CNN, (e) FuNet-C,(f) MDGCN, (g) FF-DT, and (h) ground truth.
Figure 5. Classification results map of various algorithms on the Salinas dataset: (a) RBF-SVM, (b) EMP-SVM, (c) 2D-CNN, (d) 3D-CNN, (e) FuNet-C,(f) MDGCN, (g) FF-DT, and (h) ground truth.
Entropy 23 00020 g005
Figure 6. Speed comparison of 3D-CNN and the proposed FF-DT.
Figure 6. Speed comparison of 3D-CNN and the proposed FF-DT.
Entropy 23 00020 g006
Table 1. Classification results of the Pavia University dataset.
Table 1. Classification results of the Pavia University dataset.
Classr-SVME-SVMCNN3D-CNNFuNet-CMDGCNFF-DT
10.87900.99100.99600.98600.94920.98960.9980
20.88300.87700.98400.97200.99170.99630.9990
30.70900.99800.83100.97101.00000.89760.966
40.94700.99900.83600.98200.97820.95090.9930
50.99901.00000.97801.00001.00000.97281.0000
60.86800.97400.91501.00000.99900.97401.0000
70.83400.99600.98700.99800.85920.98040.9980
80.83900.99400.93600.99800.90250.96350.9980
91.00001.00000.88200.89100.99930.90390.9840
OA0.87800.93400.95300.98100.97200.98810.9980
AA0.88400.98100.92700.97100.95910.97580.9970
KAPPA0.83400.91100.93800.97200.96290.98410.9980
Table 2. 5% Pavia University dataset.
Table 2. 5% Pavia University dataset.
Macro AvgWeighted Avg
PrecisionRecallF1-ScorePrecisionRecallF1-Score
EMP-SVM98.1%98.4%98.2%98.8%98.8%98.7%
RBF-SVM98.1%97.1%97.6%98.5%98.4%98.4%
CNN98.2%97.7%97.9%98.6%98.6%98.5%
3D-CNN95.2%93.8%94.3%95.0%94.8%94.7%
FF-DT99.8%99.7%99.7%99.8%99.8%99.8%
Table 3. Classification results of the Indian Pines dataset.
Table 3. Classification results of the Indian Pines dataset.
Classr-SVME-SVMCNN3D-CNNFuNet-CMDGCNFF-DT
10.14320.92960.97230.80530.87930.88570.7863
20.66630.88390.87320.90420.76720.92750.9286
30.60530.89240.91130.87960.82560.94340.9943
40.54740.76620.85910.60230.73940.95530.9766
50.85370.85350.69400.89310.92710.93520.8323
60.97470.97130.96670.97400.97350.98031.0000
700.57720.52160.91290.95900.81761.0000
80.99571.00001.00000.96450.98410.99390.9769
90.16770.55610.45280.82361.00000.80580.9843
100.73770.88370.83460.9630.79470.89970.7714
110.85660.91280.93770.9490.87670.97760.9798
120.62230.82420.87680.75240.76410.94170.9890
130.99520.99580.92490.91250.99360.98240.8083
140.96930.99720.97640.98460.94330.98110.9981
150.4640.92240.95951.0000.67380.95550.7786
160.91710.88110.45620.96320.95120.81750.9711
OA0.78750.91130.91230.92560.87970.96500.9721
AA0.65720.86740.82640.84220.90330.94930.9233
KAPPA0.75570.89810.89330.91420.86290.86010.9662
Table 4. 5% Indian Pines dataset.
Table 4. 5% Indian Pines dataset.
Macro AvgWeighted Avg
PrecisionRecallF1-ScorePrecisionRecallF1-Score
EMP-SVM94.9%96.2%95.4%93.6%93.5%93.5%
RBF-SVM84.9%81.4%82.4%96.0%96.4%96.1%
CNN92.1%92.5%92.0%96.2%96.3%96.1%
3D-CNN91.3%92.6%91.7%91.4%91.3%91.2%
FF-DT96.0%92.3%93.9%97.1%97.0%97.0%
Table 5. Classification results of the Salinas dataset.
Table 5. Classification results of the Salinas dataset.
Classr-SVME-SVMCNN3D-CNNFuNet-CMDGCNFF-DT
10.99911.00000.97800.98600.99510.97111.0000
20.99111.00001.00001.00000.99850.89830.9998
30.96420.99831.00001.00000.96900.98790.9570
40.98630.99660.97770.95380.99050.97410.9988
50.99540.99931.00000.96700.95690.98240.8834
61.00001.00000.99411.00000.99900.99470.9866
70.99941.00000.99721.00000.99830.99710.9957
80.74530.73690.86940.98260.86550.84901.0000
90.99140.99751.00000.99870.98170.98780.9985
100.83131.00000.98160.99540.96760.98530.9797
110.94141.00000.98551.00000.96350.99160.9994
120.97151.00000.99990.97281.00000.99571.0000
130.94931.00001.00000.99830.99750.99441.0000
140.97950.99970.99140.99470.94730.99440.9990
150.80130.94160.99930.91930.78460.96161.0000
160.99851.00000.98920.99750.98860.97710.9993
OA0.90230.92250.97360.98130.94220.95640.9956
AA0.94650.97950.98550.98580.96860.98010.9870
KAPPA0.88830.91140.97260.97850.93560.95150.9941
Table 6. 5% Salinas dataset.
Table 6. 5% Salinas dataset.
Macro AvgWeighted Avg
PrecisionRecallF1-ScorePrecisionRecallF1-Score
EMP-SVM98.8%98.5%98.6%97.1%97.1%97.1%
RBF-SVM98.4%98.7%98.5%96.7%96.7%96.7%
CNN98.7%98.8%98.7%98.4%98.9%98.5%
3D-CNN94.9%94.3%94.0%90.8%89.8%88.9%
FF-DT99.4%98.7%99.0%99.5%99.5%99.5%
Table 7. Classification accuracy (OA) of single and fusion features.
Table 7. Classification accuracy (OA) of single and fusion features.
FeaturesPavia USalinasIndian Pines
EMP (30)0.90740.99160.9689
Linear MSSC0.95220.84330.8510
Non-Linear MSSC0.99610.98760.9675
Feature Fusion0.99840.99550.9707
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, S.; Liu, S.; Wang, H.; Chen, W.; Zhang, F.; Xiao, Z. A Hyperspectral Image Classification Approach Based on Feature Fusion and Multi-Layered Gradient Boosting Decision Trees. Entropy 2021, 23, 20. https://doi.org/10.3390/e23010020

AMA Style

Xu S, Liu S, Wang H, Chen W, Zhang F, Xiao Z. A Hyperspectral Image Classification Approach Based on Feature Fusion and Multi-Layered Gradient Boosting Decision Trees. Entropy. 2021; 23(1):20. https://doi.org/10.3390/e23010020

Chicago/Turabian Style

Xu, Shenyuan, Size Liu, Hua Wang, Wenjie Chen, Fan Zhang, and Zhu Xiao. 2021. "A Hyperspectral Image Classification Approach Based on Feature Fusion and Multi-Layered Gradient Boosting Decision Trees" Entropy 23, no. 1: 20. https://doi.org/10.3390/e23010020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop