Next Article in Journal
CFD Simulation of Solid Suspension for a Liquid–Solid Industrial Stirred Reactor
Next Article in Special Issue
Deeply-Supervised 3D Convolutional Neural Networks for Automated Ovary and Follicle Detection from Ultrasound Volumes
Previous Article in Journal
Investigating the Micro/Macro-Texture Performance of Roller-Compacted Concrete Pavement under Simulated Traffic Abrasion
Previous Article in Special Issue
Supervised Learning Based Peripheral Vision System for Immersive Visual Experiences for Extended Display
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Advanced Spectral–Spatial Classification Framework for Hyperspectral Imagery Based on DeepLab v3+

1
National Engineering Research Center for Optical Instruments, Centre for Optical and Electromagnetic Research, Zhejiang University, Hangzhou 310058, China
2
Ningbo Research Institute, Zhejiang University, Ningbo 315100, China
3
Research Institute of Zhejiang University-Taizhou, Taizhou 318000, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2021, 11(12), 5703; https://doi.org/10.3390/app11125703
Submission received: 23 March 2021 / Revised: 16 May 2021 / Accepted: 7 June 2021 / Published: 19 June 2021
(This article belongs to the Special Issue Deep Image Semantic Segmentation and Recognition)

Abstract

:
DeepLab v3+ neural network shows excellent performance in semantic segmentation. In this paper, we proposed a segmentation framework based on DeepLab v3+ neural network and applied it to the problem of hyperspectral imagery classification (HSIC). The dimensionality reduction of the hyperspectral image is performed using principal component analysis (PCA). DeepLab v3+ is used to extract spatial features, and those are fused with spectral features. A support vector machine (SVM) classifier is used for fitting and classification. Experimental results show that the framework proposed in this paper outperforms most traditional machine learning algorithms and deep-learning algorithms in hyperspectral imagery classification tasks.

1. Introduction

With the development of remote sensing technology in the late 20th century, the emergence of hyperspectral imagery (HSI) technology has attracted widespread attention in various fields. Unlike natural images and multispectral images, the resolution of hyperspectral images in the spectral range is very high, generally in the order of 10−2λ. A hyperspectral imager can record hundreds of channels of spectral band information from ultraviolet to midinfrared. This rich spectral band information has become the focus of many fields such as land monitoring, crop growth, petroleum exploration, and military target recognition.
Hyperspectral imagery classification (HSIC) is more difficult than natural image classification for two primary reasons [1]:
(a)
The spectral bands of hyperspectral images have high dimensionality, which increases the computational complexity. Moreover, according to the Hughes phenomenon, as the dimension number increases, the classification accuracy first increases and then decreases.
(b)
The spatial resolution of hyperspectral images is very low, and one pixel typically represents a distance of tens of meters. Therefore, the number of pictures in each dataset is very small, and extracting spatial features is more difficult.
At present, there are several different ideas to solve the problem of excessive dimensionality of hyperspectral datasets. First, the dimensionality of the spectral band can be reduced using one of the two broad strategies: projection and manifold learning. Principal component analysis [2] is a typical linear dimensionality reduction method in projection. The purpose is to map high-dimensional data to a low-dimensional space through a certain linear projection. In some high-dimensional images with more noise, the effect of principal component analysis is not satisfactory. Specifically, when the variance of the noise contained in a principal component with a large amount of information is greater than the variance of the signal, the quality of the image formed by the principal component is poor. Green [3] et al. addressed this problem using the MNF (minimum noise fraction) rotation. It essentially includes two stacked principal component transformations. When faced with some manifold data such as Swiss rolls, projection-based dimensionality reduction methods cannot solve the problem well; therefore, manifold learning is a better choice. The main methods are Isomap algorithm [4], Laplacian eigenmaps [5], and so on. In addition to dimensionality reduction methods, 3D-CNN in deep learning has also been applied to hyperspectral datasets. Chen [6] and Ying [7] used 3D-CNN to directly extract advanced features in hyperspectral images and realized the classification of hyperspectral images, cleverly avoiding the high dimensionality of hyperspectral images. However, the computational complexity of 3D-CNN is relatively large, and the training time of neural networks using 3D-CNN is relatively long.
To fully extract the effective feature information in the hyperspectral image, people try to combine different methods to think about the features in the spectral domain and spatial domain. In traditional machine-learning methods, some typical classifiers, such as k-nearest neighbors [8], support vector machines [9,10], etc., used the characteristic information of the spectral domain as the division standard and adopted different methods to effectively classify hyperspectral images. The powerful feature extraction ability makes deep learning attractive for traditional optical problems [11,12,13] and hyperspectral classification problems. Stacked AutoEncoders (SAE) [14] used the fusion feature information of the spectral domain and the spatial domain as training features for the classification of hyperspectral images. This is the first time that deep learning has been applied to the classification of hyperspectral images. Convolutional neural network (CNN) has local perception and parameter sharing and can learn different levels of spatial feature information. This weight-sharing network structure reduces the complexity of the network model, reduces the number of weights, and can effectively learn the corresponding features, while avoiding the complex feature extraction process. However, continuous convolution and pooling operations in the network destroy the original spatial information and reduce the resolution, which decreases the classification accuracy of picture. The fully convolutional network (FCN) [15] converts the fully connected layer at the end of the convolutional neural network into a convolutional layer. This makes it possible for the convolutional neural network to achieve pixel-level classification. Combining fully convolutional neural network and ELM (extreme learning machine) [16] has achieved the pixel-level classification of hyperspectral images.
We propose a classification structure based on the DeepLab v3+ network for the problem of how to use the limited hyperspectral image data to fully mine the feature information, thereby improving the accuracy of hyperspectral image classification. The main contributions of this article are summarized as follows:
(1)
We utilize DeepLab v3+ as the neural network structure to extract spatial domain features and merge them with spectral domain features. DeepLab v3+ is one of the fourth-generation DeepLab series of semantic segmentation networks developed by Google and has the best comprehensive performance so far. We are the first to apply the latest version of the DeepLab network to the hyperspectral image classification task.
(2)
We use the PCA method to reduce the dimensionality of the original hyperspectral image. In the spatial feature extraction stage, we select the first three principal components and the first principal component as training data and labels [16] to solve the dimensionality problem of the hyperspectral imagery.
(3)
We select different classifiers such as SVM, KNN, etc., to complete the task of hyperspectral image classification. We test and compare the classification accuracy under different conditions.

2. DeepLab v3+ and Proposed Framework

2.1. DeepLab v3+

DeepLab v3+ [17] is the latest work in the DeepLab series neural networks of semantic segmentation. Its predecessors include DeepLab v1, v2, and v3. The DeepLab series of neural networks use deep convolutional neural networks (DCNNs), which have excellent translation invariance so that they have good image-level classification capabilities. However, it is difficult for DCNNs to deal with the pixel-level classification problem, which is also an urgent problem for semantic segmentation. Therefore, in light of the following main problems, the author [17] has proposed solutions in different versions of DeepLab.
Continuous convolution and pooling in DCNNs inevitably lead to a decrease in the resolution of feature maps, which affects the final segmentation accuracy. Inspired by the paper [18], the author of DeepLab used the atrous convolution in the v2 version. The mathematical reasoning process of the hole convolution is described in detail [18] elsewhere; we discuss it briefly. Figure 1 represents the comparison of convolution principles in two different ways.
Let F : Z 2 R be a discrete function, and Ω r = [ r , r ] 2 Z 2   ( r > 0 ) . Let the convolution kernel be k : Ω r R , then the size of the convolution kernel is (2r + 1)2. The definition of discrete convolution can be obtained from the above conditions:
( F k ) ( p ) = s + t = p F ( s ) k ( t )
when considering the atrous convolution, let l be the hole coefficient. The atrous convolution l with the atrous coefficient l can be defined as:
( F l k ) ( p ) = s + l t = p F ( s ) k ( t )
Secondly, another major problem with DCNNs is that due to the fixed parameters of the fully connected layer at the end of the network, the input images need to have a fixed size. The usual solution is to crop or scale the input image. However, in the process of cropping and scaling, the target object in the picture is deformed, resulting in a decrease in recognition accuracy. To address this problem, atrous spatial pyramid pooling (ASPP) is added to the DeepLab series neural network combining the design ideas of atrous convolution and spatial pyramid pooling (SPP) [19], which are represented in Figure 2.
The spatial pyramid pooling structure is generally placed before the last fully connected layer of the network. When the feature map of the size ( w , h ) output by the previous layer is passed into the SPP, it is divided multiple times. The same image is divided into m × n image blocks of size ( w / m , h / n ) by the division scale ( m , n ) . Each image block in each divided image is put into the pooling layer separately and finally linked to a fully connected layer after joining. The biggest advantage of the structure is that it solves the size problem of the input picture. Due to the flexibility of the input size, the network can extract the convergence features under different pooling window sizes to improve the classification accuracy. Similarly, each feature map is separately convolved by convolution kernels with different dilation rates and then concatenated together in an atrous spatial pyramid pooling structure.
In the v3+ version, in order to integrate multiscale information, the v3 network is designed as an Encoder, and then the spatial resolution is restored in the Decoder structure. The Encoder–Decoder model is commonly used in semantic segmentation. During the test, DeepLab v3+ showed good segmentation performance. In the PASAL VOC 2012 dataset, it produced an mIOU index of 89.0, which is the best performance so far. Therefore, we choose DeepLab v3+ as the main network for feature extraction in our research. Figure 3a shows the network structure of DeepLab v3+.

2.2. Proposed Framework

Figure 3b shows the framework we proposed of hyperspectral image classification. The framework can be roughly divided into three parts: PCA dimensionality reduction, feature extraction and fusion, and the use of classifiers to achieve pixel-level classification. The specific details and mathematical principles of the framework are as follows.
The unprocessed hyperspectral image is regarded as a series of samples. Assume that the matrix size represented by the hyperspectral image is h × w × n , where h × w represents the resolution of the image and n represents the number of spectral channels. If x i R n is represented by a column vector, then the dimensionality reduction goal is to reduce the dimensionality of X R n × h × w composed of h × w samples to X ^ R k × h × w ( k n ) . The process of principal component analysis is as follows.
Let u i be the i t h principal component. The optimization is:
argmax u i ( j = 1 n | p r o j ( x j , u i ) | ) ,   ( x j , u i R n )
and the projection length of x j on u i :
p r o j ( x j , u i ) = x j u i || u i || 2 = x j T u i || u i || 2
Let || u i || = 1 , and we already know X = [ x 1 , x 2 , , x h × w ] . Hence, it can be shown:
( j = 1 n | p r o j ( x j , u i ) | ) = argmax u i ( u i T X X T u i )
The matrix X X T R n × n is positive definite. Therefore, u i T X X T u i is positive definite quadratic form, and it has a maximum value. Under the constraint of u i T u i = 1 , we obtain X X T u i = α u i by Lagrange multiplier method. α is the eigenvalue of the matrix X X T , and the objective function takes the maximum value when u i takes the eigenvector corresponding to α. We can get:
argmax u i ( u i T X X T u i ) = α u i T u i = α
where u i is the i t h principal component. Taking the eigenvectors of the first k eigenvalues and the corresponding matrix, U = [ u 1 , u 2 , , u k ] R n × k . Finally, a dataset reduced from n dimensions to k dimensions can be obtained by
X ^ = U T X R k × h × w
Using an established method [16], the dimensionality of original hyperspectral dataset is reduced by PCA. As a method of dimensionality reduction, PCA only reduces the dimensionality of the data while there is very little information loss. The introduction of PCA is to avoid the use of three-dimensional convolution in the feature extraction process which is computationally intensive. If we do not use PCA, three-dimensional convolution has to be used in all data, including training set and test set generally. Therefore, the introduction of PCA theoretically does not lead to biased results.
We take the first three principal components, namely k = 3 , as the dataset, and take the first principal component, namely k = 1 , as the label. All data processed by PCA are converted into uint8 format; therefore, in theory, they are all integers in the range of [ 0 , 255 ] . The main reason for using the first principal component as the label is that the data after dimensionality reduction remove the redundant information in the spectral domain and only a single channel is retained. The process retains the feature relationship of the object represented by each pixel in space, which is also an essential element for extracting spatial features.
To expand the number of samples so as to extract the spatial feature information abundantly, we use an established method [20] to crop the reduced-dimensional images and labels. The size of the cropping window is 45 × 45 pixels, the stride is 5 pixels, and the pictures and labels are cropped from left to right and top to bottom. Then, the cropped pictures and corresponding labels are used as the training set.
In the spatial feature extraction stage, we choose resnet50 as the backbone of DeepLab v3+ and set the label number to 256. The crossentropy loss function is the optimization target, as follows:
H ( p , q ) = i = 1 n p ( x i ) l o g ( q ( x i ) )
where p is the expected output and q is the actual result. Crossentropy is used to evaluate the difference between the current training probability distribution and the true distribution. Therefore, it is reasonable to reduce the loss function so that the probability distribution of the actual output and the probability distribution of the expected output are as close as possible. Other related training parameters are given in Table 1.
After the DeepLab v3+ network is trained, the dimensionality-reduced and not cropped picture with the size of W × H × 3 is input into the network, and the spatial feature with the size of W × H × 256 can be obtained without resolution loss. The pixel matrix with the size of W × H × K without dimensionality reduction is considered as a spectral domain feature and fused with spatial features. Since the ultimate purpose is to classify each pixel, it is necessary to fuse the spatial and spectral features of each pixel. The number of pixels in the picture is W × H . After normalizing the spatial and spectral features, respectively, the two features are stitched and integrated in the third dimension to complete the feature fusion process. In addition, the size of fusion feature is W × H × ( 256 + K ) . Finally, we select different classifiers, such as SVM, KNN, etc., and input the fusion features into the classifier for fitting and comparing.
In terms of evaluation, hyperspectral imagery classification generally uses overall accuracy (OA), average accuracy (AA), and κ coefficients, which are all calculated based on the confusion matrix. Among them, OA refers to the ratio of the number of correctly classified category pixels to the total number of pixels, AA refers to the average value of the recall of each category, and κ coefficient is an indicator of consistency.

3. Experimental Results

We selected three hyperspectral datasets: Pavia University [21], Kennedy Space Center [22], and HyRANK [23]. These hyperspectral datasets are public and open source. Pavia University is part of the hyperspectral data imaged in 2003 in the Italian city of Pavia. The size of the hyperspectral image is 610 × 340 pixels, with a total of 115 bands. After removing water vapor noise, there are 103 bands, in the wavelength range of 0.43–0.86 μm. The data contain nine types of labels, a total of 42,776 pixels. The Kennedy Space Center dataset was captured by the AVIRIS sensor on 23 March 1996. It contains 176 spectral bands after removing water vapor noise and 13 types of labels. The HyRANK dataset was acquired by the Hyperion sensor, and each image has 176 spectral bands. To test the generalizability of this feature extraction framework, two pictures with real labels in the dataset, Loukia and Dioni, were selected. The former includes 14 types of labels, and the latter includes 12 types of them. We randomly selected 5% of the pixels of each category in Pavia University and Kennedy Space Center, and 10% in HyRANK as training set, and used the rest as the test set.
Each pixel in the hyperspectral remote sensing image represents a ground distance from about tens to hundreds of meters. In other words, each pixel can be regarded as a sample, and an image contains tens of thousands of such samples. Each sample contains the information of spectral band, and there is also a spatial dimension relationship between them. Therefore, the data of hyperspectral image are sufficient.
The influence of different classifiers and feature types on the classification results are verified separately by controlling variables. In this process, we use the Pavia University dataset and the Kennedy Space Center dataset and compare the experimental results with other literature. To verify whether the proposed framework is equally effective for the new dataset in the feature extraction stage, we use the image Loukia in the HyRANK dataset as training set of DeepLab v3+ and the image Dioni as test set.

3.1. Comparison Between Different Classifiers

In this study, five machine learning classifiers, KNN, logistic regression, decision tree, naive Bayesian model, and SVM were selected, and the established classification algorithms [14,16], such as SRC-T (SRC classifier with diagonal weight matrix T), ELM, SVM-RBF (SVM classifier with radial basis function), SAE-LR (stacked AutoEncoders with logistic regression), and CNN, were compared with our framework. Table 2 and Table 3, respectively, show the classification accuracy of different classifiers for each category for the Pavia University dataset and Kennedy Space Center dataset. The results show that compared to the other four classifiers, SVM classifier performs better in both single category and overall indicators.
Different classifiers in machine learning have different theoretical basis, and their classification accuracy is also restricted by many external factors. For example, random forest classifier does not perform well in fitting data with many categories of features. In addition, the naive Bayes algorithm is often used in text classification, which is very sensitive to the expression form of the input data; therefore, it performs poorly in the task of HSIC. SVM classifier is different from other classification algorithms, since it does not involve probability measures and law of large numbers (LNN). The support vector plays a decisive role in the classification. Therefore, its own optimization goal is to minimize structured risk rather than experience risk, avoiding the traditional process from induction to deduction. By grabbing key samples to complete its inference process, SVM shows superior classification performance and robustness in hyperspectral imagery classification tasks. SVM is selected as the optimal classification algorithm of the framework proposed in this paper, and the classifier variables are fixed in the following comparative experiments.
To show the superiority of the framework, a horizontal comparison experiment is needed. Table 4 and Table 5, respectively, represent the differences in classification accuracy between this framework and some previous algorithms [14,16] for the Pavia University datasets and Kennedy Space Center datasets. Within the known range, the accuracy of the algorithm in this paper has reached the state of the art.

3.2. Comparison of Spectral Features, Spatial Features and Fusion Features

One of the important reasons why the framework can achieve high classification accuracy is the idea of feature fusion. To verify the impact of different features on hyperspectral imagery classification, we use spectral features, spatial features, and fusion features as feature vectors, respectively, to train the SVM classifier and check the differences in classification accuracy. Figure 4 shows the classification results obtained by different feature types for the Pavia University datasets and Kennedy Space Center dataset. When the fusion feature is used as a feature vector, the best classification effect can be achieved. However, when there are only spectral features or spatial features, the classification accuracy decreases to varying degrees. Generally, it is reasonable to believe that when the feature information of a certain category is more fully mined, the resulting higher-dimensional feature vector has a more positive impact on the classification and therefore improves the classification accuracy. For different types of label, the classification accuracy of fusion feature has less fluctuation, which enhances the robustness of the framework.
The spatial features extracted by the DeepLab v3+ network have a positive impact on subsequent classification. The classification framework based on DeepLab v3+ surpassed CNN, ELM, and other algorithms in this experiment. It extracted the explicit or implicit relationship in space between each category of pixels and other pixels. The spatial features obtained satisfy the requirements of classification tasks, which also indirectly reflects the powerful spatial feature extraction capabilities of DeepLab v3+ and its wide applicability.

3.3. Generalization Verification

In an ideal situation, the image after the principal component extraction would be fed to the DeepLab v3+ neural network, and the spatial feature information would be obtained after training. However, in real scenes, many data to be classified are unlabeled, which requires that the trained neural network can also extract spatial feature from unseen images. Therefore, two labeled pictures in the HyRANK dataset are used to verify the generalization of the feature extraction network. Among them, the picture Loukia is used as a known dataset, and the picture Dioni is used as an unknown dataset. After feature extraction and fusion, the classification results are obtained. Table 6 represents the classification result of the known dataset Loukia and the unknown dataset Dioni using different classifiers.
The results show that the classification accuracy of Loukia is relatively mediocre. The reason is that the same feature category is scattered and each area is small. Different types of feature categories are mixed together, which brings great difficulty to the classification. Apart from this, the difference in spectral characteristics between different types of ground features is not obvious enough, which is also a reason for the poor classification.
However, although Dioni is a new dataset for the feature extraction network, the framework can still perform the classification task excellently. Facing the same environmental conditions, the framework has a strong classification ability for hyperspectral data collected by the same sensor. Therefore, the framework has strong generalization ability and application prospects.

3.4. Visualization

To intuitively understand the impact of different factors on the classification, we also have done a series of data visualization. First, we use the visualization tool TSNE to reduce the dimensionality of the feature vector of each pixel to two dimensions, and project it to a 2-D coordinate system. By observing the degree of dispersion in each category, we can judge the difficulty of classification.
Figure 5 represents the distribution effects of pixels with spectral, spatial, and fusion features for the Pavia University dataset, Kennedy Space Center dataset, and HyRANK dataset. As we can see from the figure, the pixels with fusion features have the best degree of aggregation in each category, and the overlapping area between different categories is the least, making it is easy to be classified. This shows the importance and efficiency of the feature fusion method in hyperspectral imagery classification. Similarly, the degree of aggregation of Dioni’s pixel distribution is better than that of Loukia.
Figure 6 and Figure 7, respectively, represent the prediction classification maps for the Pavia University dataset and Kennedy Space Center dataset with different factors. After the spatial feature extraction of the DeepLab v3+ network, the feature fusion of spatial and spectral features, and the use of SVM classifiers for fitting and prediction produce the best classification map. Figure 8 represents the prediction classification maps for Loukia and Dioni. Compared to Dioni, the distribution of categories in Loukia is more scattered and complex; therefore, it is more difficult to be classifier.

4. Conclusions

In this paper, a hyperspectral imagery classification framework based on DeepLab v3+ is proposed. In the framework, DeepLab v3+ neural network is used as the spatial feature extraction method, and the spatial and spectral features are used for feature fusion. Finally, the SVM classifier is selected as the classification method among several classifiers. Compared with traditional machine learning algorithms and convolutional neural network algorithms in the same kind of dataset, our proposed framework not only significantly improves the classification accuracy but also improves the classification efficiency. Experimental results show that DeepLab v3+ has excellent spatial feature extraction capabilities and applicability in hyperspectral imagery classification, and the feature fusion method can effectively improve the classification accuracy of hyperspectral imagery. Experiments have shown that the framework proposed in this paper has good generalization, and the classification accuracy is better than other traditional machine-learning algorithms and deep-learning algorithms.

Author Contributions

Y.S. (Yifan Si) proposed the idea of the paper and was responsible for the preparation of related codes. D.G. was responsible for the writing of the paper. Y.G. and X.Z. participated in the guidance of related algorithms. Q.H. and J.E. polished the paper. S.H. and Y.S. (Yaoran Sun) are the people in charge of this project. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [the National Key Research and Development Program of China] grant number [2018YFC1407506], [Key Research and Development Program of Zhejiang Province] grant number [2021C03178], [the National Natural Science Foundation of China] grant number [11621101], [the Fundamental Research Funds for the Central Universities (Zhejiang University NGICS Platform)], [Ningbo Science and Technology Project] grant number [2020Z077 and 2020G012].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The authors are grateful to Tengfei Ma and Yiran Wu for valuable discussions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Harsanyi, J.C.; Chang, C.I. Hyperspectral image classification and dimensionality reduction: An orthogonal subspace projection approach. IEEE Trans. Geosci. Remote Sens. 1994, 32, 779–785. [Google Scholar] [CrossRef] [Green Version]
  2. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear versus nonlinear PCA for the classification of hyperspectral data based on the extended morphological profiles. IEEE Geosci. Remote Sens. Lett. 2011, 9, 447–451. [Google Scholar] [CrossRef] [Green Version]
  3. Green, A.A.; Craig, M.D.; Shi, C. The application of the minimum noise fraction transform to the compression and cleaning of hyper-spectral remote sensing data. Int. Geosci. Remote Sens. Symp. IEEE 1988, 3, 1807. [Google Scholar]
  4. Balasubramanian, M.; Schwartz, E.L.; Tenenbaum, J.B.; de Silva, V.; Langford, J.C. The lsomap algorithm and topological stability. Science 2002, 295, 7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Belkin, M.; Niyogi, P. Laplacian eigenmaps and spectral techniques for embedding and clustering. Adv. Neural Inf. Process. Syst. 2001, 14, 585–591. [Google Scholar]
  6. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  7. Ying, L.; Haokui, Z.; Qiang, S. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar]
  8. Samaniego, L.; Bárdossy, A.; Schulz, K. Supervised classification of remotely sensed imagery using a modified k-NN technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  9. Hasanlou, M.; Samadzadegan, F.; Homayouni, S. SVM-based hyperspectral image classification using intrinsic dimension. Arab. J. Geosci. 2015, 8, 477–487. [Google Scholar] [CrossRef]
  10. Kang, X.; Li, S.; Benediktsson, J.A. Spectral-spatial hyperspectral image classification with Edge-Preserving Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  11. Chen, X.; Wei, Z.; Li, M.; Rocca, P. A review of deep learning approaches for inverse scattering problems (Invited Review). Prog. Electromagn. Res. 2020, 167, 67–81. [Google Scholar] [CrossRef]
  12. Ma, T.; Lyu, H.; Liu, J.; Xia, Y.; Qian, C.; Evans, J.; Xu, W.; Hu, J.; Hu, S.; He, S. Distinguishing bipolar depression from major depressive disorder using fNIRS and deep neural network. Prog. Electromagn. Res. 2020, 169, 73–86. [Google Scholar] [CrossRef]
  13. Fajardo, J.E.; Galván, J.; Vericat, F.; Carlevaro, C.M.; Irastorza, R.M. Phaseless microwave imaging of dielectric cylinders: An artificial neural networks-based approach. Prog. Electromagn. Res. 2019, 166, 95–105. [Google Scholar] [CrossRef] [Green Version]
  14. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 7, 2094–2107. [Google Scholar] [CrossRef]
  15. Jonathan, L.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
  16. Li, J.; Zhao, X.; Li, Y.; Du, Q.; Xi, B.; Hu, J. Classification of hyperspectral imagery using a new fully convolutional neural network. IEEE Geosci. Remote Sens. Lett. 2018, 15, 292–296. [Google Scholar] [CrossRef]
  17. Chen, L.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. ECCV 2018, 2018, 801–818. [Google Scholar]
  18. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. In Proceedings of the International Conference on Learning Representations 2016, San Juan​, Puerto Rico, 2–4 May 2016; pp. 397–410. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Niu, Z.; Liu, W.; Zhao, J.; Jiang, G. DeepLab-based spatial feature extraction for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 251–255. [Google Scholar] [CrossRef]
  21. Pavia University Hyperspectral Satellite Dataset from the Telecommunications and Remote Sensing Laboratory, Ipavia University (Italy). Available online: https://www.kaggle.com/abhijeetgo/paviauniversity (accessed on 21 June 2018).
  22. Kennedy Space Center Hyperspectral Satellite Dataset from the NASA AVIRIS (Airborne Visible/Infrared Imaging Spectrometer) Instrument. Available online: http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 23 March 1996).
  23. HyRANK Hyperspectral Satellite Dataset I (Version v001) [Data Set]. Available online: http://doi.org/10.5281/zenodo.1222202 (accessed on 20 April 2008).
Figure 1. Illustrations of (a) convolution and (b) atrous convolution.
Figure 1. Illustrations of (a) convolution and (b) atrous convolution.
Applsci 11 05703 g001
Figure 2. Illustration of (a) spatial pyramid pooling (SPP) structure and (b) atrous spatial pyramid pooling (ASPP) structure in DeepLab v3+.
Figure 2. Illustration of (a) spatial pyramid pooling (SPP) structure and (b) atrous spatial pyramid pooling (ASPP) structure in DeepLab v3+.
Applsci 11 05703 g002
Figure 3. Schematic depictions of the architecture of (a) DeepLab v3+ and overall architecture of the (b) proposed framework.
Figure 3. Schematic depictions of the architecture of (a) DeepLab v3+ and overall architecture of the (b) proposed framework.
Applsci 11 05703 g003aApplsci 11 05703 g003b
Figure 4. Classification accuracy versus label with fusion feature, spectral feature, and spatial feature for the (a) Pavia University dataset and (b) Kennedy Space Center dataset.
Figure 4. Classification accuracy versus label with fusion feature, spectral feature, and spatial feature for the (a) Pavia University dataset and (b) Kennedy Space Center dataset.
Applsci 11 05703 g004
Figure 5. Feature distribution of pixels with fusion feature, spectral feature and spatial feature for the Pavia University dataset and Kennedy Space Center, and fusion feature for HyRANK dataset. (a) Fusion feature-PaviaU; (b) spatial feature-PaviaU; (c) spectral feature-PaviaU; (d) fusion feature-KSC; (e) spatial feature-KSC; (f) spectral feature-KSC; (g) fusion feature-Loukia; and (h) fusion feature-Dioni.
Figure 5. Feature distribution of pixels with fusion feature, spectral feature and spatial feature for the Pavia University dataset and Kennedy Space Center, and fusion feature for HyRANK dataset. (a) Fusion feature-PaviaU; (b) spatial feature-PaviaU; (c) spectral feature-PaviaU; (d) fusion feature-KSC; (e) spatial feature-KSC; (f) spectral feature-KSC; (g) fusion feature-Loukia; and (h) fusion feature-Dioni.
Applsci 11 05703 g005
Figure 6. Feature maps achieved by using different classifiers and features for the Pavia University dataset. (a) DT—fusion feature; (b) KNN—fusion feature; (c) LR—fusion feature; (d) NBM—fusion feature; (e) SVM—fusion feature; (f) SVM—spatial feature; and (g) SVM—spectral feature.
Figure 6. Feature maps achieved by using different classifiers and features for the Pavia University dataset. (a) DT—fusion feature; (b) KNN—fusion feature; (c) LR—fusion feature; (d) NBM—fusion feature; (e) SVM—fusion feature; (f) SVM—spatial feature; and (g) SVM—spectral feature.
Applsci 11 05703 g006
Figure 7. Feature maps achieved by using different classifiers and features for the Kennedy Space Center dataset. (a) DT—fusion feature, (b) KNN—fusion feature, (c) LR—fusion feature, (d) NBM—fusion feature; (e) SVM—fusion feature; (f) SVM—spatial feature; and (g) SVM—spectral feature.
Figure 7. Feature maps achieved by using different classifiers and features for the Kennedy Space Center dataset. (a) DT—fusion feature, (b) KNN—fusion feature, (c) LR—fusion feature, (d) NBM—fusion feature; (e) SVM—fusion feature; (f) SVM—spatial feature; and (g) SVM—spectral feature.
Applsci 11 05703 g007
Figure 8. Feature maps produced using SVM classifiers and fusion features for the HyRANK dataset. (a) The ground truth of Loukia. (b) The predicted feature map for Loukia. (c) The ground truth of Dioni. (d) The predicted feature map for Dioni.
Figure 8. Feature maps produced using SVM classifiers and fusion features for the HyRANK dataset. (a) The ground truth of Loukia. (b) The predicted feature map for Loukia. (c) The ground truth of Dioni. (d) The predicted feature map for Dioni.
Applsci 11 05703 g008
Table 1. Training parameters in DeepLab v3+.
Table 1. Training parameters in DeepLab v3+.
Output stride16
Learning rate0.001
Learning rate scheduler policyPoly
Weight decay0.0001
Table 2. Classification accuracy (%) for the Pavia university dataset via different classifiers we used in this paper.
Table 2. Classification accuracy (%) for the Pavia university dataset via different classifiers we used in this paper.
CLASSKNNLogistic RegressionDecision TreeNaive Bayesian ModelSVM
198.7293.7992.1696.1799.19
297.8198.3494.9898.0499.48
396.6997.0584.6785.9299.30
498.5797.0781.2547.6697.53
598.0099.9293.4884.5299.92
699.3998.4392.0686.6099.79
796.6390.2084.6877.1798.90
898.0189.3782.1978.5396.91
9100.0098.0499.22100.0099.78
OA98.1696.5191.4985.3399.10
AA98.2095.8089.6483.8598.98
Kappa97.5695.3688.7381.4098.81
Table 3. Classification accuracy (%) for the Kennedy Space Center dataset via different classifiers we used in this paper.
Table 3. Classification accuracy (%) for the Kennedy Space Center dataset via different classifiers we used in this paper.
CLASSKNNLogistic RegressionDecision TreeNaive Bayesian ModelSVM
193.8996.6586.0793.4798.89
280.0798.6548.6086.59100
398.1210076.5095.2497.98
497.0210052.3680.7599.13
599.3210045.1896.5896.20
610010052.9799.43100
710093.4640.13100100
899.7498.0287.8867.1091.42
996.3099.8084.3178.97100
1099.7210069.6599.6899.47
1195.3894.7593.1696.2198.74
1298.0999.3385.5892.1097.88
1310097.8999.5499.55100
OA96.6798.2279.5290.4998.47
AA96.7498.3570.9291.2198.44
Kappa96.2998.0277.2189.4098.29
Table 4. Classification accuracy (%) for the Pavia university dataset via different algorithms.
Table 4. Classification accuracy (%) for the Pavia university dataset via different algorithms.
CLASSSRC-TELMSVM-RBFCNNFCNProposed Framework
191.2064.4582.2194.9793.5899.19
296.7080.1177.4196.4496.7099.48
370.2070.6271.6884.6990.7899.30
493.4096.4094.7697.3996.7297.53
5100.0098.6099.9299.1488.5599.92
669.1075.9686.4094.7797.3399.79
772.4078.2086.2588.9092.3898.90
877.9079.3389.0584.1190.8396.91
992.8053.67100.00100.0088.2099.78
OA88.7077.7682.4994.3595.1199.10
Kappa84.8371.9380.7992.4993.2298.81
Table 5. Classification accuracy (%) for the Kennedy Space Center dataset via different algorithms.
Table 5. Classification accuracy (%) for the Kennedy Space Center dataset via different algorithms.
CLASSSAE-LRLinear SVMPCA
RBF-SVM
RBF-SVMProposed Framework
OA96.7395.5295.3596.5198.47
AA94.0891.9791.5793.9598.44
Kappa96.3695.0194.8296.1198.29
Table 6. Classification accuracy (%) for Loukia and Dioni.
Table 6. Classification accuracy (%) for Loukia and Dioni.
LoukiaDioni
ClassKNNLogistic RegressionDecision TreeNaive Bayesian ModelSVMKNNLogistic RegressionDecision TreeNaive Bayesian ModelSVM
168.0640.1518.8818.0367.8285.6679.1949.0440.4092.81
21001005010010093.8994.8931.33100100
383.8173.0351.4540.9490.498.5574.1449.5854.1095.73
459.3272.7310.8710010081.4488.8919.3778.26100
581.4966.1145.2935.7382.8794.6378.6061.8762.6094.57
681.4846.0732.0215.47100\\\\\
777.3366.6739.3240.3494.0897.4392.1481.9563.89100
872.9867.1650.4653.8482.19\\\\\
974.2271.0862.8580.1078.0297.2690.8287.7988.4595.34
1083.0669.763.5373.3173.296.7887.9083.6487.3493.84
1189.6676.6949.7849.4096.2595.4284.0160.5872.5390.79
1297.3184.5469.6057.9395.5199.5596.8482.3388.99100
1310099.6095.9110010098.5199.6697.8999.04100
1410099.0179.3391.8210099.7096.8291.7459.77100
OA81.4573.9260.7052.7682.3996.0688.0077.2676.9094.92
AA83.4873.7551.3861.2190.0394.9088.6666.4374.6196.92
Kappa77.8068.7253.4347.6678.6995.1085.0671.9271.8093.66
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Si, Y.; Gong, D.; Guo, Y.; Zhu, X.; Huang, Q.; Evans, J.; He, S.; Sun, Y. An Advanced Spectral–Spatial Classification Framework for Hyperspectral Imagery Based on DeepLab v3+. Appl. Sci. 2021, 11, 5703. https://doi.org/10.3390/app11125703

AMA Style

Si Y, Gong D, Guo Y, Zhu X, Huang Q, Evans J, He S, Sun Y. An Advanced Spectral–Spatial Classification Framework for Hyperspectral Imagery Based on DeepLab v3+. Applied Sciences. 2021; 11(12):5703. https://doi.org/10.3390/app11125703

Chicago/Turabian Style

Si, Yifan, Dawei Gong, Yang Guo, Xinhua Zhu, Qiangsheng Huang, Julian Evans, Sailing He, and Yaoran Sun. 2021. "An Advanced Spectral–Spatial Classification Framework for Hyperspectral Imagery Based on DeepLab v3+" Applied Sciences 11, no. 12: 5703. https://doi.org/10.3390/app11125703

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop