Next Article in Journal
Design and Verification of Humidity Sensors Based on Magnesium Oxide Micro-Arc Oxidation Film Layers
Next Article in Special Issue
Best Water Vapor Information Layer of Himawari-8-Based Water Vapor Bands over East Asia
Previous Article in Journal
Super Wide Band, Defected Ground Structure (DGS), and Stepped Meander Line Antenna for WLAN/ISM/WiMAX/UWB and other Wireless Communication Applications
Previous Article in Special Issue
Comparative Study on Planetary Magnetosphere in the Solar System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of CNN Algorithms on Hyperspectral Image Classification in Agricultural Lands

Graduate Institute of Communication Engineering, National Taiwan University, Taipei 10617, Taiwan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(6), 1734; https://doi.org/10.3390/s20061734
Submission received: 17 February 2020 / Revised: 12 March 2020 / Accepted: 18 March 2020 / Published: 20 March 2020
(This article belongs to the Special Issue Simulation Studies on Remote Sensing Scenarios)

Abstract

:
Several versions of convolutional neural network (CNN) were developed to classify hyperspectral images (HSIs) of agricultural lands, including 1D-CNN with pixelwise spectral data, 1D-CNN with selected bands, 1D-CNN with spectral-spatial features and 2D-CNN with principal components. The HSI data of a crop agriculture in Salinas Valley and a mixed vegetation agriculture in Indian Pines were used to compare the performance of these CNN algorithms. The highest overall accuracy on these two cases are 99.8% and 98.1%, respectively, achieved by applying 1D-CNN with augmented input vectors, which contain both spectral and spatial features embedded in the HSI data.

1. Introduction

Hyperspectral images (HSIs) contain abundant spectral-spatial information for versatile applications in agriculture, mineralogy, surveillance, physics, astronomy, chemical imaging and environmental sciences [1]. Classifications based on HSIs face the challenges of redundant features, limited number of available training samples and high dimensionality of data [2]. Many methods have been applied to classify HSIs, including random forest (RF) [3], k-nearest neighbors [4], multinomial logistic regression (MLR) [5], support vector machine (SVM) [6], convolutional neural network (CNN) [7], deep belief network (DBN) [8] and stacked auto-encoder (SAE) [9]. Among these methods, deep CNN has the potential of learning high-level spatial and spectral features [10,11].
Feature extraction and feature selection approaches were proposed to curtail the redundancy of information among hyperspectral bands [12]. By feature extraction, a projection matrix is used to map the original spectral data to a feature space while holding the dominant spectral information [13]. Typical feature extraction algorithms include principal component analysis (PCA) [14], linear discriminant analysis (LDA) [15], manifold learning [16], nonnegative matrix factorization (NMF) [17] and spatial-spectral feature extraction [18].
Feature selection means selecting part of the original bands based on proper criterion [19]. Typical feature selection algorithms include multitask sparsity pursuit [20], structure-aware [21], support vector machine [22], hypergraph model [23], sparse Hilbert–Schmidt independence criterion [24] and nonhomogeneous hidden Markov chain model [25]. Different measures were used to select preferred bands, including mutual information [12], information divergence [13], variance [26] and local spatial information [27]. However, these algorithms are time-consuming because the classifiers must be trained and tested again as the set of selected bands is changed. Pixels in an HSI are usually spatially correlated to their adjacent pixels [28], which can be exploited to complement the spectral information to achieve higher classification rate [29]. In [30], a spectral-spatial semisupervised training set construction was proposed to mitigate the problem of labeled data scarcity, in which unlabeled pixels are recruited for a class training subset if they belong to the same spectral cluster and are in the spatial neighborhood of a labeled pixel.
Deep spectral features embedded in the HSIs of agricultural vegetation can be physically related to the photosynthetic pigment absorption in the wavelengths of 400–700 nm, large spectral slope in 700–750 nm [31,32], liquid water inflection point in 1080–1170 nm [33], absorption by various leaf waxes and oils in 1700–1780 nm and cellulose absorption around 2100 nm, to name a few. The spectral features relevant to soil properties mainly appear in 2100–2300 nm [34,35]. These spectral features can be exploited for applications like precision agriculture [36,37], noxious weed mapping for rangeland management [38,39], forest health monitoring [40,41], vegetation stress analysis [42,43] and carbon sequestration site monitoring [44].
CNN has the potential of exploiting deep-level features embedded in its input data for classification, making it suitable for terrain classification with HSI data that contain both spatial and spectral features. Although CNNs have been widely used for classification in agricultural lands, there are always some outliers or misclassification between similar classes which contain similar spatial and spectral features. In this work, we present several versions of CNN, each taking different types of input vectors that include more feature information, to resolve these issues. These CNNs were trained and tested on the HSIs of Salinas Valley and Indian Pines, respectively. The former is a crop agriculture, the latter contains two-thirds crop agriculture and one-third forest or other natural perennial vegetation. The rest of this article is organized as follows. The 1D-CNN with input vector composed of pixelwise spectral data and spectral-spatial data are presented in Section 2 and Section 3, respectively, The 2D-CNN with input layer of principal components is presented in Section 4, simulation results are presented and analyzed in Section 5, and some conclusions are drawn in Section 6.

2. 1D-CNN with Pixelwise Spectral Data

Figure 1 shows an HSI cube which is composed of N x × N y pixels and each pixel contains spectral data in N bands. A one-dimensional (1D) input vector is prepared for each pixel by extracting the spectral data from that pixel. The input vectors from a selected set of training pixels are used to train the 1D-CNN shown in Figure 2, then the input vectors from another set of testing pixels are used to evaluate the performance of the 1D-CNN.
In the schematic of a 1D-CNN shown in Figure 2, the compositions of convp- n ( 1 × 2 ) and convp- 20 ( 20 × 2 ) are shown in Figure 3 and Figure 4, respectively, and FC ( 20 N × M ) is a fully connected layer shown in Figure 5.
Figure 3 shows the schematic of a convp- 20 ( 1 × 2 ) layer, where conv- n ( 1 × 2 ) is a convolutional layer composed of n filters of kernel size two, taking one input vector. The outputs of conv- n ( 1 × 2 ) are processed with batch normalization (BN), rectified linear unit (ReLU) activation function and maxpooling, MP(2), in sequence. The BN is used to make the learning process less sensitive to initialization. The input to a BN is a mini-batch of M input vectors, x ¯ m = [ x m 1 , x m 2 , , x m N ] t with 1 m M . The mean value and variance of the nth band in the bth mini-batch are computed as [45]
μ b n = 1 M m = 1 M x m n , σ b n 2 = 1 M m = 1 M ( x m n μ b n ) 2
Then, the original input vectors in the bth mini-batch are normalized as
x ˜ m n = x m n μ b n σ b n 2 + ϵ
where ϵ is a regularization constant to avoid divergence when σ b n 2 is too small. To further increase the degree-of-freedom in the subsequent convolutional layers, the normalized input vectors are scaled and shifted as
y m n = γ n x ˜ m n + β n
where the offset β n and the scaling factor γ n are updated during the training phase. The ReLU activation function is defined as y = max { 0 , x } , with input x and output y. The maxpooling function, MP(), reduces the computational load by picking the maximum from input data, which preserves the main characteristics of feature maps at the cost of coarser resolution.
Figure 4 shows the composition of a convp- 20 ( 20 × 2 ) layer, where conv- n ( 20 × 2 ) is a convolutional layer composed of n filters of kernel size two, taking 20 input feature maps.
Figure 5 shows the schematic of a fully connected layer, FC ( 20 N × M ) , which connects the feature maps from the last convolutional layer, convp- 20 ( 20 × 2 ) , to the input of the softmax function for final classification. The softmax function is a normalized exponential function defined as
softmax ( x ¯ ) = m = 1 M e x m 1 e x ¯
which enhances the largest component in x ¯ while making the sum of all its output components equal to one.

Band Selection Approach

Figure 6 shows the flowchart of a band selection (BS) approach based on CNN (BSCNN), which selects the best combination of spectral bands for classification. A CNN was first trained by using all the N spectral bands of training pixels, and its configuration remained the same during the subsequent band selection process. The BS process was executed L times. In each time, N bands ( N < N ) were randomly selected and the data in the other ( N N ) bands were reset to zero in the input vector. Among all these L combinations, the N bands delivering the highest overall accuracy is adopted for retraining the CNN for classification.
Figure 7 shows the preparation of input vectors with the selected N bands to retrain the 1D-CNN, as configured in Figure 2. The newly trained 1D-CNN is be used for classification.

3. 1D-CNN with Spectral-Spatial Data

Figure 8 shows the preparation of an augmented input vector by concatenating the spectral bands of the target pixel with the PCA data surrounding the target pixel, exploiting the spatial correlation between neighboring pixels. The PCA is first applied to all the spectral bands of each pixel to extract the first Q principal components. Then, the first Q principal components of all the R × R pixels surrounding the target pixel are collected, vectorized and concatenated to the original N bands of the target pixel to form an augmented input vector of dimension N + R × R × Q , to be input to the 1D-CNN.

4. 2D-CNN with Principal Components

Figure 9 shows the preparation of input layers, composed of principal components from each pixel, to be input to the 2D-CNN shown in Figure 10. The PCA is first applied to all the N spectral bands of each pixel to extract the first Q principal components [46]. The Q principal components from each of the R × R pixels surrounding the target pixel form an input layer associated with the target pixel. The PCA extracts the main features in the spectral dimension and exploits the spatial features embedded in the hyperspectral data.
Figure 10 shows the schematic of the 2D-CNN used in this work, where the compositions of convp- n ( 1 × 2 × 2 ) and convp- 20 ( 20 × 2 × 2 ) are shown in Figure 11 and Figure 12, respectively, FC ( 20 R × R × M ) is a fully connected layer shown in Figure 13. Cascading four convp- 20 ( 1 × 2 × 2 ) layers makes the resulting 2D-CNN highly nonlinear and enables it to recognize more abstract spatial-spectral features embedded in the hyperspectral data.
Figure 11 shows the schematic of a convp- 20 ( 1 × 2 × 2 ) layer, where conv- n ( 1 × 2 × 2 ) is a convolutional layer composed of n filters of kernel size 2 × 2 , taking one input layer. The outputs of conv- n ( 1 × 2 × 2 ) are processed with BN, ReLU activation function and MP( 2 × 2 ) in sequence.
Figure 12 shows the composition of a convp- 20 ( 20 × 2 × 2 ) layer, where conv- n ( 20 × 2 × 2 ) is a convolutional layer composed of n filters of kernel size 2 × 2 , taking 20 input feature maps.
Figure 13 shows the composition of a fully connected layer, FC ( 20 ( R × R ) × M ) , which connects the feature maps from the last convolutional layer, convp- 20 ( 20 × 2 × 2 ) , to the input of softmax function for final classification.

5. Simulations and Discussions

5.1. Salinas Valley HSI

Figure 14 shows the image and ground truth, respectively, in Salinas Valley, which were acquired with the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor in October 1998 [47]. The HSI is composed of 512 × 217 pixels, with spatial resolution of 3.7 m. The spectral data in wavelengths of 400–2500 nm were recorded in 224 bands, among which bands 108–112, 154–167 and 224 are removed for the concern of dense water vapor and atmospheric effects, leaving 204 more reliable bands. Table 1 lists the ground truth of 54,129 pixels in 16 classes [47].
The main objective of the AVIRIS project was to identify, measure and monitor the composition of Earth surface and atmosphere, based on the signatures of molecular absorption and particle scattering. Research with AVIRIS data were focused on understanding the processes related to global environment and climate change [48].
In the training phase of this work, 50% of pixels were randomly selected, labeled by ground-truth data, to determine the weights and biases associated with each neuron. Mini-batches of size M = 16 were used over 200 training epochs. The other 50% of pixels were then used in the testing phase to evaluate the classification performance.
Figure 15 shows the mean value and standard deviation of all the pixels in each of the 16 classes, over all 224 bands. The bands of dense water vapor and atmospheric effects are marked by a grey shade.
Figure 16a shows the classification image with 1D-CNN applied to 204 selected bands. The overall accuracy (OA) is 91.8%, which is the ratio of pixels correctly classified and the total number of testing pixels. Table 2a lists the producer accuracy (PA) of each class, which is the ratio of pixels correctly classified to a specific class and the total number of pixels which are classified to that specific class. The PAs of classes #8 and #15 are 91.9% and 54.8%, respectively, consistent with the observation on Figure 16a that classes #8 and #15 are apparently misclassified. Also notice that some spectral curves in Figure 15 nearly overlap in certain bands, which may cause errors in classification.
Figure 17 shows the effect of band number on the overall accuracy, which indicates that choosing N = 70 bands renders the highest overall accuracy. Table 2b lists the PAs of all 16 classes on the Salinas Valley HSI by applying BSCNN with 70 bands. By comparing with the results in Table 2a, the PA on class #8 decreases from 91.9% to 86.1% and that on class #15 increases from 54.8% to 74.8%. Figure 16b shows the classification image by applying the BSCNN.
Figure 8 shows the preparation of an augmented input vector by concatenating the N spectral bands of the target pixel and Q principal components from each of the R × R pixels surrounding the target pixel. By choosing N = 204 , Q = 1 and R = 21 , the augmented input vector has the dimension of N = 645 . With additional spatial information, the accuracy rate is expected to improve [49]. Figure 16c shows the classification image with augmented input vectors of 645 bands. Table 2c lists the PAs of all 16 classes, and the overall accuracy is 99.8%. By comparing Table 2a with Table 2b, the PAs on classes #8 and #15 are significantly increased to 99.8% and 99.7%, respectively.
Figure 16d shows the classification image by using 2D-CNN, with one principal component from each pixel to form an input layer. Table 2d lists the PAs of all 16 classes, and the overall accuracy is 99%.
In summary, the OA of 1D-CNN with 204 selected bands is the lowest of 91.8%, that of BSCNN with 70 selected bands is 93.2%, that of 1D-CNN with augmented input vectors of 645 bands is the highest of 99.8% and that of 2D-CNN with one principal component from each pixel is 99%.

5.2. Indian Pines HSI

Figure 18 shows a testing site in the Indian Pines, which was recorded on 12 June 1992, with AVIRIS sensors over the Purdue University Agronomy farm, northwest of West Lafayette. The image is composed of 145 × 145 pixels, each containing 220 spectral reflectance bands in wavelengths of 400–2500 nm. The number of bands is reduced to 200 after removing bands 104–108, 150–163 and 220, which suffer significant water absorption. Two-thirds of the test site was covered with agriculture land and one-third with forest or other natural perennial vegetation. There were also two dual-lane highways, a rail line, some low-density housings, other man-made structures and local roads. At the time of recording, some crops were growing, corn and soybeans were in their early stage of growth, with less than 5% coverage. Table 3 lists the available ground truth in 16 classes, which are not mutually exclusive.
In the training phase of this work, 50% of pixels were randomly selected, labeled by ground-truth data, to determine the weights and biases associated with each neuron. Mini-batches of size M = 16 were used over 200 training epochs. The other 50% of pixels were then used in the testing phase to evaluate the classification performance.
Figure 19 shows the mean value and standard deviation of all the pixels in each of the 16 classes, over the 200 selected bands.
Figure 20a shows the classification image with 1D-CNN applied to 200 selected bands, and the overall accuracy is 83.4%. Table 4a lists the PAs of all 16 classes, in which the lowest PAs are 55.2%, 57.1% and 62.2%, on classes #3, #7 and #15, respectively.
Figure 20b shows the classification image with augmented input vectors of 641 bands, where we choose N = 200 , Q = 1 and R = 21 . Table 4b lists the PAs of all 16 classes, with an overall accuracy of 95.4%. Compared to Table 4a, the PAs on classes #3, #7 and #15 are improved to 94%, 94.7% and 99.5%, respectively.
Figure 20c shows the classification images by applying 2D-CNN, with the input layer composed of one principal component from each pixel. Table 4c lists the PAs of all 16 classes, with the overall accuracy of 91.5%.
In summary, the OA of 1D-CNN with 200 selected bands is the lowest of 83.4%, that of 1D-CNN with augmented input vectors of 641 bands is 95.4%, and that of 2D-CNN with one principal component from each pixel is 91.5%.
Figure 21 shows the overall accuracy of 2D-CNN with different numbers of principal components. The highest OA is slightly below 98%, with 4, 30 or 60 principal components. Figure 20d shows the classification image by applying 1D-CNN with augmented input vectors of 1964 bands, where we choose N = 200 , Q = 4 and R = 21 . Table 4d lists the PAs of all 16 classes, with the overall accuracy of 98.1%. Compared to Table 4a, the PAs on classes #3, #7 and #15 are significantly improved to 98.6%, 100% and 99.5%, respectively.
The computational time for training and testing these CNNs as well as the resulting accuracy were affected by the size of input vector, input layer, convolution kernel, batch and epoch. Table 5 lists the CPU time for each CNN developed in this work, on a desktop PC with Intel® Core™ i7-8700 processor 3.2 GHz.
Both sets of HSI data used in this work were recorded in about 200 bands and classified into 16 labels. The pixel numbers are 54,129 and 21,025, respectively. The complexity of the CNNs adopted in this work seems suitable to these HSI datasets. It is conjectured that more complicated CNN configuration should be considered if more bands or more labels are involved. The results on these two cases show that the overall accuracy of the 1D-CNN with augmented input vectors is higher than 1D-CNN, BSCNN and 2D-CNN. The results of 2D-CNN turn out to be more accurate than conventional 1D-CNN, indicating that the spatial feature embedded in the spectral data can be useful. A small percentage of misclassification between similar classes can be resolved by applying the 1D-CNN with augmented input vectors which contain both the spatial and spectral features embedded in the HSI data.

6. Conclusions

Both the spectral and spatial features of HSI are exploited to increase the overall accuracy of image classification with several versions of 1D-CNN and 2D-CNN. The PCA was applied to extract significant spectral information while reducing the data dimension. These CNNs were applied to the HSI data on Salinas Valley and Indian Pines, respectively, to compare their accuracy rates of classification. The selection of band number and principal components was investigated by simulations. The highest OA on the Salinas Valley HSI is 99.8%, achieved by applying 1D-CNN to augmented input vectors of 645 bands, with one principal component from 21 × 21 pixels surrounding the target pixel. The highest OA on the Indian Pines HSI is 98.1%, achieved by applying 1D-CNN to augmented input vectors of 1964 bands, with four principal components from 21 × 21 pixels surrounding the target pixel. Possible misclassification between similar labels can be resolved by augmenting the input vectors to include more spatial and spectral features embedded in the HSI data.

Author Contributions

Conceptualization, T.-H.H. and J.-F.K.; Data curation, T.-H.H.; Formal analysis, T.-H.H. and J.-F.K.; Funding acquisition, J.-F.K.; Investigation, T.-H.H. and J.-F.K.; Methodology, T.-H.H. and J.-F.K.; Resources, J.-F.K.; Software, T.-H.H.; Supervision, J.-F.K.; Validation, J.-F.K.; Visualization, T.-H.H.; Writing—original draft, T.-H.H.; Writing—review and editing, J.-F.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Ministry of Science and Technology, Taiwan, under contract MOST 108-2221-E-002-030.

Acknowledgments

This work is partly supported by the donation from Pixart Imaging Inc. for promoting science and technology.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  2. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  3. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef] [Green Version]
  4. Ma, L.; Crawford, M.M.; Tian, J. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  5. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral-spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  6. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  8. Chen, Y.; Zhao, X.; Jia, X. Spectral-spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  9. Mughees, A.; Tao, L. Efficient deep auto-encoder learning for the classification of hyperspectral images. In Proceedings of the International Conference on Virtual Reality and Visualization, Hangzhou, China, 24–26 September 2016. [Google Scholar]
  10. Farooq, A.; Hu, J.; Jia, X. Weed classification in hyperspectral remote sensing images via deep convolutional neural network. In Proceedings of the IEEE International Symposium Geoscience and Remote Sensing (IGARSS), Valencia, Spain, 22–27 July 2018. [Google Scholar]
  11. Chova, L.G.; Tuia, D.; Moser, G.; Valls, G.C. Multimodal classification of remote sensing images: A review and future directions. Proc. IEEE 2015, 103, 1560–1584. [Google Scholar] [CrossRef]
  12. Martinez-Uso, A.; Pla, F.; Sotoca, J.M.; Garcia-Sevilla, P. Clustering-based hyperspectral band selection using information measures. IEEE Trans. Geosci. Remote Sens. 2007, 45, 4158–4171. [Google Scholar] [CrossRef]
  13. Chang, C.-I.; Wang, S. Constrained band selection for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1575–1585. [Google Scholar] [CrossRef]
  14. Rodarmel, C.; Shan, J. Principal component analysis for hyperspectral image classification. Surv. Land Inf. Sci. 2002, 62, 115. [Google Scholar]
  15. Joy, A.A.; Hasan, M.A.M.; Hossain, M.A. A comparison of supervised and unsupervised dimension reduction methods for hyperspectral image classification. In Proceedings of the 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), Cox’s Bazar, Bangladesh, 7–9 February 2019. [Google Scholar]
  16. Liao, D.; Qian, Y.; Tang, Y.Y. Constrained manifold learning for hyperspectral imagery visualization. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2018, 11, 1213–1226. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, X.; Zhong, Y.; Zhang, L.; Xu, Y. Spatial group sparsity regularized nonnegative matrix factorization for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6287–6304. [Google Scholar] [CrossRef]
  18. Feng, J.; Chen, J.; Liu, L.; Cao, X.; Zhang, X.; Jiao, L.; Yu, T. CNN-based multilayer spatial-spectral feature fusion and sample augmentation with local and nonlocal constraints for hyperspectral image classification. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2019, 12, 1299–1313. [Google Scholar] [CrossRef]
  19. Kong, Y.; Wang, X.; Cheng, Y. Spectral-spatial feature extraction for HSI classification based on supervised hypergraph and sample expanded CNN. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2018, 11, 4128–4140. [Google Scholar] [CrossRef]
  20. Yuan, Y.; Zhu, G.; Wang, Q. Hyperspectral band selection by multitask sparsity pursuit. IEEE Trans. Geosci. Remote Sens. 2015, 53, 631–644. [Google Scholar] [CrossRef]
  21. Zhu, G.; Huang, Y.; Lei, J.; Bi, Z.; Xu, F. Unsupervised hyperspectral band selection by dominant set extraction. IEEE Trans. Geosci. Remote Sens. 2016, 54, 227–239. [Google Scholar] [CrossRef]
  22. Patra, S.; Modi, P.; Bruzzone, L. Hyperspectral band selection based on rough set. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5495–5503. [Google Scholar] [CrossRef]
  23. Bai, X.; Guo, Z.; Wang, Y.; Zhang, Z.; Zhou, J. Semisupervised hyperspectral band selection via spectral-spatial hypergraph model. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2015, 8, 2774–2783. [Google Scholar] [CrossRef] [Green Version]
  24. Damodaran, B.B.; Courty, N.; Lefevre, S. Sparse Hilbert Schmidt independence criterion and surrogate-kernel-based feature selection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2385–2398. [Google Scholar] [CrossRef] [Green Version]
  25. Feng, S.; Itoh, Y.; Parente, M.; Duarte, M.F. Hyperspectral band selection from statistical wavelet models. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2111–2123. [Google Scholar] [CrossRef]
  26. Chang, C.-I.; Du, Q.; Sun, T.-L.; Althouse, M.L.G. A joint band prioritization and band-decorrelation approach to band selection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2631–2641. [Google Scholar] [CrossRef] [Green Version]
  27. Cao, X.; Xiong, T.; Jiao, L. Supervised band selection using local spatial information for hyperspectral image. IEEE Geosci. Remote Sens. Lett. 2016, 13, 329–333. [Google Scholar] [CrossRef]
  28. Mei, S.; He, M.; Wang, Z.; Feng, D. Spatial purity based endmember extraction for spectral mixture analysis. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3434–3445. [Google Scholar] [CrossRef]
  29. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in spectral-spatial classification of hyperspectral images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef] [Green Version]
  30. Ayerdi, B.; Graña, M. Hyperspectral image analysis by spectral-spatial processing and anticipative hybrid extreme rotation forest classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2627–2639. [Google Scholar] [CrossRef]
  31. Carter, G.A.; Knapp, A.K. Leaf optical properties in higher plants: Linking spectral characteristics to stress and chlorophyll concentration. Am. J. Bot. 2001, 88, 677–684. [Google Scholar] [CrossRef] [Green Version]
  32. Zarco-Tejada, P. Chlorophyll fluorescence effects on vegetation apparent reflectance: I. Leaf-level measurements and model simulation. Remote Sens. Environ. 2000, 74, 582–595. [Google Scholar] [CrossRef]
  33. Penuelas, J.; Filella, I.; Biel, C.; Save, R.; Serrano, L. The reflectance at the 950–970 nm region as an indicator of plant water status. Int. J. Remote Sens. 1993, 14, 1887–1905. [Google Scholar] [CrossRef]
  34. Galvao, L.S.; Pizarro, M.A.; Epiphanio, J.C.N. Variations in reflectance of tropical soils: Spectral-chemical composition relationships from AVIRIS data. Remote Sens. Environ. 2001, 75, 245–255. [Google Scholar] [CrossRef]
  35. Summers, D.; Lewis, M.; Ostendorf, B.; Chittleborough, D. Visible near-infrared reflectance spectroscopy as a predictive indicator of soil properties. Ecol. Indicators 2011, 11, 123–131. [Google Scholar] [CrossRef]
  36. Strachan, I.; Pattey, E.; Boisvert, J.B. Impact of nitrogen and environmental conditions on corn as detected by hyperspectral reflectance. Remote Sens. Environ. 2002, 80, 213–224. [Google Scholar] [CrossRef]
  37. Haboudane, D.; Miller, J.R.; Pattey, E.; Zarco-Tejada, P.J.; Strachan, I.B. Hyperspectral vegetation indices and novel algorithms for predicting green LAI of crop canopies: Modeling and validation in the context of precision agriculture. Remote Sens. Environ. 2004, 90, 337–352. [Google Scholar] [CrossRef]
  38. Evangelista, P.; Stohlgren, T.J.; Morisette, J.T.; Kumar, S. Mapping invasive tamarisk (tamarix): A comparison of single-scene and time-series analyses of remotely sensed data. Remote Sens. 2009, 1, 519–533. [Google Scholar] [CrossRef] [Green Version]
  39. Wang, L. Invasive species spread mapping using multi-resolution remote sensing data. Int. Arch. Photogram. Remote Sens. Spatial Info. Sci. 2008, 37, 135–142. [Google Scholar]
  40. Lawrence, R.; Labus, M. Early detection of Douglas-Fir beetle infestation with subcanopy resolution hyperspectral imagery. Western J. Appl. For. 2003, 18, 202–206. [Google Scholar] [CrossRef] [Green Version]
  41. Sampson, P.H.; Zarco-Tejada, P.J.; Mohammed, G.H.; Miller, J.R.; Noland, T.L. Hyperspectral remote sensing of forest condition: Estimating chlorophyll content in tolerant hardwoods. For. Sci. 2003, 49, 381–391. [Google Scholar]
  42. Carter, G.A.; Miller, R.L. Early detection of plant stress by digital imaging within narrow stress-sensitive wavebands. Remote Sens. Environ. 1994, 50, 295–302. [Google Scholar] [CrossRef]
  43. Smith, K.L.; Steven, M.D.; Colls, J.J. Use of hyperspectral derivative ratios in the red-edge region to identify plant stress responses to gas leaks. Remote Sens. Environ. 2004, 92, 207–217. [Google Scholar] [CrossRef]
  44. Bellante, G.J.; Powell, S.L.; Lawrence, R.L.; Repasky, K.S.; Dougher, T.A. Aerial detection of a simulated CO2 leak from a geologic sequestration site using hyperspectral imagery. Int. J. Greenhouse Gas Control 2013, 13, 124–137. [Google Scholar] [CrossRef]
  45. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  46. Jolliffe, I.T.; Cadima, J. Principal component analysis: A review and recent developments. Phil. Trans. R. Soc. A 2016, 374, 20150202. [Google Scholar] [CrossRef]
  47. Hyperspectral Remote Sensing Dataset. Available online: http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 15 August 2019).
  48. Airborne Visible/Infrared Imaging Spectrometer. Available online: https://aviris.jpl.nasa.gov/ (accessed on 15 August 2019).
  49. Zhu, Z.; Woodcock, C.E.; Rogan, J.; Kellndorfer, J. Assessment of spectral, polarimetric, temporal, and spatial dimensions for urban and peri-urban land cover classification using Landsat and SAR data. Remote Sens. Environ. 2012, 117, 72–82. [Google Scholar] [CrossRef]
  50. Baumgardner, M.F.; Biehl, L.L.; Landgrebe, D.A. 220 Band AVIRIS Hyperspectral Image Data Set: June 12, 1992 Indian Pine Test Site 3; Purdue University Research Repository: West Lafayette, ID, USA, 2015. [Google Scholar]
Figure 1. The 1D input vector retrieved from a hyperspectral image (HSI) cube.
Figure 1. The 1D input vector retrieved from a hyperspectral image (HSI) cube.
Sensors 20 01734 g001
Figure 2. Schematic of a 1D convolutional neural network (CNN).
Figure 2. Schematic of a 1D convolutional neural network (CNN).
Sensors 20 01734 g002
Figure 3. Schematic of convp- 20 ( 1 × 2 ) layer.
Figure 3. Schematic of convp- 20 ( 1 × 2 ) layer.
Sensors 20 01734 g003
Figure 4. Schematic of convp- 20 ( 20 × 2 ) .
Figure 4. Schematic of convp- 20 ( 20 × 2 ) .
Sensors 20 01734 g004
Figure 5. Schematic of FC ( 20 N × M ) .
Figure 5. Schematic of FC ( 20 N × M ) .
Sensors 20 01734 g005
Figure 6. Flowchart of band selection (BS) approach based on CNN (BSCNN).
Figure 6. Flowchart of band selection (BS) approach based on CNN (BSCNN).
Sensors 20 01734 g006
Figure 7. Input vectors to the 1D-BSCNN.
Figure 7. Input vectors to the 1D-BSCNN.
Sensors 20 01734 g007
Figure 8. Preparation of an augmented input vector by concatenating pixelwise spectral bands and the principal component analysis (PCA) data surrounding the target pixel.
Figure 8. Preparation of an augmented input vector by concatenating pixelwise spectral bands and the principal component analysis (PCA) data surrounding the target pixel.
Sensors 20 01734 g008
Figure 9. Preparation of input layers for 2D-CNN.
Figure 9. Preparation of input layers for 2D-CNN.
Sensors 20 01734 g009
Figure 10. Schematic of the 2D-CNN.
Figure 10. Schematic of the 2D-CNN.
Sensors 20 01734 g010
Figure 11. Schematic of convp- 20 ( 1 × 2 × 2 ) .
Figure 11. Schematic of convp- 20 ( 1 × 2 × 2 ) .
Sensors 20 01734 g011
Figure 12. Schematic of convp- 20 ( 20 × 2 × 2 ) .
Figure 12. Schematic of convp- 20 ( 20 × 2 × 2 ) .
Sensors 20 01734 g012
Figure 13. Schematic of FC ( 20 ( R × R ) × M ) .
Figure 13. Schematic of FC ( 20 ( R × R ) × M ) .
Sensors 20 01734 g013
Figure 14. Images of Salinas Valley, centered at 36 ° 40 39.8568 N, 121 ° 39 19.8072 W [47], area size is 1895 × 803 m, (a) grey-tone image, (b) ground truth, (c) legends of 16 classes.
Figure 14. Images of Salinas Valley, centered at 36 ° 40 39.8568 N, 121 ° 39 19.8072 W [47], area size is 1895 × 803 m, (a) grey-tone image, (b) ground truth, (c) legends of 16 classes.
Sensors 20 01734 g014
Figure 15. (a) Mean value and (b) standard deviation of all the pixels in each of the 16 classes, over all 224 bands in Salinas Valley HSI.
Figure 15. (a) Mean value and (b) standard deviation of all the pixels in each of the 16 classes, over all 224 bands in Salinas Valley HSI.
Sensors 20 01734 g015
Figure 16. Classification images on Salinas Valley HSI, (a) 1D-CNN with 204 bands, overall accuracy (OA) is 91.8%, (b) BSCNN with 70 selected bands, OA is 93.2%, (c) 1D-CNN with augmented input vectors of 645 ( 21 × 21 + 204 ) bands, OA is 99.8%, (d) 2D-CNN with input layer composed of one principal component from each pixel, OA is 99%, color code the same as in Figure 14.
Figure 16. Classification images on Salinas Valley HSI, (a) 1D-CNN with 204 bands, overall accuracy (OA) is 91.8%, (b) BSCNN with 70 selected bands, OA is 93.2%, (c) 1D-CNN with augmented input vectors of 645 ( 21 × 21 + 204 ) bands, OA is 99.8%, (d) 2D-CNN with input layer composed of one principal component from each pixel, OA is 99%, color code the same as in Figure 14.
Sensors 20 01734 g016
Figure 17. Effect of band number in BSCNN on overall accuracy.
Figure 17. Effect of band number in BSCNN on overall accuracy.
Sensors 20 01734 g017
Figure 18. Images of Indian Pines, centered at 40 ° 27 37.9 N, 87 ° 00 32.7 W [50], area size is 3.2 × 3.2 km, (a) false color image, (b) ground truth, black area is unlabeled, (c) legends of 16 classes.
Figure 18. Images of Indian Pines, centered at 40 ° 27 37.9 N, 87 ° 00 32.7 W [50], area size is 3.2 × 3.2 km, (a) false color image, (b) ground truth, black area is unlabeled, (c) legends of 16 classes.
Sensors 20 01734 g018
Figure 19. (a) Mean value and (b) standard deviation of all the pixels in each of 16 classes over all 200 bands in Indian Pines HSI.
Figure 19. (a) Mean value and (b) standard deviation of all the pixels in each of 16 classes over all 200 bands in Indian Pines HSI.
Sensors 20 01734 g019
Figure 20. Classification images on Indian Pines HSI: (a) 1D-CNN with 200 bands, OA is 83.4%, (b) 1D-CNN with augmented input vectors of 641 ( 21 × 21 + 200 ) bands, OA is 95.4%, (c) 2D-CNN with input layer composed of one principal component from each pixel, OA is 91.5%, (d) 1D-CNN with augmented input vectors of 1964 ( ( 21 × 21 ) × 4 + 200 ) bands, OA is 98.1%, color code the same as in Figure 18.
Figure 20. Classification images on Indian Pines HSI: (a) 1D-CNN with 200 bands, OA is 83.4%, (b) 1D-CNN with augmented input vectors of 641 ( 21 × 21 + 200 ) bands, OA is 95.4%, (c) 2D-CNN with input layer composed of one principal component from each pixel, OA is 91.5%, (d) 1D-CNN with augmented input vectors of 1964 ( ( 21 × 21 ) × 4 + 200 ) bands, OA is 98.1%, color code the same as in Figure 18.
Sensors 20 01734 g020
Figure 21. Overall accuracy versus number of principal components in 2D-CNN.
Figure 21. Overall accuracy versus number of principal components in 2D-CNN.
Sensors 20 01734 g021
Table 1. Summary of ground truth in Figure 14.
Table 1. Summary of ground truth in Figure 14.
#ClassSample Number
1broccoli green weeds 12009
2broccoli green weeds 23726
3fallow1976
4fallow rough plow1394
5fallow smooth2678
6stubble3959
7celery3579
8grapes untrained11,271
9soil vineyard develop6203
10corn senesced green weeds3278
11lettuce romaine 4 weeks1068
12lettuce romaine 5 weeks1927
13lettuce romaine 6 weeks916
14lettuce romaine 7 weeks1070
15vineyard untrained7268
16vineyard vertical trellis1807
Table 2. Producer accuracy (%) on Salinas Valley HSI: (a): 1D-CNN with 204 bands, (b): BSCNN with 70 selected bands, (c): 1D-CNN with augmented input vectors of 645 bands, (d): 2D-CNN with input layer composed of one principal component from each pixel.
Table 2. Producer accuracy (%) on Salinas Valley HSI: (a): 1D-CNN with 204 bands, (b): BSCNN with 70 selected bands, (c): 1D-CNN with augmented input vectors of 645 bands, (d): 2D-CNN with input layer composed of one principal component from each pixel.
#1#2#3#4#5#6#7#8#9#10#11#12#13#14#15#16OA
(a)99.899.899.899.998.110099.891.999.698.29299.810097.554.899.791.8
(b)99.899.895.699.498.399.999.886.110097.599.310099.199.174.899.393.2
(c)99.899.910099.999.910099.999.899.999.710099.999.810099.799.999.8
(d)99.710099.810099.610010097.699.799.910010010010097.199.899
Table 3. Summary of ground truth in Figure 18 [50].
Table 3. Summary of ground truth in Figure 18 [50].
#ClassSample Number
1alfalfa46
2corn-notill1428
3corn-mintill830
4corn237
5grass-pasture483
6grass-trees730
7grass-pasture-mowed28
8hay-windrowed478
9oats20
10soybean-notill972
11soybean-mintill2455
12soybean-clean593
13wheat205
14woods1265
15buildings-grass-trees-drives386
16stone-steel-towers93
Table 4. Producer accuracy (%) on Indian Pines HSI: (a): 1D-CNN with 200 bands, (b): 1D-CNN with augmented input vectors of 641 bands, (c): 2D-CNN with input layer composed of one principal component from each pixel, (d): 1D-CNN with augmented input vectors of 1964 bands.
Table 4. Producer accuracy (%) on Indian Pines HSI: (a): 1D-CNN with 200 bands, (b): 1D-CNN with augmented input vectors of 641 bands, (c): 2D-CNN with input layer composed of one principal component from each pixel, (d): 1D-CNN with augmented input vectors of 1964 bands.
#1#2#3#4#5#6#7#8#9#10#11#12#13#14#15#16OA
(a)83.376.155.28294.496.557.198.777.873.387.284.99996.662.291.783.4
(b)95.292.89497.589.199.294.799.671.489.797.588.410098.999.510095.4
(c)90.582.287.49290.796.510010058.382.195.18810097.796.997.791.5
(d)10094.998.697.599.299.710010090.996.798.897.29999.299.598.298.1
Table 5. Computational time for training and testing CNNs.
Table 5. Computational time for training and testing CNNs.
HSI on Salinas ValleyCPU TimeHSI on Indian PinesCPU Time
1D-CNN with 204 bands1 h 43 min1D-CNN with 200 bands18 min
BSCNN with 70 selected bands1 h 35 min1D-CNN with augmented input vectors of 641 bands26 min
1D-CNN with augmented input vectors of 645 bands6 h 58 min2D-CNN with input layer composed of one principal component from each pixel17 min
2D-CNN with input layer composed of one principal component from each pixel1 h 32 min1D-CNN with augmented input vectors of 1964 bands1 h 27 min

Share and Cite

MDPI and ACS Style

Hsieh, T.-H.; Kiang, J.-F. Comparison of CNN Algorithms on Hyperspectral Image Classification in Agricultural Lands. Sensors 2020, 20, 1734. https://doi.org/10.3390/s20061734

AMA Style

Hsieh T-H, Kiang J-F. Comparison of CNN Algorithms on Hyperspectral Image Classification in Agricultural Lands. Sensors. 2020; 20(6):1734. https://doi.org/10.3390/s20061734

Chicago/Turabian Style

Hsieh, Tien-Heng, and Jean-Fu Kiang. 2020. "Comparison of CNN Algorithms on Hyperspectral Image Classification in Agricultural Lands" Sensors 20, no. 6: 1734. https://doi.org/10.3390/s20061734

APA Style

Hsieh, T. -H., & Kiang, J. -F. (2020). Comparison of CNN Algorithms on Hyperspectral Image Classification in Agricultural Lands. Sensors, 20(6), 1734. https://doi.org/10.3390/s20061734

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop