Next Article in Journal
Evaluation of the Ocean Surface Wind Speed Change following the Super Typhoon from Space-Borne GNSS-Reflectometry
Previous Article in Journal
Rheology of the Zagros Lithosphere from Post-Seismic Deformation of the 2017 Mw7.3 Kermanshah, Iraq, Earthquake
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification

1
Harbin Institute of Technology, Shenzhen 518055, China
2
Shenzhen Key Laboratory of Internet Information Collaboration, Shenzhen 518055, China
3
Department of Information Systems, City University of Hong Kong, Hong Kong 999077, China
4
School of Computer Science and Engineering, Nanyang Technological University, Singapore 628798, Singapore
5
School of Information Engineering, East China Jiaotong University, Nanchang 330000, China
*
Authors to whom correspondence should be addressed.
Remote Sens. 2020, 12(12), 2033; https://doi.org/10.3390/rs12122033
Submission received: 23 May 2020 / Revised: 13 June 2020 / Accepted: 15 June 2020 / Published: 24 June 2020

Abstract

:
Accurate hyperspectral image classification has been an important yet challenging task for years. With the recent success of deep learning in various tasks, 2-dimensional (2D)/3-dimensional (3D) convolutional neural networks (CNNs) have been exploited to capture spectral or spatial information in hyperspectral images. On the other hand, few approaches make use of both spectral and spatial information simultaneously, which is critical to accurate hyperspectral image classification. This paper presents a novel Synergistic Convolutional Neural Network (SyCNN) for accurate hyperspectral image classification. The SyCNN consists of a hybrid module that combines 2D and 3D CNNs in feature learning and a data interaction module that fuses spectral and spatial hyperspectral information. Additionally, it introduces a 3D attention mechanism before the fully-connected layer which helps filter out interfering features and information effectively. Extensive experiments over three public benchmarking datasets show that our proposed SyCNNs clearly outperform state-of-the-art techniques that use 2D/3D CNNs.

Graphical Abstract

1. Introduction

With the rapid development of optics and photonics, hyperspctral sensors have been installed in many satellites. Hyperspectral image classification is a fundamental and yet challenging task whose purpose is to label each pixel contained in a hyperspectral image. Based on the rich spatial–spectral information preserved in hyperspectral images, it enables us to distinguish different objects of interest in the scene. They have been widely used in a variety of fields such as precise agriculture, environmental surveillance, and astronomy [1]. For example, Brown et al. [2] proposed a linear mixture model by combining various absorption band methods on CRISM to determine the mineralogy of the surface on Mars.
Hyperspectral images classification has attracted more and more research attention. Conventional image classification techniques such as support vector machine (SVM) [3] and K-nearest neighbor (KNN) classifier have achieved reasonable performance for this task as they take into account rich spectral information [4] captured in the hyperspectral images. Wang et al. [5] classified the hyperspectral images by using the Locality Adaptive Discriminant Analysis (LADA) algorithm to reduce the dimensionality of hyperspectral images. There also exist several other approaches to cope with this issue. For example, Wang et al. [6] proposed a dimensionality reduction method for hyperspectral image classification by utilizing the manifold ranking algorithm to perform band selection. Furthermore, Yuan et al. [7] proposed a novel dual clustering-based band selection approach for hyperspectral image classification. While these methods have demonstrated superior classification performance, they are not effective to classify hyperspectral images under complex scenarios.
Recently, with the big success of deep learning, the convolutional neural networks (CNNs) [8,9,10,11] based approaches have achieved excellent performance for various image analysis related tasks, e.g., image classification and object recognition. To classify hyperspectral images, both the spectral and spatial perspectives should be taken into account. Intuitively, hyperspectral image consists of hundreds of “images”, each of which represents a narrow wavelength band (visible or none-visible) of the electromagnetic spectrum, also known as spectral perspective. Meanwhile, the spatial perspective refers to the 2D spatial information about the objects contained in the hyperspectral images. Thus, hyperspectral images are usually represented as the 3D spectral–spatial data. Consequently, several approaches have been proposed in the literature. However, existing CNN-based approaches [12,13], which focus on either spatial or spectral features alone, inevitably overlook the interweaving relationships between the spatial and the spectral perspectives of objects captured in the hyperspectral images. Essentially, the interweaving information could be leveraged to further improve classification performance due to the following reasons. First, the interweaving relationships can further enrich the feature space for the later classification task. Second, if spectral–spatial information could be used simultaneously, the classifier could be co-supervised by different perspectives of the labeled data. As a result, a relatively good CNN-based classifier can be trained by using a small number of labeled 3D spectral–spatial images only.
To simultaneously model spectral–spatial information, some pioneer attempts have been made along this line [14,15,16,17,18,19,20,21,22,23]. These approaches perform stacked convolution operations over spatial and spectral feature space in a layer by layer manner, called 3D CNN models. Obviously, the advantage of this kind of 3D CNN model lies in the generated rich feature maps. However, the main disadvantages of these approaches are threefold. First, it is difficult to generate a deeper 3D CNN model. The reason is that the solution space exponentially increases with the increasing number of 3D convolution operations, which limits the depth of the model and the interpretation ability of the model. Second, the memory cost is too high if a large number of 3D convolution operations are invoked. Third, more training examples are needed to train a deeper 3D CNN model which is not practical as the public hyperspectral image datasets are rather small. To cope with aforementioned challenges, this paper is proposed to design a novel 3D CNN model which only needs few 3D convolution operations but can generate richer feature maps.
In this paper, we propose the Synergistic 2D/3D Convolutional Neural Network (SyCNN) to classify hyperspectral images. In the proposed SyCNN, the 2D CNN component and the 3D CNN component are mixed together. Different from the conventional 3D CNNs that stack up 3D convolution layer by layer, the proposed SyCNN, as shown in Figure 1, integrates 3D CNNs with 2D CNNs to learn the salient features. Then, a data interaction module is proposed which fuses the 2D features and 3D features together. Last but not least, a 3D attention mechanism is designed to drop out the less important features. Experimental results on three benchmark datasets have demonstrated that the proposed SyCNNs outperform a number of state-of-the-art hyperspectral image classifiers. Our major research contributions are summarized as follows.
  • First, we design a basic SyCNN model (SyCNN-S) that comprises 2D CNNs and 3D CNNs to extract rich feature maps. To classify the hyperspectral images, 2D CNNs and 3D CNNs are separately trained in the SyCNN-S model before the final-connected layer. Then, the 2D output features and 3D output features are concatenated and sent to the final-connected layer to label the pixel by the softmax function.
  • Second, we design an end-to-end hyperspectral image classification framework—deep SyCNN (SyCNN-D) by integrating data interaction module into the mixture model to generate richer feature maps. At each round of the model fusion process, the 2D CNNs and 3D CNNs are equally fused together to generate deeper and more sophisticated feature maps, and then data interaction is invoked to support the cross-domain transfer.
  • Third, we propose a more sophisticated SyCNN model, namely SyCNN-ATT, by incorporating the 2D and 3D attention mechanisms into the SyCNN model to drop out the less relevant features.
  • Finally, a number of extensive experiments have been conducted based on three different datasets, and the experimental results have demonstrated that the proposed SyCNN model is superior to other state-of-the-art hyperspectral image classifiers.
The rest of the paper is organized as follows. Section 2 briefly reviews the related work. In Section 3, we first carefully design the architecture of the proposed SyCNN as well as the data interaction module, and then we propose the deep SyCNNs for hyperspectral image classification task. Experimental results are reported in Section 4 and we conclude the paper in Section 5.

2. Related Work

Hyperspectral image classification has long been studied in the literature. However, most existing works are based on the conventional techniques. In this section, we only review the latest deep learning-based approaches and these approaches can be classified into two categories, i.e., Traditional Classification Methods and Deep Learning Models.

2.1. Traditional Classification Methods

A number of hyperspectral images classification methods had been proposed recently. For example, Bandos et al. [24] proposed a linear discriminant method to classify hyperspectral images. However, it cannot perform feature selection on low spectral resolution images. To automatically separate overlapping absorption bands, Bandos et al. [25] further developed a noise-insensitive method. There also exist some methods such as quadratic discriminant analysis and logarithmic discriminant analysis, which are developed to address the nonlinearity. While these methods have achieved good results, they suffer from the Huges phenomenon. For example, Wang et al. [5] proposed a novel dimensionality reduction method, named LADA to classify the hyperspectral images. Wang et al. [6] also proposed a manifold ranking based salient selection method for hyperspectral images classification. This method first finds some representative bands by employing an evolution algorithm, and then selects salient bands by using a manifold ranking strategy.
Many researchers developed kernel-based methods to improve the model performance. The kernel-based methods firstly projected samples into a high dimensional space, and then the projected samples of different classes are linearly separable. Camps et al. [26] introduced the SVM as a kernel trick in the hyperspectral image classification. Rakotomamonjy et al. [27] proposed a multiple kernel learning (MKL) method to classify hyperspectral images, which can learn a kernel and a classification predictor at the same time. However, the kernel-based methods cloud not explicitly exploit a spatial context. The composite kernel (CK) method was proposed to address this problem. In [28], the CK method was generalized by utilizing extended multi-attribute profiles (EMAP). A generalized composite kernel (GCK) was proposed to exploit both extended multi-attribute profiles and raw hyperspectral features in [29].

2.2. Deep Learning Models

Recently, deep learning models have achieved a significant performance on image classification task. Deep learning based methods are introduced into hyperspectral images classification problem to learn hierarchical representations from the hyperspectral remote sensing images.

2.2.1. 2D CNN-Based Approaches

LeCun firstly deigned the CNN [30,31] model on 1996. Due to the back propagation (BP) mechanism, the proposed CNN model can be trained easily and achieves a very good performance in handwritten digit recognition. A lot of deep learning based methods are developed in many fields. For example, Glorot et al. [32] improved the CNNs by introducing the Rectified Linear Units (ReLU) as the activation function. It can alleviate the vanishing gradient problem and the ineffective exploration problem caused by the BP mechanism. Krizhevsky et al. [8] designed the AlexNet network with ReLU activation. However, the deeper of the deep learning network will lead to the overfitting problem. To address this phenomenon, Srivastava et al. [33] proposed the dropout mechanism after the convolutional operation to reduce the noisy samples. At the same time, Szegedy et al. [10] proposed the GoogLeNet model to classify the images, which is a deep CNN model with each layer comprising multi-scale convolutional operations. He et al. [11] introduced the residual mechanism in the network, and proposed a deep residual CNN model, namely ResNet.
There also have a lot of works for hyperspectral image classification based on CNN. To extract the spectral–spatial information contained in hyperspectral images, a 2D CNN-based approach was proposed in [12], where 2D CNN was utilized to explore the band selection results generated by the adapted adaBootst-based SVM classifier. Based on the band selection results, several methods were proposed to fuse certain 2D CNN networks for hyperspectral images classification. For example, Liu et al. [34] utilized a deep belief netowkrs (DBNs) to extract deep spectral features. Ran et al. [35] proposed a deep band sensitive convolutional network to classify the hyperspectral images and the proposed network has two sub-components. One component is to extract the spectral features and another one is to aggregate there features from different bands. In [36], the author proposed an attention-based convolutional neural network to perform band-selection for hyperspectral images. Furthermore, Liu et al. [13] proposed a semi-supervised 2D CNN model including the encoder, the corrupted encoder and the decoder component. The classification model is trained to minimize the reconstruction error between the input images and the mixture of both labeled and unlabeled images. One of the state-of-the-art approach is proposed in [14]. In this paper, a deep 2D CNN model is proposed to directly label each pixel in the hyperspectral images. Alternatively, the two-streamlike 2D CNN approaches [37,38] were proposed with each stream of the proposed network was a pre-trained 2D CNN model. While these approaches could achieve comparably good model performance, a large training dataset was a prerequisite which is unaffordable in most real-world applications.

2.2.2. Three-Dimensional CNN-Based Approaches

In [39], the 3D CNN approach was first proposed to learn discriminative features to recognize object actions from the spatial-temporal data. Later, Chen et al. [17] proposed a deep 3D CNN network which stacked up several 3D convolutional layers to extract spectral–spatial feature maps for classification. Similarly, a deep fully convolutional network (FCN) with a focus on 3D data was proposed by Lee et al. [16]. Different from [16], Li et al. [18] proposed a 3D CNN network which stacked up 3D convolutional layers without the pooling layer. The proposed model can well capture the changes of local signals contained in the spectral–spatial data. Hamida et al. [15] designed a deep 3D CNN network for hyperspectral image classification, where the pooling layers were replaced by the spectral–spatial 3D convolutional layers. Furthermore, there were some hybrid models that combined 2D CNNs with 3D CNNs. Obviously, the 3D CNN-based approaches have more parameters than the 2D CNN-based approaches. Accordingly, both the model complexity and the memory consumption of the 3D CNN-based approaches are much higher than that of the 2D CNN-based approach. As a result, Sun et al. [40] tried to replace the 3D convolutional layer with a mixture of a 2D spatial convolutional layer and a 1D temporal convolutional layer which could largely alleviate the aforementioned problem. Unfortunately, this approach extracted the spatial features and spectral features separately and thus ignored the interweaving relationships between the spatial and the spectral perspectives of the hyperspectral images.
Apparently, these 3D CNN-based approaches cannot be directly adapted to model such interweaving relationships due to the following reason. Most existing 3D CNN-based approaches have several stacked 3D convolutional layers and thus is hard to directly optimize the estimation loss through such nonlinear structure. To take the advantages of both the 2D and 3D CNN-based approaches, this paper proposed to fuse the features extracted from 2D and 3D convolution operations in an iterative manner and synergistically trained a hybrid 2D/3D CNN model with relatively few training examples.

3. The Proposed SyCNN and Deep SyCNN Network

In this section, we first introduce the 3D convolution operation. Then, we introduce the proposed SyCNN with more technical details. Finally, we propose a deep SyCNN network for hyperspectral image classification task.

3.1. 3D Convolution Operation

Generally speaking, a 3D spectral–spatial hyperspectral image can be represented by a tensor and its size is S × H × W × C , where S denotes the spectral domain (comprising a range of wavelengths), H and W are the height and width in the spatial domain, and C represents the number of channels. Different from the 2D convolution operation, the kernels of a 3D convolution operation are formulated as a 4D tensor ψ n k × s k × h k × w k , where s k , h k , w k denote the kernel size in S , H , and W, respectively, and n k is the number of kernels. The 3D convolution operation is illustrated in Figure 2. In Figure 2, it is easy to observe that the 3D convolution operation takes the 3D spectral–spatial features I = I s , h , w as the inputs and produces the 3D feature maps O = o s , h , w by executing the convolution operation along both the spatial and spectral dimensions of the inputs. Accordingly, the 3D convolution operation can be formulated as follows:
O = ψ I , o s 0 , h 0 , w 0 = [ q s 0 , h 0 , w 0 1 , q s 0 , h 0 , w 0 2 , , q s 0 , h 0 , w 0 n k ] T , q t 0 , h 0 , w 0 n = Σ t , w , h ψ n , t , h , w · I s , w , h t 0 h 0 w 0 ,
where I s , w , h s 0 h 0 w 0 denotes that the kernel starts from the position ( s 0 , h 0 , w 0 ) of the input image I and ends at the position ( s , w , h ) of the input I. The q s 0 , h 0 , w 0 n is the value at the cell of ( s 0 , h 0 , w 0 ) on the n-th output feature map by the n-th kernel.

Data Interaction

The proposed data interaction module for generating 2D and 3D feature maps is illustrated in Figure 3. This module consists of the cross-domain transfer operation and the local cross-domain connection operation. Let O 2 denote the 2D output feature maps, and O 3 denote the 3D output feature map. Moreover, t is the time slot, then we have:
O 2 t = ϕ ( I 2 t ) + I 2 t , O 3 t = ψ ( I 3 t ) + I 3 t ,
where I 2 t h × w and I 3 t s × h × w are the 2D input features and the 3D input features at time t, respectively. ϕ denotes the 2D convolution and ψ is the 3D convolution. As shown in Figure 3, the 2D features and 3D features are interacted by using the cross-domain transfer module and the local cross-domain connection module.
We first transfer the 3D features obtained by the 3D convolution to the new 2D features O 2 t ´ , and then transfer the 2D features obtained by the 2D convolution to the new 3D features O 3 t ´ , and we have:
O 2 t ´ = τ ( O 3 t ) , O 3 t ´ = β ( O 2 t ) ,
where β denotes the transfer operation from 2D features to 3D features, and τ denotes the transfer operation from 3D features to 2D features. The transferred features are then concatenated with the corresponding 2D or 3D features to form the final output features, written as:
O 2 t = O 2 t ´ + O 2 t , O 3 t = O 3 t ´ + O 3 t .
It is easy to observe that the final output 2D features O 2 t contain not only the 2D spatial feature information but also the 3D spectral feature information. Similarly, the output 3D features O 3 t also contain spatial and spectral feature information. The new 2D features and 3D features will be fed into the next 2D convolutional and 3D convolutional operations as the input, respectively. By doing so, we already generate more training samples to train the 3D CNNs, and the feature maps of 2D CNNs are also spectral–spatial information. Thus the data interaction module not only reduces the complexity of the proposed SyCNN in learning 3D convolution kernels, but also generates rich feature maps to train the SyCNN. Due to its prowerful learning and extraction abilities, the proposed SyCNN can greatly alleviate the overfitting problem.

3.2. Deep SyCNN Network

To perform hyperspectral image classification, we design a hybrid deep model by stacking 2D convolution and 3D convolution which is synergistically trained. As shown in Figure 4a, the proposed simple mixed model consists of three 2D/3D convolutional operations, which directly connect to the fully-connected layer to classify unlabeled images. We design a simple yet efficient deep SyCNN network by stacking the SyCNNs together. The proposed SyCNN model is an end-to-end network and takes the hyperspectral images as input. The proposed deep SyCNN network consists of three SyCNNs which only involves 3D convolutions, and the model is illustrated in Figure 4b. To further improve the model performance, we introduce the attention mechanism with a focus on generating salient features to facilitate hyperspectral image classification task. The proposed deep SyCNN-attention model is plotted in Figure 4c. Furthermore, it is a none-trivial task to pre-process the input data. Note that the BN-inception [41] and the ReLU function [32] are applied after each convolutional block in the proposed model, which is used to address the overfitting problem caused by the sparse training data and limited training samples in hyperspectral images. For simplicity reason, these BN and ReLu layers are not plotted in the figure. In order to allow the input images of any length, we use a global pooling layer as the last layer of the network.
The difference between the proposed approach and the state-of-the-art 3D CNN models [14,19] is that the deep SyCNN requires fewer 3D convolution operations for the spectral–spatial fusion stage, and yet it can generate deeper and richer feature maps. Moreover, different from the conventional 3D CNN based models, the deep SyCNN approach can take full advantage of 2D CNN approaches, and yet it can be trained using a much smaller image data set.

4. Experimental Results

To evaluate the performance of the proposed deep SyCNN, rigorous experiments are performed on three benchmark hyperspectral image datasets. We implemented the proposed approaches as well as several state-of-the-art approaches on these datasets. The experimental settings as well as the evaluation criteria are given in the following subsections. The proposed approaches are called deep simple SyCNN network (SyCNN-S), deep SyCNN network with data interaction module (SyCNN-D), and deep SyCNN-attention network (SyCNN-ATT), respectively.

4.1. Experimental Settings

In these experiments, three widely adopted benchmark hyperspectral image datasets, i.e., Indian Pines Scene, Botswana Scene, and Kennedy Space Center datasets, are chosen for model performance evaluation.
The Indian Pines Scene dataset: This dataset is collected by the AVIRIS sensors located at north-western of India, U.S., on 1992. It contains 145 × 145 pixels in spatial dimension in the image, and 200 pixels in spectral dimension. Due to the presence of 20 noisy bands, we only used 200 hyperspectral bands in this experiment. Specifically, it removes the bands covering the regions of water absorption, i.e., [104–108], [150–163], 220. The ground truth available includes 16 classes. The false-color image and the corresponding ground reference map are shown in Figure 5. The numbers of the training and test samples per class are listed in Table 1.
The Botswana Scene dataset: This dataset is collected by the Hyperion sensors on the NASA EO-1 satellite over the Okavango Delta on 31 May 2001. It has 1476 × 1476 pixels in the spatial dimension and 145 corrected bands in spectral dimension. Following with the Indian Pines Scene dataset, we removed the noisy bands to produce an experimental data set containing 145 bands. This image dataset contains 14 categories. The false-color image and the corresponding ground reference map are shown in Figure 6. The numbers of the training and test samples per class are listed in Table 2.
The Kennedy Space Center (KSC) dataset: This dataset is captured by using the AVIRIS sensor over Florida on 23 March 1996. It has 512 × 614 × 172 pixels in spectral–spatial dimension. Dut to remove 48 noisy bands, it obtains 172 spectral bands. The image dataset contains 13 labeled classes. The false-color image and the corresponding ground reference map are shown in Figure 7. The numbers of the training and test samples per class are listed in Table 3.
In the experiments, these datasets are randomly partitioned into training and testing datasets at the ratio of 70 % to 30 % . The proposed approaches are implemented on a NVIDIA GTX1080ti GPU machine using Pytorch [42]. The convolutional operation with the same color share the same parameters of each proposed model. We adopt the Adam gradient descent optimizer with an initial learning rate of 1 × 10 4 to train our models. The mini-batch size is set to 100. The drop out ratio and weight decay rate are respectively set to 0.5 and 5 × 10 4 .

4.2. Baseline Models and Evaluation Criteria

We compare our approaches with both conventional and the state-of-the-art approaches which are illustrated as follows.
  • SVM-Grid A support vector machine based approach optimized by the stochastic gradient descent algorithm. This conventional approach is one of the most widely adopted algorithms for hyperspectral image classification.
  • NN A simple neural network (NN) model with 4 fully connected layers with dropout, and this approach is considered as the baseline deep learning model.
  • Sharma A 2D CNN model consists of three 2D convolutional operations with band selection [12], originally proposed for hyperspectral face recognition. It is a dimension reduction method, where 1D convolutional operation is introduced to lead to a drastic reduction in the number of parameters used.
  • Liu is a semi-supervised 2D CNN model [13] for hyperspectral image classification, and it includes one convolutional operation, one clean encoder, one corrupted encoder, and one decoder.
  • Hamida is a 3D deep learning approach for hyperspectral image classification consisting of four 3D convolutions [15].
  • Lee is a fully convolutional network (FCN) [43] which does not use any subsampling layer with arbitrary sizes. It contains two 3D convolutions and eight 2D convolutions [16].
  • Chen is a deep 3D CNN model consisting of three 3D convolutions [17].
  • Li is a 3D CNN model having two 3D convolutions and a fully-connected layer for image classification [18]. Different from other 3D CNNs, the 3D convolution in this work has a fixed spatial size and only slightly changes its spectral depth.
  • He is a multi-scale 3D deep learning network [19] consisting of two 3D convolutions and two 3D convolution-blocks, and each block contains four 3D convolutions.

4.3. Evaluation Criteria

Three widely adopted evaluation criteria are chosen in the experiments to evaluate the model performance which are overall accuracy (OA), the F1 score and Kappa coefficient (K). The OA is the mean accuracy of each category. The F1 score is also called balanced F score which is harmonic average value of precision and recall value of each category. The Kappa coefficient (K) is an index value measuring whether the classification results are consistent with the underlying true classes or not. For hyperspectral image classification, we mainly check whether such inconsistency is caused by “accidental” factors or “inevitable” factors. These evaluation criteria are calculated as follows.
P r e c i s o n = T P T P + F P , R e c a l l = T P T P + F N , A c c u r a c y = T P + T N T P + T N + F N + F P , O A = ( 1 n k A c c u r a c y k ) , F 1 = ( 1 n k 2 · P r e c i s o n k · R e c a l l k P r e c i s o n k + R e c a l l k ) 2 , K = N i = 1 n x i i i = 1 n ( x i + × x + i ) N 2 i = 1 n ( x i + × x + i ) .
where T P denotes the true positive value, F P denotes the false positive value, and F N denotes the false negative value. n is the number of categories, and N is the total number of data samples. x i i denotes the number of categories on the diagonal of the confusion matrix, x i + is the total number of the i-th row, and x + i denotes the total number of the i-th column.

4.4. Experimental Results

4.4.1. Experimental Results on the Indian Pines Scene Dataset

The comparative model performance based on the Indian Pines Scene dataset with respect to OA, F1 and K are depicted in Table 4. From this table, it is obvious that the model performance of our proposed approaches is superior to both baseline methods and the state-of-the-art approaches. Among all proposed approaches, the SyCNN is the best one with respect to all evaluation criteria and this is consistent with our expectation. Moreover, the proposed deep SyCNN-attention model can achieve a comparably good model performance to the proposed SyCNN. Among all compared methods, the Sharma achieves the best classification results. It is also well noticed that the performance of 2D CNN based approaches is better than that of 3D CNN based approaches. Apparently, the reason is that the number of training examples is rather limited which only favors the less complicate models such as the Sharma.
To further evaluate the model performance, the experimental results of all approaches are plotted in Figure 8. It can be seen from Figure 8 that the proposed SyCNN model performs the best result among all methods. The possible reasons are as follows: (i) It is a synergistic trained model, which consists of 2D convolutions and 3D convolutions to generate deeper and richer features; (ii) it makes full use of the 2D and 3D features by the data interaction module. Since the parameters of the deep 3D CNN models need more training samples, we find that the Sharma and Liu methods produce the second better results than the methods based on 3D CNNs, such as Hamida, Lee, Li, Chen, and He. It’s observed that the worst model is NN, which is a baseline deep learning model for hyperspectral image classification. The reasons are that the NN model only has one 2D convolution but 4 fully connected layers for prediction task and the 2D convolution is not sufficient enough to generate the deep and rich features.
According to the Table 4, the Sharma approach can achieve the best model performance among all compared models. However, the model performance of all proposed approaches is better than that of the Sharma. Among all proposed approaches, the results of the evaluation criteria, i.e., OA, F1 and K, of the proposed SyCNN-S is respectively increased by 0.26%, 0.36% and 0.2% when compared with the Sharma. The SyCNN-D can further improve the model performance by 0.228%, 0.24% and 0.3% with respect to OA, F1 and K. The SyCNN-ATT is the best one among all proposed approaches. Apparently, the 3D CNN based approaches cannot achieve desired better model performance than 2D CNN based approaches which could also be found in the visualized results plotted in Figure 8.

4.4.2. Experimental Results Based on the Botswana Scene Dataset

Table 5 reports the evaluation results based on the Botswana Scene dataset. Botswana Scene dataset only has 3248 samples, and the labels are unbalanced. Similarly, the proposed approaches can achieve the best classification performance. Due to the data interaction module, the proposed SyCNN-D could extract the deeper and richer features from the hyperspectral images. The 2D convolutional operation could extract much more spatial information by treating the whole spectral information as channels, but it loses the relationship between different spectral bands. While the 3D convolutional operation cloud extract much more spatial–spectral fusion information by operating on spectral information, it could only combine a few spectral information with a fixed size and need more training samples for the next operation. In the data interaction module, the 2D extracted features could be projected into 3D features and the projected 3D features are fused in the 3D extracted features. Thus, the new 3D features combine more spectral information and make full use of spatial information. In addition, the new 3D features are sufficient to train the next 3D convolutional operation. So as the new 2D features that have more spectral information are also sufficient to train the next 2D convolutional operation. The Sharma and Liu methods are the second best models, and the rest 3D CNN based approaches are the worst models except for the NN model.
We also visualize the corresponding experimental results evaluated based on Botswana Scene dataset in Figure 9. Clearly, the SyCNN-ATT model is the best one among all the experimental methods.

4.4.3. Experimental Results Based on the Kennedy Space Center Dataset

Table 6 reports the evaluation results based on the Kennedy Space Center dataset. Similar observations of the proposed methods are obtained, that is, they achieve better classification performance than other baseline methods. Among all proposed methods, the SyCNN-ATT method is the best one. Again, the outstanding performance of the proposed deep SyCNN-ATT model can be easily observed based on the visualized results that are plotted in Figure 10.

4.4.4. Effects on the Ratio of Training Examples

Apart from the adopted evaluation criteria, we also examined how the ratio of training examples could affect a model’s classification performance. In this experiment, we varied the ratio of training examples from 10 % to 70 % and the experimental results based on all datasets are plotted in Figure 11. It is obvious that the performance of all models improves with the increasing number of training examples. This experimental result suggests that the deep neural network-based hyperspectral image classifiers generally need more training data to achieve better classification performance. It is also observed that the proposed approaches can successfully alleviate overfitting and outperform other approaches when only 10 % training data was used. Based on the result of this experiment, we can conclude that the proposed three SyCNN-based methods can achieve good classification performance even a small training set is provided. Such a characteristic of the SyCNN-based methods is promising as the available training data for most real-world image classification applications is usually quite limited. It’s more capable to handle spatial–spectral information and each part of the SyCNN also helps to improve performance. We note that the design of the SyCNN makes it possible to produce a promising result based on a small dataset.

4.4.5. Comparison of Parameters and Times of Different Methods

The comparison results about execution time of different models are reported in Table 7. The experiments are implemented on Botswana dataset. It could be seen that the proposed methods have a number of network parameters, including parameters of 2D CNNs, data interact module, and 3D CNNs. However, it does not significantly increase the model complexity. The execution time is not as high as Liu’s. The possible reason is that the model has more operations such as convolutional operations. It also can be seen that the proposed methods outperform all methods.

5. Conclusions

In this paper, we first propose a simple synergistic trained deep learning model, which is constructed by mixing 2D CNNs and 3D CNNs to extract deeper spatial–spectral features with fewer 2D/3D convolutions. We further present a deep SyCNN network for hyperspectral image classification, which introduces a data interaction module into the simple synergistic trained 2D/3D model. Experiment results show that our proposed deep SyCNN model has obtained a robust, good result and outperforms state-of-the-art methods for image classification. We also introduce 3D attention module into the deep SyCNN model to gain a state-of-the-art performance.
One explanation of the performance improvement of the SyCNN network could be that a deep network generally benefits by deepening the model. Another is based on the observation that there is stronger spectral redundancy than a spatial one in hyperspectral images. Thus, we propose a synergistic trained deep learning model that consists of 2D CNNs and 3D CNNs to further enhance the abstraction ability on the spatial domain and explore more useful information on the spectral domain. In the future, we will seek to design a optimal deeper SyCNN network for hyperspectral image classification. On the other hand, we will introduce a relation network into the SyCNN to learn the relationship between 2D and 3D output features, which is proposed to enhance the feature extraction process.

Author Contributions

X.Y., this paper first author, write the manuscript and correct it. X.Z., the Correspondence author and correct this paper. Y.Y., R.Y.K.L., S.L., X.L. and X.H., these are co-authors and also correct this paper. X.Z., Y.Y., R.Y.K.L., X.L. and X.H. also provide the funding to help me. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Key R&D Program of China, 2018YFB0504900, 2018YFB0504905 and the Shenzhen Science and Technology Program under Grant JCYJ20170811160212033, Grant JCYJ20160330163900579, Grant JCYJ20180507183823045, and Grant JCYJ20170413105929681. Zhang’s work was supported by the National Science Foundation of China under Grant No.61872108, Shenzhen Science and Technology Program under Grant No.201708113000098. Lau’s work was supported in part by the Research Grant Council of the Hong Kong SAR (Projects: CityU 11502115 and CityU 11525716), the National Natural Science Foundation of China (NSFC) Basic Research Program (Project 71671155), and CityU Shenzhen Research Institute. X. Huang was supported in part by Natural Science Foundation of Jiangxi Province under Grant No.20192ACBL21006.

Acknowledgments

This work was partially supported by the National Key R&D Program of China, 2018YFB0504900, 2018YFB0504905 and the Shenzhen Science and Technology Program under Grant JCYJ20170811160212033, Grant JCYJ20160330163900579, Grant JCYJ20180507183823045, and Grant JCYJ20170413105929681. Zhang’s work was supported by the National Science Foundation of China under Grant No.61872108, Shenzhen Science and Technology Program under Grant No.201708113000098. Lau’s work was supported in part by the Research Grant Council of the Hong Kong SAR (Projects: CityU 11502115 and CityU 11525716), the National Natural Science Foundation of China (NSFC) Basic Research Program (Project 71671155), and CityU Shenzhen Research Institute. X. Huang was supported in part by Natural Science Foundation of Jiangxi Province under Grant No.20192ACBL21006.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fauvel, M.; Benediktsson, J.A.; Boardman, J.; Brazile, J.; Bruzzone, L.; Camps-Valls, G.; Chanussot, J.; Gamba, P.; Gualtieri, A.; Marconcini, M.; et al. Recent Advances in Techniques for Hyperspectral Image Processing. Remote Sens. Environ. 2007, 113, S110–S120. [Google Scholar]
  2. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  3. Scholkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2001. [Google Scholar]
  4. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef] [Green Version]
  5. Wang, Q.; Meng, Z.; Li, X. Locality adaptive discriminant analysis for spectral–spatial classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2077–2081. [Google Scholar] [CrossRef]
  6. Wang, Q.; Lin, J.; Yuan, Y. Salient band selection for hyperspectral image classification via manifold ranking. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1279–1289. [Google Scholar] [CrossRef]
  7. Yuan, Y.; Lin, J.; Wang, Q. Dual-clustering-based hyperspectral band selection by contextual analysis. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1431–1445. [Google Scholar] [CrossRef]
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  9. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  10. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  12. Sharma, V.; Diba, A.; Tuytelaars, T.; Van Gool, L. Hyperspectral CNN for Image Classification & Band Selection, with Application to Face Recognition; KU Leuven, ESAT: Leuven, Belgium, 2016. [Google Scholar]
  13. Liu, B.; Yu, X.; Zhang, P.; Tan, X.; Yu, A.; Xue, Z. A semi-supervised convolutional neural network for hyperspectral image classification. Remote Sens. Lett. 2017, 8, 839–848. [Google Scholar] [CrossRef]
  14. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.; Zhang, X.; Huang, X. Hyperspectral Image Classification with Deep Learning Models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  15. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D Deep Learning Approach for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  16. Lee, H.; Kwon, H. Contextual deep CNN based hyperspectral classification. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 3322–3325. [Google Scholar]
  17. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  18. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef] [Green Version]
  19. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  20. Luo, Y.; Zou, J.; Yao, C.; Zhao, X.; Li, T.; Bai, G. HSI-CNN: A Novel Convolution Neural Network for Hyperspectral Image. In Proceedings of the 2018 International Conference on Audio, Language and Image Processing, Shanghai, China, 16–17 July 2018; pp. 464–469. [Google Scholar]
  21. Fan, J.; Chen, T.; Lu, S. Superpixel Guided Deep-Sparse-Representation Learning for Hyperspectral Image Classification. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 3163–3173. [Google Scholar] [CrossRef]
  22. Fan, J.; Chen, T.; Lu, S. Unsupervised Feature Learning for Land-Use Scene Recognition. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2250–2261. [Google Scholar] [CrossRef]
  23. Fan, J.; Tan, H.L.; Toomik, M.; Lu, S. Spectral–spatial hyperspectral image classification using super-pixel-based spatial pyramid representation. In Proceedings of the Image and Signal Processing for Remote Sensing XXII, Edinburgh, UK, 26–29 September 2016; Volume 10004, pp. 315–321. [Google Scholar]
  24. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of hyperspectral images with regularized linear discriminant analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  25. Brown, A.J. Spectral curve fitting for automatic hyperspectral data analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1601–1608. [Google Scholar] [CrossRef] [Green Version]
  26. Camps-Valls, G.; Bruzzone, L. Kernel-based methods for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1351–1362. [Google Scholar] [CrossRef]
  27. Rakotomamonjy, A.; Bach, F.R.; Canu, S.; Grandvalet, Y. Simplemkl. J. Mach. Learn. Res. 2008, 9, 2491–2521. [Google Scholar]
  28. Dalla Mura, M.; Atli Benediktsson, J.; Waske, B.; Bruzzone, L. Extended profiles with morphological attribute filters for the analysis of hyperspectral data. Int. J. Remote Sens. 2010, 31, 5975–5991. [Google Scholar] [CrossRef]
  29. Li, J.; Marpu, P.R.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A. Generalized composite kernel framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  30. LeCun, Y.; Boser, B.E.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.E.; Jackel, L.D. Handwritten digit recognition with a back-propagation network. In Advances in Neural Information Processing Systems; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1990; pp. 396–404. [Google Scholar]
  31. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  32. Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the AISTATS Fourteenth International Conference on Artificial Intelligence and Statistics, Ft. Lauderdale, FL, USA, 11–13 April 2011; Volume 15, p. 275. [Google Scholar]
  33. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  34. Liu, P.; Zhang, H.; Eom, K.B. Active Deep Learning for Classification of Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 712–724. [Google Scholar] [CrossRef] [Green Version]
  35. Ran, L.; Zhang, Y.; Wei, W.; Yang, T. Bands sensitive convolutional network for hyperspectral image classification. In Proceedings of the International Conference on Internet Multimedia Computing and Service, Xi’an, China, 19–21 August 2016; pp. 268–272. [Google Scholar]
  36. Lorenzo, P.R.; Tulczyjew, L.; Marcinkiewicz, M.; Nalepa, J. Band Selection from Hyperspectral Images Using Attention-based Convolutional Neural Networks. arXiv 2018, arXiv:1811.02667. [Google Scholar]
  37. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Van Gool, L. Temporal segment networks: Towards good practices for deep action recognition. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 20–36. [Google Scholar]
  38. Diba, A.; Sharma, V.; Van Gool, L. Deep temporal linear encoding networks. In Proceedings of the 2017 CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 1541–1550. [Google Scholar]
  39. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 221–231. [Google Scholar] [CrossRef] [Green Version]
  40. Sun, L.; Jia, K.; Yeung, D.Y.; Shi, B.E. Human action recognition using factorized spatio-temporal convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4597–4605. [Google Scholar]
  41. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  42. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic differentiation in PyTorch. In Proceedings of the Neural Information Processing System-Workshop, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  43. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
Figure 1. The Synergistic Convolutional Neural Network (SyCNN) framework which integrates 2D CNN, 3D CNN and the data interaction module for hyper-spectral image classification.
Figure 1. The Synergistic Convolutional Neural Network (SyCNN) framework which integrates 2D CNN, 3D CNN and the data interaction module for hyper-spectral image classification.
Remotesensing 12 02033 g001
Figure 2. Illustration of a 3D convolution operation. The convolution kernels slide along both the spatial and spectral dimensions of the input 3D images and generate the 3D spectral–spatial features.
Figure 2. Illustration of a 3D convolution operation. The convolution kernels slide along both the spatial and spectral dimensions of the input 3D images and generate the 3D spectral–spatial features.
Remotesensing 12 02033 g002
Figure 3. SyCNN with data interaction module that consists of the local cross-domain connection and cross-domain transfer. 2D spectral–spatial fusion is achieved by transferring the 3D inputs to 2D outputs and concatenating these outputs with the 2D inputs. On the other hand, the 3D spectral–spatial fusion is obtained by transferring the 2D inputs to 3D outputs and concatenating these 3D outputs with the 3D inputs.
Figure 3. SyCNN with data interaction module that consists of the local cross-domain connection and cross-domain transfer. 2D spectral–spatial fusion is achieved by transferring the 3D inputs to 2D outputs and concatenating these outputs with the 2D inputs. On the other hand, the 3D spectral–spatial fusion is obtained by transferring the 2D inputs to 3D outputs and concatenating these 3D outputs with the 3D inputs.
Remotesensing 12 02033 g003
Figure 4. Illustration of the proposed models: (a) The deep simple SyCNN model, (b) the deep SyCNN model, (c) the deep SyCNN with attention module. Yellow blocks and blue blocks refer to 3D convolution and 2D convolution. Green blocks refer to data interaction module.
Figure 4. Illustration of the proposed models: (a) The deep simple SyCNN model, (b) the deep SyCNN model, (c) the deep SyCNN with attention module. Yellow blocks and blue blocks refer to 3D convolution and 2D convolution. Green blocks refer to data interaction module.
Remotesensing 12 02033 g004
Figure 5. Indian Pines dataset. (a) Three-band false-color composite. (b) Ground-truth map.
Figure 5. Indian Pines dataset. (a) Three-band false-color composite. (b) Ground-truth map.
Remotesensing 12 02033 g005
Figure 6. Botswana Scene dataset. (a) Three-band false-color composite. (b) Ground-truth map.
Figure 6. Botswana Scene dataset. (a) Three-band false-color composite. (b) Ground-truth map.
Remotesensing 12 02033 g006
Figure 7. Kennedy Space Center dataset. (a) Three-band false-color composite. (b) Ground-truth map.
Figure 7. Kennedy Space Center dataset. (a) Three-band false-color composite. (b) Ground-truth map.
Remotesensing 12 02033 g007
Figure 8. Visualization of the experimental results based on Indian Pines: (a) Indian Pines, (b) Ground truth, (c) SVM-Grid, (d) NN, (e) Sharma, (f) Liu, (g) Hamida, (h) Lee, (i) Li, (j) Chen, (k) He, (l) SyCNN-S, (m) SyCNN-D, (n) SyCNN-ATT. It is observed that the outputs produced by our proposed models are quite close to the Ground truth.
Figure 8. Visualization of the experimental results based on Indian Pines: (a) Indian Pines, (b) Ground truth, (c) SVM-Grid, (d) NN, (e) Sharma, (f) Liu, (g) Hamida, (h) Lee, (i) Li, (j) Chen, (k) He, (l) SyCNN-S, (m) SyCNN-D, (n) SyCNN-ATT. It is observed that the outputs produced by our proposed models are quite close to the Ground truth.
Remotesensing 12 02033 g008
Figure 9. Visualization of the experimental results based on Botswana: (a) Indian Pines, (b) Ground truth, (c) SVM-Grid, (d) NN, (e) Sharma, (f) Liu, (g) Hamida, (h) Lee, (i) Li, (j) Chen, (k) He, (l) SyCNN-S, (m) SyCNN-D, (n) SyCNN-ATT. It is observed that the outputs produced by our proposed models are quite close to the Ground truth.
Figure 9. Visualization of the experimental results based on Botswana: (a) Indian Pines, (b) Ground truth, (c) SVM-Grid, (d) NN, (e) Sharma, (f) Liu, (g) Hamida, (h) Lee, (i) Li, (j) Chen, (k) He, (l) SyCNN-S, (m) SyCNN-D, (n) SyCNN-ATT. It is observed that the outputs produced by our proposed models are quite close to the Ground truth.
Remotesensing 12 02033 g009
Figure 10. Visualization of the experimental results based on KSC: (a) Indian Pines, (b) Ground truth, (c) SVM-Grid, (d) NN, (e) Sharma, (f) Liu, (g) Hamida, (h) Lee, (i) Li, (j) Chen, (k) He, (l) SyCNN-S, (m) SyCNN-D, (n) SyCNN-ATT. It is observed that the outputs produced by our proposed models are quite close to the Ground truth.
Figure 10. Visualization of the experimental results based on KSC: (a) Indian Pines, (b) Ground truth, (c) SVM-Grid, (d) NN, (e) Sharma, (f) Liu, (g) Hamida, (h) Lee, (i) Li, (j) Chen, (k) He, (l) SyCNN-S, (m) SyCNN-D, (n) SyCNN-ATT. It is observed that the outputs produced by our proposed models are quite close to the Ground truth.
Remotesensing 12 02033 g010
Figure 11. Visualization of the influence of training samples proportion for different methods based on the three datasets: (a) Indian Pines, (b) Botswana, (c) KSC. It can be observed that the results of our proposed models are very stable and better than the comparison methods.
Figure 11. Visualization of the influence of training samples proportion for different methods based on the three datasets: (a) Indian Pines, (b) Botswana, (c) KSC. It can be observed that the results of our proposed models are very stable and better than the comparison methods.
Remotesensing 12 02033 g011
Table 1. Sample size for Indian Pines Scene.
Table 1. Sample size for Indian Pines Scene.
NO.Class NameTraining SamplesTesting Samples
1Alfalfa3214
2Corn-notill1003424
3Corn-mintill585245
4Corn16869
5Grass-pasture340143
6Grass-trees512216
7Grass-pasture-mowed208
8Hay-windrowed335143
9Oats146
10Soybean-notill683289
11Soybean-mintill1721733
12Soybean-clean417174
13Wheat14461
14Woods888374
15Buildings-Grass-Trees-Drives272113
16Stone-Steel-Towers6528
Total72003040
Table 2. Sample size for Botswana Scene.
Table 2. Sample size for Botswana Scene.
NO.Class NameTraining SamplesTesting Samples
1Water18981
2Hippo grass7130
3Floodplain grasses 117675
4Floodplain grasses 215164
5Reeds18881
6Riparian18881
7Fire scar18376
8Island interior14360
9Acacia woodlands22094
10Acacia shrunblands17474
11Acacia grasslands21491
12Short mopane12754
13Mixed mopane18979
14Exposed soils6728
Total2280968
Table 3. Samples size for Kennedy Space Center scene.
Table 3. Samples size for Kennedy Space Center scene.
NO.Class NameTraining SamplesTesting Samples
1Scrub533228
2Willow swamp17073
3Cabbage palm hammock17977
4Cabbage palm/oak hammock17676
5Slash pine11348
6Oak/broadleaf hammock16069
7Hardwood swamp7431
8Graminoid marsh302129
9Spartina marsh364156
10Cattail marsh284120
11Salt marsh294125
12Mud flats352151
13Water649278
Total36501561
Table 4. Evaluation results on the Indian Pines Scene dataset.
Table 4. Evaluation results on the Indian Pines Scene dataset.
ClassMethods
#SVM-GridNNSharmaLiuHamidaLeeLiChenHeSyCNN-SSyCNN-DSyCNN-ATT
188.0092.3010066.7010063.6010096.30100100100100
285.5084.2099.4088.5084.9082.4096.6095.9089.0010099.6099.20
380.8080.3093.6086.1079.7080.4090.6091.1086.0091.3091.7096.10
479.2071.9010083.6093.6082.8095.8097.2095.0097.1010099.30
595.2093.7094.6091.6093.8093.7094.5092.6095.7096.1095.3097.20
696.7096.1010098.4099.1097.5099.8099.1010010099.50100
780.0093.3010040.0076.9082.4010093.30100100100100
899.0099.3010096.3099.0097.3099.7099.70100100100100
966.70100100090.9028.3010010050.00100100100
1080.6082.4097.0089.3076.2084.2096.1090.5091.9097.7097.0098.30
1184.9087.9097.5091.5087.4089.8096.2094.7092.6098.1098.0098.70
1289.3083.1097.2089.7087.7079.7096.2096.9093.8098.3098.9098.60
1310010010010010096.10100100100100100100
1495.8093.5099.7098.4098.9094.2099.7098.9099.10100100100
1576.7069.1084.2077.2076.3072.0082.2083.7085.8086.8091.1087.40
1610098.2096.4010098.2010010098.2096.4010098.20100
OA87.9387.5795.6489.5686.9987.8794.2293.2091.8795.9096.1397.31
F187.4089.0797.4881.0890.1683.4296.7195.5192.2197.8498.0898.43
K86.2085.8095.1088.1085.2086.1093.4092.3090.8095.3095.6096.90
Table 5. Comparative evaluation based on the Botswana Scene dataset.
Table 5. Comparative evaluation based on the Botswana Scene dataset.
ClassMethods
#SVM-GridNNSharmaLiuHamidaLeeLiChenHeSyCNN-SSyCNN-DSyCNN-ATT
199.4010099.40100100100100100100100100100
298.4096.8010096.6092.9098.4010010010098.40100100
398.6092.9010010096.6091.70100100100100100100
497.7090.3010010092.7069.3099.2091.4097.70100100100
592.4081.5098.8095.2091.0083.7094.9097.7093.4098.80100100
687.4074.4099.4092.1072.6076.1091.0092.9091.30100100100
798.7099.4010010010099.4010089.30100100100100
810097.5010010097.6076.9010098.10100100100100
990.2085.1010096.1064.7061.7010093.8097.40100100100
1090.3083.9099.3097.9097.4081.7010094.40100100100100
1194.4093.1099.5097.9095.5095.6010091.50100100100100
1291.4095.4098.1099.1097.1090.6096.2096.2099.1095.2097.1099.10
1393.9087.7010010099.4010010099.40100100100100
1410096.7010010090.6088.5010092.60100100100100
OA94.6690.2599.4897.4991.3885.9497.9495.3898.2599.3899.6999.79
F195.2091.0599.6199.6892.0186.3498.3695.5298.4999.4699.7999.93
K94.2089.4099.4097.8090.7084.8097.8095.0098.1099.3099.7099.80
Table 6. Classification results of the Salinas scene.
Table 6. Classification results of the Salinas scene.
ClassMethods
#SVM-GridNNSharmaLiuHamidaLeeLiChenHeSyCNN-SSyCNN-DSyCNN-ATT
157.8087.6089.4081.5069.4094.5087.4054.1091.4094.4094.7097.80
272.5071.3094.5080.0061.8068.8091.2088.6093.4098.6099.3099.30
3062.4092.2068.307.5072.2069.2091.4095.1096.70100100
4024.1076.6060.9040.707.3070.3091.5075.2094.6098.70100
5053.7088.9061.7031.6067.4082.2010086.3095.9097.9096.80
6030.8067.3042.202.9064.2076.3098.5087.8098.6099.30100
7054.2092.5085.305.6035.0090.0096.8098.4098.50100100
841.6061.4094.3072.7070.2080.0081.4082.8097.2098.0096.4097.70
973.4083.1099.4084.5085.3095.7091.1093.3099.40100100100
1083.5088.1099.6096.6077.1086.3096.7010099.20100100100
1196.0098.4010010099.6098.8099.60100100100100100
1274.8085.3099.7095.0087.1085.0093.5010099.0099.70100100
1396.7099.5010099.6099.8097.30100100100100100100
OA67.3979.7993.3584.2776.2284.1488.6286.8394.6997.4497.7698.92
F145.8967.3791.8879.1062.8273.2786.8492.0894.0398.4698.9599.35
K62.8077.4092.6082.5073.1082.3087.4085.5094.1097.2097.5098.80
Table 7. Comparison of parameters and times of all methods.
Table 7. Comparison of parameters and times of all methods.
MethodsNumber of ParametersExecution Time (s)OA
SVM-Grid2.98 MB-67.39
NN65.52 MB20.5779.79
Sharma5.9 KB25.8693.35
Liu4.08 MB8282.7884.27
Hamida472 KB46.8876.22
Lee1.22 MB128.2984.14
Li2.38 MB37.9088.62
Chen3.27 MB285.6386.83
He8.18 KB98.6794.69
SyCNN-S21.14 MB117.0397.44
SyCNN-D29.46 MB146.5297.76
SyCNN-ATT34.98 MB154.1598.92

Share and Cite

MDPI and ACS Style

Yang, X.; Zhang, X.; Ye, Y.; Lau, R.Y.K.; Lu, S.; Li, X.; Huang, X. Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sens. 2020, 12, 2033. https://doi.org/10.3390/rs12122033

AMA Style

Yang X, Zhang X, Ye Y, Lau RYK, Lu S, Li X, Huang X. Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sensing. 2020; 12(12):2033. https://doi.org/10.3390/rs12122033

Chicago/Turabian Style

Yang, Xiaofei, Xiaofeng Zhang, Yunming Ye, Raymond Y. K. Lau, Shijian Lu, Xutao Li, and Xiaohui Huang. 2020. "Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification" Remote Sensing 12, no. 12: 2033. https://doi.org/10.3390/rs12122033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop