Next Article in Journal
AT2ES: Simultaneous Atmospheric Transmittance-Temperature-Emissivity Separation Using Online Upper Midwave Infrared Hyperspectral Images
Previous Article in Journal
Deep Learning Classification of Cheatgrass Invasion in the Western United States Using Biophysical and Remote Sensing Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiple Spectral Resolution 3D Convolutional Neural Network for Hyperspectral Image Classification

College of Computer Science, South-Central University for Nationalities, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(7), 1248; https://doi.org/10.3390/rs13071248
Submission received: 25 February 2021 / Revised: 21 March 2021 / Accepted: 22 March 2021 / Published: 25 March 2021

Abstract

:
In recent years, benefiting from the rapid development of deep learning technology in the field of computer vision, the study of hyperspectral image (HSI) classification has also made great progress. However, compared with ordinary RGB images, HSIs are more like 3D cubes; therefore, it is necessary and beneficial to explore classification methods suitable for the very special data structure of HSIs. In this paper, we propose Multiple Spectral Resolution 3D Convolutional Neural Network (MSR-3DCNN) for HSI classification tasks. In MSR-3DCNN, we expand the idea of multi-scale feature fusion and dilated convolution from the spatial dimension to the spectral dimension, and combine 3D convolution and residual connection; therefore, it can better adapt to the 3D cubic form of hyperspectral data and make efficient use of spectral information in different bands. Experimental results on four benchmark datasets show the effectiveness of the proposed approach and its superiority as compared with some state-of-the-art (SOTA) HSI classification methods.

Graphical Abstract

1. Introduction

Hyperspectral image (HSI) is a hot topic in remote sensing data analysis [1]. By combining rich spectral information with spatial information, HSI can support very accurate land use and land cover classifications for the ground surface, which is of great significance in agriculture [2], forestry [3], and other fields [4]. Land use and land cover classification tasks based on hyperspectral images are usually denoted as HSI classification. Rather than an ordinary image classification task, HSI classification is more similar to the semantic segmentation task in computer vision [5]. With the proposed model frameworks such as convolutional neural network (CNN) [6], the fully convolutional networks (FCN) [5], and residual network (ResNet) [7], deep learning technology has made great achievements in the field of RGB image processing. Deep learning technology has also been introduced to solve HSI classification tasks [4], such as the widely cited research work of Chen et al. [8] and Zhao et al. [9]. In recent years, more and more deep learning-based HSI classification approaches with different designs have been proposed, and huge improvements in classification accuracy have been reported [10,11,12,13,14].
However, as compared to the RGB images in a common semantic segmentation task, the structures of HSIs are much more complicated. For example, in the past, the Hyperion [15] sensor could provide 220 spectral bands, covering a spectrum range from 0.4 to 2.50 µm. At present, the more representative PRISMA [16] sensor can provide 240 spectral bands, covering a spectrum range from 0.4 to 2.45 µm. The newer Environmental Mapping and Analysis Program (EnMAP) [17] sensor provides 228 spectral bands, covering a spectrum range from 0.42 to 2.40 µm. It is clear that the number of channels in HSIs is relatively large relative to their spatial sizes, which makes HSI more like a three-dimensional cube of data. Therefore, some semantic segmentation methods designed for ordinary RGB images are no longer applicable to HSIs. Thus, the key to solving the classification problem of HSIs is to make efficient use of the rich spectral information in HSIs and to process 3D cubic data more efficiently. Inspired by the research work in [18], we use 3D convolution as the basis of the network to make the classification more suitable to the 3D structure of HSIs. As compared with the 2D convolutional kernel, the 3D convolutional kernel has one more dimension, which can extract spectral information more effectively.
Recently, we have noticed that models using multi-scale spatial features corresponding to different receptive fields can improve the semantic segmentation accuracy of ordinary RGB images. For example, PSPnet fuses features under four different scales and separates the feature map into different subregions forming pooled representation for different locations and size. Different levels of features are then concatenated as the final pyramid pooling global feature [19]. Similarly, the Inception Module uses different size convolution kernels to extract information at different levels, and the output feature maps are connected [20]. In our study, in order to enlarge the widths of the spectrums, we propose Spectral Dilated Convolutions (SDC) inspired by a Dilated Residual Network (DRN) [21]. At the same time, we extend this multi-scale feature fusion to the spectral dimension. Based on SDC and spectral multi-scale feature fusion, we propose a Multiple Spectral Resolution (MSR) module to extract the rich spectral information of HSIs. The module consists of multiple different 3D convolution branches corresponding to multiple different spectrum widths and can extract multi-scale spectral features of HSIs, respectively.
In the proposed network, we cascade two MSR modules as the essential component of the network. At the same time, in order to avoid the problem of gradient disappearance and gradient explosion, we refer to the idea of the residual network (ResNet) and add jump connection to the proposed network [7]. In general, we combine the above-mentioned MSR module, SDC, and 3D convolutional and residual connection to propose a Multiple Spectral Resolution 3D Convolutional Neural Network (MSR-3DCNN) for HSI classification. The main contributions of our work are summarized as follows:
(1)
For the cubic structure of HSI, we propose a fully 3D CNN HSI classification model.
(2)
Referring to the spatial Dilated Residual Network (DRN), we propose Spectral Dilated Convolutions (SDC). Referring to spectral multi-scale feature fusion, we propose spectral multi-scale feature fusion.
(3)
Based on SDC and spectral multi-scale feature fusion, we propose a Multiple Spectral Resolution (MSR) module. Combining the MSR module, SDC, and 3D convolution and residual connection, we propose Multiple Spectral Resolution 3D Convolutional Neural Network (MSR-3DCNN) for HSI classification.
(4)
Experimental results on four benchmark datasets show the effectiveness of MSR-3DCNN and its superiority as compared with some SOTA HSI classification methods.
The rest of the paper is organized as follows. Section 2 reviews the related work. Section 3 describes the proposed classification method. Section 4 introduces experimental settings and analyses the experimental results. Finally, conclusions and suggestions are provided in Section 5.

2. Related Work

In this section, we will introduce some basic knowledge and related work, including 3D convolutional neural network, multi-scale feature fusion, residual connection, and dilated convolution.

2.1. 3D Convolutional Neural Network

In a 2D CNN, convolution is applied to the two-dimensional feature graph, and the feature is calculated only from the spatial dimension. Formally, the value at position ( x , y ) in the jth feature map in the ith layer, denoted as v i j x y , is given by
v i j x y = Φ b i j + m p = 0 P i 1 q = 0 Q i 1 w i j m p q × v ( i 1 ) m ( x + p ) ( y + q )
where Φ is the activation function, b i j is the bias parameter for this feature map, m indexes over the set of feature maps in the ( i 1 ) th layer connected to the current feature map, w i j m p q is the weight parameter at the position ( p , q ) of the kernel connected to the mth feature map, and P i and Q i are the height and width of the kernel, respectively.
In a 3D CNN, 3D convolution kernels [18] are applied to three-dimensional image cubes like HSIs, and the image features are calculated from both the spatial dimension and the spectral dimension simultaneously. Formally, the value at position ( x , y , z ) on the jth feature map in the ith layer is given by
v i j x y z = Φ b i j + m p = 0 P i 1 q = 0 Q i 1 r = 0 R i 1 w i j m p q r × v ( i 1 ) m ( x + p ) ( y + q ) ( z + r )
where R i is the depth of the 3D convolutional kernel, w i j m p q r is the weight parameter at the position ( p , q , r ) of the kernel connected to the mth feature map in the previous layer, and the other parameters are the same as in the 2D convolutions. The parameters of CNN, such as the bias b and the kernel weight w, are usually trained using supervised approaches with the help of a gradient descent optimization technique [6,22]. 3D convolution is the base of some very powerful HSI classification models, such as hybrid spectral CNN (HybridSN) proposed by Roy et al. in [11] and Double-Branch Multi-Attention Mechanism Network (DBMA) proposed by Ma et al. in [12].

2.2. Multi-Scale Feature Fusion

Researchers have found that the fusion of multi-scale semantic features can improve the precision of semantic segmentation of common optical images. For example, in order to make use of spatial features corresponding to multiple scales, Zhao et al. proposed pyramid pooling module [19]. As illustrated in Figure 1, the pyramid pooling module integrates four levels of different pyramid scales. The first level gets a single value feature map through global pooling, and the next three levels get a feature map of different scales through pooling of different scales. A 1 × 1 convolution is used after each feature map to reduce the channel of these feature maps. Then, bilinear interpolation is used to upsample these feature maps to the original spatial resolution of the input. Finally, all the feature maps are concatenated with the input of the module. The size of the final output is decided by the size of the input.
The Inception module [20] illustrated in Figure 2 can be considered as another implementation of this multi-scale spatial feature fusion idea. Convolution kernels of different sizes are used to extract spatial features corresponding to different scales, and the aggregation is also based on feature map concatenation.

2.3. Residual Connections

The increase in model depth usually can improve the performance of a neural network model [23]. However, the increase in the depth of the model often causes the problem of gradient disappearance or gradient explosion. He et al. suggest that the residual connection solves this problem well [7] and greatly increases the depth of the networks. Therefore, at present, most convolutional neural networks use residual network (ResNet) as backbones.
As shown in Figure 3, the “shortcut connection” in a residual block is basically a direct connection between the input and the output of the block, while the “hidden block” usually contains several convolution layers along with batch normalization and activation.
ResNet allows input information to pass directly and quickly to subsequent layers. The shortcut connections can be seen as identity mapping. In ResNet, the output of the l-th block be computed as
X l = F x l 1 + h x l 1
In general, through the residual connection, the original function can be divided into F ( x ) and h ( x ) . F ( x ) is almost the same as the original network. h ( x ) can skip connections through shortcut connection.

2.4. Dilated Convolution

It is important to enlarge the receptive field of a segmentation model when we have to perform the classification and localization tasks simultaneously [24]. A direct way to obtain a large receptive field is to use large convolutional kernels; however, this means more model parameters and longer training time. An alternative is to use dilated convolutions [21], which can enlarge the receptive field without increasing the size of the model.
As shown in Figure 4, the red dot represents the convolution kernel and the blue base represents the receptive field. As shown in Figure 4a, it is a 3 × 3 dilated convolution with the dilation rate set to 1, which is the same as a regular convolution. As shown in Figure 4b, it is a 3 × 3 dilated convolution with the dilation rate set to 2. As shown in the red dots in the figure, the actual size of the convolution kernel is still 3 × 3 ; however, the receptive field of the kernel is expanded to 7 × 7 . As shown in Figure 4c, when the dilation rate is set to 4, a 3 × 3 dilated convolution kernel can hold a 15 × 15 receptive field. Therefore, dilated convolution is helpful to expand the receptive field without increasing the number of parameters and original structure of the model.

3. The Proposed Approach

In this section, we illustrate in detail the proposed approach for HSI classification. First, we introduce Spectral Dilated Convolution (SDC). On the basis of SDC, we explain the Multiple Spectral Resolution (MSR) module which is an essential component of the proposed approach. Finally, we give a general illustration of Multiple Spectral Resolution 3D Convolutional Neural Network (MSR-3DCNN).

3.1. Spectral Dilated Convolutions

As described in Section 2.4, dilated convolution can expand the receptive field without changing the number of parameters and the original structure of the model. However, traditional 2D dilated convolutions can only be applied to enlarge the receptive field along spatial dimensions. Inspired by this, we propose the 3D spectral dilated convolution (SDC) to expand the receptive field of convolution kernels along the spectral dimension.
As shown in Figure 5, the black cuboids represent the convolution kernel and the red line cuboids cover the receptive field. As shown in Figure 5a, it is a 3 × 3 × 3 spectral dilated convolution (SDC) with the spectral dilation rate (SDR) set to 1, which is the same as a regular 3D convolution kernel. As shown in Figure 5b, it is a 3 × 3 × 3 SDC with the SDR set to 2 and the receptive field of the SDC kernel reaches 3 × 3 × 7 . As shown in Figure 5c, it is a 3 × 3 × 3 SDC with the SDR set to 3. Without increase the size of the kernel, the receptive field is further enlarged to 3 × 3 × 11 . The receptive field of SDC is decided by the spectral dilation rate as
R f = 2 × S D R 1
K e q u = K r e a l + K r e a l 1 × S D R 1
R F i + 1 = R F i + K e q u 1 × S P i
S P i = i = 1 i S t r i d e i
where R f represents the receptive field of a single convolution kernel in spectral depth. S D R represents the spectral dilation rate (SDR). K e q u represents the number of the equivalent convolution kernels at spectral depth. K r e a l represents the number of the real convolution kernels at spectral depth. R F i + 1 represents the receptive field of the current layer in spectral depth. R F i represents the receptive field of the previous layer in spectral depth. S P i represents the product of all the previous layers stride, and the current layer is not included. S t r i d e i represents the ith layer stride.

3.2. Multiple Spectral Resolution Module

Inspired by multi-scale feature fusion, we propose to construct a Multiple Spectral Resolution (MSR) module to make better usage of the rich spectral information in HSIs. As shown in Figure 6, the Multiple Spectral Resolution (MSR) module consists of multiple different convolution branches, and each branch consists of multiple SDCs. SDCs in the same branch are exactly the same, and all the SDCs share the same height (H) and width (W). In all the branches, the SDR values S1, S2 ... increase successively, so the SDC depths (D) of different branches are different.
We keep the size of the output of each SDC the same as the input. The input passes through different branches, and then the output results of the branches are added. Finally, the sum is taken as the output of the MSR model. This design makes the branches have the same receptive field in the spatial dimension and different receptive field in spectral depth. If the parameters are set properly, the receptive fields of different branches in spectral depth can differ by many times. Therefore, multiple spectral resolution features can be extracted and fused by the proposed MSR model.

3.3. Multiple Spectral Resolution 3D Convolutional Neural Network

The MSR modules based on SDC in our proposed network can extract and fuse features corresponding to multiple spectral resolutions, and therefore the rich spectral information can be more effectively used. We also employ residual blocks to deepen the proposed model while avoid the vanishing gradient problem. Both the MSR module and the residual block are composed of 3D convolutions; therefore, we named the proposed model as Multiple Spectral Resolution 3D Convolutional Neural Network (MSR-3DCNN).
As shown in Figure 7, MSR-3DCNN consists of a set of individual SDCs, two MSR modules cascaded in the form of residual connection, two 3D Residual modules, and a set of fully connected layers. The first two layers of the model are composed of SDCs, with the same size of 1 × 1 × 3 and the same dilation rate of 3. There are 32 SDC kernels in the first layer and 8 SDC kernels in the second one. Following the individual SDC layers are two identical MSR modules. Each MSR module consists of three types of SDCs, which share the same size as 3 × 3 × 7 , while their dilation rates are set to 1, 7, and 11, respectively. According to Equations (5)–(7), we calculate the receptive fields of the three branches in the spectral depth as 14, 129, and 216, respectively. The two 3D residual convolution blocks are quite similar. They share the same structure, with only some differences in the size of the 3D convolution kernels used in the main path of each block. At the end of both modules is a 2 × 2 × 2 3D pooling. Finally, there are two fully connected layers which contain 256 and 128 neurons, respectively. In order to avoid overfitting, we use Dropout [25] in the fully connected layers and add batch normalization [26] after them.
The model takes H × W × 100 image patches as inputs. A preprocessing based on principal component analysis (PCA) [27,28] is applied beforehand to reduce the original spectral dimensions of HSIs to 100. No padding is applied along the spectral dimension in the first two individual SDC layers; therefore, we get feature cubes with the size of H × W × 88 ; 8 . Then, in both the MSR modules, the size of the feature cubes is kept constant; therefore, skip connections can be used to directly connect the outputs and the inputs of these two modules. The data flow can be expressed as follows:
M s e c o n d = f b r a 1 M f i r s t + f b r a 2 M f i r s t + f b r a 3 M f i r s t + M f i r s t = f M S R M f i r s t + M f i r s t
M t h i r d = f M S R M s e c o n d + M s e c o n d + M f i r s t
where M f i r s t represents the input feature map of the first MSR. M s e c o n d represents the output feature map of the first MSR. M t h i r d represents the output feature map of the second MSR. f b r a 1 , f b r a 2 , and f b r a 3 represent the arithmetic processing of three branch in MSR. f M S R represents the arithmetic processing of MSR. Through the second part of the network, two MSR modules and MSR-3DCNN achieve further spectral feature extraction and multiple spectral resolution fusion. Although same paddings are used in all the 3D convolution kernels in the two residual convolution blocks of the model, the size of the feature cubes are reduced to H / 4 × W / 4 × 22 by the 3D poolings at the end of each residual block. At last, these feature cubes are flattened and further processed by the fully connected layers.
The main idea of the proposed MSR-3DCNN is based on two novel designs: spectral dilated convolution and multiple spectral resolution fusion. Spectral dilated convolution is an extension of the classical dilated convolution from the spatial dimension to the spectral dimension, and the concept of multiple spectral resolution fusion is inspired by the spatial dimension technique, multi-scale fusion. These two designs give the proposed model three advantages: First of all, operations based on SDC are quite efficient to extract spectral features in HSIs. Then, the MSR modules in the proposed model can effectively make use of the rich spectral information in HSIs. As more attention has been paid to the processing of spectral information, MSR-3DCNN shows less reliance on spatial information and therefore is more robust to images with different spatial structures.

4. Experiments and Discussion

4.1. Datasets Description

In the experiments, four widely used HSI datasets are used to test the proposed approach: the Indian Pines (IP) dataset, the Salinas scene (SA) dataset, the Pavia University scene (PU) dataset, and the Botswana (BO) dataset. These datasets are quite complex some as there are many different land cover types in these scenes, such as corn, soybean, woods, fallow, stubble, celery, gravel, trees, occasional swamps, and many other types. Some of these land cover types are quite similar and difficult to distinguish. Furthermore, these data sets are obviously different from each other. The IP, SA, and BO data sets represent three different agriculture and forest areas, and the PU data set mainly consists of urban objects. These datasets are available and have been shared among researchers for years (These datasets can be found at http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes.). Table 1 shows the summary of the characteristic of the datasets.
The Indian Pines (IP) dataset [29] was gathered by the airborne sensor, Airborne Visible Infrared Imaging Spectrometer (AVIRIS), from an agricultural and forest region, in the northwestern of India. The HSI consists of 145 × 145 pixels with a ground sampling distance (GSD) of 20 m per pixel (m/pixel) and 224 spectral bands in the wavelength range 0.4 to 2.5 µm. The 24 water-absorbing spectral bands have been discarded, considering the remaining 200 bands for the experiments. The ground truth available is 10,249 pixels from a total of 21,025 and designated into 16 classes, e.g., corn, soybean, woods, etc.
The Salinas scene (SA) dataset [29] was also gathered by the AVIRIS sensor from an agricultural field of the Salinas Valley, California, USA. The HSI consists of 512 × 217 pixels with a GSD of 3.7 m/pixel and 224 spectral bands in the wavelength range 0.4 to 2.5 µm. The 20 water-absorbing spectral bands have been discarded, considering the remaining 204 bands for the experiments. The ground truth available is 54,129 pixels from a total of 111,104 and designated into 16 classes, e.g., fallow, stubble, celery, etc.
The Pavia University scene (PU) dataset [30] was gathered by the Reflective Optics System Imaging Spectrometer (ROSIS) sensor on airborne vehicle from the urban area of Pavia University, in northern Italy. The HSI consists of 610 × 340 pixels with a GSD of 1.3 m/pixel and 103 spectral bands in the wavelength range 0.43 to 0.86 µm. The ground truth available is 42,776 pixels from a total of 207,400 and designated into 9 classes, e.g., asphalt, gravel, trees, etc.
The Botswana (BO) dataset [31] was gathered by the Hyperion sensor on the Earth Observing One (EO-1) satellite from the vegetation and swamps, Okavango Delta, Botswana. The HSI consists of 1476 × 256 pixels with a GSD of 30 m/pixel and 242 spectral bands in the wavelength range 0.4 to 2.5 µm. Uncalibrated and noisy bands that cover water absorption features were removed, considering the remaining 145 bands for the experiments. The ground truth available is 3248 pixels from a total of 377,856 and designated into 14 classes, e.g., seasonal swamps, occasional swamps, drier woodlands located, etc.

4.2. Experiment Settings

Figure 8 illustrates the whole framework of our Experiment. First, preprocessings, including principal component analysis (PCA) and cropping, are performed on the HSI dataset. PCA changes the spectral band from the original to 100. After the PCA process, N × N image patches are cropped from the dimension reduced HSI, and the ground truth label corresponding to the center pixel of each patch is assigned to the patch. Therefore, a set of labeled patches are obtained. Then, the labeled patches are randomly divided into two groups: the training set and test set. The training set is used to train network parameters and the test set is used to evaluate the effectiveness of the classification method.
The proposed architecture is implemented using TensorFlow [32]. The hardware device is a server equipped with a NVIDIA GeForce RTX 3060 GPU and 32 GB of RAM. The source code can be found at github (https://github.com/shouhengx/MSR-3DCNN). Appendix A gives some brief descriptions about the environment in which the code runs and how it runs. During the training stage, the learning rate is initialized as 0.001 and the Adam algorithm is adopted to optimize the learning rate after each iteration. During the evaluation stage, three metrics, i.e., overall accuracy (OA) [33], average accuracy (AA) [34], and Kappa coefficient [34], are used to quantitatively evaluate the overall performance of the classification. OA refers to the ratio of the number of correct classifications to the total number of pixels to be classified. AA refers to the average accuracy of all classes. Kappa coefficients are used for consistency testing and can also be used to measure classification accuracy. In detail, the F1 score [35] is used to quantitatively evaluate the classification of each class. The F1 score combines the precision and recall metrics. At the same time, McNemar’s test [36] is conducted to analyze the statistical significance of the proposed method with the other methods. This test is based upon the standardized normal test statistic:
Z = f 12 f 21 f 12 + f 21
where f 12 indicates the number of samples classified correctly by classifier 1 and incorrectly by classifier 2. Similarly, f 21 indicates the number of samples classified correctly by classifier 2 and incorrectly by classifier 1. The difference in accuracy between classifiers 1 and 2 is said to be statistically significant if | Z | > 1.96 . The sign of Z indicates whether classifier 1 is more accurate than classifier 2 (Z > 0), and vice versa (Z < 0).

4.3. Comparisons with State-of-the-Art Methods

We preprocess the datasets through PCA and cropping as mentioned in Section 4.2. Then, using exactly the same testing protocol, we reproduce several HSI classification methods and compare them with the proposed method using 9 × 9 , 11 × 11 and 13 × 13 patch sizes, respectively. Ten percent of the labeled samples is used to train the models. In addition, we fuse the Predicted results of three independent classifiers under different patch sizes by sum rule [37]. The compared methods include contextual deep CNN [10], hybrid spectral CNN (HybridSN) [11], Double-Branch Multi-Attention mechanism network (DBMA) [12], and the most recent methods Fast 3D CNN [13] and Multi-scale Residual Network (MSRN) [14].
The overall accuracy results on the BO, IP, SA, and PU datasets are reported in Table 2. Figure 9 shows the overall accuracy changes of the BO, IP, SA, and PU datasets with 10% training samples and different patch sizes. Among the compared methods, the methods incorporating 3D convolution to extract spectral information, i.e., HybridSN, DBMA, Fast 3D CNN, MSRN, and MSR-3DCNN, have better OAs than the method using only 2D convolution, i.e., contextual deep CNN. Due to the fact that more attention is paid to the processing of spectral dimension information, MSR-3DCNN achieved good results on all experimental patch sizes, and this advantage is especially obvious in the case of small patch size. This indicates the robustness of MSR-3DCNN to the change of patch size. Furthermore, the proposed method performs well on all the four datasets, and its performance is especially good on the SA dataset. On the contrary, the spectral information processing takes a much less import role in the model of MSRN. This can be the reason for the unstable performances of MSRN on different data sets. For small data sets with uneven sample distribution (such as BO and IP), appropriate patch size is required for MSRN to perform well.
Table 3, Table 4, Table 5 and Table 6 show the performance of the different classification methods using 9 × 9 patch size and 10% of samples on different datasets in more details. Overall, MSR-3DCNN achieved good performance on different data sets using 9 × 9 patch size. On the BO and the IP datasets, the best values for OA, AA, and Kappa Coefficient are all achieved by MSR-3DCNN. On the SA datasets, the OA and Kappa Coefficient of MSR-3DCNN are also the best and the AA is the second best result. On the PU dataset, the OA and Kappa Coefficient of MSR-3DCNN are the second best. At the same time, MSR-3DCNN shows better reliability on basically all the classes, and the advantage is quite obvious for some classes with very few samples for training such as Class 10 of BO dataset and Class 9 of IP dataset.
Table 7, Table 8, Table 9 and Table 10 show the statistical significance of the classification results obtained by McNemar’s test. It can be seen that most of the absolute value of Z between MSR-3DCNN and other classification methods are greater than 1.96 [36], indicating that there is a significant difference between MSR-3DCNN and other classification methods.
Figure 10, Figure 11, Figure 12 and Figure 13 show classification results of different methods conducted on the BO, IP, SA and PU datasets using 10% training samples and 9 × 9 patch size. From the classification result maps, “salt-and-pepper” noise is quite obvious in the classification image produced by contextual deep CNN due. This can be ascribed to the lack of spectral information extraction in the model. The quality of the classifications produced by MSR-3DCNN is fairly good, as some small patch of samples can be correctly classified.
Figure 14, Figure 15, Figure 16 and Figure 17 show the confusion matrices [38,39] between ground truth on the experimental datasets and the predicted results of MSR-3DCNN using 10% training samples and 9 × 9 patch size. These confusion matrices show the classification status of each category on each dataset of MSR-3DCNN, which is consistent with the visual evaluation of the classification results. MSR-3DCNN can obtain accurate classifications for basically all the classes, including these with very limited amount of samples.

5. Conclusions

In this paper, we propose a Multiple Spectral Resolution 3D Convolutional Neural Network for HSI classification. As compared to the other researches about HSI classification models, more attention has been paid to the spectral dimension of HSIs in our model. The idea of dilated convolution and multi-scale fusion has been expanded from the spatial dimension to the spectral dimension; therefore, two novel designs—spectral dilated convolution and multiple spectral resolution fusion—have been proposed. Using these two novel designs, the proposed MSR-3DCNN can theoretically have better spectral information extraction and analysis capabilities, and therefore may be less demanding on the spatial information of HSIs.The theoretical advantage of MSR-3DCNN has been verified in our comparative experiments. As shown by the experimental results, MSR-3DCCNN is much more insensitive to the change of patch size and has better performance under small patch size. Note that MSR-3DCNN showed advantages in various experimental datasets, showing a strong applicability, which is of great significance for the application of a new generation of hyperspectral sensors. In the future, we plan to further improve this approach by reducing the calculations as well as applying it to the new generation of hyperspectral systems.

Author Contributions

Conceptualization, H.X., W.Y. and L.C.; methodology, H.X.; software, H.X.; writing—original draft preparation, H.X.; writing—review and editing, W.Y. and B.L.; supervision, W.Y.; funding acquisition, W.Y., L.C. and B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was jointly funded by the Natural Science Foundation of China under Grants 61976226, the State Scholarship Fund of China under Grant 201908420071, the Fundamental Research Funds for the Central Universities, South-Central University for Nationalities (CZT20021), and the Innovation and Entrepreneurship Training Program Funded by South-Central University for Nationalities.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes.

Acknowledgments

The authors would also like to thank the peer researchers who made their source codes available to the whole community and all the open sources of the benchmark HSI data sets.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HSIhyperspectral image
MSR-3DCNNMultiple Spectral Resolution 3D Convolutional Neural Network
MSRMultiple Spectral Resolution
SDCSpectral Dilated Convolution
SDRSpectral Dilation Rate
OAoverall accuracy
AAaverage accuracy
PCAprincipal component analysis
CNNconvolutional neural network
GSDground sampling distance
IPIndian Pines dataset
SASalinas scene dataset
PUPavia University scene dataset
BOBotswana dataset
SOTAState-of-the-art

Appendix A. Experimental Code Running Details

The source code for our proposed method can be found at https://github.com/shouhengx/MSR-3DCNN. This code needs to run in Python 3.7+ and Jupyter Notebook. This code also requires some Python Package: tensorflow-gpu 2.1.0+, scikit-learn 0.21.3+, numpy 1.19.5+, matplotlib 3.1.1+, scipy 1.3.1+, spectral 0.22.1+. The datasets for the experiment can be found at http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes. The datasets needs to be placed in a folder named data with the same path as the code. When you are ready, run the code using Jupyter Notebook.

References

  1. Signoroni, A.; Savardi, M.; Baronio, A.; Benini, S. Deep learning meets hyperspectral image analysis: A multidisciplinary review. J. Imaging 2019, 5, 52. [Google Scholar] [CrossRef] [Green Version]
  2. Teke, M.; Deveci, H.S.; Haliloğlu, O.; Gürbüz, S.Z.; Sakarya, U. A short survey of hyperspectral remote sensing applications in agriculture. In Proceedings of the 2013 6th International Conference on Recent Advances in Space Technologies (RAST), Istanbul, Turkey, 12–14 June 2013; pp. 171–176. [Google Scholar] [CrossRef]
  3. Adão, T.; Hruška, J.; Pádua, L.; Bessa, J.; Peres, E.; Morais, R.; Sousa, J.J. Hyperspectral Imaging: A Review on UAV-Based Sensors, Data Processing and Applications for Agriculture and Forestry. Remote Sens. 2017, 9, 1110. [Google Scholar] [CrossRef] [Green Version]
  4. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  5. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  7. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  8. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  9. Zhao, W.; Du, S. Spectral–Spatial Feature Extraction for Hyperspectral Image Classification: A Dimension Reduction and Deep Learning Approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  10. Lee, H.; Kwon, H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [Green Version]
  11. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  12. Ma, W.; Yang, Q.; Wu, Y.; Zhao, W.; Zhang, X. Double-branch multi-attention mechanism network for hyperspectral image classification. Remote Sens. 2019, 11, 1307. [Google Scholar] [CrossRef] [Green Version]
  13. Ahmad, M. A fast 3D CNN for hyperspectral image classification. arXiv 2020, arXiv:2004.14152. [Google Scholar]
  14. Zhang, X.; Wang, T.; Yang, Y. Hyperspectral Images Classification Based on Multi-scale Residual Network. arXiv 2020, arXiv:2004.12381. [Google Scholar]
  15. Datt, B.; McVicar, T.R.; Van Niel, T.G.; Jupp, D.L.; Pearlman, J.S. Preprocessing EO-1 Hyperion hyperspectral data to support the application of agricultural indexes. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1246–1259. [Google Scholar] [CrossRef] [Green Version]
  16. Pignatti, S.; Palombo, A.; Pascucci, S.; Romano, F.; Santini, F.; Simoniello, T.; Umberto, A.; Vincenzo, C.; Acito, N.; Diani, M.; et al. The PRISMA hyperspectral mission: Science activities and opportunities for agriculture and land monitoring. In Proceedings of the 2013 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Melbourne, Australia, 21–26 July 2013; pp. 4558–4561. [Google Scholar]
  17. Guanter, L.; Kaufmann, H.; Segl, K.; Foerster, S.; Rogass, C.; Chabrillat, S.; Kuester, T.; Hollstein, A.; Rossner, G.; Chlebek, C.; et al. The EnMAP spaceborne imaging spectroscopy mission for earth observation. Remote Sens. 2015, 7, 8830–8857. [Google Scholar] [CrossRef] [Green Version]
  18. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [Green Version]
  19. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  20. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  21. Yu, F.; Koltun, V.; Funkhouser, T. Dilated residual networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 472–480. [Google Scholar]
  22. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  23. Bengio, Y.; LeCun, Y. Scaling learning algorithms towards AI. Large-Scale Kernel Mach. 2007, 34, 1–41. [Google Scholar]
  24. Peng, C.; Zhang, X.; Yu, G.; Luo, G.; Sun, J. Large kernel matters–improve semantic segmentation by global convolutional network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4353–4361. [Google Scholar]
  25. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  26. Santurkar, S.; Tsipras, D.; Ilyas, A.; Madry, A. How does batch normalization help optimization? arXiv 2018, arXiv:1805.11604. [Google Scholar]
  27. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  28. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  29. Ozdemir, A.; Polat, K. Deep learning applications for hyperspectral imaging: A systematic review. J. Inst. Electron. Comput. 2020, 2, 39–56. [Google Scholar] [CrossRef]
  30. Vali, A.; Comai, S.; Matteucci, M. Deep learning for land use and land cover classification based on hyperspectral and multispectral earth observation data: A review. Remote Sens. 2020, 12, 2495. [Google Scholar] [CrossRef]
  31. Audebert, N.; Le Saux, B.; Lefèvre, S. Deep learning for classification of hyperspectral data: A comparative review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef] [Green Version]
  32. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  33. Alberg, A.J.; Park, J.W.; Hager, B.W.; Brock, M.V.; Diener-West, M. The use of “overall accuracy” to evaluate the validity of screening or diagnostic tests. J. Gen. Intern. Med. 2004, 19, 460–465. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Fung, T.; LeDrew, E. For change detection using various accuracy. Photogramm. Eng. Remote Sens. 1988, 54, 1449–1454. [Google Scholar]
  35. Goutte, C.; Gaussier, E. A probabilistic interpretation of precision, recall and F-score, with implication for evaluation. In Proceedings of the European Conference on Information Retrieval, Santiago de Compostela, Spain, 21–23 March 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar]
  36. Wang, Y.; Yu, W.; Fang, Z. Multiple Kernel-Based SVM Classification of Hyperspectral Images by Combining Spectral, Spatial, and Semantic Information. Remote Sens. 2020, 12, 120. [Google Scholar] [CrossRef] [Green Version]
  37. Polikar, R. Ensemble learning. In Ensemble Machine Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–34. [Google Scholar]
  38. Deng, X.; Liu, Q.; Deng, Y.; Mahadevan, S. An improved method to construct basic probability assignment based on the confusion matrix for classification problem. Inf. Sci. 2016, 340, 250–261. [Google Scholar] [CrossRef]
  39. Foody, G.M. Status of land cover classification accuracy assessment. Remote Sens. Environ. 2002, 80, 185–201. [Google Scholar] [CrossRef]
Figure 1. Structure of pyramid pooling module.
Figure 1. Structure of pyramid pooling module.
Remotesensing 13 01248 g001
Figure 2. Structure of Inception module.
Figure 2. Structure of Inception module.
Remotesensing 13 01248 g002
Figure 3. Structure of residual connection.
Figure 3. Structure of residual connection.
Remotesensing 13 01248 g003
Figure 4. Schematic of dilated convolution.
Figure 4. Schematic of dilated convolution.
Remotesensing 13 01248 g004
Figure 5. Schematic of spectral dilated convolution.
Figure 5. Schematic of spectral dilated convolution.
Remotesensing 13 01248 g005
Figure 6. Structure of multiple spectral resolution module.
Figure 6. Structure of multiple spectral resolution module.
Remotesensing 13 01248 g006
Figure 7. Schematic of Multiple Spectral Resolution 3D Convolutional Neural Network.
Figure 7. Schematic of Multiple Spectral Resolution 3D Convolutional Neural Network.
Remotesensing 13 01248 g007
Figure 8. The training procedure of our method.
Figure 8. The training procedure of our method.
Remotesensing 13 01248 g008
Figure 9. Changes of the overall accuracy of predictions on the four datasets, using 10% samples for training and different patch sizes.
Figure 9. Changes of the overall accuracy of predictions on the four datasets, using 10% samples for training and different patch sizes.
Remotesensing 13 01248 g009
Figure 10. Results of the BO dataset using 10% samples for training and 9 × 9 patch size.
Figure 10. Results of the BO dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g010
Figure 11. Results of the IP dataset using 10% samples for trainingand 9 × 9 patch size.
Figure 11. Results of the IP dataset using 10% samples for trainingand 9 × 9 patch size.
Remotesensing 13 01248 g011
Figure 12. Results of the SA dataset using 10% samples for training and 9 × 9 patch size.
Figure 12. Results of the SA dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g012
Figure 13. Results of the PU dataset using 10% samples for training and 9 × 9 patch size.
Figure 13. Results of the PU dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g013
Figure 14. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the BO dataset using 10% samples for training and 9 × 9 patch size.
Figure 14. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the BO dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g014
Figure 15. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the IP dataset using 10% samples for training and 9 × 9 patch size.
Figure 15. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the IP dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g015
Figure 16. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the SA dataset using 10% samples for training and 9 × 9 patch size.
Figure 16. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the SA dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g016
Figure 17. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the PU dataset using 10% samples for training and 9 × 9 patch size.
Figure 17. The confusion matrix between the predicted results of MSR-3DCNN and ground truth on the PU dataset using 10% samples for training and 9 × 9 patch size.
Remotesensing 13 01248 g017
Table 1. Summary of the characteristics of the IP, SA, PU, and BO datasets.
Table 1. Summary of the characteristics of the IP, SA, PU, and BO datasets.
DescriptionDatasets
IPSAPUBO
SensorAVIRISAVIRISROSISHyperion
Pixels 145 × 145 512 × 217 610 × 340 1476 × 256
GSD20 m/pixel3.7 m/pixel1.3 m/pixel30 m/pixel
Spectral range0.4 to 2.5 µm0.4 to 2.5 µm0.43 to 0.86 µm0.4 to 2.5 µm
Original band  ( 1 ) 224224103242
Processed band  ( 2 ) 200204103145
Ground truth  ( 3 ) 10,24954,12942,7763248
Class  ( 4 ) 1616914
ModeAirborneAirborneAirborneSatellite
(1) number of spectral bands in the original HSI. (2) number of spectral subsetted from the raw dataset. (3) number of pixels for ground-truthing. (4) number of land cover and land use classes.
Table 2. Overall Accuracy (%) of the predictions by contextual deep CNN, HybridSN, DBMA, Fast 3D CNN, MSRN, and MSR-3DCNN using different patch size and the fuse the prediction results of these predictions obtained by sum rule.
Table 2. Overall Accuracy (%) of the predictions by contextual deep CNN, HybridSN, DBMA, Fast 3D CNN, MSRN, and MSR-3DCNN using different patch size and the fuse the prediction results of these predictions obtained by sum rule.
DatasetsPatch SizeModels
Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
BO 9 × 9 67.3797.2697.3098.5095.4298.80
IP60.6596.4198.0397.0495.8398.10
SA99.8299.6599.9199.8899.9499.96
PU97.6099.2899.5899.3799.8499.62
BO 11 × 11 74.3899.1598.8797.9586.9798.60
IP64.2397.3998.6097.4791.1497.56
SA99.9599.8299.9599.8199.9699.97
PU97.6099.0099.6499.6099.8599.60
BO 13 × 13 79.2199.5999.0897.9591.9399.15
IP72.4497.7797.0898.5197.2197.30
SA99.9599.7999.9199.9399.9299.99
PU98.3899.6099.5499.1999.8499.72
BOFusion85.9499.0899.9799.4591.2199.62
IP73.0598.3398.6898.3598.2398.41
SA99.9799.9399.9599.9199.9799.98
PU98.7399.5999.7299.5299.9099.76
Table 3. F1 Score (%) for Each Class, OA (%), AA (%), and Kappa Coefficient Values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the BO datasets.
Table 3. F1 Score (%) for Each Class, OA (%), AA (%), and Kappa Coefficient Values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the BO datasets.
ClassTrainingTestModels
Total3242924Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
12724383.48100.0099.3899.1899.7999.38
2109162.82100.0098.38100.0094.3099.45
32522693.0599.12100.0098.91100.00100.00
42119495.3193.27100.0098.9899.4899.74
52724254.9592.4899.1795.9196.0395.54
62724245.8291.8699.5997.7394.1796.45
72623378.87100.0099.7999.36100.00100.00
82018382.58100.0099.46100.0099.1999.19
93128353.5598.4199.6596.7398.5797.75
102522351.6999.3384.7599.5590.4299.55
113027575.2799.4590.16100.0085.8099.10
121816346.9399.0897.3199.3894.7799.38
132724145.5299.7998.1099.5991.2099.59
14108530.5175.9197.7089.1793.83100.00
OA67.3797.2697.3098.5095.4298.80
AA64.9195.8697.5397.8695.4798.89
Kappa coefficient0.64580.97030.97070.98370.95030.9870
Training time19.3517.6650.0214.2851.86224.18
Test time0.740.731.990.481.547.09
Table 4. F1 Score (%) for each class OA (%), AA (%), and Kappa coefficient values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the IP dataset.
Table 4. F1 Score (%) for each class OA (%), AA (%), and Kappa coefficient values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the IP dataset.
ClassTrainingTestModels
Total10249225Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
154122.64100.0080.00100.0042.3189.19
2143128527.8495.9297.3896.7193.2297.74
38374722.1494.3098.2692.6292.5997.88
42421360.9695.3598.6188.3181.6795.63
54843588.3097.9397.9297.8398.2799.08
67365781.0199.7098.4698.9498.7999.00
732527.59100.00100.00100.0096.15100.00
84843092.76100.0099.08100.0097.0799.08
921812.9068.9753.3380.0028.5797.14
109787524.3892.0497.8396.8697.2097.33
11245221071.9096.2297.8897.5897.1398.43
125953430.7993.7997.1395.2694.4796.16
132018580.00100.0099.7399.7399.7398.93
14126113991.0199.0899.5699.1798.5399.39
153934743.1297.5198.4096.7595.5197.35
1698442.7596.4796.3995.2994.1293.57
OA60.6596.4198.0397.0495.8398.10
AA50.3194.4593.0594.7785.7896.40
Kappa coefficient0.54470.95910.97750.96630.95230.9784
Training time39.0933.45126.4030.99117.89667.04
Test time1.511.335.231.023.8019.30
Table 5. F1 Score (%) for each class, OA (%), AA (%), and Kappa Coefficient values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the SA dataset.
Table 5. F1 Score (%) for each class, OA (%), AA (%), and Kappa Coefficient values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the SA dataset.
ClassTrainingTestModels
Total541248,717Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
1201180899.75100.00100.00100.00100.00100.00
23723354100.00100.00100.00100.00100.00100.00
3197177999.9299.97100.00100.00100.00100.00
4139125599.2999.8499.4899.68100.0099.76
5268241099.5099.9299.7399.83100.0099.88
6396356399.97100.00100.00100.00100.00100.00
7358322199.92100.00100.00100.00100.00100.00
8112710,14499.8699.3099.8999.7999.8799.98
9620558399.9299.89100.0099.98100.0099.96
10328295099.4699.6699.8399.8199.9599.92
1110796199.8499.84100.0099.95100.00100.00
12193173499.8099.9199.9799.97100.0099.97
1391825100.0099.8899.94100.00100.0099.82
1410796399.5399.6999.8499.6499.9599.84
15727654199.7998.9599.8299.6999.7999.97
16181162699.8899.88100.00100.0099.94100.00
OA99.8299.6599.9199.8899.9499.96
AA99.8199.7899.8999.8999.9799.95
Kappa coefficient0.99790.99610.99900.99860.99930.9996
Training time174.42153.89650.92144.36614.323476.82
Test time7.766.3325.184.3717.30100.28
Table 6. F1 Score(%) for each class, OA (%), AA (%), and Kappa Coefficient values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the PU dataset.
Table 6. F1 Score(%) for each class, OA (%), AA (%), and Kappa Coefficient values for the entire dataset, training time (s), and test time (s) using 9 × 9 patch size and 10% samples for training on the PU dataset.
ClassTrainingTestModels
Total427738,499Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
1663596895.9399.3199.6299.4699.8699.72
2186516,78499.9399.9899.9999.95100.0099.98
3210188994.6395.2398.2096.2298.8597.28
4306275894.5899.5398.9299.4099.9899.51
5134121199.9299.9299.63100.00100.0099.96
6503452699.4299.9499.9399.87100.0099.99
7133119794.1699.2999.5899.8399.96100.00
8368331492.9496.5498.5196.8999.1298.36
99585288.8599.4798.9499.8899.9499.11
OA97.5999.2899.5899.3799.8499.62
AA94.9898.6499.0798.9899.7299.13
Kappa coefficient0.96810.99040.99450.99170.99790.9949
Training time137.17124.86526.00113.83485.232747.81
Test time5.444.5320.803.6613.8875.55
Table 7. Standardized normal test statistic (Z) for the BO dataset using 9 × 9 patch size and 10% samples for training.
Table 7. Standardized normal test statistic (Z) for the BO dataset using 9 × 9 patch size and 10% samples for training.
Class 2Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
Class 1
contextual deep CNN0.00−29.56−29.58−30.17−28.60−30.32
HybridSN29.560.00−0.08−5.694.38−5.42
DBMA29.580.080.00−3.425.53−4.54
Fast 3D CNN30.175.693.420.008.22−1.67
MSRN28.60−4.38−5.53−8.220.00−9.08
MSR-3DCNN30.325.424.541.679.080.00
Table 8. Standardized normal test statistic (Z) for the IP dataset using 9 × 9 patch size and 10% samples for training.
Table 8. Standardized normal test statistic (Z) for the IP dataset using 9 × 9 patch size and 10% samples for training.
Class 2Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
Class 1
contextual deep CNN0.00−57.44−58.72−57.94−56.96−58.78
HybridSN57.440.00−9.93−5.094.00−10.77
DBMA58.729.930.006.5813.19−0.69
Fast 3D CNN57.945.09−6.580.009.21−8.00
MSRN56.96−4.00−13.19−9.210.00−13.79
MSR-3DCNN58.7810.770.698.0013.790.00
Table 9. Standardized normal test statistic (Z) for the SA dataset using 9 × 9 patch size and 10% samples for training.
Table 9. Standardized normal test statistic (Z) for the SA dataset using 9 × 9 patch size and 10% samples for training.
Class 2Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
Class 1
contextual deep CNN0.006.39−5.29−3.64−6.79−8.43
HybridSN−6.390.00−10.22−9.99−11.79−12.09
DBMA5.2910.220.002.66−1.82−3.54
Fast 3D CNN3.649.99−2.660.00−3.69−5.98
MSRN6.7911.791.823.690.00−1.81
MSR-3DCNN8.4312.093.545.981.810.00
Table 10. Standardized normal test statistic (Z) for the PU dataset using 9 × 9 patch size and 10% samples for training.
Table 10. Standardized normal test statistic (Z) for the PU dataset using 9 × 9 patch size and 10% samples for training.
Class 2Contextual Deep CNNHybridSNDBMAFast 3D CNNMSRNMSR-3DCNN
Class 1
contextual deep CNN0.00−25.32−27.46−26.17−29.41−27.91
HybridSN25.320.00−8.39−3.36−14.76−11.00
DBMA27.468.390.006.02−8.85−1.41
Fast 3D CNN26.173.36−6.020.00−13.34−7.73
MSRN29.4114.768.8513.340.007.98
MSR-3DCNN27.9111.001.417.73−7.980.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, H.; Yao, W.; Cheng, L.; Li, B. Multiple Spectral Resolution 3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sens. 2021, 13, 1248. https://doi.org/10.3390/rs13071248

AMA Style

Xu H, Yao W, Cheng L, Li B. Multiple Spectral Resolution 3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sensing. 2021; 13(7):1248. https://doi.org/10.3390/rs13071248

Chicago/Turabian Style

Xu, Hao, Wei Yao, Li Cheng, and Bo Li. 2021. "Multiple Spectral Resolution 3D Convolutional Neural Network for Hyperspectral Image Classification" Remote Sensing 13, no. 7: 1248. https://doi.org/10.3390/rs13071248

APA Style

Xu, H., Yao, W., Cheng, L., & Li, B. (2021). Multiple Spectral Resolution 3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sensing, 13(7), 1248. https://doi.org/10.3390/rs13071248

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop