Next Article in Journal
JointNet: A Common Neural Network for Road and Building Extraction
Next Article in Special Issue
Convolutional Neural Network and Guided Filtering for SAR Image Denoising
Previous Article in Journal
Domain Transfer Learning for Hyperspectral Image Super-Resolution
Previous Article in Special Issue
Detection of Fir Trees (Abies sibirica) Damaged by the Bark Beetle in Unmanned Aerial Vehicle Images with Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multiscale Deep Middle-level Feature Fusion Network for Hyperspectral Classification

1
School of Computer Science, Shenyang Aerospace University, Shenyang 110136, China
2
College of Mathematics and Computer Science, Yan’an University, Yan’an 716000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(6), 695; https://doi.org/10.3390/rs11060695
Submission received: 17 February 2019 / Revised: 19 March 2019 / Accepted: 19 March 2019 / Published: 22 March 2019
(This article belongs to the Special Issue Convolutional Neural Networks Applications in Remote Sensing)

Abstract

:
Recently, networks consider spectral-spatial information in multiscale inputs less, even though there are some networks that consider this factor, however these networks cannot guarantee to get optimal features, which are extracted from each scale input. Furthermore, these networks do not consider the complementary and related information among different scale features. To address these issues, a multiscale deep middle-level feature fusion network (MMFN) is proposed in this paper for hyperspectral classification. In MMFN, the network fully fuses the strong complementary and related information among different scale features to extract more discriminative features. The training of network contains two stages: the first stage obtains the optimal models corresponding to different scale inputs and extracts the middle-level features under the corresponding scale model. It can guarantee the multiscale middle-level features are optimal. The second stage fuses the optimal multiscale middle-level features in the convolutional layer, and the subsequent residual blocks can learn the complementary and related information among different scale middle-level features. Moreover, the idea of identity mapping in residual learning can help the network obtain a higher accuracy when the network is deeper. The effectiveness of our method is proved on four HSI data sets and the experimental results show that our method outperforms the other state-of-the-art methods especially with small training samples.

1. Introduction

Hyperspectral Imaging (HSI) has hundreds of continuous spectral bands and high spatial correlation, so it contains abundant spectral and spatial information which is useful for the classification of different materials. HSI has been applied to many fields, including environment management [1], geological mapping [2], mineral exploitation [3], and scene recognition [4]. Although HSI contains rich spectral information, it is difficult to obtain enough training samples in practice, which often leads to the “curse of dimensionality”. In addition, the neighboring bands of HSI are of high correlation, which means that only a few bands play a critical role. This increases the computational complexity and affects the following classification process. Therefore, dimensionality reduction (DR) is necessary for HSI classification preprocessing, which can reduce computational complexity and retain useful information of classification [5,6,7,8]. Feature selection and feature extraction are the traditional methods to implement DR [9]. Feature selection aims to find more discriminative bands from the raw HSI data to represent the entire image and this method can remain the physical meaning of original data [10,11,12,13,14,15]. Some clustering-based methods [16,17] and ranking-based methods [18,19] find the representative bands to classify distinct classes. Compared with feature selection, feature extraction [20,21,22,23,24,25,26] finds more useful features through mathematical transformation to improve the classification accuracy. These features are, for example, multilinear principal component ananlysis (PCA) [27] and Fisher’s linear discriminant analysis [28] etc., but these methods can only extract low-level features which have limited representation capacity to express the abundant information of spectral and spatial features.
Recently, about the above issue, many deep learning models have been proposed and they can learn more distinguished features with the goal of high classification accuracy [29,30]. In the typical deep learning model, stacked autoencoders (SAEs) can extract spatial and spectral information, then combine these features for HSI classification [10]. The potential of deep belief networks (DBN) [31] and restricted Boltzmann machines [32] is used to combine the spatial and spectral information to classify the image. These methods are intended for 1-D input and the input data misses the spatial structure information, which is important for HSI classification. A deep convolutional neural network (CNN) [33] is adopted to get the spatial feature and it has no requirement for the input. 3-D CNN is used to extract spectral-spatial features from the original image directly and gets better classification accuracy. Reference [34] proposes an end-to-end framework to learn the spectral and spatial features and this method can exploit the correlation between the spectral and spatial domains. But in this framework, the input of spectral data is 1-D dimension. It is missing the neighborhood information of spatial dimension. And the classification accuracy of these deep learning models will decrease when the network is deeper. Reference [35] proposes a supervised spectral–spatial residual network and the idea of identity mapping in residual blocks mitigates the decreasing-accuracy phenomenon, but this network firstly learns the spectral features that are used as the input to extract the spatial information, so the spatial features are found from data that has been transformed and so misses the original spatial correlation. Reference [36] applies CNN to extract multiple spatial features and then stacks with spectrum to generate the spectral-spatial feature. This method would have better performance if the spectral feature was extracted by multiscale. Song W proposes a deep fusion feature network [37] for classification. In this network, the features from the low layer, the middle layer and the high layer are respectively extracted by the residual network, and the features of different layers are fused in the fully convoluted layer to classify the image. Although the network considers the influence of different layer features on the classification, it does not consider the spectral-spatial fusion features and it directly extracts the features from the original image. Moreover, the features of fusion at the fully connected layer cannot enable the entire network to fully use the fusion features to learn more discriminative features. Most of the proposed deep learning models now consider the spectral-spatial fusion feature under single scale input and do not consider the abundant correlation between the spectral and spatial in multiscale inputs. Even though some models consider multiscale inputs, however, they cannot guarantee that each scale feature is optimal. Furthermore, these models cannot make full use of the strong complementary and related information among the multiscale fusion features because the features are fused in the fully connected layer to directly classify the image.
To solve these problems and extract more discriminative fusion features, we propose a multiscale deep middle-level feature fusion (MMFN) network for hyperspectral image classification. The training of the network contains two stages: in the first stage, each scale input is used to train a model and the optimal model is saved. The middle-level feature is extracted from the corresponding scale model and it can guarantee the multiscale middle-level features are optimal. In the second stage, the multiscale middle-level features are fused in the convolution layer and the subsequent residual learning block can fully use the strong complementary and related information among multiscale fusion features to extract more discriminative and higher-level features for classification. Furthermore, the residual learning [38] can help the network maintain a higher accuracy when the layer is deeper and make the network more robust.
The three major contributions of this paper include:
(1) The idea of multiscale features fusion is proposed, and this is an idea that contains more abundant neighborhood correlation and low-level features, such as spatial structure, and texture features, which are beneficial for classification.
(2) The training of the network consists of two stages, the first stage obtains the optimal models corresponding to different scales, and extracts the middle-level features under the corresponding scale model. It can ensure the multiscale middle-level features are optimal, which is helpful for the subsequent training stage extracting more discriminative features. The second stage fuses the optimal multiscale middle-level features in the convolutional layer to train a new model for final classification.
(3) Different scale features have strong complementary and related information. Compared with the features that are fused directly in the fully connected layer to classify the image, the multiscale deep middle-level features are fused in the convolutional layer, which can enable the network to make full use of the strong complementary and related information among multiscale fusion features. Moreover, the subsequent residual learning modules can learn the multiscale fusion features to extract more discriminative and higher-level features for classification and can help the network maintain a higher accuracy with deeper layers.
The rest of this paper is organized as follows. Section 2 introduces the detailed architecture of our method. Section 3 presents the results of classification accuracy on the four data sets, and shows the performance of all methods. Finally, the conclusion is provided in Section 4.

2. Methodology

A deep network can be regarded as a process of feature learning, which is a step-by-step abstract representation of the original input through a hidden layer. It can learn the original input data structure and find more useful features. Through feature combination, it transfers the original input into the low-layer features, middle-level features, high-level features up to the final mission objectives. Deep learning through the learning of hierarchical features can extract features from the texture information in the low-level features to the local information in the middle layer to the object information in the high-level layer. From this process, it is not difficult to find the connection between the original input and the low-level features, and the connection between the middle-level layer features and the high-level features, it is difficult to cross directly from the original input to the high-level features. In the MMFN framework, it consists of two training stages. The first stage mainly obtains the optimal model corresponding to each scale, and extracts the features of the last residual block of the corresponding scale in the optimal model. The second stage mainly fuses the multiscale features from the first stage in the convolution layer to train a new model for final classification. Because the multiscale features are extracted from the residual block, which are neither the low-level features that are close to the original image or nor the high-level features that are close to the fully connected layer, the features are defined as the middle-level features in the MMFN network, and the multiscale fusion features are used as inputs to a new model in the second stage to learn more discriminative features for classification.

2.1. Extracting Multiscale Deep Middle-Level Features

HSI data can be denoted as R M × N × L , R i M × N is i the band image, M , N , L denote that the Hyperspectral Image has M × N pixels, and L bands, respectively. The main purpose of first training stage on MMFN is to extract optimal multiscale deep middle-level features and each scale 3-D data cube is used to train the corresponding model. The model contains a spectral and spatial learning module with different size of convolution filters. Let x be the input of a convolutional layer and x i is the i th feature map of x . Supposing that the convolutional layer has k filters denoted as W and the bias parameter is b, the j output of the convolutional layer can be represented:
z j = i = 1 L R ( x i W j + b j ) j = 1 , 2 ..... , k ,
The features from the spectral and spatial module are fused as spectral-spatial fusion features, the fusion operation is defined as:
Y f u s i o n = g { W i [ y s p e c t r a l y s p a t i a l ] + b i } ,
g ( x ) = max ( 0 , x ) ,
g represents the relu function and it is a rectified linear unit activation function which sets elements with negative numbers to zero. y s p e c t r a l and y s p a t i a l represent the outputs which are found from the spectral and spatial learning module, respectively. The subsequent residual learning module can use the spectral-spatial fusion feature to learn more discriminative features and the structure of residual learning block is showed in Figure 1.
In Figure 1, x represents the input of the first residual block, Y is the function learned through a two convolution layers and it is defined as:
Y = F ( x , { W } ) + x ,
W is the convolutional kernel, F is residual function and can be written as:
F = R ( R ( x W 1 + b 1 ) W 2 + b 2 ) ,
W 1 , W 2 is parameters of the first and second convolution kernel, respectively, b 1 , b 2 is the next and the next two layers bias of the input layer, respectively. In residual learning, we use the batch normalization (BN) operation to regularize the learning process for every convolutional operation and BN is formulated as:
Y n = R ( y n 1 W n + b n ) ,
and Y n represents the output of n th layer after BN operation, W n , b n mean the convolutional kernels and bias, respectively on the n th layer. And the y n 1 is defined as:
y n 1 = Y n 1 E ( Y n 1 ) V a r ( Y n 1 ) ,
which Y n 1 is the output of (n−1) th layer after BN operation. After the residual block layer, the average pooling operation is done for the output of the residual block and the average pooling operation is formulated as:
z = 1 C ( i , j ) S x i j ,
We suppose the S is the filter size and C is the number of elements of S, x i j is the value of the corresponding position ( i , j ) in the input data x , and in this paper, we use global average pooling. After the average pooling, the feature is sent to the softmax layer for HSI classification. The predicted value of the framework is a vector y ^ = [ y ^ 1 , y ^ 2 , .... , y ^ c ] , and the truth label vector y = [ y 1 , y 2 , .... , y c ] , c is the number of land-cover categories. The parameters of the framework are updated through back propagating the gradients of the cross-entropy objective function which is defined as:
c r o s s ( y ^ , y ) = i = 1 L y i ( log j = 1 L e y ^ j y ^ i ) ,
For each scale input data, the above operation is done for training a corresponding scale model. In order to get an optimal model, we use the classification accuracy of a validation set to see whether it is improved in some training epochs to determine whether the model is optimal. Through this method, one can guarantee every scale input corresponds to an optimal model, and the feature from the last residual block is extracted on every model as the deep middle-level feature. These multiscale features are calculated by the weights from the optimal trained model, so these multiscale features are the best, which is helpful for the final classification.

2.2. Fusing Multiscale Deep Middle-Level Features

In the first training stage of MMFN, we can get the optimal multiscale middle-level features and these features have different spatial sizes because of the different scales of inputs. Before fusing these features, the spatial size of features should be same. For example, there have been three different size of features which are F 1 , F 2 , F 3 respectively, the spatial sizes are 5 × 5, 7 × 7, 9 × 9. For the size of 7 × 7, 9 × 9, and we can use 3 × 3 and 5 × 5 filters to make the features with same size of 5 × 5, then these three features are fused. The fusion operation is formulated as:
X = g { W [ F 1 A 2 ( F 2 ) A 3 ( F 3 ) ] + b } ,
X represents the tensor after fusing the multiscale middle-level features in the convolution layer, A 2 , A 3 is the different convolution operations to guarantee the features with the same spatial size. is concatenating the outputs from the multiscale features, W , b denote the convolutional kernels and bias in the convolution layer respectively. After getting the multiscale middle-level fusion feature, the residual block is used to learn higher-level and discriminative features, which are sent to the softmax layer for the final classification.

2.3. Classifying HSI Based on the MMFN

We take the IN Data Set as an example to describe the architecture of our method in Figure 2 and Figure 3. Figure 2 shows the first training stage of the MMFN, the size of 7 × 7 × 200 as the input data are sent to the spatial learning and spectral learning module with the size of 3 × 3,128 and 1 × 1, 128, respectively and the features with the size of 7 × 7 × 128 are obtained. Then these features are concatenated as the spectral-spatial fusion features to do the next convolutional and BN operation, and the size of features is constant with 7 × 7 × 128. In a residual learning module, it contains two residual blocks and every block uses the size of 3 × 3 × 128, 24 filters to extract features from the spectral-spatial fusion feature tensor, the feature size of 5 × 5 × 24 is generated after residual learning and the BN operation is done after every convolutional layer, which can regularize the learning process and improve the classification performance. What the feature tensor gets from the residual block is an input to the average pooling layer and it can obtain a 1 × 1 × 24 vector that is sent to the softmax layer for the final classification. After several epochs of training, we can get the optimal model. For the other scale of 9 × 9 × 200 and 11 × 11 × 200 are the inputs of the network, they are done the same as the above operation and we can get the corresponding scale of the model. From the optimal models corresponding to multiscale inputs, the features, of which the sizes are 5 × 5 × 24, 7 × 7 × 24, 9 × 9 × 24, respectively, are extracted from the last residual block as the deep middle-level features.
Figure 3 shows the second training stage of the whole network and this stage fuses the multiscale middle-level features and sends them to the residual learning block to learn the higher-level and discriminative features for the final classification. Because each input scale corresponds to a separately trained model, we save the parameters of the optimal model by the classification accuracy of the validation set, and in this way, we can guarantee that the middle-level feature is calculated under the optimal parameters when the features are extracted from the corresponding scale model. It means that the multiscale middle-level features are optimal which can improve the classification performance. The middle-level features have a size of 5 × 5 × 24, 7 × 7 × 24, 9 × 9 × 24, and we can use 3 × 3, 24 and 5 × 5, 24 kernels to convolute the features tensors of 7 × 7 × 24, 9 × 9 × 24, respectively and make them with the size of 5 × 5 × 24. Then, these three different scales of middle-level features are concatenated to generate a tensor with the size of 5 × 5 × 72. Through the size of 3 × 3 filters, the fused feature with size of 5 × 5 × 72 will be transformed into a tensor with size of 5 × 5 × 24 as an input to the residual learning module which contains two residual blocks and every block consists of 3-D convolution operation with size of 3 × 3 × 128, 24. Finally we can get the higher-level feature with size of 5 × 5 × 24 as the input of the average pooling layer and a tensor with size of 1 × 1 × 24 is generated for the final classification.

3. Experimental Results

3.1. Data Description and Experimental Settings

In this section, the effectiveness of our method is proved in four real-world hyperspectral remote sensing data sets which contain the Indian Pines (IN) Data Set, Pavia University (UP) Data Set, Kennedy Space Center (KSC) Data Set and Salinas Valley (Salinas) Data Set and the proposed method is compared with other state-of-the-art methods. The overall accuracy (OA) and the average accuracy (AA) are the classification metrics used to assess the classification performance of all the methods.
The Indian Pines Data Set (IN) was collected by AVIRIS in 1992 in northwestern Indiana. This commonly used data set has 16 vegetation classes and 224 bands. The spatial size is 145×145 and the spatial resolution is 20 m per pixel. To avoid the negative influence on classification due to water absorption and noise, some bands are discarded and the remaining 200 bands are adopted for analysis.
The Pavia University Data Set (UP) was captured by a Reflective Optics System Imaging Spectrometer optical sensor over an urban area surrounding the University of the Pavia. The image is of size 610 × 340 × 115 with a resolution of 1.3 m per pixel and 9 urban land-cover classes are considered in this experiment. The number of remaining bands is 103 after discarding the useless bands.
The KSC Data Set was collected by AVIRIS in 1996 in Florida, and contains 512 × 614 pixels with spatial resolution of 18 m per pixel and the ground-truth classes are 13. After removing the noise bands, 176 bands are retained and used for the experiment.
The Salinas Data Set which was gathered by AVIRIS and it consists of 224 bands with spatial size of 512 × 217 pixels. The spatial resolution of the data is 3.7 m per pixel and the ground-truth classes are 16. Twenty noisy bands are removed, and 204 bands are left for the next experiment.
The information of data sets is shown in Table 1, Table 2, Table 3 and Table 4. The corresponding the false-color image and ground-truth map are shown in Figure 4, Figure 5, Figure 6 and Figure 7. For all data sets, the number of experiments was twenty times to reduce the influence of random effects, which are caused by randomly choosing different training samples every time. Through verifying whether the accuracy of validation set is improved in some epochs to determine whether the network model is optimal, the optimal weight values of each model were saved. We made the average effects the final results to evaluate the classification accuracy of every method. We evaluated the performance of all methods on the small training samples to prove that our proposed MMFN has strong robustness and generalization. Furthermore, the MMFN had a better performance on classification accuracy when training samples were small. In four data sets which contained IN, UP, KSC and Salinas, we split the data into training, validation, testing set, and the ratio was 5%, 10%, 85%, respectively on these four data sets.
In our implementation, the training epoch was set to 100 and the optimizer adopted the standard stochastic gradient descent method. The batch size was set to 64, the optimum learning rates in IN, UP, KSC, Salinas data set were fixed as 0.0003, 0.0001, 0.0001, 0.0003, respectively, and the momentum was set to 0.9.
The proposed method was compared with some state-of-the-art methods including the SVM [39], ResNet [38], SAE [10], 3-D CNN [23] and Two-CNN [34], SSRN [35]. In order to compare fairly, we used the SVM using spatial information through the Gaussian filter. The framework of ResNet adopts the same residual blocks as our method and it does not contain the spectral-spatial learning module.

3.2. Influence of Parameters

3.2.1. The Selection of Multiscale Inputs

The spatial scale of input data was changed on the four data sets, and the appropriate multiscale inputs were determined through the classification accuracy achieved by the different scale inputs. In the four data sets, the data was split into 5%, 10% and 85% to comprise the training set, validation sets, and test sets, respectively. The classification results of different spatial scale inputs are shown in Table 5. In most cases, as the spatial scale of the input became larger, the classification accuracy achieved higher results on the four data sets. It was proved that when the spatial scale was within a certain range, the input data contained more spatial structure information with a larger spatial scale, which was helpful for the network learning more discriminative features and obtaining a higher classification accuracy.
From Table 5, when the spatial scale of input was larger than or equal to 11 × 11, the improvement of classification accuracy on the four data sets was relatively small and basically stable. And when the spatial scale of the input was 7 × 7, the classification results on the four data sets reached a higher accuracy, which had obvious advantages over the classification accuracy achieved by 3 × 3 and 5 × 5 as spatial input scales. Therefore, in order to select a relatively small spatial scale of input and achieve a higher classification accuracy, the selection of spatial scales were 7 × 7, 9 × 9, 11 × 11 as the multiscale inputs of the whole network.

3.2.2. The Effectiveness of Multiscale Inputs

In order to validate the suggestion that multiscale inputs were more beneficial for HSI classification than a single input, some experiments were done on these four data sets. The experiment results are showed in Table 6, and the number of training samples for the four data sets ranges from 3% to 6% for each class, and the classification accuracy is evaluated by the overall classification accuracy (OA). It can be seen from the Table 6 that the input with a large spatial scale has a higher classification accuracy than the input with a small spatial scale. Although the number of training samples increased the classification accuracy of the input with large spatial scale and the input with small spatial scale were both improving, however the accuracy of larger spatial scale input was higher in all cases than the smaller spatial scale input. When the training samples of the four data sets were 3%, the classification accuracy obtained by multiscale inputs was higher than the single scale input. Especially in the IN data set, which was difficult to classify, the classification accuracy achieved by the multiscale inputs had an obvious advantage compared with the single scale input. From the Table 6, the bolded classification accuracies of multiscale inputs are higher than the single input, no matter how many training samples in the four data sets. It proves the multiscale inputs are more useful than single input for classification.
Although the classification accuracy did not improve significantly in the UP and KSC data sets, and the single scale input and multiscale inputs were both reaching a higher accuracy because the UP and KSC data sets both have a higher spatial resolution, the advantages of multiscale inputs were not obvious, but in most cases the classification performance of multiscale inputs had better generalization and effectiveness with small labeled samples. Multiscale inputs of the network can generate multiscale spectral-spatial fusion features that contain abundant the correlation between spatial and spectral, spatial structure information and texture information compared with the spectral-spatial fusion feature generated by single scale input. This information can help the network learn more discriminative features for better classification. With increase of training samples, the classification accuracy was improved in most cases, and the multiscale inputs achieved better classification results than single scale input. The experimental results also proved that the idea of multiscale inputs was more suitable for deep network classification and could improve the final classification accuracy.

3.2.3. The Selection of Number of Residual Block

In the second stage of the MMFN, the residual learning module was used to learn higher-level features from the middle-level fusion features. The selection of the number of residual blocks in the network was determined by experiments in this section. In four data sets that were IN, UP, KSC, Salinas, the suitable number of residual blocks was selected through the classification accuracy achieved by changing the number of residual blocks and training samples. The result is shown in Table 7. It can be seen that by increasing the number of training samples, the classification accuracy was improved regardless of the number of residual blocks in the four data sets. It was also proved that more labeled samples are helpful for improving classification results. In comparing the different number of residual blocks in the case of the same training samples, the classification accuracy was higher when the network contained two residual blocks than zero, one and three residual blocks on the four data sets.
Although on the UP and Salinas data sets, when the number of training samples reaches 6%, the network contains two residual blocks and the classification accuracy is slightly lower than the accuracy of three residual blocks, because the UP and Salinas data sets have a high spatial resolution with 1.3 m and 3.7 m respectively, which helps the network to achieve a high accuracy even with small training samples. So, when the number of training samples is 6% per class, the MMFN has an excellent performance regardless of the network contains zero, one, two or three residual blocks. The results of classification accuracy show that the network contained residual blocks and achieved higher accuracy than the network which did not use residual block in the four data sets and it proves that the residual block can help the network to improve classification accuracy. In most cases, the classification performance was better than other number of residual blocks when the network contained two residual blocks from the bolded classification accuracies. The result also shows that the network layer was shallow with one residual block, and the features that may be learned were not discriminative, and the three residual blocks fused features from too many layers which may introduce too much redundant information resulting in a reduction of classification accuracy. Through the experimental results, we chose two residual blocks to form the structure of MMFN in the second training stage.

3.3. Experiment Results and Analysis

In order to prove the superiority of the proposed network MMFN in the case of small label samples, we compared MMFN with other state-of-the-art methods on the four data sets, and the classification results are shown in Figure 8. Changing the number of training samples from 3% to 6% each class. Figure 8a shows the classification performance of each method on the IN data set. It can be seen from the Figure 8a that MMFN has a distinct advantage over other methods when the number of training samples was 3%, it showed that the MMFN network can learn more discriminative features to help with classifying the image even with small training samples. Figure 8b is the classification result of each method on the UP data set. Although the classification accuracy curve of the MMFN network was close to the curves of the SSRN when the number of training samples increased. However, when the number of training samples was small, the classification performance of the MMFN network was better than other methods. Figure 8c shows the performance of all methods in the KSC data set.
MMFN has obvious advantages over other methods at most cases. It shows that the optimal middle-level features are helpful for the second training stage extracting more discriminative features, and the multiscale middle-level features are fused in the convolution layer can make the network to learn strong complementary and related information among multiscale features. Figure 8d is the classification result on the Salinas data set. Because the spatial resolution of the Salinas data set is high, the accuracies achieved by the MMFN, and SSRN networks are high in the case of small training samples, however it can be seen from the figure that MMFN still has obvious advantages in classification.
From the experimental results of the four data sets, the MMFN network fused the extracted multiscale middle-level features in the convolutional layer, which helped the residual network to learn more discriminative and higher-level features, however in Two-CNN, the fusion features were fused in the fully connected layer and this made the network use these fusion features only in the classification, which may have reduced the classification accuracy.
Table 8 shows the classification accuracies of different methods on the four data sets which contain IN, UP, KSC, and Salinas. Table 9, Table 10, Table 11 and Table 12 list the class-specific accuracies of different methods on the four data sets. The training set, validation set, and test set are split into 5%, 10%, and 85%, respectively. It can be seen from the bolded classification accuracied in these tables that MMFN performs the better than other methods in OA and AA in most cases, and it proves the effectiveness of the network. MMFN achieved higher classification accuracy than the ResNet network on the four data sets, because the MMFN made full use of the multiscale middle-level features, and sent the fused features to the residual block instead of learning directly from the original image like ResNet.
This also shows that the optimal middle-level feature obtained by each scale input in the first stage of MMFN was beneficial for classification. This also proves the validity of extracting middle-level features in MMFN. Compared with the idea of spectral-spatial fusion in SSRN, MMFN introduces the idea of multiscale inputs, which provides the network with abundant complementary and related information among different scale features, and the spectral and spatial learning module in MMFN is based on original image, it can extract more primitive and accurate spatial structure information.
The input of the spatial learning module in the SSRN network is based on the features extracted from the spectral learning module, the spatial learning will miss the spatial information in the original image, so the classification accuracy is lower than the MMFN. It can be seen from the table that the variance value of MMFN network classification result is smaller than other methods in most cases, which shows the stability of the network.
The training and testing times provide a direct measure of computational efficiency for MMFN. All experiments were conducted on an HP z620 workstation with GT 980Ti graphical processing unit (GPU). The loss function values on the four data sets were 0.2520(IN),0.2318(UP),0.0653(KSC) and 0.1457(Salinas), respectively. Table 13 shows the results of training and test times for all methods on four different data sets. The training set, validation set, and test set of all methods on the four data sets were split into 5%, 10%, and 85%, respectively. Two-CNN, ResNet, SAE, CNN, SSRN, MMFN were iterated 100 times and the SVM was trained 20 times. It can be seen from Table 13 that the training of the MMFN network takes the longest time, because the training of the network is divided into two phases, and the network has multiscale inputs to increase the computational time, although MMFN is longer than the training time of SSRN about 1–6 minutes in the larger data sets such as UP and Salinas, but when MMFN has small labeled samples, the classification accuracy of the network is higher than SSRN, especially in the IN data set that is difficult to classify, the advantage of the MMFN is obvious. In other methods, ResNet contains two residual blocks, so the training time of it is longer than Two-CNN, SAE, and CNN, but its classification accuracy is higher than these methods.

3.4. Discussions

In this section, we briefly discuss the experimental results presented earlier. First, we found that the performance of MMFN model on all four data sets was generally better than other models. There are three possible reasons for such a performance improvement: 1) the multiscale model effectively fuses more abundant neighborhood correlation and low-level feature. 2) the middle-level features fusion structure can better exploit strong complementary and related information among multiscale fusion features than a high-level features fusion structure. 3) the residual learning modules can extract more discriminative and higher-level features and make deep learning models much easier to train. As can be seen in Figure 8, a residual-based network model achieved a better performance on all four data sets compared with SAE, CNN and Two-CNN. As can be seen in Figure 8 and Table 8, MMFN achieved higher classification accuracy than the ResNet and SSRN on the four data sets. MMFN made full use of the multiscale middle-level features, and sent the fused features to the residual block instead of learning directly from the original image like ResNet. SSRN also adopted residual connections, and treated spectral features and spatial features separately in two consecutive blocks, however, if the input of the spatial block is based on the spectral block, the spatial learning will miss the spatial information.
Second, two aspects will influence the HSI classification accuracy: 1) the number and spatial size of input into the network; 2) the number of training labeled samples. MMFN uses the suitable number and spatial size of inputs through the experiments and achieves a higher accuracy. Multiscale inputs with relatively larger spatial size contained more useful and abundant information which can boost the classification performance. MMFN performs better with relatively small labeled samples and this network can be generalized to other remote-sensing scenarios because of its deep feature learning capacity.
Finally, the disadvantage of the MMFN model is that the training time is relatively long, which is mainly because the training of the network is divided into two stages, and the multiscale input increases the corresponding time. As can be seen in Table 13, the training time of MMFN is about 1–6 minutes longer than that of SSRN and 2–12 minutes longer than that of CNN, which means that MMFN is more computationally expensive than the SSRN and the CNN. Fortunately, the adoption of GPU has largely alleviated the extra computational costs and reduced the training times.

4. Conclusions

In this paper, a novel MMFN deep learning method is proposed for hyperspectral image classification. Compared with the previous networks, MMFN consists of two independent training stages. The first stage is to fuse the spectral-spatial information and effectively extract the spectral-spatial fusion feature. The second stage fuses the complementary yet correlated information among the regions of different scales and effectively extracts the multiscale fusion feature, which can further improve the classification accuracy. In addition, multiscale deep middle-level feature fusion network works better for spectral–spatial feature learning as compared to single scale network. The multiscale deep middle-level features are fused in the convolutional layer rather than the fully connected layer, which can enable the network to make full use of the strong complementary and related information among multiscale fusion features. Finally, the residual-based network model achieved a better performance on all four data sets. The residual learning modules can extract more discriminative features for classification and can help the network maintain a higher accuracy with deeper layers. The experimental results show the superiority of the proposed method with small labeled samples on the four data sets over other state-of-the-art methods. In future work, we can research a better fusion strategy and fuse different data sets from other sensors to improve the classification accuracy.

Author Contributions

Conceptualization, Z.L.; Methodology, Z.L., L.H. and J.H.; Writing – original draft, Z.L., L.H. and J.H.

Funding

This research was funded by the Natural Science Foundation of Liaoning, grant number 20180550337, the China Postdoctoral Science Foundation, grant number No. 2018M633585, and the Natural Science Basic Research Plan in Shaanxi Province of China, grant number No.2018JQ6060.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Laurin, G.V.; Chan, J.C.W.; Chen, Q.; Lindsell, J.A.; Coomes, D.A.; Guerriero, L.; Del Frate, F.; Miglietta, F.; Valentini, R. Biodiversity mapping in a tropical West African forest with airborne hyperspectral data. PLoS ONE 2014, 9, e97910. [Google Scholar]
  2. Van der Meer, F.D.; Van der Werff, H.M.; Van Ruitenbeek, F.J.; Hecker, C.A.; Bakker, W.H.; Noomen, M.F.; van der Meijde, M.; Carranza, E.J.M.; de Smeth, J.B.; Woldai, T. Multi-and hyperspectral geologic remote sensing: A review. Int. J. Appl. Earth Observ. Geoinf. 2012, 14, 112–128. [Google Scholar] [CrossRef]
  3. Yokoya, N.; Chan, J.C.-W.; Segl, K. Potential of Resolution-Enhanced Hyperspectral Data for Mineral Mapping Using Simulated EnMAP and Sentinel-2 Images. Remote Sens. 2016, 8, 172. [Google Scholar] [CrossRef]
  4. Lu, X.; Li, X.; Mou, L. Semi-Supervised Multitask Learning for Scene Recognition. IEEE Trans. Cybern. 2015, 45, 1967–1976. [Google Scholar] [PubMed]
  5. Zhong, P.; Zhang, P.; Wang, R. Dynamic learning of SMLR for feature selection and classification of hyperspectral data. IEEE Geosci. Remote Sens. Lett. 2008, 5, 280–284. [Google Scholar] [CrossRef]
  6. Khodr, J.; Younes, R. Dimensionality reduction on HSIs: A comparative review based on artificial datas. In Proceedings of the 4th 2011 International Congress on Image and Signal Processing, Shanghai, China, 15–17 October 2011; pp. 1875–1883. [Google Scholar]
  7. Plaza, A.; Martinez, P.; Plaza, J.; Perez, R. Dimensionality reduction and classification of hyperspectral image data using sequences of extended morphological transformations. IEEE Trans. Geosci. Remote Sens. 2005, 43, 466–479. [Google Scholar] [CrossRef]
  8. Lu, X.; Zheng, X.; Yuan, Y. Remote Sensing Scene Classification by Unsupervised Representation Learning. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5148–5157. [Google Scholar] [CrossRef]
  9. Serpico, B.S.; Bruzzone, L. A new search algorithm for feature selection in hyperspectral remote sensing images. IEEE Trans. Geosci. Remote Sens. 2011, 39, 1360–1367. [Google Scholar] [CrossRef]
  10. Tao, C.; Pan, H.; Li, Y.; Zou, Z. Unsupervised spectral-spatial feature learning with stacked sparse autoencoder for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2438–2442. [Google Scholar]
  11. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–12. [Google Scholar]
  12. Persello, C.; Bruzzone, L. Kernel-Based Domain-Invariant Feature Selection in Hyperspectral Images for Transfer Learning. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2615–2626. [Google Scholar] [CrossRef]
  13. Imbiriba, T.; Bermudez, J.C.; Richard, C. Band selection for nonlinear unmixing of hyperspectral images as a maximal clique problem. IEEE Trans. Image Process. 2017, 99, 2179–2191. [Google Scholar] [CrossRef] [PubMed]
  14. Yuan, Y.; Zheng, X.; Lu, X. Discovering Diverse Subset for Unsupervised Hyperspectral Band Selection. IEEE Trans. Image Process. 2016, 26, 51–64. [Google Scholar] [CrossRef]
  15. Damodaran, B.B.; Courty, N.; Lefèvre, S. Sparse Hilbert Schmidt Independence Criterion and Surrogate-Kernel-Based Feature Selection for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2385–2398. [Google Scholar] [CrossRef]
  16. Yuan, Y.; Lin, J.; Wang, Q. Dual-Clustering-Based Hyperspectral Band Selection by Contextual Analysis. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1431–1445. [Google Scholar] [CrossRef]
  17. Jia, S.; Ji, Z.; Qian, Y.; Shen, L. Unsupervised band selection for hyperspectral imagery classification without manual band removal. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2012, 5, 531–543. [Google Scholar] [CrossRef]
  18. Chang, C.I.; Wang, S. Constrained band selection for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1575–1585. [Google Scholar] [CrossRef]
  19. Sun, K.; Geng, X.; Ji, L. A new sparsity-based band selection method for target detection of hyperspectral image. IEEE Geosci. Remote Sens. Lett. 2015, 12, 329–333. [Google Scholar]
  20. Jimenez-Rodrguez, L.O.; Arzuaga-Cruz, E.; Velez-Reyes, M. Unsupervised linear feature-extraction methods and their effects in the classification of high-dimensional data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 469–483. [Google Scholar] [CrossRef]
  21. Zhong, Z.; Fan, B.; Duan, J.; Wang, L.; Ding, K.; Xiang, S.; Pan, C. Discriminant Tensor Spectral–Spatial Feature Extraction for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2017, 12, 1028–1032. [Google Scholar] [CrossRef]
  22. Wong, W.K.; Lai, Z.; Wen, J. Low rank embedding for robust image feature extraction. IEEE Trans. Image Process. 2017, 26, 2905–2917. [Google Scholar] [CrossRef]
  23. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  24. Lu, X.; Yuan, Y.; Zheng, X. Joint Dictionary Learning for Multispectral Change Detection. IEEE Trans. Cybern. 2017, 47, 884–897. [Google Scholar] [CrossRef]
  25. Xu, L.; Wong, A.; Li, F.; Clausi, D.A. Intrinsic Representation of Hyperspectral Imagery for Unsupervised Feature Extraction. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1118–1130. [Google Scholar] [CrossRef]
  26. Zhao, W.; Du, S. Spectral–Spatial Feature Extraction for Hyperspectral Image Classification: A Dimension Reduction and Deep Learning Approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  27. Guo, X.; Huang, X.; Zhang, L.; Zhang, L.; Plaza, A.; Benediktsson, J.A. Support tensor machines for classification of hyperspectral remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3248–3264. [Google Scholar] [CrossRef]
  28. Yuan, S.; Xia, M.; Chen, L. Multilinear Spatial Discriminant Analysis for Dimensionality Reduction. IEEE Trans. Image Process. 2017, 26, 2669–2681. [Google Scholar] [CrossRef]
  29. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  30. Li, Z.; Huang, L.; Zhang, D.; Liu, C.; Wang, Y.; Shi, X. A Deep Network Based on Multiscale Spectral-Spatial Fusion for Hyperspectral Classification. In Proceedings of the International Conference on Knowledge Science, Engineering and Management, Changchun, China, 17–19 August 2018; pp. 283–290. [Google Scholar]
  31. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  32. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  33. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  34. Yang, J.; Zhao, Y.Q.; Chan, C.W. Learning and Transferring Deep Joint Spectral–Spatial Features for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4729–4742. [Google Scholar] [CrossRef]
  35. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral-Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  36. Zhao, W.; Du, S. Learning multiscale and deep representations for classifying remotely sensed imagery. ISPRS J. Photogramm. Remote Sens. 2016, 113, 155–165. [Google Scholar] [CrossRef]
  37. Song, W.; Li, S.; Fang, L.; Lu, T. Hyperspectral Image Classification with Deep Feature Fusion Network. IEEE Trans. Geosci. Remote Sens. 2018, 99, 3173–3184. [Google Scholar] [CrossRef]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 770–778. [Google Scholar]
  39. Cristianini, N.; Shawetaylor, J. An Introduction to Support Vector Machines and Other Kernel-based Learning Methods. Kybernetes 2001, 32, 1–28. [Google Scholar]
Figure 1. The structure of residual block.
Figure 1. The structure of residual block.
Remotesensing 11 00695 g001
Figure 2. The framework of first training stage in a multiscale deep middle-level feature fusion network (MMFN).
Figure 2. The framework of first training stage in a multiscale deep middle-level feature fusion network (MMFN).
Remotesensing 11 00695 g002
Figure 3. The framework of second training stage in MMFN.
Figure 3. The framework of second training stage in MMFN.
Remotesensing 11 00695 g003
Figure 4. (a) False-color image of the Indian Pines data, (b) Ground truth of the Indian Pines data.
Figure 4. (a) False-color image of the Indian Pines data, (b) Ground truth of the Indian Pines data.
Remotesensing 11 00695 g004
Figure 5. (a) False-color image of the Pavia University data, and (b) Ground truth of the Pavia University data.
Figure 5. (a) False-color image of the Pavia University data, and (b) Ground truth of the Pavia University data.
Remotesensing 11 00695 g005
Figure 6. (a) False-color image of the Ground-truth map of the KSC data, and (b) Ground truth of the KSC data.
Figure 6. (a) False-color image of the Ground-truth map of the KSC data, and (b) Ground truth of the KSC data.
Remotesensing 11 00695 g006
Figure 7. (a) False-color image of the Salinas data, (b) Ground truth of the Salinas data.
Figure 7. (a) False-color image of the Salinas data, (b) Ground truth of the Salinas data.
Remotesensing 11 00695 g007
Figure 8. The Overall Accuracy of changing the percentage of training samples by all methods on the four data sets. (a) Overall Accuracy on IN. (b) Overall Accuracy on UP. (c) Overall Accuracy on KSC. (d) Overall Accuracy on Salinas.
Figure 8. The Overall Accuracy of changing the percentage of training samples by all methods on the four data sets. (a) Overall Accuracy on IN. (b) Overall Accuracy on UP. (c) Overall Accuracy on KSC. (d) Overall Accuracy on Salinas.
Remotesensing 11 00695 g008aRemotesensing 11 00695 g008b
Table 1. Land cover classes and numbers of samples in the Indian Pines data set.
Table 1. Land cover classes and numbers of samples in the Indian Pines data set.
No.Class NameNumbers of Samples
1Alfalfa46
2Corn-notill1428
3Corn-mintill830
4Corn237
5Grass-pasture483
6Grass-tree730
7Grass-pasture-mowed28
8Hay-windrowed478
9Oats20
10Soybean-notill972
11Soybean-mintill2455
12Soybean-clean593
13Wheat205
14Woods1265
15Buildings-Grass-Trees-Drives386
16Stone-Steel-Towers93
Total10,249
Table 2. Land cover classes and numbers of samples in the Pavia University data set.
Table 2. Land cover classes and numbers of samples in the Pavia University data set.
No.Class NameNumbers of Samples
1Asphalt6631
2Meadows18,649
3Gravel2099
4Trees3064
5Painted metal sheets1345
6Bare Soil5029
7Bitumen1330
8Self-Blocking Bricks3682
9Shadows947
Total42,776
Table 3. Land cover classes and numbers of samples in the Kennedy Space Center (KSC) data set.
Table 3. Land cover classes and numbers of samples in the Kennedy Space Center (KSC) data set.
No.Class NameNumbers of Samples
1Scrub761
2Willow swamp243
3CP hammock256
4Slash pine252
5Oak/Broadleaf161
6Hardwood229
7Grass-pasture-mowed105
8Graminoid marsh431
9Spartina marsh520
10Cattail marsh404
11Salt marsh419
12Mud flats503
13Water927
Total5211
Table 4. Land cover classes and numbers of samples in the Salinas data set.
Table 4. Land cover classes and numbers of samples in the Salinas data set.
No.Class NameNumbers of Samples
1Brocoli_green_weeds_12009
2Brocoli_green_weeds_23726
3Fallow1976
4Fallow_rough_plow1394
5Fallow_smooth2678
6Stubble3959
7Celery3579
8Grapes_untrained11,271
9Soil_vinyard_develop6203
10Corn_senesced_green_weeds3278
11Lettuce_romaine_4wk1068
12Lettuce_romaine_5wk1927
13Lettuce_romaine_6wk916
14Lettuce_romaine_7wk1070
15Vinyard_untrained7268
16Vinyard_vertical_trellis1807
Total54,129
Table 5. Classification results (OA%) of our methods on the four data sets with different input sizes.
Table 5. Classification results (OA%) of our methods on the four data sets with different input sizes.
Spatial SizeINUPKSCSalinas
3 × 378.6494.2486.3495.17
5 × 584.2596.4990.2996.29
7 × 787.3298.5395.2197.42
9 × 990.4299.1795.8998.27
11 × 1194.1599.3996.7699.24
13 × 1394.8799.4096.9399.28
15 × 1594.9499.4397.1499.31
Table 6. Classification results (OA%) of our methods on the four data sets with different input sizes when the percentage of training samples is changing.
Table 6. Classification results (OA%) of our methods on the four data sets with different input sizes when the percentage of training samples is changing.
Data SetSize of Input3%4%5%6%
IN7 × 786.5786.8987.3291.24
9 × 986.9488.1490.4293.6
11 × 1188.0691.2694.1594.63
7 × 7, 9 × 9,11 × 1191.8193.8796.5997.16
UP7 × 796.0497.3898.5398.87
9 × 996.1798.4899.1799.24
11 × 1198.6898.8899.3999.43
7 × 7, 9 × 9,11 × 1199.499.7399.7599.84
KSC7 × 792.2193.0495.2195.86
9 × 992.7994.0695.8996.52
11 × 1194.6295.4896.7697.48
7 × 7, 9 × 9,11 × 1197.1598.2498.5098.97
salinas7 × 794.7595.7897.4297.7
9 × 996.5498.0498.2798.62
11 × 1197.2599.299.2499.33
7 × 7, 9 × 9,11 × 1198.3799.6599.6999.73
Table 7. The number of residual blocks selection experiment.
Table 7. The number of residual blocks selection experiment.
Data SetThe Number of Residual Block3%4%5%6%
INno residual block88.4490.2893.3795.59
one residual block91.2392.6594.1796.71
two residual block91.8193.8796.5997.16
three residual block91.3493.1695.5196.66
UPno residual block99.0799.2599.3299.49
one residual block99.5999.6499.6799.72
two residual block99.4099.7399.7599.84
three residual block99.1299.6899.7199.86
KSCno residual block93.6596.2197.0297.95
one residual block97.0698.0398.1298.61
two residual block97.1598.2498.598.97
three residual block96.9197.9497.9898.44
salinasno residual block96.2696.8997.3198.28
one residual block97.7998.6899.2899.31
two residual block98.3799.6599.6999.73
three residual block98.2599.3699.4499.75
Table 8. Classification accuracies (%) of different methods on the four data sets.
Table 8. Classification accuracies (%) of different methods on the four data sets.
Data SetMethodOAAA
INTwo-CNN76.78 ± 0.4775.23 ± 0.56
SAE74.58 ± 0.5675.27 ± 0.62
SVM77.01 ± 0.5468.73 ± 0.53
CNN77.89 ± 0.4377.14 ± 0.36
ResNet86.25 ± 0.3785.93 ± 0.42
SSRN94.97 ± 0.3194.52 ± 0.22
MMFN96.59 ± 0.2696.13 ± 0.25
UPTwo-CNN94.63 ± 0.2793.31 ± 0.22
SAE92.27 ± 0.3592.58 ± 0.29
SVM94.83 ± 0.2893.37 ± 0.32
CNN96.89 ± 0.1796.75 ± 0.25
ResNet97.48 ± 0.2097.03 ± 0.23
SSRN99.42 ± 0.2699.05 ± 0.32
MMFN99.75 ± 0.1999.68 ± 0.23
KSCTwo-CNN83.47 ± 0.3884.32 ± 0.43
SAE90.72 ± 0.3589.51 ± 0.38
SVM84.37 ± 0.4182.18 ± 0.37
CNN88.65 ± 0.3786.39 ± 0.42
ResNet93.29 ± 0.3192.89 ± 0.36
SSRN97.68 ± 0.4397.17 ± 0.36
MMFN98.50 ± 0.2998.32 ± 0.31
salinasTwo-CNN91.38 ± 0.3689.74 ± 0.43
SAE93.69 ± 0.4193.21 ± 0.47
SVM92.73 ± 0.4491.58 ± 0.48
CNN91.17 ± 0.3792.16 ± 0.40
ResNet95.82 ± 0.3094.56 ± 0.34
SSRN99.22 ± 0.2699.31 ± 0.33
MMFN99.69 ± 0.3199.43 ± 0.31
Table 9. Class-specific accuracies (%) of the Indian Pines Data Set.
Table 9. Class-specific accuracies (%) of the Indian Pines Data Set.
ClassTraining/TestTwo-CNNSAESVMCNNResNetSSRNMMFN
12/4473.8875.2530.0583.5483.2492.594.63
271/135757.7168.6752.1166.4863.1996.4595.3
342/78860.2862.5160.1765.8585.3498.695.11
412/22580.9262.4871.2285.4789.1593.1399.28
524/45974.0981.4269.1985.4694.9499.2798.8
637/65777.4681.7688.6381.5494.3499.8498.73
71/2780.073.6440.2176.4970.560.3282.77
824/45486.4585.3881.5482.6795.5100100
91/1910078.2649.4681.4390.594.36100
1049/92362.2267.5776.9271.6285.390.7195.68
11123/233267.8172.5276.9262.6994.4795.598.41
1230/56359.0971.3557.4465.8365.896.3798.54
1310/19586.4983.6492.8284.3894.8910096.62
1463/120282.7780.3596.9283.5194.6799.999.53
1519/36788.4683.7580.9286.9495.295.3797.59
165/8866.1975.8976.1370.3677.9110087.23
OA-76.7874.5877.0177.8986.2594.9796.59
Kappa-0.71550.72340.73650.76280.84250.93920.9458
AA-75.2375.2768.7377.1485.9394.5296.13
Table 10. Class-specific accuracies (%) of the UP Data Set.
Table 10. Class-specific accuracies (%) of the UP Data Set.
ClassTraining/TestTwo-CNNSAESVMCNNResNetSSRNMMFN
1332/629993.2794.5993.3698.4599.8210099.67
2932/1771797.1496.4494.6796.3799.6299.9899.96
3105/199485.4684.5783.9296.2194.8894.9799.88
4153/291198.5897.3797.5897.5896.2498.9299.22
567/127899.5699.699.9296.369599.82100
6251/477890.5893.3993.8896.5199.8398.7499.72
767/126388.3488.5794.6797.2399.3100100
8184/349887.1785.6693.5793.4793.7399.0298.76
947/90099.7493.0494.0598.5794.87100100
OA-94.6392.2794.8396.8997.4899.4299.75
Kappa-0.92160.92350.93140.96470.98530.99430.9972
AA-93.3192.5893.3796.7597.0399.0599.68
Table 11. Class-specific accuracies (%) of the KSC Data Set.
Table 11. Class-specific accuracies (%) of the KSC Data Set.
ClassTraining/TestTwo-CNNSAESVMCNNResNetSSRNMMFN
138/72390.1293.1499.1792.45100100100
212/23191.2892.0489.6190.4895.4699.4999.47
312/24490.2285.4990.7687.1694.6310098.29
413/23980.872.0263.5980.3891.4491.5498.7
58/15375.5582.359.6380.9678.9185.7889.46
611/21891.6683.1567.4383.2190.9110096.96
75/10080.5976.4658.3183.2683.1192.6393.69
822/40993.9994.189.6688.2694.297.1299.46
926/49487.9194.5790.3989.0494.0198.67100
1020/38471.0298.9189.8484.8595.5599.12100
1121/39871.6298.1493.9683.8994.7396.7398.26
1225/47889.1896.4284.3190.4398.12100100
1346/88182.2497.8391.7188.796.55100100
OA-83.4790.7284.3788.6593.2997.6898.5
Kappa-0.81040.91210.81620.86570.93910.95340.9737
AA-84.3289.5199.1786.3992.8997.1798.32
Table 12. Class-specific accuracies (%) of the Salinas Data Set.
Table 12. Class-specific accuracies (%) of the Salinas Data Set.
ClassTraining/TestTwo-CNNSAESVMCNNResNetSSRNMMFN
1100/190993.9492.7494.7394.67100100100
2186/354093.4695.1394.3295.5296.27100100
399/187790.993.6290.4892.7493.8310099.67
470/136490.5895.688.3994.6192.7399.6898.77
5134/254492.594.3995.2794.69100100100
6198/376191.8195.7393.7594.8296.4899.97100
7179/340094.1496.4795.6697.2894.37100100
8564/1070773.5791.5884.3887.6993.3499.8699.31
9310/589389.7595.4694.3790.6194.8799.84100
10164/311488.7785.2787.4686.5494.0199.8699.83
1153/101586.9292.5293.7589.4892.2497.9196.48
1296/183193.3593.7494.3192.7494.8899.7799.94
1346/87092.4192.7393.6490.6892.92100100
1454/101692.7195.2694.3193.1794.79100100
15363/690577.7886.4776.7585.7289.8691.9199.43
1690/171793.3294.7193.7693.7492.3610099.84
OA-91.3893.6992.7391.1795.8299.2299.69
Kappa-0.90830.91710.91330.90580.96060.97330.9961
AA-89.7493.2191.5892.1694.5699.3199.43
Table 13. Comparison of training and test time for each method on four data sets.
Table 13. Comparison of training and test time for each method on four data sets.
INUPKSCSalinas
Two-CNNTrain (m)3.613.12.112.7
Test (s)5.215.62.116.9
SAETrain (m)3.512.81.712.3
Test (s)3.815.41.618.4
SVMTrain (m)1.68.00.88.4
Test (s)1.49.70.810.3
CNNTrain (m)3.614.91.913.1
Test (s)4.818.62.027.2
ResNetTrain (m)4.016.32.214.2
Test (s)4.718.42.326.5
SSRNTrain (m)5.018.52.619.4
Test (s)6.324.73.236.4
MMFNTrain (m)8.223.53.825.4
Test (s)6.725.63.537.6

Share and Cite

MDPI and ACS Style

Li, Z.; Huang, L.; He, J. A Multiscale Deep Middle-level Feature Fusion Network for Hyperspectral Classification. Remote Sens. 2019, 11, 695. https://doi.org/10.3390/rs11060695

AMA Style

Li Z, Huang L, He J. A Multiscale Deep Middle-level Feature Fusion Network for Hyperspectral Classification. Remote Sensing. 2019; 11(6):695. https://doi.org/10.3390/rs11060695

Chicago/Turabian Style

Li, Zhaokui, Lin Huang, and Jinrong He. 2019. "A Multiscale Deep Middle-level Feature Fusion Network for Hyperspectral Classification" Remote Sensing 11, no. 6: 695. https://doi.org/10.3390/rs11060695

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop