Next Article in Journal
Comparison of Remote Sensing Techniques for Geostructural Analysis and Cliff Monitoring in Coastal Areas of High Tourist Attraction: The Case Study of Polignano a Mare (Southern Italy)
Previous Article in Journal
MDPrePost-Net: A Spatial-Spectral-Temporal Fully Convolutional Network for Mapping of Mangrove Degradation Affected by Hurricane Irma 2017 Using Sentinel-2 Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Local to Global: Class Feature Fused Fully Convolutional Network for Hyperspectral Image Classification

1
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
2
School of Engineering and Information Technology, The University of New South Wales, Canberra, ACT 2600, Australia
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(24), 5043; https://doi.org/10.3390/rs13245043
Submission received: 22 October 2021 / Revised: 29 November 2021 / Accepted: 9 December 2021 / Published: 12 December 2021
(This article belongs to the Section AI Remote Sensing)

Abstract

:
Current mainstream networks for hyperspectral image (HSI) classification employ image patches as inputs for feature extraction. Spatial information extraction is limited by the size of inputs, which makes networks unable to perform effective learning and reasoning from the global perspective. As a common component for capturing long-range dependencies, non-local networks with pixel-by-pixel information interaction bring unaffordable computational costs and information redundancy. To address the above issues, we propose a class feature fused fully convolutional network (CFF-FCN) with a local feature extraction block (LFEB) and a class feature fusion block (CFFB) to jointly utilize local and global information. LFEB based on dilated convolutions and reverse loop mechanism can acquire the local spectral–spatial features at multiple levels and deliver shallower layer features for coarse classification. CFFB calculates global class representation to enhance pixel features. Robust global information is propagated to every pixel with low computational cost. CFF-FCN considers a fully global class context and obtains more discriminative representation by concatenating high-level local features and re-integrated global features. Experimental results conducted on three real HSI data sets demonstrate that the proposed fully convolutional network is superior to multiple state-of-the-art deep learning-based approaches, especially in the case of a small number of training samples.

Graphical Abstract

1. Introduction

Hyperspectral image (HSI) classification is an analysis technique that characterizes ground objects by abundant spectral and spatial information. Its main task is to assign a class label to each pixel of HSIs and generate specific classification maps. Motivated by the rapid development of deep learning in computer vision, natural language processing and other fields, classification methods based on deep neural networks have become a research hotspot in HSI domain. Unlike traditional methods that rely on expert experience and parameter adjustment to manually design feature, deep learning methods extract highly robust and discriminative semantic features from hyperspectral data in a data-driven and automatic parameter tuning manner. Generally speaking, shallower layers capture structured features, such as texture and edge information, and deeper layers acquire more complex semantic features.
Typical deep learning architectures, including deep belief networks (DBNs), stacked auto-encoders (SAEs), recurrent neural networks (RNNs), and convolutional neural networks (CNNs) have been widely used for HSI classification. In [1], Chen et al. introduced deep neural network to hyperspectral image classification for the first time, and proposed a joint spectral–spatial classification method based on deep stack auto-encoder. It employs SAEs to learn deep spectral–spatial features in neighborhoods and sends them to a logistic regression (LR) model to obtain classification results. This method shows the great potential of the deep learning framework for HSI classification. In HSI data analysis, DBNs, as a variant of auto-encoders (AEs), perform feature extraction in an unsupervised manner [2]. For example, Liu et al. [3] proposed an effective classification framework based on DBNs and active learning to extract deep spectral features and iteratively select high-quality labeled samples as training samples. As a method of processing sequence information, RNNs are extremely suitable for analyzing hyperspectral data, each pixel of which can be regarded as a continuous spectral vector. Reference [4] applied the concept of RNN to spectral classification by developing a new activation function and an improved gated recurrent unit (GRU) to effectively analyzing hyperspectral pixels as sequence data. The above methods only accept vector inputs, which leads to the inability to preserve the spatial structure characteristics of HSIs. In contrast, CNNs process spectral and spatial features flexibly, and become the most popular deep learning model for HSI classification.
According to the types of extracted features, CNN-based models can be divided into spectral CNNs and spectral–spatial CNNs. Spectral CNNs are designed to extract deep features in the spectral domain. For example, Hu et al. proposed a deep learning model composed of 1D CNNs [5], which extract deep features and implement the classification in the spectral domain, and achieved excellent performance that is superior to support vector machines (SVM). Li et al. presented a 1D CNN models to capture pixel pair features (PPFs), which employs a combination of the center pixel and neighborhood pixels as the input and explores the spectral similarity between pixels [6]. The input of pixel pairs significantly increases the training samples and relieves the pressure of network training. In addition, Wu et al. [7] proposed a combination model of 1D CNN and RNN, exploiting the deep spectral features extracted by 1D CNN as the input of RNN to further improve the discrimination of features. Previous studies on HSI classification have proved that spatial information contributes to further improve the classification accuracy. It is due to the spatial consistency of the ground truth distribution.
The spectral–spatial CNNs take spatial neighbor patches as inputs and employ 2D or 3D CNN to extract joint deep spectral–spatial features. Spectral-spatial CNNs with stronger representation capability and classification performance than spectral CNNs, become more popular by researchers. Aptoula et al. combined powerful spatial-spectral attribute profiles (APs) and a CNN with nonlinear representation capacity to achieve remarkable classification performances [8]. He et al. exploited multi-scale covariance maps obtaining by local spectral–spatial information of different window sizes to train a CNN model [9]. Besides using handcrafted spectral–spatial features training networks to implement spectral–spatial CNNs, combining CNNs with other models is an alternative strategy to extract spectral and spatial features separately. For example, Zhao et al. proposed a classification method based on spectral–spatial features (SSFC), which designs a balanced local discriminant embedding (BLDE) algorithm to extract spectral features and uses 2D CNN to learn spatial features [10]. In [11], a two-branch network composed of 1D CNN and 2D CNN was proposed to extract spectral and spatial information respectively. References [12,13] also proposed two-branch classification networks based on 3D CNN and attention mechanism, which capture spectral and spatial features and utilize the attention mechanism to further refine feature maps. Further, some studies attempt to extract spectral–spatial features simultaneously. HSIs are 3D data cubes composed of 1D spectral dimensions and 2D spatial dimensions. The application of 3D CNN is a natural fashion to acquire high-level spectral–spatial features for effective classification without any pre- and post-processing procedures. For example, Chen et al. first introduced 3D CNN for HSI classification to achieve the effective spectral–spatial feature extraction [14]. Reference [15] presented an attention-based adaptive spectral–spatial kernel improved residual network (A2S2K-ResNet) to achieve the automatic selection of spectral–spatial 3D convolutional kernels and the adaptive adjustment of receptive filed. However, in the face of limited labeled samples, it is challenging to train deeper 3D-CNNs. To this end, the HybridSN model combining 3D CNN and 2D CNN was proposed to learn spectral–spatial features by using 3D CNN and 2D CNN in turns [16]. Compared with using 3D CNN alone, the hybrid model reduces the complexity. Zhang et al. also proposed a 3D lightweight convolutional network to alleviate the small sample problem [17].
The above patch-based CNNs can take full advantage of spectral–spatial information. However, these methods have unavoidable redundant computations caused by overlapped areas of adjacent pixels and limited receptive fields. To address the above problems, researchers try to introduce the fully convolution framework into the field of hyperspectral classification. The fully convolutional network (FCN) proposed by Lang et al. provides a novel idea to get rid of fixed input size and limitation of the receptive fields for pixel-level tasks [18]. Replacing the fully connected layer (FC) 1 × 1 convolutional layer as a classifier, FCN can accept inputs of any size and produce corresponding outputs through effective reasoning and learning. In [19], a multi-scale spectral–spatial CNN (HyMSCN) was proposed to expand the receptive field and extract multi-scale spatial features by dilated convolutions of different sizes. Zheng et al. presented a fast patch-free global learning framework (FPGA), which contains an encode-decoder-based FCN and adopts a global random hierarchical sampling (GS2) strategy to ensure fast and stable convergence [20]. To solve the sample problem of insufficiency and imbalance, Zhu et al. designed a spectral–spatial dependent global learning (SSDGL) framework based on FPAG [21]. Although the above methods all employ FCN as the basis for image-based classification and achieve significant classification performances, they still exist limitations on the extraction of global information. Xu et al. proposed a two-branch spectral–spatial fully convolutional networks (SSFCN-CRF), which extracts spectral and spatial features separately, and introduces the dense conditional random field (dense CRF) to further balance the local and global information [22]. However, it is unable to achieve end-to-end training and difficult to optimize globally.
From the previous analysis, there are some problems that limit the improvement of hyperspectral classification performance. First, the input size of patch-based frameworks limits the learning of local information. Generally, a larger input size will make the network learn more robust features and improve the classification accuracy. At the same time, a large number of overlapping areas between different input patches lead higher calculation costs. Second, current networks mainly rely on local information. When the number of training samples is small, the structured neighborhood features provided by the samples are limited, resulting in weak classification performance. Many complex spatial structures cannot be learned by the network. Pixels located at the boundary of different classes are easily misclassified. Third, most networks ignore the global representation of HSIs and lack the ability to adaptively capture global information.
To solve the above problems, we propose a fully convolutional network based on class feature fusion (CFF-FCN) containing the local feature extraction block (LFEB) and the class feature fusion block (CFFB), which receives the whole HSI as the input and takes the local spatial structures and the global context into full consideration at the low computational cost. Specifically, LFEB based on dilated convolutions and reverse loop mechanism is designed to extract multi-level spectral–spatial information. Inspired by the idea of coarse-to-fine segmentation [23,24], the coarse classification is utilized to obtain the key class representations for CFFB. CFFB calculates the affinity matrix by the similarity between each pixel and each class feature, which are served to reconstruct the more robust global features. The global class features obtained by CFFB are aggregated with the local features further refined by reverse loop mechanism to enhance the discrimination of classification features. The main contributions of our study can be summarized as follows.
  • A coarse-to-fine CFF-FCN with LEFB and CFFB is proposed for HSI classification, which is based on the fully convolutional layer to avoid data preprocessing and redundant calculations. The LFEB with the reverse loop mechanism not only retains low-level detailed information but also extracts high-level abstract features to provides semantic information of multiple levels and local features of multiple receptive fields and further improves the discrimination of features.
  • The proposed CFFB can capture the global class information by coarse classification and propagate it to each pixel according to the similarity. Compared with the pixel-by-pixel global calculation, the CFFB allows several key class features to update the pixel features that reduces computational costs and redundant information and enhances the representation ability of classification features.
  • Experiments on three real hyperspectral data sets demonstrate that the proposed CFF-FCN achieves the most reliable and robust classification with fewer network parameters and lower computational time, and improves the misclassification phenomenon of category boundaries.
The remainder of this paper is organized as follows. Section 2 briefly reviews the related works. Section 3 describes the proposed framework. Section 4 and Section 5 provides the details and the discussion of the experimental results. Finally, Section 6 concludes this paper.

2. Related Works

2.1. Dilated Convolution

Dilated convolution is a generalization of Kronecker-factored Convolutional Filters [25] that can expand the receptive field while maintaining the resolution. The dilation rate d is a parameter controlling the dilation scale of convolution kernel. The dilated convolution is equivalent to extending the filter of the standard convolution in the light of dilation rate and filling the empty elements with zero. In other words, the standard convolution can be seen as a dilated convolution with the dilation rate of 1. When the dilation rate is greater than 1, the convolution layer can extract non-adjacent pixel information. Assuming convolutional kernels of size k × k , the size of corresponding dilated convolutional filter is k d × k d , where k d = k + ( k 1 ) · ( d 1 ) . Figure 1 shows an example of dilated convolutional filter. The stacking dilated convolution layers enlarge the receptive field exponentially, while the number of parameters grows linearly [26].
Dilated convolutions aims to capture long-range information. However, adopting dilated convolutions with large dilation rates may lose local information. In particular, neighbor pixels are highly correlated on account of HSIs with spatial consistency. Meanwhile, continuous stacking of multiple dilated convolutions with the same dilation rate will cause a gridding effect [27]. Due to the discontinuity of dilated convolution kernels, not all pixels in the coverage of convolution kernels participate in the calculation. The continuity of local information is destroyed, which is unfavorable for pixel-level dense prediction tasks. Therefore, how to design a dilated convolutional network to balance the relationship between local information and long-range information and ensure the continuity of information deserves attention. Wang et al. proposed a strategy: using the hybrid dilated convolutions (HDC) with incremental dilation rates to avoid gridding effect while ensuring that the receptive field remains unchanged [27].

2.2. Non-Local Module

As the attention mechanism has successively gained a dominant position in the fields of natural language processing and computer vision, more and more researchers are focusing on how to capture critical features of the current task from abundant information. One of the most concerned works is the non-local neural network that can capture long-range dependencies by learning the similarity relationship between any two pixels in the image to maintain global information [28]. Although the stacking of convolutional layers can enlarge the receptive field, the scope of image information extraction is still limited to local neighborhoods rather than whole image. In order to capture long-range dependencies, Wang et al. proposed a non-local operation to calculate the global response at any position, which obtains the non-local response value by weighted summation. The core operations of non-local modules can be expressed as:
y i = 1 C ( x ) j f x i , x j g x j ,
where i and j are position indexes of pixels and x i is the input feature of the position i. The function f computes the similarity of x i and x j . The unary function g calculates the pixel representation of the position j. C ( x ) denotes a normalization factor. Only considering g as a linear function, g can be written as:
g x j = W g x j ,
where W g , as a learnable weigh matrix, can be implemented by 1 × 1 convolution operation. There are many options for the relation estimation, which are described in detail in [28]. Here, the Embedded Gaussian form is chosen to calculate similarity in the embedded space. The specific form is as follows:
f x i , x j = e θ x i T ϕ x j ,
where θ ( x i ) = W θ x i and ϕ ( x j ) = W ϕ x j are transformation functions implemented by 1 × 1 convolution. C ( x ) is set as j f x i , x j . Therefore, Equation (1) can be rewritten as:
y = softmax x T W θ T W ϕ x x T W g T .
To insert non-local operation into the existing structures, the residual connection is incorporated into the non-local module, which is defined as:
z = W z y + x .
z is the output of the non-local block, and W z is a 1 × 1 convolution to make the dimension of non-local information consistent with the input. The structure of a nonlocal module can be represented in Figure 2.
Non-local blocks directly capture long-range dependencies by calculating the relation between any two positions. Compared with convolution operation, the perceptive scope of non-local block is no longer limited to adjacent pixels. Meanwhile, the non-local block as a flexible unit can be combined with other existing networks to build a more powerful hierarchical structure that aggregates non-local and local information. However, non-local operation mainly relies on the weighted average of the affinity matrix and the corresponding features. The specific implementation includes a large number of matrix multiplications and linear transformations. As the image size increases, the demand for computing resources will also increase drastically. Some studies try to apply non-local networks in the field of HSI classification. Sun et al. designed an end-to-end spectral–spatial classification network with non-local modules, which extracts non-local information to improve classification performance of HSI classification [29]. However, it is still a patch-based method, and needs expensive computational costs to learn the non-local features. Wang et al. combined the FCN and the non-local module to utilize long-range dependents of HSI [30]. However, this method only accepts 96 × 96 sub-images as input, resulting in the long-range information to remain locally restricted.
According to the distribution characteristics of land cover in HSI, the pixels of the same category are often distributed continuously in space or dominate in a certain region. Pixels that are not spatially adjacent may also have a high correlation. However, not all pixels are highly correlated with the current pixel. Extracting global information in a pixel-by-pixel way may introduce a large number of irrelevant features disturbing classification. Therefore, we design a deep FCN classification framework to capture contextual information with more efficient and representative class features while considering local and global information.

3. Methodology

The architecture of CFF-FCN is shown in Figure 3. Assuming the HSI X R N × H × W × B as the input of CFF-FCN, the output Y R N × H × W × C can be obtain that indicates the class probabilities of each pixel. Here, N, H, W, B and C represent the batch size, height, width, number of spectral bands and number of land-cover classes, respectively. Since HSI classification is the pixel-level task based on single image, the batch size N of CFF-FCN is 1. In Figure 3, the original HSI after dimensionality reduction by 1 × 1 C o n v f i r s t is send to LFEB and CFFB to extract the local and global information, which are aggregated and delivered to the classifier consisting of the convolutional classification layer C o n v c l a s s and softmax layer. The following subsections describe the details of each block.

3.1. Local Feature Extraction Block

As the backbone of CFF-FCN, LFEB is a two-stage network based on dilated convolutions with dense connections and reverse loop mechanism. The stacking of dilated convolutions with different dilation rates fully considers the local information of different scales. Dense connections ensure the maximum information transmission in the network [31]. Let X L F E be the input of LFEB, which contains five dilated convolutions with convolution kernels of 3 × 3 . The specific architecture is displayed in Figure 4. To ensure the continuity of local information while avoiding gridding effect, we employ a set of incremental dilation rates that set as d = [ 1 , 2 , 3 , 4 , 5 ] . The nonlinear transformation H d ( · ) of each convolutional layer is composed of instance normalization (IN) [32], rectified linear unit (ReLU) and convolution (Conv).
The network structure of the multi-layer stack yields the deviation of the data distribution. To reduce the impact of the distribution changes, normalization strategy is usually used in neural networks to map the data to a certain interval. Neural networks taking patches as input generally use batch normalization (BN) [33] to normalize data in a whole batch of images. While the batch size of the FCN-based classification network is 1, BN is not applicable. Therefore, we use IN for normalization in the channel dimension. Normalization operation is performed within single sample. The specific operation can be expressed as:
Z ˜ = γ Z μ ( Z ) σ ( Z ) + β ,
where Z R 1 × H × W × K represents the tensor to be normalized, K is the number of channels. Z ˜ denotes the normalized result. γ and β are parameters to be learned. μ ( Z ) and σ ( Z ) represent the mean and variance of Z, which are calculated as follows:
μ k ( Z ) = h = 1 H w = 1 W Z k h w ,
σ k ( Z ) = 1 H W h = 1 H w = 1 W Z k h w μ k ( Z ) 2 + ε .
Here, h, w, and k represent the index subscript of the spatial location and the channel dimension, respectively. IN can be seen as an adjustment of global information to ensure fast convergence of networks.
In the first stage (Stage-I), each layer takes the outputs of all previous layers as input, and its output is also directly delivered to all subsequent layers. Let X i I be the output of i-th convolutional layer in Stage I, the calculation of LFEB can be represented as:
X i I = H i X L E F , X 1 I , , X i 1 I .
[ · ] denotes the cascade operation of features. For each dilated convolutional layer, the corresponding maximum and minimum receptive field can be calculated as:
R F m a x i = R F max i 1 + k i 1 · s t r i d e ,
R F m i n i = k i · s t r i d e .
s t r i d e is set to 1. Receptive fields cover 3 × 3 to 31 × 31 after Stage-I. In Stage-I, each layer is directly connected to the subsequent layers. The network learns and retains richer features by feature reuse at different levels, which enhances the feature transfer from shallow to deep. At the same time, the implementation of multi-scale receptive fields can adaptively capture and fuse semantic information in different scale, which is conducive to retaining local details and obtaining more complex and robust feature maps. The output of each layer is aggregated as the first stage feature of LFEB:
F e a t s I = i = 1 5 X i I .
means continuous cascade operation for feature maps.
Inspired by the alternate update clique [33], we introduce the loop mechanism based on parameter reuse into LFEB and further refine F e a t s I on the basis of the first-stage network, which modulate low-level features with high-level semantic information. The specific unfold of LFEB is shown in Figure 4. The second stage network (Stage-II) uses the latest output results of other layers to update weights of the current layer. The output of Stage-II can be expressed as:
X i I I = H i d = 1 i 1 X d I I , d = i + 1 5 X d I .
In Stage-II, second stage features of all the previous layers and first stage features of all subsequent layers are concatenated to update the current layer. The higher-level semantic information is fed back into low-level layers to refine features of each layer in an alternate update manner.
Table 1 displays the forward propagation of LEFB. W i j denotes parameters of convolution from X i to X j . Stage-I can be regarded as a layer-by-layer initialization process. After Stage-II, any layer has two-way connections to other layers. The update layer receives the most advanced feedback information.It realizes top-down refinement. The feedback structure maximizes communication between all layers. Outputs of all layers in Stage-II are connected as the second feature of LFEB:
F e a t s I I = i = 1 5 X i I I .
Forward propagation networks usually employ the highest-level semantic features as the classification feature. In fact, fine-grained features captured by low-level layers can provide more structural information for classification. In order to take full advantage of different levels of features, we deliver F e a t s I into CFFB to extract global class information, which is aggregated with F e a t s I I to generate the final classification feature.

3.2. Class Feature Fusion Block

Non-local neural networks explore the similarity relationship between all pixels and update pixel features with global information. Taking into account the characteristics of HSIs, the pixel-by-pixel global information not only contains a large amount of highly correlated redundant information, but also brings high calculation cost. In response to this problem, we design CFFB to reconstruct pixel information by more representative global class context features, which reduces the computational cost while improving the discrimination of pixel features. The specific structure is shown in Figure 5, which mainly includes two parts: class feature extraction and pixel feature fusion.
The most intuitive way to obtain the class representation in HSI is to take the average of all pixel features of the current category as the class center. In actual classification, the labels of all pixels is not available in advance. To this end, we employ a coarse classification mechanism to classify by the intermediate features and obtain the class probability of each pixel in a supervised manner. Through the weighted average of the pixel features and the corresponding class probability, the approximate class center is obtained as the class representation. Let the input of CFFB be X C F R R N × H × W × L , where L is the number of input channels. The class probabilities of pixels on the spatial position i ( 1 i H W ) is calculated as follows:
p i = softmax f c l s x i .
x i represents the input feature of pixel i. f c l s denotes the classifier that maps the input from R K 0 to R C . The class representation of the c-th class can be defined by the weighted average with features of all pixels and their probability belonging to the c-th class:
F c = Σ i = 1 H W p i c f p x i i = 1 H W p i c ,
where p i c is the c-th element of p i , and f p aims to calculate the pixel feature. f c l s = W c l s x i and f p = W p x i are implemented by 1 × 1 convolution. If the coarse classification result shows that the pixel belongs to the c-th class, the pixel has the largest contribution to the c-th class representation. Class features allow the network to capture the overall representation of each class from a global view. Coarse classification provides powerful supervision information in training phase to learn more discriminative features of each class.
Pixel feature fusion aims to learn relations between pixels and class representations, and exploit global class context information to enhance the features of each pixel. We utilize the relationship estimation method of non-local block and calculate relations between each pixel and each class (affinity matrix):
A i c = e θ x i T ϕ F c ,
where A R H W × C is affinity matrix, and A i c is an element of A that indicates the contribution degree of the class feature F c to the reconstructed pixel i. θ ( x i ) = W θ x i and ϕ ( F c ) = W ϕ F c are 1 × 1 convolutions. The global estimation of the pixel i can be expressed as:
y i = c = 1 C A i c g F c c = 1 C A i c ,
g ( F c ) = W g F c is a class feature transform function implemented by 1 × 1 convolution. Therefore, the output of CFFB can be described as:
Y C F R = softmax X C F R T W θ T W ϕ F F T W g T .
The global class representation learned in a supervised manner is obtained from highly correlated pixel features of the same category. The process of coarse classification can be seen as clustering of pixels, which can adaptively gather semantically related pixels and enlarge the differences between classes. Pixel feature mapping by category features and similarity matrix allow the network to re-estimate pixel features with key semantic information, which reduces intra-class differences and comprehensively improve the discrimination of fusion features.
Finally, we concatenate fusion features with second-stage features of the backbone network to obtain the final classification feature [ F e a t s I I , Y C F R ] that is sent to the classification layer. Due to existing unlabeled samples in hyperspectral images, we only calculate the loss function for labeled training samples. In this paper, cross entropy loss is used for coarse classification and final classification. The cross entropy loss is described as:
L softmax = i = 1 m t i c log p i c ,
where m is the total number of training samples, t i c is the c-th value of a label vector for pixel i, and p i c is the the corresponding probability. The loss function of CFF-FCN can be written as:
L C F R = L f i n a l + λ L c o a r s e .
λ is the weight factors to balance the two loss functions.

4. Experimental Results

4.1. Experimental Data Sets

To evaluate the performance of the proposed method, we conducted experiments and analyses on three widely used public data sets, including Indian Pines (IP), University of Pavia (UP), and Kennedy Space Center (KSC). The details of these three data sets are described as follows. Table 2 lists category names and the number of corresponding labeled samples.
IP data sets were the earliest test data used for HSI classification. They were collected by the Airborne Visible Infrared Imaging Spectrometer (AVIRIS) in Northwest Indiana, USA. A subset of the whole data set with a size of 145 × 145 is cut and annotated for HSI classification. The wavelength range is 0.4–2.5 μm, including 224 reflection bands. After removing the noise and water absorption bands, the remaining 200 bands are used as research objects. The spatial resolution is about 20 m, which is easy to produce mixed pixels and brings difficulty to classification. The IP data set includes 16 classes of ground truth. Moreover, 9 categories with labeled samples larger than 400 were selected for the experiment.
The UP data set was collected by the Reflective Optics System Imaging Spectrometer (ROSIS) from the University of Pavia, Italy. It contains 115 bands in the wavelength range of 0.43–0.86 μm, with a spatial resolution of 1.3 m. Moreover, 12 bands were eliminated due to the influence of noise. The size of UP was 610 × 340 , covering 9 classes, including trees, asphalt roads, bricks, pastures, etc.
The KSC data set was acquire by the AVIRIS at the Kennedy Space Center in Florida, USA. The image contains data of 224 bands from 0.4–2.5 μm, with a spatial resolution of 18 m. After removing the water absorption and noise bands, the remaining 176 bands were used for experiments. KSC contains 512 × 614 pixels and 13 land cover types.
The experimental results including Figure S1 and Table S1 of the Bostwana data set can be found in the Supplementary Materials.

4.2. Experimental Setup

Each data set was divided into a training set, validation set, and test set. The training set was served for back propagation to optimize model parameters. The validation set did not participate in the training, but evaluated the temporary model in the training phase. The test set evaluated the classification ability of the final model. For IP and UP data sets, 30 labeled samples of each class were randomly selected for training and validation. The remaining labeled samples were used for testing. For the KSC data set, 10 labeled samples for each class were randomly selected for training and verification.
We adopted class-based accuracy, overall accuracy (OA), average accuracy (AA), and kappa coefficient to the evaluation classification performance. The experiment was implemented on the PyTorch framework. The hardware platform included an Intel E5-2650 CPU, an NVIDIA GTX2080Ti GPU. The number of training iterations was 1000, and the learning rate was 0.0003.
We compared the proposed CFF-FCN with several state-of-the-art deep learning-based classifiers, including 3D-CNN [34], the contextual deep CNN (CDCNN) [35], the deep feature fusion network (DFFN) [36], the spectral–spatial residual network (SSRN) [37], and the spectral–spatial fully convolutional network (SSFCN-CRF) [22]. These methods all employ spectral–spatial features to implement HSI classification. Among them, 3D-CNN, CDCNN, DFFN, and SSRN are patch-based CNNs, and SSFCN-CRF is an image-based fully convolutional framework. The network settings of the above methods follow the corresponding references. Meanwhile, we also conducted experiments on two-stage networks (TSFCN) without CFFB, which can be regarded as the backbone of our proposed CFF-FCN. The specific numerical results and full classification maps are listed in Table 3, Table 4 and Table 5 and Figure 6, Figure 7 and Figure 8.

4.3. Classification Maps and Categorized Results

4.3.1. Indian Pines

Figure 6 and Table 3 respectively show the classification maps and numerical results of different methods in the IP data set. The TSFCN and CFF-FCN proposed in this paper are significantly better than other methods in three overall indicators. TSFCN achieves 100% classification accuracy on class 8 (Hay-windrowed), and CFF-FCN maintains the best accuracy on other classes. The introduction of global class context information makes OA increase by 2.31%. Due to the limitation of the number of training samples, 3DCNN, CDCNN and SSFCN-CRF cannot extract reliable spectral–spatial features, and the classification performance is relatively weak. Classification accuracies of DFFN and SSRN is preceded only to the proposed methods. Figure 6c,d,f also indicate that 3DCNN, CDCNN, and SSFCN-CRF have obvious class mixing and salt and pepper noise. Classification map of DFFN is smoother than SSRN’s. However, DFFN seems to have a certain degree of overfitting. The classification maps of TSFCN and CFF-FCN show the best visual effects, especially the use of global information makes the category edges of CFF-FCN smoother.

4.3.2. University of Pavia

Table 4 reports the classification accuracies for the UP data set. CFF-FCN shows the best performance in most categories and overall indexes. CFF-FCN and SSRN have achieved 100% class accuracy in class 5 (metal sheets). Compared with TSFCN, the OA of CFF-FCN is increased by 0.45%, indicating the effectiveness of CFFB. Figure 7 display the classification map of different methods. The 3DCNN, CDCNN, and SSFCN-CRF all have large regions of misclassification, and the class edges are not smooth. DFFN can observe obvious overfitting phenomenon in the UP data set. Compared with other methods, classification maps of TSFCN and CFF-FCN obviously have fewer misclassification points, and the class boundaries are clearer and smoother. Comparing TSFCN and CFF-FCN, it can be observed that global class information can alleviate the confusion of category boundaries to a certain extent.

4.3.3. Kennedy Space Center

Table 5 and Figure 8 present the experimental results of different methods on the KSC data set. Table 5 indicates that the performance of our proposed TSFCN and CFF-FCN is significantly better than other methods, achieving 100% classification accuracies for classes 7, 11, and 13. Compared with TSFCN, the OA of CFF-FCN increases by 1.56%. The classification accuracy of DFFN is second only to the proposed TSFCN and CFF-FCN. Table 5 shows that class 4 (CP) and class 6 (Oak) are difficult to distinguish correctly. Compared with other classifiers, CFF-FCN has achieved better accuracy in these two categories. From Figure 8b, it can be seen that the labeled samples of KSC only occupy a very small part of the entire image and are loosely distributed. In the classification map of CDCNN, there exists a lot of salt and pepper noise. DFFN still exhibits overfitting. SSFCN-CRF can detect all objects on the water surface. However, its classification map does not have clear category boundaries. Figure 8h also shows that the classification map of CFF-FCN is closer to the ground-truth map.

5. Discussion

In this section, we further discuss several factors affecting the performance of CFF-FCN. First, different numbers of training samples are input into the network to evaluate the performance in the case of limited numbers of training samples. Second, the effect of loss balance parameter λ with different values on classification results. Third, the effect of reverse loop mechanism and classification feature fusion module on classification accuracy is studied in an ablation manner. Finally, we compare and analyze the running times and the numbers of network parameters to evaluate the running efficiencies of different methods.

5.1. Investigation of the Proportion of Training Samples

Figure 9 shows the OAs and their standard deviations of different methods under different numbers of training samples. The number of training samples for the IP and UP data sets varies from 10 to 50. Due to the KSC label sample limitation, the number of training samples ranges from 3 to 20. In the case of precious few training samples, CDCNN and SSFCN-CRF are unable to provide reliable classification results on the three data sets. SSRN with higher model complexity can obtain relatively better classification performances in most cases. Figure 9c indicates that when the number of training samples on KSC is greater than 10, the classification performance of SSRN is not as good as 3DCNN. DFFN achieved competitive classification performances on both IP and KSC data sets. Of the three data sets, the OAs provided by TSFCN and CFF-FCN is better than other algorithms. In particular, in the case of fewer training samples, CFF-FCN can obtain a more reliable classification effect than TSFCN, reflecting the effectiveness of CFFB and the robustness of CFF-FCN. The standard deviation reflects the stability of the network performance. With the three data sets, the stability of TSFCN and CFF-FCN is better than other methods. Especially with the increase of training samples, the proposed models become more stable.

5.2. Investigation of λ

λ is the weighting factor used to balance the final classification loss L f i n a l and the coarse classification loss L c o a r s e in Equation (21). To study the influence of λ on the classification performance of CFF-FCN, we conducted experiments with different values of λ on three data sets, and the results are reported in Figure 10. It can be seen that CFF-FCN on the IP data set is most sensitive to changes in λ . For IP and KSC data sets, the model performances are best when λ is 1. The UP data set is the least sensitive to changes in λ , and OA changes very little under different λ . Combining the classification results of the three data sets, we set λ to 1.

5.3. Ablation Study

The ablation experiments in this section mainly discuss three issues: (1) the effectiveness of CFFB; (2) the effectiveness of the second stage of LFEB; (3) the contributions of F e a t s I and F e a t s I I to HSI classification. To this end, we conducted experiments on six types of networks according to different settings, and named them Model 1–6, among which Model 2 and Model 5 correspond to the TSFCN and CFF-FCN mentioned above. The specific model settings and classification accuracy are shown in Table 6.
Models 1–3 do not employ CFFB module. When LFEB does not reuse the second-stage parameters, the classification accuracies of Model 1 on IP and UP are significantly lower than that of Model 2 and 3. It demonstrates that the further refinement of the features from top to bottom in the second stage is conducive to improve the representation ability of the network. The similar conclusion can be drawn from Models 4–6. Comparing Model 2 and Model 3, it can be found that although F e a t s I I is a higher-level feature, F e a t s I shows stronger classification ability on UP and KSC. Parameters reused in Stage-II are also involved in calculation of F e a t s I . In Stage-II, high-level features are fed back to modulate low-level information, and F e a t s I is also refined to present the classification ability different from F e a t s I I . In order to increase feature complexity and discrimination, we use F e a t s I as the input of CFFB to reconstruct pixel features. The comparison of the results of Models 5 and 6 also show that the joint features at different stages can indeed improve the representation ability of the network and obtain more accurate classification results.

5.4. Investigation of Running Efficiency

Table 7 displays training times, testing times and numbers of network parameters for different methods. Among them, training times and test times correspond to the experimental settings in Table 3, Table 4 and Table 5, and network parameters are counted based on the IP data set. For patch-based methods, training times, and testing times are proportional to the number of samples, since they are pixel-by-pixel processing methods. For FCN-based methods, training times and testing times are proportional to the image size, because they need to traverse the entire image in each epoch. For the IP data set, patch-based methods require longer training times, and the training efficiency is not as good as image-based classification methods. For the UP data set, patch-based methods except DFFN and image-based methods have little difference in training time. For data sets with few labeled samples and large image size, the training time of FCNs is generally longer than that of patch-based methods. In the test phase, FCNs only need to traverse the image once, so the test time is generally shorter than patch-based methods. DFFN is a deeper classification network based on residual connections, resulting in the longest time in both the training and testing phases. In addition, compared with the SSRN whose classification performance is closest to CFF-FCN, our parameter amount is only 1/3 of SSRN. In practical applications, CFF-FCN can balance running efficiency and classification accuracy.

6. Conclusions

In this study, we propose a novel deep learning framework (CFF-FCN) with LEFB and CFFB to acquire the local spatial information and global class context for HSI classification. By dilated convolutions and reverse loop mechanism, LEFB provides multi-level semantic information covering different receptive fields. CFFB clustering correlated pixels, by supervised information, captures the global class representation and reconstructs pixel feature with class context. CFFB, different from pixel-by-pixel non-local extraction, replaces redundant pixel information with more representative class features in the reconstruction process, which decreases the demand for computing resources. Local spatial information and global reconstructed features are aggregated to implement classification. Extensive experiments and ablation study over three HSI data sets demonstrate that the proposed method provides promising classification performance compared with several state-of-the art deep learning methods. However, there are some hard samples at the boundaries between different classes that are always misclassified. How to further solve this problem will be the focus of our follow-up research.

Supplementary Materials

The following are available at https://www.mdpi.com/article/10.3390/rs13245043/s1, Figure S1: Classification maps for Botswana image, Table S1: Classification accuracies for Botswana using different classifiers.

Author Contributions

Conceptualization, Q.L.; methodology, Q.L.; formal analysis, Q.L.; writing—original draft, Q.L.; writing—review and editing, X.J., Z.W. (Zebin Wu) and Y.X.; supervision, Z.W. (Zhihui Wei); funding acquisition, Z.W. (Zebin Wu). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61772274, grant 61701238, and Grant 61671243, in part by the Jiangsu Provincial Natural Science Foundation of China under grant BK20180018, and grant BK20170858, in part by the Fundamental Research Funds for the Central Universities under grant 30917015104, grant 30919011103, and grant 30919011402, and in part by the China Postdoctoral Science Foundation under grant 2017M611814.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  2. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  3. Liu, P.; Zhang, H.; Eom, K.B. Active Deep Learning for Classification of Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 712–724. [Google Scholar] [CrossRef] [Green Version]
  4. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  5. Wei, H.; Yangyu, H.; Li, W.; Fan, Z.; Hengchao, L. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015, 258619. [Google Scholar]
  6. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  7. Wu, H.; Saurabh, P. Convolutional Recurrent Neural Networks forHyperspectral Data Classification. Remote Sens. 2017, 9, 298. [Google Scholar] [CrossRef] [Green Version]
  8. Aptoula, E.; Ozdemir, M.C.; Yanikoglu, B. Deep Learning With Attribute Profiles for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1970–1974. [Google Scholar] [CrossRef]
  9. He, N.; Paoletti, M.E.; Haut, J.M.; Fang, L.; Li, S.; Plaza, A.; Plaza, J. Feature Extraction With Multiscale Covariance Maps for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 755–769. [Google Scholar] [CrossRef]
  10. Zhao, W.; Du, S. Spectral—Spatial Feature Extraction for Hyperspectral Image Classification: A Dimension Reduction and Deep Learning Approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  11. Zhang, H.; Li, Y.; Zhang, Y.; Shen, Q. Spectral-spatial classification of hyperspectral imagery using a dual-channel convolutional neural network. Remote Sens. Lett. 2017, 8, 438–447. [Google Scholar] [CrossRef] [Green Version]
  12. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of Hyperspectral Image Based on Double-Branch Dual-Attention Mechanism Network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef] [Green Version]
  13. Ma, W.; Yang, Q.; Wu, Y.; Zhao, W.; Zhang, X. Double-Branch Multi-Attention Mechanism Network for Hyperspectral Image Classification. Remote Sens. 2019, 11, 1307. [Google Scholar] [CrossRef] [Green Version]
  14. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  15. Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-Based Adaptive Spectral–Spatial Kernel ResNet for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7831–7843. [Google Scholar] [CrossRef]
  16. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3D-2D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, H.; Li, Y.; Jiang, Y.; Wang, P.; Shen, Q.; Shen, C. Hyperspectral Classification Based on Lightweight 3D-CNN With Transfer Learning. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5813–5828. [Google Scholar] [CrossRef] [Green Version]
  18. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
  19. Cui, X.; Zheng, K.; Gao, L.; Zhang, B.; Ren, J. Multiscale Spatial-Spectral Convolutional Network with Image-Based Framework for Hyperspectral Imagery Classification. Remote Sens. 2019, 11, 2220. [Google Scholar] [CrossRef] [Green Version]
  20. Zheng, Z.; Zhong, Y.; Ma, A.; Zhang, L. FPGA: Fast Patch-Free Global Learning Framework for Fully End-to-End Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5612–5626. [Google Scholar] [CrossRef]
  21. Zhu, Q.; Deng, W.; Zheng, Z.; Zhong, Y.; Guan, Q.; Lin, W.; Zhang, L.; Li, D. A Spectral-Spatial-Dependent Global Learning Framework for Insufficient and Imbalanced Hyperspectral Image Classification. IEEE Trans. Cybern. 2021, 1–15. [Google Scholar] [CrossRef] [PubMed]
  22. Xu, Y.; Du, B.; Zhang, L. Beyond the Patchwise Classification: Spectral-Spatial Fully Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Big Data 2019, 6, 492–506. [Google Scholar] [CrossRef]
  23. Zhang, F.; Chen, Y.; Li, Z.; Hong, Z.; Ding, E. ACFNet: Attentional Class Feature Network for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  24. Yuan, Y.; Chen, X.; Wang, J. Object-Contextual Representations for Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), 24–28 August 2020; pp. 173–190. [Google Scholar]
  25. Zhou, S.; Wu, J.N.; Wu, Y.; Zhou, X. Exploiting Local Structures with the Kronecker Layer in Convolutional Networks. arXiv 2015, arXiv:1512.09194. [Google Scholar]
  26. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Garcia-Rodriguez, J. A Review on Deep Learning Techniques Applied to Semantic Segmentation. arXiv 2017, arXiv:1704.06857. [Google Scholar]
  27. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding Convolution for Semantic Segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  28. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local Neural Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  29. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral—Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3232–3245. [Google Scholar] [CrossRef]
  30. Wang, C.; Bai, X.; Zhou, L.; Zhou, J. Hyperspectral Image Classification Based on Non-Local Neural Networks. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 584–587. [Google Scholar]
  31. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  32. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance Normalization: The Missing Ingredient for Fast Stylization. arXiv 2016, arXiv:1607.08022. [Google Scholar]
  33. Yang, Y.; Zhong, Z.; Shen, T.; Lin, Z. Convolutional Neural Networks with Alternately Updated Clique. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2413–2422. [Google Scholar] [CrossRef] [Green Version]
  34. Ying, L.; Zhang, H.; Qiang, S. Spectral—Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar]
  35. Lee, H.; Kwon, H. Going Deeper With Contextual CNN for Hyperspectral Image Classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Song, W.; Li, S.; Fang, L.; Lu, T. Hyperspectral Image Classification With Deep Feature Fusion Network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3173–3184. [Google Scholar] [CrossRef]
  37. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral—Spatial Residual Network for Hyperspectral Image Classification: A 3D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
Figure 1. Receptive fields of 3 × 3 dilated convolution at different dilation rates. (a) d = 1 . (b) d = 2 . (c) d = 3 .
Figure 1. Receptive fields of 3 × 3 dilated convolution at different dilation rates. (a) d = 1 . (b) d = 2 . (c) d = 3 .
Remotesensing 13 05043 g001
Figure 2. The structure of the non-local block.
Figure 2. The structure of the non-local block.
Remotesensing 13 05043 g002
Figure 3. An overview of the proposed CFF-FCN.
Figure 3. An overview of the proposed CFF-FCN.
Remotesensing 13 05043 g003
Figure 4. The structure of LFEB.
Figure 4. The structure of LFEB.
Remotesensing 13 05043 g004
Figure 5. The structure of LFEB.
Figure 5. The structure of LFEB.
Remotesensing 13 05043 g005
Figure 6. For Indian Pines image. (a) RGB composite image of three bands. (b) Ground-truth map. (c) 3DCNN. (d) CDCNN. (e) DFFN. (f) SSRN. (g) SSFCN-CRF. (h) TSFCN. (i) CFF-FCN.
Figure 6. For Indian Pines image. (a) RGB composite image of three bands. (b) Ground-truth map. (c) 3DCNN. (d) CDCNN. (e) DFFN. (f) SSRN. (g) SSFCN-CRF. (h) TSFCN. (i) CFF-FCN.
Remotesensing 13 05043 g006
Figure 7. For University of Pavia image. (a) RGB composite image of three bands. (b) Ground-truth map. (c) 3DCNN. (d) CDCNN. (e) DFFN. (f) SSRN. (g) SSFCN-CRF. (h) TSFCN. (i) CFF-FCN.
Figure 7. For University of Pavia image. (a) RGB composite image of three bands. (b) Ground-truth map. (c) 3DCNN. (d) CDCNN. (e) DFFN. (f) SSRN. (g) SSFCN-CRF. (h) TSFCN. (i) CFF-FCN.
Remotesensing 13 05043 g007
Figure 8. For Kennedy Space Center image. (a) RGB composite image of three bands. (b) Ground-truth map. (c) 3DCNN. (d) CDCNN. (e) DFFN. (f) SSRN. (g) SSFCN-CRF. (h) TSFCN. (i) CFF-FCN.
Figure 8. For Kennedy Space Center image. (a) RGB composite image of three bands. (b) Ground-truth map. (c) 3DCNN. (d) CDCNN. (e) DFFN. (f) SSRN. (g) SSFCN-CRF. (h) TSFCN. (i) CFF-FCN.
Remotesensing 13 05043 g008
Figure 9. The overall accuracy (OA) of different methods for different numbers of training samples in three HSI data sets. (a) Indian Pines. (b) University of Pavia. (c) Kennedy Space Center.
Figure 9. The overall accuracy (OA) of different methods for different numbers of training samples in three HSI data sets. (a) Indian Pines. (b) University of Pavia. (c) Kennedy Space Center.
Remotesensing 13 05043 g009
Figure 10. The overall accuracy of CFF-FCN with different λ .
Figure 10. The overall accuracy of CFF-FCN with different λ .
Remotesensing 13 05043 g010
Table 1. Forword propagation of LEFB.
Table 1. Forword propagation of LEFB.
InputWeightOutput
X L E F W 01 X 1 I
[ X L E F , X 1 I ] [ W 02 , W 12 ] X 2 I
[ X L E F , X 1 I , X 2 I ] [ W 03 , W 13 , W 23 ] X 3 I
[ X L E F , X 1 I , X 2 I , X 3 I ] [ W 04 , W 14 , W 24 , W 34 ] X 4 I
[ X L E F , X 1 I , X 2 I , X 3 I , X 4 I ] [ W 05 , W 15 , W 25 , W 35 , W 45 ] X 5 I
[ X 2 I , X 3 I , X 4 I , X 5 I ] [ W 21 , W 31 , W 41 , W 51 ] X 1 I I
[ X 1 I I , X 3 I , X 4 I , X 5 I ] [ W 12 , W 32 , W 42 , W 52 ] X 2 I I
[ X 1 I I , X 2 I I , X 4 I , X 5 I ] [ W 13 , W 23 , W 43 , W 53 ] X 3 I I
[ X 1 I I , X 2 I I , X 3 I I , X 5 I ] [ W 14 , W 24 , W 34 , W 54 ] X 4 I I
[ X 1 I I , X 2 I I , X 3 I I , X 4 I I ] [ W 15 , W 25 , W 35 , W 45 ] X 5 I I
Table 2. The samples for each category of IP, UP, and KSC.
Table 2. The samples for each category of IP, UP, and KSC.
IPUPKSC
ClassTotalClassTotalClassTotal
1Alfalfa54Asphalt6631Scrub716
2No-till corn1434Meadow18,649Willow swamp243
3Min. till corn834Gravel2099CP hammock256
4Corn234Trees3064CP/Oak hammock252
5Pasture grass497Metal sheets1345Slash pine161
6Grass-trees747Bare soil5029Oak/Broadleaf hammock229
7Grass-pasture-mowed26Bitumen1330Hardwood swamp105
8Hay-windrowed489Bricks3682Graminoid marsh431
9Oats19Shadows947Spartina marsh520
10No-till soybean968 Cattail marsh404
11Min. till soybean2468 Salt marsh419
12Clean till soybean614 Mud flats503
13Wheat212 Water927
14Woods1294
15Buildings–grass–trees380
16Woods95
10,366 42,776 5211
Table 3. Classification accuracy(%) for Indian Pines using different classifiers.
Table 3. Classification accuracy(%) for Indian Pines using different classifiers.
CLASS3DCNNCDCNNDFFNSSRNSSFCN-CRFTSFCNCFF-FCN
251.0357.0281.8178.3356.9482.388.09
355.4450.1190.090.1459.7593.7195.68
585.7887.5693.2395.4883.5196.7297.6
694.1690.6693.699.1294.0399.399.65
898.5698.0699.1399.8998.810099.96
1054.2255.6686.6485.3672.7790.1591.25
1163.6561.9283.5978.2463.6586.3889.79
1247.7151.5493.8988.9466.1892.4594.61
1491.690.594.2898.5392.5499.8699.91
OA68.1868.288.4287.1972.691.393.61
AA71.3571.4590.6990.4576.4693.4395.17
κ × 100 62.6862.9186.4885.0568.1689.8392.53
Table 4. Classification accuracy(%) for University of Pavia using different classifiers.
Table 4. Classification accuracy(%) for University of Pavia using different classifiers.
CLASS3DCNNCDCNNDFFNSSRNSSFCN-CRFTSFCNCFF-FCN
183.5383.1279.095.9577.7696.9998.04
291.4591.5889.9494.6881.5797.2997.94
384.9879.5589.0494.5678.5798.1998.34
489.1894.5886.5395.8488.699.1199.05
599.5799.9397.3510096.8399.99100
658.2274.3984.3597.7987.2199.4899.13
784.387.2293.9299.2787.9499.9299.95
891.9176.7390.0987.9785.999999.42
999.1799.589.5998.1496.3799.7999.91
OA86.0786.987.6595.1283.3698.0498.49
AA86.9287.488.8796.0286.7698.8699.09
κ × 100 81.5182.7483.8593.678.6297.4298.01
Table 5. Classification accuracy(%) for Kennedy Space Center using different classifiers.
Table 5. Classification accuracy(%) for Kennedy Space Center using different classifiers.
CLASS3DCNNCDCNNDFFNSSRNSSFCN-CRFTSFCNCFF-FCN
191.9792.5694.594.2782.3698.9998.38
281.556069.8787.4267.3891.593.99
391.7173.1787.8567.488.2196.5996.5
451.6523.3971.4526.6954.4263.673.43
566.8955.3684.7771.3253.7174.4488.48
662.0157.5884.8936.7166.2195.4897.21
794.2187.6899.8974.3296.21100100
877.4870.8667.7779.4147.659293.54
989.7393.6990.878.4779.296.9497.49
1083.8351.2990.0584.0653.3295.6995.63
1198.794.0899.4198.4187.53100100
1281.583792.0586.9269.4791.0595.17
1397.7485.9799.6599.96100100100
OA86.0872.7889.582.9476.2494.4395.99
AA82.2367.8987.1575.872.7492.0294.6
κ × 100 84.569.6588.380.9373.5493.7995.53
Table 6. Effectiveness analysis (OA%) of CRFB and LFEB.
Table 6. Effectiveness analysis (OA%) of CRFB and LFEB.
ModelsCFFBStage-IIInput of CFFBOutput of LFEBData Sets
F e a t s I F e a t s I I F e a t s I F e a t s II IPUPKSC
1×××××89.497.7394.47
2××××91.398.0494.43
3××××90.4398.4594.85
4×××90.7998.1595.01
5××93.6198.695.99
6××92.8298.595.82
Table 7. Comparison of running efficiency between different methods.
Table 7. Comparison of running efficiency between different methods.
ModelsTraining Time(s)Testing Time(s)Parameters(M)
IPUPKSCIPUPKSC
3DCNN155.85154.3979.301.476.670.830.2
CDCNN179.25177.49104.751.486.840.841.06
DFFN784.10957.53548.5038.08225.6322.600.4
SSRN226.64169.16103.494.9022.512.730.36
SSFCN-CRF25.65114.04282.170.010.020.020.11
TSFCN76.15159.62294.710.020.080.170.12
CFF-FCN102.78186.51310.290.020.080.160.12
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, Q.; Wu, Z.; Jia, X.; Xu, Y.; Wei, Z. From Local to Global: Class Feature Fused Fully Convolutional Network for Hyperspectral Image Classification. Remote Sens. 2021, 13, 5043. https://doi.org/10.3390/rs13245043

AMA Style

Liu Q, Wu Z, Jia X, Xu Y, Wei Z. From Local to Global: Class Feature Fused Fully Convolutional Network for Hyperspectral Image Classification. Remote Sensing. 2021; 13(24):5043. https://doi.org/10.3390/rs13245043

Chicago/Turabian Style

Liu, Qian, Zebin Wu, Xiuping Jia, Yang Xu, and Zhihui Wei. 2021. "From Local to Global: Class Feature Fused Fully Convolutional Network for Hyperspectral Image Classification" Remote Sensing 13, no. 24: 5043. https://doi.org/10.3390/rs13245043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop