Next Article in Journal
Impact of Tides and Surges on Fluvial Floods in Coastal Regions
Previous Article in Journal
Using Apparent Electrical Conductivity to Delineate Field Variation in an Agroforestry System in the Ozark Highlands
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multibranch Crossover Feature Attention Network for Hyperspectral Image Classification

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610041, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(22), 5778; https://doi.org/10.3390/rs14225778
Submission received: 19 October 2022 / Revised: 8 November 2022 / Accepted: 9 November 2022 / Published: 16 November 2022

Abstract

:
Recently, hyperspectral image (HSI) classification methods based on convolutional neural networks (CNN) have shown impressive performance. However, HSI classification still faces two challenging problems: the first challenge is that most existing classification approaches only focus on exploiting the fixed-scale convolutional kernels to extract spectral–spatial features, which leads to underutilization of information; the second challenge is that HSI contains a large amount of redundant information and noise, to a certain extent, which influences the classification performance of CNN. In order to tackle the above problems, this article proposes a multibranch crossover feature attention network (MCFANet) for HSI classification. The MCFANet involves two primary submodules: a cross feature extraction module (CFEM) and rearranged attention module (RAM). The former is devised to capture joint spectral–spatial features at different convolutional layers, scales and branches, which can increase the diversity and complementarity of spectral–spatial features, while the latter is constructed to spontaneously concentrate on recalibrating spatial-wise and spectral-wise feature responses, meanwhile exploit the shifted cascade operation to rearrange the obtained attention-enhanced features to dispel redundant information and noise, and thus, boost the classification performance. Compared with the state-of-the-art classification methods, massive experiments on four benchmark datasets demonstrate the meliority of our presented method.

1. Introduction

Hyperspectral image (HSI), as an image-spectrum merging technology, combinates subdivisional spectroscopy with imaging technology, which contains abundant spatial distribution information of surface targets and hundreds or even thousands of contiguous narrow spectral bands [1,2]. In terms of recognition and classification, which benefit from luxuriant spatial and spectral features, HSI not only has an inherent preponderance over natural images but also can efficaciously distinguish different land-cover categories and objects. Therefore, HSI plays a crucial role in multifarious fields, such as military defense [3], atmospheric science [4], urban planning [5], vegetation ecology [6,7] and environmental monitoring [8,9]. Among the hyperspectral community, one of the most vibrant research applications is HSI classification. However, HSI classification also faces many formidable challenges, such as extensive redundant spectral information interference, available labeled samples deficiency and high intra-class variability.
Initially, traditional HSI classification methods, such as kernel-based and machine-learning approaches [10,11], are composed of two primary parts: feature extraction and classifier optimization. Representative algorithms are band selection (BS) [12], sparse representation classifier (SRC) [13], multinomial logistic regression (MLR) [14], principal components analysis (PCA) [15] and support vector machine (SVM) [16], which utilize rich spectral features to implement HSI classification. However, the above classification approaches only exploit spectral information and do not take full advantage of spatial information. In order to improve the classification performance, many spectral–spatial-based methods have been developed, which incorporate spatial context information into classifiers. For example, 3D morphological profiles [17] and 3D Gabor filters [18] were designed to obtain spectral–spatial features. Li et al. presented multiple kernel learning (MKL) to excavate spectral and spatial information of HSI [19]. Fang et al. constructed a novel local covariance matrix (CM) representation approach, which can capture the relation of spatial information and different spectral bands of HSI [20]. These conventional classification approaches depend on handcrafted features, which leads to the discriminative features’ insufficient extraction and poor robust ability.
In recent years, with the breakthrough of deep learning, HSI classification methods based on deep learning have demonstrated superior performance [21,22,23,24,25]. Chen et al. first applied a convolution neural network (CNN) to HSI classification [26]. Hu et al. raised a framework based on CNN, which comprised five convolutional layers to perform classification tasks [27]. Zhao et al. utilized a 2D CNN to obtain spatial features and then integrated them with spectral information [28]. In order to capture joint spectral–spatial information, Zou et al. constructed a 3D fully convolutional network, which can further obtain high-level semantic features [29]. In order to obtain promising information, Zhang et al. devised a diverse region-based CNN to extract semantic context-aware information [30]. Ge et al. presented a lower triangular network to fuse spectral and spatial features and thus achieved high-dimension semantic information [31]. Nie et al. proposed a multiscale spectral–spatial deformable network, which employed a spectral–spatial joint network to obtain low-level features composed of spatial and spectral information [32]. An effective and efficient CNN-based spectral partitioning residual network was built, which utilized cascaded parallel improved residual blocks to achieve spatial and spectral information [33]. A dual-path siamese CNN was designed by Huang et al., which integrated extended morphological profiles and siamese network with spectral–spatial feature fusion [34]. In order to obtain more prominent spectral–spatial information, Gao et al. designed a multiscale feature extraction module [35]. Shi et al. devised densely connected 3D convolutional layers to capture preliminary spectral–spectral features [36]. Chan et al. utilized spatial and spectral information to train a novel framework for classification [37].
The attention mechanism plays a crucial part in the HSI classification task and focuses on significant information related to the classification task [38,39,40,41,42]. Zhu et al. proposed a spatial attention block to adaptively choose a useful spatial context and a spectral attention block to emphasize necessary spectral bands [43]. In order to optimize and refine the obtained feature maps, Li et al. built a spatial attention block and a channel attention block [44]. Gao et al. proposed a channel–spectral–spatial attention block to enhance the important information and lessen unnecessary ones [45]. Xiong et al. utilized the dynamic routing between attention initiation modules to learn the proposed architecture adaptively [46]. Xi et al. constructed a hybrid residual attention module to enhance vital spatial-spectral information and suppress unimportant ones [47].
Inspired by the above successful classification methods, this article proposes a multibranch crossover feature attention network (MCFANet) for HSI classification. The MCFANet is composed of two primary submodules: a crossover feature extraction module (CFEM) and rearranged attention module (RAM). CFEM is designed to capture spectral–spatial features at different convolutional layers, scales and branches, which can boost the discriminative representations of HSI. Specifically speaking, CFEM consists of three parallel branches with multiple available, receptive fields, which can increase the diversity of spectral–spatial features. Each branch utilizes three additive link units (ALUs) to extract spectral and spatial information while introducing the cross transmission into ALU to take full advantage of spectral–spatial feature flows between different branches. Moreover, each branch also employs dense connections to combinate shallow and deep information and achieve strong related and complementary features for classification. RAM, including a spatial attention branch and a spectral attention branch, is constructed to not only adaptively pay attention to recalibrating spatial-wise and spectral-wise feature responses but also exploit the shifted cascade operation to rearrange the obtained attention-enhanced features to dispel redundant information and noise and, thus, enhance the classification accuracy. The main contributions of this article can be summarized as follows:
(1)
In order to decrease training parameters and accelerate model convergence, we designed an additive link unit (ALU) to replace the conventional 3D convolutional layer. For one, ALU utilizes the spectral feature extraction factor and spatial feature extraction factor to capture joint spectral–spatial features; for another, it also introduces the cross transmission to take full advantage of spectral–spatial feature flows between different branches;
(2)
In order to tackle the fixed-scale convolutional kernels that are difficult to sufficiently extract spectral–spatial features, a crossover feature extraction module (CFEM) was constructed, which can obtain spectral–spatial features at different convolutional scales and branches. CFEM not only utilizes three parallel branches with multiple available, receptive fields to increase the diversity of spectral–spatial features but also applies the dense connection to each branch to incorporate shallow and deep features and, thus, realize robust complementary features for classification;
(3)
In order to dispel the interference of redundant information and noise, we devised a rearranged attention module (RAM) to adaptively concentrate on recalibrating spatial-wise and spectral-wise feature response while exploiting the shifted cascade operation to realign the obtained attention-enhanced features, which are beneficial to boost the classification performance.
The remaining part of this article can be summarized as follows: Section 2 describes in detail our developed MCFANet, Section 3 provides experimental results and discussion and Section 4 gives the conclusion part of this article.

2. Methodology

The proposed MCFANet comprises two significant submodules: crossover feature extraction module (CFEM) and rearranged attention module (RAM), as shown in Figure 1. The former utilizes a multibranch intersection dense connection structure to enhance the representation ability of multiscale spectral–spatial information. The latter can not only focus on adaptively reweighting the significance of spectral-wise and spatial-wise features but also introduce the shifted cascade operation to replume the obtained attention-enhanced features to achieve more discriminative spectral–spatial features while dispelling redundant information and noise, thus, improving the classification performance.

2.1. Additive Link Unit

Generally, low-level features are more concrete and include more spatial details without discrimination, while high-level features are more abstract and contain more semantic information but without a high resolution. In recent years, many studies have indicated that integrating features of different layers is helpful for HSI classification. Zhang et al. designed a deep feature aggregation network to exploit the low-, middle- and high-level features in HSI [48]. Li et al. constructed a multi-attention fusion network to obtain complementary information from different levels [49]. Ge et al. adopted the early fusion method to aggregate features from the front and middle parts of the network and then utilized the fused features to further learn high-level semantic features [31]. Inspired by the above research, we devised an innovative additive link unit (ALU), which introduces the cross transmission to integrate spectral–spatial features of different scales and branches and strengthens spectral–spatial feature representation ability. The structure of the proposed ALU is shown in Figure 2.
As exhibited in Figure 2, the ALU can be partitioned into three parts: spectral feature extraction factor, cross transmission and spatial feature extraction factor. In order to reduce the training parameters, we utilized the spectral feature extraction factor and spatial feature extraction factor instead of the 3D conventional convolution layer. The spectral feature extraction factor was devised to obtain spectral information, including 1 × 1 × n convolution operation, BN layer and Relu activation function. The spatial feature extraction factor was designed to capture spatial information, including n × n × 1 convolution operation, BN layer and Relu activation function, where n represents the convolution kernel size, BN layer and Relu activation function are used to accelerate the network regularization. In addition, considering that if the three branches are independent of each other, it is difficult to exchange and fuse features at different scales. Therefore, we introduced cross transmission between the spectral feature extraction factor and spatial feature extraction factor to make it easier to achieve feature exchange at different scales and branches, thus improving the richness of spectral–spatial information. It can be calculated as follows:
s = δ ( w 0 X )
z = s + s 1 + s 2
y = δ ( w 1 z )
where X represents the input features of ALU; s , s 1 and s 2 represents the output features of spectral feature extraction factor of ALU at the same position on three branches; z represents the intermediate fusion features; and y represents the output features of ALU. w 0 and w 1 are the weights of spectral feature extraction factor and spatial feature extraction factor, respectively. δ is ReLU activation function.

2.2. Crossover Feature Extraction Module

Compared with single-scale feature extraction, integrating spectral–spatial information at different scales is a crucial method for enhancing HSI classification performance. Gao et al. employed mixed depthwise convolution to extract features at different scales from each feature map [50]. Xue et al. built a hierarchical residual network to capture multiscale spectral and spatial features [51]. Nie designed a multiscale spectral–spatial deformable network for HSI classification [32]. DenseNet enabled each layer to receive raw information generated by preceding layers, which was effective for the exploration of new features [52]. Inspired by the aforesaid advantages of approaches, we raised a crossover feature extraction module (CFEM), which not only adopts the multiple available, receptive fields to enrich the diversity of spectral–spatial features but also utilizes cross transmission and dense connection to explore deeper and newer features that are more conducive to HSI classification.
First, CFEM is composed of three parallel branches, and each branch contains three ALUs with diverse convolution kernels to obtain different scale spectral and spatial features of the input image, involving 1 × 1 × 3 , 3 × 3 × 1 , 1 × 1 × 5 , 5 × 5 × 1 , 1 × 1 × 7 and 7 × 7 × 1 . The structure of each branch is shown in Figure 3. Second, the cross transmission of ALU in each branch can maximize the use of local feature flows between different scales and branches. Furthermore, to increase the complementarity of spectral–spatial features, we applied the dense connection to three ALUs at the same branch. Finally, feature information of different scales and branches was fused by element-wise addition, which can be expressed as follows:
Y = a d d ( H ( X 0 , X 1 , , X i 1 ) , H 1 ( X 0 , X 1 , , X i 1 ) , H 2 ( X 0 , X 1 , , X i 1 ) )
where Y denotes the output features of CFEM; a d d ( · ) denotes the element-wise summation; X denotes the output features after fusion; and H 1 ( · ) , H 2 ( · ) and H 3 ( · ) denote the output features of each branch.

2.3. Rearranged Attention Module

The attention mechanism showed unexceptionable potential in boosting the model performance, which can focus on the information related to the current task and suppress ones. Common attention mechanisms can be divided into two categories: spectral attention mechanism and spatial attention mechanism. The former is to seek which feature map is important to learn and redistribute the weight to different channels, while the latter is to identify which area is worth paying attention to and redistribute the weight to the spatial content. Inspired by the above ideas, we designed a rearranged attention module (RAM). RAM concentrates on adaptively recalibrating spatial-wise and spectral-wise responses to obtain attention-enhanced features. Additionally, RAM utilizes the shifted cascade operation to realign the obtained attention-enhanced features to achieve more discriminative features for HSI classification while dispelling redundant information and noise, thus, promoting classification accuracy. The structure of our proposed RAM is exhibited in Figure 4.
As shown in Figure 4, the input feature maps of RAM are indicated by X R H × W × C , where H and W are the height and width of the spatial domain, and C is the spectral band number. We divided the input feature maps X into two sets: X 1 R H × W × ( C / 2 ) and X 2 R H × W × ( C / 2 ) , and transmitted them to the spectral attention branch and spatial branch attention, respectively. In the spectral attention branch, first, we adopted the global average pooling layer (GAP) to transform X 1 R H × W × ( C / 2 ) to X 1 R 1 × 1 × ( C / 2 ) . Then, a simple squeeze-and-excitation mechanism, including two fully connected layers (FCs) with different activation functions, was utilized to calculate the weights of each spectral band and obtain the spectral-wise dependencies F s p e . Finally, the input feature maps X 1 were multiplied by the channel weights F s p e to obtain the spectral attention-enhanced output feature maps X s p e , which can be expressed as follows:
F s p e = F C 1 ( F C 0 ( G A P ( X 1 ) ) ) = σ ( w 1 δ ( w 0 G A P ( X 1 ) ) )
X s p e = X 1 F s p e
where w 0 and w 1 represent the weight matrices of two FCs, respectively; δ represents the ReLU activation function; and σ represents the sigmoid activation function.
In the spatial attention branch, first, we used an average pooling layer (AP) and max pooling layer (MP) to transform X 2 R H × W × ( C / 2 ) to X 2 R H × W × 1 . Second, a concatenation operation was utilized to aggregate the average pooling feature and max pooling feature. Then, the aggregated features were sent to a 3 × 3 convolution layer to calculate the weights of the spatial domain and obtain the spatial-wise dependencies F s p a . Finally, the input feature maps X 2 were multiplied by the spatial weights F s p a to obtain the spatial attention-enhanced output feature maps X s p a . It can be expressed as follows:
F s p a = C o n v ( C o n c a t e [ A P ( X 2 ) , M P ( X 2 ) ] )
X s p a = X 2 F s p a
Although the outputs of two branches make use of the desired spectral and spatial features, these features are processed over the same feature space, which may impede the information flow. Therefore, we introduce the shifted cascade operation to alleviate this problem. First, we divided the spectral attention-enhanced feature maps X s p e and spatial attention-enhanced feature maps X s p a into two sunsets, respectively, denoted X s p e i R H × W × ( C / 2 ) , i { 1 , 2 } , X s p a j R H × W × ( C / 2 ) , j { 1 , 2 } . Then, the shifted cascade operation was utilized to combine X s p e 1 with X s p a 2 , and X s p e 2 with X s p a 1 to explore the new mixed features X 3 and X 4 . Finally, for X 3 and X 4 , we integrated them by the shifted cascade operation to obtain more conspicuous and representative features F for improving the classification performance. It can be expressed as follows:
X 1 = C o n c a t e [ X s p e 1 , X s p a 2 ]
X 2 = C o n c a t e [ X s p e 2 , X s p a 1 ]
F = C o n c a t e [ X 3 , X 4 ]
where C o n c a t e [ ] denotes the shifted cascade operation, which can explore some new features in other ways and enhance spectral–spatial feature representation ability.

2.4. Framework of the Proposed MCFANet

We utilized the Indian Pines dataset as an example to illustrate the developed method, as shown in Figure 5. First, HSI contains hundreds of spectral bands, and they are highly correlated with each other, leading to the existence of redundant information that is not conducive to the classification task. Therefore, we performed PCA on the raw HSI to eliminate unnecessary information and maintain 20 spectral bands that include the most important features. Second, to fully exploit the property of HSI containing both spectral and spatial data, we extracted a 3D image cube denoted by X R 11 × 11 × 20 as the input of our proposed MCFANet, where 11 × 11 is the spatial size and 20 is the number of spectral bands. Third, the input data were fed into CFEM comprising three parallel branches to capture multiscale spectral–spatial features and explore deeper and newer features for classification. The branch1 uses three ALUs, which are built by the kernel size of 1 × 1 × 3 × 16 and 3 × 3 × 1 × 16 in series, to excavate spectral and spatial features, respectively. Three ALUs were linked to each other by dense connection. Similarly, branch2 uses three ALUs which are built by the kernel size of 1 × 1 × 5 × 16 and 5 × 5 × 1 × 16 in series, to excavate spectral and spatial features, respectively. Three ALUs are linked to each other by dense connection. Branch3 uses three ALUs, which are built by the kernel size of 1 × 1 × 7 × 16 and 7 × 7 × 1 × 16 to excavate spectral and spatial features, respectively. Three ALUs were linked to each other by dense connection. Then, we adopted the element-wise addition to integrate the output features of three branches and acquire multiscale spectral–spatial features with the size of 11 × 11 × 16 . Next, the multiscale spectral–spatial features were sent to RAM to adaptively recalibrate spatial-wise and spectral-wise responses to obtain spectral attention-enhanced features with the size of 11 × 11 × 8 and spatial attention-enhanced features with the size of 11 × 11 × 8 . Furthermore, we adopted the shifted cascade operation to rearrange the obtained attention-enhanced features to achieve more discriminative features with the size of 11 × 11 × 16 for HSI classification. Finally, by using global average pooling, two FC layers and two dropout layers were used to convert the feature maps into a one-dimensional matrix, and class labels were generated via the softmax function.

3. Experimental Results and Discussion

This section introduces in detail the benchmark datasets used, the experimental setup, a series of parameter analyses, and the discussion of experimental results.

3.1. Datasets Description

The Pavia University (UP) dataset was gathered by the Reflective Optics Spectrographic Imaging System (ROSIS-03) sensor in 2003 during a flight campaign in Pavia, Northern Italy. The wavelength range is 0.43–0.86 μm. This dataset contained nine ground truth categories and has 610 × 340 pixels with a spatial resolution of 1.3 m. After excluding 12 bands due to noise, the remaining 103 bands are generally utilized for the experiment.
The Indian Pines (IP) dataset was collected by the airborne visible/infrared imaging spectrometer (AVIRIS) in 1992 in northwestern Indiana, USA. The wavelength range is 0.4–2.5 μm. This dataset contains 16 ground truth categories and has 145 × 145 pixels with a spatial resolution of 20 m. Since 20 bands cannot be reflected by water, the remaining 200 bands are generally utilized for the experiment.
The Salinas Valley (SA) dataset was captured by an AVIRIS sensor in 2011 for the Salinas Valley in California, USA. The wavelength range is 0.36–2.5 μm. This dataset contains 16 ground truth categories and has 512 × 217 pixels with a spatial resolution of 3.7 m. After eliminating bands that cannot be reflected by water, the remaining 204 bands are generally utilized for the experiment.
For the UP and SA datasets, we chose at random 10% labeled samples of each category for training and the remaining 90% labeled samples for testing. For the IP dataset, we select at random 20% labeled samples of each category as the training set, and the remaining 80% labeled samples as the testing set. Table 1, Table 2 and Table 3 list the land-over category details, sample numbers of three datasets and corresponding colors of each category.

3.2. Experimental Setup

All experiments were conducted on a system with an NVIDIA GeForce RTX 2060 SUPER GPU and 6 GB of RAM. The software environment of the system is TensorFlow 2.3.0, Keras 2.4.3, and Python 3.6.
The batch size was set to 16, and the number of training epochs was set to 200. Moreover, we adopted the RMSprop as an optimizer to update the parameters during the training process, and the learning rate was set to 0.0005. Three evaluation indices were used to evaluate the classification performance, i.e., Kappa coefficient (Kappa), average accuracy (AA) and overall accuracy (OA). The Kappa measures the consistency between the ground truth and the classification results. The AA represents the ratio between the total sample numbers of each category and the correctly classified sample numbers. The OA is the proportion of correctly classified samples in the total samples. In theory, the closer these evaluation indices utilized in this article are to 1, the better the classification performance will be.

3.3. Parameter Analysis

In this section, we mainly discussed five vital parameters that impact the classification results of our proposed MCFANet, i.e., the spatial sizes, training sample ratios, principal component numbers, the convolution kernel numbers in additive link unit and the number of additive link units. All experiments used the control variable method to analyze the influence of the aforementioned five important parameters.

3.3.1. Effect of the Spatial Sizes

Different HSI datasets have different feature distributions, and different spatial sizes may generate different classification results. Small size results in insufficient receptive fields, whereas large size results in more noise, which is to the disadvantage of HSI classification. Therefore, we fixed other parameters and set the spatial size to 7 × 7 , 9 × 9 , 11 × 11 , 13 × 13 and 15 × 15 to analyze their effects on the classification results of our proposed MCFANet for three datasets. The experimental results are provided in Figure 6. According to Figure 6, for the UP and IP datasets, the optimal spatial size is 11 × 11 . For the SA dataset, as the spatial size is 13 × 13 and 15 × 15 , three evaluation indices under the two conditions are the same. By considering the number of training parameters and time, the optimal spatial size was set to 13 × 13 for the SA dataset.

3.3.2. Effect of the Training Sample Ratios

The training sample ratios have a great effect on the HSI classification performance. In order to evaluate the robustness and generalization of the proposed MCFANet, we randomly choose 1%, 3%, 5%, 7%, 10%, 20% and 30% labeled samples for training and the remaining labeled samples for testing. The experimental results are provided in Figure 7. According to Figure 7, for three experimental datasets, it can be seen that as the training sample ratio is 1%, three evaluation indices are the lowest. With the increase in training sample ratios, three evaluation indices gradually improve. For the UP and SA datasets, as the training sample ratio is 10%, three evaluation indices reach a relatively stable level. For the IP dataset, as the training sample ratio is 20%, three evaluation indices reach a relatively stable level. This is because the UP and SA datasets have sufficient labeled samples, so even though the training sample ratio is small, our proposed method can still obtain high classification accuracies. In contrast, the IP dataset contains relatively small labeled samples, which means that the training ratio needs to be large for the proposed method to achieve good classification results. Therefore, the best training sample ratio for the UP and SA datasets is 10%, and the best training sample ratio for the IP dataset is 20%.

3.3.3. Effect of the Principal Component Numbers

HSI contains hundreds of spectral bands, and they are highly correlated with each other, leading to the existence of redundant information that is not conducive to the classification task. We performed PCA on the raw HSI dataset to reduce the training parameters of the proposed method by reducing the number of spectral bands of the original HSI dataset. We set the number of principal components to 5, 10, 20, 30 and 40 to evaluate their effects on the classification results of our proposed MCFANet for three datasets. The experimental results are provided in Figure 8. According to Figure 8, for the IP and SA datasets, as the number of the principal components is 20, three evaluation indices are obviously superior to other conditions. For the UP dataset, as the principal components numbers are 20, three evaluation indices are 99.82%, 99.62% and 99.80%. As the number of principal components is 40, three evaluation indices are 99.87%, 99.63% and 99.83%. Although the three evaluation indices of the former are 0.05%, 0.01% and 0.03% lower than those of the latter, in contrast, the former needs fewer training parameters and training time. Therefore, the optimal number of principal components for three datasets is 20.

3.3.4. Effect of the Number of Convolutional Kernels in Additive Link Units

The ALU is composed of a spatial feature extraction factor and a spectral feature extraction factor, which are similar to traditional 3D convolutional operations. Therefore, the number of output feature maps of ALU directly impacts the complexity and classification performance of the proposed method. We set the number of convolutional kernels in ALU to 2, 4, 8, 16, 32 and 64 to evaluate their effects on the classification results for three datasets. The experimental results are provided in Figure 9. According to Figure 9, for three datasets, it is clear that as the convolutional kernel number is 16, three evaluation indices are the most advantageous, and the proposed method has the best classification performance. Hence, the optimal number of convolutional kernels for three datasets is 16.

3.3.5. Effect of the Number of Additive Link Units

Our proposed MCFANet is composed of three parallel branches, and each branch includes multiple ALUS. A small amount of ALU leads to insufficient feature extraction, whereas too many ALUs lead to cause problems such as overfitting, a more complex structure of the model and gradient explosion, which are not conducive to HSI classification. Therefore, we set the number of ALUs to 2, 3, 4 and 5 to evaluate their effects on the classification results of our proposed MCFANet for three datasets. The experimental results are provided in Figure 10. According to Figure 10, the optimal number of ALUs for UP and SA datasets is 3. For the IP dataset, as the number of ALUs is 3, three evaluation indices are 99.61%, 99.72% and 99.55%. As the number of ALUs is 5, three evaluation indices are 99.68%, 99.69% and 99.64%. Although the OA and Kappa of the former are 0.07% and 0.09% lower than those of the latter, in contrast, the former needs fewer training parameters and training time. All things considered, the optimal number of ALUs for the IP dataset is 3.

3.4. Ablation Study

Our presented MCFANet involves two submodules: CFEM and RAM. In order to more comprehensively prove the validity of each module, ablation studies were implemented on three benchmark datasets, i.e., not using CFEM (named Netwok1), not using RAM (named Network 2) and using the combination CFEM and RAM (named Network 3). The experimental results are provided in Figure 11. From Figure 11, it is clear that Network 1 has the worst classification accuracies. For example, for the IP dataset, three evaluation indices are 2.37%, 3.13% and 3.05% lower than those of Network 3. In contrast, the classification performance of Network 2 has remarkable improvement. For example, for the UP dataset, three evaluation indices of Network 2 are 2.02%, 3.94% and 2.69% higher than those of Network 1. By comparison, Network 3 has superior classification results, which indicates that our devised CFEM has a greater effect on the classification performance of the presented model, and RAM can further enhance the classification results.

3.5. Comparison Methods Discussion

In order to verify the performance of our presented MCFANet, we utilized eleven classical methods for comparison experiments, which can be divided into two categories. One is based on traditional machine learning: SVM, RF, KNN and GuassianNB, which take all spectral bands as input. Another is based on deep learning: spectral–spatial residual network (SSRN) [53] is a 3D CNN, which uses spatial and spectral residual blocks to capture spectral–spatial features; fast, dense spectral–spatial convolution network (FDSSC) [54] utilizes different 3D convolutional kernel sizes based on dense connection to extract spatial and spectral features separately; HybridSN [55] combines one 2D convolutional layer and three 3D convolutional layers; Hybrid 3D/2D CNN (3D_2D_CNN) [56] is similar to HybridSN, which splices together 2D CNN components with 3D CNN components; multibranch 3D-dense attention network (MBDA) [57] exploits 3D CNNs to obtain spectral–spatial information and designs spatial attention mechanisms to enhance the spatial feature representations; multiscale residual network (MSRN) [50] constructs a multiscale residual block with mixed depthwise convolution to achieve multiscale feature learning. Table 4, Table 5 and Table 6 provide the classification results of different methods on three experimental datasets.
First, as shown in Table 4, Table 5 and Table 6, it can be seen that our developed MCFANet achieves excellent classification performance and has the highest classification accuracies in most categories. For the UP dataset, compared with other methods, the OA, AA and Kappa of the proposed MCFANet have an increase of approximately 0.34–33.98%, 0.47–33.15% and 0.32–43.17%, respectively. For the IP dataset, the OA, AA and Kappa of the proposed MCFANet have an increase of approximately 1.4–51.77%, 3.15–49.19% and 1.5~58.33%, respectively. For the SA dataset, the OA, AA and Kappa of the proposed MCFANet have an increase of approximately 0.14–23.20%, 0.09–13.7% and 0.15–25.51%, respectively. Overall, compared with seven deep learning-based classification approaches, SVM, RF, KNN and GuassianNB have lower classification accuracies, of which GuassianNB performs the worst. This is because they only capture features in the spectral domain and neglect ample spatial features. In addition, they need to rely on the prior information of experienced experts, leading to inferior robustness and generalization ability. Due to the hierarchical structure, seven methods based on deep learning can extract low-, middle- and high-level spectral–spatial features automatically and obtain decent classification accuracies.
Second, MBDA, MSRN and our proposed MCFANet adopted the multiscale feature extraction strategy. MBDA uses three parallel branches with a convolutional kernel with different sizes to capture multiscale spectral–spatial features. MSRN utilizes a multiscale residual block to perform multiscale feature extraction, which is composed of depthwise separable convolution with mixed depthwise convolution. It is clear from Table 4, Table 5 and Table 6 of the three methods that our proposed MCFANet obtains superior classification performance. For the UP dataset, three evaluation indices of the proposed MCFANet are 0.34%, 0.47% and 0.32% higher than those of MBDA and are 12.63%, 10.76% and 16.42% higher than those of MSRN. For the IP dataset, three evaluation indices of the proposed MCFANet are 1.4%, 3.15% and 1.59% higher than those of MBDA and are 2.34%, 6.90% and 2.67% higher than those of MSRN. For the SA dataset, three evaluation indices of the proposed MCFANet are 0.34%, 0.33% and 0.39% higher than those of MBDA and are 1.29%, 0.92% and 1.44% higher than those of MSRN. This is because compared with MBDA and MSRN, our proposed MCFANet can introduce not only multiple available, receptive fields to capture multiscale spectral–spatial features but also utilize dense connection and cross transmission to aggregate spectral–spatial features from different layers and branches.
Third, according to Table 4, Table 5 and Table 6, we also can obviously see that the evaluation indices of MBDA occupy the second place. This is because MBDA builds a spatial attention module to focus on spatial features that are related to HSI classification. Our proposed MCFANet also constructs a RAM to enhance spectral–spatial features and achieve the greatest classification accuracies. These indicate that the attention mechanism can boost classification performance to a certain degree. For the UP dataset, three evaluation indices of the proposed MCFANet are 0.34%, 0.47% and 0.32% higher than those of MBDA. For the IP dataset, three evaluation indices of the proposed MCFANet are 1.4%, 3.15% and 1.59% higher than those of MBDA. For the SA dataset, three evaluation indices of the proposed MCFANet are 0.34%, 0.33% and 0.39% higher than those of MBDA. This is because our designed RAM can not only focus on adaptively reweighting the significance of spectral-wise and spatial-wise features but also introduce the shifted cascade operation to replume the obtained attention-enhanced features to achieve more discriminative spectral–spatial features while dispelling redundant information and noise and thus, improving the classification performance.
Furthermore, as shown in Table 4, Table 5 and Table 6, it is clear that compared with six DL-based classification approaches, the FLOPs and test time of our proposed MCFANet are not the least, which indicates that our presented method still has some shortcomings. This could be because our designed CFEM includes different parallel branches, and the spectral–spatial information between branches is shared with each other; the spectral–spatial features at different scales and branches are effectively integrated, but the structure of our developed model and the test time is relatively large and not superior. Therefore, how to shorten the test time and reduce the complexity of the proposed MCFANet is still a problem worth studying.
Moreover, Figure 12, Figure 13 and Figure 14 provide the classification visual result maps of eleven comparison methods on the three experimental datasets. In contrast, SVM, RF, KNN and GuassianNB have coarse classification maps and contain vast noise and high misclassification rates. The classification maps of SSRN, FDSSC, 3D_2D_CNN, HybridSN, MBAN, MSRN and MCFANet have significantly improved and become clear. By comparison, the classification maps of our developed MCFANet are smoother and more accurate and are highly consistent with the ground-truth map on the three public datasets.

4. Conclusions

In this article, we developed a multibranch crossover feature attention network (MCFANet) for HSI classification. The MCFANet includes two main functional submodules: a cross feature extraction module (CFEM) and rearranged attention module (RAM). The former is designed to capture spectral–spatial features at different convolutional layers, scales and branches, which can increase the diversity and complementarity of spectral–spatial information. While the latter is constructed to not only adaptively pay attention to recalibrating spatial-wise and spectral-wise feature responses but also exploit the shifted cascade operation to rearrange the obtained attention-enhanced features to dispel redundant information and noise and, thus, improve the classification accuracy. The experimental results of plentiful comparison experiments and ablation studies demonstrate that our proposed MCFANet obtains superior classification accuracies on three benchmark datasets and has robust generalization ability while it can be extended to other HSI datasets.
Labeling the HSI pixels is expensive and time-consuming, and HSI contains limited labeled samples, therefore in future work, we will apply the semi-supervised or unsupervised classification approaches and data enhancement techniques to HSI classification to deal with the above problem of lack of labeled samples.

Author Contributions

Conceptualization, D.L.; validation, P.L., H.Y. and G.H.; formal analysis, D.L.; investigation, D.L., P.L., H.Y. and G.H.; original draft preparation, D.L.; review and editing, D.L., Y.W., P.L., Q.L., H.Y., D.C., Z.L. and G.H.; funding acquisition, G.H. and D.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Department of Science and Technology of Jilin Province under Grant number 20210201132GX.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, Y.; Chanussot, J.; Jia, X.; Benediktsson, J.A. Multiple Kernel learning for hyperspectral image classification: A review. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6547–6565. [Google Scholar] [CrossRef]
  2. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  3. Xu, Y.; Wu, Z.; Chanussot, J.; Wei, Z. Joint reconstruction and anomaly detection from compressive hyperspectral images using mahalanobis distance-regularized tensor RPCA. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2919–2930. [Google Scholar] [CrossRef]
  4. Pyo, J.; Duan, H.; Ligaray, M.; Kim, M.; Baek, S.; Kwon, Y.S.; Lee, H.; Kang, T.; Kim, K.; Cha, Y.; et al. An integrative remote sensing application of stacked autoencoder for atmospheric correction and cyanobacteria estimation using hyperspectral imagery. Remote Sens. 2020, 12, 1073. [Google Scholar] [CrossRef] [Green Version]
  5. Ghamisi, P.; Dalla Mura, M.; Benediktsson, J.A. A survey on spectral classification techniques based on attribute profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  6. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Process. Mag. 2014, 31, 45–54. [Google Scholar] [CrossRef] [Green Version]
  7. Ghiyamat, A.; Shafri, H.Z. A review on hyperspectral remote sensing for homogeneous and heterogeneous forest biodiversity assessment. Int. J. Remote Sens. 2010, 31, 1837–1856. [Google Scholar] [CrossRef]
  8. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, P.; Nasrabadi, N.; Nasrabadi, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  9. Malthus, T.J.; Mumby, P.J. Remote sensing of the coastal zone: An overview and priorities for future research. Int. J. Remote Sens. 2003, 24, 2805–2815. [Google Scholar] [CrossRef]
  10. Camps-Valls, G.; Bruzzone, L. Kernel-based methods for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1351–1362. [Google Scholar] [CrossRef]
  11. Gewali, U.B.; Monteiro, S.T.; Saber, E. Machine learning based hyperspectral image analysis: A survey. arXiv 2018, arXiv:1802.08701. [Google Scholar]
  12. Du, H.; Qi, H.; Wang, X.; Ramanath, R.; Snyder, W.E. Band selection using independent component analysis for hyperspectral image processing. In Proceedings of the 32nd Applied Imagery Pattern Recognition Workshop, Washington, DC, USA, 15–17 October 2003; pp. 93–98. [Google Scholar]
  13. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification via kernel sparse representation. IEEE Trans. Geosci. Remote Sens. 2013, 51, 217–231. [Google Scholar] [CrossRef] [Green Version]
  14. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised hyperspectral image segmentation using multinomial logistic regression with active learning. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4085–4098. [Google Scholar] [CrossRef] [Green Version]
  15. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J.A. PCA-based edge preserving features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  16. Mercier, G.; Lennon, M. Support vector machines for hyperspectral image classification with spectral-based kernels. In Proceedings of the 2003 IEEE International Geoscience and Remote Sensing Symposium (IGARSS 2003), Toulouse, France, 21–25 July 2003; Volume 6, pp. 288–290. [Google Scholar]
  17. Zhu, J.; Hu, J.; Jia, S.; Jia, X.; Li, Q. Multiple 3-D feature fusion framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1873–1886. [Google Scholar] [CrossRef]
  18. Huo, L.-Z.; Tang, P. Spectral and spatial classification of hyperspectral data using SVMs and Gabor textures. In Proceedings of the 2011 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2011, Vancouver, BC, Canada, 24–29 July 2011; pp. 1708–1711. [Google Scholar]
  19. Li, J.; Marpu, P.R.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A. Generalized composite kernel framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  20. Fang, L.; He, N.; Li, S.; Plaza, A.J.; Plaza, J. A new spatial–spectral feature extraction method for hyperspectral images using local covariance matrix representation. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3534–3546. [Google Scholar] [CrossRef]
  21. Mou, L.; Zhu, X. Learning to pay attention on spectral domain: A spectral attention module-based convolutional network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 110–122. [Google Scholar] [CrossRef]
  22. He, M.; Li, B.; Chen, H. Multi-scale 3d deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  23. Sellami, A.; Farah, M.; Farah, I.R.; Solaiman, B. Hyperspectral imagery classification based on semi-supervised 3-D deep neural network and adaptive band selection. Expert Syst. Appl. 2019, 129, 246–259. [Google Scholar] [CrossRef]
  24. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Li, J. Visual Attention-Driven Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8065–8080. [Google Scholar] [CrossRef]
  25. Fang, B.; Li, Y.; Zhang, H.; Chan, J.C.W. Hyperspectral Images Classification Based on Dense Convolutional Networks with Spectral-Wise Attention Mechanism. Remote Sens. 2019, 11, 159. [Google Scholar] [CrossRef]
  26. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  27. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef] [Green Version]
  28. Zhao, W.; Du, S. Spectral-spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  29. Zou, L.; Zhu, X.; Wu, C.; Liu, Y.; Qu, L. Spectral–Spatial Exploration for Hyperspectral Image Classification via the Fusion of Fully Convolutional Networks. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 659–674. [Google Scholar] [CrossRef]
  30. Zhang, M.; Li, W.; Du, Q. Diverse Region-Based CNN for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 27, 2623–2634. [Google Scholar] [CrossRef]
  31. Ge, Z.; Cao, G.; Zhang, Y.; Li, X.; Shi, H.; Fu, P. Adaptive Hash Attention and Lower Triangular Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 59, 5509119. [Google Scholar] [CrossRef]
  32. Nie, J.; Xu, Q.; Pan, J.; Guo, M. Hyperspectral Image Classification Based on Multiscale Spectral–Spatial Deformable Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5500905. [Google Scholar] [CrossRef]
  33. Zhang, X.; Shang, S.; Tang, X.; Feng, J.; Jiao, L. Spectral Partitioning Residual Network with Spatial Attention Mechanism for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5507714. [Google Scholar] [CrossRef]
  34. Huang, L.; Chen, Y. Dual-Path Siamese CNN for Hyperspectral Image Classification with Limited Training Samples. IEEE Geosci. Remote Sens. Lett. 2021, 18, 518–522. [Google Scholar] [CrossRef]
  35. Gao, H.; Zhang, Y.; Chen, Z.; Li, C. A Multiscale Dual-Branch Feature Fusion and Attention Network for Hyperspectral Images Classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 13, 8180–8192. [Google Scholar] [CrossRef]
  36. Shi, H.; Cao, G.; Zhnag, Y.; Ge, Z.; Liu, Y.; Fu, P. H2A2Net: A Hybrid Convolution and Hybrid Resolution Network with Double Attention for Hyperspectral Image Classification. Remote Sens. 2022, 14, 4235. [Google Scholar] [CrossRef]
  37. Chan, R.H.; Li, R. A 3-Stage Spectral-Spatial Method for Hyperspectral Image Classification. Remote Sens. 2022, 14, 3998. [Google Scholar] [CrossRef]
  38. Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene classification with recurrent attention of VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1155–1167. [Google Scholar] [CrossRef]
  39. Yang, K.; Sun, H.; Zou, C.; Lu, X. Cross-Attention Spectral–Spatial Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5518714. [Google Scholar] [CrossRef]
  40. Xiang, J.; Wei, C.; Wang, M.; Teng, L. End-to-End Multilevel Hybrid Attention Framework for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2019, 57, 1155–1167. [Google Scholar] [CrossRef]
  41. Huang, H.; Luo, L.; Pu, C. Self-Supervised Convolutional Neural Network via Spectral Attention Module for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6006205. [Google Scholar] [CrossRef]
  42. Tu, B.; He, W.; He, W.; Ou, X.; Plaza, A. Hyperspectral Classification via Global-Local Hierarchical Weighting Fusion Network. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 15, 182–200. [Google Scholar] [CrossRef]
  43. Zhu, M.; Jiao, L.; Liu, F.; Yang, F.; Wang, J. Residual Spectral–Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 449–462. [Google Scholar] [CrossRef]
  44. Li, R.; Zheng, S.; Chen, D.; Yang, Y.; Wang, X. Classification of Hyperspectral Image Based on Double-Branch Dual-Attention Mechanism Network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef] [Green Version]
  45. Gao, H.; Miao, Y.; Cao, X.; Li, C. Densely Connected Multiscale Attention Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 2563–2576. [Google Scholar] [CrossRef]
  46. Xiong, Z.; Yuan, Y.; Wang, Q. AI-NET: Attention inception neural networks for hyperspectral image classification. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 2647–2650. [Google Scholar]
  47. Xi, B.; Li, J.; Li, Y.; Song, R.; Shi, Y.; Liu, S.; Du, Q. Deep Prototypical Networks with Hybrid Residual Attention for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 3683–3700. [Google Scholar] [CrossRef]
  48. Zhang, C.; Li, G.; Lei, R.; Du, S.; Zhang, X.; Zheng, H.; Wu, Z. Deep Feature Aggregation Network for Hyperspectral Remote Sensing Image Classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 5314–5325. [Google Scholar] [CrossRef]
  49. Li, Z.; Zhao, X.; Xu, Y.; Li, W.; Zhai, L.; Fang, Z.; Shi, X. Hyperspectral Image Classification with Multiattention Fusion Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5503305. [Google Scholar] [CrossRef]
  50. Gao, H.; Yang, Y.; Li, C.; Gao, L.; Zhangm, B. Multiscale Residual Network With Mixed Depthwise Convolution for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3396–3408. [Google Scholar] [CrossRef]
  51. Xue, X.; Yu, X.; Liu, B.; Tan, X.; Wei, X. HResNetAM: Hierarchical Residual Network With Attention Mechanism for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 3566–3580. [Google Scholar] [CrossRef]
  52. Huang, G.; Liu, Z.; Weinberger, K.Q.; van der Maaten, L. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  53. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral-Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  54. Wu, W.; Dou, S.; Jiang, Z.; Sun, L. A Fast Dense Spectral-Spatial Convolution Network Framework for Hyperspectral Images Classification. Remote Sens. 2018, 10, 1068. [Google Scholar]
  55. Kumar Roy, S.; Krishna, G.; Ram Dubey, S.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar]
  56. Ahmad, M.; Shabbir, S.; Aamir Raza, P.; Mazzara, M.; Distefano, S.; Mehmood Khan, A. Hyperspectral Image Classification: Artifacts of Dimension Reduction on Hybrid CNN. arXiv 2021, arXiv:2101.10532v. [Google Scholar]
  57. Yin, J.; Qi, C.; Huang, W.; Chen, Q.; Qu, J. Multibranch 3D-Dense Attention Network for Hyperspectral Image Classification. IEEE Access 2022, 10, 71886–71898. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the developed MCFANet model.
Figure 1. The overall architecture of the developed MCFANet model.
Remotesensing 14 05778 g001
Figure 2. The structure diagram of additive link unit (ALU).
Figure 2. The structure diagram of additive link unit (ALU).
Remotesensing 14 05778 g002
Figure 3. The structure diagram of each branch of cross feature extraction module (CFEM).
Figure 3. The structure diagram of each branch of cross feature extraction module (CFEM).
Remotesensing 14 05778 g003
Figure 4. The structure diagram of rearranged attention module (RAM).
Figure 4. The structure diagram of rearranged attention module (RAM).
Remotesensing 14 05778 g004
Figure 5. Overall network model.
Figure 5. Overall network model.
Remotesensing 14 05778 g005
Figure 6. The effect of spatial sizes.
Figure 6. The effect of spatial sizes.
Remotesensing 14 05778 g006
Figure 7. The effect of training sample proportions.
Figure 7. The effect of training sample proportions.
Remotesensing 14 05778 g007
Figure 8. The effect of principal component numbers.
Figure 8. The effect of principal component numbers.
Remotesensing 14 05778 g008
Figure 9. The effect of convolutional kernel numbers in ALU.
Figure 9. The effect of convolutional kernel numbers in ALU.
Remotesensing 14 05778 g009
Figure 10. The effect of ALU numbers.
Figure 10. The effect of ALU numbers.
Remotesensing 14 05778 g010
Figure 11. The classification results of ablation study.
Figure 11. The classification results of ablation study.
Remotesensing 14 05778 g011
Figure 12. The visual classification maps of diverse approaches on the UP dataset.
Figure 12. The visual classification maps of diverse approaches on the UP dataset.
Remotesensing 14 05778 g012
Figure 13. The visual classification maps of diverse approaches on the IP dataset.
Figure 13. The visual classification maps of diverse approaches on the IP dataset.
Remotesensing 14 05778 g013
Figure 14. The visual classification maps of diverse approaches on the SA dataset.
Figure 14. The visual classification maps of diverse approaches on the SA dataset.
Remotesensing 14 05778 g014
Table 1. The information on UP dataset, including number of training and test samples.
Table 1. The information on UP dataset, including number of training and test samples.
No.ColorClassTrainTest
1Remotesensing 14 05778 i001Asphalt6645967
2Remotesensing 14 05778 i002Meadows186516,784
3Remotesensing 14 05778 i003Gravel2101889
4Remotesensing 14 05778 i004Trees3072757
5Remotesensing 14 05778 i005Metal sheets1351210
6Remotesensing 14 05778 i006Bare Soil5034526
7Remotesensing 14 05778 i007Bitumen1331197
8Remotesensing 14 05778 i008Bricks3693313
9Remotesensing 14 05778 i009Shadows95852
Total428138,495
Table 2. The information on IP dataset, including number of training and test samples.
Table 2. The information on IP dataset, including number of training and test samples.
No.ColorClassTrainTest
1Remotesensing 14 05778 i010Alfalfa1036
2Remotesensing 14 05778 i011Corn-notill2861142
3Remotesensing 14 05778 i012Corn-mintill166664
4Remotesensing 14 05778 i013Corn48189
5Remotesensing 14 05778 i014Grass-pasture97386
6Remotesensing 14 05778 i015Grass-trees146584
7Remotesensing 14 05778 i016Grass-pasture-mowed622
8Remotesensing 14 05778 i017Hay-windrowed96382
9Remotesensing 14 05778 i018Oats416
10Remotesensing 14 05778 i019Soybean-notill195777
11Remotesensing 14 05778 i020Soybean-mintill4911964
12Remotesensing 14 05778 i021Soybean-clean119474
13Remotesensing 14 05778 i022Wheat41164
14Remotesensing 14 05778 i023Woods2531012
15Remotesensing 14 05778 i024Buildings-Grass-Tree78308
16Remotesensing 14 05778 i025Stone-Steel-Towers1974
Total20558194
Table 3. The information of SA dataset, including number of training and test samples.
Table 3. The information of SA dataset, including number of training and test samples.
No.ColorClassTrainTest
1Remotesensing 14 05778 i026Broccoli-green-weeds-12012825
2Remotesensing 14 05778 i027Broccoli-green-weeds-23733353
3Remotesensing 14 05778 i028Fallow1981178
4Remotesensing 14 05778 i029Fallow-rough-plow140154
5Remotesensing 14 05778 i030Fallow-smooth2682410
6Remotesensing 14 05778 i031Stubble-trees3963563
7Remotesensing 14 05778 i032Celery3583221
8Remotesensing 14 05778 i033Grapes-untrained112810,143
9Remotesensing 14 05778 i034Soil-vineyard-develop6215582
10Remotesensing 14 05778 i035Corn-senseced-green-weeds3282950
11Remotesensing 14 05778 i036Lettuce-romaine-4 week107961
12Remotesensing 14 05778 i037Lettuce-romaine-5 week1931734
13Remotesensing 14 05778 i038Lettuce-romaine-6 week92824
14Remotesensing 14 05778 i039Lettuce-romaine-7 week107963
15Remotesensing 14 05778 i040Vineyard-untrained7276541
16Remotesensing 14 05778 i041Vineyard-vertical-trellis1811626
Total541848,711
Table 4. The classification accuracies of comparison approaches on the UP dataset.
Table 4. The classification accuracies of comparison approaches on the UP dataset.
No.SVMRFKNNGuassianNBSSRNFDSSC2D_3D_CNNHybridSNMBDAMSRNMCFANet
176.5291.6089.7196.5099.6798.3897.9898.7899.6298.9999.88
280.5989.0886.4679.1498.8999.7799.2999.9799.8399.9499.94
381.0380.0168.5130.4196.3085.0497.3993.30100.0099.4699.95
495.8593.3296.3344.0299.72100.0099.4796.5598.9993.1699.53
599.5999.3399.6679.4799.92100.0099.7795.9299.7640.5899.51
694.2588.8882.8239.1599.7299.3698.8599.96100.00100.0099.91
70.0084.7076.3238.9185.39100.0096.8197.15100.0033.7899.92
864.7379.0577.2970.0396.0799.9793.8391.5199.2691.3899.67
999.88100.00100.00100.0096.15100.0099.3399.2595.1574.9899.29
OA80.2188.8486.0165.8798.2398.7198.4298.2399.6187.2299.85
AA66.4785.3582.3472.5797.8998.2997.4395.8099.1588.8699.62
Kappa×10072.3984.9381.0556.6197.7098.2997.9097.6599.4883.3899.80
FLOPs (×106)----1.390.510.5110.240.480.321.13
Test Time (s)----9.3311.890.8710.3210.8610.8817.82
Table 5. The classification accuracies of comparison approaches on the IP dataset.
Table 5. The classification accuracies of comparison approaches on the IP dataset.
No.SVMRFKNNGuassianNBSSRNFDSSC2D_3D_CNNHybridSNMBDAMSRNMCFANet
10.0083.3311.7642.50100.000.0097.6290.9197.6286.3694.74
261.2172.3348.7340.9594.9895.3888.1797.0497.0991.2799.56
378.3179.5756.6723.2897.1696.1087.6199.4698.4296.9899.85
481.8273.3352.599.0489.77100.0087.8098.20100.00100.00100.00
593.5889.0183.332.7599.0691.6594.0694.5397.5295.81100.00
679.0979.8077.5867.5997.7699.6993.9299.5399.8499.2299.83
70.00100.0088.33100.00100.000.0095.4585.00100.0078.57100.00
884.8692.1088.3283.13100.0091.1097.5096.41100.00100.00100.00
90.000.0050.0011.1166.670.00100.00100.00100.00100.00100.00
1072.8399.3765.2323.3497.6399.7690.8599.40100.0098.2899.36
1156.7775.5970.6963.7394.4899.4190.7399.0599.4599.9599.49
1245.4859.9664.4615.7996.6493.8490.9389.4390.8194.1499.57
1387.4391.9580.3087.9298.9298.9280.5398.40100.00100.0098.80
1485.9090.6991.8575.1497.7099.6598.7596.6798.4498.87100.00
1581.8276.4059.8662.7189.3899.1492.4398.7999.7199.71100.00
1698.4698.4698.48100.00100.0097.5394.1294.0580.6185.4294.87
OA68.8979.0069.4547.8496.0497.4591.6597.4598.2197.2799.61
AA53.0967.7361.9850.5393.4279.6185.3092.5296.5892.8099.72
Kappa × 10063.5175.8065.0141.2295.4897.0990.4497.0997.9696.8899.55
FLOPs (×106)----0.173.810.5210.240.480.320.80
Test Time (s)-—---2.235.650.312.437.400.886.52
Table 6. The classification accuracies of comparison approaches on the SA dataset.
Table 6. The classification accuracies of comparison approaches on the SA dataset.
No.SVMRFKNNGuassianNBSSRNFDSSC2D_3D_CNNHybridSNMBDAMSRNMCFANet
1100.0099.8999.8399.87100.00100.00100.00100.00100.00100.00100.00
299.0899.6799.2699.5399.69100.0099.92100.00100.00100.00100.00
392.1495.0591.0689.4699.5792.6099.95100.00100.00100.00100.00
497.9498.5797.5097.0999.2498.8197.9999.7796.1590.3199.92
597.8798.9698.2697.2899.2199.9699.4199.84100.00100.0099.92
699.9299.92100.0099.97100.00100.00100.00100.00100.00100.00100.00
798.0299.5799.1697.78100.00100.0099.85100.0099.97100.00100.00
870.1277.5672.8874.3699.4799.9297.3699.6799.7398.31100.00
998.7898.8799.0998.90100.0099.5899.81100.00100.00100.00100.00
1087.8693.9189.8959.8499.1399.8698.97100.00100.0097.6799.90
1192.9693.6891.8131.1698.0296.7597.0098.1695.5699.3199.38
1294.9596.8895.1491.7699.5699.8999.40100.0099.8999.89100.00
1392.5096.8394.4594.62100.00100.0098.31100.0099.89100.0099.88
1497.4597.5996.6367.9099.60100.00100.0099.6195.5896.69100.00
1581.9978.5660.2845.8884.2789.3497.3999.7199.8396.56100.00
1698.9798.8199.0085.7999.82100.00100.00100.00100.0099.94100.00
OA88.5491.2287.2976.7797.2097.9398.8199.8399.6398.6899.97
AA92.6095.1693.1386.2698.7598.9999.1199.8799.6399.0499.96
Kappa × 10087.1890.2185.8574.4696.8997.7098.6799.8299.5898.5399.97
FLOPs (×106)----4.153.930.5210.240.480.321.13
Test Time (s)----16.5725.814.7015.5812.6914.5930.91
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, D.; Wang, Y.; Liu, P.; Li, Q.; Yang, H.; Chen, D.; Liu, Z.; Han, G. A Multibranch Crossover Feature Attention Network for Hyperspectral Image Classification. Remote Sens. 2022, 14, 5778. https://doi.org/10.3390/rs14225778

AMA Style

Liu D, Wang Y, Liu P, Li Q, Yang H, Chen D, Liu Z, Han G. A Multibranch Crossover Feature Attention Network for Hyperspectral Image Classification. Remote Sensing. 2022; 14(22):5778. https://doi.org/10.3390/rs14225778

Chicago/Turabian Style

Liu, Dongxu, Yirui Wang, Peixun Liu, Qingqing Li, Hang Yang, Dianbing Chen, Zhichao Liu, and Guangliang Han. 2022. "A Multibranch Crossover Feature Attention Network for Hyperspectral Image Classification" Remote Sensing 14, no. 22: 5778. https://doi.org/10.3390/rs14225778

APA Style

Liu, D., Wang, Y., Liu, P., Li, Q., Yang, H., Chen, D., Liu, Z., & Han, G. (2022). A Multibranch Crossover Feature Attention Network for Hyperspectral Image Classification. Remote Sensing, 14(22), 5778. https://doi.org/10.3390/rs14225778

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop