Next Article in Journal
A Global-Information-Constrained Deep Learning Network for Digital Elevation Model Super-Resolution
Next Article in Special Issue
Active Learning-Driven Siamese Network for Hyperspectral Image Classification
Previous Article in Journal
Synthetic Aperture Radar (SAR) Meets Deep Learning
Previous Article in Special Issue
Strip Attention Networks for Road Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low-Rank Constrained Attention-Enhanced Multiple Spatial–Spectral Feature Fusion for Small Sample Hyperspectral Image Classification

PLA Strategic Support Force Information Engineering University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(2), 304; https://doi.org/10.3390/rs15020304
Submission received: 2 December 2022 / Revised: 28 December 2022 / Accepted: 2 January 2023 / Published: 4 January 2023

Abstract

:
Hyperspectral images contain rich features in both spectral and spatial domains, which bring opportunities for accurate recognition of similar materials and promote various fine-grained remote sensing applications. Although deep learning models have been extensively investigated in the field of hyperspectral image classification (HSIC) tasks, classification performance is still limited under small sample conditions, and this has been a longstanding problem. The features extracted by complex network structures with large model size are redundant to some extent and prone to overfitting. This paper proposes a low-rank constrained attention-enhanced multiple feature fusion network (LAMFN). Firstly, factor analysis is used to extract very few components that can describe the original data using covariance information to perform spectral feature preprocessing. Then, a lightweight attention-enhanced 3D convolution module is used for deep feature extraction, and the position-sensitive information is supplemented using a 2D coordinate attention. The above widely varying spatial–spectral feature groups are fused through a simple composite residual structure. Finally, low-rank second-order pooling is adopted to enhance the convolutional feature selectivity and achieve classification. Extensive experiments were conducted on four representative hyperspectral datasets with different spatial–spectral characteristics, namely Indian Pines (IP), Pavia Center (PC), Houston (HU), and WHU-HongHu (WHU). The contrast methods include several advanced models proposed recently, including residual CNNs, attention-based CNNs, and transformer-based models. Using only five samples per class for training, LAMFN achieved overall accuracies of 78.15%, 97.18%, 81.35%, and 87.93% on the above datasets, which has an improvement of 0.82%, 1.12%, 1.67%, and 0.89% compared to the second-best model. The running time of LAMFN is moderate. For example, the training time of LAMFN on the WHU dataset was 29.1 s, and the contrast models ranged from 3.0 s to 341.4 s. In addition, ablation experiments and comparisons with some advanced semi-supervised learning methods further validated the effectiveness of the proposed model designs.

Graphical Abstract

1. Introduction

Hyperspectral images (HSI) exhibit rich features in the 2D spatial context domain and the 1D spectral reflectance domain, which support accurate recognition of objects with similar colors and textures. With the rapid development of portable sensors and unmanned aerial vehicle (UAV) remote sensing platforms, the acquired hyperspectral images show remarkable characteristics of high spatial resolution, which further promote precision agriculture, environmental monitoring, forestry, and other fields [1,2,3]. Therefore, hyperspectral image classification (HSIC) is the cornerstone of many fine-grained remote sensing applications and has received continued attention. However, the fine-grained classification of HSI still needs to be improved due to inherent data complexity and small sample dilemma.
HSIC research spans several decades, and some consensus has been developed on the difficulties of accurate classification. First, feature extraction and representation are the core steps since hyperspectral data share high dimensionality and high nonlinearity [4]. Prior to the dense usage of deep learning methods, HSIC research focused on well-designed features and machine learning classifiers. Secondly, the artificial fusion of multi-types of features effectively improves classification accuracy. Introducing spatial contextual information for an individual pixel can enhance the richness of spatial–spectral features and boost classification accuracy [5,6]. Finally, hyperspectral data annotation is time-consuming and laborious, thus HSIC encounters the significant small sample dilemma [7]. In addition, to cope with high data redundancy, dimensionality reduction (DR) methods have raised considerable concern and become popular feature extraction methods and pre-processing steps [8]. Principal component analysis (PCA) and its variants, such as kernel principal component analysis (KPCA), are such traditional DR methods that can filter data redundancy in the spectral domain [9]. Recently, novel spatial–spectral DR methods have been paid considerable attention [10,11]. Focusing on the specific high dimensional data structure, manifold learning-based methods are also investigated to evolve spatial correlation in the DR process [12]. Shi et al. proposed spatial–spectral multiple manifold discriminant analysis (SSMMDA) for DR in HSIC tasks, and the multi-manifold structure of HSI data has been elaborately considered [13]. The above consensus, namely multi-feature fusion, spatial–spectral joint feature extraction, and various DR methods, are seen as crucial prior knowledge for HSIC tasks. Due to sufficient experimental verification or a solid theoretical basis, the above multiple prior knowledge plays an essential role in the HSIC tasks, which will always be a research hotspot.
Deep learning methods are extensively investigated for HSIC due to their ability to extract abstract and robust features in a data-driven manner, which can obtain better classification performance than traditional methods [14]. CNN is one of the mainstream methods for HSIC due to its inherent applicability to 3D data structures. Continuous, in-depth and extensive research on CNN-based hyperspectral classification has been performed, in which 2D-CNN and 3D-CNN are commonly used models for HSIC [15,16,17,18,19]. In particular, 2D-CNN focuses on learning spatial features, while 3D-CNN has a good pertinency for hyperspectral data cube [20]. Roy et al. proposed a hybrid convolutional network (HybridSN) by fusing 3D-CNN and 2D-CNN, which effectively combines the advantages of both types of networks [21]. Building on 3D-CNN and 3D-2D-CNN, residual learning and dense connections have been investigated to optimize the network structure and boost the feature robustness [22,23]. Song et al. proposed cross-layer feature aggregation in a 2D residual network and verified the effectiveness of composite residual structure for hyperspectral classification [24]. Feng et al. proposed R-HybridSN, which uses residual connections to fuse different scales and different types of features and introduces depth-separable convolutions to significantly reduce the number of parameters [25]. Although deep 3D-CNNs can learn robust features with sufficient training data, they still suffer from overfitting problems in small sample conditions. Meanwhile, feature redundancy and limited perceptual fields are inherent limitations of 3D-CNN models in real-world small sample HSIC tasks.
To address the inherent limitations of deep CNN models, the visual attention mechanism was introduced for convolutional feature refinement. Visual attention is a network resource allocation mechanism that allows the network to focus more on features that are useful for classification through adaptive weight assignment [26]. Attention-based models, as a rising network paradigm, have developed gradually deeper integration with CNN models and received much attention in HSIC [27]. Mou et al. introduced spectral attention to hyperspectral data before feeding them into the CNN model. Band-wise adaptive calibration has been performed to improve classification accuracy and model interpretability [28]. Zhu et al. applied spectral–spatial attention to both raw hyperspectral data and residual modules in the network, which expanded the form of the convolution-attention fusion pattern and improved the robustness of spatial–spectral features [29]. Zhang et al. proposed the attention-dense hybrid convolutional network (AD-HybridSN), which introduces channel attention after each 3D convolution layer and spatial attention after each 2D convolution layer to achieve layer-wise spatial–spectral feature refinement [30]. Dong et al. designed cooperative spectral–spatial attention based on parallel learning of spatial weights and spectral weights and enhanced the feature expression ability of attention modules with dense connections [31]. To sum up, the integration of attention modules with CNN models has experienced three stages, namely simple embedding, stage-wise fusion, and layer-wise fusion. In addition, attention modules can also be directly used for spatial–spectral feature extraction [32,33].
Unlike convolution-attention fusion models, Transformer is a representative pure attention paradigm that learns global features by stacking self-attention modules. Vision transformer (VIT) has achieved comparable results with advanced CNN in various image processing tasks [34]. Recently, transformer-based models have been primarily explored in HSIC tacks, with a focus on specific data organization and feature fusion patterns. Hong et al. proposed spectral-former based on VIT. The group-wise embedding was adopted to strengthen the neighborhood spectral feature learning, and the cross-layer adaptive fusion module was used to enhance the feature interaction level [35]. Sun et al. proposed SSFTT by integrating 3D convolution, 2D convolution, and self-attention modules, in which the features extracted from the convolution layer were used as the prior knowledge for the self-attention layers [36]. Xue et al. introduced spatial partition restore (SPR) in local transformer to ensure pixel constraints, which strengthened local spatial dependency [37]. Although transformer-based models can obtain comparable results with advanced CNN models using sufficient training samples, they usually take longer time to fit the data. Meanwhile, transformer-based models pay more attention to global feature learning, which often leads to a large parameter number and overfitting in the small sample classification tasks. The transformer-based models may be further improved by drawing on some of the properties of CNNs.
The CNN models are seen as multi-layer local feature learning frameworks, and they are also benefiting from global feature learning ideas, such as large kernels and graph learning. First, the inherent limitations have been enhanced with redesigned models featuring large convolutional kernels and elegant structures, such as ConvNeXt and RepLKNet [38,39]. Second, graph CNN (GCN) has been proposed to combine graph learning with CNN to learn long-range features. Recently, Zhu et al. proposed a multi-scale short- and long-range graph convolutional network (MSLGCN) for HSIC. Multi-scale spatial embeddings and global spectral features are deeply explored by an elegant multi-stage structure [40]. Third, holistic image representation can be better learned by collecting second-order statistics of convolutional features. Feng et al. extracted spatial–spectral features using factor analysis and a cascade feature fusion architecture with parallel convolutional kernels of different sizes. Second-order pooling and L2 normalization are utilized to exploit the higher-order statistics of features extracted by CNN models [41]. In addition, the underlying convolution operation is being revisited and improved. Roy et al. designed a generalized gradient-centered 3D convolution, improving the feature learning power for HSI from the gradient level [42].
Unsupervised feature learning methods have a good reputation as they are impressive solutions to small sample HSIC tasks. The core idea of unsupervised learning is to construct additional supervised signals and evolve unlabeled data in the training process [43]. Zhu et al. proposed a self-supervised contrastive efficient asymmetric dilated network (SC-EADNet) and performed contrastive learning with the help of random crop augmentation [44]. Liu et al. proposed multi-view unsupervised contrastive learning framework using different PCA-reduced band intervals as contrastive sample pairs [45]. Facing cross-domain small sample HSIC tasks, Liu et al. proposed deep few-shot learning (DFSL) [46]. The deep residual network is used to learn a metric space in pre-collected datasets, and then a very small number of samples (five for each class) from the target domain are used for feature extraction and classification separately. Gao et al. proposed an end-to-end deep relation network-based few-shot classification (RN-FSC) framework, which can be seen as an improved version of DFSL [47]. Recently, an unsupervised metric learning framework with multi-view constraints was proposed and further promoted cross-domain HSIC tasks [48]. Most of the above methods are pioneering works that have normalized the hyperspectral un-supervised learning paradigm to some extent. However, there are still some shortcomings. For example, unsupervised learning with only several bands does not excavate the spectral features well in DFSL and RN-FSC. In addition, there is a lack of research on the applicability of model structure and unsupervised learning.
Facing the difficulties in fine-grained small sample hyperspectral classification, a framework based on low-rank constrained attention-enhanced multiple spatial–spectral feature fusion was designed in this paper. The aim of our research is to simplify the existing complex models, find out useful features with low redundancy, and make the best use of the extracted spatial–spectral embeddings. The proposed framework and main contributions are summarized as follows:
  • To fit the small sample dilemma, a few components that can model the hyperspectral data are first extracted using factor analysis to reduce data redundancy. One single vanilla 3D convolutional layer is utilized to primarily investigate spatial–spectral features, and the features are then further refined based on an attention-like manner.
  • Complementary position-sensitive and channel correlation information are provided by coordinate attention. The multi-level features extracted by 3D convolution and 2D attention are then unified and aggregated by a simple and effective composite residual structure. Finally, a compact vector of second-order statistics is learned for each class using low-rank second-order pooling to achieve classification.
  • Extensive experiments have been performed using four representative hyperspectral datasets with varying spatial–spectral characteristics. Using only five samples of each class for training, the LAMFN outperforms several recently released well-designed models with smaller parameter size and shorter running time.

2. Related Research

2.1. Mixed CNN Model

2D convolution is a commonly used feature extraction operator for RGB images. Assuming that the input feature dimension is H × W × B and the padding strategy is used, the output dimension of p 2D convolution kernels will be H × W × p . Obviously, the features extracted by each 2D convolution kernel can be regarded as a weighted average of the B channel features [15]. This manner leads to spectral feature loss and can only highlight the contributions of different bands using multiple convolution kernels. In contrast, 3D convolution extends explicit feature extraction to the spectral domain in HSIC tasks. The p 3D convolutional kernel will output H × W × B × p feature maps, and each 3D kernel convolves the input data from three directions [20]. However, when using an equal number of kernels, the required computational resources by 3D convolution are much larger than that of 2D convolution. The complexity increase is particularly noticeable when large kernel sizes are adopted, such as 7 × 7 × 7 .
The single-type CNN models can be improved by using a combination of 3D convolution and 2D convolution to form a mixed CNN model [25,30,41]. The basic structure of a mixed convolutional network can be divided into four basic building blocks, namely, 3D convolution, reshape, 2D convolution, and fully connected layers [21]. The 3D convolutional layer learns the spatial–spectral features, and the output dimension is H × W × B × C , where C is the number of 3D kernels. It should be noted that spectral features will be preserved and integrated during the 3D convolution process. The reshape operation reorganizes the four-dimensional tensor into a three-dimensional tensor with dimensions H × W × B C , which greatly enhances the spectral feature discriminability. The 2D convolution further learns spatial features, and a fully connected layer is utilized for classification.
Although the mixed CNN model is a new feature extraction paradigm, its basic structure has some shortcomings. First, this basic structure is shallow, and the classification accuracy significantly degrades when simply increasing the number of CNN layers [18]. Second, the contribution of multi-type convolutional features to the final classification has not been fully considered. Third, the reshape operation leads to excessive channel number and feature redundancy. The above problems will be fully considered and investigated in this paper, and the corresponding solutions will be described in Section 3.

2.2. Attention Mechanism

The attention mechanism is a method designed to adaptively refine the learned features by various backbone networks in a data-driven manner and is now widely used in many remote sensing image processing tasks [27]. The basic form of the attention mechanism can be represented by Equation (1):
A t t e n t i o n ( x ) = f ( g ( x ) , x )
where x denotes the input features; g ( ) is a function that calculates the weights based on the input features x ; and f ( ) represents a function that uses the weights to refine the input features. According to the processed dimension, attention methods can be broadly classified as spatial attention, channel attention, and spatial-channel mixed attention [26]. Hyperspectral features extracted by 3D-CNN contain rich spatial and spectral information, so multiple attention mechanisms can theoretically be used to calibrate which features are more valuable for classification.
Feature adaptive selection is commonly performed by attention methods such as squeeze-and-excitation attention (SE) and convolutional block attention module (CBAM). SE is a pioneer channel attention method, and the proposed pooling-fully connected paradigm is widely used and further improved to efficiently learn feature channel weights [49]. Squeeze-and-excitation attention can be calculated using Equations (2) and (3):
g ( x ) = F C e ( F C s ( G A P s p a ( x ) ) )
S E ( x ) = g ( x ) x
where F C e and F C s denote the fully connected layers used to compress and restore the channel-wise features, respectively; G A P s p a represents the global average pooling, which is performed on the spatial domain. CBAM extends the SE with spatial attention, learning spatial weights using a 7 × 7 convolution and using sequential two attentions to achieve feature refinement [50]. SE and CBAM, as classical attention methods, regulate the basic attentional form to some extent. Many improved attention methods since then can be seen as variants or combinations of SE and CBAM.
Since the receptive field of a single convolutional layer is limited, attention mechanisms can compensate for the lack of spatial–spectral features by learning global features of a particular dimension through fully connected layers. In addition, attention and convolution are often fused with residual structures for end-to-end training [51]. In our proposed model, the attention-enhanced spatial–spectral features will be extracted in a lightweight manner. And the complementary position-aware information will be considered and evolved in the mixed CNN model.

3. Methods

3.1. Network Structure

The hyperspectral classification framework, LAMFN, proposed in this paper is shown in Figure 1. First, factor analysis is performed to conduct feature transformation and extract a small number of components that can describe the hyperspectral data in terms of feature covariance. Second, the joint spatial–spectral features are extracted using two 3D convolutional layers. Meanwhile, the features are refined in a lightweight attention-enhanced manner (denoted as LA-3D-Conv in Figure 1). Third, the position-sensitive and direction-aware information are learned using a 2D coordinate attention module. Then, a multi-stage feature fusion strategy is adopted through a composite residual structure. One 2D depth separable convolution layer (2D-Sep-Conv) is adopted to integrate the multi-stage fusion features. Finally, a compact class vector is learned for each class using low-rank second-order pooling for final classification.
The classification process of the proposed framework consists of three parts: data pre-processing, spatial–spectral feature learning, and convolutional feature post-processing. The effective fusion of multi-type features with sufficient variance and low redundancy is the key to the proposed LAMFN. The vanilla 3D convolution layers are used to build a 4D spatial–spectral feature space. The LA-3D-Conv modules are designed to further integrate the learned features and improve the receptive field. The 3D spatial–spectral features with a large number of channels can be obtained by reshape operation. The 2D coordinate attention is used to investigate the unconsidered latent in 3D convolution, namely position-aware information and cross-channel correlation. Based on the effective fusion of multi-type features, the 2D-Sep-Conv and LRSOP can learn a holistic image representation.
The parameter settings for each module in LAMFN are shown in Table 1, and the parameter number takes the Indian Pines (IP) dataset as an example, which will be introduced in Section 4.1: Experimental Datasets. The patch size, the number of bands, kernel sizes, and other related hyperparameters are set referring to our previous research in the literature [22,27,35]. Tedious hyperparameter tunning is not considered in this paper. The details of each module will be described in the following chapters.

3.2. Hyperspectral Data Pre-Processing Based on Factor Analysis

Hyperspectral images have a 3D cube structure and share strong correlations in both spatial and spectral domains. Hughes phenomenon occurred between the feature dimension, the number of samples, and the classification accuracy in HSIC tasks. In other words, under a certain sample size, the classification accuracy will first increase and then decrease as the feature dimension increases. The above prior knowledge within hyperspectral data structure provides the feasibility and necessity for dimension reduction (DR). In addition, recently acquired hyperspectral images exhibit high spatial resolution and further enhance the texture details, which presents both opportunities and challenges for HSIC tasks. The literature [41] has validated the effectiveness of factor analysis (FA) as a DR method using some classical hyperspectral datasets. In this paper, FA will be adopted on more hyperspectral datasets with high spatial resolution to further validate its heterogeneous noise modeling capability. Detailed analysis using different components will also be included.
FA is closely related to PCA in terms of the calculation process. The underlying difference is that PCA focuses on the total variance of the variables, while FA focuses on the error covariance [52]. Due to the specific data structure of HSI, using FA as a dimensionality reduction method is verified helpful in improving inter-class selectivity [41]. Suppose the hyperspectral image data is represented as X = ( x 1 , x 2 , , x n ) T . Then, a latent variable model can be used to express the hyperspectral data mathematically:
X = W H + μ + ε
where H = ( h 1 , h 2 , , h n ) T is the random hidden variables and called the factors of X ; W represents the coefficient matrix to be estimated, and was known as the factor loading matrix; μ = ( μ 1 , μ 2 , μ n ) T is an offset vector that satisfies E ( x i ) = μ i ; ε = ( ε 1 , ε 2 , , ε n ) T is the noise item, which is usually assumed to obey the normal distribution. Suppose the error variance is expressed as D ( ε ) = d i a g ( σ 1 2 , σ 2 2 , , σ n 2 ) d e f ψ . If the variance of each component is equal, i.e., ψ = σ 2 I ( I is the unit matrix), then the hypothesis will lead to the probabilistic PCA model [53]; if the variance of each component is not equal, the hypothesis will lead to the FA model. In summary, FA takes into account the feature variance of the input data, which can effectively deal with heterogenous noise and help to improve inter-class differentiability. It is believed that the heterogeneous noise modeling capability of FA can improve the classification accuracy of HSI with high spatial resolution in complex scenarios.
In order to make full use of the joint spatial–spectral features, the spectral vectors within a specific range neighborhood of the pixel to be classified are extracted to form an image patch. For each pixel, its neighborhood spectral vector will be extracted to form an image patch of which dimension is S × S × D , where S is the window size. S and D are set to 15 and 16 in this paper as a compromise between classification accuracy and complexity.

3.3. Light-Weight Attention-Enhanced Spatial–Spectral Feature Learning

There are two drawbacks to 3D convolution. On the one hand, the commonly used convolution kernel size, such as 3 × 3 × 3 , has a limited field of perception. On the other hand, the four-dimensional tensor extracted by 3D convolution demands high computational resources. To further refine the features extracted by 3D convolution, the lightweight attention-enhanced 3D convolution (LA-3D-Conv) structure is proposed.
Let the hyperspectral image dimension be H × W × B , where H and W are spatial dimensions, and B is the spectral dimension. First, a vanilla 3D convolution layer is adopted to primarily investigate the joint spatial–spectral features and build 4D spatial–spectral feature space, F 3 D , with dimension H × W × B × C . It should be noted that the 4D feature space is essential since the spectral dimension is preserved, which is the main difference between 3D convolution and 2D convolution. To achieve lightweight feature refinement, the average pooling is performed along the channel dimension, and the pooled feature dimension is H × W × B × 1 . The spatial features are refined using two convolutional layers with kernels size 7 × 7 × 1 , and spectral features are further integrated. Zero padding is adopted to keep the feature dimension constant, and ReLU is served as the activation function. The feature map dimension is adjusted to H × W × B × 1 using one convolutional layer with kernel size 1 × 1 × 1 , and the sigmoid function is adopted to map the learned features to (0, 1) as weights. The final refined features will be obtained by broadcast multiplication between F 3 D and spatial–spectral feature weight. The above operation can be expressed by the following equation:
A t t e n t i o n ( F 3 D ) = f 111 ( f 771 2 ( G A P C ( F 3 D ) ) )
F r e f i n e d = F 3 D A t t e n t i o n ( F 3 D )
where F r e f i n e d represents the final refined features; represents tensor multiplication with a broadcast mechanism. The complete LA-3D-Conv structure is shown in Figure 2. To make the network training smooth, a residual structure is adopted to fuse the original features and the features refined by LA-3D-Conv, which is shown in Figure 1. In addition, a batch normalization layer is added after the residual structure to improve the feature robustness.
Our proposed lightweight attention-enhanced 3D convolution will facilitate spatial–spectral feature learning in two aspects. First, the dimensionality of the 3D convolutional feature map is reduced from H × W × B × C to H × W × B × 1 so that channel-wise redundancy has been filtered. Second, global spatial features are enhanced, and spectral features are further integrated during attention weight generation.

3.4. Spatial Feature Refinement Based on Coordinate Attention

The 4D spatial–spectral features extracted by 3D convolution in a mixed CNN model are reshaped into 3D tensors and contain feature groups with large differences, which cannot be effectively explored by vanilla 2D convolution. Aiming at the above problems, a spatial feature refinement unit based on coordinate attention (CA) is adopted. The original coordinate attention is proposed in the literature [54] and is aimed to enhance mobile network design by supplying direction-aware and position-sensitive attention maps. In this paper, we apply CA to obtain long-range dependencies across the restructured channel dimension and positional information, which are difficult to capture by 3D convolution with limited kernel size. It should be noted that BN is not adopted within CA in this paper. The coordinate attention schematic is shown in Figure 3.
The 3D convolution extracts features of dimension H × W × B C after reshaping. In order to introduce coordinate information, the feature map is pooled in both X-direction and Y-direction with two pooling kernels of ( 1 ,   W ) and ( H ,   1 ) , respectively. The pooling layers yield features of dimensions H × 1 × B C and 1 × W × B C . To further learn the channel importance, the two sets of features are aligned and connected to obtain a fused feature of dimension ( W + H ) × 1 × ( B C ) . The channel dimension is then reduced, and the position information is embedded. The process of feature concatenation and dimensionality reduction is shown in Equation (7):
f d r = a c t ( F 1 × 1 ( [ e w , e h ] ) )
where f d r represents the dimension reduced features, and the channel dimension is reduced to ( B C ) / k ; F 1 × 1 represents the 1 × 1 convolution; a c t ( ) represents the non-linear activation function, and H-Swish is adopted according to the literature [55]. It should be noted that the feature interaction in different directions is achieved since each spatial feature map of dimension ( W + H ) × 1 shares convolution weights. Before the channel dimension is recovered, f d r is separated into f w and f h along the spatial dimension, with dimensions W × 1 × ( B C ) / r and H × 1 × ( B C ) / r , respectively. The dimension of f w is adjusted to 1 × W × ( B C ) / r , keeping the same order as the initial pooling dimension. The f w and f h were then restored to their initial dimensions using two 1 × 1 convolution layers, and the feature values were mapped to (0, 1) using the sigmoid activation function as the attention weights. The above calculation process is shown in the following equation:
g w = s ( F w ( f w ) )
g h = s ( F h ( f h ) )
where g w and g h are the X- and Y-direction attention weights with dimensions 1 × W × ( B C ) and H × 1 × ( B C ) , respectively; s ( ) is the sigmoid function. The refined spatial–spectral features are obtained by tensor multiplication of g w , g h , and the input features.
Through coordinate attention, the global dependencies of spectral-channel features can be fully explored, and the position-sensitive information is embedded, which facilitates the network to find the spatial locations and valuable band intervals that are important for HSIC.

3.5. Low-Rank Constrained Second-Order Pooling for Final Classification

Compared to the commonly used fully connected (FC) layer or global pooling, second-order pooling can provide global image representation by computing the inner product of all given feature pairs. As for diverse spatial–spectral feature groups extracted by LA-3D-Conv and coordinate attention, all pairwise interactions among the multi-stage features can be considered. However, this second-order manner produces a high-dimension feature vector, which brings large parameter numbers and increases computational resources. In fact, it is found that in a high-dimensional feature space built by CNN models, less than 5% of the total features are informative [56].
Inspired by the literature [57,58], this paper introduces a simple low-rank constraint on the spatial–spectral features. A compact low-dimension vector will be learned for each category based on the second-order features. The adopted low-rank constraint method can be seen as a simplification of the method in the literature [58], and the diagram is shown in Figure 4.
Suppose the dimension of the multi-stage spatial–spectral features, which is denoted as F m u l , is S h × S w × M . First, a depth separable convolutional layer is adopted for low-dimension feature projection. Depth separable convolution is divided into depth-wise convolution and point-wise convolution. Depth-wise convolution can learn the unique spatial features of each channel without any interaction between channels. Point-wise convolution acts as a learnable nonlinear position-wise feature embedding tool. Let the embedded feature dimension be S h × S w × C , which is represented by F s s . By computing the inner product of each spatial location feature vector of F s s , which is denoted as x i R C ( 1 i S h × S w ) , and its transposed vector, the second-order feature representation, F s o p , with dimension C × C can be obtained. The above calculation process can be expressed succinctly in matrix form as follows:
F s o p = F s s T F s s
To enhance the feature robustness, L2 normalization is adopted after the second-order feature calculation.
The commonly used parameterization of the multiclass output layer with softmax activation function is known as Y = s o f t m a x ( F O W O + b O ) , where F O represents the extracted high-level features; W O and b O are the corresponding weight and bias of the output layer. Second-order pooling describes inter-channel correlation in detail and inevitably introduces redundant channels. Therefore, a low-rank constraint is introduced to learn a compact low-dimensional vector for each class. Let the parameters to be learned for each class based on second-order features be the weights W and biases b , where W R C × C and corresponds to the dimensions of F s o p . Suppose W * is the optimal parameter matrix. The literature [57] has proved that W * is a symmetric matrix, and the eigen-decomposition of W * can be expressed in the following formula:
W * = Ψ Σ Ψ T = Ψ + Σ + Ψ + T Ψ | Σ | Ψ T
where Σ + and Σ are diagonal matrices containing positive and negative eigenvalues, respectively; Ψ + and Ψ are the corresponding eigenvectors. Denote U + = Ψ + Σ + 1 2 and U = Ψ | Σ | 1 2 , and then the above decomposition can be written as follows:
W * = U + U + T U U T
Equation (12) indicates that W * may have a low-rank decomposition. Suppose the matrix rank of W * is r and satisfies r a n k ( W ) = r C , i.e., U + R C × r / 2 , U R C × r / 2 , then W * can be approximated by two low-rank matrices. In this case, only U + and U need to be parameterized and serve as the weights in the output layer, not the original weight matrix W .
HSIC tasks are fine-grained classification problems, thus N classes need to learn N sets of weights W and biases b . The low-rank weight parameter matrix of the i th class can be obtained from Equation (12) as W i = u + i u + i T u i u i T , where W i R C × C . The probability value of each class can be calculated by the following formula:
P = s o f t m a x ( vec ( F s o p ) vec ( W ) + b )
where P = [ p 1 , p 2 , , p N ] is the probability of each category; W = [ W 1 , W 2 , , W N ] is the sparse low-rank weight parameter matrix list computed through W i = u + i u + i T u i u i T ; b = [ b 1 , b 2 , , b N ] ; v e c ( ) represents vectorization operations. The classification result is obtained by Equation (13), and the parameter number will be C r N . Since r C , the parameter number and the computation burden can be remarkably reduced. Considering the balance of model size and classification accuracy, r is set to 16 in the proposed LAMFN.

4. Experimental Results and Analysis

4.1. Experimental Datasets

To verify the effectiveness and generalization of the proposed method for different types of HSI datasets, four challenging hyperspectral datasets with various spatial and spectral resolutions were used for classification experiments. The experimental datasets include classical and widely adopted Indian Pines (IP), Pavia Center (PC), Houston (HU), and the latest released WHU-HongHu (WHU) dataset with high spatial resolution. The detailed information of the experimental datasets is shown in Table 2, and the brief introduction is as follows:
  • IP is a classical dataset that is widely used to evaluate HSIC algorithms. Since its spatial resolution is low and many mixed pixels exist, the IP dataset can test the effectiveness of the algorithm in extracting robust features. The remaining three datasets have large image sizes and high spatial resolution.
  • PC has an original size of 1096 × 1096 . However, some areas have no information and they have been discarded. The processed PC dataset still has a large image size, which is 1096 × 715 . This scene contains many objects with similar textures and colors that often appear in urban areas.
  • The HU dataset was originally released during the 2013 Data Fusion Contest by the Image Analysis and Data Fusion Technical Committee of the IEEE Geoscience and Remote Sensing Society, so it is also named “2013_IEEE_GRSS_DF_Contest”. This dataset is challenging due to its large image size and sparse distribution of available samples.
  • The WHU dataset is a recently released hyperspectral dataset collected by unmanned aerial vehicles (UAVs) with extremely high spatial resolution [59]. The ground objects include 22 plant species with similar colors and textures, which greatly challenges the classification algorithm.
To sum up, the experimental datasets are mainly hyperspectral datasets with high spatial resolution. The ground objects include artificial objects in urban areas and various crops and plants in rural areas. In the following experiments, only five samples of each class were randomly selected for model training. The generalization of the proposed model for small sample HSIC tasks will be fully validated in different hyperspectral scenes.

4.2. Contrast Models

In order to assess the effectiveness of the proposed model, several advanced deep learning models published in recent years were used as contrast models, including R-HybridSN [25], MSSFN [41], A2S2K-ResNet [60], AD-HybridSN [30], SSFTT [36], and SPRLT [37]. The above contrast models can be broadly classified into three categories and will be briefly described as follows:
  • R-HybridSN and MSSFN are both residual CNN models, which are focused on small sample HSIC tasks. R-HybridSN adopts residual learning and depth separable convolution to refine the feature learning process. MSSFN further improves the classification accuracy using cascaded feature fusion patterns and second-order pooling.
  • A2S2K-ResNet and AD-HybridSN are two convolution-attention fusion models. AD-HybridSN is a mixed CNN model. Attention modules are inserted after each convolutional layer to achieve spatial–spectral feature refinement. A2S2K-ResNet adopted selective kernel attention in the first layer and efficient channel feature attention in the successive layers.
  • SSFTT and SPRLT are two very recently proposed transformer-based HSIC models. SPRLT adopts a spatial partition restore (SPR) module that is designed to split the input patch into several overlapping continuous sub-patches as sequential. SSFTT use 3D convolution and 2D convolution to build prior spatial–spectral feature space for subsequent transformer block. Both of them improve the applicability of transformer block in HSIC tasks by designing a novel feature organization approach.

4.3. Experimental Setup

The experimental hardware environment is computer equipped with an Intel i9-11900K CPU, an RTX 3090 GPU, and the RAM is 128G. The main experimental environment is Python 3.7, Tensorflow 2.8, and Pytorch 1.8. The model in this paper is based on Keras, which is a high-level deep learning framework running on top of Tensorflow.
LAMFN adopts Adam as the optimizer. The learning rate is set to 0.001, and the number of training epochs is set to 100. It should be noted that we found that LAMFN can obtain better results in the four datasets by monitoring training loss instead of training accuracy. In addition, almost the same experimental results can be obtained by directly testing the model obtained in the last epoch.
As for the contrast models, the patch size, dimension reduction setting, number of bands, optimizer, and learning rate are set according to the original papers. The training and testing process of SSFTT, SPRLT, and A2S2K-ResNet is completed on Pytorch deep learning framework. In our experiments, we used some open-source codes to process experimental data, train and test the contrast models, and the codes can be found at https://github.com/zgr6010/HSI_SSFTT and https://github.com/ZhaohuiXue/SPRLT-Net (accessed on 1 October 2022).
To ensure that the models can be well trained, the training epochs of A2S2K-ResNet and SPRLT are set to 200, and the training epochs of other contrast models are set to 100. The experimental datasets were randomly divided into a training set and a testing set using the same random seed in all experiments to guarantee a fair comparison. To be specific, we use Numpy library to divide training and testing samples randomly, and the seed number was set to 0. In all the experiments, five samples of each class are selected to train the model.

4.4. Comparison with the Contrast Models

In order to verify the effectiveness of the proposed method, classification experiments are performed using the IP, PC, HU, and WHU datasets. The contrast methods cover the advanced models proposed in recent years. The experiments were run continuously ten times, and the average values were taken as the experimental result. The accuracy of each class, OA, AA, and kappa, which are commonly used in hyperspectral classification, are used to measure the classification performance. The standard deviation of the last three statistics was given after ± to measure the accuracy fluctuation degree. Table 3, Table 4, Table 5 and Table 6 show the experimental results of IP, PC, HU, and WHU datasets, respectively.
By analyzing the quantitative experimental results from Table 3, Table 4, Table 5 and Table 6, the following findings can be drawn:
  • Various methods show distinct generalization properties for different datasets. For example, the experimental results of R-HybridSN and SPRLT with larger model sizes are not ideal for the four datasets, which indicates the over-fitting phenomenon is prominent. On the contrary, MSSFN and SSFTT performed relatively better. It is believed that the parameter number is the superficial reason, while the real reason is the difference in feature extraction ability and parameter redundancy caused by the model structure.
  • Two transformer-based models, SSFTT and SPRLT, can achieve comparable results with other models in PC and WHU, but the results in IP and HU are not ideal. Since patch-based structures are adopted, the global feature learning ability of the self-attention module is limited to a large extent. Therefore, the VIT-based approaches may consider continued improvements in data organization.
  • Quantitative experimental results demonstrate the effectiveness of LAMFN. In the four datasets, the proposed method achieved the highest OA and kappa. Compared with the second-best model, MSSFN, the OA of LAMFN has improved by 0.82%, 1.12%, 1.67%, and 0.89% in IP, PC, HU, and WHU datasets, respectively. The AA value of LAMFN was slightly lower than MSSFN in the WHU dataset.
  • Although the LAMFN achieved the best OA value in four datasets, it achieved the lowest standard deviation in only two datasets, namely PC and WHU. The stability of the LAMFN for some datasets needs to be further improved.
Figure 5, Figure 6, Figure 7 and Figure 8 show the predicted maps for all models, which can be seen as a visual reference to model performance assessment. The areas obtained by LAMFN with less noise have been highlighted to show the effectiveness of our proposed model. Overall, the proposed LAMFN has fewer misclassifications and omissions and generalizes well to different datasets with only five samples per class.

4.5. Running Time Analysis

In this paper, the algorithm complexity has been paid considerable concern at the beginning of the model design. Table 7 lists the input data size, parameter number (takes IP dataset as an example), training time, and testing time of LAMFN and all the contrast models. The WHU dataset with the largest data size is selected for running time display.
The model in this paper has good lightweight characteristics. First, the proposed model is comparable to MSSFN and SSFTT in terms of parameter number, while its accuracy exceeds those two models. This phenomenon indicates that our model can learn more robust HSI features with lower parameter redundancy. In terms of running time, the shorter running time indicates that fewer computational resources are required. The running time of A2S2K-ResNet is significantly longer than the other models since many 3D convolutional layers arranged in complex structures are used to extract features directly from hyperspectral images. The proposed model has moderate running time compared to the contrast models.

5. Discussion

5.1. Model Design Analysis of LAMFN

The overall effectiveness of the proposed method is verified in Section 4.4: Comparison with the Contrast Models. In order to further evaluate the detailed designs in LAMFN for small sample HSIC tasks, ablation experiments were conducted to test the improvement of FA, LA-3D-Conv, CA, and low-rank SOP on each dataset.
For FA, PCA was used for contrast experiments, and the other settings remained the same. For SSF, two-layer 3D convolution with residual structure was used to build a contrast model. The number of convolution kernels is consistent with the method in this paper, and the size of convolution kernels is set as 3 × 3 × 3 . For CA, blank control was performed, which means CA will be removed from LAMFN. For LRSOP, global average pooling (GAP) and original second-order pooling (SOP) will be used for comparison. Other experimental settings are consistent with the experiments in Section 4.4: Comparison with the Contrast Models, and the OA results are shown in Table 8.
The effectiveness of FA has been verified through the comparison between the results of LAMFN and C1. In order to further explore the influence of the number of bands on PCA and FA, a more detailed comparative experiment was conducted, with the number of bands ranging from 12 to 28. The experimental results are shown in Table 9. In addition, there is a big difference between FA and PCA in terms of dimensionality reduction time. Taking the WHU dataset as an example, Table 10 shows the running time of PCA and FA with different number of bands.
By analyzing the experimental results in Table 9 and Table 10, three conclusions can be drawn as follows:
  • When the number of bands is 16, FA achieves the overall best accuracy (measured in average) in the four datasets. By increasing the number of bands, only the accuracy in the IP data set continues to improve, while the accuracy of the other datasets tends to decrease. Compared with PCA, the FA dimension reduction method achieves the optimal factor number earlier.
  • FA significantly outperforms the PCA method when the number of bands is 16. In both IP and WHU datasets, FA has an absolute advantage. However, FA achieves slightly lower accuracy in the PC dataset than using PCA. As the number of bands increases, the accuracy obtained by PCA improves significantly in IP and to a lesser extent in other datasets. Overall, there is a large gap between PCA and FA methods.
  • PCA significantly outperforms FA in terms of resolving time. When the band number is 16, the resolving time of FA is 28.0 s slower than that of PCA. When the number of bands is taken as 24, the gap increases to 59.8 s. The slowdown is due to the fact that the FA takes into account the covariance of all input pixels. The extended solution time when the number of bands is taken as 16 is acceptable, given the resulting accuracy improvement.
Through the above analysis, the effectiveness and applicability of FA in hyperspectral small sample classification is further validated. On the one hand, the data covariance can effectively utilize the sample information and improve classification accuracy. On the other hand, for the trade-offs of resolving time, classification accuracy, and memory occupation, FA is sufficient to keep a small number of factors in the classification task.
The results in Table 8 completely verified all the refined designs for small sample HSIC tasks. The comparison between LAMFN, C2, and C3 demonstrates the improvements of lightweight SPS and CA modules. The comparison between LAMFN, C4, and C5 verified the improvement of low-rank second-order pooling compared to first-order pooling and direct second-order pooling. It should be noted that the LRSOP and SOP share one limitation compared to the GAP. When adopted for feature extraction, the dimension of the feature vector extracted by GAP will be much smaller than that of SOP or LRSOP since the explicit second-order feature calculation would square the channel number. Therefore, when being used for unsupervised feature extraction, such as in contrastive learning tasks, the number of channels needs to be further reduced. And this can be flexibly achieved by changing the kernel number of the 2D-Sep-Conv layer.
The motivation of LAMFN is to minimize the repetition of various features and reduce network redundancy. Multi-feature fusion is a widely adopted network design principle, but one of the key points may be ignored, namely, feature redundancy caused by the repetition of feature types. The concise feature extraction module with considerable variance is of great help to improve the classification accuracy. Based on this idea, LA-3D-Conv is designed to extract spatial–spectral features in a lightweight attention-based manner. CA is used to supplement positional information and cross-channel features ignored by 3D convolution. Then, different types of features are aggregated in a composite residual structure. In terms of feature utilization, LRSOP introduces low-rank constraints to reduce the number of parameters in the output layer while maintaining the selectivity advantage of the second-order features on different feature channels.

5.2. Comparison with Advanced Semi-Supervised Methods

In order to further verify and position the superiority of the proposed method, we compared LAMFN with some advanced semi-supervised methods proposed in recent years. Methods used for comparison include DFSL [46], DMVL [45], UM2L [48], and SC-EADNet [44]. The above methods include unsupervised feature learning and supervised fine-tunning, and they are briefly described below.
  • The DFSL is a cross-domain meta-learning method. Firstly, the model is pre-trained in the source domain dataset to obtain the metric space. Then, the well-trained model extracts robust features, and the final classification can be achieved by a simple linear classifier using very few samples.
  • The DMVL is a contrastive learning method. The two successive spectral vectors within each pixel are used for generating positive and negative samples for contrastive pretraining.
  • The UM2L is an end-to-end cross-domain meta-learning method. The pretraining phase was performed in an unsupervised manner with multi-view constrained contrast learning.
  • SC-EADNet combines contrastive learning with well-designed multiscale residual depth-separable convolutional networks.
The comparison results are shown in Table 9 (blank indicates that the dataset is not included in the above papers). For all the models, five samples from each class were selected to supervise the model training or fine-tuning process to ensure a fair comparison.
The comparison results in Table 11 show that, as a supervised learning method, the classification performance of LAMFN has exceeded several advanced semi-supervised algorithms, such as DFSL, DMVL, and UM2L. Since simple residual backbones, such as ResNet-50 and its variants, are used to conduct semi-supervised learning in the above three frameworks, the good performance of LAMFN should be attributed to the superiority of the proposed model structure. The OA of SC-EADNet is better than our method in the IP dataset and lower in the HU dataset, which also shows that the well-designed network can better learn spatial–spectral features in unsupervised manners. However, which kind of combination of model structure and semi-supervised strategy contributes more to semi-supervised hyperspectral classification has no prior knowledge and needs further study.

6. Conclusions

To boost the classification performance for small sample HSIC tasks, the LAMFN classification framework was derived from dimension reduction, spatial–spectral feature extraction, and convolution feature post-processing. The effective fusion of multi-type features with sufficient variance and low redundancy is the core idea of the proposed LAMFN. Firstly, FA is used to reduce the dimensionality of hyperspectral data, and the spectral features are pre-learned from the perspective of data covariance. The attention-enhanced feature learning unit, LA-3D-Conv, is used to further refine the hyperspectral features. The LA-3D-Conv can effectively map the dimensionally reduced hyperspectral data to the four-dimensional feature space and highlight the valuable spatial–spectral feature intervals for classification in a lightweight manner. Then, coordinate attention is used to extract direction-sensitive and channel-aware information from the reshaped 3D features. A composite residual structure is used to aggregate the features learned by different stages. Finally, the classification is realized by using second-order pooling with low-rank constraints to learn compact class vectors for each class.
Extensive classification and module ablation experiments are carried out using four hyperspectral datasets with large differences, namely, IP, PC, HU, and WHU. In the experiments, the proposed method achieves the highest classification accuracy, and the running time is moderate, which verifies the effectiveness of the proposed method. In addition, we will try to combine the method in this paper with the idea of contrastive learning to enhance the method’s applicability in more scenarios. Due to the highly modular network design principle, we hope our model can promote the small sample hyperspectral classification research and also benefit other types of remote sensing image processing tasks.

Author Contributions

Conceptualization, F.F. and J.Z.; methodology, F.F.; software, F.F.; validation, F.F. and J.Z.; formal analysis, F.F. and J.Z.; investigation, F.F.; writing—original draft preparation, F.F.; writing—review and editing, F.F., J.Z., Y.Z. and B.L.; supervision, Y.Z.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 42071340 and Program of Song Shan Laboratory (Included in the management of Major Science and Technology Program of Henan Province) under Grant 221100211000-01.

Data Availability Statement

Publicly available datasets are analyzed in this study, which can be found at http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 1 October 2022) and http://rsidea.whu.edu.cn/resource_WHUHi_sharing.htm (accessed on 1 October 2022).

Acknowledgments

The authors would like to thank the Hyperspectral Image Analysis group and the NSF Funded Center for Airborne Laser Mapping (NCALM) at the University of Houston for providing the datasets used in this study, and the IEEE GRSS Data Fusion Technical Committee for organizing the 2013 Data Fusion Contest. We also thank the anonymous reviewers for their constructive comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bouguettaya, A.; Zarzour, H.; Kechida, A.; Taberkit, A.M. Deep learning techniques to classify agricultural crops through UAV imagery: A review. Neural Comput. Appl. 2022, 34, 9511–9536. [Google Scholar] [CrossRef] [PubMed]
  2. Wambugu, N.; Chen, Y.; Xiao, Z.; Tan, K.; Wei, M.; Liu, X.; Li, J. Hyperspectral image classification on insufficient-sample and feature learning using deep neural networks: A review. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102603. [Google Scholar] [CrossRef]
  3. Vali, A.; Comai, S.; Matteucci, M. Deep Learning for Land Use and Land Cover Classification Based on Hyperspectral and Multispectral Earth Observation Data: A Review. Remote Sens. 2020, 12, 2495. [Google Scholar] [CrossRef]
  4. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature Extraction for Hyperspectral Imagery: The Evolution from Shallow to Deep: Overview and Toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  5. Ghamisi, P.; Maggiori, E.; Li, S.; Souza, R.; Tarablaka, Y.; Moser, G.; Giorgi, A.D.; Fang, L.; Chen, Y.; Chi, M.; et al. New Frontiers in Spectral–Spatial Hyperspectral Image Classification: The Latest Advances Based on Mathematical Morphology, Markov Random Fields, Segmentation, Sparse Representation, and Deep Learning. IEEE Geosci. Remote Sens. Mag. 2018, 6, 10–43. [Google Scholar] [CrossRef]
  6. Li, X.; Liu, B.; Zhang, K.; Chen, H.; Cao, W.; Liu, W.; Tao, D. Multi-view learning for hyperspectral image classification: An overview. Neurocomputing 2022, 500, 499–517. [Google Scholar] [CrossRef]
  7. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Xu, M.; Yu, S. A survey: Deep learning for hyperspectral image classification with few labeled samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  8. Ahmad, M.; Shabbir, S.; Raza, R.A.; Mazzara, M.; Distefano, S.; Khan, A.M. Artifacts of different dimension reduction methods on hybrid CNN feature hierarchy for Hyperspectral Image Classification. Optik 2021, 246, 167757. [Google Scholar] [CrossRef]
  9. Mohan, A.; Venkatesan, M. HybridCNN based hyperspectral image classification using multiscale spatiospectral features. Infrared Phys. Technol. 2020, 108, 103326. [Google Scholar] [CrossRef]
  10. Luo, F.L.; Du, B.; Zhang, L.P.; Zhang, L.F.; Tao, D.C. Feature Learning Using Spatial-Spectral Hypergraph Discriminant Analysis for Hyperspectral Image. IEEE Trans. Cybern. 2019, 49, 2406–2419. [Google Scholar] [CrossRef]
  11. Fu, H.; Sun, G.; Ren, J.; Zhang, A.; Jia, X. Fusion of PCA and Segmented-PCA Domain Multiscale 2-D-SSA for Effective Spectral-Spatial Feature Extraction and Data Classification in Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5500214. [Google Scholar] [CrossRef]
  12. Huang, H.; Shi, G.Y.; He, H.B.; Duan, Y.L.; Luo, F.L. Dimensionality Reduction of Hyperspectral Imagery Based on Spatial-spectral Manifold Learning. IEEE Trans. Cybern. 2020, 50, 2604–2616. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Shi, G.; Huang, H.; Liu, J.; Li, Z.; Wang, L. Spatial-Spectral Multiple Manifold Discriminant Analysis for Dimensionality Reduction of Hyperspectral Imagery. Remote Sens. 2019, 11, 2414. [Google Scholar] [CrossRef] [Green Version]
  14. Ahmad, M.; Shabbir, S.; Roy, S.K.; Hong, D.; Wu, X.; Yao, J.; Khan, A.M.; Mazzara, M.; Distefano, S.; Chanussot, J. Hyperspectral Image Classification—Traditional to Deep Models: A Survey for Future Prospects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 968–999. [Google Scholar] [CrossRef]
  15. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  16. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  17. Li, Z.; Wang, T.; Li, W.; Du, Q.; Wang, C.; Liu, C.; Shi, X. Deep Multilayer Fusion Dense Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1258–1270. [Google Scholar] [CrossRef]
  18. Roy, S.K.; Chatterjee, S.; Bhattacharyya, S.; Chaudhuri, B.B.; Platoš, J. Lightweight Spectral–Spatial Squeeze-and-Excitation Residual Bag-of-Features Learning for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5277–5290. [Google Scholar] [CrossRef]
  19. Liu, B.; Yu, A.; Gao, K.; Wang, Y.; Yu, X.; Zhang, P. Multiscale nested U-Net for small sample classification of hyperspectral images. J Appl. Remote Sens. 2022, 16, 016506. [Google Scholar] [CrossRef]
  20. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  21. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  22. Lee, H.; Kwon, H. Going Deeper with Contextual CNN for Hyperspectral Image Classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [Green Version]
  23. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A Fast Dense Spectral–Spatial Convolution Network Framework for Hyperspectral Images Classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef] [Green Version]
  24. Song, W.; Li, S.; Fang, L.; Lu, T. Hyperspectral Image Classification with Deep Feature Fusion Network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3173–3184. [Google Scholar] [CrossRef]
  25. Feng, F.; Wang, S.; Wang, C.; Zhang, J. Learning Deep Hierarchical Spatial-Spectral Features for Hyperspectral Image Classification Based on Residual 3D-2D CNN. Sensors 2019, 19, 5276. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Guo, M.-H.; Xu, T.-X.; Liu, J.-J.; Liu, Z.-N.; Jiang, P.-T.; Mu, T.-J.; Zhang, S.-H.; Martin, R.R.; Cheng, M.-M.; Hu, S.-M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  27. Ghaffarian, S.; Valente, J.; van der Voort, M.; Tekinerdogan, B. Effect of Attention Mechanism in Deep Learning-Based Remote Sensing Image Processing: A Systematic Literature Review. Remote Sens. 2021, 13, 2965. [Google Scholar] [CrossRef]
  28. Mou, L.; Zhu, X.X. Learning to Pay Attention on Spectral Domain: A Spectral Attention Module-Based Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 110–122. [Google Scholar] [CrossRef]
  29. Zhu, M.; Jiao, L.; Liu, F.; Yang, S.; Wang, J. Residual Spectral–Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 449–462. [Google Scholar] [CrossRef]
  30. Zhang, J.; Wei, F.; Feng, F.; Wang, C. Spatial-Spectral Feature Refinement for Hyperspectral Image Classification Based on Attention-Dense 3D-2D-CNN. Sensors 2020, 20, 5191. [Google Scholar] [CrossRef]
  31. Dong, Z.; Cai, Y.; Cai, Z.; Liu, X.; Yang, Z.; Zhuge, M. Cooperative Spectral–Spatial Attention Dense Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 866–870. [Google Scholar] [CrossRef]
  32. Xue, Z.; Zhang, M.; Liu, Y.; Du, P. Attention-Based Second-Order Pooling Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9600–9615. [Google Scholar] [CrossRef]
  33. Cheng, S.; Wang, L.; Du, A. Asymmetric coordinate attention spectral-spatial feature fusion network for hyperspectral image classification. Sci. Rep. 2021, 11, 17408. [Google Scholar] [CrossRef] [PubMed]
  34. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  35. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  36. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  37. Xue, Z.; Xu, Q.; Zhang, M. Local Transformer with Spatial Partition Restore for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4307–4325. [Google Scholar] [CrossRef]
  38. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  39. Ding, X.; Zhang, X.; Zhou, Y.; Han, J.; Ding, G.; Sun, J. Scaling Up Your Kernels to 31 × 31: Revisiting Large Kernel Design in CNNs. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  40. Zhu, W.; Zhao, C.; Feng, S.; Qin, B. Multiscale short and long range graph convolutional network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5535815. [Google Scholar] [CrossRef]
  41. Feng, F.; Zhang, Y.; Zhang, J.; Liu, B. Small Sample Hyperspectral Image Classification Based on Cascade Fusion of Mixed Spatial-Spectral Features and Second-Order Pooling. Remote Sens. 2022, 14, 505. [Google Scholar] [CrossRef]
  42. Roy, S.K.; Kar, P.; Hong, D.; Wu, X.; Plaza, A.; Chanussot, J. Revisiting Deep Hyperspectral Feature Extraction Networks via Gradient Centralized Convolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5516619. [Google Scholar] [CrossRef]
  43. Mou, L.; Ghamisi, P.; Zhu, X.X. Unsupervised Spectral–Spatial Feature Learning via Deep Residual Conv–Deconv Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 391–406. [Google Scholar] [CrossRef] [Green Version]
  44. Zhu, M.; Fan, J.; Yang, Q.; Chen, T. SC-EADNet: A Self-Supervised Contrastive Efficient Asymmetric Dilated Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5519517. [Google Scholar] [CrossRef]
  45. Liu, B.; Yu, A.; Yu, X.; Wang, R.; Gao, K.; Guo, W. Deep Multiview Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7758–7772. [Google Scholar] [CrossRef]
  46. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  47. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep Relation Network for Hyperspectral Image Few-Shot Classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef] [Green Version]
  48. Gao, K.; Liu, B.; Yu, X.; Yu, A. Unsupervised Meta Learning with Multiview Constraints for Hyperspectral Image Small Sample Set Classification. IEEE Trans. Image Process. 2022, 31, 3449–3462. [Google Scholar] [CrossRef]
  49. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [Green Version]
  50. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  51. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual Attention Network for Image Classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar]
  52. Li, N.; Zhao, H.; Jia, G. Dimensional reduction method based on factor analysis model for hyperspectral data. J. Image Graph. 2011, 16, 2030–2035. [Google Scholar]
  53. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006. [Google Scholar]
  54. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar]
  55. Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  56. Gao, Z.; Wu, Y.; Zhang, X.; Dai, J.; Jia, Y.; Harandi, M. Revisiting Bilinear Pooling: A Coding Perspective. In Proceedings of the AAAI 2020—34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; AAAI Press: Palo Alto, CA, USA, 2020; Volume 34, pp. 3954–3961. [Google Scholar]
  57. Kong, S.; Fowlkes, C. Low-Rank Bilinear Pooling for Fine-Grained Classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7025–7034. [Google Scholar]
  58. Xue, Z.; Zhang, M. Multiview Low-Rank Hybrid Dilated Network for SAR Target Recognition Using Limited Training Samples. IEEE Access 2020, 8, 227847–227856. [Google Scholar] [CrossRef]
  59. Zhong, Y.; Hu, X.; Luo, C.; Wang, X.; Zhao, J.; Zhang, L. WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2) benchmark datasets and classifier for precise crop identification based on deep convolutional neural network with CRF. Remote Sens. Environ. 2020, 250, 112012. [Google Scholar] [CrossRef]
  60. Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-Based Adaptive Spectral–Spatial Kernel ResNet for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7831–7843. [Google Scholar] [CrossRef]
Figure 1. Hyperspectral classification diagram based on LAMFN.
Figure 1. Hyperspectral classification diagram based on LAMFN.
Remotesensing 15 00304 g001
Figure 2. Schematic diagram of LA-3D-Conv structure.
Figure 2. Schematic diagram of LA-3D-Conv structure.
Remotesensing 15 00304 g002
Figure 3. Schematic diagram of spatial–spectral feature refinement based on coordinate attention.
Figure 3. Schematic diagram of spatial–spectral feature refinement based on coordinate attention.
Remotesensing 15 00304 g003
Figure 4. Schematic diagram of low-rank constrained classification based on second-order pooling.
Figure 4. Schematic diagram of low-rank constrained classification based on second-order pooling.
Remotesensing 15 00304 g004
Figure 5. Classification maps of all models for IP dataset.
Figure 5. Classification maps of all models for IP dataset.
Remotesensing 15 00304 g005
Figure 6. Classification maps of all models for PC dataset.
Figure 6. Classification maps of all models for PC dataset.
Remotesensing 15 00304 g006
Figure 7. Classification maps of all models for HU dataset.
Figure 7. Classification maps of all models for HU dataset.
Remotesensing 15 00304 g007
Figure 8. Classification maps of all models for WHU dataset.
Figure 8. Classification maps of all models for WHU dataset.
Remotesensing 15 00304 g008
Table 1. The implementation details of LAMFN.
Table 1. The implementation details of LAMFN.
ModuleOutput ShapeKernel SizeFilters or Other Parameter
Input (15, 15, 16, 1)
Stage 1: Spatial–spectral feature learning3D-Conv-BN-Relu(15, 15, 16, 16)(3, 3, 3)16
LA-3D-Conv(15, 15, 16, 16)(7, 7, 1) × 2
(1, 1, 1) × 1
16 × 2
1
Residual structure and BN layer(15, 15, 16, 16)
Stage 2: Spatial–spectral feature learning3D-Conv-BN-Relu(15, 15, 16, 16)(3, 3, 3)16
LA-3D-Conv(15, 15, 16, 16)(7, 7, 1) × 2
(1, 1, 1) × 1
16 × 2
1
Residual structure and BN layer(15, 15, 16, 16)
Stage 3: Complementary position information embeddingReshape(15, 15, 256)
Coordinate attention(15, 15, 256)(1, 1) × 364, 256, 256
Residual structure and BN layer(15, 15, 256)
Multi-stage fusionConcatenation
(Stage 1 and Stage2 also reshaped)
(15, 15, 768)
Stage 4: Low-rank constrained classification based on second-order featuresSep-Conv-BN-Relu(15, 15, 64)(3, 3)64
Reshape(225, 64)
Second-order pooling(64, 64)
L2 normalization(64, 64)
Low-rank constrained classification layerNumber of classes rank = 16
Parameter number: 157,826 (takes IP dataset as an example)
Table 2. Detailed information of the experimental datasets.
Table 2. Detailed information of the experimental datasets.
IPPCHoustonWHU
Spectral Range (μm)0.4–2.50.43–0.860.38–1.050.4–1.0
Number of Bands200102144270
Data Size145 × 1451096 × 715349 × 1905940 × 475
Spatial Resolution (m)201.32.50.043
Number of Classes1691522
Number of Labeled Data10,249148,15215,029386,693
Number of Training Data804575110
Table 3. Classification results (%) of different models in IP dataset.
Table 3. Classification results (%) of different models in IP dataset.
No.Residual CNNAttention-Based CNNTransformer ModelsProposed
R-
HybridSN
MSSFNA2S2K-
ResNet
AD-
HybridSN
SSFTTSPRLTLAMFN
196.1099.7696.3498.0598.2996.8398.78
250.3969.2840.7754.3449.0959.7264.40
348.6267.1448.1769.0256.2860.4066.02
480.4393.4581.4282.2490.3993.0690.95
575.2582.4369.2776.4675.6781.8083.37
692.8796.4788.8494.4492.7695.9096.72
7100.00100.0099.13100.00100.00100.0099.57
896.0098.2086.1795.5295.5092.0999.70
9100.0099.3399.33100.00100.0098.67100.00
1063.4269.6056.8555.1872.0368.7974.01
1155.0363.9845.0347.7857.2954.3667.36
1253.7467.1152.1347.7949.8866.8980.43
1399.6599.5597.9099.2598.5099.3598.55
1481.1395.0284.6984.6584.3889.3293.01
1576.3092.4179.9779.4882.2692.3486.90
1698.6498.8696.9396.8290.9199.3298.30
Kappa62.32 ± 2.5374.58 ± 3.2157.05 ± 6.7362.62 ± 2.8564.86 ± 3.6468.70 ± 2.4975.41 ± 3.44
OA66.34 ± 2.4277.33 ± 2.8961.35 ± 6.2866.40 ± 2.7968.68 ± 3.4671.95 ± 2.3778.15 ± 3.14
AA79.22 ± 1.4487.04 ± 1.5976.43 ± 5.3280.06 ± 1.9280.83 ± 1.9384.30 ± 1.1387.38 ± 1.55
Table 4. Classification results of different models in PC dataset.
Table 4. Classification results of different models in PC dataset.
No.Residual CNNAttention-Based CNNTransformer ModelsProposed
R-
HybridSN
MSSFNA2S2K-
ResNet
AD-
HybridSN
SSFTTSPRLTLAMFN
198.0099.4099.1499.5298.3799.1699.60
279.9884.0784.6684.6181.7984.5884.73
387.4981.9096.0587.0688.2883.0688.96
486.0998.6193.1397.9395.8172.3195.24
580.5382.8788.7084.8088.9385.9390.47
686.4391.7196.0693.8883.4887.6298.41
790.0391.0883.3691.0891.0987.9389.72
894.5298.3896.7796.7998.6495.7198.78
977.0786.1298.0388.4492.0191.8991.59
Kappa90.69 ± 2.5594.44 ± 1.8294.49 ± 1.7194.42 ± 1.2893.68 ± 1.6492.40 ± 2.7896.01 ± 0.84
OA93.34 ± 1.9196.06 ± 1.3096.09 ± 1.2396.04 ± 0.9295.51 ± 1.1794.59 ± 2.0397.18 ± 0.60
AA86.68 ± 1.0790.46 ± 2.8592.88 ± 1.4791.57 ± 1.4590.93 ± 2.1287.58 ± 3.6493.05 ± 1.12
Table 5. Classification results of different models in HU dataset.
Table 5. Classification results of different models in HU dataset.
No.Residual CNNAttention-Based CNNTransformer ModelsProposed
R-
HybridSN
MSSFNA2S2K-
ResNet
AD-
HybridSN
SSFTTSPRLTLAMFN
172.0584.9182.1077.4476.3786.9386.89
278.4984.4578.5981.6881.0080.3283.96
391.4598.0388.1597.1298.2499.5498.37
479.4888.3590.6681.4481.3286.1990.40
589.3092.7879.8996.9898.0199.6998.48
688.3885.0977.0388.3888.2293.4786.97
741.7670.8962.9167.1766.8056.3571.10
852.4652.9745.8149.5042.9958.0754.87
947.8969.0359.3761.1150.2237.6767.64
1058.3676.3841.8763.5362.1262.5273.40
1157.8373.7247.4670.6774.1964.6784.37
1268.4574.9045.8271.9565.1357.5673.08
1387.7882.5090.9194.7090.6067.8992.72
14100.00100.0097.0799.9399.76100.00100.00
1599.0596.4192.4199.8898.0999.5396.05
Kappa67.18 ± 1.9078.04 ± 2.1965.47 ± 3.9474.50 ± 2.9572.20 ± 2.4670.97 ± 2.0979.84 ± 2.82
OA69.56 ± 1.7779.68 ± 2.0368.01 ± 3.6376.35 ± 2.7574.26 ± 2.2973.11 ± 1.9581.35 ± 2.61
AA74.18 ± 1.4982.03 ± 1.6972.00 ± 3.7780.10 ± 2.3178.20 ± 1.9076.69 ± 1.7183.89 ± 2.17
Table 6. Classification results of different models in WHU dataset.
Table 6. Classification results of different models in WHU dataset.
No.Residual CNNAttention-Based CNNTransformer ModelsProposed
R-
HybridSN
MSSFNA2S2K-
ResNet
AD-
HybridSN
SSFTTSPRLTLAMFN
183.7775.4473.0780.6382.2381.0285.62
257.6275.8868.9880.7971.0367.9973.99
375.1981.1981.6475.2283.5277.8883.27
486.4194.9269.5491.4984.5679.0695.71
560.5093.5468.3586.2378.7174.7492.79
690.2493.0386.3889.9789.1278.0593.61
751.4966.9840.2063.9446.7042.3868.29
860.1162.9737.1466.0658.8428.8056.19
993.0192.8990.2992.7590.6092.3694.21
1042.6083.7565.1666.8447.4852.2980.02
1157.0073.5746.4259.1843.8826.0874.74
1244.2874.6760.3352.2449.0652.6873.47
1343.6257.5552.5052.4849.2444.8362.05
1477.4183.4971.2588.7265.2760.2186.35
1596.0499.7191.3595.8898.7786.6198.01
1684.3295.5882.8586.3489.9876.0995.13
1786.4290.9977.5277.8883.3669.1989.62
1891.7497.8387.0895.8089.7858.1495.87
1982.7092.6178.8087.6473.4847.6885.53
2080.8295.9977.4988.5059.8879.9993.66
2175.2897.4378.1689.5283.0867.6997.30
2276.4994.6286.3386.7885.5987.0394.14
Kappa71.88 ± 2.6283.84 ± 2.1264.27 ± 4.1378.47 ± 2.6971.16 ± 3.9864.05 ± 5.5684.90 ± 1.82
OA77.06 ± 2.2387.04 ± 1.7969.71 ± 3.9882.61 ± 2.2676.38 ± 3.6870.07 ± 5.1787.93 ± 1.49
AA72.59 ± 2.2885.21 ± 1.2971.40 ± 2.7479.77 ± 2.7172.92 ± 1.4465.03 ± 3.5984.98 ± 1.31
Table 7. Running related parameters of the LAMFN and the contrast models.
Table 7. Running related parameters of the LAMFN and the contrast models.
ModelR-
HybridSN
MSSFNA2S2K-
ResNet
AD-
HybridSN
SSFTTSPRLTLAMFN
Input size15 × 15 × 1615 × 15 × 169 × 9 × 20015 × 15 × 1613 × 13 × 309 × 9 × 20015 × 15 × 16
Parameter719,112159,012373,184366,662148,488839,728157,826
Training Time (s)16.624.8341.420.03.0239.329.1
Testing Time (s)49.260.7116.076.516.9184.0103.9
Table 8. OA results (%) of ablation model experiments.
Table 8. OA results (%) of ablation model experiments.
ModelDescriptionIPPCHoustonWHU
LAMFNproposed78.1597.1881.3587.93
C1Replacing FA with PCA74.7497.2781.1484.36
C2Replacing LA-3D-Conv with residual 3 × 3 × 3 3D-Conv77.0796.9880.2387.02
C3Without CA77.0697.0580.9486.96
C4Replacing LRSOP with GAP77.3995.6679.0187.67
C5Replacing LRSOP with SOP77.3596.9980.8786.56
Table 9. OA results (%) of different band number using FA and PCA.
Table 9. OA results (%) of different band number using FA and PCA.
BandIPPCHoustonWHU
FAPCAFAPCAFAPCAFAPCA
1277.7072.6297.1897.2781.1880.3187.2784.08
1678.1574.7497.1897.2781.3581.1487.9384.36
2077.6371.8796.8697.2481.7780.8986.8083.80
2478.0773.0097.0796.8380.5981.1687.0684.16
2877.4172.4796.9496.5680.4979.8786.7386.14
Table 10. Running time (second) of PCA and FA with different number of bands in WHU dataset.
Table 10. Running time (second) of PCA and FA with different number of bands in WHU dataset.
1216202428
PCA3.03.03.75.04.0
FA34.831.049.064.863.9
Table 11. Comparison results with several advanced semi-supervised methods.
Table 11. Comparison results with several advanced semi-supervised methods.
ModelIPPCHouston
LAMFN78.1597.1881.35
DFSL63.1194.9069.29
DMVL78.01 78.55
UM2L72.0994.27
SC-EADNet82.69 80.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, F.; Zhang, Y.; Zhang, J.; Liu, B. Low-Rank Constrained Attention-Enhanced Multiple Spatial–Spectral Feature Fusion for Small Sample Hyperspectral Image Classification. Remote Sens. 2023, 15, 304. https://doi.org/10.3390/rs15020304

AMA Style

Feng F, Zhang Y, Zhang J, Liu B. Low-Rank Constrained Attention-Enhanced Multiple Spatial–Spectral Feature Fusion for Small Sample Hyperspectral Image Classification. Remote Sensing. 2023; 15(2):304. https://doi.org/10.3390/rs15020304

Chicago/Turabian Style

Feng, Fan, Yongsheng Zhang, Jin Zhang, and Bing Liu. 2023. "Low-Rank Constrained Attention-Enhanced Multiple Spatial–Spectral Feature Fusion for Small Sample Hyperspectral Image Classification" Remote Sensing 15, no. 2: 304. https://doi.org/10.3390/rs15020304

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop