Next Article in Journal
A Multi-Scale Spatio-Temporal Fusion Network for Occluded Small Object Detection in Geiger-Mode Avalanche Photodiode LiDAR Systems
Previous Article in Journal
Contrastive Feature Disentanglement via Physical Priors for Underwater Image Enhancement
Previous Article in Special Issue
Tensor-Based Few-Shot Learning for Cross-Domain Hyperspectral Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Feature Enhanced Gaussian Weighted Network for Hyperspectral Image Classification

1
College of Communication and Electronic Engineering, Qiqihar University, Qiqihar 161000, China
2
College of Information Engineering, Huzhou University, Huzhou 313000, China
3
College of Information and Communication Engineering, Dalian Nationalities University, Dalian 116000, China
4
College of Computer and Control Engineering, Qiqihar University, Qiqihar 161000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(5), 763; https://doi.org/10.3390/rs17050763
Submission received: 25 December 2024 / Revised: 15 February 2025 / Accepted: 20 February 2025 / Published: 22 February 2025
(This article belongs to the Special Issue Deep Learning for Spectral-Spatial Hyperspectral Image Classification)

Abstract

:
Recently, research on hyperspectral image classification (HSIC) methods has made significant progress. However, current models commonly only focus on the primary features, overlooking the valuable information contained in secondary features that can enhance the model’s learning capabilities. To address this issue, an adaptive feature enhanced gaussian weighted network (AFGNet) is proposed in this paper. Firstly, an adaptive feature enhancement module (AFEM) was designed to evaluate the effectiveness of different features and enhance those that are more conducive to model learning. Secondly, a gaussian weighted feature fusion module (GWF2) was constructed to integrate local and global feature information effectively. Finally, a multi-head collaborative attention (MHCA) mechanism was proposed. MHCA enhances the feature extraction capability of the model for sequence data through direct interaction and global modeling. Extensive experiments were conducted on five challenging datasets. The experimental results demonstrate that the proposed method outperforms several SOTA methods.

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) are cubic data that simultaneously capture spatial and spectral information of target objects [1]. By combining imaging technology with spectroscopy, hyperspectral images generate continuous and narrow-band spectral image data, enabling a more comprehensive and detailed revelation of ground features [2]. Due to their unique spatial spectral integration characteristics, hyperspectral images are widely used in various fields such as medical imaging [3,4], agriculture [5,6], and mineral exploration [7].
HSIC is the process of analyzing the information within the image to assign category labels to each pixel in the image [8]. Early popular HSIC techniques primarily relied on the unique spectral information of different ground objects for classification. Some representative methods include K-nearest neighbor classifiers [9], support vector machines [10,11,12], and sparse representation classification [13]. Although these methods are simple and easy to implement, they only utilize spectral information and do not consider the continuity of spatial land distribution, resulting in incomplete feature extraction. In addition, traditional classification methods overly rely on expert experience, making it difficult to ensure the robustness and generalization of the methods.
In recent years, deep learning has shown strong application potential in the field of HSIC [14,15,16,17,18]. Convolutional neural network (CNN) [19,20], deep belief networks (DBNs) [21], and stacked autoencoders (SAEs) [22] have achieved good results in HSIC. However, both DBNs and SAEs require input data to be 1D vectors, which leads to significant limitations in the application of these models. CNN, with its characteristics of local perception and parameter sharing, can extract deeper data features while avoiding a sharp increase in model parameters, thus receiving widespread attention. Zhang et al. [23] proposed a spectral partitioning residual network (SPRN) based on 2D convolution kernels. This method divides the input spectrum into multiple non-overlapping continuous sub-bands by equivalently using grouped convolution, and employs cascaded parallel improved residual blocks to extract spectral–spatial features from these sub-bands separately [24]. Roy et al. [25] proposed a hybrid spectral CNN (HybridSN). Compared with the method that only uses 2D convolution kernels, this method combines more suitable 3D convolution kernels for hyperspectral data based on 2D convolution kernels, thus learning more abstract feature representations. Li et al. [26] proposed a central vector oriented self-similarity network (CVSSN), aiming to address the issue that existing CNN-based models ignore the potential relationship between the central pixel and its neighboring pixels when processing HSIs. Su et al. [27] proposed a method based on normalized spectral clustering with kernel-based learning (NSCKL). This method adopts a normalized spectral clustering algorithm, which can learn new features under the Manifold Hypothesis, and combines clustering features with extreme learning machines through kernel learning methods to achieve semi-supervised classification. Mei et al. [28] proposed a step activation network with binary weight (SAWB). This method replaces floating-point operations with integers, thereby accelerating the inference process of the network and reducing computational costs.
In traditional methods, treating each pixel equally makes it difficult to focus on valuable pixels [29,30,31]. To alleviate this problem, Zhang et al. [32] proposed a local–global cross fusion network with Gaussian initialization position prompts (LGGNet). Yang et al. [33] proposed a cross-attention spectral–spatial network (CASSN). This method utilizes cross-spectral and cross-spatial attention components to generate frequency band weights and spectral–spatial features, aiming to alleviate the problem of previous methods being less robust to HSI rotation. Wu et al. [34] proposed a network combining cross-channel dense connection and multi-scale dual aggregated attention (CDC_MDAA) to alleviate the difficulty of labeling data in HSIC. Zhang et al. [35] proposed a spectral–spatial self-attention network (SSSAN). This network can adaptively combine the local features of the pixels to be classified with long-distance dependencies.
Despite the numerous significant achievements of CNN in the field of HSIC, the limited receptive field has remained a persistent challenge. This problem makes it relatively difficult for networks to learn the dependency relationships of long-distance spatial distributions. Although this problem can be alleviated by using convolution kernels of different sizes, it also increases the complexity of the model and may lead to overfitting during network training.
Transformer, a deep learning model initially proposed by Vaswani et al. in 2017, was originally applied in the field of NLP and achieved remarkable results [36]. However, the early application of Transformer in the field of CV faced various challenges, including computational efficiency, model adaptability, and the complexity of training and inference. With the emergence of research achievements such as ViT, LeViT, DeepViT, and others, the application of Transformer in the CV field has been greatly expanded [37,38,39,40,41]. In view of this, some scholars have begun to apply Transformer in the field of HSIC. Hong et al. [42] proposed a backbone network specifically designed for HSIC, known as SpectralFormer (SF). This method can not only learn local spectral representations but also effectively transfers similar components from shallow to deep layers. Mei et al. [43] proposed a group-aware hierarchical Transformer (GAHT), which extracts more detailed local spatial spectral relationships through a group embedding module. Zhong et al. [44] proposed a spectral spatial Transformer network (SSTN), which comprises spatial attention and spectral correlation modules to overcome the limitations of the limited receptive field of convolutional kernels.
The multi-head self-attention mechanism (MHSA) in Transformer can calculate the cross-correlation between all elements in the input sequence, but this also makes it difficult to fully focus on local features. To overcome this limitation, some researchers have attempted to combine Transformer with other models. Song et al. [45] proposed a bottleneck spatial spectral Transformer (BS2T), which combines CNN and Transformer to achieve feature extraction and capture long-distance dependencies. Specifically, CNN is utilized for feature compression and expansion, while Transformer enhances the feature representation by capturing long-distance dependencies. Sun et al. [46] proposed a spectral spatial feature tokenization transformer (SSFTT), which extracts spectral spatial features through CNN and employs a feature tokenizer to perform feature transformation to obtain advanced semantic features. Jiang et al. [47] proposed a graph generative structure aware transformer (GraphGST). This method proposes an absolute position encoding to obtain the absolute position sequence of pixels and integrate it into the Transformer architecture. Feng et al. [48] proposed a central attention transformer (CAT). This method improves classification performance by using superpixel sampling and multi-level random sampling mechanisms, combined with spatial spectral tokens and central attention structures.
In addition, Wang et al. [49] proposed a capsule attention network (CAN). CAN combines activity vectors with attention mechanisms to improve HSIC. Convolution and Transformer consume a lot of computing resources during training. Therefore, Jamali et al. [50] proposed a novel backbone network called HybridKAN. This method achieves faster convergence speed by fusing 1D, 2D, and 3D KAN modules.
Although the above methods have achieved good classification performance, there are still some challenges, including the following:
  • Existing models often only focus on the primary features, neglecting the potential value of secondary features in model learning.
  • The local receptive field of the convolution kernel makes it difficult to obtain global information.
To address these issues, an adaptive feature enhanced gaussian weighted network (AFGNet) is proposed. Firstly, an adaptive feature enhancement module (AFEM) was designed. AFEM can accurately evaluate the impact of different features on classification performance and dynamically enhance features that are more conducive to model learning. Secondly, a Gaussian weighted feature fusion module (GWF2) was proposed to effectively fuse local and global features by establishing mapping relationships between local features. Finally, a multi-head collaborative attention (MHCA) mechanism was developed. Through direct interaction and global modeling, it enhances the model’s ability to extract key features from the sequence data. Extensive experiments are conducted on five challenging datasets. The experimental results demonstrate that the proposed method outperforms several SOTA methods.
The main contributions of this paper can be summarized as follows:
  • An adaptive feature enhancement module (AFEM) is proposed. AFEM can adaptively enhance features that are more conducive to model learning.
  • A Gaussian weighted feature fusion module (GWF2) is proposed. GWF2 effectively integrates local and global features by extracting and constructing mapping relationships between local features.
  • A multi-head collaborative attention (MHCA) mechanism is designed. By direct interaction and global modeling, MHCA can more adequately capture the key features in the input sequence.
The remaining parts of this paper are arranged as follows: Section 2 introduces the three modules of the proposed method in detail. Section 3 first introduces the datasets and experimental settings. Then, detailed experimental verification was conducted on the proposed method. Section 4 provides conclusions and prospects for future research directions.

2. Methodology

Original HSI data are denoted as I h × w × l , where h × w is the spatial size and l is the number of spectral bands.

2.1. Overall Structure

The overall structure of the proposed AFGNet is shown in Figure 1, which mainly consists of three modules: an AFEM module for adaptively enhancing features that are more beneficial for model learning, a GWF2 module for extracting and fusing “spatial-spectral” features, and multiple Transformer encoders (TE) incorporating MHCA to learn relationships among global features. Detailed introductions about all three modules are presented in Section 2.2, Section 2.3 and Section 2.4.
First, the HSI data I are preprocessed. The HSI data I contain a vast amount of spectral bands, offering valuable spectral information while introducing redundancy. To mitigate this, PCA [51] is applied to reduce the dimensionality of the original HSI data, thereby decreasing computational complexity and preserving essential spectral features. The HSI data after PCA dimension reduction are denoted as I pca h × w × b , where b is the number of spectral bands after PCA. Next, 3D-patch extraction is performed on the HSI data I pca . Each patch P s × s × b covers a window size of s × s . The central pixel of each patch is set to x c 1 × 1 × b . The true label of a patch is determined by the label of its central pixel.

2.2. AFEM Module

The features with strong correlation in the input feature map are considered as the primary features. The features other than these are considered as secondary features. In the field of HSIC, the phenomenon of different objects with similar spectra leads to the existing attention mechanisms inevitably introducing noise from inter-class spectral similarities while enhancing the representation of primary spectral features of target objects. Moreover, the coexisting phenomenon of the same object with different spectra further exacerbates the issue of feature confusion. Due to the suppression effect of attention mechanisms on secondary discriminative features, it is difficult for the model to effectively capture spectral variations within the same category of objects caused by factors such as lighting conditions and phenological changes. To address this issue, an AFEM is proposed, with its detailed structure shown in Figure 2.
Firstly, the central spectral pixel x c is extracted from the input patch P . Next, cosine similarity is employed to assess the similarity between the central pixel and the surrounding pixels in patch P , resulting in an initial attention map M s × s . This process can be represented as
M = F cos x c , P = x c × x i , j x c × x i , j
Here, x i , j denotes the surrounding pixel, where 0 < i , j < s . F cos represents the cosine similarity function.
Then, a threshold matrix with the same shape as the initial attention map M is constructed. The value of this matrix is determined by the product of the mean of M and the preset threshold coefficient λ. Next, based on this threshold matrix, the initial attention map is divided into a primary feature attention map M h and a secondary feature attention map M l . Following this, an adaptive adjustment strategy is adopted to process the two types of attention maps. Specifically, we introduce two parameters, α and β, and incorporate them into the backpropagation process of the network.
By calculating gradients, we can quantify the contribution of different features to model learning. The positive gradient indicates that this feature contributes to the improvement of model performance, thereby prompting the model to pay higher attention to it in subsequent training.
Finally, the two types of attention maps are fused again to obtain an adaptively enhanced attention map, denoted as M . This process can be represented as
M h , M l = Split M λ × AvgPool M
M = α M h + β M l
Here, α and β are two parameters with gradients. Based on attention map M , the weighted feature map X s × s × b can be expressed as
X = σ M P + P = 1 1 + e M P + P
Here, σ denotes the Sigmoid function and denotes the element-wise multiplication.

2.3. GWF2 Module

The limited receptive field of convolution kernels means that each kernel can only capture local information within its direct area of effect, which restricts the network’s ability to capture global information. To address this issue, a Gaussian weighted feature fusion module (GWF2) was designed, and its detailed structure is shown in Figure 1.
The data enhanced by AFEM will be used as input for this section. Firstly, X are processed by 3D convolutional block to jointly extract “spatial-spectral” features. The convolution outputs feature maps with c channels. These feature maps will be further abstracted and fused in subsequent operations to obtain higher-level representations. Afterwards, the channel dimension and spectral dimension are merged, denoted as X s × s × q , where q = b × c . This process can be expressed as
X = Reshape F τ F BN F 3 × 3 X
Here, F 3 × 3 represents a 3D convolutional layer with a kernel size of 3, F BN represents a batch normalization layer, and F τ represents the Mish function.
We then model the relationships between channels using a mapping operation, allowing features from each channel to be weighted by features from other channels. Subsequently, we fuse the original input with the weighted result and normalize the fused features using a Sigmoid function. The result is denoted as A s × s × q . This process can be expressed as
X = Dropout ReLU F spe X
A   = Reshape 1 1 + e X + X
Here, X represents the data after global spectral fusion. F spe represents a linear layer initialized with a Gaussian distribution [52]. Next, 2D convolutional block is used to extract spatial features from the feature map. Then, the spatial dimension is flattened. The result is denoted as B p × q . This process can be represented as
B = ReLU F BN F 3 × 3 A
B = Flatten Dropout B
Here, B represents the features after convolution processing, and B represents the data after B is flattened. F 3 × 3 represents a 2D convolutional layer with a kernel size of 3.
Similarly, we model the relationships between spatial positions using a mapping operation, enabling features at each position to be weighted by features from other positions. Following this, we perform two feature fusions on the original input and merge the fused result with the weighted result. The result is denoted as T p × p . This process can be represented as
B = Dropout ReLU F spa B
A = Reshape F spa F spe A
T = Reshape A + B
Here, B represents the result after global spatial fusion, and A represents the skip connection data after global spectral and spatial fusion. F spa adopts the same weight initialization as F spe .
GWF2 extracts the local structure of feature maps through convolution, while establishing dependencies between local features using mapping operations, thereby achieving more effective feature extraction.

2.4. MHCA Mechanism

By dividing the sequence into multiple heads and computing attention maps in parallel, multi-head self-attention (MHSA) can obtain multiple forms of representations, but the sequence information that each head focuses on is limited. To alleviate this problem, a multi-head collaborative attention (MHCA) mechanism is proposed, and its detailed structure is shown in Figure 3.
The sequence T processed by GWF2 will be used as the input for this section. We concatenate this sequence with a learnable semantic token T cls , which will be used for the final classification prediction. Next, mark the position information of each element in the sequence through positional encoding. Afterwards, layer normalization (LN) is utilized to ensure the stability of the feature distribution. This result is recorded as T in p + 1 × p . This process can be expressed as
T in = F LN ( Concat ( T cls , T 1 , T 2 , , T p ) + PE )
Here, F LN denotes the LN layer and PE represents the positional embedding.
To preserve more original information in the Query, we adopt an independent mapping strategy, where Key and Value are segmented from the mapping result of the same linear layer. Next, we perform a dot-product operation between Query and the transpose of Key, and then scale by the square root of the dimension of Key. Following this, the Softmax function is employed to convert the results into attention scores. Finally, we perform a matrix multiplication between the attention scores and Value to obtain the output of a single head. By concatenating the outputs of all heads, the result is denoted as T p + 1 × p . This process can be represented as follows:
Q , K , V Linear ( T in ) , Linear ( T in )
S A = Attention ( Q , K , V ) = Softmax Q K T d K V
T = Multi - head ( Q , K , V ) = Concat ( S A 1 , S A 2 , , S A h ) W
Here, SA represents the weighted result with an attention head. W represents the parameter matrix. To better capture the global dependencies in the sequence, we use global max pooling and global average pooling to compress the channel dimensions in the original sequence, which can be expressed by the following formula:
U = Reshape MaxPool T in + AvgPool T in
The result is denoted as U . Then, we project the compressed feature into a low-dimensional space using a fully connected layer and introduce non-linearity with the Mish function. Subsequently, a linear transformation is applied to restore the original dimension, and a sigmoid gating function is employed to obtain attention weights, thereby enabling fine-grained modeling of global information. Finally, we perform an element-wise multiplication of sequence and attention weights, resulting in the final output. This part is inspired by the work of SE, and the process can be represented as:
U = Linear ( F τ ( Linear ( U ) ) )
T = T Reshape 1 1 + e U
After these operations, each attention head not only captures more sequence information but also undergoes a quadratic re-weighting process involving all sequence elements, enabling the model to capture global dependencies more accurately. Subsequently, T undergoes (LN), a multi-layer perceptron MLP, and residual connection, as illustrated in Figure 1. After repeating this encoder block multiple times, we extract T cls for final classification.

3. Experiments

To verify the effectiveness of the proposed method, extensive experiments were conducted on five publicly available classic datasets. To avoid the randomness of the experiment, the average results of 10 repeated experiments were used for all experimental data.

3.1. Datasets

To validate the performance of the proposed method, this paper conducted extensive experiments on five publicly available datasets: Indian Pines (Indiana, USA), Pavia University (Pavia, Italy), Houston 2013 (Houston, TX, USA), Longkou (Jingzhou, China), and LaoYuHe (Kunming, China).
The Indian Pines (IP) dataset was imaged in June 1992 in Indiana, USA, and captured by an airborne visible infrared imaging spectrometer (AVIRIS) (NASA, Washington, DC, USA). The size of the dataset is 145 × 145, with a spatial resolution of approximately 20 m and a wavelength range of 0.4–2.5 nm. It contains 16 ground object categories and 224 continuous bands, of which 20 absorbing bands (104–108, 150–163, 220) are removed, leaving 200 bands for training.
The Pavia University (UP) dataset was acquired during a flight over Pavia, northern Italy, using the ROSIS sensor. The dataset has a size of 610 × 340, with a spatial resolution of 1.3 m, and contains 9 ground-object categories and 103 bands.
The Houston 2013 (HT) dataset [53] was imaged in Houston, Texas, and surrounding rural areas in the United States, using the compact airborne spectrographic imager (CASI) 1500 sensor (ITRES, Calgary, AB, Canada). The dataset has a size of 349 × 1905, with a spatial resolution of 2.5 m, and contains 15 ground object categories and 144 bands.
The LongKou (LK) dataset [54,55] was imaged on 17 July 2018 in Longkou Town, Hubei Province, China, using a DJI M600 Pro drone (DJI, Shenzhen, China) equipped with an 8 mm hyperspectral sensor. The size of this dataset is 550 × 400, with a band range of 400–1000 nm and a spatial resolution of 0.463 m. It contains six crop categories and 270 bands.
The LaoYuHe (LYH) dataset [56] was imaged on May 2024 in Laoyuhe Wetland Park, Kunming City, Yunnan Province, China, and collected by OHS satellite. The size of this dataset is 391 × 591, with a wavelength range of 0.4 to 1 μm and a spatial resolution of 10 m. It contains 8 land cover categories and 32 bands.
Table 1 and Table 2 list the land cover class names, the number of training samples, and the number of testing samples for each dataset. For the Indian Pines and Houston 2013 datasets, the number of training samples is 5% of the total samples; for the Pavia University and LaoYuHe datasets, it is 1%; and for the Longkou dataset, it is 0.2%.

3.2. Experiment Details

  • The proposed method was implemented using PyTorch 1.10.1 and its performance was evaluated on an NVIDIA GeForce RTX 3090 GPU (NVIDIA, Santa Clara, CA, USA).
  • This paper uses three common evaluation indicators, namely overall accuracy (OA), average accuracy (AA), and Kappa coefficient (κ × 100), to compare the performance of all methods.
  • To ensure the fairness of the experiment, all compared methods followed the optimal parameter configuration in their respective papers. Figure 4 shows the impact of different patches and learning rates on the classification performance of the proposed method. For the IP and UP datasets, the patch size was set to 13 × 13, and the learning rate was set to 1 × 10−3. For the HT dataset, the patch size was set to 13 × 13, and the learning rate was set to 5 × 10−4. For the LK dataset, the patch size was set to 15 × 15, and the learning rate was set to 5 × 10−3. For the LYH dataset, the patch size was set to 15 × 15, and the learning rate was set to 5 × 10−4. In addition, the StepLR strategy was adopted to adjust the learning rate during training, that is, the learning rate was multiplied by 0.9 for every tenth of the total training epoch. Finally, the network was trained using the Adam optimizer, and the training epoch and batch size were set to 100 and 64, respectively. The threshold coefficient λ in the AFEM module was set to 1.05.

3.3. Ablation for GWF2

In Section 2.3, GWF2 achieves effective extraction of spectral and spatial features by combining a linear layer with a Gaussian weight of the convolution kernel. To verify the impact of spectral and spatial feature fusion on model performance, this section designs a set of comparative experiments. The experimental results are shown in Figure 5. When no feature fusion mechanism is included, i.e., GWF2 consists only of basic convolutional layers, the model’s classification performance on all datasets is at the lowest level. When any feature fusion mechanism is introduced, the classification performance of the model is significantly improved. Furthermore, when the GWF2 module adopts a complete structure, the model achieves the best classification effect. This experimental result demonstrates the effectiveness of spectral/spatial feature fusion.

3.4. Ablation for AFEM and MHCA

This section verifies the effectiveness of AFEM and MHCA. Taking the GWF2 module as the baseline model, the changes in model performance were compared by removing different components. The experimental results are shown in Table 3.
  • In the proposed AFEM module, we achieved adaptive enhancement of features by differentially processing different features. In Case 1 of Table 3, the model only contained the GWF2 module, and the classification performance of the model was the lowest at this time. In Case 2, AFEM was introduced into the model, and it can be observed that the classification performance of the model has been significantly improved. Compared with Case 1, the classification performance of the model at this time improved by 0.49%, 0.32%, 0.39, 0.71%, and 0.63%, respectively, on all datasets. This result strongly proves the effectiveness of AFEM.
  • In the proposed MHCA module, the global dependencies in the features were fully captured through the collaborative work of multiple attentions. In Case 3 of Table 3, the MHCA module was introduced into the model. At this time, the model obtained performance improvements comparable to Case 2. OA was improved by 0.33%, 0.40%, 0.51%, 0.89%, and 0.75% on all datasets, respectively. This result proves the effectiveness of MHCA. In addition, in Case 4, the model included all three modules. At this time, the classification performance of the model was also the best, which proves that the proposed modules can work together effectively.

3.5. Analysis of Learnable Parameters in AFEM

In the AFEM module, by introducing learnable parameters α and β, the input features are differentiated. To verify the effectiveness of the mechanism, 5 sets of experiments were conducted, and the results are shown in Figure 6, where α corresponds to the primary feature and β corresponds to the secondary feature.
It can be observed that the change trends of these parameters show obvious differences in different datasets. Specifically, on the IP and UP datasets, parameter α showed an upward trend, while parameter β related to the secondary feature gradually decreased, which is consistent with the performance of most attention mechanisms. However, the situation is different on the HT dataset. Compared with parameter α related to the primary feature, parameter β obtained a higher weight, indicating that on this dataset, the secondary features that are beneficial to the classification performance of the model were enhanced.
In addition, on the LK dataset, although parameter α obtained a significant gain, parameter β was not overly suppressed, indicating that on this dataset, the model effectively utilizes the important information in the secondary features while maintaining attention to the primary features. Finally, the change trends of these two parameters on the LYH dataset were similar to those on the IP dataset.
These experimental results not only demonstrate that the AFEM module can flexibly process and enhance the most beneficial features according to the characteristics of the current input features, but also emphasize the importance of secondary features in the model learning process because they also contain information that is crucial to improving model performance.

3.6. Quantitative Evaluation

This section compares the proposed AFGNet with some state-of-the-art HSIC methods, including HybridSN [24], SSSAN [35], SPRN [23], CVSSN [26], GAHT [43], ViT [37], SF [42], SSTN [44], SSFTT [46], CAN [49], and HybridKAN [50]. The experimental results are shown in Table 4, Table 5, Table 6, Table 7 and Table 8. Among them, HybridSN, SSSAN, SPRN, and CVSSN adopt CNN-based frameworks; GAHT, SF, and SSTN are based on Transformer; and SSFTT and the proposed method integrate CNN and Transformer to form a hybrid structure. Particularly, SPRN and CVSSN also utilize the idea of adaptive feature enhancement.
Firstly, HybridSN combines 3D convolution and 2D convolution to explore spectral spatial features. However, due to the limited receptive field of traditional convolutional kernels, they did not achieve good classification accuracy on all five datasets. Secondly, SSSAN introduces self-attention mechanisms on top of convolutions, enabling the network to focus on longer-range information. However, its fusion strategy does not fully take effect, resulting in lower classification performance. Afterwards, SPRN divides the input spectrum into multiple sub-bands by equivalently using grouped convolutions. In addition, this method also adopts cosine similarity as a metric to enhance the features of samples that are more similar to the center vector. It can be observed that its classification performance is better among the comparison methods. Furthermore, CVSSN adopts multiple metrics such as Euclidean distance and cosine similarity to enhance the main features, and it also achieves good classification performance. Furthermore, GAHT and SF fully explore the spectral information in HSI, providing new ideas for HSIC. However, due to the lack of extraction of spatial information, its classification performance is not ideal. ViT is a baseline model for GHAT and SF, and its performance and problems are like those of GHAT and SF. SSTN is a recent Transformer-based HSIC method that effectively explores spectral–spatial information through a structured search framework, achieving good classification performance. Finally, methods with hybrid structures generally outperform those that use only CNNs or only Transformers. SSFTT extracts shallow information through CNN, then fully leverages the ability of Transformer to extract high-level semantic features through feature tokens. These two methods fully utilize the advantages of CNN and Transformer, so they both achieve good classification performance. CAN and HybridKAN are new backbone networks in the HSIC field, which provide new perspectives for research in this field. However, their problems are also obvious. When the sample size of categories is unbalanced, the understanding ability of these two methods for some categories is not ideal.
Although the above methods have achieved good classification performance, the proposed AFGNet still achieves better classification performance than other methods on all five datasets. Specifically, on the IP dataset, compared with SPRN, the method with the highest classification accuracy, the proposed method is 0.5%, 1.01%, and 0.57% higher in OA, AA, and κ × 100, respectively; on the UP dataset, compared with SSFTT, the method with the highest classification accuracy, the proposed method is 0.61%, 0.99%, and 0.8% higher in OA, AA, and κ × 100, respectively; on the HT dataset, compared with SSFTT, the method with the highest classification accuracy, the proposed method is 0.92%, 0.79%, and 1% higher in OA, AA, and κ × 100, respectively; on the LK dataset and LYH dataset, the comparison method with the highest classification accuracy is still SSFTT, but the proposed method still maintains a lead in OA, AA, and κ × 100. These experimental results fully demonstrate the superiority of the proposed AFGNet in classification performance.

3.7. Comparison of Classification Maps

To more intuitively display the results of quantitative experiments, this section visualizes the classification maps of all methods on different datasets, and the specific results are shown in Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11, along with the ground truth maps of each dataset. In CNN-based methods, the influence caused by the limitation of the receptive field of convolutional kernels can be clearly observed, resulting in many mixed and discontinuous patches in the classification maps. This is particularly evident in the IP dataset. Additionally, due to the limitations of its fusion strategy, the visualization results of SSSAN are not satisfactory. SPRN exhibits good visualization results on the IP dataset, but performs averagely on other datasets. Particularly, on the LK dataset, there are relatively obvious misclassifications between the “cotton” (green) category and the “broad-leaf soybean” (yellow) category.
In Transformer-based methods, due to the failure of GAHT and SF to fully utilize spatial information, the classification maps of these two methods contain more noise, especially in the IP and UP datasets. The classification maps of ViT also have many misclassifications similar to salt and pepper noise. In contrast, SSTN achieves more ideal classification maps by utilizing its unique structured search strategy. The classification maps of CAN have a large number of misclassifications. HybridKAN has similar problems to CAN, which is consistent with the results of quantitative experiments.
As for the methods with hybrid structures, from the perspective of visualization effects, they are generally superior to the above methods, which is consistent with the quantitative results in Table 4, Table 5, Table 6, Table 7 and Table 8. In particular, the proposed AFGNet method has a classification effect that is closer to the ground truth maps, and the frequency and area of discontinuous patches in unlabeled regions are minimized. On the LK dataset, the classification map generated by the AFGNet method has the clearest edges, further verifying the effectiveness and superiority of this method.

3.8. Comparison of Different Sample Proportions

In this section, OA was used as the evaluation index, and five groups of experiments with different training ratios were conducted on the five datasets of IP, UP, HT, LK, and LYH for the proposed method and some comparison methods. The experimental results are shown in Figure 12. Among them, for the IP and HT datasets, 2.5%, 5%, 7.5%, and 10% of the samples from each category were selected as the training set, respectively; for the UP and LYH datasets, the selected ratios were 1%, 2%, 3%, 4%, and 5%; and for the LK dataset, 0.2%, 0.4%, 0.6%, 0.8%, and 1% of the samples were selected as the training set.
As the sample size increased, the classification performance of most methods improved, but the performance of individual methods showed unexpected fluctuations. Specifically, due to the failure to effectively utilize spatial information, the performance of GAHT and SF methods was limited, and their performance ranked at the bottom among all five datasets. In contrast, the change curve of SSSAN maintained an upward trend, indicating that its fusion strategy is more suitable for training scenarios with larger sample sizes. The change curves of SPRN showed unexpected fluctuations, especially SPRN, whose classification performance on multiple datasets did not improve despite the increase in sample size. The proposed method and SSFTT both performed stably on all datasets and consistently ranked in the top two. It is particularly noteworthy that on the LK dataset, the proposed method showed significant and sustained performance gains. On all five datasets, the proposed method was able to maintain the optimal classification level, whether in scenarios with extremely limited sample sizes or in situations with relatively ample sample sizes, which fully demonstrates the superiority of the proposed method.

3.9. t-SNE Feature Visualization

To comprehensively evaluate the performance of the proposed method, this section adopts the t-SNE [57] algorithm to visually analyze the feature distributions learned by AFGNet and three comparison methods (SPRN, SSTN, SSFTT). Test samples from each dataset were selected for display, and the results are shown in Figure 13 and Figure 14.
Among the comparison methods, SSFTT exhibited the best feature visualization results, while SPRN and SSTN showed limitations to varying degrees. Specifically, SPRN displayed significant intra-class feature dispersion on multiple datasets, while SSTN suffered from inter-class feature confusion. This phenomenon indicates that these two methods have deficiencies in feature discrimination ability.
In contrast, the proposed method demonstrated significant advantages in visualization effects. On the IP dataset, AFGNet achieved the largest distance between different classes and the mildest intra-class dispersion. On the UP dataset, the proposed method also exhibited the lowest level of inter-class confusion, and features of the same class are basically distributed in the same region, unlike the dispersion observed in SPRN’s performance on the UP dataset. These visualization results not only intuitively demonstrate the superiority of AFGNet in feature learning but also further validate the effectiveness of this method in improving classification accuracy and discrimination.

3.10. Heat Maps Feature Visualization

In this section, heat maps were utilized to visualize the features learned by the model at different stages. The experimental results are shown in Figure 15. In the heat map, compared with the blue area, the red area means that the model pays more attention to this area.
First, the baseline model was used, which does not adopt any modules proposed in this paper. In Figure 15a, it can be observed that the model does not show obvious areas of interest. Then, the AFEM module was introduced. As seen in Figure 15b, more concentrated thermal responses appear in some areas, indicating that the model paid more attention to these areas. Next, the GWF2 module was further merged, which is shown Figure 15c. At this stage, the model not only paid attention to local structures, but also significantly increased its attention to long-distance features. Finally, the model containing all modules was adopted. As shown in Figure 15d, the heat map clearly demonstrates that the model has the strongest feature learning ability.

3.11. Robustness

Spectral data are often affected by various degradations, noises, and changes during the imaging process, which can seriously reduce the data quality and thus affect the performance of the classifier. This section verified the robustness of the proposed method in a noisy environment by adding different proportions of Gaussian noise to the UP dataset. The experimental results are shown in Figure 16, and OA is used as the evaluation indicator. It can be observed that with the increase of the noise ratio, the performance of all methods is affected by varying degrees. Among them, the performance of SSRN and SSFTT is most significantly affected. In contrast, SSTN has the least performance degradation after 10%. For the proposed method, its classification accuracy is obviously also affected to a certain extent, but its performance is still better than other methods.

4. Conclusions

This paper proposes an AFGNet method for HSIC task. Firstly, an AFEM module is proposed. By accurately evaluating the impact of different features, AFEM can dynamically enhance the most beneficial features, leading to improved classification performance. Secondly, a GWF2 is proposed. GWF2 not only effectively extracts local detail information but also efficiently fuses information from different channels, thereby assisting the model in more accurately comprehending features. Finally, we propose a MHCA mechanism. By directly interacting with and globally modeling the input sequence, MHCA can more effectively extract key features. Extensive experimental results demonstrate that, compared to some state-of-the-art methods, the proposed method can provide superior classification performance.
Nevertheless, this method still has some limitations. For example, in the GWF2 module, although global context information is obtained through spatial mapping and spectral mapping to assist feature extraction, this process also introduces additional computational costs. In future work, we will build on this work and explore more lightweight network structures.

Author Contributions

Conceptualization, C.S.; Data curation, C.S. and F.Z.; Formal analysis, L.W.; Methodology, C.S.; Software, F.Z.; Validation, H.P. and F.Z.; Writing—original draft, F.Z.; Writing—review and editing, C.S. and L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 42271409, in part by the Science and Technology Plan Project of Huzhou under Grant 2024GZ36, and in part by the Fundamental Research Funds in Heilongjiang Provincial Universities under Grant 145409207.

Data Availability Statement

Data and code are available at: https://github.com/Isee-max/AFGNet (accessed on 8 February 2025).

Acknowledgments

We would like to thank the handling editor and the anonymous reviewers for their careful reading and helpful remarks.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in Hyperspectral Image and Signal Processing: A Comprehensive Overview of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2018, 5, 37–78. [Google Scholar] [CrossRef]
  2. Mei, S.; Geng, Y.; Hou, J.; Du, Q. Learning hyperspectral images from RGB images via a coarse-to-fine CNN. Sci. China Inf. Sci. 2021, 65, 152102. [Google Scholar] [CrossRef]
  3. Siddique, N.; Paheding, S.; Elkin, C.P.; Devabhaktuni, V. U-Net and Its Variants for Medical Image Segmentation: A Review of Theory and Applications. IEEE Access 2021, 9, 82031–82057. [Google Scholar] [CrossRef]
  4. Khan, U.; Paheding, S.; Elkin, C.P.; Devabhaktuni, V.K. Trends in Deep Learning for Medical Hyperspectral Image Analysis. IEEE Access 2021, 9, 79534–79548. [Google Scholar] [CrossRef]
  5. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef]
  6. Yang, G.; Huang, K.; Sun, W.; Meng, X.; Mao, D.; Ge, Y. Enhanced mangrove vegetation index based on hyperspectral images for mapping mangrove. ISPRS J. Photogramm. Remote. Sens. 2022, 189, 236–254. [Google Scholar] [CrossRef]
  7. Yokoya, N.; Chan, J.C.-W.; Segl, K. Potential of Resolution-Enhanced Hyperspectral Data for Mineral Mapping Using Simulated EnMAP and Sentinel-2 Images. Remote. Sens. 2016, 8, 172. [Google Scholar] [CrossRef]
  8. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef]
  9. Song, W.; Li, S.; Kang, X.; Huang, K. Hyperspectral image classification based on KNN sparse representation. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 2411–2414. [Google Scholar]
  10. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote. Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  11. Pal, M.; Foody, G.M. Feature Selection for Classification of Hyperspectral Data by SVM. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2297–2307. [Google Scholar] [CrossRef]
  12. Yin, H.; Hu, W.; Li, F.; Lou, J. One-step multi-view spectral clustering by learning common and specific nonnegative embeddings. Int. J. Mach. Learn. Cybern. 2021, 12, 2121–2134. [Google Scholar] [CrossRef]
  13. Yuan, Y.; Lin, J.; Wang, Q. Hyperspectral Image Classification via Multitask Joint Sparse Representation and Stepwise MRF Optimization. IEEE Trans. Cybern. 2015, 46, 2966–2977. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph Information Aggregation Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1912–1925. [Google Scholar] [CrossRef]
  15. Li, C.; Sun, W.; Peng, J.; Ren, K. Deep Dynamic Adaptation Network Based on Joint Correlation Alignment for Cross-Scene Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5532312. [Google Scholar] [CrossRef]
  16. Zhang, M.; Li, W.; Zhang, Y.; Tao, R.; Du, Q. Hyperspectral and LiDAR Data Classification Based on Structural Optimization Transmission. IEEE Trans. Cybern. 2022, 53, 3153–3164. [Google Scholar] [CrossRef] [PubMed]
  17. Su, Y.; Chen, J.; Gao, L.; Plaza, A.; Jiang, M.; Xu, X.; Sun, X.; Li, P. ACGT-Net: Adaptive Cuckoo Refinement-Based Graph Transfer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5521314. [Google Scholar] [CrossRef]
  18. Shi, C.; Zhang, X.; Wang, L.; Jin, Z. A lightweight convolution neural network based on joint features for Remote Sensing scene image classification. Int. J. Remote. Sens. 2023, 44, 6615–6641. [Google Scholar] [CrossRef]
  19. Huang, Y.; Peng, J.; Zhang, G.; Sun, W.; Chen, N.; Du, Q. Adversarial Domain Adaptation Network with Calibrated Prototype and Dynamic Instance Convolution for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5514613. [Google Scholar] [CrossRef]
  20. Shi, C.; Chen, J.; Wang, L. Hyperspectral image classification based on a novel Lush multi-layer feature fusion bias network. Expert Syst. Appl. 2024, 247. [Google Scholar] [CrossRef]
  21. Chen, Y.; Zhao, X.; Jia, X. Spectral–Spatial Classification of Hyperspectral Data Based on Deep Belief Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  22. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  23. Zhang, X.; Shang, S.; Tang, X.; Feng, J.; Jiao, L. Spectral Partitioning Residual Network with Spatial Attention Mechanism for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2021, 60, 5507714. [Google Scholar] [CrossRef]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  25. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  26. Li, M.; Liu, Y.; Xue, G.; Huang, Y.; Yang, G. Exploring the Relationship Between Center and Neighborhoods: Central Vector Oriented Self-Similarity Network for Hyperspectral Image Classification. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1979–1993. [Google Scholar] [CrossRef]
  27. Su, Y.; Gao, L.; Jiang, M.; Plaza, A.; Sun, X.; Zhang, B. NSCKL: Normalized Spectral Clustering with Kernel-Based Learning for Semisupervised Hyperspectral Image Classification. IEEE Trans. Cybern. 2022, 53, 6649–6662. [Google Scholar] [CrossRef]
  28. Mei, S.; Chen, X.; Zhang, Y.; Li, J.; Plaza, A. Accelerating Convolutional Neural Network-Based Hyperspectral Image Classification by Step Activation Quantization. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502012. [Google Scholar] [CrossRef]
  29. Yin, H.; Hu, W.; Zhang, Z.; Lou, J.; Miao, M. Incremental multi-view spectral clustering with sparse and connected graph learning. Neural Netw. 2021, 144, 260–270. [Google Scholar] [CrossRef]
  30. Shi, C.; Yue, S.; Wang, L. Attention Head Interactive Dual Attention Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5523720. [Google Scholar] [CrossRef]
  31. Shi, C.; Wu, H.; Wang, L. A Feature Complementary Attention Network Based on Adaptive Knowledge Filtering for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5527219. [Google Scholar] [CrossRef]
  32. Zhang, X.; Zhang, R.; Li, L.; Li, W. Local–Global Cross Fusion Network with Gaussian-Initialized Learnable Positional Prompting for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5532216. [Google Scholar] [CrossRef]
  33. Yang, K.; Sun, H.; Zou, C.; Lu, X. Cross-Attention Spectral–Spatial Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2021, 60, 5518714. [Google Scholar] [CrossRef]
  34. Wu, H.; Shi, C.; Wang, L.; Jin, Z. A Cross-Channel Dense Connection and Multi-Scale Dual Aggregated Attention Network for Hyperspectral Image Classification. Remote. Sens. 2023, 15, 2367. [Google Scholar] [CrossRef]
  35. Zhang, X.; Sun, G.; Jia, X.; Wu, L.; Zhang, A.; Ren, J.; Fu, H.; Yao, Y. Spectral–Spatial Self-Attention Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2021, 60, 5512115. [Google Scholar] [CrossRef]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  37. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  38. Heo, B.; Yun, S.; Han, D.; Chun, S.; Choe, J.; Oh, S.J. Rethinking Spatial Dimensions of Vision Transformers. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 11916–11925. [Google Scholar]
  39. Graham, B.; El-Nouby, A.; Touvron, H.; Stock, P.; Joulin, A.; Jegou, H.; Douze, M. LeViT: A Vision Transformer in ConvNet’s Clothing for Faster Inference. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 12239–12249. [Google Scholar]
  40. Zhou, D.; Kang, B.; Jin, X.; Yang, L.; Lian, X.; Jiang, Z.; Hou, Q.; Feng, J. DeepViT: Towards Deeper Vision Transformer. arXiv 2021, arXiv:2103.11886. [Google Scholar]
  41. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Jiang, Z.; Tay, F.E.H.; Feng, J.; Yan, S. Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 558–567. [Google Scholar]
  42. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification With Transformers. IEEE Trans. Geosci. Remote. Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  43. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral Image Classification Using Group-Aware Hierarchical Transformer. IEEE Trans. Geosci. Remote. Sens. 2022, 60, 5539014. [Google Scholar] [CrossRef]
  44. Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.-S. Spectral–Spatial Transformer Network for Hyperspectral Image Classification: A Factorized Architecture Search Framework. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5514715. [Google Scholar] [CrossRef]
  45. Song, R.; Feng, Y.; Cheng, W.; Mu, Z.; Wang, X. BS2T: Bottleneck Spatial–Spectral Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2022, 60, 5532117. [Google Scholar] [CrossRef]
  46. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  47. Jiang, M.; Su, Y.; Gao, L.; Plaza, A.; Zhao, X.-L.; Sun, X.; Liu, G. GraphGST: Graph Generative Structure-Aware Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5504016. [Google Scholar] [CrossRef]
  48. Feng, J.; Wang, Q.; Zhang, G.; Jia, X.; Yin, J. CAT: Center Attention Transformer with Stratified Spatial–Spectral Token for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5615415. [Google Scholar] [CrossRef]
  49. Wang, N.; Yang, A.; Cui, Z.; Ding, Y.; Xue, Y.; Su, Y. Capsule Attention Network for Hyperspectral Image Classification. Remote. Sens. 2024, 16, 4001. [Google Scholar] [CrossRef]
  50. Jamali, A.; Roy, S.K.; Hong, D.; Lu, B.; Ghamisi, P. How to Learn More? Exploring Kolmogorov–Arnold Networks for Hyperspectral Image Classification. Remote Sens. 2024, 16, 4015. [Google Scholar] [CrossRef]
  51. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear Versus Nonlinear PCA for the Classification of Hyperspectral Data Based on the Extended Morphological Profiles. IEEE Geosci. Remote Sens. Lett. 2012, 9, 447–451. [Google Scholar] [CrossRef]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  53. Debes, C.; Merentitis, A.; Heremans, R.; Hahn, J.; Franfiadakis, N.; van Kasteren, T.; Liao, W.; Bellens, R.; Pižurica, A.; Gautama, S.; et al. Hyperspectral and LiDAR Data Fusion: Outcome of the 2013 GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2405–2418. [Google Scholar] [CrossRef]
  54. Zhong, Y.; Hu, X.; Luo, C.; Wang, X.; Zhao, J.; Zhang, L. WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2) benchmark datasets and classifier for precise crop identification based on deep convolutional neural network with CRF. Remote Sens. Environ. 2020, 250, 112012. [Google Scholar] [CrossRef]
  55. Zhong, Y.; Wang, X.; Xu, Y.; Wang, S.; Jia, T.; Hu, X.; Zhao, J.; Wei, L.; Zhang, L. Mini-UAV-Borne Hyperspectral Remote Sensing: From Observation and Processing to Applications. IEEE Geosci. Remote. Sens. Mag. 2018, 6, 46–62. [Google Scholar] [CrossRef]
  56. Wang, Q.; Huang, J.; Meng, Y.; Shen, T. DF2Net: Differential Feature Fusion Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 10660–10673. [Google Scholar] [CrossRef]
  57. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Overall structure of proposed AFGNet. First, the PCA algorithm is used to remove redundant bands. Then, the proposed AFEM module is constructed to distinguish the primary features from the secondary features. Next, the proposed GWF2 module is designed to extract local features with the assistance of global features. Finally, the proposed MHCA module is utilized to extract global features and complete classification prediction.
Figure 1. Overall structure of proposed AFGNet. First, the PCA algorithm is used to remove redundant bands. Then, the proposed AFEM module is constructed to distinguish the primary features from the secondary features. Next, the proposed GWF2 module is designed to extract local features with the assistance of global features. Finally, the proposed MHCA module is utilized to extract global features and complete classification prediction.
Remotesensing 17 00763 g001
Figure 2. Structure of AFEM. First, the central pixel is extracted from the input spectral cube. After that, the similarity matrix between the surrounding pixels and the central pixel is obtained using cosine similarity. Next, the similarity matrix is split into primary and secondary features by a preset threshold and multiplied with the adaptive coefficient. Finally, the attention map is merged and the original feature map is weighted.
Figure 2. Structure of AFEM. First, the central pixel is extracted from the input spectral cube. After that, the similarity matrix between the surrounding pixels and the central pixel is obtained using cosine similarity. Next, the similarity matrix is split into primary and secondary features by a preset threshold and multiplied with the adaptive coefficient. Finally, the attention map is merged and the original feature map is weighted.
Remotesensing 17 00763 g002
Figure 3. Structure of MHCA. The input sequence is divided into two branches. First, the first branch uses maximum pooling and mean pooling to obtain the attention map. Then, the second branch maps the Q branch and the K.V branch independently and uses scaled dot product attention to obtain the weighted feature map. Finally, the attention map of the first branch is used to re-weight the result of the second branch to obtain the final feature map.
Figure 3. Structure of MHCA. The input sequence is divided into two branches. First, the first branch uses maximum pooling and mean pooling to obtain the attention map. Then, the second branch maps the Q branch and the K.V branch independently and uses scaled dot product attention to obtain the weighted feature map. Finally, the attention map of the first branch is used to re-weight the result of the second branch to obtain the final feature map.
Remotesensing 17 00763 g003
Figure 4. Impact of different patch sizes and learning rates on classification performance, (ae): Indian Pines, Pavia University, Houston 2013, Longkou, and LaoYuHe. For the IP and UP datasets, the patch size was set to 13 × 13, and the learning rate was set to 1 ×10−3. For the HT dataset, the patch size was set to 13 × 13, and the learning rate was set to 5 ×10−4. For the LK dataset, the patch size was set to 15 × 15, and the learning rate was set to 5 ×10−3. For the LYH dataset, the patch size was set to 15 × 15, and the learning rate was set to 5 ×10−4.
Figure 4. Impact of different patch sizes and learning rates on classification performance, (ae): Indian Pines, Pavia University, Houston 2013, Longkou, and LaoYuHe. For the IP and UP datasets, the patch size was set to 13 × 13, and the learning rate was set to 1 ×10−3. For the HT dataset, the patch size was set to 13 × 13, and the learning rate was set to 5 ×10−4. For the LK dataset, the patch size was set to 15 × 15, and the learning rate was set to 5 ×10−3. For the LYH dataset, the patch size was set to 15 × 15, and the learning rate was set to 5 ×10−4.
Remotesensing 17 00763 g004
Figure 5. Ablation for GWF2. Fspe and Fspa represent spectral feature fusion and spatial feature fusion, respectively. Markings in the upper left corner represent, from top to bottom, not using Fspe and Fspa, only using Fspa, only using Fspe, and complete GWF2 module (including Fspe and Fspa).
Figure 5. Ablation for GWF2. Fspe and Fspa represent spectral feature fusion and spatial feature fusion, respectively. Markings in the upper left corner represent, from top to bottom, not using Fspe and Fspa, only using Fspa, only using Fspe, and complete GWF2 module (including Fspe and Fspa).
Remotesensing 17 00763 g005
Figure 6. Impact of different patch sizes and learning rates on classification performance, (ae): Indian Pines, Pavia University, Houston 2013, Longkou, and LaoYuHe. Alpha and beta represent adaptive scaling coefficients of primary and secondary features, respectively.
Figure 6. Impact of different patch sizes and learning rates on classification performance, (ae): Indian Pines, Pavia University, Houston 2013, Longkou, and LaoYuHe. Alpha and beta represent adaptive scaling coefficients of primary and secondary features, respectively.
Remotesensing 17 00763 g006
Figure 7. Classification maps obtained by different methods on the IP dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, AFGNet, respectively.
Figure 7. Classification maps obtained by different methods on the IP dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, AFGNet, respectively.
Remotesensing 17 00763 g007
Figure 8. Classification maps obtained by different methods on the UP dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, AFGNet, respectively.
Figure 8. Classification maps obtained by different methods on the UP dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, AFGNet, respectively.
Remotesensing 17 00763 g008
Figure 9. Classification maps obtained by different methods on the HT dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, AFGNet, respectively.
Figure 9. Classification maps obtained by different methods on the HT dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, AFGNet, respectively.
Remotesensing 17 00763 g009
Figure 10. Classification maps obtained by different methods on LK dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, and AFGNet, respectively.
Figure 10. Classification maps obtained by different methods on LK dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, and AFGNet, respectively.
Remotesensing 17 00763 g010
Figure 11. Classification maps obtained by different methods on LYH dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, and AFGNet, respectively.
Figure 11. Classification maps obtained by different methods on LYH dataset, (am): Ground Truth, HybridSN, SSSAN, SPRN, CVSSN, GHAT, SF, SSTN, SSFTT, and AFGNet, respectively.
Remotesensing 17 00763 g011
Figure 12. Comparison of different sample proportions; (ae) IP, UP, HT, LK, and LYH datasets, respectively.
Figure 12. Comparison of different sample proportions; (ae) IP, UP, HT, LK, and LYH datasets, respectively.
Remotesensing 17 00763 g012
Figure 13. t-SNE visualization results obtained by different methods on IP dataset (ad): SPRN, SSTN, SSFTT, Proposed.
Figure 13. t-SNE visualization results obtained by different methods on IP dataset (ad): SPRN, SSTN, SSFTT, Proposed.
Remotesensing 17 00763 g013
Figure 14. t-SNE visualization results obtained by different methods on UP dataset (ad): SPRN, SSTN, SSFTT, Proposed.
Figure 14. t-SNE visualization results obtained by different methods on UP dataset (ad): SPRN, SSTN, SSFTT, Proposed.
Remotesensing 17 00763 g014
Figure 15. Heat map visualization of features obtained by model at different stages on UP dataset. Compared with the blue area, the red area means that the model pays more attention to this area. (a) Baseline. (b) AFEM. (c) AFEM + GWF2. (d) AFGNet.
Figure 15. Heat map visualization of features obtained by model at different stages on UP dataset. Compared with the blue area, the red area means that the model pays more attention to this area. (a) Baseline. (b) AFEM. (c) AFEM + GWF2. (d) AFGNet.
Remotesensing 17 00763 g015
Figure 16. Impact of different proportions of Gaussian noise on each method on UP dataset: SPRN, SSTN, SSFTT, Proposed.
Figure 16. Impact of different proportions of Gaussian noise on each method on UP dataset: SPRN, SSTN, SSFTT, Proposed.
Remotesensing 17 00763 g016
Table 1. Training and test samples numbers for Indian Pines, Pavia University, and Houston 2013.
Table 1. Training and test samples numbers for Indian Pines, Pavia University, and Houston 2013.
NoIndian PinesPavia UniversityHouston 2013
Class NameTrainingTestClass NameTrainingTestClass NameTrainingTest
1Alfafa244Asphalt866545Healthy Grass631188
2Corn N.711357Meadows18618,463Stressed Grass631191
3Corn M.41789Gravel212078Synthetic Grass35662
4Corn12225Trees313033Trees621182
5Grass P.24459Painted M. S.131332Soil621180
6Grass T.37693Bare soil504979Water16309
7Grass P. M.127Bitumen131317Residential631205
8Hay W.24454S. B. B.373645Commerical621182
9Oats119Shadows10937Road631189
10Soybean N.49923 42742,349Highway611166
11Soybean M.1232332 Railway621173
12Soybean C.30563 Parking Lot 1621171
13Wheat10195 Parking Lot 223446
14Woods631202 Tennis Court21407
15B. G. T. D.19367 Running Track33627
16S. S. T.588
-Total5129737Total42742,349Total75114,278
Table 2. Training and test samples numbers for Longkou and LaoYuHe.
Table 2. Training and test samples numbers for Longkou and LaoYuHe.
NoLongkouLaoYuHe
Class NameTrainingTestClass NameTrainingTest
1Corn6934,442Metasequoia555452
2Cotton178357Other Tree Species292853
3Sesame63025Greenhouse Farmland676599
4Broad-leaf soybean12663,086Bare Land212135
5Narrow-leaf soybean84143Water Bodies777625
6Rice2411,830Buildings242432
7Water13466,922Asphalt686735
8Roads and houses147110Pitches2166
9Mixed weed115218
-Total409204,133 34333,997
Table 3. Ablation experiments. (OA is adopted as evaluation indicator, the optimal performance is bolded).
Table 3. Ablation experiments. (OA is adopted as evaluation indicator, the optimal performance is bolded).
CaseComponentsDataset
AFEMGWF2MHCAIPUPHTLKLYH
197.2797.9498.1797.9996.31
297.7898.2698.5698.7296.94
397.6098.2898.5898.5896.68
498.1298.3498.6898.8897.06
Table 4. Classification performance obtained by different methods for Indian Pines dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–16 represents accuracy of each category).
Table 4. Classification performance obtained by different methods for Indian Pines dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–16 represents accuracy of each category).
No.HybridSNSSSANSPRNCVSSNGHATViTSFSSTNSSFTTCANHybridKANProposed
192.1077.9799.7588.6377.9213.9998.5799.5359.770.9322.8088.40
286.6580.3898.0393.0666.7262.2863.9995.0494.7571.0845.8395.88
385.3780.7097.2995.9359.6757.8263.5397.5197.1372.9842.7598.25
491.3580.7297.4893.5676.4849.5368.0794.7991.6423.0133.5396.80
593.3392.9697.3094.8285.2375.5188.3798.0899.8290.9458.7299.85
696.6295.2498.2098.1378.2180.2684.1098.9199.2499.1972.8199.49
788.0074.2695.6576.0775.1954.4282.9972.8997.4025.6031.3997.78
894.0997.1510099.1090.7884.7689.5699.0899.6999.7481.5699.67
948.3856.2679.7578.4462.4240.5773.1989.2478.4260.009.4291.58
1086.7385.9994.8290.6669.6468.1672.7993.4697.0164.5939.7997.26
1191.1789.6798.3094.5175.7769.7569.9397.7999.0376.4757.7298.84
1289.7967.9599.4488.4757.8648.7951.1792.8691.1547.6632.7694.14
1397.3594.1696.8299.4983.9974.1583.5298.7499.9498.67808999.79
1497.7896.1697.2396.8787.6686.0888.9998.4299.1795.0178.3599.95
1597.4585.2898.5290.2268.5747.1674.7394.9692.9973.1944.2598.26
1684.0994.6991.0690.4795.2391.8793.9495.4686.4775.5177.2099.89
OA91.0787.1997.6294.1374.4768.9673.4796.5197.0276.8956.9198.12
AA88.7684.3596.2391.7875.7162.8278.0794.8092.7367.1650.6197.24
κ × 10089.7985.3897.2993.3170.7464.4969.3196.0296.6173.5750.2197.86
Table 5. Classification performance obtained by different methods for Pavia University dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–9 represents accuracy of each category).
Table 5. Classification performance obtained by different methods for Pavia University dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–9 represents accuracy of each category).
No.HybridSNSSSANSPRNCVSSNGHATViTSFSSTNSSFTTCANHybridKANProposed
194.7793.3591.1795.6685.3585.3684.3497.8897.2095.1262.9398.21
299.1697.3499.6798.6190.5288.5193.8197.9799.9297.8684.4799.98
389.7381.9998.0887.9661.5263.2069.7798.1388.330.0048.1387.08
496.1198.5298.6897.5296.5386.7198.8095.9994.082.4776.9194.88
595.6895.4999.9596.6498.4093.7499.0597.9499.4598.8096.6199.90
699.5892.6898.6196.7482.4573.6889.0297.9699.7313.4774.2099.60
794.9186.3179.1091.4171.7369.2468.4999.1299.930.0044.0899.97
890.9986.7487.2689.7377.5678.4577.4387.3092.9389.0850.1596.90
991.5598.9698.9897.5998.0196.2699.6910093.7266.8743.7597.69
OA96.6894.1296.1996.2086.6783.6589.1296.8797.7371.4274.0798.34
AA94.7292.3894.6194.6584.6881.6886.7196.9296.1451.5264.5897.13
κ × 10095.5992.1994.9194.9682.0978.4785.4295.8497.0058.8664.6797.80
Table 6. Classification performance obtained by different methods for Houston 2013 dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–15 represents accuracy of each category).
Table 6. Classification performance obtained by different methods for Houston 2013 dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–15 represents accuracy of each category).
No.HybridSNSSSANSPRNCVSSNGHATViTSFSSTNSSFTTCANHybridKANProposed
196.8594.8397.8196.9394.5191.3296.1890.9397.2485.6292.7499.35
298.3995.1697.1497.7996.7094.0397.7696.7399.3083.5296.3099.08
398.7096.7899.9398.7697.6899.4996.3099.3199.6884.1699.03100
493.2795.0298.2996.2194.9399.0295.2999.2798.7992.0394.3899.78
598.8599.2499.0298.9797.6996.6497.8299.48100.0089.2596.56100
698.2999.2599.2797.0387.3296.1391.0699.5796.7911.2096.4298.05
787.5887.7896.6091.6782.9482.4485.5194.1896.8578.3486.4295.56
897.1992.1098.8298.6188.4878.1183.3196.6692.6551.5788.5895.12
991.8093.6795.8194.6085.7074.5384.5494.8094.3668.8686.5799.33
1095.1678.8293.1794.4979.3082.9183.1986.0499.3151.2089.1899.66
1197.6296.3798.3995.8282.5678.8182.7897.2599.5765.0990.74100
1297.6885.7592.3595.3283.0376.9489.5894.2998.1947.2487.7199.12
1393.1490.8496.8795.8385.9562.5287.5890.3694.820.2291.5793.87
1499.2787.1310099.0388.2491.8395.4298.4010064.1394.3999.95
1597.3699.0310097.5995.3996.1295.9497.8199.4470.6296.44100
OA95.6092.0397.0196.2288.9486.0790.0495.0097.7668.1991.7298.68
AA96.0892.8297.5696.5889.3686.7290.8295.6797.8062.8792.4798.59
κ × 10095.2591.3996.7795.9288.0384.9389.2394.6097.5765.5191.0498.57
Table 7. Classification performance obtained by different methods for Longkou dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–9 represents accuracy of each category).
Table 7. Classification performance obtained by different methods for Longkou dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–9 represents accuracy of each category).
No.HybridSNSSSANSPRNCVSSNGHATViTSFSSTNSSFTTCANHybridKANProposed
198.4698.5398.2399.1998.9694.9294.7898.7599.5999.3879.9299.95
294.2465.1491.1494.7077.2749.8966.8889.2798.5097.5757.5298.53
391.9284.4899.1791.7075.0445.0578.7588.9397.4123.8069.2798.21
498.2594.8698.1497.2894.0185.3493.9897.6999.6994.0080.5799.63
587.8545.4579.1991.3352.8225.5947.6478.8673.810.0560.7985.68
694.7096.5584.6798.6496.4887.9794.0499.6199.0999.7779.1099.25
799.4999.5299.9799.6699.7998.9699.8699.3099.5299.9996.4599.94
889.0391.2395.0590.4169.6569.2664.6983.3285.7682.1762.4589.61
986.7182.6688.4889.1891.5462.7875.5884.6592.9584.2343.5492.42
OA97.3594.2794.1597.6893.7788.1892.1996.7498.3893.7484.0298.88
AA93.4184.2792.6794.6883.9568.8679.5891.1594.0875.6669.9695.91
κ × 10096.5292.4592.6196.9591.8184.2889.7595.7597.8791.7978.5098.53
Table 8. Classification performance obtained by different methods for LaoYuHe dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–8 represents accuracy of each category).
Table 8. Classification performance obtained by different methods for LaoYuHe dataset. (The optimal performance of OA, AA, and κ × 100 are bolded, No. 1–8 represents accuracy of each category).
No.HybridSNSSSANSPRNCVSSNGHATViTSFSSTNSSFTTCANHybridKANProposed
194.8894.5895.9995.1290.2590.2391.8697.8295.1996.4576.1796.70
294.1995.5698.2196.4990.4688.2895.2596.0694.6097.1280.6994.55
397.9393.3894.1294.8484.1973.9577.1797.4494.6891.1273.6499.39
489.8172.2495.7179.7862.4453.1360.4892.2990.4841.1143.6494.07
597.6897.7699.2697.0897.3096.1594.9097.7498.3497.6288.4799.10
685.0977.6973.4582.3956.8351.1259.8686.2488.1450.7841.6191.67
789.5884.2690.8386.5069.0262.6764.7385.8396.8267.0953.2397.47
879.7462.8380.5980.8948.6738.6362.2993.9360.240.1215.4166.08
OA93.8890.1193.5291.8682.4577.7980.7493.8695.1082.7170.7697.06
AA91.1284.7991.0289.1474.8969.2775.8292.1789.8167.6859.1192.38
κ × 10092.6488.1092.2190.2078.7673.2076.6492.5994.0779.0664.5096.46
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, F.; Shi, C.; Wang, L.; Pan, H. An Adaptive Feature Enhanced Gaussian Weighted Network for Hyperspectral Image Classification. Remote Sens. 2025, 17, 763. https://doi.org/10.3390/rs17050763

AMA Style

Zhu F, Shi C, Wang L, Pan H. An Adaptive Feature Enhanced Gaussian Weighted Network for Hyperspectral Image Classification. Remote Sensing. 2025; 17(5):763. https://doi.org/10.3390/rs17050763

Chicago/Turabian Style

Zhu, Fei, Cuiping Shi, Liguo Wang, and Haizhu Pan. 2025. "An Adaptive Feature Enhanced Gaussian Weighted Network for Hyperspectral Image Classification" Remote Sensing 17, no. 5: 763. https://doi.org/10.3390/rs17050763

APA Style

Zhu, F., Shi, C., Wang, L., & Pan, H. (2025). An Adaptive Feature Enhanced Gaussian Weighted Network for Hyperspectral Image Classification. Remote Sensing, 17(5), 763. https://doi.org/10.3390/rs17050763

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop