Next Article in Journal
Assessing Future Ecological Sustainability Shaped by Shared Socioeconomic Pathways: Insights from an Arid Farming–Pastoral Zone of China
Previous Article in Journal
Analysis of Despeckling Filters Using Ratio Images and Divergence Measurement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MGCET: MLP-mixer and Graph Convolutional Enhanced Transformer for Hyperspectral Image Classification

by
Mohammed A. A. Al-qaness
1,2,*,
Guoyong Wu
1 and
Dalal AL-Alimi
3,4
1
College of Physics and Electronic Information Engineering, Zhejiang Normal University, Jinhua 321004, China
2
Zhejiang Institute of Optoelectronics, Jinhua 321004, China
3
School of Computer Science, China University of Geosciences, Wuhan 430074, China
4
Faculty of Engineering, Sana’a University, Sana’a 12544, Yemen
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(16), 2892; https://doi.org/10.3390/rs16162892
Submission received: 19 June 2024 / Revised: 21 July 2024 / Accepted: 6 August 2024 / Published: 8 August 2024

Abstract

:
The vision transformer (ViT) has demonstrated performance comparable to that of convolutional neural networks (CNN) in the hyperspectral image classification domain. This is achieved by transforming images into sequence data and mining global spectral-spatial information to establish remote dependencies. Nevertheless, both the ViT and CNNs have their own limitations. For instance, a CNN is constrained by the extent of its receptive field, which prevents it from fully exploiting global spatial-spectral features. Conversely, the ViT is prone to excessive distraction during the feature extraction process. To be able to overcome the problem of insufficient feature information extraction caused using by a single paradigm, this paper proposes an MLP-mixer and a graph convolutional enhanced transformer (MGCET), whose network consists of a spatial-spectral extraction block (SSEB), an MLP-mixer, and a graph convolutional enhanced transformer (GCET). First, spatial-spectral features are extracted using SSEB, and then local spatial-spectral features are fused with global spatial-spectral features by the MLP-mixer. Finally, graph convolution is embedded in multi-head self-attention (MHSA) to mine spatial relationships and similarity between pixels, which further improves the modeling capability of the model. Correlation experiments were conducted on four different HSI datasets. The MGEET algorithm achieved overall accuracies (OAs) of 95.45%, 97.57%, 98.05%, and 98.52% on these datasets.

Graphical Abstract

1. Introduction

Instead of the conventional panchromatic and multispectral images used for remote sensing, hyperspectral images (HSI) have a wealth of spectral information [1]. This makes it easier to determine the unique features of different objects. Consequently, HSIs are extensively employed in a multitude of fields, including target identification in geological exploration [2], agriculture [3], urban ecological research [4], and mineral exploration [5]. Hyperspectral data are usually represented in the hyperspectral domain by a 3D cube X C × H × W , in which C, H, and W are the spectral dimension, height, and width of the hyperspectral data, respectively. Typically, each pixel in an HSI has hundreds of spectral dimensions. This high-dimensional information represents both the improvement in classification accuracy and the phenomenon known as the “curse of dimensionality”.
HSI classification represents a crucial aspect of hyperspectral remote sensing Earth observation technology. In recent decades, a multitude of HSI classification methods have emerged. Early machine learning methods primarily focused on pixel classification by mining spectral information from HSIs. Examples include K-nearest neighbor methods (KNN) [6,7], support vector machine (SVM) with radial basis function (SVM-RBF) [8,9,10], logistic regression (LR) [11,12,13], sparse representation (SR) [14,15], maximum likelihood estimation (MLE) [16], and other methods. However, most of the above methods require manual design of the algorithms, and thus the quality of the features depends heavily on the experience and expertise of the designer. Fortunately, deep learning overcomes the shortcomings of the above methods to provide a more desirable approach for HSI classification. Deep learning methods such as auto-encoders (AEs) [17], convolutional neural networks (CNNs) [18,19,20,21], recurrent neural networks (RNNs) [22], graph convolutional networks (GCNs) [23,24,25], and Transformer [26,27] are widely used in HSI classification.
CNNs have become the most popular HSI classification method [28,29], due to their properties, such as parameter sharing, local awareness, translation invariance, and multiple kernels. Hu et al. [30] designed a model with five layers of one-dimensional convolutional. However, since one-dimensional convolutional only focuses on the spectral dimension, and not spatial ones, Hamida et al. [31] proposed a HSI classification model using three-dimensional convolutional that takes into account both spectral and spatial extractions. In some cases, deep CNN models are more effective at extracting spectral-spatial predictors in HSI. However, the danger of overfitting and gradient vanishing is elevated when the model’s depth is increased. To solve the aforementioned problem, Zhong et al. [32] designed spectral and spatial residual blocks and then proposed a residual CNN model that fuses spectral and spatial information. In [33], the authors suggested a residual pyramid network that increases the diversity of spectral-spatial features from shallow to deep. Typically, CNNs are based on patches as inputs, so they cannot fully utilize the global information and multi-scale features of HSIs. For this reason, Meng et al. [34] suggested a fully dense multi-scale fusion network (FDMFN). In order to pay more attention to the more important information in the spectral and spatial domains, Zhu et al. [35] applied an attention mechanism to residual CNNs for both spectral and spatial attention, to eliminate useless bands and focus on important features. Li et al. [36] suggested a double-brand dual-attention mechanism network, which uses spectral and spatial attention modules, and finally performs feature fusion. Additionally, there have also been other unsupervised methods used for HSI classification, such as spatial-spectral masked auto-encoder [37], nearest neighbor-based contrastive learning [38], and adaptive dropblock-enhanced generative adversarial [39] networks.
Transformer was initially applied to the NLP field and quickly became the dominant model in this field, due to its powerful modeling capabilities. Its success has received significant attention from a large number of researchers. Vision transformer (ViT) [40] treats a picture as a complete sentence, and each patch is a word; in this way, it has successfully migrated from the NLP field into the CV field, while achieving excellent classification results. As a result, a large number of ViT-based approaches have emerged in the domain of HSI classification. Hong et al. [41] proposed a new ViT backbone network from the timing perspective, which was able to learn and capture the local spectral sequence information from HSI neighboring bands and achieved better classification results than the traditional ViT. A spectral-spatial feature tokenization transformer (SSFTT) [42] transforms surface spatial-spectral features acquired by CNN networks into high-level semantic features to improve classification accuracy. The group-aware hierarchical transformer (GAHT) proposed by Mei et al. [43] employs grouped pixel embedding (GPE) to extract multi-dimensional spatial-spectral features and combines these with multi-head self-attention (MHSA) to overcome the shortcomings of the original MHSA, which has too much dispersed attention. Inception transformer (IFomer) proposed by Ren et al. [44] uses a channel segmentation mechanism to fully utilize the local information in the high-frequency channel and the global information in the low-frequency channel. Yang et al. [45] linearly combined GCN with transformer to fully use the contextual information of input data, while fusing local spatial-spectral features. In [46], the authors presented a multiple vision architecture-based hybrid network (MVAHN). This model effectively combined multiple types of feature information through the flexible combination of GCN, CNN, and ViT.

Motivation and Contribution

Although the above model architectures achieved excellent HSI classification results, there is still room for discussion and improvements. Whether a CNN-based model or a ViT-based model, they usually focus on a single type of feature and cannot fully utilize the rich spectral information in HSIs. For instance, the capacity to obtain features in CNNs is constrained by the dimensions of the receptive field. In contrast, ViT is more inclined to prioritize global features and exhibits less sensitivity to the degree of similarity between sequential data. The majority of contemporary hybrid networks merely combine diverse network paradigms, without access to crucial spatial-spectral features. Furthermore, they fail to fully consider the characteristics of different network paradigms and the deep integration between them. This results in an inability to fully utilize the advantages of various paradigms.
To address these issues, this study proposes a novel hybrid network architecture that effectively harnesses the strengths of MLP, GCN, and ViT to enhance HSI classification. This hybrid network excels at mining essential spatial-spectral data and significantly impacts classification outcomes by utilizing an MLP-mixer and incorporating graph convolution into self-attention to establish interconnections between sequential data. The model consists of three primary modules: a spatial-spectral extraction block (SSEB), an MLP-mixer, and a graph convolutional enhanced transformer (GCET). The SSEB module initiates the process by extracting deep spectral-spatial feature information from the HSI data. The MLP-mixer module follows this by extracting crucial spatial-spectral features and suppressing less significant information, without altering the shape of the feature map. In the GCET module, the GCN, CNN, and ViT are deeply integrated, with each model leveraging its respective strengths to identify the correlations between local and global features, as well as sequence data.
The novelty of this study lies in its comprehensive integration of diverse network paradigms to fully utilize their respective strengths, ensuring a more effective and thorough exploitation of spatial-spectral features for improved HSI classification. The main contributions can be highlighted as follows:
  • A spatial-spectral extraction block (SSEB) module for efficient extraction of spatial-spectral features is suggested, to extract deep spatial-spectral features and localized spatial-spectral features using a 3D-convolution module and a 2D-convolution module, respectively.
  • The spatial spectral features obtained from the SSEB are further mined using the token-mixing MLP module and channel-mixing MLP module of the MLP-mixer, respectively, and feature maps containing more information are obtained under the condition that the output feature maps are guaranteed to be unchanged from the output.
  • The graph convolutional enhanced transformer (GCET) introduces graph convolutional into the ViT, which overcomes the shortcomings of MHSA’s overly dispersed attention, while fully exploiting the spatial relationships and similarities between pixels.
  • MGCET was subjected to a large number of ablation and comparison experiments on four datasets. The results showed that MGCET had good classification accuracy compared to several state-of-the-art approaches.

2. Methodology

In this section, the MLP-mixer and graph convolutional enhanced transformer for hyperspectral image classification are introduced. Section 2.1 describes the overall network structure. Section 2.2, Section 2.3 and Section 2.4 provide the basic structure and principles of the spatial-spectral extraction block module, MLP-mixer module, and graph convolutional enhanced transformer, respectively. The flowcharts, algorithms, and working steps of the models can be found in Section 2.5.

2.1. Overview of MLP-mixer and Graph Convolutional Enhanced Transformer

In this study, a multi-paradigm fusion network (i.e., MGCET) is presented. The MGCET consists of three modules: the SSEB, the MLP-mixer, and the GCET. The SSEB and MLP-mixer are employed to acquire local-global spatial-spectral features, and the GCET improves the modeling of similar relationships between pixels by embedding graph convolution. The specific structure of the MGCET is presented in Figure 1. Taking Indian Pines as an example, first, a certain percentage (e.g., 5%) of pixels from each type of object in the HSI is selected according to a class-randomization-based method, and based on this, the HSI is segmented into 3D-patches with the same size, and then the local spatial-spectral features are extracted by the SSEB. Second, the 3D data are transformed to 2D data by the rearrange and flatten layer, then projected onto the input dimension of the MLP-mixer using a linear layer. The position information of the sequence data is obtained by adding position embedding, and then different types of features are successively extracted by the MLP-mixer and GCET. Finally, a classifier is employed to obtain the final classification results.

2.2. Spatial-Spectral Extraction Block

To extract deep localized spatial-spectral features, we designed a spatial-spectral extraction block (SSEB), as shown in Figure 2. Specifically, the SSEB is divided into a 3D convolution block and a 2D convolution block. The input patch X is first entered into the 3D convolutional layer, where the channel dimension is increased through the use of a 11 × 3 × 3 3D convolutional layer. The spatial-spectral features are then extracted through the use of a 3 × 3 × 3 3D convolutional layer. Finally, the previously obtained features are reused by stacking the previously obtained features. The expression can be written as follows:
Z o u t = 3 D C o n v 3 C o n c a t 3 D C o n v 1 X , 3 D C o n v 2 3 D C o n v 1 X
where Z o u t denotes the output of the 3D convolutional model. Concat denotes the stacking of features on the channel dimension. The three 3D convolutional layers 3 D C o n v 1 , 3 D C o n v 2 , and 3 D C o n v 3 are the three 3D convolutional layers in Figure 2a, in that order.
In the case of the subsequent 2D convolution, the output of the 3D convolution must be rearranged in order to transform its dimensions. Similarly to in the 3D convolution module, the spectral dimensions of the input features are first increased using 1 × 1 convolutional layers. Secondly, local features are extracted using 3 × 3 depth-wise convolution (DWConv), and finally, the local spatial-spectral features are fully extracted by stacking the previously obtained features. The expression is as follows:
Z P W = P W C o n v 1 r e a a r a n g e Z o u t Z D W = D W C o n v 1 Z P W Z s s e b = P W C o n v 2 C o n c a t Z P W , Z D W
where Z s s e b , Z P W , and Z D W represent the output of SSEB, PW layer, and DW layer, respectively. r e a r r a n g e refers to the transform input feature dimension. P W C o n v 1 denotes the point-wise convolution layer with output channel 256. P W C o n v 2 output channel is the same as the spectral dimension of the patch. C o n c a t indicates that the outputs of P W C o n v 1 and D W C o n v are stacked in the channel dimension.

2.3. MLP-mixer

Deep learning network architectures contain two main types of feature extraction layers: (i) mixing features for a given spatial location, (ii) mixing features between different spatial locations, or both at the same time are functions of different layers in modern deep learning architectures. For instance, CNNs can be implemented with N × N (N > 1) convolution and pooling for (ii), with 1 × 1 convolution for (i), or using larger convolution kernels for (i) and (ii). In ViT or other approaches based on attentional mechanisms, the realization of (i) and (ii) can be achieved through self-attention. The particular structure can be visualized in Figure 3.
The MLP-mixer takes sequence data X N × D as input data, where N refers to the sequence data length and D denotes the number of sequence data channels. The multilayer perceptron (MLP) component within the MLP-mixer model comprises two fully connected layers and an activation function designated as GELU. Two MLP modules are employed: a token-mixing MLP to act on the columns of X to map R N to R N , and a channel-mixing MLP to act on the rows of X to map R D to R D . Additional components include skip-connections, dropout, and layer normalization applied to the module. The MLP-mixer expression is expressed as follows:
Z t o k e n = X + W 2 σ W 1 L a y e r N o r m X Z m i x e r = Z t o k e n + W 4 σ W 3 L a y e r N o r m Z t o k e n
where W 1 , W 2 , W 3 , and W 4 denote the four fully connected layers, and σ denotes the activation function GELU. X, Z t o k e n , and Z m i x e r denote the input data, the output of the token-mixing MLP, and the output of the MLP-mixer, respectively.

2.4. Graph Convolutional Enhanced Transformer

An undirected graph G = V , E can be used in GCN to describe the relationship between HSI pixels, effectively establishing the spatial relationship between pixels, which is a good complement to the transformer. The graph convolution module we designed mainly consists of the GConv layer and 1-D convolution layer, as depicted in Figure 4e. The spatial similarity between the input sequence data is first mined by the GConv layer, and then the spectral features are extracted using the 1-D convolutional layer. This process can be briefly represented as
Z g c m = C o n v 1 d G C o n v Z i n p u t
where Z g c m and Z i n p u t denote the output results and input data of the graph convolution module, respectively.
The primary structure of the module is presented in Figure 4b, where the input image patches X = X 1 , , X d are initially transformed into sequence data P = P 1 , , P d R D × C by executing a flatten and linear projection operation, a process that can be represented by
Z s e q = P r o j e c t i o n F l a t t e n Z s s e b
where D refers to the length of the sequence data and C indicates the number of channels of the sequence data. Z s e q and Z s s e b are the transformed sequence data and the output of the SSEB module, respectively. The location information of the image patch is of paramount importance for ViT, necessitating the addition of location information to the sequence data P, in order to obtain the input data Z m i x e r = M i x e r Z s e q + Z p o s . Here, M i x e r denotes the MLP-mixer module. The output of MHSAG can be expressed in the following form:
Z g e t = M H S A G L N Z m i x e r + Z m i x e r
where MHSAG represents the multi-head self-attention (MHSA) of the embedded graph convolution, as shown in Figure 4c. Graph convolution embedding attention mechanisms (GCEA) improve the original self-attention (SA) layer of ViT by utilizing a GCN to boost the feature extraction capability of ViT, and the specific structure is illustrated in Figure 4d. Specifically, the input data P are first projected to higher channel dimensions using convolution and then averaged into four feature vectors having the same channel dimensions, Q, K, V, and G. The GCEA layer is expressed as
Z a t t n = G C E A Q , K , V , G = S o f t max Q K T d + Z g c m
where Z a t t n denotes the output of GCEA. Multiple GCEAs can be spliced into a larger attention matrix, e.g., when the heads are equal to 4, this can be denoted as Z a t t n h e a d s h e a d s = 1 , , 4 . To maintain the consistency between the input data and the output data, Z a t t n h e a d s can be mapped to match the dimensions of input features using a 1-D pooling layer.
Lastly, the RBB (residual bottleneck block) is utilized to replace the original MLP module in ViT. The feature vectors at each position are nonlinearly transformed so that more nonlinear features are introduced, while reducing the parameters of the network to enhance the effect of feature extraction. The specific structure is shown in Figure 4a. The complete feature extraction process of the graph convolutional enhanced transformer encoder can be summarized as follows:
Z g e t = R B B L N Z g e t + Z g e t
The output of the entire graph convolutional enhanced transformer encoder is represented by Z g e t . In summary, the module offers the following enhancements compared to pure ViT: (1) graph convolution is introduced in the MHSA layer to enhance the feature extraction capabilities of ViT; (2) convolutional projection, instead of pure linear transformation, is performed for the input sequence data; (3) the use of a residual bottleneck block in place of an MLP decreases the number of network parameters; and (4) the use of a 1-D pooling layer ensures that the dimensions of the inputs and outputs features are consistent.

2.5. System Model

The specific parameter settings and flowchart of MGCET are presented in Table 1 and Figure 5, respectively. The overall workflow of MGCET is summarized into the following eight steps:
1: 
Split HSI dataset X R H × W × C into 3D-patches with the same size.
2: 
Spatial-spectral extraction block is utilized to obtain feature Z s s e b .
3: 
Flattened and projected feature Z s s e b to obtain sequence data Z s e q .
4: 
The spatial and channel features are further extracted using a token-mixer MLP and channel-mixer MLP, respectively, to obtain the feature Z m i x e r .
5: 
The graph convolution module embedded in MHSA is employed to obtain a feature Z g c m , which contains the similarity between the sequence data.
6: 
Output features Z a t t n of MHSAG: S o f t m a x Q K T d + Z g c m .
7: 
Feature Z g e t is the final output of the entire model.
8: 
The final classification result Z o u t is obtained through AvgPooling and a linear classifier.

3. Experiments

This section is divided into four main parts. Section 3.1 describes several well-known public HSI datasets used in the experiment, as well as the specific division of the training and test sets. Section 3.2 describes the model used to conduct the comparison experiments and the associated experimental setup details. Section 3.3 introduces the metrics used to measure the performance of the model. Section 3.4 details the relevant comparison experiments.

3.1. Dataset Description

In order to validate the effectiveness of the proposed model, correlation experiments were conducted on four common hyperspectral datasets. The four datasets utilized in this study were the Indian Pines (IP), Pavia University (PU), Salinas Valley (SA), and Kennedy Space Center (KSC) datasets.

3.1.1. Indian Pines (IP) Dataset

The IP dataset is an HSI image dataset that was collected using an airborne visible infrared imaging spectrometer (AVIRIS) over a site in northwestern Indiana, USA, in June 1992. It covers 145 × 145 (spatial dimensions) for an agricultural area and has 200 spectral bands with a 0.4 to 2.5 μ m wavelength range and 10,249 labeled pixels. The study area has 16 classes, including corn, oats, wheat, and wood. We randomly selected 5% samples from each category of the IP dataset as the training set, with the remainder as the test set. The number of training and testing samples for each category of the IP dataset can be found in Table 2; the colored background indicates the color of the category in the image.

3.1.2. Pavia University (PU) Dataset

The PU dataset covers the city of Pavia and its surrounding areas in Italy. It was captured using a different sensor called the Reflective Optics Spectrographic Imaging System (ROSIS) onboard an airborne sensor platform. The spatial size of the Pavia University dataset is 610 × 340, and its approximate spatial resolution is 1.3 m. Outside of the background pixels (unlabeled pixels), there are 42,776 labeled pixels covering nine classification objects. The dataset has 103 bands (after removing 12 high noise bands). Table 3 gives specific information about the PU dataset.

3.1.3. Salinas Valley (SA)

The Salinas dataset was acquired using the AVIRIS sensor over the Salinas Valley, California. The spatial resolution of the dataset is 3.7 m, its data size is 512 × 217, and it contains 204 continuous bands. There are 16 feature types, including vineyard fields, vegetables, and bare soils. Table 4 details the number of training and testing sets selected, as well as the colors represented by the different categories in the classification map.

3.1.4. Kennedy Space Center (KSC)

AVIRIS acquired the KSC dataset for the Kennedy Space Center in Florida, United States of America, 1996. The spatial dimensions of this image are 512 by 614 (spatial dimensions), and its resolution is 18 m. The precise count of bands is 176 (spectral bands) after eliminating the low signal-to-noise and absorption bands. The studied area comprises thirteen distinct classes, such as slash pine, cabbage palm hammock, willow swamp, and scrub. The category information in the dataset is presented in Table 5, in which the colored background indicates the color of the category in the classification map.

3.2. Experimental Settings

The settings of the experiments were modified to be 0.001, 200, 100, and 11 × 11 for the learning rate, epoch, batch size, and patch size, respectively. All methods used in the experimental comparison were trained ten times to ensure the fairness of the experiments. In the random division step, for small sample datasets such as IP and KSC, in order to ensure that the number of training samples met the training requirements, we randomly selected 5% samples from each category of the IP and KSC datasets as the training set, with the remaining samples as the test set. Similarly, for large sample datasets such as PU and SA, 1% of the samples were used as training objects, with the remaining 99% as test objects [47,48,49]. Furthermore, the means and standard deviations of overall accuracy (OA), average accuracy (AA), and kappa coefficient (kappa) were obtained for evaluation. The cross-entropy loss function was employed to assess the degree of correspondence between the predicted and actual values throughout the training phase. In order to circumvent overfitting, we opted to uniformly utilize adaptive moment estimation (Adam) as the optimizer. The MGCET algorithm was tested against 10 HSI classification algorithms for the classification performance under four publicly available HSIs. All of the above experiments were implemented on a computer equipped with an i5-13400F, RTX 3090, and 32 GB of RAM. The experiments were conducted 10 times using the PyTorch deep learning framework, and the average was taken as the result of the experiments.
(1) 
SVM-RBF: As a traditional machine learning method, it employs a nonlinear mapping through radial basis function (RBF) for extracting the spectral information of HSI and then selects the optimal penalty coefficients and kernel function parameter through a grid search.
(2) 
3D-CNN: The 3D-CNN model comprises two 3D-CNN with 3 × 3 × 3 convolution kernels, three 3D-CNN with 3 × 1 × 1 convolution kernels, one 3D-CNN with 2 × 3 × 3 convolution kernel, four ReLU activation layers, and a fully connected layer.
(3) 
RSSAN (residual spectral-spatial attention network) is comprised of two primary modules: the spectral-spatial attention learning module, and the spectral-spatial feature module. The former incorporates spatial and spectral attention, while the latter comprises two residual spectral spatial attention modules (containing two convolutional layers and one attention layer).
(4) 
DBDA (double-branch dual-attention mechanism network) mainly consists of spectral and spatial branches. The former comprises a series of three-dimensional convolutional layers and channel attention, while the latter comprises a series of 3D convolutional layers and spatial attention.
(5) 
SSRN (spectral-spatial residual network) is comprised of two blocks for spectral residuals and the same for spatial residuals. The former comprises 3D convolutional layers with a 1 × 1 × 7 kernel size and residuals. Similarly, the latter comprises 3D convolutional layers with a 3 × 3 × 128 kernel size and residuals.
(6) 
ViT (vision transformer) divides an image into patches, subsequently employing the linear embedding sequences of these image blocks as input to the transformer for the purpose of training an image classification model in a supervised manner.
(7) 
SSFTT (spectral-spatial feature tokenization transformer) is comprised of three modules: a spectral-spatial feature extraction, a Gaussian weighted feature tokenizer, and a transformer encoder. The MHSA in the transformer encoder has four heads with an embedding dimension of 256.
(8) 
GAHT: Group-aware hierarchical transformer comprises three grouped pixel embedding and three transformer encoders. The heads of the MHSA in the transformer encoder are 8, 4, 2, with embedding dimensions of 256, 128, 64.
(9) 
SpectralFormer: SpectralFormer primarily comprises two modules: the groupwise spectral embedding and the cross-layer adaptive fusion. The embedding dimensions are 256.
(10) 
IFormer: IFormer primarily consists of a ghost module and inception transformer encoder. The inception transformer encode includes high-frequency feature extraction, low-frequency feature extraction, and feature fusion. The embedding dimensions of the transformer encode are 256.
(11) 
MGECT: MLP-mixer and transformer fusion network is mainly composed of a joint GCN and transformer (JGT) structure, the MLP-mixer module, and a spatial-spectral feature extraction module. The JGT consists of a residual bottleneck block and a transformer encoder with graph convolution embedding. The MHSA in the transformer encoder has four heads, with an embedding dimension of 256.

3.3. Performance Evaluation Indicators

We assessed the performance of the different models by computing OA, AA, and kappa via confusion matrices, and the classification effect was positively correlated with these three indices (OA, AA, kappa ∈ [0, 1]). In the following expressions for OA, AA, and kappa, C denotes the number of species sampled in the HSI dataset, M denotes the number of samples, A denotes the confusion matrix for the prediction results, A i i denotes the value on the diagonal in the confusion matrix (the number of correct classifications), and A i j is the number of samples classifying i as j.
Overall accuracy (OA) indicates the proportion of correctly classified samples out of the total samples, reflecting the algorithm’s overall classification performance. The formula for OA is as follows:
O A = 1 M i = 1 C A i i
Average accuracy (AA) represents the average classification accuracy for each category in the dataset, reflecting the model’s performance on individual categories. The formula for AA is as follows:
A A = 1 C i = 1 C A i j / j = 1 C A i j
With an unbalanced number of samples of different classes in the dataset, Kappa is able to comprehensively evaluate the classification effects in terms of both model stability and validity. The expression can be written as follows:
K a p p a = M i = 1 C A i i i = 1 C A i j i = 1 C A j i M 2 i = 1 C i = 1 C A i j i = 1 C A j i

3.4. Experimental Results and Analysis

3.4.1. Results of Comparative Experiments

The performance of MGCET with traditional machine learning, deep learning methods, and new methods proposed in recent years on HSI datasets such as IP, SA, PU, and KSC is presented in Table 6, Table 7, Table 8 and Table 9. Figure 6, Figure 7, Figure 8 and Figure 9 illustrate the classification maps of different models on the same data. As an illustration, Figure 6a depicts the ground truth map for the IP dataset, while Figure 6b–l show the classification maps obtained by different models.
For instance, using the IP dataset, we proportionally selected 5% (512 samples) as the training object and used the remaining 9737 samples as the test set. As illustrated in Table 6, MGCET exhibited a classification accuracy of 97.97% for the 11th category (Soybean-mintill) with the largest number of samples, which was approximately 12% higher than that of RSSAN and GAHT and approximately 5% higher than that of spectral former. Similarly, it was approximately 10% higher than RSSAN and spectral former and approximately 4% higher than GAHT and DBDA in the ninth category (Oats) with the lowest number of samples. The reason why MGCET’s accuracy was better than that of the comparative objects for most of the classified objects was that it was able to fully exploit global and local features by effectively combining multiple network structures, thus improving the recognition ability for difficult-to-classify objects.
Overall, our proposed MGCET outperformed the other methods for OA, AA, and Kappa. Firstly, the SVM and 3D-CNN achieved 76.96% and 77.83%, respectively, which were lower than the other models, indicating the lack of competitiveness of these classical models compared to the other new models. The classification accuracies of ViT and spectral former were 81.96% and 87.76%, respectively, which were approximately 11% and 5% lower than those of hybrid models such as DBDA and SSFTT. This discrepancy was due to the fact that, while ViT has a stronger global feature extraction capability than traditional methods, it lacks the ability to acquire some localized features, which is the reason for the bottleneck in the performance of ViT. The residual CNN models, including RSSAN, DBDA, and SSRN, achieved classification accuracies of 91.66%, 91.75%, and 92.47%, respectively. In comparison, MGCET demonstrated superior classification results. Furthermore, for the other ViT hybrid models (e.g., iFormer, GAHT), MGCET employed the MLP-mixer module to extract the spatial and channel relationship between input pixels, thereby achieving a higher Kappa, OA, and AA.

3.4.2. Image Patch Size Analysis

Figure 10a–d illustrates the classification accuracy of the MGCET in comparison to the benchmark model across a range of patch sizes on the IP, SA, PU, and KSC datasets. As shown, the MGCET had a higher OA than the other models in most cases. On most datasets, the OA of most methods did not show a positive correlation with patch size; instead showing a certain degree of decrease. This discrepancy can be attributed to the fact that, while a larger patch size facilitates the acquisition of spatial information, thereby enhancing classification, the same patch may also contain pixels that are elevated in a multiplicative manner, rendering the identification of the target pixel more challenging. It is noteworthy that the MGCET exhibited the highest OA on the IP and PU datasets when the patch size was 11 × 11 and performed well on the remaining datasets.

3.4.3. Analysis of Training Samples

As a further analysis, we compared the performance of the MGCET with the other models with a different percentage of training samples. The training samples for the IP dataset and the KSC dataset ranged from 3% to 11% of the dataset, respectively, with an interval of 1%. Similarly, the training samples for the PU and SA datasets ranged from 0.4% to 1.6% of the dataset, respectively, with an interval of 0.15%. Figure 11 shows the OA curves of all models on the IP, PU, SA, and KSC datasets.
As illustrated in Figure 11, an increase in the number of training samples for the four datasets was accompanied by an enhancement in the OA for all models. Upon reaching a specific threshold of training samples, the OA of the majority of models began to fluctuate and exhibit a decline or, in some cases, a complete absence of growth. In general, when the number of training samples was held constant, the OA of MGCET (red dot) was consistently higher than that of the other models, particularly for single CNN models (e.g., RSSAN, 3D-CNN) and pure ViT (e.g., ViT, spectral former). This is primarily attributable to the capacity of MGCET to fully extract both local and global spatial-spectral features through the integration of multiple network modules, which compensated for the inherent limitations of a single network structure in terms of feature extraction capabilities. Furthermore, it is worth mentioning that the MGCET possesses certain advantages over other hybrid models, which can be attributed to the integration of the MLP-mixer.
Figure 12 illustrates the accuracy trends for the four datasets during the training phase of the compared models. Unlike the other models, the proposed model demonstrated remarkable stability and maintained a high accuracy throughout the entire training process. This was particularly evident with the more complex IP and PU datasets, where its performance consistently outpaced that of the other models. The sustained high accuracy and stability suggest that the proposed model did not experience overfitting.

4. Discussion

In this section, we designed and conducted various experiments to analyze the rationality of the proposed MGCET from different perspectives.

4.1. Analysis of Different Modules

The Mlp-mixer and graph convolutional enhanced transformer (GCET) are the key modules of the MGCET. In order to demonstrate that the Mlp-mixer and GCET could effectively improve the results of the model, we performed sufficient ablation experiments on the four datasets introduced above and then evaluated the effectiveness of the added modules using the evaluation indicators. Different modules made up these networks. (1)Net-1: ViT; (2) Net-2: ViT + Mlp-mixer; (3) Net-3: GCET; (4) Net-4: GCET + Mlp-mixer.
The evaluation results of the above model are presented in Table 10, where the various added modules had different degrees of gain for MGCET. First, Net-1 obtained the lowest OA, AA, and Kappa on all datasets. Second, Net-2 performed significantly better than Net-1 on the four datasets, and the classification accuracy of the pure ViT was significantly improved after the introduction of the MLP-mixer, which, side by side, illustrates the contribution of the MLP-mixer to the improvement in classification accuracy. This was due to the fact that the MLP-mixer contains both a token-mixing MLP and channel-mixing MLP, which enables the fusion and interaction of spatial and channel features. Net-3 enhances the feature information acquisition ability of the ViT by replacing the encoder of the ViT with a graph-convolution embedded encoder and adding residual bottleneck blocks, thus achieving an excellent classification accuracy. Finally, Net-4 further improves the classification effect based on Net-2 and Net-3. This is primarily due to the fact that the MGCET is capable of extracting both rich spatial-spectral features and local-global features in HSIs with the assistance of multiple paradigms, such as CNN, Transformer, MLP, and GCN, and it organically fuses them to boost the classification accuracy even further.
To ascertain the contribution of graph convolution to the classification accuracy of the MGCET algorithm, this paper conducted relevant comparison experiments between MHSA and MHSAG on four HSI datasets. As illustrated in Table 11, the incorporation of graph convolution into MHSA led to a notable enhancement in the OA, AA, and kappa values for all datasets. For instance, the OA was elevated by approximately 0.5% by MHSAG on the IP dataset. This evidence suggests that MGCET is capable of capturing the inherent similarities between sequence data through the embedding of graph convolution, thereby enhancing the model’s capacity to represent features effectively.

4.2. Analysis of Attention Mechanism

Figure 13 demonstrates the classification accuracy of MGCET under different numbers of encoder layers and multiple self-attention layers. On most of the datasets, the OA did not show a corresponding increase with the increase in the number of encoder layers but instead showed a certain degree of decrease. This suggests that model depth does not contribute to the classification accuracy of the MGCET but rather inhibits it to some extent. The number of heads in MHSA did not significantly impact the classification accuracy of the MGCET. In the majority of cases, the classification accuracy of MGCET initially improved and subsequently declined as the number of heads was increased. Overall, the OA exhibited fluctuations as a function of the number of encoder layers and MHSA, yet all stabilized within a stable range. This was observed for the IP dataset (94.83% to 95.45%), the SA dataset (96.67% to 97.57%), the PU dataset (97.36% to 98.05%), and the KSC dataset (98.11% to 98.52%). As illustrated in Figure 13, the MGCET achieved the most optimal classification outcomes on the IP, PU, and KSC datasets when one layer and four heads were selected and on the SA dataset when one layer and five heads were selected. In order to ensure the consistency of training results, the MGCET selected a single layer and four heads in a uniform manner.

4.3. Analysis of Model Architecture

The MLP-mixer and GCET are the key modules of the network architecture, and the way they are connected to each other, as well as the feature fusion method, have an important impact on the model classification effect. In order to prove the rationality of the network architecture proposed in this paper, we proposed four different network architectures based on MLP-mixer and GCET:
  • Structure-1: parallel and additive fusion;
  • Structure-2: parallel and multiplicative fusion;
  • Structure-3: series and MLP-mixer preceded GCET;
  • Structure-4: Series and GCET preceded MLP-mixer.
Where Structure-3 is the network architecture used in this paper.
As can be seen in Table 12, the four network architectures achieved very good classification results on all four publicly available datasets, demonstrating that the combination of GCET and MLP-mixer possesses strong feature extraction capabilities regardless of whether a parallel or serial network architecture is used. In general, the serial network architecture demonstrated superior performance compared to the parallel network architecture. For example, on the IP dataset, Structure-3 and Structure-4 achieved 95.45% and 95.14%, respectively, while Structure-1 and Structure-2 achieved 94.57% and 94.29%. This phenomenon may be attributed to the fact that the multi-branch architectural design diverts the model’s attention from the crucial information that influences the accuracy of the classification, thereby impairing the model’s classification efficacy. The main role of the MLP-mixer is to mine the key spatial-spectral information that affects the classification effect, while GCET is able to continue to acquire the local-global combination of features on the basis of the information acquired by the MLP-mixer, as well as to capture the similarity between the sequence information with the help of graph convolution. In comparison to Structure-3, Structure-4 differs in the sequence of the MLP-mixer and GCET, which has an impact on the model’s classification accuracy. The Structure-3 used in this paper achieved the best performance on all four public datasets, justifying the network architecture proposed in this paper in terms of classification accuracy. In comparison to Structure-1, Structure-2, and Structure-4, Structure-3, as utilized in this paper, exhibited the optimal performance on all four public datasets, thereby substantiating the proposed network architecture from the standpoint of classification accuracy.

5. Conclusions

In order to compensate for the lack of feature extraction capabilities of any single network paradigm, this paper proposed a hybrid network MLP-mixer and graph convolutional enhanced transformer (MGCET) for HSI classification. The network combines four currently dominant network models, including MLP, GCN, CNN, and ViT. Specifically, the CNN and MLP modules are first utilized for spatial-spectral feature extraction. Then, the feature modeling capability of ViT is improved by embedding graph convolution into MHSA to further enhance the classification accuracy of the model. Experimental evaluation on several HSI datasets showed that the combination of multiple network architectures could indeed boost the accuracy of the method, while having a higher classification accuracy compared to the other methods.
A frequent challenge in multi-network architecture models is the competition between different paradigms, which can negatively impact classification accuracy. To address this issue, future research could investigate enhanced network fusion methods and explore incorporating innovative network architectures such as generative adversarial networks (GAN) and the stacked autoencoder (SAE). Alternatively, replacing the original network architectures may offer a potential solution to mitigate this problem. How to reduce the computational complexity of the model through lightweighting techniques is also an important research direction for us.

Author Contributions

Methodology, M.A.A.A.-q. and G.W.; Software, G.W.; Validation, M.A.A.A.-q. and D.A.-A.; Formal analysis, M.A.A.A.-q. and D.A.-A.; Investigation, M.A.A.A.-q. and D.A.-A.; Data curation, G.W.; Writing—original draft, G.W.; Writing—review & editing, M.A.A.A.-q. and D.A.-A.; Visualization, G.W.; Supervision, M.A.A.A.-q.; Funding acquisition, M.A.A.A.-q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets used in this paper are publicly available, as described in the main text.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Feng, X.; He, L.; Cheng, Q.; Long, X.; Yuan, Y. Hyperspectral and multispectral remote sensing image fusion based on endmember spatial information. Remote Sens. 2020, 12, 1009. [Google Scholar] [CrossRef]
  2. Gao, A.F.; Rasmussen, B.; Kulits, P.; Scheller, E.L.; Greenberger, R.; Ehlmann, B.L. Generalized unsupervised clustering of hyperspectral images of geological targets in the near infrared. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4294–4303. [Google Scholar]
  3. Zhao, X.; Li, W.; Zhang, M.; Tao, R.; Ma, P. Adaptive iterated shrinkage thresholding-based lp-norm sparse representation for hyperspectral imagery target detection. Remote Sens. 2020, 12, 3991. [Google Scholar] [CrossRef]
  4. Sun, G.; Jiao, Z.; Zhang, A.; Li, F.; Fu, H.; Li, Z. Hyperspectral image-based vegetation index (HSVI): A new vegetation index for urban ecological research. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102529. [Google Scholar] [CrossRef]
  5. Ma, Y.; Zhang, Y.; Mei, X.; Dai, X.; Ma, J. Multifeature-based discriminative label consistent K-SVD for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4995–5008. [Google Scholar] [CrossRef]
  6. Ma, L.; Crawford, M.M.; Tian, J. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  7. Ge, H.; Pan, H.; Wang, L.; Liu, M.; Li, C. Self-training algorithm for hyperspectral imagery classification based on mixed measurement k-nearest neighbor and support vector machine. J. Appl. Remote Sens. 2021, 15, 042604. [Google Scholar] [CrossRef]
  8. Okwuashi, O.; Ndehedehe, C.E. Deep support vector machine for hyperspectral image classification. Pattern Recognit. 2020, 103, 107298. [Google Scholar] [CrossRef]
  9. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  10. Ye, Q.; Huang, P.; Zhang, Z.; Zheng, Y.; Fu, L.; Yang, W. Multiview learning with robust double-sided twin SVM. IEEE Trans. Cybern. 2021, 52, 12745–12758. [Google Scholar] [CrossRef] [PubMed]
  11. Haut, J.; Paoletti, M.; Paz-Gallardo, A.; Plaza, J.; Plaza, A.; Vigo-Aguiar, J. Cloud implementation of logistic regression for hyperspectral image classification. In Proceedings of the 17th International Conference on Computational and Mathematical Methods in Science and Engineering, CMMSE, Rota, Spain, 4–8 July 2017; Volume 3, pp. 1063–2321. [Google Scholar]
  12. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2011, 50, 809–823. [Google Scholar] [CrossRef]
  13. Khodadadzadeh, M.; Li, J.; Plaza, A.; Bioucas-Dias, J.M. A subspace-based multinomial logistic regression for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2105–2109. [Google Scholar] [CrossRef]
  14. Ghamisi, P.; Maggiori, E.; Li, S.; Souza, R.; Tarablaka, Y.; Moser, G.; De Giorgi, A.; Fang, L.; Chen, Y.; Chi, M.; et al. New frontiers in spectral-spatial hyperspectral image classification: The latest advances based on mathematical morphology, Markov random fields, segmentation, sparse representation, and deep learning. IEEE Geosci. Remote Sens. Mag. 2018, 6, 10–43. [Google Scholar] [CrossRef]
  15. Luo, F.; Huang, H.; Ma, Z.; Liu, J. Semisupervised sparse manifold discriminative analysis for feature extraction of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6197–6211. [Google Scholar] [CrossRef]
  16. Peng, J.; Li, L.; Tang, Y.Y. Maximum likelihood estimation-based joint sparse representation for the classification of hyperspectral remote sensing images. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1790–1802. [Google Scholar] [CrossRef] [PubMed]
  17. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  18. Hong, Q.; Zhong, X.; Chen, W.; Zhang, Z.; Li, B. Hyperspectral Image Classification Network Based on 3D Octave Convolution and Multiscale Depthwise Separable Convolution. ISPRS Int. J. Geo-Inf. 2023, 12, 505. [Google Scholar] [CrossRef]
  19. Li, H.; Xiong, X.; Liu, C.; Ma, Y.; Zeng, S.; Li, Y. SFFNet: Staged Feature Fusion Network of Connecting Convolutional Neural Networks and Graph Convolutional Neural Networks for Hyperspectral Image Classification. Appl. Sci. 2024, 14, 2327. [Google Scholar] [CrossRef]
  20. Zahisham, Z.; Lim, K.M.; Koo, V.C.; Chan, Y.K.; Lee, C.P. 2SRS: Two-stream residual separable convolution neural network for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  21. Meng, Z.; Zhang, J.; Zhao, F.; Liu, H.; Chang, Z. Residual dense asymmetric convolutional neural network for hyperspectral image classification. In Proceedings of the IGARSS 2022 —2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 3159–3162. [Google Scholar]
  22. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  23. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Cai, W.; Yu, C.; Yang, N.; Cai, W. Multi-feature fusion: Graph neural network and CNN combining for hyperspectral image classification. Neurocomputing 2022, 501, 246–257. [Google Scholar] [CrossRef]
  24. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plaza, A.; Chanussot, J. Graph convolutional networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5966–5978. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Ding, Y.; Zhao, X.; Siye, L.; Yang, N.; Cai, Y.; Zhan, Y. Multireceptive field: An adaptive path aggregation graph neural framework for hyperspectral image classification. Expert Syst. Appl. 2023, 217, 119508. [Google Scholar] [CrossRef]
  26. Wang, X.; Sun, L.; Lu, C.; Li, B. A novel transformer network with a CNN-enhanced cross-attention mechanism for hyperspectral image classification. Remote Sens. 2024, 16, 1180. [Google Scholar] [CrossRef]
  27. Qing, Y.; Liu, W.; Feng, L.; Gao, W. Improved transformer net for hyperspectral image classification. Remote Sens. 2021, 13, 2216. [Google Scholar] [CrossRef]
  28. Dalal, A.A.; Cai, Z.; Al-Qaness, M.A.; Dahou, A.; Alawamy, E.A.; Issaka, S. Compression and reinforce variation with convolutional neural networks for hyperspectral image classification. Appl. Soft Comput. 2022, 130, 109650. [Google Scholar]
  29. Dalal, A.A.; Cai, Z.; Al-qaness, M.A.; Alawamy, E.A.; Alalimi, A. ETR: Enhancing transformation reduction for reducing dimensionality and classification complexity in hyperspectral images. Expert Syst. Appl. 2023, 213, 118971. [Google Scholar]
  30. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  31. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef]
  32. Zhong, Z.; Li, J.; Ma, L.; Jiang, H.; Zhao, H. Deep residual networks for hyperspectral image classification. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2107; IEEE: Piscataway, NJ, USA, 2017; pp. 1824–1827. [Google Scholar]
  33. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.J.; Pla, F. Deep pyramidal residual networks for spectral–spatial hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 740–754. [Google Scholar] [CrossRef]
  34. Meng, Z.; Li, L.; Jiao, L.; Feng, Z.; Tang, X.; Liang, M. Fully dense multiscale fusion network for hyperspectral image classification. Remote Sens. 2019, 11, 2718. [Google Scholar] [CrossRef]
  35. Zhu, M.; Jiao, L.; Liu, F.; Yang, S.; Wang, J. Residual spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 449–462. [Google Scholar] [CrossRef]
  36. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  37. Lin, J.; Gao, F.; Shi, X.; Dong, J.; Du, Q. SS-MAE: Spatial–spectral masked autoencoder for multisource remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–14. [Google Scholar] [CrossRef]
  38. Wang, M.; Gao, F.; Dong, J.; Li, H.C.; Du, Q. Nearest neighbor-based contrastive learning for hyperspectral and LiDAR data classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  39. Wang, J.; Gao, F.; Dong, J.; Du, Q. Adaptive DropBlock-enhanced generative adversarial networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5040–5053. [Google Scholar] [CrossRef]
  40. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  41. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  42. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  43. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral image classification using group-aware hierarchical transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  44. Ren, Q.; Tu, B.; Liao, S.; Chen, S. Hyperspectral image classification with iformer network feature extraction. Remote Sens. 2022, 14, 4866. [Google Scholar] [CrossRef]
  45. Yang, A.; Li, M.; Ding, Y.; Hong, D.; Lv, Y.; He, Y. GTFN: GCN and transformer fusion with spatial-spectral features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  46. Zhao, F.; Zhang, J.; Meng, Z.; Liu, H.; Chang, Z.; Fan, J. Multiple vision architectures-based hybrid network for hyperspectral image classification. Expert Syst. Appl. 2023, 234, 121032. [Google Scholar] [CrossRef]
  47. Cui, Y.; Shao, C.; Luo, L.; Wang, L.; Gao, S.; Chen, L. Center weighted convolution and GraphSAGE cooperative network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  48. Shi, C.; Yue, S.; Wang, L. A dual branch multiscale Transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–20. [Google Scholar] [CrossRef]
  49. Zhuo, R.; Guo, Y.; Guo, B. A hyperspectral image classification method based on 2-d compact variational mode decomposition. IEEE Trans. Geosci. Remote Sens. 2023, 20, 1–5. [Google Scholar] [CrossRef]
Figure 1. Overview illustration of the proposed MLP-mixer and graph convolutional enhanced transformer (MGCET) for hyperspectral image classification.
Figure 1. Overview illustration of the proposed MLP-mixer and graph convolutional enhanced transformer (MGCET) for hyperspectral image classification.
Remotesensing 16 02892 g001
Figure 2. Detailed structure of the spatial-spectral extraction block. (a) 3D convolution block. (b) 2D convolution block.
Figure 2. Detailed structure of the spatial-spectral extraction block. (a) 3D convolution block. (b) 2D convolution block.
Remotesensing 16 02892 g002
Figure 3. Detailed structure of the MLP-mixer.
Figure 3. Detailed structure of the MLP-mixer.
Remotesensing 16 02892 g003
Figure 4. Detailed structure of the Graph Convolutional Enhanced Transformer. (a) Residual bottleneck block. (b) Graph Convolutional Enhanced Transformer. (c) The Multi-head self-attention with GConv (MHSAG). (d) Graph Convolution Embedding Attention Mechanism. (e) Graph Convolution Module.
Figure 4. Detailed structure of the Graph Convolutional Enhanced Transformer. (a) Residual bottleneck block. (b) Graph Convolutional Enhanced Transformer. (c) The Multi-head self-attention with GConv (MHSAG). (d) Graph Convolution Embedding Attention Mechanism. (e) Graph Convolution Module.
Remotesensing 16 02892 g004
Figure 5. Flowchart of the proposed MGCET.
Figure 5. Flowchart of the proposed MGCET.
Remotesensing 16 02892 g005
Figure 6. Different network classification maps on IP dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) Spectral Former. (k) iFormer. (l) MGCET.
Figure 6. Different network classification maps on IP dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) Spectral Former. (k) iFormer. (l) MGCET.
Remotesensing 16 02892 g006
Figure 7. Classification maps of different networks on the SA dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) spectral former. (k) iFormer. (l) MGCET.
Figure 7. Classification maps of different networks on the SA dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) spectral former. (k) iFormer. (l) MGCET.
Remotesensing 16 02892 g007
Figure 8. Classification maps of different networks on the PU dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) spectral former. (k) iFormer. (l) MGCET.
Figure 8. Classification maps of different networks on the PU dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) spectral former. (k) iFormer. (l) MGCET.
Remotesensing 16 02892 g008
Figure 9. Classification maps of different networks on the KSC dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) spectral former. (k) iFormer. (l) MGCET.
Figure 9. Classification maps of different networks on the KSC dataset. (a) Ground truth. (b) SVM-RBF. (c) 3D-CNN. (d) RSSAN. (e) RSSN. (f) DBDA. (g) ViT. (h) SSFTT. (i) GAHT (j) spectral former. (k) iFormer. (l) MGCET.
Remotesensing 16 02892 g009
Figure 10. OA results with several patch sizes. (ad) are the OA bar graphs on IP, SA, PU, and KSC, respectively.
Figure 10. OA results with several patch sizes. (ad) are the OA bar graphs on IP, SA, PU, and KSC, respectively.
Remotesensing 16 02892 g010
Figure 11. OA curve of different training samples. (ad) are IP, SA, PU, and KSC datasets, respectively.
Figure 11. OA curve of different training samples. (ad) are IP, SA, PU, and KSC datasets, respectively.
Remotesensing 16 02892 g011
Figure 12. OA curve of different training epochs. (ad) are IP, SA, PU, and KSC datasets, respectively.
Figure 12. OA curve of different training epochs. (ad) are IP, SA, PU, and KSC datasets, respectively.
Remotesensing 16 02892 g012aRemotesensing 16 02892 g012b
Figure 13. OA of MGCET with different encoder layers and multi-head self-attention. (ad) Three-dimensional column plots on the IP, SA, PU, and KSC. The deeper the color, the higher the OA.
Figure 13. OA of MGCET with different encoder layers and multi-head self-attention. (ad) Three-dimensional column plots on the IP, SA, PU, and KSC. The deeper the color, the higher the OA.
Remotesensing 16 02892 g013
Table 1. Configuration details of the graph convolutional enhanced transformer.
Table 1. Configuration details of the graph convolutional enhanced transformer.
ModuleLayerConfiguration [Output Size] (Kernel)(Stride) (Padding)
3D Convolution Module3-D Conv-1[8, 38, 11, 11]; (11, 3, 3); (5, 1, 1); (0, 1, 1)
3-D Conv-2[8, 38, 11, 11]; (3, 3, 3); (1, 1, 1); (1, 1, 1)
3-D Conv-3[8, 38, 11, 11]; (1, 1, 1); (1, 1, 1); (0, 0, 0)
Rearrange[304, 11, 11](–)(–)(–)
2D Convolution Module2-D Conv-1[256, 11, 11]; (1 × 1); (1, 1); (0, 0)
2-D DWConv-2[256, 11, 11]; (3 × 3); (1, 1); (0, 0)
2-D Conv-3[ H S I b a n d s , 11, 11]; (1 × 1); (1, 1); (0, 0)
Patch Embeddingsnn.Flatten()[121, H S I b a n d s ](–)(–)(–)
nn.Linear()[121, 256](–)(–)(–)
Postional Embeddingsnn.Parameter()[121, 256](–)(–)(–)
MLP-mixernn.LayerNorm()[121, 256](–)(–)(–)
token-mixing MLP[121, 256](–)(–)(–)
nn.LayerNorm()[121, 256](–)(–)(–)
channel-mixing MLP[121, 256](–)(–)(–)
Graph convolutional
Enhanced Transformer
Encoder
nn.LayerNorm()[121, 256](–)(–)(–)
GConv[121, 512](–)(–)(–)
MHSAG[121, 512](–)(–)(–)
1-D AdaptiveAvgPool[121, 256](–)(–)(–)
nn.LayerNorm()[121, 256](–)(–)(–)
Residual Bottleneck Block2-D Conv-5[11, 11, 64]; (1 × 1); (1 × 1); (0, 0)
2-D DWConv-6[11, 11, 64]; (3 × 3); (1 × 1); (1, 1)
2-D Conv-7[11, 11, 256]; (1 × 1); (1 × 1); (0, 0)
ClassifierDropoutOutput Size: [121, 256]
Avg Pooling[1, 256](–)(–)(–)
nn.LayerNorm()[1, 256](–)(–)(–)
nn.Linear()[1, Classes](–)(–)(–)
Table 2. The sample numbers for each class in the experiment stages (training and testing) for the IP dataset with the corresponding color background of each class.
Table 2. The sample numbers for each class in the experiment stages (training and testing) for the IP dataset with the corresponding color background of each class.
Class No.Class NameTrainingTesting
1Alfalfa244
2Con-notill711357
3Con-mintill41789
4Corn12225
5Grass-pasture24459
6Grass-trees37693
7Grass-pasture-mowed127
8Hay-windrowed24454
9Oats119
10Soybean-notill49923
11Soybean-mintill1232332
12Soybean-clean30563
13Wheat10195
14Woods631203
15Buildings-grass-trees-drivers19367
16Stone-steel-towers588
Total 5129737
Table 3. The sample numbers for each class in the experiment stages (training and testing) for the PU dataset, with the corresponding color background of each class.
Table 3. The sample numbers for each class in the experiment stages (training and testing) for the PU dataset, with the corresponding color background of each class.
Class No.Class NameTrainingTesting
1Asphalt666565
2Meadows18618,463
3Gravel212078
4Trees313033
5Painted metal sheets131332
6Bare Soil504979
7Bitumen131317
8Self-Blocking Bricks373645
9Shadows10937
Total 42742,349
Table 4. The sample numbers for each class in the experiment stages (training and testing) for each class for SA dataset with the corresponding colored background in the classification map.
Table 4. The sample numbers for each class in the experiment stages (training and testing) for each class for SA dataset with the corresponding colored background in the classification map.
Class No.Class NameTrainingTesting
1Brocoli-green-weeds-1201989
2Brocoli-green-weeds-2373689
3Fallow201956
4Fallow-rough-plow141380
5Fallow-smooth272651
6Stubble393920
7Celery363543
8Grapes-untrained11311,158
9Soil-vinyard-develop626141
10Corn-senesced-green-weeds333245
11Lettuce-romaine-4wk111057
12Lettuce-romaine-5wk191908
13Lettuce-romaine-6wk9907
14Lettuce-romaine-7wk111059
15Vinyard-untrained727196
16Vinyard-vertical-trellis181789
Total 54153,588
Table 5. The sample numbers for each class in the experiment stages (training and testing) for KSC dataset with the corresponding colored background in the classification map.
Table 5. The sample numbers for each class in the experiment stages (training and testing) for KSC dataset with the corresponding colored background in the classification map.
Class No.Class NameTrainingTesting
1Scrub38723
2Willow Swamp12231
3CP Hammock13243
4Slash Pine13239
50ak/Broadleaf8153
6Hardwood swamp11218
7Swap5100
8Graminoid Marsh22409
9Spartina Marsh26494
10Cattail Marsh20384
11Salt Marsh21398
12Mud Flats25478
13Water46881
Total 2604951
Table 6. IP dataset classification results.
Table 6. IP dataset classification results.
Class No.SVM-RBF3D-CNNRSSANDBDARSSNViTSSFTTGAHTSpectral FormeriFormerMGCET
171.36 ± 3.6581.01 ± 3.1078.29 ± 1.3476.00 ± 2.0175.91 ± 2.7474.52 ± 4.0677.11 ± 0.7079.63 ± 0.3270.52 ± 3.1576.48 ± 3.3579.08 ± 1.03
268.16 ± 7.9971.18 ± 6.8187.69 ± 1.4483.28 ± 9.1784.61 ± 7.0281.89 + 3.7289.08 ± 1.8791.30 ± 0.8886.89 ± 5.5393.19 ± 4.1494.70 ± 1.51
361.12 ± 5.5266.92 ± 8.7289.87 ± 0.4588.17 ± 6.2181.39 ± 5.3179.46 ± 8.1992.87 ± 3.2193.59 ± 0.8691.30 ± 3.8392.51 ± 0.2393.30 ± 1.61
460.90 ± 11.7568.89 ± 2.4090.98 ± 1.99100.00 ± 0.0096.42 ± 0.9185.51 ± 4.8998.55 ± 0.4197.71 ± 1.6195.53 ± 2.3487.29 ± 3.0394.48 ± 0.93
573.18 ± 0.6973.36 ± 6.9393.23 ± 1.2995.52 ± 5.2197.05 ± 0.1785.75 ± 3.1697.12 ± 1.8496.19 ± 1.0887.51 ± 4.0195.47 ± 3.6195.25 ± 1.01
691.13 ± 1.1488.29 ± 4.1596.61 ± 0.3797.37 ± 1.1397.42 ± 1.1181.81 ± 7.7196.37 ± 0.9396.71 ± 1.1599.32 ± 0.1399.16 ± 0.6699.36 ± 0.39
764.18 ± 10.3775.00 ± 2.1176.87 ± 3.6283.24 ± 1.4779.30 ± 0.7175.48 ± 4.0179.70 ± 0.3176.87 ± 1.6784.81 ± 5.0184.81 ± 11.0883.69 ± 0.04
895.79 ± 1.3597.02 ± 1.0097.00 ± 0.0098.77 ± 0.78100.00 ± 0.0073.76 ± 5.7298.47 ± 0.4899.05 ± 0.2183.18 ± 7.7899.76 ± 0.3297.90 ± 2.12
958.17 ± 1.3778.00 ± 0.7175.00 ± 0.3379.90 ± 0.0477.29 ± 1.2083.76 ± 4.7277.36 ± 4.2779.78 ± 3.6773.76 ± 1.9477.55 ± 2.5184.27 ± 3.02
1066.76 ± 2.6977.93 ± 3.7388.39 ± 1.1386.24 ± 4.5397.63 ± 0.5989.30 ± 2.1992.24 ± 2.3394.04 ± 1.7193.77 ± 0.1397.36 ± 1.2896.68 ± 1.82
1173.59 ± 3.4159.64 ± 3.7385.79 ± 0.7683.03 ± 3.3981.21 ± 6.7177.68 ± 6.2484.21 ± 4.9187.18 ± 1.4293.17 ± 0.4296.99 ± 2.1197.97 ± 1.26
1256.22 ± 5.8762.94 ± 5.6689.16 ± 0.8886.98 ± 7.1288.30 ± 4.1986.36 ± 2.9192.98 ± 3.7290.70 ± 0.7182.48 ± 10.7293.29 ± 6.1396.61 ± 2.42
1386.97 ± 0.5189.05 ± 0.2797.00 ± 1.1096.61 ± 1.5397.37 ± 2.1294.62 + 2.1698.61 ± 1.5399.02 ± 0.3490.82 ± 4.7399.94 ± 0.0799.95 ± 0.03
1494.43 ± 0.8283.13 ± 0.6092.58 ± 0.1397.78 ± 0.0796.67 ± 2.6493.77 ± 1.4198.18 ± 0.7799.03 ± 0.3981.49 ± 7.1798.10 ± 1.1497.69 ± 1.77
1561.70 ± 2.4780.43 ± 6.9188.59 ± 0.6998.17 ± 0.5196.90 ± 1.1396.51 ± 2.3899.17 ± 0.8998.16 ± 0.7191.26 ± 3.6197.91 ± 1.7395.35 ± 4.46
1683.17 ± 8.52100.00 ± 0.0092.41 ± 2.1098.77 ± 0.35100.00 ± 0.0094.30 ± 3.2695.47 ± 2.3597.47 ± 1.25100.00 ± 0.0096.78 ± 2.0697.72 ± 1.09
OA(%)76.96 ± 2.2577.83 ± 1.7891.66 ± 3.6391.75 ± 3.8492.47 ± 2.4681.96 ± 3.2192.55 ± 2.8193.87 ± 0.2787.76 ± 1.8194.28 ± 0.7695.45 ± 0.87
AA(%)72.30 ± 1.6679.12 ± 1.2193.13 + 1.0894.21 ± 2.4793.17 ± 0.7881.55 ± 2.0793.71 + 1.7495.23 ± 0.4887.81 ± 2.0193.22 ± 2.0695.35 ± 1.96
Kappa(%)84.81 ± 3.2274.38 ± 1.9989.32 ± 0.7891.03 ± 3.4780.03 ± 3.8385.71 ± 3.2993.33 ± 3.6794.15 ± 1.2384.19 ± 4.2293.47 ± 2.7395.22 ± 0.91
Table 7. SA dataset classification results.
Table 7. SA dataset classification results.
Class No.SVM-RBF3D-CNNRSSANDBDARSSNViTSSFTTGAHTSpectral FormeriFormerMGCET
179.73 ± 6.6580.47 ± 4.3197.81 ± 1.3698.80 ± 0.51100.00 ± 0.0085.47 ± 7.16100.00 ± 0.0099.84 ± 0.1793.17 ± 2.0598.42 ± 0.6597.14 ± 3.13
289.09 ± 4.9989.59 ± 3.7197.97 ± 1.0798.21 ± 0.1999.91 ± 0.0592.59 + 1.9698.66 ± 0.8196.12 ± 1.0397.48 ± 0.5999.08 ± 0.61100.00 ± 0.00
382.50 ± 6.5288.17 ± 6.0296.80 ± 1.4198.07 ± 0.2996.61 ± 3.1188.17 ± 4.4994.79 ± 2.1297.59 ± 1.0195.19 ± 0.8698.31 ± 0.4899.46 ± 0.36
487.20 ± 6.7588.89 ± 4.4195.72 ± 3.0197.63 ± 1.0695.81 ± 1.4188.81 ± 3.1999.22 ± 0.3297.91 ± 1.0692.55 ± 3.0498.48 ± 0.0798.73 ± 1.08
587.59 ± 3.6990.99 ± 3.5195.14 ± 4.1996.21 ± 3.4196.51 ± 1.9787.99 ± 5.0695.59 ± 0.8198.48 ± 1.0391.20 ± 2.7195.47 ± 1.1098.94 ± 1.13
688.69 ± 3.2493.65 ± 3.1998.65 ± 1.3799.07 ± 0.3397.99 ± 1.6193.65 ± 2.1798.69 ± 1.03100.00 ± 0.0096.42 ± 1.3698.15 ± 0.0699.86 ± 0.15
789.95 ± 1.3793.61 ± 5.2298.99 ± 0.6396.89 ± 2.0198.45 ± 0.7989.61 ± 2.1199.10 ± 0.1198.18 ± 1.7792.25 ± 2.9199.19 ± 0.2899.96 ± 0.03
875.58 ± 7.1586.04 ± 7.0192.27 ± 4.8894.77 ± 3.0493.33 ± 2.7091.04 ± 1.0285.59 ± 3.0892.18 ± 3.0181.97 ± 3.0193.14 ± 3.2994.02 ± 2.52
992.73 ± 4.3796.46 ± 1.0299.36 ± 0.0499.95 ± 0.0398.60 ± 0.2889.46 ± 2.0297.91 ± 2.0798.97 ± 0.1691.76 ± 3.9199.15 ± 1.1599.87 ± 0.12
1085.95 ± 3.2991.52 ± 2.3595.63 ± 1.7196.01 ± 3.3898.34 ± 0.9992.51 ± 2.5193.51 ± 3.1397.08 ± 1.7486.64 ± 6.0193.47 ± 2.8497.54 ± 1.68
1190.58 ± 2.4193.73 ± 2.0396.57 ± 1.7194.23 ± 4.0397.01 ± 2.7995.73 ± 0.5496.70 ± 1.9299.26 ± 0.4987.05 ± 1.4997.62 ± 1.6999.18 ± 0.36
1293.24 ± 3.1596.92 ± 2.3699.32 ± 0.7198.98 ± 0.0298.19 ± 0.2994.92 ± 1.1998.88 ± 1.0299.44 ± 0.6195.47 ± 0.7499.06 ± 0.5399.95 ± 0.06
1394.37 ± 1.3197.18 ± 0.8199.81 ± 0.0897.79 ± 1.0497.57 ± 1.0894.88 + 0.1296.88 ± 0.5895.79 ± 1.2492.61 ± 0.8295.28 ± 1.37100.00 ± 0.00
1495.35 ± 1.0290.82 ± 2.1195.97 ± 2.0396.08 ± 1.0298.57 ± 0.6690.82 ± 3.1091.80 ± 1.0797.57 ± 1.5997.97 ± 1.1696.51 ± 1.4899.20 ± 0.51
1573.66 ± 8.0768.08 ± 4.9985.05 ± 3.0387.17 ± 3.5985.68 ± 6.2375.08 ± 9.0884.42 ± 6.0991.28 ± 3.7279.07 ± 8.3190.65 ± 2.3792.76 ± 2.75
1696.08 ± 3.2296.86 ± 0.1199.91 ± 0.1092.72 ± 0.4198.46 ± 0.4494.76 ± 0.2696.09 ± 1.5596.87 ± 1.6587.12 ± 3.0594.13 ± 2.7199.37 ± 0.61
OA(%)86.64 ± 4.0589.20 ± 3.1794.74 ± 2.9395.25 ± 1.8995.73 ± 1.0690.20 ± 1.2893.53 ± 1.1396.48 ± 1.2291.32 ± 1.8195.28 ± 0.3797.52 ± 0.37
AA(%)84.46 ± 4.6190.91 ± 2.7595.77 + 2.1296.01 ± 1.3896.17 ± 0.7892.19 ± 3.0195.08 + 0.7496.89 ± 1.0893.48 ± 2.0195.74 ± 1.2098.52 ± 0.34
Kappa(%)85.24 ± 1.5290.19 ± 4.0995.76 ± 2.1895.26 ± 0.7891.11 ± 2.3894.01 ± 1.9296.44 ± 0.6995.34 ± 1.2390.34 ± 3.8994.68 ± 0.7197.04 ± 0.41
Table 8. PU dataset classification results.
Table 8. PU dataset classification results.
Class No.SVM-RBF3D-CNNRSSANDBDARSSNViTSSFTTGAHTSpectral FormeriFormerMGCET
182.40 ± 1.8183.02 ± 2.7791.93 ± 0.9893.54 ± 1.0594.24 ± 1.0889.45 ± 2.8195.61 ± 1.4995.54 ± 0.2388.73 ± 0.6196.01 ± 0.4697.37 ± 1.64
283.34 ± 3.0588.12 ± 4.1794.63 ± 0.2796.43 ± 0.6298.03 ± 0.8990.54 ± 4.3596.36 ± 0.6397.43 ± 0.7195.88 ± 0.7297.75 ± 0.8299.55 ± 0.53
386.37 ± 3.0184.61 ± 3.8288.68 ± 2.4989.44 ± 1.5190.44 ± 1.1285.05 ± 6.9590.38 ± 2.1292.44 ± 1.3789.40 ± 2.0184.11 ± 7.1293.65 ± 3.77
482.92 ± 1.9190.11 ± 1.4496.91 ± 1.0594.93 ± 1.1294.52 ± 0.5194.77 ± 0.7196.36 ± 0.8296.93 ± 1.3095.97 ± 1.2497.92 ± 0.2797.36 ± 1.14
592.03 ± 1.2593.48 ± 0.9297.69 ± 0.6998.07 ± 0.3598.57 ± 0.6194.12 ± 0.5397.68 ± 1.2198.07 ± 0.7194.88 ± 1.7998.39 ± 0.7799.12 ± 1.10
694.21 ± 0.3281.94 ± 9.1296.32 ± 2.1297.72 ± 1.2896.62 ± 1.4789.91 ± 1.6792.37 ± 2.4797.12 ± 0.4485.29 ± 3.0687.26 ± 8.1397.96 ± 0.42
793.81 ± 1.6486.38 ± 4.0297.40 ± 1.3398.71 ± 0.2397.04 ± 0.9694.29 ± 2.7294.49 ± 1.6198.51 ± 0.3587.45 ± 3.1497.12 ± 1.0898.10 ± 3.51
890.46 ± 1.7295.97 ± 1.5996.39 ± 1.0897.37 ± 0.1997.17 ± 0.6297.01 ± 1.0695.94 ± 0.3198.17 ± 0.1794.38 ± 1.0194.96 ± 0.1296.44 ± 2.49
989.82 ± 0.9893.62 ± 0.9297.65 ± 0.0897.97 ± 0.4398.27 ± 0.5994.84 ± 1.8297.31 ± 0.7698.27 ± 0.1195.25 ± 1.1296.95 ± 0.4897.88 ± 1.70
OA(%)85.15 ± 1.0887.74 ± 1.7395.47 ± 0.7296.07 ± 0.3396.57 ± 0.2391.37 ± 3.1795.32 ± 0.9196.47 ± 0.3192.00 ± 1.4294.56 ± 0.9798.05 ± 0.41
AA(%)89.89 ± 1.1186.70 ± 3.1295.91 + 0.9595.98 ± 0.4796.38 ± 1.0192.12 ± 1.9994.39 ± 1.3596.28 ± 0.2691.37 ± 1.9293.41 ± 1.0396.99 ± 1.19
Kappa(%)83.04 ± 1.7883.69 ± 3.4193.71 ± 0.5895.24 ± 0.9295.64 ± 1.0290.80 ± 3.0893.94 ± 3.2395.74 ± 0.4092.96 ± 1.9191.23 ± 0.7197.56 ± 0.81
Table 9. KSC dataset classification results.
Table 9. KSC dataset classification results.
Class No.SVM-RBF3D-CNNRSSANDBDARSSNViTSSFTTGAHTSpectral FormeriFormerMGCET
188.40 ± 4.0093.40 ± 1.3098.20 ± 0.36100.00 ± 0.0099.86 ± 0.1292.08 ± 0.50100.00 ± 0.0099.82 ± 0.2194.31 ± 0.51100.00 ± 0.00100.00 ± 0.00
282.28 ± 7.6986.28 ± 4.6996.46 ± 1.0897.42 ± 2.9897.02 ± 1.8287.21 ± 3.6898.66 ± 0.8199.05 ± 0.7287.63 ± 2.0299.24 ± 0.7196.10 ± 2.72
385.80 ± 6.5285.80 ± 3.7197.60 ± 2.1292.61 ± 3.0597.33 ± 1.5190.08 ± 7.0094.79 ± 2.1298.16 ± 0.7186.59 ± 1.7297.97 ± 2.1798.68 ± 1.21
470.71 ± 2.6475.71 ± 4.7791.32 ± 3.6287.32 ± 8.2792.67 ± 2.6277.02 ± 9.1494.22 ± 0.3288.04 ± 10.0992.10 ± 3.2291.32 ± 6.9092.55 ± 3.46
576.22 ± 6.8569.22 ± 4.6484.41 ± 6.0294.41 ± 4.7786.86 ± 7.0584.96 ± 8.3395.59 ± 0.8190.72 ± 2.2172.52 ± 8.2690.55 ± 4.1888.72 ± 3.21
693.34 ± 2.2087.34 ± 3.8593.45 ± 4.0195.01 ± 4.0893.74 ± 5.5991.69 ± 3.6098.69 ± 1.0392.41 ± 4.1294.54 ± 2.8398.54 ± 2.3297.43 ± 0.81
789.34 ± 6.4094.34 ± 4.2092.37 ± 5.0687.37 ± 2.4697.92 ± 0.5293.75 ± 0.9399.10 ± 0.1194.00 ± 1.2493.12 ± 3.2488.62 ± 0.1799.41 ± 0.39
887.75 ± 1.4889.75 ± 2.0995.90 ± 0.8296.90 ± 1.1998.21 ± 1.6286.88 ± 4.6690.59 ± 3.0899.95 ± 0.1996.90 ± 0.9797.33 ± 0.9399.80 ± 0.08
991.01 ± 2.0995.01 ± 3.4096.51 ± 1.1097.01 ± 0.6299.72 ± 0.2691.99 ± 3.4897.91 ± 2.07100.00 ± 0.00100.00 ± 0.0099.19 ± 0.6799.95 ± 0.07
1091.63 ± 4.3893.03 ± 1.08100.00 ± 0.0099.76 ± 0.3899.53 ± 0.5091.48 ± 0.6993.51 ± 3.1399.91 ± 0.29100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
1193.93 ± 2.4195.33 ± 2.1998.14 ± 0.5299.61 ± 0.4698.84 ± 0.7092.78 ± 4.2396.70 ± 1.9298.28 ± 0.0295.49 ± 0.1799.29 ± 0.1599.94 ± 0.10
1295.56 ± 3.1592.02 ± 4.3896.87 ± 1.3898.87 ± 1.0297.15 ± 3.0292.24 ± 2.9898.88 ± 1.02100.00 ± 0.0094.23 ± 2.8599.56 ± 0.3399.03 ± 1.04
1396.32 ± 2.1398.23 ± 0.8199.20 ± 0.8199.82 ± 0.12100.00 ± 0.0097.35 ± 3.1296.88 ± 0.58100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
OA(%)88.96 ± 1.2390.97 ± 1.0796.02 ± 0.4196.59 ± 0.5397.03 ± 0.6590.71 ± 0.8397.52 ± 0.4897.77 ± 0.6093.64 ± 0.8297.59 ± 0.5898.52 ± 0.78
AA(%)86.61 ± 1.8888.01 ± 2.7594.56 + 1.1294.06 ± 1.2495.26 ± 1.4088.94 ± 1.1795.62 + 1.0996.45 ± 1.1892.58 ± 1.9196.88 ± 1.4797.38 ± 1.60
Kappa(%)89.94 ± 1.3791.21 ± 0.2996.04 ± 1.0496.21 ± 0.5996.82 ± 0.7389.66 ± 0.9296.95 ± 0.5397.30 ± 1.2392.59 ± 0.6896.02 ± 0.9398.47 ± 0.86
Table 10. The ablation results of GCET and Mlp-mixer.
Table 10. The ablation results of GCET and Mlp-mixer.
NetworkGCETMLP-mixerMetricIPSAPUKSC
Net-1OA % 83.82 ± 4.7688.80 ± 1.1382.22 ± 6.7190.26 ± 0.71
AA % 78.18 ± 5.1689.24 ± 1.5884.55 ± 4.5789.76 ± 2.32
Kappa % 82.20 ± 5.3288.69 ± 1.3282.59 ± 4.7991.25 ± 1.66
Net-2OA % 93.36 ± 2.0895.23 ± 0.5095.52 ± 2.1196.91 ± 0.42
AA % 92.05 ± 2.2996.17 ± 1.5196.33 ± 0.9394.78 ± 0.11
Kappa % 91.57 ± 2.1595.98 ± 0.6897.34 ± 1.2697.12 ± 0.30
Net-3OA % 94.41 ± 1.1896.35 ± 0.7197.21 ± 1.1797.72 ± 0.18
AA % 92.45 ± 1.8896.45 ± 0.3195.75 ± 0.8396.30 ± 0.70
Kappa % 92.49 ± 1.8495.77 ± 0.8696.02 ± 1.0296.91 ± 0.12
Net-4OA % 95.45 ± 0.9197.57 ± 0.4298.05 ± 0.4198.52 ± 0.07
AA % 93.99 ± 1.2897.22 ± 0.6097.78 ± 0.2097.74 ± 0.34
Kappa % 94.73 ± 1.0697.49 ± 0.6897.49 ± 0.6898.04 ± 0.55
Table 11. Ablation analysis of MHSAG and MHSA.
Table 11. Ablation analysis of MHSAG and MHSA.
NetworkMetricIPSAPUKSC
MHSAOA % 94.92 ± 0.7696.88 ± 0.8397.82 ± 0.7198.36 ± 0.41
AA % 93.58 ± 0.5297.04 ± 0.0896.55 ± 1.0796.76 ± 0.52
Kappa % 93.32 ± 1.1296.69 ± 0.4196.29 ± 0.7297.75 ± 0.16
MHSAGOA % 95.45 ± 0.9197.57 ± 0.4298.05 ± 0.4198.52 ± 0.07
AA % 93.99 ± 1.2897.22 ± 0.6097.78 ± 0.2097.74 ± 0.34
Kappa % 94.73 ± 1.0697.49 ± 0.6897.49 ± 0.6898.04 ± 0.55
Table 12. Analysis of Different network architectures.
Table 12. Analysis of Different network architectures.
StructureMetricIPSAPUKSC
Structure-1OA % 94.57 ± 0.7696.97 ± 0.4797.59 ± 0.2197.79 ± 0.65
AA % 93.51 ± 0.8297.11 ± 0.2495.97 ± 0.5195.97 ± 1.38
Kappa % 94.13 ± 0.8996.52 ± 0.5296.81 ± 0.2897.53 ± 0.72
Structure-2OA % 94.29 ± 0.6096.68 ± 0.8397.25 ± 0.6098.01 ± 0.27
AA % 92.74 ± 1.1797.31 ± 0.2595.52 ± 1.0296.50 ± 1.83
Kappa % 92.89 ± 0.6996.31 ± 0.9396.36 ± 0.8097.79 ± 1.09
Structure-3OA % 95.45 ± 0.9197.57 ± 0.4298.05 ± 0.4198.52 ± 0.07
AA % 93.99 ± 1.2897.22 ± 0.6097.78 ± 0.2097.74 ± 0.34
Kappa % 94.73 ± 1.0697.49 ± 0.6897.49 ± 0.6898.04 ± 0.55
Structure-4OA % 95.14 ± 0.3197.18 ± 0.2697.78 ± 0.2698.11 ± 0.41
AA % 93.66 ± 2.2197.59 ± 0.1696.51 ± 0.4297.35 ± 0.95
Kappa % 94.80 ± 0.3596.86 ± 0.2896.66 ± 0.1897.78 ± 0.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-qaness, M.A.A.; Wu, G.; AL-Alimi, D. MGCET: MLP-mixer and Graph Convolutional Enhanced Transformer for Hyperspectral Image Classification. Remote Sens. 2024, 16, 2892. https://doi.org/10.3390/rs16162892

AMA Style

Al-qaness MAA, Wu G, AL-Alimi D. MGCET: MLP-mixer and Graph Convolutional Enhanced Transformer for Hyperspectral Image Classification. Remote Sensing. 2024; 16(16):2892. https://doi.org/10.3390/rs16162892

Chicago/Turabian Style

Al-qaness, Mohammed A. A., Guoyong Wu, and Dalal AL-Alimi. 2024. "MGCET: MLP-mixer and Graph Convolutional Enhanced Transformer for Hyperspectral Image Classification" Remote Sensing 16, no. 16: 2892. https://doi.org/10.3390/rs16162892

APA Style

Al-qaness, M. A. A., Wu, G., & AL-Alimi, D. (2024). MGCET: MLP-mixer and Graph Convolutional Enhanced Transformer for Hyperspectral Image Classification. Remote Sensing, 16(16), 2892. https://doi.org/10.3390/rs16162892

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop