Next Article in Journal
Environmental Influences on the Detection of Buried Objects with a Ground-Penetrating Radar
Next Article in Special Issue
Joint Classification of Hyperspectral and LiDAR Data Based on Adaptive Gating Mechanism and Learnable Transformer
Previous Article in Journal
GNSS/5G Joint Position Based on Weighted Robust Iterative Kalman Filter
Previous Article in Special Issue
MADANet: A Lightweight Hyperspectral Image Classification Network with Multiscale Feature Aggregation and a Dual Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Image Classification with the Orthogonal Self-Attention ResNet and Two-Step Support Vector Machine

College of Information and Communication Engineering, Dalian Minzu University, Dalian 116600, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(6), 1010; https://doi.org/10.3390/rs16061010
Submission received: 8 February 2024 / Revised: 4 March 2024 / Accepted: 7 March 2024 / Published: 13 March 2024
(This article belongs to the Special Issue Recent Advances in the Processing of Hyperspectral Images)

Abstract

:
Hyperspectral image classification plays a crucial role in remote sensing image analysis by classifying pixels. However, the existing methods require more spatial–global information interaction and feature extraction capabilities. To overcome these challenges, this paper proposes a novel model for hyperspectral image classification using an orthogonal self-attention ResNet and a two-step support vector machine (OSANet-TSSVM). The OSANet-TSSVM model comprises two essential components: a deep feature extraction network and an improved support vector machine (SVM) classification module. The deep feature extraction network incorporates an orthogonal self-attention module (OSM) and a channel attention module (CAM) to enhance the spatial–spectral feature extraction. The OSM focuses on computing 2D self-attention weights for the orthogonal dimensions of an image, resulting in a reduced number of parameters while capturing comprehensive global contextual information. In contrast, the CAM independently learns attention weights along the channel dimension. The CAM autonomously learns attention weights along the channel dimension, enabling the deep network to emphasise crucial channel information and enhance the spectral feature extraction capability. In addition to the feature extraction network, the OSANet-TSSVM model leverages an improved SVM classification module known as the two-step support vector machine (TSSVM) model. This module preserves the discriminative outcomes of the first-level SVM subclassifier and remaps them as new features for the TSSVM training. By integrating the results of the two classifiers, the deficiencies of the individual classifiers were effectively compensated, resulting in significantly enhanced classification accuracy. The performance of the proposed OSANet-TSSVM model was thoroughly evaluated using public datasets. The experimental results demonstrated that the model performed well in both subjective and objective evaluation metrics. The superiority of this model highlights its potential for advancing hyperspectral image classification in remote sensing applications.

1. Introduction

Compared with traditional RGB images, hyperspectral images offer nanometre-level spectral resolution and the ability to capture information from hundreds of bands. This provides them with a more robust feature recognition capability and higher classification reliability. Hyperspectral images contain rich spatial and spectral radiation information, which makes them comprehensive carriers of multiple data types. Quantified continuous spectral curve data lends itself favourably to image classification, making hyperspectral image classification a core technology in remote sensing applications. This task assigns unique category labels to each pixel based on the spectral information. Accurate classification results provide reliable data support for subsequent applications and decision-making. Over the past few decades, numerous hyperspectral image classification algorithms have been proposed and have been widely applied to land monitoring [1,2], vegetation identification [3], and lake water quality detection [4].
Hyperspectral image classification algorithms can be categorised into two types: traditional and deep-learning-based methods. Conventional methods include the k-nearest neighbour classification (KNN) [5], decision trees [6,7], and support vector machines (SVM) [8,9]. Among these methods, SVM with kernel techniques has been widely studied and applied to classify high-dimensional and strongly correlated spectral data accurately. Cui et al. [10] proposed a recursive edge-preserving filtering algorithm as the preprocessing step. This algorithm combines bilateral and recursive filtering to remove noise while preserving the edge details. Liu et al. [11] introduced a nonparallel support vector machine algorithm with a constrained bias term and additional empirical risk minimisation to improve the generalisation ability of the model for unknown data. Mahendra et al. [12] proposed an efficient hardware architecture for the SVM classification of hyperspectral remote sensing data using high-level synthesis (HLS), enabling real-time classification systems for remote sensing applications. Although traditional SVM algorithms have achieved good classification performance and robustness for data with few samples, high dimensionality, and clear class boundaries, they are sensitive to noise and outliers, are prone to overfitting, and require manual parameter tuning, resulting in poor generalisation ability.
Deep neural networks (DNN) have remarkable capabilities in terms of non-linear expressions and autonomous learning parameters. These capabilities enable the extraction of feature information from large-scale datasets, making them widely applicable to computer vision. When it comes to hyperspectral image classification tasks, the mainstream deep backbone networks include convolutional neural networks (CNN) [13], recurrent neural networks (RNN) [14], generative adversarial networks (GAN) [15], and transformers [16]. Ghaderizadeh et al. [17] proposed a network for hyperspectral image classification that integrates 3D and 2D convolution operations. Three-dimensional (3D) convolution is employed to extract spectral–spatial information, while two-dimensional (2D) convolution further enhances spatial information, avoiding the loss of spectral information. Feng et al. [18] introduced multi-complementary GANs with contrastive learning (CMC-GAN) based on contrastive learning, utilising a network structure incorporating coarse-grained and fine-grained GAN modules to achieve complementary multiscale spatial–spectral features. Zhong et al. [19] presented a hyperspectral image classification algorithm based on a transformer structure consisting of spectral feature extraction and spatial attention modules. Combining these two modules maximises the utilisation of spectral–spatial information in hyperspectral images and enhances the interaction ability among image pixels. He et al. [20] proposed a generative adversarial network with a transformer structure. This model fully attends to global contextual information among pixels through a self-attention mechanism, thereby avoiding the loss of low-level texture information. Yang et al. [21] introduced a ViT [22] structure with convolutional operations into hyperspectral image classification. They introduced the SACP module to capture spectral–spatial fusion information, leading to more accurate classification results. However, deep-learning-based hyperspectral image classification algorithms rely on data-driven methods that require many samples. Zhong et al. [23] introduced an end-to-end spectral–spatial residual ResNet (SSRN) that uses a series of 3D convolutions in the respective residual blocks to extract discriminative joint representation. Roy [24] presents an attention-based adaptive spectral–spatial kernel improved residual network (A2S2K-ResNet) with spectral attention, which learns selective 3D convolutional kernels to jointly extract spectral–spatial features using improved 3D ResBlocks and adopts an efficient feature recalibration (EFR) mechanism to boost the classification performance. In contrast, the need for more hyperspectral image data often leads to the poor performance of trained models in different scene classification tasks. Furthermore, deep-learning methods lack interpretability, which limits their application in specific scenarios.
There are many traditional methods and deep learning methods in hyperspectral image classification. Still, few algorithms combine the two, and poor data compatibility makes improving the indicators after incorporating them challenging. To address this challenge, Chen et al. [25] introduced SVMFLE, a novel feature line embedding algorithm based on SVM that enables data dimension reduction and enhances the performance of generative adversarial networks in hyperspectral image classification. Kalaiarasi et al. [26] combined the proximal support vector machine (PSVM) algorithm with deep learning methods to develop a new SVM model, ensuring accurate classification outcomes for hyperspectral images. Even with the integration of traditional and deep learning methods in hybrid networks, further improvements are required in the deep learning component or traditional approaches to exploit the spatial–spectral information specific to hyperspectral images fully.
Therefore, this paper takes advantage of the respective advantages of the two, and we only use the powerful feature extraction ability of the deep network while leaving the classification task to the support vector machine module, effectively avoiding confusion during data fusion. Of course, in future research, we will explore the parameter space of the two methods to achieve further data fusion, which has profound research value for hyperspectral image classification. To merge the interpretability of traditional methods with the robust feature extraction capability of deep-learning techniques, this study introduced the orthogonal self-attention module (OSM) and the channel attention module (CAM). These components enable more accurate feature extraction from hyperspectral images using deep residual networks and subsequent classification using the two-step support vector machine (TSSVM) classifier. The critical contributions of the proposed hyperspectral image classification algorithm, which combines the OSM with a two-stage SVM, are as follows:
(1)
This paper presents an OSM that performs global self-attention computation along two orthogonal dimensions: image length and width. This module not only enhances the interaction of spatial information within the image but also prevents feature confusion during the training process, thereby improving the generalisation performance of the network. The proposed approach enhances feature representation and improves network robustness by incorporating OSM.
(2)
This study introduces a lightweight CAM integrated into a basic residual block. This CAM module dynamically adjusts the channel weights based on the significance of different spectral channels, enabling the network to focus on essential spectral information while reducing interference from redundant information and noise. Incorporating the CAM enhances the quality and robustness of feature representation in the proposed approach, contributing to improved classification performance in hyperspectral image analysis.
(3)
This study proposes a TSSVM classifier to integrate and optimise the model by leveraging the discriminant function of the subclassifier. TSSVM is well suited for handling complex datasets and cases with high noise interference and avoids information loss that may arise from traditional SVM methods. Incorporating TSSVM, the proposed approach achieved superior classification performance in hyperspectral image analysis, thereby validating the effectiveness of the method in dealing with challenging scenarios.

2. Materials and Methods

2.1. Overall Framework

This paper proposes a novel model for hyperspectral image classification using an orthogonal self-attention ResNet and two-step support vector machine (OSANet-TSSVM), a hyperspectral image classification model that combines a residual network with an OSM and an SVM classifier. The overall framework of the proposed model is illustrated in Figure 1. Initially, the image preprocessing stage involved slicing the spatial dimensions of the original hyperspectral image into P × P sizes, resulting in P × P × C image blocks. Subsequently, a 3 × 3 convolution operation was applied to perform preliminary feature extraction on each image block while normalising the data for suitable deep network processing. Subsequently, three cascading layer structures were utilised, with multiple underlying residual blocks within each structure. Each layer structure is composed of three residual blocks, but the internal structure is slightly different. Considering that downsampling the feature map too early may cause detail loss, Layer 1 does not perform downsampling operation. The first convolution stride in L2 and L3 was set to 2, which downsamples the feature map twice and increases the receptive field of the convolution operation, promoting better feature fusion in the network. In addition, the OSM is incorporated after each layer structure, enabling the network to capture spatial global context information three times during forward propagation, thereby highlighting the features of critical pixels in space. Next, the feature maps are transformed into one-dimensional data using fully connected layers to match the data format required by the SVM classifier. Finally, the classifier processed the data to obtain the final output.

2.2. Basic Residual Block

The basic residual block serves as the fundamental building block of the residual network, and its detailed implementation is shown in Figure 2. The two convolutional layers in the basic residual block follow a bottleneck structure that initially compresses the feature dimension and then recovers it. This technique effectively reduces the computational cost of convolution while maintaining high performance. Furthermore, the batch normalisation layer standardised the results of each convolution operation to prevent data overflow. The activation layer adds a non-linear expression capability, boosting the network performance. To augment the spatial dimensions of the feature map, the CAM was introduced, positioned before the output of the basic residual block. CAM acts as a channel attention mechanism, dynamically adjusting the channel weights based on the significance of different spectral channels, resulting in improved feature representation and robust classification performance.
Given that hyperspectral images often consist of numerous spectral bands, it is impractical to input all the bands into the network for feature extraction. Doing so requires substantial computational resources because these bands often exhibit redundant relationships. To address this issue, attention mechanisms are commonly employed to adjust the weight of each band. The spectral attention mechanism, also known as the channel attention mechanism, focuses on extracting features from the spectral dimensions of an image. In this study, a lightweight CAM is integrated into a basic residual block. This integration effectively reduces the computational cost while enabling more detailed weight learning for spectral information. The implementation details are as follows. First, the input undergoes global average pooling to obtain the initial channel-wise weights. Unlike the original channel attention algorithm, global maximum pooling is not employed. This decision is motivated by the fact that such an operation only considers the maximum value of each pixel in the spatial dimension, thereby disregarding the neighbourhood relationship between pixels. Consequently, it is necessary to enhance the weight learning capability and add unnecessary computations. Subsequently, two cascaded convolution layers and normalisation operations perform secondary learning on the attention scores. This process facilitated the acquisition of more accurate spectral weight information.

2.3. Orthogonal Self-Attention Module

Deep convolutional neural networks utilise convolutional kernels of varying sizes to extract local features from images. Although widening the kernel size or deepening the network can increase the receptive field, doing so introduces many parameters and increases the image redundancy. Self-attention mechanisms are highly effective for sequence information interaction and parallel computation. Initially applied in natural language processing, these mechanisms have gained widespread use in computer vision in recent years. Visual self-attention mechanisms can search for correlations in an image to obtain global information. However, as the image size increases, the computational complexity of spatial pixel self-attention increases exponentially, severely limiting the application of this mechanism in computer vision. To mitigate the issue of high computational complexity, the OSM is proposed, which utilises depth-wise separable convolution operations to obtain weights and performs self-attention calculations solely within the orthogonal dimensions of image length and width. This approach significantly reduces the number of parameters while effectively capturing the long-range dependencies between pixels in the input image, thereby enabling the extraction of global spatial features.
The implementation details of the OSM are shown in Figure 3. Initially, the input underwent depth-wise separable convolution operations to derive the initial weights Q, K, and V. Subsequently, a reshaping operation was applied to Q and K, followed by element-wise multiplication to generate an attention map with the exact dimensions of the original image width. This attention map is then multiplied by the Reshaped V, yielding the result of the self-attention operation in the width direction. This process is aligned with the original self-attention calculation method. Next, self-attention calculations are conducted in the length direction of the obtained output. By performing these calculations on the orthogonal dimensions of image length and width, the interaction of spatial global information within the image is effectively facilitated.
The computational process of the OSM can be represented as follows:
Q , K , V = D W C o n v ( X )
row-Att ( Q , K , V ) = s o f t m a x ( Q K T α ) V
col-Att ( Q , K , V ) = s o f t m a x ( K T Q α ) V
O S A = row-Att ( L N ( Col-Att ( X ) ) ) + X
where X denotes the input, DWConv denotes the depth-wise separable convolution operation, α represents the scaling factor, row-Att ( · ) and col-Att ( · ) are the components of the OSM in the length and width dimensions, respectively. L N ( · ) indicates the layer normalisation operation and OSA represents the expression for orthogonal self-attention.

2.4. Two-Step Support Vector Machine

The working mechanism of SVM can be summarised as follows. It aims to find a classification hyperplane that effectively separates the two classes of sample points in the training set while maximising the margin to this plane. In cases in which classes are linearly inseparable, a kernel function is used to map the data from a low-dimensional input space to a higher-dimensional space. This transformation converts the original linearly inseparable problem in a low-dimensional space into a linearly separable problem in a higher-dimensional space. SVM have a solid theoretical foundation and strong generalisation ability, providing unique advantages in addressing small-sample, non-linear, and high-dimensional pattern recognition problems. The combination of hyperspectral technology and SVM for qualitative substance classification has been successfully demonstrated and applied to various research fields.
A traditional SVM constructs a classifier for binary classification problems to obtain discriminant results. However, multi-class classification problems are typically transformed into multiple binary classification problems. A series of binary classifiers are constructed to obtain discriminant results, commonly employing One-vs-One or One-vs-All classifiers. The final classification is determined by voting or the maximum probability. Although this discriminant approach has been used for decades, outliers and subclassifiers can easily influence the results. In recent years, researchers have focused on constructing kernel functions and optimising various algorithms to enhance the performance of SVM, leading to notable advancements in theory and applications. However, challenges and limitations persist. Because the SVM involves solving a quadratic programming problem, the computation time and memory requirements significantly increase with larger datasets, thereby posing challenges in handling large-scale datasets. Therefore, effectively addressing large-scale data sets remains an active area of research. In addition, noise and outliers may cause model overfitting during the training process, thereby affecting the accuracy of the classification results. Consequently, improving the robustness of SVMs to enhance the noise and outlier tolerance is a significant problem that requires attention.
To overcome these challenges, this paper proposes a novel decision method for multiclass SVM classification. In the first phase of the SVM, multiple subclassifiers are trained, and a second phase of training is introduced to address the limitations of the individual classifiers. The discriminant results obtained by the one-step subclassifiers were retained and remapped as new features. A second SVM training was performed on the newly established mappings, resulting in the TSSVM. By integrating the results of multiple classifiers, this approach compensates for the shortcomings of individual classifiers and allows complex classification problems to be solved effectively. Figure 4 illustrates the research roadmap adopted in this study.
Given the original training samples x i , i = 1 , 2 , , k with N classes and M sub-discriminators, M discriminant functions f 1 ( · ) ,   f 2 ( · ) , ,   f M ( · ) without binarisation are obtained through the one-step SVM training using the training samples and their class labels. For instance, in One-vs-All classifiers, a 0–1 value is obtained by determining the class attribution of unknown samples, representting the possibility of the test sample being in that class. The traditional approach involves inputting an unknown sample into each of the M subclassifiers to obtain M discriminant results, taking the maximum value as the highest possibility in that subclassifier, as listed in Table 1.
Although the traditional SVM method is simple and computationally efficient, its scalability and classification performance become more robust, especially in cases of overlapping categories or when non-linear separability is present. To address these limitations, the TSSVM model concatenates multiple SVMs and fully utilises the advantages of various subclassifiers to solve complex classification problems. Specifically, the M discriminant functions obtained from the one-step SVM training were retained as new feature vectors x i ˜ = ( f 1 ( x i ) ,   f 2 ( x i ) ,   ,   f m ( x i ) ) ,   i = 1 ,   2 ,   ,   k . Using the remapped feature vectors and original label data as training data, M two-step sub-classifiers are trained and obtained, denoted as d f 1 ( · ) ,   d f 2 ( · ) ,   d f 3 ( · ) ,   d f 4 ( · ) . The two-step subclassifiers determine the result vector of the one-step subclassifiers and consider the maximum value as the result of the two-step subclassifier, as listed in Table 2. Each sub-classifier obtained from the one-step training process had different features or data subsets. Simultaneously, TSSVM integrates and optimises the discriminant results of the subclassifiers to obtain more accurate classification results.
The TSSVM model proposed in this study optimised its classification performance through secondary training of the discriminant functions obtained from the one-step SVM training. It then makes decisions by voting or probability weighting. By integrating the results of multiple classifiers, the effects of noise and outliers on the final decision can be reduced, thereby improving the accuracy and robustness of classification.

3. Experimental Analysis

The experiments were conducted on a computer with an Intel i7 processor and 32 GB of memory. The algorithm model was implemented using Python 3.7 and developed in the PyCharm integrated development environment. A pyramid residual network model was employed to construct a deep backbone network, which was implemented using the PyTorch framework. The training was performed for 150 epochs on an NVIDIA GeForce RTX 4090 GPU with a batch size of 200 epochs. The SGD optimiser with an initial learning rate of 0.1 was utilised. The manufacturer of the GPU was NVIDIA and the server motherboard generator was ASUS, both made in China. The SVM classification algorithm was trained using a nonlinear Gaussian kernel function with 10-fold cross-validation. A non-linear Gaussian kernel function was employed for the SVM classification algorithm, and training was conducted using a 10-fold cross-validation.
This section presents the experimental results obtained from the Indian Pine, Salinas, Pavia University, and Kennedy Space Center datasets. The classification results of seven models, namely SVM, AENSVM, TSSVM, ResNet, SSRN [23], A2S2K-ResNet [24], and OSANet-TSSVM, were compared. The experimental results demonstrate that the proposed OSANet-TSSVM model performs better in terms of subjective and objective evaluation metrics.

3.1. Evaluating Indicator

A confusion matrix is an essential tool for evaluating the performance of the classification models. This provides detailed information on the classification results of the model for each class. Utilising the confusion matrix, the performance of the classification model for different classes can be intuitively observed. Its form is:
M = [ P 11 P 12 P 1 n P 21 P 22 P 2 n P n 1 P n 2 P n n ]
The overall accuracy (OA), average accuracy (AA), and Kappa coefficient (k) were calculated using a confusion matrix to evaluate the overall performance of the model comprehensively. The OA represents the proportion of correctly classified samples to the total number of samples in the entire dataset. AA represents the average accuracy rate for each class, serving as a more comprehensive metric by accounting for variations among different classes and evaluating the classification ability of the model across each class. The Kappa coefficient is a statistical measure of the agreement between the classification model and random chance. It considers the differences between the classification accuracy of the model and the accuracy of the random selection. These three evaluation metrics are defined in Formulas (6)–(8), where n represents the number of classes and the number of samples of class i is predicted as class j.
O A = i = 1 n P i i i = 1 n j = 1 n P i j
A A = i = 1 n P i i j = 1 n P i j n
K a p p a = i = 1 n P i i E i = 1 n j = 1 n P i j E E = ( i = 1 n j = 1 n P i j ) · ( i = 1 n j = 1 n P j i ) ( i = 1 n j = 1 n P i j ) 2

3.2. Indian Pines Dataset

The Indian Pines dataset was collected using the AVIRIS sensor over the state of Indiana in the central northern region of the United States. The image size was 144 × 144 pixels with 224 spectral bands, a spatial resolution of 20 m, and a wavelength range of 400–500 nm. After removing the 4 zero bands and 20 water absorption bands, the remaining 200 bands were used for the experiments. The Indian Pines dataset consists of 10,249 ground objects labelled into 16 classes. Due to significant variations in sample sizes among different classes in this dataset, 10% of the samples from each class were randomly selected as training data, while the remaining 90% constituted the test set. Figure 5 shows the corresponding high-resolution pseudocolour image and ground truth map. Table 3 lists the number of samples in the training and test sets used in the experiments.
A well-divided training set was used to train the classification models: SVM, AENSVM, TSSVM, ResNet, SSRN, A2S2K-ResNet, and OSANet-TSSVM. Table 4 presents the overall accuracy (OA), average accuracy (AA), and Kappa coefficient for each class across different models. The highest accuracy values are highlighted in bold. Figure 6 presents a visual comparison of different classification models.
The results show that the classification effect of the improved AENSVM and TSSVM is improved compared with the traditional SVM. Still, these conventional methods only perform simple spectral feature processing, which can easily lead to the misclassification phenomenon. The proposed method achieved the best classification accuracy with an OA of 98.19%. For the three comprehensive evaluation indicators, AA, OA, and Kappa coefficient of the Indian Pines dataset, the proposed method outperformed the second-best A2S2K-ResNet method by 0.65%, 0.62%, and 0.89%, respectively. In classes with fewer training samples, such as alfalfa, corn, grass-pasture-mowed, and oats, the classification accuracy significantly improved to 97.56%, 97.65%, 92.00%, and 99.81%, respectively. Figure 6 shows that, compared with other models, the OSANet-TSSVM and A2S2K-ResNet models appear to have better classification performance, overcoming the problem of smooth margin transition.
Moreover, the OSANet-TSSVM model exhibited significantly better classification performance than the other models in the alfalfa and soybean-no-till classes, with more apparent visual boundary contours closer to the ground truth. This indicates that the proposed OSANet-TSSVM can help the model better capture the relationships between different bands and positions, leading to more accurate image classification. Not only did this improve the overall accuracy, but the generator also generated diverse samples in classes with few training samples, learning the essential features of these samples.

3.3. Pavia University Dataset

The Pavia University dataset is a hyperspectral remote sensing image dataset obtained from Pavia, Italy, using a ROSIS sensor from Germany. It comprises 115 bands covering wavelengths from 0.43 to 0.86 µm. After filtering out the noisy bands, the dataset comprised 102 bands with a spatial resolution of 1.3 m. The image dimensions were 610 × 340 pixels, totalling 42,776 pixels. In addition to the background, the dataset contained nine land cover classes, including residential areas, grasslands, and trees. Because of its high spatial resolution and diverse urban landscapes, it has gained popularity in hyperspectral remote sensing image classification research.
In this experiment, a subset of 100 samples per class was randomly selected for training, and the remaining samples were used for testing. The detailed quantities of the samples for each class, training, and test sample are presented in Table 5. Figure 7 shows the pseudocolour images and the overall sample label classification map.
Table 6 presents the results of the OSANet-TSSVM and six other comparative algorithms. Along with overall accuracy (OA), average accuracy (AA), and Kappa values, the table displays the accuracy values for each class. The highest precision values are highlighted in bold text in the Experimental Results section. Figure 8 illustrates the classification results of the seven methods on the Pavia University dataset.
The analysis in Table 6 reveals that OSANet-TSSVM achieved the best classification results on the Pavia University dataset, demonstrating significant advantages over traditional algorithms and residual neural networks. It slightly outperformed the second-best A2S2K-ResNet method with OA, AA, and Kappa values of 0.17%, 0.33%, and 0.22%, respectively. Furthermore, the OSANet-TSSVM excels in meadows, trees, painted metal sheets, bare soil, and self-blocking rice classes. The results also show that the improved AEN-SVM and TSSVM methods based on traditional SVM have surpassed ResNet regarding classification performance. Specifically, AEN-SVM improved by 2.78% in the OA metric, while TSSVM attained an AA metric of 91.04%. This suggests that SVM continues to exhibit great potential for hyperspectral image classification. Figure 8 presents the classification results of the seven methods on the Pavia University dataset, showing OSANet-TSSVM with the fewest misclassified labels aligned closer to the ground truth. This demonstrates that the OSANet-TSSVM prioritises crucial spectral information and enhances spectral feature extraction capabilities.

3.4. Salinas Dataset

The Salinas dataset is a hyperspectral remote sensing image dataset from Salinas Valley, California, USA. Collected using NASA’s AVIRIS sensor, it comprises 224 spectral channels with a spatial resolution of 3.7 m and an image size of 512 × 217 pixels. The dataset includes 16 land cover classes, including various vegetables, orchards, and vineyards, making it a challenging dataset for hyperspectral image classification and target detection. Owing to its high resolution and complex land cover types, the Salinas dataset has become an essential benchmark in this field. The land cover information is presented in Table 7, and Figure 9 depicts a band image and ground truth map from the Salinas dataset.
Table 8 provides a detailed comparison of the experimental accuracy between OSANet-TSSVM and the six other classification models. The highest accuracy values are highlighted in bold. Figure 10 shows the classification results listed in Table 8.
Table 8 shows that compared to other methods, the proposed OSANet-TSSVM achieves the highest classification performance, with an OA of 98.32%, an AA of 99.29%, and a Kappa coefficient of 98.11%. These results indicate that OSANet-TSSVM achieves the highest accuracy in most categories and reaches 100% classification accuracy in seven categories. In contrast, the remaining algorithms struggle to accurately classify the Grapes_untrained and Vinyard_untrained classes, while the proposed model significantly improves the accuracy, surpassing the second-best A2S2K-ResNet model by 2.80% and 4.84%, respectively. Additionally, comparison images demonstrate that the experimental model exhibits good classification performance, especially in the Grapes_untrained, Vinyard_untrained, and Corn_senesced_green_weeds classes, with fewer misclassifications. These results confirm the effectiveness of the proposed OSANet-TSSVM in extracting spectral feature information and achieving more accurate feature classification.

3.5. Kennedy Space Center Dataset

The Kennedy Space Center (KSC) dataset is a hyperspectral remote sensing image dataset collected by NASA’s AVIRIS sensor at the Kennedy Space Center in the United States. This dataset contains 224 spectral bands with a spatial resolution of 18 m. After removing the water absorption bands and low signal-to-noise ratio bands, the actual bands used for training are 176. The image size is 512 × 614 pixels, covering 13 land cover categories such as wetlands, woodland, and beaches. The KSC dataset is widely used in hyperspectral remote sensing image classification and ecological research due to its rich wetland ecosystem and complex surface characteristics. Table 9 shows the ground object category information, and Figure 11 shows a certain band image and ground truth map from the KSC dataset.
Table 10 presents a detailed comparison of the experimental accuracy of the OSANet-SVM model against other classification models, including random forest (RF), support vector machine (SVM), AEN-SVM, ResNet, SSRN, and A2S2K-ResNet. The highest accuracy values among the compared models are highlighted in bold text. Figure 12 sequentially displays the recovery maps obtained from the classification of the RF, SVM, AEN-SVM, ResNet, SSRN, and A2S2K-ResNet models, corresponding to the detailed classification results presented in Table 10.
Figure 12 shows that among the seven experimental sets, OSANet-SVM exhibits significantly fewer classification errors than other algorithms. Particularly in CP hammock, oak/broadleaf, and graminoid marsh categories, the clarity of the classification image boundaries and the notably reduced number of misclassified points are apparent, aligning closely with the ground truth maps. As demonstrated in Table 10, OSANet-SVM achieves the best classification results, with the three comprehensive evaluation metrics of AA, OA, and Kappa coefficients reaching 98.75%, 97.94%, and 98.63%, respectively. Furthermore, the results indicate that the KSC dataset comprises 13 categories, and OSANet-SVM achieves a classification accuracy of 100% in six of these categories. This demonstrates that OSANet-SVM effectively captures the feature relationships in the input images, optimises classifier performance, and enables the extraction of spatial global features from the imagery.

3.6. Ablation Experiment

Ablation experiments were conducted to validate the effectiveness of the OSAM, CAM, and TSSVM modules for hyperspectral land classification tasks; the results are presented in Table 11. The highest accuracy values are highlighted in bold. In the table, (1) represents the pyramid residual backbone network, and (2) and (3) indicate the addition of OSAM and CAM modules to the backbone network, respectively, (4) and (5) represent the removal of the CAM and TSSVM modules, respectively, and (6) represents the proposed OSANet-TSSVM model in this study.
From the experimental results, it is evident that all the proposed modules enhance the classification performance. Taking Indian Pine as an example, adding the OSAM module increased OA, AA, and Kappa by 0.52%, 0.61%, and 0.24%, respectively. Furthermore, the inclusion of both OSAM and TSSVM modules further improved classification performance. Incorporating the OSAM and CAM modules yielded evaluation indicators of 98.06%, 97.97%, and 98.36%, respectively.
These findings demonstrate that the attention mechanism in the OSAM and CAM modules successfully extracted more accurate spatial–spectral information. In addition, the TSSVM module optimised the discriminative results of feature extraction using the deep network, significantly enhancing the hyperspectral land classification performance of the model.

4. Conclusions

This paper presents a novel hyperspectral image classification algorithm called OSA-SVM, which combines deep learning and SVM. The algorithm enhances the deep feature extraction network and the SVM classification module to achieve more accurate classification outcomes. Based on the window sliding convolution operation, the convolutional neural network model considers only local information correlation within the image. However, deepening the network to obtain a larger receptive field introduces numerous redundant parameters, leading to an overfitting. In contrast, the visual transformer model, which utilises the self-attention mechanism, can capture global interdependence among pixel information. Nonetheless, the computational complexity of this method increases exponentially with image resolution, restricting its practical application. To address this problem, this study proposes a lightweight orthogonal self-attention mechanism that captures global features. This approach exclusively performs global self-attention calculations in two orthogonal spatial dimensions, thereby reducing the number of parameters while preserving the interaction of global contextual information. Additionally, a channel attention mechanism was introduced to improve the extraction of spectral feature information, further enhancing classification accuracy.
In the classification module, the traditional SVM often performs poorly in scenarios involving overlapping categories or non-linear separability. In addition, noise and outliers can lead to the overfitting of the model. To address these limitations, this paper presents an enhanced SVM model that incorporates a two-step training process. In the first stage, a discriminant function was obtained through SVM training. This function was then retained as a new feature for the second-level training, resulting in the final TSSVM model. By jointly optimising the two classifiers and leveraging comprehensive feature information, the shortcomings of a single classifier are overcome, surpassing the performance of traditional SVM algorithms. Although the proposed method demonstrates excellent performance on multiple public datasets, it is essential to note that data compatibility between deep networks and traditional methods may affect the stability of the algorithm. As a result, future research will focus on further reducing the model parameter quantity and computational complexity, as well as improving the robustness of the hybrid model algorithm.
Our model is feasible for application in real-world scenarios. Although all experiments in this paper were conducted on small hardware devices and limited datasets, the deployment in practical scenarios can be achieved. The lightweight design proposed in this paper will significantly reduce inference time to apply the model to real-time processing scenarios. Of course, hyperspectral images collected in real-world scenarios contain more complex image degradation problems, such as blurring, occlusion, etc. To overcome this problem, targeted preprocessing should be performed on the images before the classification task, such as image denoising, image super-resolution, etc. Suppose the model is to be applied in particular environments, such as underwater, high temperature, etc. In that case, different hardware materials should be selected and continuously tested to ensure the stability of the model.

Author Contributions

Conceptualisation, L.W.; methodology, H.S.; formal analysis, Y.S.; writing, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (grant number 62071084), and by the Leading Talents Project of the State Ethnic Affairs Commission.

Data Availability Statement

All datasets can be obtained at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 20 May 2011).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jun, G.; Ghosh, J. Semisupervised Learning of Hyperspectral Data with Unknown Land-Cover Classes. IEEE Trans. Geosci. Remote Sens. 2013, 51, 273–282. [Google Scholar] [CrossRef]
  2. Karalas, K.; Tsagkatakis, G.; Zervakis, M.; Tsakalides, P. Land Classification Using Remotely Sensed Data: Going Multilabel. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3548–3563. [Google Scholar] [CrossRef]
  3. Lewis, M.; Jooste, V.; de Gasparis, A.A. Discrimination of arid vegetation with airborne multispectral scanner hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 1471–1479. [Google Scholar] [CrossRef]
  4. Sun, D.; Li, Y.; Wang, Q.; Gao, J. Hyperspectral Remote Sensing of the Pigment C-Phycocyanin in Turbid Inland Waters, Based on Optical Classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 3871–3884. [Google Scholar] [CrossRef]
  5. Ranjan, S.; Nayak, D.R.; Kumar, K.S.; Dash, R.; Majhi, B. Hyperspectral Image Classification: A k-Means Clustering Based Approach. In Proceedings of the 2017 4th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 6–7 January 2017. [Google Scholar]
  6. Xu, S.; Liu, S.; Wang, H.; Chen, W.; Zhang, F.; Xiao, Z. A hyperspectral image classification approach based on feature fusion and multi-layered gradient boosting decision trees. Entropy 2020, 23, 20. [Google Scholar] [CrossRef]
  7. Pradhan, B. A comparative study on the predictive ability of the decision tree, support vector machine and neuro-fuzzy models in landslide susceptibility mapping using GIS. Comput. Geosci. 2013, 51, 350–365. [Google Scholar] [CrossRef]
  8. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  9. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  10. Cui, B.; Ma, X.; Xie, X.; Ren, G.; Ma, Y. Classification of visible and infrared hyperspectral images based on image segmentation and edge-preserving filtering. Infrared Phys. Technol. 2017, 81, 79–88. [Google Scholar] [CrossRef]
  11. Liu, G.; Wang, L.; Liu, D.; Fei, L.; Yang, J. Hyperspectral Image Classification Based on Non-Parallel Support Vector Machine. Remote Sens. 2022, 14, 2447. [Google Scholar] [CrossRef]
  12. Mahendra, H.N.; Mallikarjunaswamy, S. An efficient classification of hyperspectral remotely sensed data using support vector machine. Int. J. Electron. Telecommun. 2022, 68, 141280. [Google Scholar] [CrossRef]
  13. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 12. [Google Scholar] [CrossRef]
  14. Zhang, X.; Sun, Y.; Jiang, K.; Li, C.; Jiao, L.; Zhou, H. Spatial sequential recurrent neural network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4141–4155. [Google Scholar] [CrossRef]
  15. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; p. 30. [Google Scholar]
  17. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral Image Classification Using a Hybrid 3D-2D Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar] [CrossRef]
  18. Feng, J.; Gao, Z.; Shang, R.; Zhang, X.; Jiao, L. Multi-Complementary Generative Adversarial Networks with Contrastive Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5520018. [Google Scholar] [CrossRef]
  19. Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.-S. Spectral–Spatial Transformer Network for Hyperspectral Image Classification: A Factorized Architecture Search Framework. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5514715. [Google Scholar] [CrossRef]
  20. He, Z.; Xia, K.; Ghamisi, P.; Hu, Y.; Fan, S.; Zu, B. HyperViTGAN: Semisupervised Generative Adversarial Network with Transformer for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 6053–6068. [Google Scholar] [CrossRef]
  21. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral Image Transformer Classification Networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528715. [Google Scholar] [CrossRef]
  22. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterhiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  23. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  24. Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-Based Adaptive Spectral–Spatial Kernel ResNet for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7831–7843. [Google Scholar] [CrossRef]
  25. Chen, Y.N.; Thaipisutikul, T.; Han, C.C.; Liu, T.-J.; Fan, K.-C. Feature line embedding based on support vector machine for hyperspectral image classification. Remote Sens. 2021, 13, 130. [Google Scholar] [CrossRef]
  26. Kalaiarasi, G.; Maheswari, S. Deep proximal support vector machine classifiers for hyperspectral images classification. Neural Comput. Appl. 2021, 33, 13391–13415. [Google Scholar] [CrossRef]
Figure 1. The overall framework of OSANet-TSSVM.
Figure 1. The overall framework of OSANet-TSSVM.
Remotesensing 16 01010 g001
Figure 2. Details of basic residual block and channel attention module.
Figure 2. Details of basic residual block and channel attention module.
Remotesensing 16 01010 g002
Figure 3. Details of orthogonal self-attention module.
Figure 3. Details of orthogonal self-attention module.
Remotesensing 16 01010 g003
Figure 4. The classification process of TSSVM.
Figure 4. The classification process of TSSVM.
Remotesensing 16 01010 g004
Figure 5. Indian Pines dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Figure 5. Indian Pines dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Remotesensing 16 01010 g005
Figure 6. Indian Pines dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Figure 6. Indian Pines dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Remotesensing 16 01010 g006
Figure 7. Pavia University dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Figure 7. Pavia University dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Remotesensing 16 01010 g007
Figure 8. Pavia University dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Figure 8. Pavia University dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Remotesensing 16 01010 g008
Figure 9. Salinas dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Figure 9. Salinas dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Remotesensing 16 01010 g009
Figure 10. Salinas dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Figure 10. Salinas dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Remotesensing 16 01010 g010
Figure 11. KSC dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Figure 11. KSC dataset. (a) Pseudocolour image for HSI. (b) Ground truth map.
Remotesensing 16 01010 g011
Figure 12. KSC dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Figure 12. KSC dataset. (a) Ground truth; (b) SVM; (c) AENSVM; (d) TSSVM; (e) ResNet; (f) SSRN; (g) A2S2K-ResNet; and (h) OSANet-TSSVM.
Remotesensing 16 01010 g012
Table 1. One-vs-All one-step SVM classification discrimination.
Table 1. One-vs-All one-step SVM classification discrimination.
Hyperspectral Feature VectorSub-Classifier
f 1
Sub-Classifier
f 2
Sub-Classifier
f 3
Sub-Classifier
f M
Classification
x 1 f 1 ( x 1 ) f 2 ( x 1 ) f 3 ( x 1 ) f M ( x 1 ) Max
x 2 f 1 ( x 2 ) f 2 ( x 2 ) f 3 ( x 2 ) f M ( x 2 ) Max
x k f 1 ( x k ) f 2 ( x k ) f 3 ( x k ) f M ( x k ) Max
Table 2. One-vs-All two-step SVM classification discrimination.
Table 2. One-vs-All two-step SVM classification discrimination.
Hyperspectral Feature VectorSub-Classifier
d f 1
Sub-Classifier
d f 2
Sub-Classifier
d f 3
Sub-Classifier
d f 4
x 1 ˜ d f 1 ( x 1 ˜ ) d f 2 ( x 1 ˜ ) d f 3 ( x 1 ˜ ) d f 4 ( x 1 ˜ )
x 2 ˜ d f 1 ( x 2 ˜ ) d f 2 ( x 2 ˜ ) d f 3 ( x 2 ˜ ) d f 4 ( x 2 ˜ )
x k ˜ d f 1 ( x k ˜ ) d f 2 ( x k ˜ ) d f 3 ( x k ˜ ) d f 4 ( x k ˜ )
Table 3. Number of samples, test set, and training set for each class in the Indian Pines dataset.
Table 3. Number of samples, test set, and training set for each class in the Indian Pines dataset.
No.ClassSampleTrainingTesting
1Alfalfa54549
2Corn-notill14341431291
3 Corn-mintill83483751
4Corn23424210
5Grass-pasture49748449
6Grass-trees74773674
7Grass-pasture-mowed26323
8Hay-windrowed48948441
9 Oats20218
10Soybean-notill96897871
11Soybean-mintill24682452223
12Soybean-clean61459555
13Wheat21220192
14Woods12941261168
15Buildings-Grass-Trees-Drives30039261
16Stone-Steel-Towers95986
Total10,28610249262
Table 4. Classification results of different algorithms on Indian Pines dataset.
Table 4. Classification results of different algorithms on Indian Pines dataset.
Class No.SVMAENSVMTSSVMResNetSSRNA2S2K-
ResNet
OSANet-
SVM
177.7880.9574.1980.4977.3595.1297.56
276.5575.3580.7388.0293.3397.3597.90
376.6580.1777.6883.5395.6697.5298.26
464.6268.9066.4960.0992.5688.7397.65
593.5895.9994.6888.2892.2799.2199.31
695.3791.0694.3392.0994.6099.8599.54
760.0086.6777.7836.0055.5680.0092.00
894.1697.7393.0398.8499.7799.6299.81
952.6347.6271.435.5680.0088.89100.00
1071.6474.3377.9778.2995.2195.7797.49
1179.0186.0582.6788.3396.4099.0099.28
1281.9784.8486.1866.2992.5792.1394.76
1395.2693.9795.8892.4392.06100.00100.00
1490.9593.7093.9992.3699.8298.0798.42
1579.3069.4778.6774.6489.8297.1297.05
1693.3398.7095.7170.2492.4297.6296.43
OA (%)82.0584.3384.5785.3396.2497.5398.19
AA (%)82.4482.8483.2980.2186.9497.1197.73
K × 10079.4682.1482.9483.2694.1197.2498.13
Table 5. Number of samples, test set, and training set for each class in the Pavia University dataset.
Table 5. Number of samples, test set, and training set for each class in the Pavia University dataset.
No.ClassSampleTrainingTesting
1Asphalt66311006531
2Meadows18,64910018,549
3 Gravel20991001999
4Trees30641002964
5Painted metal sheets13451001245
6Bare Soil50291004929
7Bitumen13301001230
8Self-Blocking Bricks36821003582
9 Shadows947100847
Total42,77690041,876
Table 6. Classification results of different algorithms on Pavia University dataset.
Table 6. Classification results of different algorithms on Pavia University dataset.
Class No.SVMAENSVMTSSVMResNetSSRNA2S2K-
ResNet
OSANet-
SVM
196.6294.6796.2791.0999.5498.7999.49
296.6696.5997.4092.0999.3698.8999.56
365.2273.2767.4959.1883.6798.9595.65
477.9490.1584.1495.8899.2097.5099.33
596.4099.6899.30100.00100.00100.00100.00
668.5478.3275.8573.1697.6698.2398.35
759.4870.7060.2863.0993.2398.3795.25
876.0887.2078.6089.9296.5995.7395.77
9100.00100.00100.0097.40100.0099.6599.60
OA (%)85.2290.5288.1487.7497.1798.4798.64
AA (%)89.3487.8491.0487.8897.6798.0198.34
K × 10080.8987.4884.4083.5797.2997.9598.17
Table 7. Number of samples, test set, and training set for each class in the Salinas dataset.
Table 7. Number of samples, test set, and training set for each class in the Salinas dataset.
No.ClassSampleTrainingTesting
1Brocoli_green_weeds_120091001909
2Brocoli_green_weeds_237261003626
3 Fallow19761001876
4Fallow_rough_plow13941001294
5Fallow_smooth26781002578
6Stubble39591003859
7Celery35791003479
8Grapes_untrained11,27110011,171
9 Soil_vinyard_develop62031006103
10Corn_senesced_green_weeds32781003178
11Lettuce_romaine_4wk1068100968
12Lettuce_romaine_5wk19271001827
13Lettuce_romaine_6wk916100816
14Lettuce_romaine_7wk1070100970
15Vinyard_untrained72681007168
16Vinyard_vertical_trellis18071001707
Total54,129160052,529
Table 8. Classification results of different algorithms on the Salinas dataset.
Table 8. Classification results of different algorithms on the Salinas dataset.
Class No.SVMAENSVMTSSVMResNetSSRNA2S2K-
ResNet
OSANet-
SVM
199.95100.0099.8489.79100.00100.0099.54
299.4299.9499.3799.4599.12100.00100.00
399.2199.0599.4799.9596.1298.4299.81
498.3198.3298.5497.9196.4199.0199.82
596.3398.7397.1197.4099.8099.78100.00
699.74100.0099.9598.55100.00100.00100.00
799.9199.6899.8898.7999.4399.56100.00
882.0579.4582.4687.1492.2894.1896.98
999.3698.9998.9299.5999.00100.00100.00
1092.7195.1993.5994.6599.4798.3599.97
1197.9696.9898.0799.2594.4298.6799.87
1298.3899.2498.9799.5199.8498.32100.00
1399.7599.2699.75100.0099.15100.00100.00
1490.5794.6592.5798.3598.4899.7999.94
1566.1868.6469.4580.2590.6093.1297.96
1698.1999.2998.8299.2499.1399.3099.73
OA (%)90.2190.6391.0393.4095.4997.7398.32
AA (%)95.9095.4696.1695.3697.2197.4599.29
K × 10089.1089.5490.0092.5896.0898.7198.11
Table 9. Number of samples, test set, and training set for each class in the KSC dataset.
Table 9. Number of samples, test set, and training set for each class in the KSC dataset.
No.ClassSampleTrainingTesting
1Scrub76176685
2Willow swamp24324219
3 CP hammock25626230
4Slash pine25225227
5Oak/Broadleaf16116145
6Hardwood22923206
7Swamp1051194
8Graminoid marsh43143388
9 Spartina marsh52052468
10Cattail marsh40440364
11Salt marsh41942377
12Mud flats50350453
13Water92793834
Total52115214690
Table 10. Classification results of different algorithms on the KSC dataset.
Table 10. Classification results of different algorithms on the KSC dataset.
Class No.SVMAENSVMTSSVMResNetSSRNA2S2K-
ResNet
OSANet-
SVM
195.4793.6695.8299.8596.24100.0099.62
293.0291.3695.1968.0495.8197.72100.00
390.3482.2687.6753.0488.8990.5298.32
471.2370.7176.8942.7386.0972.6984.66
576.0381.1375.9420.6980.8093.8196.55
672.1286.9882.0897.5789.33100.0098.12
785.2686.5284.6296.8186.6789.36100.00
894.4392.1394.5279.1294.1298.2099.67
994.0792.6196.2499.3696.2499.2399.73
1096.6996.0397.8192.0394.9299.24100.00
1195.3299.7396.44100.0099.65100.00100.00
1297.4494.0497.4789.1898.53100.00100.00
1399.05100.0099.46100.0099.54100.00100.00
OA (%)92.8092.6593.0987.2295.1998.4598.75
AA (%)88.9189.7890.2380.2192.9197.4397.94
K × 10091.9991.8192.4285.8994.6597.1598.63
Table 11. Ablation experiment.
Table 11. Ablation experiment.
Model(1)(2)(3)(4)(5)(6)
OSM--
CAM---
TSSVM----
Indian PineOA97.2797.7997.5598.0198.0698.19
AA96.4397.0496.2397.8397.9797.73
Kappa97.3697.6197.7797.6698.3698.13
PaviaUOA97.3197.4097.3597.8198.3998.64
AA96.2996.6996.8397.0897.5098.34
Kappa97.2197.8996.3497.8598.4898.17
SalinasOA97.6497.7897.7097.9798.1898.32
AA98.6598.7598.0499.0599.0199.29
Kappa97.1297.9497.1797.2997.4198.11
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, H.; Wang, L.; Liu, H.; Sun, Y. Hyperspectral Image Classification with the Orthogonal Self-Attention ResNet and Two-Step Support Vector Machine. Remote Sens. 2024, 16, 1010. https://doi.org/10.3390/rs16061010

AMA Style

Sun H, Wang L, Liu H, Sun Y. Hyperspectral Image Classification with the Orthogonal Self-Attention ResNet and Two-Step Support Vector Machine. Remote Sensing. 2024; 16(6):1010. https://doi.org/10.3390/rs16061010

Chicago/Turabian Style

Sun, Heting, Liguo Wang, Haitao Liu, and Yinbang Sun. 2024. "Hyperspectral Image Classification with the Orthogonal Self-Attention ResNet and Two-Step Support Vector Machine" Remote Sensing 16, no. 6: 1010. https://doi.org/10.3390/rs16061010

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop