Next Article in Journal
Information Storage in a Black Hole’s Gravitational Field
Previous Article in Journal
Are Human Judgments of Real and Fake Faces Quantum-like Contextual?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CViT Weakly Supervised Network Fusing Dual-Branch Local-Global Features for Hyperspectral Image Classification

1
School of Information and Communication, Guilin University of Electronic Technology, Guilin 541004, China
2
Beidou Navigation Technology Center, Guangxi Institute of Industrial Technology for Space-Time Information, Nanning 530201, China
*
Authors to whom correspondence should be addressed.
Entropy 2025, 27(8), 869; https://doi.org/10.3390/e27080869
Submission received: 18 June 2025 / Revised: 30 July 2025 / Accepted: 12 August 2025 / Published: 15 August 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

In hyperspectral image (HSI) classification, feature learning and label accuracy play a crucial role. In actual hyperspectral scenes, however, noisy labels are unavoidable and seriously impact the performance of methods. While deep learning has achieved remarkable results in HSI classification tasks, its noise-resistant performance usually comes at the cost of feature representation capabilities. High-dimensional and deep convolution can capture rich deep semantic features, but with high complexity and resource consumption. To deal with these problems, we propose a CViT Weakly Supervised Network (CWSN) for HSI classification. Specifically, a lightweight 1D-2D two-branch network is used for local generalization and enhancement of spatial–spectral features. Then, the fusion and characterization of local and global features are achieved through the CNN-Vision Transformer (CViT) cascade strategy. The experimental results on four benchmark HSI datasets show that CWSN has good anti-noise ability and ensures the robustness and versatility of the network facing both clean and noisy training sets. Compared to other methods, the CWSN has better classification accuracy.

Graphical Abstract

1. Introduction

Hyperspectral image (HSI) is a three-dimensional data cube with narrow spectral bands and continuous spectral curves. It has the advantage of merging images and spectrum, resulting in superior spectral discrimination. Compared to remote sensing data such as panchromatic and multispectral images, HSI offers more significant advantages in identifying accurate subtle differences of objects, extracting complex characteristics of objects, and providing detailed spectral information of objects. HSI classification is a critical part of the application of hyperspectral remote sensing technology, and its classification results have important application value in humanities and natural domains, such as fine agriculture [1,2], resource exploration [3,4], and ecological monitoring [5,6].
With the continuous update of hyperspectral sensors, the corresponding image processing and information classification and extraction methods have also advanced significantly [7,8,9,10]. HSI interpretation in the supervised paradigm is divided into two phases: feature extraction [11,12] and object classification. Among them, the feature extraction stage progresses from using spatial or spectral characteristics alone to combining spatial and spectral features [13,14,15,16,17,18,19]. In terms of classification algorithms, the most representative traditional machine learning methods include decision tree classification [20,21], support vector machines [22,23], sparse representation [24,25], and multiple classifier systems [26,27]. However, they mainly rely on manually designed shallow features, and their performance is limited for data with complicated nonlinear relationships and rich feature information. Deep learning methods, with their robust hierarchical feature representation capabilities and autonomous learning qualities, broaden the application scope in remote sensing. CNNs, for example, offer a wide range of effective solutions for HSI classification. Three-dimensional CNNs can perform joint spatial–spectral feature extraction, and with the increase in the number of layers and the number of convolutional kernels, they can capture rich deep semantic features, which is correspondingly accompanied by higher computational resources, time costs, and training sample requirements. Convolutional neural networks, such as group convolution and depthwise separable convolution, can retain their performance while reducing the number of parameters. Due to their limited receptive field, traditional CNNs are unable to learn the global contextual information of the HSI cube. To address the issue of a limited receptive field, researchers have proposed several methods to improve the understanding of global information by CNNs. One method is to increase the number of network layers, which can naturally expand the receptive field. However, the inherent problem of vanishing or exploding gradients was not overcome until the emergence of residual networks (ResNets). The residual connection not only makes the construction of deep networks possible, but also provides a better training environment for the network, ensures the effective transfer of information and the stability of the training process, and improves the expressiveness and overall robustness of the model. Second, the size and structure of the convolution kernel are designed for feature fusion at different scales, as seen with the introduction of inflated convolution and the inception module. Third, long-range information transfer and global dependence modelling are achieved by calculating the interactions between any two positions in the input feature map, such as the introduction of non-local blocks, spatial–spectral dual-attention strategy, and self-attention mechanism. In particular, the Transformer [28,29,30], a representative model of the self-attention mechanism, has been gradually extended and derived into the field of computer vision following its achievement in Natural Language Processing (NLP) due to its powerful modelling capability. For example, Vision Transformer (ViT), CNN-Transformer, inherits the core components of Transformer on top of CNNs and uses the multi-head self-attention (MHSA) mechanism to allow each position to interact with all the other positions, helping the model to better understand the local details and global structure of the HSI cube.
These classification methods based on supervised paradigms need to be based on high-confidence training samples [31,32,33,34]. However, due to the influence of environmental interference and manual labeling errors, noise samples are unavoidable in the training set, which seriously affects the performance of the classification method. Therefore, related research has been carried out, for example, designing deep learning network architectures to improve robustness, but deep or high-dimensional convolution suffers from high resource consumption and optimization difficulties. The noise label removal strategy tends to over-clean the dataset, greatly reduces the number of training samples, and also results in the loss of real samples and feature information.
To improve the information suppression ability of semantic feature extraction and noisy training samples for deep learning in HSI classification, this paper proposes a CViT Weakly Supervised Network (CWSN). First, the model creates a spectral and spatial information deep extraction network using one-dimensional and two-dimensional convolutional functions, and then efficiently fuses the shallow features with the high-level features by residual connections to alleviate the gradient vanishing problem, and further enhance the expression of discriminative spatial spectral features. The spatial–spectral features are then processed using CNN-Vision Transformer (CViT) multi-head self-attention to capture the long-range dependence between image elements. In addition, a noise suppression loss function is created using the cross-entropy loss function theory to effectively guide the network in learning and utilizing the prior knowledge. The main contributions of this work are listed below.
(1)
In this paper, we propose a CWSN that integrates a lightweight dual-branch feature enhancement module and a CNN-Vision Transformer, while organically integrating deep semantic feature extraction and noisy sample processing into a deep learning framework.
(2)
A Dual-Branch Local Induction Module (DBLIM) is designed, which has a simple architecture, a small number of parameters, and a high generalization capacity. This module can enhance the discriminative and divisible nature of different classes of feature information, and mitigate the gradient vanishing of the depth model.
(3)
Local and global deep semantic features are generalized and characterized using CViT, together with Noise Suppression Loss (NSL), which enhances the robustness of the model and makes it stable in the face of both clean and noisy training sets.
The rest of the paper is organized as follows. In Section 2, some related work is presented. Section 3 describes the proposed algorithm in detail. Section 4 includes the design of the model components, a comparative evaluation of the classification performance, and an application analysis. Finally, conclusions are drawn in Section 5.

2. Related Work

2.1. HSI Supervised Classification

Over the last two decades, there have been numerous classification methods used in remote sensing image classification research, ranging from manual visual interpretation to manual repair based on automatic computer classification to automatic computer classification using supervised and unsupervised algorithms. In particular, new machine learning and deep learning methods have evolved in recent years, allowing for significant advances in classification accuracy and efficiency, with HSI classification mixed with artificial intelligence algorithms serving as the current research frontier. Trias-Sanz et al. [35] summarized and investigated the combination of color and texture features using quantitative evaluation of hierarchical segmentation algorithms to achieve the best results for the task of segmentation of high-resolution multispectral aerial images in rural areas. However, the use of spatial features alone does not apply to lower spatial resolution remote sensing image classification. Xia et al. [36] proposed an integrated method, Rotation Random Forest via KPCA (RoRF-KPCA), in which KPCA extracts information from each subset of the original feature space, and the resulting feature sets are merged and fed into the RF classifier. However, traditional spectral classifiers have low classification accuracy values for feature classes with similar spectral responses. Therefore, many researchers began to develop extraction and optimization techniques using spatial–spectral features, with deep learning being the most representative. Sun et al. [37] proposed Spatial–Spatial Feature Tokenization Transformer (SSFTT) to extract shallow spectral and spatial features through a designed convolutional module and transformed them using Gaussian-weighted feature tokenization. He et al. [38] proposed the Spatial–Spectral Transformer (SST) to extract spatial and spectral features, with a multilayer perceptron performing the final classification task. HSI classification methods using spatial–spectral features have achieved considerable results on reliable training datasets; however, these methods lack robustness when dealing with noisy labels.

2.2. HSI Weakly Supervised Classification

In order to mitigate the impact of noisy labels on HSI classification, researchers have successively developed solutions from different perspectives and proposed a noisy label learning strategy.
For the correction of noisy labels, Kang et al. [39] developed a mislabeled sample detection and correction method based on edge preserving filtering and spectral constraints, which utilizes spectral constraints to identify and correct mislabeled samples in training samples while preserving the edge features of the image. Li et al. [40] proposed the Adaptive Selective Loss Propagation (ASLP), which first identifies an adaptive number of samples with lower loss values from the untrusted set to supplement the trusted set, and then performs label propagation based on the enlarged trusted set to achieve label correction. Leng et al. [41] designed the Sparse Graph-Based Adaptive Label Propagation (SALP) algorithm, which relies on the assumption that the noise has a Gaussian distribution and provides an effective mechanism for detecting noisy labels in the training set. The above label correction methods perform better in simple noise environments (with random noisy labels), and all of them are combined with machine learning methods to complete the classification. It remains to be investigated if they are applied to complex noise environments or deep learning.
For noisy label removal, Tu et al. [42] first developed a Density Peak-based Noisy Label Detection (DPNLD) method that detects and removes noisy labels based on distance and local density. Later, a noisy label detection framework [43] was proposed, which combines Superpixel-to-Pixel Weighting Distance (SPWD) and density peak clustering to accurately detect and remove noisy labels from the training set before HSI classification. Fang C et al. [8] proposed a deep reinforcement learning framework method for learning each label’s correctness and removing different types of label noise. Class methods for noisy label removal tend to over-clean the dataset, resulting in the loss of real samples and feature information.
In terms of noisy label suppression, the robustness of the classifier is mainly improved by improving the structure of the deep learning network, creating the anti-noise loss function, and evaluating the label confidence, etc. To extract the spectral spatial features, Ghaderizadeh et al. [44] combined a 3D fast-learning block and a 2D convolutional neural network. This 3D hybrid convolutional neural network (3DCNN) reduces the model complexity while also being effective against noise and a limited number of training samples. Wang et al. [45] proposed a Dual Spectral–Spatial Residual Adaptive (DSSRA) network with an Adaptive Noise-Robust Loss (ANRL) function in the network to mitigate the effects of learning with noisy labels. Zhang et al. [46] proposed a Triple Contrastive Representation Learning (TCRL) framework from a deep clustering perspective to learn feature information while reducing classifier overfitting to noisy labels. Yue et al. [47] developed a novel spectral-feature-based HSI multi-source label noise simulation method that evaluates the overall credibility of different label sources and the credibility of specific categories, aiming to mitigate the impact of multi-source noisy labels. To improve robustness, noise suppression methods reduce the representational capacity of its high-level semantic features, which degrades the classification performance when dealing with credible datasets.
Compared to traditional shallow machine learning methods, deep learning training requires a larger number of high-confidence training samples, making sample mislabeling a more serious issue. Therefore, while improving the ability of the deep learning model to cope with noise, it is important to achieve the full extraction and utilization of spatial–spectral information, as well as to ensure high interpretation accuracy even on small and reliable sample sets.

3. Proposed Model

We created CWSN, a novel noise suppression HSI classification model, which aims to achieve deep extraction and characterization of spatial spectral features locally and globally, as well as mitigating the effect of noisy training sets on HSI classification. Figure 1 shows the model framework, which is made up of two modules, DBLIM and CViT, and incorporates the NSL. Since the training samples for HSI classification are usually cubic in format, several 3D cubes are randomly selected from the HSI as inputs to the CWSN, and the cubes are used as an example to illustrate the proposed network structure.

3.1. Dual-Branch Local Induction Module

DBLIM is used to extract deep spatial-spectral features, which consist of a 1DCNN-based spectral feature extraction channel and a 2DCNN-based spatial feature extraction channel. Unlike the standard convolution, DBLIM employs a dual-branch structure for local induction and enhancement of spatial and spectral features. In addition, residual connections are added to the spectral branches. This is because the spatial branch has a lower residual dependence due to its localization. Spectral branching has a high information dimension and strong correlations between bands; therefore, deep networks are needed to model the complex relationships and global dependencies among bands. Residual connection aids deep networks in stabilizing training, adequately extracting nonlinear features, mitigating gradient vanishing, and improving classification accuracy. Additionally, since 1D convolution has fewer parameters, residual blocks do not considerably increase the computational amount. The combination of multilayer 1D convolution and residual blocks enhances feature extraction capability and stability, and the input image block achieves a robust spectral representation while preserving the original spatial information.

3.1.1. Spectral Feature Extraction Channel

The structure of the spectral feature extraction channel is shown in Figure 2, which is based on 1DCNN to achieve a lightweight design and reduce the number of parameters. The input feature F i n R w × h × m is first pooled globally on average in order to obtain a one-dimensional vector feature F m of size 1 × 1 × m. The expression of F m is as follow:
F m = G A P ( F i n ) , F m R 1 × 1 × m
The 1D convolution kernel is set to 1 × 1 × m to efficiently extract spectral features without affecting the original spatial features. This size setting ensures that features are not aggregated spatially, but cover all bands in the spectral dimension. The advantages are as follows: avoiding interference from spatial noise, such as mixed image elements, while simultaneously highlighting the spectral characteristics; and capturing global features of the spectral curve and avoiding loss of spectral continuity in segmental processing. In the first convolution layer, 64 convolution kernels of 1 × 1 × m size and a sampling step of (1, 1, m) size are used to perform a convolution operation on the input w × h × m 3D cube to obtain the w × h × m 3D cube features, eliminate redundant spectral features, and better focus on the features that are crucial in classification. Then, a residual block consisting of two consecutive convolutional layers is used for deep feature extraction and enhancement of the first layer of convolutional features, emphasizing the important regions of the image and improving the spectral robustness when dealing with the noisy labels. A batch normalization layer and an activation function are applied after each convolutional operation in order to obtain the spectral dimensional feature weighting information ω s p e .
ω s p e = σ ( c o n v 1 D k ( F m ) )
where σ ( ) and c o n v 1 D k ( ) denote the ReLU activation function and one-dimensional convolution function.
Following the residual block, the input feature is dot-multiplied with the spectral weight information to obtain the output feature F o u t R w × h × 1 enhanced by spectral features, which can discriminatively extract the spectral features while maintaining the separability between spectral information classes. The expression is as follows.
F o u t = ω s p e F m , F o u t R w × h × 1

3.1.2. Spatial Feature Extraction Channel

For spatial feature extraction, a 2D convolution module is used to extract the spatial correlation features of the center pixel and its surrounding 8-neighborhood, as shown in Figure 3. Similarly, assuming that the input feature is F i n R w × h × m , in this paper, we first performed mean pooling and maximum pooling along the spectral dimension, and performed feature stacking with the following expression:
F m 1 = M E A N P ( F i n ) , F m 1 R w × h × 1
F m 2 = M A X P ( F i n ) , F m 2 R w × h × 1
F m 3 = C a t ( 3 , F m 1 , F m 2 ) , F m 3 R w × h × 3
where M E A N P ( ) , M A X P ( ) , and C a t ( ) denote the mean pooling operation, maximum pooling operation, and spectral dimension feature stacking operation, respectively.
The 1 × 1 convolution cannot capture spatial context information, while a larger kernel (e.g., 5 × 5) will increase the computation and lead to over-smoothing of features. Therefore, the 2D convolution kernel size is set to 3 × 3. This setting balances the receptive field and computational efficiency, and can effectively capture local spatial features (e.g., edges, points). Next, the first layer utilizes 128 convolution kernels of size 3 × 3 to perform a convolution operation on F m 3 Then, the second layer uses two convolution kernels of the same size to perform successive convolutions, and the strides are (1, 1). At the same time, the ReLU activation function is utilized to obtain the information about the spatial weights ω s p a of the image with the following expression:
ω s p a = σ ( c o n v 2 D 7 × 7 ( F m 3 ) )
Finally, the input feature is dot-multiplied with the spatial weight information to obtain the spatial feature-enhanced output feature F o u t 3 R w × h × m , so as to retain and emphasize important spatial information. The expression is as follows:
F o u t 3 = ω s p a F m 3 , F o u t 3 R w × h × m

3.2. CViT Hybrid Structure

3.2.1. CViT

CNNs have the advantage of local feature induction, but are unable to characterize the long-range dependence, while ViT [48] has an excellent performance in long-range correlation characterization. Thus, the CNN and Transformer fusion structure can effectively capture the spatial and spectral features of HSI. Therefore, in this paper, a CViT hybrid structure was used for deep splicing and characterization of spectral and spatial features.
Spectral and spatial features are first spliced and stacked by spectral dimension. The stacked features are then projected nonlinearly using a feature mapping layer built with one-dimensional convolution, the ReLU activation function, and layer normalization. Specifically, assuming that F i n R w × h × m is the input feature of the CViT module, thirty-two convolutional kernels of size 1 × 1 are first used for the feature mapping of the input features to obtain the coded features Y i n , and its expression is shown below:
Y i n = C o n v 1 D ( F i n )
In order to capture the spatial long-range feature dependencies in an image block, position information I p o s is added to the coded features Y i n with the following expression:
Y ^ i n = Y i n + I p o s
After convolutional coding, spectral and spatial features are fused to achieve local and global deep features. A multi-head self-attention (MHSA) block, a multi-layer perceptron (MLP), and two layer normalization (LN) processes are used to model the depth relationship of the semantically tagged feature Y ^ i n . Notably, two residual connections are utilized before the MHSA block and the MLP layer to prevent information loss. In order to learn the deep semantic features, three learnable weight matrices W Q , W K , and W V are defined, and Y ^ i n is linearly mapped to query information Q, key information K, and value information V, respectively, based on the above three learnable weight matrices. Then, the attention score is obtained based on the product of Q and K, and the weight of the score is obtained by the SoftMax function. The expression of the self-attention mechanism SA is defined as follow:
S A = A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d ) V
where d is the dimension of K.
MHSA uses multiple sets of weight matrices to map Q, K, and V using the same process. Output projections of features are performed using modules with the same structure as the mapping layer. Thus, the final self-attention results are stacked together for output, utilizing the expression defined below:
M H S A ( Q , K , V ) = C a t ( S A 1 , S A 2 , , S A h ) W
where h is the number of heads, and W is a linear transformation matrix that ensures that the output feature channels are consistent.
Finally, the globally characterized features Y ^ i n are sequentially input to the MLP layer and LN. It is worth mentioning that the MLP layer consists of two fully connected layers with an activation function in the middle. The fully connected layers are used to obtain the final class information. Furthermore, layer normalization can prevent gradient explosion or vanishing while also accelerating network convergence.

3.2.2. NSL Function

This paper focuses on HSI weakly supervised classification; hence, the mechanism and definition of noisy label generation in the experiments are presented initially. For each class, n samples with true labels are randomly selected to form the initial training set. Then, uniformly select m samples from other classes and label them as the current class. After that, all the samples in each class are mixed to ensure that the noisy samples are fully incorporated into the initial training set. Then, a class C classification training set with noisy labels can be defined as S c = ( r 1 , l 1 T ) , , ( r n , l n T ) , ( ( r n + 1 , l n + 1 T ) , , ( r n + m , l n + m T ) , where r denotes the training samples and l 1 , 2 , , C denotes the labels of the training samples (containing noisy labels). In addition, l * is introduced to denote the possible correct class labels. Based on the independent homogeneous distribution assumption, there are:
P ( l = c l = j , r ) = P ( l = c l = j )
Because all samples are randomly disrupted, each label has the probability θ = m / ( n + m ) of being mislabeled. Therefore, the probability that sample j from class i has a potentially clean label can be expressed as:
θ i j = θ C 1 , i j θ i j = 1 θ , i = j
where θ i j = P ( l = i l = j ) .
The anti-noise loss function’s central notion is to ensure that, when training on noisy samples, the optimal model parameters are also the optimal parameters when all samples are labeled correctly. To better demonstrate how the anti-noise loss function works, for the noisy training set described above, the following equation is used to describe the process of noisy label suppression:
E r , l L ( f ( r ) , l ) E r , l L ( f ( r ) , l ) , f F
where F is the possible parameter space of the model, f is the possible solution, and f is the global minimum of the empirical risk when training with potentially clean labels l .
We united the normalization cross-entropy (NCE) loss and reverse cross-entropy (RCE) loss to form a noise loss function NSL [45]. The NSL is used after the MLP layer in order to effectively suppress the noisy samples in the training set and obtain the optimal weight parameters of the network model. It is worth mentioning that the RCE loss exchanges the positions of the predicted value p ( l r ) and the true value q ( l r ) in the ordinary cross-entropy (CE) loss. The expressions of NCE, RCE, and NSL are shown below:
N C E ( p , q ) = c = 1 C q ( l r ) log p ( l r ) j = 1 C c = 1 C q ( l = j r ) log p ( l = j r )
R C E ( p , q ) = c = 1 C q ( c | r ) log p ( c | r )
N S L ( p , q ) = ( N C E ( p , q ) + R C E ( p , q ) ) / 2

4. Experiments and Analysis of Results

4.1. Datasets and Experimental Setup

4.1.1. Experimental Datasets

The component design, classification performance evaluation, and application analysis of the model were performed on four hyperspectral datasets using three evaluation metrics, Overall Accuracy (OA), Average Accuracy (AA), and Kappa coefficient. In the experiment, two noise label levels were set up to test the stability of the model under different noise intensities. Specifically, the low noise rate was set at about 20% for modeling light labeling noise, and the medium noise rate was set around 30% for modeling errors such as mixed pixels. The noise settings for each dataset are shown in Table 1. In addition, the experiment was repeated 10 times in each experimental environment, and the average value was used as the evaluation result.
(1)
University of Pavia (UP): The ROSIS 03 sensor collected data from the University of Pavia campus in Italy. The spatial resolution is 1.3 m, with a spectral coverage of 0.43–0.86 μm. Nine classes of ground cover are included in the datasets.
(2)
Washington DC (WDC): The Washington DC dataset was collected using the Hyperspectral Digital Imaging Experiment sensor over Washington DC. Seventy-eight spectral bands were selected from the 400–1000 nm spectral range to form the HSI, and the corresponding RGB images were acquired using a Sentinel-2 SRF. The dataset contains 480 scan lines with 307 pixels on each scan line.
(3)
Salinas Valley (SV): The Salinas Valley dataset was collected by the Airborne Visible Infrared Imaging Spectrometer (AVIRIS), a sensor imaging the Salinas Valley region of California, USA. The spatial resolution is 3.7 m, and the number of bands is 224, covering an area with an image size of 512 rows and 217 columns, containing 16 object classes. Twenty water vapor absorption and noise bands (numbers 108–112, 145–167, and 224) have been removed from the data, leaving 204 active bands.
(4)
Kennedy Space Centre (KSC): The Kennedy Space Centre dataset was collected by the AVIRIS instrument during an overflight of the Kennedy Space Centre in the USA, although with a low spatial resolution of 18 m. These data include raw spatial dimensions of 512 × 614 pixels, with 48 bands removed due to absorption and low signal-to-noise ratios, and 176 spectral bands used for analysis. It contains 13 classes representing different object classes.

4.1.2. Experimental Settings

In the spatial feature extraction stage, a two-dimensional convolution function with a convolution kernel size of 3 was used to execute three consecutive spatial feature extractions, with 128, 32, and 32 convolution kernels each. The spatial attention mechanism used a convolution kernel with a size of 7. In the spectral feature extraction stage, a one-dimensional convolution function with a convolution kernel size of 1 was used to perform four consecutive spectral feature extractions, in which the numbers of convolution kernels each time were 64, 64, 64, and 128, respectively. The spectral attention mechanism’s convolution kernel had a size of 3. In addition, the number of feature dimensions, the number of multi-heads, and the feature learning depth in the ViT module were set to 32, 1, and 1, respectively.
In the training stage, this paper used the image block and its class label corresponding to the center pixel with a window size of 9 × 9 as samples to construct the training sample set. Each batch contained 16 training samples, with the learning rate set to 0.0001. In addition, this paper adopted the gradient optimization algorithm based on the Adaptive Moment Estimation (Adam) to train the CWSN. It is worth mentioning that the network hyperparameter settings in the inference phase are the same as in the training phase. The network model was run by the Pytorch framework (https://pytorch.org/) in the Windows 11 environment with 16 GB RAM and an Nvidia RTX 2080 GPU (ASUS, Shenzhen, China).
In the simulation stage, the experiment was performed 10 times in each set of experimental conditions, and finally, the average value was used as the evaluation result.

4.2. Parameter Sensitivity Analysis

To explore the effect of parameter variations on the performance of the CWSN model, this section conducts a sensitivity analysis of two hyperparameters in the CViT structure: feature dimension K and number of attention heads h. The parameters K and h have values in the ranges of [16, 32, 64, 128, 256] and [1, 2, 3, 4, 5]. The parameter K determines the dimensionality of each layer’s feature vectors inside the model, with higher feature dimensions providing richer feature representations and helping the model in capturing more information, details, and feature relationships. The parameter h determines the number of attention heads that are processed in parallel in the multi-head attention mechanism; different attention heads can capture different types of dependencies, such as local or global dependencies. This diversity and complementarity allow the model to gain a more complete understanding of the data. In particular, increasing K increases the feature dimension of the data processed by CWSN, and increasing h motivates CWSN to use multi-segment features to compute long-range dependencies, but it also reduces the interpretability of the model and increases computational complexity and memory consumption.
Figure 4 shows how changing the parameters K and h affects the CWSN model’s classification accuracy. Figure 4a shows that, on the KSC dataset, the OA generally increases and then decreases as the values of K and h increase, reaching the optimum when K = 64 and h = 3, indicating that the CWSN model performance is gradually optimized with changes in the values of K and h. The results in Figure 4b–d also prove that changes in values of K and h affect the performance of the CWSN model, with the optimal values of K and h being close to K = 64 and h = 3. Therefore, in this paper, based on the experimental results, K = 64 and h = 3 were set as the default parameters of the CWSN model.

4.3. Ablation Analysis

In this section, ablation experiments on the KSC dataset are used to evaluate the performance contribution of each module in the CWSN model. For ease of description, the convolutional part of the CWSN model is denoted as SBL (CWSN Baseline), the spectral feature extraction channel is denoted as SpeA, the spatial feature extraction channel is denoted as SpaA, and the ViT structure is denoted as TF. Table 2 shows the experimental results obtained on the KSC dataset using SBL, SpeA, SpaA, TF, and CWSN, with the highest accuracy highlighted in black bold. In terms of overall classification accuracy, the OA, AA, and Kappa metrics are the lowest for SBL and the highest for CWSN. And TF comes in second to CWSN. The three metric values of TF differ from CWSN by 2.33%, 2.34%, and 2.59% when the number of mislabeled samples is 6. The three metric values of TF differ from CWSN by 0.54%, 0.95%, and 061% when the number of mislabeled samples is increased to 13. The classification accuracy of CWSN is the most consistent across all object classes. As the number of mislabeled samples increases, the classification accuracy of CWSN fluctuates less and performs better in coping with noise.
To analyze the classification effect of the different components in a more intuitive way, the regions in the image that are prone to misclassification were boxed and enlarged. Figure 5 shows that the classification effect gradually becomes better from left to right: the classification maps of SBL, SpeA, and SpaA have not only lower accuracy, but also have a large number of over-smoothing and misclassification phenomena. In contrast, the classification maps of the CWSN model, which combines all the core components, have neater and clearer boundaries within the region, indicating that the CWSN model is able to effectively extract the spectral and spatial features through the synergistic effect of the components, thus obtaining the best visualized classification effect.
To better demonstrate the necessity and effectiveness of each component, in addition to the individual effects above, we conducted experiments on partial combinations of components. Four classes were randomly selected from the thirteen classes of the KSC dataset, and 100 samples from each class were selected for the feature visualization of each component. The four selected feature classes were C2, C3, C4, and C5. Figure 6b shows that the sample features have low interclass separability and high intraclass similarity when the feature extraction is performed on the raw data from the four feature classes using only SBL. Furthermore, when these core components are introduced progressively to the network structure, the representation of the sample features gradually improves in separability. When the TF module is added, the sample features show high interclass separability and intraclass similarity. This indicates that these core components play a significant role in improving the accuracy of feature classification. Further analysis shows that, when the TF module is introduced, its powerful feature representation and self-attention mechanism significantly improve the model’s capacity to differentiate and consistently recognize sample features, resulting in excellent visual effects.

4.4. Analysis of Anti-Noise Strategies

To support the value of NSL losses, NSL was compared with existing loss functions. Table 3 shows the results of the quantitative analyses on UP. According to the data in the table, it can be seen that CE has a high sensitivity to noisy samples, which leads to its low classification accuracy on some categories, for example, the classification accuracy on C3 is only 59.27%. In terms of OA accuracy, when the numbers of mislabeled samples are 14 and 27, the RCE values are 82.68% and 82.94%, and the NSL values are 84.81% and 83.46%, respectively. This indicates that NCE has better noise immunity, but the effect on a high percentage of noisy datasets is still limited. Overall, NSL performs the best on the OA, AA, and Kappa. By combining the RCE and NCE losses, NSL has good rejection of noisy samples, which makes the model more robust and stable when dealing with noisy data.
Figure 7 illustrates the qualitative classification results on UP. For the two regions boxed in Figure 7a, CE performs the worst, with a lot of pretzel noise as well as unsmooth classification. RCE and NCE outperform the CE loss function in classification and show better noise immunity characteristics. Classification is the best when NSL is employed. In summary, by weighting RCE and NCR, the NSL loss function can effectively constrain the network learning and improve the ability to classify in noisy environments.

4.5. Computational Efficiency Analysis

On the KSC, UP, SV, and WDC datasets, this section analyses and compares the running time of CWSN with comparative methods, including mislabeled sample detection and image classification time. In this experiment, all the running time statistics experiments were conducted on the same computer. During the training process, all the datasets were trained for 100 epochs. The experimental results are shown in Table 4. As can be seen from the results in the table, compared to other methods, CWSN is not superior in terms of computational efficiency. The main reason for the high time consumption is due to the following two aspects: the Vision Transformer structure and the feature extraction process of DBLIM.

4.6. Comparison of Classification Performance with Low-Confidence Samples

To investigate the robustness of the CWSN model under different numbers of mislabeled samples, DPNLD, SPWD, 3DCNN [49], ViT, and CWSN were compared on the KSC, UP, SV, and WDC datasets. Figure 8 demonstrates the classification accuracy of each method under different conditions. Taking the experimental results on the KSC dataset as an example, when there are five mislabeled samples, the OA, AA, and Kappa metrics of the CWSN model can reach 95%, 85%, and 90%, respectively, while the DPNLD and SPWD methods achieve 87%, 81%, and 86%, and 3DCNN and ViT are even lower, with an accuracy of less than 85%. When the number of mislabeled samples increased to 13, OA, AA, and Kappa metrics of the CWSN model decreased to 90%, 84%, and 89%, respectively, while the DPNLD and SPWD methods decreased to 86%, 80%, and 85%, and both 3DCNN and ViT fell below 65%. As the number of mislabeled samples increased from 5 to 13, the values of the OA, AA, and Kappa metrics generally fluctuated and decreased for each method, with the CWSN model having the smallest range of change. The above analysis is also supported by the experimental results from the other three datasets. In summary, firstly, the increase in the number of mislabeled samples has a negative impact on the performance of all classification methods, especially 3DCNN and ViT, due to the decrease in the number of clean labels, which can easily lead to underfitting of the deep learning algorithms. Secondly, the advantage of the interpretation effect of CWSN is obvious under different noise levels and datasets, suggesting that CWSN is better able to extract advanced features due to its hybrid multilevel network structure and can also be used to resist noise to a certain extent. In addition, the NSL loss function also plays an important role in noise suppression.

4.7. Comparison of Classification Performance with High-Confidence Samples

The anti-noise performance of deep learning algorithms comes at the expense of feature extraction and representation capability. In order to prove that CWSN is equally applicable to clean training sets, the classification performance of 3DCNN, ViT, and CWSN was evaluated on four datasets in a noise-free environment. The results are shown in Table 5. On the four datasets of KSC, UP, SV, and WDC, CWSN has the highest values of OA, AA, and Kappa metrics. For example, on the KSC dataset, the three metrics of CWSN are 3.64%, 4.07%, and 4.06% higher than those of 3DCNN, with the next best performance. On the SV dataset, 3DCNN has the best performance, but CWSN differs very little from it. Overall, the CWSN model excels in all three categories.
Figure 9 shows the classification result maps of the three methods, 3DCNN, ViT, and CWSN, on four datasets in a noise-free environment. In the KSC image in Figure 9a, the CWSN model demonstrates its strong classification ability, which is the closest to the ground truth. In contrast, the 3DCNN and ViT methods exhibit more misclassification in the boxed region of the image, resulting in a large gap between the classification results and the ground truth. The results of the dataset in Figure 9b (UP) show that the classification results of the 3DCNN and ViT methods have a large amount of salt-and-pepper noise, whereas the classification results of the CWSN are the smoothest, and are significantly better than those of the other methods, although a small amount of misclassification still exists. The classification results of the SV and WDC datasets also lead to similar conclusions. Especially in the circled part, the CWSN model performs significantly better than the other methods, and the classification results are the closest to the ground-truth images.
The reasons for this large disparity are as follows: Although ViT performs well in the image classification domain, it fails to effectively integrate spectral and spatial information when dealing with HSI. When the training data contains noisy labels or an insufficient number of labels, 3DCNN struggles to extract useful information. However, by utilizing the spectral and spatial attention mechanism as well as the Transformer structure, the CWSN model can fully capture the spectral and spatial features of HSI, improving the accuracy and efficiency of feature extraction and fusion. This multilevel feature extraction not only improves the final classification accuracy, but also enhances its robustness in complex scenes. In summary, the CWSN model also has good stability and generalization ability in the task of HSI classification under noise-free conditions.

5. Conclusions and Future Work

The accuracy of HSI classification under the supervised paradigm is affected by the feature extraction accuracy and sample confidence. To address the problems of deep semantic feature extraction and noisy sample detection, we propose the CWSN model, which is evaluated based on a variety of evaluation metrics in terms of model component design, classification performance, and application examples. The DBLIM adopts a lightweight dual-branch structure to enhance spatial and spectral features, with the residual connection extracting nonlinear features and mitigating the gradient vanishing in the deep network. Compared to Transformer, ViT has more advantages in image processing. So, the CViT cascade strategy combines the advantages of CNN in local feature induction and the excellent performance of Vision-Transformer in long-range correlation characterization, enhancing the robustness of the model and improving noise suppression. CWSN, which is based on a hybrid hierarchical structure, has lower computational complexity and fewer training samples required than 3DCNN, in addition to good robustness and generalizability. Extensive experimental results on five datasets demonstrate that the CWSN model outperforms other weakly supervised classification models.
When CWSN performs feature extraction and classification, the next challenge is to improve the adaptability of the ViT structure so that it can independently fully mine features of object class attributes from spectral and spatial information perspectives, thereby improving feature discrimination accuracy under small sample conditions. In addition, the Transformer structure that relies on local chunking operations has a high computational cost in computing query and key value information. By using localized objects instead of fixed-window sliding chunks, it helps to reduce computational complexity and improve the computational efficiency of the Transformer. In terms of data, spaceborne hyperspectral sensors are different from airborne sensors; currently, their spatial resolution is 30 m or more per pixel. The lower spatial resolution makes the model challenging to classify. Methods need to be explored and improved in the future, in order to achieve spaceborne hyperspectral classification application.

Author Contributions

Methodology, Y.J.; Data curation, J.Z.; Writing—original draft, W.F. and X.Z.; Project administration, W.F. and X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Guangxi Province, China, grant number 2024GXNSFBA010265. In addition, this research was also supported by the Guangxi Science and Technology Project, grant number Gui Ke AA22068001.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We express our deepest appreciation to all the members who generously contributed their time and efforts to this work, making it possible to produce meaningful results.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, W.D.; Li, Z.X.; Li, G.H.; Zhuang, P.X.; Hou, G.J.; Zhang, Q.; Li, C.Y. GACNet: Generate Adversarial-Driven Cross-Aware Network for Hyperspectral Wheat Variety Identification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 3347745. [Google Scholar] [CrossRef]
  2. Fang, L.Y.; Li, S.T.; Duan, W.H.; Ren, J.C.; Benediktsson, J.A. Classification of Hyperspectral Images by Exploiting Spectral-Spatial Information of Superpixel via Multiple Kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef]
  3. Meyer, J.M.; Kokaly, R.F.; Holley, E. Hyperspectral remote sensing of white mica: A review of imaging and point-based spectrometer studies for mineral resources, with spectrometer design considerations. Remote Sens. Environ. 2022, 275, 113000. [Google Scholar] [CrossRef]
  4. Siebels, K.; Goïta, K.; Germain, M. Estimation of Mineral Abundance from Hyperspectral Data Using a New Supervised Neighbor-Band Ratio Unmixing Approach. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6754–6766. [Google Scholar] [CrossRef]
  5. Cao, Q.; Yu, G.L.; Sun, S.J.; Dou, Y.; Li, H.; Qiao, Z.Y. Monitoring Water Quality of the Haihe River Based on Ground-Based Hyperspectral Remote Sensing. Water 2022, 14, 22. [Google Scholar] [CrossRef]
  6. Liu, C.; Xing, C.Z.; Hu, Q.H.; Wang, S.S.; Zhao, S.H.; Gao, M. Stereoscopic hyperspectral remote sensing of the atmospheric environment: Innovation and prospects. Earth-Sci. Rev. 2022, 226, 103958. [Google Scholar] [CrossRef]
  7. Pal, M. Deep learning algorithms for hyperspectral remote sensing classifications: An applied review. Int. J. Remote Sens. 2024, 45, 451–491. [Google Scholar] [CrossRef]
  8. Fang, C.H.; Zhang, G.F.; Li, J.; Li, X.P.; Chen, T.F.; Zhao, L. Intelligent agent for hyperspectral image classification with noisy labels: A deep reinforcement learning framework. Int. J. Remote Sens. 2024, 45, 2939–2964. [Google Scholar] [CrossRef]
  9. Pour, A.B.; Guha, A.; Crispini, L.; Chatterjee, S. Editorial for the Special Issue Entitled Hyperspectral Remote Sensing from Spaceborne and Low-Altitude Aerial/Drone-Based Platforms—Differences in Approaches, Data Processing Methods, and Applications. Remote Sens. 2023, 15, 5119. [Google Scholar] [CrossRef]
  10. Datta, D.; Mallick, P.K.; Bhoi, A.K.; Ijaz, M.F.; Shafi, J.; Choi, J. Hyperspectral Image Classification: Potentials, Challenges, and Future Directions. Comput. Intell. Neurosci. 2022, 2022, 3854635. [Google Scholar] [CrossRef]
  11. Kumar, B.; Dikshit, O.; Gupta, A.; Singh, M.K. Feature extraction for hyperspectral image classification: A review. Int. J. Remote Sens. 2020, 41, 6248–6287. [Google Scholar] [CrossRef]
  12. Sawant, S.S.; Manoharan, P.; Loganathan, A. Band selection strategies for hyperspectral image classification based on machine learning and artificial intelligent techniques–Survey. Arab. J. Geosci. 2021, 14, 646. [Google Scholar] [CrossRef]
  13. Gaetano, R.; Scarpa, G.; Poggi, G. Hierarchical Texture-Based Segmentation of Multiresolution Remote-Sensing Images. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2129–2141. [Google Scholar] [CrossRef]
  14. Yuan, J.Y.; Wang, D.L.; Li, R.X. Remote Sensing Image Segmentation by Combining Spectral and Texture Features. IEEE Trans. Geosci. Remote Sens. 2014, 52, 16–24. [Google Scholar] [CrossRef]
  15. Aptoula, E. Remote Sensing Image Retrieval with Global Morphological Texture Descriptors. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3023–3034. [Google Scholar] [CrossRef]
  16. Wu, W.; Yang, Q.P.; Lv, J.K.; Li, A.D.; Liu, H.B. Investigation of Remote Sensing Imageries for Identifying Soil Texture Classes Using Classification Methods. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1653–1663. [Google Scholar] [CrossRef]
  17. He, L.; Li, J.; Liu, C.Y.; Li, S.T. Recent Advances on Spectral-Spatial Hyperspectral Image Classification: An Overview and New Guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  18. Peng, Y.S.; Zhang, Y.W.; Tu, B.; Li, Q.M.; Li, W.J. Spatial-Spectral Transformer with Cross-Attention for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5537415. [Google Scholar] [CrossRef]
  19. Li, C.H.; Kuo, B.C.; Lin, C.T.; Huang, C.S. A Spatial-Contextual Support Vector Machine for Remotely Sensed Image Classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 784–799. [Google Scholar] [CrossRef]
  20. Maurya, K.; Mahajan, S.; Chaube, N. Decision tree (DT) and stacked vegetation indices based mangrove and non-mangrove discrimination using AVIRIS-NG hyperspectral data: A study at Marine National Park (MNP) Jamnagar, Gulf of Kutch. Wetl. Ecol. Manag. 2023, 31, 805–823. [Google Scholar] [CrossRef]
  21. Feng, J.F.; Wang, D.N.; Gu, Z.J. Bidirectional Flow Decision Tree for Reliable Remote Sensing Image Scene Classification. Remote Sens. 2022, 14, 3943. [Google Scholar] [CrossRef]
  22. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  23. Gu, Y.F.; Wang, C.; You, D.; Zhang, Y.H.; Wang, S.Z.; Zhang, Y. Representative Multiple Kernel Learning for Classification in Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2012, 50, 2852–2865. [Google Scholar] [CrossRef]
  24. Xue, Z.H.; Nie, X.Y.; Zhang, M.X. Incremental Dictionary Learning-Driven Tensor Low-Rank and Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5544019. [Google Scholar] [CrossRef]
  25. Tang, Y.Y.; Yuan, H.L.; Li, L.Q. Manifold-Based Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7606–7618. [Google Scholar] [CrossRef]
  26. Tan, Q.; Guo, B.; Hu, J.; Dong, X.; Hu, J. Object-oriented remote sensing image information extraction method based on multi-classifier combination and deep learning algorithm. Pattern Recognit. Lett. 2021, 141, 32–36. [Google Scholar] [CrossRef]
  27. Yang, J.M.; Kuo, B.C.; Yu, P.T.; Chuang, C.H. A Dynamic Subspace Method for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2840–2853. [Google Scholar] [CrossRef]
  28. Kumar, V.; Singh, R.; Rambabu, M.; Dua, Y. Deep learning for hyperspectral image classification: A survey. Comput. Sci. Rev. 2024, 53, 100658. [Google Scholar] [CrossRef]
  29. Li, S.T.; Song, W.W.; Fang, L.Y.; Chen, Y.S.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  30. Li, X.T.; Ding, H.H.; Yuan, H.B.; Zhang, W.W.; Pang, J.M.; Cheng, G.L.; Chen, K.; Liu, Z.W.; Loy, C.C. Transformer-Based Visual Segmentation: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10138–10163. [Google Scholar] [CrossRef] [PubMed]
  31. Zheng, Z.; Ermon, S.; Kim, D.; Zhang, L.P.; Zhong, Y.F. Changen2: Multi-Temporal Remote Sensing Generative Change Foundation Model. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 725–741. [Google Scholar] [CrossRef] [PubMed]
  32. Thoreau, R.; Achard, V.; Risser, L.; Berthelot, B.; Briottet, X. Active Learning for Hyperspectral Image Classification: A Comparative Review. IEEE Geosci. Remote Sens. Mag. 2022, 10, 256–278. [Google Scholar] [CrossRef]
  33. Chen, L.; He, J.F.; Shi, H.; Yang, J.Y.; Li, W. SWDiff: Stage-Wise Hyperspectral Diffusion Model for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 3485483. [Google Scholar] [CrossRef]
  34. Li, H.F.; Dou, X.; Tao, C.; Wu, Z.X.; Chen, J.; Peng, J.; Deng, M.; Zhao, L. RSI-CB: A Large-Scale Remote Sensing Image Classification Benchmark Using Crowdsourced Data. Sensors 2020, 20, 1594. [Google Scholar] [CrossRef]
  35. Trias-Sanz, R.; Stamon, G.; Louchet, J. Using colour, texture, and hierarchial segmentation for high-resolution remote sensing. ISPRS J. Photogramm. Remote Sens. 2008, 63, 156–168. [Google Scholar] [CrossRef]
  36. Xia, J.S.; Falco, N.; Benediktsson, J.A.; Du, P.J.; Chanussot, J. Hyperspectral Image Classification with Rotation Random Forest Via KPCA. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1601–1609. [Google Scholar] [CrossRef]
  37. Sun, L.; Zhao, G.R.; Zheng, Y.H.; Wu, Z.B. SpectralSpatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  38. He, X.; Chen, Y.S.; Lin, Z.H. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  39. Kang, X.D.; Duan, P.H.; Xiang, X.L.; Li, S.T.; Benediktsson, J.A. Detection and Correction of Mislabeled Training Samples for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5673–5686. [Google Scholar] [CrossRef]
  40. Li, Z.N.; Yang, X.R.; Meng, D.Y.; Cao, X.Y. An Adaptive Noisy Label-Correction Method Based on Selective Loss for Hyperspectral Image-Classification Problem. Remote Sens. 2024, 16, 2499. [Google Scholar] [CrossRef]
  41. Leng, Q.M.; Yang, H.O.; Jiang, J.J. Label Noise Cleansing with Sparse Graph for Hyperspectral Image Classification. Remote Sens. 2019, 11, 1116. [Google Scholar] [CrossRef]
  42. Tu, B.; Zhang, X.F.; Kang, X.D.; Zhang, G.Y.; Li, S.T. Density Peak-Based Noisy Label Detection for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1573–1584. [Google Scholar] [CrossRef]
  43. Tu, B.; Zhou, C.L.; He, D.B.; Huang, S.Y.; Plaza, A. Hyperspectral Classification with Noisy Label Detection via Superpixel-to-Pixel Weighting Distance. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4116–4131. [Google Scholar] [CrossRef]
  44. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral Image Classification Using a Hybrid 3D-2D Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar] [CrossRef]
  45. Wang, L.Q.; Zhu, T.C.; Kumar, N.; Li, Z.W.; Wu, C.L.; Zhang, P.Y. Attentive-Adaptive Network for Hyperspectral Images Classification with Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5505514. [Google Scholar] [CrossRef]
  46. Zhang, X.Y.; Yang, S.Y.; Feng, Z.X.; Song, L.L.; Wei, Y.T.; Jiao, L.C. Triple Contrastive Representation Learning for Hyperspectral Image Classification with Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2023, 61, 500116. [Google Scholar] [CrossRef]
  47. Yue, X.; Liu, A.; Chen, N.; Xia, S.; Yue, J.; Fang, L. Hypermll: Toward robust hyperspectral image classification with multi-source label learning. IEEE Trans. Geosci. Remote Sens. 2024, 62, 3441095. [Google Scholar] [CrossRef]
  48. Bazi, Y.; Bashmal, L.; Al Rahhal, M.M.; Al Dayil, R.; Al Ajlan, N. Vision Transformers for Remote Sensing Image Classification. Remote Sens. 2021, 13, 516. [Google Scholar] [CrossRef]
  49. Li, Y.; Zhang, H.K.; Shen, Q. Spectral-Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
Figure 1. Framework of the CWSN model.
Figure 1. Framework of the CWSN model.
Entropy 27 00869 g001
Figure 2. Spectral feature extraction.
Figure 2. Spectral feature extraction.
Entropy 27 00869 g002
Figure 3. Spatial feature extraction.
Figure 3. Spatial feature extraction.
Entropy 27 00869 g003
Figure 4. Effect of parameters K and h on the classification performance of CWSN with different numbers of mislabeled samples.
Figure 4. Effect of parameters K and h on the classification performance of CWSN with different numbers of mislabeled samples.
Entropy 27 00869 g004
Figure 5. Classification maps of each component of the CWSN on the KSC dataset (first and second rows with noisy sample sizes of 6 and 13, respectively).
Figure 5. Classification maps of each component of the CWSN on the KSC dataset (first and second rows with noisy sample sizes of 6 and 13, respectively).
Entropy 27 00869 g005
Figure 6. Visualization of feature extraction results for each component of CWSN on the KSC dataset.
Figure 6. Visualization of feature extraction results for each component of CWSN on the KSC dataset.
Entropy 27 00869 g006aEntropy 27 00869 g006b
Figure 7. Classification results of CWSN on UP using different loss functions (54 T + 14 M in the first row and 54 T + 27 M in the second row).
Figure 7. Classification results of CWSN on UP using different loss functions (54 T + 14 M in the first row and 54 T + 27 M in the second row).
Entropy 27 00869 g007
Figure 8. Classification accuracy of each method under different noisy sample conditions. (ac) 26 T; (df) 54 T; (gi) 32 T; (jl) 28 T.
Figure 8. Classification accuracy of each method under different noisy sample conditions. (ac) 26 T; (df) 54 T; (gi) 32 T; (jl) 28 T.
Entropy 27 00869 g008aEntropy 27 00869 g008b
Figure 9. Classification maps for each method on the four datasets when using a clean sample set (from left to right in each dataset, the ground-truth, 3DCNN, ViT, and CWSN classification maps).
Figure 9. Classification maps for each method on the four datasets when using a clean sample set (from left to right in each dataset, the ground-truth, 3DCNN, ViT, and CWSN classification maps).
Entropy 27 00869 g009
Table 1. Settings for the number of training samples on the KSC, UP, SV, and WDC datasets.
Table 1. Settings for the number of training samples on the KSC, UP, SV, and WDC datasets.
DatasetKSCUPSVWDC
Number of classes139166
Sample settingSetting 1Setting 2Setting 1Setting 2Setting 1Setting 2Setting 1Setting 2
26 + 626 + 1354 + 1454 + 2732 + 832 + 1628 + 728 + 14
Total number of samples338 + 78338 + 169486 + 126486 + 243512 + 128512 + 256168 + 42168 + 84
In sample setting, T + M represents the total number of samples for each class; T and M represent the number of correct and mislabeled samples for each class, respectively.
Table 2. Classification accuracy (%) of each component on the KSC dataset (Bold numbers indicate the highest accuracy).
Table 2. Classification accuracy (%) of each component on the KSC dataset (Bold numbers indicate the highest accuracy).
Class26 T + 6 M26 T + 13 M
SBLSpeASpaATFCWSNSBLSpeASpaATFCWSN
C198.2695.3490.8291.7394.5487.8989.8991.6990.5894.60
C297.6287.1895.9688.9996.3892.5489.4482.5786.4893.03
C393.5783.6195.7699.1292.2396.4688.6295.2395.5495.54
C482.2274.2754.9369.7769.2276.7362.7964.7164.0186.22
C591.1483.3088.5998.9893.0075.2877.0382.7380.2376.61
C678.4378.6384.2080.0895.4663.8188.5571.4382.5784.50
C797.1095.7189.32100.0095.5396.2891.35100.00100.0084.12
C896.4093.3494.7593.5397.6691.5082.4873.7193.7984.48
C999.1895.4097.04100.0097.28100.00100.0092.6795.96100.00
C1097.7497.5697.4699.9193.9899.3794.8595.3495.0895.63
C11100.0096.4497.1499.8397.2399.6595.69100.0093.5296.10
C1296.8185.5081.8491.5895.4174.9389.6692.0992.4391.59
C1364.1496.1483.40100.00100.0098.5393.9793.28100.00100.00
OA87.7691.7389.9992.6094.9387.9190.1589.4792.1992.73
AA89.5789.4290.5591.3593.6985.6988.0287.3490.0190.96
Kappa84.6890.9887.9891.9794.5686.8489.3489.5691.6192.22
Table 3. Classification accuracy of each loss algorithm on UP.
Table 3. Classification accuracy of each loss algorithm on UP.
Class54 T + 14 M54 T + 27 M
CERCENCENSLCERCENCENSL
C188.0384.1891.6486.4979.0990.0990.6089.75
C269.6976.6979.7986.2587.6885.5883.1575.66
C359.2785.8598.4288.9870.8190.2887.0791.61
C496.7797.0593.3394.3889.7695.7690.3194.62
C599.9799.9799.7999.7995.6599.97100.00100.00
C684.4874.2573.5870.5663.0362.4074.3883.54
C793.6097.2096.9196.9686.1694.2697.7495.42
C893.3993.2347.6073.4880.1490.8988.0184.31
C997.8199.8999.5187.2196.3399.3899.9299.96
OA79.9682.6881.5484.8182.5182.9482.6783.46
AA87.0089.8186.7387.1283.1889.8590.1390.54
Kappa74.9177.8376.3480.0777.1378.9681.3681.53
Table 4. Running time of DPNLD, SPWD, 3DCNN, ViT, and CWSN in the noise training set (s).
Table 4. Running time of DPNLD, SPWD, 3DCNN, ViT, and CWSN in the noise training set (s).
MethodKSCUPSVWDC
26 T+ 6 M26 T+ 13 M54 T+ 14 M54 T+ 27 M32 T + 8 M32 T+ 16 M28 T+ 7 M28 T + 14 M
DPNLD14.1919.3029.0925.7331.0575.362.269.49
SPWD19.1126.1633.4821.6043.7284.803.334.51
3DCNN856.82796.044014.013663.196429.896152.87533.75520.64
ViT1825.771952.515439.614814.678369.648979.921105.731087.98
CWSN3472.073040.378155.478943.9911,816.9411,597.851865.951982.46
Table 5. Classification accuracies (%) obtained by 3DCNN, ViT, and CWSN models on four datasets using the CE loss function.
Table 5. Classification accuracies (%) obtained by 3DCNN, ViT, and CWSN models on four datasets using the CE loss function.
DatasetKSCUP
3DCNNViTCWSN3DCNNViTCWSN
OA91.7991.6695.4386.0790.5791.75
AA89.3089.1493.3788.6290.9192.48
Kappa90.8490.6994.9081.9287.5390.10
DatasetSVWDC
3DCNNViTCWSN3DCNNViTCWSN
OA88.1584.4686.7285.4788.9689.60
AA93.9892.2793.2286.9391.3892.01
Kappa86.8182.7385.2482.0886.4187.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, W.; Sun, X.; Zhang, X.; Ji, Y.; Zhang, J. CViT Weakly Supervised Network Fusing Dual-Branch Local-Global Features for Hyperspectral Image Classification. Entropy 2025, 27, 869. https://doi.org/10.3390/e27080869

AMA Style

Fu W, Sun X, Zhang X, Ji Y, Zhang J. CViT Weakly Supervised Network Fusing Dual-Branch Local-Global Features for Hyperspectral Image Classification. Entropy. 2025; 27(8):869. https://doi.org/10.3390/e27080869

Chicago/Turabian Style

Fu, Wentao, Xiyan Sun, Xiuhua Zhang, Yuanfa Ji, and Jiayuan Zhang. 2025. "CViT Weakly Supervised Network Fusing Dual-Branch Local-Global Features for Hyperspectral Image Classification" Entropy 27, no. 8: 869. https://doi.org/10.3390/e27080869

APA Style

Fu, W., Sun, X., Zhang, X., Ji, Y., & Zhang, J. (2025). CViT Weakly Supervised Network Fusing Dual-Branch Local-Global Features for Hyperspectral Image Classification. Entropy, 27(8), 869. https://doi.org/10.3390/e27080869

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop