Next Article in Journal
A 1.8 m Class Pathfinder Raman LIDAR for the Northern Site of the Cherenkov Telescope Array Observatory—Performance
Previous Article in Journal
Risk Assessment of the 2022 Nigerian Flood Event Using Remote Sensing Products and Climate Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial-Spectral Linear Extrapolation for Cross-Scene Hyperspectral Image Classification

School of Electronic Information Engineering, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(11), 1816; https://doi.org/10.3390/rs17111816
Submission received: 17 February 2025 / Revised: 12 May 2025 / Accepted: 19 May 2025 / Published: 22 May 2025
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

In realistic hyperspectral image (HSI) cross-scene classification tasks, it is ideal to obtain target domain samples during the training phase. Therefore, a model needs to be trained on one or more source domains (SD) and achieve robust domain generalization (DG) performance on an unknown target domain (TD). Popular DG strategies constrain the model’s predictive behavior in synthetic space through deep, nonlinear source expansion, and an HSI generation model is usually adopted to enrich the diversity of training samples. However, recent studies have shown that the activation functions of neurons in a network exhibit asymmetry for different categories, which results in the learning of task-irrelevant features while attempting to learn task-related features (called “feature contamination”). For example, even if some intrinsic features of HSIs (lighting conditions, atmospheric environment, etc.) are irrelevant to the label, the neural network still tends to learn them, resulting in features that make the classification related to these spurious components. To alleviate this problem, this study replaces the common nonlinear generative network with a specific linear projection transformation, to reduce the number of neurons activated nonlinearly during training and alleviate the learning of contaminated features. Specifically, this study proposes a dimensionally decoupled spatial spectral linear extrapolation (SSLE) strategy to achieve sample augmentation. Inspired by the weakening effect of water vapor absorption and Rayleigh scattering on band reflectivity, we simulate a common spectral drift based on Markov random fields to achieve linear spectral augmentation. Further considering the common co-occurrence phenomenon of patch images in space, we design spatial weights combined with label determinism of the center pixel to construct linear spatial enhancement. Finally, to ensure the cognitive unity of the high-level features of the discriminator in the sample space, we use inter-class contrastive learning to align the back-end feature representation. Extensive experiments were conducted on four datasets, an ablation study showed the effectiveness of the proposed modules, and a comparative analysis with advanced DG algorithms showed the superiority of our model in the face of various spectral and category shifts. In particular, on the Houston18/Shanghai datasets, its overall accuracy was 0.51%/0.83% higher than the best results of the other methods, and its Kappa coefficient was 0.78%/2.07% higher, respectively.

1. Introduction

Hundreds of thousands of continuous, narrow spectral bands from ultraviolet to infrared can be acquired simultaneously using hyperspectral imaging technology, creating a continuous spectral curve. Hyperspectral remote sensing image classification technology has become more popular and widely used as a result of its improved information detection capabilities. When it comes to classifying hyperspectral image data that are independently and identically distributed, convolutional neural networks (CNNs) have demonstrated exceptional performance [1,2,3,4,5,6]. Considering that CNNs are not good at modeling long-range dependencies, transformer-based classification algorithms further improve classification performance by fully exploiting the sequential properties of spectral features. Reference [7] first proposed a new backbone network called SpectralFormer, which can learn spectral local sequence information from adjacent bands of HS images to achieve grouped spectral embedding. Reference [8] designed a new type of DenseTransformer, which uses dense connections to strengthen feature propagation, and constructed two classification frameworks based on a DenseTransformer to improve HSI classification performance under limited training samples. Reference [9] proposed a network called HyperSFormer, which combines transformer and semantic segmentation, and enhances the SegFormer architecture by replacing its encoder with an improved Swin Transformer, while retaining the SegFormer decoder, solving the limitations of traditional CNN and RNN frameworks in expressing global information entropy due to insufficient contextual information.
Unseen regions that require evaluation are common in online applications with realistic settings. The atmospheric environment, imaging equipment, and lighting conditions are some of the factors that can cause data shifts. Applying CNN/transformer-based algorithms directly to out-of-distribution (OOD) data often results in high generalization errors, which makes transfer learning difficult [10,11].
The domain adaptation approach was initially developed to facilitate effective migration of OOD samples [12,13,14]. The various methods for cross-scene hyperspectral image classification using domain adaptation include probability distribution alignment (as discussed in [15,16]), feature selection (as explored in [17,18]), subspace learning (as presented in [19]), and deep DA (as proposed in [20]). Probability distribution alignment reduces domain shift by aligning the data distribution between the source domain and target domain using Maximum Mean Discrepancy (MMD) as a common metric for quantifying the disparity between the distributions. Feature selection helps to extract cross-domain invariant knowledge by detecting features shared between the source and target domains. Subspace learning assumes that the distributions of the source and target domains are similar in the converted subspace. Domain adaptation is achieved by minimizing distribution differences between the source and target domains within a given subspace. Deep domain adaptation uses deep neural networks to automatically extract representations from both the source and destination domains. This technique achieves consistent knowledge learning across multiple domains by matching the SD and TD data distributions.
DA algorithms have demonstrated exceptional accuracy on a variety of datasets, but they rely on an ideal condition in which data from both the SDs and TD are available for training. Because the TD is not always accessible in reality, recent computer vision (CV) tasks involve a more difficult, but more generic, setup termed domain generalization [21]. The DG method can perform admirably in transferring information to unseen TDs by training exclusively with SDs. The purpose of DG is to use one or more separate but related SDs to construct a reliable model [22]. There are three types of DG algorithms now in use: learning approaches, representation learning, and data manipulation. Data production [23,24] and data augmentation [25,26] are two techniques for data manipulation that improve training data by altering and enriching samples. Domain-invariant feature learning aims to decouple features between domains. Examples of domain-invariant feature learning include kernel methods [27], explicit feature alignment [28,29], domain adversarial training [30,31], and invariant risk minimization (IRM) [32]. In these cases, learning domain-invariant representations enables models to accurately fit various domains. A learning strategy involves integrating mature learning models into multi-domain training, such as ensemble learning [33] and meta-learning [34]. Furthermore, DG can be classified as single-DG (SDG) or multi-DG (MDG) depending on the number of domains in the SD. SDG indicates that the SD has just one data source, which is more challenging than MDG and usually necessary, owing to limitations in accessible data. MDG’s core idea is to generalize information obtained from several SDs to unknown TDs.
Single-domain generalization (SDG) is frequently required, due to data restrictions. Using various approaches to broaden the range of domains and extract invariant features is a popular strategy for addressing SDG. To imitate the variety of samples in the TD, the study in [35] proposed a progressive domain expansion network with accurate geometric and photometric adjustments. The study in [36] introduced a style-complement module that uses mutual information restrictions to generate images that closely resemble the source domain. The paper by [37] proposed the use of regularized training to adjust non-causal elements, without explicitly recovering the causal feature, while keeping the causal basics intact.
To address the SDG problem in cross-scenario HSIC, common approaches include broadening the sample or feature space, while minimizing multiple empirical risks. For instance, SDEnet [10] introduced a semantic encoder that incorporates spatial, spectral, and morphological information. It uses supervised contrastive learning to align the source and extended domains. SSECNet [38] improved causal-invariant feature reliability by introducing a contribution alignment module based on spatial and spectral domain enhancement branches. FDGNet [39] employed frequency disentanglement using a spectral-spatial encoder to maintain semantic consistency while simulating domain gaps, reducing the impact of randomization on sample semantics. These algorithms resample HSIs through different nonlinear structures to achieve reliable distribution expansion. This type of nonlinear mapping reprojects (encodes and decodes) the source domain samples into extended samples of the same dimension, which can be regarded as data augmentation (for HSI data, high-dimensional learning almost always corresponds to extrapolation, that is, interpolating samples beyond the convex hull boundaries of the source dataset [40]). However, since the neurons for nonlinear representation often show asymmetric activation for different categories, there is a tendency to learn task-irrelevant features when learning task-related features (called “feature contamination”), and this tendency makes it difficult for neural networks to generalize to scenarios with distribution shift [41]. For instance, when collecting HSI under arid climate conditions, the spectral features often show a statistical association between low soil moisture and certain classes. This spurious pattern may be learned by the deep neural network, which may cause the model to have prediction bias when the HSI comes from rainy season areas. Such background components (including lighting conditions, soil conditions, atmospheric environment, etc.) are irrelevant to the output label, and neurons tend to produce non-zero gradient projections for them, which leads to feature contamination.
To address this issue, this study attempts to achieve HSI source distribution expansion through linear sample augmentation. First of all, it is worth noting that existing related work on CV, such as example interpolation techniques, can be directly used for HSIC DG tasks. These algorithms obtain a wider data space by interpolating between the input and labels of multiple real samples. Since they do not involve neuron operations, they can also avoid introducing feature contamination in a generation. Representative algorithms include Mixup [42], MixStyle [43], StyleHall [44], and other mixed sample data augmentation (MSDA) strategies. However, this study conducted experiments on multiple cross-scenario HSI datasets and found that the improvement in generalization performance of applying such algorithms under the basic CNN backbone classifier was quite limited and even had a negative impact on specific datasets. Considering the difference in generalization gap between HSI and RGB images, where HSI is mainly affected by spectral drift, while RGB images focus on the changes in pixel spatial distribution and are insensitive to the category information characteristics of the spectral dimension, this can explain the failure of MSDA to migrate to the HSIC task in the CV field to a certain extent. Secondly, inspired by this, this study implements HSI linear extrapolation from the perspective of dimensional decoupling and implements MSDA by performing non-interconnected linear transformations on the one-dimensional spectrum and the two-dimensional patch space, respectively, so as to constrain the prediction behavior of the model in the synthetic space, thereby improving the generalization performance. The following is a summary of the main contributions of the proposed method.
1.
This study proposes a HSI linear data augmentation framework, which is designed with a complementary spatial-spectral linear transformation, to reduce the risk of feature contamination from nonlinear neurons and promote the generalization of neural networks to target domains with distribution shift.
2.
Considering the information distribution characteristics of HSIs, and inspired by the weakening effect of water vapor absorption and Rayleigh scattering on band reflectivity, a spectral linear extrapolation based on a MRF is designed to simulate spectral drift. Further considering the co-occurrence phenomenon [45] of patch images in space, spatial weights are designed in combination with the label certainty of the central pixel to construct a linear spatial extrapolation. Benefiting from the above spatial-spectral linear transformation, the sample diversity outside the convex hull boundary of the source domain is enriched.
3.
To prevent a CNN-based discriminator from learning redundant or meaningless features for linear synthetic domains, this study introduces feature-level supervised contrastive learning to unify the invariant representation by constraining the aggregate distribution of low-dimensional features between domains.
The remainder of this study is structured as follows: Section 2 introduces feature contamination and mixed sample data augmentation. Section 3 describes the SSLE in in greater detail. Section 4 describes the comprehensive experiments and analyses. Section 5 concludes with the conclusions. Supplementally, for the convenience of reading, the important abbreviations are summarized in Table 1.

2. Related Work

2.1. Feature Contamination

From the perspective of optimization goals, feature learning of neural networks is a “byproduct” of task-driven learning, and its purpose is to minimize training errors. Intuitively, neural networks should extract “task-related” features from data, while the remaining “task-irrelevant” features are equivalent to data noise. Therefore, since neural networks have the characteristic of “not learning unless necessary” (more precisely, simplicity bias [46]), neural networks should tend not to learn irrelevant features. This is also a common view in the current literature [47].
Specifically, it is assumed that the input of the network is a linear combination of two features: “core features” and “background features”. Among them, the distribution of core features depends on the category label, while the distribution of background features is independent of the label. In order to eliminate the interference of other factors, the following assumptions are made for these two types of features: a. Background features are unrelated to labels (excluding the failure mode caused by spurious correlations). b. The core features can predict the labels with 100% accuracy (excluding the failure mode caused by insufficient features in the training set). c. The core features and background features are distributed in orthogonal subspaces (excluding the failure mode caused by difficulty in decoupling different features).
Based on the above simple learning scenario, the study in [41] found that a neural network (two-layer ReLU network binary classification problem) will still learn background features that are completely unrelated to the task while learning the core features. Due to the coupling of these two features in the network weight space, a distribution shift of the background features will also increase the error of the neural network, thereby reducing the generalization of the network. This feature learning preference of a neural network is called “feature contamination”.
Through further research, feature contamination was found to actually be related to the asymmetric activation of neurons in the neural network for different categories [41]. Specifically, it can be proved that after enough SGD iterations, at least a considerable number of neurons in the network will tend to maintain a positive correlation with samples of one category (called the positive sample of the neuron), and a negative correlation with samples of another category (called the negative sample of the neuron), which will cause the activation of these neurons to have category asymmetry.
In order to further illustrate the relationship between feature contamination and the nonlinear activation function of neural networks, the study in [41] proved that after removing the nonlinearity of the neural network, feature contamination will no longer occur. Therefore, to a certain extent, it can be considered that reducing the nonlinear activation of the network during the learning process can alleviate the learning bias of contaminated features.
It is worth noting that the current single-source domain generalization network usually introduces a variety of generator networks (including 3D-CNN-based, ViT-based, and ResNet-based, etc.) in the training stage. Although this approach enriches the diversity of samples through the network’s nonlinear re-sampling of the source domain, it also increases the risk of feature contamination. Therefore, in this study, we replace the common nonlinear generator network with a specific linear projection transformation to reduce the number of neurons with asymmetric activation for different categories during training (these neurons only exist in the feature extraction module of the discriminator). The proposed linear transformation achieves spatial-spectral feature decoupling through mutually independent spatial-spectral linear expansion.
Furthermore, although the work in [41] pointed out the problem of feature contamination, it did not provide a clear solution. However, some of their subsequent works showed that similar problems also occur when fine-tuning large models, and it was also found that some gradient adjustment-based methods can indeed alleviate this problem, thereby significantly improving the generalization ability of the fine-tuned model.

2.2. Mixed Sample Data Augmentation

Mixed Sample Data Augmentation (MSDA) is a data augmentation method based on interpolation. It obtains a wider data space by interpolating between the input and labels of multiple real samples, and can constrain the prediction behavior of a model in synthetic space. Representative methods include Mixup [42], MixText  [48], and MixStyle [43]. Mixup is a weighted sum of two different real samples and categories to obtain new data. By mixing the new data with the original data in a certain proportion for training, the generalization performance of the model can be improved. MixText is a semi-supervised method for text classification that combines labeled data and unlabeled data. It uses certain techniques to label a large amount of unlabeled text and then generates new data with real labeled data and the above labeled unsupervised data. MixStyle is based on instance-level feature statistics of cross-source domain training samples via probabilistic mixing. By mixing the styles of the bottom layer of the CNN, new domains can be implicitly synthesized to increase the domain diversity of the source.

3. Proposed Domain Generalization Method

As shown in Figure 1, SSLE includes spatial-spectral linear expansion as well as inter-domain feature alignment. Let X = X s o i = 1 N represent the data obtained from the SD. The data consist of a sample of 13 × 13 × N b a n d spatial patches in the HSI, denoted as X s o . Additionally, let Y = y i i = 1 N represent the associated class labels. Here, N b a n d represents the number of spectral channels, and N is the number of source samples.

3.1. Spectral Linear Transformation

In the Earth’s atmosphere, atmospheric gases and aerosols scatter and absorb radiation. When radiation interacts with atmospheric particles, the charge in the particles begins to oscillate. This causes the charge to radiate in all directions [49], resulting in distortions in remote measurements of radiation reflected from surfaces. Another process that distorts sensor radiation is absorption, which is a function of wavelength. Unlike scattering, absorption represents the conversion of radiation into another form of energy. The spectral range over which radiation is absorbed by atmospheric components is called the absorption signature [50].
For HSIs, water vapor or Rayleigh scattering can weaken or flatten the surface reflectance under the influence of adjacent pixels [51], resulting in a reduction in the discriminant information of subsequent classification tasks. Atmospheric correction is usually used to process weakened or unstable bands, but it is theoretically impossible to perform complete atmospheric correction. Even after atmospheric correction, the influence of water vapor or Rayleigh scattering may still exist. Inspired by this, this study uses the weakening (flattening) of the spectrum from this effect to imitate the inter-domain gap, to expand the diversity of sources. In particular, this type of spectral transformation uses a Markov random field (MRF) [52] model to achieve linearization, which ensures the overall continuity and intrinsic trend of the spectrum, while making the results controllable.
To obtain generated spectral vector, R s p e = [ r 1 s p e , r 2 s p e , r 3 s p e , , r N b a n d s p e ] by R s o . R s o is the normalized reflectance of a single pixel ( R s o = [ r 1 s o , r 2 s o , r 3 s o , , r N b a n d s o ] ) from X s o in Figure 1 and Figure 2. The objective of MRF is to reduce the following energy function E for calculating R s p e :
E ( R ) = λ s Ω ϕ s r s + ( 1 λ ) ( s , t ) N ψ s t r s , r t
where the variable R = [ r 1 , r 2 , r 3 , , r N b a n d ] , ϕ s r s is called the data term, ψ s t r s , r t is called the smoothness term, and Ω is the set of spectral vectors r n . N represents the set of neighboring wave bands, and λ is a balance parameter. The settings of ϕ s r s and ψ s t r s , r t are shown in Equations (2) and (3). In Equation (2), this function is employed to compute the L 2 distance between the r n and the r n s o . Equation (3) is used to calculate the local smoothness between r n 1 , r n and r n + 1 .
ϕ s r s = 1 2 ( r n r n s o ) 2
ψ s t r s , r t = 1 2 ( r n 1 2 r n + r n + 1 ) 2
The gradient descent algorithm in Algorithm 1 is adopted to calculate R s p e from Equation (4). In this process, the initial value is designated as R s o , and the step size η is set as 0.1. The number of iterations K is set to 31. The augmented patch cube X s p e (13 × 13 × N b a n d ) is obtained by applying Equation (5) to all pixels from X s o (where u and v represent the horizontal and vertical coordinates in the image patch, respectively).
min E ( R ) = 1 2 n = 1 N b a n d λ ( r n r n s o ) 2 + ( 1 λ ) ( r n 1 2 r n + r n + 1 ) 2
X s p e ( u , v ) = R s p e u , v , u [ 1 , 13 ] , v [ 1 , 13 ]
Algorithm 1 Calculating R s p e
Input: 
Constant: R s o = [ r 1 s o , r 2 s o , r 3 s o , , r N b a n d s o ] ,
Output: 
R s p e = [ r 1 s p e , r 2 s p e , r 3 s p e , , r N b a n d s p e ]
  1:
Definition 1: Variable- R = [ r 1 , r 2 , r 3 , , r N b a n d ] , Step size- η , Iteration number-K
  2:
Definition 2: Energy function- E ( R ) = 1 2 n = 1 N b a n d λ ( r n r n s o ) 2 + ( 1 λ ) ( r n 1 2 r n + r n + 1 ) 2 , derivative function- E ( R ) = E ( R ) R
  3:
Initialization: R 0 = R s o
  4:
for each k [ 1 , K ]  do
  5:
     R k : = R k 1 η E ( R k 1 )
  6:
end for
  7:
return   R s p e = R K

3.2. Spatial Linear Transformation

For patch-wise classification algorithms, common co-occurrence phenomena in spatial distribution can mislead model learning. For example, in some HSI patches, the pixels tend to appear in multiple categories at the same time, such as weeds and soil around trees, and roads around buildings. This makes it easy for the model to learn co-occurrence features that are irrelevant to the target itself (central pixel) during training, rather than focusing on the true features of the target. In addition, when a single-category image patch appears, the model may be overly dependent on co-occurrence associations, resulting in inaccurate predictions, such as tending to predict multiple targets rather than a single target.
In order to alleviate the above problems, this study considers adding a special type of sample data to the training domain to deliberately distinguish the central pixel from other pixels in Figure 3. A discrete Gaussian operator is adopted using Equation (6), where u = ( i k ) and v = ( j l ) , and ( i , j ) and ( k , l ) represent the coordinates of the central pixel and the neighborhood pixel of the image patch, respectively.
Maximum and minimum normalization are applied using Equation (7). The augmented patch cube X s p a (13 × 13 × N b a n d ) is obtained by applying Equation (8) to all spectral channels from X s o . X s p a can be regarded as a spatial weighted resampling. β is a balancing parameter, which is set so as not to interfere with the model’s learning of the original spatial-spectral joint distribution.
w ( u , v ) = 1 2 π σ 2 e u 2 + v 2 / 2 σ 2
w n o r m ( u , v ) = w ( u , v ) m i n ( w ) m a x ( w ) m i n ( w )
X s p a ( u , v , n ) = β w n o r m ( u , v ) X s o ( u , v , n ) + ( 1 β ) X s o ( u , v , n )

3.3. Training Phase

Classifier: As illustrated in Figure 4, the fundamental 2D-CNN classifier consists of the following components for feature extraction: convolution, ReLU activation, and maximal pooling (CRM). Fully connected layers are utilized for learning class boundaries.
Regularization: Two generated domains ( X s p a , X s p e ) and source domains X s o were acquired for training. This study employed supervised contrastive loss [53] in Equation (9) to improve the network’s ability to perceive category centers across various domains. With a number of | P ( i ) | , the samples ( z i ) of the same class in X s p a , X s p e and X s o are considered positive ( z p + ), while the samples from the other classes are considered negative ( z n ). The positive and negative sets are denoted by P ( i ) and N ( i ) .
L c o n = i = 0 N 1 | P ( i ) | p P ( i ) log exp S z i , z p + / τ a N ( i ) exp S z i , z n / τ
Loss Function: Regularization terms L c o n are included in the total loss function Equation (12) to address multi-domain feature alignment with cross-entropy in Equations (10) and (11). The X s o , X s p e and X s p a are treated as three training domains with equal importance, to avoid the discriminator learning asymmetric activations between domains in Equations (9) and (11).
L CE p i , y i = c y i c log p i c
CE = 1 N i L CE p X s o i , y i + L CE p X s p e i , y i + L CE p X s p a i , y i
min θ C R M L = CE + α L c o n

3.4. Generalization Performance

Figure 5 shows the visualization results of the features extracted by CRM on X s o , X s p a and X s p e following t-SNE [54] using PaviaU as the SD and PaviaC as the TD. As illustrated in Figure 5a,d,g, the better aggregation and separability of the feature projection of X s o , X s p a , and X s p e indicate that the classifier was capable of learning a domain-invariant feature mapping. The cluster centers of the same category in Figure 5b,e,h are close together and do not overlap, showing that the X s p a and X s p e could successfully preserve domain-specific information. Although the model was transferred directly to the TD, which is substantially distinct from the SD, the classifier still showed a reliable generalization performance, as shown in Figure 5j,k.

4. Experimental Results and Discussion

To illustrate the effectiveness of our DG method, we conducted experiments with four cross-scene hyperspectral imaging datasets: Houston 2013 → Houston 2018; PaviaU → PaviaC; Hangzhou → Shanghai; Dioni → Loukia (HyRANK). The Houston/Pavia/HyRANK datasets were downloaded from https://github.com/YuxiangZhang-BIT/Data-CSHSI (accessed on 5 April 2024). The Shanghai-Hangzhou Dataset was generously provided by Dr. Qin [39].
Random flips and random radiation noise (illumination) in the Houston dataset were used to strengthen the SDs four times, whereas the others remained untreated. A total of 80% of the labeled data from SD = Houston13/PaviaU was allocated for training, while the remaining 20% was reserved for validation. Meanwhile, 5% of the Hangzhou data were used for training, and the remaining data were used for validation. All samples in the TD were used for testing. The base learning rate, regularization parameters α , and coefficient λ and σ were selected from the sets: { 1 × 10 5 , 1 × 10 4 , 1 × 10 3 , 1 × 10 2 , 1 × 10 1 } , { 1 × 10 2 , 1 × 10 1 , 1 × 10 + 0 , 1 × 10 + 1 , 1 × 10 + 2 } , { 0.1 , 0.3 , 0.5 , 0.7 , 0.9 } , and { 1 , 3 , 5 , 7 , 9 } . Cross-validation was used to determine the best parameters. For the three datasets, the selected base learning rate was 1 × 10 3 . The parameters α / λ / σ were set to 1/0.5/7 for all datasets.
Three metrics were calculated to assess the categorization performance: the Kappa coefficient (Kappa), the average accuracy (AA), and the overall accuracy (OA). The mean ± standard deviation was used to display the test results. Five repeats of training and testing were used to carry out the evaluation. The objective was to assess the uncertainty related to a number of variables, such as the random initialization of the network and the random division of datasets.
The input was specified using a 13 × 13 patch size, and 64 is selected as the batch size. For both the discriminator and generator, Adaptive Moment Estimation (Adam) [55] was the optimization method. The default value for the weight decay for 2 -norm regularization was 1 × 10 4 for all modules. The epoch with the highest OA on the validation set was utilized for early stopping after the loss function curve became stable (200 training epochs for the Houston dataset, 50 training epochs for the Pavia/HyRANK dataset, and 10 training epochs for the S-H dataset). A 12GB Nvidia GTX 3060 GPU (Manufacturer: Colorful, Shanghai, China) and Pytorch V2.0.0 were the fundamental hardware and software platforms.

4.1. Experimental Data

4.1.1. Houston Dataset

The Houston2013 [56] data were acquired using the ITRES CASI-1500 sensor (sizes: 349*1905, spatial resolution: 2.5 m, spectral region: 364 nm to 1046 nm, 144 bands, Fifteen categories). The data obtained from the 2018 IEEE GRSS Data Fusion competition were published as Houston2018 [57] (spatial resolution: 1.0 m, size: 2384*601). We extracted 48 spectral bands (wavelength range: 0.38–1.05 μ m) from the Houston2013 dataset by selecting the overlapped area of 209 × 955. The samples and their respective classifications are presented in Table 2. Figure 6 illustrates the false-color and ground truth maps.

4.1.2. Pavia Dataset

The datasets were obtained from Pavia University and Pavia Center using a ROSIS sensor with a spectral range of 430–860 nm and 115 bands, and 103 and 102 bands remained after preprocessing the data from Pavia University (610*340, spatial resolution: 1.3 m) and Pavia Center (1096*715, spatial resolution: 1.3 m), respectively. The last Pavia University band was eliminated. Table 2 shows the seven categories that are shared by both datasets. Figure 7 illustrates their false-color images and ground-truth maps.

4.1.3. Shanghai-Hangzhou Dataset

The EO-1 Hyperion hyperspectral sensor was utilized to acquire the Shanghai-Hangzhou datasets, which comprise 198 bands in total. In comparison to the Hangzhou dataset, which has a dimension of 590 × 230, the Shanghai dataset measures 1600 × 230. A distance of 162 km separates the two cities, Shanghai and Hangzhou. Three common land cover classes for cross-scene categorization are detailed in Table 2; Figure 8 depicts their pseudo-color representations and ground-truth maps.

4.1.4. Hyrank Dataset

The HyRANK dataset [59] includes two regions: Dioni and Loukia. The dimensions of Dioni are 250 × 1376, the dimensions of Loukia are 249 × 945, and the spatial resolution is 30 m. The Hyperion sensor acquired satellite hyperspectral images including 176 spectral bands and 14 uniform classifications. This research employed 176 spectral bands and 12 unique item categories. The land cover classifications utilized for cross-scene categorization are detailed in Table 3. Figure 9 illustrates their pseudo-color representations with the corresponding ground-truth maps.

4.2. Ablation Study

The spatial-spectral linear transformation for MSDA is the key module of the proposed method, and the main approach for optimizing is supervised contrastive alignment. An ablation analysis was performed to analyze the contribution of the core elements by adding/deleting each component from the overall framework in Table 4. A check mark indicates that the module was adopted, and an uncheck mark indicates that the module was removed.
This study analyzed four variants: (1) “CRM Backbone”: The model of the testing stage in Figure 1 was trained on SD and directly applied to TD, (2) “ X s p e ”: Linear spectrum expansion module, (3) “ X s p a ”: Linear spatial expansion module, (4) “Align”: the L c o n term in Equation (12).
The ablation experiment emphasized the importance of each individual component. The contribution of the different components to three datasets varied. First, the improvement in the generalization performance of the CNN backbone using only X s p e (only spectral linear expansion) varied significantly among the different datasets. For SD = houston13, the OA increased by 0.44%, and for SD = PaviaU/Hangzhou, OA increased by 4.19%/3.24%. Considering that X s p e was used to simulate the weakening effect of the atmospheric window on the band reflectivity, the reasons for the above differences may have resulted from two factors: first, the Houston dataset was less affected by the atmospheric window than the other datasets (or as IEEE GRSS data fusion competition data, it achieved better atmospheric correction); second, the Houston dataset only extracted 48 bands for cross-scene classification, and the out-of-distribution information introduced by spectral linear expansion was insufficient.
Secondly, the improvement in the generalization performance of the CNN backbone using only X s p a (only spatial linear expansion) also varied significantly among the different datasets. For SD = houston13, OA increased by 0.09%, and for SD = PaviaU/Hangzhou, OA increased by 6.71%/3.75%. Considering that spatial linear expansion was used to alleviate the model learning bias caused by the co-occurrence of image patches, the Houston13 training set was more dispersed in spatial distribution than the PaviaU/Hangzhou training set, and the clustering areas of object categories in each location were far away from each other, and the co-occurrence of image patches was quite rare. This can explain to some extent that spatial linear expansion resulted in limited improvement in the cross-scene generalization performance for the Houston dataset.
In particular, when spatial-spectral linear expansion was simultaneously adopted, compared with the CNN backbone, for SD = houston13, the OA increased by 0.84%, and for SD = PaviaU/Hangzhou, OA increased by 2.76%/4.81%. Except for SD = PaviaU, spatial-spectral linear expansion was better than spatial or spectral linear expansion alone. This abnormal superposition effect can be explained by the aggravation of label shift using simple data expansion. For SD = PaviaU, the proportion of training samples was quite unbalanced. The simultaneous spatial and spectral linear expansion of the training domain, to a certain extent, made the optimization gradient of cross-entropy be biased towards the large proportion of categories, which in turn led to the preference of model learning for specific categories. Nevertheless, adding supervised contrast regularization terms on the basis of linear expansion can effectively alleviate this problem. An important reason behind this is that by aggregating the features of samples of the same type and increasing the feature distance of samples of different classes through L c o n , the ideal domain-invariant features are further approached, while promoting a uniform distribution of optimization energy. As shown in the last row of Table 4, the complete algorithm increased by 7.25% compared to the CNN backbone, for SD = PaviaU.

4.3. Performance in Cross-Scene HSI Classification

Relevant algorithms, including mixed-based data augmentation methods: Mixup [42], MixStyle [43], and StyleHall [44]; SDG methods in computer vision: SagNet [60], LDSDG [36], and PDEN [35]; and SDG Methods for HSI: SDEnet [10], FDGNet [39], D3Net [61], and ISDGSNet [62] were used for comparison.
LDSDG and SagNet used patch sizes of 32 × 32 to match the ResNet18 input size, while the other networks used a uniform 13 × 13 patch size. The results for SD = Houston13/PaviaU were as follows: The training and test data were consistent and the dataset division and preprocessing procedure were unchanged, ref. [10] presents the results for SagNet, LDSDG, and PDEN in Table 5 and Table 6 (the classification results of each category in the original paper [10] related to computer vision methods do not report standard deviations). The other results in Table 5, Table 6, Table 7 and Table 8 were reproduced on the author’s software and hardware platform.
Compared with the best mix-based data augmentation method, Mixup, for SD = Houston13, the OA of the SSLE improved by 2.43%, and for SD = PaviaU/Hangzhou, the OA improved by 2.91%/2.49%. This shows that the spatial-spectral complementary linear expansion adopted in this study was better than the basic MSDA strategy of computer vision under the supervised alignment constraint. In addition, when alignment regularization was not used and only the data-level expansion was retained, as shown in the fourth row of Table 4, the OA under the three cross-scene datasets was lower than Mixup. One of the important reasons for this was that Mixup is a data-independent enhancement method. Compared with the specific spatial-spectral linear expansion, it is less likely to introduce strong inductive bias and thus overfit the source domain when used alone. In addition, it is worth noting that the MixStyle method showed negative enhancement of SD = houston13/PaviaU, and the OA was significantly lower than that when using only the CNN backbone. To a certain extent, this is because MixStyle probabilistically mixes instance-level feature statistics between training samples and synthesizes new domains at the feature level to indirectly achieve cross-domain generalization. This feature perturbation of the discriminator can easily lead to unexpected information loss, especially for datasets such as HSIs where the training samples are not large enough and the differences between some categories are small.
Compared with the best SDG methods in computer vision, PDEN, for SD = Houston13, the OA of the SSLE improved by 3.45%, and for SD = PaviaU/Hangzhou, the OA improved by 2.23%/0.83%. These SDG networks mainly focus on RGB image classification, especially on spatial information. The backbone feature extraction backbone of the SSLE refers to the popular network structure for HSIC, which has been widely adopted in related studies such as [38,63] and is specifically used for HSIs with rich spectral information. In addition, although these SDG strategies also adopt certain data enhancement strategies, they ignore the spectral drift problem of HSIs, because they rely on resampling the spatial distribution.
Compared to the HSI domain extension method SDEnet, the OA of the SSLE for TD = Houston18 increased by 0.81%, and the OA for TD = PaviaC/Loukia increased by 0.22%/0.92%. Considering that SDENet uses a spatial-spectral codec generator to learn out-of-distribution information in the source domain, although it uses a nonlinear neural network with a certain risk of feature contamination, its generalization performance was not greatly affected, because its network constrains the learnable feature space with a specific morphological encoder (similarly, FDGNet uses a frequency domain encoder for feature constraints). The SSLE uses a fixed spatial-spectral linear extension and also introduces out-of-distribution spatial-spectral information for the learning of the discriminator. Compared with this type of learnable nonlinear transformation, although it has a lower theoretical improvement upper limit and there was no significant difference between the results on most datasets, the standard deviation of the results was smaller overall, indicating that it had certain advantages in predictive stability.
In addition, the tables also report the standard deviation of the results of the different algorithms under repeated tests. This indicator can reflect the robustness of the algorithms to a certain extent. For SD = Houston13, the OA standard deviation of the proposed algorithm was only 0.87, which was smaller than all other algorithms. For SD = PaviaU/Hangzhou/Dioni, the OA standard deviation of the proposed algorithm was 1.41/1.81/0.95, which was medium to low. Overall, the stability of SSLE’s classification results was relatively high. In particular, for the four types of land cover types corresponding to the four pairs of datasets, the generalization performance of SSLE had significant differences. For SD = Hangzhou, thanks to more training samples and fewer categories, SSLE’s OA was as high as 90.85%. For SD = Dioni, affected by significant class imbalance and complex types, SSLE’s OA was only 62.69%, and the recognition results for categories 4, 5, and 6 were poor.
Furthermore, this study compared the confusion matrices of the HSIC DG methods. As shown in Figure 10, for SD = Houston13, the number of true positive examples predicted by SDENet and FDGNet for easily confused residential buildings and non-residential buildings was lower than that of the SSLE. For SD = PaviaU/Hangzhou, the SSLE had basically no preference for certain specific categories compared to SDENet and FDGNet (for SD = PaviaU, SDENet highlights the confusion between Asphalt and Bitumen, Meadow and Bare soil, and for SD = Hangzhou, FDGNet highlights the confusion between Land/building and Plant).
The prediction results are depicted in Figure 11, Figure 12 and Figure 13. The spatial distribution of SSLE is clear and free of noticeable noise, with distinct boundaries for each category. This demonstrates its ability to identify spatial correlations and its sensitivity in extracting class-specific information. The number of trainable parameters and the testing time with a batch size of 256 are presented in Figure 14 to demonstrate the computational cost of the various methods. Taking advantage of linear spectral-spatial augmentation that does not necessitate training, and feature alignment without involving a testing step, the adopted network had the fewest trainable parameters and quite a short testing time.

4.4. Parameter Sensitivity Analysis

A set of experiments were conducted to evaluate the parameter sensitivity of the spatial spectral linear expansion algorithm on the three TDs. The parameters λ / σ in Figure 15 and step size in Figure 16 were regarded as adjustable hyperparameters and were selected from { 0 , 0.25 , 0.5 , 0.75 , 1 } / { 1 , 3 , 5 , 7 , 9 } and { 1 × 10 5 , 1 × 10 4 , 1 × 10 3 , 1 × 10 2 , 1 × 10 1 } , respectively.
The terms λ and σ have clear interpretability, as shown in Figure 17 and Figure 18. The “roughness” of the enhanced spectral vector R s p e is determined by λ . The reciprocal of λ grows larger as the enhanced R s p e from X s p e gets closer to the R s o from X s o . As shown in Figure 15, a reduction in OA occurred when λ was either too high or too low. As λ = 1 led to a redundant expansion, λ = 0 led to a substantial loss of high-frequency information and an excessive loss of absorption features.
The centralization of spatial weight w n o r m ( u , v ) was affected by σ , as shown in Figure 15. A higher σ signifies less centralization and closer proximity to an all-one matrix. Similarly to λ , an extremely high or low value will lead to bias in OA. For σ = 1 , the X s p a primarily represents the spectral properties of the central pixel. Conversely, for σ = 9 , the X s p a becomes almost indistinguishable from the X s o and results in invalid expansion.

5. Conclusions

In order to alleviate the feature contamination problem of source expansion in nonlinear neural networks, this study proposed spatial-spectral linear extrapolation for mixed sample data augmentation, where spectral transformation was used to simulate the flattening weakening of band reflectivity by atmospheric windows to reduce the data gap between source and target domains, and spatial linear transformation applied spatial weights to address the misleading of model learning caused by the co-occurrence of image patches. In addition, inter-domain feature contrast regularization was employed for learning in multiple domains, which promoted a uniform distribution of optimized energy and prevented the gradient bias of cross-entropy caused by an increase in training domains with the same label distribution. In addition, this was also conducive to learning robust domain invariant features.
Ablation experiments showed the effectiveness of the proposed components, emphasizing the potential impact of data quality, spectral information, spatial distribution of training image patches, and label statistical characteristics on the generalization performance of each module. Comparative experiments showed that the SSLE was superior to representative mix-based strategies and DG methods in computer vision, and its generalization performance was comparable to that of advanced nonlinear expansion-based HSIC algorithms. In particular, compared with these learnable SDG algorithms, although the theoretical performance upper limit of the SSLE was lower, the standard deviation of the experimental results was smaller overall, indicating that it had a higher prediction stability. Furthermore, from the confusion matrix, it can be seen that the SSLE had less obvious misclassifications of specific easily confused categories.
The limitation of SSLE is that its expansion results are restricted to a specific linear manifold, which makes it difficult to fully simulate the spectral shift under various environmental factors. Future work will focus on discovering and learning the intrinsic components of HSIs, and internalizing the imaging mechanism into the model structure and optimization strategy as prior knowledge.

Author Contributions

Methodology, H.Z.; Software, J.W.; Formal analysis, Z.Z.; Resources, L.L.; Writing—original draft, H.Z.; Writing—review & editing, L.L.; Visualization, S.G.; Supervision, L.L.; Funding acquisition, L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xu, Y.; Du, B.; Zhang, L. Self-Attention Context Network: Addressing the Threat of Adversarial Attacks for Hyperspectral Image Classification. IEEE Trans. Image Process. 2021, 30, 8671–8685. [Google Scholar] [CrossRef] [PubMed]
  2. Wu, H.; Prasad, S. Semi-Supervised Deep Learning Using Pseudo Labels for Hyperspectral Image Classification. IEEE Trans. Image Process. 2018, 27, 1259–1270. [Google Scholar] [CrossRef] [PubMed]
  3. Dong, Y.; Liu, Q.; Du, B.; Zhang, L. Weighted Feature Fusion of Convolutional Neural Network and Graph Attention Network for Hyperspectral Image Classification. IEEE Trans. Image Process. 2022, 31, 1559–1572. [Google Scholar] [CrossRef] [PubMed]
  4. Zhou, S.; Wu, H.; Xue, Z. Grouped subspace linear semantic alignment for hyperspectral image transfer learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4412116. [Google Scholar] [CrossRef]
  5. Xu, Y.; Wu, S.; Wang, B.; Yang, M.; Wu, Z.; Yao, Y.; Wei, Z. Two-stage fine-grained image classification model based on multi-granularity feature fusion. Pattern Recognit. 2024, 146, 110042. [Google Scholar] [CrossRef]
  6. Zhang, M.; Wang, Z.; Wang, X.; Gong, M.; Wu, Y.; Li, H. Features kept generative adversarial network data augmentation strategy for hyperspectral image classification. Pattern Recognit. 2023, 142, 109701. [Google Scholar] [CrossRef]
  7. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  8. He, X.; Chen, Y.; Lin, Z. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  9. Xie, J.; Hua, J.; Chen, S.; Wu, P.; Gao, P.; Sun, D.; Lyu, Z.; Lyu, S.; Xue, X.; Lu, J. HyperSFormer: A Transformer-Based End-to-End Hyperspectral Image Classification Method for Crop Classification. Remote Sens. 2023, 15, 3491. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Li, W.; Sun, W.; Tao, R.; Du, Q. Single-Source Domain Expansion Network for Cross-Scene Hyperspectral Image Classification. IEEE Trans. Image Process. 2023, 32, 1498–1512. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Zhang, M.; Li, W.; Wang, S.; Tao, R. Language-Aware Domain Generalization Network for Cross-Scene Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–12. [Google Scholar] [CrossRef]
  12. Peng, J.; Sun, W.; Ma, L.; Du, Q. Discriminative transfer joint matching for domain adaptation in hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 972–976. [Google Scholar] [CrossRef]
  13. Ma, L.; Luo, C.; Peng, J.; Du, Q. Unsupervised manifold alignment for cross-domain classification of remote sensing images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1650–1654. [Google Scholar] [CrossRef]
  14. Ma, L.; Crawford, M.M.; Zhu, L.; Liu, Y. Centroid and covariance alignment-based domain adaptation for unsupervised classification of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 2305–2323. [Google Scholar] [CrossRef]
  15. Matasci, G.; Volpi, M.; Kanevski, M.; Bruzzone, L.; Tuia, D. Semisupervised Transfer Component Analysis for Domain Adaptation in Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3550–3564. [Google Scholar] [CrossRef]
  16. Sun, B.; Feng, J.; Saenko, K. Return of Frustratingly Easy Domain Adaptation. arXiv 2015, arXiv:1511.05547. [Google Scholar] [CrossRef]
  17. Ye, M.; Qian, Y.; Zhou, J.; Tang, Y.Y. Dictionary Learning-Based Feature-Level Domain Adaptation for Cross-Scene Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1544–1562. [Google Scholar] [CrossRef]
  18. Persello, C.; Bruzzone, L. Kernel-Based Domain-Invariant Feature Selection in Hyperspectral Images for Transfer Learning. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2615–2626. [Google Scholar] [CrossRef]
  19. Liu, T.; Zhang, X.; Gu, Y. Unsupervised Cross-Temporal Classification of Hyperspectral Images With Multiple Geodesic Flow Kernel Learning. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9688–9701. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Li, W.; Zhang, M.; Qu, Y.; Tao, R.; Qi, H. Topological structure and semantic information transfer network for cross-scene hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 2817–2830. [Google Scholar] [CrossRef]
  21. Wang, P.; Zhang, Z.; Lei, Z.; Zhang, L. Sharpness-Aware Gradient Matching for Domain Generalization. In Proceedings of the IEEE Comput Soc Conf Comput Vision Pattern Recognit, Vancouver, BC, Canada, 18–22 June 2023; pp. 3769–3778. [Google Scholar]
  22. Wang, J.; Lan, C.; Liu, C.; Ouyang, Y.; Qin, T.; Lu, W.; Chen, Y.; Zeng, W.; Yu, P. Generalizing to unseen domains: A survey on domain generalization. IEEE Trans. Knowl. Data Eng. 2022, 35, 8052–8072. [Google Scholar] [CrossRef]
  23. Qiao, F.; Zhao, L.; Peng, X. Learning to learn single domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12556–12565. [Google Scholar]
  24. Liu, A.H.; Liu, Y.C.; Yeh, Y.Y.; Wang, Y.C.F. A unified feature disentangler for multi-domain image translation and manipulation. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
  25. Yue, X.; Zhang, Y.; Zhao, S.; Sangiovanni-Vincentelli, A.; Keutzer, K.; Gong, B. Domain randomization and pyramid consistency: Simulation-to-real generalization without accessing target domain data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 November–2 October 2019; pp. 2100–2110. [Google Scholar]
  26. Shankar, S.; Piratla, V.; Chakrabarti, S.; Chaudhuri, S.; Jyothi, P.; Sarawagi, S. Generalizing across domains via cross-gradient training. arXiv 2018, arXiv:1804.10745. [Google Scholar]
  27. Blanchard, G.; Deshmukh, A.A.; Dogan, Ü.; Lee, G.; Scott, C. Domain generalization by marginal transfer learning. J. Mach. Learn. Res. 2021, 22, 46–100. [Google Scholar]
  28. Ghifary, M.; Kleijn, W.B.; Zhang, M.; Balduzzi, D. Domain generalization for object recognition with multi-task autoencoders. In Proceedings of the IEEE international Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2551–2559. [Google Scholar]
  29. Motiian, S.; Piccirilli, M.; Adjeroh, D.A.; Doretto, G. Unified deep supervised domain adaptation and generalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5715–5725. [Google Scholar]
  30. Ganin, Y.; Lempitsky, V. Unsupervised domain adaptation by backpropagation. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 1180–1189. [Google Scholar]
  31. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  32. Arjovsky, M.; Bottou, L.; Gulrajani, I.; Lopez-Paz, D. Invariant risk minimization. arXiv 2019, arXiv:1907.02893. [Google Scholar]
  33. D’Innocente, A.; Caputo, B. Domain generalization with domain-specific aggregation modules. In Proceedings of the Pattern Recognition: 40th German Conference, GCPR 2018, Stuttgart, Germany, 9–12 October 2018; Proceedings 40. Springer: Berlin/Heidelberg, Germany, 2019; pp. 187–198. [Google Scholar]
  34. Li, D.; Yang, Y.; Song, Y.Z.; Hospedales, T. Learning to generalize: Meta-learning for domain generalization. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  35. Li, L.; Gao, K.; Cao, J.; Huang, Z.; Weng, Y.; Mi, X.; Yu, Z.; Li, X.; Xia, B. Progressive Domain Expansion Network for Single Domain Generalization. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 224–233. [Google Scholar] [CrossRef]
  36. Wang, Z.; Luo, Y.; Qiu, R.; Huang, Z.; Baktashmotlagh, M. Learning to Diversify for Single Domain Generalization. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 814–823. [Google Scholar] [CrossRef]
  37. Wang, R.; Yi, M.; Chen, Z.; Zhu, S. Out-of-Distribution Generalization With Causal Invariant Transformations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 375–385. [Google Scholar]
  38. Dong, L.; Geng, J.; Jiang, W. Spectral–Spatial Enhancement and Causal Constraint for Hyperspectral Image Cross-Scene Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–13. [Google Scholar] [CrossRef]
  39. Qin, B.; Feng, S.; Zhao, C.; Xi, B.; Li, W.; Tao, R. FDGNet: Frequency Disentanglement and Data Geometry for Domain Generalization in Cross-Scene Hyperspectral Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2024, 1–14. [Google Scholar] [CrossRef]
  40. Balestriero, R.; Pesenti, J.; LeCun, Y. Learning in high dimension always amounts to extrapolation. arXiv 2021, arXiv:2110.09485. [Google Scholar]
  41. Zhang, T.; Zhao, C.; Chen, G.; Jiang, Y.; Chen, F. Feature Contamination: Neural Networks Learn Uncorrelated Features and Fail to Generalize. arXiv 2024, arXiv:2406.03345. [Google Scholar]
  42. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond Empirical Risk Minimization. arXiv 2018, arXiv:1710.09412. [Google Scholar] [CrossRef]
  43. Zhou, K.; Yang, Y.; Qiao, Y.; Xiang, T. Domain Generalization with MixStyle. arXiv 2021, arXiv:2104.02008. [Google Scholar] [CrossRef]
  44. Zhao, Y.; Zhong, Z.; Zhao, N.; Sebe, N.; Lee, G.H. Style-hallucinated dual consistency learning for domain generalized semantic segmentation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 535–552. [Google Scholar]
  45. Lei, M.; Wu, H.; Lv, X.; Wang, X. ConDSeg: A General Medical Image Segmentation Framework via Contrast-Driven Feature Enhancement. arXiv 2024, arXiv:2412.08345. [Google Scholar] [CrossRef]
  46. Shah, H.; Tamuly, K.; Raghunathan, A.; Jain, P.; Netrapalli, P. The pitfalls of simplicity bias in neural networks. Adv. Neural Inf. Process. Syst. 2020, 33, 9573–9585. [Google Scholar]
  47. Pezeshki, M.; Kaba, O.; Bengio, Y.; Courville, A.C.; Precup, D.; Lajoie, G. Gradient starvation: A learning proclivity in neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 1256–1272. [Google Scholar]
  48. Chen, J.; Yang, Z.; Yang, D. Mixtext: Linguistically-informed interpolation of hidden space for semi-supervised text classification. arXiv 2020, arXiv:2004.12239. [Google Scholar]
  49. Bohren, C.F.; Huffman, D.R. Absorption and Scattering of Light by Small Particles; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
  50. Bhatia, N.; Tolpekin, V.A.; Reusen, I.; Sterckx, S.; Biesemans, J.; Stein, A. Sensitivity of reflectance to water vapor and aerosol optical thickness. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3199–3208. [Google Scholar] [CrossRef]
  51. Zhang, H.; Han, X.; Deng, J.; Sun, W. How to Evaluate and Remove the Weakened Bands in Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 1–15. [Google Scholar] [CrossRef]
  52. Blake, A.; Kohli, P.; Rother, C. Markov Random Fields for Vision and Image Processing; MIT Press: Cambridge, MA, USA, 2011. [Google Scholar]
  53. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised contrastive learning. Adv. Neural Inf. Process. Syst. 2020, 33, 18661–18673. [Google Scholar]
  54. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  55. Kingma, D.P. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  56. Debes, C.; Merentitis, A.; Heremans, R.; Hahn, J.; Frangiadakis, N.; van Kasteren, T.; Liao, W.; Bellens, R.; Pižurica, A.; Gautama, S.; et al. Hyperspectral and LiDAR Data Fusion: Outcome of the 2013 GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2405–2418. [Google Scholar] [CrossRef]
  57. Le Saux, B.; Yokoya, N.; Hansch, R.; Prasad, S. 2018 IEEE GRSS Data Fusion Contest: Multimodal Land Use Classification [Technical Committees]. IEEE Geosci. Remote Sens. Mag. 2018, 6, 52–54. [Google Scholar] [CrossRef]
  58. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph Information Aggregation Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1912–1925. [Google Scholar] [CrossRef]
  59. Karantzalos, K.; Karakizi, C.; Kandylakis, Z.; Antoniou, G. HyRANK Hyperspectral Satellite Dataset I, Version 001; Zenodo: Geneva, Switzerland, 2018. [CrossRef]
  60. Nam, H.; Lee, H.; Park, J.; Yoon, W.; Yoo, D. Reducing Domain Gap by Reducing Style Bias. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8686–8695. [Google Scholar] [CrossRef]
  61. Chu, M.; Yu, X.; Dong, H.; Zang, S. Domain-Adversarial Generative and Dual Feature Representation Discriminative Network for Hyperspectral Image Domain Generalization. IEEE Trans. Geosci. Remote Sens. 2024, 32, 5533213. [Google Scholar] [CrossRef]
  62. Gao, J.; Ji, X.; Ye, F.; Chen, G. Invariant semantic domain generalization shuffle network for cross-scene hyperspectral image classification. Expert Syst. Appl. 2025, 273, 126818. [Google Scholar] [CrossRef]
  63. Zhao, H.; Zhang, J.; Lin, L.; Wang, J.; Gao, S.; Zhang, Z. Locally linear unbiased randomization network for cross-scene hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5526512. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method. The training stage includes spatial-spectral linear expansion and inter-domain feature alignment. A fundamental CNN classifier backbone (marked in gray, combined with 2D convolution, ReLU activation, maximal pooling, and fully connected layer) with feature and category projection used for training. X s p a and X s p e represent two image patches generated by X s o from the SD. R s o and R s p e are pixels (one-dimensional spectrum) from X s o and X s p e , respectively. The embedding features of X s o , X s p a , and X s p e output by the projection head are shown as a square, trapezoid, and regular hexagon. In the test stage, a CNN classifier backbone with category projection is used for the prediction of the image patch from the TD.
Figure 1. Flowchart of the proposed method. The training stage includes spatial-spectral linear expansion and inter-domain feature alignment. A fundamental CNN classifier backbone (marked in gray, combined with 2D convolution, ReLU activation, maximal pooling, and fully connected layer) with feature and category projection used for training. X s p a and X s p e represent two image patches generated by X s o from the SD. R s o and R s p e are pixels (one-dimensional spectrum) from X s o and X s p e , respectively. The embedding features of X s o , X s p a , and X s p e output by the projection head are shown as a square, trapezoid, and regular hexagon. In the test stage, a CNN classifier backbone with category projection is used for the prediction of the image patch from the TD.
Remotesensing 17 01816 g001
Figure 2. Flowchart of the spectral linear transformation.
Figure 2. Flowchart of the spectral linear transformation.
Remotesensing 17 01816 g002
Figure 3. Flowchart of the spatial linear transformation.
Figure 3. Flowchart of the spatial linear transformation.
Remotesensing 17 01816 g003
Figure 4. Flowchart of the classifier network.
Figure 4. Flowchart of the classifier network.
Remotesensing 17 01816 g004
Figure 5. T-SNE of class separability of the SSLE using the Pavia dataset, where SD is the PaviaU data, TD is the PaviaC data, and the number represents the class index. (a) Raw sample of SD ( X s o ). (b) SD ( X s o ) feature vectors by feature projection (P). (c) SD ( X s o ) feature vectors by feature projection (P) without L c o n . (d) Samples of X s p a . (e) X s p a feature vectors by P. (f) X s p a feature vectors by P without L c o n . (g) Samples of X s p e . (h) X s p e feature vectors by P. (i) X s p e feature vectors by P without L c o n . (j) Raw samples of TD. (k) TD feature vectors by P. (l) TD feature vectors by P without L c o n .
Figure 5. T-SNE of class separability of the SSLE using the Pavia dataset, where SD is the PaviaU data, TD is the PaviaC data, and the number represents the class index. (a) Raw sample of SD ( X s o ). (b) SD ( X s o ) feature vectors by feature projection (P). (c) SD ( X s o ) feature vectors by feature projection (P) without L c o n . (d) Samples of X s p a . (e) X s p a feature vectors by P. (f) X s p a feature vectors by P without L c o n . (g) Samples of X s p e . (h) X s p e feature vectors by P. (i) X s p e feature vectors by P without L c o n . (j) Raw samples of TD. (k) TD feature vectors by P. (l) TD feature vectors by P without L c o n .
Remotesensing 17 01816 g005
Figure 6. Pseudo-color image and ground truth map [20,58] of Houston dataset: (a) Pseudo-color image of Houston 2013, (b) Pseudo-color image of Houston 2018, (c) Ground truth map of Houston 2013, (d) Ground truth map of Houston 2018.
Figure 6. Pseudo-color image and ground truth map [20,58] of Houston dataset: (a) Pseudo-color image of Houston 2013, (b) Pseudo-color image of Houston 2018, (c) Ground truth map of Houston 2013, (d) Ground truth map of Houston 2018.
Remotesensing 17 01816 g006
Figure 7. Pseudo-color image and ground truth map [20,58] of Pavia dataset: (a) Pseudo-color image of University of Pavia, (b) Ground truth map of University of Pavia, (c) Pseudo-color image of Pavia Center, (d) Ground truth map of Pavia Center.
Figure 7. Pseudo-color image and ground truth map [20,58] of Pavia dataset: (a) Pseudo-color image of University of Pavia, (b) Ground truth map of University of Pavia, (c) Pseudo-color image of Pavia Center, (d) Ground truth map of Pavia Center.
Remotesensing 17 01816 g007
Figure 8. Pseudocolor image and ground-truth map of S-H [20]. (a) Pseudocolor image of Shanghai. (b) Pseudocolor image of Hangzhou. (c) Ground-truth map of Shanghai. (d) Ground-truth map of Hangzhou.
Figure 8. Pseudocolor image and ground-truth map of S-H [20]. (a) Pseudocolor image of Shanghai. (b) Pseudocolor image of Hangzhou. (c) Ground-truth map of Shanghai. (d) Ground-truth map of Hangzhou.
Remotesensing 17 01816 g008
Figure 9. Pseudocolor image and ground-truth map of HyRANK [59]. (a) Pseudocolor image of Dioni. (b) Pseudocolor image of Loukia. (c) Ground-truth map of Dioni. (d) Ground-truth map of Loukia.
Figure 9. Pseudocolor image and ground-truth map of HyRANK [59]. (a) Pseudocolor image of Dioni. (b) Pseudocolor image of Loukia. (c) Ground-truth map of Dioni. (d) Ground-truth map of Loukia.
Remotesensing 17 01816 g009
Figure 10. Confusion matrix with different methods for three TDs. (a) SDEnet (77.56%). (b) FDGNet (78.76%). (c) SSLE (78.96%). (d) SDEnet (83.07%). (e) FDGNet (82.47%). (f) SSLE (83.49%). (g) SDEnet (90.49%). (h) FDGNet (82.47%). (i) SSLE (90.94%).
Figure 10. Confusion matrix with different methods for three TDs. (a) SDEnet (77.56%). (b) FDGNet (78.76%). (c) SSLE (78.96%). (d) SDEnet (83.07%). (e) FDGNet (82.47%). (f) SSLE (83.49%). (g) SDEnet (90.49%). (h) FDGNet (82.47%). (i) SSLE (90.94%).
Remotesensing 17 01816 g010
Figure 11. Data visualization and classification maps for target scene Houston 2018 data obtained with different methods, including (a) Mixup (75.53%), (b) Mixstyle (60.21%), (c) StyleHall (76.55%), (d) SagNet (73.64%), (e) LDSDG (73.55%), (f) PDEN (75.98%), (g) SDEnet (77.56%), (h) FDGNet (78.76%), (i) SSLE (78.96%).
Figure 11. Data visualization and classification maps for target scene Houston 2018 data obtained with different methods, including (a) Mixup (75.53%), (b) Mixstyle (60.21%), (c) StyleHall (76.55%), (d) SagNet (73.64%), (e) LDSDG (73.55%), (f) PDEN (75.98%), (g) SDEnet (77.56%), (h) FDGNet (78.76%), (i) SSLE (78.96%).
Remotesensing 17 01816 g011
Figure 12. Data visualization and classification maps for target scene Pavia Center data obtained with different methods, including (a) Mixup (74.96%), (b) Mixstyle (71.63%), (c) StyleHall (83.69%), (d) SagNet (69.90%), (e) LDSDG (71.02%), (f) PDEN (80.87%), (g) SDEnet (81.76%), (h) FDGNet (82.47%), (i) SSLE (83.49%).
Figure 12. Data visualization and classification maps for target scene Pavia Center data obtained with different methods, including (a) Mixup (74.96%), (b) Mixstyle (71.63%), (c) StyleHall (83.69%), (d) SagNet (69.90%), (e) LDSDG (71.02%), (f) PDEN (80.87%), (g) SDEnet (81.76%), (h) FDGNet (82.47%), (i) SSLE (83.49%).
Remotesensing 17 01816 g012
Figure 13. Data visualization and classification maps for target scene shanghai data obtained with different methods, including (a) Mixup (87.47%), (b) Mixstyle (89.74%), (c) StyleHall (87.37%), (d) SagNet (82.72%), (e) LDSDG (87.43%), (f) PDEN (84.12%), (g) SDEnet (90.49%), (h) FDGNet (82.47%), (i) SSLE (90.94%).
Figure 13. Data visualization and classification maps for target scene shanghai data obtained with different methods, including (a) Mixup (87.47%), (b) Mixstyle (89.74%), (c) StyleHall (87.37%), (d) SagNet (82.72%), (e) LDSDG (87.43%), (f) PDEN (84.12%), (g) SDEnet (90.49%), (h) FDGNet (82.47%), (i) SSLE (90.94%).
Remotesensing 17 01816 g013
Figure 14. Comparison of total trainable parameter size and testing time. (a) SD = Houston13. (b) SD = PaviaU. (c) SD = HangZhou.
Figure 14. Comparison of total trainable parameter size and testing time. (a) SD = Houston13. (b) SD = PaviaU. (c) SD = HangZhou.
Remotesensing 17 01816 g014
Figure 15. Parameter tuning of λ and σ for three TDs. (a) TD = Houston18, (b) TD = Pavia Center, (c) TD = ShangHai.
Figure 15. Parameter tuning of λ and σ for three TDs. (a) TD = Houston18, (b) TD = Pavia Center, (c) TD = ShangHai.
Remotesensing 17 01816 g015
Figure 16. Parameter tuning of step size for MRF using the three TDs.
Figure 16. Parameter tuning of step size for MRF using the three TDs.
Remotesensing 17 01816 g016
Figure 17. Visual comparison results of spectral linear transformation under different λ . (a) The solid line represents a “Road” pixel (X-167, Y-130), and the dashed line represents a “Water” pixel (X-66, Y-472). (b) The solid line represents a “Bitumen” pixel (X-326, Y-121), the dotted line represents a “Meadow” pixel (X-328, Y-195).
Figure 17. Visual comparison results of spectral linear transformation under different λ . (a) The solid line represents a “Road” pixel (X-167, Y-130), and the dashed line represents a “Water” pixel (X-66, Y-472). (b) The solid line represents a “Bitumen” pixel (X-326, Y-121), the dotted line represents a “Meadow” pixel (X-328, Y-195).
Remotesensing 17 01816 g017
Figure 18. Visualization results of normalized spatial weights w n o r m ( u , v ) of 13*13 size under different standard deviations σ . The darker the color, the smaller the weight. (a) σ = 1, (b) σ = 3, (c) σ = 5, (d) σ = 7, (e) σ = 9.
Figure 18. Visualization results of normalized spatial weights w n o r m ( u , v ) of 13*13 size under different standard deviations σ . The darker the color, the smaller the weight. (a) σ = 1, (b) σ = 3, (c) σ = 5, (d) σ = 7, (e) σ = 9.
Remotesensing 17 01816 g018
Table 1. Abbreviation description.
Table 1. Abbreviation description.
AbbreviationDescription
HSICHyperspectral image classification
SDSource domain
TDTarget domain
DADomain adaptation
DGDomain generalization
SDGSingle domain generalization
SSLESpatial-spectral linear extrapolation
CRMConv2D-Relu-Maxpool2D
MRFMarkov random field
PProjection head
Table 2. Number of source and target samples for the three datasets.
Table 2. Number of source and target samples for the three datasets.
Houston DatasetPavia DatasetHangzhou-Shanghai Dataset
Class Number of Samples Class Number of Samples Class Number of Samples
No. Name Houston 2013
(Source)
Houston 2018
(Target)
No. Name PaviaU
(Source)
PaviaC
(Target)
No. Name Hangzhou
(Source)
Shanghai
(Target)
1Grass healthy34513531Tree30647598
2Grass stressed36548882Asphalt66319248
3Trees36527663Brick368226851Water18,043123,123
4Water285224Bitumen133072872Land/Building77,450161,689
5Residential buildings31953475Shadow94728633Plant40,20783,188
6Non-residential buildings40832,4596Meadow18,6493090
7Road44363657Bare soil50296584
Total253053,200 Total39,33239,355 Total135,700368,000
Table 3. Number of source and target samples for the three datasets.
Table 3. Number of source and target samples for the three datasets.
HyRANK Dataset
Class Number of Samples
No. Name Dioni
(Source)
Loukia
(Target)
C1Dense Urban Fabric1262206
C2Mineral Extraction Sites20454
C3Non Irrigated Arable Land614426
C4Fruit Trees15079
C5Olive Groves17681107
C6Coniferous Forest361422
C7Dense Sclerophyllous Vegetation50352996
C8Sparce Sclerophyllous Vegetation63742361
C9Sparsely Vegetated Areas1754399
C10Rocks and Sand492453
C11Water16121393
C12Coastal Water398421
Total20,02410,317
Table 4. Ablation comparison of each variant of the proposed method.
Table 4. Ablation comparison of each variant of the proposed method.
CRM
Backbone
X spe X spa AlignHouston2018
(Target)
PaviaCenter
(Target)
ShangHai
(Target)
74.43 ± 2.3675.25 ± 5.5183.81 ± 3.09
74.87 ± 1.1579.44 ± 5.9287.05 ± 3.18
74.52 ± 0.9281.96 ± 2.6187.56 ± 2.54
75.27 ± 0.9178.01 ± 3.7688.62 ± 1.47
78.85 ± 0.8782.50 ± 1.4190.85 ± 1.81
Table 5. Quantitative classification results of different methods for the target domain Houston 2018.
Table 5. Quantitative classification results of different methods for the target domain Houston 2018.
Classification Algorithms
Mixed-Based Data Augmentation SDG Methods in Computer Vision SDG Methods for Cross-Scene HSI Classification
Class Mixup MixStyle StyleHall SagNet LDSDG PDEN SDENet FDGNet D3Net ISDGSNet Proposed
141.88 ± 15.1852.56 ± 2.0247.23 ± 12.4725.7910.1346.4974.08 ± 22.4249.42 ± 17.2820.12 ± 10.3648.25 ± 18.2751.34 ± 11.43
289.30 ± 2.5668.80 ± 2.1879.84 ± 8.9862.7962.9777.6075.16 ± 3.9976.18 ± 6.5982.9 ± 6.1873.10 ± 11.1877.47 ± 7.41
372.44 ± 2.8263.75 ± 10.0465.58 ± 2.6348.6660.8159.7360.59 ± 2.9862.89 ± 2.3653.17 ± 7.5164.06 ± 7.4467.12 ± 2.29
4100 ± 0.0099.09 ± 26.48100 ± 0.0081.8281.82100100.00 ± 0100 ± 089.09 ± 9.96100.00 ± 0100 ± 0.00
581.74 ± 3.5648.62 ± 6.6677.60 ± 4.8259.5745.6549.6250.63 ± 11.9768.44 ± 6.6466.48 ± 8.8773.51 ± 8.8966.14 ± 6.76
681.44 ± 2.3761.76 ± 2.2079.85 ± 1.2389.2889.2284.9890.01 ± 3.1787.80 ± 3.6491.82 ± 1.5388.14 ± 2.6588.15 ± 0.86
745.42 ± 10.3858.38 ± 6.6959.84 ± 3.6434.9944.1564.2150.61 ± 1.4952.85 ± 16.3524.66 ± 4.4752.18 ± 8.4953.99 ± 6.93
OA (%)76.42 ± 0.8460.56 ± 2.1975.66 ± 1.1773.01 ± 1.8174.23 ± 1.4975.40 ± 1.7678.04 ± 1.4178.34 ± 1.6876.58 ± 1.0978.72 ± 1.9178.85 ± 0.87
Kappa61.71 ± 1.2742.05 ± 2.4961.08 ± 1.7955.29 ± 3.5055.42 ± 2.2755.87 ± 2.9261.86 ± 1.9862.88 ± 4.5857.30 ± 2.7863.67 ± 3.8763.66 ± 1.97
Table 6. Quantitative classification results of different methods for the target domain Pavia center.
Table 6. Quantitative classification results of different methods for the target domain Pavia center.
Classification Algorithms
Mixed-Based Data Augmentation SDG Methods in Computer Vision SDG Methods for Cross-Scene HSI Classification
Class Mixup MixStyle StyleHall SagNet LDSDG PDEN SDENet FDGNet D3Net ISDGSNet Proposed
191.10 ± 13.9871.33 ± 8.5684.44 ± 4.0698.3591.0985.9386.27 ± 1.3986.44 ± 1.3090.41 ± 4.8489.87 ± 4.7487.51 ± 2.67
282.42 ± 9.0073.71 ± 9.4682.44 ± 2.5259.7673.5188.5682.80 ± 3.0183.25 ± 0.5690.55 ± 1.9988.36 ± 3.8187.55 ± 2.22
345.87 ± 8.1468.28 ± 14.5172.52 ± 13.365.402.2361.3469.35 ± 6.2270.83 ± 7.0970.13 ± 14.9251.89 ± 8.266.64 ± 9.02
463.50 ± 2.0372.40 ± 6.4087.22 ± 0.9687.0371.7285.4985.35 ± 1.884.94 ± 1.3478.21 ± 6.1270.44 ± 18.8278.75 ± 8.71
582.15 ± 8.6670.37 ± 10.2580.28 ± 12.6093.1971.0487.9589.43 ± 2.8386.54 ± 3.2782.30 ± 8.4791.11 ± 2.0389.70 ± 3.55
670.92 ± 5.1874.25 ± 3.5373.59 ± 3.1749.8157.1279.2679.10 ± 1.2772.71 ± 1.7464.14 ± 5.3871.92 ± 2.9874.89 ± 3.01
772.93 ± 3.9365.74 ± 6.6984.96 ± 5.9557.9478.1364.7577.23 ± 5.0381.27 ± 0.9080.38 ± 9.2174.53 ± 9.3880.70 ± 4.52
OA (%)79.59 ± 4.6571.68 ± 1.4282.60 ± 1.5869.28 ± 1.1170.88 ± 1.5580.27 ± 1.3882.28 ± 1.4882.41 ± 0.5982.47 ± 2.8379.44 ± 4.7282.50 ± 1.41
Kappa70.55 ± 5.7866.23 ± 1.5479.11 ± 1.8663.04 ± 1.6764.06 ± 2.0776.81 ± 1.9278.76 ± 1.7778.88 ± 0.6978.85 ± 3.3675.25 ± 5.6378.98 ± 1.70
Table 7. Quantitative classification results of different methods for the target domain Shanghai.
Table 7. Quantitative classification results of different methods for the target domain Shanghai.
Classification Algorithms
Mixed-Based Data Augmentation SDG Methods in Computer Vision SDG Methods for Cross-Scene HSI Classification
Class Mixup MixStyle StyleHall SagNet LDSDG PDEN SDENet FDGNet D3Net ISDGSNet Proposed
194.16 ± 1.0194.32 ± 1.3894.81 ± 0.7293.01 ± 1.6989.26 ± 4.0393.58 ± 1.3493.89 ± 1.0493.81 ± 0.4793.73 ± 2.3293.59 ± 1.0694.13 ± 0.93
284.19 ± 7.4180.49 ± 1.2779.32 ± 11.2473.10 ± 4.2472.33 ± 7.0877.00 ± 6.0373.75 ± 7.5259.65 ± 10.8483.15 ± 8.8479.41 ± 5.8188.02 ± 4.06
385.65 ± 8.5899.58 ± 0.0597.71 ± 4.7599.57 ± 0.1999.40 ± 0.6599.49 ± 0.2299.57 ± 0.1999.95 ± 0.0591.36 ± 13.6499.11 ± 1.0693.30 ± 8.70
OA (%)88.36 ± 2.4589.43 ± 0.5186.96 ± 3.5588.56 ± 1.4887.00 ± 1.0890.02 ± 2.1789.07 ± 2.3680.19 ± 4.8188.54 ± 2.7988.61 ± 2.3090.85 ± 1.81
Kappa82.44 ± 3.5783.99 ± 0.7580.46 ± 5.1578.62 ± 2.7576.75 ± 2.7881.23 ± 4.2579.52 ± 4.4770.92 ± 6.6482.48 ± 4.0782.77 ± 3.3386.06 ± 2.66
Table 8. Quantitative classification results of different methods for the target domain Loukia.
Table 8. Quantitative classification results of different methods for the target domain Loukia.
Classification Algorithms
Mixed-Based Data Augmentation SDG Methods in Computer Vision SDG Methods for Cross-Scene HSI Classification
Class Mixup MixStyle StyleHall SagNet LDSDG PDEN SDEnet FDGNet D3Net ISDGSNet Proposed
C121.52 ± 9.970 ± 027.86 ± 10.120.97 ± 1.290 ± 01.45 ± 1.4626.69 ± 12.3926.37 ± 9.8326.80 ± 24.1320.97 ± 13.2621.94 ± 7.97
C242.59 ± 39.0217.28 ± 17.6014.81 ± 15.870 ± 00 ± 00 ± 070.37 ± 48.1516.04 ± 21.4634.08 ± 23.1960.37 ± 26.658.52 ± 13.13
C356.88 ± 3.9919.79 ± 18.7645.30 ± 12.3740.14 ± 34.950 ± 00 ± 010.95 ± 8.0351.4 ± 15.1437.51 ± 22.5332.49 ± 25.5464.79 ± 13.44
C412.65 ± 1.270 ± 08.60 ± 4.6811.39 ± 19.730 ± 00 ± 05.06 ± 5.8018.14 ± 6.372.02 ± 3.39.87 ± 13.5333.67 ± 17.71
C50.57 ± 0.990.39 ± 0.604.71 ± 3.350.18 ± 0.312.8 ± 4.240.63 ± 0.8137.48 ± 6.500.42 ± 0.296.85 ± 8.8717.98 ± 18.0731.76 ± 13.8
C635.78 ± 2.1710.66 ± 9.731.75 ± 4.6025.12 ± 12.690 ± 00.23 ± 0.4129.78 ± 17.0133.09 ± 6.7018.96 ± 16.6425.83 ± 16.099.48 ± 5.59
C778.41 ± 3.1484.42 ± 1.8775.05 ± 1.7873.74 ± 15.0784.47 ± 4.8183.52 ± 1.3077.67 ± 9.1575.13 ± 0.2376.52 ± 3.0374.33 ± 1.5469.72 ± 1.52
C868.78 ± 7.8562.12 ± 8.0272.32 ± 3.7154.3 ± 18.6461.6 ± 6.6566.38 ± 5.0752.69 ± 8.3973.13 ± 1.6866.32 ± 1.9763.31 ± 6.0960.86 ± 4.66
C932.83 ± 24.0928.23 ± 42.8355.94 ± 19.1722.39 ± 7.0042.35 ± 27.1464.82 ± 10.1874.77 ± 4.753.46 ± 24.6073.13 ± 8.7679.85 ± 4.7176.39 ± 9.25
C100 ± 01.10 ± 1.919.36 ± 20.921.473 ± 2.550 ± 026.34 ± 31.0317.66 ± 26.840 ± 09.01 ± 8.624.24 ± 8.4111.52 ± 10.12
C11100 ± 0100 ± 0100 ± 0100 ± 0100 ± 0100 ± 0100 ± 099.97 ± 0.04100.00 ± 0100.00 ± 0100.00 ± 0
C12100 ± 04.51 ± 7.82100 ± 0100 ± 098.57 ± 1.5663.51 ± 54.59100 ± 0100 ± 099.81 ± 0.4299.86 ± 0.32100.00 ± 0.42
OA (%)61.99 ± 1.4759.94 ± 1.0062.88 ± 1.4455.16 ± 2.3458.09 ± 0.8560.52 ± 1.6062.50 ± 1.6462.47 ± 0.7562.03 ± 0.7462.06 ± 1.2262.69 ± 0.95
Kappa52.83 ± 1.7243.01 ± 1.4454.11 ± 1.9144.04 ± 2.7445.99 ± 1.3750.38 ± 1.7554.35 ± 2.1853.33 ± 08652.93 ± 0.9153.49 ± 1.7154.72 ± 1.27
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, L.; Zhao, H.; Gao, S.; Wang, J.; Zhang, Z. Spatial-Spectral Linear Extrapolation for Cross-Scene Hyperspectral Image Classification. Remote Sens. 2025, 17, 1816. https://doi.org/10.3390/rs17111816

AMA Style

Lin L, Zhao H, Gao S, Wang J, Zhang Z. Spatial-Spectral Linear Extrapolation for Cross-Scene Hyperspectral Image Classification. Remote Sensing. 2025; 17(11):1816. https://doi.org/10.3390/rs17111816

Chicago/Turabian Style

Lin, Lianlei, Hanqing Zhao, Sheng Gao, Junkai Wang, and Zongwei Zhang. 2025. "Spatial-Spectral Linear Extrapolation for Cross-Scene Hyperspectral Image Classification" Remote Sensing 17, no. 11: 1816. https://doi.org/10.3390/rs17111816

APA Style

Lin, L., Zhao, H., Gao, S., Wang, J., & Zhang, Z. (2025). Spatial-Spectral Linear Extrapolation for Cross-Scene Hyperspectral Image Classification. Remote Sensing, 17(11), 1816. https://doi.org/10.3390/rs17111816

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop