Next Article in Journal
DRGNet: Enhanced VVC Reconstructed Frames Using Dual-Path Residual Gating for High-Resolution Video
Previous Article in Journal
Regression-Based Docking System for Autonomous Mobile Robots Using a Monocular Camera and ArUco Markers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Statistical Difference Representation-Based Transformer for Heterogeneous Change Detection

1
School of Artificial Intelligence, Guangzhou Huashang College, Guangzhou 511300, China
2
Graduate School, St. Paul University Philippines Tuguegarao, Cagayan 3500, Philippines
3
College of Information Science and Engineering, Guilin University of Technology, Guilin 541004, China
4
School of Mathematics and Computer Science, Yan’an University, Yan’an 716000, China
5
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3740; https://doi.org/10.3390/s25123740 (registering DOI)
Submission received: 2 April 2025 / Revised: 8 June 2025 / Accepted: 9 June 2025 / Published: 15 June 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Heterogeneous change detection refers to using image data from different sensors or modalities to detect change information in the same region by comparing images of the same region at different time periods. In recent years, methods based on deep learning and domain adaptation have become mainstream, which can effectively improve the accuracy and robustness of heterogeneous image change detection through feature alignment and multimodal data fusion. However, a lack of credible labels has stopped most current learning-based heterogeneous change detection methods from being put into application. To overcome this limitation, a weakly supervised heterogeneous change detection framework with a structure similarity-guided sample generating ( S 3 G 2 ) strategy is proposed, which employs differential structure similarity to acquire prior information for iteratively generating reliable pseudo-labels. Moreover, a Statistical Difference representation Transformer (SDFormer) is proposed to lower the influence of modality difference between bitemporal heterogeneous imagery and better extract relevant change information. Extensive experiments have been carried out to fully investigate the influences of inner manual parameters and compare them with state-of-the-art methods in several public heterogeneous change detection data sets. The experimental results indicate that the proposed methods have shown competitive performance.

1. Introduction

Change detection (CD) leverages multi-temporal remote sensing data with image processing and pattern recognition methods to extract change information, quantitatively analyze to determine the characteristics and processes of surface changes, and identify changes in surface objects or phenomena in time by comparing remote sensing images from different periods [1,2,3], which has been applied in urban planning [4], land resource supervision [5], disaster monitoring [6,7], and emergency response [8].
Traditional CD techniques predominantly hinge on the comparative analysis of multi-temporal remote sensing imagery to discern change information, which is achieved with the construction of difference images, feature transformation, and the application of unsupervised or supervised classification methodologies [9,10,11]. While these approaches have been extensively utilized in the realm of low- and medium-resolution imagery, they are fraught with several limitations [12,13]. Notably, they are characterized by complex procedural steps, a low degree of automation, and stringent requirements regarding data quality and scene complexity [3,14]. Furthermore, the efficacy in handling high-resolution and intricate scenes is notably constrained [15]. Conversely, deep learning-based CD techniques leverage sophisticated models such as convolutional neural networks (CNNs) and Transformers to automatically extract intricate, non-linear, and multi-scale features from multi-temporal remote sensing images, thereby achieving enhanced performance in CD [16,17,18]. Utilizing an end-to-end learning framework, these approaches are adept at managing change detection tasks within high-resolution and complex scenarios, thereby markedly elevating the precision and robustness of the detection process [19].
However, the majority of existing research in this domain is predicated on single-modality approaches. Single-modal CD relies on a single type of remote sensing data, which has drawbacks such as data limitations, susceptibility to noise, limited feature extraction capability, restricted applicability, lack of robustness, and difficulty in handling multitasking requirements, which limit its application effectiveness and detection accuracy in complex scenes [20,21,22]. Compared to single-modal CD, Heterogeneous CD (HCD), also known as multimodal CD, overcomes the limitations of single-modality data by integrating multiple remote sensing data sources [23,24]. HCD is particularly effective in handling change detection tasks in complex environments, thereby enhancing the capacity for feature learning and generalization [25]. Moreover, it reduces the impact of noise and error, and is better equipped to meet the demands of complex scenes and multitasking scenarios [14,26]. Consequently, HCD demonstrates substantial application value in remote sensing image CD [27].
Overall, the current challenges in HCD are primarily characterized by the significant intermodal disparities that complicate feature extraction and fusion, the scarcity of labeled data that constrains model training, the difficulty in accurately detecting change information within complex scenes, the need to balance computational resources with efficiency, and the insufficient generalization capability of existing models [28,29]. The contemporary research landscape of HCD is characterized by the increasing adoption of deep learning and domain adaptation techniques, which have facilitated the development of methods centered on feature alignment and multimodal data fusion [30,31]. These advancements have demonstrably enhanced the accuracy and robustness of change detection by mitigating the disparities between different modalities [32]. Despite these progressions, HCD continues to confront substantial challenges. These include the complexities associated with feature extraction and fusion due to significant intermodal differences, the limitations imposed by the scarcity of labeled training data, the intricacies of accurately identifying change information within complex scenes, the need to optimize the balance between computational resources and efficiency, and the persistent issue of insufficient model generalization capability [33,34,35]. To address these multifaceted challenges, researchers are actively engaged in exploring weakly supervised and unsupervised learning paradigms, devising more efficient data fusion strategies, and developing novel approaches to bolster model robustness and generalization ability [21]. These efforts are aimed at driving the continued evolution and refinement of HCD technologies.
However, HCD involves using image data from different sensors or different modalities to detect change information in the same region. However, significant differences between modalities (e.g., optical vs. radar images) can severely affect the effectiveness of feature extraction and fusion, leading to a decrease in the accuracy and robustness of change detection. Completely unsupervised methods can alleviate this problem, but they often fail to meet the diverse scene requirements and are susceptible to the quality of pseudo-labeling. While the existing HCD methods [36,37,38] show obvious limitations when dealing with complex scenes, such as sensitivity to noise, limited feature extraction capability, and insufficient generalization ability.
In this paper, to address the problems of large modal differences, scarcity of labeled data, and insufficient detection accuracy in complex scenarios in HCD, a heterogeneous change detection framework based on weakly supervised learning is proposed, which combines a Structural Similarity-Guided Sample Generation strategy ( S 3 G 2 ) and a Statistical Difference representation Transformer (SDFormer). The framework generates reliable pseudo-labeled samples iteratively to expand the training set, and takes advantage of the Transformer architecture to reduce the influence of modal differences and enhance the extraction of change information, thus significantly improving the accuracy and robustness of heterogeneity change detection while optimizing the utilization efficiency of computational resources. Overall, the main contributions of this paper are as follows:
(1)
The Statistical Difference representation Transformer (SDFormer) is proposed to reduce modal differences and enhance the ability of change information extraction through feature-level statistical analysis.
(2)
A weakly supervised framework combined with structural similarity-guided sample generation strategy ( S 3 G 2 ) is designed to iteratively generate reliable pseudo-labels to expand the training set and improve the model performance.
(3)
A statistical difference tokenization scheme within the Transformer architecture is developed to explicitly mitigate modality discrepancies while leveraging global contextual awareness, enhancing accuracy and robustness in complex HCD scenarios.
The remainder of this article is structured as follows. Section 2 describes the related work. Section 3 introduces the proposed method. Results and discussion are presented in Section 4 and Section 5, respectively. Section 6 concludes this paper.

2. Related Works

Heterogeneous change detection (HCD) refers to the use of data from different modalities (e.g., different sensors, data sources, or types of information) for change analysis [39,40,41]. In HCD, the aim is to improve the accuracy and robustness by combining information from multiple data sources [42]. The features and data of different modalities are usually complementary, so combining multimodal information can help overcome the limitations of a single data source, such as noise, low resolution, and missing information [43,44].
Graphics-based HCD combines computer graphics techniques and multimodal data analysis methods to identify and analyze changes in images, involving geometric modeling, reconstruction, and texture mapping [43,45,46]. Sun et al. [47] proposed an Improved Non-Local Patch-based Graph (INLPG) method for unsupervised HCD, which detects changes by comparing the structure of the image rather than the pixel values, which is robust and applicable to a wide range of change detection scenarios, including CD in homogeneous and heterogeneous remote sensing images. In [48], an Iterative Robust Graph and Markovian co-Segmentation method (IRG-McS) is designed, which exploits the self-similarity of images to construct robust K-nearest-neighbor graphs to represent the structure of each image, and compares these graphs by graph mapping to compute forward and backward difference images to avoid the confusion of heterogeneous data. In addition, IRG-McS [48] fuses the forward and backward difference images for change detection via Markov co-segmentation model and feeds back the change regions into the graph construction process. In [49], a Sparse-Constrained Adaptive Structure Consistency-based method (SCASC) is proposed to represent the structure of the pre-event image by constructing an adaptive similarity map, utilize the structural consistency to map the pre-event image into the domain of the post-event image, and compute the difference image and employ the super-pixel-based Markov Random Field Segmentation model to segment the difference image into changed and unchanged categories. Moreover, Sun et al. [50] proposed an unsupervised image regression method based on structural graphs (called GIR-MRF) to capture local and global structural information by learning the structural graphs of images and using structural consistency to map pre-event images to post-event image domains, introducing global self-expression constraints and local similarity constraints, as well as a Markov Random Field Segmentation model combining the change information and the spatial information, thus realizing heterogeneous CD. Moreover, a Locality-Preserving Energy Model (LPEM) [51] is designed to generate change maps by leveraging modality-invariant topological relationships between super-pixel pairs and enforcing structural consistency in feature space and label continuity in geographic space, without intermediate difference image computation. Han et al. [52] proposed an improved training method with Hierarchical Extreme Learning Machine (HELM), which addresses challenges like noise sensitivity and manual sample selection by introducing automatic training sample extraction, image smoothing for noise reduction, and hierarchical feature learning.
Deep learning-based HCD improves the accuracy and robustness due to the powerful expressive capability that enables it to automatically learn useful features from different modal data and efficiently perform multimodal data fusion [12,14]. Liu et al. [53] proposed an unsupervised Symmetric Convolutional Coupling Network (SCCN) for heterogeneous optical and radar images, which computes a disparity map and enables change detection by transforming the input images into a feature space with more consistent feature representations. In [54], two deep convolutional neural network-based architectures (X-Net and ACE-Net) are proposed. X-Net [54] consists of an architecture of two fully convolutional networks, each of which is separately responsible for mapping data from one domain to another. Specifically, one network maps data from the optical image domain to the Synthetic Aperture Radar (SAR) image domain, and the other performs the opposite mapping. X-Net [54] is trained with weighted translation loss and cyclic consistency loss to ensure that the mapping is as accurate as possible in the invariant region, while not enforcing mapping consistency in the changing region. ACE-Net [54] introduces a latent space as an intermediate representation between the two domains. ACE-Net [54] consists of two autoencoders, each consisting of an encoder and a decoder, which are used to map the data from the original domain to the latent space, and from the latent space back to the original domain. In addition, ACE-Net includes a discriminator for adversarial training to ensure that the code spaces of the two autoencoders in the potential space are aligned. Wu et al. [55] proposed an unsupervised Commonality Autoencoder CD method (CACD), which extracts features by means of a convolutional autoencoder, converts the representation of one heterogeneous image to that of another, and is able to efficiently identify the common features of unchanged regions and generate a difference map for change detection. Chen et al. [21] proposed an unsupervised HCD method, which represents the structural information of a multimodal image by constructing a structural graph and learns the structural information in the graph with a Structural Relation Graph Convolutional AutoEncoder (SR-GCAE) to measure the similarity between local and non-local structural relations. Difference images are generated by calculating the similarity levels of the two structural relations, an adaptive fusion strategy is used to fuse these difference images, and finally, morphological filtering post-processing is used to optimize the detection results. In [56], the content cleansing network is proposed to mitigate pseudo-changes in cross-platform and multi-temporal VHR images by decoupling content-style features with multi-resolution encoding and image restoration. Han et al. [57] designed an unsupervised graph autoencoder for multimodal change detection by compensating structural differences, trained via reconstruction–sparsity–structural losses. Recently, Liu et al. proposed a commonality feature representation learning (CFRL) for HCD [24], and achieved the current state-of-the-art performance. These approaches have advanced the progress of heterogeneous CDs.
In terms of Transformer-based architecture, Liu et al. [58] integrates digital surface model via cross-attention, enforces multitask consistency between semantic changes and height-derived pseudo-changes with soft-thresholding constraints. Moreover, Zou et al. [59] proposed an improved multi-scale and spectral-wise Transformer for hyperspectral image change detection, integrating a multi-scale feature fusion module, locality self-attention to mitigate spectral embedding challenges. In [60], a novel two-stream attention-in-attention vision architecture is proposed, called SimPool, which integrates SimPool and ResMLP.
While existing HCD methods have advanced capabilities, critical limitations remain unaddressed. Graph-based approaches such as IRG-McS [48] and GIR-MRF [50], while effective in capturing structural consistency, rely on handcrafted similarity metrics that struggle to adapt to non-linear feature interactions inherent in complex multimodal data. Deep learning methods like CACD [55], AEKAN [23], and CFRL [24] demonstrate enhanced modality-independent feature learning capabilities between heterogeneous features, but are still severely limited by the limitation that the lack of reliable supervision information leads to unstable learning of the methods. Though Transformer-based architectures [58] show promise in modeling global contextual relationships, their pseudo-label generation strategies lack structural guidance, leading to error propagation during iterative training, while their feature fusion mechanisms insufficiently address statistical distribution discrepancies between heterogeneous modalities. This analysis reveals three interconnected gaps: the absence of frameworks that jointly address feature-level statistical divergence and annotation scarcity, the underutilization of differential structural consistency as prior knowledge in pseudo-label refinement, and the limited exploitation of Transformers for cross-modal statistical alignment. These insights directly inform our dual innovations—the S 3 G 2 strategy leverages differential structural similarity to iteratively refine reliable pseudo-labels under sparse supervision, while SDFormer’s statistical difference tokenization enables Transformer-driven harmonization of cross-modal representations, thereby simultaneously resolving modality gaps, label scarcity, and complex scene challenges through unified statistical–geometric learning.

3. Methodology

3.1. Overview

Benefiting from the outstanding ability of change information extraction, deep learning-based techniques and methods have been widely applied in change detection tasks [16]. Based on sufficient supervised learning, these data-driven deep neural networks can effectively identify and annotate changed areas according to the need of change detection applications [61]. However, these supervised deep learning CD methods usually rely on massive labeled data to acquire acceptable CD ability. To overcome this limitation, unsupervised deep learning-based frameworks are invented to face the situations with limited training samples. In order to overcome this limitation, unsupervised deep learning frameworks are proposed to deal with the limited training samples [62], which is often the case in heterogeneous image change detection applications.
However, totally unsupervised frameworks sometimes cannot fulfill varied CD requirements and image modalities in HCD applications, since most of them possess fixed strategies to generate pseudo-training-samples or directly detect change information. Based on these facts, a weakly-supervised HCD framework with a Statistical Difference representation Transformer (SDFormer) is proposed, which utilizes a limited number of labeled data and differential structure similarity to iteratively generate training samples and deal with varied HCD scenes with acceptable accuracy. The proposed heterogeneous image change detection framework can be demonstrated as two main stages, initialization and iteration, as shown in Figure 1.

3.2. Initialization

In this stage, one percentage of pixels from current bitemporal heterogeneous imagery are randomly selected and labeled for the first epoch of training, which build initial HCD ability for the proposed SDFormer, thus facilitating the following structure similarity-guided sample generating ( S 3 G 2 ) strategy in the iteration stage. In the practical application, initial training samples, i.e., generation 0 sample set, can be manually labeled to fit varied change detection requirements. After the initial training samples gained, they are utilized to train the proposed SDFormer for several rounds until the training loss stops dropping. Considering that heterogeneous imagery usually has varied patterns and modalities, the differential information is directly used to extract structure similarity for pseudo-sample generating. To achieve this goal, the Euclidean distance is firstly used to directly measure pixel-wise change intensity between bitemporal heterogeneous images as shown below:
D I i , j = T 1 i , j T 2 i , j
where D I i , j , T 1 i , j , and T 2 i , j indicate each pixel of difference intensity map and corresponding bitemporal heterogeneous imagery, respectively, and · is the vector-wise Euclidean distance. Afterwards, the Simple Linear Iterative Clustering (SLIC) is employed to acquire the super-pixel map according to the difference intensity map, which uses amorphous homogeneous regions to describe the spatial areas sharing similar differential pattern, thus extracting differential structure similarity. To sum up, at the end of initialization stage, a super-pixel map carrying differential structure similarity, an initially trained SDFormer, and corresponding raw prediction are acquired, which will be utilized to expand the scale of sample set for better change detection performance in the following iteration stage.

3.3. Iteration

At the iteration stage, the proposed structure similarity-guided sample generating strategy is firstly applied to generate pseudo-samples based on the similarity of differential patterns and the existing initial samples with ground truth, in order to expand the size of the training set. Subsequently, the enlarged sample set with both genuine and pseudo-labels will be used to train the proposed SDFormer, resulting in better HCD performance. After each iteration, as the number of samples increases, the network performance will also improve accordingly, if there is a rational sample generating strategy.

Structure Similarity-Guided Sample Generating

A large number of weakly supervised and unsupervised change detection methods rely on reliable iterative pseudo-label-generating strategies to gain stable performance improvement, since poison pseudo-label will downgrade the model performance after training. To generate credible pseudo-samples, the super-pixels is firstly utilized to describe similar change patterns, since similar change categories lead to consistent difference intensities, which can be captured by SLIC algorithm and clustered into super-pixels. Combining the super-pixels with ground truth in the initial samples, the proposed S 3 G 2 strategy can be informed with prior change information. That is, by identifying which super-pixel each sample belongs to, it can be preliminarily determined that the pixels contained in that super pixel are highly likely to have the same label as the sample. However, considering that super pixels cannot be clustered completely correct, the difference intensity map brought by the raw prediction of SDFormer is employed to find and annotate the pixel with the highest credibility within the super-pixel. After each initial sample point is queried and processed, the size of the sample set will double, thus helping the proposed SDFormer reach higher performance (see Figure 2).

3.4. Statistical Difference Representation Transformer

Transformer models have been widely used in remote sensing fields [15]. Compared to conventional CNN models, Transformers usually extract features in a different way with richer context and global information, which can better fit the HCD requirements, since this can lower the influence brought by difference in modality, thus improving the recognition for relevant changes in bitemporal heterogeneous imagery [63]. In the proposed SDFormer, the influence of modality differences is further reduced by leveraging feature-level statistical analysis, capitalizing on the advantages of Transformers, as shown in Figure 3.
The proposed SDFormer can be divided into four different parts, i.e., a CNN encoder, a statistical difference tokenizer, a Transformer encoder and decoder, and a Multi-Layer Perceptron (MLP), for better introduction. Firstly, bitemporal heterogeneous image patches with the spatial size of 15 × 15 are input into a Siamese CNN encoder with one convolutional layer with the kernel size of 7 × 7 and two convolutional layers with the kernel size of 3 × 3, which project heterogeneous imagery to a common feature space for better change extraction. All these convolutional layers have batch normalization layers and rectified linear units to better extract bitemporal features. This process is demonstrated as follows:
f 1 = Encoder CNN i 1
f 2 = Encoder CNN i 2
where f 1 R B × C × H × W , f 2 R B × C × H × W , and i 1 , i 2 represent bitemporal heterogeneous feature maps and image patches, respectively, and B, C, H, and W indicate the batch size, channel size, height, and width of feature maps, respectively. To better enhance valid change information and lower the influence of modality gap between bitemporal heterogeneous features, the feature-level statistical analysis is employed, the statistical difference tokenizer, to represent and tokenize the bitemporal features. More specifically, channel-wise maximum, minimum, average, and standard deviation are acquired as bitemporal heterogeneous statistical tokens. Then, the statistical difference token can be obtained by subtracting the bitemporal statistical tokens. The detailed process is represented as follows:
token 1 = max f 1 , min f 1 , avg f 1 , std f 1
token 2 = max f 2 , min f 2 , avg f 2 , std f 2
token d = token 1 token 2
where · and · are channel-wise concatenation and absolute value generator, respectively, and t o k e n d R B × C × 4 is the statistical difference token with less modality information and more valid change information.
Based on the success of Transformer models, the Transformer encoder and decoder with cross-attention mechanisms is employed to better detect valid changes between bitemporal heterogeneous imagery, which shares a similar structure with [15]. Different from the conventional bitemporal Transformers, statistical difference information is utilized to directly enhance and refine feature-level differential information rather than bitemporal features, which further lower the influence of modality difference, as shown below:
f d = Decoder transfomer f 1 f 2 , Encoder transfomer token d
Finally, refined differential feature f d R B × C × H × W is processed by MLP to predict the annotation of current central pixel from the input bitemporal heterogeneous image patches. To sum up, the differential statistical information is leveraged in the proposed SDFormer to better represent valid change information and further get rid of the interference of modality difference between bitemporal heterogeneous imagery, thus improving the performance.

4. Experiments and Results

In this section, our experiments and results are presented on three public HCD data sets. First, experimental data sets are described in detail. Second, the comparison methods and evaluation metrics selected are provided in this experiment. Then, the deployment details of the proposed method and other comparison methods are given. Finally, the experimental results of the proposed approach with other comparison methods are compared and analyzed to verify its effectiveness and superiority.

4.1. Data Set Descriptions

Data Set #1: As shown in Figure 4, data set #1 consists of a T1-time image from Landsat-5 and a T2-time from Google Earth. They were captured in Sardinia, Italy in September 1995 and July 1996, with sizes of 300 × 412 × 1 and 300 × 412 × 3 , respectively. Their spatial resolution is 30 m. The change events described in this data are mainly lake expansion. The difficulty of this data set is that the spectral differences and modality differences are large, and it is challenging to detect fine change details.
Data Set #2: The T1-time and T2-time images of data set #2 are composed of SAR and optical images acquired from Radarsat-2 and Google Earth, respectively, as presented in Figure 5. They were captured in June 2008 and September 2010 at the Yellow River, China respectively, with both sizes of 343 × 291 × 1 . Their spatial resolution is 8 m. The change events depicted in this data are mainly embankment. The main challenge of this data set is the huge modal difference between optical images and SAR images, which makes it difficult to accurately detect changes because SAR images contain more coherent speckle noise than optical images.
Data Set #3: As given in Figure 6, data set #3 consists of a T1-time image from Spot and a T2-time from NDVI. They were captured in 1999 and 2000 at Gloucester, UK, with sizes of 900 × 554 × 3 and 900 × 554 × 1 , respectively. Their spatial resolution is about 25 m. Flooding is the main change event described in this data. The difficulty of this data set is similar to that of data set #1. There are huge spectral and modal differences, making it difficult to effectively identify changes through comparison.

4.2. Comparative Approaches and Evaluation Indicators

4.2.1. Comparative Approaches

In the experiment, the popular and advanced traditional graph analysis methods are selected to compare with the proposed SDFormer method, including IRG-McS [48], SCASC [49], and GIR-MRF [50]. These methods are one of the current mainstream HCD methods and have achieved excellent detection performance on related tasks. Therefore, several convincing traditional graph analysis methods are selected for comparison. In addition, popular and latest deep learning-based methods are also selected to compare with the proposed method, including X-Net [54], AEC-Net [54], CACD [55], SR-GCAE [21], AEKAN [23], and CFRL [24]. These methods are representative methods among HCD methods based on deep learning. It is worth noting that AEKAN [23] and CFRL [24] are the latest methods and have achieved state-of-the-art (SOTA) performance on multiple data sets. Based on the superiority of these methods, it is meaningful to select these methods to compare with the proposed method.

4.2.2. Evaluation Indicators

In our experiments, the popular evaluation metrics in HCD is adopted to calculate the quantitative accuracy of each method, namely, Overall Accuracy (OA), Kappa Coefficient (KC), and F1-Score (F1). Based on the binary confusion matrix, these metrics can be calculated as follows:
OA = TP + TN TP + TN + FP + FN ,
KC = 2 × PRE × REC PRE + REC ,
where
PRE = TP TP + FP ,
and
REC = TP TP + FN .
KC = OA p e 1 p e ,
where
p e = ( TP + FN ) × ( TP + FP ) + ( FP + TN ) × ( FN + TN ) ( TP + TN + FP + FN ) 2 ,
where TP, TN, FP, and FN represent the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) in the binary confusion matrix. These three evaluation indicators can comprehensively evaluate each approach.

4.3. Implementation Details

In the experiments, IRG-McS [48], SCASC [49], and GIR-MRF [50] followed the original settings of the paper. X-Net [54] and ACE-Net [54] could not be deployed due to low code versions, so the results of some comparison methods are obtained from other papers. For CACD [55], except for data set #1, the redeployed code is used to obtain the results for the rest of the data. In addition, SR-GCAE [21], AEKAN [23], and CFRL [24] were deployed using the original code and the best results were obtained through trial and error.
In the proposed SDFormer, the proposed method is deployed in the PyTorch 1.8 deep learning framework and trained using a NVIDIA 3090 Graphics Card (ASUS, Taiwan, China). Specifically, the Adam is used as the optimizer, the learning rate is set to 0.0001, and the weight decay is set to 0.00001. The batch size of the proposed method is set to 128. Notably, in the proposed framework, the iterative process terminates after 50 epochs for all experiments. In addition, in the comparative experiments, our parameter settings on the three data sets are set as follows: The number of super-pixels is 1000, the patch size is 15, and the ratio of training samples is 1%. Based on the above settings, the comparative experiments are constructed to verify the effectiveness and superiority of the proposed method. Notably, the number of parameters and computational cost of the proposed SDFormer are 3.3 M and 101.26 M FLOPs, respectively. The proposed method requires fewer computational resources, and no comparison of computational resources is provided in the comparison methods. Therefore, this experiment does not compare and discuss the computational cost of different methods.

4.4. Comparison of DIs with Different Methods

In order to verify the quality of the difference images (DIs) detected by different approaches, the Receiver Operating Characteristic (ROC) curve and Precision–Recall (PR) curve are first selected to compare and analyze the selected methods. Meanwhile, we also quantitatively compare the area under the ROC curve (AUR) and the area under the PR curve (AUP) of different methods to verify the advantages of the proposed approach. The experimental results and analysis on the three data sets are as follows.

4.4.1. Comparison Based on ROC

Figure 7 shows the ROC curves of the proposed SDFormer and all compared methods on three data sets #1–#3. Overall, the proposed SDFormer method based on structural similarity-guided sample generation presents better DI performance. For data set #1, the curve of the proposed SDFormer is closer to the upper left corner, as shown in Figure 7a. The second closest to the upper left corner is CACD [55], and the gap with the proposed SDFormer is very significant. This shows that the DI of the proposed approach shows better separability; that is, it can achieve higher detection accuracy with fewer false detections. Similarly, for data set #2, the proposed SDFormer and CACD [55] obtained the best and second-best results, respectively. The difference is that the proposed SDFormer obtained a slightly better performance than AEKAN [23] on data set #3; that is, the degree of proximity of the two to the upper left corner is almost the same. In addition, as shown in Table 1, the area under the ROC curve (AUR) of the proposed SDFormer is significantly higher than that of other compared methods. The above results and analysis present that the proposed SDFormer has achieved competitive performance on all three data sets.

4.4.2. Comparison Based on PR

In addition to the comparison of ROC curves, the comparison of PR curves is depicted in Figure 8. For data set #1, the proposed SDFormer is closer to the upper right corner than all other compared methods. Compared with the proposed approach, CACD [55] and AEKAN [23] are significantly inferior. For data sets #2 and #3, the proposed SDFormer also achieved the best performance. IRG-McS [48] and AEKAN [23] achieve the second-best results on data sets #2 and #3, respectively. As presented in Table 1, the area under the PR curve (AUP) of the proposed SDFormer performs better than other methods. This is because the proposed SDFormer can represent the changes between heterogeneous images through statistical difference feature learning. At the same time, the proposed method uses the structural similarity sample generation strategy to effectively obtain reliable labeled samples to train our SDFormer. This shows that our SDFormer is more likely to obtain accurate and comprehensive detections than other methods, proving the reliability of our method in obtaining binary detections.

4.5. Comparison of BCIs with Different Methods

In this section, the binary change images (BCIs) of different comparison methods are analyzed on three data sets #1–#3 from the perspectives of quantitative and visual results to demonstrate the effectiveness and superiority of the proposed SDFormer. The results of each data are analyzed as follows.

4.5.1. Results on Data Set #1

The BCIs of the proposed SDFormer and seven compared methods are obtained. The comparisons of quantitative accuracy and visual BCIs are shown in Table 2 and Figure 9. From the quantitative comparison, the proposed SDFormer achieves the best accuracy, and CFRL [24] and CACD [55] achieve the second-best and third-best accuracy. For example, for CFRL [24], the proposed SDFormer achieves 0.005, 0.039, and 0.036 improvements in OA, KC, and F1, respectively. Compared with CACD [55], our SDFormer improves OA, KC, and F1 by 0.004, 0.048, and 0.045, respectively. Compared with the IRG-McS [48], our proposed SDFormer boosts by 0.008, 0.085, and 0.081 in terms of OA, KC, and F1, respectively; The reason is that the proposed method proposes a Transformer network based on statistical differences to learn the differences between heterogeneous images to relatively alleviate the impact of modality differences. At the same time, a reliable training sample is obtained by guiding the sample generation strategy through structural similarity to supervise the proposed SDFormer. This is conducive to improving the detection accuracy of HCD in a supervised manner. In addition to quantitative comparison, the visual results show the advantages of the proposed method more clearly. From a visual point of view, except for the proposed SDFormer, CFRL [24], and CACD [55], other methods demonstrate more false detections and missed detections. The proposed approach and CFRL [24] obtain the results closest to the ground reference truth, especially the proposed method, which has the least false detections and missed detections. Therefore, visual BCIs further prove the superiority of the proposed approach.

4.5.2. Results on Data Set #2

Table 3 and Figure 10 display the quantitative accuracy and visual results of the proposed approach and the comparison methods on data set #2. From the quantitative accuracy comparison results, the proposed SDFormer and CFRL [24] achieve the best accuracy in three indicators, while IRG-McS [48] and GIR-MRF [50] acquire the second-best performance. Moreover, the latest AEKAN [23] achieve the third-best accuracy. Specifically, the proposed SDFormer slightly outperforms CFRL+FLICM [24] in terms of F1 indicator. The proposed method improves the IRG-McS [48] and GIR-MRF [50] by 0.001, 0.021, and 0.021 in terms of OA, KC, and F1, respectively. Compared with the AEKAN [23], the OA, KC, and F1 of the proposed method are improved by 0.005, 0.127, and 0.125, respectively. This also illustrates the advantages of the proposed approach; that is, the proposed SDFormer can perceive the changes between heterogeneous images by intuitively counting the difference information, and the samples obtained based on structural similarity to train SDFormer to better guide the proposed method to achieve effective HCD. Similarly, the visual results comparison also presents the same conclusion. The proposed SDFormer is closer to the ground truth reference than other compared methods and shows fewer false and missed pixels.

4.5.3. Results on Data Set #3

Table 4 and Figure 11 present the quantitative accuracy and visual results of the proposed SDFormer and the comparison approaches on data set #3. The accuracy comparison in Table 4 demonstrates that the proposed SDFormer reaches the best accuracy in all three evaluation indicators, while AEKAN [23] and CFRL+FLICM [24] achieve the second-best and third-best performances. Concretely, compared with the second-best AEKAN [23], the proposed SDFormer obtains improvements of 0.010, 0.042, and 0.035 in OA, KC, and F1, respectively. The proposed SDFormer exceeds the third-best CFRL+FLICM [24] by 0.011, 0.044, and 0.037 in OA, KC, and F1, respectively. The visual result comparison in Figure 11 also presents that the proposed method obtains more accurate and complete detection results. Overall, the experimental results of data set #3 are the same as the results of data sets #1 and #2. It once again proves the effectiveness and superiority of our proposed SDFormer.

5. Discussion

In this section, the ablation studies and parameters analysis are performed to further explore the effectiveness of the proposed SDFormer. First, the ablation experiments are conducted on S 3 G 2 and SDFormer in the proposed method to verify the effectiveness of each part. Second, a series of parameter sensitivity analysis experiments are conducted, including training sample ratio, patch size, and number of super-pixels to analyze the performance of our approach under different parameter settings. The details are as follows.

5.1. Ablation Study for Different Components

In this subsection, in order to test the effectiveness of S 3 G 2 and SDFormer proposed in this study, these two components are separated from the proposed SDFormer for experiments, respectively. Based on the above settings, as shown in Table 5 and Figure 12, the quantitative and comparative results of ablation experiments on data sets #1–#3 are obtained. From Table 5, it can be found that the performance of the proposed SDFormer is very limited when the proposed S 3 G 2 or SDFormer is not added, while the performance of the proposed method is significantly improved when these two components are deployed simultaneously, thus achieving the best performance. Specifically, the average accuracy of our method on the three data sets reaches 0.980, 0.837, and 0.848 in terms of OA, KC, and F1, respectively. The average accuracy on the three data sets is reduced by 0.029, 0.152, and 0.138 in terms of OA, KC, and F1, respectively, when S 3 G 2 is not deployed. Similarly, the average performance on the three data sets is reduced by 0.014, 0.066, and 0.058 in terms of OA, KC, and F1, respectively, when SDFormer is not adopted. Figure 12 presents the impact of different components in the proposed SDFormer on the performance more clearly and intuitively. In general, the above experiments demonstrate the effectiveness of the proposed S 3 G 2 and SDFormer.

5.2. Sensitivity Analysis of Parameters

In this subsection, the three parameters involved in the proposed SDFormer are analyzed to test the impact of these parameters on the proposed approach. Specifically, the proposed method includes three parameters, namely, the training sample ratio, the patch size, and the number of super-pixels. Among them, the ratio of training samples determines the degree of dependence of the proposed method on the sample; the patch size is the analysis unit of our network—a larger patch can provide more information, while a smaller patch may not provide enough information; in the structural similarity-guided sample generation strategy, the number of super-pixels determines the scale of segmentation, which has an important impact on the quality of sample generation. To this end, the parameter analysis experiments on data set #3 are conducted as follows.

5.2.1. Sensitivity Analysis of Ratio of Training Samples

In order to test the impact of different training ratios on the proposed SDFormer, four different ratios are set for testing, namely: 0.25%, 0.5%, 0.75%, and 1%. Based on this setting, the performance of the proposed SDFormer is acquired under different training sample ratios, and the relationship between them is shown in Figure 13a. Figure 13a shows that as the training sample ratio increases, the Overall Accuracy also shows an upward trend; until the training ratio reaches 1%, OA, KC, and F1 all reach the highest. This illustrates that more training samples are more conducive to improving the detection performance of the proposed SDFormer. In addition, the experiment also shows that the proposed method can still achieve good detection performance when relying on only 0.25% of training samples.

5.2.2. Sensitivity Analysis of Patch Size

The patch size is set to [11, 13, 15, 17, 19] to test the impact of patch size on the detection performance of the proposed SDFormer. The relationship curve between different patch sizes and the three accuracies of OA, KC, and F1 is shown in Figure 13b. We can find that as the patch size gradually increases from 11 to 19, the detection accuracy gradually enhances until it reaches the highest when the patch size is 15, then decreases and then increases. The reason is that a patch that is too large may introduce other objects and introduce inaccurate statistical difference information, resulting in poor accuracy, while a patch that is too small may limit the detection performance due to insufficient information. Therefore, in our proposed SDFormer, choosing an appropriate size is more beneficial to obtain better results in practical application.

5.2.3. Sensitivity Analysis of Number of Super-Pixels

In our proposed SDFormer, a parameter for the number of super-pixels needs to be set, which aims to use structural similarity based on super-pixels to guide the sample generation to select reliable training samples. Reliable training samples are crucial to the proposed approach. In fact, the number of super-pixels determines the scale of super-pixel segmentation. A large number of super-pixels indicates a smaller segmentation scale, while a small number of super-pixels indicates a larger segmentation scale. Therefore, to test the effect of the number of super-pixels on the performance of the proposed SDFormer, the number of super-pixels is set to [250, 500, 750, 1000] to perform sensitivity analysis of this parameter. The experimental results are exhibited in Figure 13c. With the increase in the number of super-pixels, the proposed SDFormer has different degrees of improvement in the three indicators of OA, KC, and F1. This demonstrates that the proposed method has a good performance at different super-pixel scales, and it also presents that our proposed structural similarity-guided sample generation strategy can still obtain reliable training samples when the super-pixel scale is small. Therefore, a relatively moderate segmentation scale is selected to obtain a better performance in practical applications.

6. Conclusions

In this paper, a SDFormer with weakly supervised sample generating is proposed for HCD to overcome the limitation due to the lack of reliable labels. In the proposed approach, a structure similarity-guided sample generating strategy is devised, which can acquire reliable pseudo-labels by employing prior information based on differential structure similarity in an iterative generation manner. Moreover, a Statistical Difference representation Transformer is constructed to alleviate the effect of modality differences between heterogeneous images, thereby improving the detection performance in heterogeneous images. Extensive comparative experiments on the three HCD data sets demonstrate the superiority of the proposed approach. Furthermore, sufficient ablation studies and parameter sensitivity analysis experiments validate the effectiveness and reliability of the proposed SDFormer. Nevertheless, our study still has the following two limitations that need further research and improvement. Although the current method can achieve good detection results by relying on less than 1% of samples through weak supervision, its applicability to larger and more complex scenes still needs further testing and exploration.
While the current framework effectively generates pseudo-labels via structural similarity and reduces modality gaps with SDFormer, limitations persist in noisy environments, computational efficiency, and generalization across diverse sensor combinations. Future efforts could focus on integrating uncertainty quantification to enhance pseudo-label robustness, designing lightweight Transformers or dynamic sparse attention for edge deployment, and advancing cross-modal adaptation techniques to improve scalability for multi-source heterogeneous data under limited supervision.

Author Contributions

Conceptualization, X.C., M.D., and H.Z.; methodology, X.C., J.G., and H.Z.; validation, X.C., X.L., and J.G.; investigation, M.D., X.L., and J.G.; writing—original draft preparation, X.C.; writing—review and editing, X.C., M.D., X.L., J.G., and H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is funded by the Guangzhou Huashang University School-Level Project (No. 2024HSDS26).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xu, X.; Zhou, Y.; Lu, X.; Chen, Z. FERA-net: A building change detection method for high-resolution remote sensing imagery based on residual attention and high-frequency features. Remote Sens. 2023, 15, 395. [Google Scholar] [CrossRef]
  2. Lv, Z.; Liu, T.; Benediktsson, J.A.; Falco, N. Land cover change detection techniques: Very-high-resolution optical images: A review. IEEE Geosci. Remote Sens. Mag. 2021, 10, 44–63. [Google Scholar] [CrossRef]
  3. Zheng, H.; Gong, M.; Liu, T.; Jiang, F.; Zhan, T.; Lu, D.; Zhang, M. HFA-Net: High frequency attention siamese network for building change detection in VHR remote sensing images. Pattern Recognit. 2022, 129, 108717. [Google Scholar] [CrossRef]
  4. Chen, Z.; Zhou, Y.; Wang, B.; Xu, X.; He, N.; Jin, S.; Jin, S. EGDE-Net: A building change detection method for high-resolution remote sensing imagery based on edge guidance and differential enhancement. ISPRS J. Photogramm. Remote Sens. 2022, 191, 203–222. [Google Scholar] [CrossRef]
  5. Li, J.; Gong, M.; Liu, H.; Zhang, Y.; Zhang, M.; Wu, Y. Multiform ensemble self-supervised learning for few-shot remote sensing scene classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4500416. [Google Scholar] [CrossRef]
  6. Huang, X.; Cao, Y.; Li, J. An automatic change detection method for monitoring newly constructed building areas using time-series multi-view high-resolution optical satellite images. Remote Sens. Environ. 2020, 244, 111802. [Google Scholar] [CrossRef]
  7. Liu, T.; Gong, M.; Jiang, F.; Zhang, Y.; Li, H. Landslide inventory mapping method based on adaptive histogram-mean distance with bitemporal VHR aerial images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 3003005. [Google Scholar] [CrossRef]
  8. Saha, S.; Bovolo, F.; Bruzzone, L. Building change detection in VHR SAR images via unsupervised deep transcoding. IEEE Trans. Geosci. Remote Sens. 2020, 59, 1917–1929. [Google Scholar] [CrossRef]
  9. Lu, D.; Mausel, P.; Brondizio, E.; Moran, E. Change detection techniques. Int. J. Remote Sens. 2004, 25, 2365–2401. [Google Scholar] [CrossRef]
  10. Shi, W.; Zhang, M.; Zhang, R.; Chen, S.; Zhan, Z. Change detection based on artificial intelligence: State-of-the-art and challenges. Remote Sens. 2020, 12, 1688. [Google Scholar] [CrossRef]
  11. Walter, V. Object-based classification of remote sensing data for change detection. ISPRS J. Photogramm. Remote Sens. 2004, 58, 225–238. [Google Scholar] [CrossRef]
  12. Daudt, R.C.; Le Saux, B.; Boulch, A. Fully convolutional siamese networks for change detection. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 4063–4067. [Google Scholar]
  13. Li, J.; Li, H.; Liu, Y.; Gong, M. Multi-fidelity evolutionary multitasking optimization for hyperspectral endmember extraction. Appl. Soft Comput. 2021, 111, 107713. [Google Scholar] [CrossRef]
  14. Liu, T.; Gong, M.; Lu, D.; Zhang, Q.; Zheng, H.; Jiang, F.; Zhang, M. Building change detection for VHR remote sensing images via local–global pyramid network and cross-task transfer learning strategy. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4704817. [Google Scholar] [CrossRef]
  15. Chen, H.; Qi, Z.; Shi, Z. Remote sensing image change detection with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4704817. [Google Scholar] [CrossRef]
  16. Shafique, A.; Cao, G.; Khan, Z.; Asad, M.; Aslam, M. Deep learning-based change detection in remote sensing images: A review. Remote Sens. 2022, 14, 871. [Google Scholar] [CrossRef]
  17. Ghaderpour, E.; Pagiatakis, S.D.; Hassan, Q.K. A survey on change detection and time series analysis with applications. Appl. Sci. 2021, 11, 6141. [Google Scholar] [CrossRef]
  18. Li, J.; Gong, M.; Wei, J.; Zhang, Y.; Zhao, Y.; Wang, S.; Jiang, X. Evolutionary multitasking cooperative transfer for multiobjective hyperspectral sparse unmixing. Knowl.-Based Syst. 2024, 285, 111306. [Google Scholar] [CrossRef]
  19. Li, J.; Gong, M.; Li, W.; Zhang, M.; Zhang, Y.; Wang, S.; Wu, Y. MSANet: Multiscale Self-Attention Aggregation Network for Few-Shot Aerial Imagery Segmentation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4500216. [Google Scholar] [CrossRef]
  20. Saidi, S.; Idbraim, S.; Karmoude, Y.; Masse, A.; Arbelo, M. Deep-Learning for Change Detection Using Multi-Modal Fusion of Remote Sensing Images: A Review. Remote Sens. 2024, 16, 3852. [Google Scholar] [CrossRef]
  21. Chen, H.; Yokoya, N.; Wu, C.; Du, B. Unsupervised multimodal change detection based on structural relationship graph representation learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5635318. [Google Scholar] [CrossRef]
  22. Li, J.; Gong, M.; Liu, Z.; Wang, S.; Zhang, Y.; Zhou, Y.; Gao, Y. Towards Multi-Party Personalized Collaborative Learning in Remote Sensing. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4503616. [Google Scholar]
  23. Liu, T.; Xu, J.; Lei, T.; Wang, Y.; Du, X.; Zhang, W.; Lv, Z.; Gong, M. AEKAN: Exploring Superpixel-based AutoEncoder Kolmogorov-Arnold Network for Unsupervised Multimodal Change Detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5601114. [Google Scholar] [CrossRef]
  24. Liu, T.; Zhang, M.; Gong, M.; Zhang, Q.; Jiang, F.; Zheng, H.; Lu, D. Commonality Feature Representation Learning for Unsupervised Multimodal Change Detection. IEEE Trans. Image Process. 2025, 34, 1219–1233. [Google Scholar] [CrossRef] [PubMed]
  25. Pu, Y.; Gong, M.; Liu, T.; Zhang, M.; Gao, T.; Jiang, F.; Hu, X. Adversarial feature equilibrium network for multimodal change detection in heterogeneous remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4512617. [Google Scholar] [CrossRef]
  26. Gong, M.; Liu, T.; Zhang, M.; Zhang, Q.; Lu, D.; Zheng, H.; Jiang, F. Context–content collaborative network for building extraction from high-resolution imagery. Knowl.-Based Syst. 2023, 263, 110283. [Google Scholar] [CrossRef]
  27. Lv, Z.Y.; Liu, T.F.; Zhang, P.; Benediktsson, J.A.; Lei, T.; Zhang, X. Novel adaptive histogram trend similarity approach for land cover change detection by using bitemporal very-high-resolution remote sensing images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9554–9574. [Google Scholar] [CrossRef]
  28. Liu, M.; Lin, S.; Zhong, Y.; Shi, Q.; Li, J. A Memory Guided Network and A Novel Dataset for Cropland Semantic Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4410013. [Google Scholar] [CrossRef]
  29. Lang, C.; Cheng, G.; Wu, J.; Li, Z.; Xie, X.; Li, J.; Han, J. Toward Open-World Remote Sensing Imagery Interpretation: Past, present, and future. IEEE Geosci. Remote Sens. Mag. 2024, 2–38. [Google Scholar] [CrossRef]
  30. Liu, T.; Pu, Y.; Lei, T.; Xu, J.; Gong, M.; He, L.; Nandi, A.K. Hierarchical Feature Alignment-based Progressive Addition Network for Multimodal Change Detection. Pattern Recognit. 2025, 162, 111355. [Google Scholar] [CrossRef]
  31. Chen, H.; Yokoya, N.; Chini, M. Fourier domain structural relationship analysis for unsupervised multimodal change detection. ISPRS J. Photogramm. Remote Sens. 2023, 198, 99–114. [Google Scholar] [CrossRef]
  32. Hu, M.; Wu, C.; Du, B.; Zhang, L. Binary change guided hyperspectral multiclass change detection. IEEE Trans. Image Process. 2023, 32, 791–806. [Google Scholar] [CrossRef] [PubMed]
  33. Hu, M.; Wu, C.; Zhang, L. HyperNet: Self-supervised hyperspectral spatial–spectral feature understanding network for hyperspectral change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5543017. [Google Scholar] [CrossRef]
  34. Radke, R.J.; Andra, S.; Al-Kofahi, O.; Roysam, B. Image change detection algorithms: A systematic survey. IEEE Trans. Image Process. 2005, 14, 294–307. [Google Scholar] [CrossRef] [PubMed]
  35. Lei, T.; Wang, J.; Ning, H.; Wang, X.; Xue, D.; Wang, Q.; Nandi, A.K. Difference enhancement and spatial–spectral nonlocal network for change detection in VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4507013. [Google Scholar] [CrossRef]
  36. Chen, X.; Vierling, L.; Deering, D. A simple and effective radiometric correction method to improve landscape change detection across sensors and across time. Remote Sens. Environ. 2005, 98, 63–79. [Google Scholar] [CrossRef]
  37. Wang, L.; Zhang, J.; Bruzzone, L. MixCDNet: A Lightweight Change Detection Network Mixing Features across CNN and Transformer. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4411915. [Google Scholar] [CrossRef]
  38. Bruzzone, L.; Bovolo, F. A novel framework for the design of change-detection systems for very-high-resolution remote sensing images. Proc. IEEE 2012, 101, 609–630. [Google Scholar] [CrossRef]
  39. Wang, X.; Du, J.; Tan, K.; Ding, J.; Liu, Z.; Pan, C.; Han, B. A high-resolution feature difference attention network for the application of building change detection. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102950. [Google Scholar] [CrossRef]
  40. Gong, M.; Li, J.; Zhang, Y.; Wu, Y.; Zhang, M. Two-path aggregation attention network with quad-patch data augmentation for few-shot scene classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4511616. [Google Scholar] [CrossRef]
  41. Marin, C.; Bovolo, F.; Bruzzone, L. Building change detection in multitemporal very high resolution SAR images. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2664–2682. [Google Scholar] [CrossRef]
  42. Wang, S.; Li, J.; Liu, Z.; Gong, M.; Zhang, Y.; Zhao, Y.; Deng, B.; Zhou, Y. Personalized Multi-Party Few-Shot Learning for Remote Sensing Scene Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4506115. [Google Scholar]
  43. Shuai, W.; Jiang, F.; Zheng, H.; Li, J. MSGATN: A superpixel-based multi-scale Siamese graph attention network for change detection in remote sensing images. Appl. Sci. 2022, 12, 5158. [Google Scholar] [CrossRef]
  44. Liu, Y.; Li, J.; Gong, M.; Liu, H.; Sheng, K.; Zhang, Y.; Tang, Z.; Zhou, Y. Collaborative Self-Supervised Evolution for Few-Shot Remote Sensing Scene Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4509215. [Google Scholar] [CrossRef]
  45. Ding, Q.; Shao, Z.; Huang, X.; Altan, O. DSA-Net: A novel deeply supervised attention-guided network for building change detection in high-resolution remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102591. [Google Scholar] [CrossRef]
  46. Yu, S.; Tao, C.; Zhang, G.; Xuan, Y.; Wang, X. Remote Sensing Image Change Detection Based on Deep Learning: Multi-Level Feature Cross-Fusion with 3D-Convolutional Neural Networks. Appl. Sci. 2024, 14, 6269. [Google Scholar] [CrossRef]
  47. Sun, Y.; Lei, L.; Li, X.; Tan, X.; Kuang, G. Structure consistency-based graph for unsupervised change detection with homogeneous and heterogeneous remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4700221. [Google Scholar] [CrossRef]
  48. Sun, Y.; Lei, L.; Guan, D.; Kuang, G. Iterative robust graph for unsupervised change detection of heterogeneous remote sensing images. IEEE Trans. Image Process. 2021, 30, 6277–6291. [Google Scholar] [CrossRef]
  49. Sun, Y.; Lei, L.; Guan, D.; Li, M.; Kuang, G. Sparse-constrained adaptive structure consistency-based unsupervised image regression for heterogeneous remote-sensing change detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4405814. [Google Scholar] [CrossRef]
  50. Sun, Y.; Lei, L.; Tan, X.; Guan, D.; Wu, J.; Kuang, G. Structured graph based image regression for unsupervised multimodal change detection. ISPRS J. Photogramm. Remote Sens. 2022, 185, 16–31. [Google Scholar] [CrossRef]
  51. Sun, Y.; Lei, L.; Guan, D.; Kuang, G.; Li, Z.; Liu, L. Locality preservation for unsupervised multimodal change detection in remote sensing imagery. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 6955–6969. [Google Scholar] [CrossRef]
  52. Han, T.; Tang, Y.; Yang, X.; Lin, Z.; Zou, B.; Feng, H. Change detection for heterogeneous remote sensing images with improved training of hierarchical extreme learning machine (HELM). Remote Sens. 2021, 13, 4918. [Google Scholar] [CrossRef]
  53. Liu, J.; Gong, M.; Qin, K.; Zhang, P. A deep convolutional coupling network for change detection based on heterogeneous optical and radar images. IEEE Trans. Neural Netw. Learn. Syst. 2016, 29, 545–559. [Google Scholar] [CrossRef] [PubMed]
  54. Luppino, L.T.; Kampffmeyer, M.; Bianchi, F.M.; Moser, G.; Serpico, S.B.; Jenssen, R.; Anfinsen, S.N. Deep image translation with an affinity-based change prior for unsupervised multimodal change detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4700422. [Google Scholar] [CrossRef]
  55. Wu, Y.; Li, J.; Yuan, Y.; Qin, A.K.; Miao, Q.G.; Gong, M.G. Commonality autoencoder: Learning common features for change detection from heterogeneous images. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 4257–4270. [Google Scholar] [CrossRef]
  56. Cheng, M.; He, W.; Li, Z.; Yang, G.; Zhang, H. Harmony in diversity: Content cleansing change detection framework for very-high-resolution remote-sensing images. ISPRS J. Photogramm. Remote Sens. 2024, 218, 1–19. [Google Scholar] [CrossRef]
  57. Han, T.; Tang, Y.; Chen, Y.; Yang, X.; Guo, Y.; Jiang, S. SDC-GAE: Structural Difference Compensation Graph Autoencoder for Unsupervised Multimodal Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5622416. [Google Scholar] [CrossRef]
  58. Liu, B.; Chen, H.; Li, K.; Yang, M.Y. Transformer-based multimodal change detection with multitask consistency constraints. Inf. Fusion 2024, 108, 102358. [Google Scholar] [CrossRef]
  59. Zou, C.; Liang, W.; Liu, L.; Zou, C. Hyperspectral image change detection based on an improved multi-scale and spectral-wise transformer. Int. J. Remote Sens. 2024, 45, 1904–1925. [Google Scholar] [CrossRef]
  60. Roy, S.K.; Jamali, A.; Chanussot, J.; Ghamisi, P.; Ghaderpour, E.; Shahabi, H. SimPoolFormer: A two-stream vision transformer for hyperspectral image classification. Remote Sens. Appl. Soc. Environ. 2025, 37, 101478. [Google Scholar] [CrossRef]
  61. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
  62. Cheng, G.; Huang, Y.; Li, X.; Lyu, S.; Xu, Z.; Zhao, H.; Zhao, Q.; Xiang, S. Change detection methods for remote sensing in the last decade: A comprehensive review. Remote Sens. 2024, 16, 2355. [Google Scholar] [CrossRef]
  63. Lv, Z.; Huang, H.; Li, X.; Zhao, M.; Benediktsson, J.A.; Sun, W.; Falco, N. Land cover change detection with heterogeneous remote sensing images: Review, progress, and perspective. Proc. IEEE 2022, 110, 1976–1991. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed bitemporal heterogeneous imagery change detection framework.
Figure 1. Overview of the proposed bitemporal heterogeneous imagery change detection framework.
Sensors 25 03740 g001
Figure 2. Processing of the proposed structure similarity-guided sample generating. Note: Note: The red dots represent the initial sample points, and the blue dots represent the newly added samples.
Figure 2. Processing of the proposed structure similarity-guided sample generating. Note: Note: The red dots represent the initial sample points, and the blue dots represent the newly added samples.
Sensors 25 03740 g002
Figure 3. Overview of the proposed SDFormer. Note: “−” refers to the subtraction operation.
Figure 3. Overview of the proposed SDFormer. Note: “−” refers to the subtraction operation.
Sensors 25 03740 g003
Figure 4. Data set #1: (a) T1 image; (b) T2 image; (c) ground truth image.
Figure 4. Data set #1: (a) T1 image; (b) T2 image; (c) ground truth image.
Sensors 25 03740 g004
Figure 5. Data set #2: (a) T1 image; (b) T2 image; (c) ground truth image.
Figure 5. Data set #2: (a) T1 image; (b) T2 image; (c) ground truth image.
Sensors 25 03740 g005
Figure 6. Data set #3: (a) T1 image; (b) T2 image; (c) ground truth image.
Figure 6. Data set #3: (a) T1 image; (b) T2 image; (c) ground truth image.
Sensors 25 03740 g006
Figure 7. Comparison of ROC curves of different methods on the data sets #1–#3 for DIs: (a) data set #1, (b) data set #2, (c) data set #3.
Figure 7. Comparison of ROC curves of different methods on the data sets #1–#3 for DIs: (a) data set #1, (b) data set #2, (c) data set #3.
Sensors 25 03740 g007
Figure 8. Comparison of PR curves of different methods on the data sets #1–#3 for DIs: (a) data set #1, (b) data set #2, (c) data set #3.
Figure 8. Comparison of PR curves of different methods on the data sets #1–#3 for DIs: (a) data set #1, (b) data set #2, (c) data set #3.
Sensors 25 03740 g008
Figure 9. Binary change results of different methods on data set #1: (a) IRG-McS [48], (b) SCASC [49], (c) GIR-MRF [50], (d) X-Net [54], (e) ACE-Net [54], (f) CACD [55], (g) SR-GCAE [21], (h) AEKAN [23], (i) CFRL+Otsu [24], (j) CFRL+FLICM [24], (k) Ours, and (l) Ground truth image. Note: White and black represent TP pixels and TN pixels, respectively; sky blue and magenta represent FP and FN pixels, respectively.
Figure 9. Binary change results of different methods on data set #1: (a) IRG-McS [48], (b) SCASC [49], (c) GIR-MRF [50], (d) X-Net [54], (e) ACE-Net [54], (f) CACD [55], (g) SR-GCAE [21], (h) AEKAN [23], (i) CFRL+Otsu [24], (j) CFRL+FLICM [24], (k) Ours, and (l) Ground truth image. Note: White and black represent TP pixels and TN pixels, respectively; sky blue and magenta represent FP and FN pixels, respectively.
Sensors 25 03740 g009
Figure 10. Binary change results of different methods on data set #2: (a) IRG-McS [48], (b) SCASC [49], (c) GIR-MRF [50], (d) X-Net [54], (e) ACE-Net [54], (f) CACD [55], (g) SR-GCAE [21], (h) AEKAN [23], (i) CFRL+Otsu [24], (j) CFRL+FLICM [24], (k) Ours, and (l) Ground truth image. Note: White and black represent TP pixels and TN pixels, respectively; sky blue and magenta represent FP and FN pixels, respectively.
Figure 10. Binary change results of different methods on data set #2: (a) IRG-McS [48], (b) SCASC [49], (c) GIR-MRF [50], (d) X-Net [54], (e) ACE-Net [54], (f) CACD [55], (g) SR-GCAE [21], (h) AEKAN [23], (i) CFRL+Otsu [24], (j) CFRL+FLICM [24], (k) Ours, and (l) Ground truth image. Note: White and black represent TP pixels and TN pixels, respectively; sky blue and magenta represent FP and FN pixels, respectively.
Sensors 25 03740 g010
Figure 11. Binary change results of different methods on data set #3: (a) IRG-McS [48], (b) SCASC [49], (c) GIR-MRF [50], (d) X-Net [54], (e) ACE-Net [54], (f) CACD [55], (g) SR-GCAE [21], (h) AEKAN [23], (i) CFRL+Otsu [24], (j) CFRL+FLICM [24], (k) Ours, and (l) Ground truth image. Note: White and black represent TP pixels and TN pixels, respectively; sky blue and magenta represent FP and FN pixels, respectively.
Figure 11. Binary change results of different methods on data set #3: (a) IRG-McS [48], (b) SCASC [49], (c) GIR-MRF [50], (d) X-Net [54], (e) ACE-Net [54], (f) CACD [55], (g) SR-GCAE [21], (h) AEKAN [23], (i) CFRL+Otsu [24], (j) CFRL+FLICM [24], (k) Ours, and (l) Ground truth image. Note: White and black represent TP pixels and TN pixels, respectively; sky blue and magenta represent FP and FN pixels, respectively.
Sensors 25 03740 g011
Figure 12. Comparison results of ablation experiments of the proposed SDFormer on three data sets: (a) Data set #1; (b) Data set #2; (c) Data set #3.
Figure 12. Comparison results of ablation experiments of the proposed SDFormer on three data sets: (a) Data set #1; (b) Data set #2; (c) Data set #3.
Sensors 25 03740 g012
Figure 13. Parameter analysis results of the proposed SDFormer: (a) relationship curve between training sample ratio and accuracy; (b) relationship curve between patch size and accuracy; (c) relationship curve between the number of super-pixels and accuracy.
Figure 13. Parameter analysis results of the proposed SDFormer: (a) relationship curve between training sample ratio and accuracy; (b) relationship curve between patch size and accuracy; (c) relationship curve between the number of super-pixels and accuracy.
Sensors 25 03740 g013
Table 1. Quantitative comparison of AUR and AUP of DIs generated by different methods on data sets #1 to #3.
Table 1. Quantitative comparison of AUR and AUP of DIs generated by different methods on data sets #1 to #3.
MethodsData Set #1Data Set #2Data Set #3Average
AURAUPAURAUPAURAUPAURAUP
CACD [55]0.9470.7630.9780.7420.9020.4590.9420.655
SR-GCAE [21]0.8880.3390.9610.7190.9570.7950.9350.618
IRG-McS [48]0.8990.6430.9730.7730.9460.7350.9390.717
SCASC [49]0.8860.4560.9420.5940.9380.7580.9220.603
GIR-MRF [50]0.8880.4570.9320.3480.9390.7590.9200.521
AEKAN [23]0.9500.6460.9510.7120.9840.8650.9620.741
CFRL [24]0.9850.8760.9940.8590.9800.7730.9860.836
Proposed SDFormer0.9900.9150.9960.8860.9930.9450.9930.915
Table 2. Evaluation indicator comparison of different methods on data set #1. In bold are the best results.
Table 2. Evaluation indicator comparison of different methods on data set #1. In bold are the best results.
MethodsOAKCF1
IRG-McS [48]0.9710.7390.754
SCASC [49]0.9470.5930.621
GIR-MRF [50]0.9570.6740.697
X-Net [54]0.9180.3400.443
ACE-Net [54]0.9350.5490.582
CACD [55]0.9750.7760.790
SR-GCAE [21]0.9370.5460.579
AEKAN [23]0.9550.6600.684
CFRL+Otsu [24]0.9730.7800.795
CFRL+FLICM [24]0.9740.7850.799
Ours0.9790.8240.835
Table 3. Evaluation indicator comparison of different methods on data set #2. In bold are the best results.
Table 3. Evaluation indicator comparison of different methods on data set #2. In bold are the best results.
MethodsOAKCF1
IRG-McS [48]0.9860.7880.795
SCASC [49]0.9760.6230.636
GIR-MRF [50]0.9860.7880.795
X-Net [54]0.9180.2320.268
ACE-Net [54]0.9280.2970.329
CACD [55]0.9670.6140.630
SR-GCAE [21]0.9810.6940.704
AEKAN [23]0.9820.6820.691
CFRL+Otsu [24]0.9850.7870.795
CFRL+FLICM [24]0.9870.8090.815
Ours0.9870.8090.816
Table 4. Evaluation indicator comparison of different methods on data set #3. In bold are the best results.
Table 4. Evaluation indicator comparison of different methods on data set #3. In bold are the best results.
MethodsOAKCF1
IRG-McS [48]0.9360.7040.740
SCASC [49]0.9500.7760.804
GIR-MRF [50]0.9370.7340.770
X-Net [54]0.9090.6370.688
ACE-Net [54]0.9280.6590.701
CACD [55]0.7980.4170.516
SR-GCAE [21]0.8850.5860.649
AEKAN [23]0.9640.8370.858
CFRL+Otsu [24]0.9600.8220.845
CFRL+FLICM [24]0.9630.8350.856
Ours0.9740.8790.893
Table 5. Quantitative results of ablation experiments of the proposed SDFormer on data sets #1–#3. The best results are shown in bold.
Table 5. Quantitative results of ablation experiments of the proposed SDFormer on data sets #1–#3. The best results are shown in bold.
MethodsData Set #1Data Set #2Data Set #3Average
OAKCF1OAKCF1OAKCF1OAKCF1
w/o S 3 G 2 0.9480.6710.6970.9640.6250.6410.9410.7580.7910.9510.6850.710
w/o SDFormer0.9570.7080.7310.9850.7890.7970.9570.8170.8420.9660.7710.790
Ours0.9790.8240.8350.9870.8090.8160.9740.8790.8930.9800.8370.848
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, X.; Dong, M.; Liu, X.; Gong, J.; Zheng, H. Statistical Difference Representation-Based Transformer for Heterogeneous Change Detection. Sensors 2025, 25, 3740. https://doi.org/10.3390/s25123740

AMA Style

Cao X, Dong M, Liu X, Gong J, Zheng H. Statistical Difference Representation-Based Transformer for Heterogeneous Change Detection. Sensors. 2025; 25(12):3740. https://doi.org/10.3390/s25123740

Chicago/Turabian Style

Cao, Xinhui, Minggang Dong, Xingping Liu, Jiaming Gong, and Hanhong Zheng. 2025. "Statistical Difference Representation-Based Transformer for Heterogeneous Change Detection" Sensors 25, no. 12: 3740. https://doi.org/10.3390/s25123740

APA Style

Cao, X., Dong, M., Liu, X., Gong, J., & Zheng, H. (2025). Statistical Difference Representation-Based Transformer for Heterogeneous Change Detection. Sensors, 25(12), 3740. https://doi.org/10.3390/s25123740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop