Next Article in Journal
Prototype of a Multimodal and Multichannel Electro-Physiological and General-Purpose Signal Capture System: Evaluation in Sleep-Research-like Scenario
Previous Article in Journal
Detection and Localization of Rotor Winding Inter-Turn Short Circuit Fault in DFIG Using Zero-Sequence Current Component Under Variable Operating Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic-Aware Remote Sensing Change Detection with Multi-Scale Cross-Attention

1
College of Design and Engineering, National University of Singapore, Singapore 119077, Singapore
2
School of Information Science and Engineering, Hunan Institute of Science and Technology, Yueyang 414000, China
3
College of Electronics and Information Engineering, Sichuan University, Chengdu 610065, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(9), 2813; https://doi.org/10.3390/s25092813
Submission received: 19 March 2025 / Revised: 25 April 2025 / Accepted: 28 April 2025 / Published: 29 April 2025
(This article belongs to the Section Environmental Sensing)

Abstract

:
Remote sensing image change detection plays a vital role in diverse real-world applications such as urban development monitoring, disaster assessment, and land use analysis. As deep learning strives, Convolutional Neural Networks (CNNs) have shown their effects in image processing applications. There are two problems in old-school change detection techniques: First, the techniques do not fully use the effective information of the global and local features, which causes their semantic comprehension to be less accurate. Second, old-school methods usually simply rely on differences and computation at the pixel level without giving enough attention to the information at the semantic level. To address these problems, we propose a multi-scale cross-attention network (MSCANet) based on a CNN in this paper. First, a multi-scale feature extraction strategy is employed to capture and fuse image information across different spatial resolutions. Second, a cross-attention module is introduced to enhance the model’s ability to comprehend semantic-level changes between bitemporal images. Compared to the existing methods, our approach better integrates spatial and semantic features across scales, leading to more accurate and coherent change detection. Experiments on three public datasets (LEVIR-CD, CDD, and SYSU-CD) demonstrate competitive performance. For example, the model achieves an F1-score of 96.19% and an IoU of 92.67% on the CDD dataset. Additionally, robustness tests with Gaussian noise show that the model maintains high accuracy under input degradation, highlighting its potential for real-world applications. These findings suggest that our MSCANet effectively improves semantic awareness and robustness, offering a promising solution for change detection in complex and noisy remote sensing environments.

1. Introduction

Remote sensing image change detection (RSICD) is an essential technique that researchers use to analyze differences among images of the same area taken at different times, which has been playing a role in environmental monitoring, urban planning, agricultural management, and disaster response. By comparing images taken at different times, change detection can reveal dynamic changes on the ground surface and land use, giving people valuable information to understand environmental changes.
In the early stage of the change detection field, due to the weak points of sensor and satellite technology, the resolution of remote sensing images was low, so the change detection techniques were relatively simple. People first performed basic algebraic operations on image pixel pairs to produce difference maps and then set threshold values to classify image changes. Gillespie et al. [1] used the Euclidean Distance (ED) as the calculation method to obtain the magnitude metric of change detection. Celik [2] proposed a method for multi-temporal remote sensing images, which distinguishes changed pixels from unchanged pixels by dividing the feature vectors of pixel pairs into two clusters. Nielsen et al. used the Multivariate Alteration Detection (MAD) method [3] to extract features from bi-phasic hyperspectral images. Hao et al. [4] proposed a method that applied the fuzzy c-mean clustering algorithm and Markov random field in Change Vector Analysis (CVA) [3,5] for analyzing bi-phasic hyperspectral images. It generates the initial change map and calculates the probability of belonging to specific clusters. At last, it obtains the change detection map. Li et al. [6] proposed a feature selection method for hyperspectral remote sensing images based on the HSIAO framework, which optimizes band selection using the Jeffries–Matusita distance and evolutionary algorithms, significantly improving the efficiency and accuracy of downstream classification tasks.
With the progress in machine learning and its emergence in applications, conventional machine learning methods have been exploited in remote change detection. Touazi et al. [7] regarded changed and unchanged detection as a context-independent binary classification problem and proposed a scheme based on K-Nearest Neighbor (k-NN) to update the change detection decision from the feedforward neural network. Nemmour et al. [8] used a support vector machine (SVM) in the application of mapping urban extensions. Volpi et al. [9] implemented supervised change detection by reading the contextual information with support vector machines, which collected spatial and contextual information from a local texture distribution. Liu et al. [10] proposed difference representation learning with the application of stacked constrained Boltzmann machines for change detection in remote sensing pictures. This method analyzes the different images and detects changes between pixel pairs of multi-temporal remote sensing images.
In recent years, some change detection algorithms have been improved, and deep learning methods are becoming increasingly popular. Hyperspectral images gradually become mainstream data sources. Some classical deep learning frameworks applied to change detection are Convolutional Neural Network (CNN) [11], Long Short-Term Memory (LSTM) [12], Recurrent Neural Network (RNN) [13], etc. Du et al. [14] proposed an Unsupervised Deep Slow Feature Analysis (DSFA) network for multi-temporal hyperspectral image change detection, which effectively utilizes the excellent performance of deep networks in the fields of feature extraction and projection. Li et al. [15] proposed an unsupervised end-to-end framework based on deep learning for hyperspectral image change detection. Wang et al. [16] proposed a Siamese CNN with Spectral–Spatial-Wise Attention (SSA-SiamNet) for hyperspectral image change detection. This method affords more attention to informative channels and locations to improve spectral and spatial features adaptively. To deal with the inability of the existing methods to handle the details of complex objects at different scales in hyperspectral images, Yang et al. [17] proposed a deep multi-scale pyramidal network with spatial–spectral residual attention enhancement for hyperspectral image change detection. The network was employed to effectively utilize the excellent performance of deep networks in feature extraction and projection.
Transformer structure models, with their attention mechanism, have shown their capability of capturing complex features in the field of natural language processing. In recent years, this model has also been introduced into image detection tasks to help improve CNNs in identifying long-range dependencies. The transformer model has a self-attention module, which is effective at memorizing global information when processing an image. This is particularly important for recognizing changing regions in an image. Thus, the transformer model can effectively capture subtle differences between two images, even if these differences are spatially far apart. For example, the Position–Time-Aware Transformer (PT-Former) significantly improves change detection accuracy by modeling the position and time relationships in dual-temporal images. In addition, the transformer model performs better in handling problems such as pseudo-changes and incomplete edges when dealing with multi-temporal remote sensing images. This is because the transformer model can model positional and temporal correlations in dual-temporal RS images. The transformer model shows impressive potential in change detection tasks by aggregating contextual information of ground objects through fusion blocks and reconstructing spatial relationships guided by dual-temporal features. Fang et al. [18] proposed a deformable convolution-enhanced hierarchical transformer model for hyperspectral image classification (HSIC), known as SClusterFormer. This model incorporates a unique clustering attention mechanism that enhances the representation of homogeneous local details in 3D hyperspectral images (HSIs) and 2D morphological data while improving the discrimination of non-local structures. Zhang et al. [19] designed SwinSUNet using a transformer module based on the shift window mechanism. ChangeFormer [20] applies a hierarchical dual transformer encoder to create a model that efficiently acquires multi-scale remote sensing details. BIT-CD [21] utilizes ResNet for feature extraction and enhances the features extracted with an additional transformer module. Wang et al. [22] proposed a framework, UVACD, for the detection of dual-temporal image changes, which combines a CNN and a visual transformer. MSTDSNet-CD [23] introduced deep supervision in a CD network constructed with a CNN and a Swin transformer module for the detection of urban area changes. ICIF-Net [24] extracts semantic features of images of the dual-temporal phase in parallel. This method then feeds these features into a cross-scale fusion module, which provides integrated global details for change detection of remote sensing image pairs.
In various CV tasks like image classification and semantic segmentation, multi-scale feature extraction is able to capture the information of an image at both the local scale and global scale, thus improving the performance and generalization of the model. In recent years, multi-scale feature extraction techniques have been widely used and rapidly developed in several fields. Lei et al. [25] proposed a CNN module known as pyramid pooling. The module consists of three different sizes of convolutional kernels, in which each kernel takes charge of features within a certain scale range; thus, the model can observe objects of different sizes in images. SLDDNet [26] introduces a central aggregation strategy, fusing global semantic information with local detailed features. MSCCA-Net first introduces a bottom-up hierarchical fusion strategy and optimizes the feature fusion process by using the multi-scale cascade cross-attention (MSCCA) mechanism, which redistributes the spatial and semantic information in images. In this way, the accuracy of detecting small targets is significantly improved. Thus, this method outperforms the other existing techniques regarding comprehensive performance.
Although CNNs have achieved remarkable results, the existing methods are still deficient in dealing with the fusion of global and local feature information as well as the mining of deep semantic information. Concerning these issues, this paper proposes an innovative CNN-based multi-scale cross-attention network. The following points summarize our research contributions:
(1)
Multi-scale feature extraction mechanism: We introduce a multi-scale feature extraction mechanism, which achieves an effective fusion of image information among multiple scales; thus, the model’s effectiveness in capturing image details is enhanced.
(2)
Cross-attention module: This module improves the model’s ability to understand semantic information. In this way, the model can better recognize the key change parts of the image.
(3)
Experimental validation: Our experiments are conducted on three public datasets. The results show that the algorithm proposed in this paper outperforms some current representative change detection algorithms in several key performance metrics, which validates the effectiveness of our method.

2. Methods

2.1. Network Structure

The overall structure of MSCA is shown in Figure 1, a cross-attention hierarchical network. The architecture processes dual-temporal remote sensing data with three input channels and generates two-channel change detection maps. A weight-shared two-path encoder progressively downsamples the input bitemporal remote sensing imagery through hierarchical compression. The downsampling process is realized through a patch merging layer. After the feature maps are downsampled to half the size of the original image, they are input into the semantic map generation module in the cross-attention module, which generates low-spatial-resolution semantic maps through nonlinear projection. These semantic maps are computed with the feature images for cross-attention. When the spatial resolution is reduced to 1/4 of the original image, the encoder concatenates the feature maps extracted from its dual-branch part, and the number of channels is modified for dual-branch feature fusion. The fused feature map passes through two encoding stages and then enters the decoding stage, where the change information is gradually recovered by the cross-attention and convolution modules. Finally, the change map is recovered to the same spatial resolution as the input image, and the number of channels is adjusted to two by the convolution module to output the final change map. The working procedure of MSCA initially extracts low-level features from remote-sensing images via the convolution module, followed by downsampling to reduce the spatial resolution and computational burden of the CS module. The semantic map is generated by nonlinear projection and represents a highly discriminative feature matrix. The spatial resolution of the semantic map is adjusted to half of the minimum resolution level. This cuts down the self-attention’s complexity from O(n2) to O(n) yet still enables efficient global information aggregation and precise error refinement. The C i values in Figure 1 represent the channel values of different stage features and semantic maps. By adjusting the value of C i , the model can deal with different channel change scales.

2.2. Semantic Generation Module

The CA module utilizes two matrices with distinct spatial resolutions as input data. Therefore, a semantic generation module is introduced in the CA module to create an initial semantic mapping. This module uses convolutional layers to extract features, followed by initial semantic mapping generation by linear mapping. The two mappings are the starting inputs of the CA module. Figure 2 presents the detailed structure characterizing the semantic generation module.
In this process, the features X f are first assigned to two different subspaces X f and X w . This ensures that the spatial resolution of both subspaces remains consistent with the original features while adjusting the channel count of X w to match the product of the semantic map’s length and width. Next, the semantic mapping of X w is reshaped and processed by the softmax function. The processed X w becomes a weight matrix when operating with the reshaped X f matrix. This step resigns the weights of information in X f , and generates a low-resolution semantic map that facilitates information aggregation.

2.3. Cross-Attention

The main structure of cross-attention is shown in Figure 3, which is based on the transformer structure and built on multi-head attention (MHSA) and Fully Connected Layer (FFN). In vision tasks, in order to comply with the transformer, each feature map input is flattened from X f R H × W × d to X f R n × d . H is the height of the feature map, W is the width of the feature map, and d is the dimension of the feature map, n = H × W . In order to compute the attention of X t , it is projected to the query vector query Q R n × d , the key vector key K R n × d , and the value vector value V R n × d . The self-attention equation is calculated as below:
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d ) V
Here, Q K T denotes the dot product of the transpose of the query matrix Q and the key matrix K. d is a scaling factor to control the scale of the dot product, which prevents the gradient from disappearing before the softmax calculation when the dot product is too large.
Figure 3. Cross-attention network structure.
Figure 3. Cross-attention network structure.
Sensors 25 02813 g003
The multi-head attention mechanism is the core component of the transformer model, which learns information from different representation subspaces. It has multiple attention heads working in parallel to capture dependencies of different positions in the sequence and thus learn information from different representation subspaces to enhance its expression ability. The formula of multi-head attention is as follows:
M H S A ( Q , K , V ) = C o n c a t ( h e a d 1 , , h e a d n ) W O
h e a d i = A t t e n t i o n ( Q W i Q , K W i K , V W i V )
W i Q R d m × d w , W i K R d m × d k , W i V R d m × d v are learnable parameters. d K and d V are the hidden dimensions of the projected subspace. d m is the embedding dimension. In the cross-attention module, the feature extraction and representation capabilities are improved by enhancing the interaction between the feature graph and the semantic graph, thus capturing their cross-module relationships. The idea is to optimize the performance of downstream tasks by focusing on critical regions in the feature graph and semantic graph while suppressing irrelevant information through self-attention and cross-attention mechanisms. The input to the module is the feature map, which is downsampled and input to the semantic generation module to generate semantic vectors. The feature graph generates query vectors and value vectors through depth-separable convolution operation, respectively; the semantic graph generates query vectors and value vectors in the same way. Subsequently, these query vectors and value vectors are rearranged, and the attention weights between the feature graph and the semantic graph are computed using the self-attention mechanism. The process adjusts the attention computation by a scaling factor to ensure the stability of the computation. The specific calculation formula is as follows:
C r o s s A t t e n t i o n ( Q , K ¯ , V ¯ ) = s o f t m a x ( Q K ¯ T d ) V ¯
The resulting attention graph is used for weighted merged feature and semantic graph representations, and the cross-attention scores of the feature and semantic graphs are normalized by the softmax function to generate a weighted feature representation. To further improve the stability of the model, a dropout layer is added to the cross-attention part to avoid overfitting. The output of the feature map is further optimized by the PCM module, which is jump-connected to the feature map weighted by the attention matrix, which extracts more stable and robust features through a series of convolutional layers, batch normalization, and activation functions to further improve the representation of the model. The final output of the feature map is projected to the desired output dimension by generating the corresponding output features through convolutional layers. The whole process, through the combination of the two-way attention mechanism and the positional conditioning module, captures the correlation between regions and improves the model’s performance in tasks such as image segmentation and target detection.

2.4. Loss Function

Change detection (CD) task requires delicate predictions for each pixel. To compute the loss during the training process, we used the weighted cross-entropy (WCE) loss function [27] and the Dice loss [28] function. The cross-entropy (CE) loss function was introduced to address the problem of sample imbalance in the change and non-change domains. We model the predicted change map Y ^ , which is a set containing H × W pixel points, where H and W represent the height and width of the image. These pixel points constitute the predicted change scenarios for subsequent loss calculations.
The WCE loss function can be defined in the following form:
L w c e = 1 H × W i = 1 H × W w e i g h t [ c l a s s ] · log e y ^ i [ c l a s s ] e ( y ^ i [ 0 ] + y ^ i [ 1 ] )
y ^ i [ class ] belongs to 0, 1, representing change and non-change. The Dice coefficient describes the similarity of the set and takes values between [0, 1]. Y denotes the set of pixels of the true label, and Y ^ denotes the set of pixels of the predicted map. The Dice coefficient is calculated as follows:
D i c e C o e f f i c i e n t = 2 | Y Y ^ | | Y | + | Y ^ |
| Y Y ^ | represents the number of pixels that are predicted correctly. Dice coefficient is defined as below:
L d i c e = 1 D i c e C o e f f i c i e n t
The smaller Dice loss is, the more similar the ground truth and predictions are. The loss in total is defined as follows:
L l o s s = L w c e + L d i c e

3. Experiments

3.1. Datasets

The three publicly available datasets used in this work cover diverse regions and resolutions, as detailed in Table 1. This includes urban and rural areas, different years, and both aerial and satellite imagery, making them suitable benchmarks for evaluating the robustness of change detection models.
  • CDD [29]: The CDD dataset, collected from Google Earth and consisting of 16,000 remote sensing image pairs, includes a variety of change scenarios ranging from synthetic images with no or minor relative object movement to real-world scenes exhibiting seasonal variations. The image size is 256 × 256. The dataset is divided into a training set containing 10,000 images and a validation set of 3000 images.
  • LEVIR-CD [30]: This dataset comes from 20 different areas of several cities in Texas, USA, which is a large remote sensing building change detection dataset, including houses, apartments, garages, and other types of buildings. The image size is 1024 × 1024.
  • SYSU-CD [31]: This dataset contains 20,000 pairs of 0.5 m aerial images with a size of 256 × 256, which were taken in Hong Kong between 2007 and 2014. The classes included in the dataset are recently built buildings, suburban expansion, changes in plantings, and road expansion.

3.2. Baselines

We have selected six different deep learning change detection methods for comparison as follows:
STANet [29]: This method captures the spatial–temporal relationship between images through a self-attention mechanism to extract more practical features. The network combines basic and pyramid-structured attention modules to handle changes at different scales.
ChangeFormer [20]: This method proposes a transformer-based dual network for remote sensing image change detection. It extracts multi-scale features through a hierarchical transformer encoder and fuses them through an MLP decoder to predict changes, achieving better performance than traditional CNN-based methods.
BIT-CD [21]: This method converts dual-temporal images into a small number of semantic tokens and models the spatial–temporal context by tokens using a transformer encoder. This method converts the context tokens back into pixel space via a transformer decoder to refine the features. This method improves prediction accuracy while keeping efficiency.
HANet [32]: This method effectively integrates multi-scale features and refines detailed features through a progressive foreground-balanced sampling strategy and a lightweight self-attention mechanism, HAN module. This method adopts a hybrid loss function to successfully alleviate the data imbalance problem, which is validated on two extremely imbalanced datasets.
SNUNet [27]: This method enhances the accuracy of change detection by employing a densely connected Siamese network architecture, which mitigates the loss of localization information in deep networks through compact information transfer between the encoders and decoders, as well as between decoders themselves. Furthermore, it introduces an Ensemble Channel Attention Module (ECAM) for deep supervision. The ECAM allows for the refinement of the most representative features from different semantic levels, which are then utilized for the final classification.
SEIFNet [33]: This method proposes a spatiotemporal enhancement and interlevel fusion network (SEIFNet) for remote sensing image change detection. It enhances the representation of changing objects by capturing global and local information from bitemporal feature maps using a spatiotemporal difference enhancement module (ST-DEM). Additionally, SEIFNet integrates interlevel features through an adaptive context fusion module (ACFM) and refines the change detection results with a progressive decoder. This approach effectively highlights changed regions while suppressing noise interference and improving boundary details.

3.3. Metrics

In order to quantitatively identify the performance of models, we select the metrics that are mainstream in change detection tasks, including precision, recall, OA, IoU, and F1. The formulas for these metrics are shown below:
precision = TP TP + FP
recall = TP TP + FN
OA = TP + TN TP + FN + TN + FP
IoU = TP TP + FP + FN
F 1 = 2 × precision × recall precision + recall

3.4. Comparison Experiments

3.4.1. Results Analysis on CDD Datasets

As shown in Table 2, MSCA performs well in all evaluation metrics. Specifically, MSCA achieves the highest values in precision (96.39), F1-score (96.19), and IoU (92.67). The overall accuracy (OA) also achieves a high value of 99.10, which demonstrates that MSCA reduces the false prediction rate while accurately identifying the change regions. MSCA achieves a better balance between precision and recall than baseline methods such as ChangeFormer and BIT-CD. The superior performance of MSCA is mainly attributed to its advanced multi-scale feature extraction mechanism and cross-attention module. This design enables MSCA to capture global and local contextual information effectively, so MSCA is particularly outstanding in dealing with scenes with complex spatial structures and small changes. The performance of traditional methods, such as HANet and STANet, is inferior, e.g., the F1-score and IoU of STANet are 91.56 and 84.44. The IoU of HANet is even lower, only 79.19. In Figure 4, compared with ground truth, the change detection results output by MSCA are highly consistent in boundary details and overall integrity and can accurately describe the shape and edges of the changed area. In contrast, other models like HANet and STANet have obvious omissions in complex scenes. BIT-CD has incomplete segmentation dealing with complex change regions, although it performs well in some details. Especially in detecting small targets and delicate changes, MSCA, with its multi-scale feature extraction and cross-attention mechanism, shows excellent robustness and generalization ability. The detection results are almost the same as those of ground truth in both small-scale details and large-scale structural changes, further proving the technical advantages of MSCA in remote sensing image change detection.

3.4.2. Results Analysis on LEVIR-CD Datasets

On the LEVIR-CD dataset, MSCA also demonstrates strong detection capability. As shown in Table 3, it achieves an F1-score of 91.02, which is the most outstanding among all experiments. Other metrics are also close to the optimal level, and it has an advantage in overall detection balance. The performance of MSCA indicates its ability to effectively handle complex urban and infrastructure change detection tasks. Figure 5 shows the detection results of the different methods on the LEVIR-CD dataset. HANet and STANet have remarkable errors on some images. In the fourth row of building change detection, HANet and STANet omitted huge areas. STANet is more ambiguous in the boundary processing, and the detection results of some areas are missing. In contrast, MSCA can accurately segment the change region, and its boundary integrity is highly consistent with ground truth.

3.4.3. Results Analysis on SYSU-CD Datasets

The change scenarios are more complex in the SYSU-CD dataset. Nevertheless, MSCA achieves the highest F1-score (80.76) and IoU (67.72), as shown in Table 4. MSCA’s significant performance improvement is due to the effective integration of its cross-attention mechanism. This mechanism effectively alleviates the problem that other methods fall into. For example, STANet tends to ignore changes on a small scale. The F1-score of STANet is only 78.12, and the IoU is only 64.10, which is relatively weak. In Figure 6, the change detection results of the models on the SYSU-CD dataset show large differences. In some complex scenarios (e.g., the second and third rows), MSCA can accurately segment the change regions, whereas the other methods have obvious false positives in these scenarios. The results of HANet especially have more background noise, while STANet has a blurring phenomenon in boundary processing.

3.5. Ablation Study

To validate the effectiveness of the proposed cross-attention (CA) module, we conduct a series of ablation experiments. Specifically, we compare the complete model with two variants: one in which the CA module is replaced by standard convolution layers and another baseline version without both the semantic perception and cross-attention modules.
The quantitative results are presented in Table 5, evaluated on three benchmark datasets: LEVIR-CD, CDD, and SYSU-CD. From the results, it can be observed that the integration of the CA module consistently improves the F1-score and IoU across all datasets. On the CDD dataset, the F1-score increases from 95.28% to 96.19%, and IoU rises from 91.00% to 92.67%, demonstrating that CA can better capture fine-grained and structural changes. On the more challenging SYSU-CD dataset, the CA module brings a significant boost in recall (from 74.87% to 78.95%) and improves the overall accuracy slightly. Notably, the improvements on the CDD and SYSU-CD datasets highlight the CA’s effectiveness in modeling complex change scenarios.
To more intuitively demonstrate the impact of the CA module, we also provide visual comparisons of the ablation experiments, as shown in Figure 7, Figure 8 and Figure 9. For clarity, we present these results separately. Each figure includes input images A and B, the ground truth change map, the output of the ablation model, and the output of the full model with MSCA. From these visualizations, it is evident that the CA-enhanced model produces more accurate and cleaner change maps, particularly in preserving object boundaries, suppressing false positives, and detecting fine-grained structural changes. In contrast, the ablation variant often suffers from noise, missed detections, or fragmented regions—especially in challenging scenes with complex textures or subtle changes. This highlights the importance of both semantic feature extraction and the cross-attention mechanism in change detection tasks.
In summary, the ablation study demonstrates that the proposed CA module significantly improves the model’s discriminative power and overall accuracy, proving its practical value in enhancing change detection performance.

3.6. Impact of Input Noise and Model Robustness

To evaluate the robustness of the proposed MSCA framework under noisy conditions, we conduct experiments by adding Gaussian noise ( σ = 0.1) to the input images. We carry out experiments on two representative datasets: CDD, which covers a wide range of seasonal and structural changes, and SYSU-CD, which contains diverse urban development scenarios with fine-grained annotations. Quantitative results are shown in Table 6, and visual comparisons are presented in Figure 10 and Figure 11.
Table 6 shows that MSCA maintains high performance even under noisy conditions. For instance, when Gaussian noise is added to the CDD dataset (denoted as CDD+gn), the F1-score only drops slightly from 96.19% to 95.98%, and the OA remains above 99%. On the SYSU-CD dataset, the F1-score decreases marginally from 80.76% to 80.03% after noise perturbation, confirming the model’s resilience.
Figure 10 and Figure 11 visualize the results. Despite noise interference, the predicted change maps align closely with the ground truth, indicating that MSCA effectively captures essential semantic changes and suppresses irrelevant noise.
These results demonstrate that our model exhibits strong robustness against common image degradations, which is essential for real-world remote sensing applications where image quality may be compromised due to atmospheric or sensor-induced distortions.

3.7. Visualization and Interpretability of Cross-Attention Mechanism

To enhance the interpretability of attention mechanisms, particularly the cross-attention (CA) module, we provide a series of visualizations in the form of heatmaps (as shown in Figure 12). These heatmaps highlight the regions of the input images (Image A and Image B) that the model focuses on during different stages of the feature fusion and decoding process.
From the visualizations, it is evident that the CA blocks progressively enhance the attention to semantically meaningful structures such as roads, buildings, and edges. This attention is not random; rather, it corresponds to areas critical for accurate prediction, as shown in the final segmentation output. By visualizing intermediate features, we can better understand how the model integrates change information from bitemporal images.
These heatmaps serve as a valuable interpretability tool, offering insights into the decision-making process of the model. This is particularly important in high-stakes applications such as urban planning or disaster management, where understanding the model’s reasoning can build trust and guide further human-in-the-loop analysis.

3.8. Generalization and Overfitting Mitigation

Although the proposed MSCA framework achieves strong performance on three widely used benchmark datasets (CDD, LEVIR-CD, and SYSU-CD), concerns may arise regarding its ability to generalize to unseen domains due to potential dataset-specific overfitting. To address this, we adopt several strategies to enhance the model’s robustness and generalization capability:
  • Data Augmentation: To increase data diversity and reduce the risk of overfitting, we employ various data augmentation techniques during training, including random horizontal and vertical flips, rotations, color jittering, and scaling. These augmentations simulate different spatial arrangements and lighting conditions, helping the model to generalize beyond the training distribution.
  • Dropout Regularization: Dropout layers are incorporated within the attention modules to prevent the co-adaptation of neurons and reduce reliance on specific feature patterns. This regularization technique encourages the model to learn more robust, distributed representations, thereby improving its generalizability.
  • Semantic Abstraction in Cross-Attention: The semantic map projection embedded in the cross-attention module allows the model to operate on high-level abstractions rather than raw pixel differences. This semantic-level reasoning helps the model to identify meaningful changes across different domains and reduces the risk of overfitting to low-level, domain-specific noise.
Through this combination of regularization, augmentation, and architectural design, the MSCA model demonstrates strong potential for real-world deployment across varied and unseen remote sensing scenarios.

4. Conclusions

In order to enhance the interaction between local and global features, this paper proposes a CNN-based multi-scale cross-attention network (MSCANet) specifically for remote sensing image change detection (DC). The model incorporates the concept of serial–parallel correlation. The CA module handles global feature extraction, while the PCM module focuses on local feature capture to supplement the detailed information that may be lost due to aggregation. The MSCANet employs an intermediate-stage fusion strategy, distinguished from the traditional early or late fusion methods. This strategy effectively fuses global and local information and enhances the model’s understanding of image semantics. We conducted experiments on three publicly available datasets: CDD, LEVIR-CD, and SYSU-CD. The experimental results show that our method outperforms the traditional methods. Specifically, MSCANet achieved an F1-score of 96.19% and IoU of 92.67% on the CDD dataset, 91.02% F1-score on LEVIR-CD, and 80.76% F1-score on SYSU-CD, demonstrating its strong capability in handling complex change scenarios. Although MSCANet has demonstrated strong robustness, the large number of parameters may still limit its application in resource-constrained environments. Future research could focus on optimizing the network architecture, reducing the unnecessary computational burden, and exploring more efficient model compression techniques. More practical change detection solutions may be deployed by reducing computational costs and storage requirements while ensuring detection accuracy.

Author Contributions

Writing—original draft, Methodology, Investigation, Conceptualization, X.Z. and X.L.; Writing—review and editing, Validation, Supervision, X.O.; Writing—review and editing, L.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Xizang Key Research and Development Program (No. XZ202501ZY0064), the Chengdu Major Technology Application Demonstration Project (No. 2023-YF09-00019-SN), and the Scientific Research Fund of Education Department of Hunan Province (Grant Number: 23B0644, 22A0471).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and analysed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Carvalho, O.A., Jr.; Guimarães, R.F.; Gillespie, A.R.; Silva, N.C.; Gomes, R.A.T. A New Approach to Change Vector Analysis Using Distance and Similarity Measures. Remote Sens. 2011, 3, 2473–2493. [Google Scholar] [CrossRef]
  2. Celik, T. Unsupervised Change Detection in Satellite Images Using Principal Component Analysis and k-Means Clustering. IEEE Geosci. Remote Sens. Lett. 2009, 6, 772–776. [Google Scholar] [CrossRef]
  3. Nielsen, A.A.; Müller, A.; Plads, R.P. Change detection by the MAD method in hyperspectral image data. In Proceedings of the the 3rd EARSeL Workshop on Imaging Spectroscopy, Oberpfaffenhofen, Germany, 13–16 May 2003. [Google Scholar]
  4. Hao, M.; Zhang, H.; Shi, W.; Deng, K. Unsupervised change detection using fuzzy c-means and MRF from remotely sensed images. Remote Sens. Lett. 2013, 4, 1185–1194. [Google Scholar] [CrossRef]
  5. Malila, W.A. Change Vector Analysis: An Approach for Detecting Forest Changes with Landsat. Change 1980, 1, 1. [Google Scholar]
  6. Li, H.; Qi, A.; Chen, H.; Chen, S.; Zhao, D. HSIAO Framework in Feature Selection for Hyperspectral Remote Sensing Images Based on Jeffries-Matusita Distance. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5101121. [Google Scholar] [CrossRef]
  7. Touazi, A.; Bouchaffra, D. A k-Nearest Neighbor approach to improve change detection from remote sensing: Application to optical aerial images. In Proceedings of the 2015 15th International Conference on Intelligent Systems Design and Applications (ISDA), Marrakech, Morocco, 14–16 December 2015; pp. 98–103. [Google Scholar] [CrossRef]
  8. Nemmour, H.; Chibani, Y. Multiple support vector machines for land cover change detection: An application for mapping urban extensions. ISPRS J. Photogramm. Remote Sens. 2006, 61, 125–133. [Google Scholar] [CrossRef]
  9. Volpi, M.; Tuia, D.; Bovolo, F.; Kanevski, M.; Bruzzone, L. Supervised change detection in VHR images using contextual information and support vector machines. Int. J. Appl. Earth Obs. Geoinf. 2013, 20, 77–85. [Google Scholar] [CrossRef]
  10. Liu, J.; Gong, M.; Zhao, J.; Li, H.; Jiao, L. Difference representation learning using stacked restricted Boltzmann machines for change detection in SAR images. Soft Comput. 2014, 20, 4645–4657. [Google Scholar] [CrossRef]
  11. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR 2014, abs/1409.1556. [Google Scholar]
  12. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  13. Medsker, L.; Jain, L.C. Recurrent Neural Networks: Design and Applications; CRC Press: Boca Raton, FL, USA, 1999. [Google Scholar]
  14. Du, B.; Ru, L.; Wu, C.; Zhang, L. Unsupervised Deep Slow Feature Analysis for Change Detection in Multi-Temporal Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9976–9992. [Google Scholar] [CrossRef]
  15. Li, Q.; Gong, H.; Dai, H.; Li, C.; He, Z.; Wang, W. Unsupervised Hyperspectral Image Change Detection via Deep Learning Self-Generated Credible Labels. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2021, 14, 9012–9024. [Google Scholar] [CrossRef]
  16. Wang, L.; Wang, L.; Wang, Q.; Atkinson, P.M. SSA-SiamNet: Spectral–Spatial-Wise Attention-Based Siamese Network for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5510018. [Google Scholar] [CrossRef]
  17. Yang, Y.; Qu, J.; Xiao, S.; Dong, W.; Li, Y.; Du, Q. A Deep Multiscale Pyramid Network Enhanced with Spatial–Spectral Residual Attention for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5525513. [Google Scholar] [CrossRef]
  18. Fang, Y.; Sun, L.; Zheng, Y.; Wu, Z. Deformable Convolution-Enhanced Hierarchical Transformer with Spectral-Spatial Cluster Attention for Hyperspectral Image Classification. IEEE Trans. Image Process. 2025, 34, 701–716. [Google Scholar] [CrossRef] [PubMed]
  19. Zhang, C.; Wang, L.; Cheng, S.; Li, Y. SwinSUNet: Pure Transformer Network for Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5224713. [Google Scholar] [CrossRef]
  20. Bandara, W.G.C.; Patel, V.M. A Transformer-Based Siamese Network for Change Detection. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 207–210. [Google Scholar] [CrossRef]
  21. Chen, H.; Qi, Z.; Shi, Z. Remote Sensing Image Change Detection with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5607514. [Google Scholar] [CrossRef]
  22. Wang, G.; Li, B.; Zhang, T.; Zhang, S. A Network Combining a Transformer and a Convolutional Neural Network for Remote Sensing Image Change Detection. Remote Sens. 2022, 14, 2228. [Google Scholar] [CrossRef]
  23. Song, F.; Zhang, S.; Lei, T.; Song, Y.; Peng, Z. MSTDSNet-CD: Multiscale Swin Transformer and Deeply Supervised Network for Change Detection of the Fast-Growing Urban Regions. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6508505. [Google Scholar] [CrossRef]
  24. Feng, Y.; Xu, H.; Jiang, J.; Liu, H.; Zheng, J. ICIF-Net: Intra-Scale Cross-Interaction and Inter-Scale Feature Fusion Network for Bitemporal Remote Sensing Images Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4410213. [Google Scholar] [CrossRef]
  25. Lei, T.; Zhang, Y.; Lv, Z.; Li, S.; Liu, S.; Nandi, A.K. Landslide Inventory Mapping From Bitemporal Images Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2019, 16, 982–986. [Google Scholar] [CrossRef]
  26. Fu, Z.; Li, J.; Ren, L.; Chen, Z. SLDDNet: Stagewise Short and Long Distance Dependency Network for Remote Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 3000319. [Google Scholar] [CrossRef]
  27. Fang, S.; Li, K.; Shao, J.; Li, Z. SNUNet-CD: A Densely Connected Siamese Network for Change Detection of VHR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 8007805. [Google Scholar] [CrossRef]
  28. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016. [Google Scholar]
  29. Lebedev, M.; Vizilter, Y.V.; Vygolov, O.; Knyaz, V.A.; Rubis, A.Y. Change Detection in Remote Sensing Images Using Conditional Adversarial Networks. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 42, 565–571. [Google Scholar] [CrossRef]
  30. Chen, H.; Shi, Z. A Spatial-Temporal Attention-Based Method and a New Dataset for Remote Sensing Image Change Detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
  31. Shi, Q.; Liu, M.; Li, S.; Liu, X.; Wang, F.; Zhang, L. A Deeply Supervised Attention Metric-Based Network and an Open Aerial Image Dataset for Remote Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5604816. [Google Scholar] [CrossRef]
  32. Han, C.; Wu, C.; Guo, H.; Hu, M.; Chen, H. HANet: A Hierarchical Attention Network for Change Detection with Bitemporal Very-High-Resolution Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 3867–3878. [Google Scholar] [CrossRef]
  33. Huang, Y.; Li, X.; Du, Z.; Shen, H. Spatiotemporal Enhancement and Interlevel Fusion Network for Remote Sensing Images Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5609414. [Google Scholar] [CrossRef]
Figure 1. MSCAnetwork structure.
Figure 1. MSCAnetwork structure.
Sensors 25 02813 g001
Figure 2. Semantic mapping generation module.
Figure 2. Semantic mapping generation module.
Sensors 25 02813 g002
Figure 4. Prediction results of all methods on CDD datasets. (A refers to T1 image, B refers to T2 image).
Figure 4. Prediction results of all methods on CDD datasets. (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g004
Figure 5. Prediction results of all methods on LEVIE-CD datasets.(A refers to T1 image, B refers to T2 image).
Figure 5. Prediction results of all methods on LEVIE-CD datasets.(A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g005
Figure 6. Prediction results of all methods on SYSU-CD. (A refers to T1 image, B refers to T2 image).
Figure 6. Prediction results of all methods on SYSU-CD. (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g006
Figure 7. Ablation experiments on the SYSU-CD dataset. (A refers to T1 image, B refers to T2 image).
Figure 7. Ablation experiments on the SYSU-CD dataset. (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g007
Figure 8. Ablation experiments on the CDD dataset. (A refers to T1 image, B refers to T2 image).
Figure 8. Ablation experiments on the CDD dataset. (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g008
Figure 9. Ablation experiments on the LEVIR-CD dataset. (A refers to T1 image, B refers to T2 image).
Figure 9. Ablation experiments on the LEVIR-CD dataset. (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g009
Figure 10. Experiment results on CDD dataset with Gaussian noise (Noise_A, Noise_B). (A refers to T1 image, B refers to T2 image).
Figure 10. Experiment results on CDD dataset with Gaussian noise (Noise_A, Noise_B). (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g010
Figure 11. Experiment results on SYSU-CD dataset with Gaussian noise (Noise_A, Noise_B). (A refers to T1 image, B refers to T2 image).
Figure 11. Experiment results on SYSU-CD dataset with Gaussian noise (Noise_A, Noise_B). (A refers to T1 image, B refers to T2 image).
Sensors 25 02813 g011
Figure 12. Network visualization taking bitemporal RS images from LEVIR-CD as an example. Dashed lines indicate skip connections.
Figure 12. Network visualization taking bitemporal RS images from LEVIR-CD as an example. Dashed lines indicate skip connections.
Sensors 25 02813 g012
Table 1. Summary of the dataset characteristics.
Table 1. Summary of the dataset characteristics.
DatasetResolution (m/pixel)Patch Size (pixels)Data SourceRegion
CDD [29]∼0.5–1.0256 × 256Google Earth imageryGlobal urban/rural
LEVIR-CD [30]0.51024 × 1024Google Earth (Texas, USA)20 urban areas in Texas
SYSU-CD [31]0.5256 × 256Aerial imagery (HK Gov.)Hong Kong (2007–2014)
Table 2. Detection results on CDD datasets.
Table 2. Detection results on CDD datasets.
ModelPreRecF1IouOA
SNUNet95.6295.2495.4290.3799.10
STANet88.9794.3191.5684.4487.95
ChangeFormer95.2793.8294.5489.6598.72
BIT-CD95.8694.5995.0090.8898.88
HANet92.4484.6888.3979.1997.12
SEIFNet95.8592.6294.2089.0498.66
MSCA96.3995.9996.1992.6799.10
Table 3. Detection results on LEVIR-CD datasets.
Table 3. Detection results on LEVIR-CD datasets.
ModelPreRecF1IouOA
SNUNet89.1887.1788.1678.8398.82
STANet80.9991.9285.7975.1298.46
ChangeFormer91.5388.8690.1782.1099.01
BIT-CD90.9588.5790.2382.1999.02
HANet89.5687.1388.3379.1098.83
SEIFNet89.6785.8789.1480.4198.93
MSCA90.6587.4691.0280.2298.90
Table 4. Detection results on SYSU-CD datasets.
Table 4. Detection results on SYSU-CD datasets.
ModelPreRecF1IouOA
SNUNet74.5565.1567.6861.1188.78
STANet82.3674.3078.1264.1090.18
ChangeFormer84.9970.9377.3363.0490.19
BIT-CD79.0476.7177.8663.7589.71
HANet78.7176.1477.4163.1489.52
SEIFNet79.4778.2878.8765.1190.10
MSCA82.6578.9580.7667.7291.13
Table 5. Performance comparison of baseline and MSCA on three datasets.
Table 5. Performance comparison of baseline and MSCA on three datasets.
DatasetMethodPre (%)Rec (%)F1 (%)IoU (%)OA (%)
Levir-CDBaseline92.4588.4290.3982.4799.04
MSCA90.6587.4691.0280.2298.90
CDDBaseline95.0795.5195.2891.0098.89
MSCA96.3995.9996.1992.6799.10
SYSU-CDBaseline85.0274.8779.6266.1590.96
MSCA82.6578.9580.7667.7291.13
Table 6. Experiment results of MSCA on two datasets with added Gaussian noise.
Table 6. Experiment results of MSCA on two datasets with added Gaussian noise.
MethodDatasetPre (%)Rec (%)F1 (%)Iou (%)OA (%)
MSCACDD96.3995.9996.1992.6799.10
CDD+gn96.3595.6195.9892.2799.05
SYSU-CD82.6578.9580.7667.7291.13
SYSU-CD+gn86.7274.2980.0366.7291.26
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, X.; Lin, X.; Qing, L.; Ou, X. Semantic-Aware Remote Sensing Change Detection with Multi-Scale Cross-Attention. Sensors 2025, 25, 2813. https://doi.org/10.3390/s25092813

AMA Style

Zheng X, Lin X, Qing L, Ou X. Semantic-Aware Remote Sensing Change Detection with Multi-Scale Cross-Attention. Sensors. 2025; 25(9):2813. https://doi.org/10.3390/s25092813

Chicago/Turabian Style

Zheng, Xingjian, Xin Lin, Linbo Qing, and Xianfeng Ou. 2025. "Semantic-Aware Remote Sensing Change Detection with Multi-Scale Cross-Attention" Sensors 25, no. 9: 2813. https://doi.org/10.3390/s25092813

APA Style

Zheng, X., Lin, X., Qing, L., & Ou, X. (2025). Semantic-Aware Remote Sensing Change Detection with Multi-Scale Cross-Attention. Sensors, 25(9), 2813. https://doi.org/10.3390/s25092813

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop