Next Article in Journal
SAC-MS: Joint Slice Resource Allocation, User Association and UAV Trajectory Optimization with No-Fly Zone Constraints
Previous Article in Journal
A Dual-Segmentation Framework for the Automatic Detection and Size Estimation of Shrimp
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DynaNet: A Dynamic Feature Extraction and Multi-Path Attention Fusion Network for Change Detection

1
College of Information Engineering, Inner Mongolia University of Technology, Huhhot 010080, China
2
Inner Mongolia Key Laboratory of Intelligent Perception and System Engineering, Hohhot 010080, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(18), 5832; https://doi.org/10.3390/s25185832
Submission received: 17 August 2025 / Revised: 15 September 2025 / Accepted: 17 September 2025 / Published: 18 September 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Existing change detection methods often struggle with both inadequate feature fusion and interference from background noise when processing bi-temporal remote sensing imagery. These challenges are particularly pronounced in building change detection, where capturing subtle spatial and semantic dependencies is critical. To address these issues, we propose DynaNet, a dynamic feature extraction and multi-path attention fusion network for change detection. Specifically, we design a Dynamic Feature Extractor (DFE) that leverages a cross-temporal gating mechanism to amplify relevant change signals while suppressing irrelevant variations, enabling high-quality feature alignment. A Contextual Attention Module (CAM) is then employed to incorporate global contextual information, further enhancing the discriminative capability of change regions. Additionally, a Multi-Branch Attention Fusion Module (MBAFM) is introduced to model inter-scale semantic relationships through self- and cross-attention mechanisms, thereby improving the detection of fine-grained structural changes. To facilitate robust evaluation, we present a new benchmark dataset, Inner-CD, comprising 800 pairs of 256 × 256 bi-temporal satellite images with 0.5–2 m spatial resolution. Unlike existing datasets, Inner-CD features abundant buildings in both temporal images, with changes manifested as subtle morphological variations. Extensive experiments demonstrate that DynaNet achieves state-of-the-art performance, obtaining F1-scores of 90.92% on Inner-CD, 92.38% on LEVIR-CD, and 94.35% on WHU-CD.

1. Introduction

Remote sensing change detection involves analyzing images taken at various times from the same locations to recognize alterations on the Earth’s surface [1]. This technology has a wide range of applications in multiple fields, including monitoring urban expansion [2], analysis of land-use change [3], agricultural resource surveys [4], protection of the ecological environment [5], and assessment of natural disasters [6]. With continuous advancements in remote sensing technology, change detection techniques have gradually expanded to include a variety of data sources, including high-resolution optical, SAR, and hyperspectral data. The availability of different data sources has facilitated the development and innovation of corresponding change detection methods [7,8,9,10,11]. In recent years, advancements in optical sensor technology have significantly increased the number of acquisitions of dual-temporal remote sensing image data, promoting in-depth research into automatic change detection methods and achieving a significant improvement in model performance [12,13].
Change detection in remote sensing images can be considered a binary semantic segmentation task. In this process, each pixel is assigned a binary label to indicate whether the object of interest in that area has undergone any changes. In practical applications, various factors—including seasonal illumination changes, structural modifications of buildings, and differences in sensor and imaging conditions—often result in large non-change regions, presenting significant challenges for remote sensing change detection (RSCD). Additionally, the area of change regions is typically much smaller than that of non-change regions, requiring fine spatial details to improve detection accuracy. As shown in Figure 1, the left image presents a sample from our proposed Inner-CD dataset, while the right image illustrates the imbalanced distribution of change and non-change pixels in Inner-CD compared with three other datasets.
Conventional methods for change detection include Change Vector Analysis (CVA) [14] and Principal Component Analysis (PCA) [15]. CVA determines the type and extent of changes by analyzing the differences in pixel values and played an essential role in early studies. Conversely, PCA extracts the main components from images through linear transformation to highlight change areas. However, these methods have limitations, such as reliance on manually set thresholds, poor generalizability, and susceptibility to false positives or false negatives in complex environments. With advancements in machine learning techniques, sample-based approaches like Support Vector Machines (SVM) [16] and Random Forests (RF) [17] have been integrated into change detection. These methods can automatically learn change features, thus improving detection accuracy. However, SVM faces low computational efficiency when handling large-scale data and has limited classification ability for complex nonlinear change features. In contrast, RF can handle high-dimensional data but its generalization ability is still insufficient when dealing with complex land cover types and change patterns, limiting its adaptability to new datasets [18].
Recent advances in deep learning have driven substantial progress in remote sensing change detection. Among them, Convolutional Neural Networks (CNNs) have demonstrated remarkable capabilities in image processing, owing to their effective feature representation, and are widely adopted in this field [19,20,21,22,23,24,25,26,27]. CNNs automatically learn local features from images through convolutional layers, effectively capturing information such as texture and shape of land cover. For example, in land-use change detection, CNNs can accurately identify changes in different land covers, such as farmland, forests, and urban areas. Additionally, CNNs can process large-scale remote sensing data, demonstrating high computational efficiency and scalability.
However, CNNs also have limitations. Due to their local receptive field, CNNs mainly focus on regional features in images and cannot capture global contextual information. In some complex change scenarios, such as urban expansion or post-disaster change detection, relying solely on local features may not accurately assess the type and extent of changes. To address this issue, researchers are increasingly investigating the integration of Transformer architecture into remote sensing change detection [28,29,30,31,32,33,34]. The Transformer, through its self-attention mechanism, can simultaneously capture global and local information, providing excellent global modeling ability. For example, in flood disaster monitoring, the Transformer can precisely identify large submerged areas and detailed changes in buildings, roads, and other features within these regions.
Despite significant progress in existing remote sensing change detection (CD) methods, challenges persist in fine-grained change detection, as the effectiveness of deep learning models heavily relies on the quality of training datasets. Numerous prior studies have focused on creating high-quality datasets for change detection, such as the Land Cover Change Dataset (CLCD) [35], LEVIR-CD [23], and SYSU-CD [36]. However, most current change detection datasets are based on scenarios where one temporal image lacks buildings and the other contains dense buildings, with little focus on cases where both temporal images have many buildings that have undergone spatial or morphological changes. This poses a challenge to the adaptability of existing methods. Therefore, we propose a new Inner-CD dataset, specifically designed for research on building morphological and spatial structural change detection tasks, aiming to enhance the model’s ability to identify complex building change patterns.
To address these issues, this paper proposes DynaNet, a method combining dynamic feature change extraction with a multi-branch attention feature fusion network, and introduces a high-resolution building change detection dataset (Inner-CD). Specifically, DynaNet employs ResNet to extract multi-scale features from bi-temporal images, uses a Dynamic Feature Extractor (DFE) with cross-time gating to enhance change-related signals, applies a Contextual Attention Module (CAM) to leverage global contextual information, and integrates self- and cross-attention through a Multi-Branch Attention Fusion Module (MBAFM) to strengthen multi-scale feature relationships and improve building change detection. The main contributions are summarized as follows:
  • We propose a Dynamic Feature Extractor (DFE) that enhances change-related features and suppresses background noise through a trans-temporal gating mechanism, enabling accurate extraction of fine-grained changes from bi-temporal images.
  • A Contextual Attention Module (CAM) is designed to leverage global context, enhancing focus on key change regions while suppressing background noise, thereby improving the accuracy and robustness of fine change detection in complex environments.
  • We introduce a Multi-Branch Attention Fusion Module (MBAFM) that models long-range dependencies and fuses multi-level features. By integrating self- and cross-attention, it enhances the structural relationships between buildings and their surroundings, improving change region recognition, boundary clarity, and robustness to noise.
  • We provide a high-resolution Inner-CD dataset containing 600 pairs of 256 × 256 pixel images with a spatial resolution of 0.5–2 m to facilitate future research in building change detection.

2. Related Works

The rapid advancement of deep learning, particularly the extensive use of Convolutional Neural Networks (CNNs) in various domains, has greatly enhanced remote sensing image change detection technology. CNNs have become indispensable in change detection tasks due to their strong local feature extraction capabilities. Daudt et al. [19] proposed three basic models based on Fully Convolutional Networks (FCN) [37], which identify change regions through simple feature concatenation or differential calculation. Peng et al. [38] addressed error accumulation in feature difference computation by improving the UNet++ architecture and proposing an end-to-end change detection network with attention mechanisms and multiscale feature fusion, significantly improving detection performance. In addition, researchers have explored various optimization strategies for CNN structural improvements. For instance, Fang et al. [20] introduced a dense connection twin network based on UNet++ and implemented the Channel Attention Mechanism (CAM) to enhance the representation of fine-grained features. Zhang et al. [24] designed a deep supervision image fusion network that effectively combines spatial and guided attention mechanisms to improve feature representation. Han et al. [26] proposed a lightweight and efficient self-attention mechanism to integrate and optimize multi-scale feature representations. In change detection tasks, difference enhancement strategies are commonly employed to optimize network performance. For example, Zhang et al. [39] developed a differential pyramid to extract multi-scale difference feature maps. Lei et al. [40] developed a differential enhancement module that suppresses irrelevant noise and highlights actual change regions through CAM and residual connections. Chen and Shi [23] generated difference maps by calculating the Euclidean distance between features, effectively reducing false change phenomena and improving detection accuracy.
Meanwhile, lightweight networks have made significant progress in balancing computational efficiency with detection performance. Lei et al. [41] adopted multi-scale decoupled convolutions and a spatial–spectral feature cooperative strategy to significnatly enhance feature extraction efficiency and accuracy. Li et al. [31] improved feature aggregation and change recognition capabilities based on Mobile Networks, achieving excellent detection results while reducing computational costs. Liu et al. [42] proposed the EFICNN network, which uses an iterative VGG16 backbone with a Siamese structure combined with a 3D differential enhancement module, edge-guided attention module, and flow-guided fusion module, achieving excellent performance in change detection tasks.
Transformers, with their self-attention mechanism, have demonstrated significant advantages in modeling global dependencies and have gradually become the mainstream architecture in remote sensing change detection [43,44]. For example, Bandara et al. [29] proposed the ChangeFormer network, which significantly enhances multi-scale detail representation by combining hierarchical Transformer encoders and multilayer perceptron decoders. Zhang et al. [21] proposed the ChangeStar method, which uses object changes in unpaired images as supervisory signals, effectively improving model performance. Liu et al. [44] designed the Swin Transformer architecture, further strengthening global modeling abilities. Nevertheless, Transformers face challenges in local feature extraction and preserving fine details, particularly when processing high-resolution remote-sensing images. To address these shortcomings, researchers have attempted to combine CNNs and Transformers to create hybrid models. These models leverage the local feature extraction capabilities of CNNs while utilizing the global feature modeling advantages of Transformers, thus reducing computational complexity. For example, Chen et al. [45] proposed a hybrid model combining a ResNet encoder with a self-attention decoder, which achieved good results in change detection tasks. Transformer-based change detection networks have shown strong capabilities in modeling long-term dependencies. Liu et al. [35] proposed LGPNet, which extracts multi-scale building features through a local–global feature pyramid and further improves change detection performance by integrating attention mechanisms and transfer learning strategies. Chen et al. [33] proposed a method that transforms bi-temporal images into semantic tokens and models their relationships using a Transformer framework. Bandara and Patel [29] designed a Transformer-based change detection network that employs a twin Transformer encoder and multilayer perceptron decoder. Yan et al. [46] introduced a fully Transformer-based change detection network that enhances feature representation ability through global perception feature extraction and pyramid fusion. Zhang et al. [47] further designed a full Transformer network with a twin U-shape architecture to overcome CNN limitations in global change modeling.
Additionally, hybrid methods combining CNNs and Transformers have gained increasing attention. Xia et al. [48] proposed an edge-guided network that integrates ResNet with a Transformer encoder to optimize change feature extraction. Li et al. [49] combined Transformers with U-Net to extract global and local change information. Li et al. [31] proposed ConvTransNet, which integrates global and local features through a parallel CNN-Transformer structure and introduces a multi-scale framework to improve the detection of small-scale change regions. Notably, large foundational models such as Vision Transformer (ViT) [50] and CLIP [28] have achieved significant success in computer vision tasks and demonstrated strong generalization abilities in change detection tasks. Continued optimization of Transformer architectures and their integration with CNNs is expected to further expand their application potential in remote sensing change detection.

3. Method

3.1. Overall Architecture

The proposed DynaNet adopts a typical encoder–decoder architecture consisting of three main components: the Dynamic Feature Extractor (DFE), the Contextual Attention Module (CAM), and the Multi-Branch Attention Fusion Module (MBAFM). The overall structure is illustrated in Figure 2. First, the bi-temporal remote sensing images T1 and T2 undergo feature extraction through ResNet-18—a shallower architecture within the Residual Network series, consisting of 18 layers with residual blocks and skip connections that mitigate gradient vanishing and decay issues. This step captures multi-scale spatial and semantic information from the bi-temporal images to form robust representations, thereby enhancing the perception of change regions. Next, these multi-scale features are input into DFE, where a differential operation generates a coarse change representation, which is then refined by deep DE convolution (DEConv) to capture finer spatial details. The extracted coarse change features are then processed by the CAM, which applies weighting based on global contextual information. This module enhances important regions while suppressing background noise, ensuring the model focuses on the most discriminative change areas. Subsequently, the bi-temporal features are further fused through the MBAFM, where the self-attention head captures intra-class relationships between different change regions, while the cross-attention head analyzes the interaction between change regions and the background. The self-attention head models the relationships between multiple change regions, while the cross-attention head focuses on the interaction between buildings and their surrounding environments, thereby improving the accuracy of building change detection. Finally, the fused features are optimized through a multilayer perceptron, which integrates global information to generate the final single-channel change prediction map.

3.2. Dynamic Feature Extractor

Dual-temporal remote sensing imagery often contains complex ground objects and inconsistencies in imaging conditions (e.g., variations in weather, seasons, illumination changes, and temporary activities). These variations cause objects with the same semantics to exhibit different spectral features and generate multiple spectral variations. These factors give rise to a large number of meaningless variations, which increases the difficulty of dual-temporal feature fusion for change detection. Most existing methods rely on feature splicing or feature subtraction to reduce the interference of irrelevant changes. As a result, these methods perform poorly when dealing with different remote sensing change detection (RSCD) targets. To address this problem, we propose a DFE that employs a trans-temporal gating mechanism to guide the processing of change-specific features, selectively enhancing relevant changes while suppressing irrelevant variations. In this way, the DFE achieves high-quality dual-temporal feature fusion. As shown in Figure 3, the dual-temporal features G 1 and G 2 first undergo a differencing operation to obtain a rough change representation G c .
G c = G 1 G 2
where − denotes elemental subtraction. G c is elementally spliced with G 1 and G 2 , and then DEConv is performed to obtain G c 1 and G c 2 as follows:
G c 1 = D E C o n v G c G 1
G c 2 = D E C o n v G c G 2
Note that we chose Deep DE Convolution (DEConv) for two main reasons: (1) DEConv is able to significantly reduce the number of parameters of the model, thus reducing computational consumption; (2) DEConv captures finer spatial details by calculating the difference between the input feature maps and their feature maps processed with different convolution kernels. This allows the model to learn more details in the spatial dimension and thus generate more accurate weights. The weights are calculated as follows:
W 1 = S i g m o i d ( D W c o n v ( G c 1 ) )
W 2 = S i g m o i d ( D W c o n v ( G c 2 ) )
Finally, the previous rough change Fc is spliced through the Contextual Attention Module with the weighted features of the dual temporal phase to get our final output, where through the Contextual Attention Module, the model is able to weight the critical regions in the feature map to enhance focus on the change regions and suppress irrelevant information. It also ensures that the model can focus on areas that are critical for change detection, thus improving the detection accuracy.
G o u t = D E c o n v ( ( W 1 G 1 ) ( W 2 G 2 ) ( C A M ( G c ) ) )
where ⊗ denotes elemental multiplication.

3.3. Contextual Attention Module

In remote sensing change detection, due to complex ground objects and diverse imaging conditions, changes between two-time points are often subtle and noisy. Therefore, simple feature differentiation may lead to unnecessary interference. To address this issue, we propose a CAM that can identify the regions most relevant to change detection in the image and apply global contextual information to weight these regions. Through the global attention mechanism, the model not only focuses on the details of local regions but also integrates global information from the image, ensuring the capture of the most discriminative features. Specifically, CAM computes the weights for each position in the feature map, thus significantly improving change detection accuracy.
Figure 4 illustrates the detailed structure of the Global Transformer Module. The input feature F R C × H × W first undergoes a 3 × 3 convolution and dimensional reconstruction to generate the query Q F R C 1 × 1 , key K F R H W × C 1 , and value V F R C 1 × H W , as shown in the following equation:
V F , K F , Q F Re ( C o n v 3 × 3 ( F ) )
The process begins by combining Q F and K F and passing them through a softmax layer to compute the feature weights. These weights are then multiplied by the values to create the weighted feature map. In the final step, a 1 × 1 convolutional layer returns the features to their original dimensionality, and they are then combined with the original input features via concatenation to form the final result. The following illustrates this sequence of operations:
G ( F ) = K ( F ) T Q ( F )
H ( F ) = V ( F ) s o f t max ( G ( F ) )
F o u t = F + α G ( F )
where T denotes the transpose operation, α represents the 1 × 1 convolution, G F R H W × 1 denotes the fused features, H F R C 1 × 1 represents the weighted features, and F o u t R C × H × W is the output feature.
The Q in CAM acts as a channel query rather than a position query. The query vector is the feature selection from the key matrix channels. CAM can be roughly described as an attention mechanism that integrates channel and spatially weighted features with the input information.

3.4. Multi-Branch Attention Fusion Module

Compared to other parts of remote sensing image change detection, the regions of changed objects are typically small, making the restoration of spatial details of the change representation crucial [51]. During the feature fusion process, high-level change representations contain rich semantic information but have lower boundary quality, while low-level change representations capture spatial details but lack background information. Therefore, the common strategy adopted so far is to densely connect change representations at different scales to obtain both semantic information and spatial details. However, boundary and background information may introduce interference, leading to a decrease in change detection performance.
For this reason, the MBAFM in this paper consists of self-attention headers and cross-attention headers, which are used to capture intra-class relationships between features at different levels and feature information highlighting the change region, respectively, as shown in Figure 5. Each header contains three trainable linear embedding layers (realized by 3 × 3 convolution and reshape operations), which are used as query, key and value generators, respectively, and the pair-by-pair formulas for query and key in the self-attention header and cross-attention header are as follows:
F s 1 ( F 1 ) = C o n c a t ( Q s ( F 1 ) , K s ( F 1 ) )
F c 1 ( F 1 , F 2 ) = C o n c a t ( Q c ( F 1 ) , K c ( F 2 ) )
where the subscripts s and c represent the self-attention and cross-attention heads, respectively. F s 1 ( F 1 ) and F c 1 ( F 1 , F 2 ) represent the initial self-attention features and cross features after pairwise computation of F 1 and F 2 using the query and key; Q s ( · ) and K s ( · ) are used to generate the queries and keys in the self-attention mechanism; Q c ( · ) and K c ( · ) are used to generate the queries and keys in the cross-attention mechanism. Next, the individual attention features of the two heads are computed as follows:
θ s ( F 1 ) = V s ( F 1 ) s o f t max ( F s 1 ( F 1 ) )
θ c ( F 1 , F 2 ) = V c ( F 2 ) s o f t max ( F c 1 ( F 1 , F 2 ) )
where the Softmax function is a commonly used activation function that transforms real-valued vectors into a probability distribution, where each element is between 0 and 1, and the sum of all elements equals 1. H is used to generate feature values in both the self-attention mechanism and the cross-attention mechanism. Residual learning is applied to each head, and the output is obtained as follows:
F s = W s θ s ( F 1 ) F 1
F c = W c θ c ( F 1 , F 2 ) F 2
where W s and W c are linear embeddings realized by 1 × 1 convolution, and the ⊕ operation is realized by residual concatenation through element-by-element addition.
The self-attention head captures long-range dependencies by calculating a weighted sum of features from all positions for a given position. This allows it to exchange information between multiple change regions, regardless of their distance, thereby modeling the intra-class relationships of these regions. At the same time, the self-attention mechanism helps to distinguish between different types of changes and further refine the boundaries of large-scale change areas, improving detection accuracy and enabling precise segmentation of building change areas. On the other hand, the cross-attention head extracts global structural information from land cover features, merging the interactions between buildings and their surrounding environments. Considering the close relationship between buildings and their environments, cross-attention helps to identify newly added or demolished buildings more accurately while reducing false detection caused by image noise or non-building structures (such as trees and road changes). In this paper, the features extracted by the self-attention head are fused with those extracted by the cross-attention head to form the final result:
F o u t = c o n c a t ( F c , F s )

3.5. Loss Functions

In this paper, we use the strategy of combining focal and Dice loss to address the problem of category inequality in change detection and enhance the accuracy of edge detection [52]. Together, these two loss functions complement each other, improving both robustness and detection accuracy. The combined loss function is shown below:
L o s s = F o c a l L o s s + D i c e L o s s
Specifically, the formula for loss of focus is as follows:
F o c a l L o s s = α ( 1 p t ) γ log ( p t )
p t = p , if y = 1 ( 1 p ) , otherwise
where, α and γ are two hyperparameters, α governs the weighting of positive and negative samples, while adjusts the model’s emphasis on difficult-to-classify samples (empirically set to α = 0.2 and γ = 3). The term denotes the predicted probability, and is the ground-truth binary label, with 0 signifying unchanged pixels and 1 indicating changed pixels. Furthermore, the expression for the Dice loss is as follows:
D ice L o s s = 1 2 · P G + smoo thing P + G + smoo t h i n g
where P represents the region predicted by the model and G represents the ground truth region. |P| and |G| are their respective areas (or pixel counts). The smoothing term is a small positive number used to ensure that the denominator is not zero when both P and G are empty regions, thus avoiding computational errors. Adding the smoothing term in cases of small areas or imbalanced classes helps improve numerical stability and prevents large fluctuations in the loss value.

4. Experimental Results and Analysis

4.1. Dataset Introduction

The efficacy of the proposed DynaNet was validated through experiments on two widely used public building change detection (BCD) datasets, LEVIR-CD [23](LEarning, VIsion and Remote sensing laboratory, Beijing, China) and WHU-CD [53](Wuhan University Sigma Laboratory, Wuhan, China), alongside our proprietary Inner-CD dataset. For consistency with our small-batch learning strategy, all datasets were uniformly segmented into non-overlapping image patches of 256 × 256 pixels and subsequently divided at random into training, validation, and testing subsets. Figure 6 presents visual examples of bi-temporal input images and their corresponding change labels from each dataset.
(1) Inner-CD (Ours): Inner-CD is a newly curated high-resolution urban change detection dataset, meticulously developed to address fundamental limitations in existing datasets such as LEVIR-CD and WHU-CD. These traditional benchmarks typically simplify the change detection task to binary presence or absence scenarios, failing to adequately represent finer-grained structural evolutions and nuanced morphological alterations commonly encountered in real-world urban settings.
To bridge this critical gap and foster the advancement of structure-aware and content-sensitive change detection methodologies, Inner-CD features 600 meticulously curated bi-temporal image pairs, each with a resolution of 256 × 256 pixels. These images were extracted from multi-temporal Google Maps satellite imagery covering diverse urban and suburban landscapes in Hohhot and Ulanqab, Inner Mongolia, China. The imagery captures temporal spans ranging from 2 to 5 years, providing an authentic representation of urban morphological dynamics. The spatial resolutions within Inner-CD vary between 0.5 and 2 m, reflecting realistic variations in data quality and geographical conditions.
Distinctively, Inner-CD is characterized by high temporal symmetry, where the primary categories—House, Land, Road, and Forest—remain present across both temporal captures. Unlike conventional datasets that focus on simple appearance or disappearance events, Inner-CD emphasizes subtle intra-object variations, such as partial demolitions and incremental expansions in houses, nuanced alterations in land utilization, road modifications, and slight morphological changes in forested areas. Such fine-grained changes challenge models to move beyond straightforward detection tasks and engage in intricate structural comparisons, detailed spatial analysis, and semantic interpretation (as illustrated in Figure 7).
Each image pair within Inner-CD underwent rigorous annotation conducted by expert annotators using advanced GIS tools. Annotators precisely delineated change areas at a pixel-level, categorizing them into one of the four specified classes: House, Land, Road, and Forest, ensuring high-quality, accurate, and consistent annotations. To guarantee annotation reliability and minimize errors, a comprehensive multi-step validation approach was employed, incorporating cross-verification by multiple annotators and periodic oversight by senior quality assurance experts.
By significantly elevating the semantic complexity and spatial nuance, Inner-CD compels change detection models toward more sophisticated and robust capabilities, including nuanced morphological reasoning and intricate feature correlation analyses. This rigorous approach facilitates the development and benchmarking of advanced algorithms designed to operate reliably in realistic and morphologically complex scenarios.
Ultimately, Inner-CD serves not merely as another dataset but as a foundational resource intended to propel the field of urban change detection toward achieving higher standards of accuracy and robustness in recognizing intricate urban dynamics. It provides an ideal benchmark to rigorously assess models like DynaNet, particularly in their ability to effectively handle delicate structural changes, nuanced spatial transformations, and sophisticated inter-temporal correlations essential for practical real-world deployment.
(2) LEVIR-CD [23]: The LEVIR-CD dataset contains 637 high-resolution (1024 × 1024) image pairs with a 0.5-m spatial resolution, collected across multiple cities in Texas, USA. The temporal gap ranges from 5 to 14 years. Building types include villas, apartments, garages, and warehouses, but change patterns are typically large-scale insertions of buildings into previously empty regions, making detection tasks more binary and localized.
(3) WHU-CD [53]: This dataset consists of two large aerial images with dimensions of 32507×15354 pixels and a 0.3-m resolution. Similar to LEVIR-CD, it focuses on significant structural additions or removals, with relatively limited intra-object semantic variation. As such, although it supports general change detection, it lacks the temporal symmetry and spatial nuance present in Inner-CD.

4.2. Implementation Details and Evaluation Metrics

All experiments were conducted on a 64-bit Ubuntu 20.04 system equipped with an Intel Xeon(R) Gold 6430 CPU @ 16 GHz, an NVIDIA RTX 4090 GPU, and a deep learning environment configured with Python 3.10, PyTorch 2.2.0, and CUDA 11.8. (All equipment and environment configurations are managed on a computing power leasing platform developed by Shituo Cloud China (Nanjing) Technology Co., Ltd., Nanjing, China) We adopted Stochastic Gradient Descent (SGD) as the optimizer, owing to its stability and reliable convergence on large-scale datasets. A momentum of 0.9 was employed to accelerate convergence and smooth gradient updates. The initial learning rate was set to 0.05 with a weight decay of 0.00005, ensuring global search capability while mitigating overfitting. To simplify training and maintain stability, a fixed single-stage learning rate schedule was applied. The batch size was set to 8, considering the high resolution of remote sensing images and the resulting large GPU memory consumption. To compensate for the relatively small batch size, the training process was extended to 200 epochs. Overall, these parameter choices were made to balance convergence speed, model complexity, and generalization ability under the constraints of remote sensing change detection tasks and available hardware resources.
To assess the performance of the proposed algorithm, we employed a suite of standard metrics for change detection, namely precision (Pr), recall (Re), F1-score, and intersection over union (IOU). High precision helps reduce false positives, while high recall helps minimize false negatives. These metrics are composite indicators, with higher values indicating better performance. The formulas for calculating these metrics are as follows:
F 1 = 2 × P r × R e P r + R e
P r = T P T P + F P
R e = T P T P + F N
I O U = T P T P + F N + F P
where TP and TN denote the number of pixels correctly classified as changed and unchanged, respectively. Conversely, FN refers to the count of pixels that are actually changed but were incorrectly predicted as unchanged, while FP is the count of pixels that are actually unchanged but were incorrectly predicted as changed.

4.3. Comparison Experiments

Comparison Methods: To benchmark the performance of DynaNet on high-resolution remote sensing change detection tasks, we conducted a comparative analysis against 13 established methods, which were grouped into three distinct classes. The initial group comprises models based on Convolutional Neural Networks (CNNs), such as FC-Siam-Diff [54], FC-Siam-Conc [19], SNUNet [20], StarCD-Net [54], and DTCDSCN [44]. The FC-Siam-Diff and FC-Siam-Conc models both utilize CNN architectures; however, FC-Siam-Conc further integrates Siamese networks and feature fusion strategies to improve its change detection capabilities. SNUNet, also a CNN-based method, adopts the U-Net architecture, effectively extracting local features, thereby improving detection accuracy. StarCD-Net is a CNN-based method that leverages StarNet and differential operators to enhance edge-aware change detection. The second category consists of Transformer-based methods, including DMATNet [55], BIT [33], Changeformer [29], and STANet [23]. BIT uses a Transformer architecture to capture global dependency information; Changeformer is designed explicitly for change detection in remote sensing images; and STANet improves model performance by introducing a temporal modeling mechanism. The third category includes hybrid frameworks combining CNN and Transformer, such as ConvTransNet [31], WNet [32] and LGPNet [35], which adopt strategies combining CNNs and Transformers. RDSF-Net [56] integrates CNN and Mamba—a state-space-model-based efficient global modeling structure—and employs residual wavelet transform for downsampling.
Quantitative Analysis: Table 1, Table 2 and Table 3 present the experimental results of DynaNet and other state-of-the-art methods on three datasets.
For the LEVIR-CD dataset, DynaNet outperforms all other evaluation metrics, achieving the highest F1-score (92.38 %), precision (93.45%), recall (91.46%), and IOU (84.57%). Compared to the second-ranked RDSF-Net, DynaNet achieves an F1-score that is 1.11% higher, precision that is 2.29% higher, recall that is 1.01% higher, and IOU that is 0.78% higher. This demonstrates that DynaNet offers a clear performance advantage on the LEVIR-CD dataset, detecting changes more accurately while maintaining high precision and recall.
For the WHU-CD dataset, DynaNet performs the best across three evaluation metrics, although it slightly trails Changeformer in precision. Specifically, DynaNet leads in F1-score (94.35%), recall (92.46%), and IOU (87.57%), while Changeformer outperforms in precision. However, its lower recall results in an overall performance that is still inferior to DynaNet. Compared to ConvTransNet, DynaNet leads in precision, recall, F1-score, IOU, and OA (Overall Accuracy) by 1.22%, 1.13%, 1.18%, 2.04%, and 0.07%, respectively, further confirming the superiority of DynaNet on the WHU-CD dataset.
For the Inner-CD dataset, DynaNet also performs strongly, achieving an F1-score of 90.92%, precision of 93.80%, recall of 87.84%, and IOU of 80.37%. DynaNet excels in F1-score, precision, recall, and IOU compared to existing methods. These results demonstrate that the proposed framework has excellent generalization and adaptability when handling building change detection tasks. This is primarily due to the general knowledge learned from two pre-trained models, allowing the model to effectively handle complex scenarios where buildings are abundant in both temporal images, and some undergo spatial position or morphological changes. This capability is particularly important for Inner-CD, where the change patterns are more complex and diverse than those in conventional datasets.
Visualization Analysis: In the visualization analysis, black represents TN (True Negative), white represents TP (True Positive), green represents FN (False Negative), and red represents FP (False Positive). Figure 8 shows the visualization results of change detection using different methods on the LEVIR-CD dataset. Figure 9 and Figure 10 display similar results on the WHU-CD and Inner-CD datasets. In the visualization images, the sequence from left to right shows Temporal A, Temporal B, labels, BIT, SNUNet, Changeformer, LGPNet, ConvTransNet, and the experimental results of the proposed DynaNet. Figure 6 demonstrates the qualitative experimental results of DynaNet compared to other state-of-the-art methods on the WHU-CD dataset.
Figure 8 shows that our method significantly reduces the regions of false positives (FP) and false negatives (FN). Specifically, DynaNet effectively avoids the interference of shadows and environmental factors in the detection results. In challenging cases where building colors closely match the surrounding environment, other methods often struggle to detect the disappearance of buildings, leading to missed detections (green areas). In contrast, DynaNet effectively reduces missed detections and accurately locates the change regions. These findings strongly validate the superior performance of DynaNet on the WHU-CD dataset, demonstrating its outstanding ability in complex change scenarios.
Figure 9 shows the visual comparison of DynaNet with other methods on the LEVIR-CD dataset. The LEVIR-CD dataset contains many dense urban building changes, with buildings having irregular shapes. To verify the model’s effectiveness, we selected several images with different object shapes and distributions for comparison. From Figure 9, it can be seen that DynaNet performs better in detecting irregular change targets and can accurately detect small targets at the edges of images. Moreover, our method exhibits strong adaptability in handling simple and complex objects, significantly improving the precision of change detection.
Figure 10 presents the visual comparison results of DynaNet and other methods on the Inner-CD dataset. Unlike datasets such as LEVIR-CD and WHU-CD, the bi-temporal images in Inner-CD contain many buildings, and some buildings undergo spatial position or morphological changes. Such complex change patterns pose a more significant challenge to existing methods. From the visualization results, DynaNet performs exceptionally well in detecting complex building changes and can more accurately capture buildings’ spatial displacement and shape changes, significantly reducing false positives (FP) and false negatives (FN). In comparison, other methods tend to generate misdetections or missed detections in densely built-up areas, particularly lacking in fine-grained structural change detection. Additionally, DynaNet excels in handling edge details, maintaining the integrity of buildings, and avoiding boundary-blurring issues.

4.4. Ablation Experiments and Visualiztion

We conducted ablation experiments on three datasets, focusing on validating the contribution of each module in DynaNet. The experiments were performed through module splitting and replacement, aiming to assess the impact of each module on the model’s performance. As shown in Table 4, NO.1 represents the original baseline model. Compared to NO.2, NO.2, which introduces the DFE, significantly improves the model’s performance, with F1-scores and IOU values increasing by 2.24%, 2.89%, 1.40% and 3.84%, 6.97%, 3.74% across the three datasets, respectively. This indicates that the DFE module can effectively capture the change features between bi-temporal images, enhancing the model’s ability to recognize complex change areas. NO.7, based on NO.2, adds the CAM, resulting in F1-score improvements of 4.00% and 9.11% and an increase in IOU values of 6.27% and 10.92%. The CAM module leverages global contextual information to enhance detail recognition and noise suppression, significantly improving the model’s detection accuracy. Further, NO.8, built upon NO.7, introduces the MBAFM, enhancing the change detection performance. The F1-scores increased by 1.70% and 1.69%, and the IOU values rose by 2.09% and 3.06%. The MBAFM module strengthens the fusion of multi-scale features, improving the model’s ability to recognize building change boundaries and reducing false detection occurrences.In summary, each module in DynaNet plays a key role in improving the model’s overall performance, reducing false detections, and minimizing noise interference, further validating the effectiveness of these modules in complex change detection tasks.
Table 5 details the outcomes of our ablation analysis on various configurations of the DEConv and CAM modules within the TVRE framework. The findings indicate that both the DEConv and CAM components enhance the model’s effectiveness through unique mechanisms, and integrating them produces the optimal outcome. Specifically, the DEConv module enhances the model’s ability to extract and refine feature maps, improving the detection of subtle changes. As a result, it leads to an increase in both precision and recall, particularly in the presence of fine-grained changes. On the other hand, CAM focuses on global contextual information, enabling the model to capture broader relationships across the image. This helps suppress background noise and improve the detection of change regions. However, relying solely on CAM may lead to a slight reduction in precision when fine details are more important. When DEConv and CAM are combined, their strengths complement each other. DEConv refines local features, while CAM optimizes global context, increasing model robustness to small and large changes. This combination enhances the model’s overall performance, resulting in higher F1- and IOU scores across all datasets, as it addresses both local feature extraction and global contextual understanding. Thus, the combination of DEConv and CAM achieves the best performance by balancing local detail enhancement with global context integration.
Figure 11 shows the prediction results of DynaNet with different ablation modules on the LEVIR-CD, WHU-CD and Inner-CD datasets. From top to bottom, the examples for each dataset are presented. (a) and (b) represent remote sensing images before and after the change, (c) show the ground-truth change labels, and (d) display the prediction results of the baseline model. From (e) to (h), key components such as CAM, MBAFM and DFE are progressively introduced. As these modules are added, the detection results of the building change areas become more complete and the impact of false changes is effectively suppressed. This shows that each module plays a positive role in enhancing the change detection performance.

4.5. Efficiency Comparison

As indicated by the experiments, our proposed method surpasses the accuracy of the latest building change detection techniques on all three datasets. In this study, we employ the number of model parameters and floating-point operations per second (FLOPs) as metrics for model complexity and computational cost. The number of parameters reflects a model’s ability to learn and represent features, while FLOPs measure the computational effort required for a single forward pass, indicating the model’s efficiency and potential inference speed. Together, these metrics provide a balanced assessment of a model’s representational capability and computational efficiency—particularly crucial for constructing change detection tasks involving high-resolution images and timely processing. Table 6 presents these findings, with all methods standardized to an input size of 256 × 256 × 3 to ensure a consistent evaluation.
The experimental results show that DynaNet improves the F1-score for building change detection accuracy and maintains a relatively low model complexity. This advantage is primarily attributed to the collaborative effect of the DFE, the CAM, and the MBAFM, which enable the model to more precisely extract and fuse multi-scale features, thereby enhancing the ability to recognize building change areas. Although methods such as ConvTransNet and DTCDSCN have slightly lower computational costs than DynaNet, their detection accuracy is relatively lower. Furthermore, DynaNet does not significantly increase the number of trainable parameters or floating-point operations, indicating that it achieves high accuracy while maintaining good computational efficiency. DynaNet enhances building change detection accuracy through efficient feature extraction and fusion strategies without significantly increasing computational complexity. We compare computational complexity and parameter count in Figure 12 to visually demonstrate the model complexity of different methods.

5. Discussion

This paper proposes DynaNet, a novel deep neural network approach aimed at identifying changes in buildings from remote sensing data. Extensive experiments have demonstrated that DynaNet performs exceptionally well on a variety of datasets. It not only surpasses traditional CNN-based methods but also outperforms various Transformer models with significantly larger parameter counts, demonstrating promising capabilities for real-world usage and further advancements. The excellent performance of DynaNet can be attributed to three key design features, outlined as follows:
(1)
By introducing the DFE module, DynaNet effectively filters out irrelevant features and focuses on detecting meaningful changes. The cross-temporal gating mechanism within DFE allows the model to selectively enhance relevant changes while suppressing background noise, resulting in improved change detection accuracy. This contributes significantly to the model’s robustness in handling complex change scenarios, especially in cases where objects undergo shape or spatial changes.
(2)
The CAM module brings global context into feature fusion, greatly enhancing the model’s capacity to concentrate on key areas of change. Global attention helps the network overcome challenges like environmental interference and subtle changes within remote sensing imagery that are often hard to identify. The global context CAM provides ensures that the model effectively captures important change signals, boosting detection precision and recall.
(3)
The dual-branch attention fusion mechanism leverages self-attention and cross-attention, allowing the model to model long-range dependencies and interactions between scales and regions. This capability is particularly important for building change detection, as buildings and surrounding structures often interact, and their changes may span various spatial scales. These attention mechanisms enable DynaNet to precisely identify building changes, even in challenging scenarios with complex building layouts and diverse environmental contexts.
Despite these strengths, several limitations and avenues for future work remain:
(1)
Performance in complex environments: While DynaNet performs exceptionally well in typical building change detection tasks, its performance may be challenged in highly complex environments, such as areas with dynamic weather conditions, seasonal vegetation changes, or rapid urban development. Future work could explore integrating more advanced pre-trained vision models or multi-modal data to further enhance robustness and accuracy without altering the core structure of DynaNet.
(2)
Model complexity and computational cost: DynaNet achieves superior performance by combining modules such as DFE, CAM, and MBAFM. However, this modular design also increases computational complexity and memory requirements, which may limit real-time applications or deployment on resource-constrained devices. Future research could focus on designing more lightweight architectures, pruning redundant connections, or optimizing attention mechanisms to reduce computational cost while maintaining performance.
(3)
Generalization across datasets: Although DynaNet demonstrates strong performance on the Inner-CD dataset, its generalization capability across different geographic regions or imaging conditions has not been fully validated. Future studies could investigate cross-dataset evaluation and domain adaptation techniques to ensure broader applicability.

6. Conclusions

This paper proposes a new dataset for change detection, Inner-CD, and the DynaNet architecture. The introduction of Inner-CD provides high-quality data for deep learning-based building change detection models, especially for handling more complex and dynamic building transformations. It aims to address the limitations of current datasets, where publicly available datasets typically contain one time point with dense buildings and another with sparse buildings. DynaNet utilizes the DFE module to enhance dynamic feature extraction from bi-temporal images, effectively filtering out irrelevant features and improving change detection accuracy. The CAM combines global contextual information to suppress background noise while amplifying focus on the change areas, significantly improving detection accuracy. The MBAFM module further captures the intrinsic relationships between multi-scale features through self-attention and cross-attention mechanisms, optimizing the segmentation of change areas. Extensive experiments conducted on multiple high-resolution remote sensing datasets show that DynaNet outperforms existing methods, achieving exceptional performance in qualitative and quantitative evaluations. Ablation studies confirm each module’s significant contribution to the model’s overall performance, particularly in capturing fine-grained changes and reducing background noise interference. In conclusion, DynaNet offers a novel and effective solution for the challenge of detecting changes in remote sensing data, making change detection more precise in real-world applications.
However, DynaNet is not without its drawbacks. Future research will focus on incorporating more recent and sophisticated pre-trained models to enhance its capabilities in particularly challenging change scenarios—for instance, those with dynamic environmental factors or rapidly evolving landscapes. Additionally, model optimization techniques such as pruning or distillation will be investigated to reduce computational complexity, making DynaNet suitable for real-time deployment on resource-constrained edge devices while maintaining high detection accuracy.

Author Contributions

Methodology, X.L. and D.L.; validation, X.L. and J.F.; investigation, D.L.; writing—review and editing, X.L. and X.F.; writing—original draft preparation, D.L. and X.L.; data curation, J.F. and X.F.; funding acquisition, D.L. and J.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (62361049), the Fundamental Research Funds for the Central Universities (JY20250087) and the Inner Mongolia Science and Technology Plan Project (2025SKYPT0035, RZ2500000575).

Institutional Review Board Statement

The study did not require ethical approval.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to thank the editor and reviewers for their contributions to this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cheng, G.; Huang, Y.; Li, X.; Lyu, S.; Xu, Z.; Zhao, H.; Zhao, Q.; Xiang, S. Change Detection Methods for Remote Sensing in the Last Decade: A Comprehensive Review. Remote Sens. 2024, 16, 2355. [Google Scholar] [CrossRef]
  2. Lu, Y.; Coops, N.C.; Hermosilla, T. Estimating urban vegetation fraction across 25 cities in Pan-Pacific using Landsat time series data. ISPRS J. Photogramm. Remote Sens. 2017, 126, 11–23. [Google Scholar] [CrossRef]
  3. Bruzzone, L.; Prieto, D.F. Automatic Analysis of the Difference Image for Unsupervised Change Detection. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1171–1182. [Google Scholar] [CrossRef]
  4. Feranec, J.; Hazeu, G.W.; Christensen, S.; Jaffrain, G. Corine Land Cover Change Detection in Europe (case studies of the Netherlands and Slovakia). Land Use Policy 2007, 24, 234–247. [Google Scholar] [CrossRef]
  5. Hamidi, E.; Peter, B.G.; Muñoz, D.F.; Moftakhari, H.; Moradkhani, H. Fast flood extent monitoring with SAR change detection using Google Earth Engine. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–19. [Google Scholar] [CrossRef]
  6. Gong, M.; Zhao, J.; Liu, J.; Miao, Q.; Jiao, L. Change detection in synthetic aperture radar images based on deep neural networks. IEEE Trans. Neural Networks Learn. Syst. 2016, 27, 125–138. [Google Scholar] [CrossRef]
  7. Yu, Q.; Zhang, M.; Yu, L.; Wang, R.; Xiao, J. SAR Image Change Detection Based on Joint Dictionary Learning with Iterative Adaptive Threshold Optimization. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2022, 15, 5234–5249. [Google Scholar] [CrossRef]
  8. Amitrano, D.; Guida, R.; Iervolino, P. Semantic Unsupervised Change Detection of Natural Land Cover with Multitemporal Object-Based Analysis on SAR Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5494–5514. [Google Scholar] [CrossRef]
  9. Wang, L.; Wang, L.; Wang, Q.; Atkinson, P.M. SSA-SiamNet: Spectral-Spatial-Wise Attention-Based Siamese Network for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5510018. [Google Scholar] [CrossRef]
  10. Luo, F.; Zhou, T.; Liu, J.; Guo, T.; Gong, X.; Ren, J. Multiscale Diff-Changed Feature Fusion Network for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5502713. [Google Scholar] [CrossRef]
  11. Yang, Z.; Xu, W.; Chen, N.; Chen, Y.; Wu, K.; Xie, M.; Xu, H.; Zheng, E. SDA-YOLO: Multi-Scale Dynamic Branching and Attention Fusion for Self-Explosion Defect Detection in Insulators. Electronics 2025, 14, 3070. [Google Scholar] [CrossRef]
  12. Khelifi, L.; Mignotte, M. Deep Learning for Change Detection in Remote Sensing Images: Comprehensive Review and Meta-Analysis. IEEE Access 2020, 8, 126385–126400. [Google Scholar] [CrossRef]
  13. Zhang, S.; Zhang, Y.; Yu, Z.; Yang, S.; Kang, H.; Xu, J. MCA-GAN: A Multi-Scale Contextual Attention GAN for Satellite Remote-Sensing Image Dehazing. Electronics 2025, 14, 3099. [Google Scholar] [CrossRef]
  14. Johnson, R.D.; Kasischke, E.S. Change Vector Analysis: A Technique for the Multispectral Monitoring of Land Cover and Condition. Int. J. Remote Sens. 1998, 19, 411–426. [Google Scholar] [CrossRef]
  15. Byrne, G.F.; Crapper, P.F.; Mayo, K.K. Monitoring Land-Cover Change by Principal Component Analysis of Multitemporal Landsat Data. Remote Sens. Environ. 1980, 10, 175–184. [Google Scholar] [CrossRef]
  16. Nemmour, H.; Chibani, Y. Multiple Support Vector Machines for Land Cover Change Detection: An Application for Mapping Urban Extensions. ISPRS J. Photogramm. Remote Sens. 2006, 61, 125–133. [Google Scholar] [CrossRef]
  17. Im, J.; Jensen, J. A Change Detection Model Based on Neighborhood Correlation Image Analysis and Decision Tree Classification. Remote Sens. Environ. 2005, 99, 326–340. [Google Scholar] [CrossRef]
  18. Habib, T.; Inglada, J.; Mercier, G.; Chanussot, J. Support Vector Reduction in SVM Algorithm for Abrupt Change Detection in Remote Sensing. IEEE Geosci. Remote Sens. Lett. 2009, 6, 606–610. [Google Scholar] [CrossRef]
  19. Daudt, R.C.; Le Saux, B.; Boulch, A. Fully Convolutional Siamese Networks for Change Detection. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 4063–4067. [Google Scholar] [CrossRef]
  20. Fang, S.; Li, K.; Shao, J.; Li, Z. SNUNet-CD: A Densely Connected Siamese Network for Change Detection of VHR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  21. Zheng, Z.; Ma, A.; Zhang, L.; Zhong, Y. Change is Everywhere: Single-Temporal Supervised Object Change Detection in Remote Sensing Imagery. In Proceedings of the Proceedings of IEEE/CVF Int. Conf. Comput. Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 15173–15182. [Google Scholar] [CrossRef]
  22. Fang, S.; Li, K.; Li, Z. Changer: Feature Interaction is What You Need for Change Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–11. [Google Scholar] [CrossRef]
  23. Chen, H.; Shi, Z. A Spatial-Temporal Attention-Based Method and a New Dataset for Remote Sensing Image Change Detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
  24. Zhang, C.; Yue, P.; Tapete, D.; Jiang, L.; Shangguan, B.; Huang, L.; Liu, G. A Deeply Supervised Image Fusion Network for Change Detection in High Resolution Bi-temporal Remote Sensing Images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 183–200. [Google Scholar] [CrossRef]
  25. Feng, Y.; Jiang, J.; Xu, H.; Zheng, J. Change Detection on Remote Sensing Images Using Dual-Branch Multilevel Intertemporal Network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  26. Han, C.; Wu, C.; Guo, H.; Hu, M.; Chen, H. HANet: A Hierarchical Attention Network for Change Detection with Bitemporal Very-High-Resolution Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2023, 16, 3867–3878. [Google Scholar] [CrossRef]
  27. J, E.; Wang, L.; Zhao, C.; Mathiopoulos, P.; Ohtsuki, T.; Adachi, F. SAR Images Change Detection Based on Attention Mechanism-Convolutional Wavelet Neural Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 1, 12133–12149. [Google Scholar] [CrossRef]
  28. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 3–7 May 2021. [Google Scholar] [CrossRef]
  29. Bandara, W.G.C.; Patel, V.M. A Transformer-Based Siamese Network for Change Detection. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 207–210. [Google Scholar] [CrossRef]
  30. Wang, Q.; Jing, W.; Chi, K.; Yuan, Y. Cross-Difference Semantic Consistency Network for Semantic Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–12. [Google Scholar] [CrossRef]
  31. Li, W.; Xue, L.; Wang, X.; Li, G. ConvTransNet: A CNN–Transformer Network for Change Detection with Multiscale Global–Local Representations. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  32. Tang, X.; Zhang, T.; Ma, J.; Zhang, X.; Liu, F.; Jiao, L. WNet: W-Shaped Hierarchical Network for Remote-Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–14. [Google Scholar] [CrossRef]
  33. Chen, H.; Qi, Z.; Shi, Z. Remote Sensing Image Change Detection with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  34. Zhou, J.; Zhang, P.; Zhou, Z. CIMF-Net: A Change Indicator-enhanced Multiscale Fusion Network for Remote Sensing Change Detection. IEEE Access 2025, 1, 66843–66854. [Google Scholar] [CrossRef]
  35. Liu, M.; Chai, Z.; Deng, H.; Liu, R. A CNN-transformer network with multi-scale context aggregation for fine-grained cropland change detection. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2022, 15, 4297–4306. [Google Scholar] [CrossRef]
  36. Shi, Q.; Liu, M.; Li, S.; Liu, X.; Wang, F.; Zhang, L. A Deeply Supervised Attention Metric-Based Network and an Open Aerial Image Dataset for Remote Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5604816. [Google Scholar] [CrossRef]
  37. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  38. Peng, D.; Zhang, Y.; Guan, H. End-to-End Change Detection for High-Resolution Satellite Images Using Improved UNet++. Remote Sens. 2019, 11, 1382. [Google Scholar] [CrossRef]
  39. Zhang, X.; Yue, Y.; Gao, W.; Yun, S.; Su, Q.; Yin, H.; Zhang, Y. DifUnet++: A Satellite Images Change Detection Network Based on Unet++ and Differential Pyramid. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  40. Lei, T.; Wang, J.; Ning, H.; Wang, X.; Xue, D.; Wang, Q.; Nandi, A.K. Difference Enhancement and Spatial-Spectral Nonlocal Network for Change Detection in VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4507013. [Google Scholar] [CrossRef]
  41. Lei, T.; Geng, X.; Ning, H.; Lv, Z.; Gong, M.; Jin, Y.; Nandi, A.K. Ultralightweight Spatial–Spectral Feature Cooperation Network for Change Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4402114. [Google Scholar] [CrossRef]
  42. Liu, Y.; Li, S.; He, Z.; Liu, K. Edge and Flow Guided Iterative CNN for Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 4406422. [Google Scholar] [CrossRef]
  43. Tan, H.; He, L.; Du, W.; Liu, H.; Chen, H.; Zhang, Y.; Yang, H. PRX-Change: Enhancing Remote Sensing Change Detection through Progressive Feature Refinement and Cross-Attention Interaction. Int. J. Appl. Earth Obs. Geoinf. 2024, 132, 104008. [Google Scholar] [CrossRef]
  44. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar] [CrossRef]
  45. Chen, H.; Wu, C.; Du, B.; Zhang, L.; Wang, L. Change Detection in Multisource VHR Images via Deep Siamese Convolutional Multiple-Layers Recurrent Neural Network. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2848–2864. [Google Scholar] [CrossRef]
  46. Yan, T.; Wan, Z.; Zhang, P.; Cheng, G.; Lu, H. TransY-Net: Learning Fully Transformer Networks for Change Detection of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–12. [Google Scholar] [CrossRef]
  47. Zhang, C.; Wang, L.; Cheng, S.; Li, Y. SwinSUNet: Pure Transformer Network for Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  48. Xia, L.; Chen, J.; Luo, J.; Zhang, J.; Yang, D.; Shen, Z. Building Change Detection Based on an Edge-Guided Convolutional Neural Network Combined with a Transformer. Remote Sens. 2022, 14, 4524. [Google Scholar] [CrossRef]
  49. Li, Q.; Zhong, R.; Du, X.; Du, Y. TransUNetCD: A Hybrid Transformer Network for Change Detection in Optical Remote-Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
  50. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models from Natural Language Supervision. In Proceedings of the International Conference on Machine Learning (ICML), Virtual, 18–24 July 2021. [Google Scholar] [CrossRef]
  51. Li, X.; Li, D.; Fang, J.; Feng, X. Remote Sensing Image Change Detection Method Based on Change Guidance and Bidirectional Mamba Network. Acta Opt. Sin. 2025, 45, 0528002. [Google Scholar]
  52. Milletari, F.; Navab, N.; Ahmadi, S. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  53. Ji, S.; Wei, S.; Lu, M. Fully Convolutional Networks for Multisource Building Extraction From an Open Aerial and Satellite Imagery Data Set. IEEE Trans. Geosci. Remote Sens. 2019, 57, 574–586. [Google Scholar] [CrossRef]
  54. Zeng, C.; Xue, X.; Li, C.; Xu, X.; Zhao, S.; Xu, Y. StarCD-Net: A Remote Sensing Change Detection Method Combining StarNet and Differential Operators. IEEE Access 2025, 13, 65466–65477. [Google Scholar] [CrossRef]
  55. Song, X.; Hua, Z.; Li, J. Remote Sensing Image Change Detection Transformer Network Based on Dual-Feature Mixed Attention. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  56. Wang, S.; Cheng, D.; Yuan, G.; Li, J. RDSF-Net: Residual Wavelet Mamba-based Differential Completion and Spatio-Frequency Extraction Remote Sensing Change Detection Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 1, 1–16. [Google Scholar] [CrossRef]
Figure 1. Partial samples of the Inner-CD dataset and the pixel imbalance distribution of other datasets. The left image presents a subset of the proposed Inner-CD dataset, while the right image illustrates the imbalanced distribution of changed and unchanged pixels across three datasets.
Figure 1. Partial samples of the Inner-CD dataset and the pixel imbalance distribution of other datasets. The left image presents a subset of the proposed Inner-CD dataset, while the right image illustrates the imbalanced distribution of changed and unchanged pixels across three datasets.
Sensors 25 05832 g001
Figure 2. The overall architecture of DynaNet comprises four key components: the ResNet-18 backbone, the Dynamic Feature Extractor, the Contextual Attention Module, and the Multi-Branch Attention Fusion Module.
Figure 2. The overall architecture of DynaNet comprises four key components: the ResNet-18 backbone, the Dynamic Feature Extractor, the Contextual Attention Module, and the Multi-Branch Attention Fusion Module.
Sensors 25 05832 g002
Figure 3. Structure of the proposed DFE.
Figure 3. Structure of the proposed DFE.
Sensors 25 05832 g003
Figure 4. Structure of the proposed CAM.
Figure 4. Structure of the proposed CAM.
Sensors 25 05832 g004
Figure 5. Structure of the proposed MBAFM.
Figure 5. Structure of the proposed MBAFM.
Sensors 25 05832 g005
Figure 6. Selected samples from the three datasets.
Figure 6. Selected samples from the three datasets.
Sensors 25 05832 g006
Figure 7. Illustrative examples from the Inner-CD dataset, depicting representative bi-temporal image pairs across the four primary change categories: House, Land, Road, and Forest. Subtle yet critical morphological variations are highlighted, including partial demolitions, expansions, land-use modifications, road alterations, and minor structural adjustments in forested regions. These examples underscore the dataset’s emphasis on fine-grained structural evolution and nuanced urban dynamics.
Figure 7. Illustrative examples from the Inner-CD dataset, depicting representative bi-temporal image pairs across the four primary change categories: House, Land, Road, and Forest. Subtle yet critical morphological variations are highlighted, including partial demolitions, expansions, land-use modifications, road alterations, and minor structural adjustments in forested regions. These examples underscore the dataset’s emphasis on fine-grained structural evolution and nuanced urban dynamics.
Sensors 25 05832 g007
Figure 8. The LEVIR-CD dataset displays example outputs of DynaNet and other comparison methods. Pixels are visualized with color encoding, where white represents true positives (TP), black represents true negatives (TN), red represents false positives (FP), and green represents false negatives (FN).
Figure 8. The LEVIR-CD dataset displays example outputs of DynaNet and other comparison methods. Pixels are visualized with color encoding, where white represents true positives (TP), black represents true negatives (TN), red represents false positives (FP), and green represents false negatives (FN).
Sensors 25 05832 g008
Figure 9. The WHU-CD dataset displays example outputs of DynaNet and other comparison methods. Pixels are visualized with color encoding, where white represents true positives (TP), black represents true negatives (TN), red represents false positives (FP), and green represents false negatives (FN).
Figure 9. The WHU-CD dataset displays example outputs of DynaNet and other comparison methods. Pixels are visualized with color encoding, where white represents true positives (TP), black represents true negatives (TN), red represents false positives (FP), and green represents false negatives (FN).
Sensors 25 05832 g009
Figure 10. The Inner-CD dataset displays example outputs of DynaNet and other comparison methods. Pixels are visualized with color encoding, where white represents true positives (TP), black represents true negatives (TN), red represents false positives (FP), and green represents false negatives (FN).
Figure 10. The Inner-CD dataset displays example outputs of DynaNet and other comparison methods. Pixels are visualized with color encoding, where white represents true positives (TP), black represents true negatives (TN), red represents false positives (FP), and green represents false negatives (FN).
Sensors 25 05832 g010
Figure 11. Ablation of Different Modules of DynaNet on Three Datasets. (a) Temporal A, (b) Temporal B, (c) Label, (d) Base, (e) Base + CAM, (f) Base + MBAFM, (g) Base + CAM + DFE, (h) Base + CAM + DFE + MBAFM.
Figure 11. Ablation of Different Modules of DynaNet on Three Datasets. (a) Temporal A, (b) Temporal B, (c) Label, (d) Base, (e) Base + CAM, (f) Base + MBAFM, (g) Base + CAM + DFE, (h) Base + CAM + DFE + MBAFM.
Sensors 25 05832 g011
Figure 12. Visualization comparison of the complexity and number of parameters across different models.
Figure 12. Visualization comparison of the complexity and number of parameters across different models.
Sensors 25 05832 g012
Table 1. Comparison of different change detection methods on the LEVIR-CD dataset, with bold values representing the best experimental results.
Table 1. Comparison of different change detection methods on the LEVIR-CD dataset, with bold values representing the best experimental results.
MethodsF1 (%)Pre (%)Rec (%)IOU (%)
FC-Sima-conc [19]86.3189.5383.3177.21
FC-Sima-diff [19]87.3589.5382.4575.14
STANet [23]89.1790.6887.7080.45
SNUNet [20]88.3091.2585.5579.06
Changeformer [29]89.8291.8587.8881.52
BIT [33]89.9691.7488.2581.76
StarCD-Net [54]90.2391.4390.1582.30
LGPNet [35]89.3792.1386.3280.74
DMATNet [55]89.9790.7889.1781.83
DTCDSCN [44]87.6788.5386.8378.05
ConvTransNet [31]90.4391.4787.6482.56
WNet [32]90.6791.1690.1882.93
RDSF-Net [56]91.2391.3490.4583.79
DynaNet (Ours)92.3893.4591.4684.57
Table 2. Comparison of different change detection methods on the WHU-CD dataset, with bold values representing the best experimental results.
Table 2. Comparison of different change detection methods on the WHU-CD dataset, with bold values representing the best experimental results.
MethodsF1 (%)Pre (%)Rec (%)IOU (%)
FC-Sima-conc [19]84.7683.6286.4573.56
FC-Sima-diff [19]85.7786.5385.0275.07
STANet [23]88.2389.4087.1078.85
SNUNet [20]89.5087.6088.4979.06
Changeformer [29]89.4793.3383.7980.52
BIT [33]89.9692.2488.2583.48
StarCD-Net [54]92.3591.4390.7485.30
LGPNet [35]87.0790.8485.5379.74
DMATNet [55]90.8891.3889.6881.70
DTCDSCN [44]71.9585.1385.1277.06
ConvTransNet [31]92.1192.6691.5785.38
WNet [32]91.2592.3790.1583.91
RDSF-Net [56]92.1494.6590.5385.67
DynaNet (Ours)94.3593.2892.4687.57
Table 3. Comparison of different change detection methods on the Inner-CD dataset, with bold values representing the best experimental results.
Table 3. Comparison of different change detection methods on the Inner-CD dataset, with bold values representing the best experimental results.
MethodsF1 (%)Pre (%)Rec (%)IOU (%)
FC-Sima-conc [19]75.9473.9977.9961.26
FC-Sima-diff [19]78.7977.5478.3268.07
STANet [23]83.1690.0478.0871.78
SNUNet [20]86.3687.6085.1675.25
Changeformer [29]86.5489.2683.9876.27
BIT [33]85.1587.4684.9375.68
StarCD-Net [54]85.4185.0786.2676.53
LGPNet [35]85.3584.7586.4874.52
DMATNet [55]87.4892.2483.1977.75
DTCDSCN [44]85.0084.8785.3273.91
ConvTransNet [31]86.7388.5285.2976.79
WNet [32]87.7889.7385.9278.23
RDSF-Net [56]87.2489.6784.7577.96
DynaNet (Ours)90.9293.8087.8480.37
Table 4. Ablation study results on the LEVIR-CD, WHU-CD, and Inner-CD Datasets. Bold values represent the best experimental results. ✓ indicates that the module is introduced into the baseline model.
Table 4. Ablation study results on the LEVIR-CD, WHU-CD, and Inner-CD Datasets. Bold values represent the best experimental results. ✓ indicates that the module is introduced into the baseline model.
NO.DFECAMMBAFMLEVIR-CDWHU-CDInner-CD
F1(%) IOU(%) F1(%) IOU(%) F1(%) IOU(%)
1 83.2971.3777.6563.5878.7165.42
2 85.5375.2180.5470.5580.1169.16
3 86.7175.0081.9472.3681.4568.42
4 87.0177.5984.3475.9183.2770.34
5 89.0179.5988.6283.8986.7973.17
6 90.2181.4890.9481.2887.9475.36
7 90.7282.2192.2985.6988.8578.10
892.3884.5794.3587.5790.9280.37
Table 5. Ablation experimental results of DFE on three datasets, with bold values representing the best experimental results. NO.1 in the table corresponds to NO.6 in Table 4.
Table 5. Ablation experimental results of DFE on three datasets, with bold values representing the best experimental results. NO.1 in the table corresponds to NO.6 in Table 4.
NO.DEConvCAMLEVIR-CDWHU-CDInner-CD
F1(%) IOU(%) F1(%) IOU(%) F1(%) IOU(%)
1 90.2181.4890.9481.2887.9475.36
2 91.9683.5693.1386.2588.7378.82
3 90.6782.0192.3683.7288.4277.54
492.3884.5794.3587.5790.9280.37
Table 6. Experimental comparison of the complexity and number of parameters across different models.
Table 6. Experimental comparison of the complexity and number of parameters across different models.
MethodsParams (M)Flops (G)ΔF1 (%)
FC-Sima-conc [19]1.554.86−14.98
FC-Sima-diff [19]1.354.73−12.13
STANet [23]37.5986.77−7.76
SNUNet [20]12.0346.79−5.13
Changeformer [29]41.03202.86−4.38
BIT [33]11.898.71−5.77
LGPNet [35]70.99125.79−3.57
StarCD-Net [54]7.4615.74−5.51
DTCDSCN [44]41.0714.42−5.92
ConvTransNet [31]7.1330.53−3.19
RDSF-Net [56]27.3015.60−3.68
DynaNet (Ours)12.728.360
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Li, D.; Fang, J.; Feng, X. DynaNet: A Dynamic Feature Extraction and Multi-Path Attention Fusion Network for Change Detection. Sensors 2025, 25, 5832. https://doi.org/10.3390/s25185832

AMA Style

Li X, Li D, Fang J, Feng X. DynaNet: A Dynamic Feature Extraction and Multi-Path Attention Fusion Network for Change Detection. Sensors. 2025; 25(18):5832. https://doi.org/10.3390/s25185832

Chicago/Turabian Style

Li, Xue, Dong Li, Jiandong Fang, and Xueying Feng. 2025. "DynaNet: A Dynamic Feature Extraction and Multi-Path Attention Fusion Network for Change Detection" Sensors 25, no. 18: 5832. https://doi.org/10.3390/s25185832

APA Style

Li, X., Li, D., Fang, J., & Feng, X. (2025). DynaNet: A Dynamic Feature Extraction and Multi-Path Attention Fusion Network for Change Detection. Sensors, 25(18), 5832. https://doi.org/10.3390/s25185832

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop