Next Article in Journal
Optimizing Age of Information in Internet of Vehicles over Error-Prone Channels
Previous Article in Journal
Real-Time Detection of IoT Anomalies and Intrusion Data in Smart Cities Using Multi-Agent System
Previous Article in Special Issue
An Enhanced Feature-Fusion Network for Small-Scale Pedestrian Detection on Edge Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Language-Guided Semantic Clustering for Remote Sensing Change Detection

B-DAT and CICAEET, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 7887; https://doi.org/10.3390/s24247887
Submission received: 24 October 2024 / Revised: 4 December 2024 / Accepted: 8 December 2024 / Published: 10 December 2024
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection: 2nd Edition)

Abstract

:
Existing learning-based remote sensing change detection (RSCD) commonly uses semantic-agnostic binary masks as supervision, which hinders their ability to distinguish between different semantic types of changes, resulting in a noisy change mask prediction. To address this issue, this paper presents a Language-guided semantic clustering framework that can effectively transfer the rich semantic information from the contrastive language-image pretraining (CLIP) model for RSCD, dubbed LSC-CD. The LSC-CD considers the strong zero-shot generalization of the CLIP, which makes it easy to transfer the semantic knowledge from the CLIP into the CD model under semantic-agnostic binary mask supervision. Specifically, the LSC-CD first constructs a category text-prior memory bank based on the dataset statistics and then leverages the CLIP to transform the text in the memory bank into the corresponding semantic embeddings. Afterward, a CLIP adapter module (CAM) is designed to fine-tune the semantic embeddings to align with the change region embeddings from the input bi-temporal images. Next, a semantic clustering module (SCM) is designed to cluster the change region embeddings around the semantic embeddings, yielding the compact change embeddings that are robust to noisy backgrounds. Finally, a lightweight decoder is designed to decode the compact change embeddings, yielding an accurate change mask prediction. Experimental results on three public benchmarks including LEVIR-CD, WHU-CD, and SYSU-CD demonstrate that the proposed LSC-CD achieves state-of-the-art performance in terms of all evaluated metrics.

1. Introduction

Remote sensing change detection (RSCD) involves taking a pair of bi-temporal images as input that differ in time but share the same spatial location, aiming to identify the regions that have changed between the bi-temporal images. RSCD has been widely applied in various surface monitoring scenarios such as disaster assessment [1,2,3], urban planning [4,5,6], and agricultural management [7,8,9].
The traditional methods in RSCD [10,11,12,13,14,15,16,17] identify changed regions in an unsupervised manner by analyzing handcrafted features. However, due to the limited discriminability of handcrafted features, traditional methods struggle to handle increasingly complex RSCD scenarios. Afterward, with the rapid development of deep learning (DL), the data-driven DL paradigm has demonstrated superior performance in learning discriminative feature representations, breaking the representative limitations of the handcrafted features, thereby attracting much attention in RSCD. Daudt et al. [18] introduced a Siamese architecture, consisting of two parallel subnetworks for RSCD. These subnetworks simultaneously extract the features from bi-temporal images and directly analyze changed features. Inspired by this structure, numerous change detection models have emerged that can be divided into two major categories: the convolutional neural network (CNN)-based and the transformer-based methods. The CNN-based methods [18,19,20,21,22,23,24,25,26] aim to enhance local feature discrimination capabilities by using the convolutional architecture [27]. In contrast, to reduce the inductive bias of the CNN architecture, the transformer-based methods [28,29,30,31,32,33,34] focus on improving the long-range modeling abilities via introducing the self-attention mechanism into transformer architectures [35].
Despite the demonstrated success of these methods, researchers have only used a semantic-agnostic binary ground-truth mask as supervision to train their model, which may suffer from land cover classification ambiguity issues. This is due to the fact that in highly complicated and dynamic varying environment scenarios, remote sensing images often contain hardly classified regions consisting of a variety of different land cover types with similar color or texture appearances, especially for regions across land cover boundaries. As shown in Figure 1, the input images contain interferences of changed land cover in different categories but have similar texture and color appearances. This is prone to causing the visual representations of the different land cover types learned by these methods to be indistinguishable in the visual space, leading to irrelevant pseudo-changes in the change mask (see Figure 1a). To address this issue, it is necessary to explore an effective framework for learning more discriminative representations in RSCD.
Recently, vision-language representation learning has gained significant attention in computer vision [38,39,40], focusing on learning robust representations from image–text pairs. This paradigm introduces multimodal information to assist visual representation learning. Following this paradigm, the widely adopted CLIP [38] uses a contrastive learning strategy on large-scale image–text pairs to project the text and visual features into the same latent space, termed as visual-semantic space. The CLIP has demonstrated outstanding zero-shot generalization for various downstream tasks, such as Text4point [41], Gridclip [42], and Clip2scene [43]. This motivated us to utilize the pretrained CLIP to construct visual-semantic space to assist in representation learning in RSCD.
To this end, this paper presents a language-guided semantic clustering framework for RSCD, which uses clustering with semantic embeddings to modulate change region embeddings, effectively categorizing different types of changes in visual-semantic space. Specifically, the text superior memory bank is first constructed based on common classes in RSCD datasets, and the CLIP is used to map this memory bank into semantic embeddings. Meanwhile, bi-temporal images are encoded into changed region embeddings through an encoder [44]. Then, a lightweight CLIP adapter module (CAM) is designed to modulate the semantic embeddings, aligning them adaptively with the changed region embeddings. Next, the semantic clustering module (SCM) is applied to group the change region embeddings around the semantic embeddings. Finally, a lightweight decoder decodes the modulated change region embeddings, with the entire process constrained by a loss function. Extensive experiments on three public benchmarks including LEVIR-CD [45], WHU-CD [23], and SYSU-CD [46] demonstrate the superiority of the proposed LSC-CD over various state-of-the-art methods.
The main contributions of this paper can be summarized as follows:
  • We propose a language-guided semantic clustering framework for RSCD that achieves strong semantic modeling capabilities by introducing the CLIP.
  • We design an SCM that modulates change embeddings from a semantic-guided perspective using semantic clustering.
  • We develop a CAM to fine-tune the CLIP text encoder, aligning its encoded semantic embeddings to the change region embeddings.

2. Related Work

This section primarily explores the recent research related to our proposed method, focusing on two key components: semantic segmentation and RSCD. In semantic segmentation, we primarily discuss relevant studies in the field of computer vision, which inspired our approach. In RSCD, we review the existing mainstream change detection methods, highlighting their strengths and limitations, and introduce the research gap to emphasize the significance of our proposed method.

2.1. Semantic Segmentation

Semantic segmentation is a fundamental task in computer vision, aiming to assign a class label to every pixel in an image. RSCD is an important downstream task of semantic segmentation. Currently, most semantic segmentation models follow the structure of FCN [47], which employs an encoder–decoder framework. Subsequent works have introduced various optimizations. For example, some studies [48,49,50] have replaced the encoder with architectures from classification tasks, such as VGG [51], ResNet [52], and HRNet [53]. With the emergence of transformer [35], several works [54,55,56,57] have incorporated attention mechanisms to enhance the ability to capture long-range dependencies. Broadly, these methods can be classified into two types: the class-label-based semantic segmentation and the pretrained semantic segmentation models.
The class-label-based semantic segmentation models [58,59,60,61,62] rely on large amounts of pixel-level annotations as supervision, using encoder–decoder structures to directly learn semantic information from images in a data-driven manner. However, collecting pixel-level annotations requires significant resources. To address this issue, pretrained semantic segmentation methods have emerged. These methods [63,64,65,66,67,68] typically leverage pretrained weights to assist semantic segmentation, significantly reducing the need for extensive pixel-level annotations and cutting costs. However, the performance of these methods is often limited by the pretrained weights. The early approaches [63,64,65] mostly utilized the large-scale ImageNet [69] classification datasets as pretrained data, using image-level labels to aid semantic presentation learning. Due to the limitations of ImageNet data, models pr-trained on ImageNet often exhibit incomplete semantic modeling capabilities for segmentation tasks. The emergence of large-scale models such as the CLIP [38] has partially mitigated these challenges. Subsequent works [66,67,68] have begun applying large-scale models to semantic segmentation tasks, achieving superior performance.
In this work, we applied the CLIP to RSCD. Specifically, we leveraged the pretrained CLIP text encoder to encode semantic information and used a lightweight and effective CLIP adapter module to align the semantic information with the changed features.

2.2. Remote Sensing Change Detection

Remote sensing change detection methods can be categorized into traditional methods and deep learning-based methods.
Traditional methods include algebra-based methods, statistics-based methods, and trans -formation-based methods [70]. Algebra-based methods generate a change map using techniques such as image difference [10], image regression [11], image ratioing [12], and change vector analysis (CVA) [13]. The performance of these methods relies heavily on the selection of empirical thresholds. Statistics-based methods analyze the statistical properties of the image or parts of it, deriving a change map based on pixel distribution and the corresponding attributes. However, the effectiveness of these methods is constrained by the reliability of the statistical properties. Transformation-based methods employ techniques such principal component analysis (PCA) [14], multivariate alteration detection (MAD) [15], Gramm–Schmidt transformation [16], and tasseled cap transformation [17] to transform the images in a way that enhances change regions while suppressing unchanged areas. Nonetheless, these methods still depend on empirical thresholds. For example, PCA requires manually set parameters, limiting its ability to adaptively learn changed features. Overall, these traditional methods are overly reliant on handcrafted features, lack robustness, and struggle to effectively address the increasingly complex scenarios in remote sensing change detection.
Existing state-of-the-art methods are mostly based on deep learning in RSCD, which breaks through the limitations of traditional methods by adopting a data-driven strategy to adaptively learn changed features. Throughout the development of deep learning technologies, two main paradigms have emerged: CNN-based methods and transformer-based methods. The CNN-based methods leverage the powerful nonlinear modeling capabilities of convolutional neural networks (CNNs) to directly extract changed features from bi-temporal images and decode a corresponding change map, breaking the limitations of traditional handcrafted feature-based methods. Wang et al. [19] introduced faster region-based convolutional neural network (Faster R-CNN) into change detection and achieved notable success. Amin et al. [20] further improved performance by incorporating a Siamese architecture into CNN-based approaches. Building on this, Daudt et al. [18] proposed three fully convolutional change detection models that refined the Siamese structure. Liu et al. [21] combined the UNet [71] architecture with Siamese networks, interpreting change detection as an image translation problem. Fang et al. [25] optimized the fusion of deep and shallow features in the Siamese framework, designing the ECAM module to guide the model in learning effective representations. Yin et al. [22] introduced attention mechanisms into CNN-based models, further enhancing their performance. The aforementioned CNN-based methods have significantly advanced remote sensing change detection, but they are limited by their lack of long-range modeling capabilities, particularly in complex scenarios.
To overcome this limitation, transformer-based methods, recognized for their superior long-range modeling abilities, have emerged as promising solutions in bi-temporal remote sensing change detection problems. Chen et al. [28] introduced a transformer encoder to model spatial contexts based on semantic tokens, enhancing long-range modeling for bi-temporal remote sensing change detection. Yu et al. [34] proposed a global context-aware transformer structure, improving global context learning to address bi-temporal remote sensing change detection. Bandara et al. [36] designed a transformer-based Siamese network architecture, incorporating multi-scale change features to solve the bi-temporal remote sensing change detection problem. Zhang et al. [33] combined a transformer-based Siamese network with a UNet architecture to extract effective global information for this problem. Li et al. [30] proposed a hybrid transformer structure, aiming to leverage the strengths of both CNNs and transformers to learn local-global context features for bi-temporal remote sensing change detection.
While transformer-based methods have broken through the inherent limitations of CNNs due to their superior long-range modeling capabilities, they often rely on binary masks as supervision, lacking effective semantic label guidance. As a result, the semantic representations learned by these methods are constrained. To address this, we propose a semantic-guided transformer-based Siamese network. By introducing semantic clustering centers at the input stage, our approach guides semantic representation learning, compensating for the absence of semantic label guidance at the output stage.

3. Methods

This section presents the technical details of the proposed LSC-CD, which consists of four main sections: Overview, CLIP-Adapter Module, Semantic Clustering Module, and Loss Function. In the Overview Section, we introduce the overall pipeline of LSC-CD. In the CLIP-Adapter Module and Semantic Clustering Module Section, we begin by explaining the motivation behind the design of each module, followed by a detailed description of their implementation. In the Loss Function Secction, we explain the supervised loss function we developed for LSC-CD, based on supervised RSCD.

3.1. Overview

Figure 2 illustrates the pipeline of the proposed LSC-CD. First, the bi-temporal images I 1 , I 2 R H × W × 3 are encoded into multi-scale change features F 1 i , F 2 i i = 1 4 using a Segformer encoder [44], where H denotes the height of the image, W denotes the width of the image, and i refers to the i-th scale of the features. Meanwhile, the text superior memory bank T = t 1 , t 2 , , t k is encoded into semantic embeddings S R K × C through a pretrained CLIP text encoder, where K denotes the number of prior texts, and C represents the channel dimension. Next, the CAM is applied to adaptively modulate the semantic embeddings S to generate the modulated embeddings S ^ R K × C . Then, S ^ and F 1 i , F 2 i are fed into the SCM, where S ^ serve as clustering centers to modulate F 1 i , F 2 i , resulting in the modulated change embeddings F ^ 1 i , F ^ 2 i . Finally, the concatenated F ^ 1 i , F ^ 2 i is passed through a lightweight decoder [36] to generate a change map.

3.2. CLIP-Adapter Module (CAM)

To obtain semantic information that guides the change detection model, we designed a CAM based on CLIP, aiming to directly incorporate the outstanding semantic knowledge from the CLIP into RSCD. In the CAM, considering the complexity of the RSCD scenario, we encounter two challenges. First, it is impractical to anticipate all possible text descriptions of subcategories. Second, overly detailed subcategories may be subject to the fluctuations influenced by environmental factors. For example, the distinction between dense forests and sparse forests may arise from seasonal changes rather than real changes that have occurred. Consequently, fine-grained prior change detection text memory may not necessarily benefit the change detection model in identifying changes. Therefore, we establish a coarse-grained text memory bank T , which is described in Table 1, and use a pretrained CLIP text encoder to encode it into corresponding semantic embeddings. In the encoding process, we focus solely on encoding the category information while avoiding the detailed encoding of each subcategory to prevent excessive description. This process can be formalized as
S k = E n c o d e r t ( t k ) ,
where E n c o d e r t represents the text encoder of the CLIP, S k R 1 × C denotes the semantic embedding corresponding to the k-th text, and k represents the index of the coarse-grained text memory bank T .
Although the CLIP can directly encode semantic information by introducing a coarse-grained text memory bank T , there is a misalignment between the semantic embeddings S and the change embeddings F 1 i , F 2 i . To align the semantic embeddings from the CLIP with the change embeddings, we designed a lightweight CLIP adapter module that adaptively modulates the encoded semantic embeddings to align with the change embeddings. The CLIP adapter module consists of two layers of MLP [72] and a residual structure [52]. This process can be formulated as
S ^ = α R e l u ( S W 1 ) W 2 + ( 1 α ) S ,
where R e l u ( · ) denotes the ReLU function [73], S ^ R K × C represents the modulated semantic embeddings, K denotes the length of the text memory, C denotes channel dimension. W 1 , W 2 denote the adaptive weights of the MLP [72], and α signifies the residual ratio, which is set to 0.7.

3.3. Semantic Clustering Module (SCM)

To fully integrate the modulated semantic embeddings S ^ obtained from the CAM with the change embeddings F 1 i , F 2 i , we designed the SCM based on clustering, as illustrated in Figure 3. In SCM, the semantic embeddings S ^ serve as clustering centers, while the F 1 i , F 2 i are treated as the features to be clustered. Given that we utilize a coarse-grained text memory bank T , there may be misalignment between the semantic embeddings of categories and the visual embeddings of specific subcategories. To address this issue, we developed the SEUM to adjust the S ^ within the SCM. The structure of the SEUM is illustrated in Figure 4.
In the SEUM, we designed an interactive attention mechanism that utilizes the change embeddings F 1 i , F 2 i to update the S ^ . This process can be formalized as
S ^ j i = M L P ( N o r m ( A t t e n t i o n ( S ^ , F j i , F j i ) ) ) ,
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T ) V ,
s o f t m a x ( z i ) = e x p ( z i ) j e x p ( z j ) ,
where A t t e n t i o n ( · ) represents the cross-attention mechanism, and N o r m ( · ) denotes normalization [35]. M L P ( · ) refers to a single-layer multi-layer perceptron [72]. s o f t m a x ( · ) represents the activation function, i 1 , 2 , 3 , 4 , and j 1 , 2 .
After the SEUM, to fuse the semantic information and change embeddings F 1 i , F 2 i , we use cosine similarity as the metric to cluster F 1 i , F 2 i , with the modulated semantic embeddings S ^ serving as the clustering centers. This process can be expressed as
F j i = β s i m ( F j i , S ^ j i ) S j i T + ( 1 β ) F j i ,
s i m ( A , B ) = A B A B ,
where β is the modulation ratio, set to 0.5. s i m ( · , · ) represents the cosine similarity.

3.4. Loss Funcation

The proposed LSC-CD is a supervised change detection method, where the entire learning process is guided by binary semantic-agnostic masks. During training, we employ the binary cross-entropy loss b c e [74] to supervise the learning of changed features.
In this paper, the binary cross-entropy loss can be formalized as
B C E = 1 H × W i H j W [ y i log y ^ i + ( 1 y i ) log ( 1 y ^ i ) ] ,
where H and W represent the height and width of the image, respectively; y i denotes the ground truth (GT) label of the i-th pixel, and y ^ i refers to the predicted value for the i-th pixel in the LSC-CD output.

4. Experiments

This section primarily describes the experimental process, including six subsections: Datasets, Evaluation Metrics, Implementation Details, Quantitative Analysis, Qualitative Analysis, and Ablation Studies. In the Datasets Section, we provide a comprehensive description of three public datasets. In the Evaluation Metrics Section, we present five metrics used to assess the experimental results. In the Implementation Details Section, we outline the experimental environment and the settings for hyperparameters. In the Quantitative Analysis Section, we analyze the differences between our method and other methods from a qualitative perspective. In the Qualitative Analysis Section, we quantitatively examine the differences between our method and other methods, incorporating the evaluation metrics. Finally, in the Ablation Studies Section, we analyze the impact of different modules on our method.

4.1. Datasets

  • LEVIR-CD. The LEVIR-CD is a large-scale remote sensing dataset for building change detection models, released by Beihang University in 2020. It consists of 637 pairs of very-high-resolution Google Earth images with a resolution of 1024 × 1024 pixels, where each pixel represents 0.2 m. The dataset primarily focuses on building changes, including building growth and demolition, with change areas labeled as 1 and unchanged areas labeled as 0, without semantic labels. It contains many irrelevant changes due to seasonal variations. The buildings in LEVIR-CD range from villas, small garages, and high-rise apartments to large warehouses, with a total of 31,333 changed buildings. Following [36], we split the original data into 256 × 256 patches without overlap, resulting in 10,192 images, of which 7120 are used for training, 2048 for testing, and 1024 for validation.
  • WHU-CD. The WHU-CD is another building change detection dataset for remote sensing, similar to LEVIR-CD. It contains two aerial images with a resolution of 32,507 × 15,354 pixels, where each pixel represents 0.3 m. Like LEVIR-CD, it does not provide additional semantic labels. This dataset has been widely used as a benchmark for change detection studies. However, due to the large resolution of the image pairs, they need to be divided into smaller patches for model input, and there is no standard method for this splitting. To ensure a fair comparison, we followed [37] for splitting the dataset. Specifically, we divided the original images into 256 × 256 pixel patches, resulting in 7620 image pairs. Of these, 6096 pairs are used for training, 762 pairs for testing, and 762 pairs for validation.
  • SYSU-CD. The SYSU-CD contains 20,000 pairs of high-resolution images, each with 256 × 256 pixels, collected in Hong Kong between 2007 and 2014, where each pixel represents 0.5 m. Unlike LEVIR-CD and WHU-CD, SYSU-CD includes a variety of changes, including newly built urban buildings, suburban expansion, groundwork before construction, vegetation changes, road expansion, and sea construction. Change areas are labeled as 1 and unchanged areas as 0, without semantic labels. The dataset consists of 12,000 pairs for training, 4000 pairs for testing, and 4000 pairs for validation.

4.2. Evaluation Metrics

In the experiment, we adopted five common evaluation metrics to measure the model’s ability to segment changed regions, including precision (Pr), recall (Re), overall accuracy (OA), F1 score (F1), and the intersection over union (IoU). Among these, F1 is used as the primary evaluation metric, with the other four metrics being considered in conjunction with F1. These five metrics can be formulated as follows:
P r = T P T P + F P ,
R e = T P T P + F N ,
O A = T P + T N T P + F P + T N + F N ,
F 1 = 2 × P r × R e P r + R e ,
I o U = T P T P + F P + F N ,
where TP represents true positives, indicating the number of pixels correctly predicted as changed. FP represents false positives, indicating the number of pixels incorrectly predicted as changed when no change occurred. TN represents true negatives, indicating the number of pixels correctly predicted as unchanged. FN represents false negatives, indicating the number of pixels incorrectly predicted as unchanged when a change occurred.

4.3. Implementation Details

We deployed our proposed LSC-CD within a PyTorch framework and used an NVIDIA RTX 3090 GPU. The RTX 3090 is designed and manufactured by NVIDIA Corporation, whose headquarters are located in Santa Clara, CA, USA. During the training, to ensure fairness, we applied four commonly used data augmentation techniques: random flips, Gaussian blurs, random cropping, and random color dithering. The network took 256 × 256 images as the input, with a batch size of 8. The image encoder utilized the Segformer encoder’s pretrained weights, while the rest of the network was randomly initialized. The entire process was supervised by a binary cross-entropy loss function. Adam [75] was used as the optimizer, with a learning rate of 0.0001, decaying by 0.00002 every 50 epochs, for a total of 200 training epochs.

4.4. Quantitative Comparison

To ensure fairness, we conducted comparisons on three public datasets: LEVIR-CD, WHU-CD, and SYSU-CD. We evaluated various methods across these datasets. Due to limitations regarding the availability of open-source code, for some methods, we directly referenced the results reported in the original papers. As a result, some data may be missing, and we marked these with the symbol “-”. We compared 15 different methods, including FC-EF [18], FC-Siam-Di [18], FC-Siam-Conc [18], STANet [23], IFNet [24], SNUNet-CD [25], BIT [28], TransUNetCD [30], ChangeFormer [36], StransUNet [31], ICIF-Net [29], AMCA [32], DMINet [26], GCD-DDPM [37], and GCFormer [34], with GCD-DDPM and GCFormer being the latest methods.
Table 2 lists the performance of the various change detection methods on the LEVIR-CD dataset. The results showed that our proposed LSC-CD achieved state-of-the-art performance. LSC-CD attained an F1 score of 92.01, which was the best performance. In addition, LSC-CD performed well in terms of precision, IoU, and overall accuracy. It is worth noting that LSC-CD did not achieve the best performance in recall. This is because the recall metric considers both true positives and false negatives, and some change detection models tended to predict more background, which reduces false negatives and increased recall. However, This tendency does not align with practical demands, making recall alone an insufficient reflection of a change detection model’s true performance. Therefore, a comprehensive evaluation of all metrics is necessary.
Table 3 lists the performance of various change detection methods on the WHU-CD dataset. The results show that our proposed LSC-CD achieved the best performance in F1, recall, IoU, and overall accuracy, with an F1 score of 94.75. This demonstrates that the introduction of effective semantic modeling in LSC-CD led to outstanding results. Notably, LSC-CD only achieved the second-best performance in precision. Similar to recall, precision is influenced by true positives and false positives. A high precision indicates that the change detection model tends to predict more foreground, which can result in many irrelevant changes being detected. This tendency is evident in earlier methods like IFNet, which achieved the best precision but did not perform as well overall compared to current state-of-the-art methods. In contrast, LSC-CD delivered the best overall performance, further validating the effectiveness of our proposed approach.
Table 4 lists the comparative results of several methods on SYSU-CD. Since SYSU-CD is a multi-class change detection dataset, it is more challenging compared to LEVIR-CD and WHU-CD, as reflected in the performance metrics. Due to the unavailability of the code for some recent works, we only compared algorithms with accessible code. From the results of the comparison, similar to LEVIR-CD and WHU-CD, our proposed LSC-CD demonstrates outstanding performance across several metrics, achieving an F1 score of 81.29, achieves state-of-the-art performance. The strong performance on SYSU-CD indicates that the integration of CLIP semantic knowledge into LSC-CD significantly enhances the overall performance of the change detection model.

4.5. Qualitative Comparison

To ensure fairness, we conducted the qualitative comparison on three public datasets: LEVIR-CD, WHU-CD, and SYSU-CD, similar to the quantitative comparison. We visualized selected test samples from five different methods, including FC-EF [18], BIT [28], ChangeFormer [36], ICIFNet [29], and DMINet [26].
Figure 5 demonstrates the qualitative results of the different methods on the LEVIR-CD dataset. LEVIR-CD is a building change detection dataset, where only building changes are considered changes, which can be observed from the ground truth binary mask (GT). This dataset primarily tests the robustness of change detection models in handling no-task-specific changes. It is apparent from the findings that our proposed LSC-CD yielded masks that were nearly identical to the ground truth binary masks (GT). Notably, LSC-CD showed the fewest false positives, as reflected by the smallest red areas in Figure 5. This demonstrates that by incorporating effective semantic guidance, LSC-CD can effectively suppress the interference of no-task-specific changes.
Figure 6 illustrates a quantitative comparison of different methods on the WHU-CD dataset. The results show that our proposed LSC-CD predicted the fewest false positives while closely aligning with the ground truth, indicating that LSC-CD effectively suppresses no-task-specific changes.
Figure 7 illustrates a quantitative comparison of different methods on the SYSU-CD dataset. Unlike LEVIR-CD and WHU-CD, which focus on single-class change detection, SYSU-CD is a multi-class change dataset containing more irrelevant disturbances. The results show that other methods introduced a significant amount of no-task-specific changes. In contrast, our method effectively overcame this issue, demonstrating that LSC-CD, by incorporating semantic information to guide change feature learning, can effectively suppress no-task-specific changes in multi-class change scenarios.

4.6. Ablation Studies

To validate the effectiveness of the proposed CAM and SCM, we conducted ablation experiments on the LEVIR-CD dataset. We compared five variants: the baseline (ChangeFormer [36]), which used the same encoder and decoder as our approach; the baseline with the CLIP method; the baseline with the CLIP and CAM method; the baseline with the CLIP method, CAM, and SCM without the SEUM method; and the full LSC-CD method. The comparison results are listed in Table 5. From the results, the baseline with the CLIP method achieved an improvement in F1 and IoU scores by 0.44 and 0.74, respectively, compared to the baseline, demonstrating the importance of semantic information for the RSCD task. The baseline with the CLIP and CAM method further improved the F1 and IoU scores by 0.24 and 0.4, respectively, indicating that modulating semantic embeddings to align change embeddings is necessary. The method with SCM integrated the modulated semantic embeddings with the change embeddings, resulting in significant increases of 0.61 and 1.04 in F1 and IoU, respectively, showing that using coarse-grained clustering facilitates the fusion of semantic and change embeddings. Finally, the full LSC-CD, which incorporated the SEUM module to refine clustering centers, boosted the F1 and IoU scores by 0.32 and 0.49, respectively, indicating that the introduction of the SEUM module helps improve the RSCD performance.
Figure 8 presents a comparison of the attention distributions between our proposed LSC-CD and ChangeFormer. The results indicate that our proposed LSC-CD, by introducing CAM and SCM to guide change feature clustering with semantic information, achieved a more refined attention distribution for both the heatmaps during the encoding process and the final concatenated heatmap compared to the baseline (ChangeFormer).
To the best of our knowledge, we are not the first to apply the CLIP to the RSCD task. Before the development of LSC-CD, ChangeCLIP [76] was another viable approach. Both methods share the use of the prior text memory bank but differ in that, on the one hand, the proposed LSC-CD uses only coarse-grained categories as clustering centers without the need for fine-grained text prompts. On the other hand, LSC-CD incorporates semantic guidance for change embeddings encoding through clustering in the encoder without fusing at the decoder. Table 6 compares the performance of both methods, using the same CLIP, on the LEVIR-CD dataset. The results demonstrate that LSC-CD generally outperforms ChangeCLIP, validating the effectiveness of introducing semantic information into the encoder.

5. Conclusions

In this paper, we proposed the LSC-CD, a language-guided semantic cluster framework for RSCD. First, we introduced a category text memory bank based on dataset statistics. Then, the CLIP text encoder was utilized to transform the bank into the corresponding semantic embeddings. Afterward, we designed the CAM to fine-tune the semantic embeddings to align with the change region embeddings encoded from the bi-temporal images. Next, we developed the SCM to cluster the change region embeddings around the semantic embeddings in the visual-semantic space. Finally, we introduced a lightweight decoder to decode the change region embeddings after the SCM to yield the change mask prediction. The promising performance on three public RSCD benchmarks demonstrated that language-guided semantic clustering helps RSCD to distinguish noisy masks.
Although the proposed LSC-CD achieved excellent performance to a certain extent, we used additional text semantic information and introduced extra parameters. This kind of text semantic information needs to be provided in advance. We plan to utilize weak semantic segmentation to extract reliable semantic information from the bi-temporal images in a lightweight manner without additional information.

Author Contributions

Conceptualization, S.H.; methodology, S.H., H.S. and K.Z.; software, S.H., Y.B. and B.C.; validation, S.H., Y.B. and B.C.; formal analysis, S.H., H.S. and K.Z.; investigation, S.H., Y.B. and H.S.; resources, H.S. and K.Z.; writing—original draft preparation, S.H., Y.B. and B.C.; writing—review and editing, S.H., H.S. and K.Z.; visualization, Y.B.; funding acquisition, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

The authors express their gratitude to the editor and the reviewers for their constructive feedback, which significantly enhanced the quality of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Qin, D.; Zhou, X.; Zhou, W.; Huang, G.; Ren, Y.; Horan, B.; He, J.; Kito, N. MSIM: A change detection framework for damage assessment in natural disasters. Expert Syst. Appl. 2018, 97, 372–383. [Google Scholar] [CrossRef]
  2. Michel, U.; Thunig, H.; Ehlers, M.; Reinartz, P. Rapid change detection algorithm for disaster management. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 1, 107–111. [Google Scholar] [CrossRef]
  3. Zheng, Z.; Zhong, Y.; Wang, J.; Ma, A.; Zhang, L. Building damage assessment for rapid disaster response with a deep object-based semantic change detection framework: From natural disasters to man-made disasters. Remote Sens. Environ. 2021, 265, 112636. [Google Scholar] [CrossRef]
  4. Bolorinos, J.; Ajami, N.K.; Rajagopal, R. Consumption change detection for urban planning: Monitoring and segmenting water customers during drought. Water Resour. Res. 2020, 56, e2019WR025812. [Google Scholar] [CrossRef]
  5. Du, P.; Liu, S.; Gamba, P.; Tan, K.; Xia, J. Fusion of difference images for change detection over urban areas. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1076–1086. [Google Scholar] [CrossRef]
  6. Liu, X.; Lathrop, R., Jr. Urban change detection based on an artificial neural network. Int. J. Remote Sens. 2002, 23, 2513–2518. [Google Scholar] [CrossRef]
  7. Tarimo, B.; Mtalo, E.; Liwa, E. Land use change detection and impact assessment on an agricultural area. J. Sustain. Dev. 2013, 6, 55. [Google Scholar] [CrossRef]
  8. Prishchepov, A.V.; Radeloff, V.C.; Dubinin, M.; Alcantara, C. The effect of Landsat ETM/ETM+ image acquisition dates on the detection of agricultural land abandonment in Eastern Europe. Remote Sens. Environ. 2012, 126, 195–209. [Google Scholar] [CrossRef]
  9. Malinverni, E.S.; Rinaldi, M.; Ruggieri, S. Agricultural crop change detection by means of hybrid classification and high resolution images. EARSeL EProc. 2012, 11, 132. [Google Scholar]
  10. Bruzzone, L.; Prieto, D.F. Automatic analysis of the difference image for unsupervised change detection. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1171–1182. [Google Scholar] [CrossRef]
  11. Luppino, L.T.; Bianchi, F.M.; Moser, G.; Anfinsen, S.N. Unsupervised image regression for heterogeneous change detection. arXiv 2019, arXiv:1909.05948. [Google Scholar] [CrossRef]
  12. Ayhan, B.; Kwan, C.; Zhou, J. A new nonlinear change detection approach based on band ratioing. In Proceedings of the Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XXIV, Orlando, FL, USA, 17–19 April 2018; SPIE: Bellingham, WA, USA, 2018; Volume 10644, pp. 300–309. [Google Scholar]
  13. Chen, J.; Gong, P.; He, C.; Pu, R.; Shi, P. Land-use/land-cover change detection using improved change-vector analysis. Photogramm. Eng. Remote Sens. 2003, 69, 369–379. [Google Scholar] [CrossRef]
  14. Deng, J.; Wang, K.; Deng, Y.; Qi, G. PCA-based land-use change detection and analysis using multitemporal and multisensor satellite data. Int. J. Remote Sens. 2008, 29, 4823–4838. [Google Scholar] [CrossRef]
  15. Nielsen, A.A. The regularized iteratively reweighted MAD method for change detection in multi-and hyperspectral data. IEEE Trans. Image Process. 2007, 16, 463–478. [Google Scholar] [CrossRef] [PubMed]
  16. Hashim, F.; Dibs, H.; Jaber, H.S. Adopting gram-schmidt and brovey methods for estimating land use and land cover using remote sensing and satellite images. Nat. Environ. Pollut. Technol. 2022, 21, 867–881. [Google Scholar] [CrossRef]
  17. Han, T.; Wulder, M.A.; White, J.C.; Coops, N.C.; Alvarez, M.; Butson, C. An efficient protocol to process Landsat images for change detection with tasselled cap transformation. IEEE Geosci. Remote Sens. Lett. 2007, 4, 147–151. [Google Scholar] [CrossRef]
  18. Daudt, R.C.; Le Saux, B.; Boulch, A. Fully convolutional siamese networks for change detection. In Proceedings of the 2018 25th IEEE international conference on image processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 4063–4067. [Google Scholar]
  19. Wang, Q.; Zhang, X.; Chen, G.; Dai, F.; Gong, Y.; Zhu, K. Change detection based on Faster R-CNN for high-resolution remote sensing images. Remote Sens. Lett. 2018, 9, 923–932. [Google Scholar] [CrossRef]
  20. El Amin, A.M.; Liu, Q.; Wang, Y. Convolutional neural network features based change detection in satellite images. In Proceedings of the First International Workshop on Pattern Recognition, Tokyo, Japan, 11–13 May 2016; SPIE: Bellingham, WA, USA, 2016; Volume 10011, pp. 181–186. [Google Scholar]
  21. Liu, T.; Li, Y.; Xu, L. Dual-channel convolutional neural network for change detection of multitemporal SAR images. In Proceedings of the 2016 International Conference on Orange Technologies (ICOT), Melbourne, Australia, 17–20 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 60–63. [Google Scholar]
  22. Yin, M.; Chen, Z.; Zhang, C. A CNN-Transformer Network Combining CBAM for Change Detection in High-Resolution Remote Sensing Images. Remote Sens. 2023, 15, 2406. [Google Scholar] [CrossRef]
  23. Chen, H.; Shi, Z. A spatial-temporal attention-based method and a new dataset for remote sensing image change detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
  24. Zhang, C.; Yue, P.; Tapete, D.; Jiang, L.; Shangguan, B.; Huang, L.; Liu, G. A deeply supervised image fusion network for change detection in high resolution bi-temporal remote sensing images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 183–200. [Google Scholar] [CrossRef]
  25. Fang, S.; Li, K.; Shao, J.; Li, Z. SNUNet-CD: A densely connected Siamese network for change detection of VHR images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  26. Feng, Y.; Jiang, J.; Xu, H.; Zheng, J. Change detection on remote sensing images using dual-branch multilevel intertemporal network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  27. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  28. Chen, H.; Qi, Z.; Shi, Z. Remote sensing image change detection with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar] [CrossRef]
  29. Feng, Y.; Xu, H.; Jiang, J.; Liu, H.; Zheng, J. ICIF-Net: Intra-scale cross-interaction and inter-scale feature fusion network for bitemporal remote sensing images change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  30. Li, Q.; Zhong, R.; Du, X.; Du, Y. TransUNetCD: A hybrid transformer network for change detection in optical remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
  31. Yuan, J.; Wang, L.; Cheng, S. STransUNet: A siamese TransUNet-based remote sensing image change detection network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 9241–9253. [Google Scholar] [CrossRef]
  32. Xu, X.; Yang, Z.; Li, J. AMCA: Attention-guided multiscale context aggregation network for remote sensing image change detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–19. [Google Scholar] [CrossRef]
  33. Zhang, C.; Wang, L.; Cheng, S.; Li, Y. SwinSUNet: Pure transformer network for remote sensing image change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  34. Yu, W.; Zhuo, L.; Li, J. GCFormer: Global Context-aware Transformer for Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–12. [Google Scholar] [CrossRef]
  35. Vaswani, A. Attention Is All You Need. In Advances in Neural Information Processing Systems; Curran Associates Inc.: New York, NY, USA, 2017. [Google Scholar]
  36. Bandara, W.G.C.; Patel, V.M. A transformer-based siamese network for change detection. In Proceedings of the IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 207–210. [Google Scholar]
  37. Wen, Y.; Ma, X.; Zhang, X.; Pun, M.O. GCD-DDPM: A generative change detection model based on difference-feature guided DDPM. IEEE Trans. Geosci. Remote Sens. 2024, 1, 1–17. [Google Scholar] [CrossRef]
  38. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Online, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  39. Gu, T.; Yang, K.; An, X.; Feng, Z.; Liu, D.; Cai, W.; Deng, J. RWKV-CLIP: A Robust Vision-Language Representation Learner. arXiv 2024, arXiv:2406.06973. [Google Scholar]
  40. Lan, M.; Chen, C.; Ke, Y.; Wang, X.; Feng, L.; Zhang, W. Clearclip: Decomposing clip representations for dense vision-language inference. arXiv 2024, arXiv:2407.12442. [Google Scholar]
  41. Huang, R.; Pan, X.; Zheng, H.; Jiang, H.; Xie, Z.; Wu, C.; Song, S.; Huang, G. Joint representation learning for text and 3d point cloud. Pattern Recognit. 2024, 147, 110086. [Google Scholar] [CrossRef]
  42. Lin, J.; Gong, S. Gridclip: One-stage object detection by grid-level clip representation learning. arXiv 2023, arXiv:2303.09252. [Google Scholar]
  43. Chen, R.; Liu, Y.; Kong, L.; Zhu, X.; Ma, Y.; Li, Y.; Hou, Y.; Qiao, Y.; Wang, W. Clip2scene: Towards label-efficient 3d scene understanding by clip. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17 –24 June 2023; pp. 7020–7030. [Google Scholar]
  44. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  45. Ji, S.; Wei, S.; Lu, M. Fully convolutional networks for multisource building extraction from an open aerial and satellite imagery data set. IEEE Trans. Geosci. Remote Sens. 2018, 57, 574–586. [Google Scholar] [CrossRef]
  46. Shi, Q.; Liu, M.; Li, S.; Liu, X.; Wang, F.; Zhang, L. A deeply supervised attention metric-based network and an open aerial image dataset for remote sensing change detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  47. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  48. Sugirtha, T.; Sridevi, M. Semantic segmentation using modified u-net for autonomous driving. In Proceedings of the 2022 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS), Toronto, ON, Canada, 1–4 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–7. [Google Scholar]
  49. Seong, S.; Choi, J. Semantic segmentation of urban buildings using a high-resolution network (HRNet) with channel and spatial attention gates. Remote Sens. 2021, 13, 3087. [Google Scholar] [CrossRef]
  50. Zhuang, J.; Yang, J.; Gu, L.; Dvornek, N. Shelfnet for fast semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 847–856. [Google Scholar]
  51. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  53. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3349–3364. [Google Scholar] [CrossRef] [PubMed]
  54. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.; et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 6881–6890. [Google Scholar]
  55. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 7262–7272. [Google Scholar]
  56. Chu, X.; Tian, Z.; Wang, Y.; Zhang, B.; Ren, H.; Wei, X.; Xia, H.; Shen, C. Twins: Revisiting the design of spatial attention in vision transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 9355–9366. [Google Scholar]
  57. Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  58. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  59. Chen, L.C. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  60. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  61. Yuan, Y.; Chen, X.; Wang, J. Object-contextual representations for semantic segmentation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part VI 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 173–190. [Google Scholar]
  62. Zhang, H.; Zhang, H.; Wang, C.; Xie, J. Co-occurrent features in semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 548–557. [Google Scholar]
  63. Huang, Z.; Wang, X.; Wang, J.; Liu, W.; Wang, J. Weakly-supervised semantic segmentation network with deep seeded region growing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7014–7023. [Google Scholar]
  64. Kirillov, A.; Girshick, R.; He, K.; Dollár, P. Panoptic feature pyramid networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6399–6408. [Google Scholar]
  65. Xiao, T.; Liu, Y.; Zhou, B.; Jiang, Y.; Sun, J. Unified perceptual parsing for scene understanding. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 418–434. [Google Scholar]
  66. Zhou, Z.; Lei, Y.; Zhang, B.; Liu, L.; Liu, Y. Zegclip: Towards adapting clip for zero-shot semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 11175–11185. [Google Scholar]
  67. He, W.; Jamonnak, S.; Gou, L.; Ren, L. Clip-s4: Language-guided self-supervised semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 11207–11216. [Google Scholar]
  68. Ma, X.; Wu, Q.; Zhao, X.; Zhang, X.; Pun, M.O.; Huang, B. Sam-assisted remote sensing imagery semantic segmentation with object and boundary constraints. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  69. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 248–255. [Google Scholar]
  70. Parelius, E.J. A review of deep-learning methods for change detection in multispectral remote sensing images. Remote Sens. 2023, 15, 2092. [Google Scholar] [CrossRef]
  71. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; proceedings, part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  72. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  73. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics. JMLR Workshop and Conference Proceedings, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  74. Zhu, J.; Liao, S.; Yi, D.; Lei, Z.; Li, S.Z. Multi-label cnn based pedestrian attribute learning for soft biometrics. In Proceedings of the 2015 International Conference on Biometrics (ICB), Phuket, Thailand, 19–22 May 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 535–540. [Google Scholar]
  75. Kingma, D.P. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  76. Dong, S.; Wang, L.; Du, B.; Meng, X. ChangeCLIP: Remote sensing change detection with multimodal vision-language representation learning. ISPRS J. Photogramm. Remote Sens. 2024, 208, 53–69. [Google Scholar] [CrossRef]
Figure 1. Structural comparison between semantic-agnostic dominant methods [26,28,29,34,36,37] and the proposed LSC-CD. Compared to the disorder of the dominant methods in visual space, the LSC-CD obtains more orderly and compact semantic embeddings in visual-semantic space through clustering.
Figure 1. Structural comparison between semantic-agnostic dominant methods [26,28,29,34,36,37] and the proposed LSC-CD. Compared to the disorder of the dominant methods in visual space, the LSC-CD obtains more orderly and compact semantic embeddings in visual-semantic space through clustering.
Sensors 24 07887 g001
Figure 2. The pipeline of the proposed LSC-CD. The Transformer Block is the multi-scale transformer encoder from Segformer [44].
Figure 2. The pipeline of the proposed LSC-CD. The Transformer Block is the multi-scale transformer encoder from Segformer [44].
Sensors 24 07887 g002
Figure 3. Architecture of the proposed SCM.
Figure 3. Architecture of the proposed SCM.
Sensors 24 07887 g003
Figure 4. Architecture of the SEUM in the SCM.
Figure 4. Architecture of the SEUM in the SCM.
Sensors 24 07887 g004
Figure 5. Qualitative comparison results of different CD methods on LEVIR-CD datasets: the black represents true negative, the white represents true positive, the red represents false positive and the green represents false negative.
Figure 5. Qualitative comparison results of different CD methods on LEVIR-CD datasets: the black represents true negative, the white represents true positive, the red represents false positive and the green represents false negative.
Sensors 24 07887 g005
Figure 6. Qualitative comparison results of different CD methods on WHU-CD datasets: the black represents true negative, the white represents true positive, the red represents false positive and the green represents false negative.
Figure 6. Qualitative comparison results of different CD methods on WHU-CD datasets: the black represents true negative, the white represents true positive, the red represents false positive and the green represents false negative.
Sensors 24 07887 g006
Figure 7. Qualitative comparison results of different CD methods on SYSU-CD datasets: the black represents true negative, the white represents true positive, the red represents false positive, and the green represents false negative.
Figure 7. Qualitative comparison results of different CD methods on SYSU-CD datasets: the black represents true negative, the white represents true positive, the red represents false positive, and the green represents false negative.
Sensors 24 07887 g007
Figure 8. Heatmap comparison results. TBlock1–TBlock4 represent attention maps at four different scales from the encoder [44]. The 1st and 3rd rows show the pre-change and post-change heatmaps of the baseline (ChangeFormer), while the 2nd and 4th rows show the pre-change and post-change heatmaps of LSC-CD.
Figure 8. Heatmap comparison results. TBlock1–TBlock4 represent attention maps at four different scales from the encoder [44]. The 1st and 3rd rows show the pre-change and post-change heatmaps of the baseline (ChangeFormer), while the 2nd and 4th rows show the pre-change and post-change heatmaps of LSC-CD.
Sensors 24 07887 g008
Table 1. Common categories and subcategories of remote sensing images.
Table 1. Common categories and subcategories of remote sensing images.
CategorySubcategory
BuildingsResidential buildings, Commercial buildings, Factory buildings,
Church, Building
ForestTree, Forest Regrowth, Riparian Forest, Planted Forest,
Evergreen Trees, Deciduous Trees, Sparse Forest, Dense Forest
TransportationTruck, Cars, Ship, Trains, Airplane
Surface waterLake, Sea, River, Reservoir, Wetlands, Canal
RoadwayBridge, Freeway, Harbor, Runway, Railway, Road, Highway.
SportsBasketball Court, Ground Track Field, Stadium, Tennis Court, Golf Course
OthersContainer, Island, Snow Land, Pond, Fertile Land, Beach,
Square, Parking Lot, Park
Table 2. The quantitative comparison of results of different methods on the LEVIR-CD dataset.
Table 2. The quantitative comparison of results of different methods on the LEVIR-CD dataset.
MethodYearStructureP ↑R ↑F1 ↑IoU ↑OA ↑
FC-EF [18]2018CNN86.3783.5482.3571.8096.97
FC-Siam-Di [18]2018CNN89.5383.3186.3175.9298.67
FC-Siam-Conc [18]2018CNN91.9976.7783.6971.9698.49
STANet [23]2020CNN+Attention83.8191.0087.2677.4098.66
IFNet [24]2020CNN83.7780.3282.2970.9798.61
SNUNet-CD [25]2021CNN + Attention89.1887.1788.1678.8398.82
BIT [28]2021CNN + Transformer89.2489.3789.3180.6898.62
TransUNetCD [30]2022CNN + Transformer92.4389.8291.1183.67-
ChangeFormer [36]2022Transformer92.0588.8090.4082.4899.04
StransUNet [31]2022CNN + Transformer92.3090.5591.4184.1999.13
ICIF-Net [29]2023CNN + Transformer87.7980.8883.6571.8998.73
AMCA [32]2023CNN + Transformer91.8290.6791.4883.3998.73
DMINet [26]2023CNN92.5289.9490.7182.9999.07
GCD-DDPM [37]2024CNN90.6891.2490.9683.5699.14
GCFormer [34]2024CNN + Transformer89.3492.5090.8383.2199.08
LSC-CD-Transformer93.3190.6992.0185.1599.19
Red represents the best performance, blue represents the second-best performance, and all metrics are expressed as percentages. The arrow ↑ indicates that a larger value corresponds to better performance.
Table 3. A quantitative comparison of the results of different methods on the WHU-CD dataset.
Table 3. A quantitative comparison of the results of different methods on the WHU-CD dataset.
MethodYearStructureP ↑R ↑F1 ↑IoU ↑OA ↑
FC-EF [18]2018CNN83.5086.3384.8973.7498.87
FC-Siam-Di [18]2018CNN90.8684.6987.6778.0499.13
FC-Siam-Conc [18]2018CNN84.0287.7285.8375.1898.94
STANet [23]2020CNN+Attention79.3785.5082.3269.9598.52
IFNet [24]2020CNN96.9173.1983.4071.5298.83
SNUNet-CD [25]2021CNN + Attention85.6081.4983.5071.6798.71
BIT [28]2021CNN + Transformer82.0489.7485.7175.0098.62
TransUNetCD [30]2022CNN + Transformer93.5989.6091.6084.42-
ChangeFormer [36]2022Transformer91.8388.0289.8881.6399.12
StransUNet [31]2022CNN + Transformer93.2190.1591.6584.5999.29
ICIF-Net [29]2023CNN + Transformer90.7987.5889.1680.4399.01
AMCA [32]2023CNN + Transformer91.3781.9086.3875.1395.97
DMINet [26]2023CNN93.8486.2588.6979.6898.97
GCD-DDPM [37]2024CNN92.7992.2692.5486.5299.39
GCFormer [34]2024CNN91.8190.1590.9783.4399.29
LSC-CD-Transformer96.3293.1894.7589.9899.59
Red represents the best performance, blue represents the second-best performance, and all metrics are expressed as percentages. The arrow ↑ indicates that a larger value corresponds to better performance.
Table 4. The quantitative comparison of the results of different methods on the SYSU-CD dataset.
Table 4. The quantitative comparison of the results of different methods on the SYSU-CD dataset.
MethodYearStructureP ↑R ↑F1 ↑IoU ↑OA ↑
FC-EF [18]2018CNN76.4775.1775.8161.0488.69
FC-Siam-Di [18]2018CNN76.2875.3075.7961.0188.65
FC-Siam-Conc [18]2018CNN73.6776.7575.1860.2388.05
IFNet [24]2020CNN82.4472.3876.3861.85-
SNUNet-CD [25]2021CNN + Attention83.5875.8779.5466.0290.79
BIT [28]2021CNN + Transformer81.4277.9078.3264.3790.26
TransUNetCD [30]2022CNN + Transformer82.5977.7380.0966.7990.88
ChangeFormer [36]2022Transformer81.7072.3876.7662.2976.76
ICIF-Net [29]2023CNN + Transformer83.3778.5180.7468.1291.24
LSC-CD-Transformer88.6475.0681.2968.4791.85
Red represents the best performance, blue represents the second-best performance, and all metrics are expressed as percentages. The arrow ↑ indicates that a larger value corresponds to better performance.
Table 5. Results of ablation experiments on the LEVIR-CD dataset.
Table 5. Results of ablation experiments on the LEVIR-CD dataset.
MethodP ↑R ↑F1 ↑IoU ↑OA ↑
baseline92.0588.8090.4082.4899.04
baseline + CLIP91.7889.9290.8483.2299.08
baseline + CLIP + CAM91.3690.8091.0883.6299.09
baseline + CLIP + CAM + SCM93.4490.0891.6984.6699.16
baseline + CLIP + CAM + SCM + SEUM93.3190.6992.0185.1599.19
Red represents the best performance, blue represents the second-best performance, and all metrics are expressed as percentages. The arrow ↑ indicates that a larger value corresponds to better performance.
Table 6. Comparison between LSC-CD and ChangeCLIP.
Table 6. Comparison between LSC-CD and ChangeCLIP.
MethodYearP ↑R ↑F1 ↑IoU ↑OA ↑
ChangeCLIP(ViT-B/16)202493.6889.0491.3083.9999.14
LSC-CD(ViT-B/16)202493.3190.6992.0185.1599.19
Red represents the best performance, and all metrics are expressed as percentages. The arrow ↑ indicates that a larger value corresponds to better performance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, S.; Bian, Y.; Chen, B.; Song, H.; Zhang, K. Language-Guided Semantic Clustering for Remote Sensing Change Detection. Sensors 2024, 24, 7887. https://doi.org/10.3390/s24247887

AMA Style

Hu S, Bian Y, Chen B, Song H, Zhang K. Language-Guided Semantic Clustering for Remote Sensing Change Detection. Sensors. 2024; 24(24):7887. https://doi.org/10.3390/s24247887

Chicago/Turabian Style

Hu, Shenglong, Yiting Bian, Bin Chen, Huihui Song, and Kaihua Zhang. 2024. "Language-Guided Semantic Clustering for Remote Sensing Change Detection" Sensors 24, no. 24: 7887. https://doi.org/10.3390/s24247887

APA Style

Hu, S., Bian, Y., Chen, B., Song, H., & Zhang, K. (2024). Language-Guided Semantic Clustering for Remote Sensing Change Detection. Sensors, 24(24), 7887. https://doi.org/10.3390/s24247887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop