TIMA-Net: A Lightweight Remote Sensing Image Change Detection Network Based on Temporal Interaction Enhancement and Multi-Scale Aggregation
Abstract
1. Introduction
- (1)
- A new lightweight change detection method, TIMA-Net, is proposed. Based on MobileNetV2 as the backbone network, Combining the Cross-Level Aggregation Fusion Module (CAFM) and the Temporal Interaction Enhancement Module (TIEM), Lightweight Multi-Scale Convolution Unit (LMCU), and Decoding Module (DM) achieves a higher performance with fewer parameters.
- (2)
- We proposed CAFM (Cross-Level Aggregation Fusion Module) to address the issue of disconnection between deep semantic features and shallow detail features in lightweight backbones (such as MobileNetV2), introducing a neighborhood aggregation and dynamic attention mechanism. By aggregating adjacent feature layers through sliding windows (reducing computational complexity from O(N2) to O(N)) and introducing adaptive weight allocation, this module alleviates the limitation of “dimensionality explosion caused by simple concatenation” in traditional multi-scale fusion, thereby enhancing feature representation.
- (3)
- We proposed TIEM (Temporal Interaction Enhancement Module), which adopts a dual-branch structure to learn global and local multi-scale information of bi-temporal features respectively. It uses a joint metric of Euclidean distance and cosine similarity to replace traditional subtraction operations and introduces a Coordinate Attention (CA) mechanism to reduce noise and pseudo-changes, highlighting change regions.
- (4)
- We designed a lightweight progressive decoder based on multi-scale convolution, including LMCU (Lightweight Multi-Scale Convolution Unit) and DM (Decoding Module). We constructed LMCU through channel splitting and multi-directional convolutions (1 × 9/9 × 1 convolutions to capture long-range dependencies), combined with a DM module with self-attention reweighting. This decoder can extract multi-scale contextual information with fewer parameters to adapt to change features in different regions.
2. Related Work
2.1. Traditional RSCD Methods
2.2. Deep Learning-Based RSCD Methods
2.3. Lightweight RSCD Methods
3. Methods
3.1. Overall Framework
3.2. Cross-Level Aggregation Fusion Module (CAFM)
- (1)
- F2 preserves the richest spatial details (e.g., building edges, road cracks), maximizing high-frequency information utilization;
- (2)
- It avoids checkerboard artifacts caused by multi-level independent interpolation.
3.3. Temporal Interaction Enhancement Module (TIEM)
3.4. Lightweight Multi-Scale Convolution Unit (LMCU)
3.5. Decoding Module (DM)
3.6. Loss Function
4. Results
4.1. Configuration
4.1.1. Implementation Details
4.1.2. Datasets
4.1.3. Evaluation Metrics
4.2. Comparative Studies of SOTA Methods
4.2.1. Quantitative Comparison
4.2.2. Qualitative Evaluation
- (1)
- Advantage in reducing the occurrence of pseudo-changes: In the fourth row of Figure 6, many methods fail to distinguish pseudo-changes on roads with appearances similar to buildings. In contrast, the proposed method can effectively identify the boundaries and main bodies of building changes. In the third row of Figure 7, areas near changed buildings exhibit different colors at different times, leading to pseudo-changes irrelevant to actual building changes. Many methods such as DMINet, STFF-GA, SEIFNet, and RaHFF cannot eliminate these pseudo-changes, while the method in this paper handles pseudo-changes caused by the diversity of land cover appearances effectively. In the first row of Figure 8, due to seasonal differences, roads show different colors at different times. Methods like BIT, ICIF-Net, DMINet, STFF-GA, and SEIFNet fail to identify these pseudo-changes, whereas the proposed method demonstrates the best recognition effect for it. This advantage can be attributed to the TIEM method, which enhances temporal difference features based on a dual-branch path, reducing false detections.
- (2)
- Advantage in detecting edges and details: In the second row of Figure 6, the changing objects are small in area and densely arranged. Methods such as FC-Diff and DMINet have difficulty identifying details like the edges between densely clustered buildings. In the fourth row of Figure 7, methods such as DMINet and RaHFF struggle to recognize the differences between roads and buildings with similar colors. Although DMINet and RaHFF achieve relatively good change detection results, changes in the edge areas are easily overlooked.In contrast, our DMTI-Net can effectively identify the boundaries and objects. These results can be attributed to the multi-scale context fusion capability of the LMCU and the inverse change map operation of the DM, which enhance the completeness of the content and improve the accuracy of boundary detection. In the visualization results of the LEVIR-CD dataset (Figure 6), to more prominently show that our method has the best detection performance for large-sized, small-sized, or densely clustered objects, we stitched 16 images of 256 × 256 into an image of 1024 × 1024 for display, demonstrating the overall optimal detection performance of our method.
- (3)
- Advantage in more complete detection of large objects: As shown in the first row of Figure 7, first and second rows of Figure 8, compared with the limitation of the comparison methods that are prone to miss detections in the recognition of complex semantic changes, this method can achieve a more complete change detection effect. This is mainly due to the following three aspects: Firstly, through the adjacent temporal feature fusion mechanism, the spatial semantic information in the temporal dimension is effectively integrated. Secondly, a multi-level temporal difference representation system is constructed, and the hierarchical feature capture mechanism is used to extract context features at different abstraction levels. Thirdly, based on the attention-guided feature refinement strategy, a progressive fusion method from global to local is adopted, and finally, a RSCD map with a fine edge structure is generated. This multi-level progressive fusion architecture ensures the complete perception of complex object changes in the RSCD task.
4.2.3. Comparison of Efficiency
4.3. Ablation Experiments
- (1)
- Effectiveness of CAFM
- (2)
- Effectiveness of TIEM
- (3)
- Effectiveness of LMCU
- (4)
- Effectiveness of the DM
- (5)
- Effectiveness of the Loss Function
- (6)
- Comparison of Different Backbone Networks
5. Discussion
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- de Bem, P.P.; de Carvalho Júnior, O.A.; Guimarães, R.F.; Gomes, R.A.T. Change detection of deforestation in the Brazilian Amazon using Landsat data and convolutional neural networks. Remote Sens. 2020, 12, 901. [Google Scholar] [CrossRef]
- Zheng, Z.; Zhong, Y.; Wang, J.; Ma, A.; Zhang, L. Building damage assessment for rapid disaster response with a deep object-based semantic change detection framework: From natural disasters to man-made dis-asters. Remote Sens. Environ. 2021, 265, 112636–112657. [Google Scholar] [CrossRef]
- Chen, J.; Liu, H.; Hou, J.; Yang, M.; Deng, M. Improving building change detection in VHR remote sensing imagery by combining coarse location and co-segmentation. ISPRS Int. J. Geo-Inf. 2018, 7, 213. [Google Scholar] [CrossRef]
- Ke, L.; Lin, Y.; Zeng, Z.; Zhang, L.; Meng, L. Adaptive change detection with significance test. IEEE Access 2018, 6, 27442–27450. [Google Scholar] [CrossRef]
- Coppin, P.R.; Bauer, M.E. Digital change detection in forest ecosystems with remote sensing imagery. Remote Sens. Rev. 1996, 13, 207–234. [Google Scholar] [CrossRef]
- Lambin, E.F.; Strahlers, A.H. Change-vector analysis in multitemporal space: A tool to detect and categorize land-cover change processes using high temporal-resolution satellite data. Remote Sens. Environ. 1994, 48, 231–244. [Google Scholar] [CrossRef]
- Celik, T. Unsupervised change detection in satellite images using principal component analysis and k-means clustering. IEEE Geosci. Remote Sens. Lett. 2009, 6, 772–776. [Google Scholar] [CrossRef]
- Crist, E.P. A TM tasseled cap equivalent transformation for reflectance factor data. Remote Sens. Environ. 1985, 17, 301–306. [Google Scholar] [CrossRef]
- Gong, J.; Hu, X.; Pang, S.; Li, K. Patch matching and dense CRF-based co-refinement for building change detection from bi-temporal aerial images. Sensors 2019, 19, 1557. [Google Scholar] [CrossRef]
- Im, J.; Jensen, J. A change detection model based on neighborhood correlation image analysis and decision tree classification. Remote Sens. Environ. 2005, 99, 326–340. [Google Scholar] [CrossRef]
- Verdier, G.; Ferreira, A. Adaptive Mahalanobis distance and k-nearest neighbor rule for fault detection in semiconductor manufacturing. IEEE Trans. Semicond. Manuf. 2011, 24, 59–68. [Google Scholar] [CrossRef]
- Seo, D.K.; Kim, Y.H.; Eo, Y.D.; Lee, M.H.; Park, W.Y. Fusion of SAR and multispectral images using random forest regression for change detection. ISPRS Int. J. Geo-Inf. 2018, 7, 401. [Google Scholar] [CrossRef]
- Jiang, W.; Sun, Y.; Lei, L.; Kuang, G.; Ji, K. Change detection of multisource remote sensing images: A review. Int. J. Digit. Earth 2024, 17, 2398051. [Google Scholar] [CrossRef]
- Li, X.; He, M.; Li, H.; Shen, H. A combined loss-based multiscale fully convolutional network for high-resolution remote sensing image change detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
- Li, X.; Du, Z.; Huang, Y.; Tan, Z. A deep translation (GAN) based change detection network for optical and SAR remote sensing images. ISPRS J. Photogramm. Remote Sens. 2021, 179, 14–34. [Google Scholar] [CrossRef]
- Bai, B.; Fu, W.; Lu, T.; Li, S. Edge-guided recurrent convolutional neural network for multitemporal remote sensing image building change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
- Chen, H.; Qi, Z.; Shi, Z. Remote sensing image change detection with transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
- Peng, D.; Zhang, Y.; Guan, H. End-to-end change detection for high resolution satellite images using improved UNet++. Remote Sens. 2019, 11, 1382. [Google Scholar] [CrossRef]
- Zhu, Q. Land-use/land-cover change detection based on a Siamese global learning framework for high spatial resolu-tion remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2022, 184, 63–78. [Google Scholar] [CrossRef]
- Fang, S.; Li, K.; Shao, J.; Li, Z. SNUNet-CD: A densely connected Siamese network for change detection of VHR images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
- Tan, M.; Le, Q. EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
- Fang, S.; Li, K.; Li, Z. Changer: Feature interaction is what you need for change detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–11. [Google Scholar] [CrossRef]
- Liang, S.; Hua, Z.; Li, J. Enhanced feature interaction network for remote sensing change detection. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–15. [Google Scholar] [CrossRef]
- Sun, Y.; Lei, L.; Li, Z.; Kuang, G. Similarity and dissimilarity relationships based graphs for multimodal change detection. ISPRS J. Photogramm. Remote Sens. 2024, 208, 70–88. [Google Scholar] [CrossRef]
- Zhang, H.; Lin, M.; Yang, G.; Zhang, L. ESCNet: An end-to-end superpixel-enhanced change detection network for very-high-resolution remote sensing images. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 28–42. [Google Scholar] [CrossRef]
- Li, Q.; Zhong, R.; Du, X.; Du, Y. TransUNetCD: A hybrid transformer network for change detection in optical remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
- Daudt, R.; Le Saux, B.; Boulch, A. Fully convolutional siamese networks for change detection. In Proceedings of the 2018 25th IEEE International Conference on Image Processing, ICIP, Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 4063–4067. [Google Scholar]
- Zhang, Z.; Wang, X.; Jung, C. DCSR: Dilated convolutions for single image super-resolution. IEEE Trans. Image Process. 2018, 28, 1625–1635. [Google Scholar] [CrossRef]
- Feng, Y.; Xu, H.; Jiang, J.; Liu, H.; Zheng, J. ICIF-Net: Intra-scale cross-interaction and inter-scale feature fusion network for bitemporal remote sensing images change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
- Zhang, C.; Wang, L.; Cheng, S.; Li, Y. SwinSUNet: Pure transformer network for remote sensing image change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
- Yan, L.; Jiang, J. A hybrid siamese network with spatiotemporal enhancement and two-level feature fusion for remote sensing image change detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–17. [Google Scholar] [CrossRef]
- Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
- Li, Z.; Tang, C.; Liu, X.; Zhang, W.; Dou, J.; Wang, L.; Zomaya, A.Y. Lightweight remote sensing change detection with progressive feature aggregation and supervised attention. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–12. [Google Scholar] [CrossRef]
- Lei, T.; Geng, X.; Ning, H.; Lv, Z.; Gong, M.; Jin, Y.; Nandi, A.K. Ultralightweight spatial–spectral feature cooperation network for change detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–14. [Google Scholar] [CrossRef]
- Codegoni, A.; Lombardi, G.; Ferrari, A. TINYCD: A (not so) deep learning model for change detection. Neural Comput. Appl. 2023, 35, 8471–8486. [Google Scholar] [CrossRef]
- Huang, Y.; Li, X.; Du, Z.; Shen, H. Spatiotemporal Enhancement and Interlevel Fusion Network for Remote Sensing Images Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–14. [Google Scholar] [CrossRef]
- Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 13708–13717. [Google Scholar]
- Milletari, F.; Navab, N.; Ahmadi, S. V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
- Chen, H.; Shi, Z. A spatial–temporal attention-based method and a new dataset for remote sensing image change detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
- Shi, Q.; Liu, M.; Li, S.; Liu, X.; Wang, F.; Zhang, L. A deeply supervised attention metric-based network and an open aerial image dataset for remote sensing change detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
- Feng, Y.; Jiang, J.; Xu, H.; Zheng, J. Change detection on remote sensing images using dual-branch multilevel intertemporal network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
- Wei, H. Spatio-Temporal Feature Fusion and Guide Aggregation Network for Remote Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5642216. [Google Scholar] [CrossRef]
- Wang, B.; Zhao, K.; Xiao, T.; Qin, P.; Zeng, J. RaHFF-Net: Recall-Adjustable Hierarchical Feature Fusion Network for Remote Sensing Image Change Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 176–190. [Google Scholar] [CrossRef]
Methods | LEVIR (%) | SYSU (%) | BCDD (&) | FLOPs (G) | Params (M) | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
IoU | F1 | Rec | Pre | IoU | F1 | Rec | Pre | IoU | F1 | Rec | Pre | |||
FC-Siam-diff | 75.18 | 85.84 | 84.59 | 87.12 | 52.06 | 68.47 | 55.89 | 88.36 | 71.94 | 83.68 | 77.76 | 90.57 | 9.42 | 1.35 |
BIT | 82.76 | 90.57 | 89.40 | 91.77 | 65.84 | 79.40 | 76.68 | 82.32 | 88.38 | 93.83 | 92.03 | 95.70 | 16.88 | 3.01 |
ICIF-Net | 82.03 | 90.13 | 88.89 | 91.40 | 62.20 | 76.70 | 76.99 | 76.39 | 88.45 | 93.88 | 92.12 | 95.61 | 25.41 | 23.84 |
DMINet | 82.92 | 90.67 | 88.89 | 92.54 | 68.11 | 81.19 | 78.10 | 84.19 | 88.72 | 93.95 | 92.30 | 95.65 | 14.55 | 6.24 |
STFF-GA | 83.39 | 90.93 | 91.34 | 90.55 | 69.45 | 81.97 | 80.14 | 83.89 | 89.11 | 94.24 | 92.68 | 95.82 | 8.22 | 12.28 |
SEIFNet | 83.40 | 90.95 | 89.46 | 92.49 | 69.96 | 82.32 | 79.98 | 84.81 | 88.95 | 94.10 | 92.55 | 95.75 | 8.37 | 27.91 |
RaHFF | 80.85 | 89.41 | 89.35 | 89.46 | 66.73 | 80.04 | 79.79 | 80.30 | 81.86 | 90.03 | 86.64 | 93.68 | 33.95 | 16.48 |
AFRNet (ours) | 83.96 | 91.28 | 89.97 | 92.62 | 70.49 | 83.01 | 80.31 | 85.88 | 89.43 | 94.42 | 93.65 | 95.84 | 6.05 | 4.31 |
NO. | Variants | LEVIR(%) | SYSU(%) | ||||||
---|---|---|---|---|---|---|---|---|---|
IoU | F1 | Rec | Pre | IoU | F1 | Rec | Pre | ||
#01 | TIMA-Net | 83.96 | 91.28 | 89.97 | 92.62 | 70.06 | 82.39 | 80.49 | 84.38 |
Cross-Level Aggregation Fusion Module (CAFM) | |||||||||
#02 | w/o CAFM | 83.04 | 90.42 | 88.95 | 91.93 | 68.53 | 81.02 | 79.11 | 83.02 |
#03 | CAFM w/o FW | 83.28 | 90.71 | 89.03 | 92.45 | 69.03 | 81.67 | 79.87 | 83.51 |
Temporal Interaction Enhancement Module (TIEM) | |||||||||
#04 | Sub | 83.13 | 90.65 | 89.32 | 92.01 | 69.14 | 81.78 | 80.04 | 83.62 |
#05 | Cat | 82.95 | 90.41 | 88.93 | 91.94 | 68.71 | 81.28 | 79.42 | 83.25 |
#06 | w/o DConv | 83.31 | 90.89 | 89.84 | 91.98 | 69.18 | 81.91 | 80.35 | 83.52 |
Lightweight Multi-Scale Convolution Unit (LMCU) | |||||||||
#07 | w/o LMCU | 82.98 | 90.42 | 88.96 | 91.93 | 68.14 | 80.66 | 78.67 | 82.76 |
Decoder Module (DM) | |||||||||
#08 | w/o DM | 83.19 | 90.71 | 89.25 | 92.19 | 68.45 | 81.01 | 78.34 | 83.87 |
Loss Function | |||||||||
#09 | BCE | 83.29 | 90.74 | 89.17 | 92.36 | 68.18 | 81.09 | 78.89 | 83.41 |
#10 | Dice | 83.54 | 90.93 | 89.46 | 92.45 | 68.85 | 81.71 | 79.24 | 84.33 |
NO. | Modules | IoU | F1 | Rec | Pre | FlOPs (G) | Params (M) |
---|---|---|---|---|---|---|---|
#01 | CAFM(Ours) | 83.96 | 91.28 | 89.97 | 92.62 | 6.05 | 4.31 |
#02 | NAM(A2Net) | 82.57 | 90.45 | 89.26 | 91.68 | 6.21 | 4.52 |
NO. | Backbones | IoU | F1 | Rec | Pre | FlOPs (G) | Params (M) |
---|---|---|---|---|---|---|---|
#01 | ResNet18 | 83.72 | 91.15 | 90.28 | 92.03 | 23.41 | 19.13 |
#02 | EfficientNet | 83.69 | 91.08 | 90.04 | 92.14 | 7.49 | 6.21 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhou, Z.; Zhang, X.; Luo, X.; Wang, L.; Yu, W.; Xu, S.; Wang, L. TIMA-Net: A Lightweight Remote Sensing Image Change Detection Network Based on Temporal Interaction Enhancement and Multi-Scale Aggregation. Remote Sens. 2025, 17, 2332. https://doi.org/10.3390/rs17142332
Zhou Z, Zhang X, Luo X, Wang L, Yu W, Xu S, Wang L. TIMA-Net: A Lightweight Remote Sensing Image Change Detection Network Based on Temporal Interaction Enhancement and Multi-Scale Aggregation. Remote Sensing. 2025; 17(14):2332. https://doi.org/10.3390/rs17142332
Chicago/Turabian StyleZhou, Zhijun, Xuejie Zhang, Xiaoliang Luo, Lvchun Wang, Wei Yu, Shufang Xu, and Longbao Wang. 2025. "TIMA-Net: A Lightweight Remote Sensing Image Change Detection Network Based on Temporal Interaction Enhancement and Multi-Scale Aggregation" Remote Sensing 17, no. 14: 2332. https://doi.org/10.3390/rs17142332
APA StyleZhou, Z., Zhang, X., Luo, X., Wang, L., Yu, W., Xu, S., & Wang, L. (2025). TIMA-Net: A Lightweight Remote Sensing Image Change Detection Network Based on Temporal Interaction Enhancement and Multi-Scale Aggregation. Remote Sensing, 17(14), 2332. https://doi.org/10.3390/rs17142332