This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
TMAFNet: A Transformer-Based Multi-Level Adaptive Fusion Network for Remote Sensing Change Detection
by
Yushuai Yuan
Yushuai Yuan 1,
Zhiyong Fan
Zhiyong Fan 2,3,*,
Shuai Zhang
Shuai Zhang 2,
Min Xia
Min Xia 2,3
and
Yalu Huang
Yalu Huang 1
1
School of Artificial Intelligence (Future Technology College), Nanjing University of Information Science and Technology, Nanjing 210044, China
2
School of Automation, Nanjing University of Information Science and Technology, Nanjing 210044, China
3
Jiangsu Key Laboratory of Big Data Analysis Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2026, 18(8), 1143; https://doi.org/10.3390/rs18081143 (registering DOI)
Submission received: 2 March 2026
/
Revised: 7 April 2026
/
Accepted: 9 April 2026
/
Published: 12 April 2026
Abstract
High-resolution remote sensing imagery encompasses complex land cover types and rich textural details, whilst temporal variations often manifest as subtle feature differences and unstable structural patterns. This renders traditional change detection methods ineffective at accurately characterizing genuine alterations, frequently leading to underdetection, false positives, and ambiguous boundaries. To address these challenges, this paper proposes a Transformer-Based Multi-level Adaptive Fusion Network. It is built upon the DeepLabV3+ encoder–decoder framework, in which a shared-weight ResNet-101 is adopted as the backbone for dual-temporal feature extraction, with the final residual block of layer 4 cropped to extract deeper semantic features at a higher spatial resolution. The Adaptive Window–Attention Feature Fusion Module (AWAFM) adaptively models local and global differences across temporal phases, enhancing sensitivity to genuine changes. The Dual Strip Pool Fusion Module (DSPFM) enhances sensitivity to directional structural variations through horizontal and vertical strip pooling. The Progressive Multi-Scale Feature Fusion Module (PMFFM) progressively aggregates deep and shallow features via semantic residual transmission. To further suppress misleading suppression caused by complex textures, the Transformer-Enhanced Reverse Attention Fusion Module (TRAFM) explicitly models long-range dependencies, effectively mitigating false change responses. On the LEVIR-CD dataset, it achieves state-of-the-art performance, with a PA and an IoU of 92.36% and 90.13%, respectively. On the SYSU-CD dataset, PA and IoU reach 88.96% and 86.15%, demonstrating TMAFNet’s stability and superiority in scenarios involving complex ground surface disturbances, weak textural variations, and large-scale structural changes.
Share and Cite
MDPI and ACS Style
Yuan, Y.; Fan, Z.; Zhang, S.; Xia, M.; Huang, Y.
TMAFNet: A Transformer-Based Multi-Level Adaptive Fusion Network for Remote Sensing Change Detection. Remote Sens. 2026, 18, 1143.
https://doi.org/10.3390/rs18081143
AMA Style
Yuan Y, Fan Z, Zhang S, Xia M, Huang Y.
TMAFNet: A Transformer-Based Multi-Level Adaptive Fusion Network for Remote Sensing Change Detection. Remote Sensing. 2026; 18(8):1143.
https://doi.org/10.3390/rs18081143
Chicago/Turabian Style
Yuan, Yushuai, Zhiyong Fan, Shuai Zhang, Min Xia, and Yalu Huang.
2026. "TMAFNet: A Transformer-Based Multi-Level Adaptive Fusion Network for Remote Sensing Change Detection" Remote Sensing 18, no. 8: 1143.
https://doi.org/10.3390/rs18081143
APA Style
Yuan, Y., Fan, Z., Zhang, S., Xia, M., & Huang, Y.
(2026). TMAFNet: A Transformer-Based Multi-Level Adaptive Fusion Network for Remote Sensing Change Detection. Remote Sensing, 18(8), 1143.
https://doi.org/10.3390/rs18081143
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article metric data becomes available approximately 24 hours after publication online.