Next Article in Journal
Spatiotemporal Variations in Vegetation Phenology in the Qinling Mountains and Their Responses to Climate Variability
Previous Article in Journal
Quantifying Broad-Leaved Korean Pine Forest Structure Using Terrestrial Laser Scanning (TLS), Changbai Mountain, China
Previous Article in Special Issue
Fine-Grained Multispectral Fusion for Oriented Object Detection in Remote Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

DSEPGAN: A Dual-Stream Enhanced Pyramid Based on Generative Adversarial Network for Spatiotemporal Image Fusion

School of Geophysics and Geomatics, China University of Geosciences, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(24), 4050; https://doi.org/10.3390/rs17244050
Submission received: 14 October 2025 / Revised: 5 December 2025 / Accepted: 15 December 2025 / Published: 17 December 2025

Abstract

Many deep learning-based spatiotemporal fusion (STF) methods have been proven to achieve high accuracy and robustness. Due to the variable shapes and sizes of objects in remote sensing images, pyramid networks are generally introduced to extract multi-scale features. However, the down-sampling operation in the pyramid structure may lead to the loss of image detail information, affecting the model’s ability to reconstruct fine-grained targets. To address this issue, we propose a novel Dual-Stream Enhanced Pyramid based on Generative Adversarial Network (DSEPGAN) for the spatiotemporal fusion of remote sensing images. The network adopts a dual-stream architecture to separately process coarse and fine images, tailoring feature extraction to their respective characteristics: coarse images provide temporal dynamics, while fine images contain rich spatial details. A reversible feature transformation is embedded in the pyramid feature extraction stage to preserve high-frequency information, and a fusion module employing large-kernel and depthwise separable convolutions captures long-range dependencies across inputs. To further enhance realism and detail fidelity, adversarial training encourages the network to generate sharper and more visually convincing fusion results. The proposed DSEPGAN is compared with widely used and state-of-the-art STF models in three publicly available datasets. The results illustrate that DSEPGAN achieves superior performance across various evaluation metrics, highlighting its notable advantages for predicting seasonal variations in highly heterogeneous regions and abrupt changes in land use.
Keywords: spatiotemporal fusion; deep learning; pyramid network; invertible neural network; large kernel convolution spatiotemporal fusion; deep learning; pyramid network; invertible neural network; large kernel convolution

Share and Cite

MDPI and ACS Style

Zhou, D.; Xu, L.; Wu, K.; Liu, H.; Jiang, M. DSEPGAN: A Dual-Stream Enhanced Pyramid Based on Generative Adversarial Network for Spatiotemporal Image Fusion. Remote Sens. 2025, 17, 4050. https://doi.org/10.3390/rs17244050

AMA Style

Zhou D, Xu L, Wu K, Liu H, Jiang M. DSEPGAN: A Dual-Stream Enhanced Pyramid Based on Generative Adversarial Network for Spatiotemporal Image Fusion. Remote Sensing. 2025; 17(24):4050. https://doi.org/10.3390/rs17244050

Chicago/Turabian Style

Zhou, Dandan, Lina Xu, Ke Wu, Huize Liu, and Mengting Jiang. 2025. "DSEPGAN: A Dual-Stream Enhanced Pyramid Based on Generative Adversarial Network for Spatiotemporal Image Fusion" Remote Sensing 17, no. 24: 4050. https://doi.org/10.3390/rs17244050

APA Style

Zhou, D., Xu, L., Wu, K., Liu, H., & Jiang, M. (2025). DSEPGAN: A Dual-Stream Enhanced Pyramid Based on Generative Adversarial Network for Spatiotemporal Image Fusion. Remote Sensing, 17(24), 4050. https://doi.org/10.3390/rs17244050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop