This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
HDCGAN+: A Low-Illumination UAV Remote Sensing Image Enhancement and Evaluation Method Based on WPID
by
Kelly Chen Ke
Kelly Chen Ke ,
Min Sun
Min Sun *
,
Xinyi Wang
Xinyi Wang
,
Dong Liu
Dong Liu and
Hanjun Yang
Hanjun Yang
Institute of Remote Sensing and Geographic Information Systems, Peking University, 5 Summer Palace Road, Beijing 100871, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2026, 18(7), 999; https://doi.org/10.3390/rs18070999 (registering DOI)
Submission received: 28 January 2026
/
Revised: 16 March 2026
/
Accepted: 24 March 2026
/
Published: 26 March 2026
Abstract
Remote sensing images acquired by UAVs under nighttime or low-illumination conditions suffer from insufficient illumination, leading to degraded image quality, detail loss, and noise, which restrict their application in public security and disaster emergency scenarios. Although existing machine learning-based enhancement methods can recover part of the missing information, they often cause color distortion and texture inconsistency. This study proposes an improved low-illumination image enhancement method based on a Weakly Paired Image Dataset (WPID), combining the Hierarchical Deep Convolutional Generative Adversarial Network (HDCGAN) with a low-rank image fusion strategy to enhance the quality of low-illumination UAV remote sensing images. First, YCbCr color channel separation is applied to preserve color information from visible images. Then, a Low-Rank Representation Fusion Network (LRRNet) is employed to perform structure-aware fusion between thermal infrared (TIR) and visible images, thereby enabling effective preservation of structural details and realistic color appearance. Furthermore, a weakly paired training mechanism is incorporated into HDCGAN to enhance detail restoration and structural fidelity. To achieve objective evaluation, a structural consistency assessment framework is constructed based on semantic segmentation results from the Segment Anything Model (SAM). Experimental results demonstrate that the proposed method outperforms state-of-the-art approaches in both visual quality and application-oriented evaluation metrics.
Share and Cite
MDPI and ACS Style
Ke, K.C.; Sun, M.; Wang, X.; Liu, D.; Yang, H.
HDCGAN+: A Low-Illumination UAV Remote Sensing Image Enhancement and Evaluation Method Based on WPID. Remote Sens. 2026, 18, 999.
https://doi.org/10.3390/rs18070999
AMA Style
Ke KC, Sun M, Wang X, Liu D, Yang H.
HDCGAN+: A Low-Illumination UAV Remote Sensing Image Enhancement and Evaluation Method Based on WPID. Remote Sensing. 2026; 18(7):999.
https://doi.org/10.3390/rs18070999
Chicago/Turabian Style
Ke, Kelly Chen, Min Sun, Xinyi Wang, Dong Liu, and Hanjun Yang.
2026. "HDCGAN+: A Low-Illumination UAV Remote Sensing Image Enhancement and Evaluation Method Based on WPID" Remote Sensing 18, no. 7: 999.
https://doi.org/10.3390/rs18070999
APA Style
Ke, K. C., Sun, M., Wang, X., Liu, D., & Yang, H.
(2026). HDCGAN+: A Low-Illumination UAV Remote Sensing Image Enhancement and Evaluation Method Based on WPID. Remote Sensing, 18(7), 999.
https://doi.org/10.3390/rs18070999
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article Access Statistics
For more information on the journal statistics, click
here.
Multiple requests from the same IP address are counted as one view.