This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach
Dipartimento di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, 20133 Milano, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(10), 4737; https://doi.org/10.3390/app16104737 (registering DOI)
Submission received: 30 January 2026
/
Revised: 17 April 2026
/
Accepted: 5 May 2026
/
Published: 10 May 2026
Abstract
Low-light enhancement technologies are of great significance for visual driver assistance applications and autonomous driving systems. Infrared vision can improve nighttime visibility but also faces challenges of low resolution and lack of color information. This paper presents a unified framework for RGB-guided infrared super-resolution and infrared-visible fusion that achieves high-resolution output under limited computational resources. Our approach employs a U-Net architecture with novel triple-grouped window attention (TGWA) encoding that captures global dependencies through grouped attention while reducing computational overhead, and adaptive multi-dilated convolutional (AMDC) decoding that adaptively selects optimal dilation rates using mixture-of-experts-inspired routing. Experiments on multiple datasets achieve competitive super-resolution and fusion results with minimal computational complexity, while real-world downstream object detection validation confirms robust performance in challenging nighttime scenarios. Quantitatively, the proposed method achieves 28.744 dB/0.872 SSIM on PBVS24 and 31.424 dB/0.882 SSIM on HDRT-Night for 8× infrared super-resolution, reaches competitive fusion quality on both MSRS and HDRT-Night, and attains 69.4% mAP@0.5 in downstream object detection on FLIR_aligned, while requiring only 1.12 M parameters and 85.44 G FLOPs. This work provides new possibilities for seeing clearly in the dark.
Share and Cite
MDPI and ACS Style
Lu, J.; Desantis, V.; Paracchini, M.B.M.; Marcon, M.
Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach. Appl. Sci. 2026, 16, 4737.
https://doi.org/10.3390/app16104737
AMA Style
Lu J, Desantis V, Paracchini MBM, Marcon M.
Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach. Applied Sciences. 2026; 16(10):4737.
https://doi.org/10.3390/app16104737
Chicago/Turabian Style
Lu, Jiajie, Viviana Desantis, Marco Brando Mario Paracchini, and Marco Marcon.
2026. "Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach" Applied Sciences 16, no. 10: 4737.
https://doi.org/10.3390/app16104737
APA Style
Lu, J., Desantis, V., Paracchini, M. B. M., & Marcon, M.
(2026). Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach. Applied Sciences, 16(10), 4737.
https://doi.org/10.3390/app16104737
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article Access Statistics
For more information on the journal statistics, click
here.
Multiple requests from the same IP address are counted as one view.