Next Article in Journal
Stability of High Stopes and Optimization of Combined Mining: A Case Study of the Dongguashan Copper Mine
Previous Article in Journal
Performance Investigation of Novel Desiccant Evaporative Cooling Systems Integrating with Shallow Geothermal Energy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach

Dipartimento di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, 20133 Milano, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(10), 4737; https://doi.org/10.3390/app16104737 (registering DOI)
Submission received: 30 January 2026 / Revised: 17 April 2026 / Accepted: 5 May 2026 / Published: 10 May 2026
(This article belongs to the Special Issue Recent Advances in Hyperspectral Imaging Technology)

Abstract

Low-light enhancement technologies are of great significance for visual driver assistance applications and autonomous driving systems. Infrared vision can improve nighttime visibility but also faces challenges of low resolution and lack of color information. This paper presents a unified framework for RGB-guided infrared super-resolution and infrared-visible fusion that achieves high-resolution output under limited computational resources. Our approach employs a U-Net architecture with novel triple-grouped window attention (TGWA) encoding that captures global dependencies through grouped attention while reducing computational overhead, and adaptive multi-dilated convolutional (AMDC) decoding that adaptively selects optimal dilation rates using mixture-of-experts-inspired routing. Experiments on multiple datasets achieve competitive super-resolution and fusion results with minimal computational complexity, while real-world downstream object detection validation confirms robust performance in challenging nighttime scenarios. Quantitatively, the proposed method achieves 28.744 dB/0.872 SSIM on PBVS24 and 31.424 dB/0.882 SSIM on HDRT-Night for 8× infrared super-resolution, reaches competitive fusion quality on both MSRS and HDRT-Night, and attains 69.4% mAP@0.5 in downstream object detection on FLIR_aligned, while requiring only 1.12 M parameters and 85.44 G FLOPs. This work provides new possibilities for seeing clearly in the dark.
Keywords: infrared-visible image fusion; low-light imaging; multimodal vision; image enhancement; deep learning infrared-visible image fusion; low-light imaging; multimodal vision; image enhancement; deep learning

Share and Cite

MDPI and ACS Style

Lu, J.; Desantis, V.; Paracchini, M.B.M.; Marcon, M. Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach. Appl. Sci. 2026, 16, 4737. https://doi.org/10.3390/app16104737

AMA Style

Lu J, Desantis V, Paracchini MBM, Marcon M. Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach. Applied Sciences. 2026; 16(10):4737. https://doi.org/10.3390/app16104737

Chicago/Turabian Style

Lu, Jiajie, Viviana Desantis, Marco Brando Mario Paracchini, and Marco Marcon. 2026. "Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach" Applied Sciences 16, no. 10: 4737. https://doi.org/10.3390/app16104737

APA Style

Lu, J., Desantis, V., Paracchini, M. B. M., & Marcon, M. (2026). Deep Learning for Low-Light Vision: An Efficient Infrared–Visible Fusion Approach. Applied Sciences, 16(10), 4737. https://doi.org/10.3390/app16104737

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop