A Cascaded Enhancement-Fusion Network for Visible-Infrared Imaging in Darkness
Abstract
1. Introduction
2. Imaging Methods
2.1. Framework of the Proposed Method
2.2. Loss Function
2.2.1. Loss from Low-Light Enhancement
2.2.2. Fusion Loss
3. Results and Discussions
3.1. Experimental Setup
3.2. Experimental Results
3.3. Generalization Ability of the Method
3.4. Ablation Study
3.4.1. Smoothing Loss Analysis

3.4.2. Intensity Loss Analysis
3.4.3. Gradient Loss Analysis
3.5. Model Complexity Analysis and High-Level Visual Task Evaluation
4. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Jin, S.; Zhang, Y.; Xu, Z.; Qin, Z.; Jiang, X.; Liu, H.; Xu, M. Adaptive dual-domain-guided semantics diffusion model for image defogging. Opt. Laser Technol. 2025, 190, 113081. [Google Scholar] [CrossRef]
- Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
- Wei, Q.; Bioucas-Dias, J.; Dobigeon, N.; Tourneret, J.Y. Hyperspectral and Multispectral Image Fusion Based on a Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3658–3668. [Google Scholar] [CrossRef]
- Liu, C.H.; Qi, Y.; Ding, W.R. Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 2017, 83, 94–102. [Google Scholar] [CrossRef]
- Mou, J.; Gao, W.; Song, Z. Image fusion based on non-negative matrix factorization and infrared feature extraction. In Proceedings of the 2013 6th International Congress on Image and Signal Processing (CISP), Hangzhou, China, 16–18 December 2013; pp. 1046–1050. [Google Scholar]
- Chen, J.; Li, X.; Luo, L.; Ma, J. Multi-Focus Image Fusion Based on Multi-Scale Gradients and Image Matting. IEEE Trans. Multimed. 2022, 24, 655–667. [Google Scholar] [CrossRef]
- Zhang, H.; Xu, H.; Tian, X.; Jiang, J.; Ma, J. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
- Liu, J.; Fan, X.; Jiang, J.; Liu, R.; Luo, Z. Learning a Deep Multi-Scale Feature Ensemble and an Edge-Attention Guidance for Image Fusion. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 105–119. [Google Scholar] [CrossRef]
- Jin, S.; Xu, Z.; Xu, M.; Liu, H. Time-gated imaging through dense fog via physics-driven Swin transformer. Opt. Express 2024, 32, 18812–18830. [Google Scholar] [CrossRef]
- Li, H.; Wu, X.J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
- Liu, H.; Wang, P.; He, X.; Chen, M.; Liu, M.; Xu, Z.; Jiang, X.; Peng, X.; Xu, M. PI-NLOS: Polarized infrared non-line-of-sight imaging. Opt. Express 2023, 3, 44113–44126. [Google Scholar] [CrossRef] [PubMed]
- Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2021, 70, 1–14. [Google Scholar] [CrossRef]
- Tang, W.; He, F.; Liu, Y.; Duan, Y.; Si, T. DATFuse: Infrared and Visible Image Fusion via Dual Attention Transformer. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3159–3172. [Google Scholar] [CrossRef]
- Tang, W.; He, F.; Liu, Y. YDTR: Infrared and Visible Image Fusion via Y-Shape Dynamic Transformer. IEEE Trans. Multimed. 2023, 25, 5413–5428. [Google Scholar] [CrossRef]
- Tang, W.; He, F.; Liu, Y. ITFuse: An interactive transformer for infrared and visible image fusion. Pattern Recognit. 2024, 156, 110822. [Google Scholar] [CrossRef]
- Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1777–1786. [Google Scholar]
- Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond Brightening Low-light Images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
- Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 3496–3504. [Google Scholar]
- Tang, L.; Yuan, J.; Zhang, H.; Jiang, X.; Ma, J. PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 2022, 83–84, 79–92. [Google Scholar] [CrossRef]
- Li, H.; Wu, X.-J.; Kittler, J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
- Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef]
- Zhang, H.; Ma, J. SDNet: A Versatile Squeeze-and-Decomposition Network for Real-Time Image Fusion. Int. J. Comput. Vis. 2021, 129, 2761–2785. [Google Scholar] [CrossRef]
- Cui, G.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 2015, 341, 199–209. [Google Scholar] [CrossRef]
- Eskicioglu, A.M.; Fisher, P.S. Image quality measures and their performance. IEEE Trans. Commun. 1995, 43, 2959–2965. [Google Scholar] [CrossRef]
- Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 23522. [Google Scholar]
- Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 313–315. [Google Scholar] [CrossRef]
- Piella, G.; Heijmans, H. A new quality metric for image fusion. In Proceedings of the 2003 International Conference on Image Processing, Barcelona, Spain, 14–17 September 2003; IEEE: New York, NY, USA, 2003. [Google Scholar]
- Han, Y.; Cai, Y.; Cao, Y.; Xu, X. A new image fusion performance metric based on visual information fidelity. Inf. Fusion 2013, 14, 127–135. [Google Scholar] [CrossRef]
- Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware Dual Adversarial Learning and a Multi-scenario Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5792–5801. [Google Scholar]
- Toet, A. The TNO Multiband Image Data Collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef] [PubMed]






| DenseFuse | DDcGAN | GANMcC | RFN-Nest | SDNet | MFEIF | Ours | |
|---|---|---|---|---|---|---|---|
| MI | 2.558 ± 0.772 | 1.813 ± 0.663 | 2.321 ± 0.844 | 1.854 ± 1.172 | 1.981 ± 0.873 | 1.954 ± 1.443 | 3.812 ± 1.625 |
| VIF | 0.726 ± 0.114 | 0.624 ± 0.273 | 0.695 ± 0.287 | 0.479 ± 0.286 | 0.347 ± 0.202 | 0.425 ± 0.246 | 1.017 ± 0.184 |
| SF | 0.018 ± 0.006 | 0.032 ± 0.010 | 0.021 ± 0.009 | 0.018 ± 0.009 | 0.018 ± 0.007 | 0.023 ± 0.008 | 0.043 ± 0.008 |
| EN | 5.395 ± 0.976 | 6.498 ± 0.305 | 4.728 ± 1.731 | 3.464 ± 2.301 | 4.196 ± 0.851 | 2.997 ± 2.262 | 6.637 ± 0.933 |
| 0.360 ± 0.042 | 0.264 ± 0.125 | 0.355 ± 0.134 | 0.276 ± 0.169 | 0.285 ± 0.152 | 0.298 ± 0.177 | 0.557 ± 0.158 | |
| AG | 1.526 ± 0.618 | 2.716 ± 0.801 | 1.557 ± 0.895 | 1.019 ± 0.676 | 1.400 ± 0.560 | 1.235 ± 0.800 | 3.355 ± 1.179 |
| DenseFuse | DDcGAN | GANMcC | RFN-Nest | SDNet | MFEIF | Ours | |
|---|---|---|---|---|---|---|---|
| MI | 3.092 ± 0.772 | 2.797 ± 0.651 | 2.946 ± 0.656 | 3.182 ± 1.012 | 3.171 ± 1.039 | 3.487 ± 0.983 | 3.246 ± 1.079 |
| VIF | 0.781 ± 0.203 | 0.759 ± 0.240 | 0.582 ± 0.177 | 0.974 ± 0.355 | 0.773 ± 0.148 | 0.904 ± 0.223 | 1.322 ± 0.291 |
| SF | 0.019 ± 0.009 | 0.037 ± 0.016 | 0.022 ± 0.011 | 0.018 ± 0.007 | 0.030 ± 0.015 | 0.022 ± 0.012 | 0.058 ± 0.016 |
| EN | 6.433 ± 0.513 | 6.767 ± 0.305 | 5.006 ± 1.083 | 6.777 ± 0.525 | 6.680 ± 0.300 | 6.746 ± 0.510 | 6.812 ± 0.527 |
| 0.370 ± 0.040 | 0.454 ± 0.112 | 0.294 ± 0.092 | 0.347 ± 0.092 | 0.508 ± 0.085 | 0.428 ± 0.126 | 0.514 ± 0.093 | |
| AG | 1.660 ± 0.748 | 2.942 ± 1.449 | 1.693 ± 0.869 | 1.676 ± 0.778 | 2.576 ± 1.232 | 1.934 ± 0.950 | 4.217 ± 1.652 |
| DenseFuse | DDcGAN | GANMcC | RFN-Nest | SDNet | MFEIF | Ours | |
|---|---|---|---|---|---|---|---|
| MI | 2.910 ± 0.573 | 2.462 ± 0.485 | 2.734 ± 0.519 | 2.730 ± 0.778 | 2.766 ± 0.747 | 2.966 ± 0.813 | 3.438 ± 0.901 |
| VIF | 0.762 ± 0.139 | 0.713 ± 0.184 | 0.620 ± 0.152 | 0.806 ± 0.254 | 0.628 ± 0.119 | 0.741 ± 0.169 | 1.218 ± 0.202 |
| SF | 0.019 ± 0.006 | 0.035 ± 0.011 | 0.032 ± 0.008 | 0.018 ± 0.006 | 0.026 ± 0.010 | 0.022 ± 0.008 | 0.053 ± 0.011 |
| EN | 6.080 ± 0.474 | 6.676 ± 0.226 | 4.911 ± 0.926 | 5.651 ± 0.856 | 5.835 ± 0.351 | 5.471 ± 0.840 | 6.553 ± 0.471 |
| 0.367 ± 0.030 | 0.389 ± 0.085 | 0.315 ± 0.076 | 0.523 ± 0.084 | 0.432 ± 0.076 | 0.384 ± 0.103 | 0.529 ± 0.082 | |
| AG | 1.614 ± 0.537 | 2.865 ± 0.994 | 1.647 ± 0.649 | 1.453 ± 0.563 | 2.176 ± 0.835 | 1.696 ± 0.683 | 3.924 ± 1.162 |
| DenseFuse | DDcGAN | GANMcC | RFN-Nest | SDNet | MFEIF | Ours | |
|---|---|---|---|---|---|---|---|
| MI | 3.092 ± 0.772 | 2.797 ± 0.651 | 3.021 ± 1.068 | 3.182 ± 1.012 | 3.171 ± 1.039 | 3.487 ± 0.983 | 3.246 ± 1.079 |
| VIF | 0.781 ± 0.203 | 0.759 ± 0.240 | 0.798 ± 0.232 | 0.974 ± 0.355 | 0.773 ± 0.148 | 0.904 ± 0.223 | 1.017 ± 0.391 |
| SF | 0.019 ± 0.009 | 0.037 ± 0.016 | 0.022 ± 0.009 | 0.018 ± 0.007 | 0.030 ± 0.015 | 0.022 ± 0.012 | 0.058 ± 0.026 |
| EN | 6.433 ± 0.513 | 6.767 ± 0.305 | 6.674 ± 0.151 | 6.777 ± 0.525 | 6.680 ± 0.300 | 6.746 ± 0.510 | 6.812 ± 0.595 |
| 0.370 ± 0.040 | 0.454 ± 0.112 | 0.336 ± 0.106 | 0.347 ± 0.092 | 0.508 ± 0.085 | 0.428 ± 0.126 | 0.479 ± 0.123 | |
| AG | 1.660 ± 0.748 | 2.942 ± 1.449 | 1.874 ± 0.838 | 1.676 ± 0.778 | 2.576 ± 1.232 | 1.934 ± 0.950 | 4.522 ± 2.039 |
| DenseFuse | DDcGAN | GANMcC | RFN-Nest | SDNet | MFEIF | Ours | |
|---|---|---|---|---|---|---|---|
| Parameters (M) | 0.074 | 10.93 | 1.864 | 10.93 | 0.067 | 0.373 | 0.167 |
| FLOPs (G) | 233.55 | 4232.1 | 7075.4 | 4444.5 | 254.32 | 1848.3 | 449.56 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Huang, H.; Liu, H.; Wang, H.; Yang, Y.; Guo, C.; Chen, M.; Han, K. A Cascaded Enhancement-Fusion Network for Visible-Infrared Imaging in Darkness. Photonics 2025, 12, 1231. https://doi.org/10.3390/photonics12121231
Huang H, Liu H, Wang H, Yang Y, Guo C, Chen M, Han K. A Cascaded Enhancement-Fusion Network for Visible-Infrared Imaging in Darkness. Photonics. 2025; 12(12):1231. https://doi.org/10.3390/photonics12121231
Chicago/Turabian StyleHuang, Hanchang, Hao Liu, Hailu Wang, Yunzhuo Yang, Chuan Guo, Minsun Chen, and Kai Han. 2025. "A Cascaded Enhancement-Fusion Network for Visible-Infrared Imaging in Darkness" Photonics 12, no. 12: 1231. https://doi.org/10.3390/photonics12121231
APA StyleHuang, H., Liu, H., Wang, H., Yang, Y., Guo, C., Chen, M., & Han, K. (2025). A Cascaded Enhancement-Fusion Network for Visible-Infrared Imaging in Darkness. Photonics, 12(12), 1231. https://doi.org/10.3390/photonics12121231

