Single-Exposure HDR Image Translation via Synthetic Wide-Band Characteristics Reflected Image Training
Abstract
1. Introduction
- Synthetic NIR Image Generation: A lightweight U-Net-based autoencoder generates a synthetic NIR image that replicates the characteristics of real NIR images. The encoder extracts features from the input visible image, while the decoder synthesizes the synthetic NIR image in the latent space.
- Image Fusion and Enhancement: The synthetic NIR image and the visible image are fused using a synthesis module. To maximize encoder performance, an Expand, Shuffle, and Merge Cardinality structure, inspired by YOLOv7’s E-ELAN, expands the CNN’s width and depth. Additionally, a Modified Convolutional Block Attention Module (CBAM) incorporating a Difference Map (diff-CBAM) is applied to emphasize key regions during learning.
- A two-step process is proposed, where a synthetic NIR image is first generated to replicate NIR characteristics and then fused with a visible image to enhance HDR synthesis.
- Experimental results confirm that the first-stage training effectively learns NIR characteristics from visible images, enabling the generation of high-quality synthetic NIR images.
- The second-stage training demonstrates that incorporating NIR characteristics significantly improves HDR synthesis efficiency. Additionally, a lightweight architecture, combined with cardinality and attention mechanisms, optimizes the module structure while reducing computational complexity.
2. Related Works
2.1. Traditional HDR Methods
2.1.1. Single Information-Based HDR
2.1.2. Multi-Information Based HDR
2.2. Image Processing with CNN
2.2.1. Extended Efficient Layer Aggregation Network
2.2.2. Convolutional Block Attention Module
3. Proposed Method
3.1. Synthetic NIR Image Generation
3.2. HDR Image Generation
4. Simulation and Results
4.1. Ablation Experiments
- Case 1: consists of a DSU-Net for generating synthetic NIR images and an Autoencoder-based HDR image generator that synthesizes synthetic NIR images with Visible images.
- Case 2: adds E-ELAN to the encoder structure of the HDR image generator, enabling an analysis of the impact of E-ELAN by comparing it to Case 1.
- Case 3: introduces Diff-CBAM to both the encoder and decoder of the HDR image generator, facilitating an evaluation of the effect of Diff-CBAM by comparing it to Case 2.
- Case 4: includes Perceptual Quality Optimization, which uses BRISQUE and NIQE metrics to guide the training process. By comparing Case 4 with Case 3, the impact of this training strategy on performance can be assessed.
4.2. Performance Comparisons
4.2.1. Qualitative Evaluation
4.2.2. Quantitative Evaluation
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Jinno, T.; Okuda, M. Multiple Exposure Fusion for High Dynamic Range Image Acquisition. IEEE Trans. Image Process. 2012, 21, 358–365. [Google Scholar] [CrossRef]
- Simons, R.H. Comment 2 on ‘Evaluation of High Dynamic Range Photography as a Luminance Data Acquisition System’ by M.N. Inanici. Light. Res. Technol. 2006, 38, 135. [Google Scholar] [CrossRef]
- Banterle, F.; Ledda, P.; Debattista, K.; Chalmers, A. Inverse Tone Mapping. In Proceedings of the 4th International Conference on Computer Graphics and Interactive Techniques in Australasia and Southeast Asia (GRAPHITE 2006), Kuala Lumpur, Malaysia, 29 November–2 December 2006; ACM: New York, NY, USA, 2006; pp. 349–356. [Google Scholar] [CrossRef]
- Eilertsen, G.; Kronander, J.; Denes, G.; Mantiuk, R.K.; Unger, J. HDR Image Reconstruction from a Single Exposure Using Deep CNNs. ACM Trans. Graph. 2017, 36, 178. [Google Scholar] [CrossRef]
- García-Rey, R.M.; García-Olmo, J.; De Pedro, E.; Quiles-Zafra, R.; Luque de Castro, M.D. Prediction of Texture and Colour of Dry-Cured Ham by Visible and near Infrared Spectroscopy Using a Fiber Optic Probe. Meat Sci. 2005, 70, 357–363. [Google Scholar] [CrossRef]
- Liu, Y.; Lyon, B.G.; Windham, W.R.; Realini, C.E.; Pringle, T.D.; Duckett, S. Prediction of Color, Texture, and Sensory Characteristics of Beef Steaks by Visible and near Infrared Reflectance Spectroscopy—A Feasibility Study. Meat Sci. 2003, 65, 1107–1115. [Google Scholar] [CrossRef]
- Liu, Y.; Chen, Y.; Ozaki, Y. Two-Dimensional Visible/Near-Infrared Correlation Spectroscopy. Appl. Spectrosc. 2000, 54, 901–908. [Google Scholar] [CrossRef]
- Chudnovsky, A.; Ben-Dor, E. Application of Visible, Near-Infrared, and Short-Wave Infrared (400–2500 nm) Reflectance Spectroscopy in Quantitatively Assessing Settled Dust in the Indoor Environment: Case Study in Dwellings and Office Environments. Sci. Total Environ. 2008, 393, 198–213. [Google Scholar] [CrossRef] [PubMed]
- Yang, T.; Li, Y.; Ruichek, Y.; Yan, Z. Performance Modeling a Near-Infrared ToF LiDAR Under Fog: A Data-Driven Approach. IEEE Trans. Intell. Transp. Syst. 2022, 23, 11227–11236. [Google Scholar] [CrossRef]
- Kil, T.; Cho, N.I. Image Fusion Using RGB and Near Infrared Image. J. Broadcast Eng. 2016, 21, 515–524. [Google Scholar] [CrossRef]
- Lee, J.; Oh, G.; Jeon, B. Perceptual Image Fusion Technique of RGB and NIR Images. In Proceedings of the 34th International Technical Conference on Circuits/Systems, Computers and Communications (ITC-CSCC 2019), JeJu, Republic of Korea, 23–26 June 2019; IEEE: New York, NY, USA, 2019; pp. 127–130. [Google Scholar] [CrossRef]
- Soundari, V.D.; Rabin, G.; Rishiyugan, S.; Pravinkumar, P. Image Processing for LDR to HDR Image Conversion Using Deep Learning. In Proceedings of the 5th International Conference on Inventive Research in Computing Applications (ICIRCA 2023), Coimbatore, India, 5–7 July 2023; IEEE: New York, NY, USA, 2023; pp. 146–150. [Google Scholar] [CrossRef]
- Raipurkar, P.; Pal, R.; Raman, S. HDR-CGAN: Single LDR to HDR Image Translation Using Conditional GAN. In Proceedings of the 2021 ACM International Conference on Multimedia Asia (MMAsia ’21), Gold Coast, Australia, 1–3 December 2021; ACM: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
- Bhatt, D.; Patel, C.; Talsania, H.; Patel, J.; Vaghela, R.; Pandya, S.; Modi, K.; Ghayvat, H. CNN Variants for Computer Vision: History, Architecture, Application, Challenges and Future Scope. Electronics 2021, 10, 2470. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision (ECCV 2018), Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar] [CrossRef]
- Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-Attention Generative Adversarial Networks. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), Long Beach, CA, USA, 9–15 June 2019; PMLR: London, UK, 2019; pp. 7354–7363. [Google Scholar]
- Kwon, H.; Lee, S.; Lee, G.; Sohng, K. Enhanced High Dynamic-Range Image Rendering Using a Surround Map Based on Edge-Adaptive Layer Blurring. IET Comput. Vis. 2016, 10, 689–699. [Google Scholar] [CrossRef]
- Kim, Y.J.; Son, D.M.; Lee, S.H. Retinex Jointed Multiscale CLAHE Model for HDR Image Tone Compression. Mathematics 2024, 12, 1541. [Google Scholar] [CrossRef]
- Kuang, J.; Johnson, G.M.; Fairchild, M.D. ICAM06: A Refined Image Appearance Model for HDR Image Rendering. J. Vis. Commun. Image Represent. 2007, 18, 406–414. [Google Scholar] [CrossRef]
- Reinhard, E.; Stark, M.; Shirley, P.; Ferwerda, J. Photographic Tone Reproduction for Digital Images. ACM Trans. Graph. 2002, 21, 267–276. [Google Scholar] [CrossRef]
- Liang, Z.; Xu, J.; Zhang, D.; Cao, Z.; Zhang, L. A Hybrid L1-L0 Layer Decomposition Model for Tone Mapping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2018), Salt Lake City, UT, USA, 18–23 June 2018; IEEE: New York, NY, USA, 2018; pp. 4758–4766. [Google Scholar] [CrossRef]
- Lee, M.J.; Rhee, C.H.; Lee, C.H. HSVNet: Reconstructing HDR Image from a Single Exposure LDR Image with CNN. Appl. Sci. 2022, 12, 2370. [Google Scholar] [CrossRef]
- An, J.; Lee, S.H.; Kuk, J.G.; Cho, N.I. A Multi-Exposure Image Fusion Algorithm without Ghost Effect. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2011), Prague, Czech Republic, 22–27 May 2011; IEEE: New York, NY, USA, 2011; pp. 1565–1568. [Google Scholar] [CrossRef]
- Merianos, I.; Mitianoudis, N. Multiple-Exposure Image Fusion for HDR Image Synthesis Using Learned Analysis Transformations. J. Imaging 2019, 5, 32. [Google Scholar] [CrossRef]
- Mitianoudis, N.; Stathaki, T. Optimal Contrast Correction for ICA-Based Fusion of Multimodal Images. IEEE Sens. J. 2008, 8, 2016–2026. [Google Scholar] [CrossRef]
- Mertens, T.; Kautz, J.; Van Reeth, F. Exposure Fusion. In Proceedings of the 15th Pacific Conference on Computer Graphics and Applications (PG’07), Maui, HI, USA, 29 October–2 November 2007; IEEE: New York, NY, USA, 2007; pp. 382–390. [Google Scholar] [CrossRef]
- Hu, X.; Shen, L.; Jiang, M.; Ma, R.; An, P. LA-HDR: Light Adaptive HDR Reconstruction Framework for Single LDR Image Considering Varied Light Conditions. IEEE Trans. Multimed. 2023, 25, 4814–4829. [Google Scholar] [CrossRef]
- Son, D.M.; Kwon, H.J.; Lee, S.H. Visible and Near Infrared Image Fusion Using Base Tone Compression and Detail Transform Fusion. In Proceedings of the 2022 International Conference on Electrical, Computer and Energy Technologies (ICECET 2022), Prague, Czech Republic, 20–22 July 2022; IEEE: New York, NY, USA, 2022; pp. 1–6, ISBN 978-8253950725. [Google Scholar]
- Kwon, H.J.; Lee, S.H. Visible and Near-Infrared Image Acquisition and Fusion for Night Surveillance. Chemosensors 2021, 9, 75. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; IEEE: New York, NY, USA, 2016; pp. 770–778. [Google Scholar] [CrossRef]
- Zagoruyko, S.; Komodakis, N. Wide Residual Networks. In Proceedings of the British Machine Vision Conference (BMVC 2016), York, UK, 19–22 September 2016; BMVA Press: Durham, UK, 2016; pp. 87.1–87.12. [Google Scholar] [CrossRef]
- Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; IEEE: New York, NY, USA, 2017; pp. 5987–5995. [Google Scholar] [CrossRef]
- Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
- Wang, C.-Y.; Liao, H.-Y.M.; Yeh, I.-H. Designing Network Design Strategies through Gradient Path Analysis. arXiv 2022, arXiv:2211.04800. [Google Scholar] [CrossRef]
- Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; IEEE: New York, NY, USA, 2017; pp. 1800–1807. [Google Scholar] [CrossRef]
- Mortimer, P. TAS-NIR: A VIS + NIR Dataset for Fine-Grained Semantic Segmentation in Unstructured Outdoor Environments. Data 2023, 8, 72. [Google Scholar] [CrossRef]
- Soria, X.; Sappa, A.D.; Akbarinia, A. Multispectral single-sensor RGB-NIR imaging: New challenges and opportunities. In Proceedings of the 7th International Conference on Image Processing Theory, Tools and Applications (IPTA), Montreal, QC, Canada, 28 November–1 December 2017; pp. 1–6. [Google Scholar] [CrossRef]
- EPFL IVRL. RGB–NIR Scene Dataset. [Online]. Available online: https://www.epfl.ch/labs/ivrl/research/downloads/rgb-nir-scene-dataset/ (accessed on 8 August 2025).
- MIT CSAIL. MIT Adobe 5K Dataset. [Online]. Available online: https://data.csail.mit.edu/graphics/fivek/ (accessed on 8 August 2025).
- Chen, X.; Liu, Y.; Zhang, Z.; Qiao, Y.; Dong, C. HDRUNet: Single Image HDR Reconstruction with Denoising and Dequantization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW 2021), Nashville, TN, USA, 19–25 June 2021; IEEE: New York, NY, USA, 2021; pp. 354–363. [Google Scholar] [CrossRef]
- Marnerides, D.; Bashford-Rogers, T.; Hatchett, J.; Debattista, K. ExpandNet: A Deep Convolutional Neural Network for High Dynamic Range Expansion from Low Dynamic Range Content. Comput. Graph. Forum 2018, 37, 37–49. [Google Scholar] [CrossRef]
- Guo, C.; Jiang, X. LHDR: HDR Reconstruction for Legacy Content Using a Lightweight DNN. In Proceedings of the 17th Asian Conference on Computer Vision (ACCV 2022), Macau, China, 4–8 December 2022; pp. 306–322. [Google Scholar] [CrossRef]
- Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
- Wang, Z.; Sheikh, H.R.; Bovik, A.C. No-Reference Perceptual Quality Assessment of JPEG Compressed Images. In Proceedings of the IEEE International Conference on Image Processing (ICIP 2002), Rochester, NY, USA, 22–25 September 2002; pp. 477–480. [Google Scholar] [CrossRef]
- Liu, L.; Liu, B.; Huang, H.; Conrad, A. No-Reference Image Quality Assessment Based on Spatial and Spectral Entropies. Signal Process. Image Commun. 2014, 29, 856–863. [Google Scholar] [CrossRef]
- Yang, S.; Wu, T.; Shi, S.; Lao, S.; Gong, Y. MANIQA: Multi-Dimension Attention Network for No-Reference Image Quality Assessment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022), New Orleans, LA, USA, 18–24 June 2022; pp. 1191–1200. [Google Scholar] [CrossRef]
- Kang, L.; Ye, P.; Li, Y.; Doermann, D. Convolutional Neural Networks for No-Reference Image Quality Assessment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2014), Columbus, OH, USA, 23–28 June 2014; pp. 1733–1740. [Google Scholar] [CrossRef]
Name | Components of Each Stage |
---|---|
Case 1 | Synthetic NIR Generator + HDR Image Generator |
Case 2 | Synthetic NIR Generator + HDR Image Generator + E-ELAN |
Case 3 | Synthetic NIR Generator + HDR Image Generator + E-ELAN + Diff-CBAM |
Case 4 | Synthetic NIR Generator + HDR Image Generator + E-ELAN + Diff-CBAM + Perceptual Quality Optimization |
Name/Metric | Icam06 | L1L0 | Reinhard | Kwon | Kim | HDRUNet | ExpandNet | LHDR | Proposed |
---|---|---|---|---|---|---|---|---|---|
SSEQ (↓) | 37.476 | 35.918 | 37.270 | 35.763 | 33.237 | 43.601 | 32.454 | 32.897 | 27.203 |
CNNIQA (↓) | 25.729 | 26.281 | 30.041 | 36.771 | 26.542 | 40.347 | 23.158 | 22.222 | 23.933 |
BRISQUE (↓) | 29.398 | 29.726 | 32.304 | 28.340 | 26.945 | 40.716 | 30.929 | 31.763 | 25.298 |
NRPQA (↑) | 7.9362 | 8.2634 | 7.6140 | 7.6629 | 7.9967 | 9.1712 | 9.4736 | 9.6453 | 9.3543 |
MANIQA (↑) | 0.3699 | 0.3682 | 0.3356 | 0.2988 | 0.5275 | 0.3453 | 0.5616 | 0.5539 | 0.5521 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lee, S.H.; Lee, S.H. Single-Exposure HDR Image Translation via Synthetic Wide-Band Characteristics Reflected Image Training. Mathematics 2025, 13, 2644. https://doi.org/10.3390/math13162644
Lee SH, Lee SH. Single-Exposure HDR Image Translation via Synthetic Wide-Band Characteristics Reflected Image Training. Mathematics. 2025; 13(16):2644. https://doi.org/10.3390/math13162644
Chicago/Turabian StyleLee, Seung Hwan, and Sung Hak Lee. 2025. "Single-Exposure HDR Image Translation via Synthetic Wide-Band Characteristics Reflected Image Training" Mathematics 13, no. 16: 2644. https://doi.org/10.3390/math13162644
APA StyleLee, S. H., & Lee, S. H. (2025). Single-Exposure HDR Image Translation via Synthetic Wide-Band Characteristics Reflected Image Training. Mathematics, 13(16), 2644. https://doi.org/10.3390/math13162644