Next Article in Journal
Avalanche Photodiodes and Silicon Photomultipliers of Non-Planar Designs
Previous Article in Journal
Improvement in the Post-Processing of Wave Buoy Data Driven by the Needs of a National Coast and Sea Monitoring Agency
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploiting Light Polarization for Deep HDR Imaging from a Single Exposure †

1
Department of Environmental Sciences, Informatics and Statistics, Ca’ Foscari University of Venice, 155, Via Torino, 30170 Venice, Italy
2
Institute of Visual Computing & Human-Centered Technology, TU Wien, Favoritenstr. 9-11/E193-02, 1040 Vienna, Austria
*
Author to whom correspondence should be addressed.
This paper is an extended version of the paper published in Fatima, T.; Pistellato, M.; Torsello, A.; Bergamasco, F. One-Shot HDR Imaging via Stereo PFA Cameras. In Proceedings of the Image Analysis and Processing—ICIAP 2022: 21st International Conference, Lecce, Italy, 23–27 May 2022; Part II, pp. 467–478.
These authors contributed equally to this work.
Sensors 2023, 23(12), 5370; https://doi.org/10.3390/s23125370
Submission received: 20 March 2023 / Revised: 19 May 2023 / Accepted: 30 May 2023 / Published: 6 June 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In computational photography, high dynamic range (HDR) imaging refers to the family of techniques used to recover a wider range of intensity values compared to the limited range provided by standard sensors. Classical techniques consist of acquiring a scene-varying exposure to compensate for saturated and underexposed regions, followed by a non-linear compression of intensity values called tone mapping. Recently, there has been a growing interest in estimating HDR images from a single exposure. Some methods exploit data-driven models trained to estimate values outside the camera’s visible intensity levels. Others make use of polarimetric cameras to reconstruct HDR information without exposure bracketing. In this paper, we present a novel HDR reconstruction method that employs a single PFA (polarimetric filter array) camera with an additional external polarizer to increase the scene’s dynamic range across the acquired channels and to mimic different exposures. Our contribution consists of a pipeline that effectively combines standard HDR algorithms based on bracketing and data-driven solutions designed to work with polarimetric images. In this regard, we present a novel CNN (convolutional neural network) model that exploits the underlying mosaiced pattern of the PFA in combination with the external polarizer to estimate the original scene properties, and a second model designed to further improve the final tone mapping step. The combination of such techniques enables us to take advantage of the light attenuation given by the filters while producing an accurate reconstruction. We present an extensive experimental section in which we validate the proposed method on both synthetic and real-world datasets specifically acquired for the task. Quantitative and qualitative results show the effectiveness of the approach when compared to state-of-the-art methods. In particular, our technique exhibits a PSNR (peak signal-to-noise ratio) on the whole test set equal to 23 dB, which is 18 % better with respect to the second-best alternative.

1. Introduction

Conventional cameras typically record intensity information with a limited range of possible detectable values. Such limitations are not negligible when acquiring scenes exhibiting high dynamic content, as conventional cameras fail to capture the full dynamic range. In such cases, the result will either be overly bright or excessively dark, leading to the loss of information in certain areas. These areas are commonly referred to as overexposed or underexposed regions. Such acquisitions result in so-called low dynamic range (LDR) images. High dynamic range (HDR) imaging aims to recover missing information in those problematic scene regions, thereby increasing the final dynamic range [1]. Some of the key areas where HDR is applicable include scene understanding [2], robotics [3], autonomous driving [4], medical imaging [5,6], agriculture [7], and spacecraft imaging [8].
The literature proposes many HDR imaging techniques, most of which rely on taking multiple shots of the same scene using multiple exposures and merging them together. These techniques usually suffer from global or local alignment issues: the former arises when the camera is moving between acquisitions, while the latter is due to moving subjects in the scene [9]. To overcome this issue, researchers have been trying to recover missing pixel information using just one input image: such techniques are called inverse tone mapping (ITM) [10]. Other works investigated HDR reconstruction from single or multiple LDR images, exploiting several deep learning approaches [11,12].
PFA cameras are also shown to be useful for HDR imaging [3,13,14,15]: these cameras can capture four images simultaneously with polarizers at different orientations. Since each linear polarizer reduces the observed intensity according to the degree of polarization of incoming light, the filter acts as an exposure time reduction and, thus, an equivalent exposure time can be computed to perform HDR reconstruction. In principle this works, but only when light is highly polarized, otherwise all four filters will essentially receive similar responses. Since the degree of polarization in a real-world scene is generally low, the usage of a plain PFA camera is limited. In [15], we proposed an HDR reconstruction method with a PFA stereo camera setup. In these settings, we used two PFA cameras, with only one of them equipped with an extra linear polarizer positioned outside the lenses. The rationale of adopting such a stereo setup is to use the external filter to make the incoming light highly polarized, and exploit the resulting intensity attenuation, while acquiring the original scene as-is with the regular PFA camera. This is required because the polarization state of the incoming light is needed to correctly compensate the external filter and, thus, compute the correct equivalent exposure time. Experimental results show that the idea of using an additional polarizer in such a setting produces better quality HDR compared to the single PFA camera approach, but at the cost of increased complexity.
In this paper, we build on top of the method proposed in [15], with the goal of removing one camera and obtaining a simpler yet effective system to perform HDR reconstruction using one PFA camera. We propose a novel HDR reconstruction setup involving a single PFA camera with an additional linear polarizer outside the lenses, paired with a tailored data-driven model, which is trained to simulate the response of the missing camera (i.e., the one with no filter on). The proposed model estimates the scene AoLP (angle of linear polarization) and DoLP (degree of linear polarization) as if we have an actual second camera, and the resulting data are merged with the actual image with the additional filter. As a result, the proposed system performs HDR reconstruction from a single polarimetric image without the requirement of having a second camera. We can summarize the main contributions of this work as follows: we propose a novel data-driven approach that is able to effectively recover the original AoLP and DoLP of a scene from a single picture taken with an additional external polarizer. To our knowledge, this is the first work proposing such an “inverse problem” formulation for PFA cameras, exploiting a network architecture that takes into account the mosaiced pattern to characterize the acquired pictures. In this way, we obtain a virtual PFA stereo setup: the image coming from the physical camera and the network output can be merged together to compose an HDR picture, taking into account the equivalent exposure times associated with the scene features. Finally, a refinement step takes the computed HDR image and performs tone mapping via a specifically designed CNN architecture. The proposed reconstruction approach is shown in Figure 1. An extensive experimental study supports the effectiveness of the proposed method, including a comparison with state-of-the-art techniques. The rest of the paper is organized as follows: Section 2 describes the most relevant works connected to our method, then Section 3 briefly summarizes polarimetric basics and the stereo HDR reconstruction presented in [15]. The proposed technique is then presented in detail in Section 4, and finally both qualitative and quantitative experimental results are discussed in Section 5.

2. Related Work

2.1. Multiple LDR to HDR

The most common approach for HDR imaging is to take multiple images from the same scene with increasing exposure, which is usually referred to as “multi-exposure images”. The camera response function (CRF) is estimated using these images, and the inverse of CRF is applied to linearize the multi-exposure images. These images are aligned and merged together to produce the final HDR image. The following are state-of-the-art approaches designed to create HDR using multiple exposures.

2.1.1. Classical Approaches

Numerous examples can be found in the literature, where researchers reconstructed HDR using multi-exposure images [16,17,18,19,20,21,22,23,24]. One of the earliest efforts in creating HDR images using multiple exposures was made by [16]; the authors introduced a novel method to recover the CRF as well as an HDR radiance map using multi-exposure images. Similarly, [17] proposed an HDR reconstruction method for handheld cameras since images taken by hand are more prone to artifacts, such as blurs. Images are first registered using the MTB algorithm, and then maximum likelihood is used to find the blur kernel and CRF. Mobile cameras have small sensor pixels, which means they can gather a very small number of photons per pixel. Due to this fact, images captured by mobile cameras have more noise and offer a very limited dynamic range. In [18], the authors introduced a rather unconventional way to take multiple LDR images with automatically adjusted exposures. This idea looks promising but it brings many problems, for example, when selecting auto-exposure for different real-world scenes, many hand-crafted features are used, which does not seem feasible, and it is an ’overhead’ to store such information. This pipeline is also time-consuming and may fail to produce desirable results in extreme cases (underexposed/overexposed).
Exposure fusion is an alternative approach to multi-exposure HDR imaging. Exposure fusion was introduced to simplify the pipeline for HDR imaging. This technique does not require CRF estimation and does not produce an intermediate HDR image, thus eliminating the tone mapping step. Although this technique does not provide an increase in the dynamic range, it produces visually pleasing results. Exposure fusion works well for static scenes while it suffers from artifacts in the case of moving objects or a shaking camera. Reference [1] proposed such a method in which multi-exposure images are merged in multi-resolution fashion on the basis of simple measures, such as saturation, well-exposedness, and contrast. The results produced by different measures look visually different, and each introduces its own artifacts.

2.1.2. Learning-Based Techniques

Nowadays, learning-based techniques are employed in a variety of applications, being able to solve even complex problems [25,26,27]. Researchers have also investigated the creation of HDR using multiple LDR images by designing deep learning methods. To fuse a set of multiple exposed images together and generate an artifact-free HDR image, [28] designed the first learning-based approach. This is a CNN that takes two images (underexposed and overexposed) as input, with three components: in the first component low-level features are extracted, then these features are fused in the second layer. Finally, a reconstruction component generates the output HDR image. In [29], the use of GANs for HDR generation is proposed for the first time. The generator is formed by three parts: self-attention, local details, and merge block. The discriminator has a rather simple architecture and tries to discriminate between ground truth and generated HDR in terms of probability ( 0 1 ) . Reference [30] proposed a CNN-based approach to generate ghost-free HDR for dynamic scenes. They aligned three images of the dynamic scene taken with increasing exposure using a classical optical flow algorithm. These aligned images were later merged using the CNN-based approach to produce HDR. Instead of an end-to-end mapping, [31] splits the HDR creation from dynamic input images [30] into two sequential tasks. The first is an encoder–decoder named AlignNet, which aligns three images with low, mid, and high exposures. These exposures are merged together by the next CNN called MergeNet, and the final HDR image is estimated. References [32,33] are also specifically designed to estimate HDR outputs for dynamic scenes and tested on the Kalantari [30] dataset. A self-supervised method is proposed in [34], where a set of three bracketed–exposed LDR images was used to create HDR patches for self-supervision based on static and well-exposed areas of the image. FlexHDR [35] takes an arbitrary number of LDR images as input and computes the optical flow between these differently exposed images using a flow network. After flow estimation, the FlexHDR model addresses uncertainties caused by exposure and alignment via an attention network, and the final HDR is generated by a multi-stage fusion-based merging network. Other learning-based methods can be found at [36,37,38,39,40,41,42,43,44,45].

2.2. Single LDR to HDR

The above-mentioned classical and learning-based techniques work well when the camera is mounted on a tripod and the scene is static, but if this is not the case, the output could suffer from the following problems. First, it may produce global (ghost-like) artifacts due to camera motion, and local artifacts can be caused by the movement of objects captured in the scene. Second, multiple LDR and HDR techniques cannot be applied to the existing single LDR image legacy content. To overcome the issues posed by multi-exposure methods, inverse tone mapping (ITM) is introduced with the aim of reconstructing the HDR image from a single exposure. Alignment and merging steps are no longer required while the HDR radiance map is reconstructed directly from the input image. These techniques simplify the HDR reconstruction pipeline and are useful for converting legacy data from the LDR domain to HDR. However, reconstructing HDR through a single exposure is an ill-posed problem and ITM techniques are underperforming when compared to exposure bracketing.

2.2.1. Classical

A huge amount of image and video data is produced for visualization on outdated display devices. The widespread adoption of HDR displays has created a need for converting this historical data from LDR to HDR and making it compatible with these new devices. In this case, traditional HDR image generation algorithms fail, hence, the idea of ITM was introduced by researchers to enhance LDR content in order to display it on HDR devices. In [46], the authors proposed a support vector machine (SVM)-based approach for both images and videos. In the proposed solution, the ITM operator entirely depends on the decision of the SVM, and in some cases, the SVM fails to identify the scene type, which can result in erroneous decisions, eventually yielding unexpected results in the final HDR. Similar ITM methods [10,47,48] require parameter tuning and manual input; for this reason, in such cases, it is not easy to apply these methods because they could lead to the unrealistic output of low-quality HDR images. In general, conventional ITM methods have limited performance because the information is lost in the saturated and underexposed areas of the images, which makes the recovery of the correct CRF from a single image nearly impossible. As a result of this information loss, ITM-based methods produce HDR images with limited quality.

2.2.2. Learning Based

Learning-based ITM HDRCNN proposed in [49] is a CNN autoencoder aimed at improving saturated areas of arbitrary exposed images. One limitation of the proposed technique is that it fails in the case of underexposed regions, and may introduce artifacts if large overexposed areas are present in the image. Reference [50] designed an encoder–decoder network, which takes the LDR image of the random exposure as input, and instead of directly producing the HDR image, similar to other ITM methods, it generates bracketed images with changing exposures. Due to the unavailability of appropriate datasets for training, the authors synthesized multi-exposure LDR images from existing datasets. Their technique can introduce artifacts with real-life data, especially in the case of highly overexposed regions. ITM-net [51] is another CNN proposing an ITM method, which takes an image in the LDR domain and predicts the tone-mapped HDR image. The LDR and HDR domain images are different, but in ITM-net, the authors did not take into account this fact and did not perform the domain transfer. ExpandNet [11] is an end-to-end CNN ITM method that consists of three branches: local, dilation, and global; each one targets a different level of information in the image. The outputs from the three branches are merged at the end and the final HDR is generated. ExpandNet seems to work well to recover the overexposed areas but fails to recover missing information in the dark regions of the image and, overall, produces darker results. HSVNet [52] is a U-Net-like model that converts arbitrarily exposed LDR images to HDR images. The proposed model works in the HSV (hue saturation value) color domain and introduces custom loss with masks computed over S and V channels to differentiate between underexposed and overexposed areas. Reference [53] suggests using a weakly supervised network that essentially recovers CRF using a single input image, which can generate any number of bracketed exposure images. This method avoids the alignment and merging steps of the classical multi-exposure HDR pipeline, but could introduce hallucination artifacts in some challenging regions of the image. More deep learning-based example methods include [12,54,55,56,57].

2.3. Novel Sensors for HDR

Spatially varying exposure (SVE) [58] is a specialized camera sensor, which is designed to simultaneously capture multiple exposures. The works [59,60] are based on a similar idea as SVE [58] but with different sensor pattern arrangements. These aforementioned methods overcome the issue of image alignment but they all suffer from spatial resolution reduction. There are cameras specifically designed to capture HDR content, for instance, Arri Alexa, Sony F65, and Omron [61]. SpheronVR and PanoScan cameras are designed to capture HDR panoramas [10], although these HDR cameras produce high-quality content, they are still very expensive.

2.4. PFA Camera-Based HDR

The work presented in [3] made the first effort to create single-shot HDR images using PFA cameras. The authors treated the four channels as images taken with different exposures, and the final HDR was reconstructed using the existing method [16]. An extension of this work was proposed in [13], where they adopted an 18-level autoencoder architecture similar to the one introduced in [50]. The major drawback of these approaches is that real-world scenes rarely exhibit high DoLP, and this is required to correctly map a polarimetric image into an image with equivalent exposure time. To mitigate this issue, the authors proposed another technique, named DPHR [14]. In order to improve poorly exposed pixels with low DoLP, [14] generated HDR with a PFA camera by using the degree of polarization as the cue in a CNN (similar to U-Net [62]) to create a feature mask. Image areas with high DoLP were reconstructed using a traditional approach (as in [3]), while pixels with moderate DoLP were reconstructed with both CNN and traditional techniques. Similar to the aforementioned single LDR to HDR techniques, PFA-based HDR methods also encounter the challenge of poor performance in underexposed areas. In [15], we introduced a novel stereo PFA camera setup, which provides a significant increase in the dynamic range of the HDR image, compared to using single PFA cameras. The HDR images computed with [15] are visually more pleasing and closer to the ground truth when compared to other existing methods. Despite these benefits, the stereo setup involves certain constraints, including being expensive, being complicated to manage, and image alignment issues. In Section 3, we introduce the basic concepts and equations for PFA-based HDR reconstruction, provide more details about the stereo setup and the model we employed to compute the final reconstruction with two cameras, and discuss the mentioned drawbacks.

3. HDR with Stereo PFA Cameras

Polarization is a fundamental property of light that arises from its vector nature [63]. When captured, this feature can be exploited for several applications, such as quality control [64], material classification [65], 3D reconstruction [66,67,68], and remote sensing [69,70], to name a few. The method used to acquire the polarization state of a scene often consists of placing a linear polarizer in front of an ordinary camera and rotating it to take pictures with different angles. Considering a single pixel, the observed intensities can be used to compute the Stokes vector S = ( S 0 , S 1 , S 2 , S 3 ) , which is a mathematical structure representing the polarization state of a light beam. From the values in S, we can easily compute the AoLP (angle of linear polarization) and DoLP (degree of linear polarization) as follows:
DoLP = S 1 2 + S 2 2 S 0 , AoLP = 1 2 arctan S 2 S 1 .
Note that S 3 corresponds to circular polarization, which cannot be observed without the use of a retarder, which is, in general, rarely observable in nature [71]. As in multi-shot HDR, the described setup with a rotating polarizer brings several disadvantages related to the need for image alignment and filter calibration. To avoid such problems, polarimetric filter array (PFA) cameras have been introduced. This device mounts an internal filter that allows taking polarized images with four pre-defined filter orientations ( 0 , 45 , 90 , and 135 ) in a single shot.
If we focus on a single camera pixel, a polarizer filter with an angle α reduces the incoming image irradiance I ^ , according to the following equation, depending on the incoming light polarization state:
I ^ = 1 2 I ^ 1 + D cos ( 2 θ 2 α )
where D [ 0 , 1 ] is the incoming DoLP for that pixel and θ [ π , π ] is the AoLP for the same pixel, while α [ π , π ] is the orientation of the linear polarizer with respect to the camera’s x-axis. In the case of PFA cameras, we observe macro-pixels, where each macro-pixel consists of four channels corresponding to four different filter orientations, meaning that α is going to assume fixed values ( 0 , 45 , 90 , or 135 ), as mentioned. As a consequence of this attenuation, the usage of such filters can be related to a reduced shutter speed and, thus, exploited to perform HDR reconstruction from the four images. However, Equation (2) clearly shows that the amount of attenuation is proportional to the DoLP D . In general, high DoLP is seldom observed in a general scene, as shown in Figure 2, where the DoLP and AoLP of an outdoor scene are displayed. Indeed, the DoLP image shows quite low values, except for the sky region, which is partially polarized with respect to the rest of the scene.
The top row of Figure 3 shows the same scene captured by the four channels of a PFA camera (filters oriented at 0 , 45 , 90 , and 135 ). Since there is some polarized light coming from the sky, the intensity in this area exhibits a different response across the different filters, but only in one image. Since the rest of the scene has DoLP 0 , there are no variations and, thus, in this area, HDR reconstruction would be useless. Fatima et al. [15] proposed a solution to overcome this issue by mounting an additional external linear polarizer on the PFA camera, so that incoming light is polarized ideally with DoLP 1 . Figure 3 (bottom row) displays the same scene captured with such an additional linear filter in front of the lenses. This time, the response observed in the four channels is similar to what would be obtained through exposure bracketing. Since the external filter overrides the polarization state of the original scene (needed to recover the equivalent exposure time), in [15], a stereo setup is proposed, such that a second PFA is used to acquire the original image (i.e., the AoLP and DoLP). This camera setup results in a model through which the per-pixel exposure time can be effectively estimated.
Suppose we observe a scene with irradiance I , AoLP θ , and DoLP D . Employing a regular PFA camera to capture the scene, according to Equation (2), we obtain four different intensity images I 0 , I 45 , I 90 , I 135 , one for each internal polarizer orientation α i , where i { 0 , 45 , 90 , 135 } . Note that θ and D are intrinsic features of the scene since they depend on the polarization state of the acquired light, and can be computed from the four I i . When the same scene is captured by another PFA camera, denoted as c a m f , with an additional externally mounted linear polarizer at an angle α f , we obtain different AoLP and DoLP, denoted, respectively, as θ f and D f . The combination of the external and internal filters can be modeled by applying the relationship in Equation (2), producing the following irradiance for the i-th filter:
I ^ i = 1 4 I 1 + D cos ( 2 θ 2 α f ) 1 + D f cos ( 2 θ f 2 α i ) .
This relation is exploited to compute equivalent exposure times for each pixel and proceed with HDR reconstruction. Experiments from the original paper show that by merging the responses of both cameras, the dynamic range of images is significantly increased as compared to using a single PFA camera. Despite these benefits, the described setup has some limitations:
  • It requires two PFA cameras mounted on a rigid structure, which is expensive.
  • The stereo setup is complicated to manage and not easy to use (e.g., shutter synchronization or the calibration of the camera’s response function).
  • Although two cameras have a minimum baseline, there is still a need to align the responses of both cameras, and mapping can be erroneous, especially for objects near the cameras.

4. Single-Shot HDR with the PFA Camera

The mentioned issues can be resolved by reducing the stereo’s camera setup to a single physical camera with an external filter; simulating the response of the other camera to obtain the actual AoLP and DoLP of the scene can be obtained to proceed with the HDR reconstruction, as described in the previous section. In this work, we present a technique that effectively allows discarding the second camera, thus reducing the disadvantages of the previous stereo configuration.
The idea is to acquire the scene, employing only one PFA camera with the additional linear polarizer, and then using a data-driven model to simulate the outcome of an unfiltered PFA camera based on the captured image. To this end, we trained a neural network to predict the actual AoLP and DoLP of a scene from a filtered image, observing that (i) some information is still preserved after passing through the linear filter (which is not perfect) and (ii) the training process demonstrates the network’s ability to generalize to typical configurations, to devise the expected values for the angle and degree. In this way, we are able to create a virtual stereo camera setup composed of the real camera and the output of the discussed model. Then, these two images can be combined, as already discussed in the previous section, and presented in [15]. Note that the computed HDR image exhibits a broader range of values compared to those captured by a regular camera; indeed, we obtain real values that should typically be quantized to 256 levels to visualize the results. Usually, this step involves a tone mapping technique; therefore, we also propose a second neural model that is designed to take the computed HDR image and compute the tone-mapped version, which improves the final output. Figure 1 presents an overview of the complete pipeline presented in this paper. We can describe the whole HDR reconstruction process with a single PFA camera through the following main blocks:
  • The recovery network is in charge of synthesizing the response of a regular PFA camera (in terms of AoLP and DoLP) from an image taken with an additional external linear polarizer. The network architecture includes some components, referred to as mosaiced convolutions from the PFA demosaicing network (PFADN) [72]; these components have been shown to be effective for demosaicing the PFA raw camera images to obtain a full-resolution intensity image and AoLP.
  • The image acquired via the PFA camera with the additional filter is merged with the recovery network output using the camera model presented in [15] to produce an HDR image. Note that this step is deterministic and not learning-based since it is a direct application of the theory presented in the previous section.
  • Finally, the tone mapping network takes as input the computed HDR image and is designed to make the output perceptually similar to the ground truth.
In the following part of the current section, we will describe in detail the mentioned components, namely, the recovery network, HDR generation, and tone mapping network.

4.1. Recovery Network

As already discussed, we acquire the scene with a single PFA camera (denoted as c a m f ) with an external linear polarizer mounted with an angle α f . As a result, the camera captures a mosaiced image I f , where the four channels (associated with each internal filter) are interleaved. Suppose that θ f and D f are the angle and degree of polarization of the acquired data and let θ and D be the actual AoLP and DoLP of the incoming light, which would have been captured by the camera without the additional filter. The recovery network (RecNet) takes as input the mosaiced image I f and aims at recovering AoLP ( θ ) and DoLP ( D ) , which is essentially equivalent to the removal of the additional polarizer filter.
RecNet is built using mosaiced convolution (MConv) operations. These convolutions are designed to extract features from a polarimetric mosaic input by taking into account the PFA pattern; details of MConv can be found in [72]. Figure 4 shows the complete architecture of the proposed network. It takes a tensor of size w × h × 1 as the input, and then two MConv blocks extract features from I f , giving a feature map of size w × h × 16 , where 16 is the total number of computed filters (or depth). In the following layer, such a tensor is demosaiced into a ( w / 2 ) × ( h / 2 ) × 64 tensor, concatenating the four channels depthwise. After performing a 1 × 1 2D convolution followed by ReLU activation, we obtain a ( w / 2 ) × ( h / 2 ) × 12 tensor. After that, the network splits into two parts: the first branch of the network is responsible for recovering the DoLP D of a scene, while the second branch estimates the AoLP θ .
Regarding the first part, we know that D [ 0 , 1 ] ; therefore, it can be considered a grayscale image normalized between 0 and 1. To produce D , 2D convolutions with ReLU activations are applied in each layer to produce a sequence of 64, 32, 12, 8, 8, 1 features maps. After the last convolution, the network produces an ( w / 2 ) × ( h / 2 ) intensity image that represents the predicted DoLP of the scene without the filter.
The second network branch aims at producing the scene’s original AoLP θ . In order to do that, it performs three 2D convolutions, obtaining tensors with depths of 12 , 8 , 1 . Unlike [72], we are not interested in the full-resolution output, so we do not perform upscaling, and directly output the two vector components ( 2 u , 2 v ) of the AoLP θ of the scene, such that:
θ = 1 2 atan 2 ( 2 v , 2 u ) .
The loss function employed for RecNet is as follows:
L R e c N e t ( D , D ^ , A , A ^ ) = γ L D ( D , D ^ ) + ( 1 γ ) L A ( A , A ^ )
where we set γ = 0.5 , while L D and L A are, respectively, the loss functions for the degree and the angle:
L D ( D , D ^ ) = ( 1 β ) | D D ^ | + β ( 1 SSIM ( D , D ^ ) ) L A ( A , A ^ ) = A A ^ 2 2
where SSIM is the structural similarity function [73] between the predicted and ground truth DoLP, while A ^ is the two-channel image ( 2 u , 2 v ) and A = ( cos ( 2 θ ) , sin ( 2 θ ) ) . The weighting parameter β in our settings has been set to 0.85 .

4.2. HDR Generation

The four intensity values I 0 , I 45 , I 90 , I 135 captured by c a m f with the exposure time t c are related to the actual irradiance I ^ i (defined in Equation (3)), as follows:
g ( I i ) = I ^ i t c
where g ( · ) is the inverse camera response function (ICRF). Then, expanding I ^ i as in Equation (3), we can easily substitute I ^ i t c with I t i , where
t i = 1 4 t c 1 + D cos ( 2 θ 2 α f ) 1 + D f cos ( 2 θ f 2 α i ) .
Indeed, we can define the t i (with i { 0 , 45 , 90 , 135 } ) as equivalent exposure times. As a consequence, the acquired images I 0 , I 45 , I 90 , I 135 can be seen as the intensities one would have observed by exposing the image, respectively, with times t 0 , t 45 , t 90 , t 135 , without the polarizers. Since each pixel is an independent observation, the equivalent exposure times vary across the image and, thus, are considered per-pixel.
Let t c be the shutter speed of c a m f . After RecNet produces the scene’s AoLP θ and DoLP D , these are combined with the angle θ f and degree D f (i.e., observed by c a m f ) to compute the equivalent exposure time as in Equation (8). Finally, the intensities from the four filters I 0 , I 45 , I 90 , I 135 , along with the equivalent exposure times t i , are used to compute HDR values via the following weighted average:
I H D R = i T w ( I i ) I i t i i T w ( I i )
where T = { 0 , 45 , 90 , 135 } is the set of filter orientations and w is the Gaussian weight function, defined as w ( x ) = exp ( ( x 0.5 ) 2 2 σ 2 ) . The idea behind w is to give more importance to values that are closer to the middle of the response and, consequently, assign a higher weight to the properly exposed pixels. By computing the value I H D R for each observed pixel, we obtain the complete HDR image H .

4.3. Tone Mapping Network

The tone mapping network (TMNet) takes H as the input and performs a tone mapping operation, producing the final image that can be visualized. As shown in Figure 5, the network consists of five 2D convolutions layers, which extract 64, 32, 16, 8, 1 feature maps, respectively; the last one corresponds to the tone-mapped HDR image H ^ . In each layer, the receptive field is progressively decreased, having kernels of sizes 9, 5, 3, 1, 1, respectively. This results in mapping only important features to the output layer. The ReLU function is used to activate all layers, except the output. The loss function of TMNet is designed to take into account the overall quality of the output image; for this reason, we opted for a linear combination of the SSIM and a perceptual loss:
L T M N e t ( H ¯ , H ^ ) = α 1 SSIM ( H ¯ , H ^ ) + ( 1 α ) L p e r ( H ¯ , H ^ )
where H ¯ is the ground truth HDR image and α is a balancing value that we empirically set to 0.5 after some preliminary tests. The function L p e r is a perceptual loss function that takes as input the two HDR images, H ¯ and H ^ , and computes the sum of the squared differences of the output of the intermediate feature maps ϕ l from the predefined layers of the VGG-19 architecture [74]. In particular:
L p e r ( H ¯ , H ^ ) = l L ( ϕ l ( H ¯ ) ϕ l ( H ^ ) ) 2
where L = { p o o l 1 , p o o l 2 , p o o l 3 } are the VGG layer identifiers.

5. Experimental Section

In this section, we first describe the process we employed to acquire a suitable dataset that was used to train both RecNet and TMNet, along with the training process. Then, we evaluated the quality of the proposed method by comparing it with state-of-the-art methods, both quantitatively and qualitatively.

5.1. Dataset Acquisition

We used a single PFA camera setup to acquire the data needed to train the model and assess the quality of the final HDR reconstruction. We employed a FLIR Blackfly monochrome PFA camera with the Sony IMX250MZR sensor (Tokyo, Japan), which gave raw images that were 2448 × 2048 pixels (before demosaicing). The camera ICRF was calibrated using [75], and applied to all of the input images as a preprocessing step.
We positioned the camera on a stable support for each unique scene. First, we captured a single picture with the additional external filter. Following that, we captured a sequence of 30 images without the filter, gradually increasing the exposure time. The whole acquisition process lasted a couple of seconds, and we captured the static scenes in order to be able to perform reasonable comparisons with other methods. For each set of images, we demosaiced all 30 exposures to obtain the scene intensities, which were used to compute the HDR ground truth for the observed scenes using the method described in [16]. In total, 20 high-dynamic indoor and outdoor scenes with different ambient conditions were captured. Finally, the acquired images were divided into blocks of size 128 × 128 , resulting in a total of 5100 extracted patches.

5.2. Training Procedure

Our HDR reconstruction pipeline was composed of two independent networks, which were trained using different portions of datasets. We recall that the objective of the recovery network is to take as input a mosaiced image from a PFA camera with an external linear polarizer and estimate the original scene features in terms of AoLP and DoLP. In order to do this, we specifically designed a dataset to simulate the outcome of such acquisitions in real-world settings. Indeed, given a raw mosaiced image taken with a PFA camera, it is quite straightforward to simulate the effect of an additional filter outside the lenses. Recalling Equation (3), the observed intensity is connected to the other parameters as follows:
I ^ i = 1 4 S 0 1 + D cos ( 2 θ 2 α f ) 1 + D f cos ( 2 θ f 2 α i )
where S 0 is the intensity captured by the PFA camera (the first Stokes parameter), D and θ are the actual DoLP and AoLP of the scene, and α i is the angle of the internal polarizer filter array α i { 0 , 45 , 90 , 135 } . The values D f and θ f depend on the external filter and can be easily generated to simulate different acquisition conditions to simulate a high number of training images without physically adding and removing the filter. Therefore, since RecNet requires several images for the training process, we synthetically generated filtered images starting from real PFA camera data. Specifically, we used 50 % of our data and added the dataset presented in [76] (“sunny outdoor” acquisitions). The ground truth AoLP and DoLP are directly calculated from the captured dataset, while the network input is synthesized by applying Equation (11). We simulated random D f between 0.8 and 1 and randomized the value θ f to avoid completely dark images in individual channels.
The tone mapping network was trained exclusively on our dataset, with the HDR ground truth computed from exposure bracketing, as previously explained; tone mapping was applied as described in [77]. We chose random exposures from each scene, avoiding complete dark or white images so that the presented model could learn to reconstruct the HDR irrespective of the exposure time.
For both networks, datasets were divided into 70/30 for training and testing, respectively, and the Adam optimizer was used with a 10 3 learning rate and trained up to convergence (∼30 epochs for each network). The model was implemented using TensorFlow and NumPy libraries and the process ran on an NVIDIA GeForce RTX 2080 Ti GPU with a batch size set to 64.

5.3. Quantitative Analysis

In order to evaluate the performance of the proposed technique, we designed a set of experiments to first assess the effectiveness of RecNet and then to test the final HDR reconstruction by comparing the outcomes of several approaches.
As already discussed, we trained RecNet with synthesized filters, starting from our dataset and other datasets. Figure 6 shows two example outputs of RecNet from the test set; the input is the mosaiced image with the (simulated) external filter. The degree of linear polarization was successfully recovered by the network; the angle results were almost good in the first row, while in the second, we can see some differences. This is actually not a problem for our application because we are interested in recovering the original scene angles for regions with high DoLP, where the angle is relevant. Indeed, in regions where the DoLP is very close to zero, the camera model described by Equation (8) discards the angle value since the term has a minimal effect. As a consequence, we aim to recover only the angles for which the degree is quite high. The overall average angle loss for the test set was 36 .
The performance of the presented method was verified by comparing it to some existing single-image-based HDRI techniques. In total, seven state-of-the-art methods were selected for comparison; among these, five were learning-based and two were algorithmic approaches. In particular, KO [48] is an inverse tone mapping technique while Wu et al. [3] presented an HDR technique specifically designed for PFA cameras. The learning-based HDRI methods we compared with are the two-stage HDR [78], HDRCNN [49], Deep-HDR [79], ExpandNet [11], and DPHDR [13]. For all of the listed methods, we used the implementations of the authors as well as pre-trained weights where possible. Note that Wu et al. [3] and DPHDR [13] are methods specifically designed to generate HDR images by taking a PFA image as input, the rest of the techniques are not designed for PFA cameras; hence, intensity images are provided as input. We used our acquired dataset for a comparison on the test set, and Reinhard tone mapping [80] was applied to the HDR outputs only for methods that do not directly output a tone-mapped image. In all of the experiments, the ground truth was computed as previously described, and the output of our proposed method was generated by using a single image captured with an additional linear polarizer. Moreover, in order to show the effectiveness of the tone mapping network proposed in our approach, we also compared the results obtained from our model without the TMNet. In other words, we directly tone-mapped the HDR output with [80], excluding the second network.
Table 1 shows the comparison results. For all techniques, we computed the peak signal-to-noise ratio (PSNR) and multi-scale structural similarity (MS-SSIM) [81]. PSNR is computed as follows:
PSNR ( X , X G T ) = 20 l o g 10 M A X I MSE ( X , X G T )
where M A X I is the maximum possible value among image pixels, and MSE denotes the mean squared error between the ground truth X G T and the predicted image X. MS-SSIM is an extension of the SSIM [73] function, which calculates and combines the SSIM of an image at different scales. Given a pair of ground truth X G T and predicted images X, MS-SSIM is computed as follows:
MS - SSIM ( X , X G T ) = [ l M ( X , X G T ) ] α M . j = 1 M [ c j ( X , X G T ) ] β j [ s j ( X , X G T ) ] γ j
where M is the number of scales, l j , c j , and s j refer to the luminance, contrast, and similarity measures of the j-th scale, respectively. Parameters α j , β j , and γ j are used to assign weights to different measurements for the j-th scale.
We can see that our method produced the highest average PSNR (dB) and MS-SSIM for the test dataset. Although Wu et al. [3] and DPHDR are built for PFA cameras, they do not provide any significant improvements, for two reasons. First: The DoLP of the scenes is generally very small; hence, it fails to take full advantage of polarimetric data. Second: DPHDR is a learning-based method but the model provided by the authors is not suitable to work with such data or it overfits some patterns and does not generalize. Moreover, DPHDR works on patches of size 512 × 512 and does not provide a way to connect these patches to produce a full-resolution output of ( 1024 × 1224 ) . We also notice that our method without TMNet (second row) still offers results that are better when compared to other methods, meaning that the output provided by RecNet is effective in simulating the second camera with the additional filter. The rest of the data-driven techniques (based on a single image) presented in the table offer lower outcomes: the best is the two-stage HDR ( 19.59 dB), which can be compared to our model without TMNet ( 20.51 dB).

5.4. Qualitative Analysis

As the last part of our analysis, we qualitatively compared some of the results from our method with respect to other approaches. The first thing that we wanted to compare is the final result with or without the contribution of the tone mapping network. Figure 7 shows three example images that were obtained by applying the full proposed pipeline with TMNet (second column) and directly applying the tone mapping [80] after HDR generation (third column). We can notice that excluding TMNet leads to the worst results and saturated regions, especially in the sky area, whereas our full approach manages to recover better details in such challenging settings. Figure 8 compares the quality of all of the proposed methods from a scene with a simulated additional filter. Our method demonstrates superior quality results compared to other methods, even though we employed a synthetic filter approach to train RecNet. This shows that the proposed training procedure is indeed effective in simulating real-environment conditions when employing the external filter in our method.
Finally, Figure 9 and Figure 10 are produced with our acquired dataset and physical filter mounted on the camera. In particular, Figure 10 shows a qualitative comparison of all the methods, and Figure 9 shows a close-up of the three best methods, according to Table 1. In detail, the third column in Figure 9 shows the PFA-based HDRI Wu et al. [3] and the last one is the single image HDR reconstruction method, denoted as the two-stage HDR method. In general, we notice that Wu et al. failed to deal with the oversaturated regions of the image, resulting in overall brighter results. Moreover, in Figure 10, the results generated by [3] have overexposed regions; this works better when compared to other techniques except ours. The technique proposed by [3] has been specifically designed for PFA cameras, but due to the lower DoLP of the scenes, the four filter images taken by PFA cameras have, approximately, the same exposure times and, consequently, there is not enough increase in the dynamic range of the resulting HDR image. On the other hand, the output of the two-stage HDR in the overexposed regions is somehow similar to ground truth, but it could not recover low-level details in complicated areas. Our method produced better quality HDR images, which exhibited a contrast similar to the ground truth while retaining fine details, which is evident in the zoomed-in patches.
Qualitative results of other competing techniques can be found in Figure 10. The techniques have their own shortcomings and they all suffer from different kinds of artifacts, leading to poor results. The under-performance of deep learning-based techniques, such as ExpandNet, HDRCNN, and Deep-HDR, can be explained by different reasons. One possible reason is the training process, as these methods are trained on specific datasets, and due to overfitting, it is difficult for these pre-trained networks to reproduce better-quality results. Another reason could be the lack of a standard protocol for HDR reconstruction problems, such as dataset and evaluation conditions. These methods might report good performances with their set conditions but they fail when dealing with different scenes.

6. Conclusions

In this paper, we proposed an HDR reconstruction method based on a data-driven virtual stereo PFA camera setup, requiring only a single shot with an additional external linear polarizer. The idea is based on a previous method that uses a physical stereo setup with two synchronized PFA cameras to generate HDR images that exploit linear polarizers. The virtual setup proposed in this work overcomes the drawbacks of the previous approach by discarding one camera while taking advantage of the polarimetric imaging model. To do so, we introduced a novel CNN architecture, denoted as the recovery network (RecNet), which is designed to eliminate the need for a second camera. Indeed, such a model is trained to predict the original angle and degree of polarization from the raw mosaiced image taken with the additional filter. In this way, the recovery network output is directly used to compute the HDR image. Finally, the HDR image is further improved through a second CNN model that performs optimal tone mapping on the input data. In addition, since there are no publicly available datasets for this kind of task, we created a new dataset that is designed for learning-based polarimetric HDR imaging. We carefully performed the data acquisition procedure so that we could effectively acquire challenging scenes, and at the same time, we compared the results obtained with our method and other methods designed for single-shot HDRI. Experiments show that the described two-stage neural network pipeline increases the range of estimated exposure times and improves the final image quality even further when applying the second network. As part of future work, we will aim to improve the two CNN model performances in terms of both the angle recovery and tone mapping, so that the proposed system can be applied in more challenging environments.

Author Contributions

Conceptualization, M.P.; data curation, T.F.; investigation, M.P. and T.F.; methodology, M.P.; resources, M.W.; software, T.F.; supervision, M.P. and M.W.; validation, M.P. and M.W.; visualization, T.F.; writing—original draft, M.P. and T.F.; writing—review and editing, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mertens, T.; Kautz, J.; Van Reeth, F. Exposure fusion: A simple and practical alternative to high dynamic range photography. In Proceedings of the Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2009; Volume 28, pp. 161–171. [Google Scholar]
  2. Ladas, N.; Chrysanthou, Y.; Loscos, C. Improving tracking accuracy using illumination neutralization and high dynamic range imaging. In High Dynamic Range Video; Elsevier: Amsterdam, The Netherlands, 2017; pp. 203–213. [Google Scholar]
  3. Wu, X.; Zhang, H.; Hu, X.; Shakeri, M.; Fan, C.; Ting, J. HDR reconstruction based on the polarization camera. IEEE Robot. Autom. Lett. 2020, 5, 5113–5119. [Google Scholar] [CrossRef]
  4. Seger, U. Hdr imaging in automotive applications. In High Dynamic Range Video; Elsevier: Amsterdam, The Netherlands, 2016; pp. 477–498. [Google Scholar]
  5. Ramponi, G.; Badano, A.; Bonfiglio, S.; Albani, L.; Guarnieri, G. An Application of HDR in Medical Imaging. In High Dynamic Range Video; Elsevier: Amsterdam, The Netherlands, 2016; pp. 499–518. [Google Scholar]
  6. Wu, J.C.H.; Lin, G.S.; Hsu, H.T.; Liao, Y.P.; Liu, K.C.; Lie, W.N. Quality enhancement based on retinex and pseudo-HDR synthesis algorithms for endoscopic images. In Proceedings of the 2013 Visual Communications and Image Processing (VCIP), Kuching, Malaysia, 17–20 November 2013; pp. 1–5. [Google Scholar] [CrossRef]
  7. Suh, H.K.; Hofstee, J.W.; Van Henten, E.J. Improved vegetation segmentation with ground shadow removal using an HDR camera. Precis. Agric. 2018, 19, 218–237. [Google Scholar] [CrossRef] [Green Version]
  8. Karr, B.; Chalmers, A.; Debattista, K. High dynamic range digital imaging of spacecraft. In High Dynamic Range Video; Elsevier: Amsterdam, The Netherlands, 2016; pp. 519–547. [Google Scholar]
  9. Khan, E.A.; Akyuz, A.O.; Reinhard, E. Ghost removal in high dynamic range images. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 2005–2008. [Google Scholar]
  10. Banterle, F.; Ledda, P.; Debattista, K.; Chalmers, A. Inverse tone mapping. In Proceedings of the 4th International Conference on Computer Graphics and Interactive Techniques in Australasia and Southeast Asia, Kuala Lumpur, Malaysia, 29 November–2 December 2006; pp. 349–356. [Google Scholar]
  11. Marnerides, D.; Bashford-Rogers, T.; Hatchett, J.; Debattista, K. Expandnet: A deep convolutional neural network for high dynamic range expansion from low dynamic range content. In Proceedings of the Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2018; Volume 37, pp. 37–49. [Google Scholar]
  12. Khan, Z.; Khanna, M.; Raman, S. Fhdr: Hdr image reconstruction from a single ldr image using feedback network. In Proceedings of the 2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Ottawa, ON, Canada, 11–14 November 2019; pp. 1–5. [Google Scholar]
  13. Ting, J.; Wu, X.; Hu, K.; Zhang, H. Deep snapshot HDR reconstruction based on the polarization camera. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1769–1773. [Google Scholar]
  14. Ting, J.; Shakeri, M.; Zhang, H. Deep Polarimetric HDR Reconstruction. arXiv 2022, arXiv:2203.14190. [Google Scholar]
  15. Fatima, T.; Pistellato, M.; Torsello, A.; Bergamasco, F. One-Shot HDR Imaging via Stereo PFA Cameras. In Proceedings of the International Conference on Image Analysis and Processing, Lecce, Italy, 23–27 May 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 467–478. [Google Scholar]
  16. Debevec, P.E.; Malik, J. Recovering high dynamic range radiance maps from photographs. In Proceedings of the ACM SIGGRAPH 2008 Classes, Los Angeles, CA, USA, 11–15 August 2008; pp. 1–10. [Google Scholar]
  17. Lu, P.Y.; Huang, T.H.; Wu, M.S.; Cheng, Y.T.; Chuang, Y.Y. High dynamic range image reconstruction from hand-held cameras. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 509–516. [Google Scholar]
  18. Hasinoff, S.W.; Sharlet, D.; Geiss, R.; Adams, A.; Barron, J.T.; Kainz, F.; Chen, J.; Levoy, M. Burst photography for high dynamic range and low-light imaging on mobile cameras. ACM Trans. Graph. (ToG) 2016, 35, 1–12. [Google Scholar] [CrossRef]
  19. Zhang, W.; Cham, W.K. Gradient-directed composition of multi-exposure images. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 530–536. [Google Scholar] [CrossRef]
  20. Sun, N.; Mansour, H.; Ward, R. HDR image construction from multi-exposed stereo LDR images. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, 26–29 September 2010; pp. 2973–2976. [Google Scholar] [CrossRef] [Green Version]
  21. Mann, S.; Picard, R. On being “undigital” with digital cameras: Extending dynamic range by combining differently exposed pictures. In Proceedings of the IS&T 48th Annual Conference Society for Imaging Science and Technology Annual Conference, Washington, DC, USA, 7–11 May 1995. [Google Scholar]
  22. Kirk, K.; Andersen, H.J. Noise Characterization of Weighting Schemes for Combination of Multiple Exposures. In Proceedings of the British Machine Vision Conference, Edinburgh, UK, 4–7 September 2006; Volume 3, pp. 1129–1138. [Google Scholar]
  23. Granados, M.; Ajdin, B.; Wand, M.; Theobalt, C.; Seidel, H.P.; Lensch, H.P. Optimal HDR reconstruction with linear digital cameras. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 215–222. [Google Scholar]
  24. Pistellato, M.; Cosmo, L.; Bergamasco, F.; Gasparetto, A.; Albarelli, A. Adaptive Albedo Compensation for Accurate Phase-Shift Coding. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2450–2455. [Google Scholar] [CrossRef]
  25. Umair, M.B.; Iqbal, Z.; Faraz, M.A.; Khan, M.A.; Zhang, Y.D.; Razmjooy, N.; Kadry, S. A Network Intrusion Detection System Using Hybrid Multilayer Deep Learning Model. In Big Data; Mary Ann Liebert, Inc.: New Rochelle, NY, USA, 2022. [Google Scholar]
  26. Huang, Q.; Ding, H.; Razmjooy, N. Optimal deep learning neural network using ISSA for diagnosing the oral cancer. Biomed. Signal Process. Control 2023, 84, 104749. [Google Scholar] [CrossRef]
  27. Gasparetto, A.; Ressi, D.; Bergamasco, F.; Pistellato, M.; Cosmo, L.; Boschetti, M.; Ursella, E.; Albarelli, A. Cross-Dataset Data Augmentation for Convolutional Neural Networks Training. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 910–915. [Google Scholar] [CrossRef] [Green Version]
  28. Ram Prabhakar, K.; Sai Srikar, V.; Venkatesh Babu, R. Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4714–4722. [Google Scholar]
  29. Xu, H.; Ma, J.; Zhang, X.P. MEF-GAN: Multi-exposure image fusion via generative adversarial networks. IEEE Trans. Image Process. 2020, 29, 7203–7216. [Google Scholar] [CrossRef]
  30. Kalantari, N.K.; Ramamoorthi, R. Deep high dynamic range imaging of dynamic scenes. ACM Trans. Graph. 2017, 36, 144. [Google Scholar] [CrossRef] [Green Version]
  31. KS, G.R.; Biswas, A.; Patel, M.S.; Prasad, B.P. Deep multi-stage learning for hdr with large object motions. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 4714–4718. [Google Scholar]
  32. Yan, Q.; Gong, D.; Shi, Q.; Hengel, A.v.d.; Shen, C.; Reid, I.; Zhang, Y. Attention-guided network for ghost-free high dynamic range imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1751–1760. [Google Scholar]
  33. Pu, Z.; Guo, P.; Asif, M.S.; Ma, Z. Robust high dynamic range (hdr) imaging with complex motion and parallax. In Proceedings of the Asian Conference on Computer Vision, Virtual, 30 November–4 December 2020. [Google Scholar]
  34. Nazarczuk, M.; Catley-Chandar, S.; Leonardis, A.; Pellitero, E.P. Self-supervised HDR Imaging from Motion and Exposure Cues. arXiv 2022, arXiv:2203.12311. [Google Scholar]
  35. Catley-Chandar, S.; Tanay, T.; Vandroux, L.; Leonardis, A.; Slabaugh, G.; Pérez-Pellitero, E. FlexHDR: Modeling Alignment and Exposure Uncertainties for Flexible HDR Imaging. IEEE Trans. Image Process. 2022, 31, 5923–5935. [Google Scholar] [CrossRef]
  36. Nejati, M.; Karimi, M.; Soroushmehr, S.R.; Karimi, N.; Samavi, S.; Najarian, K. Fast exposure fusion using exposedness function. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 2234–2238. [Google Scholar]
  37. Li, Z.; Wei, Z.; Wen, C.; Zheng, J. Detail-enhanced multi-scale exposure fusion. IEEE Trans. Image Process. 2017, 26, 1243–1252. [Google Scholar] [CrossRef]
  38. Ma, K.; Duanmu, Z.; Zhu, H.; Fang, Y.; Wang, Z. Deep guided learning for fast multi-exposure image fusion. IEEE Trans. Image Process. 2019, 29, 2808–2819. [Google Scholar] [CrossRef] [PubMed]
  39. Lecouat, B.; Eboli, T.; Ponce, J.; Mairal, J. High dynamic range and super-resolution from raw image bursts. arXiv 2022, arXiv:2207.14671. [Google Scholar] [CrossRef]
  40. Shaw, R.; Catley-Chandar, S.; Leonardis, A.; Pérez-Pellitero, E. HDR Reconstruction from Bracketed Exposures and Events. arXiv 2022, arXiv:2203.14825. [Google Scholar]
  41. Yoon, H.; Uddin, S.N.; Jung, Y.J. Multi-Scale Attention-Guided Non-Local Network for HDR Image Reconstruction. Sensors 2022, 22, 7044. [Google Scholar] [CrossRef] [PubMed]
  42. Song, M.; Tao, D.; Chen, C.; Bu, J.; Luo, J.; Zhang, C. Probabilistic exposure fusion. IEEE Trans. Image Process. 2011, 21, 341–357. [Google Scholar] [CrossRef] [PubMed]
  43. Li, Z.G.; Zheng, J.H.; Rahardja, S. Detail-enhanced exposure fusion. IEEE Trans. Image Process. 2012, 21, 4672–4676. [Google Scholar] [PubMed]
  44. Tico, M.; Gelfand, N.; Pulli, K. Motion-blur-free exposure fusion. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 3321–3324. [Google Scholar]
  45. Zhang, W.; Cham, W.K. Reference-guided exposure fusion in dynamic scenes. J. Vis. Commun. Image Represent. 2012, 23, 467–475. [Google Scholar] [CrossRef]
  46. Kuo, P.H.; Tang, C.S.; Chien, S.Y. Content-adaptive inverse tone mapping. In Proceedings of the 2012 Visual Communications and Image Processing, San Diego, CA, USA, 27–30 November 2012; pp. 1–6. [Google Scholar]
  47. Kovaleski, R.P.; Oliveira, M.M. High-quality brightness enhancement functions for real-time reverse tone mapping. Vis. Comput. 2009, 25, 539–547. [Google Scholar] [CrossRef]
  48. Kovaleski, R.P.; Oliveira, M.M. High-quality reverse tone mapping for a wide range of exposures. In Proceedings of the 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images, Rio de Janeiro, Brazil, 26–30 August 2014; pp. 49–56. [Google Scholar]
  49. Eilertsen, G.; Kronander, J.; Denes, G.; Mantiuk, R.K.; Unger, J. HDR image reconstruction from a single exposure using deep CNNs. ACM Trans. Graph. (TOG) 2017, 36, 1–15. [Google Scholar] [CrossRef] [Green Version]
  50. Endo, Y.; Kanamori, Y.; Mitani, J. Deep reverse tone mapping. ACM Trans. Graph. 2017, 36, 177. [Google Scholar] [CrossRef]
  51. Kinoshita, Y.; Kiya, H. ITM-Net: Deep inverse tone mapping using novel loss function considering tone mapping operator. IEEE Access 2019, 7, 73555–73563. [Google Scholar] [CrossRef]
  52. Lee, M.J.; Rhee, C.h.; Lee, C.H. HSVNet: Reconstructing HDR Image from a Single Exposure LDR Image with CNN. Appl. Sci. 2022, 12, 2370. [Google Scholar] [CrossRef]
  53. Le, P.H.; Le, Q.; Nguyen, R.; Hua, B.S. Single-Image HDR Reconstruction by Multi-Exposure Generation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 4063–4072. [Google Scholar]
  54. Gharbi, M.; Chen, J.; Barron, J.T.; Hasinoff, S.W.; Durand, F. Deep bilateral learning for real-time image enhancement. ACM Trans. Graph. (TOG) 2017, 36, 1–12. [Google Scholar] [CrossRef]
  55. Moriwaki, K.; Yoshihashi, R.; Kawakami, R.; You, S.; Naemura, T. Hybrid loss for learning single-image-based HDR reconstruction. arXiv 2018, arXiv:1812.07134. [Google Scholar]
  56. Wu, G.; Song, R.; Zhang, M.; Li, X.; Rosin, P.L. LiTMNet: A deep CNN for efficient HDR image reconstruction from a single LDR image. Pattern Recognit. 2022, 127, 108620. [Google Scholar] [CrossRef]
  57. Cao, G.; Zhou, F.; Liu, K.; Wang, A.; Fan, L. A decoupled kernel prediction network guided by soft mask for single image HDR reconstruction. ACM Trans. Multimed. Comput. Commun. Appl. 2023, 19, 1–23. [Google Scholar] [CrossRef]
  58. Nayar, S.K.; Mitsunaga, T. High dynamic range imaging: Spatially varying pixel exposures. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2000, Hilton Head Island, SC, USA, 15 June 2000; Volume 1, pp. 472–479. [Google Scholar]
  59. Cho, H.; Kim, S.J.; Lee, S. Single-shot High Dynamic Range Imaging Using Coded Electronic Shutter. In Proceedings of the Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2014; Volume 33, pp. 329–338. [Google Scholar]
  60. Gu, J.; Hitomi, Y.; Mitsunaga, T.; Nayar, S. Coded rolling shutter photography: Flexible space-time sampling. In Proceedings of the 2010 IEEE International Conference on Computational Photography (ICCP), Cambridge, MA, USA, 29–30 March 2010; pp. 1–8. [Google Scholar]
  61. Banterle, F.; Artusi, A.; Debattista, K.; Chalmers, A. Advanced High Dynamic Range Imaging; AK Peters/CRC Press: Wellesley, MA, USA, 2017. [Google Scholar]
  62. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  63. Collett, E. Field Guide to Polarization; SPIE: Bellingham, WA, USA, 2005. [Google Scholar]
  64. Ferraton, M.; Stolz, C.; Morel, O.; Meriaudeau, F. Quality control of transparent objects with polarization imaging. In Proceedings of the Eighth International Conference on Quality Control by Artificial Vision, 2007, Napoli, Italy, 29 May 2007; Volume 6356, pp. 54–61. [Google Scholar]
  65. Wolff, L.B. Polarization-based material classification from specular reflection. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 1059–1071. [Google Scholar] [CrossRef]
  66. Morel, O.; Meriaudeau, F.; Stolz, C.; Gorria, P. Polarization imaging applied to 3D reconstruction of specular metallic surfaces. In Proceedings of the Machine Vision Applications in Industrial Inspection XIII; SPIE: Napoli, Italy, 2005; Volume 5679, pp. 178–186. [Google Scholar]
  67. Pistellato, M.; Albarelli, A.; Bergamasco, F.; Torsello, A. Robust joint selection of camera orientations and feature projections over multiple views. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancún, Mexico, 4–8 December 2016; pp. 3703–3708. [Google Scholar] [CrossRef]
  68. Pistellato, M.; Bergamasco, F.; Albarelli, A.; Torsello, A. Dynamic optimal path selection for 3D Triangulation with multiple cameras. In Proceedings of the Image Analysis and Processing—ICIAP 2015: 18th International Conference, Genova, Italy, 7–11 September 2015; Volume 9279, pp. 468–479. [Google Scholar] [CrossRef] [Green Version]
  69. Zappa, C.J.; Banner, M.L.; Schultz, H.; Corrada-Emmanuel, A.; Wolff, L.B.; Yalcin, J. Retrieval of short ocean wave slope using polarimetric imaging. Meas. Sci. Technol. 2008, 19, 055503. [Google Scholar] [CrossRef]
  70. Pistellato, M.; Bergamasco, F.; Torsello, A.; Barbariol, F.; Yoo, J.; Jeong, J.Y.; Benetazzo, A. A physics-driven CNN model for real-time sea waves 3D reconstruction. Remote Sens. 2021, 13, 3780. [Google Scholar] [CrossRef]
  71. Cronin, T.W.; Marshall, J. Patterns and properties of polarized light in air and water. Philos. Trans. R. Soc. B Biol. Sci. 2011, 366, 619–626. [Google Scholar] [CrossRef] [Green Version]
  72. Pistellato, M.; Bergamasco, F.; Fatima, T.; Torsello, A. Deep Demosaicing for Polarimetric Filter Array Cameras. IEEE Trans. Image Process. 2022, 31, 2017–2026. [Google Scholar] [CrossRef] [PubMed]
  73. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  74. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  75. Mitsunaga, T.; Nayar, S.K. Radiometric self calibration. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), Fort Collins, CO, USA, 23–25 June 1999; Volume 1, pp. 374–380. [Google Scholar]
  76. Ba, Y.; Gilbert, A.; Wang, F.; Yang, J.; Chen, R.; Wang, Y.; Yan, L.; Shi, B.; Kadambi, A. Deep shape from polarization. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Part XXIV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 554–571. [Google Scholar]
  77. Reinhard, E.; Stark, M.; Shirley, P.; Ferwerda, J. Photographic tone reproduction for digital images. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 23–26 July 2002; pp. 267–276. [Google Scholar]
  78. A Sharif, S.M.; Naqvi, R.A.; Biswas, M.; Kim, S. A two-stage deep network for high dynamic range image reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 20–25 June 2021; pp. 550–559. [Google Scholar]
  79. Santos, M.S.; Ren, T.I.; Kalantari, N.K. Single image HDR reconstruction using a CNN with masked features and perceptual loss. arXiv 2020, arXiv:2005.07335. [Google Scholar] [CrossRef]
  80. Reinhard, E.; Heidrich, W.; Debevec, P.; Pattanaik, S.; Ward, G.; Myszkowski, K. High Dynamic Range Imaging: Acquisition, Display, and Image-Based Lighting; Morgan Kaufmann: San Francisco, CA, USA, 2010. [Google Scholar]
  81. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
Figure 1. Our method’s pipeline overview. The mosaic input ( I ^ ) is the image taken from the camera with the linear polarizer mounted in front ( c a m f ) and is passed to the recovery network; this network simulates the response of the normal camera by predicting the actual angle ( θ ) and the degree ( D ) of polarization of the scene. θ f and D f are the angle and degree captured by c a m f , computed from I ^ . HDR is reconstructed using polarization information and is further improved via the tone mapping network.
Figure 1. Our method’s pipeline overview. The mosaic input ( I ^ ) is the image taken from the camera with the linear polarizer mounted in front ( c a m f ) and is passed to the recovery network; this network simulates the response of the normal camera by predicting the actual angle ( θ ) and the degree ( D ) of polarization of the scene. θ f and D f are the angle and degree captured by c a m f , computed from I ^ . HDR is reconstructed using polarization information and is further improved via the tone mapping network.
Sensors 23 05370 g001
Figure 2. An outdoor high dynamic scene. The left image is the DoLP [ 0 , 1 ] and the right is the AoLP [ π 2 , π 2 ] , recorded with the PFA camera.
Figure 2. An outdoor high dynamic scene. The left image is the DoLP [ 0 , 1 ] and the right is the AoLP [ π 2 , π 2 ] , recorded with the PFA camera.
Sensors 23 05370 g002
Figure 3. Images taken with the stereo camera setup introduced in [15]. The top row refers to demosaiced polarized filter images taken with a PFA camera: all channels essentially look the same. The bottom row shows the same scene taken with the second PFA camera and external linear polarizer; it produces images similar to exposure bracketing.
Figure 3. Images taken with the stereo camera setup introduced in [15]. The top row refers to demosaiced polarized filter images taken with a PFA camera: all channels essentially look the same. The bottom row shows the same scene taken with the second PFA camera and external linear polarizer; it produces images similar to exposure bracketing.
Sensors 23 05370 g003
Figure 4. RecNet architecture: the input is a 128 × 128 mosaiced image taken by a PFA camera with an external filter. Two subsequent mosaic convolutions extract a feature map of size 128 × 128 × 16 , which is demosaiced into a 64 × 64 × 64 cube. The network then computes the degree of linear polarization D and angle of polarization θ in parallel.
Figure 4. RecNet architecture: the input is a 128 × 128 mosaiced image taken by a PFA camera with an external filter. Two subsequent mosaic convolutions extract a feature map of size 128 × 128 × 16 , which is demosaiced into a 64 × 64 × 64 cube. The network then computes the degree of linear polarization D and angle of polarization θ in parallel.
Sensors 23 05370 g004
Figure 5. Improvement network: The HDR images generated using the technique described in Section 4.2 were first converted into the LDR domain via tone mapping. The tone-mapped images were then used as input to the network, where a sequence of 2D convolutional layers extracted features and mapped them to the final output.
Figure 5. Improvement network: The HDR images generated using the technique described in Section 4.2 were first converted into the LDR domain via tone mapping. The tone-mapped images were then used as input to the network, where a sequence of 2D convolutional layers extracted features and mapped them to the final output.
Sensors 23 05370 g005
Figure 6. Outputs from RecNet. From left to right: input image with the synthesized external filter, ground truth AoLP, predicted AoLP (both in radians), ground truth DoLP, predicted DoLP. In the first row, we can see good reconstruction in terms of both the angle and degree of polarization. In the second row, the angle is not always correct, but the predicted degree is good. This does not affect the HDR outcome since, for high degrees, the corresponding angle is correct, and for lower degrees (the majority of this scene), the original angle is actually discarded.
Figure 6. Outputs from RecNet. From left to right: input image with the synthesized external filter, ground truth AoLP, predicted AoLP (both in radians), ground truth DoLP, predicted DoLP. In the first row, we can see good reconstruction in terms of both the angle and degree of polarization. In the second row, the angle is not always correct, but the predicted degree is good. This does not affect the HDR outcome since, for high degrees, the corresponding angle is correct, and for lower degrees (the majority of this scene), the original angle is actually discarded.
Sensors 23 05370 g006
Figure 7. Qualitative comparison from our test set with and without TMNet. From left to right: ground truth HDR image, output of our proposed complete pipeline, tone-mapped output without applying our TMNet.
Figure 7. Qualitative comparison from our test set with and without TMNet. From left to right: ground truth HDR image, output of our proposed complete pipeline, tone-mapped output without applying our TMNet.
Sensors 23 05370 g007
Figure 8. Qualitative results for all compared techniques against ground truth from our test data (with synthetic filter). Wu et al. refers to [3].
Figure 8. Qualitative results for all compared techniques against ground truth from our test data (with synthetic filter). Wu et al. refers to [3].
Sensors 23 05370 g008
Figure 9. Best HDR methods with details. From left to right: ground truth (GT, produced using classical multi-exposure techniques), our proposed method, the PFA-based method Wu et al. [3], and learning-based single image HDRI Two-stage HDR, respectively. For better visualization, zoomed-in patches of rows 1 and 3 are shown in rows 2 and 4, respectively.
Figure 9. Best HDR methods with details. From left to right: ground truth (GT, produced using classical multi-exposure techniques), our proposed method, the PFA-based method Wu et al. [3], and learning-based single image HDRI Two-stage HDR, respectively. For better visualization, zoomed-in patches of rows 1 and 3 are shown in rows 2 and 4, respectively.
Sensors 23 05370 g009
Figure 10. Qualitative comparison of the test set captured with a real filter outside the PFA camera lenses.
Figure 10. Qualitative comparison of the test set captured with a real filter outside the PFA camera lenses.
Sensors 23 05370 g010aSensors 23 05370 g010b
Table 1. Comparison between HDR methods with Reinhard tone mapping [80] applied. Bold font indicates the best value in each column.
Table 1. Comparison between HDR methods with Reinhard tone mapping [80] applied. Bold font indicates the best value in each column.
MethodPSNR (dB)MS-SSIM
Ours23.0670 ± 2.13620.9756 ± 0.0145
Ours w/o TMnet20.5114 ± 3.71500.9741 ± 0.0121
Wu et al. [3]18.9489 ± 4.11960.9706 ± 0.0188
KO [48]12.9069 ± 1.69900.5067 ± 0.1880
DPHDR [13]12.5211 ± 2.48260.6465 ± 0.0786
Two-stage HDR [78]19.5877 ± 3.19490.9658 ± 0.0230
Deep-HDR [79]16.6983 ± 1.86590.7840 ± 0.1401
ExpandNet [11]14.6473 ± 2.51560.8113 ± 0.0788
HDRCNN [49]14.8472 ± 2.86040.7202 ± 0.1661
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pistellato, M.; Fatima, T.; Wimmer, M. Exploiting Light Polarization for Deep HDR Imaging from a Single Exposure. Sensors 2023, 23, 5370. https://doi.org/10.3390/s23125370

AMA Style

Pistellato M, Fatima T, Wimmer M. Exploiting Light Polarization for Deep HDR Imaging from a Single Exposure. Sensors. 2023; 23(12):5370. https://doi.org/10.3390/s23125370

Chicago/Turabian Style

Pistellato, Mara, Tehreem Fatima, and Michael Wimmer. 2023. "Exploiting Light Polarization for Deep HDR Imaging from a Single Exposure" Sensors 23, no. 12: 5370. https://doi.org/10.3390/s23125370

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop