Next Article in Journal
Traditional Chinese Medicine Pulse Diagnosis on a Smartphone Using Skin Impedance at Acupoints: A Feasibility Study
Previous Article in Journal
ARES: Automated Risk Estimation in Smart Sensor Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low-Light Image Brightening via Fusing Additional Virtual Images

1
Institute of Robotics and Intelligent Systems, School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan 430081, China
2
Institute for Infocomm Research, Singapore 138632, Singapore
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(16), 4614; https://doi.org/10.3390/s20164614
Submission received: 7 August 2020 / Revised: 13 August 2020 / Accepted: 14 August 2020 / Published: 17 August 2020
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Capturing high-quality images via mobile devices in low-light or backlighting conditions is very challenging. In this paper, a new, single image brightening algorithm is proposed to enhance an image captured in low-light conditions. Two virtual images with larger exposure times are generated to increase brightness and enhance fine details of the underexposed regions. In order to reduce the brightness change, the virtual images are generated via intensity mapping functions (IMFs) which are computed using available camera response functions (CRFs). To avoid possible color distortion in the virtual image due to one-to-many mapping, a least square minimization problem is formulated to determine brightening factors for all pixels in the underexposed regions. In addition, an edge-preserving smoothing technique is adopted to avoid noise in the underexposed regions from being amplified in the virtual images. The final brightened image is obtained by fusing the original image and two virtual images via a gradient domain guided image filtering (GGIF) based multiscale exposure fusion (MEF) with properly defined weights for all the images. Experimental results show that the relative brightness and color are preserved better by the proposed algorithm. The details in bright regions are also preserved well in the final image. The proposed algorithm is expected to be useful for computational photography on smart phones.

1. Introduction

Exposure time and ISO values are usually carefully tuned by professional photographers to capture well-posed images under various lighting conditions [1,2]. A pairing of short exposure time and a small ISO value can be adopted to reduce motion blur and overexposed regions from an image captured under low-light or backlighting conditions [3]. Since ghost removal is the Achilles’ heel of high dynamic range (HDR) imaging, this method is also efficient to capture images for HDR scenes, especially in the presence of moving objects. However, the visibility of underexposed regions in the captured image is poor. It is necessary to increase the brightness of the underexposed regions in many applications such as tracking ships [4] and capturing CT images [5].
Directly brightening the low-light image is probably the most intuitive and simplest way to improve the visibility of dark regions. Unfortunately, such an intuitive way causes another problem, i.e., relatively bright regions might be saturated and thus lose corresponding details [6]. To address the problem, many single image brightening algorithms were proposed in the past decade. A lightness-order-error measure was proposed in [7] to achieve overenhancement and preserve natural illumination. Based on an observation that an inverted low-light image looks like a haze image, Dong et al. [8] designed an interesting single image brightening algorithm by dehazing the inverted low-light images. The algorithm was further improved in [9] by first segmenting the input image and then adaptively denoising different segments. Even though the above dehazing-like methods can provide visually pleasing results, the basic model lacks physical explanation. A low-lighting image enhancement (LIME) was proposed in [10] to address such a concern by extending the concept of Max-RGB [11] to the pixel level. Particularly, the illumination of each pixel is first estimated individually by finding the maximum value in R, G, and B channels. Then, a structure prior is imposed on initial illumination as the final illumination map. The enhanced image is finally obtained by using the well-constructed illumination map. The framework on the illumination map estimation [10] is similar to the transmission map estimation in [12]. A weighted variational model was proposed for simultaneous reflectance and illumination estimation (SRIE) in [13]. With the estimated reflectance and illumination, the original image can be enhanced by manipulating the illumination. Two virtual images were produced to directly brighten different regions of the input image. All the virtual images and the input image were fused via a multiscale exposure fusion (MEF) algorithm to produce the desired image in [3,14].
Both the LIME- [10] and the MEF-based algorithms [3,14] assume that the brightness relationships of all the pixels at the corresponding position between the two differently exposed images are linear, i.e., the intensity mapping functions (IMFs) between the differently exposed images are linear functions. This is not always true. As shown in Figure 1b, brightness change is introduced in the red block of the enhanced image if this is not true [15]. To reduce the brightness change, a novel enhancement framework was proposed in [15] by estimating response characteristics of cameras. A reasonable camera response model and its parameters are first determined. Illumination estimation techniques are then adopted to estimate the exposure ratio for each pixel. Finally, the selected camera response model is used to adjust each pixel to the desired exposure according to the estimated exposure ratio map. On the other hand, it is very difficult to estimate the camera response functions (CRFs) using a single input image. Moreover, it is very challenging to use an IMF-based algorithm to brighten a dark image. One possible issue for the IMF-based brightening algorithm is color distortion in the brightened image, as illustrated in Figure 1c. This is because the true relationship of intensity from a dark image to a bright one is a one-to-many mapping for all underexposed pixels [16]. As such, the color distortion could appear in virtual images, especially in underexposed regions of the input image. Furthermore, bright regions could become overexposed when an image is brightened. This is also an issue for single image brightening. This issue is ignored by most deep-learning-based single image brightening algorithms. Another issue is the amplification of noise in the underexposed regions, as shown in Figure 1. The BM3D (Block-Matching and 3D filtering) [17] was used in the LIME as a postprocessing procedure to reduce the noise in the brightened image. As noticed in [15], such a postprocessing method will reduce the efficiency of the image brightening algorithm, especially when it is embedded in mobile devices with limited computational resource. A maximum exposure value is set for pixels with very low illumination in [15] to avoid amplifying noise in the underexposed regions. However, it will lead to the decline of visibility in the very dark regions by setting the maximum exposure ratio. There are also date-driven algorithms that intend to deal with these problems, such as [18,19]. In summary, it is a challenge to keep naturalness while improving the visibility of dark regions in the existing methods. Hence, it is desired to design a simple, single image brightening algorithm which can avoid brightness change, overexposure, and color distortion as well as noise amplification from appearing in the brightened image.
In this paper, a new single image brightening algorithm is introduced to achieve the above objective, and the proposed algorithm is supposed to be embedded in mobile devices such as digital cameras and smartphones. Unlike the algorithm in [15], it is assumed that the CRFs are available. This is not an issue because a set of differently exposed images can be captured in advance to estimate the CRFs by using the algorithm in [20]. To enhance dark regions while preserving the details in bright regions, the enhanced image is obtained from two virtual images and one dark image by using multiscale exposure fusion. Instead of multiplying the dark image by constant ratios, as in [3,14], the virtual images are generated by using the IMFs. Different from [21], the virtual images are generated from a single low-light image, and then the IMFs between the input image and the virtual images are computed from the CRFs. As such, the brightness change is prevented from appearing in the virtual images. The IMFs are used directly to brighten pixels that are not underexposed due to reliability of the IMFs for these pixels. However, color distortions will be brought if the IMFs are directly used to brighten a pixel with one underexposed color channel. To handle this problem, a linear method will be provided to brighten the pixel with at least one underexposed color channel. In other words, the brightened pixel is obtained by multiplying the pixel by a constant ratio. The constant ratio is determined by a least square minimization problem. It is worth noting that the constant ratio is more determined by reliable pixels in order to avoid the effects of underexposed pixels. Besides the color distortion, it is also necessary to avoid amplifying noise in the underexposed regions. A simple edge-preserving denoising method is provided to address the problem. The input image is decomposed into a base layer and a detail layer by using a weighted guided image filter (WGIF), and only the base layer is brightened [16].
All the input image and two virtual images are fused via a state-of-the-art multiscale exposure fusion algorithm, as in [22], to produce the final image. Different pixels usually have different constant ratios. Besides the measurements of contrast, exposure level, and saturation in [23], one more new weighting function is introduced in the proposed algorithm. With the new weighting function, larger weights are given to all the pixels in bright regions of the input image so as to avoid them from being overexposed in the brightened image. Experimental results show that the proposed algorithm outperforms existing single image brightening algorithms. Due to the simplicity of the proposed algorithm, it can indeed be embedded into mobile devices such as an app in smart phones. In summary, three major contributions of this paper are as follows:
  • Propose a simple method to brighten a low-light or backlighting image by fusing additional virtual images;
  • Reduce the possible color distortion of the generated virtual image due to inaccurate IMF;
  • Present a simple method to alleviate the noise in virtual image.
The rest of this paper is organized as follows. Two virtual images are generated in Section 2. All the input image and virtual images are fused together to produce the final image via a multiscale exposure fusion algorithm, described in Section 3. Experimental results are provided in Section 4 to illustrate the efficiency of the proposed algorithm. Finally, the paper is concluded in Section 5.

2. Generation of Two Virtual Images

In this section, two virtual images are generated and both of them are brighter than the input image. To avoid the brightness change from appearing in the brightened images, the virtual images are generated by using the IMFs that are computed from the available CRFs as follows:
The IMFs between input image and virtual images are computed via CRFs. The CRFs are generated using the algorithm in [20] by using a large number of differently exposed images.
Let the CRFs f l ( · ) ( l = 1 , 2 , 3 ) , Z 2 , and Z 3 be two brightened images from the input Z 1 . Their exposure times are Δ t 1 , Δ t 2 , and Δ t 3 , respectively. Without loss of generality, Δ t 3 is equal to 4 Δ t 2 and Δ t 2 is equal to 4 Δ t 1 . As shown in [21], there is brightness order reversal if the ratios among exposure times are too large. The IMFs between input image and two virtual images can be expressed as follows:
Λ 1 , i , l ( z ) = f l ( f l 1 ( z ) Δ t i Δ t 1 ) ,
where f l 1 ( · ) is the inverse function of the CRF f l ( · ) .

2.1. Generation of Two Intermediate Brightened Images via IMFs

Let the input image be denoted as Z 1 and [ Z 1 , 1 ( p ) , Z 1 , 2 ( p ) , Z 1 , 3 ( p ) ] be a pixel for three color channels. Z 2 and Z 3 are two virtual images which are brightened from the input image Z 1 . Let the IMF from the c-th color channel of image Z 1 to the c-th color channel of image Z i be denoted as Λ 1 , i , c ( z ) . It can be easily verified that the function Λ 1 , i , c ( z ) has the following feature:
The function Λ 1 , i , c ( z ) is a one-to-many mapping when the value of z is smaller than a threshold ξ L and a one-to-one mapping if the value of z is above the threshold [16]. Here, the value of ξ L depends on the quality of the camera. The value of ξ L is small for a high-quality camera.
In other words, the IMF from a dark image to a bright one is not reliable if the pixel value is less than the threshold ξ L . It is thus necessary to consider the following two cases:
Case 1: None of Z 1 , c ( p ) ( 1 c 3 ) is less than ξ L . The two corresponding brightened pixels are [ Λ 1 , 2 , 1 ( ( Z 1 , 1 ( p ) ) , Λ 1 , 2 , 2 ( Z 1 , 2 ( p ) ) , Λ 1 , 2 , 3 ( Z 1 , 3 ( p ) ) ] T and [ Λ 1 , 3 , 1 ( ( Z 1 , 1 ( p ) ) , Λ 1 , 3 , 2 ( Z 1 , 2 ( p ) ) , Λ 1 , 3 , 3 ( Z 1 , 3 ( p ) ) ] T , respectively.
Case 2: Some of Z 1 , c ( p ) ( 1 c 3 ) is (are) less than ξ L . The IMFs are not reliable. The color distortion will be brought to the two virtual images if the IMFs are directly used to generate the brightened pixels as in the Case 1. Noise will also be amplified. A piecewise constant method will be first provided to brighten pixels in the underexposed regions of the input image, i.e., when the IMFs are not reliable. An edge-preserving decomposition method will then be provided to avoid noise from being amplified in the brightened images.

2.2. Brightening Pixels in Underexposed Regions

A brightening algorithm under the assumption of the linear IMFs generates a brightened pixel by multiplying a pixel to be brightened by a constant. Such a constant brightening algorithm can avoid the color distortion while it causes the brightness change. On the other hand, a brightening algorithm based on the general IMFs can avoid the brightness change while it causes the color distortion for pixels in underexposed regions of the input image. In this subsection, the advantages of both the algorithms are combined to design an algorithm to brighten the pixels in the underexposed region.
Let pixel Z 1 ( p ) be a pixel with a channel value less than ξ L . The pixel Z 1 ( p ) will be brightened as γ 1 Z 1 ( p ) and γ 2 Z 1 ( p ) , respectively.
The values of γ i ( i = 1 , 2 ) are obtained by solving the following optimization problems:
arg min γ i { c = 1 3 w ˜ ( Z 1 , c ( p ) ) ( Λ 1 , i + 1 , c ( Z 1 , c ( p ) ) γ i Z 1 , c ( p ) ) 2 } ,
where the function w ˜ ( z ) is defined according to different reliability for different pixels. The lower the pixel value, the higher the probability of being affected by noise, which is more likely to bring color distortion in the enhanced image. Therefore, the confidence interval of the pixel with a low value is smaller than the confidence interval of the pixel with a high value. It is defined as
w ˜ ( z ) = 1 ; if 0 z < ξ L 128 381 h 2 ( z ) + 254 h 3 ( z ) ; if ξ L z < ξ U 128 ; otherwise ,
where ξ U and ξ L are positive constants to determine the reliability of pixel. Their values are chosen by the quality of fused image. The function h ( z ) is defined as
h ( z ) = ξ U z ξ U ξ L .
Consider the pixel [ Z 1 , 1 ( p ) , Z 1 , 2 ( p ) , Z 1 , 3 ( p ) ] T in the input image. It can be shown from Equation (2) that the brightness of the corresponding brightened pixel in the first virtual image is determined by [ Λ 1 , 2 , 1 ( ( Z 1 , 1 ( p ) ) , Λ 1 , 2 , 2 ( Z 1 , 2 ( p ) ) , Λ 1 , 2 , 3 ( Z 1 , 3 ( p ) ) ] T and in the second virtual image by [ Λ 1 , 3 , 1 ( ( Z 1 , 1 ( p ) ) , Λ 1 , 3 , 2 ( Z 1 , 2 ( p ) ) , Λ 1 , 3 , 3 ( Z 1 , 3 ( p ) ) ] T . In other words, the brightness is determined by the general IMF-based brightening method. As such, the brightness change can be avoided. To further avoid the color distortion, the brightened pixels are generated by using the linear IMF-based method, i.e., the corresponding brightened pixels are [ γ 1 Z 1 , 1 ( p ) , γ 1 Z 1 , 2 ( p ) , γ 1 Z 1 , 3 ( p ) ] T and [ γ 2 Z 1 , 1 ( p ) , γ 2 Z 1 , 2 ( p ) , γ 2 Z 1 , 3 ( p ) ] T , respectively. The resultant brightening algorithm is a piecewise-constant brightening algorithm.
Equation (2) is a classic least squares problem. The values of γ i ( i = 1 , 2 ) can be computed through the method of finding the extremum of partial differential equation. Consequently, the γ i ( i = 1 , 2 ) are computed as
γ i = l = 1 3 w ˜ ( Z 1 , c ( p ) ) Z 1 , c ( p ) Λ 1 , i + 1 , c ( Z 1 , c ( p ) ) l = 1 3 w ˜ ( Z 1 , c ( p ) ) Z 1 , c 2 ( p ) ,
and the brightened pixels are given as
Z i + 1 ( p ) = γ i Z 1 ( p ) ( i = 1 , 2 ) .
Clearly, the values of γ i are usually different for different pixels. Therefore, the pixels in underexposed regions are brightened using a different method from that in [3,14]. As a result, both the brightness change and the color distortion are being avoided from appearing in the two brightened images.

2.3. Noise Reduction of Brightened Images

A dark pixel includes noise. If the pixel Z 1 ( p ) is directly brightened as γ ˜ i Z 1 ( p ) ( i = 1 , 2 ) , the noise is amplified in the brightened pixels. In this subsection, a simple method is provided to address the problem. To reduce noise while preserving details, the pixel Z 1 ( p ) is decomposed as
Z 1 ( p ) = Z 1 , b ( p ) + Z 1 , d ( p )
by using an edge-preserving smoothing filter, such as the WGIF in [24]. Here, Z 1 , b ( p ) and Z 1 , d ( p ) are the base layer and the detail layer of pixel Z 1 ( p ) , respectively.
The values of γ ˜ i ( i = 1 , 2 ) are obtained by minimizing the following cost function:
c = 1 3 w ˜ ( Z 1 , c ( p ) ) ( Λ 1 , i + 1 , c ( Z 1 , c ( p ) ) Z 1 , c , e ( p ) γ ˜ i Z 1 , c , b ( p ) ) 2 .
Similar to Equation (2), the values of γ ˜ i ( i = 1 , 2 ) are computed as
γ ˜ i = c = 1 3 w ˜ ( Z 1 , c ( p ) ) Z 1 , c , b ( p ) ( Λ 1 , i + 1 , c ( Z 1 , c ( p ) ) Z 1 , c , d ( p ) ) c = 1 3 w ˜ ( Z 1 , c ( p ) ) Z 1 , c , b 2 ( p ) ,
and the brightened pixels are then given as
Z i + 1 ( p ) = γ ˜ i Z 1 , b ( p ) + Z 1 , d ( p ) ( i = 1 , 2 ) .
Since the noise is almost included in the detail layer, the noise is avoided from being amplified by the brightened method (10).

3. Fusion of the Input Image and Two Virtual Images

3.1. Weights of Three Images

The input image and two virtual images look like three differently exposed images, but they are different from a set of differently exposed images which is captured by a digital camera. Therefore, a different method is adopted to define weights for the three images.
As pointed out in the introduction, a possible issue for the single image brightening is that relatively bright regions might be overexposed in the final image because of the brightening. Clearly, this can be avoided if the details in the relatively bright regions in the final image are from the corresponding regions in the input image. To achieve this objective, two different functions are adopted to define the weights for the three images.
One function is used to determine amplification factors of all the pixels in the first image. To brighten the dark regions while preserving the details in the bright regions, the bright regions are assigned high weights in the low-light image. The weight is defined as
w 1 ( z ) = 2 ; if z > η U 1 + h 3 2 ( z ) ( 3 2 h 3 ( z ) ) ; if η U z > η L 1 ; otherwise ,
where the function h 3 ( z ) is defined as
h 3 ( z ) = z η L η U η L ,
and η L and η U are two constants.
The other is to measure contrast, exposure level, and color saturation of all the pixels in the three images and it is defined the same as in [23]:
w 2 ( z i ( p ) ) = w c ( z i ( p ) ) × w s ( z i ( p ) ) × w e ( z i ( p ) ) ,
where w c ( Z i ( p ) ) , w s ( Z i ( p ) ) and w e ( Z i ( p ) ) measure contrast, color saturation, and exposure level of pixel Z i ( p ) , respectively. Particularly, let Y i be the luminance component of the image Z i ( i = 1 , 2 , 3 ) . The weight of the pixel Z 1 ( p ) is given as
W ( Z 1 ( p ) ) = w 1 ( Y 1 ( p ) ) w 2 ( Z 1 ( p ) ) ,
and the weight of the pixel Z i ( p ) ( i = 2 , 3 ) is given as
W ( Z i ( p ) ) = w 2 ( Z i ( p ) ) .
It can be shown in Equations (14) and (15) that the weight of a bright pixel in the input image is much larger than a bright pixel in the two virtual images. As such, the details in the relatively bright regions in the final image are usually from the corresponding regions in the input image. The relatively bright region in the input image is avoided from being overexposed in the final image.

3.2. Fusion of Three Images via a GGIF-Based MEF

To get the good fused image, in this subsection, the three images are fused via a gradient domain guided image filtering (GGIF) based on multiexposure images fusion (MEF) [22]. The weight in each level of pyramid is smoothed by the GGIF. To reduce the noise, the guided image is luminance component of input image.
Let W i ( l ) ( i = 1 , 2 , 3 ) be weight in l-level Gaussian pyramid and GGIF be G G . The weight in l-level is computed as
W ( Z i ( l ) ) = G G { Y i ( l ) , W i ( l ) } ,
where Y i ( l ) is the l-level Gaussian pyramid of the luminance component of the image Z i . The l-level fused image is presented as
Z f u s e d ( l ) = i = 1 3 W ( Z i ( l ) ) Z i ( l ) i = 1 3 W ( Z i ( l ) ) .
The final fused image can be reconstructed from Z i ( l ) . Overall, the proposed single image brightening algorithm is summarized as Algorithm 1.
 Algorithm 1: Single image brightening and the corresponding CRFs
 Input: a low-light image
 Output: a brightened image
 Step 1 Compute IMFs from the available CRFs
 Step 2 Generate two virtual images with larger exposure times
  Case1 the value of pixel above 5 do
    Brighten the pixel via Equation (17)
  end
  Case2 the value of pixel below 5 do
   1. Constructing weight matrix by Equation (2)
   2. Decompose input image as base layer and detail layer using WGIF
   3. Compute brighten constant γ i using Equation (8)
   4. Brighten the pixel using Equation (9)
  end
 Step3 Fuse the input image and two virtual images to produce the
         final image using Equations (10)–(16).

4. Experimental Results

In this section, extensive experimental results are provided to verify the performance of the proposed algorithm. All images are captured by us using a Nikon D7200. It should be noted that the differently exposed images are also captured at the same time.

4.1. Generation of Virtual Images

In this subsection, the effect of exposure ratios of two virtual images on the quality of enhanced image is analyzed. For the same scene, four groups of enhanced images are generated by fusing virtual images with different exposure ratios. In addition, the quality of different exposure images is also discussed. The ground truths of enhanced image are nine real exposure images and the ground truth of virtual image is the real image with the same exposure time as virtual image. The SSIM (Structural Similarity) index maps are shown in Figure 2. It demonstrates that the higher the exposure ratio of the virtual image to the dark image, the lower the quality of the virtual image is. However, it is found from Figure 3 that the quality of the image quality improved as the virtual image exposure ratios increased. This is because the virtual image with a small exposure ratio is unable to enhance the details in underexposed regions. In the rest of this section, two virtual images are generated with exposure ratios of 4 and 16.

4.2. Difference Choices of Parameters

In this subsection, the suitable parameters are selected by a well-used image quality, i.e., the MEF-SSIM in [25]. The test images are three images of different static scenes captured by ourselves. It is worth noting that the exposure ratio of the input image and two virtual images is 1:4:16. Here, the two images with the same exposure times as those of virtual images are also captured for each input image.
When analyzing the influence of ξ U and ξ L , the MEF-SSIM is adopted to assess the quality of the virtual images with the largest exposure time. Specially, the reference image is a real image with the same of exposure time as the virtual image. The 14 pairs of ( ξ U , ξ L ) and the corresponding MEF-SSIM are shown in Figure 4. Due to the high quality of the camera, the MEF-SSIM of three images decline when ξ L is below 5, as demonstrated in Figure 4. During the determination of ξ L , the value of ξ U has little effect on the MEF-SSIM. In this paper, the parameters are set as ξ L = 5 , ξ U = 60 .
Then, the influence of different pairs of ( η U , η L ) is studied on the fused image. Specially, the reference images are three real, different exposure images with the same exposure times as the three images for fusion. The results of MEF-SSIM are shown in Figure 4. When the values of η L and η U fluctuate by a small amount, the result will not be affected too much, as shown in Figure 4. In this paper, their values are selected as η L = 128 , η U = 160 .

4.3. Comparison of the Proposed Algorithm with Existing Ones

In this subsection, we compare our algorithm with seven state-of-the-art image enhancement algorithms, including two data-driven algorithms [18,19] and five model-driven algorithms—LIME [10], NPE [7], LECARM [15], SNIE [13], and Dong [8]. All the test images are taken by the Nikon D7200.
Firstly, the possible color distortion of eight algorithms is analyzed. Since a different light source may introduce luminance change that can affect the overall fluctuation of RGB values in the enhanced image, instead of calculating the absolute difference of RGB value, the angular color between two color vectors respectively produced by enhanced image and ground truth is computed as in [26,27]. As the wavelengths of purple light and blue light are short, the color of image will be biased toward yellow or even red if the exposure time is small. So, the values of pixels on the color checker board are mostly around 128 in the ground truth. Without loss of generality, there is a color checker board in every captured image, and a different color of this board can be regarded as ground truth. The color distortion is expressed as
ε n = cos 1 ( Z n · Z n g t | | Z n | | · | | Z n g t | | ) ,
where Z n is the RGB value regarded as the one of the color vectors generated by one enhanced image, and Z n g t is the corresponding color vector from the color checker board. It is worth noting that the color vector is averaged around a 3 × 3 block of the pixel.
The 24 color errors of eight algorithms are shown in Figure 5. From the box plots, our median value is the smallest and the maximum value and the minimum value are also the smallest. This means that the 24 errors of our enhanced image follow a narrow distribution and the average is the smallest. It further implies that the proposed algorithm outperforms other seven algorithms in keeping color fidelity. The comparison of the enhanced color checker board is also shown in Figure 6. The figure indicates that the Dong [8] and NPE [7,19] suffer from serious color distortion in multiple color blocks. The LIME [10] makes the image bright enough, but the enhanced image looks too sharp. The LECARM [15] and SNIE [13,18] try reducing distortion at the expense of brightness, however, the first two find it difficult to distinguish the colors of the first three blocks of the last line. By contrast, our algorithm can effectively maintain the authenticity of colors.
Next, the contrast distortion is analyzed in terms of PCQI (Patch-Based Contrast Quality Index) [28], which can predict the overall and local contrast quality of enhanced images. The ground truth is generated by fusing nine different exposure images of the same scene by Kou algorithm [22]. These nine real, different exposure images are taken with a camera from low-exposure to high-exposure, one EV value for each image interval. The assessments of each enhanced algorithm are shown in Table 1. Our algorithm first ranks by contrast quality.
The MEF-SSIM is adopted to further evaluate the performance of eight algorithms. Due to the lack of ground truth, the reference images for the MEF-SSIM are nine real, different exposure images. The nine images contain more information than one dark image, so it is fairer to use this evaluation strategy. The assessment is shown in Table 2. Clearly, our algorithm outperforms all the seven state-of-art algorithms on average.
Besides the objective evaluation, partial enlargements of brighten images are presented in Figure 7 and the large images are also given in Figure 8. Clearly, all eight algorithms can indeed brighten low-light images. However, it can be shown from the second row of Figure 7 that the enhanced image by the LIME [10] looks too sharp. It is demonstrated in the first row of Figure 7 that the LECARM [15] and Dong [8] produce ring artifacts in the enhanced images. Furthermore, there are both color and lightness distortions in the enhanced images by the Dong [8], SNIE [13], and NPE [7,19], as illustrated in the last row of Figure 7. In Figure 8, the enhanced images by the LIME [10] are too sharp and there is color distortion in the first column. There is slight color distortion in the enhanced image by the SINE [13] and the brightness of the enhanced images needs to be improved. All the problems are overcome by the LECARM [15] and the proposed algorithm. Since the proposed algorithm is on top of the CRF, the color is preserved better. This implies that the CRF can be applied to improve the quality of enhanced images.

4.4. Efficiency of Noise Reduction via WGIF-Based Smoothing Technique

In this subsection, efficiency of the noise reduction is evaluated. When the pixel value is less than 5, the constant amplification will lead to noise amplification, as shown in Figure 9.
To reduce the amplified noise and preserve the detail, the base layer of the input image is amplified by Formula (9) while the detail layer is not amplified. As shown in the enlarged yellow block, the noise is indeed reduced in the dark regions, such as the black regions on the panda doll and the shadow of the branches. Moreover, the details of the branches and panda doll have not been weakened. It is worth noting that the BM3D was used in the LIME [10] to reduce the effect of noise. The complexity of the BM3D could be an issue. A maximum exposure ratio was defined in [15] to address the effect of noise. However, setting the maximum exposure ratio will lead to the decline of visibility in the very dark regions. It can even reduce visibility in other regions if it is not defined properly.
It should be pointed out that the WGIF-based smoothing could be omitted if a stack of images is captured. Frame averaging could be used to reduce noise in the dark regions by properly detecting moving regions. If all the corresponding raw images are available, it is not difficult to detect the moving regions.

4.5. Comparison of Running Time

The running time of these six algorithms is also compared by testing three image sizes on an AMD 2600X CPU. The results are presented as a bar graph, shown in Figure 10. It should be noted that the two data-driven algorithms are not compared because of the limitations of our hardware. It can be seen in the bar graph that the Dong is the fastest, and our algorithm ranks fourth. Compared with the LECARM and LIME, the running time of our algorithm decreases rapidly with the decrease of image size because the MEF in [22] calculates weights in every image scale. The percentages on the running time of different components in our algorithm is shown in the right. It can be found that the MEF in [22] takes up most of the time. Luckily, the MEF in [22] can be replaced by the simpler MEF in [14] with negligible loss of the MEF-SSIM.

4.6. Limitation of the Proposed Algorithm

The proposed algorithm outperforms the other five algorithms. However, the proposed algorithm assumes that the accurate CRFs are available. This is not an issue if the proposed algorithm is embedded in a model device. However, this assumption is not true for those images that are downloaded from the Internet. It is very difficult to estimate the accurate CRFs if there is only one image. If the CRFs are not accurate, the IMF between the input image and virtual images are not accurate too. In other words, the brightened image could be a little poor, such as the color distortion in Figure 11. This problem will be studied in our future research.

5. Conclusions and Future Remarks

In this paper, we proposed an effective single image brightening algorithm. The proposed algorithm is supposed to be embedded in mobile devices such as digital cameras and smart phones. Two virtually brightened images are generated via intensity mapping functions (IMFs) which are computed by the camera response functions (CRFs). The input image and two virtual images are fused to preserve the natural structure. Since the three images are different from the three differently exposed images that are captured by a camera directly, the weights of these images are specially designed so as to preserve details in bright regions of the input image in the final image. Experimental results show that the proposed algorithm outperforms existing algorithms. The proposed algorithm can be embedded in mobile devices such as digital cameras and smart phones as an app.
It should be pointed out that the CRFs are assumed to be available for the proposed algorithm. Unfortunately, this is not always true. As shown in [29,30], a learning-based method could be adopted to address this issue. The proposed algorithm also assumes that details in bright regions are captured well in the input image. It is also important to enhance details in the bright regions if this assumption is not true. The concept of local inverse tone mapping in [31] could be borrowed to study this problem, and the details could also be preserved using the idea in [32]. The proposed algorithm can be improved by using a data-driven approach. All these problems will be investigated in our future research.

Author Contributions

Methodology: Y.Y., Z.L. and S.W.; Data Curation and Coding:Y.Y.; Data analysis: Y.Y., Z.L. and S.W.; Manuscript finalization: Y.Y., Z.L. and S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [the National Natural Science Foundation of China under Grants] grant number [61775172] and [61371190].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, L.; Deshpande, A.; Chen, X. Denoising vs. deblurring: HDR imaging techniques using moving cameras. In Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 522–529. [Google Scholar]
  2. Hasinoff, S.W.; Durand, F.; Freeman, W.T. Noise-optimal capture for high dynamic range photography. In Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 553–560. [Google Scholar]
  3. Li, Z.G.; Zheng, J.H. Single image brightening via exposure fusion. In Proceedings of the 2016 International Conference on Acoustics, Speech, and Signal Processing, Shanghai, China, 20–25 March 2016. [Google Scholar]
  4. Chen, X.; Wang, S.; Shi, C.; Wu, H.; Zhao, J.; Fu, J. Robust Ship Tracking via Multi-view Learning and Sparse Representation. J. Navig. 2019, 72, 176–192. [Google Scholar] [CrossRef]
  5. Wei, W.; Zhou, B.; Połap, D.; Woźniak, M. A regional adaptive variational PDE model for computed tomography image reconstruction. Pattern Recognit. 2019, 92, 64–81. [Google Scholar] [CrossRef]
  6. Celik, T.; Tjahjadi, T. Contextual and variational contrast enhancement. IEEE Trans. Image Process. 2011, 20, 3431–3441. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Wang, S.; Zheng, J.; Hu, H.; Li, B. Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 2013, 22, 3538–3578. [Google Scholar] [CrossRef] [PubMed]
  8. Dong, X.; Wang, G.; Pang, Y.; Li, W.; Wen, J.; Meng, W.; Lu, Y. Fast efficient algorithm for enhancement of low lighting video. In Proceedings of the IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011; pp. 1–6. [Google Scholar]
  9. Li, L.; Wang, R.; Wang, W.; Gao, W. A low-light image enhancement method for both denoising and contrast enlarging. In Proceedings of the IEEE Conference on Image Processing, Quebec, Canada, 27–30 September 2015; pp. 3730–3734. [Google Scholar]
  10. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2017, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
  11. Land, E.H. The Retinex theory of color vision. Sci. Am. 1997, 237, 108–128. [Google Scholar] [CrossRef] [PubMed]
  12. Li, Z.G.; Zheng, J.H. Edge-preserving decomposition-based single image haze removal. IEEE Trans. Image Process. 2015, 24, 5432–5441. [Google Scholar] [CrossRef] [PubMed]
  13. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2782–2790. [Google Scholar]
  14. Li, Z.G.; Wei, Z.; Wen, C.Y.; Zheng, J.H. Detail-enhanced multi-scale exposure fusion. IEEE Trans. Image Process. 2017, 26, 1243–1252. [Google Scholar] [CrossRef] [PubMed]
  15. Ren, Y.; Ying, Z.; Li, H.; Zheng, G.L. LECARM: Low-light image enhancement using camera response model. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 968–981. [Google Scholar] [CrossRef]
  16. Li, Z.G.; Zheng, J.H.; Zhu, Z.J.; Wu, S.Q. Selectively detail enhanced fusion of differently exposed images with moving objects. IEEE Trans. Image Process. 2014, 23, 4372–4382. [Google Scholar] [CrossRef] [PubMed]
  17. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, R.; Qing, Z.; Fu, C.; Shen, X.; Zheng, W.; Jia, J. Underexposed photo enhancement using deep illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 6849–6857. [Google Scholar]
  19. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. In Proceedings of the British Machine Vision Conference, Newcastle, UK, 3–6 September 2018; p. 1. [Google Scholar]
  20. Debevec, P.E.; Malik, J. Recovering high dynamic range radiance maps from photographs. In Proceedings of the ACM SIGGRAPH, Los Angeles, CA, USA, 3–8 August 1997; pp. 369–378. [Google Scholar]
  21. Yang, Y.; Cao, W.; Wu, S.Q.; Li, Z.G. Multi-scale fusion of two large-exposure-ratio images. IEEE Trans. Image Process. 2018, 25, 1885–1889. [Google Scholar] [CrossRef]
  22. Kou, F.; Li, Z.G.; Wen, C.Y.; Chen, W.H. Multi-scale exposure fusion via gradient domain guided image filtering. In Proceedings of the IEEE International Conference on Multimedia and Expo, Hong Kong, China, 10–14 July 2017; pp. 1105–1110. [Google Scholar]
  23. Mertens, T.; Kautz, J.; Reeth, F.V. Exposure fusion. In Proceedings of the Conference on Computer Graphics and Applications, Budmerice Castle, Slovakia, 26–28 April 2007; pp. 382–390. [Google Scholar]
  24. Li, Z.G.; Zheng, J.H.; Zhu, Z.J.; Yao, W.; Wu, S. Weighted guided image filtering. IEEE Trans. Image Process. 2014, 24, 120–129. [Google Scholar]
  25. Ma, K.D.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  26. Karaimer, H.C.; Brown, M.S. Improving color reproduction accuracy on cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6440–6449. [Google Scholar]
  27. Schiller, F.; Valsecchi, M.; Gegenfurtner, R.K. An evaluation of different measures of color saturation. Vis. Res. 2018, 151, 117–134. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, S.; Ma, K.; Yeganeh, H.; Wang, Z.; Lin, W.S. A patch-structure representation method for quality assessment of contrast changed images. IEEE Signal Process. Lett. 2015, 22, 2387–2390. [Google Scholar] [CrossRef]
  29. Cai, J.R.; Gu, S.H.; Zhang, L. Learning a Deep Single Image Contrast Enhancer from Multi-Exposure Images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef] [PubMed]
  30. Chen, C.; Chen, Q.F.; Xu, J.; Koltun, V. Learning to See in the Dark. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3291–3300. [Google Scholar]
  31. Wei, Z.; Wen, C.Y.; Li, Z.G. Local inverse tone mapping for scalable high dynamic range image coding. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 550–555. [Google Scholar] [CrossRef]
  32. Li, Z.G.; Zheng, J.H. Visual-salience-based tone mapping for high dynamic range images. IEEE Trans. Ind. Electron. 2014, 61, 7076–7082. [Google Scholar] [CrossRef]
Figure 1. The low-light image and enhanced images are some of technical books about Java. (a) is a low-exposed image; (b) is obtained by low-lighting image enhancement (LIME) [10]; (c) is obtained by the inaccurate intensity mapping function (IMF). Clearly, brightness change is possible issues for the former. Color distortion and noise amplification are possible issues for the latter.
Figure 1. The low-light image and enhanced images are some of technical books about Java. (a) is a low-exposed image; (b) is obtained by low-lighting image enhancement (LIME) [10]; (c) is obtained by the inaccurate intensity mapping function (IMF). Clearly, brightness change is possible issues for the former. Color distortion and noise amplification are possible issues for the latter.
Sensors 20 04614 g001
Figure 2. The SSIM (Structural Similarity) index maps of virtual images of different exposure times. The exposure ratios are 2, 4, 8, and 16 from right to left. The SSIMs are in the lower left corner of the images.
Figure 2. The SSIM (Structural Similarity) index maps of virtual images of different exposure times. The exposure ratios are 2, 4, 8, and 16 from right to left. The SSIMs are in the lower left corner of the images.
Sensors 20 04614 g002
Figure 3. The MEF-SSIMs of enhanced images under different virtual images of different exposure ratios.
Figure 3. The MEF-SSIMs of enhanced images under different virtual images of different exposure ratios.
Sensors 20 04614 g003
Figure 4. The MEF-SSIM values of the three images under different ξ L , ξ U , η L , and η U parameters. When ξ L is greater than 5, the SSIM value drops due to the high quality of the camera, as shown on the left. The indexes of different η L and η U parameters in terms of MEF-SSIM are shown on the right.
Figure 4. The MEF-SSIM values of the three images under different ξ L , ξ U , η L , and η U parameters. When ξ L is greater than 5, the SSIM value drops due to the high quality of the camera, as shown on the left. The indexes of different η L and η U parameters in terms of MEF-SSIM are shown on the right.
Sensors 20 04614 g004
Figure 5. The left box-plot shows the errors of 24 colors from eight algorithms, each error is displayed on the right side in order from left to right, top to bottom.
Figure 5. The left box-plot shows the errors of 24 colors from eight algorithms, each error is displayed on the right side in order from left to right, top to bottom.
Sensors 20 04614 g005
Figure 6. Brighten the color checker board. (a) The ground truth of color checker board, (b) is by using NPE [7], (c) is by using SNIE [13], (d) is by using LIME [10], (e) is by using Dong [8], (f) is by using LECARM [15], (g) is by using [18], (h) is by using [19], and (i) is by using our algorithms.
Figure 6. Brighten the color checker board. (a) The ground truth of color checker board, (b) is by using NPE [7], (c) is by using SNIE [13], (d) is by using LIME [10], (e) is by using Dong [8], (f) is by using LECARM [15], (g) is by using [18], (h) is by using [19], and (i) is by using our algorithms.
Sensors 20 04614 g006
Figure 7. The enhanced results (the images of first and second rows of images are technical books about Java) of eight algorithms. (a) shows the input images, (b) shows results by using NPE [7], (c) shows results by using SNIE [13], (d) shows results by using LIME [10], (e) shows results by using Dong [8], (f) shows results by using LECARM [15], (g) is by using [18], (h) is by using [19], and (i) is by using our algorithms.
Figure 7. The enhanced results (the images of first and second rows of images are technical books about Java) of eight algorithms. (a) shows the input images, (b) shows results by using NPE [7], (c) shows results by using SNIE [13], (d) shows results by using LIME [10], (e) shows results by using Dong [8], (f) shows results by using LECARM [15], (g) is by using [18], (h) is by using [19], and (i) is by using our algorithms.
Sensors 20 04614 g007
Figure 8. The brightened night images of eight algorithms. The first row shows the input images, the second row is performed using NPE [7], the third row using SNIE [13], the fourth row using LIME [10], the fifth row using Dong [8], the sixth row using LECARM [15], the seventh row using [18], the eighth row using [19], and the last row is by our algorithm.
Figure 8. The brightened night images of eight algorithms. The first row shows the input images, the second row is performed using NPE [7], the third row using SNIE [13], the fourth row using LIME [10], the fifth row using Dong [8], the sixth row using LECARM [15], the seventh row using [18], the eighth row using [19], and the last row is by our algorithm.
Sensors 20 04614 g008
Figure 9. Results on using weighted guided image filter (WGIF)-based noise reduction when the value of a pixel is below 5, and the results of directly amplifying pixels whose value is below 5 by using constant. The first row shows no denoised images, and the second row shown denoised images. Compare the amplified red blocks in the images. The method can effectively reduce noise and preserve details.
Figure 9. Results on using weighted guided image filter (WGIF)-based noise reduction when the value of a pixel is below 5, and the results of directly amplifying pixels whose value is below 5 by using constant. The first row shows no denoised images, and the second row shown denoised images. Compare the amplified red blocks in the images. The method can effectively reduce noise and preserve details.
Sensors 20 04614 g009
Figure 10. Comparison of the running times between six algorithms. The left shows the times of six algorithms, the right is the percentages on the running time in our algorithm.
Figure 10. Comparison of the running times between six algorithms. The left shows the times of six algorithms, the right is the percentages on the running time in our algorithm.
Sensors 20 04614 g010
Figure 11. Compare the brightened image (the second row of images is a supermarket in China) by using inaccurate camera response functions (CRFs) and accurate CRFs. The first column shows input images; the second column shows brightened images by using inaccurate CRFs; the last column shows brightened images by using accurate CRFs.
Figure 11. Compare the brightened image (the second row of images is a supermarket in China) by using inaccurate camera response functions (CRFs) and accurate CRFs. The first column shows input images; the second column shows brightened images by using inaccurate CRFs; the last column shows brightened images by using accurate CRFs.
Sensors 20 04614 g011
Table 1. PCQI of eight different algorithms.
Table 1. PCQI of eight different algorithms.
BookBuildingCoffeeVaseWindingMuseumGatePavilionMarketBridgeAvg
NPE [7]0.57560.48820.44960.62450.44350.44460.35680.35310.50800.32620.4572
SNIE0.53120.47620.45510.59660.42870.43790.36500.26220.46930.31500.4337
LIME [10]0.63330.49230.55840.64260.52040.43800.40910.36860.51950.34050.4923
LECARM [15]0.58740.49750.51730.65080.47510.44610.39250.27530.49450.32510.4662
Dong [8]0.56190.45950.46660.50400.43290.41630.38990.29900.48790.31900.4337
[18]0.52100.46130.43860.58020.49310.42580.38970.30110.45810.35890.4428
[19]0.51940.46010.40670.40480.41840.40190.35610.24480.46100.31110.3984
Our0.62290.51710.55710.67560.49820.46290.43500.31370.50950.42600.5018
Table 2. MEF-SSIM of eight different algorithms.
Table 2. MEF-SSIM of eight different algorithms.
BookVaseWindingMuseumBuildingBridgeCoffeeGatePavilionMarketAvg
NPE [7]0.85710.92370.76250.65660.81490.69160.74990.62850.73820.75590.7579
SNIE [13]0.85170.94050.78950.66840.84020.71820.79390.70970.64100.73920.7692
LIME [10]0.87300.91630.80590.63910.81110.69380.83980.69860.76660.77970.7823
LECARM [15]0.88970.94980.83000.67290.85700.73150.84590.74960.67300.77110.7971
Dong [8]0.83470.85990.73130.62790.80290.64310.76390.69540.71430.76090.7434
[18]0.75990.91350.81210.62500.81090.69050.73540.68210.72320.77100.7524
[19]0.72570.90230.70450.61340.79490.63140.71440.61800.70340.72160.7130
Our0.90610.95250.84470.67300.86010.74070.85440.75130.72010.77780.8081

Share and Cite

MDPI and ACS Style

Yang, Y.; Li, Z.; Wu, S. Low-Light Image Brightening via Fusing Additional Virtual Images. Sensors 2020, 20, 4614. https://doi.org/10.3390/s20164614

AMA Style

Yang Y, Li Z, Wu S. Low-Light Image Brightening via Fusing Additional Virtual Images. Sensors. 2020; 20(16):4614. https://doi.org/10.3390/s20164614

Chicago/Turabian Style

Yang, Yi, Zhengguo Li, and Shiqian Wu. 2020. "Low-Light Image Brightening via Fusing Additional Virtual Images" Sensors 20, no. 16: 4614. https://doi.org/10.3390/s20164614

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop