Next Article in Journal
A Randomized, Double-Blind, Placebo-Controlled Parallel Study on the Efficacy and Safety of Centella asiatica L. Extract for Reducing Alanine Transaminase (ALT) Level in Subjects with Elevated ALT
Next Article in Special Issue
Tooth Position Determination by Automatic Cutting and Marking of Dental Panoramic X-ray Film in Medical Image Processing
Previous Article in Journal
Design and Experimental Validation of Power Electric Vehicle Emulator for Testing Electric Vehicle Supply Equipment (EVSE) with Vehicle-to-Grid (V2G) Capability
Previous Article in Special Issue
Quantitative Retrieval of Soil Salinity Using Landsat 8 OLI Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Brightness Image Enhancement Algorithm

1
College of Information Science and Engineering, Xinjiang University, Ürümqi 830046, China
2
Key Laboratory of Signal Detection and Processing, Xinjiang Uygur Autonomous Region, Xinjiang University, Ürümqi 830046, China
3
Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai 200400, China
4
School of Engineering, Computing and Mathematical Sciences, Auckland University of Technology, Auckland 1020, New Zealand
5
Intelligent Systems Research Center, Ulster University, Londonderry BT48 7JL, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(23), 11497; https://doi.org/10.3390/app112311497
Submission received: 26 September 2021 / Revised: 25 November 2021 / Accepted: 29 November 2021 / Published: 4 December 2021
(This article belongs to the Special Issue Advances in Intelligent Control and Image Processing)

Abstract

:
In this paper, we introduce a tone mapping algorithm for processing high-brightness video images. This method can maximally recover the information of high-brightness areas and preserve detailed information. Along with benchmark data, real-life and practical application data were taken to test the proposed method. The experimental objects were license plates. We reconstructed the image in the RGB channel, and gamma correction was carried out. After that, local linear adjustment was completed through a tone mapping window to restore the detailed information of the high-brightness region. The experimental results showed that our algorithm could clearly restore the details of high-brightness local areas. The processed image conformed to the visual effect observed by human eyes but with higher definition. Compared with other algorithms, the proposed algorithm has advantages in terms of both subjective and objective evaluation. It can fully satisfy the needs in various practical applications.

1. Introduction

Videos are frequently used to judge and identify information. They have been widely used in public security monitoring. However, according to weather conditions, luminance conditions, capture equipment and other causes, high-brightness areas may exist in a given video, which results in the information of the target not being obtained. The existence of the bright regions cannot meet the needs of applications. In particular, strong light sources (such as car lights, streetlights and flashlights) make it challenging to enhance night-time images. However, although the existing image enhancement algorithms have apparent effects in improving the details of low-light images, they do not suppress the high-brightness areas in the picture. The existence of bright regions seriously affects the display of the highlighted area’s details and the observer’s direct visual experience.
To solve the abovementioned problems, we propose a method for video image processing based on tone mapping. The method is used to remove the bright part of the image, and it can restore the original details of the illuminated area while other areas with brightness remain.
The examined video was composed of one image frame, so we focused on the enhancement of a single image. The information of the area with uneven illumination was recovered by the image enhancement algorithm. Although the dynamic range can be compressed by global linear scaling, the result is unsatisfactory because the image structure is flattened linearly, losing most of the visual information in light areas and shadows.
Tone mapping is a technique for mapping high-dynamic range (HDR) images to display devices with low dynamic ranges. Tumblin et al. [1] introduced the concept of tone mapping to computer graphics in 1993. Larson et al. [2] modified and adjusted traditional histogram equalization technology and added some theories regarding the human visual perception system, such as contrast, sharpness and color sensitivity, which were inspired by the previous work of Ferwerda [3]. The method proposed by Ferwerda reduces the dynamic range of an image by simple linear scaling, so it cannot operate with strong dynamic range compression. Durand et al. [4] proposed separating an image into high-frequency sub-images and low-frequency sub-images by using an approximate bilateral filter and a downsampling acceleration filter. The high-frequency part is called the detail layer, and the low-frequency part is called the basic layer, which can better preserve edges. This method can be applied to any global tone mapping algorithm, but after using this method, halation is not entirely eliminated. For the improvement of this method, trilateral filters and other methods have appeared. In 2002, Reinhard et al. [5] proposed a tone mapping method based on photography. This algorithm (with a global component) mainly compresses the high-brightness part of an image. Photographic hue reproduction is a local tone mapping algorithm that retains the boundary effect and avoids halation. In addition, it has the advantage that it does not need to input corrected images. Ferradans et al. [6] proposed a global mapping method in line with the human visual model, and this algorithm enhances the local contrast of images.
In 2003, Drago et al. [7] proposed a classical global tone mapping algorithm—the adaptive logarithmic mapping algorithm—which can adaptively adjust the cardinality of logarithmic equations according to the different brightness values of pixels. This method can reduce the dynamic range of the examined image, but it easily loses local details, and its effect on improving the overall contrast of the image is not apparent. Duan et al. [8] proposed a novel hue mapping algorithm based on the fast adjustment of global histograms, which can effectively use the whole dynamic range of the display to reproduce the global contrast of high dynamic range images. However, this method cannot effectively preserve local contrast and details, which is a common weakness of global tone mapping operators. The local tone mapping methods proposed in [9,10,11,12,13] are better than the global tone mapping approach used in detail preservation and dynamic range compression. Qiao et al. [12] used local gamma correction, which is also regarded as a kind of local tone mapping algorithm. Tone mapping algorithms in [14,15,16,17,18,19,20,21] are the latest that have good processing effects in terms of image structure, detail preservation, lighting and other features. Wang [15] proposed a new blind tone mapping image (TMI) quality evaluation method based on local degradation characteristics and global statistical characteristics. The texture, structure, color and other image attributes can be local or global. The extracted local and global features are aggregated into the overall quality by regression. Most color mapping operators use the human visual system, and the scene’s dynamic range is displayed dynamically by a curve. For example, a model based on the sensitivity of the retina and an adjustment of the graph histogram was proposed in [21]. An asymmetric sigmoid curve is constructed to optimize this algorithm. Choi et al. [22] proposed a novel tone mapping method. Jung et al. [17] proposed a GAN training optimization model for the tone mapping of HDR images. The author collected a large number of tone mapping images, which existing tone mapping operators generated. An objective index is used for evaluation for all output tone mapping operators, and the algorithmic result with the highest evaluation index is selected and classified into the dataset. The model trained by the dataset can learn the best features of the tone mapping operator, so the experimental image processing effect is better than those of the existing tone mapping algorithms. Banderle et al. [23] proposed a new classification method for tone mapping based on the global and local tonal mapping methods commonly mentioned by other authors, frequency tone mapping approach (only the low-frequency part of the image) and split tone mapping approach (the image is divided into large areas and mapped separately for each area). This classification is more detailed than and different from those of other authors. For example, most authors classify the approaches in references [8,9] as local tone mapping methods, but Banderle ranked them as frequency tone mapping operators. Liang [21] proposed a new decomposition model that decomposes an HDR image into a base layer and detail layer and uses the prior information of the base layer and detail layer to minimize the halos and artifacts in the resulting image. Regarding tonal mapping image evaluation, [24] proposed a measure called the tonal mapping quality index (TMQI), which has become one of the most popular indicators for comparing the quality of tonal mapping images.
Although HDR video acquisition technology was relatively mature in approximately 2010, video tone mapping algorithms appeared as early as 15 years ago. After relevant research [25], most video tone mapping algorithms have the same problems: heavy artificial traces, contrast losses and detail losses.
Many methods for adjusting local contrast for video processing are based on optical flow prediction in the motion domain [26,27,28]. The advantage of these postoperative algorithms is that we do not need to consider how the tone mapping process is completed (as in previous algorithms), and we do not need to consider what changes need to be made to the filter or whether other operations need to be added for different tone curves. The disadvantage of this kind of method is that it is weaker than the filter method for presenting local contrast and actual video effects. Shahid et al. [29] proposed a hybrid model. As the term suggests, two different tone mapping algorithms are used to process the same frame. Each frame is processed by the method in [30], and the image is divided into regions with drastic changes and regions without drastic changes. The regions with drastic changes are mapped by the algorithm from [31], which preserves the details of the image through the local operation. The remaining region of the image is processed by a histogram correction algorithm [2]. Finally, to maintain the temporal properties of the video, the brightness of the mapped frames is scaled so that the intensity differences between consecutive frames do not exceed the preset threshold.
We propose a method that uses a window based on tone mapping and solves global optimization problems with local constraints. Specifically, our window method can operate over the whole image. In each window, a linear function limits tone reproduction to suppress strong edges while retaining weak edges naturally. An image with a high dynamic range can be compressed by solving the image-level optimization problem, which integrates all window-based constraints. A global optimization method is used to avoid the artifacts and halation caused by local tone mapping.
After using our optimized framework, the overall tone mapping effect of the given image is nonlinear and spatially variable, and it is suitable for the diverse structure of the image. Our method has a closed-form solution to achieve the global optimum. Due to this optimality, any local tone adjustment in our method causes the overall effect such that the error is minimized and distributed in the whole image. In addition, our method can flexibly adjust the image quality with only a few parameters to meet the various requirements related to the range compression ratio and the level of detail that must be retained.
Compared with those of the existing algorithms, the main contributions of the proposed algorithm are as follows:
(1)
The utilized small window can overcome the problem in which the global tone mapping algorithms cannot availably enhance the local contrast and details of the input image. The method we propose in this paper can operate globally and avoid halos and artifacts.
(2)
For the guidance map component, factors including image brightness, radiance, local average value, standard deviation and so on are considered, which have good effects on image restoration.
(3)
The experimental images selected in this paper are vehicle images, and the highlighted part blocks the information of the license plate. The background color of the license plate is blue. Therefore, the B channel in the RGB image is retained, and the R and G channels are exchanged for image preprocessing, which significantly affects the recovery of the image exposure area.
The rest of this paper is organized as follows: Section 2 shows the related work. Section 3 prepares and introduces our algorithm. Section 4 shows our experimental results. Section 5 is our conclusion.

2. Related Work

Generally, we divide tone mapping into two categories: global tone mapping and local tone mapping. Histogram-based algorithms and S-shaped transfer function-based algorithms (such as gamma correction) are the two main categories of global operators. The key for global tone mapping methods is to establish one-to-one mapping or low dynamic radiance mapping. In reference [19], the statuses and categories of tone mappings were summarized. Reference [20] proposed a multiscale image decomposition method that can preserve image edges. The results show that this type of image decomposition method is superior to bilateral filtering in producing range-compressed images.
Unlike the methods above, our algorithm does not need to decompose the input image into binary or fractional images. Therefore, it avoids the problems associated with layer decomposition. We propose a global optimization framework to constrain each pixel value naturally. In addition, our method does not need to impose any smoothness constraint on the final output.
Comparing our work with existing tone mapping algorithms, our method can significantly improve the pixel intensity in the overexposed area and balance the brightness of other areas to ensure that the overall contrast of the image will not change too much.

3. Proposed Method

We propose a window algorithm based on tone mapping, which improves the visual effect of the overexposed area in an image by compressing the image’s dynamic range and restoring the information in the exposed area. In this section, we introduce our algorithm in detail. First, the input image is linearized, and the RGB channel of the image is reconstructed; this is a simple preprocessing step. Then, a local gamma correction is performed on the image, and the local contrast of the image is further improved by color compression. The image passes through the previously set window frame, including a series of operations such as guided filtering, after which the final image is output. The algorithmic flow chart is shown in Figure 1:
The algorithm process mainly involves the following steps: (1) image input; (2) channel reconstruction; (3) gamma correction; (4) generating the guidance map c according to Equation (14); (5) restoring the RGB channels I k l in the tone mapped result by Equation (12); (6) resulted image output.
The first step of linearizing the image is crucial. The main reason for operating on the image in linear space is that this provides greater control and predictability regarding the brightness and color domain expansion. In an unknown area, it is difficult to predict how the expansion process will behave. In addition, to recover the scene attributes, such as the mean, geometric average, and standard deviation, we need to estimate the radiance of the scene accurately.
For an input image I l = f ( I h ) , I h is the radiance map of the image. f ( ) represents the operator. I l represents the radiance map of the image after the operation is completed. In short, the core idea of our algorithm is to increase smaller amplitudes and decrease larger amplitudes, this processing step should be conducted in a small area to meet the requirements of amplitude mapping in each pixel neighborhood. The linear function of window w i , centered on pixel i is expressed as follows:
I l ( j ) = m i I h ( j ) + n i , j w i
Formula (1) is a complete representation satisfying local constraints. The coefficients m and n directly affect the local contrast, and m represents the slope of the linear function. When m > 1 , the local contrast is enhanced so that the dark areas in the image are brightened; when 0 < m < 1 , the high-brightness areas in the image are improved. In the overlapping windows with multiple pixels, similar smoothing constraints are enforced in the neighborhood of the center pixel in the smooth area. The contrast is maintained reasonably in the area with abrupt amplitude changes, so there is no need to add constraints.
The defined local linear equation is applied to the window w i , and the image is processed by minimizing the solution:
i j w i [ I l ( j ) m i I h ( j ) n i ] 2
Because of the existence of a trivial solution, we need to add a constraint term to the above formula to make the minimization Formula (2) feasible. According to Formula (1), the coefficient m directly affects the local contrast of the image. We propose guiding its value in each small window to reduce the global contrast while maintaining the image’s information. We support the final objective function as follows:
f = i { j w i [ I l ( j ) m i I h ( j ) n i ] 2 + ε c i 2 ( m i c i ) 2 }
where c i is the default positive number, which is used to guide the processing of the local contrast. The collective c i in the image space form the guidance map. By setting an appropriate value for this map, the contrast of the local area can be adjusted correctly (weakening the strong-contrast areas and enhancing the low-contrast areas). The term ε c i 2 ( m i c i ) 2 is the squared relative error from the guidance map. The term c i 2 is used to normalize the difference between m i and c i . The term ε is the weight of the added item. We set its value to 0.1 in the experiments.
The importance of the guidance map c is as follows. Although the value c i affects the local contrast of the image, the accuracy c i does not need to be very high as the radiance value I h can be directly modified. In our optimization window, ε is a weight coefficient that imposes smooth constraints on the guidance map, and the value of this coefficient is very small. The radiance of each pixel is also limited by the linear equation i j w i [ I l ( j ) m i I h ( j ) n i ] 2 , which effectively maintains the image structure.

3.1. Optimization and Implementation

It is difficult to minimize Formula (3) directly because it contains several unknowns. For each group of linear coefficients m i and n i , the definitions exist only in a single window, so the minimization problem of Formula (3) can be expressed as:
arg min m ,   n ,   I l   f =   arg   min I l i arg min m i ,   n i f i   , where   f i = { j w i [ I l ( j ) m i I h ( j ) n i ] 2 + μ c i 2 ( m i c i ) 2 }
Through this expression, by computing the partial derivatives of function f i for m i and n i setting them to zero, we obtain the optimal summed solution and obtain the optimal I ^ l after solving the linear equation. The calculation process is as follows:
f i p i | m i = m i * ,   n i = n i * = 2 ε c i 2 ( m i * c i ) j ϵ w i 2 [ I l ( j ) m i * I h ( j ) n i * ] I h ( j ) = 0  
f i q i | m i = m i * ,   n i = n i * = j w i 2 [ I l ( j ) m i * I h ( j ) n i * ] = 0
Formulas (5) and (6) can be expressed as a linear relationship:
K i [ m i * n i * ] Τ = W i
where
K i = [ ε c i 2 + j w i I h ( j ) 2 j w i I h ( j ) j w i I h ( j ) j w i 1 ]
W i = [ ε c i 2 + j w i I h ( j ) I l ( j ) j w i I l ( j ) ]
By solving Formula (7), we can obtain:
[ m i * n i * ] Τ = K i 1 W i = 1 m i Δ i [ 1 μ i μ i Δ i + μ i 2 ] W i
where
m i = j w i 1
μ i = 1 x i j w i I h ( j )
Δ i = σ i 2 + ( ε c i 2 m i )
By taking the partial derivatives of the function f to I l ( k ) and setting them to zero, we have:
f I l ( k ) | I l = I ^ l = i | k w i 2 [ I ^ l ( k ) m i * I ^ h ( k ) n i * ] = 0
Combining (8) and (9), we can obtain:
f I l ( k ) | I l = I ^ l = i | k w i ( I ^ l ( k ) 1 m i Δ i ( ε c i 1 I h ( k ) + I h ( k ) j w i I h ( j ) I ^ l ( j ) μ i I h ( k ) j w i I ^ l ( j ) μ i ε c i 1 μ i j w i I h ( j ) I ^ l ( j ) + ( Δ i + μ i 2 ) j w i I ^ l ( j ) ) ) = 0
Because of the linearity of Formula (10), it can be simplified as follows:
D I ^ l = B
where
d k j = 2 f I l ( k ) I ^ l ( j ) = i | { k , j } w i ( δ k j 1 m i Δ i [ ( I h ( k ) μ i ) ( I h ( j ) μ i ) + Δ i ] )
b k = i | k w i ε m i Δ i c i [ I h ( k ) μ i ]
δ k j is a Kronecker delta. The optimal solution of Formula (3) is obtained by solving Formula (11). The matrix D in (11) is symmetric and sparse. In each row i of D , the number of nonzero elements reaches the area defined according to (11). Given an input image including N pixels, where the size of the window w i is set to k × k , D has fewer than N × ( 2 k 1 ) 2 nonzero elements (a k = 5 example is shown in Figure 2). Mathematically, the minimum possible window size is 2 × 2 , as there are three unknowns in each local window, and at least three linear constraints are needed to make the problem well-posed.
If the image has a medium resolution, we solve the linear system in (11) using symmetric L Q or the generalized minimum residual method. However, the above method needs several minutes to converge to process an image with a resolution higher than 3024 × 4032. We use a multigrid method [32] to speed up the processing time, which is the same as the method used in [33]. In our experiment, it took only a few seconds to produce satisfactory I ^ l .
The size of the local window can be adjusted in the system. We set window sizes from 3 × 3 to 11 × 11 and found that a window size of 5 × 5 could produce satisfactory results in terms of image clarity and visual effects. Considering the factors of edge definition and computational complexity, we chose a 5 × 5 window in our experiments.
The algorithm above operates on the brightness channel of the input image. After obtaining the radiance map of the image, the RGB channel was reconstructed using the algorithm mentioned in reference [34]:
I k l ( i ) = [ I k h ( i ) / I h ( i ) ] s × I ^ l ( i ) ,   k { r , g , b } ,
where I k l is one of the RGB color channels of the input image and I k h is one of the RGB color channels of the output image. The parameter s represents the saturation factor. A larger value of s produces a more saturated result. The introduction of (12) maintains the ratios between the brightness and color channels so that the color of the output image can be similar to that of the original image. In the experiment, we chose s [ 1 , 3 ] to produce images with natural color.

3.2. Guidance Map Configuration

In the linear optimization framework introduced before, the image processing quality is determined by the guidance map c . We next mainly describe how to obtain an appropriate guidance map c . The coefficient m i directly affects the local contrast of the image. To retain the visual information in the image, we reduced the value of m i where the local contrast was significant and increased the value of m i where the local contrast was slight. The standard deviation σ i of I h is a measure of the local contrast of the image in each window. However, c i is inversely proportional to σ i , which has a bad effect on the results. We defined the guidance map as:
c i = ( σ i β 2 + κ ) 1
To improve the visual quality of the resulting image, we calculated the Gaussian filtered image and the σ i of the Gaussian filtered image to reduce noise. In Figure 3b,c show the radiance map of I h ( i ) and the radiance map of the mean value μ , respectively. Since the local mean value μ and the radiance map I h ( i ) represent the radiance information of the image, we could see that the information contained in any one of the images was sufficient to contain all the information from the original image completely. If the two factors are added to the guidance image, the local contrast details of the image can be retained as much as possible. We also found that the contrast of the local window radiance is related to the absolute variance. Considering these factors, we add μ and I h ( i ) to the guidance map. We redefine c i as:
c i = [ μ i β 1 σ i β 2 I h ( i ) β 3 + κ ] 1
where κ is a small weight coefficient. We set it to 0.5 in our experiments to prevent c i from being divided by zero. Normally, β 1 [ 0.4 0.9 ] , β 2 [ 0.1 0.4 ] , β 3 = 0.1 .
Figure 3 shows the visual image of each parameter and the effect of our newly defined guidance map c i on the output image. Figure 3a is the input image with a large extensive exposure, in which background information can hardly be seen. Only the colors of some personnel’s clothes can be seen in this picture, and the visual effect is inferior. Figure 3b,c show the image brightness and mean radiance after performing 5 × 5 window processing, respectively. Figure 3d shows our measure for the local contrast based on the standard deviation of the radiance σ i . This step can preserve and strengthen the gradients and edges of the image. Figure 3h shows the image generated by the proposed algorithm, which is produced by the guidance map generated by Formula (7); its visual image is Figure 3f. We can see that compared with the output image produced by the standard deviation, average value and average brightness of a single image, the visual image of the proposed method contains all the information in the original image, which can significantly improve the restoration quality of the image. Compared with Figure 3g generated by Formula (6), Figure 3h generated by our algorithm is better in terms of its color, visual effect and detail processing. The detailed comparison is shown in Figure 3i–l.

3.3. Parameter Settings

Our settings were based on the following two points. First, the sum of the three β s affects the global dynamic compression of the input image; a large value makes the resulting image smoother. Second, increasing the value β 2 make the compressed edges with higher contrasts more obvious such that more space is left for less prominent areas.
In fact, in our algorithm, minor changes to the sum do not greatly impact the results. The values of these two parameters are small, and the ranges of the values are very small. β 3 is fixed at 0.1. For most test images, the default values are β 1 = 0.6 and β 2 = 0.2 , and most of the tested images can produce visually satisfactory results.

4. Comparative Analysis

To make the algorithm proposed in this paper closer to practical problems, we selected multiple high-brightness license plate images as the processing objects, and we compared our algorithm with several related algorithms proposed by Hui [35], Khan [36], Liang [21] and Xin [37]. This section includes two parts: part one assesses the restored overexposed images subjectively, and part two objectively compares the results produced by our method and other algorithms.
The sizes of the processed images were 3024 × 4032 pixels in the experiment. To prove the effectiveness of the proposed algorithm, all the algorithms were run on a computer with an Intel Core i5-9400 CPU @ 2.9 GHz (Acer, Taiwan, China), and the compiler software was MATLAB 2019b (9. 7. 0. 1190202, MathWorks, Natick, MA, USA, 2019). In this configuration, the experiments could be completed within 240 seconds.

4.1. Subjective Analysis

In this part, we mainly analyze the experimental results. As shown in Figure 4, Figure 5, Figure 6, Figure 7, the representative algorithms developed by Hui [35], Khan [36], Liang [21] and Xin [37] were compared with the algorithm proposed in this paper on different images.
From the perspective of visual effects, as shown in Figure 6e and Figure 7e, the algorithm of Xin [37] also improved the high-brightness area. Still, our algorithm had greater advantages in terms of the overall contrast of the image and the details of the scene. However, the other algorithms could not effectively recover the information of the key area for all the experimental images. The results showed that the effect of our algorithm regarding key information recovery was pronounced. Although the image’s hue was slightly green, especially as shown in Figure 7f (our algorithm made the vehicle tail lamp change from red to green), the critical exposure area information was recovered. To the overall improvement of the image, the visual effect was ideal, reaching expectations.
By observing the experimental results, we could find that the brightness of the original image with much light was improved significantly, and the information of the local area with much light was effectively restored. Compared with the image processed by the algorithm, the information of the high-brightness area was not recovered well.

4.2. Objective Analysis

We calculated four objective indicators, including the structural similarity (SSIM), peak signal-to-noise ratio (PSNR), image information entropy (H) and executive time. The bigger the first three indicators, the better. The specific values are shown in Table 1, Table 2, Table 3, Table 4, Table 5 below. Table 1 shows the objective values of Figure 4. Table 2 shows the objective values of Figure 5. Table 3 shows the objective values of Figure 6. Table 4 shows the objective values of Figure 7.
As shown in the tables, we can see that our algorithm ranked first, especially in terms of the SSIM and PSNR. Our algorithm was also better than that of Xin [37] in terms of the SSIM. SSIM is an index used to measure the similarity of two images. We got images with different illumination intensity at different times from the video, including lossless images and distorted images. We processed the high-brightness images through the proposed algorithm. Comparing the resulting image with the lossless, we evaluated the quality of image restoration. SSIM reflects the comprehensive similarity of two images in brightness, contrast and structure. By averaging the objective indicators of the 50 high-brightness video images, the average PSNR of our algorithm was 29.53, which was 7.22 higher than that of the algorithm by Xin [37]. In terms of the PSNR, the proposed algorithm was better than the algorithm of Xin [37]. This shows that after compression, the image quality was better than that obtained by the algorithm of Xin [37]. The larger the PSNR value, the smaller the image distortion. “H” reflects the average amount of information in the image and demonstrates the richness of image information from the perspective of information theory. Generally, the bigger the H value, the better. Our method ranked second in H. For the time part, because the generation of a guidance map takes a long time, our algorithm was not dominant in all comparative experiments. This will be our next optimization direction.
In summary, by combining a subjective evaluation and an objective analysis, we can conclude that the proposed algorithm could effectively recover the exposure area information from high-brightness video images with sound visual effects, and the objective indicators were much better than those produced by the compared algorithms.

5. Conclusions

In this paper, we proposed a window algorithm based on tone mapping. In this method, a local linear window is utilized to compress the image pixel values into a specific range. While accurately recovering the information of the exposure area, our algorithm could effectively enhance the local contrast and preserve the image details and structures. Global optimization genuinely suppresses artifacts. Therefore, our algorithm could complete the restoration of high-brightness images.

Author Contributions

Conceptualization, Y.W. and Z.J.; methodology, Y.W.; software, Y.W.; validation, Y.W. and Z.J.; formal analysis, Y.W. and Z.J.; investigation, Y.W.; resources, Z.J.; data curation, Y.W.; writing—original draft preparation, Y.W.; writing—review and editing, Z.J. and N.K.K.; visualization, Y.W.; supervision, J.Y.; funding acquisition, Z.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation of China (No. U1803261) and the International Science and Technology Cooperation Project of the Ministry of Education of the People’s Republic of China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data provided in this study are available from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interests.

Abbreviations

HDRHigh dynamic range
HInformation entropy
SSIMStructural similarity
PSNRPeak signal-to-noise ratio

References

  1. Tumblin, J.; Rushmeier, H. Tone reproduction for realistic images. IEEE Comput. Graph. Appl. 1993, 13, 42–48. [Google Scholar] [CrossRef]
  2. Larson, G.W.; Rushmeier, H.; Piatko, C. A visibility matching tone reproduction operator for high dynamic range scenes. IEEE Trans. Vis. Comput. Graph. 1997, 3, 291–306. [Google Scholar] [CrossRef] [Green Version]
  3. Ferwerda, J.A.; Pattanaik, S.N.; Shirley, P.; Green-Berg, D.P. A model of visual adaptation for realistic image synthesis. In Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, New York, NY, USA, 1 August 1996. [Google Scholar]
  4. Durand, F.; Dorsey, J. Interactive Tone Mapping, in: Rendering Techniques 2000. In Proceedings of the Eurographics Workshop, Brno, Czech Republic, 26–28 June 2000; pp. 219–230. [Google Scholar]
  5. Reinhard, E.; Stark, M.; Shirley, P. Photographic tone reproduction for digital images. ACM Trans. Graph. 2002, 21, 267–276. [Google Scholar] [CrossRef] [Green Version]
  6. Ferradans, S.; Bertalmio, M.; Provenzi, E.; Caselles, V. An analysis of visual adaptation and contrast perception for tone mapping. Pattern Anal. Mach. Intell. 2011, 33, 2002–2012. [Google Scholar] [CrossRef] [Green Version]
  7. Drago, F.; Myszkowski, K.; Annen, T.; Chiba, N. Adaptive logarithmic mapping for displaying high contrast scenes. Comput. Graph. Forum 2003, 22, 419–426. [Google Scholar] [CrossRef]
  8. Duan, J.; Bressan, M.; Dance, C.; Qui, G. Tone-mapping high dynamic range images by novel histogram adjustment. Pattern Recognit. 2010, 43, 1847–1862. [Google Scholar] [CrossRef]
  9. Durand, F.; Dorsey, J. Fast Bilateral Filtering for the Display of High-Dynamic-Range Images. ACM Trans. Gr. 2002, 21, 257–266. [Google Scholar] [CrossRef] [Green Version]
  10. Fattal, R.; Lischinski, D.; Werman, M. Gradient Domain High Dynamic Range Compression. ACM Trans. Gr. 2002, 21, 249–256. [Google Scholar] [CrossRef] [Green Version]
  11. Li, Y.; Sharan, L.; Adelson, E.H. Compressing and Companding High Dynamic Range Images with Subband Architectures. ACM Trans. Gr. 2005, 24, 836–844. [Google Scholar] [CrossRef] [Green Version]
  12. Qiao, M.; Ng, M.K. Tone mapping for high-dynamic-range images using localized gamma correction. J. Electron. Imaging 2015, 24, 013010. [Google Scholar] [CrossRef]
  13. Bruce, N.D.B. ExpoBlend: Information preserving exposure blending based on normalized log-domain entropy. Comput. Graph. 2014, 39, 12–23. [Google Scholar] [CrossRef]
  14. Kasliwal, H.; Modi, S. A Novel Approach for Reduction of Dynamic Range Based on Hybrid Tone Mapping Operator. Procedia Comput. Sci. 2015, 58, 265–271. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, X.; Jiang, Q.; Shao, F.; Gu, K.; Zhai, G.; Yang, X. Exploiting Local Degradation Characteristics and Global Statistical Properties for Blind Quality Assessment of Tone-Mapped HDR Images. IEEE Trans. Multimed. 2021, 23, 692–705. [Google Scholar] [CrossRef]
  16. Liu, T.; Liu, H.; Wu, Y.; Yin, B.; Wei, Z. Exposure Bracketing Techniques for Camera Document Image Enhancement. Appl. Sci. 2019, 9, 4529. [Google Scholar] [CrossRef] [Green Version]
  17. Jung, S.-W.; Kwon, H.-J.; Lee, S.-H. Enhanced Tone Mapping Using Regional Fused GAN Training with a Gamma-Shift Dataset. Appl. Sci. 2021, 11, 7754. [Google Scholar] [CrossRef]
  18. Shu, X.; Wu, X. Locally Adaptive Rank-Constrained Optimal Tone Mapping. ACM Trans. Graph. 2018, 37, 1–10. [Google Scholar] [CrossRef]
  19. Eilertsen, G.; Mantiuk, R.K.; Unger, J. A comparative review of tone-mapping algorithms for high dynamic range video. Comput. Graph. Forum 2017, 36, 565–592. [Google Scholar] [CrossRef] [Green Version]
  20. Dae-Hong, L.; Ming, F.; Seung-Wook, K.; Kang, M.C.; Ko, S.J. High dynamic range image tone mapping based on asymmetric model of retinal adaptation. Signal Process. Image Commun. 2018, 68, 120–128. [Google Scholar]
  21. Liang, Z.; Xu, J.; Zhang, D.; Cao, Z.; Zhang, L. A Hybrid l1-l0 Layer Decomposition Model for Tone Mapping. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  22. Choi, H.-H.; Kang, H.-S.; Yun, B.-J. Tone Mapping of High Dynamic Range Images Combining Co-Occurrence Histogram and Visual Salience Detection. Appl. Sci. 2019, 9, 4658. [Google Scholar] [CrossRef] [Green Version]
  23. Banterle, F.; Artusi, A.; Sikudova, E.; Bashford-Rogers, T.; Ledda, P.; Bloj, M.; Chalmers, A. Dynamic range compression by differential zone mapping based on psychophysical experiments. In Proceedings of the SAP 2012-ACM Symposium on Applied Perception, New York, NY, USA, 5 August 2012; pp. 39–46. [Google Scholar]
  24. Yeganeh, H.; Wang, Z. Objective Quality Assessment of Tone-Mapped Images. IEEE Trans. Image Process. 2013, 22, 657–667. [Google Scholar] [CrossRef]
  25. Eilertsen, G.; Wanat, R.; Mantiuk, R.K.; Jonas, U. Evaluation of Tone Mapping Operators for HDR-Video. Comput. Graph. Forum 2013, 32, 275–284. [Google Scholar] [CrossRef] [Green Version]
  26. Lang, M.; Wang, O.; Aydin, T.; Smolic, A.; Gross, M.; Zurich, D.R.; Zurich, E.T.H. Practical temporal consistency for image-based graphics applications. ACM Trans. Graph. 2012, 31, 1–8. [Google Scholar] [CrossRef]
  27. Dong, X.; Bonev, B.; Zhu, Y.; Alan, L.Y. Region-based temporally consistent video post-processing. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  28. Bonneel, N.; Tompkin, J.; Sunkavalli, K. Blind video temporal consistency. ACM Trans. Graph. 2015, 34, 1–9. [Google Scholar] [CrossRef] [Green Version]
  29. Shahid, H.; Li, D.; Fanaswala, M.; Pourazad, P. Nasiopoulos A new hybrid tone mapping scheme for high dynamic range (HDR) videos. In Proceedings of the IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 9–12 January 2015. [Google Scholar]
  30. Anand, S. Edge Detection using Directional Filter Bank. Auton. Neurosci. Basic Clin. 2012, 1, 10409745. [Google Scholar] [CrossRef]
  31. Kuang, J.; Johnson, G.M.; Fairchil, M.D. iCAM06: A refined image appearance model for HDR image rendering. J. Vis. Commun. Image Represent. 2007, 18, 406–414. [Google Scholar] [CrossRef]
  32. Hager, W.W. Multiscale Optimization Methods and Applications; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  33. Levin, A.; Lischinski, D.; Weiss, Y. A Closed Form Solution to Natural Image Matting. IEEE Comput. Soc. 2006, 1, 61–68. [Google Scholar]
  34. Schlick, C. A Customizable Reflectance Model for Everyday Rendering. In Proceedings of the Fourth Eurographics Workshop on Rendering, Paris, France, 14–16 June 1993; pp. 73–83. [Google Scholar]
  35. Hui, J.; Xixi, J.; Lei, Z. Clustering based content and color adaptive tone mapping. Comput. Vis. Image Underst. CVIU 2018, 168, 37–49. [Google Scholar]
  36. Khan, I.R.; Rahardja, S.; Khan, M.M.; Movania, M.M.; Fidaa, A. A Tone-Mapping Technique Based on Histogram Using a Sensitivity Model of the Human Visual System. IEEE Trans. Ind. Electron. 2018, 65, 247. [Google Scholar] [CrossRef]
  37. Xin, Y.; Jia, Z.; Yang, J.; Kasabov, N.K. Specular Reflection Image Enhancement Based on a Dark Channel Prior. IEEE Photonics J. 2021, 13, 6500211. [Google Scholar] [CrossRef]
Figure 1. Algorithmic structure of the proposed method.
Figure 1. Algorithmic structure of the proposed method.
Applsci 11 11497 g001
Figure 2. Illustration of the region containing pixel i . All windows w with size 5 × 5 that have i must be within the yellow part. Therefore, under the influence of the local window w , there are at most 80 pixels around pixel i . If the image has N pixels, the total number of nonzero elements in matrix D is less than 80 × N .
Figure 2. Illustration of the region containing pixel i . All windows w with size 5 × 5 that have i must be within the yellow part. Therefore, under the influence of the local window w , there are at most 80 pixels around pixel i . If the image has N pixels, the total number of nonzero elements in matrix D is less than 80 × N .
Applsci 11 11497 g002
Figure 3. (a) Original image; (b) original radiance map I h ( i ) ; (c) map of the mean radiance μ i ; (d) map of the standard deviation σ i ; (e) the guidance map generated by (6); (f) the visualized guidance map c i ; (g) the result obtained using the guidance map shown in (e); (h) the result obtained using the guidance map shown in (f); (il) the comparison of the details of figures (g,h), respectively.
Figure 3. (a) Original image; (b) original radiance map I h ( i ) ; (c) map of the mean radiance μ i ; (d) map of the standard deviation σ i ; (e) the guidance map generated by (6); (f) the visualized guidance map c i ; (g) the result obtained using the guidance map shown in (e); (h) the result obtained using the guidance map shown in (f); (il) the comparison of the details of figures (g,h), respectively.
Applsci 11 11497 g003
Figure 4. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Figure 4. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Applsci 11 11497 g004
Figure 5. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Figure 5. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Applsci 11 11497 g005
Figure 6. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Figure 6. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Applsci 11 11497 g006
Figure 7. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Figure 7. Experimental results: (a) original image, (b) Hui [35], (c) Khan [36], (d) Liang [21], (e) Xin [37], (f) proposed.
Applsci 11 11497 g007
Table 1. Objective values of the images shown in Figure 4.
Table 1. Objective values of the images shown in Figure 4.
MethodsHSSIMPSNRTime(s)
Hui [35]7.080.3810.91123.5
Khan [36]7.960.6516.1210.6
Liang [21]7.610.5614.3650.2
Xin [37]6.950.8722.2444.3
Proposed7.660.9530.05225.3
Table 2. Objective values of the images shown in Figure 5.
Table 2. Objective values of the images shown in Figure 5.
MethodsHSSIMPSNRTime(s)
Hui [35]7.250.5112.68125.7
Khan [36]7.920.6515.5810.7
Liang [21]7.310.6013.8653.7
Xin [37]6.840.8924.3245.2
Proposed7.480.9428.86223.2
Table 3. Objective values of the images shown in Figure 6.
Table 3. Objective values of the images shown in Figure 6.
MethodsHSSIMPSNRTime(s)
Hui [35]7.340.6514.14125.8
Khan [36]7.900.7217.1810.6
Liang [21]7.410.6714.6950.6
Xin [37]6.690.8219.6443.1
Proposed7.510.9529.65221.8
Table 4. Objective values of the images shown in Figure 7.
Table 4. Objective values of the images shown in Figure 7.
MethodsHSSIMPSNRTime(s)
Hui [35]7.260.3910.76126.3
Khan [36]7.950.4315.6211.2
Liang [21]7.330.4211.7054.3
Xin [37]6.300.9124.3244.5
Proposed7.480.9329.76224.1
Table 5. Average objective values of 50 high-brightness images.
Table 5. Average objective values of 50 high-brightness images.
MethodsHSSIMPSNRTime(s)
Hui [35]7.270.5512.76125.2
Khan [36]7.930.6615.8910.9
Liang [21]7.450.5913.9751.9
Xin [37]6.750.8722.3144.8
Proposed7.560.9429.53223.2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wei, Y.; Jia, Z.; Yang, J.; Kasabov, N.K. High-Brightness Image Enhancement Algorithm. Appl. Sci. 2021, 11, 11497. https://doi.org/10.3390/app112311497

AMA Style

Wei Y, Jia Z, Yang J, Kasabov NK. High-Brightness Image Enhancement Algorithm. Applied Sciences. 2021; 11(23):11497. https://doi.org/10.3390/app112311497

Chicago/Turabian Style

Wei, Yifei, Zhenhong Jia, Jie Yang, and Nikola K. Kasabov. 2021. "High-Brightness Image Enhancement Algorithm" Applied Sciences 11, no. 23: 11497. https://doi.org/10.3390/app112311497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop