You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

15 May 2024

Retinex Jointed Multiscale CLAHE Model for HDR Image Tone Compression

,
and
School of Electronic and Electrical Engineering, Kyungpook National University, 80 Daehak-ro, Buk-gu, Daegu 41566, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue New Advances and Applications in Image Processing and Computer Vision

Abstract

Tone-mapping algorithms aim to compress a wide dynamic range image into a narrower dynamic range image suitable for display on imaging devices. A representative tone-mapping algorithm, Retinex theory, reflects color constancy based on the human visual system and performs dynamic range compression. However, it may induce halo artifacts in some areas or degrade chroma and detail. Thus, this paper proposes a Retinex jointed multiscale contrast limited adaptive histogram equalization method. The proposed algorithm reduces localized halo artifacts and detail loss while maintaining the tone-compression effect via high-scale Retinex processing. A performance comparison of the experimental results between the proposed and existing methods confirms that the proposed method effectively reduces the existing problems and displays better image quality.

1. Introduction

The human visual system (HVS) has a color constancy that aims to perceive the color of an object consistently, even if the lighting conditions change [1]. Retinex theory, a representative tone-mapping algorithm, is designed to mimic how the human eye and brain recognize and process light and color. Retinex separates images into illumination and reflectance components, enhancing the overall brightness, luminance contrast, and detail information in images. Retinex is used to adjust luminance under low-light conditions and reduce noise [2,3]. According to the hyperparameter σ of the Gaussian filter used to extract the illumination layer, the details and noise of the output image exhibit a trade-off. The multiscale Retinex (MSR) method combines single-scale Retinex (SSR) with multiple scales to address this problem [4]. The MSR method improves noise reduction and enhances details. However, it has limitations regarding noise removal and can induce a halo effect, distorting colors along the boundaries of high-contrast objects or figures.
In addition to the Retinex technique, various tone-mapping processes have been developed for compressing images with wide dynamic ranges. For example, Reinhard et al. proposed correcting the image through a simple sigmoid transfer curve to reproduce wide dynamic range images in various display environments [5]. The sigmoid transfer curve is defined by an experimental parameter, which could make it challenging to establish the optimal level of parameters for each image. Due to the significant differences in dynamic range between the input and output caused by tone mapping, some information may be lost or artifacts can occur. Therefore, several methods effectively preserve detailed image information by processing the basic and detailed components separately [6,7]. A prominent method for base–detail separation is the bilateral filter, which smooths input images while preserving the edge components. Durand and Dorsey developed an accelerated bilateral filter, the fast bilateral filter, to expedite filtering [8]. Moreover, Kwon et al. proposed edge-adaptive layer blurring to estimate and compensate for halo areas to reduce these artifacts caused by localized tone mapping [9]. Lee et al. proposed a tone-mapping technique through visual achromatic responses to reduce artifacts caused by differences between the compressed base and detail layers [10]. The extended version of the color appearance model (CAM) based on the HVS, iCAM06, is one of the techniques to reproduce wide dynamic range images. The iCAM06 reflects characteristics of the HVS, such as color adaptation and tone compression under diverse lighting conditions [11]. In iCAM06, feature-based fusion is applied to preserve image edge structures and details. However, there may be a detail reduction problem due to feature-based fusion with fixed-edge stopping functions. Hence, Kwon et al. proposed using the characteristics of the contrast sensitivity function as a base-detail separation method and detail compensation technique to preserve iCAM06 edge components [12].
To implement HDR image, multiple LDR images with various exposures under the same conditions are required. However, acquiring multiple images under precisely identical conditions in the real world can be challenging. Therefore, a technology is required to acquire an image with wide dynamic range through tone compression from one LDR image.
The existing method of removing surround through multiple blurring uses multiple Gaussian filters, so the output image is dependent on the σ value. A larger σ value leads to decreased detail, while a smaller one can result in halo and noise issues. Learning-based methods often suffer from low tone compression effects and cause some image distortions, such as an overall white balance shift in the image. As such, existing methods have limitations in expressing a single LDR image as an HDR image.
In this paper, we aim to represent a single LDR image as a superior HDR image in various aspects. This paper proposes effective tone compression and halo artifact reduction using Retinex theory and multiscale contrast limited adaptive histogram equalization (CLAHE). High-scale SSR is advantageous for global tone compression, and CLAHE is suitable for local tone compression. Therefore, by combining the two theories, we derive an image with excellent overall tone compression.
The level of image improvement of CLAHE varies depending on the TileSize and ClipLimit parameters. The proposed method enhances global tone-mapping performance based on high-σ value Retinex while integrating low levels of TileSize and ClipLimit in CLAHE to reduce halo artifacts and noise. Additionally, to increase sharpness and enhance detail, high levels of TileSize and ClipLimit are coordinated. The proposed method applies a multiscale CLAHE to the optimal global scale Retinex image to enhance global and local toning performance. Retinex images with reduced details due to trade-offs according to the high sigma value of the Gaussian filter were compensated for using multiscale CLAHE. The multiscale CLAHE technique compensates for reduced details using localized histogram equalizations based on two hyperparameters: TileSize and ClipLimit. Finally, the resultant image is generated through the color compensation step with the color channel of the original image. Thus, the proposed method enhances the image quality via effective tone mapping, detail representation, noise reduction, and color expression. In summary, Retinex jointed multiscale CLAHE improves the tone-compression effect across the entire image by combining the two technologies, which are complementary each other in terms of global and local.

3. Proposed Method

Figure 3 illustrates the tone-mapping algorithm proposed in this paper, which employs Retinex processing and multiscale CLAHE. The proposed method aims to address image distortion problems, such as halo artifacts and noise, which are not adequately resolved by Retinex theory and other tone-mapping algorithms by introducing multiscale CLAHE based on high-scale Retinex. The proposed method also aims to enhance tone compression and detail representation capabilities. The proposed method consists of SSR, multiscale CLAHE, and color correction, as illustrated in Figure 3.
Figure 3. Flow chart for the proposed method.
As shown in Figure 3, the algorithm proceeds in the order of SSR, multiscale CLAHE, and color correction. First, to reduce halo artifacts and noise, high-scale SSR is performed. In this process, a high-σ value Gaussian filter is applied to heavily blur the surround layer, effectively removing halo artifacts and obtaining a reflection layer with superior global toning. Next, the multiscale CLAHE described in the blue box is performed. CLAHE is used to compensate for visibility, such as details lost by strong SSR. However, single-parameter CLAHE can introduce noise while compensating for details, thereby compromising overall image quality. Therefore, this paper synthesizes multiple CLAHE layers with different parameters to suppress noise induction while compensating for details, thus enhancing overall image quality. The images from each scale are weighted and synthesized. Finally, the algorithm undergoes a process to naturally restore some distorted colors, as indicated in the green box of Figure 3. Lastly, as shown in the fourth box, the synthesized luminance channel and the color-restored a and b channels are combined, and the output image is provided by converting them into RGB channels. In summary, the main progress of the algorithm is as follows:
  • To effectively address the halo artifact and noise issues that arise in conventional multi-blurring techniques, high-scale SSR is performed.
  • To compensate for visibility while maintaining image quality, we perform CLAHE on multiple scales.
  • To restore some distorted colors due to the previous procedure, the color correction method, MSRCR, is performed.

3.1. Decomposition of Reflectance and Illumination Layers

First, the input image in the RGB color space is converted to the Lab color space. Because the Lab color space separates the luminance channel and color channels, the luminance channel can be synthesized, reducing the color information distortion in the image. The SSR method is applied to the separated luminance channel to divide it into illumination and reflectance components. As depicted in Figure 1, a higher value of the standard deviation σ of the Gaussian function results in a more effective reduction of the halo artifacts and noise. Therefore, considering the compensation for details through CLAHE, even as image details decrease, σ is set to a high value to reduce noise significantly and perform high tone mapping. In this paper, considering that the computational time is more than double the standard deviation increases, σ is set appropriately to 250. The decomposition of the luminance channel into reflectance and illumination layers through the Gaussian function is performed as follows:
L s x , y = G [ L ( x , y ) ] ,
L R x , y = L x , y G L x , y ,
where G represents the Gaussian filter, L s denotes the illumination layer image obtained by applying a Gaussian filter to the luminance channel L of the input image, and L R indicates the reflectance layer image obtained by removing the illumination layer component from the luminance image L .

3.2. Multiscale CLAHE

High-scale SSR processing enables acquiring images with no noise and effective tone compression. However, the brightness contrast and detail are decreased due to the trade-off. The image enhancement method CLAHE is applied to increase brightness contrast and compensate for the decreased detail to address this problem. The degree of image enhancement varies depending on the direction in which the hyperparameters of CLAHE (ClipLimit and TileSize) are set. Figure 4 depicts images with CLAHE applied with various parameters.
Figure 4. Comparison of image enhancement with various parameter values for ClipLimit and TileSize.
Generally, as the size of ClipLimit increases, the contrast and detail compensation also increase. However, when ClipLimit becomes too large, excessive compensation leads to increased noise and results in unnatural output. As for TileSize, increasing it reduces the overall noise and improves tone compression, but excessive values lead to severe halo artifacts and image degradation, resulting in unnatural output. Therefore, an image with superior brightness contrast, detail compensation, noise reduction, and tone compression can be obtained by composing n pairs (ClipLimit and TileSize), applying weights to the images obtained with each pair, and combining them. For optimal synthesis, the diagonal matrix images from the image array in Figure 3 are adopted as the images used for composition. High-scale CLAHE achieves excellent detail representation and tone compression, whereas low-scale CLAHE reduces noise and mitigates halo artifacts. Middle-scale CLAHE serves as a connector for combining them. Superior image quality can be achieved through this synthesis. The parameters of the diagonal matrix result in each pair of parameters being (n, 2n+1). In this paper, n = 3; thus, three CLAHE layers were synthesized, and the parameter pairs (βn, Mn × Mn) for each CLAHE layer were set to (1.0, 4 × 4), (2.0, 8 × 8), and (3.0, 16 × 16), respectively.
The reflectance layer image LR of the luminance channel is divided into tiles of the corresponding size according to the TileSize parameter Mn × Mn to apply local histogram equalization. The representation of this process follows:
L R = L R 11 L R 1 j L R i 1 L R i j ( 1 i P M n , 1 j Q M n ) ,
where L R i j   represents the regions divided into tiles of size M n × M n , and P   and Q   are the dimensions of the input image. The ij-th region of the image corresponding to each tile is as follows:
L R i j = T 11 T 1 M N T M n 1 T M n M n ,
where T M n M n represents the pixel values of the ij-th tile of the input image.
Histogram equalization is applied to each tile by limiting the number of histogram bins based on the ClipLimit value. First, the probability mass function (PMF) is constructed from the pixel values of each tile to obtain the CDF, which serves as a means of applying histogram equalization, as follows:
h i j l = x = 1 M n y = 1 M n δ T x y l , l = 0 , 1 , 2 , , L 1 ,
where h i j   represents the PMF of the ijth tile, T x y denotes the pixel value at the xy-th position within the ij-th tile, and L represents the maximum pixel value. Additionally, δ is the impulse function that determines which brightness value the T x y corresponds to.
The PMF must be limited so that it does not have more than a certain number of pixels at specific brightness values to prevent excessive brightness in certain areas. Moreover, if the limit is exceeded, the excess pixels are redistributed evenly across the entire brightness range. This process is conducted according to the following equations:
N e x c e s s = k = 0 L 1 m a x ( h i j k β n , 0 ) ,
N a v e r a g e = N e x c e s s L ,
h c l i p p e d l = h i j l + N a v e r a g e ,
where N e x c e s s represents the number of pixels exceeding the ClipLimit   β n , L denotes the maximum brightness value of the pixels (256 for an 8-bit image), and N a v e r a g e indicates the number of pixels to be evenly redistributed across the entire brightness range. In addition, h c l i p p e d represents the new PMF of the ijth tile after the redistribution.
After redistribution, the process is repeated by limiting the pixel count again until N e x c e s s reaches zero. Finally, h c l i p p e d takes the following form:
h c l i p p e d l =                       β n                     h l β h i j l + N u p   h l < β .
The number of pixels exceeding the ClipLimit is constrained to β n , whereas the number of pixels less than β n increases by the amount of N u p which is the amount of the pixels for redistribution. The formula for obtaining the CDF for histogram equalization using the constrained PMF is as follows.
C i j l = 1 M n 2 k = 0 l h c l i p p e d ( k ) ,
where C i j represents the CDF to be multiplied by each pixel value of the ij-th tile. In this process, the CDF is normalized by dividing by the total number of pixels in the tile, M n 2 , to ensure that the slope of the CDF does not exceed 1 within the tile.
The CDF, C i j , obtained from Equation (18), is applied to the ij-th tile. Then, histogram equalization is performed by multiplying by the maximum pixel intensity value, L − 1, to the tile. This process is conducted according to the following equation:
L R i j x , y = C i j L R i j x , y × L 1 ,
where L R i j represents the ij-th tile of the L R image, where CLAHE was applied, and the × symbol is the multiplication operator.
After performing local histogram equalization for each tile, block artifacts occur (Figure 5). Pixel values between tile boundaries do not smoothly transition, resulting in noticeable gaps. Bilinear interpolation is applied to each pixel based on the tile centers to address this problem. Before applying the technique, pixels are categorized into internal regions, boundary regions, and corner regions based on their positions and neighboring tiles (Figure 6). Different equations are applied to each region accordingly.
Figure 5. Blocking effect due to local histogram equalization.
Figure 6. Region classification based on pixel positions in the input image: (a) Inside region; (b) Boundary region; (c) Corner region.
First, the transformation function for pixels in the internal region is calculated as follows for the inside region:
L R i j x , y = s r + s q p + q f L R i 1 j 1 + p p + q f L R i j 1 + r r + s q p + q f L R i 1 j + q p + q f L R i j
where L R i j , L R i 1 j , L R i j 1 , and L R i 1 j 1 are the center pixels of the four adjacent tiles of a desired pixel, in which each tile is subjected to contrast restriction histogram equalization. The values of p, q, s, and r represent distances between the current pixels from the centers of the four adjacent tiles. f L R i j , f L R i 1 j ,   f L R i j 1 ,   and f L R i 1 j 1 represent the value of the center pixel of each tile, and L R i j x , y indicates the pixel value after the interpolation method is applied.
The transformation functions for the boundary and edge regions, respectively, are similarly expressed as follows:
L R i j x , y = s r + s f L R i j 1 + r r + s f L R i j ,
L R i j x , y = L R i j x , y .
Hence, each output image fully processed by CLAHE with each scale is represented as follows:
L m a p p e d , n = L R 11 L R 1 j L R i 1 L R i j ( 1 i P M n , 1 j Q M n ) .
The combined image of the obtained output images at each scale is represented as follows:
L f u s e d = n = 1 N w n L m a p p e d , n ,
w h e r e   L m a p p e d represents the reflection layer image L R with the high-scale SSR and single-scale CLAHE applied, where w n denotes the weight applied to each image. In addition, N   indicates the number of images for fusion, which is set to N = 3 in this paper. Finally, L f u s e d denotes the reflection layer image with multiscale CLAHE applied.

3.3. Color Compensation

After the conversion to the Lab color space, the color information of the original image is separated into the a and b channels, whereas the luminance information is represented by the L channel. Therefore, changes in the luminance channel L due to tone adjustments do not affect the original color information. However, imbalances may occur between the modified luminance channel L and the color channels a and b, leading to unnatural color representation. Additionally, when applying Retinex theory, some colors may distort during the tone-mapping process, differing from the intended colors. Retinex theory incorporates gray world assumption (GWA) theory [4]. GWA is a technique based on the assumption that the mean values of the R, G, and B channels of an image are equal. It adjusts the gain to make the mean values of the R, G, and B channels equal. While GWA performs well in images with a uniform range of colors, it can lead to color desaturation in images where certain colors dominate, resulting in a grayscale appearance. Similarly, Retinex theory that embeds GWA can cause the same problem. As seen in Figure 7, it can distort the original colors of plants and buildings, making them gray. Therefore, a process to restore the distorted grayscale images using the color information from the original image is necessary.
Figure 7. Color desaturation caused by Retinex and different L channel ratios with a and b channels: (a) an original input image, (b) an image without color correction, and (c) an image with color correction.
Color channels that match the modified luminance channel must be obtained by correcting the color information of the output image with the color information of the original image to address these problems. The color correction method uses the MSRCR technique, which performs color restoration based on multiscale Retinex theory. The algorithm for color correction is provided as follows:
I x , y = I ( x , y ) j = 0 S I j ( x , y ) ,
C C x , y = γ l o g ( α I x , y ) ,
R G B C C x , y = C C ( x , y ) · R G B f u s e d ( x , y ) ,
w h e r e   I represents the input image and I j represents each R, G, and B channel of the input image. The term C C is the color correction function obtained for I through constants and γ . In this case, the experimentally determined appropriate values are = 125 and γ = 46. The term R G B f u s e d ( x , y ) represents the image without color correction. Finally, R G B C C x , y represents the result image with color correction.
The color correction function, obtained by applying the proposed method to the processed luminance channel L f u s e d and the original a and b channels, is multiplied by the RGB fused image, which is the image transformed into the RGB color space. However, due to the high correlation between luminance and each color channel in the RGB color space, the luminance component could be distorted. Therefore, only the color information of the image after color correction is obtained, and the luminance information is synthesized using the luminance channel L f u s e d of the proposed method. For this purpose, the R G B C C image is converted back to the Lab color space, and only the a and b color channels are extracted. Finally, the synthesized luminance channel L f u s e d and color channels   a c c and b c c , processed by the color correction method, are combined, and the final synthesized image is obtained after converting it to the RGB color space.

4. Simulations

4.1. Experimental Results

In this paper, we tested our model with various images, including benchmark datasets, i.e., CS [16] and Meylan et al. HDR [17,18] datasets, and self-images [19]. The results of several tone-mapping algorithms and the proposed method were compared to evaluate the performance of the proposed method. Low-light images for comparing tone compression in bright and dark areas, outdoor images for comparing detail compensation, and a color booth image suitable for the simultaneous comparison of color reproduction and tone compression of input images were used for the performance evaluation. The experiments and evaluation were implemented on Windows 10 Pro 64bit and an Intel(R) Core(TM) i7-7700 CPU @ 3.60 GHz, using Python 3.11.4 and MATLAB R2023a. Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 depict the comparison results for images with a wide range of brightness, from dark to bright areas. For existing methods, such as Reinhard [5] and L1L0, the tone increases across the brightness range, resulting in saturation in bright areas and reduced visibility in dark areas. Although iCAM06 and the method by Kwon et al [12]. ensure visibility in dark areas, color distortions, such as saturation, occur in some areas, and an overall reddish tone appears in the image. In contrast, the proposed method effectively compresses tones while maintaining visibility in dark areas without saturation in bright areas. Furthermore, detailed areas, such as tree trunks, leaves, tiles, and tire surfaces, display strong contrasts according to each area, effectively representing detailed information.
Figure 8. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 9. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 10. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 11. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 12. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 13. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 14. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 15 and Figure 16 present the comparison results for the night-time images. The existing methods often suffer from oversaturation, leading to saturation in some areas or decreased color contrast, causing overall haziness or reduced detail. Especially in Figure 16, areas like the sky lack accurate color reproduction and appear hazy. In contrast, the proposed method represents colors without oversaturation, maintaining a clear color contrast and accurately expressing details in the walls, patterns, and vegetation.
Figure 15. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 16. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.
Figure 17 illustrates the performance comparison of a color chart. The Reinhard and L1L0 methods effectively represent the primary colors of the color chart, but the achromatic colors exhibit saturation. In contrast, the iCAM06 and Kwon et al. methods suffer from oversaturation, resulting in an inaccurate color representation and excessive brightness in the image. In contrast, the proposed method accurately reproduces natural colors for both the primary and achromatic color charts while demonstrating great detailed representation in the buildings and floor tiles.
Figure 17. Input and result images: (a) Original, (b) Reinhard (2012) [5], (c) L1L0, (d) iCAM06, (e) Kwon et al. [12], and (f) proposed.

4.2. Objective Assessment

To conduct an objective evaluation, we performed six quantitative metrics: the local phase coherence–sharpness index (LPC_SI), cumulative probability blur detection (CPBD), maximum contrast with minimum artifacts (MCMA), tone-mapped image quality index (TMQI), natural image quality evaluator (NIQE), and perception-based image quality evaluator (PIQE) to compare image quality using 17 images. First, LPC_SI evaluates the sharpness of the image based on the weakening of the local phase coherence intensity through blur [20]. The CPBD evaluates the sharpness and blur level of the image based on a probabilistic model detecting blur at the edges from the changing contrast values [21]. In addition, MCMA is a metric similar to human perception, evaluating the visual quality enhancement and detail preservation by comparing the original and enhanced images [22]. Further, the TMQI is an evaluation metric for tone-mapped images, including a multiscale structural fidelity measure and a statistical naturalness measure [23]. The TMQI indicates the level of structural distortion and provides an overall quality score for the enhanced image compared to the original. These four metrics indicate better image quality and sharpness as their values increase. The PIQE is a no-reference evaluation method, estimating localized distortions in images and evaluating image quality through the local variance of distorted blocks [24]. Further, PIQE scores range from 0 to 100, with lower values indicating higher image quality. The NIQE provides a perceptual quality score for images based on a model computed from natural images compared to the image itself [25]. Similar to the PIQE, the NIQE scores range from 0 to 100, with lower values indicating higher image quality.
Figure 18 presents the results of the evaluation based on the scores for 17 comparison images and their averages. Table 1 represents the average values of the 17 experimental images in the specific numerical form.
Figure 18. Metric scores [5,12]: (a) local phase coherence-sharpness index (LPC_SI) score, (b) cumulative probability blur detection (CPBD) score, (c) maximum contrast with minimum artifact (MCMA) score, (d) tone-mapped image quality index (TMQI) score, (e) natural image quality evaluator (NIQE) score, and (f) perception-based image quality evaluator (PIQE) score.
Table 1. Comparison with metric scores.
While the rankings of each image may differ based on the evaluation metrics, the proposed method demonstrates excellent results across all evaluation metrics on average. Notably, the CPBD, MCMA, and TMQI metrics display significant performance improvement compared to other synthesis methods. Overall, the proposed method consistently ranks highly for all evaluation metrics without significant variation, unlike other synthesis methods that may exhibit considerable discrepancies in performance evaluation across metrics. Thus, the proposed method is outstanding in image quality performance regarding detail representation, noise reduction, and tone-compression effects compared to the existing synthesis methods.

5. Conclusions

Conventional tone-mapping algorithms can introduce unnecessary noise and halo artifacts during the tone-compression process. This paper proposes a tone-mapping algorithm that combines Retinex theory with CLAHE in a multiscale synthesis approach to resolve these problems and achieve better image quality. The proposed method enhances tone-compression performance and reduces noise and halo artifacts by performing high-σ SSR on images with a wide dynamic range. In addition, multiscale CLAHE secures the visibility and detailed information expression by synthesizing each image with exceptional detail representation and noise reduction. Finally, the algorithm enhances color representation by employing saturation compensation based on the color ratio of the original image to the synthesized image. Compared to the existing tone-mapping algorithms, such as MSR, Reinhard, L1L0, iCAM06, and the method by Kwon et al., the proposed algorithm has excellent tone-compression performance and increases local contrast, resulting in prominent clarity characteristics. Additionally, the color of the original image is well represented by minimizing the color-channel distortion due to the change in the luminance channel. In addition to subjective evaluations, objective image quality metrics confirm the superiority of the proposed method for various aspects of image quality. Thus, the proposed method outperforms existing methods in image sharpness, visibility, and overall quality performance.
Additionally, we extend the application of existing theories by employing them at a multiscale level. CLAHE has primarily been used with a single scale of two parameters. Two theories, Retinex and CLAHE, rely on their own parameters. As such, the degree of image improvement has a trade-off relationship depending on the parameter values. Furthermore, both theories do not have separate optimal parameter values and the values required for each image are different. Due to these problems, conventional methods have limitations in their application and effectiveness. By employing conventional theory at a multiscale level and integrating two theories which are complementary to each other, this paper overcomes the limitations imposed by trade-offs and enhances the resulting image quality in various ways. Ultimately, the proposed method presents a more dynamic approach to using existing theories and combines them to reproduce superior HDR images from single LDR images compared to traditional techniques.

Author Contributions

Conceptualization, S.-H.L.; methodology, S.-H.L. and Y.-J.K.; software, Y.-J.K.; validation, S.-H.L. and Y.-J.K.; formal analysis, S.-H.L. and Y.-J.K.; investigation, S.-H.L. and Y.-J.K.; resources, S.-H.L., D.-M.S., and Y.-J.K.; data curation, S.-H.L., D.-M.S., and Y.-J.K.; writing—original draft preparation, Y.-J.K.; writing—review and editing, S.-H.L.; visualization, Y.-J.K.; supervision, S.-H.L.; project administration, S.-H.L.; funding acquisition, S.-H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF), funded by the Ministry of Education, Korea (NRF-2021R1I1A3049604, 50%), and supported by the Innovative Human Resource Development for Local Intellectualization program through the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (IITP-2024-RS-2022-00156389, 50%).

Data Availability Statement

The data presented in this study are openly available in https://www.cs.cmu.edu/afs/cs/project/cil/ftp/html/v-images.html (accessed on 5 May 2024) in references [16], Meylan et al in references [17,18], and our self-images on GitHub in reference [19].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Land, E.; McCann, J.J. Lightness and Retinex theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef] [PubMed]
  2. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 2nd ed.; Prentice-Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
  3. Land, E. An Alternative Technique for The Computation of The Designator in The Retinex Theory of Color Vision. Proc. Natl. Acad. Sci. USA 1986, 83, 3078–3080. [Google Scholar] [CrossRef] [PubMed]
  4. Jobson, D.J.; Rahman, Z.-U.; Woodell, G.A. A Multiscale Retinex for Bridging the Gap Between Color Images and the Human Observation of Scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed]
  5. Reinhard, E.; Stark, M.; Shirley, P.; Ferwerda, J. Photographic Tone Reproduction for Digital Images. Semin. Graph. Pap. Push. Bound. 2002, 2, 661–670. [Google Scholar]
  6. Tumblin, J.; Turk, G. LCIS: A boundary hierarchy for detailpreserving contrast reduction. In Proceeding of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 83–90. [Google Scholar]
  7. Farbman, Z.; Fattal, R.; Lischinski, D.; Szeliski, R. Edge-preserving decompositions for multi-scale tone and detail manipulation. ACM Trans. Gr. 2008, 27, 67. [Google Scholar] [CrossRef]
  8. Durand, F.; Dorsey, J. Fast Bilateral Filtering for the Display of High-Dynamic-Range Images. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 21–26 July 2002. [Google Scholar]
  9. Kwon, H.J.; Lee, S.H.; Lee, G.Y.; Sohng, K.I. Enhanced high dynamic-range image rendering using a surround map based on edge-adaptive layer blurring. IET Comput. Vis. 2016, 10, 689–699. [Google Scholar] [CrossRef]
  10. Lee, G.Y.; Kwon, H.J.; Lee, S.H. HDR image reproduction based on visual achromatic response. Opt. Rev. 2020, 27, 361–374. [Google Scholar] [CrossRef]
  11. Kuang, J.; Johnson, G.M.; Fairchild, M.D. iCAM06: A refined image appearance model for HDR image rendering. J. Vis. Commun. Image Represent. 2007, 18, 406–414. [Google Scholar] [CrossRef]
  12. Kwon, H.J.; Lee, S.H. Contrast sensitivity based multiscale base-detail separation for enhanced HDR imaging. Appl. Sci. 2020, 10, 2513. [Google Scholar] [CrossRef]
  13. Jobson, D.J.; Rahman, Z.; Woodell, G.A. Properties and Performance of A Center/surround Retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef] [PubMed]
  14. Gonzalez, R.; Wood, R. Digital Image Processing, 3rd ed.; Pearson Education: London, UK, 2009. [Google Scholar]
  15. Zuiderveld, K. Contrast limited adaptive histogram equalization. Graphics Gems 1994, 8, 474–485. [Google Scholar]
  16. Computer Vision Test Images. Available online: https://www.cs.cmu.edu/afs/cs/project/cil/ftp/html/v-images.html (accessed on 5 May 2024).
  17. Meylan, L.; Susstrunk, S. High dynamic range image rendering with a retinex-based adaptive filter. IEEE Trans. Image Process. 2006, 15, 2820–2830. [Google Scholar] [CrossRef] [PubMed]
  18. Meylan, L. Tone Mapping for High Dynamic Range Images. Ph.D. Thesis, École Polytechnique Fédérale de Lausanne, Lausanne, Switzerland, 2016. [Google Scholar] [CrossRef]
  19. YuJoong/Retinex-Jointed-Multiscale-CLAHE. Available online: https://github.com/YuJoong/Retinex-jointed-multiscale-CLAHE/tree/main/self-image (accessed on 10 May 2024).
  20. Hassen, R.; Wang, Z.; Salama, M.M.A. Image sharpness assessment based on local phase coherence. IEEE Trans. Image Process. 2013, 22, 2798–2810. [Google Scholar] [CrossRef] [PubMed]
  21. Narvekar, N.D.; Karam, L.J. A no-reference perceptual image sharpness metric based on a cumulative probability of blur detection. In Proceedings of the 2009 International Workshop on Quality of Multimedia Experience, San Diego, CA, USA, 29–31 July 2009; pp. 87–91. [Google Scholar]
  22. Abdoli, M.; Nasiri, F.; Brault, P.; Ghanbari, M. A Quality Assessment tool for Performance Measurement of Image Contrast Enhancement Methods. IET Image Process. 2019, 13, 833–842. [Google Scholar] [CrossRef]
  23. Yeganeh, H.; Wang, Z. Objective quality assessment of tone-mapped images. IEEE Trans. Image Process. 2013, 22, 657–667. [Google Scholar] [CrossRef] [PubMed]
  24. Venkatanath, N.; Praneeth, D.; Bh, M.C.; Channappayya, S.S.; Medasani, S.S. Blind image quality evaluation using perception based features. In Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Mumbai, India, 1 March 2015; pp. 1–6. [Google Scholar]
  25. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.