Next Article in Journal
Radiological Diagnosis and Advances in Imaging of Vertebral Compression Fractures
Next Article in Special Issue
Optimizing Digital Image Quality for Improved Skin Cancer Detection
Previous Article in Journal
The Role of Plain Radiography in Assessing Aborted Foetal Musculoskeletal Anomalies in Everyday Practice
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Overview of High-Dynamic-Range Image Quality Assessment

by
Yue Liu
1,†,
Yu Tian
1,†,
Shiqi Wang
1,
Xinfeng Zhang
2 and
Sam Kwong
3,*
1
Department of Computer Science, City University of Hong Kong, Hong Kong, China
2
School of Computer Science and Technology, University of Chinese Academic of Sciences, Beijing 100190, China
3
School of Data Sciences, Lingnan University, Hong Kong, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(10), 243; https://doi.org/10.3390/jimaging10100243
Submission received: 5 August 2024 / Revised: 21 September 2024 / Accepted: 23 September 2024 / Published: 27 September 2024
(This article belongs to the Special Issue Novel Approaches to Image Quality Assessment)

Abstract

:
In recent years, the High-Dynamic-Range (HDR) image has gained widespread popularity across various domains, such as the security, multimedia, and biomedical fields, owing to its ability to deliver an authentic visual experience. However, the extensive dynamic range and rich detail in HDR images present challenges in assessing their quality. Therefore, current efforts involve constructing subjective databases and proposing objective quality assessment metrics to achieve an efficient HDR Image Quality Assessment (IQA). Recognizing the absence of a systematic overview of these approaches, this paper provides a comprehensive survey of both subjective and objective HDR IQA methods. Specifically, we review 7 subjective HDR IQA databases and 12 objective HDR IQA metrics. In addition, we conduct a statistical analysis of 9 IQA algorithms, incorporating 3 perceptual mapping functions. Our findings highlight two main areas for improvement. Firstly, the size and diversity of HDR IQA subjective databases should be significantly increased, encompassing a broader range of distortion types. Secondly, objective quality assessment algorithms need to identify more generalizable perceptual mapping approaches and feature extraction methods to enhance their robustness and applicability. Furthermore, this paper aims to serve as a valuable resource for researchers by discussing the limitations of current methodologies and potential research directions in the future.

1. Introduction

Compared to ordinary Low-Dynamic-Range (LDR) images, High-Dynamic-Range (HDR) images exhibit a wider dynamic range (i.e., a larger ratio between the highest and lowest value in the image), resulting in superior overall contrast effects and richer detail, especially in the extremely bright or dark regions. Owing to their ability to offer highly realistic visual experiences to viewers, recent years have witnessed a significant increase in the research of HDR images, including photography [1], automotive [2], medical imaging [3], and remote sensing [4]. In this context, the HDR Image Quality Assessment (IQA) becomes a fundamental topic in the field of HDR image processing tasks, which can be widely employed for optimizing parameters in various codecs and learning-based models. However, it is not suitable to directly apply LDR IQA metrics to HDR images, due to the substantial disparities in statistical distribution and luminance range between LDR and HDR images. Therefore, specialized evaluation methods and metrics tailored for HDR images are essential to assess their quality effectively.
In general, existing IQA algorithms can be classified into subjective and objective methods according to the evaluation approach. With the explosive growth of HDR content, the pressing issue at hand pertains to the storage and transmission of HDR images. Consequently, existing subjective experiments [5,6,7] are dedicated to exploring the impact of compression algorithms on the visual quality of HDR images. Nevertheless, the subjective experiments are typically time-consuming and expensive. In contrast, objective methods are considered more efficient due to their ability to directly extract features that closely align with the human visual system (HVS) from the images. Specifically, the HDR-VDP family [8,9,10] models the luminance perceived by HVS. Some methods evaluate the quality of HDR images in the perceptual domain [11,12], while others adapt existing metrics to predict quality scores [13,14]. Despite great progress, HDR IQA still encounters some challenges, particularly in the lack of large subjective databases and the difficulty of properly compressing the dynamic range of the HDR image into the perceptual domain.
The necessity for a comprehensive survey of HDR IQA methods arises from the limitations of existing surveys, which fail to cover the scope of the field comprehensively or lack summaries of the latest methodologies [15]. Therefore, this paper aims to fill this gap by providing an extensive overview of existing HDR IQA algorithms, including both subjective and objective methods. Specifically, our statistical evaluation, which includes nine LDR IQA metrics and three perceptual mapping methods, demonstrates that the choice of mapping method and dataset characteristics substantially influence performance outcomes. Furthermore, we discuss the limitations of current subjective and objective HDR IQA approaches and potential directions for improvement, hoping to provide comprehensive insight for future research.
The structure of this paper is outlined as follows: Section 2 introduces existing subjective HDR IQA databases, including detailed descriptions of the database and experimental settings. Section 3 summarizes existing objective HDR IQA algorithms. The limitations of both subjective and objective HDR IQA approaches are discussed in Section 4, together with the potential directions for future research.

2. Subjective HDR Image Quality Assessment Databases

Existing subjective HDR IQA databases focus on the evaluation of the visual quality of HDR images with compression distortions. Therefore, we first introduce the existing HDR image compression schemes in Section 2.1 to provide some basic information that can help to understand the construction and experimental settings in the subjective databases. Subsequently, the details of the databases are described in Section 2.2.

2.1. Existing HDR Image Compression Scheme

In contrast to LDR images that are typically encoded in 8-bit integers, the most commonly used formats for HDR images are Radiance RGBE and OpenEXR, where the data are stored in 16-bit or 32-bit floating-point numbers. Therefore, traditional image compression schemes like JPEG and JPEG2000, designed for LDR images, can not handle the compression of HDR images directly. Due to the widespread application of LDR codecs, converting the HDR image to a LDR image to make it compatible with existing codecs is a highly intuitive approach with practical applications. As shown in Figure 1, the existing compression scheme [16,17] for HDR images can be summarized into three steps, detailed as follows: In the pre-processing stage, the input HDR image is converted to an LDR image by a Tone-Mapping Operator (TMO). Herein, Sugiyama et al. propose to improve the quality of the tone-mapped LDR image by an optimization module [17]. Afterward, the LDR image is fed into the codec (JPEG or JPEG2000), together with the additional information such as the parameters for inverse TMO (iTMO), or the different information between the HDR and corresponding LDR image. In the post-preprocessing stage, the content to be displayed varies according to the conditions of the terminal monitor. If the monitor supports HDR content, the HDR image will be generated from the decoded LDR image by the iTMO with additional information. Otherwise, the decoded LDR image can be directly displayed. Based on the previous research works on HDR image compression with JPEG and JPEG2000, a new, dedicated compression standard, designed for HDR images, is proposed by the JPEG committee, called JPEG-XT [18], where the LDR image is compressed in the base layer, and the overall dynamic range is provided in the extension layer. The latest video coding scheme, Versatile Video Coding (VVC) [19], also supports the compression of the HDR content, where the luma mapping with chroma scaling and luma-adaptive deblocking [20] are proposed to improve the performance of the visual quality of the compressed HDR content.
With such architecture, the visual quality of the compressed HDR images has been evaluated in several subjective databases. Narwaria et al. built two databases to evaluate the influence of the optimization methods [5] in the pre-processing stage and the effectiveness of different TMOs [21] under the JPEG and JPEG2000 scheme, respectively. Korsunov et al. [6] compressed the HDR images using the JPEG-XT scheme. Valenzise et al. [7] constructed a new database including the JPEG, JPEG2000, and JPEG-XT schemes to evaluate the performance of the objective HDR IQA algorithms. Afterward, Zerman et al. [22] extended the content in [7] and add a new TMO method, Perceptual Quantization (PQ) [23], in the proposed database. Recently, Mikhailiuk et al. [24] constructed a large HDR IQA database by combining two existing HDR IQA databases [5,6], where the MOS values are aligned by the psychometric scaling method [25]. Liu et al. [26] first proposed a large database to evaluate the performance of various HDR IQA metrics on JPEG-XT and the newest compression scheme VVC. More details of the databases are summarized in Table 1 and described as follows.

2.2. Existing HDR Databases for Quality Assessment

To evaluate the performance of (1) optimization methods in HDR compression and (2) the objective IQA algorithms for compressed HDR images, Narwaria et al. generated a database [5] and conducted a subjective experiment to obtain the corresponding mean opinion score (MOS). The database is composed of 10 reference HDR images and 140 distorted HDR images, which are compressed with the JPEG scheme with seven bitrates and two optimization methods. The Mean Square Error (MSE) and Structural Similarity Index (SSIM) [27] are used to optimize the quality of the generated LDR image with the iCAM06 [28] TMO. The quality scores of the 150 HDR images are collected from 27 observers using the absolute category rating with hidden reference (ACR-HR) methodology [29]. The subjective experiment is conducted in a room with 130 cd/m2 luminance condition. The HDR images are displayed on a SIM2 Solar47 HDR monitor [30] with a maximum luminance value of up to 4000 cd/m2. From the experimental results, the SSIM does not outperform MSE when they are used to optimize the quality of tone-mapped LDR images. Furthermore, the HDR-VDP-2 [9] shows better performance on the overall results. However, when it comes to statistical analysis, the HDR-VDP-2 is indistinguishable from the SIQM method [31].
Table 1. An overview of existing HDR IQA databases. The observers, reference images, distorted images, distortion type, and subjective methodology are denoted as Obs., Ref., Dis., Dist., and Meth., respectively.
Table 1. An overview of existing HDR IQA databases. The observers, reference images, distorted images, distortion type, and subjective methodology are denoted as Obs., Ref., Dis., Dist., and Meth., respectively.
Database#Obs.#Ref.#Dis.Dist.TMOMeth.
Narwaria2013 [5]2710140JPEGiCAM06ACR-HR
Narwaria2014 [21]296210JPEG2000AL, Dur
RG, RL
Log
ACR-HR
Korshunov2015 [6]2420240JPEG-XTRG
MT [32]
DSIS
Valenzise2014 [7]15550JPEG
JPEG2000
JPEG-XT
MaiDSIS
Zerman2017 [22]15550JPEG
JPEG2000
Mai, PQDSIS
UPIQ [24]
(HDR part)
2030380JPEG
JPEG-XT
iCAM06
RG
MT [32]
-
HDRC [26]2080400JPEG-XT
VVC
RG
PQ
DSIS
In the database proposed by Narwaria et al. [21], the performance of five TMOs under the JPEG2000 compression scheme is subjectively evaluated. Herein, three local TMOs proposed by Reinhard (RL) [33], Durand (Dur) [34], and Ashikmin (AL) [35], together with two global TMOs, logarithmic TMO (Log) and the global TMO (RG), proposed in [33], are applied in the experiments. In this database, six reference HDR scenes are selected to obtain 210 distorted HDR images with 7-bit rates for each TMO. The subjective experiment settings are consistent with the conditions in [5]. The MOS values of the 216 HDR images are obtained from 29 participants with ACR-HR methodology. The experimental analysis results reveal that the AL TMO statistically outperforms the other TMOs. Furthermore, the local TMOs can help to achieve better visual quality of the compressed HDR images, while the performance of the two global TMOs does not show statistical differences.
Korsunov et al. [6] compressed the HDR image with the JPEG-XT [36] compression scheme and conducted a subjective experiment to build an HDR IQA database. In this database, 20 reference HDR images with various content are cropped and adjusted to the size of 944 × 1080 for display. A total of 240 compressed HDR images are generated by applying three profiles and 4-bit rates for each scene. The subjective experiment is conducted in a room with 20 cd/m2 background luminance value. A SIM2 HDR monitor with luminance values varying from 0.001 to 4000 cd/m2 is adopted to display the HDR images. The double-stimulus impairment scale (DSIS) methodology [37] is applied to acquire the MOS values from 24 subjects.
Valenzise et al. [7] have developed a database specifically for evaluating full-reference HDR IQA metrics. The database is composed of 5 reference HDR images and 50 distorted HDR images, compressed using the JPEG, JPEG2000, and JPEG XT schemes. The TMO used in this database is proposed by Mai et al. [38]. The subjective experiment is conducted in a room with a background luminance value of 20 cd/m2. The HDR images are presented on a SIM2 HDR47 monitor. The MOS scores are collected by the DSIS methodology from 15 subjects, where the continuous quality scores from 0 to 100 are divided into five scales, denoting the distortion level from “very annoying" to “Imperceptible”. Based on this database, they evaluate the performance of the objective IQA methods, including HDR-VDP-2 [9] tailored for HDR images, alongside PSNR and SSIM, which were adapted for the assessment of perceptual mapped HDR images by logarithmic function or perceptual uniform (PU) encoding [39]. The results show that PU-SSIM outperforms other metrics, followed by log-SSIM and HDR-VDP-2.
In order to conduct a more comprehensive assessment of the performance of full-reference HDR IQA metrics, Zerman et al. [22] introduced a novel database, building upon their previous work [7]. This database comprises 5 reference HDR images and 50 HDR images distorted by the JPEG and JPEG2000 compression schemes. Herein, two TMOs are employed to generate the LDR images: the TMO proposed by Mai et al. [32], and the Perceptual Quantizer (PQ) [23] published in SMPTE ST 2084. The subjective experiment settings are controlled to be the same as in Valenzise2014 [7]. Subsequently, the obtained MOS values are further aligned with the MOS values in [7] for further analysis. In the experiment, 25 IQA metrics are evaluated on the proposed database and four existing databases. There are three interesting conclusions that can be drawn from the results. Firstly, the Narwaria2014 [21] is a much more challenging database due to its complex distortion types. Secondly, metrics tailored for HDR content, such as HDR-VDP-2.2 [40] and HDR-VQM [11], exhibit superior performance in comparison to other IQA metrics. Nevertheless, the probability of misclassification still remains as high as 20%, indicating the imperative for the design of HDR IQA metrics that better align with HVS. Furthermore, concerning compression distortions, structural loss in the luminance domain is more perceptible to the HVS than loss in the color space.
Due to the limited number of images in the previous HDR IQA databases, Mikhailiuk et al. [24] constructed a Unified Photometric Image Quality (UPIQ) database with 380 HDR images and 3779 LDR images. The HDR images are collected from the two databases Korsunov2015 [6] and Narwaria2013 [5]. The MOS values are aligned by the psychometric scaling methodology, where the rating and ranking data are combined to capture the true underlying quality of images. Specifically, the rating scores, such as MOS, are linearly mapped to a psychometric scale, while pairwise comparisons are analyzed to determine the probability of one image being rated higher than another. The unified scale was achieved by performing additional subjective cross-dataset comparisons, which were then scaled using Thurstone’s model [41] to ensure that all datasets adhered to a common quality scale. This method enhances precision and comparability across datasets that use different scoring methods, ensuring that scores from different experimental protocols can be directly compared.
Recently, Liu et al. [26] proposed an HDR compression (HRDC) database, with the goal of examining the perceptual quality of HDR images compressed using the JPEG-XT and VVC codecs. The experiment is conducted using a Samsung QN90A TV, with 20 participants providing opinion scores based on the DSIS testing methodology. This database is the largest open-source database available for the HDR IQA task, including 400 distorted HDR images with JPEG-XT and VVC compression and 80 reference HDR images. Based on this database, they conduct extensive experiments to evaluate the effectiveness of both HDR and LDR IQA metrics on the HDR IQA task. Furthermore, they also evaluate the effects of different perceptual mapping methods, such as PQ, PU, and PU21 mapping methods, on the performance of LDR IQA metrics. Based on the experimental results, these perceptual mapping methods exhibit significant variations in performance across different databases. Moreover, no single perceptual mapping method achieves optimal performance across all databases, underscoring the challenges posed by HDR image evaluation and suggesting the need for new mapping methods to achieve stable and effective quality assessment for HDR images.

3. Objective HDR Image Quality Assessment Methods

Recent years have witnessed significant developments of efficient objective LDR IQA metrics [42,43,44,45,46,47,48,49], which can serve as crucial constraints for various high-level tasks [50,51]. Since both HDR and LDR images are 2D visual data, can we directly use LDR IQA metrics to evaluate HDR image quality? To answer this, it is crucial to first identify the differences between HDR and LDR images. In HDR images, pixel values are linearly related to the natural luminance. On the contrary, during the capture process of LDR images (as shown in Figure 2), the camera sensors first involve a clipping operation to restrict the maximum value of the image. Consequently, the nonlinear mapping process compresses the dynamic range of the image using the Camera Response Functions (CRFs). Finally, the quantization step aims to alleviate the storage burden on cameras. In Figure 3, we provide a more intuitive way to illustrate the statistical differences between the HDR and LDR images. It is evident that the HDR image has a wider dynamic range. Furthermore, there are over-exposed regions where many pixel values are 255 in the LDR image. However, there are no such regions in the HDR image. Therefore, directly applying IQA metrics specifically designed for LDR images to assess the quality of HDR images is unreasonable due to the different data distributions between HDR and LDR images. In Zerman et al. [22] and Liu et al. [26]’s works, extensive experimental results also demonstrate a significant decline in the performance of the LDR IQA algorithms when they are directly applied to HDR images without any perceptual mapping functions.
Consequently, the effective evaluation of the quality of the HDR images remains a challenging task. Existing HDR IQA methods can be roughly divided into the following three categories according to the design strategies: (1) HVS-modeling-based HDR IQA methods: directly evaluating the image quality by modeling how the human eye perceives and processes the light signals from natural scenes; (2) existing IQA-metrics-based HDR IQA methods: such algorithms aim to make full use of existing HDR or LDR IQA metrics to effectively assess the quality of HDR images; (3) PU-encoding-based HDR IQA methods: assessing the quality of HDR images specifically within the PU domain. More details of the existing HDR IQA methods are summarized in Figure 4, Table 2 and described in Section 3.1 and Section 3.2, respectively.

3.1. HVS-Modeling-Based HDR IQA Methods

The HDR-VDP family comprises a widely used and efficient set of HDR IQA algorithms, where the first version, HDR-VDP-1 [8], is proposed by Mantiuk et al., followed by the improved HDR-VDP-2 [9], HDR-VDP-2.2 [40], and the latest HDR-VDP-3 [10]. The basic idea of the HDR-VDP family is to model the HVS. Building on the Visual Difference Predictor (VDP) [53], HDR-VDP-1 first models the HVS within three stages: (1) modeling the light scattering using the Optical Transfer Function (OTF) [54]; (2) modeling the nonlinear response of the photoreceptors by the Just-Noticeable Difference (JND); (3) removing the less-sensitive frequency according to the Contrast-Sensitive Function (CSF). Subsequently, the extracted visual features are decomposed and then fed into a pooling module to calculate the visual difference map. Herein, no single quality score is provided in HDR-VDP-1. In HDR-VDP-2, the multi-band pooling strategy is proposed to obtain the quality score of the distorted HDR image. Additionally, photoreceptor response is modeled using sensitivity functions [55] rather than JND. In the pooling stage, the weighting parameters are optimized using the LIVE database [56] due to the lack of HDR IQA databases. Narwaria et al. improved the performance of HDR-VDP-2 by optimizing the parameters using two HDR databases [5,21] in HDR-VDP-2.2. Recently, the HDR-VDP-3 metric has been proposed, where the impact of different ages on visual quality is considered, together with the modeling of the local luminance. Additionally, the UPIQ database is used for parameter calibration to enable the quality evaluation of both HDR and LDR images. Despite significant success in HDR IQA, the HDR-VDP family still presents aspects worthy of improvement. Firstly, the parameters involved in simulating the HVS and the pooling stage necessitate intricate calibration processes [7]. Secondly, based on the experimental results in [5], HDR-VDP-2 appears to over-penalize the distortions for HDR images with higher quality.
Aydin et al. [57] proposed a Dynamic Range Independent Metric (DRIM) to measure the visual difference between two images with arbitrary dynamic range. Herein, they first model the HVS to extract the contrast features. Afterward, the cortex transform [58] is applied to obtain three distortion types separately, which are later used to generate the final visual difference map. Similar to HDR-VDP-1, this method tends to predict the probability of the difference between two images, yet it lacks the quantitative evaluation result, such as the MOS value. Given the same distorted image, Aydin et al. [39] found that the perceived distortion is more pronounced when viewed on an HDR display compared to an LDR display. Motivated by the above observation, Kottayil et al. [59] proposed a Deep Neural Network (DNN)-based blind IQA for HDR images. The network is composed of an E-Net for error prediction and a P-Net to emulate the perceptual resilience of the HVS, thereby quantifying the visibility of discrepancies within image blocks. The network is trained on the database that combines five HDR databases [5,6,7,21,22], where the MOS values are aligned by the Iterated Nested Least Square Algorithm (INLSA) [60]. Despite employing 5-fold cross-validation, the fact remains that only 552 distorted images are utilized for training in each fold, which is notably insufficient for a DNN model.
Figure 4. Overview of objective HDR IQA methods. The HDR images presented in this paper are tone-mapped for visualization by Gamma correction.
Figure 4. Overview of objective HDR IQA methods. The HDR images presented in this paper are tone-mapped for visualization by Gamma correction.
Jimaging 10 00243 g004

3.2. PU-Encoding-Based HDR IQA Methods

PU encoding, introduced by Aydin et al. [39], aims to convert HDR images into a perceptual domain, enabling the HDR image quality evaluation by LDR IQA metrics, which has been proven to be an effective mapping method [22]. Therefore, existing PU-encoding-based HDR IQA methods apply PU encoding first, followed by the quality evaluation in the PU domain. HDR video quality metric (HDR-VQM) [11] is the first HDR video quality assessment metric, and its efficacy in evaluating the quality of HDR images has been substantiated through subjective experiments [12,22,24]. In this metric, the PU encoding is first applied to obtain the perceived luminance map. Subsequently, the perceptual error is calculated by the log-Gabor [61] filter. The final quality score of the video is predicted by a spatial and temporal pooling strategy. The network proposed by Jia et al. [62], involved designing a Convolutional Neural Network (CNN) for blind HDR IQA. In the experiment, the network trained on PU-encoded HDR images shows significant improvement compared with the network trained on original HDR images. The PU-PieAPP proposed by Mikhailiuk et al. [24] combines the PU encoding with the PieAPP [63] for the quality evaluation of both HDR and LDR images. The network is trained on the UPIQ database [24], including 80 HDR images and 3779 LDR images. Liu et al. [12] introduce the local–global frequency feature-based model (LGFM). Herein, the odd log-Gabor filter is applied to obtain the local frequency feature, while the Butterworth filter is employed to simulate the CSF to generate the global frequency feature. The two feature maps undergo separate pooling processes and are subsequently integrated to yield the perceptual quality score.
Table 2. Overview of existing HDR IQA methods.
Table 2. Overview of existing HDR IQA methods.
MethodYearClassificationMajor Category
HDR-VDP-1 [8]2005Full-ReferenceHVS-modeling-based HDR IQA methods
DRIM [57]2008Full-ReferenceHVS-modeling-based HDR IQA methods
HDR-VDP-2 [9]2011Full-ReferenceHVS-modeling-based HDR IQA methods
HDR-VDP-2.2 [40]2015Full-ReferenceHVS-modeling-based HDR IQA methods
HDR-VQM [11]2015Full-ReferencePU-encoding-based HDR IQA methods
Kottayil et al. [59]2017No-ReferenceHVS-modeling-based HDR IQA methods
Jia et al. [62]2017No-ReferencePU-encoding-based HDR IQA methods
HDR-CQM [13]2018Full-ReferenceExisting IQA metrics based HDR IQA methods
PU-PieAPP [24]2021Full-ReferencePU-encoding-based HDR IQA methods
HDR-VDP-3 [10]2023Full-ReferenceHVS-modeling-based HDR IQA methods
LGFM [12]2023Full-ReferencePU-encoding-based HDR IQA methods
Cao et al. [14]2024Full-ReferenceExisting IQA-metrics-based HDR IQA methods

3.3. Existing IQA-Metrics-Based HDR IQA Methods

Currently, there is a limited number of HDR IQA algorithms, whereas there are numerous efficient LDR IQA methods. Therefore, it is common practice to directly employ existing LDR IQA methods for quality evaluation after mapping HDR images to the perceptual domain. In the early stages, Aydin et al. [39] introduced the PU encoding to convert HDR images into the perceptual domain. Recently, Mantiuk and Azimi proposed the PU21 [64] to model the new perceptual uniform space based on a new CSF [65]. Additionally, PQ [23] and Hybrid-Log–Gamma (HLG) [66] are two commonly used transfer functions in codecs, aiming to provide the optimal visualization results of the HDR content on the monitors. Note that PQ and HLG are not dedicated designed for HDR IQA, but they can also be adopted as the perceptual mapping functions. Although these mapping functions can effectively convert the HDR image into the perceptual domain, they are usually non-differentiable, making them unsuitable for the optimization of the learning-based model. Currently, there are four widely used differentiable mapping functions, including the logarithm function, Gamma correction, hyperbolic tangent function (tanh), and μ -law, which can be used to optimize the learning-based models [67,68,69]. To be specific, the logarithm function and Gamma correction model the nonlinear perception of light in the HVS, which are derived from the Weber–Fechner law and Stevens’ power law [70], respectively. Tanh is an extensively employed nonlinear function that can compress the dynamic range of the image into [ 1 , 1 ] . The effectiveness of the μ -law algorithm has been demonstrated in various HDR-related tasks [71,72] for dynamic range compression, which is a differentiable companding algorithm originally designed for audio signals.
In addition, Choudhury and Daly proposed to evaluate the quality of HDR images by combining the results of various HDR IQA metrics (HDR-CQM) [13]. To be specific, they first selected six HDR IQA metrics with top performance according to the experimental results in [73]. Subsequently, the Sequential Floating Forward Selection (SFFS) [74] algorithm is applied to find the combination of metrics with the best performance. Afterward, the Support Vector Machine (SVM) is trained to predict the final quality score. Cao et al. [14] recently designed an inverse display model to convert the HDR image into the LDR domain to improve the performance of existing LDR metrics. Specifically, given an input HDR image, a set of LDR images with different exposures is generated by the proposed inverse display model. Afterwards, an LDR IQA metric is used to evaluate the quality of the image set. In addition, the LDR image stacks are also used to obtain the local exposure weights. The final quality score of the HDR image is obtained after the weighting pooling stage.

4. Discussion and Conclusions

This paper provides a comprehensive overview of existing subjective HDR IQA databases and objective HDR IQA metrics. While this research represents significant advancements, HDR IQA is still a challenging task, which is evolving alongside the generation of HDR content and the development of compression schemes. Therefore, this section further discusses the limitations of existing methods, along with the potential future work, aiming to provide a comprehensive understanding of the challenges and opportunities in the field of HDR IQA.

4.1. Subjective HDR IQA Databases

For the construction of HDR IQA subjective databases, we can observe three main limitations. (1) There is a limited number of both reference and distorted HDR images in existing databases. The content diversity of a database is crucial, as it not only affects the generalizability of the objective IQA algorithm evaluated on it but also impacts the robustness of the DNNs. The DNNs, owing to their powerful feature extraction capabilities, have been applied in numerous IQA models. However, this necessitates a large amount of training data for support. Unfortunately, the current HDR IQA databases are insufficient to train generalized deep learning-based HDR IQA models. (2) The databases currently focus primarily on limited compression distortions. As compression technologies evolve, commonly used compression techniques for HDR images have transitioned from JPEG and JPEG2000 to JPEG-XT and VVC [19]. Consequently, research on the effects of new compression distortion types on the human visual system remains insufficient. Moreover, there is a lack of research on other types of distortions, including motion distortions that may occur during the capture process, noise, and contrast distortions that may arise during the display process. (3) All subjective experiments in existing HDR databases are conducted based on professional HDR monitors. While these monitors offer high brightness levels, they are relatively costly and have limited application scopes. In contrast, currently widely used commercial HDR displays, although unable to achieve the brightness levels of professional monitors, still provide satisfactory visual experiences. Therefore, there is a need for subjective experiments on commercial displays to more accurately cater to the needs of a wider range of application scenarios. Based on the insights above, generating a larger database, exploring the effects of various distortion types, as well as conducting subjective experiments on commercial HDR displays can be potential research directions in the subjective HDR IQA database construction.

4.2. Objective HDR IQA Metrics

Despite the proliferation of HDR content, there are relatively few algorithms specifically designed for HDR image quality evaluation. Therefore, it is still a valuable undertaking to design robust, reliable, and perceptually aligned mapping methods that can map the brightness and color of the natural world to the perceptual domain that can be better aligned with HVS. For the HVS-modeling-based approaches, the current challenge lies in effectively modeling the HVS. HVS modeling presents significant complexity due to the large number of parameters involved and its reliance on experimental data, which are influenced by numerous factors. For instance, recent updates to the CSF by Ashraf et al. [75] highlight the importance of color, luminance, area, spatiotemporal frequency, and eccentricity as critical components of CSF modeling, which is an important part when modeling HVS. Such factors introduce considerable variability, as they are sensitive to both individual differences and environmental conditions. The fundamental issue lies in how to unify these multiple factors into a cohesive framework that can simplify the evaluation process while maintaining the accuracy of HVS-based predictions. For the PU-encoding-based and existing-IQA-metrics-based approaches, there are mainly two factors that influence the effectiveness of the metrics, perceptual mapping functions, and the feature extraction method. Herein, we conducted a comprehensive statistical analysis using the left-tail F-test, as suggested in [76], across five publicly available databases, to evaluate the performance of three perceptual mapping methods (Gamma, PQ, PU21) combined with nine LDR IQA metrics (PSNR, SSIM [27], MS-SSIM [77], FSIM [78], GMSD [76], VIF [79], ESIM [80], GFM [81], and DISTS [43]). As shown in Figure 5, cells highlighted in green (with a value of 1) indicate that the model corresponding to the row significantly outperforms the model corresponding to the column, while a value of 0 indicates no significant difference. From these results, we can draw the following key observations: (1) The choice of perceptual mapping method affects the accuracy of IQA model evaluations. When different mapping methods are applied to these IQA metrics, the performance ranking of IQA models can change. For instance, on the HDRC database, FSIM performs better than GFM when using the Gamma mapping, but there is no significant difference when using PQ or PU21. This indicates that perceptual mapping methods have a considerable impact on how well IQA models align with subjective quality assessments, highlighting the importance of selecting an appropriate mapping strategy for accurate evaluations. This finding aligns with the experimental results observed in the HDRC database [26]. Furthermore, through additional analysis of the statistical data, we identified a new observation that (2) existing IQA models show varying performance across databases with different data distributions. When using the same perceptual mapping method to assess the IQA models on different databases, certain models exhibit significant performance differences. For example, Gamma-DISTS performs better than Gamma-GMSD on the HDRC and Narwaria2013 datasets, whereas no significant difference is observed on the other three databases. A possible explanation is that the same perceptual mapping method may project different databases into distinct feature spaces compared with their original data distributions. Some IQA algorithms may be better suited to one type of feature space, while others perform more effectively in different feature spaces. From these experiments, we can conclude two limitations of existing HDR IQA models, as follows: (1) the sensitivity of IQA models to the characteristics of different datasets, and (2) the influence of perceptual mapping methods on the consistency of quality evaluations across databases. These findings suggest the need for further optimization and validation of IQA models and perceptual mapping methods for HDR content.
Accurate IQA models are foundational for a wide range of tasks and applications. For instance, in super-resolution, robust IQA models [82] ensure that the enhanced images maintain high visual quality and fidelity, reflecting the original content. In object detection, effective IQA metrics [83] help improve the clarity and precision of detected objects, contributing to more reliable outcomes in complex scenes. Additionally, in image compression, accurate IQA models [84] are critical for evaluating and optimizing compression algorithms to preserve image quality while reducing file size. Beyond these areas, IQA models are integral to various other applications, such as image enhancement, medical imaging, and virtual reality, where they play a crucial role in assessing and improving the quality of visual content. To develop more accurate objective quality models for HDR images, future work can focus on the following aspects: (1) Pre-processing. A more perceptually aligned mapping model could be designed to enhance the consistency between the tone-mapped image and the human visual system (HVS). This would ensure that the processed images better reflect human perception. (2) Feature extraction. Existing deep learning models can be further optimized for the more efficient extraction of visual features. For instance, transformers [85], specifically vision transformers (ViTs) [86], have shown great potential in capturing global dependencies in images [87,88]. ViTs work by splitting the image into patches, treating each patch as a token, and then applying self-attention mechanisms to model the relationships between different regions of the image. This ability to capture long-range dependencies makes ViTs particularly suitable for IQA, as they can effectively represent subtle perceptual differences. (3) Learning strategy. Considering the significance of a subjective visual quality assessment, a promising approach to improve model accuracy could be the adoption of a Human-in-the-Loop strategy [83], which involves incorporating human feedback into the training loop, allowing the model to adjust based on real-time subjective input. This approach enhances the model’s ability to align with human perception and improve generalization to unseen data. To sum up, advancing HDR IQA models requires attention to pre-processing, feature extraction, and innovative training strategies. By focusing on these areas, we can build models that have a higher consistency with the HVS, thereby providing more accurate quality assessment methods for a broader range of image processing tasks.

Author Contributions

Conceptualization, Y.L. and Y.T.; investigation, Y.L.; resources, Y.T.; data curation, Y.L.; writing—original draft preparation, Y.L. and Y.T.; writing—review and editing, S.W. and X.Z.; visualization, Y.L. and Y.T.; supervision, S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HDRHigh Dynamic Range
LDRLow Dynamic Range
HVSHuman Visual System
IQAImage Quality Assessment
TMOTone-Mapping Operator
DSISDouble-Stimulus Impairment Scale
ACR-HRAbsolute Category Rating with Hidden Reference
CSFContrast-Sensitive Function
PUPerceptual Uniform
PQPerceptual Quantizer
HLGHybrid-Log-Gamma
MOSMean Opinion Score
VDPVisual Difference Predictor
OTFOptical Transfer Function
DNNDeep Neural Network

References

  1. Metzler, C.A.; Ikoma, H.; Peng, Y.; Wetzstein, G. Deep optics for single-shot high-dynamic-range imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1375–1385. [Google Scholar]
  2. Shopovska, I.; Stojkovic, A.; Aelterman, J.; Van Hamme, D.; Philips, W. High-Dynamic-Range Tone Mapping in Intelligent Automotive Systems. Sensors 2023, 23, 5767. [Google Scholar] [CrossRef] [PubMed]
  3. Sánchez, D.; Gómez, S.; Mauricio, J.; Freixas, L.; Sanuy, A.; Guixé, G.; López, A.; Manera, R.; Marín, J.; Pérez, J.M.; et al. HRFlexToT: A high dynamic range ASIC for time-of-flight positron emission tomography. IEEE Trans. Radiat. Plasma Med. Sci. 2021, 6, 51–67. [Google Scholar] [CrossRef]
  4. Wang, Z.; Chen, W.; Xing, J.; Zhang, X.; Tian, H.; Tang, H.; Bi, P.; Li, G.; Zhang, F. Extracting vegetation information from high dynamic range images with shadows: A comparison between deep learning and threshold methods. Comput. Electron. Agric. 2023, 208, 107805. [Google Scholar] [CrossRef]
  5. Narwaria, M.; Da Silva, M.P.; Le Callet, P.; Pepion, R. Tone mapping-based high-dynamic-range image compression: Study of optimization criterion and perceptual quality. Opt. Eng. 2013, 52, 102008. [Google Scholar] [CrossRef]
  6. Korshunov, P.; Hanhart, P.; Richter, T.; Artusi, A.; Mantiuk, R.; Ebrahimi, T. Subjective quality assessment database of HDR images compressed with JPEG XT. In Proceedings of the 2015 Seventh International Workshop on Quality of Multimedia Experience (QoMEX), Costa Navarino, Greece, 26–29 May 2015; pp. 1–6. [Google Scholar]
  7. Valenzise, G.; De Simone, F.; Lauga, P.; Dufaux, F. Performance evaluation of objective quality metrics for HDR image compression. In Proceedings of the Applications of Digital Image Processing XXXVII, San Diego, CA, USA, 18–21 August 2014; Volume 9217, pp. 78–87. [Google Scholar]
  8. Mantiuk, R.; Daly, S.J.; Myszkowski, K.; Seidel, H.P. Predicting visible differences in high dynamic range images: Model and its calibration. In Proceedings of the Human Vision and Electronic Imaging X, San Jose, CA, USA, 17–20 January 2005; Volume 5666, pp. 204–214. [Google Scholar]
  9. Mantiuk, R.; Kim, K.J.; Rempel, A.G.; Heidrich, W. HDR-VDP-2: A calibrated visual metric for visibility and quality predictions in all luminance conditions. ACM Trans. Graph. (TOG) 2011, 30, 40. [Google Scholar] [CrossRef]
  10. Mantiuk, R.K.; Hammou, D.; Hanji, P. HDR-VDP-3: A multi-metric for predicting image differences, quality and contrast distortions in high dynamic range and regular content. arXiv 2023, arXiv:2304.13625. [Google Scholar]
  11. Narwaria, M.; Da Silva, M.P.; Le Callet, P. HDR-VQM: An objective quality measure for high dynamic range video. Signal Process. Image Commun. 2015, 35, 46–60. [Google Scholar] [CrossRef]
  12. Liu, Y.; Ni, Z.; Wang, S.; Wang, H.; Kwong, S. High dynamic range image quality assessment based on frequency disparity. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4435–4440. [Google Scholar] [CrossRef]
  13. Choudhury, A.; Daly, S. HDR image quality assessment using machine-learning based combination of quality metrics. In Proceedings of the 2018 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Anaheim, CA, USA, 26–29 November 2018; pp. 91–95. [Google Scholar]
  14. Cao, P.; Mantiuk, R.K.; Ma, K. Perceptual Assessment and Optimization of HDR Image Rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 22433–22443. [Google Scholar]
  15. Rousselot, M.; Le Meur, O.; Cozot, R.; Ducloux, X. Quality assessment of HDR/WCG images using HDR uniform color spaces. J. Imaging 2019, 5, 18. [Google Scholar] [CrossRef]
  16. Ward, G.; Simmons, M. JPEG-HDR: A backwards-compatible, high dynamic range extension to JPEG. In ACM SIGGRAPH 2006 Courses; ACM: New York, NY, USA, 2006; pp. 3–es. [Google Scholar]
  17. Sugiyama, N.; Kaida, H.; Xue, X.; Jinno, T.; Adami, N.; Okuda, M. HDR image compression using optimized tone mapping model. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 1001–1004. [Google Scholar]
  18. Artusi, A.; Mantiuk, R.K.; Richter, T.; Hanhart, P.; Korshunov, P.; Agostinelli, M.; Ten, A.; Ebrahimi, T. Overview and evaluation of the JPEG XT HDR image compression standard. J. Real-Time Image Process. 2019, 16, 413–428. [Google Scholar] [CrossRef]
  19. Bross, B.; Wang, Y.K.; Ye, Y.; Liu, S.; Chen, J.; Sullivan, G.J.; Ohm, J.R. Overview of the versatile video coding (VVC) standard and its applications. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3736–3764. [Google Scholar] [CrossRef]
  20. Browne, A.; Ye, Y.; Kim, S. Algorithm description for Versatile Video Coding and Test Model 16 (VTM 16). Jt. Video Expert. Team (JVET) ITu-T SG 2022, 16, 3–12. [Google Scholar]
  21. Narwaria, M.; Da Silva, M.P.; Le Callet, P.; Pépion, R. Impact of tone mapping in high dynamic range image compression. In Proceedings of the VPQM, Chandler, AZ, USA, 30–31 January 2014; p. 1. [Google Scholar]
  22. Zerman, E.; Valenzise, G.; Dufaux, F. An extensive performance evaluation of full-reference HDR image quality metrics. Qual. User Exp. 2017, 2, 5. [Google Scholar] [CrossRef]
  23. Miller, S.; Nezamabadi, M.; Daly, S. Perceptual signal coding for more efficient usage of bit codes. SMPTE Motion Imaging J. 2013, 122, 52–59. [Google Scholar] [CrossRef]
  24. Mikhailiuk, A.; Pérez-Ortiz, M.; Yue, D.; Suen, W.; Mantiuk, R.K. Consolidated dataset and metrics for high-dynamic-range image quality. IEEE Trans. Multimed. 2021, 24, 2125–2138. [Google Scholar] [CrossRef]
  25. Perez-Ortiz, M.; Mikhailiuk, A.; Zerman, E.; Hulusic, V.; Valenzise, G.; Mantiuk, R.K. From pairwise comparisons and rating to a unified quality scale. IEEE Trans. Image Process. 2019, 29, 1139–1151. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, Y.; Ni, Z.; Chen, P.; Wang, S.; Kwong, S. HDRC: A subjective quality assessment database for compressed high dynamic range image. Int. J. Mach. Learn. Cybern. 2024, 15, 4373–4388. [Google Scholar] [CrossRef]
  27. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  28. Kuang, J.; Johnson, G.M.; Fairchild, M.D. iCAM06: A refined image appearance model for HDR image rendering. J. Vis. Commun. Image Represent. 2007, 18, 406–414. [Google Scholar] [CrossRef]
  29. ITU. Video Quality Assessment Methods for Multimedia Applications. ITU-T Recommendation P.910. 2008. Available online: https://www.itu.int/rec/T-REC-P.910-202310-I/en (accessed on 25 September 2024).
  30. Available online: https://hdr.sim2.it/ (accessed on 25 September 2024).
  31. Narwaria, M.; Lin, W.; McLoughlin, I.V.; Emmanuel, S.; Chia, L.T. Fourier transform-based scalable image quality measure. IEEE Trans. Image Process. 2012, 21, 3364–3377. [Google Scholar] [CrossRef] [PubMed]
  32. Mantiuk, R.; Myszkowski, K.; Seidel, H.P. A perceptual framework for contrast processing of high dynamic range images. ACM Trans. Appl. Percept. (TAP) 2006, 3, 286–308. [Google Scholar] [CrossRef]
  33. Reinhard, E.; Stark, M.; Shirley, P.; Ferwerda, J. Photographic tone reproduction for digital images. ACM Trans. Graph. 2002, 21, 267–276. [Google Scholar] [CrossRef]
  34. Durand, F.; Dorsey, J. Fast bilateral filtering for the display of high-dynamic-range images. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 23–26 July 2002; pp. 257–266. [Google Scholar]
  35. Ashikhmin, M. A tone mapping algorithm for high contrast images. In Proceedings of the 13th Eurographics workshop on Rendering, Pisa, Italy, 26–28 June 2002; pp. 145–156. [Google Scholar]
  36. Richter, T. On the standardization of the JPEG XT image compression. In Proceedings of the 2013 Picture Coding Symposium (PCS), San Jose, CA, USA, 8–11 December 2013; pp. 37–40. [Google Scholar]
  37. ITU. Methodology for the Subjective Assessment of the Quality of Television Pictures. ITU-R Recommendation BT.500. 2002. Available online: https://www.itu.int/rec/R-REC-BT.500-15-202305-I/en (accessed on 25 September 2024).
  38. Mai, Z.; Mansour, H.; Mantiuk, R.; Nasiopoulos, P.; Ward, R.; Heidrich, W. Optimizing a Tone Curve for Backward-Compatible High Dynamic Range Image and Video Compression. IEEE Trans. Image Process. 2011, 20, 1558–1571. [Google Scholar] [PubMed]
  39. Aydın, T.O.; Mantiuk, R.; Seidel, H.P. Extending quality metrics to full luminance range images. In Proceedings of the Human Vision and Electronic Imaging Xiii, San Jose, CA, USA, 28–31 January 2008; Volume 6806, pp. 109–118. [Google Scholar]
  40. Narwaria, M.; Mantiuk, R.K.; Da Silva, M.P.; Le Callet, P. HDR-VDP-2.2: A calibrated method for objective quality prediction of high-dynamic range and standard images. J. Electron. Imaging 2015, 24, 010501. [Google Scholar] [CrossRef]
  41. Thurstone, L.L. A law of comparative judgment. In Scaling; Routledge: London, UK, 2017; pp. 81–92. [Google Scholar]
  42. Pan, Z.; Yuan, F.; Lei, J.; Fang, Y.; Shao, X.; Kwong, S. VCRNet: Visual compensation restoration network for no-reference image quality assessment. IEEE Trans. Image Process. 2022, 31, 1613–1627. [Google Scholar] [CrossRef] [PubMed]
  43. Ding, K.; Ma, K.; Wang, S.; Simoncelli, E.P. Image quality assessment: Unifying structure and texture similarity. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2567–2581. [Google Scholar] [CrossRef] [PubMed]
  44. Tian, Y.; Wang, S.; Chen, B.; Kwong, S. Causal Representation Learning for GAN-Generated Face Image Quality Assessment. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7589–7600. [Google Scholar] [CrossRef]
  45. Ni, Z.; Liu, Y.; Ding, K.; Yang, W.; Wang, H.; Wang, S. Opinion-Unaware Blind Image Quality Assessment using Multi-Scale Deep Feature Statistics. IEEE Trans. Multimed. 2024. early access. [Google Scholar] [CrossRef]
  46. Chen, B.; Zhu, H.; Zhu, L.; Wang, S.; Kwong, S. Deep Feature Statistics Mapping for Generalized Screen Content Image Quality Assessment. IEEE Trans. Image Process. 2024, 33, 3227–3241. [Google Scholar] [CrossRef]
  47. Li, Y.; Chen, B.; Chen, B.; Wang, M.; Wang, S.; Lin, W. Perceptual quality assessment of face video compression: A benchmark and an effective method. IEEE Trans. Multimed. 2024, 26, 8596–8608. [Google Scholar] [CrossRef]
  48. Zhu, H.; Wu, H.; Li, Y.; Zhang, Z.; Chen, B.; Zhu, L.; Fang, Y.; Zhai, G.; Lin, W.; Wang, S. Adaptive Image Quality Assessment via Teaching Large Multimodal Model to Compare. arXiv 2024, arXiv:2405.19298. [Google Scholar]
  49. Tian, Y.; Chen, B.; Wang, S.; Kwong, S. Towards Thousands to One Reference: Can We Trust the Reference Image for Quality Assessment? IEEE Trans. Multimed. 2023, 26, 3278–3290. [Google Scholar] [CrossRef]
  50. Lei, J.; Li, D.; Pan, Z.; Sun, Z.; Kwong, S.; Hou, C. Fast intra prediction based on content property analysis for low complexity HEVC-based screen content coding. IEEE Trans. Broadcast. 2016, 63, 48–58. [Google Scholar] [CrossRef]
  51. Hong, Y.; Kwong, S.; Chang, Y.; Ren, Q. Consensus unsupervised feature ranking from multiple views. Pattern Recognit. Lett. 2008, 29, 595–602. [Google Scholar] [CrossRef]
  52. Hanji, P.; Mantiuk, R.; Eilertsen, G.; Hajisharif, S.; Unger, J. Comparison of single image HDR reconstruction methods—the caveats of quality assessment. In Proceedings of the ACM SIGGRAPH 2022 Conference Proceedings, Vancouver, BC, Canada, 7–11 August 2022; pp. 1–8. [Google Scholar]
  53. Daly, S.J. Visible differences predictor: An algorithm for the assessment of image fidelity. In Proceedings of the Human Vision, Visual Processing, and Digital Display III, San Jose, CA, USA, 10–13 February 1992; Volume 1666, pp. 2–15. [Google Scholar]
  54. Deeley, R.J.; Drasdo, N.; Charman, W.N. A simple parametric model of the human ocular modulation transfer function. Ophthalmic Physiol. Opt. 1991, 11, 91–93. [Google Scholar] [CrossRef] [PubMed]
  55. Stockman, A.; Sharpe, L.T. The spectral sensitivities of the middle-and long-wavelength-sensitive cones derived from measurements in observers of known genotype. Vis. Res. 2000, 40, 1711–1737. [Google Scholar] [CrossRef] [PubMed]
  56. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef] [PubMed]
  57. Aydin, T.O.; Mantiuk, R.; Myszkowski, K.; Seidel, H.P. Dynamic range independent image quality assessment. ACM Trans. Graph. (TOG) 2008, 27, 1–10. [Google Scholar] [CrossRef]
  58. Watson, A. The cortex transform- Rapid computation of simulated neural images. Comput. Vision Graph. Image Process. 1987, 39, 311–327. [Google Scholar] [CrossRef]
  59. Kottayil, N.K.; Valenzise, G.; Dufaux, F.; Cheng, I. Blind quality estimation by disentangling perceptual and noisy features in high dynamic range images. IEEE Trans. Image Process. 2017, 27, 1512–1525. [Google Scholar] [CrossRef] [PubMed]
  60. Pinson, M.H.; Wolf, S. An objective method for combining multiple subjective data sets. In Proceedings of the Visual Communications and Image Processing 2003, Lugano, Switzerland, 8–11 July 2003; Volume 5150, pp. 583–592. [Google Scholar]
  61. Field, D.J. Relations between the statistics of natural images and the response properties of cortical cells. JOSA A 1987, 4, 2379–2394. [Google Scholar] [CrossRef] [PubMed]
  62. Jia, S.; Zhang, Y.; Agrafiotis, D.; Bull, D. Blind high dynamic range image quality assessment using deep learning. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 765–769. [Google Scholar]
  63. Prashnani, E.; Cai, H.; Mostofi, Y.; Sen, P. Pieapp: Perceptual image-error assessment through pairwise preference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1808–1817. [Google Scholar]
  64. Mantiuk, R.K.; Azimi, M. PU21: A novel perceptually uniform encoding for adapting existing quality metrics for HDR. In Proceedings of the 2021 Picture Coding Symposium (PCS), Bristol, UK, 29 June–2 July 2021; pp. 1–5. [Google Scholar]
  65. Mantiuk, R.K.; Kim, M.; Ashraf, M.; Xu, Q.; Luo, M.R.; Martinovic, J.; Wuerger, S. Practical Color Contrast Sensitivity Functions for Luminance Levels up to 10000 cd/m2. In Proceedings of the Color and Imaging Conference. Society for Imaging Science & Technology, Online, 4–19 November 2020; Volume 28, pp. 1–6. [Google Scholar]
  66. Borer, T.; Cotton, A. A display-independent high dynamic range television system. SMPTE Motion Imaging J. 2016, 125, 50–56. [Google Scholar] [CrossRef]
  67. Huang, X.; Zhang, Q.; Feng, Y.; Li, H.; Wang, X.; Wang, Q. Hdr-nerf: High dynamic range neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18398–18408. [Google Scholar]
  68. Chen, X.; Liu, Y.; Zhang, Z.; Qiao, Y.; Dong, C. Hdrunet: Single image hdr reconstruction with denoising and dequantization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 354–363. [Google Scholar]
  69. Catley-Chandar, S.; Tanay, T.; Vandroux, L.; Leonardis, A.; Slabaugh, G.; Pérez-Pellitero, E. Flexhdr: Modeling alignment and exposure uncertainties for flexible hdr imaging. IEEE Trans. Image Process. 2022, 31, 5923–5935. [Google Scholar] [CrossRef] [PubMed]
  70. Stevens, S.S. On the psychophysical law. Psychol. Rev. 1957, 64, 153. [Google Scholar] [CrossRef]
  71. Santos, M.S.; Tsang, R.; Khademi Kalantari, N. Single Image HDR Reconstruction Using a CNN with Masked Features and Perceptual Loss. ACM Trans. Graph. 2020, 39, 80. [Google Scholar] [CrossRef]
  72. Chen, J.; Yang, Z.; Chan, T.N.; Li, H.; Hou, J.; Chau, L.P. Attention-guided progressive neural texture fusion for high dynamic range image restoration. IEEE Trans. Image Process. 2022, 31, 2661–2672. [Google Scholar] [CrossRef] [PubMed]
  73. Hanhart, P.; Bernardo, M.V.; Pereira, M.; Pinheiro, A.M.G.; Ebrahimi, T. Benchmarking of objective quality metrics for HDR image quality assessment. EURASIP J. Image Video Process. 2015, 2015, 1–18. [Google Scholar] [CrossRef]
  74. Pudil, P.; Novovičová, J.; Kittler, J. Floating search methods in feature selection. Pattern Recognit. Lett. 1994, 15, 1119–1125. [Google Scholar] [CrossRef]
  75. Ashraf, M.; Mantiuk, R.K.; Chapiro, A.; Wuerger, S. castleCSF—A contrast sensitivity function of color, area, spatiotemporal frequency, luminance and eccentricity. J. Vis. 2024, 24, 5. [Google Scholar] [CrossRef] [PubMed]
  76. Xue, W.; Zhang, L.; Mou, X.; Bovik, A.C. Gradient magnitude similarity deviation: A highly efficient perceptual image quality index. IEEE Trans. Image Process. 2013, 23, 684–695. [Google Scholar] [CrossRef]
  77. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
  78. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A feature similarity index for image quality assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef]
  79. Sheikh, H.R.; Bovik, A.C.; De Veciana, G. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans. Image Process. 2005, 14, 2117–2128. [Google Scholar] [CrossRef]
  80. Ni, Z.; Ma, L.; Zeng, H.; Chen, J.; Cai, C.; Ma, K.K. ESIM: Edge similarity for screen content image quality assessment. IEEE Trans. Image Process. 2017, 26, 4818–4831. [Google Scholar] [CrossRef]
  81. Ni, Z.; Zeng, H.; Ma, L.; Hou, J.; Chen, J.; Ma, K.K. A Gabor feature-based quality assessment model for the screen content images. IEEE Trans. Image Process. 2018, 27, 4516–4528. [Google Scholar] [CrossRef]
  82. Rehman, M.U.; Nizami, I.F.; Majid, M.; Ullah, F.; Hussain, I.; Chong, K.T. CN-BSRIQA: Cascaded network-blind super-resolution image quality assessment. Alex. Eng. J. 2024, 91, 580–591. [Google Scholar] [CrossRef]
  83. Kerdvibulvech, C.; Li, Q. Empowering Zero-Shot Object Detection: A Human-in-the-Loop Strategy for Unveiling Unseen Realms in Visual Data. In Proceedings of the Digital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management; Duffy, V.G., Ed.; Springer: Cham, Switzerland, 2024; pp. 235–244. [Google Scholar]
  84. Zhang, X.; Lin, W.; Wang, S.; Liu, J.; Ma, S.; Gao, W. Fine-grained quality assessment for compressed images. IEEE Trans. Image Process. 2018, 28, 1163–1175. [Google Scholar] [CrossRef] [PubMed]
  85. Vaswani, A. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  86. Alexey, D. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  87. Yuan, F.; Zhang, Z.; Fang, Z. An effective CNN and Transformer complementary network for medical image segmentation. Pattern Recognit. 2023, 136, 109228. [Google Scholar] [CrossRef]
  88. Liu, H.; Zhang, C.; Deng, Y.; Xie, B.; Liu, T.; Li, Y.F. TransIFC: Invariant cues-aware feature concentration learning for efficient fine-grained bird image classification. IEEE Trans. Multimed. 2023. early access. [Google Scholar] [CrossRef]
Figure 1. Overview of the HDR image compression schemes.
Figure 1. Overview of the HDR image compression schemes.
Jimaging 10 00243 g001
Figure 2. The production process of the LDR image in the sensor.
Figure 2. The production process of the LDR image in the sensor.
Jimaging 10 00243 g002
Figure 3. Average histogram of the HDR and LDR images in the SIHDR [52] database.
Figure 3. Average histogram of the HDR and LDR images in the SIHDR [52] database.
Jimaging 10 00243 g003
Figure 5. Statistical analysis using the left-tail F-test. Cells highlighted in green (with a value of 1) indicate that the model corresponding to the row significantly outperforms the model corresponding to the column, while a value of 0 indicates no significant difference.
Figure 5. Statistical analysis using the left-tail F-test. Cells highlighted in green (with a value of 1) indicate that the model corresponding to the row significantly outperforms the model corresponding to the column, while a value of 0 indicates no significant difference.
Jimaging 10 00243 g005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Tian, Y.; Wang, S.; Zhang, X.; Kwong, S. Overview of High-Dynamic-Range Image Quality Assessment. J. Imaging 2024, 10, 243. https://doi.org/10.3390/jimaging10100243

AMA Style

Liu Y, Tian Y, Wang S, Zhang X, Kwong S. Overview of High-Dynamic-Range Image Quality Assessment. Journal of Imaging. 2024; 10(10):243. https://doi.org/10.3390/jimaging10100243

Chicago/Turabian Style

Liu, Yue, Yu Tian, Shiqi Wang, Xinfeng Zhang, and Sam Kwong. 2024. "Overview of High-Dynamic-Range Image Quality Assessment" Journal of Imaging 10, no. 10: 243. https://doi.org/10.3390/jimaging10100243

APA Style

Liu, Y., Tian, Y., Wang, S., Zhang, X., & Kwong, S. (2024). Overview of High-Dynamic-Range Image Quality Assessment. Journal of Imaging, 10(10), 243. https://doi.org/10.3390/jimaging10100243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop