Next Article in Journal
Secure Surveillance Systems Using Partial-Regeneration-Based Non-Dominated Optimization and 5D-Chaotic Map
Previous Article in Journal
A Consortium Blockchain Wallet Scheme Based on Dual-Threshold Key Sharing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Curvature and Entropy Statistics-Based Blind Multi-Exposure Fusion Image Quality Assessment

1
Faculty of Information Science and Engineering, Ningbo University, Ningbo 315211, China
2
College of Science and Technology, Ningbo University, Ningbo 315211, China
3
Business School, Ningbo University, Ningbo 315211, China
*
Author to whom correspondence should be addressed.
Symmetry 2021, 13(8), 1446; https://doi.org/10.3390/sym13081446
Submission received: 28 June 2021 / Revised: 31 July 2021 / Accepted: 2 August 2021 / Published: 6 August 2021

Abstract

:
The multi-exposure fusion (MEF) technique provides humans a new opportunity for natural scene representation, and the related quality assessment issues are urgent to be considered for validating the effectiveness of these techniques. In this paper, a curvature and entropy statistics-based blind MEF image quality assessment (CE-BMIQA) method is proposed to perceive the quality degradation objectively. The transformation process from multiple images with different exposure levels to the final MEF image leads to the loss of structure and detail information, so that the related curvature statistics features and entropy statistics features are utilized to portray the above distortion presentation. The former features are extracted from the histogram statistics of surface type map calculated by mean curvature and Gaussian curvature of MEF image. Moreover, contrast energy weighting is attached to consider the contrast variation of the MEF image. The latter features refer to spatial entropy and spectral entropy. All extracted features based on a multi-scale scheme are aggregated by training the quality regression model via random forest. Since the MEF image and its feature representation are spatially symmetric in physics, the final prediction quality is symmetric to and representative of the image distortion. Experimental results on a public MEF image database demonstrate that the proposed CE-BMIQA method achieves more outstanding performance than the state-of-the-art blind image quality assessment ones.

1. Introduction

With the rapid development and popularization of multimedia information technology, high-quality images have become one of the indispensable media contents in our daily lives. Multi-exposure fusion (MEF) can be considered an effective technique to enhance the quality of the image [1]. It provides us a new choice to obtain as much natural scene information on the bright and dark regions as possible [2]. However, the weight assignment processes of MEF algorithms for image fusion will bring about the quality degradation problem that affects the symmetrical perception of the human eyes, especially the structure and detail loss by the underexposure and overexposure [3]. Therefore, it is urgent to design the specific objective quality assessment methods for MEF images. Actually, the performance of any objective quality assessment method is best gauged by its correlation with human subjective opinions of quality. Such human subjective opinions are obtained by conducting the subjective quality assessment. When the opinions are averaged, the mean opinion score (MOS) is obtained, representing the perceptual subjective quality.
Up to now, several studies have been proposed to solve the quality assessment problem for ordinary images. Different image quality assessment (IQA) methods can be divided into three categories: full-reference (FR), reduced-reference (RR), and no-reference/blind (NR) [4]. The FR and RR IQA methods need the pre-defined reference information to make the comparison, while blind IQA (BIQA) methods do not. Since the reference image is difficult to be obtained in real applications, the development of BIQA methods may be more meaningful. Generally, BIQA methods need some quality-sensitive features to learn the quality prediction model. Therefore, the core of designing BIQA methods refers to feature extraction.
Some excellent BIQA methods for ordinary images have been presented via natural scene statistics (NSS) analysis [5,6,7,8,9,10]. For example, one of the famous methods is blind/referenceless image spatial quality evaluator (BRISQUE) [5], which uses scene statistics of locally normalized luminance coefficients to quantify the naturalness loss. The specific features refer to the parameters of NSS models including the empirical distribution of locally normalized luminances and products of locally normalized luminances in the spatial domain. Moreover, Moorthy et al. [6] proposed NSS-based method, dubbed the distortion identification-based image integrity and verity evaluation (DIIVINE) method. It combines the distortion identification with distortion-specific IQA process which mainly extracts the summary statistics based on an NSS wavelet coefficient model. Different from DIIVINE, Saad et al. [7] designed the blind image notator using discrete cosine transform (DCT) statistics (BLINDS-II) method in the DCT domain, and the estimated parameters of the DCT coefficients-based NSS model were utilized to form features for perceptual quality prediction. Li et al. [8] extracted a novel effective structural feature to perceive structural degradation, which utilized the local binary pattern (LBP) to encode the gradient information (GLBP) of images, denoted as the gradient-weighted histogram of GLBP (GWH-GLBP) method. Liu et al. [9] focused on the gradient orientation which has not been deeply explored before, and also deployed a relative gradient magnitude feature accounting for perceptual masking, the method called oriented gradients IQA (OG-IQA). Oszust et al. [10] proposed a hand-crafted blind image assessment measure with local descriptor and derivative filters (SCORER), which emphasized the local features carried by image derivatives of different orders and used for the quality prediction. In addition, Liu et al. [11] constructed the CurveletQA method by utilizing the curvelet transform to extract a set of statistical features, including the coordinates of the maxima of the log-histograms of the curvelet coefficients values, and the energy distributions of both orientation and scale. Gu et al. [12] combined local and global considerations to develop a new NR image quality metric for contrast distortion (NIQMC). The basic principle behind the NIQMC method is that an image with more valuable information has better quality.
However, the above-mentioned BIQA methods are just attempting for quality assessment for ordinary images, which cannot accurately predict the quality of MEF images. Currently, with the emergence of various multi-media forms, the tone-mapped image has also attracted a lot of attention. It is transformed from high dynamic range (HDR) image, which is also a technology for quality enhancement [13]. For tone-mapped images, Gu et al. [14] established a blind tone-mapped quality index (BTMQI) method by analyzing the image information, naturalness, and structure. Kundu et al. [15] extracted the NSS features from spatial domain and HDR-specific gradient features in the gradient domain to design a method called HDR image gradient-based evaluator (HIGRADE). Although the tone-mapped HDR technology has a similar goal with MEF technology, the latter one bypasses the creation process of HDR images. Naturally, the resulting image obtained by the two technologies will present different phenomena.
Overall, the aforementioned methods can accurately evaluate the quality of ordinary and tone-mapped images, while the experimental performances validated on the available MEF database are not always excellent due to the specific distortion that existed in MEF images, such as detail information loss by overexposure and underexposure. It is desirable to propose the IQA methods for MEF images. Currently, there are some FR methods [3,16,17,18,19] for the MEF image. Such as, Ma et al. [3] proposed an FR IQA method based on the structural similarity [20]. Xing et al. [16] designed a method for MEF images by utilizing the contrast structure and contrast saturation. Deng et al. [17] extracted color, structure, and texture-related features for quality regression. Rahman et al. [18] proposed a quality map fusion approach to obtain the true reference of source MEF images for the perceptual quality assessment. Martinez et al. [19] combined multi-scale computation and structural similarities for quality prediction. However, there are almost no specific BIQA methods designed for MEF images.
To propose an appropriate BIQA method for MEF images, how to extract effective feature vectors to distinguish the difference of MEF images with good/poor representations is a rather critical issue. Motivated by these, a curvature and entropy statistics-based blind MEF image quality assessment (CE-BMIQA) method are proposed in this paper, and the main contributions are concluded as follows:
(1)
In terms of structure and detail distortion introduced by inappropriate exposure conditions, the histogram statistics features of surface type maps generated from the mean and Gaussian curvature, and entropy statistics features in the spatial and spectral domains are extracted to form the quality-aware feature vectors.
(2)
Since the contrast variation is a key factor affecting the quality of the MEF image, contrast energy weights are designed to aggregate the above curvature features. Furthermore, a multi-scale scheme is adopted to perceive the distortion of the image in different resolutions for simulating the multi-channel characteristics in the human visual system.
(3)
Considering that it is significant for multimedia applications to bridge the gap between BIQA methods and MEF images, a novel CE-BMIQA method specialized for MEF images is proposed. Experimental results on the available MEF database demonstrate the superiority of the proposed CE-BMIQA method compared with the state-of-the-art BIQA methods.
The remaining of this paper is organized as follows: the proposed CE-BMIQA method is specifically analyzed in Section 2. The experimental results and analysis including database and experimental protocols, performance comparison, impacts of individual feature set and scale number, and computational complexity are provided in Section 3. Further discussion is given in Section 4. Finally, the conclusion is given in Section 5.

2. Proposed CE-BMIQA Method

As the previous introduction mentioned, MEF technology merges a sequence of images with different exposure levels into a MEF image, which can be formulated as
I ( r ) = v = 1 k ϖ v ( r ) X v ( r )
where k is the number of the multi-exposure source images, Xv(r) and ωv(r) represent the luminance value and the weight of the r-th pixel in the v-th exposure image, respectively, I denotes the output MEF image.
After the weight assignment process, the MEF image is prone to quality degradation in some areas due to the unreasonable distributed weights among source multi-exposure images, which is usually expressed as the structure and information loss by the underexposure and overexposure. Figure 1 depicts three MEF images generated by Li’s algorithm [21], Raman’s algorithm [22], and local energy weighting [23], the corresponding mean opinion score (MOS) values are also provided, where the higher the MOS value, the better the visual quality of the MEF image. From Figure 1, the following observations can be drawn: Figure 1a is the best with its abundant detail information and proper exposure to be full of recognition; Figure 1b has lower brightness due to underexposure condition, and cannot preserve the fine details in—crowd, grass, and balloon areas; Figure 1c yields the annoying and unnatural artifacts near the edges of sky and balloon, which is regarded as a pseudo contour. Moreover, the MOS values of three MEF images are 8.74, 3.17, and 2.70, respectively. Therefore, structure and information preservation directly affect the visual perceptual quality of MEF images. Obviously, the distortion form of the MEF image is symmetric to its final prediction quality.
Generally, a BIQA method includes two stages that are training stage and testing stage, the schematic flowchart is shown in Figure 2. The specific conceptual steps are given in Table 1. To meet the quality discriminative requirement, a novel BIQA method for the MEF image is proposed, dubbed as CE-BMIQA, and Figure 3 depicts the feature extraction process of the CE-BMIQA method; refer to Figure 2 and Table 1 for other processes. Specifically, two types of effective features (i.e., the histogram statistics of surface type maps and entropy statistics) are excavated for discriminating the above-described distortion. Then, a feature-weighting approach based on contrast energy is designed to capture the structure loss of image with contrast changes. Finally, all extracted features based on a multi-scale scheme are aggregated to establish the mapping relationship with subjective scores, which is a symmetric process. The detailed implementations of the proposed CE-BMIQA method are given in the following three subsections.

2.1. Curvature Statistics

Generally, a MEF image I(x,y) can be regarded as a product with irregular concave and convex structures, and the points on the image are classified into different types according to their geometric characteristics, which are known as surface types (STs) [24]. Considering STs will change with the introduction of distortion, they are utilized as the basis for the ultimate statistic feature computation to percept structural loss. First, the mean curvature Mc and Gaussian curvature Gc of I(x,y) are calculated to determine the type of each point in the MEF image [25], which is expressed as
M c = ( 1 + g x 2 ) g y y + ( 1 + g y 2 ) g x x 2 g x g y g x y 2 ( 1 + g x 2 + g y 2 ) 3 / 2
G c = g x x g y y g x y 2 ( 1 + g x 2 + g y 2 ) 2
where gx and gy are the first-order partial derivatives, and gxx, gyy, and gxy are the second-order partial derivatives, respectively.
To calculate gx, gy, gxx, gyy, and gxy, the MEF image I(x,y) should be smoothed by a binomial filter first. It can be represented as S = ssT, the vector s is expressed as
s = 1 64 [ 1 6 15 20 15 6 1 ] T
Then, derivative estimation window masks should be calculated via
L x = l 0 l 1 T , L y = l 1 l 0 T , L x x = l 0 l 2 T , L y y = l 2 l 0 T , L x y = l 1 l 1 T
where l0, l1, l2 are given by
l 0 = 1 7 [ 1 1 1 1 1 1 1 ] T l 1 = 1 28 [ 3 2 1 0 1 2 3 ] T l 2 = 1 84 [ 5 0 3 4 3 0 5 ] T
Finally, gx, gy, gxx, gyy, and gxy can be defined as
g x ( x , y ) = L x S I ( x , y ) g y ( x , y ) = L y S I ( x , y ) g x x ( x , y ) = L x x S I ( x , y ) g y y ( x , y ) = L y y S I ( x , y ) g x y ( x , y ) = L x y S I ( x , y )
where ∗ represents the convolution operation.
Figure 4 gives three corresponding mean and Gaussian curvature maps of Figure 1, it is pretty obvious that mean curvature maps emphasize more structural details of MEF images compared with Gaussian curvature maps, and they all have distinctive appearances because of the varying distortions.
With the above-mentioned procedures, ST maps can be computed by different value combinations of Mc and Gc, while Mc and Gc just determine the graph surface shapes and the convex surface shapes, respectively. In total, there are nine STs obtained from the above curvature assignation process, and their complete definitions are listed in Table 2. From Table 2, it can be observed that each point in the MEF image can be categorized into eight fundamental STs (i.e., Peak, Ridge, Saddle Ridge, Flat, Minimal, Pit, Valley, and Saddle Valley) and one non-shaped ST (defined as “None”) for the combination of Mc = 0 and Gc > 0. Figure 5a–c shows three corresponding ST maps of Figure 1, it can be found that they are sensitive to the existing distortion among three MEF images. Therefore, it can be used as an effective way for distortion identification. To quantify the differences of ST maps clearly, histograms of them are also given in Figure 5d–f, respectively. In Figure 5d–f, the horizontal axis represents the ST label, and the vertical axis represents the number of each label. From them, corresponding STs histogram statistics will undergo variation greatly when MEF images are destroyed by complex distortion with different intensities.
Moreover, contrast variation of MEF image is also a key factor for affecting its visual quality. To simultaneously capture the structural details caused by contrast changes, a novel contrast energy weighted STs histogram statistic method is presented, which can be expressed as
h ST ( j ) = i = 1 N W C ( i ) f ST ( ST , j )
f ST ( x , y ) = 1 , x = y 0 ,   otherwise
where N is the number of pixels in a MEF image, j is the index of possible STs, and WC is the contrast energy map predicting local contrast [26], which is defined as:
W C = γ φ ( I ) φ ( I ) + γ o μ
φ ( I ) = ( I d v ) 2 + ( I d h ) 2
where dv and dh are vertical and horizontal second derivatives of Gaussian function, respectively, o is the contrast gain, γ is the maximum value of φ(I), I represents the MEF image, and μ is the noise threshold. Note that the parameters o and μ in the Equation (10) are set according to the recommendations given in [26].
Finally, as the special case (i.e., ST = 4) is invalid, the 8-dimensional curvature statistics feature set is obtained and denoted as F1.

2.2. Entropy Statistic

As can be found from Figure 1, the amount of information contained in each MEF image is usually impaired to different degrees due to underexposure and overexposure, and it can be measured by the entropy, where information loss of image will cause a low entropy value. The entropy values of MEF images in Figure 1 are 7.6106, 7.2598, 6.8534, respectively. Our basic goal is to utilize entropy as statistic properties for distorted MEF images. Obviously, compared with global entropy, local entropy can better discriminate the information distribution in the spatial domain, which just eliminates the impact of special cases that MEF images with the same global entropy may appear distinctly different in visual quality. Since the local entropy is more sensitive to quality degradation, so the entropy of the MEF image block is calculated to analyze the joint distribution of each pixel in the spatial domain, which is defined as:
E s = n p ( n ) log 2 p ( n )
where n is the value of pixel in each MEF image block, and p(n) is the probability density of pixel value.
In addition, block-based entropy in the spectral domain (i.e., after the discrete cosine transformation (DCT) operation) is also computed by using the function of the probability distribution of DCT coefficients within 8 × 8 blocks. To calculate the spectral entropy, the DCT coefficient matrix C of each block should be obtained and then normalized to produce a probability distribution map, which is expressed as:
p ( a , b ) = C ( a , b ) 2 a b C ( a , b ) 2
where a and b are the integers between 1 and 8, and a , b 1 . Then, the local spectral entropy is defined as:
E t = a b p ( a , b ) log 2 p ( a , b )
To observe the behavior of spatial and spectral entropy values against diverse distortions produced by different MEF algorithms, the above calculations are conducted on the different MEF images in Figure 1. As shown in Figure 6, different MEF algorithms bring different influences on the spatial and spectral entropy values, especially reflected in the shape of the histogram. Therefore, we use the mean and skew values of spatial and spectral entropy as the 4-dimensional entropy statistics feature set, denoted as F2.

2.3. Quality Regression

By aggregating the curvature and entropy statistics feature sets, a total of 12-dimensional feature vectors can be obtained and denoted as F = {F1, F2}. Moreover, multi-scale space of image can capture the contents from fine level to coarse level, which is inspired by the processing mechanism on the low-level retina of human visual system (HVS). Therefore, the above feature extraction process is carried out on l scales. After feature extraction, random forest (RF) is utilized to map the feature space to quality space, so that the final quality score Q can be computed by:
Q = f m ( F 1 , F 2 , , F l )
where fm(·) is the trained function for quality regression, and Fl is the extracted quality-aware feature vector on the l-th scale.

3. Experiment Results and Analysis

In this section, the experimental settings are introduced first, including database, evaluation protocols, and implementation details. Second, the proposed CE-BMIQA method is compared with state-of-the-art BIQA methods to verify its validity. Finally, the impacts of individual feature sets, the number of scales, and computational complexity are analyzed.

3.1. Database and Experimental Protocols

To verify the effectiveness of the proposed CE-BMIQA method, the comparative experiments were conducted on an available MEF subjective assessment database [3,27], which contains 136 MEF images with corresponding MOS values. Specially, all MEF images in the database are derived from 17 source multi-exposure images, and the ones belonging to the same image sequence are generated by 8 MEF algorithms, which include local energy weighted linear combination, global energy weighted linear combination, Raman09 [22], Gu12 [23], ShutaoLi12 [28], ShutaoLi13 [29], Li12 [21], and Mertens07 [30]. More details about this MEF image database are given in Figure 7 and Table 3, including the name, resolution, number, source of image sequences.
To objectively compare the performance of the proposed CE-BMIQA method with state-of-the-art BIQA methods, three commonly used evaluation criteria are utilized to reflect the relationship between the predicted objective quality and human subjective scores [5], i.e., the Pearson correlation coefficient (PLCC), Spearman rank-order correlation coefficient (SROCC), and root mean square error (RMSE). PLCC, SROCC, and RMSE measure the accuracy, monotonicity, and prediction error of BIQA methods, respectively. They are defined as:
P L C C = m = 1 U ( Q m Q ¯ ) ( Z m Z ¯ ) m = 1 U ( Q m Q ¯ ) 2 m = 1 U ( Z m Z ¯ ) 2
S R O C C = 1 6 × m = 1 U ε m 2 U × ( U 2 1 )
R M S E = 1 U m = 1 U ( Z m Q m ) 2
where Qm and Zm denote the objective predicted score and subjective score (i.e., MOS) of the m-th MEF image, respectively, while Q ¯ and Z ¯ are the mean values of Qm and Zm, respectively, εm denotes the difference between the ranks of m-th MEF images in the subjective and objective assessments, and U denotes the total number of MEF images.
The optimal evaluation performance should achieve the following results that the absolute values of PLCC and SROCC are closer to 1, RMSE is closer to 0. According to the report from video quality expert group (VQEG) [31], a five-parameter logistic function is applied to make the relationship between subjective and objective quality scores more linear before calculating PLCC and RMSE, which is defined as
Y ( Q ) = β 1 1 2 1 1 + exp β 2 Q β 3 + β 4 Q + β 5
where {β1, β2,…, β5} are the fitting parameters determined by minimizing the sum of squared errors between the objective prediction score and MOS.
During the calculation of PLCC and RMSE, Q is replaced by Y(Q).
In terms of the proposed CE-BMIQA method, there are several experimental parameters to be set in the process of feature extraction, i.e., contrast gain o and noise threshold μ in Equation (10), the block size for entropy statistics, and the number of scale l. Specifically, according to the suggestion in [26], the contrast-gain and noise threshold are fixed at 0.1 and 0.2353, respectively. The noise threshold was determined in a separate image set consisting of 1800 images [32]. The block size for entropy statistics follows the conventional block size settings in the DCT process, that is, 8 × 8. However, the number of scale l may affect the performance of the CE-BMIQA method, and it is set as 3 in this experiment. The specific scale selection process is described in the following Section 3.4.

3.2. Performance Comparison

To verify the performance of the proposed CE-BMIQA method, two categories of the state-of-the-art BIQA methods are selected to make a comparison with it on the MEF image database [27]. The first category of BIQA methods is served for ordinary images, including BRISQUE [5], DIIVINE [6], BLINDS-II [7], GWH-GLBP [8], OG-IQA [9], SCORER [10], CurveletQA [11], NIQMC [12], and the second one contains two BIQA methods for tone-mapped images, i.e., BTMQI [14], HIGRADE [15]. Considering that there is no BIQA method specialized for the MEF image up to now, the aforementioned BIQA methods are utilized for performance comparison. To avoid errors, all methods are obtained from the original released code by authors. Moreover, these ten BIQA methods and the proposed method adopt the supervised learning way to predict the quality of MEF image, so the whole benchmark MEF image database is split into training and testing subsets, and 17-fold cross validation is used to evaluate the performance. Specifically, the MEF images belonging to the same scene are taken as units and the entire database is divided into 17 subsets. Subsequently, 16 subsets and the rest subset are used for training and testing, respectively. Finally, the above train-test stage is cycled in turn until all scenes are traversed, so as to obtain the objective predicted scores of each testing subset.
Table 4 and Table 5 list the PLCC and SROCC results of the proposed CE-BMIQA method and the other comparison BIQA methods for each MEF image sequence and whole database, where the best performance is highlighted in bold. Furthermore, the corresponding hit counts of best performance are recorded to better visualize the performance differences among eleven BIQA methods. From Table 4 and Table 5, the following findings can be drawn: (1) in terms of eight BIQA methods (i.e., BRISQUE [5], DIIVINE [6], BLINDS-II [7], GWH-GLBP [8], OG-IQA [9], SCORER [10], CurveletQA [11], NIQMC [12]) for ordinary images, their performances in quality prediction for MEF images are relatively poor, and the overall PLCC and SROCC values are only 0.534 and 0.525 at most, respectively. While GWH-GLBP method has the worst performance among these methods, only 0.163 and 0.113, and it can be regarded as invalid; (2) obviously, the performance of the latter two methods (i.e., BTMQI [14], HIGRADE [15]) for tone-mapped images are slightly better than the former eight methods, especially HIGRADE method. Such performance difference is likely to be attributed to a few similarities between the distortion caused by tone mapping algorithms and MEF algorithms. However, BTMQI and HIGRADE methods still cannot meet the demands for evaluating MEF images accurately without reference images; (3) the proposed CE-BMIQA method achieves 0.640 and 0.610 on PLCC and SROCC, respectively, which outperforms all competing BIQA methods, the reason is that it takes the impact of structural degradation and information loss caused by underexposure and overexposure into account.
Moreover, the scatter plots between the objective scores and MOS are given on the whole database, as shown in Figure 8. Meanwhile, the logistic function is used to fit these scatter points, and the more points that fall on the fitting curve, the better the performance of BIQA methods. From Figure 8, it can be clearly found that the predicted scores of the proposed CE-BMIQA method have a higher correlation with MOS values compared with the other ten BIQA methods, which further illustrates the superiority of the proposed CE-BMIQA method in predicting the quality of MEF images.

3.3. Impacts of Individual Feature Set

Actually, the proposed CE-BMIQA method has been validated to achieve relatively stable performance since it considers the inherent distortion of MEF images, i.e., structure and information loss caused by underexposure and overexposure. However, the impacts of individual feature set on the final performance are still unclear, thus corresponding evaluation results from two perceptual feature sets (i.e., curvature and entropy statistics) are investigated on the MEF image database [27]. Here, the individual feature sets (i.e., F1 and F2) are used to learn the relevant quality prediction models to validate their significances via three criteria, and the results are reported in Figure 9, where F1 and F2 are the feature sets of curvature statistics and entropy statistics, respectively. From Figure 9, it can be found that F1 makes a great contribution to the final performance, while F2 also plays a significant role in the overall performance improvement. As a result, it seems that F1 is more important than F2, and it can be attributed to that HVS is sensitive to structure degradation with contrast variation. Except for the above observation, it can turn to the best performance result obviously by aggregating the two feature sets, so curvature and entropy statistics features are complementary with each other for distortion estimation.

3.4. Impacts of Scale Number and Computational Complexity

The feature extraction process in the proposed CE-BMIQA method is performed on the multi-scale space, and the scale number l may affect the performance of quality prediction. Therefore, a comparative experiment is designed to select the optimal parameter. Specifically, we assign l = {1, 2, 3, 4} and calculate three evaluation criteria (i.e., PLCC, SROCC, and RMSE) on the MEF database [27] for each possible case, respectively, and the results are showed in Figure 10. Obviously, it reaches the optimal performance when l is 3, so the scale number is set as 3 to ensure the accuracy and robustness of the proposed CE-BMIQA method.
To compare the computational complexity of the proposed CE-BMIQA method with other BIQA methods, their mean running times for 136 MEF images are reported with different resolutions on a computer with 1.80GHz Intel Core i7-8565U CPU, 8GB RAM, and NVIDIA MX 150. All BIQA methods are implemented on MATLAB 2017b (produced by the company of MathWorks, Portola Valley, CA, USA) by using the original codes provided by authors. The running time of each BIQA method is listed in Table 6. From Table 6, it can be observed that although the proposed CE-BMIQA method is in the moderate rank in terms of running time, it is acceptable and adequate for practical applications that need higher precision but are not sensitive to running time.

4. Further Discussion

In this study, a BIQA method for the MEF image has been developed, which is based on curvature and entropy statistics features. Although the proposed CE-BMIQA method has shown better performance than some comparison methods, there still exist some issues to be considered in future work:
(1)
The detail loss usually occurs in the overexposure and underexposure regions of MEF images. In general, the degree of detail preservation also varies with different environmental compositions under different exposure conditions. Therefore, accurate segmentation methods for different exposure regions and semantic segmentation methods can be applied to design BIQA methods for MEF images in the future.
(2)
Generally, the natural scene is almost colorful and moving in practice. When the quality of the MEF image is not good, the color will inevitably be affected, so color is also one of the factors that need to be considered. In future work, color analysis will be concentrated to develop more valuable BIQA methods for MEF images.

5. Conclusions

In this paper, a curvature and entropy statistics-based blind multi-exposure fusion (MEF) image quality assessment (CE-BMIQA) method is proposed for distinguishing the differences among image fusion algorithms. First, histograms of surface type maps based on mean and Gaussian curvature are calculated to form the curvature statistics feature, so as to identify the structure degradation of MEF images. Moreover, a novel contrast energy weighting approach is presented for capturing the structure loss with contrast variation. Then, the mean and skew of spatial and spectral entropy values for MEF image blocks are implemented as another type of quality-aware features. Finally, a multi-scale scheme is adopted to perceive the distortion of MEF image from the fine level to the coarse level. Experimental results on the benchmark MEF image database demonstrate that the proposed CE-BMIQA method achieves outstanding performance than other comparison methods as a whole.
In addition, the CE-BMIQA method still has different evaluation performances for the visual quality of each individual scene. The main reason is that the content of each scene is different, and some scenes may have fewer over-exposure regions and under-exposure regions. As a result, the distortion perception based on contrast and detail information in the CE-BMIQA method cannot accurately predict their quality. In this case, perhaps more attention should be paid to the color and aesthetic distortion. In future work, we will more comprehensively consider the distortion types of MEF images and integrate more considerations.

Author Contributions

Conceptualization, Z.H.; data curation, L.L.; funding acquisition, Y.S.; methodology, Z.H.; software, Z.H.; supervision, L.L.; validation, Y.S. and C.Z.; writing—original draft, Z.H.; writing—review and editing, C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural science foundation of Zhejiang province, grant number LQ20F010002 and Natural science foundation of Ningbo, grant number 2019A610101.

Data Availability Statement

Not Applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kinoshita, Y.; Kiya, H. Scene segmentation-based luminance adjustment for multi-exposure image fusion. IEEE Trans. Image Process. 2019, 28, 4101–4116. [Google Scholar] [CrossRef] [Green Version]
  2. Qi, Y.; Zhou, S.; Zhang, Z.; Luo, S.; Lin, X.; Wang, L.; Qiang, B. Deep unsupervised learning based on color un-referenced loss functions for multi-exposure image fusion. Inf. Fusion 2021, 66, 18–39. [Google Scholar] [CrossRef]
  3. Ma, K.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  4. Wang, Z.; Bovik, A.C. Reduced and no reference visual quality assessment. IEEE Signal Process. Mag. 2011, 29, 29–40. [Google Scholar] [CrossRef]
  5. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  6. Moorthy, A.K.; Bovik, A.C. Blind image quality assessment: From natural scene statistics to perceptual quality. IEEE Trans. Image Process. 2011, 20, 3350–3364. [Google Scholar] [CrossRef]
  7. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind image quality assessment: A natural scene statistics approach in the DCT domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef]
  8. Li, Q.; Lin, W.; Fang, Y. No-reference quality assessment for multiply-distorted images in gradient domain. IEEE Signal Process. Lett. 2016, 23, 541–545. [Google Scholar] [CrossRef]
  9. Liu, L.; Hua, Y.; Zhao, Q.; Huang, H.; Bovik, A.C. Blind image quality assessment by relative gradient statistics and adaboosting neural network. Signal Process. Image Commun. 2016, 40, 1–15. [Google Scholar] [CrossRef]
  10. Oszust, M. Local feature descriptor and derivative filters for blind image quality assessment. IEEE Signal Process. Lett. 2019, 26, 322–326. [Google Scholar] [CrossRef]
  11. Liu, L.; Dong, H.; Huang, H.; Bovik, A.C. No-reference image quality assessment in curvelet domain. Signal Process. Image Commun. 2014, 29, 494–505. [Google Scholar] [CrossRef]
  12. Gu, K.; Lin, W.; Zhai, G.; Yang, X.; Zhang, W.; Chen, C.W. No-reference quality metric of contrast-distorted images based on information maximization. IEEE Trans. Cybern. 2017, 47, 4559–4565. [Google Scholar] [CrossRef]
  13. Rana, A.; Singh, P.; Valenzise, G.; Dufaux, F.; Komodakis, N.; Smolic, A. Deep tone mapping operator for high dynamic 397 range images. IEEE Trans. Image Process. 2019, 29, 1285–1298. [Google Scholar] [CrossRef] [Green Version]
  14. Gu, K.; Wang, S.; Zhai, G.; Ma, S.; Yang, X.; Lin, W.; Zhang, W.; Gao, W. Blind quality assessment of tone-mapped images via analysis of information, naturalness, and structure. IEEE Trans. Multimed. 2016, 18, 432–443. [Google Scholar] [CrossRef]
  15. Kundu, D.; Ghadiyaram, D.; Bovik, A.C.; Evans, B.L. No-reference quality assessment of tone-mapped HDR pictures. IEEE Trans. Image Process. 2017, 26, 2957–2971. [Google Scholar] [CrossRef]
  16. Xing, L.; Cai, L.; Zeng, H.G.; Chen, J.; Zhu, J.Q.; Hou, J.H. A multi-scale contrast-based image quality assessment model for multi-exposure image fusion. Signal Process. 2018, 145, 233–240. [Google Scholar] [CrossRef]
  17. Deng, C.W.; Li, Z.; Wang, S.G.; Liu, X.; Dai, J.H. Saturation-based quality assessment for colorful multi-exposure image fusion. Int. J. Adv. Robot. Syst. 2017, 14, 1–15. [Google Scholar] [CrossRef]
  18. Rahman, H.; Soundararajan, R.; Babu, R.V. Evaluating multiexposure fusion using image information. IEEE Signal Process. Lett. 2017, 24, 1671–1675. [Google Scholar] [CrossRef]
  19. Martinez, J.; Pistonesi, S.; Maciel, M.C.; Flesia, A.G. Multiscale fidelity measure for image fusion quality assessment. Inf. Fusion 2019, 50, 197–211. [Google Scholar] [CrossRef]
  20. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Li, Z.; Zheng, J.; Rahardja, S. Detail-enhanced exposure fusion. IEEE Trans. Image Process. 2012, 21, 4672–4676. [Google Scholar] [PubMed]
  22. Raman, S.; Chaudhuri, S. Bilateral filter based compositing for variable exposure photography. In Proceedings of the Eurographics (Short Papers), Munich, Germany, 30 March–3 April 2009; pp. 1–3. [Google Scholar]
  23. Gu, B.; Li, W.; Wong, J.; Zhu, M.; Wang, M. Gradient field multi-exposure images fusion for high dynamic range image visualization. J. Vis. Commun. Image Represent. 2012, 23, 604–610. [Google Scholar] [CrossRef]
  24. Besl, P.J.; Jain, R.C. Segmentation through variable-order surface fitting. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 167–192. [Google Scholar] [CrossRef] [Green Version]
  25. Do Carmo, M.P. Differential Geometry of Curves and Surfaces; Prentice-Hall: Englewood Cliffs, NJ, USA, 1976. [Google Scholar]
  26. Choi, L.K.; You, J.; Bovik, A.C. Referenceless prediction of perceptual fog density and perceptual image defogging. IEEE Trans. Image Process. 2015, 24, 3888–3901. [Google Scholar] [CrossRef]
  27. Multi-exposure Fusion Image Database. Available online: http://ivc.uwaterloo.ca/database/MEF/MEFDatabase.php (accessed on 11 July 2015).
  28. Li, S.; Kang, X. Fast multi-exposure image fusion with median filter and recursive filter. IEEE Trans. Consum. Electron. 2012, 58, 626–632. [Google Scholar] [CrossRef] [Green Version]
  29. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [PubMed]
  30. Mertens, T.; Kautz, J.; Van Reeth, F. Exposure fusion: A simple and practical alternative to high dynamic range photography. Comput. Graph. Forum 2009, 28, 161–171. [Google Scholar] [CrossRef]
  31. Antkowiak, J.; Baina, T.J. Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment; ITU-T Standards Contributions COM: Geneva, Switzerland, 2000. [Google Scholar]
  32. Groen, I.I.A.; Ghebreab, S.; Prins, H.; Lamme, V.A.F.; Scholte, H.S. From image statistics to scene gist: Evoked neural activity reveals transition from low-level natural image structure to scene category. J. Neurosci. 2013, 33, 18814–18824. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. An example of three MEF images generated by different algorithms. (a) Image generated by Li’s algorithm [21] (MOS = 8.74); (b) image generated by Raman’s algorithm [22] (MOS = 3.17); (c) image generated by local energy weighting [23] (MOS = 2.70).
Figure 1. An example of three MEF images generated by different algorithms. (a) Image generated by Li’s algorithm [21] (MOS = 8.74); (b) image generated by Raman’s algorithm [22] (MOS = 3.17); (c) image generated by local energy weighting [23] (MOS = 2.70).
Symmetry 13 01446 g001
Figure 2. The schematic flowchart of MEF BIQA method.
Figure 2. The schematic flowchart of MEF BIQA method.
Symmetry 13 01446 g002
Figure 3. Feature extraction of the proposed CE-BMIQA method.
Figure 3. Feature extraction of the proposed CE-BMIQA method.
Symmetry 13 01446 g003
Figure 4. The corresponding mean and Gaussian curvature maps of MEF images in Figure 1. (ac) The mean curvature maps; (df) the Gaussian curvature maps.
Figure 4. The corresponding mean and Gaussian curvature maps of MEF images in Figure 1. (ac) The mean curvature maps; (df) the Gaussian curvature maps.
Symmetry 13 01446 g004
Figure 5. The corresponding STs maps and their histogram statistics in Figure 1. (ac) The STs maps; (df) histogram statistics of STs maps.
Figure 5. The corresponding STs maps and their histogram statistics in Figure 1. (ac) The STs maps; (df) histogram statistics of STs maps.
Symmetry 13 01446 g005
Figure 6. The corresponding histograms of spatial and spectral entropy of different MEF images in Figure 1. (a) Histograms of spatial entropy; (b) histograms of spectral entropy.
Figure 6. The corresponding histograms of spatial and spectral entropy of different MEF images in Figure 1. (a) Histograms of spatial entropy; (b) histograms of spectral entropy.
Symmetry 13 01446 g006
Figure 7. Source image sequences contained in the benchmark MEF database [3,27]. Each image sequence is represented by the fused image with the best quality in the subjective test.
Figure 7. Source image sequences contained in the benchmark MEF database [3,27]. Each image sequence is represented by the fused image with the best quality in the subjective test.
Symmetry 13 01446 g007
Figure 8. Scatter plots for the objective predicted scores and MOS values of eleven BIQA methods. (a) BRISQUE [5]; (b) DIIVINE [6]; (c) BLINDS-II [7]; (d) GWH-GLBP [8]; (e) OG-IQA [9]; (f) SCORER [10]; (g) CurveletQA [11]; (h) NIQMC [12]; (i) BTMQI [14]; (j) HIGRADE [15]; (k) CE-BMIQA.
Figure 8. Scatter plots for the objective predicted scores and MOS values of eleven BIQA methods. (a) BRISQUE [5]; (b) DIIVINE [6]; (c) BLINDS-II [7]; (d) GWH-GLBP [8]; (e) OG-IQA [9]; (f) SCORER [10]; (g) CurveletQA [11]; (h) NIQMC [12]; (i) BTMQI [14]; (j) HIGRADE [15]; (k) CE-BMIQA.
Symmetry 13 01446 g008
Figure 9. The performance results of individual feature sets.
Figure 9. The performance results of individual feature sets.
Symmetry 13 01446 g009
Figure 10. The performance results for different scale numbers.
Figure 10. The performance results for different scale numbers.
Symmetry 13 01446 g010
Table 1. The conceptual flowchart of the MEF BIQA method.
Table 1. The conceptual flowchart of the MEF BIQA method.
No.Step
1Input MEF images
2Extract distortion features
3Training stage: train a quality prediction model via random forest (which means learning a mapping relationship between feature space and subjective score). Subjective score means mean opinion score (MOS) which is provided in the subjective database.
4Testing stage: calculate the objective quality score of a MEF image via the trained quality prediction model
Table 2. ST labels obtained by the surface curvature assignation.
Table 2. ST labels obtained by the surface curvature assignation.
Gc > 0Gc = 0Gc < 0
Mc < 0Peak (ST = 1)Ridge (ST = 2)Saddle Ridge (ST = 3)
Mc = 0None (ST = 4)Flat (ST = 5)Minimal (ST = 6)
Mc > 0Pit (ST = 7)Valley (ST = 8)Saddle Valley (ST = 9)
Table 3. The detailed information about the MEF image database [3,27].
Table 3. The detailed information about the MEF image database [3,27].
No.Source SequencesSizeImage Source
1Balloons339 × 512 × 9Erik Reinhard
2Belgium house512 × 384 × 9Dani Lischinski
3Lamp1512 × 384 × 15Martin Cadik
4Candle512 × 364 × 10HDR Projects
5Cave512 × 384 × 4Bartlomiej Okonek
6Chinese garden512 × 340 × 3Bartlomiej Okonek
7Farmhouse512 × 341 × 3HDR Projects
8House512 × 340 × 4Tom Mertens
9Kluki512 × 341 × 3Bartlomiej Okonek
10Lamp2512 × 342 × 6HDR Projects
11Landscape512 × 341 × 3HDRsoft
12Lighthouse512 × 340 × 3HDRsoft
13Madison capitol512 × 384 × 30Chaman Singh Verma
14Memorial341 × 512 × 16Paul Debevec
15Office512 × 340 × 6Matlab
16Tower341 × 512 × 3Jacques Joffre
17Venice512 × 341 × 3HDRsoft
Table 4. The PLCC results of eleven BIQA methods on each MEF image sequence and the entire database.
Table 4. The PLCC results of eleven BIQA methods on each MEF image sequence and the entire database.
No. Methods
[5][6][7][8][9][10][11][12][14][15]Proposed
10.9160.8290.7740.8860.9450.9210.672 0.823 0.8830.9340.578
20.8770.8400.7030.7180.7990.8910.851 0.110 0.6770.9830.851
30.8860.7510.8480.8440.7910.7630.817 0.9410.3410.8950.923
40.8250.8840.8760.2750.5480.7790.043 0.760 0.9740.4120.880
50.1300.4960.6580.7960.6190.7880.191 0.233 0.5220.4660.538
60.0080.1370.3440.2310.8980.3490.644 0.593 0.0860.8350.995
70.7290.5480.4600.2230.9230.8320.856 0.767 0.8410.7100.848
80.6620.7600.5560.7250.5910.8910.9830.841 0.8900.9680.967
90.7920.7630.8500.6280.8030.7930.406 −0.321 −0.1200.8610.882
100.7180.8310.9020.4050.7810.6510.897 0.643 0.5430.9620.045
110.8160.7700.8620.2760.7780.5810.588 0.457 0.6670.9420.948
120.9370.9040.8710.7810.9380.6850.994 0.894 0.6470.8050.995
130.6420.7270.5160.6760.7580.7010.939 0.597 0.7500.7460.941
140.6670.6820.9550.8280.4610.816−0.935 0.826 0.4750.7830.978
150.8560.8210.8560.7030.8340.7120.868 0.884 0.0250.9010.923
160.7610.9010.5260.6360.9350.5950.699 0.532 0.6660.5010.705
170.9900.4230.8520.9140.8400.7090.866 0.549 0.2400.7360.837
All0.4140.4910.5340.1630.5230.4810.3710.5190.4520.5610.640
Hit Count11012011128
Note: [5,6,7,8,9,10,11,12,14,15] represent BRISQUE, DIIVINE, BLINDS-II, GWH-GLBP, OG-IQA, SCORER, CurveletQA, NIQMC, BTMQI, HIGRADE methods, respectively.
Table 5. The SROCC results of eleven BIQA methods on each MEF image sequence and the entire database.
Table 5. The SROCC results of eleven BIQA methods on each MEF image sequence and the entire database.
No. Methods
[5][6][7][8][9][10][11][12][14][15]Proposed
10.8330.7140.7380.6670.8810.8570.310 0.524 0.7860.7620.405
20.7310.6350.5150.5270.8260.7310.802 0.240 0.4070.8980.659
30.4760.6430.5480.3810.9050.7140.571 0.738 0.4760.7140.910
400.8100.310−0.3810.1430.762−0.286 0.429 0.1910.0950.119
50.2380.1430.3330.7140.1670.3330.119 0.381 0.64300.578
60.1910.2140.3100.2860.5950.1910.333 0.381 0.1430.8570.976
70.5950.5480.191−0.0950.7860.5950.8100.286 0.5240.7140.667
80.5000.5240.4760.6430.5240.7620.9050.786 0.7380.7860.905
90.4050.7380.7860.3100.7140.8100.357 0.333 00.7380.833
100.7140.6430.786−0.262−0.1430.5480.7860.429 0.4290.7860.095
110.1910.6670.8810.1670.6430.7380.119 −0.357 −0.5950.8100.883
120.8100.9120.1910.3810.8100.8100.9050.095 −0.5710.7860.929
130.5950.6910.3570.1430.6430.6190.9050.190 0.5950.7620.786
140.0950.4760.714−0.0710.1670.7620.905 0.595 0.4050.7860.833
150.7110.6150.7830.5300.6510.7830.566 0.783 −0.2050.8430.964
160.2860.8330.405−0.5240.9290.7380.071 0.192 0.6190.6910.738
170.252−0.1920.7190.6830.2760.6710.742 0.275 0.1080.7430.635
All0.3800.4030.3460.1130.5250.4940.3370.4040.3430.5660.610
Hit Count01112050038
Note: [5,6,7,8,9,10,11,12,14,15] represent BRISQUE, DIIVINE, BLINDS-II, GWH-GLBP, OG-IQA, SCORER, CurveletQA, NIQMC, BTMQI, HIGRADE methods, respectively.
Table 6. Comparison of running time of the proposed CE-BMIQA method versus other ones.
Table 6. Comparison of running time of the proposed CE-BMIQA method versus other ones.
MetricsBRISQUEDIIVINEBLINDS-IIGWH-GLBPOG-IQASCORER
Time (sec.)0.0416.60214.7360.0580.0310.588
MetricsCurveletQANIQMCBTMQIHIGRADEProposed
Time (sec.)2.51411.92410.0760.2600.480
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

He, Z.; Song, Y.; Zhong, C.; Li, L. Curvature and Entropy Statistics-Based Blind Multi-Exposure Fusion Image Quality Assessment. Symmetry 2021, 13, 1446. https://doi.org/10.3390/sym13081446

AMA Style

He Z, Song Y, Zhong C, Li L. Curvature and Entropy Statistics-Based Blind Multi-Exposure Fusion Image Quality Assessment. Symmetry. 2021; 13(8):1446. https://doi.org/10.3390/sym13081446

Chicago/Turabian Style

He, Zhouyan, Yang Song, Caiming Zhong, and Li Li. 2021. "Curvature and Entropy Statistics-Based Blind Multi-Exposure Fusion Image Quality Assessment" Symmetry 13, no. 8: 1446. https://doi.org/10.3390/sym13081446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop