Next Article in Journal
Nonlinear Reconstruction of Images from Patterns Generated by Deterministic or Random Optical Masks—Concepts and Review of Research
Previous Article in Journal
Evaluation of STIR Library Adapted for PET Scanners with Non-Cylindrical Geometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features

Ronin Institute, Montclair, NJ 07043, USA
J. Imaging 2022, 8(6), 173; https://doi.org/10.3390/jimaging8060173
Submission received: 17 April 2022 / Revised: 15 June 2022 / Accepted: 17 June 2022 / Published: 19 June 2022
(This article belongs to the Section Image and Video Processing)

Abstract

:
With the development of digital imaging techniques, image quality assessment methods are receiving more attention in the literature. Since distortion-free versions of camera images in many practical, everyday applications are not available, the need for effective no-reference image quality assessment algorithms is growing. Therefore, this paper introduces a novel no-reference image quality assessment algorithm for the objective evaluation of authentically distorted images. Specifically, we apply a broad spectrum of local and global feature vectors to characterize the variety of authentic distortions. Among the employed local features, the statistics of popular local feature descriptors, such as SURF, FAST, BRISK, or KAZE, are proposed for NR-IQA; other features are also introduced to boost the performances of local features. The proposed method was compared to 12 other state-of-the-art algorithms on popular and accepted benchmark datasets containing RGB images with authentic distortions (CLIVE, KonIQ-10k, and SPAQ). The introduced algorithm significantly outperforms the state-of-the-art in terms of correlation with human perceptual quality ratings.

1. Introduction

With the considerable advancements made in digital imaging and technology and the easy availability of cheap image-capturing devices, a large number of digital images are captured by non-technical users every day. As a consequence, people upload huge amounts of images and videos to the internet and extensively use streaming applications. In addition, visual information represents 85% of the information that is usable for human beings. Therefore, the quality assessment of digital images is of great importance and a hot research topic, along with several practical applications, such as benchmarking computer vision algorithms [1], monitoring the quality of network visual communication applications [2], fingerprint image evaluation [3], medical imaging applications [4], evaluating image compression [5], or denoising [6] algorithms. Image degradations may occur due to various reasons, such as noise, blurring, fading, or blocking artifacts. Further, the mentioned degradations can be introduced in all phases of the imaging process, such as acquisition, compression, transmission, decompression, storage, and display.
Traditionally, image quality assessment (IQA) algorithms are divided into three distinct classes in the literature with respect to the accessibility of the reference (distortion-free) images [7], i.e., no-reference (NR), full-reference (FR), and reduced-reference (RR). As the idioms indicate, NR-IQA algorithms have absolutely no access to the reference images, FR-IQA methods have full access to the reference images, while RR-IQA approaches have partial information about the reference images. In the literature, NR-IQA is recognized as a more difficult research task than the other two classes due to the complete lack of reference images [8].
In this paper, we introduce a novel NR-IQA model that relies on the fusion of local and global image features. Namely, many NR-IQA methods [9,10,11,12] characterize digital images with global features that are computed using the whole image. This approach is effective in case of artificial distortions (such as JPEG or JPEG2000 compression noise), since they are distributed uniformly in the image. However, authentically or naturally distorted images are often contaminated by noise locally. To improve the perceptual quality estimation of authentically distorted images, we combined novel local and global feature vectors using the statistics of local feature descriptors and some powerful perceptual features. Specifically, 93 features were employed, from which, 80 were introduced for NR-IQA in this study. It was empirically proved that the statistics of local feature descriptors are quality-aware features (since they interpret local image regions from the viewpoint of a human visual system); combining them with global features results in a higher performance than the current state-of-the-art. Our proposed method is code-named as FLG-IQA, referring to the fact that it is based on the fusion of local and global features.
The rest of this study is arranged as follows. In Section 2, related and previous papers are reviewed and outlined. Section 3 provides an overview of the used benchmark databases and evaluation metrics. Moreover, it describes the proposed approach in detail. Section 4 presents experimental results. Specifically, an ablation study is introduced to explain the effects of individual features. Moreover, a comparison to 12 other state-of-the-art methods is presented. Finally, this study is concluded in Section 5.

2. Literature Review

In this section, a review of the existing NR-IQA algorithms is given. For more comprehensive summaries about this topic, interested readers can refer to the studies by Zhai et al. [7], Mohammadi et al. [13], Yang et al. [14], and the book by Xu et al. [15].
NR-IQA methods can be further divided into training-free and machine learning-based classes. Our proposed method falls into the machine learning-based class; thus, this state-of-the-art study mainly focuses on this category. As the terminology indicates, machine learning-based algorithms incorporate some kind of machine learning algorithm to provide an estimation of the perceptual quality of a digital image, while training-free methods do not contain any training steps or rely on machine learning techniques using distorted images. For example, Venkatanath et al. [16] proposed the perception-based image quality evaluator (PIQE), which estimates the image qualities of distorted images using mean subtracted contrast normalized coefficients calculated at all pixel locations. In contrast, the naturalness image quality evaluator (NIQE) proposed by Mittal et al. [17] determines the distance between naturalness features extracted from the distorted image and the features obtained beforehand from distortion-free, pristine images to quantify perceptual image quality. NIQE was further developed in [18,19]. Namely, Zhang et al. [18] improved NIQE by using a Bhattacharyya-like distance [20] between the learned multivariate Gaussian model from the pristine images and those of the distorted images. In contrast, Wu et al. [19] boosted NIQE by more complex features to increase the prediction performance. Recently, Leonardi et al. [21] utilized deep features extracted from a pre-trained convolutional neural network to construct an opinion–unaware method using the correlations through the Gramian matrix between feature maps.
In the literature, various types of machine learning-based NR-IQA methods can be found. Namely, many machine learning-based algorithms rely on natural scene statistics (NSS), which is a powerful tool in characterizing image distortions. The main assumption of NSS is that digital images of high quality follow a sort of statistical regularity and distorted images significantly deviate from this pattern [22]. For instance, Saad et al. [23] examined the statistics of discrete cosine transform (DCT) coefficients. Specifically, a generalized Gaussian distribution (GGD) model was fitted on the DCT coefficients and its parameters were utilized as quality-aware features and mapped onto a perceptual quality score with the help of a support vector regressor (SVR). Another line of papers focused on the wavelet transform to extract quality-aware features. For example, Moorthy and Bovik [24] carried out a wavelet transform over three scales and three orientations; similar to [23], GGDs were fitted to the subband coefficients and their parameters were used as quality-aware features. This method was further improved in [25], where the correlations across scales, subbands, and orientations were also used as quality-aware features. In contrast, Tang et al. [26] extracted quality-aware features from complex pyramid wavelet coefficients. In [27], the authors used color statistics, such as NSS. Specifically, mean subtracted contrast normalized (MSCN) coefficients were created from the color channels of different color spaces. Next, GGD was fitted on these coefficients; similar to the previously mentioned methods, its parameters were used to map onto quality scores. Similar to [27], Mittal et al. [9] applied MSCN coefficients to characterize perceptual quality but they extracted them from the spatial domain. Other proposed models [28,29,30] utilized the statistics of local binary patterns [31] to characterize image texture degradation in the presence of noise and distortions. In [32], Freitas et al. presented a performance comparison of a wide range of local binary pattern types and variants for NR-IQA. Ye and Doerman [33,34] applied the first visual codebooks for NR-IQA. More specifically, Gabor filters were applied for feature extraction and visual codebook creation. Subsequently, the perceptual quality of an image was expressed as the weighted average of the codeword quality scores. In [35], an unsupervised feature learning framework was introduced where an unlabeled codebook was compiled from raw image patches using K means clustering.
With the popularity of deep learning techniques and research, many deep learning-based methods were also proposed in the literature. For instance, Kang et al. [36] trained a convolutional neural network (CNN) with a convolutional layer (using both max and min pooling) and two fully-connected layers. Specifically, the proposed CNN was trained on image patches and the predicted scores of the patches were averaged to obtain the estimated quality of the whole input image. In contrast, Bianco et al. [37] introduced the DeepBIQ architecture, which extracts deep features with the help of a pre-trained CNN from multiple image patches to compile feature vectors that were mapped onto perceptual quality with a SVR. Similarly, Gao et al. [38] relied on a pre-trained CNN but multi-level feature vectors were extracted through global average pooling layers. Sheng et al. [39] utilized the fact that visual saliency is correlated with image quality [40] and trained a saliency-guided deep CNN for NR-IQA from scratch. In contrast, Liu et al. [41] introduced the learning to rank framework [42] for NR-IQA. Specifically, the authors implemented a Siamese CNN to realize this framework. Li et al. [43] implemented a novel loss function (very similar to Pearson’s linear correlation coefficient) to provide the CNN with a shorter convergence time and a better perceptual quality estimation performance. Celona and Schettini [44] trained a novel network, which processes input images at multiple scales, trained jointly by considering NR-IQA as regression, classification, and pairwise ranking, simultaneously.

3. Materials and Methods

3.1. Materials

In this subsection, the applied IQA benchmark databases are described. Further, the applied evaluation methodology and implementation environment are given.

3.1.1. Applied IQA Databases

To evaluate our proposed methods and compare them against the state-of-the-art, three publicly available databases containing RGB images with authentic distortions were utilized—CLIVE [45], KonIQ-10k [46], and SPAQ [47]. The main properties of the applied IQA databases are given in Table 1. The empirical mean opinion score (MOS) distributions of the utilized IQA benchmark databases containing authentic distortions are depicted in Figure 1. In the field of IQA, MOS corresponds to the arithmetic mean of the collected individual quality ratings. Further, those distortions are considered authentic and were introduced to the images during the daily (usually non-expert) usage of imaging devices, such as overexposure, underexposure, camera jitter, motion blur, or noises from camera vibration [48].

3.1.2. Evaluation

The evaluation of NR-IQA algorithms relies on the measurements of the correlations between predicted and ground truth perceptual quality scores. In the literature, Pearson’s linear correlation coefficient (PLCC), Spearman’s rank-order correlation coefficient (SROCC), and Kendall’s rank-order correlation coefficient (KROCC) are widely applied and accepted for this end [49]. As recommended by Sheikh et al. [50], non-linear mapping was carried out between the predicted and the ground truth scores before the computation of PLCC using a logistic function with five parameters,
Q = β 1 1 2 1 1 + e β 2 ( Q p β 3 ) + β 4 Q p + β 5 ,
where Q p and Q indicate the predicted and mapped scores, respectively. Further, β i , i = 1 , . . . , 5 denote the fitting parameters.
As usual, in machine learning, particularly in NR-IQA, 80% of images in a database were used for training and the remaining 20% for testing. Further, median PLCC, SROCC, and KROCC values, which were measured over 100 random train–test splits, are reported in this study. PLCC, SROCC, and KROCC can be defined between vectors x and y as
P L C C ( x , y ) = i = 1 N ( x i x ¯ ) ( x i x ¯ ) i = 1 N ( x i x ¯ ) 2 i = 1 N ( x i x ¯ ) 2 ,
S R O C C ( x , y ) = 1 6 · i = 1 N d i 2 N ( N 2 1 ) ,
where
d i = r a n k ( x i ) r a n k ( y i ) ,
K R O C C ( x , y ) = N c N d N N 1 2 ,
where N stands for the length of the vectors, x ¯ and y ¯ are the means of the vectors, N c and N d indicate the number of concordant and discordant pairs between x and y , respectively.
Our experiments were carried out in MATLAB R2021a, mainly employing the functions of the image processing and machine learning and statistics toolboxes. The main characteristics of the computer configuration applied in our experiments are outlined in Table 2.

3.2. Methods

A high-level summary of the proposed method is depicted in Figure 2, while Table 3 sums up the introduced and applied quality-aware features. As it can be seen from Figure 2, feature vectors were extracted from the set of training images to obtain a quality model with the help of a machine learning algorithm. Formally, our quality model is defined by q = G ( F ) , where q is the estimated quality score, F , and G is a regression model that can be the Gaussian process regression (GPR) and support vector regressor (SVR). In the training phase, the model is optimized to minimize the distance between the estimated and ground truth quality scores. In the testing phase, the obtained quality model G was applied to estimate the perceptual quality of previously unseen test images. As already mentioned in the previous subsection, the evaluation of the quality model relies on measuring the correlation strength between the predicted and ground truth quality scores. Further, median PLCC, SROCC, and KROCC values measured over 100 random train–test splits are reported in this study.
The main (novel) contribution of the proposed method is the introduced quality-aware features outlined in Table 3. To tackle the wide variety of authentic distortions, a broad spectrum of statistical local and global features were applied. Artificial distortions (such as JPEG or JPEG2000 compression) are usually uniformly distributed in an image, which can be characterized well by global and homogeneous features [51]. However, authentic distortions often appear locally in a digital image, which can be better captured by local, non-homogeneous image features. Namely, authentic distortions are mainly introduced to the images during daily (non-expert) usage of imaging devices, such as overexposure, underexposure, camera jitter, motion blur, or noises from camera vibration [48]. In this study, the statistics of local feature descriptors supplemented by global features were introduced to construct a novel image quality model. Specifically, local feature descriptors were designed to characterize images from the human visual system’s point of view [52]. For instance, FAST [53] and Harris [54] are suitable for finding small mono-scale keypoints. On the other hand, other local feature descriptors, such as SURF [55], find multi-scale keypoints. Although they are designed to exhibit some kind of invariance against noise and illumination, perfect robustness does not exist. That is why their statistics may help characterize local image quality. Moreover, images with authentic distortions may suffer from overexposure or underexposure, which influence image content globally. The main goal of this study was to empirically prove that the fusion of local and global features can effectively estimate the perceptual quality of digital images with authentic distortions. To be specific, 93 features were applied in this study, of which, 80 were introduced for NR-IQA. The introduced features can be divided into four groups, i.e., the statistics of local feature descriptors measured on the grayscale image (f1-f35 in Table 3) and on the Prewitt-filtered image (f36-f70), Hu invariant moments computed from the binarized Sobel edge map of the input image (f66-f70), perceptual features (f78-f87), and the relative Grünwald–Letnikov derivative and gradient statistics (f88-f93). In the following subsection, each group of the features is introduced in detail.
Table 3. Summary of the applied features. Quality-aware features proposed by this paper are in bold.
Table 3. Summary of the applied features. Quality-aware features proposed by this paper are in bold.
Feature
Number
InputFeatureNumber of
Features
f1-f5SURF [55], Grayscale imagemean, median, std,
skewness, kurtosis
5
f6-f10FAST [53], Grayscale imagemean, median, std,
skewness, kurtosis
5
f11-f15BRISK [56], Grayscale imagemean, median, std,
skewness, kurtosis
5
f16-f20KAZE [57], Grayscale imagemean, median, std,
skewness, kurtosis
5
f21-f25ORB [58], Grayscale imagemean, median, std,
skewness, kurtosis
5
f26-f30Harris [54], Grayscale imagemean, median, std,
skewness, kurtosis
5
f31-f35Minimum Eigenvalue [59],
Grayscale image
mean, median, std,
skewness, kurtosis
5
f36-f40SURF [55], Filtered imagemean, median, std,
skewness, kurtosis
5
f41-f45FAST [53], Filtered imagemean, median, std,
skewness, kurtosis
5
f46-f50BRISK [56], Filtered imagemean, median, std,
skewness, kurtosis
5
f51-f55KAZE [57], Filtered imagemean, median, std,
skewness, kurtosis
5
f56-f60ORB [58], Filtered imagemean, median, std,
skewness, kurtosis
5
f61-f65Harris [54], Filtered imagemean, median, std,
skewness, kurtosis
5
f66-f70Minimum Eigenvalue [59],
Filtered image
mean, median, std,
skewness, kurtosis
5
f71-f77Binary imageHu invariant moments [60]7
f78-f87RGB imagePerceptual features10
f88GL-GM map ( α = 0.3 ) histogram variance1
f89GL-GM map ( α = 0.6 ) histogram variance1
f90GL-GM map ( α = 0.9 ) histogram variance1
f91GM map [61]histogram variance1
f92RO map [61]histogram variance1
f93RM map [61]histogram variance1

3.3. Statistics of Local Feature Descriptors

In contrast to artificially distorted images, where distortions are uniformly distributed, natural or authentic distortions are often present locally in images. This is why the statistics of local feature descriptors are proposed for quality-aware features in this study. In this study, we calculated the statistics of the following local feature detectors: SURF [55], FAST [53], BRISK [56], KAZE [57], ORB [58], Harris [54], and minimum eigenvalue [59]. Specifically, the strongest 250 interest points were detected separately with each of the above-mentioned local feature detectors and the adherent feature vectors were determined. In this study, the following statistics of the obtained feature vectors were considered as local features: mean, median, standard deviation, skewness, and kurtosis. The skewness of a vector x containing n elements is calculated as
s = 1 n i = 1 n ( x i x ¯ ) 3 1 n i = 1 n ( x i x ¯ ) 2 2 ,
while its kurtosis can be given as
k = 1 n i = 1 n ( x i x ¯ ) 4 1 n i = 1 n ( x i x ¯ ) 2 3 ,
where x ¯ stands for the arithmetic mean of x . Specifically, the statistics of the features around the feature points were calculated and their arithmetic means were considered as quality-aware features. Moreover to increase the distinctiveness of local feature detectors [62], the statistics of the local feature detectors were also extracted from the filtered version of the input image obtained by Prewitt operators [63].

3.4. Hu Invariant Moments

The moment ( m ) is a projection of a function (in image processing an image I ( x , y ) ) to the polynomial basis. Formally, it can be written as
m i , j = x = 1 N y = 1 M x i y j I ( x , y ) ,
where N and M are the width and the height of image I ( x , y ) , respectively. The order ( r ) of the moment is defined as r = i + j . It can be easily pointed out that m 0 , 0 is the mass of the image. Further, m 1 , 0 / m 0 , 0 and m 0 , 1 / m 0 , 0 are the coordinates of the center of gravity of the image. The central moment ( μ ) is defined as
μ i , j = x = 1 N y = 1 M x m 1 , 0 m 0 , 0 i y m 0 , 1 m 0 , 0 j I ( x , y ) .
It is evident that μ 0 , 0 = m 0 , 0 . Next, the normalized central moment ( η ) is given as
η i , j = μ i , j ( μ 0 , 0 ) λ ,
where
λ = i + j 2 + 1 .
Hu [60] proposed seven invariant moments, which were defined using the normalized central moment, such as
ϕ 1 = η 2 , 0 + η 0 , 2 ,
ϕ 2 = ( η 2 , 0 + η 0 , 2 ) 2 + 4 ( η 1 , 1 ) 2 ,
ϕ 3 = ( η 3 , 0 3 η 1 , 2 ) 2 + ( 3 η 2 , 1 η 0 , 3 ) 2 ,
ϕ 4 = ( η 3 , 0 + η 1 , 2 ) 2 + ( η 2 , 1 + η 0 , 3 ) 2 ,
ϕ 5 = ( η 3 , 0 3 η 1 , 2 ) ( η 3 , 0 + η 1 , 2 ) [ ( η 3 , 0 + η 1 , 2 ) 2 3 ( η 2 , 1 + η 0 , 3 ) 2 ] +   ( 3 η 2 , 1 η 0 , 3 ) ( η 2 , 1 + η 0 , 3 ) [ 3 ( η 3 , 0 + η 1 , 2 ) 2 ( η 2 , 1 + η 0 , 3 ) 2 ]
ϕ 6 = ( η 2 , 0 η 0 , 2 ) [ ( η 3 , 0 + η 1 , 2 ) 2 ( η 2 , 1 + η 0 , 3 ) 2 ] + 4 η 1 , 1 ( η 3 , 0 + η 1 , 2 ) ( η 2 , 1 + η 0 , 3 )
ϕ 7 = ( 3 η 2 , 1 η 0 , 3 ) ( η 3 , 0 + η 1 , 2 ) [ ( η 3 , 0 + η 1 , 2 ) 2 3 ( η 2 , 1 + η 0 , 3 ) 2 ]   ( η 3 , 0 3 η 1 , 2 ) ( η 2 , 1 + η 0 , 3 ) [ 3 ( η 3 , 0 + η 1 , 2 ) 2 ( η 2 , 1 + η 0 , 3 ) 2 ] .
The human visual system is very sensitive to edge and contour information, since this information gives reliable implications about the structure of an image [64,65]. To characterize the contour information of an image, the Hu invariant moments [60] of the image’s binary Sobel edge map were used as quality-aware features. The horizontal and vertical derivative approximations of input image I were determined as
G x = + 1 0 1 + 2 0 2 + 1 0 1 I ,
G y = + 1 + 2 + 1 0 0 0 1 2 1 I ,
where ∗ is the convolution operator. Next, the gradient magnitude was computed as
G = G x 2 + G y 2 .
The binary Sobel edge map was obtained from the gradient magnitude by applying a cutoff threshold corresponding to the quadruple of G ’s mean.

3.5. Perceptual Features

Some perceptual features, which are proved to be consistent with human quality judgments [66], were also built into our model. In the following, an RGB color image is denoted by I and I c ( c ( R , G , B ) ) is a color channel of input image I. Moreover, x stands for the pixel coordinate and we assume that I has N pixels.
  • Blur: It is probably the most dominant source of perceptual image quality deterioration in digital imaging [67]. To quantify the emergence of the blur effect, the blur metric of Crété-Roffet et al. [68], which is based on the measurements of intensity variations between neighboring pixels, was implemented due to its low computational costs.
  • Colorfulness: In [69], Choi et al. pointed out that colorfulness is a critical component in human image quality judgment. We determined colorfulness using the following formula proposed by Hasler and Suesstrunk [70]:
    C F = σ r g 2 + σ y b 2 + 3 10 μ r g 2 + μ y b 2 ,
    where σ and μ stand for the standard deviation and the mean of the matrices denoted in the subscripts, respectively. Specifically, these matrices are given as:
    r g = R G ,
    y b = 1 2 ( R + G ) B ,
    where R, G, and B are the red, green, and blue color channels, respectively.
  • Chroma: It is one of the relevant image features among a series of color metrics in the CIELAB color space. Moreover, chroma is significantly correlated with haze, blur, or motion blur in the image [70]. It is defined as
    C h r o m a ( x ) = a ( x ) 2 + b ( x ) 2 ,
    where a and b are the corresponding color channels of the CIELAB color space. The arithmetic mean of C h r o m a ( x ) was used as a perceptual feature in our model.
  • Color gradient: The estimated color gradient magnitude (CGM) map is defined as
    C G M ( x ) = c ( R , G , B ) ( I x c ( x ) ) 2 + ( I y c ( x ) ) 2 ,
    where I x ( x ) and I y ( x ) stand for the approximate directional derivatives in the horizontal x and vertical y directions of I ( x ) , respectively. In our study, the mean and standard deviations of C G M ( x ) are utilized as quality-aware features.
  • Dark channel feature (DCF): In the literature, Tang et al. [71] proposed DCF [72] for image quality assessment, since it can effectively identify haze effects in images. A dark channel is defined as
    I d a r k ( x ) = min y Ω ( x ) min c ( R , G , B ) I c ( y ) ,
    where Ω ( x ) denotes the image patches around the pixel location x. In our implementation, an image patch corresponds to a 15 × 15 square. Next, the DCF is defined as
    D C F = 1 | | S | | i S I d a r k ( i ) c ( R , G , B ) I c ( i ) ,
    where | | S | | is the size of image I.
  • Michelson contrast: Contrast is one of the most fundamental characteristics of an image, since it influences the ability to distinguish objects from each other in an image [73]. Thus, contrast information is built into our NR-IQA model. The Michelson contrast measures the difference between the maximum and minimum values of an image [74], defined as
    C M i c h e l s o n = c ( R , G , B ) max ( I c ( x ) ) min ( I c ( x ) ) max ( I c ( x ) ) + min ( I c ( x ) ) .
  • Root mean square (RMS) contrast is defined as
    C R M S = 1 N x = 0 N 1 ( I ( x ) I ¯ ) 2
    where I ¯ denotes the mean luminance of I ( x ) .
  • Global contrast factor (GCF): Contrary to Michelson and RMS contrasts, GCF considers multiple resolution levels of an image to estimate human contrast perception [75]. It is defined as
    G C F = i = 1 9 w i C i ,
    where C i ’s are the average local contrasts and w i ’s are the weighting factors. The authors examined nine different resolution levels, which is why the number of weighting factors are nine; w i ’s are defined as
    w i = ( 0.406385 · i 9 + 0.334573 ) · i 9 + 0.0877526 ,
    which is a result of an optimum approximation from the best fitting [75]. The local contrasts are defined as follows. First, the image of size w × h is rearranged into a one-dimensional vector using row-wise sorting. Next, the local contrast l C i in pixel location i is defined as
    l C i = | L i L i 1 | + | L i L i + 1 | + | L i L i w | + | L i L i + w | 4 ,
    where L i denotes the pixel value at location i after gamma correction ( γ = 2.2 ) . Finally, the average local contrast at resolution i (denoted by C i in Equation (31)) is determined as the average of all l C i ’s over the entire image.
  • Entropy: It is a quantitative measure of the image’s carried information [76]. Typically, an image with better quality is able to transmit more information. This is why entropy was chosen as a quality-aware feature. The entropy of a grayscale image is defined as
    E = n = 0 255 p ( n ) · log 2 p ( n ) ,
    where p ( · ) consists of the normalized histogram counts of the grayscale image.
Table 4 illustrates the average values of the perceptual features in five equal MOS intervals of CLIVE [45] from very low image quality to very high image quality. From these numerical results, it can be observed that the mean values of the applied perceptual features are roughly proportional with the perceptual quality class. For instance, the mean values of several perceptual features (the mean and the standard deviations of the color gradient, Michelson contrast, RMS contrast, GCF, and entropy) monotonically increase with the quality class. Similarly, Table 5 illustrates the standard deviation values of the perceptual features in five equal MOS intervals of CLIVE [45] from very low image quality to very high image quality. It can be seen that the standard deviation values are also roughly proportional to the perceptual quality classes. For instance, the standard deviation values of several perceptual features (color gradient-mean, DCF, Michelson contrast, RMS contrast, and entropy) exhibit a remarkable proportionality with the perceptual quality classes.

3.6. Relative Grünwald–Letnikov Derivative and Gradient Statistics

In image processing, image gradient is one of the most widely used features [77,78,79,80] and a strong predictive factor for image quality [61,81]. To characterize gradient degradation in the presence of image noise, the idea of gradient magnitude (GM), relative gradient orientation (RO), and relative gradient magnitude (RM) maps were applied from [61], on the one hand. On the other hand, the idea of GM maps was generalized and developed further using the Grünwald–Letnikov (GL) derivative [82]. Once, GM, RO, and RM maps were computed—following the recommendations of [61]—their histogram variances were used as quality-aware features. Given a normalized histogram h ( x ) , the histogram variance is defined as
V a r ( h ( x ) ) = i = 1 n ( h ( x ) h ¯ ) ,
where n is the length of h ( x ) and h ¯ is the mean of h ( x ) .
The Grünwald–Letnikov derivative, introduced by Anton Karl Grünwald and Aleksey Vasilievich Letnikov, enables differentiating a function a non-integer amount of times [83]. Generally, a one-dimensional function f ( x ) can be differentiated for any n N + using the following formula
d n f ( x ) d x n = lim h 0 j = 0 n ( 1 ) j n j f ( x j h ) h 2 ,
where
n j = n ( n 1 ) . . . ( n j + 1 ) j ! .
Grünwald and Letnikov invented an approach, which enables taking the derivative of a function by arbitrary, non-integer α -times. Formally, it has been written
D x 0 , x G L α f ( x ) = lim h 0 1 h α j = 0 [ x x 0 h ] ( 1 ) j Γ ( α + 1 ) Γ ( j + 1 ) Γ ( α j + 1 ) f ( x j h ) ,
where Γ ( · ) is the Gamma-function, D x 0 , x G L α f ( x ) is the α th order Grünwald–Letnikov derivative, and x and x 0 stand for the upper and lower bounds, respectively. For a discrete two-dimensional signal (in the context of image processing: a digital image), I ( x , y ) , the GL derivative in the x-direction can be defined as
D G L α I x ( x , y ) = I ( x , y ) α I ( x 1 , y ) + α ( α 1 ) 2 I ( x 2 , y ) .
Similarly, in the y-direction
D G L α I y ( x , y ) = I ( x , y ) α I ( x , y 1 ) + α ( α 1 ) 2 I ( x , y 2 ) .
Similar to the traditional definition, the GL derivate can be given as
D G L α I ( x , y ) = ( D G L α I x ( x , y ) ) 2 + ( D G L α I y ( x , y ) ) 2 .
Although the physical meaning of the GL fractional derivative is not absolutely understandable, it is important to notice that GL derivatives, in contrast to integer derivatives, do not have local characters [84]. Namely, their derivatives depend on the entire functions. In previous work [85], the combination of global and local variations of an image using GL derivatives (global) and image gradients (local) proved to be beneficial for FR-IQA. Note that the full-reference (FR) setting aims to evaluate an image using reference images without distortions. In this study, in order to characterize global variations of images for no reference (NR)-IQA, the computation of the GM [61] map was modified by using the equations of the GL derivative, as in Equations (39)–(41). More specifically, three GL-GM maps were computed with α = 0.3 , α = 0.6 , and α = 0.9 , respectively, and their histogram variances were taken as quality-aware features.
The GM map can be given very similarly to Equation (41):
G M ( x , y ) = ( I x ( x ) ) 2 + ( I y ( x ) ) 2
where I x ( x ) and I y ( x ) stand for the approximate directional derivatives in the horizontal x and vertical y directions of I ( x ) , respectively.
The definition of the RO map is as follows [61]. First, the gradient orientation needs to be defined:
θ ( x , y ) = arctan I y ( x , y ) I x ( x , y ) .
The RO map can be given as
R O ( x , y ) = θ ( x , y ) θ A V E ( x , y )
where θ A V E ( x , y ) is the local average orientation; the authors define as
θ A V E ( x , y ) = arctan I y , A V E ( x , y ) I x , A V E ( x , y ) ,
where the average directional derivatives are defined as
I y , A V E ( x , y ) = 1 M · N ( m , n ) W I y ( x m , y n ) ,
and
I x , A V E ( x , y ) = 1 M · N ( m , n ) W I x ( x m , y n ) ,
where W stands for the local neighborhood over the values that are computed. Similar to the RO map, the RM map can be given as
R M ( x , y ) = I x ( x , y ) I x , A V E ( x , y ) 2 + I y ( x , y ) I y , A V E ( x , y ) 2 .

4. Results

In this section, our results are presented. Specifically, Section 4.1 consists of an ablation study to analyze the performance of the proposed, individual quality-aware features. Next, Section 4.2 describes the results of a performance comparison to other state-of-the-art NR-IQA methods.

4.1. Ablation Study

In this subsection, an ablation study on CLIVE [45] is presented to reason the design choices and demonstrate the individual performances of the proposed quality-aware features. The results are summarized in Table 6. From these results, it can be observed that the statistics of feature descriptors are quality-aware features and can deliver a rather strong performance. However, the set of the applied perceptual features delivers the strongest performance. It can also be seen that combining the statistics of local feature descriptors with global features results in improved performance. Moreover, GPR with the rational quadratic kernel function outperforms SVR with the Gaussian kernel function as a regressor for all of the proposed quality-aware features. Figure 3 provides detailed results for all local feature descriptors. On their own, all local feature descriptors can provide weak or mediocre performances. However, their concatenations provide rather strong performances. We attribute this result to the ability of local feature descriptors to diversely characterize local image distortions. Based on the above observations, we used GPR with the rational quadratic kernel function as a regressor in the proposed method, which is code-named FLG-IQA, referring to the fact that it is based on the fusion of local and global features.
To demonstrate that all features are relevant in FLG-IQA, we present an experiment based on the evaluation protocol described in Section 3.1.2 and using the CLIVE [45] database, in which a given feature is eliminated from the entire feature vector. As demonstrated in Figure 4, all features are important. If one of the applied features is removed, the performance of FLG-IQA falls back. However, the statistics of SURF, KAZE, and minimum eigenvalue local feature descriptors on the Prewitt-filtered image are the most decisive to the performance of FLG-IQA from the statistics of local feature descriptors. Moreover, perceptual features have the most contributing effects to the performance of the proposed method. If we contrast the results in Figure 4 with results in Figure 3d and Table 6, the following interesting fact can be observed. Features that have strong effects on the performance of FLG-IQA, do not have always superior performance individually. For instance, if the statistics of the minimum eigenvalue local feature descriptor in the Prewitt-filtered image are removed, a significant performance drop can be observed. On the other hand, its individual performance lags behind those of the Harris statistics on the filtered image, although Harris has a rather strong individual performance. This indicates that the statistics of local feature descriptors are quality-aware feature vectors and complement each other in NR-IQA. Further, perceptual features have strong individual performances and their removal from the feature vector throw back the performance, indicating that they are very strong predictors of the perceptual image quality.
To further prove that all entries of the proposed feature vector are important, the rank importance of all predictors was investigated using the RReliefF algorithm [86]. Namely, the main idea behind RReliefF [87] is to estimate the discriminative power of features based on their ability on how well they differentiate between instances that lie near each other in the feature space. To this end, RReliefF penalizes those predictors that provide different values to neighbors with the same response values. On the other hand, it rewards those predictors that give different values to neighbors with different response values. Further, the number of examined neighbors is an input parameter of RReliefF. For all details about RReliefF, we refer to the paper by Robnik-Sikonja and Kononenko [88]. The results of the RReliefF algorithm—using 1 , 3 , 5 , and 7 nearest neighbors—on the features extracted from the images of CLIVE [45], are depicted in Figure 5. From these results, it can be seen that all entries of the proposed feature vector are important since the weights of importance are non-negative in all cases.

4.2. Comparison to the State-of-the-Art

In this subsection, the proposed FLG-IQA algorithm is compared to several state-of-the-art methods, such as BLIINDS-II [23], BMPRI [12], BRISQUE [9], CurveletQA [10], DIIVINE [25], ENIQA [89], GRAD-LOG-CP [11], GWH-GLBP [90], NBIQA [91], OG-IQA [61], PIQE [16], and SSEQ [92], whose original MATLAB source codes can be found online. The above-mentioned algorithms were evaluated in the same environment with the same evaluation protocol as ours. Since PIQE [16] is a training-free method without any machine learning technique, it was directly evaluated in the full datasets. The results are summarized in Table 7 and Table 8, where it can be seen that the proposed FLG-IQA is able to outperform all other considered methods in three large IQA databases with authentic distortions, i.e., CLIVE [45], KonIQ-10k [46], and SPAQ [47]. Table 9 illustrates the direct and weighted average performance values obtained from the achieved results on the used IQA benchmark databases. It can be observed that the proposed FLG-IQA outperforms all other examined state-of-the-art algorithms by a large margin in this comparison. Specifically, FLG-IQA surpasses the second best methods by approximately 0.05 in terms of PLCC, SROCC, and KROCC in direct and weighted averages as well. In general, all methods achieved higher values in the case of the weighted average, which implies that the examined methods tend to perform better on larger databases. As an illustration of the results, Figure 6 depicts ground truth versus predicted scores in CLIVE [45] and KonIQ-10k [46] test sets, respectively.
To prove that the achieved performance difference against the state-of-the-art in CLIVE [45], KonIQ-10k [46], and SPAQ [47] was significant, significance tests were also carried out. Specifically, one-sided t-tests were applied between the 100 SROCC values provided by the proposed FLG-IQA method and one other examined state-of-the-art algorithm. Further, the null hypothesis was that the mean SROCC values of the two sets were equal to each other at a confidence level of 95 % . The results of the significance tests are summarized in Table 10 where symbol 1 ( 1 ) denotes that the proposed FLG-IQA is significantly better (worse) than the algorithm represented in the row of the table on the IQA benchmark database represented in the column. From the presented results, it can be seen that FLG-IQA is significantly better than the state-of-the-art in the utilized IQA benchmark databases containing authentic distortions.
The effectiveness of the proposed FLG-IQA was further proved in a cross-database test where the examined state-of-the-art algorithm and the proposed method were trained on the large KonIQ-10k [46] and tested in CLIVE [45]. The results of this test are summarized in Table 11. From the presented numerical results, it can be seen that the proposed method provides a significantly higher performance than the other methods. Specifically, FLG-IQA performs better than the second-best method bu approximately 0.11 in terms of PLCC and 0.07 in terms of SROCC, respectively. Figure 7 depicts the results of FLG-IQA in the cross database in normalized ground truth scores versus a normalized predicted score scatter plot.

5. Conclusions

In this paper, a novel NR-IQA method for authentically distorted images was introduced. Specifically, a diverse set of local and global quality-aware features was proposed and applied with a GPR with the rational quadratic kernel function to obtain a perceptual quality estimator. The main idea behind the usage of local feature descriptor statistics was that these feature descriptors interpret local image regions from the human visual system’s viewpoint. The features were studied by taking into consideration their effects on the performance of perceptual quality estimation. The numerical comparison to 12 other state-of-the-art methods on three popular benchmark databases (CLIVE [45], KonIQ-10k [46], and SPAQ [47]) proved the superior performance of the proposed method. Our future work will involve boosting the quality-aware properties of the local feature descriptors by applying bio-inspired filters.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

In this paper, the following publicly available benchmark databases were used: 1. CLIVE: https://live.ece.utexas.edu/research/ChallengeDB/index.html (accessed on 16 April 2022); 2. KonIQ-10k: http://database.mmsp-kn.de/koniq-10k-database.html (accessed on 16 April 2022); 3. SPAQ: https://github.com/h4nwei/SPAQ (accessed on 16 April 2022). The source code of the proposed FLG-IQA method is available at: https://github.com/Skythianos/FLG-IQA (accessed on 12 June 2022).

Acknowledgments

We thank the academic editor and the anonymous reviewers for their careful reading of our manuscript and their many insightful comments and suggestions.

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BRIEF   binary robust independent elementary features
BRISKbinary robust invariant scalable keypoints
CGMcolor gradient magnitude
CNNconvolutional neural network
CPUcentral processing unit
DCFdark channel feature
DCTdiscrete cosine transform
DFdecision fusion
DSCdigital still camera
DSLRdigital single-lens reflex camera
FASTfeatures from accelerated segment test
FRfull-reference
GGDgeneralized Gaussian distribution
GLGrünwald–Letnikov
GMgradient magnitude
GPUgraphics processing unit
GPRGaussian process regression
IQAimage quality assessment
KROCCKendall’s rank-order correlation coefficient
MOSmean opinion score
MSCNmean subtracted contrast normalized
NIQEnaturalness image quality evaluator
NRno-reference
NSSnatural scene statistics
ORBoriented FAST and rotated BRIEF
PIQEperception-based image quality evaluator
PLCCPearson’s linear correlation coefficient
RBFradial basis function
RMrelative gradient magnitude
RMSroot mean square
ROrelative gradient orientation
RRreduced-reference
SPHNsmartphone
SROCCSpearman’s rank-order correlation coefficient
SVRsupport vector regressor

References

  1. Torr, P.H.; Zissermann, A. Performance characterization of fundamental matrix estimation under image degradation. Mach. Vis. Appl. 1997, 9, 321–333. [Google Scholar] [CrossRef]
  2. Zhao, Q. The Application of Augmented Reality Visual Communication in Network Teaching. Int. J. Emerg. Technol. Learn. 2018, 13, 57. [Google Scholar] [CrossRef]
  3. Shen, T.W.; Li, C.C.; Lin, W.F.; Tseng, Y.H.; Wu, W.F.; Wu, S.; Tseng, Z.L.; Hsu, M.H. Improving Image Quality Assessment Based on the Combination of the Power Spectrum of Fingerprint Images and Prewitt Filter. Appl. Sci. 2022, 12, 3320. [Google Scholar] [CrossRef]
  4. Esteban, O.; Birman, D.; Schaer, M.; Koyejo, O.O.; Poldrack, R.A.; Gorgolewski, K.J. MRIQC: Advancing the automatic prediction of image quality in MRI from unseen sites. PLoS ONE 2017, 12, e0184661. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Ma, K.; Yeganeh, H.; Zeng, K.; Wang, Z. High dynamic range image compression by optimizing tone mapped image quality index. IEEE Trans. Image Process. 2015, 24, 3086–3097. [Google Scholar]
  6. Goyal, B.; Gupta, A.; Dogra, A.; Koundal, D. An adaptive bitonic filtering based edge fusion algorithm for Gaussian denoising. Int. J. Cogn. Comput. Eng. 2022, 3, 90–97. [Google Scholar] [CrossRef]
  7. Zhai, G.; Min, X. Perceptual image quality assessment: A survey. Sci. China Inf. Sci. 2020, 63, 1–52. [Google Scholar] [CrossRef]
  8. Kamble, V.; Bhurchandi, K. No-reference image quality assessment algorithms: A survey. Optik 2015, 126, 1090–1097. [Google Scholar] [CrossRef]
  9. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  10. Liu, L.; Dong, H.; Huang, H.; Bovik, A.C. No-reference image quality assessment in curvelet domain. Signal Process. Image Commun. 2014, 29, 494–505. [Google Scholar] [CrossRef]
  11. Xue, W.; Mou, X.; Zhang, L.; Bovik, A.C.; Feng, X. Blind image quality assessment using joint statistics of gradient magnitude and Laplacian features. IEEE Trans. Image Process. 2014, 23, 4850–4862. [Google Scholar] [CrossRef] [PubMed]
  12. Min, X.; Zhai, G.; Gu, K.; Liu, Y.; Yang, X. Blind image quality estimation via distortion aggravation. IEEE Trans. Broadcast. 2018, 64, 508–517. [Google Scholar] [CrossRef]
  13. Mohammadi, P.; Ebrahimi-Moghadam, A.; Shirani, S. Subjective and objective quality assessment of image: A survey. arXiv 2014, arXiv:1406.7799. [Google Scholar]
  14. Yang, X.; Li, F.; Liu, H. A survey of DNN methods for blind image quality assessment. IEEE Access 2019, 7, 123788–123806. [Google Scholar] [CrossRef]
  15. Xu, L.; Lin, W.; Kuo, C.C.J. Visual Quality Assessment by Machine Learning; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  16. Venkatanath, N.; Praneeth, D.; Bh, M.C.; Channappayya, S.S.; Medasani, S.S. Blind image quality evaluation using perception based features. In Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Bombay, India, 27 February–1 March 2015; pp. 1–6. [Google Scholar]
  17. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  18. Zhang, L.; Zhang, L.; Bovik, A.C. A feature-enriched completely blind image quality evaluator. IEEE Trans. Image Process. 2015, 24, 2579–2591. [Google Scholar] [CrossRef] [Green Version]
  19. Wu, L.; Zhang, X.; Chen, H.; Wang, D.; Deng, J. VP-NIQE: An opinion-unaware visual perception natural image quality evaluator. Neurocomputing 2021, 463, 17–28. [Google Scholar] [CrossRef]
  20. Bhattacharyya, A. On a measure of divergence between two statistical populations defined by their probability distributions. Bull. Calcutta Math. Soc. 1943, 35, 99–109. [Google Scholar]
  21. Leonardi, M.; Napoletano, P.; Schettini, R.; Rozza, A. No Reference, Opinion Unaware Image Quality Assessment by Anomaly Detection. Sensors 2021, 21, 994. [Google Scholar] [CrossRef]
  22. Reinagel, P.; Zador, A.M. Natural scene statistics at the centre of gaze. Netw. Comput. Neural Syst. 1999, 10, 341. [Google Scholar] [CrossRef]
  23. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind image quality assessment: A natural scene statistics approach in the DCT domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef] [PubMed]
  24. Moorthy, A.K.; Bovik, A.C. A two-step framework for constructing blind image quality indices. IEEE Signal Process. Lett. 2010, 17, 513–516. [Google Scholar] [CrossRef]
  25. Moorthy, A.K.; Bovik, A.C. Blind image quality assessment: From natural scene statistics to perceptual quality. IEEE Trans. Image Process. 2011, 20, 3350–3364. [Google Scholar] [CrossRef] [PubMed]
  26. Tang, H.; Joshi, N.; Kapoor, A. Learning a blind measure of perceptual image quality. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 305–312. [Google Scholar]
  27. Wang, Q.; Chu, J.; Xu, L.; Chen, Q. A new blind image quality framework based on natural color statistic. Neurocomputing 2016, 173, 1798–1810. [Google Scholar] [CrossRef]
  28. Wu, J.; Lin, W.; Shi, G. Image quality assessment with degradation on spatial structure. IEEE Signal Process. Lett. 2014, 21, 437–440. [Google Scholar] [CrossRef]
  29. Freitas, P.G.; Akamine, W.Y.; Farias, M.C. No-reference image quality assessment based on statistics of local ternary pattern. In Proceedings of the 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal, 6–8 June 2016; pp. 1–6. [Google Scholar]
  30. Freitas, P.G.; Akamine, W.Y.; Farias, M.C. No-reference image quality assessment using orthogonal color planes patterns. IEEE Trans. Multimed. 2018, 20, 3353–3360. [Google Scholar] [CrossRef]
  31. Ojala, T.; Pietikäinen, M.; Harwood, D. A comparative study of texture measures with classification based on featured distributions. Pattern Recognit. 1996, 29, 51–59. [Google Scholar] [CrossRef]
  32. Garcia Freitas, P.; Da Eira, L.P.; Santos, S.S.; Farias, M.C.Q.d. On the Application LBP Texture Descriptors and Its Variants for No-Reference Image Quality Assessment. J. Imaging 2018, 4, 114. [Google Scholar] [CrossRef] [Green Version]
  33. Ye, P.; Doermann, D. No-reference image quality assessment based on visual codebook. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 3089–3092. [Google Scholar]
  34. Ye, P.; Doermann, D. No-reference image quality assessment using visual codebooks. IEEE Trans. Image Process. 2012, 21, 3129–3138. [Google Scholar]
  35. Ye, P.; Kumar, J.; Kang, L.; Doermann, D. Unsupervised feature learning framework for no-reference image quality assessment. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1098–1105. [Google Scholar]
  36. Kang, L.; Ye, P.; Li, Y.; Doermann, D. Convolutional neural networks for no-reference image quality assessment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1733–1740. [Google Scholar]
  37. Bianco, S.; Celona, L.; Napoletano, P.; Schettini, R. On the use of deep learning for blind image quality assessment. Signal Image Video Process. 2018, 12, 355–362. [Google Scholar] [CrossRef] [Green Version]
  38. Gao, F.; Yu, J.; Zhu, S.; Huang, Q.; Tian, Q. Blind image quality prediction by exploiting multi-level deep representations. Pattern Recognit. 2018, 81, 432–442. [Google Scholar] [CrossRef]
  39. Yang, S.; Jiang, Q.; Lin, W.; Wang, Y. Sgdnet: An end-to-end saliency-guided deep neural network for no-reference image quality assessment. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1383–1391. [Google Scholar]
  40. Zhang, L.; Shen, Y.; Li, H. VSI: A visual saliency-induced index for perceptual image quality assessment. IEEE Trans. Image Process. 2014, 23, 4270–4281. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Liu, X.; Van De Weijer, J.; Bagdanov, A.D. Rankiqa: Learning from rankings for no-reference image quality assessment. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1040–1049. [Google Scholar]
  42. Burges, C.; Shaked, T.; Renshaw, E.; Lazier, A.; Deeds, M.; Hamilton, N.; Hullender, G. Learning to rank using gradient descent. In Proceedings of the 22nd International Conference on Machine Learning, Bonn, Germany, 7–11 August 2005; pp. 89–96. [Google Scholar]
  43. Li, D.; Jiang, T.; Jiang, M. Norm-in-norm loss with faster convergence and better performance for image quality assessment. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 789–797. [Google Scholar]
  44. Celona, L.; Schettini, R. Blind quality assessment of authentically distorted images. JOSA A 2022, 39, B1–B10. [Google Scholar] [CrossRef]
  45. Ghadiyaram, D.; Bovik, A.C. Massive online crowdsourced study of subjective and objective picture quality. IEEE Trans. Image Process. 2015, 25, 372–387. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Lin, H.; Hosu, V.; Saupe, D. KonIQ-10K: Towards an ecologically valid and large-scale IQA database. arXiv 2018, arXiv:1803.08489. [Google Scholar]
  47. Fang, Y.; Zhu, H.; Zeng, Y.; Ma, K.; Wang, Z. Perceptual quality assessment of smartphone photography. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3677–3686. [Google Scholar]
  48. Guan, X.; Li, F.; He, L. Quality Assessment on Authentically Distorted Images by Expanding Proxy Labels. Electronics 2020, 9, 252. [Google Scholar] [CrossRef] [Green Version]
  49. Ding, Y. Visual Quality Assessment for Natural and Medical Image; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  50. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef]
  51. Varga, D. No-Reference Image Quality Assessment with Global Statistical Features. J. Imaging 2021, 7, 29. [Google Scholar] [CrossRef]
  52. Krig, S. Computer Vision Metrics; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  53. Rosten, E.; Drummond, T. Fusing points and lines for high performance tracking. In Proceedings of the Tenth IEEE International Conference on Computer Vision (ICCV’05) Volume 1, Beijing, China, 17–21 October 2005; Volume 2, pp. 1508–1515. [Google Scholar]
  54. Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988; Volume 15, pp. 10–5244. [Google Scholar]
  55. Bay, H.; Tuytelaars, T.; Gool, L.V. Surf: Speeded up robust features. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar]
  56. Leutenegger, S.; Chli, M.; Siegwart, R.Y. BRISK: Binary robust invariant scalable keypoints. In Proceedings of the 2011 International Conference on Computer Vision, Las Vegas, NV, USA, 18–21 July 2011; pp. 2548–2555. [Google Scholar]
  57. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 214–227. [Google Scholar]
  58. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Las Vegas, NV, USA, 18–21 July 2011; pp. 2564–2571. [Google Scholar]
  59. Shi, J.; Tomasi. Good features to track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar]
  60. Hu, M.K. Visual pattern recognition by moment invariants. IRE Trans. Inf. Theory 1962, 8, 179–187. [Google Scholar]
  61. Liu, L.; Hua, Y.; Zhao, Q.; Huang, H.; Bovik, A.C. Blind image quality assessment by relative gradient statistics and adaboosting neural network. Signal Process. Image Commun. 2016, 40, 1–15. [Google Scholar] [CrossRef]
  62. Moreno, P.; Bernardino, A.; Santos-Victor, J. Improving the SIFT descriptor with smooth derivative filters. Pattern Recognit. Lett. 2009, 30, 18–26. [Google Scholar] [CrossRef] [Green Version]
  63. Prewitt, J.M. Object enhancement and extraction. Pict. Process. Psychopictorics 1970, 10, 15–19. [Google Scholar]
  64. Marr, D.; Hildreth, E. Theory of edge detection. Proc. R. Soc. London Ser. B. Biol. Sci. 1980, 207, 187–217. [Google Scholar]
  65. Huang, J.S.; Tseng, D.H. Statistical theory of edge detection. Comput. Vis. Graph. Image Process. 1988, 43, 337–346. [Google Scholar] [CrossRef]
  66. Jenadeleh, M. Blind Image and Video Quality Assessment. Ph.D. Thesis, University of Konstanz, Konstanz, Germany, 2018. [Google Scholar]
  67. Kurimo, E.; Lepistö, L.; Nikkanen, J.; Grén, J.; Kunttu, I.; Laaksonen, J. The effect of motion blur and signal noise on image quality in low light imaging. In Proceedings of the Scandinavian Conference on Image Analysis, Oslo, Norway, 15–18 June; pp. 81–90.
  68. Crété-Roffet, F.; Dolmiere, T.; Ladret, P.; Nicolas, M. The blur effect: Perception and estimation with a new no-reference perceptual blur metric. In Proceedings of the SPIE Electronic Imaging Symposium Conf Human Vision and Electronic Imaging, San Jose, CA, USA, 29 January 29–1 February 2007. [Google Scholar]
  69. Choi, S.Y.; Luo, M.R.; Pointer, M.R.; Rhodes, P.A. Predicting perceived colorfulness, contrast, naturalness and quality for color images reproduced on a large display. In Proceedings of the Color and Imaging Conference, Society for Imaging Science and Technology, Portland, OR, USA, 10–14 November 2008; Volume 2008, pp. 158–164. [Google Scholar]
  70. Hasler, D.; Suesstrunk, S.E. Measuring colorfulness in natural images. In Proceedings of the Human Vision and Electronic Imaging VIII: International Society for Optics and Photonics, Santa Clara, CA, USA, 21–24 January 2003; Volume 5007, pp. 87–95. [Google Scholar]
  71. Tang, X.; Luo, W.; Wang, X. Content-based photo quality assessment. IEEE Trans. Multimed. 2013, 15, 1930–1943. [Google Scholar] [CrossRef] [Green Version]
  72. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  73. Kukkonen, H.; Rovamo, J.; Tiippana, K.; Näsänen, R. Michelson contrast, RMS contrast and energy of various spatial stimuli at threshold. Vis. Res. 1993, 33, 1431–1436. [Google Scholar] [CrossRef]
  74. Michelson, A.A. Studies in Optics; Courier Corporation: Chelmsford, MA, USA, 1995. [Google Scholar]
  75. Matkovic, K.; Neumann, L.; Neumann, A.; Psik, T.; Purgathofer, W. Global contrast factor-a new approach to image contrast. In Proceedings of the Eurographics Workshop on Computational Aesthetics in Graphics, Visualization and Imaging 2005, Girona, Spain, 18–20 May 2005; pp. 159–167. [Google Scholar]
  76. Tsai, D.Y.; Lee, Y.; Matsuyama, E. Information entropy measure for evaluation of image quality. J. Digit. Imaging 2008, 21, 338–347. [Google Scholar] [CrossRef] [Green Version]
  77. Petrovic, V.S.; Xydeas, C.S. Gradient-based multiresolution image fusion. IEEE Trans. Image Process. 2004, 13, 228–237. [Google Scholar] [CrossRef]
  78. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  79. Kobayashi, T.; Otsu, N. Image feature extraction using gradient local auto-correlations. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 346–358. [Google Scholar]
  80. Chiberre, P.; Perot, E.; Sironi, A.; Lepetit, V. Detecting Stable Keypoints from Events through Image Gradient Prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1387–1394. [Google Scholar]
  81. Liu, A.; Lin, W.; Narwaria, M. Image quality assessment based on gradient similarity. IEEE Trans. Image Process. 2011, 21, 1500–1512. [Google Scholar]
  82. Gorenflo, R.; Mainardi, F. Fractional calculus. In Fractals and Fractional Calculus in Continuum Mechanics; Springer: Berlin/Heidelberg, Germany, 1997; pp. 223–276. [Google Scholar]
  83. Sabatier, J.; Agrawal, O.P.; Machado, J.T. Advances in Fractional Calculus; Springer: Berlin/Heidelberg, Germany, 2007; Volume 4. [Google Scholar]
  84. Motłoch, S.; Sarwas, G.; Dzieliński, A. Fractional Derivatives Application to Image Fusion Problems. Sensors 2022, 22, 1049. [Google Scholar] [CrossRef] [PubMed]
  85. Varga, D. Full-Reference Image Quality Assessment Based on Grünwald–Letnikov Derivative, Image Gradients, and Visual Saliency. Electronics 2022, 11, 559. [Google Scholar] [CrossRef]
  86. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  87. Kira, K.; Rendell, L.A. A practical approach to feature selection. In Machine Learning Proceedings 1992; Elsevier: Amsterdam, The Netherlands, 1992; pp. 249–256. [Google Scholar]
  88. Robnik-Šikonja, M.; Kononenko, I. An adaptation of Relief for attribute estimation in regression. In Proceedings of the Fourteenth International Conference on Machine Learning (ICML’97), San Francisco, CA, USA, 8–12 July 1997; Volume 5, pp. 296–304. [Google Scholar]
  89. Chen, X.; Zhang, Q.; Lin, M.; Yang, G.; He, C. No-reference color image quality assessment: From entropy to perceptual quality. EURASIP J. Image Video Process. 2019, 2019, 77. [Google Scholar] [CrossRef] [Green Version]
  90. Li, Q.; Lin, W.; Fang, Y. No-reference quality assessment for multiply-distorted images in gradient domain. IEEE Signal Process. Lett. 2016, 23, 541–545. [Google Scholar] [CrossRef]
  91. Ou, F.Z.; Wang, Y.G.; Zhu, G. A novel blind image quality assessment method based on refined natural scene statistics. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–29 September 2019; pp. 1004–1008. [Google Scholar]
  92. Liu, L.; Liu, B.; Huang, H.; Bovik, A.C. No-reference image quality assessment based on spatial and spectral entropies. Signal Process. Image Commun. 2014, 29, 856–863. [Google Scholar] [CrossRef]
Figure 1. Empirical MOS distributions in the applied IQA databases: (a) CLIVE [45], (b) KonIQ-10k [46], and (c) SPAQ [47].
Figure 1. Empirical MOS distributions in the applied IQA databases: (a) CLIVE [45], (b) KonIQ-10k [46], and (c) SPAQ [47].
Jimaging 08 00173 g001
Figure 2. Workflow of the proposed NR-IQA algorithm.
Figure 2. Workflow of the proposed NR-IQA algorithm.
Jimaging 08 00173 g002
Figure 3. Comparison of the statistics of local feature descriptors as quality-aware features in CLIVE [45]. Median SROCC values were measured over 100 random train–test splits. (a) RGB image, SVR, (b) RGB image, GPR, (c) filtered image, SVR, (d) filtered image, GPR.
Figure 3. Comparison of the statistics of local feature descriptors as quality-aware features in CLIVE [45]. Median SROCC values were measured over 100 random train–test splits. (a) RGB image, SVR, (b) RGB image, GPR, (c) filtered image, SVR, (d) filtered image, GPR.
Jimaging 08 00173 g003
Figure 4. Performance in terms of SROCC of the proposed FLG-IQA in cases when a given feature is removed from the proposed feature vector. The performance of the entire feature vector is indicated by ‘X’. Median SROCC values were measured on CLIVE [45] after 100 random train–test splits.
Figure 4. Performance in terms of SROCC of the proposed FLG-IQA in cases when a given feature is removed from the proposed feature vector. The performance of the entire feature vector is indicated by ‘X’. Median SROCC values were measured on CLIVE [45] after 100 random train–test splits.
Jimaging 08 00173 g004
Figure 5. Results of the RReliefF algorithm on the features extracted from the images of CLIVE [45]. (a) k = 1 nearest neighbours, (b) k = 3 nearest neighbours, (c) k = 5 nearest neighbours, (d) k = 7 nearest neighbours.
Figure 5. Results of the RReliefF algorithm on the features extracted from the images of CLIVE [45]. (a) k = 1 nearest neighbours, (b) k = 3 nearest neighbours, (c) k = 5 nearest neighbours, (d) k = 7 nearest neighbours.
Jimaging 08 00173 g005
Figure 6. Ground truth scores versus predicted scores in (a) CLIVE [45] and (b) KonIQ-10k [46] test sets.
Figure 6. Ground truth scores versus predicted scores in (a) CLIVE [45] and (b) KonIQ-10k [46] test sets.
Jimaging 08 00173 g006
Figure 7. Normalized ground truth scores versus normalized predicted score scatter plot of FLG-IQA in the cross-database test.
Figure 7. Normalized ground truth scores versus normalized predicted score scatter plot of FLG-IQA in the cross-database test.
Jimaging 08 00173 g007
Table 1. Summary about the applied benchmark IQA databases with authentic distortions. DSLR: digital single-lens reflex camera. DSC: digital still camera. SPHN: smartphone.
Table 1. Summary about the applied benchmark IQA databases with authentic distortions. DSLR: digital single-lens reflex camera. DSC: digital still camera. SPHN: smartphone.
AttributeCLIVE [45]KonIQ-10k [46]SPAQ [47]
#Images116210,07311,125
Resolution 500 × 500 1024 × 768 4000 × 4000
#Subjects81001,467600
#Annotations14001,200,000186,400
Scale of quality scores0–1001–50–100
Subjective methodologyCrowdsourcingCrowdsourcingLaboratory
Types of camerasDSLR/DSC/SPHNDSLR/DSC/SPHNSPHN
Year of publication201720182020
Table 2. The applied computer configuration of the experiments.
Table 2. The applied computer configuration of the experiments.
Computer modelSTRIX Z270H Gaming
Operating systemWindows 10
CPUIntel(R) Core(TM) i7-7700K CPU 4.20 GHz (8 cores)
Memory15 GB
GPUNvidia GeForce GTX 1080
Table 4. Mean values of perceptual features in CLIVE [45] with respect to five equal MOS intervals.
Table 4. Mean values of perceptual features in CLIVE [45] with respect to five equal MOS intervals.
0 MOS < 20 20 MOS < 40 40 MOS < 60 60 MOS < 80 80 MOS 100
Blur0.4120.3620.3150.2850.329
Colorfulness0.0460.0380.0420.0450.072
Chroma15.51013.68114.99515.40921.977
Color gradient-mean92.801116.884154.651189.795196.287
Color gradient-std132.693163.876207.837244.420235.855
DCF0.2170.2110.1970.2200.192
Michelson contrast2.8042.8322.9112.9372.953
RMS contrast0.2010.2010.2190.2220.223
GCF5.3045.4886.6026.2646.796
Entropy6.8326.9857.1827.4137.583
Table 5. Standard deviation values of perceptual features in CLIVE [45] with respect to five equal MOS intervals.
Table 5. Standard deviation values of perceptual features in CLIVE [45] with respect to five equal MOS intervals.
0 MOS < 20 20 MOS < 40 40 MOS < 60 60 MOS < 80 80 MOS 100
Blur0.1090.0960.0750.0670.093
Colorfulness0.0500.0330.0370.0390.049
Chroma12.6988.1439.6808.92711.720
Color gradient-mean45.48066.16489.76296.28399.800
Color gradient-std58.23671.18782.10484.17978.250
DCF0.1410.1220.1170.1150.105
Michelson contrast0.3280.2520.1730.1430.140
RMS contrast0.0800.0680.0650.0560.051
GCF1.9341.6651.7611.8571.746
Entropy1.0190.9660.7480.5320.227
Table 6. Ablation study on CLIVE [45] database. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits.
Table 6. Ablation study on CLIVE [45] database. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits.
SVRGPR
MethodPLCCSROCCKROCCPLCCSROCCKROCC
Feature descriptors, RGB image0.5180.4840.3370.5780.5230.364
Feature descriptors, filtered image0.5290.4880.3380.5820.5270.364
Hu invariant moments0.3020.2950.1990.3280.3200.219
Perceptual features0.6070.5880.4200.6260.5980.425
GL and gradient statistics0.5280.4920.3430.5410.4950.343
All0.6360.6040.4280.6850.6440.466
Table 7. Comparison to the state-of-the-art in CLIVE [45] and KonIQ-10k [46] databases. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits. The best results are in bold and the second-best results are underlined.
Table 7. Comparison to the state-of-the-art in CLIVE [45] and KonIQ-10k [46] databases. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits. The best results are in bold and the second-best results are underlined.
CLIVE [45]KonIQ-10k [46]
MethodPLCCSROCCKROCCPLCCSROCCKROCC
BLIINDS-II [23]0.4730.4420.2910.5740.5750.414
BMPRI [12]0.5410.4870.3330.6370.6190.421
BRISQUE [9]0.5240.4970.3450.7070.6770.494
CurveletQA [10]0.6360.6210.4210.7300.7180.495
DIIVINE [25]0.6170.5800.4050.7090.6930.471
ENIQA [89]0.5960.5640.3760.7610.7450.544
GRAD-LOG-CP [11]0.6070.6040.3830.7050.6960.501
GWH-GLBP [90]0.5840.5590.3950.7230.6980.507
NBIQA [91]0.6290.6040.4270.7710.7490.515
OG-IQA [61]0.5450.5050.3640.6520.6350.447
PIQE [16]0.1720.1080.0810.2080.2460.172
SSEQ [92]0.4870.4360.3090.5890.5720.423
FLG-IQA0.6850.6440.4660.8060.7710.578
Table 8. Comparison to the state-of-the-art in the SPAQ [47] database. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits. The best results are in bold and the second-best results are underlined.
Table 8. Comparison to the state-of-the-art in the SPAQ [47] database. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits. The best results are in bold and the second-best results are underlined.
MethodPLCCSROCCKROCC
BLIINDS-II [23]0.6760.6750.486
BMPRI [12]0.7390.7340.506
BRISQUE [9]0.7260.7200.518
CurveletQA [10]0.7930.7740.503
DIIVINE [25]0.7740.7560.514
ENIQA [89]0.8130.8040.603
GRAD-LOG-CP [11]0.7860.7820.572
GWH-GLBP [90]0.8010.7960.542
NBIQA [91]0.8020.7930.539
OG-IQA [61]0.7260.7240.594
PIQE [16]0.2110.1560.091
SSEQ [92]0.7450.7420.549
FLG-IQA0.8500.8450.640
Table 9. Comparison to the state-of-the-art. Direct and weighted average PLCC, SROCC, and KROCC are reported based on the results measured in CLIVE [45], KonIQ-10k [46], and SPAQ [47]. The best results are in bold and the second-best results are underlined.
Table 9. Comparison to the state-of-the-art. Direct and weighted average PLCC, SROCC, and KROCC are reported based on the results measured in CLIVE [45], KonIQ-10k [46], and SPAQ [47]. The best results are in bold and the second-best results are underlined.
Direct AverageWeighted Average
MethodPLCCSROCCKROCCPLCCSROCCKROCC
BLIINDS-II [23]0.5740.5640.3970.6200.6180.443
BMPRI [12]0.6390.6130.4200.6830.6690.459
BRISQUE [9]0.6520.6310.4520.7070.6890.498
CurveletQA [10]0.7200.7040.4730.7560.7410.495
DIIVINE [25]0.7000.6760.4630.7370.7180.489
ENIQA [89]0.7230.7040.5080.7780.7650.565
GRAD-LOG-CP [11]0.6990.6940.4850.7400.7340.530
GWH-GLBP [90]0.7030.6840.4810.7550.7400.519
NBIQA [91]0.7340.7150.4940.7790.7630.522
OG-IQA [61]0.6410.6210.4680.6830.6730.516
PIQE [16]0.1970.1700.1150.2080.1940.127
SSEQ [92]0.6070.5830.4270.6610.6500.480
FLG-IQA0.7800.7530.5610.8220.8010.603
Table 10. Results of the significance tests. Symbol 1 ( 1 ) denotes that the proposed FLG-IQA algorithm is significantly ( 95 % confidence interval) better (worse) than the NR-IQA algorithm in the row on the IQA benchmark database in the column.
Table 10. Results of the significance tests. Symbol 1 ( 1 ) denotes that the proposed FLG-IQA algorithm is significantly ( 95 % confidence interval) better (worse) than the NR-IQA algorithm in the row on the IQA benchmark database in the column.
MethodCLIVE [45]KonIQ-10k [46]SPAQ [47]
BLIINDS-II [23]111
BMPRI [12]111
BRISQUE [9]111
CurveletQA [10]111
DIIVINE [25]111
ENIQA [89]111
GRAD-LOG-CP [11]111
GWH-GLBP [90]111
NBIQA [91]111
OG-IQA [61]111
PIQE [16]111
SSEQ [92]111
Table 11. Results of the cross-database test. The examined and the proposed methods were trained on KonIQ-10k [46] and tested on CLIVE [45]. The best results are in bold and the second-best results are underlined.
Table 11. Results of the cross-database test. The examined and the proposed methods were trained on KonIQ-10k [46] and tested on CLIVE [45]. The best results are in bold and the second-best results are underlined.
MethodPLCCSROCCKROCC
BLIINDS-II [23]0.1070.0900.063
BMPRI [12]0.4530.3890.298
BRISQUE [9]0.5090.4600.310
CurveletQA [10]0.4960.5050.347
DIIVINE [25]0.4790.4340.299
ENIQA [89]0.4280.3860.272
GRAD-LOG-CP [11]0.4270.3840.261
GWH-GLBP [90]0.4800.4790.328
NBIQA [91]0.5030.5090.284
OG-IQA [61]0.4420.4270.289
SSEQ [92]0.2700.2560.170
FLG-IQA0.6130.5710.399
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Varga, D. No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features. J. Imaging 2022, 8, 173. https://doi.org/10.3390/jimaging8060173

AMA Style

Varga D. No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features. Journal of Imaging. 2022; 8(6):173. https://doi.org/10.3390/jimaging8060173

Chicago/Turabian Style

Varga, Domonkos. 2022. "No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features" Journal of Imaging 8, no. 6: 173. https://doi.org/10.3390/jimaging8060173

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop