Next Article in Journal
Transfer Learning Based Semantic Segmentation for 3D Object Detection from Point Cloud
Next Article in Special Issue
Uplink vs. Downlink: Machine Learning-Based Quality Prediction for HTTP Adaptive Video Streaming
Previous Article in Journal
Interference Analysis for mmWave Automotive Radar Considering Blockage Effect
Previous Article in Special Issue
Synchronization of Acoustic Signals for Steganographic Transmission
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Salient Region Guided Blind Image Sharpness Assessment

1
Key Laboratory of Convergent Media and Intelligent Technology (Communication University of China), Ministry of Education, Beijing 100024, China
2
School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
3
School of Aerospace Science and Technology, Xidian University, Xi’an 710126, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2021, 21(12), 3963; https://doi.org/10.3390/s21123963
Submission received: 4 April 2021 / Revised: 31 May 2021 / Accepted: 4 June 2021 / Published: 8 June 2021
(This article belongs to the Collection Machine Learning for Multimedia Communications)

Abstract

:
Salient regions provide important cues for scene understanding to the human vision system. However, whether the detected salient regions are helpful in image blur estimation is unknown. In this study, a salient region guided blind image sharpness assessment (BISA) framework is proposed, and the effect of the detected salient regions on the BISA performance is investigated. Specifically, three salient region detection (SRD) methods and ten BISA models are jointly explored, during which the output saliency maps from SRD methods are re-organized as the input of BISA models. Consequently, the change in BISA metric values can be quantified and then directly related to the difference in BISA model inputs. Finally, experiments are conducted on three Gaussian blurring image databases, and the BISA prediction performance is evaluated. The comparison results indicate that salient region input can help achieve a close and sometimes superior performance to a BISA model over the whole image input. When using the center region input as the baseline, the detected salient regions from the saliency optimization from robust background detection (SORBD) method lead to consistently better score prediction, regardless of the BISA model. Based on the proposed hybrid framework, this study reveals that saliency detection benefits image blur estimation, while how to properly incorporate SRD methods and BISA models to improve the score prediction will be explored in our future work.

1. Introduction

Human vision system (HVS) is verified as sensitive to the most conspicuous regions in a visual scene, and selective attention is paid to those regions of interest [1,2]. At each moment, the human brain needs to tackle massive messages. Since the amount of information sensed is too high to be completely processed, the brain prioritizes salient regions as the most important cues for follow-up analysis [1]. Emperical and computational studies have reported evidence of a saliency map formed in cortical brain areas or before the primary visual cortex [3,4,5], and this map is used to guide human visual attention to the most relevant regions. This finding inspires increasing applications in object segmentation and pattern analysis for scene understanding [6,7,8].
To identify the most informative and useful regions or objects in an image, salient region detection (SRD) is becoming an increasingly hot topic in the field of computer vision, and many SRD methods have been proposed in the last two decades [9,10,11]. According to the backbone techniques, these SRD methods could be grouped into classic methods and deep-learning-based methods [11], and the former can be further divided into intrinsic- and extrinsic-cue-based methods in terms of the exploited attributes. Regarding the intrinsic cues, an input image is explored to highlight the potential target regions through different aggregation or optimization algorithms in multi-scale or hierarchical analysis. To enhance the feasibility and suppress distractors, various hypotheses are additionally integrated, such as center prior, color prior, objectness prior, global contrast, local contrast and center-surround contrast [12,13,14]. For instance, the most influential method comes from the biologically inspired “center-surround difference” operation, which estimates saliency maps via multi-scale analysis of image color, intensity and orientation [15]. Using a similar framework, a “center-surround divergence” method is designed in an information-theoretic way, and two distributions of visual feature occurrence, derived from a center region and its surrounding region, are compared and finally fused into a single saliency map [16]. In accordance with the understanding of HVS behavior, a center–surround interaction is developed by combining early visual feature extraction, contrast sensitivity analysis, perception decomposition and visual masking, followed by perceptual grouping to bind visual features into a meaningful and high-level structure for saliency detection [17]. However, intrinsic cues might be insufficient, and the resultant performance unsatisfactory, and, thus, extrinsic cues are incorporated, such as ground truth annotation, 3D depth maps and saliency co-occurrence, to overcome the challenges in saliency map estimation [18,19,20]. In [21], the normalized depth prior and the global-context surface orientation prior are proposed, and both priors are integrated with several intrinsic cues for the RGB-D data analysis. In [22], saliency cues are extracted from color images and corresponding depth maps, and, in particular, the feature contrast is derived from color contrast and depth contrast, and the spatial bias is extended from the center bias, both of which are combined for saliency localization and segmentation.
It is unsurpising that deep learning has updated the state-of-the-art in many fields, including, but not limited to, computer vision, precision medicine, and remote sensing [10,23,24,25,26,27]. Since 2015, deep-learning-based methods have also been applied in SRD and achieved promising results [9,10,11]. As one of the most popular tools, deep learning has revolutionized image representation through multi-level and multi-scale abstraction. The deep-learning-based SRD methods can be further categorized from various perspectives, such as backbone network archhitectures, the level of supervision, and learning paradigm [10]. Li and Yu propose a hybrid framework that integrates handcrafted and deep features for saliency detection, and the backbone network is a multi-scale convolutional network, followed by a conditional random field to fuse multiple saliency maps [28]. Based on the maximization of a posteriori principle, Zhang et al. design a novel subset optimization framework to generate a compact set of detection windows of noisy saliency maps, and the proposal of noisy maps is generated by a convolutional neural network (CNN) [29]. For the streamless approximation of saliency maps, fully convolutional networks (FCNs) are used, which perform end-to-end salient object detection. Hu et al. developed a network architecture to learn a level set function to refine the boundaries of the saliency map, and an additional superpixel-guided filter is extended for a more compact saliency output [30]. Besides CNN and FCN, other deep learning architectures, such as an autoencoder, are applied [31,32].
Automatic localization of the most visually relevant areas is useful in many applications, while it is unknown whether salient regions could improve blind image sharpness assessment (BISA). BISA is a distortion-specific task for image quality scoring and requires no reference images [33]. Among the various kinds of distortion, image blurring is more common in daily life. It might be derived from out-of-focus imaging and relative target motion, which degrades image quality and causes difficulties in scene understanding. Since image blur is frequently accompanied by changes in the edges in the spatial domain, and the according high-frequency attenuation, studies are looking into analysis of the edges, contours and image structures. Marziliano et al. compute the spread of edges in the spatial domain and measure the mean edge width (MEW) as a sharpness estimator [34]. Karam et al. first propose the concept of just-noticeable blur via the measure of local contrast in an image, then the concept is incorporated into a probability summation model to score the blurness of the image [35], and later into the cumulative probability of blur detection (CPBD) [36]. Chandler et al. quantify the attenuation of high-frequency messages in the spectral domain via the magnitude spectrum slop and the impact of the contrast component in the spatial domain via the total variation, and propose the spectral and spatial sharpness (S3) indicator [37]. Later, they weight the log-energies of the wavelet sub-band after discrete wavelet transform and design the fast image sharpness (FISH) estimator [38]. Sang et al. estimate the image blurring through the shape of the singular value curve (SVC), since singular values are decreased due to the extent of image blurring [39]. It is also observed that image blur disrupts the structue of local phase coherence (LPC), and then Hassen et al. implement LPC as a sharpness index [40]. Bahrami and Kot consider the maximum local variation (MLV) of each pixel in a small image patch for image blur measure [41]. Gu et al. advance a novel autoregressive-based image sharpness metric by analyzing the autoregressive parameter space of images [42]. Based on an over-complete dictionary of edge patterns, trained with high-quality natural images, Li and his colleagues propose a sparse representation based image sharpness (SPARISH) metric [43]. They also represent a noticeable blur with the magnitudes of discrete orthogonal moments, and design the blind image blur evaluation (BIBLE) model [44]. Furthermore, they collect a total of eleven sharpness-aware features in multi-scale spatial and spectral domains, and the robust image sharpness evaluation (RISE) model is formulated [45]. Sun et al. simulate the impact of viewing distance on blur distortion, and calculate the distribution characteristics of the local maximum gradient of multi-resolution images in the spatial domain [46]. Cai et al. design an interesting procedure that reblurs the orignial blurred image [47], and the global sharpness is estimated through inter-resolution self-similarities, since the discrepancy between an image and its reblurred version indicates the extent of blur in the image. CNN has also been used for BISA tasks. Limited by a small number of training samples, Yu et al. propose a hybrid framework to incorporate a shallow CNN for representation learning of image sharpness, and the prediction performance is further boosted by different kinds of numerical regressors [25,48]. Their further kernel visualization indicates that image sharpness is related to image edges and structures [25]. Li et al. design semantic feature aggregation for blur-dominated distortions, which alleviates the impact of image content variation, and the results suggest that deep semantic features might address image content variation in image-quality assessment [49]. In addition, Hosseini et al. synthesize the HVS response as a linear combination of finite impulse response derivative filters, and the falloff of high band frequency magnitudes is stimulated in natural imaging paradigm [50].
Several studies have used saliency detection for image quality estimation. In [51], the visual saliency map is set as a weighting function with the purpose of improving the existing image quality metrics. In [52], an image is divided into regions of interest and background regions, and computational metrics from these regions are pooled to score the image quality. In [53], visual saliency is usede to compute the local quality map and then as a weighting function to reflect the importance of a local region for final quality assessment. In [54], visual saliency information is introduced into image quality metrics, and the correlation performance can be further boosted between the predicted scores and the subjective scores. In [55], image quality degradation is modelled as saliency deviation, and both the saliency map and quality information are embedded into the proposed metric. In these studies [51,52,53,54,55], saliency maps are used as a weighting function, quantitative features, or degradation results in an implicit manner to improve the prediction of image quality.
However, the effect of salient regions on image quality assessment is still unknown. To better analyze this effect, this study concerns a specific distortion of Gaussian blurring, and it excludes the effect of other kinds of distortion variations. A hybrid salient-region-guided BISA framework is designed. This reorganizes the output of an SRD method as the input of a BISA model and, thus, the change in the BISA metric value can be directly related to the change in the BISA input. In summary, the contributions of this study are as follows:
1.
A salient region guided BISA framework is designed. It aims to observe the effect of salient region input on image blur estimation;
2.
Based on the proposed framework, the effect of SRD results on the BISA prediction is investigated. Specifically, this involves three SRD methods combined with 10 BISA models, and the experiment is conducted on three Gaussian blurring image datasets;
3.
Experimental results reveal that salient regions can help a BISA model achieve a comparable performance to one using whole-image input. In particular, the saliency optimization from robust background detection (SORBD) method is suggested.
The remainder of this paper is organized as follows. The details of the involved datasets, SRD methods, BISA models, and experiment design are presented in Section 2. Experimental results are reported in Section 3. The findings and limitations of this study are summarized in Section 5. Finally, the conclusions are included in Section 5.

2. Materials and Methods

2.1. Data Collection

Synthesized Gaussian blurring images are collected from three widely used datasets. CSIQ [56] and LIVE [57] contain 30 and 29 pristine images, respectively. In these datasets, images are distorted by five different levels of blurring, and image quality is reported using differential mean opinion scores. The third dataset TID13 [58] has 25 reference images. Each image in the database is degraded by four different levels of Gaussian blurring, and image quality is reported using mean opinion scores.
The pristine images are from a set of source images and reflect adequate diversity in terms of image content, including pictures of faces, people, animals, close-up shots, wide-angle shots, nature scenes, man-made objects, images with distinct foreground/background configurations, and images without any specific object of interest. Figure 1 demonstrates several representative images in these datasets. Interestingly, it was found that the refernce images in CSIQ are distinctive from those in the other two databases, while more than ten images have the same or similar content as that shared in LIVE and TID13; for instance, the lighthouse, as shown in Figure 1.
In the three databases, Gaussian blurring is synthesized intentionally. To each pristine image, its R, G, and B channels are filtered using a circular-symmetric 2-D Gaussian kernel of standard deviation ( δ ) of Gaussian distribution [57], as shown in Equation (1), where c is a color channel, δ stands for the standard deviation of the distribution, x and y are the location indices of pixels. The three color components of an image are blurred using a same kernel, while, to different images, the values of δ change in different ranges.
G 2 D c ( x , y , δ ) = 1 2 π δ 2 exp ( x 2 + y 2 2 δ 2 ) ,

2.2. Evaluated SRD Methods

Three SRD methods are evaluated. One detects salient regions with diffusion process on a 2-layer sparse graph (DPLSG) [59]. It constructs the diffusion matrix using sparse graph and obtains the seed vector from the spatial variance of superpixel clusters. DPLSG computes two coarse saliency maps via the foreground and the background seed vectors, and the final saliency map is generated using a manifold ranking diffusion method. The second method, the saliency optimization from robust background detection (SORBD) [60], designs a robust background measure to compute boundary connectivity, which characterizes the spatial layout of image regions. Moreover, it integrates multiple low-level cues into the optimization framework with an intuitive geometrical interpretation. The last method incorporates both salient region localization and image segmentation (SRIS) into a hybrid framework and aims to overcome image noise and artifacts [61]. It uses an adaptive level-set evolution protocol and an adaptive weight function is embedded. SRIS can balance both the internal and the external functions and tackle different kinds of image distortion, such as inhomogeneity and noise.
The reasons for using the three SRD methods are manifold. Firstly, the three methods consider saliency detection from various perspectives. DPLSG considers sparse graph and manifold ranking diffusion [59], SORBD computes the boundary connectivity and integrates low-level geometric cues for optimization [60], and SRIS designs an adaptive level-set evolutional protocol and an adaptive weight function to tackle different kinds of image distortion [61]. Secondly, the SRD methods are able to estimate the saliency maps of the Gaussian blurring images. Offline experiments were also conducted on two deep-learning-based SRD methods (BASNet [62] and PiCANet [63]). Since image contextures or boundaries are important, both methods cause failure cases on highly blurry images. Last, but not least, the codes of the three SRD methods are avaialable online, and the effort involved in algorithm implementation is relieved.

2.3. Involved BISA Models

This study involves 10 BISA models, including MEW [34], CPBD [36], S3 [37], FISH [38], SVC [39], LPC [40], MLV [41], SPARISH [43], BIBLE [44], and RISE [45]. The methods are mainly implemented with MATLAB. Codes are available online and the methods are tested without any modifications.
Notably, deep-learning-based BISA models were also explored [25,49]. Deep learning requires a large number of samples for model training. To tackle this problem, Yu et al. [25] randomly select hundreds of patches from an image, and the score of each patch is paired with that of the image. Li et al. [49] represent an image by using multiple overlapping patches with the purpose of avoiding introducing unwanted geometric deformation and retaining the model performance. However, the sizes of salient regions are decreased, and some key parameters in deep learning-based models become meaningless, such as the patch number [25] and the stride [49]. In other words, the deep-learning-based BISA models [25,49] cannot be well trained and, consequently, the effect of salient regions on deep-learning-based BISA performance becomes ambiguous. Therefore, deep-learning-based BISA models are not considered in this study.

2.4. The Proposed Hybrid Framework

The proposed hybrid framework is shown in Figure 2. It uses SRD methods to generate saliency maps to guide BISA tasks (the arrows with solid line). An SRD method computes the saliency map (B) of an input image (A). Then, the saliency map is binarized into foreground masks, and a minimum outer rectangle is generated accordingly (D). Next, two kinds of salienct-region-guided BISA experiment are conducted. One takes the image region in the rectangle as the input (G), and the other uses the binary mask with the largest area (E) to generate its outer rectangle and to obtain the image region as the input (H). For fair comparison, two additional experiments are carried out. One uses the source image as the input (I), and the other takes the center area (C) of the image as the input (F), since the center region, as a center prior, assumes salient objects are more likely to be located at the center of an image [11]. Therefore, to each BISA model, four kinds of experiments are conducted by using different image inputs (F-I). Finally, by using different kinds of image input, including center region (CR), salient region (SR), major salient region (mSR) and whole image (WI), a BISA model predicts the scores as { s C R } , { s S R } , { s m S R } and { s W I } , respectively.

2.5. Performance Evaluation

Two metrics, Pearson linear correlation coefficient (PLCC) and Spearman rank-order correlation coefficient (SRCC), are used to evaluate the performance of BISA models. The former measures the prediction accuracy, and the latter assesses the prediction monotonicity. The values of both metrics range between 0 and 1, and a higher value indicates a better prediction.
Beefore quantifying the prediction performance, a nonlinear fitting is routinely used to map the predicted scores to the range of corresponding subjective ratings. There are two fitting methods. Since little difference is found [64], this study uses the five-parameter-based curve fitting method, shown below,
Q ( s ) = q 1 ( 1 2 1 1 + exp q 2 ( s q 3 ) ) + q 4 s + q 5 ,
where s and Q ( s ) correpond to the input scores (such as { s C R } ) and the mapped scores, and q i ( i = 1 , 2 , , 5 ) are determined during the score fitting.

2.6. Experiment Design

The experiment design is described below:
1.
Experiments are conducted on different combinations of SRD methods and BISA models, and the prediction performance is illustrated;
2.
On each dataset, the BISA model with the largest score drop between the WI input and the CR input is investigated. The procedure consists of three steps. First, compute the difference in PLCC values derived from the WI and the CR input, and the BISA model with the largest score drop is determined. Then, to the BISA model, absolute difference of prediction scores between the WI input and the CR input is shown, and the images causing the three largest score drop are determined. In the end, the WI, the CR, the SR, and the mSR are also demonstrated;
3.
The results from the CR input are set as the baseline, and the difference obtained by subtracting the value (PLCC and SRCC) of the CR input from that of the corresponding SR and mSR input is shown;
4.
The size of salient regions (SR and mSR) is analyzed and illustrated to verify the effectiveness of SRD methods;
5.
The average running time of SRD methods and BISA models is reported separately.

2.7. Software and Platform

Software runs on a Windows system. The system is embedded with Intel(R) Core(TM) i7-8700 CPU (3.20 GHz), 16 GB RAM and one GPU card (Nvidia GeForce GTX 1070, Nvidia, Santa Clara, California, United States). The SORBD method is implemented by Python 3.6, and other codes of SRD methods and BISA models are implemented by Matlab R2018a. It is worth noting that some codes need executable files to run. For a fair comparison, these algorithm codes are accessible on GitHub (https://github.com/NicoYuCN/srgBISA (accessed on 5 June 2021)).

3. Results

Based on the proposed framework, the effect of salient region inputs on the BISA performance is investigated. A total of three SRD methods combined with ten BISA models are explored, and three datasets (CSIQ, LIVE, and TID13) are involved in the quality assessment of Gaussian blurring images. The prediction performance is quantified with PLCC and SRCC values. To each dataset, the results are demonstrated from three parts:
1.
The metric values are shown in Figures 3, 5 and 7. To differentiate the inputs, WI (red star ), CR (blue triangle Δ), SR (pink square ); and mSR (black dot ·) are denoted with different combinations of colors and markers.
2.
The BISA model with the largest score drop between the WI and its CR input is shown in Figures 4, 6 and 8. The reasons why the BISA model causes the lowest PLCC value are also provided. Note that SORBD is used as the default SRD method;
3.
Setting the performance of the CR input as the baseline, the difference obtained by subtracting the baseline from that of using SR and mSR inputs is shown in Table 1, Table 2 and Table 3. Accordingly, if the BISA metric value of a salient region input is smaller than the baseline performance, its sharpness metric value is negative.
Besides the three parts, the distribution of size proportions of SR and mSR input is shown in Figure 9 and Table 4, and the running time is listed in Table 5 and Table 6. In the end, the SRD method with good balance between computing efficiency and prediction performance is suggested.

3.1. On the CSIQ Gaussian Blurring Images

Figure 3 shows the BISA results on the CSIQ Gaussian blurring images. The SR input () leads to equal or comparable metric values to the WI input (), followed by the mSR input (·), and the CR input (Δ) obtains the worst BISA results. In general, SORBD and SRIS lead to slightly higher prediction scores than DPLSG for the same BISA model (except the MEW). Based on SORBD or SRIS, several BISA models, such as MEW, CPBD, S3, LPC, MLV, SPARISH and BIBLE, obtain reliable scores when the WI is replaced by SR or mSR, but not the CR input. Interestingly, among the BISA models, MEW with salient regions from DPLSG or SORBD achieves superior PLCC values in comparison to that using the WI input.
As shown in Figure 3, SVC generally achieves low PLCC scores compared to the BISA models. Subsequently, based on the SORBD method, followed by the SVC model, the absolute score difference in each image is shown in Figure 4, and the images are illustrated, besides the points of the three largest score differences. In the figure, the horizontal axis shows the image index and the vertical axis shows the score difference. It can be seen that the three images are highly blurry and the CR covers only a part of the salient region. In contrast, using the salient regions from SORBD as an input leads to a small score drop (Figure 3), because the detected salient regions retains most of edge structures and image content.
Moreover, in comparison with the BISA results, by using the CR input, the performance difference is quantified and shown in Table 1. Note that a negative metric value denoted using the salient region input results in a worse prediction performance than that using the CR input. Overall, both the SR and the mSR input could enhance the metric values. Among the SRD methods, SORBD enables the PLCC values of from 0.012 to 0.085 higher, and the SRCC values of from 0.003 to 0.051 higher. Among the BISA models, S3, MLV, SPARISH and BIBLE obtain a consistently positive performance. For instance, MLV obtains an increase of PLCC values in the range between 0.026 and 0.063 and SRCC values in the range between 0.026 and 0.054.

3.2. On the LIVE Gaussian Blurring Images

As shown in Figure 5, the performance of BISA models combined with SRD methods is evaluated on the LIVE Gaussian blurring images. This shows that the SR input () obtains competitive metric values compared with the WI input (), followed by the mSR (·) and the CR input (Δ). Among the SRD methods, when using the SR and mSR input, the DPLSG, when combined with MEW, CPBD, FISH, SVC or MLV models, causes an obvious metric decrease, and SORBD and SRIS lead to a slightly inferior performance for each BISA model. Notably, using SORBD or SRIS, most BISA models, including CPBD, S3, FISH, SVC, LPC, MLV, SPARISH, BIBLE and RISE, maintain the prediction scores well. In addition, the salient region detected by SORBD and then combined with MEW or SVC can improve the PLCC values, more than the WI input for image blur estimation.
As shown in Figure 5, the MEW model obtains the lowest PLCC score when using the CR input. Based on the SORBD method combined with the MEW model, the absolute score difference in each image is shown in Figure 6, where the horizontal axis shows the image index, and the vertical axis shows the score difference. In addition, images leading to the three largest score drops are illustrated. The figure shows that the images are blurry, and the center regions cover relatively few edge structures in comparison to the whole images, both of which are important to the MEW model in image blur estimation. In contrast, when using salient regions as input, the score drop is relatively small, because the regions retain most edge structures and scene content.
Moreover, Table 2 shows the BISA difference between the CR input and the salient region input. Note that a negative metric value denotes that the salient region input had a worse prediction performance than the CR input. In general, both the SR and the mSR input enhance the metric values. On the LIVE Gaussian blurring images, 6 out of 10 BISA models consistently improve the values of evaluation metrics. Notably, the MEW model increases the PLCC value and the SRCC value up to 0.159 and 0.130 higher, respectively.

3.3. On the TID13 Gaussian Blurring Images

The metric values obtained using BISA models combined with SRD methods are shown in Figure 7. Compared to the WI input (), the SR () maintains the BISA performance well, followed by the mSR input (·), both of which outperform the CR input (Δ). Notably, SORBD combined with BISA models (except the CPBD model), and SRIS with CPBD, S3, FISH or MLV, achieve competitive BISA results compared to the corresponding WI input, while DPLSG leads to an observable decrease in BISA scores. In addition, based on the salient regions detected by SORBD, the BISA models, such as MEW and SVC, achieve even higher metric values than the WI input.
As shown in Figure 7, the MEW model obtains the lowest PLCC score when using the CR input. Based on the SORBD method combined with the MEW model, the absolute score difference in each image is shown in Figure 8 where the horizontal axis shows the image index and the vertical axis shows the score difference. In addition, the images causing the three largest score drops are illustrated. This indicates the three images have various levels of blurring distortion. Moreover, the CR input covers limited edge structures, while the detected salient regions retain most or all of the edge structures.
Based on the TID13 Gaussian blurring images, Table 3 shows the results of a comparison between salient region inputs and the CR input from PLCC and SRCC values. By setting the CR input as the baseline, the quantitative difference indicates that most BISA models achieve a better performance. Specifically, the MEW model has a 0.128 higher PLCC value and a 0.069 higher SRCC value, and the MLV model obtains an even better improvement. Regarding the SRD methods, SORBD and SRIS outperform the DPLSG method.

3.4. Distribution of the Size Proportion of Salient Regions

The distribution of the size proportion of used salient regions over all the images is shown in Figure 9. The size proportions are divived into five bins with equal intervals, which are highlighted by different colors. On the CSIQ dataset, the size proportion of mSR input generated from the DPLSG, the SORBD and the SRIS mainly falls into the range of (0.0, 0.2], (0.4, 0.6] and (0.8, 1.0], respectively. The similar phenomenon is also observed on the other two datasets.
In addition, the average size proportion of salient regions (SR and mSR) over the whole image is summarized in Table 4. This indicates that the DPLSG method generates a relatively smaller SR and mSR, followed by the SORBD method, while the SRIS method produces an SR input of approximately the same size as the whole image.

3.5. Time Consumption

The average time consumption per image with regard to different SRD methods is shown in Table 5. It was found that the DPLSG method requires more than 3.00 s to compute the saliency map of an image, while SORBD takes less than 1.00 s.
Table 6 summarizes the time cost per input, on average, regarding different BISA models. In general, when using WI input for image blur estimation, S3 and SPARISH are the most time-consuming models, taking more than 7.00 s, RISE and LPC require less than 2.00 s, and the time cost of other models is less than 1.00 s. When using CR, SR and mSR inputs, the time consumption drops, since only a part of the image region is used to estimate image sharpness.

4. Discussion

This study proposed a hybrid framework for no-reference image blur estimation. Based on the framework, the effect of salient region input on BISA tasks was investigated. It involved three SRD methods and ten BISA models. The procedure was conducted on three Gaussian blurring image datasets. Specifically, a BISA model’s input comes from different parts of an image, including WI, CR, SR and mSR.
The BISA performance (PLCC and SRCC) based on salient regions is comparable to that using the WI input (Figure 3, Figure 5, and Figure 7). For instance, on the LIVE (Figure 5), the SORBD method enables the BISA models, such as MEW, CPBD, S3, LPC, MLV, SPARISH and RISE, retain the estimation of image quality well. Moreover, SORBD jointly with MEW, leads to even higher PLCC values compared to the WI input. A similar phenomenon is observed on the other two image datasets (Figure 3 and Figure 7). This indicates that the performance using detected salient regions could approximate that using whole images.
However, quite a bit of variation in the metric values is observed for each BISA model. This shows that the PLCC value of SVC is generally low in CSIQ (Figure 3), and the PLCC and SRCC values of MEW are generally low in LIVE (Figure 5) and TID13 (Figure 7). FTo further understand this, we used the SORBD method for saliency detection to found out why the SVC model on CSIQ and the MEW model on LIVE and on TID13 lead to inferior performance. Correspondingly, the absolute score difference on each image and the images causing the three largest score drops are illustrated in Figure 4, Figure 6 and Figure 8. The reasons for the variations in metric values are manifold. From the perspective of datasets, the images in CSIQ are distinct from those in the other two datasets, while more than ten images have the same or similar content in LIVE and TID13, such as the reference images, named lighthouse, parrots, bikes, sailing, woman, and statue in the LIVE dataset. Thus, the discovery of variation in metric values is meaningful. Secondly, from the principle of the BISA models, SVC focuses on singular value analysis, and MEW estimates the edge width. As shown in Figure 4, the center region of images is homogeneous because of limited contrast. Thus, the curve in singular values struggles to reflect the curve in the whole image, while the mean edge width of the center region might be close to that of the whole image. On the other hand, as shown in Figure 6 and Figure 8, the center regions are inhomogeneous. The curve in singular values might be close to the curve in the whole image, while a gap in edge-width values might exist between the center region and the salient regions. In addition, the center-region or the salient-region input is quite different from the whole-image input. Thus, the predicted scores are correspondingly changed, which might accumulatively affect the score nonlinear fitting and the BISA metric values.
In comparison to the CR input, corresponding salient regions (SR and mSR) lead to better BISA performance. On the LIVE images (Table 2), there are six BISA models that consistently increase the metric values. The MEW model dramatically enhances the PLCC and the SRCC up to 0.159 and 0.130 higher, respectively. Similar phenomenon are also observed on the other two databases (Table 1 and Table 3). This suggests that the use of center regions is insufficient, while using salient regions seems more appropriate in image blur estimation. Accoding to the review of SRD methods [11], center bias is the most popular bias, and important objects are composed in the center of an image. This bias is widely used in SRD, but is rare in image-quality estimation. Layek et al. [66] propose an image quality metric that uses visual saliency and contrast, and extra attention is paid to the center by increasing the sensitivity of the similarity maps between the reference image and the distorted image. In their study, an image is splitted into 3 × 3 blocks, as shown in Figure 2D, and, to enhance the prediction performance, not only the center region but also its whole image are quantified from saliency similarity and contrast similiarty. Thus, the center region is not adequate to represent the whole image in image quality estimation. In other words, center bias or center-surround priors could help in silency map localization, while, in image quality estimation, salient regions are more informative and helpful.
The difference in prediction performance is observed among BISA models combined with SRD methods. In general, the detected salient regions from the SORBD and SRIS methods are suitable for the MEW, FISH, LPC, MLV, SPARISH and BIBLE models, and the regions from the DPLSG method are helpful for the models of SPARISH and BIBLE, in the utimate goal of image blur estimation. Among the models, MEW quantifies the spread of edges [34], FISH weights the log-energies of wavelet sub-band [38], LPC analyzes the coherence of local phases [40], MLV measures the maximum local variation in pixels [41], SPARISH concerns image edge patterns from an over-complete dictionary [43], and BIBLE measures the magnitudes of orthogonal moments [44], all of which could benefit from saliecy detection to avoid unreliable edge detection, energy weighting, coherence analysis, variation measure, pattern reconstruction, and magnitue computing, respectively. For saliency detection, DPLSG computes the diffusion matrix and seed vectors are obtained from the spatial variance in superpixels [59], SORBD integrates spatial layout, low-level cues and geometric interpretation for robust background measure [60], and SRIS concerns the effect of image inhomogeneity, noise and artificats, and emphasizes level-set evolution protocol [61]. However, image blur imposes additional difficulties to the diffusion matrix computation [59], and the strategies from both [60,61] might tackle these challenges, while the generailization capacity of these methods on the three datasets needs further quantification.
Most importantly, the distribution proportion of the sizes of used salient regions verifies its practical advantage. That is, saliency detection could decrease the computational burden (Figure 9 and Table 4) and the time cost in BISA tasks (Table 6). Specifically, DPLSG requires only from 30% to 36% of the image region and SORBD needs from 71% to 87% of image region to guide the BISA task. Overall, among the three SRD methods, SORBD is identified as the best one for tradeoff saliency detection, time consumption and blur estimation (Table 5). On the other hand, DPLSG shows superiority when used in combination with several BISA metrics, such as MEW, S3, FISH, and RISE, with a slight decrease in metric values. To our knowledge, several studies [51,52,53,54,55] have verified the added value of visual saliency maps in the estimation of image quality. Suprisingly, Ref. [65] shows that the profit extent seems not to be directly relevant to the performance of human fixation prediction when using the saliency detection method. This might suggest that salient regions are different from human fixation regions. The former reveals the conspicuous regions in a visual scene, while the latter aims to process the scene for increased image understanding. Thus, the incorporation of BISA models and SRD methods should be properly designed.
There are several limitations to the current study. First, most SRD methods take advantage of data biases, prior knowledge and image contrast in the algorithm design [11], and these methods might not be suitable for Gaussian blurring. The study [54] shows that blurred images could be divided into pattern and non-pattern groups, and most salient regions are more influential on pattern images than non-pattern images. Thus, a detailed study is required to verify the capacity of SRD methods on the general natural image test. Second, image blur imposes additional challenges for saliency detection. Future development of saliency-guided BISA models should consider the impact of image blur on saliency map detection in advance. Third, no deep-learning-based SRD methods or BISA models are evaluated. Deep learning requires massive training samples with a fixed image size, while it is difficult for an SRD method to generate salient regions with a consistent output size from images with different content. Specifically, our offline experiments figured out the infeasibility of deep learning methods. In addition, this study concerns only synthesized Gaussian blurring. Other image distortions could be evaluated in the proposed framework and, in the near future, the effect of detected salient regions on image-quality estimation can be fully understood.

5. Conclusions

This study proposed a hybrid framework to investigate the effect of detected salient regions on image blur assessment. This involved three saliency detection methods and ten blind image sharpness estimators, and experiments were conducted on three Gaussian blurring image datasets. Experimental results verified that salient regions benefit from image blur estimation and, most importantly, salient regions might be used as the surrogate of whole images for image quality assessment to enhance computing efficiency. The aim of future work will be to improve the prediction performance, develop accurate and generalizable saliency detection methods and properly incorporate SRD methods and BISA models.

Author Contributions

Conceptualization, S.Y. and L.J.; methodology, S.L. and S.Y.; software, S.L., Z.T., H.Y. and S.Y.; validation, S.L., H.Y. and S.Y.; formal analysis, Y.Z. and S.Y. and L.J.; investigation, S.Y. and H.Y.; resources, S.Y. and L.J.; data curation, S.L. and S.Y.; writing—original draft preparation, S.L. and S.Y.; writing—review and editing, H.Y., S.Y. and L.J.; visualization, S.L. and S.Y.; supervision, L.J.; project administration, L.J.; funding acquisition, Z.T., Y.Z., H.Y. and L.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by grants from the National Key Research and Development Program of China (2017YFB1402203-2; 2018YFB1404100; 2020YFB1406800) and the Fundamental Research Funds for the Central Universities (CUC210B017; JB211312).

Data Availability Statement

The Gaussian blurring images used to support the findings of this study are accessible online. The CSIQ image quality assessment database is from the Oklahoma State University (https://qualinet.github.io/databases/databases/ (accessed on 5 June 2021)), the LIVE database is from the University of Texas at Austin (https://live.ece.utexas.edu/research/quality/subjective.htm (accessed on 5 June 2021)), and the TID13 database is from the Tampere University of Technology (https://qualinet.github.io/databases/image/tampere_image_database_tid2013/ (accessed on 5 June 2021)).

Acknowledgments

The authors would like to thank the editor and anonymous reviewers for their invaluable comments that have helped to improve the paper quality.

Conflicts of Interest

The authors declare there is no conflict of interest. The founding sponsors had no role in the design of this study, in the collection, analysis or interpretation of data, in the writing of this manuscript, nor in the decision to publish the experimental results.

Abbreviations

The following abbreviations are used in this manuscript:
HVShuman vision system
SRDsalient region detection
CNNconvolutional neural network
FCNfully convolutional network
BISAblind image sharpness assessment
MEWmean edge width
CPBDcumulative probability of blur detection
S3spectral and spatial sharpness
FISHfast image sharpness
SVCsingular value curve
LPClocal phase coherence
MLVmaximum local variation
SPARISHsparse representation based image sharpness
BIBLEblind image blur evaluation
RISErobust image sharpness evaluation
LIVEan image database available online
CSIQan image database available online
TID13an image database available online
DPLSGdiffusion process on a 2-layer sparse graph
SORBDsaliency optimization from robust background detection
SRISsaliency region detection and image segmentation
PLCCPearson linear correlation coefficient
SRCCSpearman rank-order correlation coefficient
CRcenter region
WIwhole image
SRsalient region
mSRmajor salient region

References

  1. Driver, J. A selective review of selective attention research from the past century. Br. J. Psychol. 2001, 92, 53–78. [Google Scholar] [CrossRef]
  2. Frintrop, S.; Rome, E.; Christensen, H.I. Computational visual attention systems and their cognitive foundations: A survey. ACM Tran. Appl. Percept. 2010, 7, 1–39. [Google Scholar] [CrossRef]
  3. Li, Z. A saliency map in primary visual cortex. Trends Cognitive Sci. 2002, 6, 9–16. [Google Scholar] [CrossRef]
  4. Zhang, X.; Li, Z.; Zhou, T.; Fang, F. Neural activities in V1 create a bottom-up saliency map. Neuron 2012, 73, 183–192. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. White, B.J.; Kan, J.Y.; Levy, R.; Itti, L.; Munoz, D.P. Superior colliculus encodes visual saliency before the primary visual cortex. Proc. Natl. Acad. Sci. USA 2017, 114, 9451–9456. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Jian, M.; Dong, J.; Ma, J. Image retrieval using wavelet-based salient regions. Imaging Sci. J. 2011, 59, 219–231. [Google Scholar] [CrossRef]
  7. Fan, Q.; Qi, C. Saliency detection based on global and local short-term sparse representation. Neurocomputing 2016, 175, 81–89. [Google Scholar] [CrossRef]
  8. Zhu, J.-Y.; Wu, J.; Xu, Y.; Chang, E.; Tu, Z. Unsupervised object class discovery via saliency-guided multiple class learning. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 862–875. [Google Scholar] [CrossRef] [PubMed]
  9. Cong, R.; Lei, J.; Fu, H.; Cheng, M.-M.; Lin, W.; Huang, Q. Review of visual saliency detection with comprehensive information. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 2941–2959. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, W.; Lai, Q.; Fu, H.; Shen, J.; Ling, H.; Yang, R. Salient object detection in the deep learning era: An in-depth survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021. [Google Scholar] [CrossRef]
  11. Borji, A.; Cheng, M.-M.; Hou, Q.; Jiang, H.; Li, J. Salient object detection: A survey. Comput. Visual Media 2019, 5, 117–150. [Google Scholar] [CrossRef] [Green Version]
  12. Yang, C.; Zhang, L.; Lu, H. Graph-regularized saliency detection with convex-hull-based center prior. IEEE Signal Process Lett. 2013, 20, 637–640. [Google Scholar] [CrossRef]
  13. Ju, R.; Liu, Y.; Ren, T.; Ge, L.; Wu, G. Depth-aware salient object detection using anisotropic center-surround difference. Signal Process. Image Commun. 2015, 38, 115–126. [Google Scholar] [CrossRef]
  14. Jian, M.; Lam, K.-M.; Dong, J.; Shen, L. Visual-patch-attention-aware saliency detection. IEEE Trans. Cybern. 2014, 45, 1575–1586. [Google Scholar] [CrossRef]
  15. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  16. Klein, D.A.; Frintrop, S. Center-surround divergence of feature statistics for salient object detection. Int. Conf. Comput. Vision 2011, 20, 2214–2219. [Google Scholar]
  17. Le Meur, O.; Le Callet, P.; Barba, D.; Thoreau, D. A coherent computational approach to model bottom-up visual attention. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 802–817. [Google Scholar] [CrossRef] [Green Version]
  18. Li, H.; Ngan, K.N. A co-saliency model of image pairs. IEEE Trans. Image Process. 2011, 20, 3365–3375. [Google Scholar] [CrossRef] [PubMed]
  19. Wu, J.; Han, G.; Liu, P.; Yang, H.; Luo, H.; Li, Q. Saliency detection with bilateral absorbing Markov chainguided by depth information. Sensors 2021, 21, 838. [Google Scholar] [CrossRef]
  20. Lu, S.; Mahadevan, V.; Vasconcelos, N. Learning optimal seeds for diffusion-based salient object detection. Proc. IEEE conf. Comput. Vision Pattern Recognit. 2014, 2790–2797. [Google Scholar] [CrossRef] [Green Version]
  21. Ren, J.; Gong, X.; Yu, L.; Zhou, W.; Ying Yang, M. Exploiting global priors for RGB-D saliency detection. Proc. IEEE conf. Comput. Vision Pattern Recognit. 2015, 25–32. [Google Scholar] [CrossRef]
  22. Cheng, Y.; Fu, H.; Wei, X.; Xiao, J.; Cao, X. Depth enhanced saliency detection method. Proc. Int. Conf. Internet Multimedia Comput. Serv. 2014, 23–27. [Google Scholar] [CrossRef]
  23. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  24. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vision 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  25. Yu, S.; Wu, S.; Wang, L.; Jiang, F.; Xie, Y.; Li, L. A shallow convolutional neural network for blind image sharpness assessment. PLoS ONE 2017, 12, e0176632. [Google Scholar] [CrossRef] [PubMed]
  26. Yu, S.; Liu, L.; Wang, Z.; Dai, G.; Xie, Y. Transferring deep neural networks for the differentiation of mammographic breast lesions. Sci. China Technol. Sci. 2019, 62, 441–447. [Google Scholar] [CrossRef]
  27. Cheng, G.; Xie, X.; Han, J.; Guo, L.; Xia, G.-S. Remote sensing image scene classification meets deep learning: Challenges, methods, benchmarks, and opportunities. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 3735–3756. [Google Scholar] [CrossRef]
  28. Li, G.; Yu, Y. Visual saliency detection based on multiscale deep CNN features. IEEE Trans. Image Process. 2016, 25, 5012–5024. [Google Scholar] [CrossRef] [Green Version]
  29. Zhang, J.; Sclaroff, S.; Lin, Z.; Shen, X.; Price, B.; Mech, R. Unconstrained salient object detection via proposal subset optimization. Proc. IEEE Conf. Comput. Vision Pattern Recognit. 2016, 5733–5742. [Google Scholar] [CrossRef] [Green Version]
  30. Hu, P.; Shuai, B.; Liu, J.; Wang, G. Deep level sets for salient object detection. Proc. IEEE Conf. Comput. Vision Pattern Recognit. 2017, 2300–2309. [Google Scholar] [CrossRef]
  31. Feng, W.; Li, X.; Gao, G.; Chen, X.; Liu, Q. Multi-scale global contrast CNN for salient object detection. Sensors 2020, 20, 2656. [Google Scholar] [CrossRef]
  32. Shi, C.; Zhang, W.; Duan, C.; Chen, H. A pooling-based feature pyramid network for salient object detection. Image Vision Comput. 2021, 107, 104099. [Google Scholar] [CrossRef]
  33. Chandler, D.M. Seven challenges in image quality assessment: Past, present, and future research. Int. Sch. Res. Not. 2013, 2013. [Google Scholar] [CrossRef]
  34. Marziliano, P.; Dufaux, F.; Winkler, S.; Ebrahimi, T. Perceptual blur and ringing metrics: Application to JPEG2000. Signal Process. Image Commun. 2004, 19, 163–172. [Google Scholar] [CrossRef] [Green Version]
  35. Ferzli, R.; Karam, L.J. A no-reference objective image sharpness metric based on the notion of just noticeable blur (JNB). IEEE Trans. Image Process. 2009, 18, 717–728. [Google Scholar] [CrossRef] [PubMed]
  36. Narvekar, N.D.; Karam, L.J. A no-reference image blur metric based on the cumulative probability of blur detection (CPBD). IEEE Trans. Image Process. 2011, 20, 2678–2683. [Google Scholar] [CrossRef] [PubMed]
  37. Vu, C.T.; Phan, T.D.; Chandler, D.M. S3: A spectral and spatial measure of local perceived sharpness in natural images. IEEE Trans. Image Process. 2011, 21, 934–945. [Google Scholar] [CrossRef]
  38. Vu, P.V.; Chandler, D.M. A fast wavelet-based algorithm for global and local image sharpness estimation. IEEE Signal Process Lett. 2012, 19, 423–426. [Google Scholar] [CrossRef]
  39. Sang, Q.; Qi, H.; Wu, X.; Li, C.; Bovik, A.C. No-reference image blur index based on singular value curve. J. Visual Commun. Image Represent. 2014, 25, 1625–1630. [Google Scholar] [CrossRef]
  40. Hassen, R.; Wang, Z.; Salama, M.M.A. Image sharpness assessment based on local phase coherence. IEEE Trans. Image Process. 2013, 22, 2798–2810. [Google Scholar] [CrossRef]
  41. Bahrami, K.; Kot, A.C. A fast approach for no-reference image sharpness assessment based on maximum local variation. IEEE Signal Process Lett. 2014, 21, 751–755. [Google Scholar] [CrossRef]
  42. Gu, K.; Zhai, G.; Lin, W.; Yang, X.; Zhang, W. No-reference image sharpness assessment in autoregressive parameter space. IEEE Trans. Image Process. 2015, 24, 3218–3231. [Google Scholar]
  43. Li, L.; Wu, D.; Wu, J.; Li, H.; Lin, W.; Kot, A.C. Image sharpness assessment by sparse representation. IEEE Trans. Multimedia 2016, 18, 1085–1097. [Google Scholar] [CrossRef]
  44. Li, L.; Lin, W.; Wang, X.; Yang, G.; Bahrami, K.; Kot, A.C. No-reference image blur assessment based on discrete orthogonal moments. IEEE Trans. Cybern. 2015, 46, 39–50. [Google Scholar] [CrossRef] [PubMed]
  45. Li, L.; Xia, W.; Lin, W.; Fang, Y.; Wang, S. No-reference and robust image sharpness evaluation based on multiscale spatial and spectral features. IEEE Trans. Multimedia 2016, 19, 1030–1040. [Google Scholar] [CrossRef]
  46. Sun, C.; Cui, Z.; Gan, Z.; Liu, F. No-reference Image Blur Assessment Based on Multi-scale Spatial Local Features. KSII Trans. Internet Inf. Syst. 2020, 14, 4060–4079. [Google Scholar]
  47. Cai, H.; Wang, M.; Mao, W.; Gong, M. No-reference image sharpness assessment based on discrepancy measures of structural degradation. J. Visual Commun. Image Represent. 2020, 71, 102861. [Google Scholar] [CrossRef]
  48. Yu, S.; Jiang, F.; Li, L.; Xie, Y. CNN-GRNN for image sharpness assessment. Asian Conf. Comput. Vision 2016, 50–61. [Google Scholar] [CrossRef]
  49. Li, D.; Jiang, T.; Lin, W.; Jiang, M. Which has better visual quality: The clear blue sky or a blurry animal? IEEE Trans. Multimedian 2018, 21, 1221–1234. [Google Scholar] [CrossRef]
  50. Hosseini, M.S.; Zhang, Y.; Plataniotis, K.N. Encoding visual sensitivity by maxpol convolution filters for image sharpness assessment. IEEE Trans. Image Process. 2019, 28, 4510–4525. [Google Scholar] [CrossRef] [Green Version]
  51. Moorthy, A.K.; Bovik, A.C. Visual importance pooling for image quality assessment. IEEE J. Sel. Top. Signal Process. 2009, 3, 193–201. [Google Scholar] [CrossRef]
  52. Engelke, U.; Zepernick, H.-J. Framework for optimal region of interest-based quality assessment in wireless imaging. J. Electron. Imaging 2010, 19, 011005. [Google Scholar] [CrossRef] [Green Version]
  53. Zhang, L.; Shen, Y.; Li, H. VSI: A visual saliency-induced index for perceptual image quality assessment. IEEE Trans. Image Process. 2014, 23, 4270–4281. [Google Scholar] [CrossRef] [Green Version]
  54. Liu, T.-J. Study of visual quality assessment on pattern images: Subjective evaluation and visual saliency effects. IEEE Access 2018, 6, 61432–61444. [Google Scholar] [CrossRef]
  55. Zhang, W.; Zou, W.; Yang, F. Linking visual saliency deviation to image quality degradation: A saliency deviation-based image quality index. Signal Process. Image Commun. 2019, 75, 168–177. [Google Scholar] [CrossRef]
  56. Larson, E.C.; Chandler, D.M. Most apparent distortion: Full-reference image quality assessment and the role of strategy. J. Electron. Imaging 2010, 19, 011006. [Google Scholar]
  57. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef] [PubMed]
  58. Ponomarenko, N.; Jin, L.; Ieremeiev, O.; Lukin, V.; Egiazarian, K.; Astola, J.; Vozel, B.; Chehdi, K.; Carli, M.; Battisti, F.; et al. Image database TID2013: Peculiarities, results and perspective. Signal Process. Image Commun. 2015, 30, 57–77. [Google Scholar] [CrossRef] [Green Version]
  59. Zhou, L.; Yang, Z.; Zhou, Z.; Hu, D. Salient region detection using diffusion process on a two-layer sparse graph. IEEE Trans. Image Process. 2017, 26, 5882–5894. [Google Scholar] [CrossRef] [PubMed]
  60. Zhu, W.; Liang, S.; Wei, Y.; Sun, J. Saliency optimization from robust background detection. Proc. IEEE Conf. Comput. Vision Pattern Recognit. 2014, 2814–2821. [Google Scholar] [CrossRef] [Green Version]
  61. Joshi, A.; Khan, M.S.; Soomro, S.; Niaz, A.; Han, B.S.; Choi, K.N. SRIS: Saliency-based region detection and image segmentation of COVID-19 infected cases. IEEE Access 2020, 8, 190487–190503. [Google Scholar] [CrossRef]
  62. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. BASNet: Boundary-aware salient object detection. Proc. IEEE/CVF Conf. Comput. Vision Pattern Recognit. 2019, 7479–7489. [Google Scholar] [CrossRef]
  63. Liu, N.; Han, J.; Yang, M.-H. Picanet: Pixel-wise contextual attention learning for accurate saliency detection. IEEE Trans. Image Process. 2020, 29, 6438–6451. [Google Scholar] [CrossRef] [Green Version]
  64. Dai, G.; Wang, Z.; Li, Y.; Chen, Q.; Yu, S.; Xie, Y. Evaluation of no-reference models to assess image sharpness. In Proceedings of the 2017 IEEE International Conference on Information and Automation (ICIA), Macau, China, 18–20 July 2017; pp. 683–687. [Google Scholar] [CrossRef]
  65. Zhang, W.; Borji, A.; Wang, Z.; Le Callet, P.; Liu, H. The application of visual saliency models in objective image quality assessment: A statistical evaluation. IEEE Trans.Neural Netw. Learn. Syst. 2015, 27, 1266–12781. [Google Scholar] [CrossRef] [Green Version]
  66. Layek, M.A.; Uddin, A.F.M.S.; Le, T.P.; Chung, T.; Huh, E.-N. Center-emphasized visual saliency and a contrast-based full reference image quality index. Symmetry 2019, 11, 296. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Several representative pristine images in the three datasets.
Figure 1. Several representative pristine images in the three datasets.
Sensors 21 03963 g001
Figure 2. The proposed hybrid framework and experiment design. Under different image input scenarios, the performance of salient-region-guided image sharpness estimation is compared.
Figure 2. The proposed hybrid framework and experiment design. Under different image input scenarios, the performance of salient-region-guided image sharpness estimation is compared.
Sensors 21 03963 g002
Figure 3. Comparison of prediction scores regarding different inputs on the CSIQ Gaussian blurring images. This indicates that using SR (pink square ) or the mSR (black dot ·) as input could achieve close results to the corresponding WI input (red star ), and their results are better than that using the CR (blue triangle Δ) input. (The figure can be enlarged for viewing).
Figure 3. Comparison of prediction scores regarding different inputs on the CSIQ Gaussian blurring images. This indicates that using SR (pink square ) or the mSR (black dot ·) as input could achieve close results to the corresponding WI input (red star ), and their results are better than that using the CR (blue triangle Δ) input. (The figure can be enlarged for viewing).
Sensors 21 03963 g003
Figure 4. Absolute difference in predicted scores on CSIQ images based on the SORBD method combined with the SVC model. The images causing the three largest score drops are illustrated next to the points. (The figure can be enlarged for viewing).
Figure 4. Absolute difference in predicted scores on CSIQ images based on the SORBD method combined with the SVC model. The images causing the three largest score drops are illustrated next to the points. (The figure can be enlarged for viewing).
Sensors 21 03963 g004
Figure 5. Comparison of BSIA results with different inputs from SRD methods on the LIVE Gaussian blurring images. This shows that the SR () or the mSR (·) input could achieve comparable results to the WI input (), and the CR (Δ) input leads to the worst prediction performance. (The figure can be enlarged for viewing).
Figure 5. Comparison of BSIA results with different inputs from SRD methods on the LIVE Gaussian blurring images. This shows that the SR () or the mSR (·) input could achieve comparable results to the WI input (), and the CR (Δ) input leads to the worst prediction performance. (The figure can be enlarged for viewing).
Sensors 21 03963 g005
Figure 6. Absolute score difference in each image in LIVE based on the SORBD method and the MEW model. Beside the points of the three largest score drops, the corresponding images are illustrated. (The figure can be enlarged for viewing).
Figure 6. Absolute score difference in each image in LIVE based on the SORBD method and the MEW model. Beside the points of the three largest score drops, the corresponding images are illustrated. (The figure can be enlarged for viewing).
Sensors 21 03963 g006
Figure 7. Comparison of BSIA performance on the TID13 Gaussian blurring images. The metric values using salient region inputs are slightly inferior to those obtained from the WI input (), while the values are much better than those obtained using the CR input (Δ). (The figure can be enlarged for view).
Figure 7. Comparison of BSIA performance on the TID13 Gaussian blurring images. The metric values using salient region inputs are slightly inferior to those obtained from the WI input (), while the values are much better than those obtained using the CR input (Δ). (The figure can be enlarged for view).
Sensors 21 03963 g007
Figure 8. Absolute score difference in each image in TID13 using the SORBD method and the MEW model. The corresponding images are given next to the points of the three largest score drops. The three images show various levels of blurring distortion. (The figure can be enlarged for view).
Figure 8. Absolute score difference in each image in TID13 using the SORBD method and the MEW model. The corresponding images are given next to the points of the three largest score drops. The three images show various levels of blurring distortion. (The figure can be enlarged for view).
Sensors 21 03963 g008
Figure 9. Distribution of the size proportion of used salient regions over the corresponding whole images. The size proportions are divived into five bins and denoted with different colors. (The figure can be enlarged for viewing).
Figure 9. Distribution of the size proportion of used salient regions over the corresponding whole images. The size proportions are divived into five bins and denoted with different colors. (The figure can be enlarged for viewing).
Sensors 21 03963 g009
Table 1. Center region (CR) input as the baseline for BISA comparison on the CSIQ.
Table 1. Center region (CR) input as the baseline for BISA comparison on the CSIQ.
MEWCPBDS3FISHSVCLPCMLVSPARISHBIBLERISE
SR inputDPLSGPLCC0.027−0.0230.0240.0240.000−0.0040.0300.0410.048−0.003
SRCC−0.021−0.0490.0330.005−0.0950.0030.0260.0210.0290.007
SORBDPLCC0.0280.0120.0290.0350.0850.0160.0620.0420.0480.014
SRCC0.0030.0210.0340.0120.0320.0290.0510.0250.0310.019
SRISPLCC0.0070.0150.0310.0460.0720.0150.0630.0460.0570.009
SRCC0.0070.0220.0330.0270.0350.0250.0540.0240.0380.016
mSR inputDPLSGPLCC0.008−0.0390.016−0.0040.012−0.0160.0260.0370.042−0.002
SRCC−0.017−0.0570.025−0.024−0.113−0.0130.0230.0200.0260.012
SORBDPLCC0.0270.0120.0290.0350.0850.0160.0620.0420.0470.016
SRCC0.0030.0210.0350.0130.0330.0290.0510.0250.0300.020
SRISPLCC0.0020.0070.0250.0440.0740.0150.0610.0410.0570.006
SRCC0.0020.0210.0330.0240.0180.0220.0540.0220.0390.010
Table 2. Center region (CR) input as the baseline for BISA comparison on the LIVE.
Table 2. Center region (CR) input as the baseline for BISA comparison on the LIVE.
MEWCPBDS3FISHSVCLPCMLVSPARISHBIBLERISE
SR inputDPLSGPLCC0.113−0.0780.0140.052−0.219−0.0040.1090.0240.0210.005
SRCC0.088−0.130−0.0240.052−0.2390.0350.1060.0220.0260.011
SORBDPLCC0.1560.0190.0130.0820.0190.0360.1190.0310.0470.014
SRCC0.1130.013−0.0140.0780.0070.0640.1200.0240.0440.016
SRISPLCC0.1550.0290.0150.0830.0130.0360.1180.0320.0450.017
SRCC0.1300.0210.0050.080-0.0040.0630.1180.0250.0430.019
mSR inputDPLSGPLCC0.125−0.1510.0050.017−0.256−0.0060.0760.0090.0090.006
SRCC0.086−0.237−0.0160.006−0.3050.0140.0740.0080.0160.011
SORBDPLCC0.1590.0190.0140.0820.0190.0360.1190.0310.0470.014
SRCC0.1190.014−0.0100.0770.0060.0630.1190.0240.0430.017
SRISPLCC0.1330.0270.0190.083−0.0660.0400.1200.0160.0280.008
SRCC0.1260.0200.0070.079−0.1170.0620.1200.0100.0240.013
Table 3. Center region (CR) input as the baseline for BISA comparison of the TID13.
Table 3. Center region (CR) input as the baseline for BISA comparison of the TID13.
MEWCPBDS3FISHSVCLPCMLVSPARISHBIBLERISE
SR inputDPLSGPLCC0.083−0.0520.0420.045−0.226−0.0570.0750.0720.076−0.013
SRCC0.038−0.0930.0190.037−0.174−0.0150.0700.0600.072−0.003
SORBDPLCC0.1260.0600.0520.0920.0820.0100.1280.0700.0850.014
SRCC0.0650.0530.0370.0830.0820.0490.1150.0580.0710.011
SRISPLCC0.1050.0990.0560.0990.0750.0090.1290.0830.0950.021
SRCC0.0620.0940.0380.0860.0690.0520.1170.0650.0840.025
mSR inputDPLSGPLCC0.068−0.0400.0460.029−0.194−0.0590.0810.0700.076−0.013
SRCC0.022−0.0830.0360.027−0.160−0.0180.0770.0580.068−0.003
SORBDPLCC0.1280.0600.0520.0920.0800.0090.1290.0710.0870.014
SRCC0.0690.0530.0410.0830.0790.0490.1160.0580.0710.011
SRISPLCC0.0190.0930.0550.0950.0290.0000.1280.0600.0650.007
SRCC0.0190.0900.0340.0900.0250.0370.1210.0410.0610.011
Table 4. The average size proportion of salient regions over all the images.
Table 4. The average size proportion of salient regions over all the images.
DPLSGSORBDSRIS
SRmSRSRmSRSRmSR
CSIQ0.360.300.710.580.960.85
LIVE0.300.250.870.790.970.75
TID20130.300.240.790.680.980.82
Table 5. Average time cost per image on salient region detection (seconds).
Table 5. Average time cost per image on salient region detection (seconds).
DPLSGSORBDSRIS
CSIQ3.930.771.87
LIVE4.010.872.72
TID133.360.661.41
Table 6. Average time cost per input on image sharpness assessment (seconds).
Table 6. Average time cost per input on image sharpness assessment (seconds).
MEWCPBDS3FISHSVCLPCMLVSPARISHBIBLERISE
CSIQWI0.110.1512.030.220.231.350.109.420.961.54
CR0.020.021.450.020.030.210.010.980.210.25
DPLSG_SR0.040.064.440.070.070.560.033.090.400.55
DPLSG_mSR0.030.053.760.060.060.480.032.560.360.50
SORBD_SR0.080.118.960.160.151.140.086.330.711.04
SORBD_mSR0.070.087.280.120.120.940.055.180.600.87
SRIS_SR0.100.1411.680.220.211.330.109.130.931.47
SRIS_mSR0.090.1310.340.180.181.270.098.080.831.28
LIVEWI0.150.2017.100.300.261.880.1213.251.271.95
CR0.020.032.060.030.030.250.021.410.240.27
DPLSG_SR0.050.065.430.090.080.650.043.810.460.62
DPLSG_mSR0.040.054.430.070.070.540.033.140.390.52
SORBD_SR0.120.1614.670.260.211.900.1010.691.101.60
SORBD_mSR0.110.1513.380.230.191.680.109.611.011.43
SRIS_SR0.140.2016.430.290.241.910.1312.731.221.83
SRIS_mSR0.100.1512.640.220.181.510.099.610.961.41
TID13WI0.070.129.580.150.151.040.087.090.751.12
CR0.020.021.160.020.030.150.010.750.190.20
DPLSG_SR0.030.042.880.050.050.370.022.100.310.37
DPLSG_mSR0.020.032.380.040.040.310.021.690.270.34
SORBD_SR0.060.087.480.120.110.960.065.310.610.87
SORBD_mSR0.050.076.440.110.100.830.044.550.530.77
SRIS_SR0.070.129.400.150.151.050.087.080.731.11
SRIS_mSR0.060.097.760.120.120.920.065.790.630.94
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, S.; Yu, S.; Zhao, Y.; Tao, Z.; Yu, H.; Jin, L. Salient Region Guided Blind Image Sharpness Assessment. Sensors 2021, 21, 3963. https://doi.org/10.3390/s21123963

AMA Style

Liu S, Yu S, Zhao Y, Tao Z, Yu H, Jin L. Salient Region Guided Blind Image Sharpness Assessment. Sensors. 2021; 21(12):3963. https://doi.org/10.3390/s21123963

Chicago/Turabian Style

Liu, Siqi, Shaode Yu, Yanming Zhao, Zhulin Tao, Hang Yu, and Libiao Jin. 2021. "Salient Region Guided Blind Image Sharpness Assessment" Sensors 21, no. 12: 3963. https://doi.org/10.3390/s21123963

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop