Next Article in Journal
Intelligent Drone Positioning via BIC Optimization for Maximizing LPWAN Coverage and Capacity in Suburban Amazon Environments
Previous Article in Journal
RFE-UNet: Remote Feature Exploration with Local Learning for Medical Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel No-Reference Quality Assessment Metric for Stereoscopic Images with Consideration of Comprehensive 3D Quality Information

1
Shanghai Institute for Advanced Communication and Data Science, Shanghai University, Shanghai 200444, China
2
Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32603, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
These authors contributed equally to this work.
Sensors 2023, 23(13), 6230; https://doi.org/10.3390/s23136230
Submission received: 7 December 2022 / Revised: 24 March 2023 / Accepted: 29 March 2023 / Published: 7 July 2023
(This article belongs to the Topic Advances in Perceptual Quality Assessment of User Generated Contents)
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Recently, stereoscopic image quality assessment has attracted a lot attention. However, compared with 2D image quality assessment, it is much more difficult to assess the quality of stereoscopic images due to the lack of understanding of 3D visual perception. This paper proposes a novel no-reference quality assessment metric for stereoscopic images using natural scene statistics with consideration of both the quality of the cyclopean image and 3D visual perceptual information (binocular fusion and binocular rivalry). In the proposed method, not only is the quality of the cyclopean image considered, but binocular rivalry and other 3D visual intrinsic properties are also exploited. Specifically, in order to improve the objective quality of the cyclopean image, features of the cyclopean images in both the spatial domain and transformed domain are extracted based on the natural scene statistics (NSS) model. Furthermore, to better comprehend intrinsic properties of the stereoscopic image, in our method, the binocular rivalry effect and other 3D visual properties are also considered in the process of feature extraction. Following adaptive feature pruning using principle component analysis, improved metric accuracy can be found in our proposed method. The experimental results show that the proposed metric can achieve a good and consistent alignment with subjective assessment of stereoscopic images in comparison with existing methods, with the highest SROCC (0.952) and PLCC (0.962) scores being acquired on the LIVE 3D database Phase I.

1. Introduction

With the development of entertainment, military and industry, stereoscopic image/video technology has been widely used [1,2,3,4]. Meanwhile, the demand for high-definition and high-quality stereoscopic images/videos has become urgent, which leads to perceptual stereoscopic image quality assessment (SIQA) being an important issue for evaluating the performance of 3D technologies [5]. Stereoscopic images consist of two images, the left image and the right image, which are captured by two cameras [6,7,8]. Two channels of information need to be dealt with simultaneously to assess the quality of stereoscopic images. SIQA is widely used to assess the performance of stereoscopic image processing algorithms, such as stereoscopic image compression and denoising. Furthermore, a good SIQA algorithm can also be utilized to guide stereoscopic image processing, as well as the implementations and optimization proceduresof the image [9].
Intrinsically, stereoscopic image quality is determined not only by the spatial domain factors, but also by the transformed domain factors and stereo perceptual factors, such as binocular disparity/depth perception and binocular rivalry/combination [4,10,11,12,13]. Such dual-image and perceptual factors should be simultaneously considered in the process of feature description in SIQA cases, in addition to cyclopean features. However, almost all the existing feature extraction-based SIQA algorithms have not taken such features into consideration, with only the cyclopean features being exploited. Many assessment methods utilize cyclopean images in a different way. In Chen et al. [14], the natural scene statistics (NSS) model is used to extract spatial features from a cyclopean image, followed by support vector regression (SVR) score fitting. Recently, also based on the generated cyclopean image, Su et al. [15] utilized the spatial domain univariate NSS features, the wavelet domain univariate NSS features, and the bivariate density and correlation NSS features to create another assessment model. In addition, Zhou et al. [16] proposed an assessment model which utilizes the complementary local patterns of binocular energy response (BER) and binocular rivalry response (BRR), both extracted from the cyclopean image, to simulate the binocular visual perception.
In this paper, a perceptual quality assessment metric for stereoscopic images is proposed. The important effort of this work is to take comprehensive dual-image/3D relevant information into consideration of the stereoscopic quality to assess its image quality. Specifically, the traditional goal of an SIQA algorithm is to predict the quality of the cyclopean image formed within an observer’s mind when left and right images are presented stereoscopically. Therefore, like many assessment methods, in our method, features of the cyclopean images in both the spatial domain and transformed domain are extracted based on NSS. Furthermore, to better comprehend the intrinsic properties of the stereoscopic image, in our method, the binocular rivalry effect and other 3D visual properties are also considered in the process of feature extraction. The main contributions of this work are as follows: (1) new 3D perceptual features including binocular rivalry-based NSS features, binocular disparity matching error-based NSS features and binocular disparity consistency-based NSS features, are extracted for binocular perception representation; (2) neighbor difference and neighbor product-based features are optimized to reflect the high correlation between surrounding pixels in the spatial domain to assess the cyclopean image’s quality; (3) gradient magnitude, phase congruency and Log-Gabor response features are designed in the transformed domain to assess the cyclopean image’s quality; and finally, (4) 2D features from the spatial domain and the transformed domain and 3D visual perceptual features are jointly utilized to assess 3D image quality, and principal component analysis (PCA) [17] is utilized to reduce the dimensions of the feature vectors and enhance the performance of the quality prediction.
The proposed paper has several differences from our previously published work, called “No-Reference Stereoscopic Image Quality Assessment Based on Image Distortion and Stereo Perceptual Information” [18]. The differences can be listed as follows: (1) this paper focuses on both 2D and 3D statistics about the given stereoscopic image pairs, such as coherent and rivaling 3D perceptions of the human visual system, while the work in [18] only emphasizes the intrinsic image distortions of 2D image pairs, such as blurriness, noisiness and blockiness; (2) in this work, to exploit more fundamental 3D distortion characteristics, before the machine learning process, a principle component analysis (PCA) is also applied to adaptively recognize the notable features, giving useful attentions for the following machine learning.
The remainder of this paper is organized as follows. Section 2 reviews previous work on SIQAs. We present our model, describing in detail the generation of cyclopean images, the NSS-based feature extraction and quality prediction mechanisms in Section 3. Experimental results are analyzed in Section 4, and finally, conclusions are drawn in Section 5.

2. Previous Work on SIQA

Typically, there are two types of stereoscopic image quality assessment (SIQA) methods, which are the subjective SIQA and the objective SIQA. For the subjective one, such assessment is conducted by human observers [19], thus being unstable and time-consuming, hindering its industrial application. Contrarily, the objective SIQA method is conducted by computer algorithms, making it an appropriate solution for industries. Basically, such methods are divided into three types: the full reference (FR) [20,21,22,23,24], the reduced reference (RR) [25,26] and the no reference (NR) [27,28,29,30]. In FR-SIQA methods, evaluations are carried out with both the distorted and the original images provided. In RR-SIQA ones, only partial information from the original image, along with the distorted image, is provided, while in NR-SIQA methods, only the distorted image is given.
For FR-SIQA metrics, four categories altogether can be classified as follows. (1) First is directly applying the traditional image quality assessment algorithms on both the left and right stereo images, and then fusing these scores into the final stereoscopic score [22,23]. Such traditional methods include the peak signal-to-noise ratio (PSNR) [31], the structural similarity index (SSIM) [32] and the MS-SSIM [33]; (2) Second is calculating the tradition IQA scores on both the left and right images and simultaneously taking the stereo perceptual information, such as stereo disparity/depth [34], into consideration. For example, You et al. [35] applied different 2D quality metrics on stereo image pairs as well as the generated disparity map, finding a better combination for good performance. Meanwhile, Benoit et al. [36] computed quality scores of both stereo images and their disparity map, combining them to produce a final score. (3) Thirdly, in addition to evaluating stereo pairs, biological models can be used to simulate binocular combination behaviors [37]. Such metrics usually generate cyclopean images based on the binocular combination model and then evaluate the quality of cyclopean images using 2D-IQA metrics, such as the stereoscopic image quality. For example, Chen et al. [38] formulated binocular rivalry problems by modeling binocular combination behaviors based on a cyclopean map and the estimated disparity map. (4) Fourth is simulating human cognition by building human perception models through the mechanism of the human vision system (HVS) or the binocular visual characteristics, such as binocular fusion, binocular rivalry and binocular suppression [39]. However, due to the original non-distorted reference being missing in most cases, FR-SIQA methods are rarely used in real-world scenarios.
Recently, only a few works have focused on the development of RR and NR-SIQA models. Hewage et al. [40] proposed an RR-SIQA algorithm, in which edges are first computed from depth maps, and then the PSNR between the reference and edge maps is used to analyze the 3D image quality. Akhter et al. [41] proposed an NR-SIQA algorithm, which extracts segmented local features of artifacts from stereo-pairs and the estimated disparity map. In ref. [42], Ryu et al. computed the perceptual blurriness and blockiness scores of left and right images. The two scores are then combined into an overall quality index using the binocular quality perception model. Shao’s methods [27,43] conduct NR tasks without learning based on human opinion scores. Shao’s method [44] specifies the task of SIQA as monocular feature encoding and binocular feature combination based on sparse coding theory. Wang’s scheme [45] extracts features from binocular energy response, and machine learning is used to learn a visual quality prediction model; however, it does not consider any 3D visual characteristics. Fang’s method [46] combines blurriness- and noisiness-based features and information weight features to assess the quality of stereo-pairs. In ref. [47], NSS-based features are extracted from two synthesized images from the cyclopean image and fed into a stacked neural network model for quality prediction.
Recently, with the growth of the deep learning technology, several convolution neural network (CNN)-based methods have been proposed in pursuit of higher performance gains. For example, Du’s method [48] adopts an unsupervised deep network to map images to hierarchical representations. In ref. [49], a general active learning framework is proposed by combining the representativeness and informativeness of samples for active sampling. In ref. [50], an iterative heterogeneous transfer learning framework is proposed that learns a shared feature space for the source and target images and utilizes a novel iterative reweighting strategy to fuse the source samples. Recently, with the advancement of CNNs, several methods also try to implicitly model the feature relations with improved CNN structures. For example, in Zhou et al. [51], a deformable feature extraction block is explored to simulate the function of the primary visual cortex cells, and in ref. [52], a new network block, called the fusion weights assignment module (FWAM), is proposed to adaptively weigh features to guide the fusion of the stereo image features. The solution in [53] analyzes the features using CNN in the wavelet domain. The method in [54] models the relationship between the extracted features and subjective scores using the general regression neural network (GRNN).
Table 1 depicts the considered binocular characteristics and some implementation details of the state-of-the-art SIQA metrics according to the characteristics of HVS. The FR metrics are in italics. “CYC”, “BD”, “BF”, “BR”, “G/P”, “RM”, “SD” and “TD” denote “Cyclopean”, “Binocular Disparity”, “Binocular Fusion”, “Binocular Rivalry”, “Global/Patch”, “Regression Method”, “Spatial Domain” and “Transformed Domain”, respectively. “Global/ Patch” means that the metric is conducted globally across whole images or in patches. “O” and “X” represent “including” and “not including” in the corresponding scheme. The FR metrics are in italics. “NN” denotes a neural network fitting method. It can be seen that the proposed method takes all the listed binocular properties into consideration, while the other metrics lack some of them.

3. The Proposed Algorithm

The flowchart of the proposed NR-SIQA method is shown in Figure 1. Binocular disparity is firstly estimated according to an improved-SSIM-based stereo-matching algorithm. Then, a cyclopean image will be generated using stereoscopic images based on a binocular combination model. Next, both 2D NSS features based on neighbor difference and neighbor product in the spatial domain and gradient magnitude, phase congruency and log-Gabor filter responses in the transformed domain are extracted from the cyclopean image. Meanwhile, 3D NSS features are extracted by considering the binocular rivalry, binocular disparity matching error and binocular consistency. Finally, PCA and machine learning are applied for image score computation through training and testing procedures on these extracted feature vectors.

3.1. Disparity Estimation

Binocular disparity is an important factor in stereo visual perception and is utilized to reflect the consistency of left and right images. In the past few years, many disparity prediction methods have been proposed, based on the Sum of Absolute Difference (SAD) [38] or SSIM [32]. For SAD-based method, the disparity is predicted by minimizing the SAD between current pixel and its shifted correspondence in the other view. For SSIM-based methods, the disparity is estimated by maximizing the SSIM scores between two such pixels. However, all these methods only achieve satisfactory results when the input quality is high. In this paper, the improved Gaussian average scheme is added to achieve a better SSIM-based disparity estimation performance in low-quality scenarios.
Specifically, both the pixel-matching condition and that for its neighbors are considered when we search one point in other image. For any point in the left image [ x l , y l ] , its corresponding point in the right image is searched within [ x l r a n g e , y l ] to [ x l + r a n g e , y l ] , and the r a n g e could be changed with specific settings. The searched-for point having the largest calculated similarity with current point in the left image is chosen to be the matched point. Specifically, the SSIM values between those two points are calculated, and all the SSIM values are merged together with a Gaussian weighted sum. The Gaussian weighted sum of SSIM values, considered as the similarity, is used to search for the best matched point. Figure 2 shows the disparity maps of reference stereo-pair ‘im20_l.bmp’ and ‘im20_r.bmp’ from the LIVE 3D IQA database Phase I [55], estimated using the three above-mentioned methods. It can be seen from Figure 2 that the disparity map estimated by the proposed method has less no-texture and disclosure areas, and the rate of error matching of discontinuous areas is also low.

3.2. Cyclopean Image

When a human processes the visual stereoscopic images, internally, these signals will be composited into a fusion image, i.e., the cyclopean image, as is analyzed by the binocular vision combination model. To analytically interpret such a process, the Gain-Control Theory model [56] is utilized to conduct binocular fusion and cyclopean perception. Meanwhile, as is suggested in [57], Gabor filter can model the visual processing mechanism of the cells in the primary visual cortex of the human eye in high performance, the normalized Gabor filters are used to weight the paired stereoscopic images. A complex 2D Gabor filter is defined as
G E ( x , y , σ x , σ y , ξ x , ξ y , θ ) = 1 2 π σ x σ y e x p ( 1 2 [ ( R 1 σ x ) 2 + ( R 2 σ y ) 2 ] ) e x p ( i ( x ξ x + y ξ y ) ) R 1 = x c o s θ + y s i n θ R 2 = x s i n θ + y c o s θ
where σ x and σ y are the standard deviations of an elliptical Gaussian envelope along x and y axes, ξ x and ξ y are the spatial frequencies of filter and θ denotes the orientation of the filter.
In this paper, the first step of cyclopean image generation is to decide the base image between the left image and the right image. In our implementation, the competitive 2D NR-IQA metric, BRISQUE [58] is utilized to evaluate the quality of the left and right images. The image with better quality is selected as the base image, while the other one is chosen as the aid image.
Case 1: The left image is selected as the base image, and the synthesized cyclopean image I ( x , y ) is calculated as
I ( x , y ) = ω L ( x , y ) · I L ( x , y ) + ω R ( x D L ( x , y ) , y ) · I R ( x D L ( x , y ) , y )
where
ω L = G E L ( x , y ) G E L ( x , y ) + G E R ( x D L ( x , y ) , y )
ω R = G E R ( x D L ( x , y ) , y ) G E L ( x , y ) + G E R ( x D L ( x , y ) , y )
Case 2: The right image is selected as base image, and the synthesized cyclopean image I ( x , y ) can be calculated as
I ( x , y ) = ω L ( x + D R ( x , y ) , y ) · I L ( x + D R ( x , y ) , y ) + ω R ( x , y ) · I R ( x , y )
where
ω L = G E L ( x + D R ( x , y ) , y ) G E L ( x + D R ( x , y ) , y ) + G E R ( x , y )
ω R = G E R ( x , y ) G E L ( x + D R ( x , y ) , y ) + G E R ( x , y )
From Equations (2)–(7), D L or D R is the disparity map estimated by fixing the left or right image and shifting right or left image, respectively. ω L and ω R are weighting maps computed using the overall response magnitudes G E L of left image and G E R of right image on all scales and all orientations, respectively.
Figure 3 shows examples of synthesized cyclopean images. The distorted stereoscopic image is asymmetrically distorted by white noise, as shown in Figure 3a,b. Since white noise distortion can increase the right stimulus strength, the cyclopean image in Figure 3c is dominated by the right view with white noise. In the other case, when the stereoscopic image is asymmetrically distorted by Gaussian blur, as shown in Figure 3d,e, the simulated cyclopean image in Figure 3f is determined by the good image, i.e., the left image. This is because blur can decrease the Gabor filter response to a great extent, which causes the weight of the right image to be much smaller than that of the left image. This phenomenon is in accord with the subjective/human perception when stereoscopic images are asymmetrically distorted by different types of distortions.

3.3. Feature Extraction

Research on NSS has demonstrated that natural images have certain statistical properties [59], i.e., the probability density distribution of luminance in an image exhibits a Gaussian-like appearance after a local non-linear pre-processing operation called mean subtracted contrast normalization (MSCN) is performed. Applying the local non-linear operation to log-contrast luminances can remove local mean displacements from zero log-contrast and normalize the local variance of the log contrast, thus producing a decorrelating effect [59]. Using these kind of NSS features, several highly competitive 2D NR-IQA models, such as BRISQUE [58], DIIVINE [60] and BLIINDS-II [61], have been developed.
Given a pixel I C C ( x , y ) of a cyclopean image, the normalized coefficient I C C ^ ( x , y ) can be calculated through such an MSCN operation:
I C C ^ ( x , y ) = I C C ( x , y ) μ ( x , y ) σ ( x , y ) + C
where C is a small constant to increase the stability of I C C ^ ( x , y ) when the denominator tends to zero. In our implementation, a 7 × 7 Gaussian weighting matrix with a sigma of 1.17 pixels is utilized to compute μ and σ , and C is fixed to 1.
A zero-mean generalized Gaussian distribution (GGD) model is used to fit the probability density distributions of I C C ^ ( x , y ) as
f ( x ; α , σ 2 ) = α 2 β Γ ( 1 / α ) e x p x β α
where β = σ Γ ( 1 / α ) Γ ( 3 / α ) and Γ ( · ) is the gamma function.
The shape parameter α controls the real ‘shape’ of the distribution, while σ 2 controls the variance. The MSCN operation can show the statistics properties of images with different shape, kurtosis and spread in the presence of distortion. This means that it is possible to predict the type of distortion as well as its perceptual quality through quantifying these changes. The parameters α and σ 2 are effective features, which can be reliably estimated using the moment-matching-based approach in [62].
It also can be found that statistical relationships exist among neighboring pixels along different orientations [58]. The empirical distribution of the statistical relationships between pixels along different orientations is not logically symmetric about zero. To better fit the empirical histograms of the relationships, the very general asymmetric generalized Gaussian distribution (AGGD) model [63] is adopted. The AGGD with zero mode is given by
f ( x ; v , σ l 2 , σ r 2 ) = v ( β l + β r ) Γ ( 1 / v ) e x p x β l v , x < 0 v ( β l + β r ) Γ ( 1 / v ) e x p x β r v , x 0
where β l = σ l Γ ( 1 / v ) Γ ( 3 / v ) and β r = σ r Γ ( 1 / v ) Γ ( 3 / v )
The shape parameter v controls the real ‘shape’ of the distribution, while σ l 2 and σ r 2 are scale parameters that control the variance on each side of the mode. The parameters of the AGGD distribution are also estimated using the moment-matching-based approach in [62]. All four parameters ( η , v , σ l 2 , σ r 2 ) of AGGD are extracted as features where η is computed by
η = ( β r β l ) Γ ( 2 / v ) Γ ( 1 / v )
Inspired by the success of NSS in image quality assessment, a large amount of images with different contents and different distortions are selected from the LIVE database [64], and a series of experiments are conducted based on these images to prove that there is identical statistical regularity between different contents of images and that the distortions will remarkably disturb the statistical distributions of the images. The reference (REF) image “building2.bmp” and its five distorted versions, including such distortions as jpeg2000 (JP2K), JPEG, white noise (WN), Gaussian blur (BLUR) and fast fading (FF), are chosen as examples in Figure 4 and Figure 5a–c to show the experimental results. The NSS feature extraction and explanation of Figure 4 and Figure 5a–c are specified as follows.

3.3.1. Neighbor Difference-Based 2D Features

There is high correlation between surrounding pixels of an image in the spatial domain [58]. The difference between neighboring pixels and the product of neighboring pixels can effectively capture information on the image quality.
The difference between adjacent MSCN coefficients I C C ^ ( x , y ) of a cyclopean image are computed at a distance of one pixel along four orientations: horizontal (H), vertical (V), main-diagonal (D1) and secondary-diagonal (D2), which is illustrated in Figure 6. Specifically,
d i f H ( i , j ) = I C C ^ ( i , j ) I C C ^ ( i , j + 1 ) d i f V ( i , j ) = I C C ^ ( i , j ) I C C ^ ( i + 1 , j ) d i f D 1 ( i , j ) = I C C ^ ( i , j ) I C C ^ ( i + 1 , j + 1 ) d i f D 2 ( i , j ) = I C C ^ ( i , j ) I C C ^ ( i + 1 , j 1 )
where i { 2 , . . . M 1 } and j { 2 , . . . N 1 } and M , N are the image width and height, respectively.
Figure 4a shows the probability density distributions of neighbor differences along the horizontal orientation. Neighbor differences along these four orientations are parameterized using the GGD model to effectively capture the image statistics in the presence of distortion. By estimating two parameters, α and σ 2 , in each orientation as the spatial domain features, eight features are obtained overall.

3.3.2. Neighbor Product-Based 2D Features

In Figure 7, we also model the statistical correlation of pairwise products of neighbor pixels at a distance of two pixels along eight orientations, i.e., 0 , 22 . 5 , 45 , 67 . 5 , 90 , 112 . 5 , 135 and 157 . 5 . Specifically,
p r o 1 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i , j + 2 ) p r o 2 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 1 , j + 2 ) p r o 3 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 2 , j + 2 ) p r o 4 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 2 , j + 1 ) p r o 5 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 2 , j ) p r o 6 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 2 , j 1 ) p r o 7 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 2 , j 2 ) p r o 8 ( i , j ) = I C C ^ ( i , j ) · I C C ^ ( i + 1 , j 2 )
where i { 3 , . . . M 2 } and j { 3 , . . . N 2 } and M , N are the image width and height, respectively.
Figure 4b shows examples of the probability density distributions of neighbor products along the horizontal orientation, i.e., p r o 1 ( i , j ) . We utilize the AGGD model to fit the empirical histograms of the paired neighbor products, and the parameters ( η , v , σ l 2 , σ r 2 ) are estimated as the spatial domain features. Finally, 32 features are extracted for these 8 paired products.

3.3.3. Gradient Magnitude-Based 2D Features

As suggested in [65], image gradient magnitudes (GM) can reflect some properties of images. Specifically, GM is computed to be the root mean square of the directional gradients in the horizontal and vertical directions. In this paper, the Scharr operators [66] are applied along the horizontal (x) and vertical (y) directions, which are defined as
h x = 1 16 3 0 3 10 0 10 3 0 3 , h y = 1 16 3 10 3 0 0 0 3 10 3
Convolving h x and h y with cyclopean image I C C produces the horizontal gradient G M x ( x ) and vertical gradient G M y ( y ) . The gradient magnitude of the cyclopean image is defined as
G M ( x ) = ( I C C h x ) 2 + ( I C C h y ) 2 = G M x 2 ( x ) + G M y 2 ( y )
where the symbol “⊗” denotes the convolution operation.
Figure 4c shows the probability density distributions of G M y ( y ) after the MSCN operation. The GGD model is utilized to fit the distributions of G M ( x ) , G M x ( x ) and G M y ( y ) after MSCN, and the parameters α and σ 2 are estimated as the NSS-based transformed domain features.

3.3.4. Phase Congruency-Based 2D Features

Phase congruency (PC) reflects the image changes in brightness or contrast, which represents the spatial and phase locations where the maximum Fourier response is received [67]. This can be considered as the coherency degree of the image’s local frequencies. In this paper, PC is computed by Kovesi’s salient work [67] on the cyclopean image.
For cyclopean image I C C , the multi-direction and multi-scale complex Gabor filters are applied. M n e and M n o denote the even-symmetric (cosine) and odd-symmetric (sine) filters at scale n, respectively, forming a quadrature pair. By convolving I C C with these filters, the responses will form a vector at position x, on scale n and orientation θ :
[ e n , θ ( x ) , o n , θ ( x ) ] = [ I C C M n e , I C C M n o ]
where the symbol “⊗” denotes the convolution operation and θ is the orientation angle of the filters. The local amplitude of point x on scale n and orientation θ is computed by
A n , θ ( x ) = e n , θ 2 ( x ) + o n , θ 2 ( x ) .
The local energy along orientation θ is computed by
E θ ( x ) = F θ 2 ( x ) + H θ 2 ( x )
where F θ ( x ) = n e n , θ ( x ) , H θ ( x ) = n o n , θ ( x ) . Then, the phase congruency of the cyclopean image at point x is defined by
P C ( x ) = j E θ ( x ) ϵ + n j A n , θ ( x )
where ϵ is a small positive constant, j is the orientation angle number and n is the scale number.
Figure 5a shows the probability density distributions of the phase congruency map after MSCN is performed. To extract features from the phase congruency map, the AGGD model is used to fit the MSCN coefficients of phase congruency and AGGD model parameters ( η , v , σ l 2 , σ r 2 ) are extracted as the transformed domain features.

3.3.5. Log-Gabor Response-Based 2D Features

Since neurons in the visual cortex selectively respond to stimuli in different orientations and frequencies, filters with multiple scales and multiple orientations are useful to simulate this physiological phenomenon [57]. These filters can closely model frequency–orientation decompositions in the primary visual cortex and can capture energy in a highly localized manner in both space and frequency domains [68]. Such filter responses of an image are also useful for generating quality-aware NR-IQA features. Here, perceptually relevant log-Gabor filters are exploited to accomplish multi-scale and multi-orientation filtering.
A 2D log-Gabor filter with multiple frequencies and multiple orientations in the Fourier domain can be expressed as
G l o g ( ω , θ ) = e x p l o g ( ω / ω 0 ) 2 2 σ r 2 · e x p ( θ θ j ) 2 2 σ θ 2
where θ j = j π / J , j = { 0 , 1 , . . . J 1 } is the orientation of the log-Gabor filter, J is the overall number of orientations, ω 0 is the center frequency of the filter, σ r determines the filter’s radial bandwidth and σ θ controls the filter’s angular bandwidth. After convolving with the cyclopean image I C C , the local responses of the real part, e n , θ j ( x ) , and imaginary part, o n , θ j ( x ) , of the log-Gabor filter are obtained at position x on scale n and orientation θ j .
The log-Gabor filter response of cyclopean image I C C is expressed as
G l o g ( x ) = n j e n , θ j 2 ( x ) + o n , θ j 2 ( x ) .
The real and imaginary parts of the log-Gabor response of cyclopean image I C C are respectively expressed as follows:
G x l o g ( x ) = n j e n , θ j 2 ( x ) G y l o g ( x ) = n j o n , θ j 2 ( x )
The log-Gabor filter phase of cyclopean image I C C is expressed as
G P l o g ( x ) = n j a r c t a n e n , θ j ( x ) , o n , θ j ( x )
Figure 5b,c show the probability density distributions of G x l o g ( x ) and G P l o g ( x ) after MSCN operation, respectively. The probability density distributions of G l o g ( x ) , G x l o g ( x ) , G y l o g ( x ) and G P l o g ( x ) after MSCN is performed can be fitted using the model of GGD, and the parameters α and σ 2 are included in the feature set.

3.3.6. Binocular Rivalry-Based 3D Features

Binocular disparity or depth is an important factor in stereo visual perception. Binocular disparity is computed by the distance between the corresponding pixels in the left and right images. The binocular disparity can be regarded as the measurement that the human brain senses the depth information with. In general, it can lead to binocular rivalry and binocular suppression. As indicated by [35,38], the disparity or depth information has an impact on stereoscopic image quality. Therefore, the binocular disparity can be used to generate the binocular rivalry properties and enhance the detection accuracy of the saliency map.
According to visual psychophysical research, binocular combination will integrate two retinal points into a single binocular perception when similar contents of the left and right images fall on the corresponding retinal points [69]. Binocular rivalry happens when the content of the left and right images is relatively different. Therefore, the left and right eyes view mismatched images at the same retinal locations. Numerous studies suggest that binocular rivalry is the result of competition between human eyes [70]. This competition is caused by the mismatched content of the left and right images. Binocular suppression is a special case of binocular rivalry. When binocular suppression happens, the entire image from one retina is partly suppressed [1].
We use the binocular disparity estimated in Section 3.1 to describe the binocular rivalry. A variety of stereo-pairs with different distortions and contents are selected from the LIVE 3D database Phase I [55] to validate the effectiveness of the binocular rivalry-based 3D features and the following 3D features. The statistical distributions of the stereo-pair “ i m 17 _ l . b m p ” and “ i m 17 _ r . b m p ” are chosen as an example in Figure 5d–f. Figure 5d shows the probability density distributions of the binocular disparity map after MSCN operation. We use parameters α and σ 2 of the GGD model to represent the statistical features of binocular disparity.

3.3.7. Binocular Disparity Matching Error-Based 3D Features

Except for the estimated binocular disparity, the disparity matching error caused by the improved SSIM-based stereo matching algorithm is also a useful feature related to the stereo image quality. One pixel fails to find the corresponding pixel in the other image in a stereo-pair when large matching error exists. The disparity matching error is defined as follows:
D L e r r o r ( x , y ) = I L ( x , y ) I R ( x D L ( x , y ) , y ) D R e r r o r ( x , y ) = I L ( x + D R ( x , y ) , y ) I R ( x , y )
where D L and D R are disparity maps estimated using the left and right images as the base image, respectively. I L ( x + D R ( x , y ) , y ) and I R ( x D L ( x , y ) , y ) are the left and right disparity-compensated images, respectively.
Figure 5e shows the probability density distribution of binocular disparity matching error map after MSCN operation. Then, the GGD model is utilized to fit the probability density distribution, and the parameters ( α , σ 2 ) are estimated as 3D features.

3.3.8. Binocular Disparity Consistency-Based 3D Features

Like normal natural images, there is high correlation between neighboring pixels, except in some edge regions in a disparity map. Thus, one point in a disparity map is usually similar to its surrounding points when the disparity map is computed from an original, undistorted stereoscopic image. We describe the disparity consistency by using a high-pass filter to convolve with the estimated disparity map D ( x , y ) . The disparity consistency is defined as
D c o n s i s t ( x , y ) = D ( x , y ) 0 1 4 0 1 4 1 1 4 0 1 4 0
The value of D c o n s i s t ( x , y ) would tend to zero if the stereoscopic image has no distortion or little distortion. Figure 5f shows the probability density distributions of disparity consistency maps after MSCN operation. These six corresponding disparity maps are extracted from the stereo-pair i m 17 . b m p and its five distorted versions from the LIVE 3D database Phase I. It can be seen that the coefficient values from a disparity consistency map mostly concentrate to zero. However, when the disparity is obtained from a distorted stereo-pair, the coefficient values of its disparity consistency map are more likely to spread out. In addition, the probability of coefficient values equal to zero is much lower.
The GGD model is used to fit the empirical distribution of D c o n s i s t ( x , y ) after the MSCN operation is performed, and the GGD parameters ( α , σ 2 ) are extracted as features.

3.4. Machine Learning

To adaptively choose a salient feature set for better regression model fitting, the principal component analysis (PCA) is applied, and the chosen salient features are sent into the support vector regressor (SVR) for the final fitting of the objective quality score. In the training stage, the generated features and the labeled subjective quality scores are sent to the regression model, and the training epoch is 1000. In the testing stage, the predicted objective scores are obtained from the trained regressor. Our implementation is based on the MATLAB platform, and the SVR is conducted using the LIBSVM package [2]. For the SVR’s settings, the epsilon-SVR with a radial basis function (RBF) kernel is used, and the SVR parameters C and γ are set to 512 and 0.015625, respectively, according to extensive experiments. The details about the parameter selection can be found in Section 4.2.

4. Experimental Results and Analysis

4.1. Databases and Evaluation Criteria

For training and testing datasets, we adopt the LIVE 3D IQA database Phase I [1] and Phase II [14] and the Waterloo IVC (UW/IVC) 3D IQA database [3] for testing the proposed algorithm and other SIQA algorithms. For LIVE 3D Phase I, 20 reference stereoscopic images and 365 distorted stereoscopic pairs are included, each having 5 types of distortions, i.e., JP2K compression distortion, JPEG compression distortion, white noise (WN), fast fading (FF) and blurring (BLUR). All distortions are symmetric in nature. Similarly, for LIVE 3D Phase II, 8 reference stereoscopic images, 120 symmetrically distorted stereoscopic pairs and 240 asymmetrically distorted stereoscopic pairs generated from those 8 references are included. The Waterloo IVC 3D IQA database also contains two data phases, created from 6 and 10 pristine stereo-pairs, respectively. Both of these phases include symmetrical and asymmetrical distortions, including additive white Gaussian noise, Gaussian blur and JPEG compression. Each type of distortion has four levels. Altogether, a total of 78 single-view images and 330 stereoscopic images are contained in Phase I, and a total of 130 single-view images and 460 stereoscopic images are included in Phase II.
To evaluate the overall accuracy of the stereoscopic objective assessment, three criteria are used: the Pearson linear correlation coefficient (PLCC), the Spearman rank order correlation coefficient (SROCC) and the root mean squared error (RMSE). Traditionally, the PLCC and RMSE are used to evaluate the prediction accuracy of SIQA metrics, and the SROCC detects the prediction’s monotonicity. Note that the higher the SIQA’s accuracy, the higher the values PLCC and SROCC will be, while the value of the RMSE should be lower. Before the evaluation, a non-linear regression analysis, suggested by the video quality experts group [4], is utilized to provide mappings between the objective scores and subjective mean opinion scores (MOSs). Specifically, a five-parameter logistic function [10] is adopted as such a non-linear mapping:
f ( x ) = β 1 · ( 1 2 1 1 + e β 2 ( x β 3 ) ) + β 4 · x + β 5
where β i = 1 , 2 , . . . , 5 are parameters determined by the subjective scores and the objective scores.

4.2. Implementation Details

The implementation of the whole procedure is conducted on MATLAB, and the support vector regressor is conducted using LibSVM. The PCA is introduced and its effectiveness is validated by various experiments. The following experiments are also conducted to select the SVR parameters (C, γ ) and prove that almost no over-fitting exists in our method.

4.2.1. PCA for Feature Dimension Reduction

In the proposed algorithm, more than 60 features are extracted from the 2D cue and 3D cue for each image. Inevitably, information redundancy may be contained in these NSS features (i.e., the shape feature). To highlight the main component and reduce the dimensions of the feature vectors, PCA is utilized, resulting in a more efficient quality predictor. Let X = [ x 1 , x 2 , . . . , x n ] d × n denote the feature matrix extracted from n images, and d is the number of features extracted from each image. By applying PCA to X , a transform matrix, Φ d × m , will be obtained by an eigenvalue decomposition conducted on the covariance matrix of the feature matrix X . The transform matrix is formed by the m ( m < d ) principle vectors associated with the m most significant eigenvalues of the covariance matrix of X . Then, the dimension of each feature vector x i can be reduced by transformation as follows:
x i = Φ T x i , x i m × 1 i = 1 , 2 , . . . , n
The PLCC with various PCA dimensions in LIVE Phase I and II is shown in Figure 8. As the number of feature dimensions is set to 44, the performance is the best. m is set to 44 in our implementation.

4.2.2. SVR Parameters (C, γ ) Selection

A cross-validation experiment is conducted to choose the value of (C, γ ). During this experiment, samples from LIVE 3D Phase I and II are randomly partitioned into two sets: the training set and the testing set. For each phase, 80% of the images are for training and the remaining 20% of the images are for testing. The training epoch is 1000, and the median results are reported for performance evaluation. Parameters (C, γ ) delivering the best median result are chosen as the best. The illustration of this experimental process can be found in Figure 9. The number of the level contour indicates the SROCC values for the cross-validation. Apparently, there exists a circular region surrounding the center, where the proposed metric obtains the best SROCC results on the two databases. This indicates that this learning model is more robust to different databases of samples. According to the results, the optimal parameters (C, γ ) are set to be 512, 0.015625.

4.2.3. Two-Fold Cross-Validation

Each database is randomly split into two sets: the training set and the testing set, with a 50:50 splitting ratio. Performance evaluations are conducted as follows. First, the optimal regression model is learned from the training set, and then tested on the training set and testing set, respectively. Next, the other regression model is learned from the testing set, and then tested on the training set and testing set, respectively. Performance evaluations on LIVE Phase I and II are listed in Table 2 (“MTrain→Testing set” means that the model is learned from the training set is tested on the testing set, and “MTest→Training set” represents that the model learned from the testing set is tested on the training set). As is seen on the table, regardless of whether the regression model is learned from the training set or the testing set, no obvious difference exists between the evaluated performance, indicating that almost no over-fitting exists in the proposed metric.

4.2.4. Features with Noise-Addition

Here, we test the impact of Gaussian white noise (WN) on the model’s fitting performance. Similarly, LIVE 3D Phase I and II are randomly partitioned into the training set and the testing set with the ratio 80:20. All the features of the samples in the training set are applied with white noise before the experiment. The variance of the white noise is set to 0.01. Performance evaluations are conducted on the training set affected or not affected by white noise, and then tested on the non-affected training set and testing set, respectively. The results are listed in Table 3, where “WN→Training set” means that the model is learned from the WN-affected training set and is tested on the non-affected training set, and “WN/F→Testing set” denotes that the model is learned from the non-affected training set and is tested on the non-affected testing set. As is seen in the table, there is no big difference between the model’s performance when tested on the training set or testing set. Moreover, the results also indicate that white noise has barely any impact on the model’s fitting process. Therefore, we can make a conclusion that almost no over-fitting exists in the proposed metric.

4.3. Performance Comparison

In the proposed method, 80% images of each database are chosen randomly as training set and the rest is for testing. The training-testing process is repeated 1000 times, and the average of performance is reported for comparison.

4.3.1. Overall Performance Comparison

The overall performance comparisons of the proposed algorithm and other NR-SIQA methods on both LIVE Phase I and II are shown in Table 4. Shao [27,44] split the LIVE Phase II into symmetrically and asymmetrically distorted dataset, and thus the overall performance on whole LIVE Phase II is unavailable, which represented by “-”. It can be seen that the proposed method achieves much higher performance than other state-of-the-art metrics. Besides, it also can be seen from Table 4 that the proposed metric and other NR-SIQA metrics achieve a better performance on Phase I than those on Phase II. The reason is that Phase I only contains symmetric distortions, while Phase II contains both symmetric and asymmetric distortions. It is much more difficult to evaluate the stereoscopic image quality with asymmetric distortions due to the limited understanding of HVS when human watch asymmetrically distorted stereoscopic images.
Table 5 presents the comparison results between the proposed method and other FR and NR metrics on Waterloo IVC 3D database. In Table 5, the FR methods are marked in italic. Specifically, Mittal et al. [58] is a 2D NR metric; therefore, its performance is evaluated on the both left and right views, resulting two objective scores. The final score for this method is the average of these two scores. It can be seen that the proposed method outperforms the other metrics. The reason that Chen’s [14] method performs worse than Mittal’s [58] may be because its stereo-matching algorithm fails to generate precise disparity.

4.3.2. Performance Comparison of Each Distortion Type

Here, we discuss the performance comparisons over different distortions. Several state-of-the-art FR-SIQA metrics (i.e., Benoit [36], You [35], Gorley [22] and Chen [38]), along with our method, are tested on LIVE Phase I and II, and the results are listed in Table 6. Note that the best scores are marked in bold in that table. According to the table, the proposed algorithm achieves a slightly higher performance than the algorithm in [44], while performing better than other FR-SIQA and NR-SIQA metrics. As FR-SIQA metrics, Gorley et al. [22] predicts the overall quality by averaging the scores of left and right views generated from 2D IQA algorithms. Its performance is much worse than other SIQA methods because it does not take the binocular visual properties into account. In addition to the dual-2D quality scores for both views, You et al. [35] and Benoit et al. [36] simultaneously consider disparity information, and the performances of these two methods are better than Gorley et al., as their SROCC and PLCC scores reach up to 0.90 on LIVE Phase I. Chen et al.’s model [38] applies the SSIM to assess the quality of cyclopean images, making its SROCC and PLCC scores reach up to 0.920 on LIVE Phase I. For NR-SIQA metrics, Shao et al. [27] may be opinion-unaware; however, its performance is inferior to other methods based on prior subjective human opinion scores. Shao et al. [44] achieves better performance than the other methods, except for the proposed one, as it lacks the binocular properties of our method. The performance of method [41] is barely satisfactory on JP2K and JPEG distortion in the LIVE Phase I database. For Chen et al.’s method [14], it extracts NSS-based features of the cyclopean image in the spatial domain, but it lacks considerations of the binocular rivalry effect and other 3D visual perceptions. Its performance is worse than our proposed algorithm, as our method sufficiently extracts 2D and 3D features, and takes 3D visual characteristics into consideration.

4.4. Performance Evaluation

4.4.1. Performance Evaluation of Each Feature Type

In our metric, features including neighbor difference (ND)-based NSS features and neighbor product (NP)-based NSS features in the spatial domain and gradient magnitude (GM)-based NSS features, phase congruency (PC)-based NSS features, log-Gabor filter response (GR)-based NSS features in the transformed domain and 3D information (3DI), such as binocular rivalry, binocular disparity matching error and binocular disparity consistency NSS features, are extracted. In order to verify whether some of these features could be ignored, we analyze the gain of each type of feature in Table 7. Note that this experiment is conducted with an incomplete feature set; thus, the PCA process is not used to reduce the feature dimensions. We can see from Table 7 that each type of feature can improve the performance of our model.

4.4.2. Cross-Database Performance Evaluation

The evaluation strategy used in Section 4.3 is inadequate to evaluate the generalization capability of a NR-SIQA model, since the random partition method on a single database leads to the training set and the testing set containing the same types of distortions. Therefore, to demonstrate the efficiency of an IQA model on other possibly unknown distortions, and to verify such methodologies are actually “completely blind”, we test the generalization capability of the proposed method by training the regression model on one database, i.e., LIVE Phase I, and testing the regression model on another database, i.e., LIVE Phase II.
In our implementation, we train the regression model on LIVE Phase I and test it on LIVE Phase II on one hand, and on the other hand, we train the model on LIVE Phase II and test it on LIVE Phase I. The results are reported in Table 8 and Table 9, from which we can draw the conclusion that the regression model that is trained on one database (i.e., LIVE Phase I) is not completely efficient for testing on the other database (i.e., LIVE Phase II), since the contents in the two databases are quite different. Phase II includes both symmetrically and asymmetrically distorted stereo-pairs, while Phase I only contains symmetrically distorted stereo-pairs.

4.4.3. Cross-Distortion Performance Evaluation

Here, we verify the cross-distortion performance comparisons between the proposed method and the other methods. Specifically, each tested method is trained on one type of distortion but tested on another type, with the same training sample sources. The results are displayed in Table 10 and Table 11. According to the row elements of the tables, we can see that the model trained on one type of distortion has some influence on the performance on other types of distortion, but with less influence on the overall performance. This demonstrates that the proposed scheme is a good metric for the perceived quality prediction of the stereoscopic image.

4.4.4. Image Content-Based Performance Evaluation

In this experiment, we discuss the database partitioning strategies, which are split by content or split randomly. During this experiment, we took 80% of the samples for training and the rest for testing. Specifically, for the LIVE 3D database Phase I, 16 reference images and their corresponding distorted images were selected as the training sets, while the others were chosen as the testing sets. For LIVE 3D Phase II, six samples were chosen for training, and the other two were used for testing. The comparison results between the content-based partitioning and random partitioning are listed in Table 12. It can be seen that the performance using image content-based partitioning is slightly worse than that using random partitioning. This is because the regression model learned from the content-based partitioned samples is a “content-aware” model, and thus the performance is a little worse than that of the “content-unaware” model.

4.5. Influence of Disparity Estimate Algorithm

In order to verify whether our metric is method-independent on disparity estimation, we compare the performance of the improved SSIM-based disparity search algorithm in our method with those of the SAD-based and SSIM-based disparity estimation algorithms. The performance comparisons are listed in Table 13. It can be seen from Table 13 that different disparity estimation algorithms lead to similar performance, which means our metric is method-independent. It also can be seen that the performance from the improved SSIM-based disparity search algorithm is slightly better.

4.6. Complexity Analysis

The time–cost ratios for each step of our algorithm are listed in Table 14. As is described in the table, the biggest share of complexity in our method is from the PCA ans SVM fitting steps, which cost more than half of our prediction time, followed by the SSIM matching step that costs 36% of our prediction time.

5. Conclusions

In this paper, we propose a novel natural scene statistics-based, no-reference quality assessment algorithm for stereoscopic images, in which both the quality of the cyclopean image is considered and the binocular rivalry and other 3D visual intrinsic properties are exploited. The proposed algorithm mainly consists of two steps. First, 2D features are extracted from the cyclopean image in the spatial domain and transformed domain, and then 3D quality-relevant perceptual features are extracted from stereo visual perceptual information, including the binocular rivalry, the disparity matching error and the disparity consistency. Next, followed by NSS modeling and adaptive PCA feature pruning, SVR is utilized to fit an optimum regression model. Compared with other FR-SIQA metrics and NR-SIQA metrics on the LIVE Phase I and LIVE Phase II databases, the proposed stereoscopic image quality metric achieves the best performance compared to other state-of-the-art SIQA metrics.

Author Contributions

Conceptualization, Y.Y. and X.G.; methodology, Y.Y. and L.S.; software, Y.Y.; validation, Y.Y.; formal analysis, Y.Y.; writing—original draft preparation, Y.Y.; writing—review and editing, Y.Y. and L.S.; visualization, Y.Y. and D.W.; supervision, L.S. and R.F.; project administration, L.S.; funding acquisition, L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China under grant No.61422111, 61671282 and U1301257 and was sponsored by the Shanghai Shuguang Program (15pjd015) and the Innovation Program of Shanghai Municipal Education Commission (17SG37).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, H.; Wang, S.; Zhao, Y.; Wei, J.; Piao, M. Large-scale elemental image array generation in integral imaging based on scale invariant feature transform and discrete viewpoint acquisition. Displays 2021, 69, 102025. [Google Scholar] [CrossRef]
  2. Deng, L.; Pu, Y. Analysis of college martial arts teaching posture based on 3D image reconstruction and wavelet transform. Displays 2021, 69, 102044. [Google Scholar] [CrossRef]
  3. Qi, S.; Ning, X.; Yang, G.; Zhang, L.; Long, P.; Cai, W.; Li, W. Review of multi-view 3D object recognition methods based on deep learning. Displays 2021, 69, 102053. [Google Scholar] [CrossRef]
  4. Wang, X.; Wang, C.; Liu, B.; Zhou, X.; Zhang, L.; Zheng, J.; Bai, X. Multi-view stereo in the Deep Learning Era: A comprehensive review. Displays 2021, 70, 102102. [Google Scholar] [CrossRef]
  5. Winkler, S.; Min, D. Stereoscopic image quality compendium. In Proceedings of the 2011 8th International Conference on Information, Communications & Signal Processing, Singapore, 13–16 December 2011; pp. 1–5. [Google Scholar]
  6. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  7. Yang, H.; Li, J. Application of multivariate statistics and 3D visualization analysis in tacit knowledge diffusion map. Displays 2021, 69, 102062. [Google Scholar] [CrossRef]
  8. Duan, Z.; Chen, Y.; Yu, H.; Hu, B.; Chen, C. RGB-Fusion: Monocular 3D reconstruction with learned depth prediction. Displays 2021, 70, 102100. [Google Scholar] [CrossRef]
  9. Brandao, T.; Queluz, M.P. No-reference image quality assessment based on DCT-domain statistics. Signal Process. 2008, 88, 822–833. [Google Scholar] [CrossRef]
  10. Lu, B.; Sun, L.; Yu, L.; Dong, X. An improved graph cut algorithm in stereo matching. Displays 2021, 69, 102052. [Google Scholar] [CrossRef]
  11. Ye, P.; Wu, X.; Gao, D.; Deng, S.; Xu, N.; Chen, J. DP3 signal as a neuro-indictor for attentional processing of stereoscopic contents in varied depths within the ‘comfort zone’. Displays 2020, 63, 101953. [Google Scholar] [CrossRef]
  12. Lu, B.; He, Y.; Wang, H. Stereo disparity optimization with depth change constraint based on a continuous video. Displays 2021, 69, 102073. [Google Scholar] [CrossRef]
  13. Gao, Z.; Zhai, G.; Deng, H.; Yang, X. Extended geometric models for stereoscopic 3D with vertical screen disparity. Displays 2020, 65, 101972. [Google Scholar] [CrossRef]
  14. Ye, P.; Kumar, J.; Kang, L.; Doermann, D. Unsupervised feature learning framework for no-reference image quality assessment. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1098–1105. [Google Scholar]
  15. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  16. You, J.; Xing, L.; Perkis, A.; Wang, X. Perceptual quality assessment for stereoscopic images based on 2D image quality metrics and disparity analysis. In Proceedings of the Fifth International Workshop on Video Processing and Quality Metrics for Consumer Electronics, Scottsdale, AZ, USA, 13–15 January 2010; pp. 4033–4036. [Google Scholar]
  17. Howard, I.P.; Rogers, B.J. Seeing in Depth; Oxford University Press: New York, NY, USA, 2008. [Google Scholar]
  18. Shen, L.; Fang, R.; Yao, Y.; Geng, X.; Wu, D. No-Reference Stereoscopic Image Quality Assessment Based on Image Distortion and Stereo Perceptual Information. IEEE Trans. Emerg. Top. Comput. Intell. 2019, 3, 59–72. [Google Scholar] [CrossRef]
  19. Lebreton, P.; Raake, A.; Barkowsky, M.; Callet, P.L. Evaluating depth perception of 3D stereoscopic videos. IEEE J. Sel. Top. Signal Process. 2012, 6, 710–720. [Google Scholar] [CrossRef] [Green Version]
  20. Mikkola, M.; Jumisko-Pyykko, S.; Strohmeier, D.; Boev, A.; Gotchev, A. Stereoscopic depth cues outperform monocular ones on autostereoscopic display. IEEE J. Sel. Top. Signal Process. 2012, 6, 698–709. [Google Scholar] [CrossRef]
  21. Tam, W.J.; Speranza, F.; Yano, S.; Shimono, K.; One, H. Stereoscopic 3D-TV: Visual comfort. IEEE Trans. Broadcast. 2011, 57, 335–346. [Google Scholar] [CrossRef]
  22. Jung, C.; Liu, H.; Cui, Y. Visual comfort assessment for stereoscopic 3D images based on salient discomfort regions. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 4047–4051. [Google Scholar]
  23. Coria, L.; Xu, D.; Nasiopoulos, P. Quality of experience of stereoscopic content on displays of different sizes: A comprehensive subjective evaluation. In Proceedings of the 2011 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 9–12 January 2011; pp. 755–756. [Google Scholar]
  24. Liu, H.; Heynderickx, I. Visual attention in objective image quality assessment: Based on eye-tracking data. IEEE Trans. Circuits Syst. Video Technol. 2011, 21, 971–982. [Google Scholar]
  25. Kim, H.; Lee, S. Transition of Visual Attention Assessment in Stereoscopic Images with Evaluation of Subjective Visual Quality and Discomfort. IEEE Trans. Multimedia 2015, 17, 2198–2209. [Google Scholar] [CrossRef]
  26. Gorley, P.; Holliman, N. Stereoscopic image quality metrics and compression. Proc. SPIE 2008, 6803, 45–56. [Google Scholar]
  27. Benoit, A.; Callet, P.L.; Campisi, P.; Cousseau, R. Quality Assessment of Stereoscopic Images. EURASIP J. Image Video Process. 2008, 2008, 659024. [Google Scholar] [CrossRef] [Green Version]
  28. Wang, S.; Shao, F.; Li, F.; Yu, M.; Jiang, G. A simple quality assessment index for stereoscopic images based on 3d gradient magnitude. Sci. World J. 2014, 2014, 890562. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Lin, Y.-H.; Wu, J.-L. Quality assessment of stereoscopic 3d image compression by binocular integration behaviors. IEEE Trans. Image Process. 2014, 23, 1527–1542. [Google Scholar] [CrossRef] [PubMed]
  30. Chen, M.J.; Su, C.C.; Kwon, D.K.; Cormack, L.K.; Bovik, A.C. Full-reference quality assessment of stereo pairs accounting for rivalry. Signal Process. Image Commun. 2013, 28, 1143–1155. [Google Scholar] [CrossRef] [Green Version]
  31. Bensalma, R.; Larabi, M.-C. A perceptual metric for stereoscopic image quality assessment based on the binocular energy. Multidimens. Syst. Signal Process. 2013, 24, 281–316. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Chandler, D.M. 3D-MAD: A Full Reference Stereoscopic Image Quality Estimator Based on Binocular Lightness and Contrast Perception. IEEE Trans. Image Process. 2015, 24, 3810–3825. [Google Scholar] [CrossRef]
  33. Li, F.; Shen, L.; Wu, D.; Fang, R. Full-reference quality assessment of stereoscopic images using disparity-gradient-phase similarity. In Proceedings of the 2015 IEEE China Summit and International Conference on Signal and Information Processing (ChinaSIP), Chengdu, China, 12–15 July 2015; pp. 658–662. [Google Scholar]
  34. Jiang, G.; Xu, H.; Yu, M.; Luo, T.; Zhang, Y. Stereoscopic Image Quality Assessment by Learning Non-negative Matrix Factorization-based Color Visual Characteristics and Considering Binocular Interactions. J. Vis. Commun. Image Represent. 2017, 46, 269–279. [Google Scholar] [CrossRef]
  35. Geng, X.; Shen, L.; Li, K.; An, P. A stereoscopic image quality assessment model based on independent component analysis and binocular fusion property. Signal Process. Image Commun. 2017, 52, 54–63. [Google Scholar] [CrossRef]
  36. Shao, F.; Li, K.; Lin, W.; Jiang, G.; Dai, Q. Learning blind quality evaluator for stereoscopic images using joint sparse representation. IEEE Trans. Multimedia 2016, 18, 2104–2114. [Google Scholar] [CrossRef]
  37. Akhter, R.; Baltes, J.; Sazzad, Z.M.P.; Horita, Y. No reference stereoscopic image quality assessment. Proc. SPIE 2010, 7524, 75240T. [Google Scholar]
  38. Chen, M.J.; Cormack, L.K.; Bovik, A.C. No-reference quality assessment of natural stereopairs. IEEE Trans. Image Process. 2013, 22, 3379–3391. [Google Scholar] [CrossRef] [PubMed]
  39. Wang, S.; Shao, F.; Jiang, G. Supporting binocular visual quality prediction using machine learning. In Proceedings of the 2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), Chengdu, China, 14–18 July 2014; pp. 1–6. [Google Scholar]
  40. Ryu, S.; Sohn, K. No-Reference Quality Assessment for Stereoscopic Images Based on Binocular Quality Perception. IEEE Trans. Circuits Syst. Video Technol. 2014, 24, 591–602. [Google Scholar]
  41. Shao, F.; Lin, W.; Wang, S.; Jiang, G.; Yu, M. Blind Image Quality Assessment for Stereoscopic Images Using Binocular Guided Quality Lookup and Visual Codebook. IEEE Trans. Broadcast. 2015, 61, 154–165. [Google Scholar] [CrossRef]
  42. Shao, F.; Lin, W.; Wang, S.; Jiang, G.; Yu, M.; Dai, Q. Learning receptive fields and quality lookups for blind quality assessment of stereoscopic images. IEEE Trans. Cybern. 2016, 46, 730–743. [Google Scholar] [CrossRef] [PubMed]
  43. Su, C.C.; Cormack, L.K.; Bovik, A.C. Oriented Correlation Models of Distorted Natural Images With Application to Natural Stereopair Quality Evaluation. IEEE Trans. Image Process. 2015, 24, 1685–1699. [Google Scholar] [CrossRef] [PubMed]
  44. Wang, J.; Zeng, K.; Wang, Z. Quality prediction of asymmetrically distorted stereoscopic images from single views. In Proceedings of the 2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), Chengdu, China, 14–18 July 2014; pp. 1–6. [Google Scholar]
  45. Wang, J.; Rehman, A.; Zeng, K.; Wang, S.; Wang, Z. Quality Prediction of Asymmetrically Distorted Stereoscopic 3D Images. IEEE Trans. Image Process. 2015, 24, 3400–3414. [Google Scholar] [CrossRef]
  46. Zhao, Y.; Zhang, Y.; Yu, L. Subjective study of binocular rivalry in stereoscopic images with transmission and compression artifacts. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013; pp. 132–135. [Google Scholar]
  47. Zhang, L.; Zhang, L.; Tao, D.; Huang, X.; Du, B. Compression of hyperspectral remote sensing images by tensor approach. Neurocomputing 2015, 147, 358–363. [Google Scholar] [CrossRef]
  48. Du, B.; Zhang, M.; Zhang, L.; Hu, R.; Tao, D. PLTD: Patch-Based Low-Rank Tensor Decomposition for Hyperspectral Images. IEEE Trans. Multimed. 2017, 19, 67–79. [Google Scholar] [CrossRef]
  49. Zhou, W.; Qiu, W.; Wu, M. Utilizing dictionary learning and machine learning for Blind Quality Assessment of 3-D Images. IEEE Trans. Broadcast. 2017, 63, 404–415. [Google Scholar] [CrossRef]
  50. Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K. Evaluating the visualization of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2660–2673. [Google Scholar] [CrossRef] [Green Version]
  51. Zhou, M.; Li, S. Deformable Convolution Based No-Reference Stereoscopic Image Quality Assessment Considering Visual Feedback Mechanism. In Proceedings of the 2021 International Conference on Visual Communications and Image Processing (VCIP), Munich, Germany, 5–8 December 2021; pp. 1–5. [Google Scholar]
  52. Jinhui, F.; Li, S.; Chang, Y. No-Reference Stereoscopic Image Quality Assessment Considering Binocular Disparity and Fusion Compensation. In Proceedings of the 2021 International Conference on Visual Communications and Image Processing (VCIP), Munich, Germany, 5–8 December 2021; pp. 1–5. [Google Scholar]
  53. Bourbia, S.; Karine, A.; Chetouani, A.; Hassoun, M.E. A Multi-Task Convolutional Neural Network For Blind Stereoscopic Image Quality Assessment Using Naturalness Analysis. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1434–1438. [Google Scholar]
  54. Sandić-Stanković, D.D.; Kukolj, D.D.; Callet, P.L. Quality Assessment of DIBR-Synthesized Views Based on Sparsity of Difference of Closings and Difference of Gaussians. IEEE Trans. Image Process. 2022, 31, 1161–1175. [Google Scholar] [CrossRef] [PubMed]
  55. Gu, K.; Tao, D.; Qiao, J.; Lin, W. Learning a no-reference quality assessment model of enhanced images with big data. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 1301–1313. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Bosse, S.; Maniry, D.; Müller, K.; Wiegand, T.; Samek, W. Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process. 2018, 27, 206–219. [Google Scholar] [CrossRef] [Green Version]
  57. Oh, H.; Ahn, S.; Kim, J.; Lee, S. Blind deep S3D image quality evaluation via local to global feature aggregation. IEEE Trans. Image Process. 2017, 26, 4923–4936. [Google Scholar] [CrossRef]
  58. Hoa, D.K.; Dung, L.; Dzung, N.T. Efficient determination of disparity map from stereo images with modified sum of absolute differences (SAD) algorithm. In Proceedings of the 2013 International Conference on Advanced Technologies for Communications (ATC 2013), Ho Chi Minh City, Vietnam, 16–18 October 2013; pp. 657–660. [Google Scholar]
  59. Howard, P.l.; Rogers, B.J. Binocular Vision and Stereopsis; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  60. Levelt, W.J.M. On Binocular Rivalry. Ph.D. Thesis, Leiden University, Leiden, The Netherlands, 1965. [Google Scholar]
  61. Hubel, D.H. The visual cortex of the brain. Sci. Am. 1963, 209, 54–63. [Google Scholar] [CrossRef]
  62. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  63. Fang, R.; Al-Bayaty, R.; Wu, D. BNB Method for No-Reference Image Quality Assessment. IEEE Trans. Circuits Syst. Video Tech. 2017, 27, 1381–1391. [Google Scholar] [CrossRef]
  64. Sheikh, H.R.; Bovik, A.C.; de Veciana, G. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans Image Process. 2005, 14, 2117–2128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Fang, Y.; Ma, K.; Wang, Z.; Lin, W.; Fang, Z.; Zhai, G. No-Reference Quality Assessment of Contrast-Distorted Images Based on Natural Scene Statistics. IEEE Signal Process. Lett. 2015, 22, 838–842. [Google Scholar] [CrossRef]
  66. Md, S.K.; Appina, B.; Channappayya, S.S. Full-Reference Stereo Image Quality Assessment Using Natural Stereo Scene Statistics. IEEE Signal Process. Lett. 2015, 22, 1985–1989. [Google Scholar]
  67. Chang, C.C.; Lin, C.J. LIBSVM: A Library for Support Vector Machines. 2001. Available online: http://www.csie.ntu.edu.tw/cjlin/libsvm/ (accessed on 1 January 2020).
  68. Moorthy, A.K.; Su, C.-C.; Mittal, A.; Bovik, A.C. Subjective evaluation of stereoscopic image quality. Signal Process. Image Commun. 2013, 28, 870–883. [Google Scholar] [CrossRef]
  69. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef] [PubMed]
  70. Gholipour, A.; Araabi, B.N.; Lucas, C. Predicting chaotic time series using neural and neurofuzzy models: A comparative study. Neural Process. Lett. 2006, 24, 217–239. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed NR-SIQA algorithm.
Figure 1. Framework of the proposed NR-SIQA algorithm.
Sensors 23 06230 g001
Figure 2. Disparities estimated by different disparity estimation algorithms: (a) SAD-based algorithm, (b) SSIM-based algorithm, (c) Gaussian average SSIM-based algorithm.
Figure 2. Disparities estimated by different disparity estimation algorithms: (a) SAD-based algorithm, (b) SSIM-based algorithm, (c) Gaussian average SSIM-based algorithm.
Sensors 23 06230 g002
Figure 3. Cyclopean images synthesized by distorted stereo-pairs. (a,b) and (d,e) are left and right views of stereo-pairs, respectively. (c,f) are their respective cyclopean images.
Figure 3. Cyclopean images synthesized by distorted stereo-pairs. (a,b) and (d,e) are left and right views of stereo-pairs, respectively. (c,f) are their respective cyclopean images.
Sensors 23 06230 g003
Figure 4. Probability density distributions of (a) neighbor difference, (b) neighbor product and (c) vertical gradient magnitude for six natural cyclopean images synthesized by a reference stereo-pair and its five distorted versions.
Figure 4. Probability density distributions of (a) neighbor difference, (b) neighbor product and (c) vertical gradient magnitude for six natural cyclopean images synthesized by a reference stereo-pair and its five distorted versions.
Sensors 23 06230 g004
Figure 5. Probability density distributions of (a) phase congruency, (b) real part of log-Gabor response, (c) log-Gabor phase response, (d) binocular disparity, (e) binocular disparity matching error and (f) binocular disparity consistency computed from a reference stereo-pair and its five distorted versions.
Figure 5. Probability density distributions of (a) phase congruency, (b) real part of log-Gabor response, (c) log-Gabor phase response, (d) binocular disparity, (e) binocular disparity matching error and (f) binocular disparity consistency computed from a reference stereo-pair and its five distorted versions.
Sensors 23 06230 g005
Figure 6. Various pairwise differences computed to quantify neighboring pixel statistical correlation. Neighboring pixel differences are computed along four orientations—horizontal (H), vertical (V), main(on)-diagonal (D1) and secondary(off)-diagonal (D2).
Figure 6. Various pairwise differences computed to quantify neighboring pixel statistical correlation. Neighboring pixel differences are computed along four orientations—horizontal (H), vertical (V), main(on)-diagonal (D1) and secondary(off)-diagonal (D2).
Sensors 23 06230 g006
Figure 7. Various pairwise products computed to quantify neighboring pixels’ statistical correlations. Neighboring pixel products are computed along eight orientations— 0 , 22 . 5 , 45 , 67 . 5 , 90 , 112 . 5 , 135 and 157 . 5 —at a distance of 2 pixels from central pixel.
Figure 7. Various pairwise products computed to quantify neighboring pixels’ statistical correlations. Neighboring pixel products are computed along eight orientations— 0 , 22 . 5 , 45 , 67 . 5 , 90 , 112 . 5 , 135 and 157 . 5 —at a distance of 2 pixels from central pixel.
Sensors 23 06230 g007
Figure 8. PLCC of the proposed metric with various feature dimensions from LIVE Phase I and Phase II.
Figure 8. PLCC of the proposed metric with various feature dimensions from LIVE Phase I and Phase II.
Sensors 23 06230 g008
Figure 9. The SVR parameters’ (C, γ ) selection process on (a) LIVE 3D Database Phase I and (b) Phase II. The number on the level contour denotes the SROCC value of the cross-validation.
Figure 9. The SVR parameters’ (C, γ ) selection process on (a) LIVE 3D Database Phase I and (b) Phase II. The number on the level contour denotes the SROCC value of the cross-validation.
Sensors 23 06230 g009
Table 1. The categorization of SIQA studies according to the considered visual characteristics and some implementation details.
Table 1. The categorization of SIQA studies according to the considered visual characteristics and some implementation details.
AlgorithmCYCNSSBDBFBRG/PRMSDTD
Gorley [22]XXXXXG-OX
You [35]XXOXXG-OX
Benoit [36]XXOXXG-OX
Chen [38]OXOOOG-OX
Chen [14]OOOOOGSVROX
Akhter [41]XXOXXP-OX
Ryu [42]XXXOOG-OO
Wang [45]XXOXXPSVRXO
Shao [27]XXOOOP-XO
Shao [43]XXOOOP-XO
Shao [44]XXXXXPSVRXO
Fang [46]OOOXOGNNOX
Karimi [47]OOOOXGNNOX
ProposedOOOOOGSVROO
Table 2. Performance on the training set and testing set as the regression model learned from the TRAINING set and TESTING set, respectively.
Table 2. Performance on the training set and testing set as the regression model learned from the TRAINING set and TESTING set, respectively.
DatabaseLIVE Phase ILIVE Phase II
Training/TestingSROCCPLCCRMSESROCCPLCCRMSE
MTrain→Training set0.9780.9793.2650.9710.9762.452
MTrain→Testing set0.9480.9584.8070.9330.9393.929
MTest→Training set0.9480.9564.8260.9340.9403.881
MTest→Testing set0.9770.9803.2820.9710.9762.482
Table 3. Performance on training set and testing set using a regression model learned from the features of a training set with Gaussian white noise added or with no noise.
Table 3. Performance on training set and testing set using a regression model learned from the features of a training set with Gaussian white noise added or with no noise.
DatabaseLIVE Phase ILIVE Phase II
Training/TestingSROCCPLCCRMSESROCCPLCCRMSE
WN→Training set0.9620.9674.2340.9500.9543.386
WN→Testing set0.9440.9544.9560.9240.9364.000
WN/F→Training set0.9760.9783.4010.9690.9732.259
WN/F→Testing set0.9520.9624.4930.9400.9503.546
Table 4. Performance on LIVE 3D database Phase I and II compared with the existing NR SIQA methods.
Table 4. Performance on LIVE 3D database Phase I and II compared with the existing NR SIQA methods.
DatabaseLIVE Phase ILIVE Phase II
AlgorithmSROCCPLCCRMSESROCCPLCCRMSE
Akhter [41]0.3830.62614.8270.5430.5689.249
Chen [14]0.8910.8957.2470.8800.8955.102
Wang [45]0.8280.8857.2380.7940.7847.326
Shao [27]0.8940.899----
Shao [44]0.9500.957----
Fang [46]0.9320.936-0.9310.936-
Proposed 0.9520.9624.4930.9400.9503.546
Table 5. Performance on Waterloo IVC 3D database Phase I and II compared with the existing FR and NR SIQA methods.
Table 5. Performance on Waterloo IVC 3D database Phase I and II compared with the existing FR and NR SIQA methods.
DatabaseUW/IVC Phase IUW/IVC Phase II
AlgorithmSROCCPLCCSROCCPLCC
You [35]0.5970.7130.5870.682
Chen [38]0.6820.7340.5780.613
Mittal [58]0.8450.8690.7940.849
Chen [14]0.7080.7150.5470.551
Proposed0.9060.9190.8520.863
Table 6. Performance on each type of distortion on LIVE 3D database Phase I and Phase II.
Table 6. Performance on each type of distortion on LIVE 3D database Phase I and Phase II.
DatabaseLIVE Phase ILIVE Phase II
AlgorithmJP2KJPEGWNBLURFFALLJP2KJPEGWNBLURFFALL
Benoit [36]0.9100.6030.9300.9310.6990.8990.7510.8670.9230.4550.7730.728
You [35]0.8600.4390.9400.8820.5880.8780.8940.7950.9090.8130.8910.786
Gorley [22]0.0150.5690.7410.7500.3660.1420.1100.0270.8750.7700.6010.146
Chen [38]0.8880.5300.9480.9250.7070.9160.8140.8430.9400.9080.8840.889
SROCCShao [27]0.9000.6070.9260.924-0.894------
Shao [44]0.9360.8180.9350.9270.8140.950------
Akhter [41]0.9140.8660.6750.5550.6400.3830.7140.7240.6490.6820.5590.543
Chen [14]0.9190.8630.6170.8780.6520.8910.9500.8670.8670.9000.9330.880
Proposed0.9370.7790.9590.9210.8510.9520.9460.9030.8310.9120.9390.940
Benoit [36]0.9390.6400.9250.9480.7470.9020.7840.8530.9260.5350.8070.748
You [35]0.8770.4870.9410.9190.7300.8810.9050.8300.9120.7840.9150.800
Gorley [22]0.4850.3120.7960.8520.3640.4510.3720.3220.8740.9340.7060.515
Chen [38]0.9120.6030.9420.9420.7760.9170.8340.8620.9570.9630.9010.900
PLCCShao [27]0.8720.5970.9160.923-0.899------
Shao [44]0.9490.7960.9380.9860.8370.957------
Akhter [41]0.9040.9050.7290.6170.5030.6260.7220.7760.7860.7950.6740.568
Chen [14]0.9170.9070.6950.9170.7350.8950.9470.8990.9010.9410.9320.895
Proposed0.9580.8010.9710.9650.8830.9620.9740.9220.8580.9770.9490.950
Benoit [36]4.4265.0226.3074.5718.2577.0616.0963.7874.02811.7636.8947.490
You [35]6.2065.7095.6215.6798.4927.7464.1864.0864.3968.6494.6496.772
Gorley [22]11.3236.21110.1977.56211.56914.6359.1136.9405.2024.9888.1559.675
Chen [38]5.3205.2165.5814.8227.8376.5335.5623.8653.3683.7474.9664.987
RMSEShao [27]------------
Shao [44]------------
Akhter [41]7.0925.4834.27311.3879.33214.8277.4166.1894.5358.4508.5059.249
Chen [14]6.4335.4024.5235.8988.3227.2473.5134.2983.3424.7254.1805.102
Proposed3.9383.9083.9124.0555.8124.4932.5753.7732.8433.2073.6063.546
Table 7. Performance with feature type increasing on LIVE 3D database Phase I and Phase II.
Table 7. Performance with feature type increasing on LIVE 3D database Phase I and Phase II.
DatabaseLIVE Phase ILIVE Phase II
Feature SetSROCCPLCCRMSESROCCPLCCRMSE
PC0.8680.8797.3150.7360.7757.057
PC+GM0.8930.9126.8120.7930.8575.852
PC+GM+GR0.9030.9226.2010.8320.8695.031
PC+GM+GR+ND0.9310.9435.8150.8740.8914.764
PC+GM+GR+ND+NP0.9380.9485.1280.9070.9184.197
PC+GM+GR+ND+NP+3DI0.9480.9574.9140.9320.9393.867
Table 8. Performance on LIVE 3D database Phase II training on Phase I.
Table 8. Performance on LIVE 3D database Phase II training on Phase I.
CriteriaJP2KJPEGWNBLURFFALL
SROCC0.8640.5730.8880.8830.8720.802
PLCC0.8590.5810.9080.9590.8890.822
RMSE5.0295.9674.6253.9485.2616.427
Table 9. Performance on LIVE 3D database Phase I training on Phase II.
Table 9. Performance on LIVE 3D database Phase I training on Phase II.
CriteriaJP2KJPEGWNBLURFFALL
SROCC0.8860.5400.8790.9150.7870.867
PLCC0.9280.5730.8890.9440.8350.872
RMSE4.8435.3597.6184.7796.8458.014
Table 10. SROCC on LIVE 3D database Phase I for cross-distortion performance evaluation.
Table 10. SROCC on LIVE 3D database Phase I for cross-distortion performance evaluation.
TestJP2K JPEGWNBLURFFALL
Train
JP2K-0.7580.9110.9360.6640.898
JPEG0.586-0.9260.7640.4980.746
WN0.8050.552-0.8880.7240.848
BLUR0.8780.5530.921-0.7370.889
FF0.8380.4700.9330.884-0.874
Table 11. SROCC on LIVE 3D database Phase II for cross-distortion performance evaluation.
Table 11. SROCC on LIVE 3D database Phase II for cross-distortion performance evaluation.
TestJP2K JPEGWNBLURFFALL
Train
JP2K-0.7570.9450.7540.9060.793
JPEG0.943-0.8120.6930.7000.758
WN0.8630.588-0.6210.8330.778
BLUR0.9450.3860.922-0.8830.818
FF0.9300.7820.9220.841-0.887
Table 12. Performance evaluation of random database partitioning and image content-based partitioning.
Table 12. Performance evaluation of random database partitioning and image content-based partitioning.
DatabaseLIVE Phase ILIVE Phase II
Partition MethodSROCCPLCCRMSESROCCPLCCRMSE
Random-based partition0.9520.9624.4930.9400.9503.546
Content-based partition0.9080.9286.7520.8670.8666.685
Table 13. Performance evaluation of different disparity estimation algorithms.
Table 13. Performance evaluation of different disparity estimation algorithms.
DatabaseLIVE Phase ILIVE Phase II
AlgorithmSROCCPLCCRMSESROCCPLCCRMSE
SAD-based0.9520.9604.5810.9270.9413.846
SSIM-based0.9460.9564.8410.9370.9453.727
Improved SSIM-based0.9520.9624.4930.9400.9503.546
Table 14. Complexity analysis for the proposed method.
Table 14. Complexity analysis for the proposed method.
StepsTime Ratios
SSIM-Based Matching and Disparity Generation36.37%
Cyclopean Generation2.48%
Spatial Domain Features1.95%
Transform Domain Features7.42%
3D Perceptual Features0.66%
PCA and SVM51.12%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, L.; Yao, Y.; Geng, X.; Fang, R.; Wu, D. A Novel No-Reference Quality Assessment Metric for Stereoscopic Images with Consideration of Comprehensive 3D Quality Information. Sensors 2023, 23, 6230. https://doi.org/10.3390/s23136230

AMA Style

Shen L, Yao Y, Geng X, Fang R, Wu D. A Novel No-Reference Quality Assessment Metric for Stereoscopic Images with Consideration of Comprehensive 3D Quality Information. Sensors. 2023; 23(13):6230. https://doi.org/10.3390/s23136230

Chicago/Turabian Style

Shen, Liquan, Yang Yao, Xianqiu Geng, Ruigang Fang, and Dapeng Wu. 2023. "A Novel No-Reference Quality Assessment Metric for Stereoscopic Images with Consideration of Comprehensive 3D Quality Information" Sensors 23, no. 13: 6230. https://doi.org/10.3390/s23136230

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop