Next Article in Journal
An Improved Forest Fire Detection Method Based on the Detectron2 Model and a Deep Learning Approach
Next Article in Special Issue
Pixel Intensity Resemblance Measurement and Deep Learning Based Computer Vision Model for Crack Detection and Analysis
Previous Article in Journal
Autonomous Intersection Management: Optimal Trajectories and Efficient Scheduling
Previous Article in Special Issue
Privacy Preserving Image Encryption with Optimal Deep Transfer Learning Based Accident Severity Classification Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Blind Video Quality Assessment for Ultra-High-Definition Video Based on Super-Resolution and Deep Reinforcement Learning

School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1511; https://doi.org/10.3390/s23031511
Submission received: 4 January 2023 / Revised: 25 January 2023 / Accepted: 26 January 2023 / Published: 29 January 2023

Abstract

:
Ultra-high-definition (UHD) video has brought new challenges to objective video quality assessment (VQA) due to its high resolution and high frame rate. Most existing VQA methods are designed for non-UHD videos—when they are employed to deal with UHD videos, the processing speed will be slow and the global spatial features cannot be fully extracted. In addition, these VQA methods usually segment the video into multiple segments, predict the quality score of each segment, and then average the quality score of each segment to obtain the quality score of the whole video. This breaks the temporal correlation of the video sequences and is inconsistent with the characteristics of human visual perception. In this paper, we present a no-reference VQA method, aiming to effectively and efficiently predict quality scores for UHD videos. First, we construct a spatial distortion feature network based on a super-resolution model (SR-SDFNet), which can quickly extract the global spatial distortion features of UHD videos. Then, to aggregate the spatial distortion features of each UHD frame, we propose a time fusion network based on a reinforcement learning model (RL-TFNet), in which the actor network continuously combines multiple frame features extracted by SR-SDFNet and outputs an action to adjust the current quality score to approximate the subjective score, and the critic network outputs action values to optimize the quality perception of the actor network. Finally, we conduct large-scale experiments on UHD VQA databases and the results reveal that, compared to other state-of-the-art VQA methods, our method achieves competitive quality prediction performance with a shorter runtime and fewer model parameters.

1. Introduction

UHD video can bring viewers a better visual experience because of its high resolution, high frame rate, high dynamic range and wide color gamut. However, the quality of UHD video is subject to various distortions and noises during the processes of signal capturing, encoding, storing, transmission and displaying. To better monitor and control the video quality in each stage, it is essential to develop an accurate and efficient UHD video-quality-assessment algorithm.
Video quality assessment (VQA) is a technology that can automatically measure video quality. Based on whether the reference video is accessible, VQA can be categorized into three types: full-reference VQA (FR-VQA) [1,2], reduced reference VQA (RR-VQA), and no-reference VQA (NR-VQA). The FR-VQA methods predict the video quality by comparing the reference and distorted videos. The RR-VQA methods employ only partial information from the reference videos for quality prediction. In contrast, the NR-VQA methods can perform quality prediction without using any information from the reference video.
NR-VQA is also known as blind VQA [3,4]. Since no reference video is needed, BVQA can better meet the needs of real-world applications, such as the video quality of experience (QoE). The ITU-T study group designed a no-reference model for UHD videos and standardized it to P.1204.5 [5]. This method uses distorted video and other information, such as the encoding format and bit rate, for quality assessment. Inspired by this method, STEVE et al. proposed the Nofu model [6], which extracts 64 features, such as colorfulness, tone, and saturation from the distorted UHD video, without additional information, and predicts the quality by the support vector regression (SVR). Moreover, most of the current NR-VQA methods are still aimed at low-resolution videos, such as [7,8]. They do not perform well when applied to UHD videos. The main reasons are as follows:
First, the VQA task is closely correlated with the perception of distortion information. However, the current BVQA methods [9,10,11] usually use backbone convolutional neural networks to extract global spatial distortion features. Such networks are designed for computer vision tasks, such as image classification, focusing more on objects rather than distortions. While these methods perform well for those videos with relatively low resolution, because viewers tend to pay more attention to objects which usually appear in the center of the screen, they do not work well for UHD videos. For UHD videos, the attention of viewers is not only focused on the objects in the screen center, but is also attracted by the content in other regions of the screen. These methods that focus on objects cannot fully extract the distortion information outside the object region and, thus, cannot effectively predict the quality of UHD videos. In addition, due to the high resolution of UHD videos, the current methods often need to divide a UHD video frame into image patches of small size, and then perform quality prediction based on image patches instead of the whole video frame. This makes it hard for the backbone neural network to fully extract the global spatial distortion features.
In contrast, the video super-resolution task is more related to the VQA task and both of them aim at how to better extract the features related to video quality. The process of video super-resolution involves dealing with the content feature and distortion feature. The content feature is extracted and used to generate high-quality and high-resolution videos, while the distortion information is not used in the super-resolution task. Because the super-resolution model can convert low-resolution video into high-resolution video, the distortion information hidden in the low-resolution video can reflect the distortion distribution of the high-resolution video well. Therefore, we can use the distortion information hidden in the low-resolution video to predict the UHD video quality by embedding the super-resolution model into the VQA model. Figure 1 shows the relationship between the quality assessment and the super-resolution task. The red box in Figure 1 shows the structure of a common super-resolution model. The model first extracts the content features through the super-resolution feature extraction network, and then the content features are used to output high-quality and high-resolution frames by an up-sampling network. The green box in Figure 1 shows the structure of a VQA model proposed in this paper in which the distortion features from the super-resolution model are used in the follow-up quality prediction. Furthermore, in this structure, the input to the feature extraction network is the low-resolution version of the UHD video frame rather than the segmented image patches, so the frame-level global spatial distortion features can be effectively extracted.
Second, due to the limitation of computing power, most VQA models are unable to continuously handle the temporal features of UHD videos with high frame rates. Instead, these models usually have to segment the video into multiple segments, predict the quality score of each segment, and then average these segment quality scores to obtain the video quality score, as shown in Figure 2a. This kind of operation breaks the temporal correlation of the video sequences and leads to discontinuities in the temporal features. For example, if the quality of the first segment is degraded, it will not affect the quality prediction of the subsequent segment. This is not consistent with the characteristics of human visual perception. Human visual perception of video quality is continuous rather than segment-by-segment, as shown in Figure 2b. To simulate this process, we use a reinforcement learning model to continuously extract the temporal features.
According to the above analysis, we propose a blind video quality assessment method for UHD video based on super-resolution and deep reinforcement learning. First, we construct a spatial distortion feature network based on a super-resolution model (SR-SDFNet) to extract the global spatial distortion features of UHD frames. Then, a time fusion network based on reinforcement learning (RL-TFNet) is employed to aggregate the spatial distortion features of each UHD frame to obtain the UHD video quality score. For the SR-SDFNet, as shown in the green box in Figure 1, the UHD video frames are downsampled before being fed into the super-resolution model for extraction of the global spatial distortion features. This significantly reduces the computational time. For the RL-TFNet, as shown in Figure 2b, the video quality score is adjusted iteratively by combining the global spatial distortion features of each frame with the history quality score. When the last frame is input, the model outputs the final video quality score.
We make three main contributions:
(1) We propose a global spatial distortion feature network based on a super-resolution model (SR-SDFNet) to efficiently and effectively extract the global spatial distortion features from the UHD frames.
(2) We propose a time fusion network based on reinforcement learning (RL-TFNet) to continuously fuse the global spatial distortion features of each frame to obtain the whole video quality score, which can ensure the continuity of the video temporal domain and improve the prediction accuracy for the UHD video quality.
(3) By combining the SR-SDFNet with the RL-TFNet, we present an NR-VQA model BVQA-SR&DRL for UHD videos. Our model is driven by distorted UHD videos without human subjective scores. We evaluate our model on two public UHD-VQA databases. The result demonstrates that our model achieves state-of-the-art performance compared with existing BIQA methods.

2. Related Work

2.1. BVQA

The current UHD video quality assessment methods usually use the VQA model designed for low-resolution video, apart from the above two methods P.1204.5 [5] and Nofu [6]. Most of these methods are based on machine learning. Zhang et al. [12] used a 3D-DCT transform to extract spatiotemporal features of videos and fed the extracted features into a convolutional neural network to regress perceptual quality scores. CNN-MR [13] and COME [14] used the 2D CNN AlexNet to extract spatial features of videos, which were combined with motion features, and then used support vector regression (SVR) to obtain the final quality score. Chen et al. proposed a concept of temporal resolution and designed RIRNet [15] to extract temporal features at different temporal resolutions for quality perception. Since RIRNet has many temporal resolution inputs, the computational complexity will be higher than for other models. Dendi et al. [16] proposed an asymmetric generalized Gaussian distribution (AGGD) to simulate the MSCN coefficients of distorted video and the statistics of the bandpass filtered output, which are used to predict the quality score by SVR. ChipQA [17] tracked and cropped video regions where motion information existed to obtain localized spatiotemporal slices, and outputted natural video statistical parameters for those extracted slices to perceive video quality. Chen et al. [18] extracted multi-scale distortion features using a VGG16 network combined with an attention mechanism and proposed a pyramid aggregation model along the temporal dimension to obtain the final video-level quality score. Xu et al. [19] extracted time-space features of UGC videos through graph convolution and attention blocks, and built a long short-term memory network to integrate distortion features and obtain the video quality score. Varga et al. [20] proposed an FLG-VQA model which extracts and integrates local and global image statistics features for quality perception. Li et al. [21] proposed a bidirectional GRU network to predict UGC frame scores and integrated the quality scores of multiple frames by constructing temporal memory blocks. According to the characteristics of compressed videos, Lin et al. [22] combined perceivable encoding artifacts (PEAs) detection and visual saliency perception to output the final quality score. However, for UHD videos with large spatial resolution and a high frame rate, the above methods have unsatisfactory performance and slow operation speed. Therefore, our method aims to achieve improvements in terms of both efficiency and effectiveness.

2.2. Deep Reinforcement Learning

DRL [23] imitates the learning process of the human brain. The model can find the optimal strategy through constant trials and errors in the environment. In addition, DRL can complete the construction of knowledge models even when environmental information is insufficient. Early works on DRL mainly focused on robotic control and game agents. In recent years, DRL has become a research hotspot in computer vision.
Zhang et al. [24] proposed a DRL model called EBSNet to automatically select appropriate exposure images and combine it with MEFNet to generate high-dynamic-range images. Sun et al. [25] achieved referring expression grounding through a reinforcement learning model, which predicts the possible orientation of the object at each iteration to localize the target object. Nauata et al. [26] combined the DRL network with the GAN network to achieve automated floorplan generation. Wang et al. [27] applied the DRL model, which reduces the computational complexity while ensuring performance, to the field of face video segmentation. Lu et al. [28] regarded coronary CT angiography vessel-level image quality assessment (CCTA VIQA) as a multi-instance learning (MIL) problem, and constructed a reinforcement learning model to gradually select key instances for subsequent quality assessment. Saeed et al. [29] proposed a meta-RL-based meta-learning model to improve the adaptability of the common quality assessment model to the task-specific IQA, such as pneumonia detection in X-ray images and other clinical applications.
However, to our knowledge, RL has never been used for video quality assessment tasks. Existing VQA methods usually divide the entire video into multiple video segments and then process them. We argue that such segmentation is inconsistent with the actual subjective perception process. Therefore, our method attempts to process the entire video with an RL model, which predicts the quality fluctuations that each video frame brings to the entire video, thereby preserving complete and continuous video temporal features.

3. Proposed Method

3.1. BVQA-SR&DRL

We propose a BVQA method BVQA-SR&DRL that aims to efficiently and accurately predict the quality score of the UHD distorted video. It consists of two components: a spatial distortion feature network based on a super-resolution model (SR-SDFNet) and a time fusion network based on reinforcement learning (RL-TFNet). Figure 3 illustrates the main steps of our method. First, we downsample the UHD video frames to obtain the low-resolution video frames and input them into the SR-SDFNet for spatial distortion feature extraction. Guided by the super-resolution network, the spatial distortion features extracted by the SR-SDFNet can reflect the distortion distribution of the UHD frames well.
Second, the proposed RL-TFNet observes the spatial features extracted by SR-SDFNet, as well as the historical quality score S n 1 , and outputs the adjust value Δ S n to increase or decrease the quality score from the first frame, and iteratively adjusts the quality score until the last frame, finally outputting the overall video quality score. The prediction process of the entire video score is shown in the line graph of Figure 3.

3.2. Spatial Distortion Feature Network Based on a Super-Resolution Model (SR-SDFNet)

To extract high-resolution spatial distortion features quickly and completely, we employ the generator network SRResNet used in the super-resolution method ESRGAN [30] to construct SR-SDFNet. Specifically, SRResNet can be divided into a feature network N f e a t and an upsample network N u p , as Figure 4 shows. The feature network N f e a t consists of four residual-in-residual dense blocks (RRDB), which are used to extract content features and distortion features. Each RRDB contains three residual networks and uses dense connections to combine.
In the task of SR, the content features F _ c o n t e n t n L H are extracted by the feature network N f e a t and used to output high-quality frames I n o u t p u t by an upsample network N u p , as shown in Formulas (1) and (2). Considering the spatial features of a frame consists of content features and distortion features, we extract the distortion features by making a difference between the frame and the content features in the task of VQA.
F _ c o n t e n t n L H = N feat I n L
I n o u t p u t = N up F _ c o n t e n t n L H
In detail, we first downsample the UHD video frame I n H (3840 × 2160) by four times to get the low-resolution video frame I n L (960 × 540). Then, I n L is fed into SR-SDFNET to extract the global spatial distortion features. In detail, SR-SDFNet diffs the I n L with the output of N f e a t to get the information F n L H , which is adopted as the distortion feature for the UHD frame, as shown in the Figure 5.
I n L = Downsampling I n H
F n L H = I n L N feat I n L
To train SR-SDFNET, we first train the super-resolution model ESRGAN based on the VQA database. The inputs to ESRGAN are the distorted frames which are downsampled by a factor of four, and the corresponding reference frames are used as the ground truth of the ESRGAN model. After the losses of ESRGAN stop falling, we use the feature network N f e a t in SRResNet as the pre-training parameter of SR-SDFNet. Most existing VQA methods use sliding window segmentation on UHD video frames to predict the quality score, which consumes a lot of time and computation. In contrast, the proposed SR-SDFNet only needs to input low-resolution video frames and, thus, most computation is performed in the low-resolution feature space. In this way, the algorithm efficiency is greatly improved. Furthermore, because the spatial distortion feature extraction is based on the video frames instead of the segmented image patches, the integrity of the spatial distortion feature is guaranteed.

3.3. Time Fusion Network Based on Reinforcement Learning (RL-TFNet)

To continuously integrate the spatial quality features extracted by SR-SDFNet in the temporal dimension, we formulate the video quality assessment task as a Markov decision process (MDP) and propose the RL-TFNet model, which simulates the process of human visual perception of video quality. As shown in Figure 6, we initially set the video quality score S 0 to one (i.e., the highest quality). The RL-TFNet model consists of the actor network and critic network, in which the actor network iteratively adjusts the quality score according to the spatial distortion features extracted and the historical quality score, as shown in the formula. In addition, the critic network outputs action values to optimize the quality perception of the actor network.
Since the subjective quality scores of each video frame used as the ground truth are hard to obtain, the training of RL-TFNet is guided by a reward generated from VAMF-4K [31], which is a widely used FR-VQA method based on VMAF. VMAF-4K mainly extracts three features, the visual quality fidelity (VIF) [32], a detail loss measure (DLM), and temporal information (TI), and integrates them to calculate the video quality score. The proposed RL-TFNet contains the following important factors: the state space, the action space and the definition of the reward function, which will be introduced in detail.

3.3.1. State Space

In order for our model to continuously fuse long-range UHD video quality features, we define the state vector as state n = F n L H , S n 1 , where F n L H represents the spatial distortion feature extracted by SR-SDFNet for the n-th frame. S n 1 means the historical quality score from the first frame to the n-1th frame, which ranges from 0 to 1, and the larger the value, the higher the video quality. The initial preset historical quality score S 0 is 1. The model integrates the spatial distortion features of the nth frame with the historical quality of the previous n-1 frames by observing F n L H and S n 1 .

3.3.2. Action Space

For the n-th video frame I n H , RL-TFNet outputs action a n by observing s t a t e n to adaptively adjust the historical quality score S n 1 , where action a n is a continuous value taken from [ 0.15 , 0.15]. When the quality of the nth frame is improved compared to the previous n-1 frames, and a n is a positive value, the quality score S n becomes higher than S n 1 . Conversely, when the quality of the nth frame is lower than the previous n − 1 frames, the output a n is between [−0.15, 0], the quality score S n becomes lower. Our model adjusts the historical quality score S n 1 through action a n 1 , so that the quality score S n is closer to the subjective perception of the previous n frame.

3.3.3. Reward Function

The reward is an essential metric for the agent to learn which action performs better in the environment. Considering the frame I n H and the current quality score S n 1 , the reward should encourage the model to increase or decrease the video quality score. The reward function in our method is shown in the formula.
r n = a n Δ S n v m a f
Δ S n v m a f = S n v m a f i = 1 n 1 S i v m a f n 1
S n v m a f represents the quality score of the nth frame calculated by the full reference model VMAF-4K. The higher the score, the higher the quality. i = 1 n 1 S i v m a f n 1 represents the average VMAF-4K score of the 1st to n−1th video frames, and  Δ S n v m a f represents the quality fluctuation of in the nth frame. When a n is closer to Δ S n v m a f , r n will be larger, otherwise, r n will be smaller.

3.3.4. Network Introduction

After defining the above components of MDP, we design and update RL-TFNet via a deep deterministic policy gradient (DDPG) [33]. RL-TFNet consists of an actor network and a critic network, as shown in Figure 7. The input of the actor network is the state vector s t a t e n = F n L H , S n 1 , and the corresponding output is the adjusted value Δ S n to the quality score. The critic network criticizes the action Δ S n based on the state s t a t e n and outputs the state value Q. Both networks have a copy network for calculating the target value—the structure is the same as the main network, but the parameter update speed is different. In each training iteration, both the actor and critic networks are optimized sequentially, as shown in Algorithm 1. In the inference stage, only the actor network is used to predict actions to adjust the quality score.
Algorithm 1 RL-TFNet Algorithm
Randomly set the initialize weights of the actor network π S t a t e θ π and critic network Q S t a t e , A θ Q .
Initialize copy network weights θ Q θ Q , θ π θ π .
Initial quality score S 0 = 1
1:
for t r a i n s t e p n [ 1 , N ] do
2:
 Set s t a t e n = F n L H , S n ;
3:
 Select action a n = π s t a t e n θ π + N o i s e ;
4:
 Execute action on quality score S n + 1 = S n + a n and get reward r n ;
5:
 Set s t a t e n + 1 = F n + 1 L H , S n + 1 ;
6:
 Store transition s t a t e n , a n , r n , s t a t e n + 1 in replay buffer;
7:
 Set y n = γ Q s t a t e n + 1 , π s t a t e n + 1 θ π θ Q
+ r n
8:
 Minimize the loss to optimize critic network:
L c r i t i c = 1 m n y n Q s t a t e n , a n θ Q 2
9:
 Optimize actor network by the cost function:
J θ π = 1 m n Q s t a t e n , a n θ Q
10:
 Update the copy network weights:
θ Q τ θ Q + ( 1 τ ) θ Q
θ π τ θ π + ( 1 τ ) θ π
11:
end for

3.4. Training Procedure

In our method, we first extract global spatial distortion features F n L H from the n-th UHD distorted frame I n H through SR-SDFNet, and the extracted features F n L H and historical quality score S n 1 are used as the input of RL-TFNet to output Δ S n which adjusts the historical quality score S n 1 and get the current quality score S n . The above process continues from the first frame to the last frame to get the entire video quality score, as shown in the formula.
F n L H = N SR - SDFNet downsampling I n H
Δ S n = N R L V Q A N e t F n L H , S n 1
S n = S n 1 + Δ S n
To train the two networks (SR-SDFNet and RL-TFNet) stably, we first train the super-resolution model ESRGAN, the distorted frames are downsampled by a factor of four as the input of ESRGAN, and the corresponding reference frames are used as the ground truth of the super-resolution model. After the ESRGAN is trained to convergence, we use the feature network N f e a t in the ESRGAN generator network SRResNet as the pre-training parameters of SR-SDFNet, and then we jointly train the two networks (SR-SDFNet and RL-TFNet).

4. Experimental Results

4.1. Databases and Evaluation Criteria

To train the proposed model, we collect 50 UHD (3840 × 2160 pixels and 50 fps) video sequences from the source materials of the UHD TV programs. These source materials are only slightly compressed, so the video quality is high enough to be used as reference videos. The collected video sequences cover a wide range of content, including outdoor, indoor, night scenery, buildings, people, sports games and so on. The duration of each sequence is 10 s to 15 s. We manually add three types of compression distortion (AVC, HEVC and VP9) to each of these UHD sequences;each type has four levels, which are set according to the MCML [34] database. Finally, our training database contains 50 reference videos and 600 distorted videos.
In order to evaluate the performance of our method, two UHD VQA databases are used for testing: the AVT-VQDB-UHD-1 [35] and the MCML [34] database. The AVT-VQDB-UHD-1 database includes 13 open-source reference videos and 432 distorted videos, which correspond to four different subjective tests. TEST_1 contains three different codecs (H.264, H.265 and VP9) and four resolutions from 360 p to 2160 p, TEST_2 considers two codecs (H.264 and H.265) and four different bits-per-pixel (bpp) values. TEST_3 uses the same bpp values as TEST_2 and encodes videos with H.265 and VP9 codecs. Test videos for TEST_4 include four frame rates from 15 fps to 60 fps and six different resolutions, which are encoded with H.264. Each distorted video has a mean opinion score (MOS), which ranges from 1 to 5. The MCML database consists of 10 reference videos and 240 distorted videos, which have two resolutions (1920 × 1080 and 3840 × 2160) and three compression distortions (AVC, HEVC and VP9) with four different levels.
We chose Spearman’s rank order correlation coefficient (SROCC), the linear correlation coefficient (PLCC) and root-mean-square error (RMSE) to measure the VQA model performance. SROCC describes the rank consistency of the predicted score and the subjective score. Where N is the total number of distorted videos, and d n represents the ranking difference between the subjective score and the predicted score of the n-th video.
SROCC = 1 6 n d n 2 N N 2 1
PLCC describes the linear correlation between the predicted score and the subjective score, where s i is the predicted quality score and p i is the video subjective score. s ¯ and p ¯ denote the average of the predicted scores and the average of the subjective scores, respectively.
PLCC = i = 1 N s i s ¯ p i p ¯ i = 1 N s i s ¯ 2 i = 1 N p i p ¯ 2
RMSE is mainly used to measure the relative error, where s p is the predicted quality score and s m is the video subjective score.
RMSE = S p S m 2 n 1
Following Zhang et al. [36] and Ma et al. [37], we estimate the parameters of a nonlinear function that transforms the prediction values to the same scales as the subjective score.
q ˜ = ( η 1 η 2 ) / ( 1 + e x p ( ( q ^ η 3 ) / η 4 ) ) + η 2

4.2. Implementation Details

In the training session, the SR model ESRGAN is first trained on the training database according to [30]. The distorted frames in our training database are downsampled by a factor of four as the input of ESRGAN. Then the SR-SDFNET and RL-TFNet are jointly trained in the training database. The video frames used for training are obtained by sampling one frame every 30 frames of each video in the training database. The Adam algorithm is used to optimize the model. The learning rate is set to 1 × 10−4 and decays by a factor of 0.5 every 10,000 iterations. We choose the model with the largest reward value as the best model. In the testing session, each test video is frame sampled every 30 frames intervals for score prediction.

4.3. Comparisons with the State of the Art

In this section, we evaluate the performances of our model and various other BVQA models on two UHD-VQA databases. The main models being compared have all been designed for video quality assessment in recent years, and include VSFA [38], TLVQM [39], VIDEVAL [40], GSTVQA [18], RAPIQUE [10], ChipQA [17], NOFU [6], HEKE [9] and HFR-BVQA [11]. Among them, VIDEVAL, ChipQA, TLVQM, RAPIQUE, and VSFA are five VQA models for UGC video. To ensure fairness of comparison, we train the nine BVQA models on the KONVID [41] database and test them on the MCML and AVT-VQDB-UHD-1 databases. Such cross-dataset experiments can also give the generalization performance of these models. The overall testing results are shown in Table 1 and Table 2. The best results are highlighted in bold.
It can be observed from Table 1 and Table 2 that our model obtains superior performance compared to other BVQA algorithms on both the two databases. Compared with previous methods, our model can extract the distortion features of large-resolution videos more completely. Moreover, the above results illustrate the strong generalizability of our model on UHD distortion databases. This is because the previous methods are supervised by MOS, whereas our model is driven by the super-resolution process without human subjective scores and can adapt to different UHD video quality databases.
To our surprise, the proposed model is weaker than HFR-BVQA for TEST_4 on the AVT-VQDB-UHD-1 database. One reason for this is that the TEST_4 of AVT mainly contains distorted videos with different frame rates. Compared with our method, HFR-BVQA is specially designed for low-frame-rate videos. In response to this phenomenon, we will make special optimizations for distorted videos with different frame rates in the future.

4.4. Performance on Computation Complexity and Runtime

To evaluate the time and spatial complexity of the proposed model, our model is compared with the other nine BVQA models: VSFA, TLVQM, VIDEVAL, GSTVQA, RAPIQUE, ChipQA, NOFU, HEKE and HFR-BVQA. These models are implemented on CPU or GPU, as requested by the authors. The hardware platform used for testing includes an NVIDIA TITAN X GPU processor and Core i7-5930K CPU @ 3.5 GHz. We summarize the runtime and network parameter size in Table 3. The runtime represents the average time to predict a UHD video with a resolution of 3840 × 2160 and 300 frames from the MCML database. In addition, Figure 8 is a scatterplot of speed and performance comparison for our model and the other nine BVQA models in MCML to visually compare.
As shown in Table 3, our model is faster than most BVQA models and the parameter size is smaller than those of the above methods. It should be noted that the previous methods take more time to extract the spatial features of large-resolution video frames. In contrast, our method downsamples the video frames from 3840 × 2160 to 960 × 540 for processing, which greatly improves the running speed. In addition, some models extract spatial features based on basic CNN models, such as GSTVQA using the VGG16 model, and RAPIQUE and HFR-BVQA using the Resnet-50 model. In contrast, our method employs a lightweight super-resolution model for feature extraction, which can predict video quality with smaller model parameters. Furthermore, although the spatial and time complexity of VSFA is slightly lower than that of our model, our model has a comprehensive advantage in both performance and complexity, as shown in Figure 8.

4.5. Visual Analysis

In order to visually verify the effectiveness of our model, we output the feature map from the spatial distortion feature network (SR-SDFNet) and show the fluctuation of video quality over time through a line graph, as shown in Figure 9, where the red dot is the quality scores calculated by the FR-VQA method VMAF-4K and the blue dots are the quality scores predicted by the proposed method. Here, we use the VMAF-4K method to reflect the real quality distribution of the distorted video. As shown in Figure 9a, the trends in the predicted quality scores in our model are similar to those of VMAF-4K, which is gradually degraded. The feature map shows that, along with the degradation of the quality scores, the distortion features are increasing. The same phenomenon can also be observed in Figure 9b. We find that the predicted scores of the proposed model are consistent with VMAF-4K. As time changes, the video quality is first degraded and then improved. Similarly, the spatial distortion features in the feature map are also increased and then reduced. This suggests that our model can perceive the distortion features and accurately predict the video quality.

4.6. Ablation Study

In this section, we investigate the functionalities of SR-SDFNet and RL-TFNet. All experiments are evaluated on the AVT-VQDB-UHD-1 and MCML databases.

4.6.1. SR-SDFNet

To verify the effectiveness of the proposed SR-SDFNet, we evaluate the performance of different spatial feature extractors. The analysis includes five kinds of networks, Resnet-50 [42], VGG16 [41], Mobilenetv2 [43], RCAN [44] and SR-SDFNet (proposed). RCAN is a super-resolution network, which is implemented by the residual-in-residual (RIR) structure and the channel attention mechanism. Table 4 shows the PLCC, SROCC and network parameter size on two databases. We can observe from the results that SR-SDFNet is superior to the other four spatial feature extractors. This is because the three extractors (Resnet-50, VGG16 and Mobilenetv2) are designed for other computer vision tasks, such as image classification. In contrast, SR-SDFNet and RCAN are built based on a video quality enhancement model, which is more suitable for the VQA task. Moreover, the RCAN network only takes a single LI loss as a loss function, and SR-SDFNet introduces perceptual loss and GAN network loss through the pre-training of ESRGAN, so that SR-SDFNet can better extract spatial distortion features with the smallest parameters than other networks.
Furthermore, Figure 10 shows some visual examples of feature maps from SR-SDFNet and Resnet-50. The results in Figure 10c are produced by the SR-SDFNet model and the results in Figure 10b are produced by the Resnet-50 model. We can intuitively find that the Resnet-50 model mainly extracts the local edge of the object, and SR-SDFNet pays more attention to the distortion area of the frames. It can be concluded that the proposed SR-SDFNet is suitable for the UHD-VQA task.

4.6.2. RL-TFNet

We use two time-process strategies to explore the effectiveness of the proposed RL-TFNet. In the first strategy, we first extract spatial features through SR-SDFNet and then regress the features of each frame to a quality score through two fully connected layers. In testing, we average the prediction scores for multiple frames as the score of the entire video. In the second strategy, the spatial features from multiple frames are fed into an LSTM network [45] to predict quality scores. The comparison results are shown in Table 5 and demonstrate that the RL-TFNet performs better than the other two strategies. This occurs because the average pooling in the first strategy only deals with spatial features and the LSTM model in the second strategy has difficulty fusing long-range temporal features.

5. Conclusions

In this paper, we present a fast and accurate blind quality assessment algorithm for UHD videos. It consists of two components: a spatial distortion feature network based on a super-resolution model (SR-SDFNet) and a time fusion network based on reinforcement learning (RL-TFNet) in which SR-SDFNet can employ a super-resolution model to efficiently extract the spatial distortion features and RL-TFNet is used to adjust the quality score in the continuous time dimension. Extensive experimental results demonstrate that the proposed model outperforms all compared competing BVQA methods on UHD quality databases. Furthermore, our model has a smaller parameter size and runs faster than other methods, which makes it more suitable for practical applications in UHD VQA.

Author Contributions

Conceptualization, D.P.; methodology, Z.Y.; software, Z.Y.; validation, Z.Y.; formal analysis, D.P.; investigation, Z.Y.; resources, P.S.; data curation, Z.Y.; writing—original draft preparation, Z.Y.; writing—review and editing, D.P.; visualization, Z.Y.; supervision, P.S.; project administration, P.S.; funding acquisition, P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key R&D Program of China (2019YFB1405902).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used were obtained from public, open-source datasets: 1. AVT-VQDB-UHD-1: https://github.com/Telecommunication-Telemedia-Assessment/AVT-VQDB-UHD-1 (accessed on 30 October 2019). 2. MCML: https://mcml.yonsei.ac.kr/downloads/4kuhdvideoquality (accessed on 10 June 2020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Madhusudana, P.C.; Birkbeck, N.; Wang, Y.; Adsumilli, B.; Bovik, A.C. St-greed: Space-time generalized entropic differences for frame rate dependent video quality prediction. IEEE Trans. Image Process. 2021, 30, 7446–7457. [Google Scholar] [CrossRef] [PubMed]
  2. Wu, J.; Liu, Y.; Dong, W.; Shi, G.; Lin, W. Quality assessment for video with degradation along salient trajectories. IEEE Trans. Multimed. 2019, 21, 2738–2749. [Google Scholar] [CrossRef]
  3. Appina, B.; Dendi, S.V.R.; Bovik, A.C. Study of subjective quality and objective blind quality prediction of stereoscopic videos. IEEE Trans. Image Process. 2019, 28, 5027–5040. [Google Scholar] [CrossRef] [PubMed]
  4. Korhonen, J.; Su, Y.; You, J. Blind natural video quality prediction via statistical temporal features and deep spatial features. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020. [Google Scholar]
  5. Raake, A.; Borer, S.; Satti, S.M.; Gustafsson, J.; Rao, R.R.R.; Medagli, S.; List, P.; Göring, S.; Bitto, R. Multi-model standard for bitstream-, pixel-based and hybrid video quality assessment of uhd/4k: Itu-t p.1204. IEEE Access 2020, 8, 193020–193049. [Google Scholar] [CrossRef]
  6. Göring, S.; Rao, R.R.R.; Feiten, B.; Raake, A. Modular framework and instances of pixel-based video quality models for uhd-1/4k. IEEE Access 2021, 9, 31842–31864. [Google Scholar] [CrossRef]
  7. Li, B.; Zhang, W.; Tian, M.; Zhai, G.; Wang, X. Blindly assess quality of in-the-wild videos via quality-aware pre-training and motion perception. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5944–5958. [Google Scholar] [CrossRef]
  8. Mahankali, N.S.; Raghavan, M.; Channappayya, S.S. No-reference video quality assessment using voxel-wise fmri models of the visual cortex. IEEE Signal Process. Lett. 2022, 29, 319–323. [Google Scholar] [CrossRef]
  9. Liu, Y.; Wu, J.; Li, L.; Dong, W.; Zhang, J.; Shi, G. Spatiotemporal representation learning for blind video quality assessment. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 3500–3513. [Google Scholar] [CrossRef]
  10. Tu, Z.; Yu, X.; Wang, Y.; Birkbeck, N.; Adsumilli, B.; Bovik, A.C. Rapique: Rapid and accurate video quality prediction of user generated content. IEEE Open J. Signal Process. 2021, 2, 425–440. [Google Scholar] [CrossRef]
  11. Zheng, Q.; Tu, Z.; Fan, Y.; Zeng, X.; Bovik, A.C. No-reference quality assessment of variable frame-rate videos using temporal bandpass statistics. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 1795–1799. [Google Scholar]
  12. Zhang, Y.; Gao, X.; He, L.; Lu, W.; He, R. Blind video quality assessment with weakly supervised learning and resampling strategy. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 2244–2255. [Google Scholar] [CrossRef]
  13. Wang, C.; Su, L.; Huang, Q. Cnn-mr for no reference video quality assessment. In Proceedings of the 2017 4th International Conference on Information Science and Control Engineering (ICISCE), Changsha, China, 21–23 July 2017; pp. 224–228. [Google Scholar]
  14. Wang, C.; Su, L.; Zhang, W. Come for no-reference video quality assessment. In Proceedings of the 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Miami, FL, USA, 10–12 April 2018; pp. 232–237. [Google Scholar]
  15. Chen, P.; Li, L.; Ma, L.; Wu, J.; Shi, G. Rirnet: Recurrent-in-recurrent network for video quality assessment. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 834–842. [Google Scholar] [CrossRef]
  16. Dendi, S.V.R.; Channappayya, S.S. No-reference video quality assessment using natural spatiotemporal scene statistics. IEEE Trans. Image Process. 2020, 29, 5612–5624. [Google Scholar] [CrossRef]
  17. Ebenezer, J.P.; Shang, Z.; Wu, Y.; Wei, H.; Sethuraman, S.; Bovik, A.C. Chipqa: No-reference video quality prediction via space-time chips. IEEE Trans. Image Process. 2021, 30, 8059–8074. [Google Scholar] [CrossRef]
  18. Chen, B.; Zhu, L.; Li, G.; Lu, F.; Fan, H.; Wang, S. Learning generalized spatial-temporal deep feature representation for no-reference video quality assessment. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 1903–1916. [Google Scholar] [CrossRef]
  19. Xu, J.; Li, J.; Zhou, X.; Zhou, W.; Wang, B.; Chen, Z. Perceptual quality assessment of internet videos. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 1248–1257. [Google Scholar]
  20. Varga, D. No-reference video quality assessment using the temporal statistics of global and local image features. Sensors 2022, 22, 9696. [Google Scholar] [CrossRef]
  21. Li, J.; Li, X. Study on no-reference video quality assessment method incorporating dual deep learning networks. Multim. Tools Appl. 2022, 82, 3081–3100. [Google Scholar] [CrossRef]
  22. Lin, L.; Zheng, Y.; Chen, W.; Lan, C.; Zhao, T. Saliency-aware spatio-temporal artifact detection for compressed video quality assessment. arXiv 2023, arXiv:2301.01069. [Google Scholar]
  23. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M.A. Playing atari with deep reinforcement learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  24. Wang, Z.; Zhang, J.; Lin, M.; Wang, J.; Luo, P.; Ren, J. Learning a reinforced agent for flexible exposure bracketing selection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1817–1825. [Google Scholar]
  25. Sun, M.; Xiao, J.; Lim, E.G. Iterative shrinking for referring expression grounding using deep reinforcement learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14060–14069. [Google Scholar]
  26. Nauata, N.; Hosseini, S.; Chang, K.-H.; Chu, H.; Cheng, C.-Y.; Furukawa, Y. House-gan++: Generative adversarial layout refinement network towards intelligent computational agent for professional architects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13627–13636. [Google Scholar]
  27. Wang, Y.; Dong, M.; Shen, J.; Wu, Y.; Cheng, S.; Pantic, M. Dynamic face video segmentation via reinforcement learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6957–6967. [Google Scholar]
  28. Lu, Y.; Fu, J.; Li, X.; Zhou, W.; Liu, S.; Zhang, X.; Jia, C.; Liu, Y.; Chen, Z. Rtn: Reinforced transformer network for coronary ct angiography vessel-level image quality assessment. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2022: 25th International Conference, Singapore, 18–22 September 2022, Proceedings, Part I; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  29. Saeed, S.U.; Fu, Y.; Stavrinides, V.; Baum, Z.M.C.; Yang, Q.; Rusu, M.; Fan, R.E.; Sonn, G.A.; Noble, J.A.; Barratt, D.C.; et al. Image quality assessment for machine learning tasks using meta-reinforcement learning. Med. Image Anal. 2022, 78, 102427. [Google Scholar] [CrossRef]
  30. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, BC, Canada, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  31. Netix. Netix Vmaf. 2021. Available online: https://github.com/Netflix/vmaf (accessed on 12 January 2021).
  32. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef]
  33. Lillicrap, T.; Hunt, J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  34. Cheon, M.; Lee, J.-S. Subjective and objective quality assessment of compressed 4k uhd videos for immersive experience. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 1467–1480. [Google Scholar] [CrossRef]
  35. Rao, R.R.R.; Göring, S.; Robitza, W.; Feiten, B.; Raake, A. Avt-vqdb-uhd-1: A large scale video quality database for uhd-1. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 17–177. [Google Scholar]
  36. Zhang, L.; Zhang, L.; Bovik, A.C. A feature-enriched completely blind image quality evaluator. IEEE Trans. Image Process. 2015, 24, 2579–2591. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Ma, K.; Liu, W.; Liu, T.; Wang, Z.; Tao, D. dipiq: Blind image quality assessment by learning-to-rank discriminable image pairs. IEEE Trans. Image Process. 2017, 26, 3951–3964. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Li, D.; Jiang, T.; Jiang, M. Quality assessment of in-the-wild videos. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2351–2359. [Google Scholar]
  39. Korhonen, J. Two-level approach for no-reference consumer video quality assessment. IEEE Trans. Image Process. 2019, 28, 5923–5938. [Google Scholar] [CrossRef]
  40. Tu, Z.; Wang, Y.; Birkbeck, N.; Adsumilli, B.; Bovik, A.C. Ugc-vqa: Benchmarking blind video quality assessment for user generated content. IEEE Trans. Image Process. 2021, 30, 4449–4464. [Google Scholar] [CrossRef]
  41. Hosu, V.; Hahn, F.; Jenadeleh, M.; Lin, H.; Men, H.; Szirányi, T.; Li, S.; Saupe, D. The konstanz natural video database (konvid-1k). In Proceedings of the 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX), Erfurt, Germany, 31 May–2 June 2017. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. Howard, A.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  44. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y.R. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  45. Varga, D.; Szirányi, T. No-reference video quality assessment via pretrained cnn and lstm networks. Signal Image Video Process. 2019, 13, 11. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The relationship between video super-resolution and quality assessment. The red box shows the super-resolution process and the green box represents the structure of our method.
Figure 1. The relationship between video super-resolution and quality assessment. The red box shows the super-resolution process and the green box represents the structure of our method.
Sensors 23 01511 g001
Figure 2. (a) Video quality score prediction process of previous NR−VQA methods. (b) Video quality score prediction process of our model.
Figure 2. (a) Video quality score prediction process of previous NR−VQA methods. (b) Video quality score prediction process of our model.
Sensors 23 01511 g002
Figure 3. Processing pipeline of the proposed model BVQA−SR&DRL. The SR−SDFNet model is used to extract spatial distortion features from the UHD video frame. Then, the RL-TFNet model observes the spatial features and the historical quality score S n 1 , which outputs Δ S n to adjust the quality score. The prediction process of the entire video score is shown in the line graph.
Figure 3. Processing pipeline of the proposed model BVQA−SR&DRL. The SR−SDFNet model is used to extract spatial distortion features from the UHD video frame. Then, the RL-TFNet model observes the spatial features and the historical quality score S n 1 , which outputs Δ S n to adjust the quality score. The prediction process of the entire video score is shown in the line graph.
Sensors 23 01511 g003
Figure 4. The network structure of the generator network SRResNet.
Figure 4. The network structure of the generator network SRResNet.
Sensors 23 01511 g004
Figure 5. The network structure of the spatial distortion feature network (SR-SDFNet).
Figure 5. The network structure of the spatial distortion feature network (SR-SDFNet).
Sensors 23 01511 g005
Figure 6. The processing pipeline of the time fusion network based on reinforcement learning (RL-TFNet).
Figure 6. The processing pipeline of the time fusion network based on reinforcement learning (RL-TFNet).
Sensors 23 01511 g006
Figure 7. (a) The network structure of the actor network. (b) The network structure of the critic network.
Figure 7. (a) The network structure of the actor network. (b) The network structure of the critic network.
Sensors 23 01511 g007
Figure 8. The speed and performance comparison for our model and the other nine BVQA models on the MCML database.
Figure 8. The speed and performance comparison for our model and the other nine BVQA models on the MCML database.
Sensors 23 01511 g008
Figure 9. Two visualization examples about the prediction quality score process of our model. (a) and (b) respectively represent two distorted videos from the AVT-VQDB-UHD-1 database and curves of their predicted quality scores.
Figure 9. Two visualization examples about the prediction quality score process of our model. (a) and (b) respectively represent two distorted videos from the AVT-VQDB-UHD-1 database and curves of their predicted quality scores.
Sensors 23 01511 g009
Figure 10. Visual comparison of spatial feature extraction. (a) shows the distorted UHD frames. (b) presents the feature maps produced by the Resnet-50 model. (c) presents the feature maps produced by the proposed SR-SDFNet model.
Figure 10. Visual comparison of spatial feature extraction. (a) shows the distorted UHD frames. (b) presents the feature maps produced by the Resnet-50 model. (c) presents the feature maps produced by the proposed SR-SDFNet model.
Sensors 23 01511 g010
Table 1. Performance comparison on the MCML databases.
Table 1. Performance comparison on the MCML databases.
MethodALLAVCHEVCVP9
SRCCPLCCRMSESRCCPLCCRMSESRCCPLCCRMSESRCCPLCCRMSE
VSFA0.6330.7081.7510.7570.8341.6130.6140.6551.8920.2950.5751.399
TLVQM0.5050.5772.0500.6070.6291.9360.4870.5262.1560.4620.4491.790
VIDEVAL0.7530.6861.8080.7510.8281.6250.8110.8341.3840.6980.7041.244
GSTVQA0.7160.8301.3820.7400.8551.5130.7740.8501.3190.5880.6821.251
RAPIQUE0.7740.7581.6170.7810.7631.8870.8180.8281.4020.7260.7791.074
ChipQA0.7480.7531.6360.7550.7381.8210.7920.7911.5650.7020.7531.068
Nofu0.8080.8031.5440.8320.8031.6740.8380.8241.5020.7260.7541.068
HEKE0.8030.8341.4870.8160.8111.6570.8100.8451.4850.7220.7551.065
HFR-BVQA0.7610.7751.6000.8080.7801.8280.8360.8471.4630.7120.6961.456
Proposed0.8360.8781.1890.8740.8411.3870.8690.8961.1100.7330.7891.050
The best results are highlighted in bold.
Table 2. Performance comparison on the AVT-VQDB-UHD-1 databases.
Table 2. Performance comparison on the AVT-VQDB-UHD-1 databases.
MethodALLTEST1TEST2TEST3TEST4
SRCCPLCCRMSESRCCPLCCRMSESRCCPLCCRMSESRCCPLCCRMSESRCCPLCCRMSE
VSFA0.6070.6010.8260.7150.6920.7680.7540.7890.6410.6960.7290.7380.5360.5960.737
TLVQM0.4880.5090.9770.5010.5010.9920.5100.5240.9630.5160.5370.9870.4690.4770.921
VIDEVAL0.7150.7240.6860.7040.7220.7380.7390.7560.6680.7470.7570.7110.6680.6790.641
GSTVQA0.6790.7010.7370.5360.6550.8030.7770.8350.5730.7800.8140.6260.6880.7310.626
RAPIQUE0.6740.7290.7080.5000.6960.7630.8370.8660.5210.7670.7970.6510.6800.7570.599
ChipQA0.7250.7450.6670.7140.7610.7160.7660.7180.7030.7400.7780.6910.6520.6890.637
Nofu0.7310.7010.7740.7980.7450.7090.7950.7460.7410.7480.6820.8230.6320.6000.803
HEKE0.7620.7750.6410.7800.8010.6800.8050.8190.6170.7600.7450.7030.6840.7070.638
HFR-BVQA0.7110.7500.6710.7500.7560.7210.7170.7640.6610.7090.7200.7570.7230.7330.615
Proposed0.8160.8030.6160.8490.8590.6520.8780.8660.5210.8070.8470.5720.7170.7290.628
The best results are highlighted in bold.
Table 3. Performance on computation complexity.
Table 3. Performance on computation complexity.
MethodImplementRuntimes (s)Parameters (Millions)
TLVQMMatlab-CPU121.20-
VIDEVALMatlab-CPU282.46-
RAPIQUEMatlab-CPU46.5496.4
HFR-BVQAMatlab-CPU58.1297.1
ChipQAPython-CPU103.40-
NofuPython-CPU117.52-
VSFAPython-GPU9.74100
GSTVQAPython-GPU30.88580.4
HEKEPython-GPU39.90377.6
ProposedPython-GPU12.086.2
Table 4. Performance comparison of different spatial feature extractors on two databases.
Table 4. Performance comparison of different spatial feature extractors on two databases.
MethodMCMLAVTParameters
SRCCPLCCSRCCPLCC
VGG160.7950.8180.7880.772528 M
Mobilenetv20.7620.7860.7230.7385.2 M
Resnet-500.8130.8540.7930.77497.8 M
RCAN0.8290.8700.8110.78959.7 M
SR-SDFNet0.8360.8780.8160.8034.9 M
Table 5. Performance comparison of different time-process strategies on two databases.
Table 5. Performance comparison of different time-process strategies on two databases.
Time Process StrategiesMCMLAVT
SRCCPLCCSRCCPLCC
Average Pooling0.7920.7950.7670.773
LSTM0.8020.8200.7810.772
RL-TFNet0.8360.8780.8160.803
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ying, Z.; Pan, D.; Shi, P. Blind Video Quality Assessment for Ultra-High-Definition Video Based on Super-Resolution and Deep Reinforcement Learning. Sensors 2023, 23, 1511. https://doi.org/10.3390/s23031511

AMA Style

Ying Z, Pan D, Shi P. Blind Video Quality Assessment for Ultra-High-Definition Video Based on Super-Resolution and Deep Reinforcement Learning. Sensors. 2023; 23(3):1511. https://doi.org/10.3390/s23031511

Chicago/Turabian Style

Ying, Zefeng, Da Pan, and Ping Shi. 2023. "Blind Video Quality Assessment for Ultra-High-Definition Video Based on Super-Resolution and Deep Reinforcement Learning" Sensors 23, no. 3: 1511. https://doi.org/10.3390/s23031511

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop