Next Article in Journal
Research on the Necessity of Lie Group Strapdown Inertial Integrated Navigation Error Model Based on Euler Angle
Previous Article in Journal
Modeling of Respiratory Motion to Support the Minimally Invasive Destruction of Liver Tumors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Temporal Prediction Model-Based Fast Inter CU Partition for Versatile Video Coding

1
College of Computer Science, University of South China, Hengyang 421001, China
2
Faculty of Physics and Electronic Information Science, Hengyang Normal University, Hengyang 421002, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(20), 7741; https://doi.org/10.3390/s22207741
Submission received: 18 August 2022 / Revised: 21 September 2022 / Accepted: 9 October 2022 / Published: 12 October 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Versatile video coding (VVC) adopts an advanced quad-tree plus multi-type tree (QTMT) coding structure to obtain higher compression efficiency, but it comes at the cost of a considerable increase in coding complexity. To effectively reduce the coding complexity of the QTMT-based coding unit (CU) partition, we propose a fast inter CU partition method based on a temporal prediction model, which includes early termination QTMT partition and early skipping multi-type tree (MT) partition. Firstly, according to the position of the current CU, we extract the optimal CU partition information of the position corresponding to the previously coded frames. We then establish a temporal prediction model based on temporal CU partition information to predict the current CU partition. Finally, to reduce the cumulative of errors of the temporal prediction model, we further extract the motion vector difference (MVD) of the CU to determine whether the QTMT partition can be terminated early. The experimental results show that the proposed method can reduce the inter coding complexity of VVC by 23.19% on average, while the Bjontegaard delta bit rate (BDBR) is only increased by 0.97% on average under the Random Access (RA) configuration.

1. Introduction

With the rapid development of display technology, new video formats such as Ultra-High Definition (UHD), High Dynamic Range (HDR), and 360 videos have provided improved visual experiences. However, their considerable data volume is a challenge for video storage and real-time transmission. To compress different emerging video formats more efficiently, the Joint Video Experts Team (JVET) formulated a new generation of video compression standards in July 2020, namely, Versatile Video Coding (VVC) [1].
Compared with High-Efficiency Video Coding (HEVC) [2], VVC reduces the coding rate by more than 30% while maintaining the same subjective video quality [3]. This is because many advanced coding techniques are introduced in VVC. To further remove the spatial coding redundancy, VVC proposes new technologies, such as extending the 35 intra prediction modes in HEVC to 67 [4], multiple reference line (MRL) prediction [5], and intra sub-partitioning (ISP) [6]. Advanced techniques such as affine motion compensation prediction, adaptive motion vector resolution, and bidirectional optical flow have been proposed to further remove the temporal coding redundancy of VVC [7]. In addition, to adapt to the different textures and motion characteristics of coding regions, VVC adopts a new coding structure, the quad-tree plus multi-type tree (QTMT) partition structure [8], which can flexibly divide the coding unit (CU) into square and rectangular regions of different sizes. Although these new coding techniques can effectively improve the compression efficiency of intra and inter frames, they also drastically increase the coding complexity of VVC intra and inter frames, which has become a major obstacle to the deployment of VVC real-time applications. In particular, the flexible QTMT partition structure process accounts for considerable coding complexity [9]. Therefore, it is necessary to ensure that fast decisions can be made in the QTMT partition process to greatly reduce the coding complexity of VVC.
The QTMT partition structure is an extension of the quad-tree (QT) partition structure in HEVC. In HEVC, the QT partition also occupies a considerable amount of coding complexity. To effectively reduce the coding complexity of HEVC, researchers have proposed many fast QT partition methods [10,11,12,13,14,15,16]. These methods can be mainly classified into two categories: correlation-based methods and learning-based methods. For example, in [12], a fast QT partition method is proposed by using the spatial and temporal correlation between the current coding CU and the coded CUs. In [15], to accelerate the QT partition decision, the prediction model is trained by the deep learning method to predict the optimal QT partition in advance. However, these fast QT partition methods for HEVC cannot be directly used in the QTMT partition of VVC, because the multi-type tree (MT) partition is added to the QTMT partition structure, which makes the CU partition more flexible. In addition, many new coding techniques are applied to the QTMT partition structure, which results in the optimal CU partition in VVC and HEVC being different in the same coding regions.
To reduce the coding complexity of the intra VVC encoder, many fast CU decision methods [17,18,19,20,21,22,23,24,25,26,27,28,29] have proposed the acceleration of intra QTMT partition. For example, in [19], the authors propose a directional gradient-based early termination CU partition method in which four directional gradients (horizontal, vertical, 45, and 135) of CUs are extracted separately to represent the relationship between the optimal CU partition and the texture characteristics of the CU. In [23], a fast decision method for QTMT partition with a cascaded decision structure composed of decision trees is proposed. In addition, the authors in [25,26,27,28,29] adopted a deep-learning-based approach for fast QTMT partition decisions. However, these fast CU decision methods for intra QTMT partition cannot be directly used for inter QTMT partition, because the optimal intra CU partition of a CTU is mainly related to the texture characteristics of the CU under the intra coding. However, the optimal inter CU partition of a CTU is not only related to the CU’s texture characteristics but also related to the motion properties of the CU. Therefore, it is necessary to propose a fast inter QTMT partition decision method to reduce the inter coding complexity of VVC.
Researchers have proposed many fast CU partition decision methods for reducing the inter coding complexity of VVC [30,31,32,33]. In [30], the difference between three frames is extracted to indicate that the CU is a motion block or a static block. If the CU is a static block, the CU is terminated for further partition early to reduce the complexity of inter coding. In [31], three binary classifiers are established by using the random forest to accelerate the inter CU partition decision. In addition, the tunable risk interval is defined to obtain different coding complexity reductions. In [32], combining the luminance, residual, and bidirectional motion information of the CU as the input information of a convolutional neural network (CNN), a multi-information fusion convolutional neural network (MF-CNN)-based early inter CU partition termination method is proposed. In addition, the Merge mode of the CU is terminated early using the residual information of the CU to further reduce the coding complexity of the inter CU. In [33], a multi-level tree convolutional neural network (MLT-CNN) is proposed for a fast CU partition decision. First, the order of the decision process of CU is split into four levels: no partition–partition, QT partition–MT partition, binary tree (BT) partition–ternary tree (TT) partition, and vertical partition–horizontal partition. The MLT-CNN is then used to predict how the CU is split in each level. However, the temporal correlation of CU partition is not fully explored in VVC inter coding. Although many methods have proposed utilizing temporal correlation for fast QT-based CU partition decisions in HEVC, they cannot be directly used in the new QTMT-based CU partition. Therefore, it is possible to explore fast QTMT partition decisions based on temporal correlations. In this paper, we propose a temporal prediction model-based approach to predict CU partition to reduce the inter coding complexity of VVC. The main contributions are summarized as follows.
  • According to the partition of the encoded CU at the corresponding position of the current CU, we established a temporal prediction model to predict the optimal CU partition.
  • To reduce prediction errors, we further extract the motion vector information of the CU to determine whether the CU partition can be terminated early.
  • Compared with deep learning methods, our proposed method does not require the establishment of additional large datasets for training and does not require additional complex training to obtain decision parameters.
We organize the remainder of this paper as follows. Section 2 presents the motivation and statistical analyses. Section 3 presents the proposed fast inter CU partition method of VVC based on the temporal prediction model. Section 4 presents the experimental results. The conclusion is in Section 5.

2. Motivation and Statistical Analyses

A new QTMT-based CU partition structure is introduced in VVC to further improve the compression efficiency, and its flexible and changeable partition method is more suitable for video content with different textures and motion characteristics. In the VVC coding standard, a coding image is firstly split into multiple Coding Tree Units (CTUs), the default size of which is 128 × 128. Afterwards, the CTU is used as the root node for further hierarchical partition. The blocks obtained by the partition process are called Coding Units (CUs). A CTU can contain one CU or be recursively split into smaller square CUs or rectangular CUs, where a square CU can be obtained by QT partition, while a rectangular CU is obtained by MT partition. MT includes horizontal BT, vertical BT, horizontal TT, and vertical TT. Figure 1 shows an example of optimal CU partition of a CTU, and the six partition methods supported by the CU. In order to obtain the optimal CTU partition, a 128 × 128 CTU will first perform a top-to-bottom multi-level hierarchical partition process until the minimum CU size is 4 × 4, and each CU needs to calculate the rate-distortion cost (RD cost). CU pruning is then performed in a bottom-up manner, and the CUs with the smallest RD cost are combined to obtain the optimal CTU partition. This method of obtaining the optimal CU partition through the brute-force search process will lead to a considerable increase in the coding complexity of VVC. Although a CTU needs to traverse CUs of different sizes, only a few CUs of different sizes are selected in the optimal CTU partition. If the optimal CU partition in the CTU can be predicted early, and the RD cost calculation of the remaining CUs can be skipped, the coding complexity of VVC can be effectively reduced.

2.1. The Distribution of Different CU Sizes

In VVC inter coding, a larger CU size is usually selected as the optimal CU partition for background or slow-moving regions, while a smaller CU size is usually selected as the optimal CU for regions with complex motion regions. To represent the distribution of optimal CU sizes in VVC inter coding, four video sequences with different motion characteristics are tested, including “MarketPlace” (1920 × 1080), “FourPeople” (1280 × 720), “BasketballDrill” (832 × 480), and “BasketballPass” (416 × 240). Test conditions are shown in Table 1. To simplify the statistical distribution of different CU sizes, we divide all optimal CU sizes into five non-overlapping categories, L64, L32, L16, L8, and L4, which correspond to the width and height of the CUs that are all greater than or equal to 64, 32, 16, 8, and 4, respectively. For example, an optimal CU with a width of 64 and a height of 32 will be classified as L32. The optimal CU size distributions for video sequences under different QPs are given in Table 2. From Table 2 we can observe the following points:
(1) For all video sequences, the average proportions of L64, L32, L16, L8, and L4 are 58.33%, 18.26%, 13.35%, 7.45%, and 2.61%, which indicates that the larger-size CUs are selected as optimal CUs for most regions. Under the same QP, as the CU size decreases, its proportion becomes smaller.
(2) With the increase of QP, the proportion of CUs in the L64 becomes larger, and the proportion of CUs in the L8 and L4 becomes smaller. When QP = 37, the maximum proportion of CU under the L64 is 89.62%, and the minimum proportion is 53.72%. The largest proportions of CUs in the L8 and L4 are only 4.89% and 1.81%. When QP = 22, the maximum proportions of CUs under L8 and L4 are 16.84% and 8.26%. This indicates that different QPs have a significant impact on the optimal CU selection for the same coding regions.
(3) For the sequence “FourPeople” containing large background regions, the proportion of CUs in the L64 exceeds 70.72%, and the proportion of CUs in the L4 is less than 2.15%. For the sequence “BasketballDrill” with complex motion regions, and the sequences “MarketPlace” and “BasketballPass” containing dynamic backgrounds and camera movements, the sum of L64 and L32 is larger. This shows that, for sequences with different motion characteristics, the distribution of the optimal CU size is different, and a larger CU size is usually selected as the optimal CU for background or slow-moving regions.
From the above observations, it can be concluded that, for different video sequences, if the CU partition can be accurately predicted in advance, it will greatly reduce the coding complexity of VVC.

2.2. Temporal Correlation of CUs

Since the content between adjacent frames is very similar, there is a high probability that the corresponding regions between adjacent frames have the same optimal CU partition. To explore the correlation of optimal CU partition between adjacent frames in the same regions, Figure 2 presents the optimal CU partition for sequences “FourPeople” and “MarketPlace” under RA configuration. In VVC, the largest CU size of 128 × 128 corresponds to a CU depth of 0. As the CU is split into smaller sub-CUs, its corresponding depth is increased by 1. Figure 2c,f show the absolute difference of CU depth between the 5th and 6th frames. In Figure 2c,f, the four colors white, blue, green, and red represent the absolute difference of zero, one, two, and three, respectively, and black represents the depth difference greater than or equal to four. We can further observe that most background regions are marked with white or blue, which indicates that the CU depth correlation between adjacent frames is very high.
To analyze the CU depth difference ( d ψ ) between adjacent frames, d ψ is defined as follows:
d ψ = CU ψ i CU ψ i + 1 , ψ { QTMT , QT , MT }
where CU ψ i and CU ψ i + 1 represent the CU depth of the i-th frame and the i + 1-th frame at the same position. The test sequences include “FourPeople” and “MarketPlace”, and the test conditions are shown in Table 1. Table 3, Table 4 and Table 5 represent the QTMT depth difference, the QT depth difference, and the MT depth difference of adjacent frames, respectively. It can be seen in Table 3, Table 4 and Table 5 that the average percentages of depth differences of QTMT, QT, and MT that are less than or equal to 1 are 77.5%, 95.1%, and 85.8%, which indicates that the CU depth difference between adjacent frames is very small. Therefore, the CU depth of adjacent coded frames can be used to predict the optimal CU depth of the current coding frame.

3. The Proposed Fast Inter CU Partition Method

In VVC, the optimal CU partition of a CTU is selected by traversing all partition sizes. Therefore, if the most probable depth of the current coding CU can be accurately determined in advance, a large amount of coding complexity can be reduced without affecting the coding quality. This section is mainly divided into two parts. Firstly, a temporal prediction model that predicts the most probable depth of the CU is established. Secondly, a fast CU partition decision method is proposed based on the temporal prediction model.

3.1. The Temporal Prediction Model

According to the analysis in Section 2.2, it can be concluded that adjacent frames have very similar texture content, which leads to a strong correlation between the optimal CU depth (CU partition) at the same positions. Therefore, based on this heuristic, we can utilize the optimal CU depth of previously coded frames to predict the optimal CU depth of the current coding frame. Figure 3 shows an example of a CU in the current coding frame and a corresponding position CU in the coded frames, D 0 and D 1 respectively represent the previously coded frame, and D c u r represents the current coding frame. In this paper, a temporal prediction model is established to predict the current CU depth according to the temporal correlation of the CU, which is defined as follows:
D p = ( D 0 + D 1 + 1 ) / 2
D 0 = arg max { D 0 w × h }
D 1 = arg max { D 1 w × h }
where D 0 and D 1 respectively represent the maximum depth of the CU in the previously encoded frames, and w and h denote the width and height of the CU. Since the optimal depth of the CU has a strong correlation with the QP, the coded frames are the nearest adjacent frames with the same or less QP as the current coding frame.

3.2. Terminating QTMT Partition Early Based on the Temporal Prediction Model

(a)
Terminating the QT partition early:
As shown in Figure 1, the optimal depth of a CU in VVC coding may include both QT depth and MT depth. Therefore, early QT depth and MT depth are both decisions in this paper. According to the temporal prediction model in Section 3.1, the optimal QT depth model for the prediction of the current CU is expressed as follows:
Q T p = ( Q T 0 + Q T 1 + 1 ) / 2
Q T 0 = arg max { QT CU 0 }
Q T 1 = arg max { QT CU 1 }
where QT 0 and QT 1 respectively represent the maximum QT depth of the coded CU in the previous frames. According to Equations (5)–(7), early QT depth decision is defined as follows:
CU QT un _ split Q T cur Q T p + Th QT split     otherwise
where QT cur represents the QT depth of the current CU, QT p denotes the prediction QT depth of the current CU according to the temporal prediction model, and Th QT is an adaptive threshold. If QT cur is less than QT p + Th QT , the current CU needs to perform the QT partition. Otherwise, the QT partition is terminated early.
(b)
Terminating the MT partition early:
Similar to the optimal QT depth model for prediction of the current CU, according to the temporal prediction model established in Section 3.1, the optimal MT depth model for the prediction of the current CU is expressed as follows:
M T p = ( M T 0 + M T 1 + 1 ) / 2
M T 0 = arg max { MT CU 0 }
M T 1 = arg max { MT CU 1 }
where MT 0 and MT 1 respectively represent the maximum depth of the BT depth and the TT depth of the coded CU in the previous frames. According to Equation (9)–(11), the early MT depth decision is defined as follows:
CU MT un _ split M T cur M T p + Th MT split     otherwise
where MT cur represents the MT depth of the current CU, MT p denotes the prediction MT depth of the current CU according to the temporal prediction model, and Th MT is an adaptive threshold. If MT cur is less than MT p + Th MT , the current CU needs to perform the MT partition. Otherwise, the MT partition is terminated early.

3.3. Prediction Error Reduction Based on Encoding Information

In Section 3.1 and Section 3.2, although the prediction depth can be used for early termination of the current CU depth, the process of early termination may lead to the accumulation of errors in depth prediction. This is because, on the one hand, the optimal depth of the current CU is predicted based on the previously coded frames, and when the next frame is coding, the CU depth of the current coding frame is used to predict the optimal CU depth of the next frame. On the other hand, when scene switching occurs between adjacent frames, directly using the CU depth information of the previous frame will not be able to accurately determine the optimal CU depth of the current frame. In order to effectively solve the above problems, the coding information of the CU is introduced, namely, motion vector difference (MVD). In inter coding, the MVD of the CU can well represent the relative motion of the CU. When the MVD of the CU is 0, the CU is in the static region or the region with the same motion direction as the adjacent CUs, and the CUs in these regions usually do not need to be further split. In order to explore the relationship between the MVD of CU and CU partition, we counted the probability that the MVD of CU is 0 when the encoding frame is the optimal CU partition. The experimental setting conditions are the same as in Section 2.1. Table 6 shows the probability that the MVD of a CU is 0. It can be seen in Table 6 that the MVD of more than 91.3% of the CUs on average is 0. Therefore, in this paper, we combine the MVD of the CU for early QTMT depth decision. Equations (8) and (12) for early QTMT depth decisions are modified as follows:
CU QT un _ split ( Q T cur Q T p + Th QT ) & ( MVD 0 ) split     otherwise
CU MT un _ split ( M T cur M T p + Th MT ) & ( MVD 0 ) split     otherwise
MVD = MV D hor + MV D ver
where “&” is the logical and. MVD hor and MVD ver represent the motion vector difference in the horizontal and vertical directions of the CU, respectively.

3.4. Skipping TT Partition Early Based on the Temporal Prediction Model

Since the MT depth may only include BT partition or less TT partition, to further reduce the coding complexity of the MT partition, only BT partition can be performed at a certain MT depth, and the TT partition can be skipped early. According to Equation (9), the early skipping of the TT partition decision is defined as follows:
C U TT skip     M T cur < M T p Th MT un _ skip otherwise
If MT cur is less than MT p - Th MT , the current CU needs to skip the TT partition. Otherwise, the TT partition is performed.

3.5. Proposed Overall Method

The basic idea of the proposed fast inter CU partition method of VVC is to use the optimal CU partition information at the corresponding position of the encoded frames to predict the current CU partition. Figure 4 shows the flowchart of the proposed method, and the specific process is summarized as follows:
(1)
Start inter CU coding and calculate the RD cost of the CU.
(2)
Calculate Q T p and M T p according to the temporal prediction model.
(3)
When the partition mode is QT, if Equation (13) is satisfied, terminate the QT partition early; otherwise, perform QT partition.
(4)
When the partition mode is BT, if Equation (14) is satisfied, terminate the BT partition early; otherwise, perform BT partition.
(5)
When the partition mode is TT, if Equation (14) or Equation (16) is satisfied, terminate/skip the TT partition early; otherwise, perform TT partition.

4. Experimental Results and Analyses

4.1. Test Conditions

To effectively evaluate the encoding performance of the proposed method, the Bjontegaard Delta Bit Rate (BDBR) [34] and coding complexity saving are adopted, where coding time saving is defined as follows:
T S = T o r i T p r o T o r i
where T o r i represents the coding time of the original reference software, and T p r o denotes the coding time after integrating the proposed method into the reference software. The experimental reference software version is VTM-6.0, the sequences used for the experimental test are the common test sequences [35], the encoding configuration is set to Random Access (RA), and four quantization parameters (QPs) (22, 27, 32, and 37) are used in the encoding. Th QT and Th MT are set to 2 and 1, respectively.

4.2. Experimental Results

Table 7 presents the coding performance of the early TT partition skipping method, the early QTMT partition termination method, and the proposed overall method. It can be observed in Table 7 that skipping the TT partition early can reduce the coding time by 15.48% on average and increase the BDBR by 0.61%. Early QTMT partition termination can reduce the coding time by 6.14% on average and increase the BDBR by 0.4%. Compared with the early QTMT partition termination method, the early TT partition skipping method can reduce more coding time. The reason for this is that the early termination conditions of the early QTMT partition termination method are stricter, resulting in fewer QTMT partitions being terminated early. The proposed overall method can reduce 8.88–42.08% of the encoding time, with an average saving of 23.19%, and BDBR increases by 0.29–1.98%, while the average increase is only 0.97%. The largest coding time saving is the sequence “Campfire”, with a 42.08% reduction in coding time and a 1.98% increase in BDBR. The smallest coding time saving is the sequence “Johnny”, with an 8.88% reduction in encoding time and a 0.29% increase in BDBR.
Figure 5 presents the time savings of four sequences under different QPs. It can be seen in Figure 5 that the time saving is smaller as the QP increases, which indicates that the proposed method can save more time under a smaller QP. This is because, as shown in Table 2, when the coding QP is small, a larger depth of the sequence (a smaller CU size) is selected as the optimal depth, and the proposed method can skip the calculation of the smaller depth early. Therefore, the proposed method can save more coding time when the QP is small.
Table 8 presents a coding performance comparison between the proposed method and the current state-of-the-art methods [30,32]. For a fair comparison, the method in [32] only includes the fast CU partition decision method and does not include the early merge mode decision method. Table 8 shows that the method in [30] reduces coding time by 8.29–38.53%, with an average saving of 26.56%, and the BDBR is increased by 0.31–10.10%, with an average increase of 3.29%. Using the method in [32], the coding time is reduced by 9.69–42.90%, with an average saving of 24.83%, and the BDBR is increased by 0.84–5.59%, with an average increase of 2.52%. Compared to these methods [30,32], the proposed method can further reduce BDBR by 2.32% and 1.55%, respectively, on average, with similar coding time savings. In addition, the proposed method has a maximum BDBR increase of 1.98%, which is much less than the 10.10% of the method in [30] and the 5.59% of the method in [32].

5. Conclusions

In this paper, a fast inter CU partition decision method is proposed for VVC based on a temporal prediction model to reduce inter coding complexity. First, a temporal prediction model is established using the temporal correlation of CUs. Afterwards, the early termination and early skipping of CU partition are performed according to the temporal prediction model. Finally, to reduce prediction errors, the MVDs of the CUs are jointly used for fast CU partition decisions. Experimental results show that the coding time saved by the proposed fast CU partition method ranges from 8.88% to 42.08%, and the BDBR increases from 0.29% to 1.98%. The comparison of experimental results shows that the BDBR can be further reduced with similar encoding time savings compared to the state-of-the-art methods [30,32].

Author Contributions

Conceptualization, Y.L. and Y.Z.; methodology, Y.L.; software, F.L.; validation, Y.L., F.L. and Y.Z.; data curation, F.L.; writing—original draft preparation, Y.L. and F.L.; writing—review and editing, Y.Z.; supervision, Y.Z.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the Natural Science Foundation of Hunan Province, China (No. 2020JJ5496) and the National Natural Science Foundation of China (No. 62001209).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. ISO/IEC 23090-3; Information Technology—Coded Representation of Immersive Media—Part 3: Versatile Video Coding. ISO: Geneva, Switzerland, 2021.
  2. Sullivan, G.J.; Ohm, J.R.; Han, W.J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  3. Bross, B.; Wang, Y.K.; Ye, Y.; Liu, S.; Chen, J.; Sullivan, G.J.; Ohm, J.R. Overview of the versatile video coding (VVC) standard and its applications. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3736–3764. [Google Scholar] [CrossRef]
  4. Pfaff, J.; Filippov, A.; Liu, S.; Zhao, X.; Chen, J.; De-Luxán-Hernández, S.; Wiegand, T.; Rufitskiy, V.; Ramasubramonian, A.K.; Van der Auwera, G. Intra prediction and mode coding in VVC. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3834–3847. [Google Scholar] [CrossRef]
  5. Chang, Y.J.; Jhu, H.J.; Jiang, H.Y.; Zhao, L.; Zhao, X.; Li, X.; Liu, S.; Bross, B.; Keydel, P.; Schwarz, H.; et al. Multiple reference line coding for most probable modes in intra prediction. In Proceedings of the 2019 Data Compression Conference (DCC), Snowbird, UT, USA, 26–29 March 2019; p. 559. [Google Scholar]
  6. De-Luxán-Hernández, S.; George, V.; Ma, J.; Nguyen, T.; Schwarz, H.; Marpe, D.; Wiegand, T. An intra subpartition coding mode for VVC. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1203–1207. [Google Scholar]
  7. Yang, H.; Chen, H.; Chen, J.; Esenlik, S.; Sethuraman, S.; Xiu, X.; Alshina, E.; Luo, J. Subblock-Based Motion Derivation and Inter Prediction Refinement in the Versatile Video Coding Standard. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3862–3877. [Google Scholar] [CrossRef]
  8. Huang, Y.W.; An, J.; Huang, H.; Li, X.; Hsiang, S.T.; Zhang, K.; Gao, H.; Ma, J.; Chubach, O. Block partitioning structure in the VVC standard. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3818–3833. [Google Scholar] [CrossRef]
  9. Bossen, F.; Sühring, K.; Wieckowski, A.; Liu, S. VVC complexity and software implementation analysis. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3765–3778. [Google Scholar] [CrossRef]
  10. Shen, L.; Zhang, Z.; Liu, Z. Effective CU size decision for HEVC intracoding. IEEE Trans. Image Process. 2014, 23, 4232–4241. [Google Scholar] [CrossRef]
  11. Min, B.; Cheung, R.C. A fast CU size decision algorithm for the HEVC intra encoder. IEEE Trans. Circuits Syst. Video Technol. 2014, 25, 892–896. [Google Scholar]
  12. Zhong, G.; He, X.; Qing, L.; Li, Y. A fast inter-prediction algorithm for HEVC based on temporal and spatial correlation. Multimed. Tools Appl. 2015, 74, 11023–11043. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Kwong, S.; Wang, X.; Yuan, H.; Pan, Z.; Xu, L. Machine learning-based coding unit depth decisions for flexible complexity allocation in high efficiency video coding. IEEE Trans. Image Process. 2015, 24, 2225–2238. [Google Scholar] [CrossRef]
  14. Zhu, L.; Zhang, Y.; Kwong, S.; Wang, X.; Zhao, T. Fuzzy SVM-based coding unit decision in HEVC. IEEE Trans. Broadcast. 2017, 64, 681–694. [Google Scholar] [CrossRef]
  15. Xu, M.; Li, T.; Wang, Z.; Deng, X.; Yang, R.; Guan, Z. Reducing complexity of HEVC: A deep learning approach. IEEE Trans. Image Process. 2018, 27, 5044–5059. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Xiong, J.; Li, H.; Meng, F.; Zhu, S.; Wu, Q.; Zeng, B. MRF-based fast HEVC inter CU decision with the variance of absolute differences. IEEE Trans. Multimed. 2014, 16, 2141–2153. [Google Scholar] [CrossRef]
  17. He, L.; Xiong, S.; Yang, R.; He, X.; Chen, H. Low-Complexity Multiple Transform Selection Combining Multi-Type Tree Partition Algorithm for Versatile Video Coding. Sensors 2022, 22, 5523. [Google Scholar] [CrossRef]
  18. Park, S.H.; Kang, J.W. Context-based ternary tree decision method in versatile video coding for fast intra coding. IEEE Access 2019, 7, 172597–172605. [Google Scholar] [CrossRef]
  19. Cui, J.; Zhang, T.; Gu, C.; Zhang, X.; Ma, S. Gradient-Based Early Termination of CU Partition in VVC Intra Coding; DCC: Hong Kong, China, 2020; pp. 103–112. [Google Scholar]
  20. Song, Y.; Zeng, B.; Wang, M.; Deng, Z. An efficient low-complexity block partition scheme for VVC intra coding. J. Real-Time Image Process. 2022, 19, 161–172. [Google Scholar] [CrossRef]
  21. Li, Y.; Yang, G.; Song, Y.; Zhang, H.; Ding, X.; Zhang, D. Early intra CU size decision for versatile video coding based on a tunable decision model. IEEE Trans. Broadcast. 2021, 67, 710–720. [Google Scholar] [CrossRef]
  22. Zhao, J.; Wu, A.; Zhang, Q. SVM-Based Fast CU Partition Decision Algorithm for VVC Intra Coding. Electronics 2022, 11, 2147. [Google Scholar] [CrossRef]
  23. Yang, H.; Shen, L.; Dong, X.; Ding, Q.; An, P.; Jiang, G. Low-complexity CTU partition structure decision and fast intra mode decision for versatile video coding. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 1668–1682. [Google Scholar] [CrossRef]
  24. Lei, M.; Luo, F.; Zhang, X.; Wang, S.; Ma, S. Look-ahead prediction based coding unit size pruning for VVC intra coding. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 4120–4124. [Google Scholar]
  25. Jin, Z.; An, P.; Shen, L.; Yang, C. CNN oriented fast QTBT partition algorithm for JVET intra coding. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  26. Zhang, Q.; Guo, R.; Jiang, B.; Su, R. Fast CU Decision-Making Algorithm Based on DenseNet Network for VVC. IEEE Access 2021, 9, 119289–119297. [Google Scholar] [CrossRef]
  27. Li, Y.; Li, L.; Fang, Y.; Peng, H.; Ling, N. Bagged Tree and ResNet-Based Joint End-to-End Fast CTU Partition Decision Algorithm for Video Intra Coding. Electronics 2022, 11, 1264. [Google Scholar] [CrossRef]
  28. Li, T.; Xu, M.; Tang, R.; Chen, Y.; Xing, Q. DeepQTMT: A deep learning approach for fast QTMT-based CU partition of intra-mode VVC. IEEE Trans. Image Process. 2021, 30, 5377–5390. [Google Scholar] [CrossRef]
  29. Zhao, J.; Wang, Y.; Li, M.; Zhang, Q. Fast coding unit size decision based on deep reinforcement learning for versatile video coding. Multimed. Tools Appl. 2022, 2022, 1–17. [Google Scholar] [CrossRef]
  30. Tang, N.; Cao, J.; Liang, F.; Wang, J.; Liu, H.; Wang, X.; Du, X. Fast CTU partition decision algorithm for VVC intra and inter coding. In Proceedings of the 2019 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS), Bangkok, Thailand, 11–14 November 2019; pp. 361–364. [Google Scholar]
  31. Amestoy, T.; Mercat, A.; Hamidouche, W.; Menard, D.; Bergeron, C. Tunable VVC frame partitioning based on lightweight machine learning. IEEE Trans. Image Process. 2019, 29, 1313–1328. [Google Scholar] [CrossRef] [PubMed]
  32. Pan, Z.; Zhang, P.; Peng, B.; Ling, N.; Lei, J. A CNN-Based Fast Inter Coding Method for VVC. IEEE Signal Process. Lett. 2021, 28, 1260–1264. [Google Scholar] [CrossRef]
  33. Yeo, W.H.; Kim, B.G. CNN-based Fast Split Mode Decision Algorithm for Versatile Video Coding (VVC) Inter Prediction. J. Multimed. Inf. Syst. 2021, 8, 147–158. [Google Scholar] [CrossRef]
  34. Bjontegaard, G. Calculation of Average PSNR Differences between RD-Curves. VCEG-M33 2001. Available online: https://cir.nii.ac.jp/crid/1572543025125831168?lang=en (accessed on 1 June 2022).
  35. Boyce, J.; Suehring, K.; Li, X.; Seregin, V. JVET common test conditions and software reference configurations. In Proceedings of the Joint Video Experts Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 10th Meeting, San Diego, CA, USA, 10–20 April 2018. [Google Scholar]
Figure 1. An example of optimal CU partition of a CTU. (a,b) denote the optimal CU partition, (c) presents six types of CU partition.
Figure 1. An example of optimal CU partition of a CTU. (a,b) denote the optimal CU partition, (c) presents six types of CU partition.
Sensors 22 07741 g001
Figure 2. Optimal CU partition of two successive frames of sequences “FourPeople” and “MarketPlace” and the absolute differences of CU depth among frames. (a,b,d,e) denote the original optimal CU partition. (c,f) represent the absolute difference between (a,b), and the absolute difference between (d,e), respectively.
Figure 2. Optimal CU partition of two successive frames of sequences “FourPeople” and “MarketPlace” and the absolute differences of CU depth among frames. (a,b,d,e) denote the original optimal CU partition. (c,f) represent the absolute difference between (a,b), and the absolute difference between (d,e), respectively.
Sensors 22 07741 g002
Figure 3. An example of current coding frame and adjacent coded frames.
Figure 3. An example of current coding frame and adjacent coded frames.
Sensors 22 07741 g003
Figure 4. The flowchart of the proposed method.
Figure 4. The flowchart of the proposed method.
Sensors 22 07741 g004
Figure 5. Encoding time saving under different QPs. (a) BasketballPass, (b) FourPeople, (c) MarketPlace, and (d) BasketballDrill.
Figure 5. Encoding time saving under different QPs. (a) BasketballPass, (b) FourPeople, (c) MarketPlace, and (d) BasketballDrill.
Sensors 22 07741 g005
Table 1. Test conditions.
Table 1. Test conditions.
Test platformVTM6.0
ConfigurationRandom Access (RA)
CTU size128 × 128
Quantization Parameter (QP)22, 27, 32, 37
Number of coded frames32
Table 2. Optimal CU size distribution under different QPs (unit: %).
Table 2. Optimal CU size distribution under different QPs (unit: %).
SequencesQPL64L32L16L8L4
2236.6725.2623.5912.242.24
MarketPlace2759.7417.8913.827.391.16
(1920 × 1080)3269.8414.8510.244.500.57
3778.2911.527.312.630.25
2270.7210.8210.485.832.15
FourPeople2782.716.366.623.071.24
(1280 × 720)3287.004.495.492.150.87
3789.623.504.811.520.55
2233.6523.8717.3816.848.26
BasketballDrill2742.3722.9617.5412.754.38
(832 × 480)3250.2622.4516.528.811.96
3758.2122.5813.514.890.81
2233.4626.0618.5614.107.82
BasketballPass2740.5126.7317.3410.574.85
(416 × 240)3246.5426.9616.497.182.83
3753.7225.8313.944.701.81
Average 58.3318.2613.357.452.61
Table 3. The distribution of the QTMT depth difference between adjacent frames (unit: %).
Table 3. The distribution of the QTMT depth difference between adjacent frames (unit: %).
SequencesQPd = 0d = 1d = 2d ≥ 3d ≤ 1
FourPeople2246.825.814.113.372.6
2766.017.38.97.883.3
3273.615.06.15.388.6
3779.212.15.13.691.3
MarketPlace2222.034.525.118.456.5
2735.634.417.112.970.0
3244.33213.89.976.3
3754.027.310.87.981.3
Average52.724.812.69.977.5
Table 4. The distribution of the QT depth difference between adjacent frames (unit: %).
Table 4. The distribution of the QT depth difference between adjacent frames (unit: %).
SequencesQPd = 0d = 1d = 2d ≥ 3d ≤ 1
FourPeople2265.127.76.60.692.8
2776.218.35.00.594.5
3281.715.03.00.396.7
3783.713.92.20.297.6
MarketPlace2242.250.27.20.492.4
2757.636.35.70.493.9
3267.128.54.20.295.6
3771.925.52.50.197.4
Average68.226.94.60.395.1
Table 5. The distribution of the MT depth difference between adjacent frames (unit: %).
Table 5. The distribution of the MT depth difference between adjacent frames (unit: %).
SequencesQPd = 0d = 1d = 2d ≥ 3d ≤ 1
FourPeople2253.529.411.06.182.9
2775.915.05.53.690.9
3281.611.94.02.593.5
3787.98.02.71.495.9
MarketPlace2230.339.821.08.970.1
2745.334.913.46.480.2
3251.832.710.45.184.5
3761.826.47.74.188.2
Average6124.89.54.785.8
Table 6. The probability that the MVD of a CU is 0 (unit: %).
Table 6. The probability that the MVD of a CU is 0 (unit: %).
QP22273237Average
Sequences
BasketballPass96.597.397.898.697.5
FourPeople87.191.992.793.691.3
MarketPlace96.998.498.899.198.3
BasketballDrill91.992.292.692.592.3
Table 7. Coding performance of proposed methods.
Table 7. Coding performance of proposed methods.
ClassSequencesEarly Skipping TT PartitionEarly Terminating QTMT PartitionProposed Overall Method
BDBR(%)TS(%)BDBR(%)TS(%)BDBR(%)TS(%)
A1Campfire1.9140.850.141.001.9842.08
FoodMarket40.1113.210.059.360.1921.10
Tango20.5916.140.567.940.9821.06
A2CatRobot10.3010.460.466.300.8618.85
DaylightRoad20.6116.370.532.231.1518.88
ParkRunning30.7029.060.100.920.7630.92
BBasketballDrive0.8620.930.554.131.2327.61
BQTerrace0.3820.680.286.480.5527.91
Cactus0.4017.100.406.670.7725.42
MarketPlace0.4611.860.422.760.8118.09
RitualDance1.2926.220.312.571.6531.79
CBasketballDrill0.8821.640.488.051.3030.12
BQMall0.7614.910.719.031.3525.25
PartyScene1.0424.240.242.481.0228.77
RaceHorsesC1.1422.180.294.391.5430.11
DBasketballPass0.313.920.8912.040.9918.63
BlowingBubbles0.378.270.6912.101.0721.88
BQSquare0.063.280.3210.250.3416.02
RaceHorses0.9413.050.8310.861.6325.48
EFourPeople0.165.030.305.430.3811.73
Johnny0.080.770.124.630.298.88
KristenAndSara0.150.480.165.350.449.57
Average0.6115.480.406.140.9723.19
Table 8. Coding performance comparison of fast inter CU partition methods.
Table 8. Coding performance comparison of fast inter CU partition methods.
ClassSequences[30][32]Proposed Method
BDBR(%)TS(%)BDBR(%)TS(%)BDBR(%)TS(%)
A1Campfire2.8728.212.8030.081.9842.08
FoodMarket41.4132.211.5942.900.1921.10
Tango23.7529.293.6834.050.9821.06
A2CatRobot14.1830.125.5930.620.8618.85
DaylightRoad24.8429.634.4329.201.1518.88
ParkRunning33.0531.541.6121.300.7630.92
BBasketballDrive5.2235.112.9632.391.2327.61
BQTerrace1.9534.640.9813.800.5527.91
Cactus2.9331.185.2025.420.7725.42
MarketPlace3.3034.193.2236.470.8118.09
RitualDance5.4938.532.9731.231.6531.79
CBasketballDrill1.7415.301.5924.381.3030.12
BQMall0.318.292.3522.411.3525.25
PartyScene2.0820.561.8414.941.0228.77
RaceHorsesC3.1520.662.2322.551.5430.11
DBasketballPass10.1024.261.5621.180.9918.63
BlowingBubbles2.2629.252.2916.971.0721.88
BQSquare3.3733.090.849.690.3416.02
RaceHorses6.2133.852.2420.331.6325.48
EFourPeople0.2711.591.7625.260.3811.73
Johnny1.9617.681.6924.920.298.88
KristenAndSara1.8715.182.1126.210.449.57
Average3.2926.562.5225.290.9723.19
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Y.; Luo, F.; Zhu, Y. Temporal Prediction Model-Based Fast Inter CU Partition for Versatile Video Coding. Sensors 2022, 22, 7741. https://doi.org/10.3390/s22207741

AMA Style

Li Y, Luo F, Zhu Y. Temporal Prediction Model-Based Fast Inter CU Partition for Versatile Video Coding. Sensors. 2022; 22(20):7741. https://doi.org/10.3390/s22207741

Chicago/Turabian Style

Li, Yue, Fei Luo, and Yapei Zhu. 2022. "Temporal Prediction Model-Based Fast Inter CU Partition for Versatile Video Coding" Sensors 22, no. 20: 7741. https://doi.org/10.3390/s22207741

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop