Next Article in Journal
Deep Learning Control for Digital Feedback Systems: Improved Performance with Robustness against Parameter Change
Next Article in Special Issue
Aspect-Based Sentiment Analysis in Hindi Language by Ensembling Pre-Trained mBERT Models
Previous Article in Journal
A Secure Protocol against Selfish and Pollution Attacker Misbehavior in Clustered WSNs
Previous Article in Special Issue
Two-Dimensional Audio Compression Method Using Video Coding Schemes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Context-Based Inter Mode Decision Method for Fast Affine Prediction in Versatile Video Coding

1
Department of Convergence IT Engineering, Kyungnam University, Changwon 51767, Korea
2
Department of Information and Communication Engineering, Kyungnam University, Changwon 51767, Korea
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(11), 1243; https://doi.org/10.3390/electronics10111243
Submission received: 23 April 2021 / Revised: 21 May 2021 / Accepted: 21 May 2021 / Published: 24 May 2021

Abstract

:
Versatile Video Coding (VVC) is the most recent video coding standard developed by Joint Video Experts Team (JVET) that can achieve a bit-rate reduction of 50% with perceptually similar quality compared to the previous method, namely High Efficiency Video Coding (HEVC). Although VVC can support the significant coding performance, it leads to the tremendous computational complexity of VVC encoder. In particular, VVC has newly adopted an affine motion estimation (AME) method to overcome the limitations of the translational motion model at the expense of higher encoding complexity. In this paper, we proposed a context-based inter mode decision method for fast affine prediction that determines whether the AME is performed or not in the process of rate-distortion (RD) optimization for optimal CU-mode decision. Experimental results showed that the proposed method significantly reduced the encoding complexity of AME up to 33% with unnoticeable coding loss compared to the VVC Test Model (VTM).

1. Introduction

As the state-of-the-art video coding standard, Versatile Video Coding (VVC) [1] was developed by Joint Video Experts Team (JVET), which was organized by ISO/IEC Moving Picture Experts Group (MPEG) and ITU-T Video Coding Experts Group (VCEG). VVC can support significant coding performance compared to High Efficiency Video Coding (HEVC) [2]. According to [3], the VVC test model (VTM) [4] accomplished twice the coding efficiency compared to HEVC test model (HM) [5] in the random-access (RA) configuration of JVET common test conditions (CTC) [6]. In order to realize this improvement, VVC has adopted new coding tools for more elaborated inter prediction, such as affine inter prediction [7] and bi-prediction with CU-level weight (BCW) [8]. In addition, VVC also has developed Adaptive Motion Vector Resolution (AMVR) [9] and Symmetric Motion Vector Difference (SMVD) [10] as well as Advanced Motion Vector Prediction (AMVP) to save the bitrate of the motion parameters. With regard to merge estimation, VVC has adopted Geometric Partitioning Mode (GPM) [11], Merge with Motion Vector Difference (MMVD) [12], Decoder-Side Motion Vector Refinement (DMVR) [13], and Combined Inter/Intra Prediction (CIIP) [14]. Although these tools can improve the coding performance, the computational complexity of the encoder was approximately increased up to 10 times under the RA configuration compared to that of HEVC [3].
One of the main differences between HEVC and VVC is the block structure scheme. Both HEVC and VVC specify the largest coding unit as coding tree unit (CTU) with an interchangeable size based on the encoder configuration. In addition, to adapt to the various block properties, CTU could be split into four coding units (CUs) using a quadtree structure (QT). HEVC specifies the multiple partition unit types, namely coding unit (CU), prediction unit (PU), and transform unit (TU). On the other hand, VVC substitutes them with another block structure named QT-based multi-type tree (MTT) structure so that MTT blocks can be partitioned into either binary trees (BTs) or ternary trees (TTs) to provide a variety of CU partitioning shapes. As shown in Figure 1, VVC specifies four MTT splitting types, which are vertical binary split (SPLIT_BT_VER), horizontal binary split (SPLIT_BT_HOR), vertical ternary split (SPLIT_TT_VER), and horizontal ternary split (SPLIT_TT_HOR). After a CTU is divided into four QTs (SPLIT_QT) by QT structure, each QT block can be further split into either sub-QTs or MTT blocks, while an MTT block can be only partitioned into sub-MTT blocks, as depicted in Figure 2. Note that a QT leaf node or a MTT leaf node is regarded as a CU, which is used as the basic unit of the prediction and transform processes without any further partitioning schemes. Therefore, a CU in VVC can have either a square or rectangular shape, while a CU in HEVC always has a square shape. It means that the block structure of VVC can provide better coding performance than HEVC by supporting flexible CU partitioning shapes.
In this paper, we define the relationship between the upper CU and current CUs. The upper CU can be either a QT node with square shape or a MTT node with square or rectangular shapes covering the area of the current CU. For example, the divided QT, BT, and TT CUs can have the same upper CU, which is the QT node with 2N × 2N, as illustrated in Figure 2a. Similarly, a MTT node is also regarded as an upper CU for further partitioned sub-MTT nodes, as depicted in Figure 2b.
In general, inter prediction searches the motion-compensated block from the previously decoded reference frames by using conventional motion estimation (CME). CME conducts a kind of block-matching algorithm based on the translational motion model. Because CME cannot efficiently describe complex motions in natural videos, such as zooming and rotation, VVC has newly adopted not only CME but also affine motion estimation (AME) to overcome the limitations of the translational motion model at the expense of higher encoding complexity. VVC provides two affine motion models in the process of AME, which are the four-parameter affine model (affine_4P) and the six-parameter (affine_6P). Because VVC performs the affine prediction in the unit of sub-block with the size of 4 × 4, the affine MVs of each sub-block are derived from a two-control-point MV (CPMV0, CPMV1) or a three-control-point MV (CPMV0, CPMV1, CPMV2) depending on affine_4P or affine_6P, respectively. As shown in Figure 3, CPMV0, CPMV1, and CPMV2 indicate the motion vectors of the top-left corner control point, the top-right corner control point, and the bottom-left corner control point, respectively. In this paper, the horizontal and vertical MVs of CPMV0, CPMV1, and CPMV2 are notated as ( m v 0 x ,   m v 0 y ) , ( m v 1 x ,   m v 1 y ) , and ( m v 2 x ,   m v 2 y ) , respectively. For the affine_4P, MV ( m v x ,   m v y ) at center sample location (x, y) for a sub-block is derived as in Equation (1).
m v x ( x , y ) = m v 1 x m v 0 x W x + m v 1 y m v 0 y W y + m v 0 x m v y ( x , y ) = m v 1 y m v 0 y W x + m v 1 x m v 0 x W y + m v 0 y
Similarly, in case of the affine_6P, MV ( m v x , m v y ) at center sample location (x, y) for a sub-block is derived as in Equation (2):
m v x ( x , y ) = m v 1 x m v 0 x W x + m v 2 x m v 0 x H y + m v 0 x m v y ( x , y ) = m v 1 y m v 0 y W x + m v 2 y m v 0 y H y + m v 0 y ,
where W and H indicate the width and the height of current CU, respectively. After computing affine MVs with 1/16 fractional MV accuracy, the predicted block of current CU is fetched from the reference frames in the unit of sub-block.
While AME enables efficient prediction of the complex motions beyond the reach of translational motions, it causes heavy computational complexity in the VVC encoder. According to the tool-off test of affine prediction [15], the coding loss and the time savings are 3.11% and 20%, respectively. Figure 4a shows the distribution between affine inter mode using AME and AMVP mode using CME in the RA configuration of JVET CTC. In addition, we investigated the distribution of affine inter mode on the reference and non-reference frames, as shown in Figure 4b. In RA configuration of VVC, it specifies group-of-picture (GOP) size to support the hierarchical encoding structure along the different temporal layers (TLs). The frames with the highest TL value are treated as the non-reference frames in VVC. Figure 4b indicates that affine inter modes rarely occurred in non-reference frames. Although affine inter mode does not frequently occur during the VVC encoding process, it contributes to improving the coding efficiency. It means that the fast encoding scheme of AME should be carefully designed in order to minimize the coding loss. In this paper, we propose a fast affine prediction method to reduce the computational complexity of AME with unnoticeable coding loss.
The remainder of this paper is organized as follows. In Section 2, we review the related fast encoding schemes to reduce the computational complexity of motion estimation in HEVC or VVC. Then, the proposed method is described in Section 3. Finally, experimental results and conclusions are given in Section 4 and Section 5, respectively.

2. Related Work

Although newly adopted VVC coding tools show the powerful compression performance compared to HEVC; it causes heavy computational complexity mainly due to the ME processes, including AME. Therefore, many researchers have studied fast encoding algorithms in the areas of ME without noticeable quality degradation. As discussed in [16,17], the increases of coding efficiency and computational complexity depend on how many reference frames are used. Therefore, reference frames should be carefully determined from the perspective of the trade-off between coding loss and complexity reduction. Pan et al. [18] reduced the number of reference frames based on the measurement with content similarity. In HEVC, a directional search pattern method [19], a rotating hexagonal pattern [20], and an adaptive search range [21] showed reasonable trade-off with the 64 × 64 search range to reduce the computational complexity of ME. VVC inherited several ME algorithms of HM, such as diamond search, raster search, and refinement processes [22], for fast ME process.
Another approach is the early termination of the ME process when the predefined conditions are satisfied at a certain ME process. For example, the ME process can be terminated with high probability because the recursive block-partitioning process has a strong correlation between motion parameters. In addition, the bi-directional ME process can be terminated with the PU correlations between QT structures with the different CU depth [23,24]. Shen et al. [25] proposed a fast CU split-decision algorithm and a CU depth-range estimation algorithm. It determines the early termination of CU partitioning using the neighboring CU-partitioning structure and motion amounts. In addition, Shen et al. [26] proposed a fast PU decision method using the motion activity. Tan et al. [27] addressed a three-step early CU-depth decision method by comparing the RD cost of the current CU with the average RD cost of the previously encoded CUs. Xiong et al. [28] presented a fast CU decision method using pyramid motion divergence, which exploits the correlation between the motion parameters and the CU split structure. Lee et al. [29] presented a fast CU size-selection method by comparing the average RD cost of the SKIP mode with an adaptive threshold.
Since AME for affine prediction was integrated on top of VTM 1.0 in 2018, several works have been developed to enhance the affine prediction in the middle of VVC standardization. In particular, Zhou et al. [30] proposed the concept of affine merge estimation to infer the affine motion parameters from the neighboring blocks without the explicit derivation of affine MVs. In [31], Zhang et al. proposed to omit the derivation of affine motion parameters in the case of chroma blocks with small block size. Those methods enable reduction of the total memory bandwidth and the encoding complexity compared to the initial affine prediction on VTM 1.0. Recently, Park et al. [32] reduced the encoding complexity of AME consisting of two steps: in the first step, if the best mode of the upper CU is encoded as SKIP mode, the AME of current CU is omitted to avoid the unnecessary RD computations caused by the affine prediction. In the second step, when the best reference frame of CME is selected as the nearest reference frame in the uni-prediction of list 0 (L0), AME is only performed on the same reference frame in the prediction of L0. While this method showed marginal coding loss compared to VTM 3.0, it slightly reduced the computational complexity of AME.

3. Proposed Method

In order to design the fast affine prediction, we investigated the context correlation between upper and current CU with regard to the affine prediction. Based on the encoding information of upper CU, we checked the reasonable conditions to skip the affine inter mode of current CU. Because the posterior probability by Bayes’ theorem shows the effectiveness of the conditions, we computed it with the likelihood and prior probability as p { ( U c b f 0   & &   U ! a f f i n e ) | C a f f i n e _ i n t e r } and p ( C a f f i n e _ i n t e r ) by Equation (3):
p { C a f f i n e _ i n t e r | ( U c b f 0   & &   U ! a f f i n e ) } = p { ( U c b f 0   & &   U ! a f f i n e ) | C a f f i n e _ i n t e r } p ( C a f f i n e _ i n t e r ) p ( U c b f 0   & &   U ! a f f i n e ) ,
where C, U, “&&”, and “!” are denoted as current CU, upper CU, AND, and NOT logic operator, respectively. In addition, cbf0 means that the transformed non-zero coefficients do not exist in the CU. In other words, if current CU is encoded as cbf0, the CU could be generally regarded as a static area with stationary motions, which might not require sophisticated coding tools, such as affine prediction. Note that affine includes affine_merge as well as affine_inter in this paper. For example, C a f f i n e _ i n t e r indicates that the best prediction mode of current CU is selected as affine inter mode in terms of rate-distortion optimization (RDO), as in Equation (4):
J = D + λ R ,
where D , λ , and R represent the distortion, Lagrange multiplier, and the required bitrates in current CU, respectively.
We obtained the prior and posterior probabilities from two full-HD (FHD) test sequences encoded by VTM 10.0 under RA configuration. Note that those statistics are derived from the different three QPs to avoid the same conditions specified by JVET CTC. Table 1 shows that the current CU has low probabilities to be encoded as affine inter mode when the upper CU is not encoded as affine and cbf0. Because the distribution of the affine prediction is affected by the motion properties of the video sequences, the prior probabilities vary according to the different sequences and QP values. On the contrary, the posterior probabilities maintain similar distributions in Table 1. For example, while the prior probability of RitualDance sequence is increased up to 27%, the posterior probability is much smaller than the prior. Those statistics imply that a VVC encoder can avoid much of the unnecessary and redundant AME processes based on the conditions of upper CU before encoding the current CU. Based on the properties of the described contexts, the proposed method is composed of two steps: one is the early termination of affine inter mode of current CU after checking the conditions of upper CU, and the other is to determine the optimal affine model between the four-parameter and six-parameter affine model. Because an upper CU can have many current CUs, the proposed method was designed to allow for current CUs to exploit the previously encoded information derived from the QTMTT block structure. In this framework, the encoding information of either the upper or current CU are used to determine whether the AME of current CU is performed or not. If it does not satisfy with the conditions to skip the AME process, the second step can be applied for further computational complexity reduction to select the optimal affine model between affine_4P and affine_6P.
Figure 5 shows the block diagram of the proposed method, which modified the existing VVC encoder with the gray-shadowed early-termination rules. Firstly, if the upper CU is not encoded as affine mode and cbf0, it means that the current CU is likely to have the translational motion. In that case, the affine inter mode of the current CU is skipped without any AME process, as shown in Figure 5. In addition, if the current frame is the non-reference frame with the highest TL, and the best mode of the current CU is not encoded as affine merge, AME is also skipped based on the distribution of Figure 4b. Secondly, we assumed that there are correlations between the current CU and upper CU in terms of the affine model for AME. Therefore, the current CU only performs the AME using affine_4P if the RD cost of affine_4P is smaller than that of affine_6P during the encoding of the upper CU, and the best mode of the current CU is encoded as the affine merge mode with affine_4P before encoding the AME process.

4. Experiment Results

The proposed method was evaluated under JVET CTC [6], as presented in Table 2, and we compared the proposed method with Park’s [32] method. Our experiments were performed on top of VTM 10.0 as an anchor and were run on the experimental environments, as in Table 3.
For comparison of the computational complexity, we measured the time saving (TS) of the AME encoding time (AMT), as in Equation (5):
T S A M T ( % ) = 1 4 Q P i { 22 , 27 , 32 , 37 } ( A M T o r g ( Q P i ) A M T f a s t ( Q P i ) A M T o r g ( Q P i ) × 100 ) ,
where A M T o r g and A M T f a s t are the AMT of the anchor and the fast method to be tested, respectively. Since the time taken for AMT can be fluctuated depending on the QP value, the TS of each test sequences was represented as the average of AMT as well as the total encoding time (TET) compared with the anchor. Similar to calculating AMT, the average TS was computed from the four TET results corresponding to the different QP values. Table 4 shows the complexity reduction between the proposed method and the previous method [32]. The proposed method reduces the AMT of the VTM up to 33%, on average, compared to the anchor. The maximum and minimum TS achieve 54% and 27% in BQTerrace and Cactus sequences with unnoticeable coding loss. Compared to the previous method, the TS of the proposed method is as fast as 15% and 3% in terms of AMT and TET, on average, respectively. According to [33], there might be a difference between the percentage of computational complexity and the actual encoding time. Table 5 shows both AMT and TET on top of VTM 10.0, which represents the sum of the encoding time from QP 22, 27, 32, and 37.
In order to evaluate the coding loss, we measured the Bjontegaard Delta Bit Rate (BDBR) [34]. In general, BDBR increase of 1% corresponds to BD-PSNR decrease of 0.05 dB where the positive increment of BDBR indicates the coding loss. Table 6 shows the coding loss both of the proposed method and the previous method compared to VTM 10.0. As shown in Table 6, the coding loss between the anchor and the proposed method is marginal for all test sequences. In particular, BDBR-U or BDBR-V shows better coding gain than the anchor in several sequences. For example, DaylightRoad2 sequence showed a 1.14% coding gain in terms of BDBR-U, and MarketPlace sequence showed a 0.76% coding gain in BDBR-V component. It implies that the proposed method can sustainably maintain the coding loss under the deployment of our fast affine scheme. In addition, Table 7 shows the percentage of CUs that perform the early termination of AME (affine_inter_off) and only AME with four parameters (affine_inter_4P), when the proposed method was integrated on top of VTM 10.0.

5. Conclusions

VVC has newly adopted an affine motion estimation (AME) method to overcome the limitations of the translational motion model at the expense of higher encoding complexity. In this paper, we proposed a context-based inter mode decision method for fast affine prediction that determines whether AME is performed or not in the process of rate-distortion (RD) optimization for optimal CU-mode decision. Because the fast encoding scheme of AME should be carefully designed in order to minimize the coding loss, we investigated the context correlation between upper and current CU with regard to the affine prediction. After we defined the relation between an upper CU and current CUs, we checked the reasonable conditions to skip the affine inter mode of current CU using the statistics of context correlations between upper CU and current CU. The proposed method was evaluated under JVET CTC, and we compared the proposed method with Park’s [32] method on top of VTM 10.0 as an anchor. For comparison of the computational complexity, we measured the time saving (TS) of the AME time (AMT). Experimental results show that the proposed method significantly reduced the encoding complexity of AME up to 33% with unnoticeable coding loss compared to VVC Test Model (VTM). In addition, the AMT of the proposed method is as fast as 15% compared to the previous method, on average.

Author Contributions

Conceptualization, S.J. and D.J.; methodology, S.J. and D.J.; software, S.J.; formal analysis, D.J.; investigation, S.J. and D.J.; data curation, S.J.; resources, D.J.; writing—original draft preparation, S.J.; writing—review and editing, D.J.; visualization, S.J.; supervision, D.J.; project administration, D.J.; funding acquisition, D.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Kyungnam University Foundation Grant, 2019.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bross, B.; Chen, J.; Liu, S.; Wang, Y. Versatile Video Coding (Draft 10). In Proceedings of the 19th Meeting Joint Video Experts Team (JVET), Teleconference (Online), 6–10 July 2020. [Google Scholar]
  2. Sze, V.; Budagavi, M.; Sullivan, G.J. High efficiency video coding (HEVC). In Integrated Circuit and Systems, Algorithms and Architectures; Springer: Berlin, Germany, 2014; pp. 1–375. [Google Scholar]
  3. Bross, F.; Li, X.; Suehring, K. AHG Report: Test Model Software Development (AHG3). In Proceedings of the 19th Meeting Joint Video Experts Team (JVET), Teleconference (Online), 6–10 July 2020. [Google Scholar]
  4. Versatile Video Coding (VVC) Test Model (VTM). Available online: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftware_VTM (accessed on 21 March 2021).
  5. Sullivan, G.; Ohm, J.; Han, W.; Wiegand, T. Overview of the High Efficiency Video Coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  6. Bossen, F.; Boyce, J.; Suehring, K.; Li, X.; Seregin, V. JVET Common Test Conditions and Software Reference Configurations for SDR Video. In Proceedings of the 12th Meeting Joint Video Experts Team (JVET), Macao, China, 3–12 October 2018. [Google Scholar]
  7. Yang, H.; Chen, H.; Zhao, Y.; Chen, J. Draft Text for Affine Motion Compensation. In Proceedings of the 11th Meeting Joint Video Experts Team (JVET), Ljubljana, Slovenia, 11–18 July 2018. Document JVET-K0565. [Google Scholar]
  8. Su, Y.-C.; Chen, C.; Huang, Y.; Lei, S.; He, Y.; Luo, J.; Xiu, X.; Ye, Y. CE4-Related: Generalized Bi-Prediction Improvements Combined. In Proceedings of the 12th Meeting Joint Video Experts Team (JVET), Macao, China, 3–12 October 2018. Document JVET-L0197, JVET-L0296, and JVET-L0646. [Google Scholar]
  9. Zhang, Y.; Han, Y.; Chen, C.; Hung, C.; Chien, W.; Karczewicz, M. CE4.3.3: Locally Adaptive Motion Vector Resolution and MVD Coding. In Proceedings of the 11th Meeting Joint Video Experts Team (JVET), Ljubljana, Slovenia, 10–18 July 2018. Document JVET-K0357. [Google Scholar]
  10. Luo, J.; He, Y. CE4-Related: Simplified Symmetric MVD Based on CE4.4.3. In Proceedings of the 13th Meeting Joint Video Experts Team (JVET), Marrakech, Morocco, 9–18 January 2019. Document JVET-M0444. [Google Scholar]
  11. Gao, H.; Esenlik, S.; Alshina, E.; Kotra, A.; Wang, B.; Liao, R.; Chen, J.; Ye, Y.; Luo, J.; Reuzé, K.; et al. Integrated Text for GEO. In Proceedings of the 17th Meeting Joint Video Experts Team (JVET), Brussels, Belgien, 7–17 January 2020. Document JVET-Q0806. [Google Scholar]
  12. Jeong, S.; Park, M.; Piao, Y.; Park, M.; Choi, K. CE4 Ultimate Motion Vector Expression. In Proceedings of the 12th Meeting Joint Video Experts Team (JVET), Macao, China, 3–12 October 2018. Document JVET-L0054. [Google Scholar]
  13. Sethuraman, S. CE9: Results of DMVR Related Tests CE9.2.1 and CE9.2.2. In Proceedings of the 13th Meeting Joint Video Experts Team (JVET), Marrakech, Morocco, 9–18 January 2019. Document JVET-M0147. [Google Scholar]
  14. Chiang, M.; Hsu, C.; Huang, Y.; Lei, S. CE10.1.1: Multi-Hypothesis Prediction for Improving AMVP Mode, Skip or Merge Mode, and Intra Mode. In Proceedings of the 12th Meeting Joint Video Experts Team (JVET), Macao, China, 3–12 October 2018. Document JVET-L0100. [Google Scholar]
  15. Chien, W.; Boyce, J. JVET AHG report: Tool Reporting Procedure (AHG13). In Proceedings of the 19th Meeting Joint Video Experts Team (JVET), Teleconference, 22 June–1 July 2020. Document JVET-S0013. [Google Scholar]
  16. Rahaman, D.; Paul, M. Virtual view synthesis for free viewpoint video and multiview video compression using gaussian mixture modelling. IEEE Trans. Image Process 2018, 27, 1190–1201. [Google Scholar] [CrossRef]
  17. Paul, M. Efficient multiview video coding using 3-D coding and saliency-based bit allocation. IEEE Trans. Broadcast. 2018, 64, 235–246. [Google Scholar] [CrossRef]
  18. Pan, Z.; Jin, P.; Lei, J.; Zhang, Y.; Sun, X.; Kwong, S. Fast reference frame selection based on content similarity for low complexity HEVC encoder. J. Vis. Commun. Represent. 2016, 40, 516–524. [Google Scholar] [CrossRef]
  19. Yang, S.-H.; Jiang, J.-Z.; Yang, H. Fast motion estimation for HEVC with directional search. Electron. Lett. 2014, 50, 673–675. [Google Scholar] [CrossRef]
  20. Nalluri, P.; Alves, L.; Navarro, A. Complexity reduction methods for fast motion estimation in HEVC. Signal Process. Image Commun. 2015, 39, 280–292. [Google Scholar] [CrossRef]
  21. Chien, W.; Liao, K.; Yang, J. Enhanced AMVP mechanism based adaptive motion search range decision algorithm for fast HEVC coding. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 3696–3699. [Google Scholar]
  22. Rosewarne, C.; Bross, B.; Naccari, M.; Sharman, K.; Sullivan, G. High Efficiency Video Coding (HEVC) Test Model 16 (HM 16) Improved Encoder Description Update 9. In Proceedings of the 28th Meeting Joint Collaborative Team on Video Coding (JCT-VC), Torino, Italy, 13–21 July 2017. Document JCTVC-AB1002. [Google Scholar]
  23. Park, S.; Lee, S.; Jang, E.; Jun, D.; Kang, J. Efficient biprediction decision scheme for fast high efficiency video coding encoding. J. Electron. Imaging 2016, 25, 063007. [Google Scholar] [CrossRef]
  24. Rhee, C.; Lee, H. Early decision of prediction direction with hierarchical correlation for HEVC compression. IEICE Trans. Inf. Syst. 2013, 96, 972–975. [Google Scholar] [CrossRef] [Green Version]
  25. Shen, L.; Zhang, Z.; An, P. Fast CU size decision and mode decision algorithm for HEVC intra coding. IEEE Trans. Consum. Electron. 2013, 59, 207–213. [Google Scholar] [CrossRef]
  26. Shen, L.; Zhang, Z.; Liu, Z. Adaptive inter-mode decision for HEVC jointly utilizing inter-level and spatiotemporal correlations. IEEE Trans. Circuits Syst. Video Technol. 2014, 24, 1709–1722. [Google Scholar] [CrossRef]
  27. Tan, H.; Liu, F.; Tan, Y.; Yeo, C. On fast coding tree block and mode decision for high-Efficiency Video Coding (HEVC). In Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012; pp. 825–828. [Google Scholar]
  28. Xiong, J.; Li, H.; Wu, Q.; Meng, F. A fast HEVC inter CU selection method based on pyramid motion divergence. IEEE Trans. Multimed. 2014, 16, 559–564. [Google Scholar] [CrossRef]
  29. Lee, J.; Kim, S.; Lim, K.; Lee, S. A fast CU size decision algorithm for HEVC. IEEE Trans. Circuits Syst. Video Technol. 2014, 25, 411–421. [Google Scholar]
  30. Zhou, M. CE4: Test Results of CE4.1.11 on Line Buffer Reduction for Affine Mode. In Proceedings of the 12th Meeting Joint Video Experts Team (JVET), Macao, China, 3–12 October 2018. Document JVET-L0045. [Google Scholar]
  31. Zhang, K.; Zhang, L.; Liu, H.; Wang, Y.; Zhao, P.; Hong, D. CE4: Affine Prediction with 4×4 Sub-blocks for Chroma Components (Test 4.1.16). In Proceedings of the 12th Meeting Joint Video Experts Team (JVET), Macao, China, 3–12 October 2018. Document JVET-L0265. [Google Scholar]
  32. Park, S.; Kang, J. Fast Affine Motion Estimation for Versatile Video Coding (VVC) Encoding. IEEE Access 2019, 7, 158075–158084. [Google Scholar] [CrossRef]
  33. Martínez-Rach, M.O.; Migallón, H.; López-Granado, O.; Galiano, V.; Malumbres, M.P. Imaging performance overview of the latest video coding proposals: HEVC, JEM and VVC. J. Imaging 2021, 7, 39. [Google Scholar] [CrossRef]
  34. Bjontegaard, G. Calculation of Average PSNR Differences between RD Curves. In Proceedings of the 13th Meeting Video Coding Experts Group (VCEG), Austin, TX, USA, 2–4 April 2001. Document VCEG-M33. [Google Scholar]
Figure 1. Four MTT splitting types in VVC.
Figure 1. Four MTT splitting types in VVC.
Electronics 10 01243 g001
Figure 2. Relations between an upper CU and current CUs: (a) Example of QT node as upper CU; (b) Example of MTT node as upper CU.
Figure 2. Relations between an upper CU and current CUs: (a) Example of QT node as upper CU; (b) Example of MTT node as upper CU.
Electronics 10 01243 g002aElectronics 10 01243 g002b
Figure 3. Affine MV derivation per a sub-block: (a) Affine MV derivation for four-parameter affine model; (b) Affine MV derivation for six-parameter affine model.
Figure 3. Affine MV derivation per a sub-block: (a) Affine MV derivation for four-parameter affine model; (b) Affine MV derivation for six-parameter affine model.
Electronics 10 01243 g003aElectronics 10 01243 g003b
Figure 4. Analysis of affine inter mode in the random-access configuration of JVET CTC: (a) Distributions between affine inter mode and AMVP mode; (b) Distributions of affine inter mode on the reference and non-reference frames.
Figure 4. Analysis of affine inter mode in the random-access configuration of JVET CTC: (a) Distributions between affine inter mode and AMVP mode; (b) Distributions of affine inter mode on the reference and non-reference frames.
Electronics 10 01243 g004
Figure 5. Block diagram of the proposed method.
Figure 5. Block diagram of the proposed method.
Electronics 10 01243 g005
Table 1. Prior and posterior probability according to the conditions of upper CU.
Table 1. Prior and posterior probability according to the conditions of upper CU.
SequenceQPPrior
Probability
Posterior
Probability
RitualDance2527.17%9.01%
3024.84%9.45%
3522.34%9.93%
BQTerrace2513.83%6.92%
3011.69%6.11%
359.52%5.86%
Total-15.80%5.54%
Table 2. Test sequences and encoding parameters used in the experiments.
Table 2. Test sequences and encoding parameters used in the experiments.
Sequences/
resolution/
frame rate
Tango2UHD
(3840 × 2160)
60 Hz
FoodMarket460 Hz
CatRobot60 Hz
DaylightRoad260 Hz
ParkRunning350 Hz
MarketPlaceFHD
(1920 × 1080)
60 Hz
RitualDance60 Hz
Cactus50 Hz
BasketballDrive50 Hz
BQTerrace60 Hz
Encoding
parameters
QP22, 27, 32, 37
Num. of reference frames2
Num. of encoding frames33
Search range384
RD optimizationEnable
Use hadamardEnable
GOP size16
Coding configurationRandom access
Table 3. Experimental environments.
Table 3. Experimental environments.
Experimental EnvironmentOptions
ProcessorInter(R) Xeon(R) Gold 6138
RAM256 GB
Operating system64-bit Windows 10
C++ compilerMicrosoft Visual C++ 2019
Table 4. Time saving between the proposed and previous method [32] on top of VTM 10.0.
Table 4. Time saving between the proposed and previous method [32] on top of VTM 10.0.
ClassSequence NamePark [32]Proposed
TSAMTTSTETTSAMTTSTET
UHDTango218%1%31%4%
FoodMarket414%2%32%6%
CatRobot23%5%29%6%
DaylightRoad217%1%31%3%
ParkRunning312%1%28%4%
Average17%2%30%5%
FHDMarketPlace19%4%28%5%
RitualDance7%0%38%8%
Cactus20%2%27%2%
BasketballDrive15%1%33%5%
BQTerrace36%5%54%3%
Average19%2%36%5%
Overall Average18%2%33%5%
Table 5. Encoding time (second) for anchor, the previous [32], and the proposed method.
Table 5. Encoding time (second) for anchor, the previous [32], and the proposed method.
ClassSequence NameVTM 10.0Park [32]Proposed
AMTTETAMTTETAMTTET
UHDTango222,085161,00618,168159,77715,264155,151
FoodMarket417,135102,89614,668101,11411,62196,835
CatRobot22,865134,10817,618127,56916,210125,670
DaylightRoad230,100188,02724,895185,32520,772181,654
ParkRunning336,030257,75931,795255,16726,053247,644
FHDMarketPlace770640,471625538,985554138,510
RitualDance740438,528687338,366455535,370
Cactus514439,340410538,580377838,579
BasketballDrive704351,514599850,858468948,899
BQTerrace329937,509209835,605151336,279
Table 6. Coding loss between the proposed method and the previous method [32] on top of VTM 10.0.
Table 6. Coding loss between the proposed method and the previous method [32] on top of VTM 10.0.
ClassSequence NamePark [32]Proposed
BDBR-YBDBR-UBDBR-VBDBR-YBDBR-UBDBR-V
UHDTango2−0.16%0.02%0.02%−0.11%−0.92%0.12%
FoodMarket4−0.02%0.36%0.33%−0.08%0.35%0.21%
CatRobot0.06%0.29%−0.11%0.07%−0.60%−0.11%
DaylightRoad20.05%−1.55%−0.21%0.06%−1.14%−0.06%
ParkRunning3−0.02%0.02%0.21%0.05%0.03%0.15%
Average−0.02%−0.17%0.05%0.00%−0.46%0.06%
FHDMarketPlace0.00%−0.08%−0.31%−0.04%0.13%−0.76%
RitualDance0.19%0.41%−0.15%0.23%0.18%−0.38%
Cactus0.10%−0.08%0.55%0.04%0.54%0.17%
BasketballDrive0.22%0.00%0.37%0.16%−1.70%0.40%
BQTerrace0.06%0.06%−0.04%0.02%−0.37%0.51%
Average0.11%0.06%0.08%0.08%−0.24%−0.01%
Overall Average0.05%−0.05%0.07%0.04%−0.35%0.03%
Table 7. Distribution of affine prediction by the proposed method.
Table 7. Distribution of affine prediction by the proposed method.
ClassSequence Nameaffine_inter_offaffine_inter_4P
UHDTango232%1%
FoodMarket430%1%
CatRobot31%2%
DaylightRoad236%5%
ParkRunning335%2%
Average33%2%
FHDMarketPlace31%4%
RitualDance37%4%
Cactus30%7%
BasketballDrive33%4%
BQTerrace55%1%
Average37%4%
Overall Average35%3%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jung, S.; Jun, D. Context-Based Inter Mode Decision Method for Fast Affine Prediction in Versatile Video Coding. Electronics 2021, 10, 1243. https://doi.org/10.3390/electronics10111243

AMA Style

Jung S, Jun D. Context-Based Inter Mode Decision Method for Fast Affine Prediction in Versatile Video Coding. Electronics. 2021; 10(11):1243. https://doi.org/10.3390/electronics10111243

Chicago/Turabian Style

Jung, Seongwon, and Dongsan Jun. 2021. "Context-Based Inter Mode Decision Method for Fast Affine Prediction in Versatile Video Coding" Electronics 10, no. 11: 1243. https://doi.org/10.3390/electronics10111243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop