Next Article in Journal
LSTM Attention Neural-Network-Based Signal Detection for Hybrid Modulated Faster-Than-Nyquist Optical Wireless Communications
Next Article in Special Issue
A Gait-Based Real-Time Gender Classification System Using Whole Body Joints
Previous Article in Journal
Estimating the Roll Angle for a Two-Wheeled Single-Track Vehicle Using a Kalman Filter
Previous Article in Special Issue
Hint-Based Image Colorization Based on Hierarchical Vision Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Study on Fast and Low-Complexity Algorithms for Versatile Video Coding

School of Computing, Gachon University, Seongnam 13120, Republic of Korea
Sensors 2022, 22(22), 8990; https://doi.org/10.3390/s22228990
Submission received: 27 October 2022 / Revised: 18 November 2022 / Accepted: 18 November 2022 / Published: 20 November 2022
(This article belongs to the Special Issue Applications of Video Processing and Computer Vision Sensor II)

Abstract

:
Versatile Video Coding (VVC)/H.266, completed in 2020, provides half the bitrate of the previous video coding standard (i.e., High-Efficiency Video Coding (HEVC)/H.265) while maintaining the same visual quality. The primary goal of VVC/H.266 is to achieve a compression capability that is noticeably better than that of HEVC/H.265, as well as the functionality to support a variety of applications with a single profile. Although VVC/H.266 has improved its coding performance by incorporating new advanced technologies with flexible partitioning, the increased encoding complexity has become a challenging issue in practical market usage. To address the complexity issue of VVC/H.266, significant efforts have been expended to develop practical methods for reducing the encoding and decoding processes of VVC/H.266. In this study, we provide an overview of the VVC/H.266 standard, and compared with previous video coding standards, examine a key challenge to VVC/H.266 coding. Furthermore, we survey and present recent technical advances in fast and low-complexity VVC/H.266, focusing on key technical areas.

1. Introduction

Many qualified media services have been provided owing to the advancements in multimedia technology, including content generation, data compression, large-data delivery, rendering technology, and real-time encoding/decoding technology. Based on such technologies, territory broadcasting, movies, on-demand videos, video-based conference calls, video-based mobile communications, video surveillance, real-time remote control, 3D videos, augmented reality videos, and virtual reality video services have generated significant interest among consumers [1].
The popularity of media services is causing network bandwidth problems due to the delivery of various types of media that require a large amount of data for qualified services. A recent Cisco report claimed that video data account for roughly 80% of all Internet data traffic [2]. This aspect is becoming increasingly entrenched because of the recent increase in non-face-to-face activities. Following the COVID-19 pandemic, for example, many activities have been replaced with non-face-to-face multimedia methods, and such video-based solutions are increasing the amount of video data traffic. With the growth in video traffic, video coding techniques capable of reducing video traffic have become more important.
Historically, video coding has been essential to the growth of the media industry. Without the use of data compression technology in the development of digital media, a digital media transformation would be impossible. MPEG-1 [3] was used on CDs, DVDs, and other storage media and played a significant role in changing the use of storage media. MPEG-2/H.262 [4] technology enabled digital broadcasting to replace analog broadcasting. MPEG-4 Advance Video Coding (AVC)/H.264 [5] is a bitstream that has been used as a core transmission format in the Internet era, accounting for more than 80% of all Internet streams worldwide [6]. High-Efficiency Video Coding (HEVC)/H.265 [7] is an ultra-high-definition (UHD) video format used in 4K terrestrial broadcasting, where it plays a leading role [8].
The aforementioned video coding technologies are typically a result of collaboration between two well-known video standard organizations, the ITU-T Video Coding Experts Group (VCEG) and the ISO/IEC Moving Picture Experts Group (MPEG), used to meet the industry needs and consumer demands for a higher resolution, higher quality, and higher frame rate within the context of constrained intra-bandwidth structures. MPEG-2/H.262 was the first video coding standard jointly created by the VCEG and MPEG groups [9]. Its goal was to enable digital television services and has been successful in this endeavor. The VCEG and MPEG groups jointly developed the AVC/H.264 video coding standard, which has been rapidly adopted in online services [10]. Since 2010, the VCEG and MPEG groups formed the Joint Collaborative Team on Video Coding (JCT-VC), which has designed the HEVC/H.265 standard for UHD services [11].
Following the completion of HEVC/H.265, there have been calls for a new video coding standard enabling the support of realistic media requiring large numbers of data, such as 8K or higher resolutions, 360-degree virtual reality, screen contents, high dynamic range, a wide chromatic gamut, and adaptive streaming [12], as shown in Figure 1. To properly support realistic media, the VCEG and MPEG groups established the Joint Video Exploration Team (JVET) and began researching important technologies for coding efficiency in 2015. After the study period, Versatile Video Coding (VVC)/H.266 [13] was launched in 2018, the primary objective of which was to achieve a compression capability that is noticeably better than that of HEVC/H.265 and to have the functionality to support a variety of applications with a single profile. The formal standardization period ran from 2018 to 2020, and in October 2020, JVET released the definition of VVC/H.266 version 1.
VVC/H.266 effectively accomplished its goal by utilizing a variety of cutting-edge coding techniques for intra- and inter-prediction, transform, quantization, in-loop filtering, and entropy coding, together with flexible block partitioning. For similar objective and subjective video qualities, VVC/H.266 showed a considerably larger bitrate reduction than HEVC/H.265 [13]. However, the encoding complexity is significantly higher than that of HEVC/H.265 because VVC/H.266 uses many coding tools with flexible partitioning, and the decision of tool on/off and the proper shape of the coding block results in numerous encoding processes. Real-time software and hardware applications may experience a great deal of strain as a result of a high encoding complexity.
To address the issue of the VVC/H.266 coding complexity, numerous studies on fast coding methods with low complexity have recently been conducted. In this study, we aim to provide a comprehensive review of the latest fast and low-complexity encoding methods for versatile video coding. The main contributions of this paper are as follows. (1) We conduct a brief review of the advancements of the VVC/H.266 standard over previous video coding standards. (2) We analyze and identify key challenges in video encoding based on VVC/H.266 coding. Finally, (3) we conduct a comprehensive survey of recent advances in fast and low-complexity video coding methods, which are classified into specific coding areas for low complexity.
The remainder of this paper is organized as follows. Section 2 provides a detailed overview of the VVC/H.266 standard and some key technologies. In Section 2, the VVC/H.266 coding performance and complexity are analyzed, and challenging issues are presented. In Section 3, recent studies on the fast and low-complexity VVC/H.266 are reviewed from key technical perspectives. Finally, Section 4 provides some concluding remarks.

2. Overview and Complexity Analysis of VVC/H.266 Standard

2.1. VVC/H.266 Standard

The VVC/H.266 standard uses a block-based hybrid coding framework, which is similar to earlier video coding standards. To improve the coding performance, the VVC/H.266 standard employs novel coding tools with a flexible block partitioning method. As with all previous video coding standards, the VVC/H.266 standard employs intra-prediction to reduce the spatial redundancy, inter-prediction to reduce the temporal redundancy, transform coding of the residuals to further reduce the spatial statistical redundancy, and in-loop filters to enhance the quality of the reconstructed video. The VVC/H.266 standard specifically employs a multi-type tree (MTT) structure as quad-tree, binary, and/or tri-tree block partitions using a 1:2 to 1:8 ratio between the width and height of the block [14,15]. In contrast to the HEVC/H.265 block partitioning method, there are no prediction units for prediction or the transformation sizes, nor are there any transform units for a transformation. In general, intra- and inter-predictions are used in block or sub-block units, followed by transformations, quantization, and entropy processes for residual coding and in-loop filter chains to improve the visual quality.
Intra-prediction is improved by increasing the number of directional angles from 35 to 93 to provide an accurate prediction using the new VVC/H.266 partitioning shapes [16]. To increase the number of original pixels in the prediction, new intra-tools such as a position-dependent prediction combination [17], multiple reference line (MRL) [18], matrix-based intra-prediction [19], cross-component linear model [20], and intra-sub-partition (ISP) [21] have been adopted. Eventually, by improving the existing prediction tools and adopting new intra-tools, the intra-coding performance has improved significantly.
To enhance the coding performance by minimizing the temporal duplication between sequential frames, many inter-prediction tools have been adopted in the VVC/H.266 standard. According to whether motion data are shared across an entire block, newly adopted inter-prediction tools in VVC/H.266 can be typically divided into two groups. For instance, all block-based inter-prediction tools include history-based motion vector prediction [22], merge with motion vector difference [23], symmetric motion vector difference [24], adaptive motion vector resolution [25], geometric partitioning mode (GPM) [26], bi-prediction with CU-level weights [27], and combined intra- and inter-prediction [28]. By contrast, sub-block-based methods include affine motion [29], sub-block-based temporal motion vector prediction [30], decoder side motion vector refinement [31], bidirectional optical flow [32], and prediction refinement with an optical flow [33]. Owing to the introduction of these cutting-edge techniques, VVC/H.266 inter-prediction has been significantly improved compared to HEVC/H.265 inter-prediction.
A multiple transform selection (MTS) [34], sub-block transform [35], and non-separable secondary transform [36] have been used toward a transform and quantization of the residual coding for both inter- and intra-compressed blocks. The quantization in VVC/H.266 adopts dependent quantization, which can be a form of sliding block vector quantization [37]. Luma mapping with chroma scaling, a cross-component adaptive loop filter, and an adaptive loop filter are all new in-loop filtering filters introduced in VVC/H.266 [38].
The VVC/H.266 standard improved all coding components of the hybrid coding structure and incorporated advanced coding tools to achieve the challenge goal of improving the coding efficiency by 50% over the prior coding standard (i.e., the HEVC/H.265 standard). Although these advancements have contributed to achieving the objectives of VVC/H.266, an increase in complexity was unavoidable, which needed to be resolved.

2.2. Complexity Analysis

During the official standard phase, JVET maintained a VVC/H.266 test model called VTM [39], which served as reference software for VVC/H.266 testing. The primary goal of VTM is to provide an exemplary reference implementation of a testbed VVC/H.266 encoder and decoder and to evaluate the proposed tools. In addition, JVET created the common test condition (CTC) [40] to evaluate the coding performance under the same testing environment. CTC specifies the test conditions based on the following scenarios that are commonly used in the real world: (1) all intra (AI), in which all frames are encoded with I slices; (2) random access (RA), which uses picture reordering with a random access picture every 1 s; (3) low delay with B slices (LB), in which frame reordering is not allowed and only the first frame is encoded using an I slice followed by B slices; and (4) low delay with P slices (LP), in which frame reordering is not allowed and only the first frame is encoded using I slices followed by P slices. Based on the four CTC scenarios, the coding performance of VVC/H.266 can be compared with the previous coding standard.
Figure 2 summarizes the performance of VTM10.0 [39] over HM16.22 [41] for all configurations in a JVET CTC environment [42]. Bjøntegaard delta bitrates (BDBR) [43] are used to compare the coding performance of VTM10.0 as a whole package to HM16.22. The runtime was used as an estimation of the codec complexity to check the encoding and decoding complexity, with the following T measurement: T = Ttest/Tanchor * 100%, where Ttest and Tanchor represent the runtimes of the tested and anchor methods, respectively. A value of 100% indicates that there is no difference in runtime.
According to Figure 2, VTM10.0 reduces the luma BDBR by 25.1%, 36.1%, 30.9%, and 33.9% compared to HM16.22. In the AI, RA, LB, and LP configurations, the encoder runtime of VTM10.0 is approximately 27, 9, 7, and 6 times slower than that of HM16.22. In all configurations, the decoder runtime of VTM10.0 is approximately 1.5 times that of HM16.22. The increase in the encoding and decoding times is primarily due to the addition of new tools that require additional rate-distortion checks for selecting the best mode for a block, and the flexible block partitioning also necessitates a more exhaustive search for finding the optimal partitioning for a coding tree unit (CTU). For instance, the encoder must compute the bit rate and distortion of all feasible combinations of the block partitions and compressible coding tools before deciding on the optimum partition and tool for a given block. Such computations used for selecting the best partition with the best coding tools result in a significant increase in the VVC/H.266 complexity.

3. Fast and Low-Complexity Coding for VVC/H.266

Several recent efforts have been made to address the complexity of VVC/H.266. The majority of studies on fast and low-complexity coding of VVC/H.266 have focused on reducing the complexity of the block partitioning. Based on our search of recent papers on this topic, we identified that more than half are related to the fast method for block partitioning, as shown in Figure 3. Considering that block partitioning involves all encoding processes associated with the coding tools, focusing on fast partitioning makes sense. As a result, in this section, we look at recent advances in fast and low-complexity video coding methods, which are divided into four categories: (1) fast methods for an early split mode decision, (2) fast methods for an early coding unit (CU) depth decision, (3) fast methods for coding tools, and (4) low-complexity platform-dependent methods.

3.1. VVC/H.266 Block Partitioning

This subsection describes the basic partitioning process used in VVC/H.266 encoding before reviewing the fast and low-complexity methods applied for partitioning. The block partitioning used in VVC/H.266 achieves a significant coding performance by allowing for a flexible block size. Such flexible partitioning can create adaptive CU partitions based on the video image characteristics using the newly introduced MTT. Figure 4 shows an example of VVC/H.266 block partitioning versus HEVC/H.265 block partitioning. As illustrated in the figure, VVC/H.266 can determine a more flexible CU than HEVC/H.265 based on the frame content in the frame. This flexibility is achieved by layering the binary tree (BT) and tri-tree (TT) partitioning on top of quadtree (QT) partitioning. This is one of the major differences between BT and TT partitioning, which can produce nonrectangular CU shapes depending on the content.
Nonetheless, the introduction of BT and TT significantly increases the encoding time. To determine the best-fitting CU block shape, an exhaustive evaluation of all possible QT, BT, and TT block shapes results in significantly more recursive calls to the coding tool functions than HEVC/H.265 using QT. Figure 5 depicts the encoding process of VVC/H.266 for determining the CU as well as the corresponding split mode of the tree at a given depth. Binary tree splits were first evaluated horizontally and vertically at the given depth, followed by tri-tree splits horizontally and vertically, and the quadtree was then evaluated. When MTT (i.e., binary tree and tri-tree) is applied to the QT leaf, only BT and TT are permitted, whereas QT is prohibited for all subsequent nodes. Each MTT node has the option of being non-split (as an MTT leaf) or divided into two child MTT nodes by a horizontal binary tree (HBT), two child MTT nodes by a vertical binary tree (VBT), three child MTT nodes by a horizontal ternary tree (HTT), or three child MTT nodes by a vertical ternary tree (VTT). The two MTT child nodes in the BT scenario were of the same size, with each being half the size of the parent MTT node. The three MTT child nodes in the TT scenario have a splitting ratio of 1:2:1 and are one-fourth, one-half, and one-quarter the size of the parent MTT node, respectively. Choosing the split mode and depth for optimal partitioning requires a significant amount of encoding time. Two approaches have been investigated to reduce the encoding time of the chosen partitioning.

3.2. Fast Method on Early Split Mode Decision

The first approach is to investigate the split mode for an early split mode determination or to skip some of the split modes in the mode evaluation. At a given depth, five split modes were evaluated in the VVC/H.266 partitioning, i.e., HBT, VBT, HTT, VTT, and QT. With this approach, the partitioning process involves a reduced number of evaluations to determine the split mode by skipping unnecessary evaluations at a given depth. This approach saves a significant amount of encoding time by reducing computations in a skipped mode evaluation. The following studies were investigated from this perspective.
Park et al. [44] proposed a simple early decision method based on a probabilistic approach that can effectively reduce the TT complexity by exploiting rate distortion (RD) costs from previously encoded CU data. The authors specifically investigated the split modes determined after the encoding process, followed by an examination of the relationship between the TT split and the texture of the contents, which can be estimated using the RD cost of the partitioning shape. The authors developed a TT decision model based on the Bayesian probability approach, and the proposed method adaptively skips the TT partition evaluation process, thereby saving a significant amount of encoding time with a marginal coding loss.
In [45], Park et al., proposed a fast TT decision method by exploiting the statistical information of coded bitstreams representing the correlation with the TT and developed two useful types of features: intra-prediction information and block information using QT, HBT, and VBT during an evaluation. The authors created a neural network and model trained using these two features. The developed model can efficiently determine whether the TT partitioning process is involved. Consequently, the proposed method efficiently reduces the encoding time related to the TT partitioning in the encoder.
In [46], Zhao et al. proposed a fast CU partitioning method that investigates the just noticeable difference (JND) model and motion state by combining such information with a decision tree to develop a CU partition decision strategy oriented toward the perceived quality of the human visual system (HVS). The authors developed a threshold that determines whether an individual split mode is evaluated at a given depth based on an analysis of the JND model and the motion state, and used the threshold to skip an individual mode evaluation.
In [47], Zhang et al., proposed a fast CU decision method based on the DenseNet network that trains a convolutional neural network (CNN) to predict the probability that the edges of the 4 × 4 blocks in each 64 × 64 block skip the computation of the unnecessary rate distortion optimization (RDO) and accelerates the coding process. The proposed CNN model analyzes the texture of four 64 × 64 blocks of content, and the produced probability of the 4 × 4 blocks in the 64 × 64 is used to determine the RDO evaluation of the individual split modes.
In [48], Saldanha et al., proposed a fast CU partitioning method using a light gradient boosting machine (LGBM) to reduce the VVC/H.266 intra-coding time. The authors trained five LGBM classifiers offline to avoid evaluating the split modes that were likely to be skipped. The generated LGBM classifiers are trained using features extracted from the texture, coding, and coding context information, and the classifiers determine whether each split mode is applied at a given depth. The authors emphasized that the proposed LGBM classifiers can handle a wide range of video characteristics and resolutions, allowing them to support many applications while requiring a relatively short encoding time.
Table 1 summarizes the fast methods for the early split-mode decision. As shown in the table, the proposed methods can reduce the encoding time by 34% to 54% with a relatively marginal coding loss by estimating the promising split mode at the early stage rather than evaluating all possible split modes.

3.3. Fast Method Applied to Early CU Depth Decision

The second approach is to determine the portioning depth at an early stage. In the VVC/H.266 partitioning process, all possible combinations of CU shapes are evaluated under the allowed depths of QT, BT, and TT. According to Figure 5, QT, HBT, VBT, HTT, and VTT are evaluated sequentially, followed by the same portion mode at the next depth with a reduced block size depending on the parent CU block. Such a recursive calling process for evaluating all possible block combinations consumes a considerable amount of the encoding time. If the encoder determines the optimal CU blocks at the early stage, it significantly reduces the useless evaluation process for further depth of partitioning. From this viewpoint, the following studies were conducted to develop a fast-encoding method for determining the CU depth at an early stage.
In [49], Zhang et al., proposed a fast-partitioning method for early depth and intra-mode decisions, and investigated a fast CU partitioning based on a random forest classifier (RFC) model and fast intra-prediction mode decision using the texture region features. First, the proposed method classifies the texture complexity of the current CU using information extracted from the mean of the absolute difference between pixels to define the difference between each pixel and its surrounding pixels for fast CU partitioning. The extracted feature is then used in the RFC to determine the coding depth by providing a split threshold. Furthermore, the authors exploited the correlation with the pixel similarity in the corresponding direction, and then proposed a fast intra-mode decision using the energy of the CU in four directions based on the texture information using the Canny operator to avoid an unnecessary intra-prediction mode evaluation.
In [50], Zhang et al., proposed a split-mode method and a fast CU depth decision approach. First, the proposed method determines whether a CU is divided by calculating the texture complexity using the rough mode decision (RMD)-based cost (JRMD) and angular second moment (ASM). The decision is made by comparing the ASM value of the current CU with the derived threshold. Second, the authors presented a fast split-mode decision method that uses a threshold generated from the SAD of each direction to determine which split modes are skipped for the current depth.
Zhang et al. [51] proposed a fast Bayes-based CU partitioning method by leveraging the relationship between JRMD and split depth determination. The proposed method determines whether a further split is required based on the threshold generated by the JRMD. Furthermore, the authors use a deblocking filter (DBF) to check the texture information of the current block and embed such information to determine which split modes can be discarded in the evaluation.
Fan et al. [52] proposed a fast CU partitioning method to determine whether to split a CU by exploiting the texture smoothness using the variance of the given block and a Sobel operation for terminating further partitions. To this end, the authors derived a threshold using the Sobel operation by comparing the variance of the block. Furthermore, the authors presented a gradient of the texture of the contents to only select one partition from the five split modes. The authors generated two additional thresholds to determine which split modes were evaluated for the proposed method.
In [53], Yang et al., proposed a fast CU partitioning method and a fast intra-mode decision method. First, the authors analyzed the statistics between the determined CU size and the texture information, and then proposed a statistical learning-based fast depth decision method derived by calculating the features to measure the texture characteristics and context correlations. These features are then used as input values to the classifiers to determine whether CU processing is required at higher depths. In addition, the authors exploited the Hadamard cost of each directional mode in the MPMs, followed by a gradient descent-based search to find the optimal intra-mode prediction. The proposed intra-prediction methods reduce the RDO computations of unnecessary intra-modes.
In [54], Tang et al., proposed a fast-partitioning method for intra- and inter-coding. The block-level-based Canny edge detector is used for intra-coding to extract the edge features to conduct an early termination of the split depth. Similarly, for inter-coding, the authors exploited the temporal correlation of consecutive frames in the video to generate a threshold based on the difference between three consecutive frames to determine whether a future split is required. Additionally, the authors presented a method for a fast split-mode decision using thresholds and a Canny edge detector skipping vertical or horizontal partition modes.
Liu et al. [55] proposed a fast-partitioning method for inter-frame coding based on spatiotemporal information by utilizing the motion features and texture complexity of the current coding block. The authors calculated the average sum of the square difference (ASSD) of the luma values of the current block and the reference block at the same position in the co-located reference frame and its derived threshold to determine whether further CU partitioning is required. Furthermore, the authors used the derived threshold to bypass the individual split mode during a split mode evaluation.
In [56], Li et al., proposed a deep MSE-CNN model that combines a conditional convolution and sub-networks with the adequate network capacity to determine the CU partition at an early stage of the CU partitioning, which can skip unnecessary evaluation processes on unused CUs. MSE-CNN with a 128 × 128 CTU input is used in the proposed method to extract a collection of 128 × 128 feature maps. The five split modes are applied sequentially using the feature maps, which are then sent into a sub-network to anticipate one of the CU split modes. The network then decides whether further splitting is necessary.
In [57], Chen et al., proposed a fast approach for VVC/H.266 intra-coding by utilizing the human visual system and a machine learning technique. To identify the visually distinguishable pixels in a CU, the author used a perceptual model of the human eyes, which shows only perceptible differences. The quantization parameter (QP) and the horizontal and vertical projections of visually distinct pixels are used as the input values of the random forest machine learning models to predict the CU partitions and remove computations regarding unnecessary split decisions.
In [58], Yea et al., proposed a CNN-based fast split mode decision method for inter-coding by utilizing the original and residual image of a CU, the picture order count, and the CU-level QP value. The proposed multi-level tree CNN method predicts which of the five split modes will be evaluated to reduce the time complexity of the inter-picture prediction mode evaluation.
Pan et al. [59] proposed an MF-CNN-based CU partitioning early termination method to streamline the CU partitioning process by utilizing texture and motion activity features for a fast inter-coding approach. The authors trained a CNN model to determine the CU depth using the luma component, residuals, and bidirectional motion field of the CU. In addition, the authors presented an early decision method for the merge mode. According to a statistical analysis, the authors discovered that determining the merge mode early can save a significant amount of encoding time. Similar to early CU partitioning, the authors applied an MF-CNN to determine whether the merge mode was the best inter-coding mode.
The method in this section uses texture or motion information to determine the early stage of depth during the CU encoding processing. Based on the observation that such information is strongly related to the determined partitions, the proposed methods apply novel algorithms developed to determine the optimal CU size and partitions at an early stage. As shown in Table 2, the fast methods for early CU depth decisions can save 12% to 58% of the total encoding time by skipping unnecessary computations when evaluating the combination of the tool and CU size.

3.4. Fast Method for Coding Tools

Many tools have been adopted in VVC/H.266 to improve the coding performance for both intra- and inter-coding. Among the newly adopted and updated tools, several computations are required to determine the best modes or motion information. In this section, we look at studies that have been conducted to reduce the number of computations required by individual coding tools. Most studies reduce the number of RDO checks by reducing the modes or determining the best mode early in the process.
Dong et al. [60] proposed a fast intra-mode decision method based on two efficient algorithms. First, the proposed method efficiently eliminates any unnecessary RDO processing by skipping the intra-block copy (IBC) and ISP tools from learning-based classifiers and excluding subsequent candidates in a complete mode list. The authors took advantage of the early depth decision to improve the speed by categorizing CUs into three groups using texture and coding information.
In [61], Tun et al., proposed a fast intra-prediction mode selection for intra-prediction mode RDO computations. The authors first examined the relationship between the RD costs of the RDO processes and SATD costs of the RMD processes and then developed a threshold that can be used to determine which intra mode will be promising. Based on the threshold, the proposed method can include only a small number of intra-predictions in the RDO evaluation process, which is time consuming.
In [62], Park et al., focused on the complexity reduction of ISP. ISP is newly adopted in VVC/H.266, which provides a flexible block shape for an intra-prediction. By referring to closely neighboring reconstructed pixels, such a flexible block shape can provide a more accurate prediction. However, ISP is one of the more time-consuming tools in VVC/H.266 intra-coding. In this paper, the authors proposed a method for skipping an ISP mode that requires an RD-based search to save the encoding time associated with the ISP coding tool by utilizing the relation between ISP and the MRL.
Tsang et al. [63] proposed a fast prediction network based on deep learning for screen content coding (SCC) tools. To handle graphically generated or mixed layered contents, SCC coding tools in VVC/H.266, such as IBC and Palette (PLT), are included in version 1. The characteristics of the contents differ from those of normal natural contents, and such SCC tools are commonly used for SCC contents. The authors created a CNN-based model that efficiently distinguishes between SCC and natural content. The proposed method can efficiently apply the encoding path associated with the content characteristics by applying the CNN model and classifying all 4 × 4 sub-blocks within each 64 × 64 CU, regardless of whether they are a natural content block or screen content block.
Park et al. [64] proposed a fast-encoding method to facilitate an affine motion estimation (AME) process by utilizing features that reflect the statistical characteristics of the CU partition and AME. AME is a newly adopted tool in VVC/H.266, which significantly improves the exit coding performance for non-translational motions in the content. One issue with AME is that it requires numerous computations to achieve a more accurate motion prediction. The authors proposed a method that uses statistical features to skip redundant AME processes by utilizing the determined inter-mode and RD costs available in the conventional motion estimation process.
In [65], Zhang et al., proposed a fast geometric prediction merge mode decision algorithm for VVC/H.266 based on the CU gradient by comparing the mean value of the gradient in four directions to determine whether the GPM can be terminated early. The GPM is a newly adopted tool in VVC/H.266 that can handle moving objects that are not rectangular in shape. The GPM also necessitates numerous computations to find the best combination of two triangular shapes in the inter-coding. The authors developed a method that uses the Sobel operator template and checked the promising directions for use of the GPM, thereby determining the promising object texture for GPM usage.
In [66], Guan et al., proposed a fast AME method based on spatial texture features and the time correlation by calculating the texture and Prewitt gradient and exploiting the best prediction mode of the current block. The enormous computational complexity of AME has motivated the development of the method proposed in this paper, as described in [1]. The proposed method calculates the texture complexity by using the histogram of the block and the texture boundary with the Prewitt operator. The generated texture complexity was used to terminate the AME early, and the accumulated information between the parent and child CUs was used to skip the inter-prediction mode evaluation.
In [67], Fu et al., proposed a fast two-stage method that uses spatial coding statistics and primary transform information to terminate the RDO process of MTS by exploiting the correlations between the RD cost of the primary transform and the RD costs of the child CUs. MTS is a newly adopted tool in VVC/H.266 that, in addition to DCT2, supports the DST7 and DCT8 kernels. By checking four additional transform modes in the core transform, MTS significantly improves the coding performance while consuming a large amount of computational complexity. To reduce the number of MTS computations, the authors presented a procedure that efficiently skips the RDO computations of the MTS mode evaluation if the sum of the RDO costs of the child CUs is greater than the sum of the RDO costs of the parent. In addition, the method employs early termination of the MTS.
In [68], Choi et al., proposed a low-complexity intra-coding scheme applying the downsampling- and upsampling-based fast method by changing the original video size for encoding/decoding and recovering the quality of the reconstructed video using the CNN-based super-resolution network. Additionally, the authors investigated the intra-coding tools in VVC/H.266 to further reduce the encoding complexity by introducing an optimal tool combination under the proposed scheme, the results of which showed significant coding savings with little coding gain.
Table 3 summarizes fast tool-based methods. Such methods can reduce the total encoding time by 4–69%, as shown in the table. Because each tool contributes a different amount to the overall encoding time, the reduction in the encoding time varies depending on the tools used. For example, the AI scenario consumes a significant amount of encoding time for intra-mode selection, and thus the reduction in complexity in intra-mode has a greater impact than other tools. However, the reduction in [62] was relatively small because ISP accounted for only a minor fraction of the total encoding time. The results of [68] are an interesting aspect of the table. Under an AI scenario, the advanced method using upsampling/downsampling with the tool selection demonstrated a relatively high reduction in the encoding time, and even a coding gain of approximately 4.6%.

3.5. Platform Dependent Low-Complexity Methods

Efforts have been made to develop a low-complexity method for VVC/H.266 implementation. The transform module was primarily targeted for hardware implementation. The transform module is the most complex module used in a decoder implementation. In particular, in VVC/H.266, the use of MTS causes the decoder to become more complex by adding DST-VII and DCT-VIII with an increased transform size of up to 64 × 64.
Kammoun et al. [69] proposed a forward inverse 2D hardware implementation of an approximate transform core for VVC/H.266 using low-cost adjustment stages on a DCT-II variant to approximate the DST-VII and DCT-VIII transform types. In this study, the authors also proposed a low-complexity-based hardware implementation of the approximate VVC/H.266 transform process.
Hamidouche et al. [70] proposed a low-complexity-based multiple-transform selection module for VVC/H.266 hardware implementation. The authors exploited the approximation of the DST-VII and DCT-VIII transforms to reduce the hardware complexity and memory requirements and thereby implement the VVC/H.266 MTS, particularly on hardware chips with reduced logic and memory resources.
Additional effort has been made to investigate low-complexity VVC/H.266 implementations in software (SW) implementations. Open-source SW-based codecs are widely used in the industry. Looking at the history of standard codecs, it is clear that the timely availability of SW-based codecs has influenced the success of standard codecs; therefore, VVC/H.266 SW-based codec development is regarded as extremely important.
In [71], Wieckowski et al., introduced VVenC as an open-source VVC/H.266 SW encoder implementation. The authors claimed that the open-source SW codec VVenC with the fastest configuration runs over 140 times faster than VTM while providing at least a 10% bitrate reduction compared to HM.
In [72], Wieckowski presented fast partitioning strategies for VVC/H.266 and their implementation in an open-source optimized SW encoder (e.g., VVenC). The authors proposed the following partitioning strategies: skip condition, skip with sub-split skip, split cost prediction adaptation, test ternary split parallel to a better binary split, and inverse split order in low-depth configurations.
In [73], Brandenburg et al., presented Pareto-optimized coding configurations for VVenC. The authors approximated the Pareto set of construct spaces during the iterative process, including search spaces that extend the encoding tool and search spaces that organize fast algorithms. Using pre-grouping tools and options, the proposed method can adaptively apply coding tools and partitions to targeted complexities and ordered sets.
Table 4 summarizes the platform-dependent low-complexity methods mentioned in this section. As shown in the table, the performance varies depending on the target platform. The number of operations and size of the memory and/or area are the most important measurement factors for hardware implementation. Accordingly, the methods in [69,70] reported performance improvements with the coding loss. The methods in [71,72,73] are aimed at producing an SW codec that can be used directly in the industry. It is critical to provide an encoding/decoding that can be applied in real time. For this purpose, the authors developed an open-source SW codec and presented an increase in speed.

4. Conclusions

In this study, we reviewed the VVC/H.266 standard against previous video coding standards and analyzed and identified key challenges in video encoding based on VVC/H.266 coding. Furthermore, we surveyed and presented recent technical advances in fast and low-complexity VVC/H.266, taking key technical areas into account. Although the latest video coding standard, VVC/H.266, achieved an improvement in coding performance of approximately 50% compared to HEVC/H.265 by incorporating new advanced technologies with flexible partitioning, the increased complexity of the encoding must be overcome before it can be made available on the market. To address the issue of VVC/H.266, significant effort has been made in developing methods for reducing the encoding/decoding of VVC/H.266. One of the attempted approaches is investigating the early CU depth and split mode decision methods during the partitioning process, which requires the majority of the VVC/H.266 encoding time. Texture analysis, statistical analysis, syntax correlation, and machine learning-based classification were used to estimate the CU depth and split mode. Some methods concentrate on increasing the speed of individual coding tools, whereas others have used practical hardware and software implementations. In conclusion, studies on fast and low-complexity VVC/H.266 algorithms are important and will lead to a promising direction for the success of this standard in academic and industrial communities.

Funding

This work was supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (IITP-2021-0-02067) and the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (NRF-2021R1F1A1060816).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

List of abbreviations used in this work.
AcronymDescriptionAcronymDescription
AIAll intra LBLow delay with B-slices
AMEAffine motion estimationLGBMLight gradient boosting machine
ASMAngular second moment LPLow delay with P-slices
ASSDAverage sum of the square difference MPEGMoving picture experts group
AVCAdvance video coding MRLMultiple reference line
BDBRBjøntegaard delta bitrates MTSMultiple transform selection
BTBinary tree MTTMulti-type tree
CNNConvolutional neural network PLTPalette mode
CTCCommon test condition QPQuantization parameter
CTUCoding tree unit QTQuadtree
CUCoding unit RARandom access
DBFDeblocking filter RDRate distortion
GPMGeometric partitioning modeRDORate distortion optimization
HBTHorizontal binary tree RFCRandom forest classifier
HEVCHigh-efficiency video coding RMDRough mode decision
HTTHorizontal ternary tree SCCScreen content coding
HVSHuman visual system SWSoftware
IBCIntra-block copy TTTri-tree
ISPIntra sub-partitionUHDUltra-high-definition
JCT-VCJoint collaborative team on video coding VBTVertical binary tree
JNDJust noticeable differenceVCEGVideo coding experts group
JRMDRough mode decision-based cost VTTVertical ternary tree
JVETJoint video exploration teamVVCVersatile video coding

References

  1. MARKETSANDMARKETS. Available online: https://www.marketsandmarkets.com/Market-Reports/augmented-reality-virtual-reality-market-1185.html (accessed on 1 September 2022).
  2. CISCO. Available online: http://www.cisco.com/c/en/us/solutions/collateral/service-provider/visual-networking-index-vni/complete-white-paper-c11-481360.html (accessed on 1 September 2022).
  3. ISO/IEC 11172-2; Information Technology—Coding of Moving Pictures and Associated Audio for Digital Storage Media at up to About 1,5Mbit/s—Part 2: Video. ISO/IEC JTC 1: Geneva, Switzerland, 1993.
  4. Recommendation ITU-T H.262 and ISO/IEC 13818-2; Information Technology—Generic Coding of Moving Pictures and Associated Audio Information: Video. ITU-T: Geneva, Switzerland; ISO/IEC JTC 1: Geneva, Switzerland, 1995.
  5. Recommendation ITU-T H.264 and ISO/IEC 14496-10 (AVC); Advanced Video Coding for Generic Audio-Visual Services. ITU-T: Geneva, Switzerland; ISO/IEC JTC 1: Geneva, Switzerland, 2003.
  6. STATISTA. Available online: https://www.statista.com/statistics/710673/worldwide-video-codecs-containers-share-online/ (accessed on 1 September 2022).
  7. Recommendation ITU-T H.265 and ISO/IEC 23008-2 (HEVC); High Efficiency Video Coding. ITU-T: Geneva, Switzerland; ISO/IEC JTC 1: Geneva, Switzerland, 2013.
  8. Ronan, P.; Eric, T.; Mickaël, R. Hybrid broadband/broadcast ATSC 3.0 SHVC distribution chain. In Proceedings of the 2018 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), Valencia, Spain, 6–8 June 2018; pp. 1–5. [Google Scholar]
  9. Haskell, B.G.; Puri, A.; Netravali, A.N. Digital Video: An Introduction to MPEG-2; Springer Science & Business Media: Berlin, Germany, 1996. [Google Scholar]
  10. Wiegand, T.; Sullivan, G.J.; Bjontegaard, G.; Luthra, A. Overview of the H. 264/AVC video coding standard. IEEE Trans. Circuits Syst. Video Technol. 2003, 13, 560–576. [Google Scholar] [CrossRef] [Green Version]
  11. Sullivan, G.J.; Ohm, J.R.; Han, W.J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  12. Bross, B.; Chen, J.; Ohm, J.R.; Sullivan, G.J.; Wang, Y.K. Developments in international video coding standardization after avc, with an overview of versatile video coding (vvc). Proc. IEEE 2021, 109, 1463–1493. [Google Scholar] [CrossRef]
  13. Recommendation ITU-T H.266 and ISO/IEC 23090-3 (VVC); Versatile Video Coding. ITU-T: Geneva, Switzerland; ISO/IEC JTC 1: Geneva, Switzerland, 2020.
  14. An, J.; Huang, H.; Zhang, K.; Huang, Y.-W.; Lei, S. Quadtree Plus Binary Tree Structure Integration with JEM Tools; doc. JVET-B0023; Joint Video Exploration Team: Geneva, Switzerland, 2016. [Google Scholar]
  15. Li, X.; Chuang, H.-C.; Chen, J.; Karczewicz, M.; Zhang, L.; Zhao, X.; Said, A. Multi-Type-Tree; doc. JVET-D0117; Joint Video Exploration Team: Geneva, Switzerland, 2016. [Google Scholar]
  16. Zhao, L.; Zhao, X.; Li, X.; Liu, S. CE3-Related: Unification of Angular Intra Prediction for Square and Non-Square Blocks; doc. JVET-L0279; Joint Video Exploration Team: Geneva, Switzerland, 2018. [Google Scholar]
  17. Van der Auwera, G.; Heo, J.; Filippov, A. CE3: Summary Report on Intra Prediction and Mode Coding; doc. JVET-L0023; Joint Video Exploration Team: Geneva, Switzerland, 2018. [Google Scholar]
  18. Bross, B.; Keydel, P.; Schwarz, H.; Marpe, D.; Wiegand, T.; Zhao, L.; Zhao, X.; Li, X.; Liu, S.; Chang, Y.-J.; et al. CE3: Multiple Reference Line Intra Prediction (Test 1.1.1, 1.1.2, 1.1.3 and 1.1.4); doc. JVET-L0283; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  19. Helle, P.; Pfaff, J.; Schäfer, M.; Rischke, R.; Schwarz, H.; Marpe, D.; Wiegand, T. Intra picture prediction for video coding with neural networks. In Proceedings of the 2019 Data Compression Conference (DCC), Snowbird, UT, USA, 26–29 March 2019; pp. 448–457. [Google Scholar]
  20. Li, J.; Wang, M.; Zhang, L.; Zhang, K.; Wang, S.; Wang, S.; Gao, W. Sub-sampled cross-component prediction for chroma component coding. In Proceedings of the 2020 Data Compression Conference (DCC), Snowbird, UT, USA, 24–27 March 2020; pp. 203–212. [Google Scholar]
  21. De-Luxán-Hernández, S.; George, V.; Ma, J.; Nguyen, T.; Schwarz, H.; Marpe, D.; Wiegand, T. An intra subpartition coding mode for VVC. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1203–1207. [Google Scholar]
  22. Zhang, L.; Zhang, K.; Liu, H.; Wang, Y.; Zhao, P.; Hong, D. CE4: History-Based Motion Vector Prediction (Test 4.4.7); doc. JVET-L0266; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  23. Jeong, S.; Park, M.W.; Piao, Y.; Park, M.; Choi, K. CE4: Ultimate Motion Vector Expression (Test 4.5.4); doc. JVET-L0054; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  24. Chen, H.; Yang, H.; Chen, J. Symmetrical Mode for Biprediction, Joint Video Experts Team; doc. JVET-J0063; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  25. Chen, J.; Chien, W.-J.; Hu, N.; Seregin, V.; Karczewicz, M.; Li, X. Enhanced Motion Vector Difference Coding; doc. JVET-D0123; Joint Video Exploration Team: Geneva, Switzerland, 2016. [Google Scholar]
  26. Gao, H.; Esenlik, S.; Alshina, E.; Steinbach, E. Geometric partitioning mode in versatile video coding: Algorithm review and analysis. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 3603–3617. [Google Scholar] [CrossRef]
  27. Su, Y.-C.; Chen, C.-Y.; Huang, Y.-W.; Lei, S.-M.; He, Y.; Luo, J.; Xiu, X.; Ye, Y. CE4-Related: Generalized Bi-Prediction Improvements Combined from JVET-L0197 and JVET-L0296; doc. JVET-L0646; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  28. Chiang, M.-S.; Hsu, C.-W.; Huang, Y.-W.; Lei, S.-M. CE10.1.1: Multi-hypothesis Prediction for Improving AMVP Mode, Skip or Merge Mode, and Intra Mode; doc. JVET-L0100; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  29. Li, L.; Li, H.; Liu, D.; Li, Z.; Yang, H.; Lin, S.; Wu, F. An efficient four-parameter affine motion model for video coding. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 1934–1948. [Google Scholar] [CrossRef] [Green Version]
  30. Chen, H.; Yang, H.; Chen, J. CE4: Separate List for Sub-Block Merge Candidates (Test 4.2.8); doc. JVET-L0369; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  31. Sethuraman, S. CE9: Results of DMVR Related Tests CE9.2.1 and CE9.2.2; JVET-M0147; Joint Video Experts Team: Geneva, Switzerland, 2019. [Google Scholar]
  32. Alshin, A.; Elshina, E. Bi-directional optical flow for future video codec. In Proceedings of the 2016 Data Compression Conference (DCC), Snowbird, UT, USA, 30 March–1 April 2016. [Google Scholar]
  33. He, Y.; Luo, J. CE4-2.1: Prediction Refinement with Optical Flow for Affine Mode; doc. JVET-O0070; Joint Video Experts Team: Geneva, Switzerland, 2019. [Google Scholar]
  34. Choi, K.; Piao, Y.; Kim, C. CE6: AMT with Reduced Transform Types (Test 1.5); doc. JVET-K0171; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  35. Zhao, Y.; Yang, H.; Chen, J. CE6: Spatially Varying Transform (Test 6.1.12.1); doc. JVET-K0139; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  36. Koo, M.; Salehifar, M.; Lim, J.; Kim, S.-H. Low frequency nonseparable transform (LFNST). In Proceedings of the 2019 Picture Coding Symposium (PCS), Ningbo, China, 12–15 November 2019. [Google Scholar]
  37. Schwarz, H.; Nguyen, T.; Marpe, D.; Wiegand, T. CE7: Transform Coefficient Coding and Dependent Quantization (Tests 7.1.2, 7.2.1); doc. JVET-K0071; Joint Video Experts Team: Geneva, Switzerland, 2018. [Google Scholar]
  38. Karczewicz, M.; Hu, N.; Taquet, J.; Chen, C.; Misra, K.; Andersson, K.; Yin, P.; Lu, T.; François, E.; Chen, J. VVC In-Loop Filters. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3907–3925. [Google Scholar] [CrossRef]
  39. VVC Reference Software. Available online: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftware_VTM/-/tags/ (accessed on 1 September 2022).
  40. Bossen, F.; Boyce, J.; Suehring, K.; Li, X.; Seregin, V. JVET Common Test Conditions and Software Reference Configurations for SDR Video; doc. JVET-N1010; Joint Video Experts Team: Geneva, Switzerland, 2019. [Google Scholar]
  41. HEVC Reference Software. Available online: https://vcgit.hhi.fraunhofer.de/jct-vc/HM/-/tags/ (accessed on 1 September 2022).
  42. Chen, W.; Chen, Y.; Chernyak, R.; Choi, K.; Hashimoto, R.; Huang, Y.; Jang, H.; Liao, R.; Liu, S. JVET AHG Report: Tool Reporting Procedure (AHG13); doc. JVET-T0013; Joint Video Experts Team: Geneva, Switzerland, 2020. [Google Scholar]
  43. Bjøntegaard, G. Improvement of BD-PSNR Model; doc. VCEG-AI11; ITU-T SG16/Q6: Geneva, Switzerland, 2008. [Google Scholar]
  44. Park, S.-H.; Kang, J.-W. Context-based ternary tree decision method in versatile video coding for fast intra coding. IEEE Access 2019, 7, 172597–172605. [Google Scholar] [CrossRef]
  45. Park, S.-H.; Kang, J. Kang. Fast multi-type tree partitioning for versatile video coding using a lightweight neural network. IEEE Trans. Multimed. 2020, 23, 4388–4399. [Google Scholar] [CrossRef]
  46. Zhao, J.; Cui, T.; Zhang, Q. Fast CU partition decision strategy based on human visual system perceptual quality. IEEE Access 2021, 9, 123635–123647. [Google Scholar] [CrossRef]
  47. Zhang, Q.; Guo, R.; Jiang, B.; Su, R. Fast CU decision-making algorithm based on DenseNet network for VVC. IEEE Access 2021, 9, 119289–119297. [Google Scholar] [CrossRef]
  48. Saldanha, M.; Sanchez, G.; Marcon, C.; Agostini, L. Configurable Fast Block Partitioning for VVC Intra Coding Using Light Gradient Boosting Machine. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 3947–3960. [Google Scholar] [CrossRef]
  49. Zhang, Q.; Wang, Y.; Huang, L.; Jiang, B. Fast CU partition and intra mode decision method for H.266/VVC. IEEE Access 2020, 8, 117539–117550. [Google Scholar] [CrossRef]
  50. Zhang, Q.; Zhao, Y.; Jiang, B.; Huang, L.; Wei, T. Fast CU partition decision method based on texture characteristics for H.266/VVC. IEEE Access 2020, 8, 203516–203524. [Google Scholar] [CrossRef]
  51. Zhang, Q.; Zhao, Y.; Jiang, B.; Wu, Q. Fast CU Partition Decision Method Based on Bayes and Improved De-Blocking Filter for H.266/VVC. IEEE Access 2021, 9, 70382–70391. [Google Scholar] [CrossRef]
  52. Fan, Y.; Chen, J.; Sun, H.; Katto, J.; Jing, M. A fast QTMT partition decision strategy for VVC intra prediction. IEEE Access 2020, 8, 107900–107911. [Google Scholar] [CrossRef]
  53. Yang, H.; Shen, L.; Dong, X.; Ding, Q.; An, P.; Jiang, G. Low complexity CTU partition structure decision and fast intra mode decision for versatile video coding. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1668–1682. [Google Scholar] [CrossRef]
  54. Tang, N.; Cao, J.; Liang, F.; Wang, J.; Liu, H.; Wang, X.; Du, X. Fast CTU partition decision algorithm for VVC intra and inter coding. In Proceedings of the IEEE Asia Pacific Conference on Circuits and Systems (APCCAS), Bangkok, Thailand, 11–14 November 2019; pp. 361–364. [Google Scholar]
  55. Liu, Z.; Qian, H.; Zhang, M. A Fast Multi-tree Partition Algorithm Based on Spatial-temporal Correlation for VVC. In Proceedings of the 2022 Data Compression Conference (DCC), Snowbird, UT, USA, 22–25 March 2022. [Google Scholar]
  56. Li, T.; Xu, M.; Tang, R.; Chen, Y.; Xing, Q. DeepQTMT: A deep learning approach for fast QTMT-based CU partition of intra-mode VVC. IEEE Trans. Image Process 2021, 30, 5377–5390. [Google Scholar] [CrossRef]
  57. Chen, M.; Lee, C.; Tsai, Y.; Yang, C.; Yeh, C.; Kau, L.; Chang, C. A fast QTMT partition decision strategy for VVC intra prediction. IEEE Access 2022, 10, 42141–42150. [Google Scholar] [CrossRef]
  58. Yeo, W.; Kim, B. CNN-based Fast Split Mode Decision Algorithm for Versatile Video Coding (VVC) Inter Prediction. J. Multimed. Inf. Syst. 2021, 8, 147–158. [Google Scholar] [CrossRef]
  59. Pan, Z.; Zhang, P.; Peng, B.; Ling, N.; Lei, J. A CNN-based fast inter coding method for VVC. IEEE Signal Process. Lett. 2021, 28, 1260–1264. [Google Scholar] [CrossRef]
  60. Dong, X.; Shen, L.; Yu, M.; Yang, H. Fast intra mode decision algorithm for versatile video coding. IEEE Trans. Multimed. 2021, 24, 400–414. [Google Scholar] [CrossRef]
  61. Tan, E.; Aramvith, S.; Onoye, T. Low complexity mode selection for H. 266/VVC intra coding. ICT Express 2021, 8, 83–90. [Google Scholar] [CrossRef]
  62. Park, J.; Kim, B.; Jeon, B. Fast VVC Intra Subpartition based on Position of Reference Pixels. In Proceedings of the 2022 International Conference on Electronics, Information, and Communication (ICEIC), Jeju, Republic of Korea, 6–9 February 2022. [Google Scholar]
  63. Tsang, S.H.; Kwong, N.W.; Chan, Y.L. FastSCCNet: Fast Mode Decision in VVC Screen Content Coding via Fully Convolutional Network. In Proceedings of the 2020 IEEE International Conference on Visual Communications and Image Processing (VCIP), Macau, China, 4 December 2020; pp. 177–180. [Google Scholar]
  64. Park, S.-H.; Kang, J.-W. Fast affine motion estimation for versatile video coding (VVC) encoding. IEEE Access 2019, 7, 158075–158084. [Google Scholar] [CrossRef]
  65. Zhang, M.; Deng, S.; Liu, Z. A fast geometric prediction merge mode decision algorithm based on CU gradient for VVC. In Proceedings of the 2022 Data Compression Conference (DCC), Snowbird, UT, USA, 22–25 March 2022. [Google Scholar]
  66. Guan, X.; Sun, X. VVC Fast ME Algorithm Based on Spatial Texture Features and Time Correlation. In Proceedings of the 2021 International Conference on Digital Society and Intelligent Systems (DSInS), Chengdu, China, 3–4 December 2021. [Google Scholar]
  67. Fu, T.; Zhang, H.; Mu, F.; Chen, H. Two-stage fast multiple transform selection algorithm for VVC intra coding. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 61–66. [Google Scholar]
  68. Choi, K.; Le, T.; Choi, Y.; Lee, J. Low-Complexity Intra Coding in Versatile Video Coding. IEEE Trans. Consum. Electron. 2022, 68, 119–126. [Google Scholar] [CrossRef]
  69. Kammoun, A.; Hamidouche, W.; Philippe, P.; Déforges, O.; Belghith, F.; Masmoudi, N.; Jean-François, N. Forward-inverse 2D hardware implementation of approximate transform core for the VVC standard. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4340–4354. [Google Scholar] [CrossRef]
  70. Hamidouche, W.; Philippe, P.; Fezza, S.; Haddou, M.; Pescador, F.; Menard, D. Hardware-Friendly Multiple Transform Selection Module for the VVC Standard. IEEE Trans. Consum. Electron. 2022, 68, 96–106. [Google Scholar] [CrossRef]
  71. Wieckowski, A.; Brandenburg, J.; Hinz, T.; Bartnik, C.; George, V.; Hege, G.; Helmrich, C.; Henkel, A.; Lehmann, C.; Stoffers, C.; et al. VVenc: An open and optimized VVC encoder implementation. In Proceedings of the 2021 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Shenzhen, China, 5–9 July 2021. [Google Scholar]
  72. Wieckowski, A.; Bross, B.; Marpe, D. Fast partitioning strategies for VVC and their implementation in an Open Optimized Encoder. In Proceedings of the 2021 Picture Coding Symposium (PCS), Bristol, UK, 29 June–2 July 2021. [Google Scholar]
  73. Brandenburg, J.; Wieckowski, A.; Henkel, A.; Bross, B.; Marpe, D. Pareto-Optimized Coding Configurations for VVenC, a Fast and Efficient VVC Encoder. In Proceedings of the 2021 IEEE 23rd International Workshop on Multimedia Signal Processing (MMSP), Tampere, Finland, 6–8 October 2021. [Google Scholar]
Figure 1. Use case of VVC standard.
Figure 1. Use case of VVC standard.
Sensors 22 08990 g001
Figure 2. VVC/H.266 performance compared to HEVC/H.265. (a) VVC/H.266 coding performance over HEVC/H.265, and (b) VVC/H.266 complexity compared to HEVC/H.265.
Figure 2. VVC/H.266 performance compared to HEVC/H.265. (a) VVC/H.266 coding performance over HEVC/H.265, and (b) VVC/H.266 complexity compared to HEVC/H.265.
Sensors 22 08990 g002
Figure 3. Classification of research into fast VVC.
Figure 3. Classification of research into fast VVC.
Sensors 22 08990 g003
Figure 4. Block partitioning: (a) QT in HEVC/H.265 and (b) QT/BT/TT in VVC/H.266.
Figure 4. Block partitioning: (a) QT in HEVC/H.265 and (b) QT/BT/TT in VVC/H.266.
Sensors 22 08990 g004
Figure 5. Flowchart of VVC/H.266 block partitioning.
Figure 5. Flowchart of VVC/H.266 block partitioning.
Sensors 22 08990 g005
Table 1. Summary of fast methods on early split mode decision.
Table 1. Summary of fast methods on early split mode decision.
PaperTech AreaKey FeatureAnchorScenarioT (%)BDBR (%)
[44]Intra partition, Fast split mode decisionBayesian probability approach, Adaptive TT skipping methodVTM4.0AI−341.02
[45]Intra partition, Fast split mode decisionCNN model, Adaptive TT skipping methodVTM4.0AI−270.44
[46]Intra partition, Fast split mode decisionJND model, Adaptive split mode skipping methodVTM7.0AI−480.79
[47]Intra partition, Fast split mode decisionCNN model, Split mode estimation VTM10.0AI−461.86
[48]Intra partition, Fast split mode decisionCNN model, Split mode estimation VTM10.0AI−541.42
Table 2. Summary of fast methods on early CU depth decision.
Table 2. Summary of fast methods on early CU depth decision.
PaperTech AreaKey FeatureAnchorScenarioT (%)BDBR (%)
[49]Intra partition, Fast depth decision, Fast split mode decisionForest classifier model, Canny operator-based texture analysis VTM4.0AI−540.93
[50]Intra partition, Fast depth decision, Deblocking filterJRMD and intra-mode analysis, SAD-based texture analysisVTM7.0AI−48.580.91
[51]Intra partition, Fast depth decision, Fast split mode decisionJRMD-based depth analysis, DBF texture information analysisVTM11.0AI−56.081.3
[52]Intra partition, Fast depth decision, Fast split mode decision, Intra-mode selectionSAD and Sobel operator-based texture analysisVTM7.0AI−49.271.63
[53]Intra partition, Inter partition, Fast depth decision, Fast split mode decisionTexture information analysis, Trained model, Gradient descent-based search VTM2.0AI−621.93
[54]Inter partition, Fast depth decision, Fast split mode decisionCanny operator-based texture analysis, Temporal correlation analysisVTM4.0AIRA−36−310.711.34
[55]Intra partition, Fast depth decisionTemporal correlation analysisVTM11.2RA−221.34
[56]Intra partition, Fast depth decisionCNN model, Split mode, and depth estimation VTM7.0AI−461.32
[57]Inter partition, Inter-mode decision, Fast depth decision, Fast split mode decisionForest classifier model, Human visual system analysisVTM7.0AI−411.14
[58]Inter partition, Inter-mode decision, Fast depth decision, Fast split mode decisionCNN model, Split mode, and depth estimation VTM11.0RA−121.01
[59]Intra partition, Fast depth decision, Fast split mode decisionCNN model, Split mode, and depth estimation VTM6.0RA−313.18
Table 3. Summary of fast tool-based methods.
Table 3. Summary of fast tool-based methods.
PaperTech AreaKey FeatureAnchorScenarioTS (%)BDBR (%)
[60]Intra-prediction, Fast depth decisionLearning-based classifier, Intra-prediction estimationVTM10.0AI−530.93
[61]Intra-modeSATD-based intra-mode estimationVTM5.0AI−210.88
[62]Intra-prediction, ISPISP and MRL analysisVTM14.0AI−40.04
[63]Intra-prediction, IBC, PLTCNN model, Local block analysisVTM9.2AI−302.42
[64]Inter-prediction, AMEStatistical analysisVTM3.0RA−370.1
[65]Inter-prediction, GPMSobel operator-based analysis, Direction analysisVTM8.0RA−140.14
[66]Inter-prediction, AMEPrewitt operator-based analysis, Histogram analysisVTM11.0RA−15.50.55
[67]Transform, MTSDCT cost analysisVTM3.0AI−230.16
[68]FrameworkDown/upsampling, Tool on/off analysisVTM12.0AI−69−4.6
Table 4. Platform-dependent low-complexity methods.
Table 4. Platform-dependent low-complexity methods.
PaperTech AreaKey FeatureAnchorScenarioPerformanceBDBR (%)
[69]Transform, Hardware implementationLow-cost DCT-II implementation, Approximate DST-VII, DCT-VIIIVTM3.0AI12% of Alms, 22% of registers, and 30% of DSP blocks0.15
[70]Transform, Hardware implementationLow-cost DCT-II implementation, Approximate DST-VII, DCT-VIIIVTM3.0AIRA5.37%, 68%, 84%, and 92% of multiplication savings with respect to transform sizes N = 8, 16, 32, and 640.090.01
[71]Software implementationFive predefined presetting different encoding speed/compression quality offsetsVTM12.0RA30 × faster12
[72]Software implementation, PartitionSplit mode and depth estimationVTM12.0RA42% speedup of encoding1.3
[73]Software implementation, Tool combinationPareto set, Pre-grouping tools and optionsHM16.22RA25% speedup of encoding−38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Choi, K. A Study on Fast and Low-Complexity Algorithms for Versatile Video Coding. Sensors 2022, 22, 8990. https://doi.org/10.3390/s22228990

AMA Style

Choi K. A Study on Fast and Low-Complexity Algorithms for Versatile Video Coding. Sensors. 2022; 22(22):8990. https://doi.org/10.3390/s22228990

Chicago/Turabian Style

Choi, Kiho. 2022. "A Study on Fast and Low-Complexity Algorithms for Versatile Video Coding" Sensors 22, no. 22: 8990. https://doi.org/10.3390/s22228990

APA Style

Choi, K. (2022). A Study on Fast and Low-Complexity Algorithms for Versatile Video Coding. Sensors, 22(22), 8990. https://doi.org/10.3390/s22228990

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop