SVM-Based Fast CU Partition Decision Algorithm for VVC Intra Coding

: As a new coding standard, Versatile Video Coding (VVC) introduces the quad-tree plus multi-type tree (QTMT) partition structure, which signiﬁcantly improves coding efﬁciency compared to High-Efﬁciency Video Coding (HEVC). The QTMT partition structure further enhances the ﬂexibil-ity of coding unit (CU) partitioning and improves the efﬁciency of VVC encoding high-resolution video, but introduces an unacceptable coding complexity at the same time. This paper proposes an SVM-based fast CU partition decision algorithm to reduce the coding complexity for VVC. First, the proportion of split modes with different CU sizes is analyzed to explore a method to effectively reduce coding complexity. Then, more reliable correlation features are selected based on the maximum ratio of the standard deviation ( SD ) and the edge point ratio ( EPR ) in sub-CUs. Finally, two SVM models are designed and trained using the selected features to provide guidance for deciding whether to divide and the direction of partition. The simulation results indicate that the proposed algorithm can save 54.05% coding time on average with 1.54% BDBR increase compared with VTM7.0. split mode of CU, we the standard deviation ( SD ) and the edge point ratio ( EPR ) in and the maximum ratio of the SD and the EPR in sub-CUs. Then, features with higher correlation are used as target features for the SVM. Finally, the proposed SVM models are ofﬂine for decision CU partition. The proposed algorithm was integrated into VTM for testing. Experimental results show that the proposed algorithm can save 54.05% coding time and increase 1.54% BDBR compared with the VTM7.0, which indicates that the proposed algorithm achieves a good compromise between complexity and efﬁciency.


Introduction
In recent years, ultra-high-definition (UHD) video has been widely used, and while video quality has been further improved, the large amount of video data has brought inconvenience in transmission and storage. The Joint Video Experts Team (JVET) has developed a new generation of video coding standards, Versatile Video Coding (VVC) [1], which significantly improves coding efficiency compared with High-Efficiency Video Coding (HEVC) [2]. According to established work, comparing the VVC test model (VTM) [3] with the HEVC test model (HM) [4], VTM can improve coding efficiency by approximately 40% while maintaining almost identical quality, but at the cost of 18-fold the coding complexity of HM for intra coding.
VVC introduces new coding techniques, among which the use of a quad-tree plus multi-type tree (QTMT) significantly improves coding efficiency. Different from the quadtree (QT) partition structure in HEVC, VVC introduces a multi-type tree (MTT) split, which allows the partition shape of CU to be not only square but also rectangular. The more flexible partition structure is more effective for encoding high-resolution video. However, the QTMT structure-based coding unit (CU) partitioning process takes up more than 97% of the coding time [5], which implies a very high level of complexity.
In VVC, the default coding tree unit (CTU) size is 128 × 128, and the minimum CU size is 4 × 4. Figure 1a shows the multiple partition modes of QTMT structure, including No Split (NS), QT, vertical BT (BTV), horizontal BT (BTH), vertical TT (TTV), and horizontal TT (TTH). Moreover, some parameters in the configuration file of VTM have limitations on the CU partition. In the default configuration, MaxCUWidth and MaxCUHeight are both 64; therefore, CTU will be divided by QT. MaxBtSize and MaxTtSize are both 32 × 32; therefore, 64 × 64 CU can only be divided by QT. MaxDepth is set to 6, and MaxMttDepth is set to 4. When the depth or the MTT depth of CU reaches the maximum value, the CU is no longer divided. In addition, the combination of different partition modes may lead to the same partition result. To limit the redundant partitions of CU, VVC prohibits BT splitting in the same direction in the middle part after TT splitting. Since MinSize is specified in the QTMT partition structure and asymmetric partitioning is allowed, CU partitioning will be limited by the CU size. TTV split is not allowed when the width of CU is 8, and BTV split is not allowed when it is 4. Similarly, TTH split is not allowed when the height of CU is 8, and BTH split is not allowed when it is 4. Figure 1b shows a CTU divided into multiple CUs with the QTMT partition structure, and different colors represent different partition depths. Same as the method for selecting the best partition result in HEVC, VVC recursively checks the RD cost of all possible CUs by brute-force RDO search and selects the combination with the minimum RD cost as the optimal partition result. VVC introduced the MTT partition structure, resulting in an average of 5781 CUs to be checked to encode a CTU [6], far exceeding the 85 CUs (1 + 4 + 4 2 + 4 3 ) that need to be checked in HEVC. The result shows that the MTT structure has become the most time-consuming part of VVC, with an average time consumption of more than 90% [7]. Therefore, the implementation of fast CU partitioning can reduce the encoding complexity of VVC.
In this paper, we propose an SVM-based fast CU partition decision algorithm to accelerate CU partitioning. First, the standard deviation (SD) and the edge point ratio (EPR) of CU are extracted to calculate the ratio of the SD and the EPR of sub-CUs. Then, the effective features are selected by calculating the F-score value to train the SVM model offline. Finally, the SVM models are integrated into VTM7.0 to predict whether the CU is divided and the direction of partition.
The rest of this paper is summarized as follows. Section 2 reviews recent work on reducing the complexity of VVC and HEVC. Section 3 is a statistical analysis of the proportion of the partition modes of CU, then proposes a fast CU partition algorithm and introduces the features of high relevance. The experimental results of the proposed algorithm are provided and compared with state-of-the-art algorithms in Section 4. Section 5 summarizes this paper.

Related Works
Recently, many researchers are working on methods to accelerate CU partitioning in H.266/VVC and H.265/HEVC.

Methods for HEVC
HEVC improves coding efficiency through a highly flexible quad-tree partition structure, but this structure takes up much encoding time. To save the coding time of HEVC, references [8][9][10][11] used traditional methods to accelerate CU partitioning. The edge complexity of CU and sub-CUs is utilized in [8] to classify CU as divided, undivided and pending. The texture complexity of the CU and the features of neighboring CUs are used in [9] to predict whether the CU is divided. In [10], the authors proposed a method based on the Bayesian decision rule, which is including early partition and pruning. Zhang et al. [11] adopted an adaptive processing method to divide the CU reasonably by adjusting the entropy and texture contrast. Many algorithms to accelerate CU partitioning have been proposed because of the wide application of machine learning and deep learning. In [12,13], the authors utilized CU texture characteristics and adjacent CU depth information to train a decision tree model to predict CU partition. Support Vector Machine (SVM) was used to classify CU partition or non-partition in HEVC in [14][15][16][17][18]. Specifically, Sun et al. [14] improved the Canny edge algorithm to extract the edge points in CU and used the EPR as a feature to train the SVM classifier. In [16], the authors introduced two fast partition models of offline training SVM and online training Bayesian probability, and the feature used to train the model is the gray-level co-occurrence matrix (GLCM). Shen et al. [17] used the RD loss caused by misclassification as a weight to reduce the influence of outliers. The convolutional neural network (CNN) is widely utilized in CU splitting in HEVC. Chen et al. [19] used shallow asymmetric-kernel CNN (AK-CNN) to extract local texture features of CU. In [20], the authors proposed a CNN structure based on depth range decision for CTU classification and depth range limitation. Xu et al. [21] proposed an early termination hierarchical CNN (ETH-CNN) predict the partition result of the CTU.

Methods for VVC
VVC introduces a QTMT partition structure, which further improves flexibility, and increases encoding complexity. Many experts still use traditional methods to accelerate CU partitioning in VVC. Fan et al. [22] proposed a fast CU partition decision based on variance, gradient, and the variance of sub-CUs variance. This method only predicts the partition mode of the 32 × 32 CU, so the effect of reducing complexity when encoding low-resolution video is limited. In [23], the authors proposed an algorithm that uses the RD cost of neighboring CUs to skip TT split. Cui et al. [24] determined the CU partition mode in advance based on the gradients in the four directions of CU and sub-CU. Many approaches [7,[25][26][27][28][29] used machine learning algorithms to accelerate CU partitioning. Thomas Amestoy et al. [25] decided on the partition mode of CU based on the random forest classifier, and the risk interval of the classifier decision is introduced to realize the tradeoff between coding complexity reduction and encoding loss. Zhang et al. [26] utilized GLCM in four directions to train the RFC model to predict CU partitioning modes. Yang et al. [27] extracted gradient information, partial texture information, and context information training decision trees. Wang et al. [28] combined two decision trees to determine CU split mode with a quadtree plus binary tree (QTBT) partition structure. Chen et al. [7] extracted entropy, texture contrast, and Haar wavelet efficiency to train SVM models online to predict the CU partition direction, since this method only determines the partition direction, it has a limited reduction in coding complexity when encoding highresolution video.Wu et al. [29] extracted the difference in variance and gradient of sub-CUs to train two SVM models offline, and the SVM models decide whether to partition the CU and the direction of CU partitioning. CNN can also be applied to VVC. Li et al. [6] proposed a multi-stage exit CNN (MSE-CNN) model with an early exit mechanism to skip the redundancy check during CU partitioning. Zhao et al. [30] classified CU into complex and homogeneous CU and utilized an adaptive CNN structure to classify CU. CNN is utilized to analyze the texture in CU with a size of 64 × 64 in [31], and the partition probability of CU is derived by predicting the vector probability of the 4 × 4 boundaries in this CU. The lightweight neural network (LNN) model was used in [32] to predict whether to skip TT split based on partition depth and shape ratio of CU.

Proposed Algorithm
In this section, we make a statistical analysis of the proportion of different sizes of CU partition type at first. Then, we propose an SVM-based fast CU partition decision algorithm, which predicts whether CU is partitioned and the direction of CU partition, and extracts the SD and the EPR of CU, the ratio of the SD and the EPR of sub-CUs. Finally, we select the effective features by calculating the F-score value to train the SVM model offline.

Statistical Analysis
To reduce the complexity of CU partition, we conduct statistics on the partition modes of CU and we have counted the CU partition modes ratio of six video sequences in total. These video sequences have different resolutions and different video contents, including "FoodMarket4" with global motion in class A1, "DaylightRoad2" with translation motion in class A2, "Cactus" with complex and irregular motion in class B, and "PartyScene" with complex texture in class C, "BlowingBubbles" with smooth textures in low-resolution class D and "Johnny" with monotonous backgrounds and stationary objects in class E. All video sequences were performed on VTM7.0 and encoded using All Intra (AI) configuration under four quantization parameter (QP) value {22, 27, 32, 37}. Figure 2 displays the ratio of the partition type of CUs with sizes ranging from 64 × 64 to 8 × 8. By analyzing the proportion of CU partition modes, we obtain the following information.

1.
NS accounts for more than 20% of the partition mode in all CU sizes, and in 32 × 8, 8 × 32, 16 × 8, and 8 × 16 CUs, the proportion of NS is more than 40%. In addition, as the CU size decreases, the proportion of NS gradually increases. By predicting whether CU is partitioned early, VVC can skip the RDO process of QT and MTT partition, thereby reducing complexity.

2.
MTT split accounts for a relatively high proportion, especially for 32 × 16 and 16 × 32 CUs, which account for more than 60%, and the proportions of horizontal partition and vertical partition are almost the same. By predicting the direction of partition early, VVC can skip the MTT split in other directions, and reduce the cross calculation in the RDO process, thereby reducing coding complexity. 3.
The proportion of QT split is relatively low, the proportion of QT split in 16 × 16 CU does not exceed 10%. Therefore, the coding complexity that can be reduced by predicting QT split is limited, and unnecessary coding performance loss increases.
According to the above analysis, this paper reduces the coding complexity by predicting whether the CU is partitioned and the partition direction.

Fast CU Partition Decision Algorithm
We model the CU partitioning process as a combined multiple binary classification problem to construct two SVM classifiers, namely S vs. NS and Hor vs. Ver. The proposed SVM classifiers are trained for different CU sizes. Specifically, 64 × 64 CU can only be divided by QT split or NS, the complexity of predicting its partitioning mode is limited, and the loss of coding performance may increase. In addition, CUs with a side length of 4 do not need to predict the partition direction, and the RDO process of CUs with small size does not occupy much coding time, the partition mode of such CUs is not predicted in advance. Therefore, we train the classifier models for CUs with sizes of 32 × 32, 32 × 16, 16 × 32, 16 × 16, 32 × 8, and 8 × 32, respectively.
The flowchart of the proposed algorithm is shown in Figure 3. If the CU size matches the predicted CU size, we extract the features of this CU, and use two SVM models to predict whether the CU is partitioned and the partition direction. Otherwise, the anchoring algorithm of VTM7.0 is used to divide the CU. The 32 × 8 CU can only BT split when partitioning in the horizontal direction, so we use the side length of the CU as the judgment condition. If the width of the CU is 8, the TT split in the vertical direction will be skipped, and if the height of the CU is 8, the TT split in the horizontal direction will be skipped. In addition, to limit the redundant partition of CU, VVC prohibits BT split in the same direction in the middle part after TT split, so the middle part after TT split is also divided by the anchor algorithm of VTM7.0. In this paper, we train SVM with the offline training mode. The offline training mode can use sufficient training samples to improve the prediction accuracy and avoid the loss of coding performance. The SVM training process is shown in Figure 4. We extract the features from the xCompressCU function in VTM, which is a recursive function that performs the CTU partitioning process from top to bottom. The labels in the training set represent whether the CU is split and the direction of the partition and these two types of labels are obtained by the code stream analysis tool (DecoderAnalyserApp) in VTM7.0.

Feature Analysis and Selection
The selection of features is crucial for training SVM classifiers, as highly correlated features can help reduce training time and improve prediction performance. By analyzing the results of CU partition, we found that CUs with complex textures are usually further split into smaller CUs for encoding, while CUs with smooth textures tend to NS. Additionally, the difference in texture complexity between adjacent blocks determines whether two blocks belong to the same CU, Therefore, we predict the partition mode of the CU by comparing the texture characteristics of the sub-CUs. Figure 5 shows the relationship between CU partition and sub-CU texture characteristics in VVC. As shown in Figure 5a, if the texture difference between two adjacent sub-CUs is large, CU tends to be divided, and vice versa. As shown in Figure 5b,c, CU tends to be horizontally divided if the difference between the texture of the two sub-CUs divided by BTH split is greater than that of the two sub-CUs divided by BTV split, and vice versa. As shown in Figure 5d, if the difference in the texture of the three sub-CUs divided by TTH split is greater than that of the three sub-CUs divided by TTV split, CU tends to be horizontally partitioned, and vice versa. According to the above analysis, we choose the following characteristics: 1.
The SD reflects the texture complexity of CU by calculating the dispersion between the luminance values in the CU and the luminance mean. A large SD indicates that the CU has a complex texture and tends to be divided into smaller CUs, and a small SD indicates that the CU has a smooth texture and tends not to be divided. The SD is expressed as: where W and H represent the width and height of the CU, respectively, f (i, j) is the luminance value with coordinates (i, j) in CU. f is the luminance mean of CU. f is expressed as: 2.
The EPR reflects the texture complexity of CU by calculating the proportion of edge points in the CU. A high EPR means that the CU has a complex texture. The most common method to extract edges is the Sobel edge detection algorithm. Firstly, the Sobel operators in the four directions are convolved with the luminance matrix of CU to obtain the gradient in each direction. The purpose of using four Sobel operators is to improve the accuracy of computing edges and anti-noise ability and to detect weak edge information accurately. The gradient is expressed as: where A is the luminance matrix of CU. Then, the absolute values of the gradients in four directions are added to obtain the edge matrix (EM). The expression of EM is: If the value in EM exceeds the threshold, it is defined as an edge point. We obtain the EPR by calculating the ratio of the number of edge points to the number of elements in EM. The expression of the EPR is: where ep num is the number of edge points in the CU, and EM num is the number of elements in EM.
We set different edge point thresholds and select the threshold with the highest correlation with the CU partition by calculating the F-score value. By calculating the Fscore for different thresholds, we set 150 and 600 as the thresholds for 8-bit and 10-bit video sequences, respectively. Figure 6 shows the result of the Sobel edge detection algorithm to detect the "BasketballPass" video sequence, and we found that CU with more edge points is divided into smaller CUs.

3.
The gradient ratio (GR) can represent the difference between the horizontal gradient and the vertical gradient of the CU. It is expressed as:

4.
The R q is the maximum difference between features of the sub-CUs divided by QT, and the features include the SD and the EPR, a large R q indicates that CU tends to be divided. It is expressed as: where f tl , f tr , f bl , and f br are the sub-CUs of top-left, top-right, bottom-left, and bottomright, respectively.

5.
The R dirB is the ratio of the difference of features in the horizontal and vertical directions in BT split, and the features include the SD and the EPR. It is expressed as: where f t and f b are the features of the top and bottom sub-CUs divided by BTH, f l and f r are the features of the left and right sub-CUs divided by BTV. 6.
The R dirT is the ratio of the difference of features in the horizontal and vertical directions in TT split, and the features include the SD and the EPR. It is expressed as: where f tt , f mh , and f bt are the feature values of the sub-CUs divided by TTH split, f lt , f mv , and f rt are the feature values of the sub-CUs divided by TTV split. Since CU of size 32 × 8 cannot be TTH split and CU of size 8 × 32 cannot be TTV split, we substitute the difference of sub-CUs divided by BT.
The F-score based on the wrapper method is used to evaluate the correlation of features, and features with high correlation are used to train the SVM model. Given a sample set x i , i ∈ [1, m], each sample contains n features and a label that is divided into positive and negative. The F-score of the jth feature is calculated as: where x j represents the mean of the jth feature, x j (+) and x j (−) are the mean of the jth feature of the samples of the two categories, respectively, P and N are the sample sizes of the two categories, x (+) i,j and x (−) i,j are the jth eigenvalues of the x i . Figure 7 shows the F-score of all features of the SVM classifiers S vs. NS and Hor vs. Ver, and the blue upper diagonal line represents the F-score of each feature of the classifier S vs. NS, and the red lower diagonal line represents the F-score of each feature of the classifier Hor vs. Ver. The F-scores of features the SD, the EPR, R q (SD), and R q (EPR) are higher than other features in the S vs. NS classifier, and the F-scores of features GR, R dirB (SD), R dirB (EPR), R dirT (SD) and R dirT (EPR) are higher than other features in the Hor vs. Ver classifier. We select features for the classifier according to the F-score descending order to avoid dimensional redundancy. We select the feature set of S vs. NS as the SD, the EPR, R q (SD), and R q (EPR). The feature set of Hor vs. Ver is GR, R dirB (SD), R dirB (EPR), R dirT (SD), and R dirT (EPR).

The Principle of SVM Algorithm and Training
SVM are generally used as models to solve binary classification problems. It can process nonlinear classification with low complexity. Given a training set {x i , y i } m i=1 , x i ∈ R n is the input features, y i ∈ {−1, +1} is the label. In this paper, we use the SD, the EPR, R q (SD), and R q (EPR) as features, CU split and not split as labels input to the S vs. NS classifier, and use GR, R dirB (SD), R dirB (EPR), R dirT (SD), R dirT (EPR) as features, horizontal partitioning and vertical partitioning as labels input to the Hor vs. Ver classifier. To solve for the hyperplane w T x + b = 0, turn the problem into: after optimization, Equation (11) is transformed into: where C is the penalty parameter and ξ i is the relaxation variable, and Φ(x i ) represents the new vector after x is mapped to a higher-dimensional space. The Lagrange function is constructed to obtain the following expression: where λ i is the Lagrange multiplier, K(x i , x j ) is the kernel function. In this paper, we use Radial Basis Function (RBF) as the kernel function, which can be applied to nonlinearly separable cases. In order to test the generalization ability of the SVM models, instead of extracting features from the video sequences of the JVET test set [33], we extracted features from the Div2K training set [34]. This training set is developed for super-resolution and had good diversity. It contains 800 images of different content types, covering a wide range of objects, plants, and animals, as well as natural scenery including the dark environments. These images were encoded by VTM7.0 using the AI configuration and four QP values (22,27,32,37). The features of CU are extracted and normalized and then combined with the labels to build a training set to train the SVM models.
We test five video sequences with different resolutions of the JVET test set to verify the accuracy of the SVM models. The video sequences are "Campfire", "BasketballDrive", "BQMall", "BlowingBubbles", and "Johnny". Figure 8 illustrates the prediction accuracy of the two SVM models for different sizes of CU. In the S vs. NS classifier, the prediction accuracy of all sizes of CU exceeded 80%, especially the prediction accuracy of 32 × 32 CU is higher than 95%. In the Hor vs. Ver classifier, the prediction accuracy of rectangular CUs is generally higher than that of square CUs, especially for smaller size CUs, and the prediction accuracy of CU of sizes 32 × 8 and 8 × 32 is higher than 90%.

Experimental Results
In experiments, the proposed algorithm was implemented in VTM7.0. We encoded 22 video sequences in the JVET test set at AI configuration and the default temporary sub-sample ratio at four QP values {22, 27, 32, 37}, and evaluated the performance of the algorithm on these video sequences. These sequences are 3 from class A1, 3 from class A2, 5 from class B, 4 from class C, 4 from class D, and 3 from class E. All experiments are run on a computer with AMD Ryzen 7 5800H CPU @ 3.20 GHz, and the windows operating system. After encoding, we use the time saving rate (TS) of the proposed algorithm compared with the original VTM to measure the complexity reduction, and use the Bjøntegaard delta bit-rate (BDBR) to evaluate the RD performance. TS is defined as: where T VTM7.0 is the total encoding time using the anchor algorithm, and T P is the encoding time using the proposed algorithm. Table 1 lists the experimental results of the proposed algorithm compared with VTM7.0. The improved coding efficiency of the proposed algorithm is relatively stable in all sequences, which can reduce the complexity by 54.05% on average with 1.54% RD performance loss compared with the VTM7.0. In terms of reducing the coding complexity, the proposed method performs well in different resolution test sequences, which indicates that the proposed method has a good generalization ability. For test sequences with smooth textures, such as "DaylightRoad2" and "ParkScene", encoding time is reduced by more than 58%, which indicates that the proposed algorithm has a better effect on texture smoothing sequences. The performance comparison of the proposed method with other fast methods [22,29,30] is shown in Table 2, where the methods proposed in [22,30] and the proposed method experiment in VTM7.0, and the method proposed in [29] experiments in VTM10.0. Since VTM10.0 and VTM7.0 use the same method in the CU partition part, it is reasonable to compare the proposed method with [29]. It has been observed that the proposed method saves an average of 54.30% encoding time with 1.54% BDBR increase. The literature [29] predicts CU partition modes based on SVM, which effectively reduces the coding complexity. Compared with the experimental results of [29], the proposed method increases the coding complexity by 9.37% with 1.31% BDBR decrease. The methods proposed in [22,30] only predict 32 × 32 CU partition modes, which can reduce the coding complexity to a certain extent. Compared with the experimental results of [30], the proposed method reduces the coding complexity by 15.08% with 0.67% BDBR increase. Compared with the experimental results of [22], the proposed method reduces the coding complexity by 5.15%. By analyzing the comparison results, the proposed method can effectively reduce the coding complexity while ensuring the video quality. To reduce coding complexity, the proposed method should take less time. We count the time consumption of extracting features, predicting the CU partition mode and the rest of the encoding process. The proportion of time consumption of different resolution sequences is shown in Figure 9. The time consumed by feature extraction is 0.4% to 3.9%, and prediction takes 4.3-8.6% of the time. In addition, high-resolution video sequences take more time to extract features and predict CU partitions than low-resolution video sequences. In order to assess the RD performance of the proposed algorithm, the video sequences "FourPeople" and "Johnny" are tested. Figure 10 shows the RD curves of the two test sequences at four QPs. It can be seen that the RD curve of the proposed algorithm and the VTM7.0 anchor algorithm almost overlap in the high QP encoding, and the RD curve slightly deviates in low QP encoding, which means that the proposed algorithm and the anchor algorithm have similar RD performance. In order to assess the RD performance of the proposed algorithm, the video sequences "FourPeople" and "Johnny" are tested. Figure 10 shows the RD curves of the two test sequences at four QPs. It can be seen that the RD curve of the proposed algorithm and the VTM7.0 anchor algorithm almost overlap in the high QP encoding, and the RD curve slightly deviates in low QP encoding, which means that the proposed algorithm and the anchor algorithm have similar RD performance.

Conclusions
In this paper, we have proposed an SVM-based fast CU partition decision algorithm to reduce the coding complexity. Based on the statistical analysis of the proportion of the split mode of CU, we extract the standard deviation (SD) and the edge point ratio (EPR) in CU and the maximum ratio of the SD and the EPR in sub-CUs. Then, features with higher correlation are used as target features for the SVM. Finally, the proposed SVM models are trained offline for decision CU partition. The proposed algorithm was integrated into VTM 7.0 for testing. Experimental results show that the proposed algorithm can save 54.05% coding time and increase 1.54% BDBR compared with the VTM7.0, which indicates that the proposed algorithm achieves a good compromise between complexity and efficiency.

Conclusions
In this paper, we have proposed an SVM-based fast CU partition decision algorithm to reduce the coding complexity. Based on the statistical analysis of the proportion of the split mode of CU, we extract the standard deviation (SD) and the edge point ratio (EPR) in CU and the maximum ratio of the SD and the EPR in sub-CUs. Then, features with higher correlation are used as target features for the SVM. Finally, the proposed SVM models are trained offline for decision CU partition. The proposed algorithm was integrated into VTM 7.0 for testing. Experimental results show that the proposed algorithm can save 54.05% coding time and increase 1.54% BDBR compared with the VTM7.0, which indicates that the proposed algorithm achieves a good compromise between complexity and efficiency.

Conflicts of Interest:
The authors declare no conflict of interest.