Spatial Correlation-Based Motion-Vector Prediction for Video-Coding Efficiency Improvement

H.265/HEVC achieves an average bitrate reduction of 50% for fixed video quality compared with the H.264/AVC standard, while computation complexity is significantly increased. The purpose of this work is to improve coding efficiency for the next-generation video-coding standards. Therefore, by developing a novel spatial neighborhood subset, efficient spatial correlation-based motion vector prediction (MVP) with the coding-unit (CU) depth-prediction algorithm is proposed to improve coding efficiency. Firstly, by exploiting the reliability of neighboring candidate motion vectors (MVs), the spatial-candidate MVs are used to determine the optimized MVP for motion-data coding. Secondly, the spatial correlation-based coding-unit depth-prediction is presented to achieve a better trade-off between coding efficiency and computation complexity for interprediction. This approach can satisfy an extreme requirement of high coding efficiency with not-high requirements for real-time processing. The simulation results demonstrate that overall bitrates can be reduced, on average, by 5.35%, up to 9.89% compared with H.265/HEVC reference software in terms of the Bjontegaard Metric.


Introduction
High-efficiency video coding (HEVC), also known as H.265, is the latest video-coding standard that was released in 2013 [1].In H.265/HEVC, the maximal size of the basic coding unit (CU) is 64 × 64, and the search range is a key parameter on search-quality control for motion estimation (ME).Compared with H.264/AVC, H.265/HEVC achieves about 50% bitrate saving, while computation complexity is significantly increased [2].
Before motion estimation in H.265/HEVC, motion-vector prediction (MVP) is introduced to define an accurate search center to save coding bits.The MVP is selected from a motion-vector (MV) candidate list that consist of a motion vector from neighboring units on the left of the current coding unit, a motion vector from the above neighboring units, and the motion vector of those spatially in the same position as the previously encoded frame.One of the MVs in the lists with minimum cost is selected as the final MVP.However, the fixed pattern of the MVP decision process without consideration of the reliability of the surrounding motion vectors makes it have lower estimation accuracy.
Recently, MV coding has been attracting much attention.Previous works have been divided into two categories: (1) based on spatial and temporal MVP candidates, (2) based on the higher-order motion model.The detailed descriptions of these methods are as follows: The main ideas based on spatial and temporal MVP candidate schemes for MV coding have one assumption in common.The motion of neighboring blocks has to be similar [3][4][5][6][7].A framework for a better MV and skip mode was proposed, and the predictors were selected by a rate-distortion criterion in Reference [3].In this method, a simple spatial median is selected by using spatial and temporal redundancies in MV fields.MV coding techniques were proposed to improve coding efficiency in Reference [4], which include a priority-based derivation algorithm for spatial and temporal motion candidates, a surroundings-based candidate list, and a parallel derivation of the candidate list.This method can achieve, on average, 3.1% bitrate saving.W.H. Peng et al. introduced an interframe prediction technique that combines two MVs derived from a template and encoding block for overlapped block-motion compensation [5].Moreover, multihypothesis prediction and motion-merge methods are used to achieve the trade-off between encoding efficiency and complexity, which achieves, on average, about 2% bitrate saving.Encoding-efficiency improvement of H.265/HEVC was proposed in Reference [6], and asymmetric motion partitioning (AMP) was used for interprediction.A new selection algorithm was proposed to improve the accuracy of prediction motion vectors in Reference [7].Furthermore, an adaptive motion search-range algorithm was designed, and bitrate saving was only 0.16% on average.A novel MVP (NMVP) method was presented to improve coding efficiency in Reference [8], but coding complexity was higher.In a conclusion, the spatial and temporal MVP candidates lack precision, and these approaches limitedly improve the performance of MV coding, with higher coding complexity.
The main idea based on the higher-order motion model is that motion can be induced by moving objects and all kinds of camera positions and zoom changes when sequence motions are neither spatially regular nor temporally consistent [9][10][11].Tok et al. describe how new motion-information coding and prediction schemes have been investigated to increase the efficiency of video coding [9,10].Springer et al. present a scheme to perform fast, reliable, and precise rotational-motion estimation (RME) on navigation sequences [11].However, the robustness of these methods is not high.
As a summary, coding-performance improvement is limited with low robustness in previous works.This work, an efficient MVP algorithm is proposed to further improve coding efficiency based on spatial-motion consistency correlation.Furthermore, a CU depth prediction algorithm is presented to reduce computation complexity based on spatial texture complexity correlation.Experiments confirm that the number of bits can be reduced with the proposed method.The proposed overall method can improve coding efficiency for the next-generation video-coding standards, and it is beyond H.265/HEVC.

Motivation for This Work
In H.265/HEVC standards, the input video is divided into a sequence of coding-tree units (CTUs), and the CTU is divided into the coding unit (CU) with a different size.The CU is a square region, each of which may be as large as 64 × 64 or as small as 8 × 8.The prediction unit (PU) is a region defined by partitioning the CU, and PU contains MV information.Current PU sizes for intercoded CUs are 2N × 2N, 2N × N, N × 2N, N × N, 2N × nU, 2N × nD, nL × 2N, and nR × 2N, where N ∈{4, 8, 16, 32}.
There are three interprediction modes: InterMode, SkipMode, and MergeMode [12].For SkipMode and InterMode, an advanced motion-vector-prediction (AMVP) technique is used to generate a motion-vector predictor among an AMVP candidate set including two spatial MVPs and one temporal MVP.For MergeMode, the Merge scheme is used to select a motion-vector predictor among a Merge candidate set containing four spatial MVPs and a temporal MVP.By using rate-distortion-optimization (RDO) processing, the encoder selects a final MVP within the candidate list for InterMode, SkipMode, or MergeMode, and transmits the index of the selected MVP to the decoder.In the case of InterMode, the sum of absolute transform differences (SATD) between the source and prediction samples is used as a distortion factor, and bits for inter_pred_ f lag, re f _idx_lX, mvd_lX, and mvp_idx_lX are set to coded bits.In the case of SkipMode, the prediction residual is not transmitted for SkipMode.The coded bits include skip_ f lag and merge_idx that signals the position of the PU that has the best motion parameters to be used for the current PU.In the case of MergeMode, the SATD between source and prediction samples is used as a distortion factor, and bits for merge_idx are set to coded bits.
As shown in Figure 1, AMVP candidates of the current PU for intercoded CUs include five spatial-motion candidates: left candidate (L), bottom-left candidate (BL), top candidate (T), top-right candidate (TR), top-left candidate (TL); and two temporal candidates (C and H).Firstly, two left spatial candidates are selected, otherwise, the top spatial candidates are checked.Secondly, one temporal candidate is checked.When the selected candidate index is no more than 2, MV(0,0) candidate is added.It is noted that BL can be used when it available, and Figure 2 shows the available BL.Moreover, CU splitting increases the computing complexity with depth 0, 1, 2, and 3.In order to speed up the HEVC encoder, two conditions (Early_SKIP and Early_CU condition) are present, as are heuristics, to reduce the computational complexity of the HEVC.The Early_SKIP condition is that the motion-vector difference (MVD) of InterMode with 2N × 2N is equal to (0, 0), and InterMode with 2N × 2N contains no nonzero transform coefficients.In the Early_SKIP case, PU mode in a current CU is determined as SkipMode at the earliest possible stage.The Early_CU condition is that the best PU mode of the current CU selects the SkipMode.In the Early_CU case, the current CU is not divided into sub-CUs in the subdepth level of the current CU.

Proposed Method
In this section, the video-coding-efficiency improvement algorithm is described.Firstly, in order to generate a more accurate motion-vector predictor, the spatial correlations-based MVP algorithm is presented to improve encoding efficiency.Then, the spatial correlations-based CU depth-prediction algorithm is proposed to reduce computation complexity.It is noted that, in H.265/HEVC, normal mode and merge/skip mode have different methods for MVP.In the proposed approach, the method is common.

Definition of Spatial-Neighborhood Set
Considering video content with strong spatial correlations, the motion-vector predictor of the current PU for intercoded CU can be generated by the surrounding PUs.Moreover, the depth level of the current CU can be predicted from neighboring CUs where there is a similar texture or there are continuous motions.
Different from fixed-pattern AMVP technology, spatial neighborhood set G is composed of all spatial neighborhood CUs.Set G is shown in Figure 3, where CU L , CU TR and CU TL denote the left, top-right, and top-left CU of the current CU, respectively.Set G is defined as On the one hand, the MVs and depth information of G can be used to predict the MVP and depth level of the current CU.CU contains one, two, or four PUs depending on partition mode, and each PU contains MV information.For set G, the surrounding PU directly connected with the current PU is selected in this work.Furthermore, the minimum surrounding PU size is set to 8 × 8, because the MVs of the 4 × 8 and 8 × 4 surrounding PUs are not regular, and the significance of reference is small for MV prediction.On the other hand, computation complexity is high by checking all information.Therefore, a relatively reliable subset should be developed for the MVP and depth prediction.In order to utilize the spatial correlation, subset M is defined as where subset M is contained in set G (M ⊂ G).The basic idea of the spatial-correlation method is to prejudge the MVP and depth of the current CU according to the MVs and depths of adjacent CUs.When subset M is available, the information of M is used to predict the MVP and depth of the current CU.In contrast, when subset M is unavailable, which means that none of spatial neighborhood CUs (CU L , CU TL , CU TR ) exist, the information of G is used to predict the MVP and depth of the current CU.In this work, the spatial correlation-based method consists of two parts: a motion-vectorprediction algorithm and CU depth-prediction algorithm.Firstly, the MVP can be selected by exploiting the spatial correlation of neighboring PUs.When there is motion consistency of the neighboring PUs, a simple MV can be selected as the optimized MV for the current PU.Otherwise, more MVs of neighboring PUs can be checked to select the optimized MV.Secondly, the depth level of the current CU can be predicted by exploiting the spatial correlation of neighboring CUs.When the texture complexity of neighboring CUs tends to simple, the content of texture of the current CU tends to be not complex.On the contrary, the texture complexity of neighboring CUs tends to be complex, and the texture content of the current CU tends to be not simple.

Spatial Correlation-Based Motion-Vector-Prediction Algorithm
The performance of motion estimation highly depends on the MVP [13][14][15][16].If the MVP is close to the calculated MV, the MVD between the MVP and the calculated MV is small, and the MVP is more accurate.However, in the H.265/HEVC standard, a total of seven spatial and temporal MVs are added to the candidate list to predict the MVP.There are two disadvantages to the current AMVP mechanism in H.265/HEVC [17].For one thing, the number of reference MVs is limited.For another, a fixed selecting pattern is not adaptive to selecting reference MVs; therefore, it is not generating an accurate MVP by using the current AMVP mechanism.In order to further improve encoding efficiency, more reference MVs can be added to the candidate list.Owing to spatial and temporal motion consistency, the MVs surrounding the current PU are useful for determining the MVP.However, too many MVP candidates may cause a large number of calculations, so it is necessary to reduce the calculation redundancy of the MVP decision.
Based on the above-mentioned views, the reference MVs of the current PU can be used to search for an accurate MVP.In neighborhood subset M, the MVs of the neighboring CUs are shown in Figure 4, where MV L , MV TR and MV TL indicate the MV candidates in the left, top right, and top left of the current PU, respectively.In H.265/HEVC, the simplified rate-distortion optimization (RDO) method is performed to estimate the motion vector [12].In the RDO process, the rate-distortion cost (RD-cost) function (J cost = D distortion + R bits × λ) is minimized by the encoder, where λ is the Lagrange multiplier, D distortion represents the distortion between the original block and reference block, and R bits represents the number of coding bits.The MVD between the MVP and the calculated MV is also signalled in the R bits .For the different texture of video content, the reliability of candidate MVs can be evaluated by the MVs of spatial neighborhood subset M: {MV L , MV TR , MV TL }.When these three MVs are equal, the MVs of adjacent CUs tend to the same direction.In this case, the reliability of candidate MVs is the highest, and a simple MV can be selected as the final MVP.That is, reference MVs satisfy as Then, MV L is selected as the optimized MVP.Furthermore, when the MV absolute difference of MV TL and MV TR is more than the MV absolute difference of MV TL and MV L , motion consistency in the top of PU is lower than motion consistency to the left of PU.In this case, the reliability of the MVs in the left of PU is higher than the reliability of the MVs in the top of PU.Otherwise, the reliability of the MVs in the top of PU is higher than the reliability of the MVs in the left of PU.Thus, when reference MVs satisfy as the MVP position tends to the left of PU.Then, MV TL and MV L are selected as MVP candidates.
Otherwise, the MVP position tends to the top of PU, and MV TR and MV TL are selected as MVP candidates.
When the MVs of subset M are not available, the reliability of candidate MVs is the lowest in the spatial domain, and it is hard to obtain an accurate MVP by using the fixed AMVP mechanism.In this case, the MVP position may tend to the left of PU, and it is also possible to tend to the top of PU.Therefore, all available MVs of spatial neighborhood set G need to be checked.In order to obtain a more accurate MVP, all surrounding MVs of G can be added to the candidate MVs, and the cost of these MVs is checked to obtain an optimized MVP by comparing one with a method.When all components of G are not available, MV (0, 0) is added to the candidate list, which is the same as that in H.265/HEVC.
It is noted that the encoder codes mvp_lx_ f lag for indicating the number of coded bits for the MVP candidates.Different from the H.265/HEVC standard, in this work, the indicating method for the codec is that mvp_lx_ f lag is designed as a variable-length code, and the length of mvp_lx_ f lag is expressed as L. The relationship between the coded bits of mvp_lx_ f lag and MVP is shown as in Table 1.When M is available, the length of mvp_lx_ f lag satisfies L = 1 bit.However, when M is not available, the maximum value of L with a different PU size is shown in Table 2.
Table 1.Relationship between the bit of mvp_lx_ f lag and motion-vector prediction (MVP).

mvp_lx_ f lag
Condition MVP When set G is selected One of the selected MVP It can be seen from Table 2 that, when the current PU size is equal to 64 × 64, the maximum value of mvp_lx_ f lag can be calculated as follows: (1) if the smaller surrounding PU size is 8 × 8, the number of coded bits that need to index the MVP candidates is log 2 (64/8 + 64/8) = 4. Thus, the length of mvp_lx_ f lag satisfies L = 4 bit.(2) if the smaller surrounding PU size is 16 × 16, the number of coded bits that need to index the MVP candidates is log 2 (64/16 + 64/16) = 3.Thus, the length of mvp_lx_ f lag satisfies L = 3 bit.(3) if the smaller surrounding PU size is 32 × 32, the number of coded bits that need to index the MVP candidates is log 2 (64/32 + 64/32) = 2. Thus, the length of mvp_lx_ f lag satisfies L = 2 bit.(4) If both surrounding PUs are 64 × 64 in size, the number of coded bits that need to index the MVP candidates is log 2 (64/64 + 64/64) = 1.Thus, the length of mvp_lx_ f lag satisfies L = 1 bit.In this case, the maximum-value length of mvp_lx_ f lag satisfies L = 4 bit.Moreover, in order to clearly specify the length of mvp_lx_ f lag, Figure 5 shows the length range of mvp_lx_ f lag with 64 × 64 PU size.For others PU size (32 × 64, 64 × 32, 48 × 64, 64 × 48, 16 × 64, and 64 × 16), the maximum value length of mvp_lx_ f lag satisfies L = 4 bit.Similarly, when the current PU size is equal to 32 × 32, the maximum value of mvp_lx_ f lag satisfies L = 3 bit.When the current PU size is equal to 16 × 16, the maximum value of mvp_lx_ f lag satisfies L= 2 bit.When the current PU size is equal to 8 × 8, and the smaller surrounding PU size is 8 × 8, the number of coded bits that need to index the MVP candidates is log 2 (8/8 + 8/8) = 1.In this case, the length of mvp_lx_ f lag satisfies L = 1 bit.In H.265/HEVC standards, the distribution of the selected spatial-motion candidates is far greater than the distribution of the temporal-motion candidates [4].In this work, the more-available spatial candidates are used to decide the MVP, and the temporal-motion candidates have little overall effect on coding efficiency.Thus, the temporal-motion candidates have been removed in the proposed method.
As per the aforementioned approaches, the spatial correlation-based motion-vector-prediction selection algorithm for interprediction is shown in Algorithm 1. Firstly, the MVP candidate list is established by using the proposed spatial-neighborhood motion vector.After that, the rate-distortionoptimal MVP is generated by executing motion estimation in the MVP candidate list, which is the search center to search for the optimal MV.Motion estimation (ME) is the process of determining a motion vector by using a block-matching algorithm [18], which is regarded as a time-consuming process.There are two advantages of this proposed method: (1) MVP accuracy was improved with the proposed method.Thus, the MVD of the current PU becomes smaller for InterMode.The length of mvp_lx_ f lag is variable.(2) Using the proposed method, the possibility that the MV and MVP of the current PU are consistent increased, and the probability that CU selects MergeMode increased.In the case when MVD is equal to zero, the majority of CUs select MergeMode.Therefore, by using the proposed algorithm, the effect of the proposed method (MVD becoming zero) and the effect of MergeMode overlap.The length of merge_idx from the merge candidate list in MergeMode is fixed, which is the same as the definition in H.265/HEVC standards.As a result, the proposed algorithm can significantly reduce the amount of bits.The main idea of this work is the sacrifice of computational complexity for higher coding efficiency.Thus, more MVs surrounding the current PU are added to the MV candidate list by the proposed method; therefore, most computational cost in this work is to search for an accurate MVP with the RDO process.

Spatial Correlation-Based CU Depth-Prediction Algorithm
The above spatial correlation-based MVP algorithm can significantly improve coding efficiency, while computation complexity is increased by a lot.There are quite a few related works that can reduce computation complexity [13,14,19].However, three important issues are carefully considered to design the conditions.Firstly, arithmetic-complexity reduction is the design motivation.Secondly, the robustness of the design condition is higher.Thirdly, owing to high availability, depth information should be used.In this paper, a spatial correlation-based CU depth-prediction algorithm is presented.In order to evaluate depth-level prediction, several experiments were performed on different conditions with different configurations.In the experiments, the accuracy rate when the predicted depth level was equal to the depth level selected by the original H.265/HEVC test model was verified.
Generally, the texture complexity of image content is directly related to the depth of the image.When the depth range of the CU is higher, the texture complexity of the CU tends to be complex.On the contrary, when CU depth range is lower, CU texture complexity tends to be simple.Based on CU depth, CU texture complexity (TC) is classified into simple or complex as where TC represents the texture complexity of a CU.D is the maximal depth of the CU in the motion-estimation processing, and default value D is set to 3 in H.265/HEVC reference software.In this verification, the test conditions have to be carefully designed.It is clear that when the TC of the left neighboring CU L , the top-right neighboring CU TR , and the top-left neighboring CU TL are simple, the TC of the current CU tends to be not complex.On the contrary, when the TC of the left neighboring CU L , the top-right neighboring CU TR , and the top-left neighboring CU TL are complex, the TC of the current CU tends to be not simple.Thus, based on the above conclusions, two conditions (C1 and C2) are proposed as follows: where D L , D TR , D TL , and D * are the maximal depth of CU L ,CU TR , CU TL , and the current CU, respectively.
In order to verify the accuracy of the two conditions, accuracy rate AR is defined as while n 1 represents the number of correct-matching test cases by using the depth of the neighboring CUs to predict the depth of the current CU, and N represents the total number of test cases.In this work, four typical sequences (PeopleOnStreet, BasketballDrill, BQSquare, Vidyo1) were applied to test with low-delay (LD) and random-access (RA) profiles.From the results of Table 3, the rates of Condition 1 and Condition 2 are about 99% and 93%, respectively.That is, the depth of CU L , CU TR , and CU TL has strong spatial correlation with the depth of the current CU.Thus, it is high availability to predict the depth of the current CU by utilizing the depth of the neighboring CUs.
Hence, the spatial correlation-based CU depth-prediction algorithm for interprediction is shown in Algorithm 2. Firstly, the predicted depth range of the current CU is determined by the depth of the neighboring CUs.Secondly, the RD-cost of the current CU is checked in the predicted depth range.The advantage of this method is simple and easy to achieve.Moreover, the robustness of this method is high.

Overall Algorithm
Based on the spatial-correlation model, the MVs of the neighboring PUs are used to obtain the optimized MVP.This method can improve coding efficiency, while computation complexity is increased by a lot.In order to achieve a better trade-off between coding efficiency and computation complexity, by jointing CU depth prediction, the overall algorithm can significantly improve coding performance.The flowchart of the overall algorithm is shown in Figure 6, which can be divided into three distinctive steps, as follows: Step 1: spatial correlation-based motion-vector prediction.The MVP is selected by using the spatial-correlation model for interprediction.Firstly, If MV L = MV TR = MV TL , MV L is selected as the optimized MVP.Secondly, If |MV TR − MV TL | > |MV TL − MV L |, MV TL and MV L are added to the candidates.Otherwise, MV TR and MV TL are added to the candidates.Lastly, If MV TR , MV TL , and MV L are invalid, All MVs surrounding the current PU are added to the candidates, and the redundant MVP candidate can be reduced by comparing one with one.Executing motion estimation is to determine the rate-distortion-optimal MVP.
Start depth prediction with the RDO method for a CU with different block partitioning.If the maximal depths of CU L , CU TR , and CU TL are less than or equal to 1, the predicted depth range of the current CU is 0, 1, and 2. Else, if the maximal depths of CU L , CU TR , and CU TL are more than 1, the predicted depth range of the current CU is 1, 2, and 3. Otherwise, the predicted depth range of the current CU is 0, 1, 2, and 3.
Step 3: If the current depth of the CU exceeds the predicted depth range, RD-cost computation is stopped.Otherwise, depth is incremented by 1 and recursively checks the RD-cost in the current depth.
It should be pointed out that the overall algorithm is a recursive process, and spatial correlation-based CU depth prediction is not applied to intra.
The proposed algorithm is able to achieve lesser quality degradation while reducing the bitrate.The benefit is from the reliability of the candidate MVs, and the improvement in MV accuracy is significant compared with the AMVP technique.In order to evaluate this opinion, some experiments were performed to count the rate in which the MVD is equal to zero, and these rates are identified as R x and R y in the X-component and Y-component, respectively.Table 5 shows the results for a typical sequence (RaceHorses) when the configuration profile is RA and QP is set to 32.It is seen from the results that the MVD of most PUs is equal to zero compared with the H.265/HEVC reference software.Thus, the accuracy of the MVP was improved for InterMode, and coding efficiency was improved with the proposed method.Moreover, the rate of MVD in a whole bitstream increases with the increase of QP. Figure 7 shows the MVD portion depending on QP for the RaceHorses sequence compared to the H.265/HEVC reference model (HM).It is noted that, at a low bitrate (high QP), motion information is a major part of the total bitstream.
In H.265/HEVC standards, MergeMode is used for the PU which MVD is zero, and only the MVP index of the selected candidate in the merge list is transmitted.In other words, MergeMode allows the MV of a PU to be copied from a neighboring PU, and no motion parameter is coded in the encoder side.Correspondingly, in the decoder side, the final MV can be directly obtained by the transmitted merging MVP index.Using large block sizes for motion compensation and MergeMode is very efficient for regions with consistent displacements.
In order to analyze the percentage of the MergeMode selected as the best prediction mode, the typical sequences (RaceHorses and BasketballDrill) are tested between the proposed method and H.265/HEVC reference software, when the configure profile is RA and QP is set to 32. Figure 8 shows the percentage of MergeMode selected as the best prediction mode in the proposed method for RaceHorses and BasketballDrill sequences, compared with HEVC reference software.It noted that 76.37% and 91.46% CUs selected MergeMode as the best PU mode in the proposed method for the RaceHorses and BasketballDrill sequences, while 54.18% and 74.47% in the H.265/HEVC reference software.Therefore, more CUs select MergeMode as the best PU mode in the proposed method compared with the H.265/HEVC reference software.
To evaluate steady performance, Figure 9 shows a typical example of the R-D curve for the RaceHorses, BasketballDrive, and BQTerrace sequences in the RA and LD profiles.Regardless of in high bitrates or in low bitrates, the coding performance of the proposed method exceeded the H.265/HEVC reference model.

Performance of Spatial Correlation-Based CU Depth Prediction Algorithm
In order to reduce computation complexity, a spatial correlation-based CU depth-prediction algorithm is proposed.The results of this method are shown in Table 6.It can be seen that the proposed method could reduce encoding time by 12.89% under an RA profile, while coding efficiency could be reduced by 0.31%.Computation complexity could be reduced by 12.69% encoding time under the LD profile, while coding efficiency can be reduced by 0.29%.Thus, the spatial correlation-based CU depth-prediction algorithm impacts complexity reduction with a slight degradation of coding efficiency.Compared with previous complexity-reduction methods, the depth information of the neighboring CU is convenient to obtain and the implementation cost is low.Moreover, the robustness of this method is high.In order to evaluate the subjective performance for CU depth prediction, subjective tests were conducted in a controlled environment.Firstly, the Double Stimulus Impairment Scale (DSIS) method was used to perform the subjective quality-assessment experiment [22].The subjects were presented with pairs of video sequences, where the first sequence was a H.265/HEVC reference video and the second sequence was a video with the proposed method.Secondly, a total of 24 naive viewers took part in the test campaign, the number of female and male viewers was 8 and 16, respectively, and the age median of the subjects was 25 years old.All viewers were screened for correct visual acuity and color vision.Thirdly, viewers were expected to mark their visual-quality score on an answer sheet with quality rating scale over a defined scale, and the scale was made of 5 levels ranging from "10" (Very annoying) to "90" (imperceptible) as shown in Table 7.The mean opinion score (MOS) was computed for each test as the mean across the rates of the valid subjects.For the RaceHorses and BasketballDrive sequences, Figures 10 and 11 show the MOS for male and female viewers with a different QP.Moreover, Figure 12 show the rate-MOS curves compared with the H.265/HEVC reference model.The results are reliable, and variations between the subjects were rather small.In Figure 12, it can be seen that the proposed method showed slightly improved visual quality over the H.265/HEVC reference model at higher bit rates, or H.265/HEVC reference model showed higher visual quality over the proposed method at a lower bit rate.As a whole, there was little difference in the subjective quality performance between the proposed method and the H.265/HEVC reference model.high bitrates or in low bitrates, the coding efficiency of the overall algorithm significantly improved coding performance.Compared with previous work [4,5], the reference results are shown in Table 9.In both high resolution and low resolution, the coding efficiency of the proposed method was higher than Lin's and Peng's method.The benefit is from an accurate MVP according to the MVs surrounding the CU.Moreover, this proposed method can achieve a better trade-off between coding efficiency and computation complexity.
It should be specially mentioned that this proposed method causes coding complexity to increase by raising coding efficiency.However, for the application that does not care about real-time encoding, and cares more about coding efficiency, it is an efficient approach for coding-efficiency improvement.

Algorithm 1 :1 3 if 6 Add 7 else 8 Add
Spatial correlation-based MVP algorithm.Start interprediction for PU 2 if subset M exist then MV TR = MV TL = MV L then 4 MV L is selected as best MVP 5 else if |MV TR − MV TL | > |MV TL − MV L | then MV TL and MV L to MV candidates MV TR and MV TL to MV candidates 9 else 10 Add all MVs of the G to MV candidates; 11 reduce redundant MV candidates 12 Execute motion estimation to determine the rate-distortion-optimal MVP.

Figure 9 .
Figure 9. R-D curve of the spatial correlation-based MVP algorithm.

Figure 13 .
Figure 13.R-D curve of the overall algorithm.

Table 2 .
Length of mvp_lx_ f lag with a different PU size.

Table 3 .
Accuracy for different conditions.

Table 5 .
Rate in which MVD is equal to zero (RaceHorses).