Video Error-Resilience Research Based on Error-Resilient Screen Content

In order to make video transmission more stable, various error-resilient mechanisms are proposed on video coding in the literature. However, the redundancy mechanism behind classical redundant coding algorithms is relatively simple and is not suitable for the network environment and video content in the context of screen content sequence with multiple abrupt frames and still frames. Motivated by this, a frame-level coding selection mechanism is proposed in this paper for the error-resilience transmission of screen content, where additional code stream or redundant information is considered to improve error-resilient performance with redundant coding and acceptable video quality is obtained in the case of frame transmission error. In addition, selective allocation redundancy is conducted to take the importance of the video frame ROI (region of interest) area into account in the co-encoding process. As a result, the redundancy insertion efficiency and the reliability are improved in return. The corresponding experiments validate the effectiveness of the schemes proposed in this paper.


Introduction
With the rapid development of multimedia technology, video resolution has been constantly improving. The storage and transmission of high-definition and ultra-high-definition video bring great challenges to video coding technology. Under the background of emerging video applications and continuous innovation of network and computer technology, high-efficiency video coding (HEVC) has further optimized the H.264 coding standard according to the characteristics of high definition (HD) video. Compared with H.264/AVC, 50% bit-rate can be saved for HEVC in terms of high profile on the premise of the same video coding quality. Consequently, HEVC is widely accepted and applied [1].
Screen content image is generally composed of natural and computer-generated images-image [2]. The encoding method of screen content image can adapt to the content of both images at the same time. At present, this coding method is widely used in remote desktop sharing, video conference, radio and television, and other video applications [3]. In modern coding standards, time-domain prediction is used to eliminate the redundancy, and video data are highly dependent on each other. Packet loss will cause serious distortion after subsequent decoding. Fault-tolerant coding algorithms can reduce the errors in the process of video transmission [4]. In order to make better use of the characteristics of screen content video and improve the encoding efficiency of screen content video, VCEG and MPEG jointly issued a proposal solicitation announcement of screen content video encoding in January 2014, and officially started the work of formulating HEVC-SCC standard [5]. After a two-year proposal collection, experimental verification, performance evaluation, and other work, the video coding standard for screen content was developed and completed in February 2016 [6]. Nevertheless, the distortion will still appear in a loss of packets for HEVC-SCC due to the bad network transmission environment, which will undermine the quality of video and affect the user experience. Therefore, the video fault tolerance mechanism is discussed to improve video quality based on HEVC-SCC in this paper.
In the literature, various error-resilient mechanisms are discussed. Carreira et al. [7] have proposed a two-stage method to improve the error robustness of HEVC by reducing the time error propagation in the case of frame loss. In the coding stage, the number of prediction units that depend on a single reference is reduced to distribute the use of reference pictures; in the streaming stage, the MV prediction mismatch at the decoder is reduced by the priority sorting algorithm of motion vectors. In another scheme [8], frame loss simulation and corresponding error concealment are proposed to find the most effective method where the decoder is used to recover lost frames. Ferre et al. [9] have proposed a redundant coding method with macroblock rate and potential distortion information. The proposed algorithm is based on the given redundancy rate constraint and can achieve error resistant wireless video transmission without relying on retransmission. Xu [10] put forward a method of macroblock classification based on texture characteristics, motion characteristics, and other information, and then redundantly encoded important macroblocks. A joint rate-distortion optimization algorithm for primary and redundant images is proposed and this method [11] traverses all coding modes and coding parameter combinations of redundant macroblocks and selects the best combination of rate-distortion as the final coding mode. In [12], two macroblock coding modes are added to enhance the transmission robustness of the coded bit stream. Tillo [13] utilized the main frame and redundant frame to generate different descriptions of MDC and analyzed the relationship between redundant frame and quantization coefficient of main frame. Because the source distortion is not related to the pixel value, and the total end-to-end distortion can be represented by the sum of single lost distortion, and the overall distortion optimization process is simplified as a separate optimization process for each frame.
Motivated by the work mentioned in [8][9][10][11][12][13], a video error-resilient scheme is proposed for error-resilient transmission of screen content coding (SCC) in this paper. In the proposed scheme, additional code stream or redundant information is considered to improve error-resilient performance with redundant coding and acceptable video quality is obtained in the case of frame transmission error. In addition, selective allocation redundancy is conducted to take the importance of the video frame ROI area into account in the co-encoding process. Consequently, the redundancy insertion efficiency and reliability are improved in return. The corresponding experiments validate the effectiveness of the schemes proposed in this paper. The main contribution of this paper is concluded as follows.

1.
This paper provides an adaptive error prevention coding mechanism for frame-level-based video transmission security, where the total end-to-end rate-distortion cost can be optimized by adjusting the number of coded redundant frames.

2.
To further optimize the bit rate and redundancy allocation, the abrupt frames are distinguished according to the characteristics of SCC, and the region of interest is considered as a redundant allocation.
The rest of this paper is organized as follows. Section 2 discusses the related work. Section 3 provides the main contribution of this paper. Simulation experiments and result analyses are accomplished in Section 4. Finally, Section 5 concludes this paper.

Related Work
In the video error recovery technology, fault-tolerant coding belongs to the forward technology. The forward technology refers to introducing redundant error correction bits to the transmission stream at the encoding end. Under certain conditions, the transmission error can be automatically corrected in the decoding process, and the BER (bit error rate) of the received stream can be reduced. According to the standard coding scheme, fault-tolerant video code stream is generated to solve the error problem in the video transmission process. Through the change of coding mode, the impact of error on the decoded video quality is suppressed as much as possible [14]. Error concealment, as a post-processing technology, repairs and conceals the data in lossy transmission according to the normal received stream information, so that the decoded video can be as close as possible to the original encoded image [15]. In the feedback-based coding transmission framework, when the receiver finds data loss, it can choose the coding unit correctly received in the reference image sequence to realize error recovery. For the coding transmission system without feedback, the redundant slice coding method can add one or several redundant representations of a coding unit in the code stream [16]. As the decoder decodes the received stream, the ordinary information is first utilized. If the received stream is lost, the added redundant slice is considered to decode and reconstruct.
When the transmission error occurs in the main frame, the redundant frame is decoded so that the decoder can obtain acceptable video quality compared with the original video. In the process of video transmission, redundancy coding can be realized in video source coding and channel coding, respectively. The additional redundant content can be either the redundant image encoded for the whole image, or the redundant slice encoded for a specific region or some macroblocks [17].
In the coding process, redundant frames often use different coding parameters from the main frame to improve the rate-distortion performance. Zhu et al. [18] propose a redundant image coding method based on reference frame selection with motion vector and error diffusion distortion to determine whether the current frame encodes redundant frames, where reference frame selection is just utilized in redundant frames. The probability of using redundant frames is only 1-p, where p denotes the probability of independent packet losses for image frames. As a result, the loss reduction with reference frame selection mechanism in redundant frames may not offset the increase of rate-distortion cost caused by the increase of code rate.
Redundant slice embedding is equivalent to the repeated encoding of pictures, which consumes more encoding bits [19]. Therefore, it is preferable to add redundancy to important areas rather than to protect all areas. For error-resilient coding with redundant frames, redundant frames are encoded to improve the robustness of I-frames, so as to further enhance the image transmission quality of video [9]. The method based on adjusting the coding parameters of the redundant frame and main frame does not consider the influence of different scene features. The adaptive redundant frame coding algorithm only uses a fixed reference frame selection mode for redundant frames and does not use other error-resilient methods for main frames, which makes the generation mechanisms of redundant images (redundant slices) simple and limits its adaptability to different network conditions and video scenes.

Proposed Method
Traditional error-resilient coding algorithms are based on a single error suppression mechanism and do not adaptively consider different error-resilient coding methods according to the application environment. For this reason, this section proposes an algorithm based on a frame-level error-resilient coding selection mechanism for the transmission security of screen content sequences. If the complexity of the current frame is greater than three times the average encoded frames, the current frame defined as a mutation (abrupt) frame. Due to the lack of correlation with the encoded frames, the mutation frame is often encoded by intra prediction, which consumes a lot of bits. By judging the mutation frame, the abrupt frame is regarded as a more important frame and more redundancy is assigned to it to achieve adaptive frame-level video error-resilient coding algorithm.
Note that the total end-to-end rate-distortion is optimized by virtue of number adjustment of coded redundant frames in this paper. In consequence, error prevention coding varies adaptively in the process of video transmission. In addition, the abrupt frames are determined based on SCC characteristics and the region of interest is considered as a redundant allocation for the proposed scheme. The flowchart of the proposed scheme is described in Figure 1. According to Figure 1, the detailed operations are concluded as follows.

Region Division
The HEVC-SCC standard is developed on the basis of the HEVC coding framework. The reliability can be ensured by judging the abrupt frames in the sequence of screen content coding and implementing the redundant coding. The judgment is made by utilizing the proportion of the complexity of the current frame in the entire sliding window. The complexity of the current frame refers to the SAD (sum of absolute differences) value of the current frame and the reference frame. represents three times the average SAD value for all previous coding frames, which is used as the judgment threshold to determine whether the current frame is a mutation frame. pre C represents the sum of the complexity of all encoded frames before the current frame, and pre N denotes the number of currently encoded frames. Note that is set due to numerical experience. 1. The proposed algorithm first creates and initializes a sliding window with size 5. 2. Then the complexity of the current frame is analyzed and compared with the threshold to determine whether the current frame is an abrupt frame.
3. If the current frame is not an abrupt frame, the original model of HEVC is used to control the bit rate.
4. If the current frame is an abrupt frame, the algorithm will judge whether the redundancy (frame) count in the slice header is greater than 0. Note that the information of the redundant frame count is in the slice header in the process of error-resilient encoding. A slice contains part or all of the data of a frame image [17]. In other words, a frame of a video image can be encoded into one or more slices. Each slice is composed of two parts, where one part is the slice header and is used to save the overall information of the slice. During the encoding process, the information of the redundant frame count is included in the slice header.
5. When a slice with a redundancy count is greater than 0, the slice is identified as a redundant slice, otherwise, it will be identified as a primary slice. In the encoding process, the encoding of the current frame is divided into two phases: primary frame encoding and redundant frame encoding. 6. The primary slice will be encoded by the primary frame encoding. The ROI of the redundant slice will be divided, then the redundant slice will be encoded with the adaptive frame-level video error-resilient coding algorithm based on ROI.
7. If the current frame is not the last frame, repeat step 2 to step 6, otherwise, the algorithm will terminate.

Region Division
The HEVC-SCC standard is developed on the basis of the HEVC coding framework. The reliability can be ensured by judging the abrupt frames in the sequence of screen content coding and implementing the redundant coding. The judgment is made by utilizing the proportion of the complexity of the current frame in the entire sliding window. The complexity of the current frame refers to the SAD (sum of absolute differences) value of the current frame and the reference frame. 3 * C pre N pre represents three times the average SAD value for all previous coding frames, which is used as the judgment threshold to determine whether the current frame is a mutation frame. C pre represents the sum of the complexity of all encoded frames before the current frame, and N pre denotes the number of currently encoded frames. Note that 3 * C pre N pre is set due to numerical experience. Since the mean value can better reflect the complexity of the already encoded frame, and test, the decision threshold can accurately identify the abrupt frame. When the complexity of the current frame is greater than three times the average value of all previously encoded frames C pre , the current frame is an abrupt frame, and vice versa. If the current frame is an abrupt frame, the current frame and subsequent frames (four frames) are subjected to redundant coding; otherwise, the code is controlled according to the original code rate control model. Figure 2 exhibits the flowchart of region division. Note that region division aims at determining whether the current frame is the abrupt frame. According to the decision results, different actions will be taken just as mentioned above.
Appl. Sci. 2020, 10, x FOR PEER REVIEW 6 of 18 Since the mean value can better reflect the complexity of the already encoded frame, and test, the decision threshold can accurately identify the abrupt frame. When the complexity of the current frame is greater than three times the average value of all previously encoded frames pre C , the current frame is an abrupt frame, and vice versa. If the current frame is an abrupt frame, the current frame and subsequent frames (four frames) are subjected to redundant coding; otherwise, the code is controlled according to the original code rate control model. Figure 2 exhibits the flowchart of region division. Note that region division aims at determining whether the current frame is the abrupt frame. According to the decision results, different actions will be taken just as mentioned above.

Error-Resilient Coding
To improve the robustness of video transmission, different error-resilient mechanisms are considered. The error-resilient code stream can be improved by providing additional code streams or information. During the encoding process, the information of the redundant frame count is included in the slice header. When a slice with a redundancy count is greater than 0, the slice is identified as a redundant slice, and a slice with a redundant frame count equal to 0 is identified as a

Error-Resilient Coding
To improve the robustness of video transmission, different error-resilient mechanisms are considered. The error-resilient code stream can be improved by providing additional code streams or information. During the encoding process, the information of the redundant frame count is included in the slice header. When a slice with a redundancy count is greater than 0, the slice is identified as a redundant slice, and a slice with a redundant frame count equal to 0 is identified as a primary slice. In the encoding process, the encoding of the current frame is divided into two phases: primary frame encoding and redundant frame encoding.
By analyzing the code rate and distortion characteristics of the current frame, an adaptive redundant coding algorithm is implemented to optimize the reference end-to-end rate-distortion cost. At the same time, the redundant coding method is adaptively selected based on the ROI region, thus completing the security-based research on video error prevention mechanisms.
The reconstructed value of the redundant pixels is made equal to the reconstructed value of the corresponding pixel in the original picture, which can reduce the error concealment bias. The embedding of redundancy can effectively enhance the error-resilient ability of the bit stream, but it also brings an increase in bit consumption. Therefore, it is necessary to strike a balance between coding efficiency and error prevention capability.
Different from classical error-resilient coding algorithms, such as [20][21][22][23], the proposed scheme in this paper provides an adaptive error prevention coding mechanism for frame level-based video transmission security. The total end-to-end rate-distortion cost can be optimized by adjusting the number of coded redundant frames. In order to further optimize the bit rate and redundancy allocation, the abrupt frames can be distinguished according to the characteristics of the screen content coding, and the region of interest can be considered as a redundant allocation. On the premise of secure transmission, it can better adapt to different network conditions and scenarios and ensure the rated performance.
In the process of image processing, the region of interest (ROI) is an image region selected from the image. This region is the focus of image analysis and is circled for further processing. Circling the target with ROI can reduce the processing time and increase the precision. In addition, a video error-resilient scheme is proposed for error-resilient transmission of screen content coding (SCC) in this paper. More specifically, the primary slice will be encoded by the primary frame encoding in the proposed scheme. The ROI of the redundant slice will be divided, then the redundant slice will be encoded with the adaptive frame-level video error-resilient coding algorithm based on ROI. Taking the importance of the video frame ROI area into account in the co-encoding process contributes to realizing the security-based research on video error prevention mechanisms. Consequently, the redundancy insertion efficiency and the reliability are improved.

Analysis of Distortion
In this subsection, the distortion situation is discussed to evaluate the performance of the error-resilience coding scheme proposed in this paper. The lower the distortion, the higher the error-resilience capacity.
In the video transmission process, the overall end-to-end distortion D total mainly comes from three parts, namely coding distortion, error concealment distortion, and potential error diffusion distortion from the reference block. D total could be described as: where D Q denotes the quantization distortion; D r represents the distortion of potential error diffusion caused by the reference frame; D ec denotes the distortion between the content of the error concealed video macroblock and the correct video macroblock; D ec depends on the current coding parameter and error concealment; D ec_r represents the potential distortion caused by hiding the wrong macroblock.
In addition, f i signifies the original video frame, ∼ f i denotes the video frame with redundancy encoding, E{·} denotes the expectation operator and p denotes the probability of independent packet losses for image frames. Note that errors may occur due to different color ranges in the process of image color conversion, and the error diffusion algorithm [24] reduces the visual errors caused by transmitting the errors to the surrounding pixels.
The distortion D pri when the redundant frame is not currently embedded can be written as: When the redundant frame is embedded, the distortion could be formulated as [21]: where D di f denotes the distortion caused by the difference between the main frame and the redundant frame, and D red_r signifies the distortion generated by the potential error diffusion of the redundant frame. If packet loss happens for the current frame with probability p, the redundant frame will be used. As a result, the redundancy of the current frame is not lost. Then the remaining frames are normal with probability p · (1 − p). The proposed redundant coding algorithm assumes that when the slice of the primary frame is lost, the decoding operation of the redundant frame on the decoding side replaces the content of the lost frame. As a result, the use of redundant frames and redundant slices reduces error propagation and improves video reliability.
Rate-distortion performance can be seen from the change rate of distortion (for a given source distribution and distortion metric, the minimum expected distortion that can be achieved at a specific bit rate) and the change of code rate: where R cur denotes the bit consumption common to the current primary frame and the redundant frame in the case of embedded redundancy, R pri represents the bit consumption of the current frame. ∆R, the amount of change in the code rate is also the code rate of the redundant frame. k signifies the order number of the pixels belonging to the rectangular ROI part. When a redundant frame is used, the probability that the redundant frame and the main frame are simultaneously lost is p 2 , it is obvious that the change rate of the distortion change and the code rate are both related to the packet loss rate p. When the packet loss rate p is small, a single redundant frame can be used to implement error-resilience. When the packet loss rate p is large, different redundant frame allocation strategies can be adopted according to the type of frame. For key frames in the screen content encoding, more redundant frames are allocated for encoding in order to adapt to different characteristics of the SCC and other scenarios, thereby achieving reliability improvement under secure transmission. Then we evaluate the rate-distortion (R-D) performance with a redundant slice, which is the ratio of the changing distortion and the changing bit consumption (the distortion change per unit bit consumption) after embedding a redundant slice: After obtaining the Lagrangian coefficient λ from the code rate, the quantization parameter is obtained based on the relationship between QP and λ.
In general, natural images can be divided into the following categories: (1) A completely static scene, that is, the image content of the current frame and the previous frame are exactly the same. (2) A continuously changing scene, that is, there is continuity between adjacent frames in the video.
The moving object or the angle of view of the shot has moved.
However, for the sequence of screen contents, in addition to the above two types of images, there are images with scene mutations appearing. In the proposed solution, several cases based on the coding of the screen content are considered. Most of the intra prediction in a picture is likely to mean scene conversion. If the motion vector of all macroblocks is large, this indicates that the features in the picture may be more complicated; and if the ROI area in a picture is large, it means that a larger protected area is needed. After the current picture is encoded, the encoding parameters are updated based on the R − λ model [25].

R-λ Model
In this section, the update of the encoding parameters is discussed based on the R − λ model [25]. Figure 3 exhibits the R − λ rate control model. Generally, the R − λ model contains two steps, bit allocation and coding parameter updating.
Appl. Sci. 2020, 10, x FOR PEER REVIEW 9 of 17 (2) A continuously changing scene, that is, there is continuity between adjacent frames in the video. The moving object or the angle of view of the shot has moved.
However, for the sequence of screen contents, in addition to the above two types of images, there are images with scene mutations appearing. In the proposed solution, several cases based on the coding of the screen content are considered. Most of the intra prediction in a picture is likely to mean scene conversion. If the motion vector of all macroblocks is large, this indicates that the features in the picture may be more complicated; and if the ROI area in a picture is large, it means that a larger protected area is needed. After the current picture is encoded, the encoding parameters are updated based on the R λ − model [25].

R-λ Model
In this section, the update of the encoding parameters is discussed based on the R λ − model [25].
where R tar is the target bit, and FR is the frame rate.

Bit Allocation
(1) Group of Picture (GOP) level bit allocation The number of encoded frames is N codee , and the number of bits used to encode these frames is R coded . The number of frames in a GOP is N GOP ; and SW is the size of the sliding window for smoothing bit allocation, which is used to make bit consumption change and the quality of encoded pictures gentler. The bit allocation of GOP level is: It is emphasized that the target code rate is achieved after SW. frames. If each frame in the SW. frames can exactly consume T AVgpic bits, then the above equation can be written as follows: (2) Frame-level bit allocation Define Coded GOP as the consumed bits of the current GOP, and ω is the bit allocation weight of each frame, then the bits assigned to the current frame is defined as: (3) Largest Coding Unit (LCU) level bit allocation Similar to the frame-level bit allocation strategy, the bits allocated to each LCU are calculated based on: where Bit header is the estimated number of all header information bits, which is estimated from the actual number of header information bits of the previously encoded pictures. ω is the bit allocation weight of each LCU.

Coding Parameters Update
According to the relationship of λ and R [25]: where R denotes the bit rate; D is the MSE (mean squared error) distortion; and are model parameters related to the characteristics of the video source [25]; λ denotes the slope of the R-D curve, and α, β signifies the parameters related to the video content. That is, λ can be directly calculated by R through α and β However, α and β are parameters related to the content characteristics of the video, and the values of different contents are significantly different. The R-λ model solves this problem by introducing the following algorithm. Firstly, the following formula is used to calculate the λ of the frame and LCU [25]: where bpp. bit rate, α and β are different for each frame and each LCU. In order to realize content adaptation, α and β will be updated continuously in the encoding process. After completing the encoding of an LCU or a frame, α and β are updated as follows: where is the actual bpp real is the actual bpp in the coding process; λ real is the actual value of λ in the coding process; δ α = 0.1 and β = 0.05 [25].
Note that bpp is defined as: where f is the frame rate, w and h are the width and height of the picture, respectively [25]. When λ is determined, QP is calculated according to [26].

Experimental Conditions
HM-16.7-SCM7.0 is an open-source encoder maintained by the HEVC standard development organization JCT-VC [27], which integrates the new coding tools in the SCC standard and greatly improves the coding efficiency of screen content sequences. The video sequences used are all selected from the test sequences recommended in the SCC official general test conditions. The frame structure of the encoder uses a low-latency profile to simulate real-time transmission. The resolution, frame rate, and target bit rate of the experimental sequence are shown in Table 1 [27]. Note that some sequences in Table 1 have the same information and bit rate, but their contents are entirely different. As a result, their names are in different forms. In addition, packet loss ratio refers to the ratio of the number of packets lost to the data groups sent during the test. Note that the packet loss ratio refers to the ratio of the number of packets lost to the data groups sent during the test. The packet loss in the channel can be simulated by the packet loss simulator. The packet loss simulation software can be obtained from [28]. Packet loss simulation software can set packet loss rate (packet loss probability) to simulate the packet loss when the video is transmitting in the channel.
The PSNR [29] can be evaluated based on: where MAX I is the maximum possible pixel value of the image. Because each pixel of an image is represented by 8 bits, the value of MAX I is 255. I(i, j) represents the pixel value of the original image at the point (i, j). K(i, j) represents the pixel value at point of the reconstructed image decoded and reconstructed by the decoder after the code stream is processed by the packet loss simulation software. PSNR is the peak signal-to-noise ratio, usually after image compression, the output image will be somewhat different from the original image. A smaller PSNR value indicates that the distortion caused by the embedded information is small. Taking PSNR as the main objective quality evaluation standard, the higher the PSNR value is, the better the video quality gets. That is, PSNR is positively correlated with the algorithm performance. As a result, PSNR is considered as the evaluation index of the proposed scheme. Figure 4a,b, respectively exhibits the PSNR value distribution with the change of the web browsing and sideshow screen content sequences at the 256kbps bit rate. According to the experiment results, the proposed scheme is able to achieve a higher overall PSNR score, which illustrates the better video encoding effect. By judging the mutation frame, the abrupt frame is regarded as a more important frame and more redundancy is assigned to it for adaptive frame-level video error-resilient coding. Additional code stream or redundant information is considered to improve error-resilient performance with redundant coding and acceptable video quality is obtained in the case of frame transmission error as a consequence. PSNR is the peak signal-to-noise ratio, usually after image compression, the output image will be somewhat different from the original image. A smaller PSNR value indicates that the distortion caused by the embedded information is small. Taking PSNR as the main objective quality evaluation standard, the higher the PSNR value is, the better the video quality gets. That is, PSNR is positively correlated with the algorithm performance. As a result, PSNR is considered as the evaluation index of the proposed scheme. Figure 4a,b, respectively exhibits the PSNR value distribution with the change of the web browsing and sideshow screen content sequences at the 256kbps bit rate. According to the experiment results, the proposed scheme is able to achieve a higher overall PSNR score, which illustrates the better video encoding effect. By judging the mutation frame, the abrupt frame is regarded as a more important frame and more redundancy is assigned to it for adaptive frame-level video error-resilient coding. Additional code stream or redundant information is considered to improve error-resilient performance with redundant coding and acceptable video quality is obtained in the case of frame transmission error as a consequence.     Figure 5 gives the average PSNR comparison without and with the proposed scheme. From Figure 5, the average PSNR score with the proposed scheme is higher than that without the proposed scheme [22]. These results accord with the analysis in Section 3, which validates the effectiveness of the proposed scheme in this paper.    To further verify the performance of the proposed scheme, the average PSNR is compared with three classical video error-resilience coding schemes [11,13,15], as it is shown in Figure 6. Obviously, the average PSNR score of the proposed scheme is the highest, which indicates the proposed adaptive frame-level video error prevention coding algorithm can improve the transmission quality of video on an error-prone network channel by judging the abrupt frames and allocating more redundancy to them.

Experimental Results
According to the simulations in Figures 4 to 6, the proposed algorithm improves bit rate error and encoding efficiency, which is very helpful for video encoding transmission in multimedia applications, especially in the case of poor network conditions like the situation of packet loss. At the same time, since the proposed method does not perform redundant coding for each frame, the bit rate is small. During redundant embedding, more redundant slices are embedded in the abrupt frame and the region of interest is considered to achieve a better error-resilience and ensure the performance of video R-D performance.
Proposed scheme [11] [13] [15]   To further verify the performance of the proposed scheme, the average PSNR is compared with three classical video error-resilience coding schemes [11,13,15], as it is shown in Figure 6. Obviously, the average PSNR score of the proposed scheme is the highest, which indicates the proposed adaptive frame-level video error prevention coding algorithm can improve the transmission quality of video on an error-prone network channel by judging the abrupt frames and allocating more redundancy to them.  To further verify the performance of the proposed scheme, the average PSNR is compared with three classical video error-resilience coding schemes [11,13,15], as it is shown in Figure 6. Obviously, the average PSNR score of the proposed scheme is the highest, which indicates the proposed adaptive frame-level video error prevention coding algorithm can improve the transmission quality of video on an error-prone network channel by judging the abrupt frames and allocating more redundancy to them.
According to the simulations in Figures 4 to 6, the proposed algorithm improves bit rate error and encoding efficiency, which is very helpful for video encoding transmission in multimedia applications, especially in the case of poor network conditions like the situation of packet loss. At the same time, since the proposed method does not perform redundant coding for each frame, the bit rate is small. During redundant embedding, more redundant slices are embedded in the abrupt frame and the region of interest is considered to achieve a better error-resilience and ensure the performance of video R-D performance.
Proposed scheme [11] [13] [15]  proposed scheme [11] [13] Proposed scheme [11] [13] Proposed scheme [11] [13]  According to the simulations in Figures 4-6, the proposed algorithm improves bit rate error and encoding efficiency, which is very helpful for video encoding transmission in multimedia applications, especially in the case of poor network conditions like the situation of packet loss. At the same time, since the proposed method does not perform redundant coding for each frame, the bit rate is small. During redundant embedding, more redundant slices are embedded in the abrupt frame and the region of interest is considered to achieve a better error-resilience and ensure the performance of video R-D performance.

Conclusions
In this paper, an error-resilience video coding scheme is proposed based on an adaptive redundancy allocation strategy for quality improvement of video transmission on error-prone network channels. By adjusting the number of coded redundant frames, the overall end-to-end rate-distortion cost is optimized. In addition, to obtain further optimization of the bit rate and redundancy allocation, the abrupt frame can be distinguished according to the characteristics of the screen content encoding while the region of interest is considered for redundant allocation. On the premise of error-resilient transmission, the proposed scheme achieves better adaptability to different network situations and scenarios and guarantees rate-distortion performance. In the future work, more optimization factors will be considered for a higher video transmission quality.