Next Article in Journal
Investigation into the Efficient Cooperative Planning Approach for Dual-Arm Picking Sequences of Dwarf, High-Density Safflowers
Previous Article in Journal
Motor Temperature Observer for Four-Mass Thermal Model Based Rolling Mills
Previous Article in Special Issue
Improvement of SAM2 Algorithm Based on Kalman Filtering for Long-Term Video Object Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Masked Feature Residual Coding for Neural Video Compression

1
School of Electrical and Electronic Engineering, Yonsei University, Seoul 03722, Republic of Korea
2
Samsung Seoul R&D Campus, Seoul 06765, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(14), 4460; https://doi.org/10.3390/s25144460
Submission received: 29 May 2025 / Revised: 9 July 2025 / Accepted: 15 July 2025 / Published: 17 July 2025

Abstract

In neural video compression, an approximation of the target frame is predicted, and a mask is subsequently applied to it. Then, the masked predicted frame is subtracted from the target frame and fed into the encoder along with the conditional information. However, this structure has two limitations. First, in the pixel domain, even if the mask is perfectly predicted, the residuals cannot be significantly reduced. Second, reconstructed features with abundant temporal context information cannot be used as references for compressing the next frame. To address these problems, we propose Conditional Masked Feature Residual (CMFR) Coding. We extract features from the target frame and the predicted features using neural networks. Then, we predict the mask and subtract the masked predicted features from the target features. Thereafter, the difference is fed into the encoder with the conditional information. Moreover, to more effectively remove conditional information from the target frame, we introduce a Scaled Feature Fusion (SFF) module. In addition, we introduce a Motion Refiner to enhance the quality of the decoded optical flow. Experimental results show that our model achieves an 11.76% bit saving over the model without the proposed methods, averaged over all HEVC test sequences, demonstrating the effectiveness of the proposed methods.

1. Introduction

As higher-resolution videos are captured by camera sensors, the strain on device storage has increased significantly. Consequently, video compression has become important for reducing file size while preserving visual quality. To reduce this load, various standardized video compression technologies have been developed, which significantly reduce file sizes while maintaining video quality. Traditional video compression standards, such as H.264 [1], H.265 [2], VP9 [3], AV1 [4], and VVC [5], include numerous hand-crafted methods like block partitioning, inter and intra prediction, transformation, in-loop filter, and entropy coding. However, since each module is optimized individually and cannot leverage complex nonlinear operations, these methods are considered sub-optimal.
Recently, neural networks have shown significant performance improvements in many image and video tasks [6,7,8,9,10,11,12,13], and they have also been integrated into video compression. The notations and terminology related to neural video compression are listed in Table 1. DVC [14] replaces all modules in the traditional video codec with neural networks and trains the model in an end-to-end manner. DVC uses a simple subtraction operation to obtain the residual between the target frame ( x t ) and the predicted frame ( x c ). This can be demonstrated in H ( x t x c ) , which is referred to as Residual Coding. However, using a simple subtraction operation is not optimal. DCVC [15] uses x c as a conditional input along with the target frame into the encoder in the feature domain ( H ( x t | x c ) ) which is called Conditional Coding. This allows the encoder to learn a more optimal operation to subtract the information in the predicted frame from the target frame. Brand et al. [16] assume that the entropy of x t x c is less than or equal to that of x t , and show that Conditional Residual Coding ( H ( x t x c | x c ) ) has less entropy than or equal to that of Conditional Coding. They feed the difference between x t and x c into the encoder along with the conditional input ( x c ). However, a simple subtraction can result in a large residual. The upper left graph in Figure 1 shows an example of the target and the prediction signal. Then, we can obtain a (a) “Simple Residual” signal, which shows a large residual. Recently, Chen et al. [17] proposed Masked Conditional Residual Coding ( H ( x t m x c | x c ) ). They predicted a soft mask, applied it to the predicted frame ( x c ), and then subtracted it from the target frame. This approach allows for a soft decision between using Conditional Coding and Conditional Residual Coding, which enables us to achieve smaller residuals. If we focus on the (b)“ Masked Residual” signal in Figure 1, assuming the mask is ideal, we can obtain a much smaller residual than that in (a). However, Masked Conditional Residual Coding has two limitations. First, applying a mask in the pixel domain and subtracting it from the target frame cannot consider the diverse and high-level representations in the images. Second, it is effective in reducing error propagation to use the reconstructed features as references to compress the subsequent frame. This is because abundant temporal context information [18,19,20,21] can be included in the reference features. However, Residual-type Conditional Coding methods require decoding the bitstream to the pixel level residuals ( x ^ r ) to obtain full information of the reconstructed target frame ( x ^ t = x ^ r + x c ). Thus, they cannot utilize the reconstructed features of the previous frame as a reference.
To address these problems, we propose Conditional Masked Feature Residual (CMFR) Coding. As shown in the signals on the bottom left in Figure 1, the target and predicted signals can be decomposed into low and high frequencies. Then, by predicting different masks for each frequency component and applying them to the prediction signals, a much smaller residual can be obtained, as shown in (c) “Decompose Residual”. However, it is difficult to find the optimal components to extract for each content of the image, and the corresponding masks are also complex. Therefore, instead of using hand-crafted methods to extract the components from the image, we use a neural network to extract various features from the target frame and predicted features. We also predict the masks for each channel of the features using another neural network. Using these neural networks, our CMFR Coding can be expressed as H ( F t ( x t ) m F c ( x c ) | x c ) , where F t and F c represent the neural networks.
Moreover, we propose the Scaled Feature Fusion (SFF) module. This module predicts the fusion features and scales using conditional information along with the target frame information in the encoder or the residual information in the decoder. We apply the scales to the fusion features and subtract or add the scaled fusion features to the target or residual information, respectively. With this structure, conditional information can be effectively subtracted or added to the target or residual information within the context encoder and decoder. Furthermore, many methods of Conditional Coding [15,18,19,20,21] focus on accurately compensating the reference frame using the decompressed motion. However, they often overlook the quality of the decompressed motion. We propose a Motion Refiner module that uses reference features and the decompressed motion to improve the quality of the motion. Through various experiments, we demonstrate that the proposed methods significantly improve the rate-distortion performance of video compression.
Our main contributions can be summarized as follows:
  • We point out the limitations of the existing Residual-type Conditional Coding, which masks and subtracts the prediction in the pixel domain.
  • To resolve this problem, we propose a CMFR Coding, which extracts the features from the image followed by masking and subtraction in the feature domain.
  • We propose the SFF module to effectively subtract or add conditional information to the target or residual information in the context encoder or decoder, respectively.
  • We introduce the Motion Refiner module to enhance the quality of the decompressed motion vectors.

2. Related Works

Recently, deep learning techniques have demonstrated significant performance improvements in video processing, and there are many attempts to apply these techniques to neural video compression. Basically, they use a neural image compression structure [22,23,24] to compress the motion and the residual information. Lu et al. [14] propose the DVC model, which replaces all modules in the traditional codec with a neural network and trains the model in an end-to-end manner. They achieve a performance comparable to x265 [25] with the “veryfast” option, which is one of the fast encoding presets. They predict the motion between the reference frame ( x ^ t 1 ) and the target frame, x t , and compress it using the neural image compression structure [24]. Then, they predict the approximation of the target frame using the motion vector. Subsequently, the difference between the target frame x t and the prediction ( x c ) is encoded using a simple subtraction operation in the pixel domain, referred to as Residual Coding ( H ( x t x c ) ). Using this residual structure, many papers [26,27,28,29] compress the residual between x t and x c .
DCVC [15] highlights that a simple subtraction operation to obtain the residual is sub-optimal. To solve this problem, DCVC proposes Conditional Coding ( H ( x t | x c ) ). DCVC extracts features from the reference frame and warps them in the feature domain to obtain the predicted feature of the current frame. Then, the target frame is fed into a context encoder, using the predicted feature as conditional input. With this approach, the context encoder can learn an optimized nonlinear function that subtracts the information in the predicted features from the target frame, achieving higher rate-distortion performance than the Residual Coding methods. Moreover, various works [18,19,20,21] that use Conditional Coding predict the approximation of the target frame in the feature domain at multi-scales, and utilize them as conditional input for the context encoder. This structure shows better performance than using only the predicted features of the original resolution.
However, the information bottleneck effect [30] can occur in the Conditional Coding approach due to information loss during the extraction of the features from the reference frame ( x t 1 ) to obtain the conditional input features of the context encoder. Consequently, it may not be possible to reduce the residual entropy to zero even if the result of motion compensation is identical to the target frame. To address this issue, Brand et al. [16] propose a new scheme to obtain the residual called Conditional Residual Coding ( H ( x t x c | x c ) ). They assume that the entropy of x t x c is smaller than or equal to that of x t , and demonstrate that Conditional Residual Coding ( H ( x t x c | x c ) ) has less entropy than or equal to that of Conditional Coding. For processing the Conditional Residual Coding, they obtain the difference between x t and x c using a simple subtraction operation. Then, the difference is fed into the context encoder along with the conditional information, x c . However, a simple subtraction operation can lead to large residuals, especially when the prediction of high-frequency components is incorrect.
To address this issue, Chen et al. [17] propose a new scheme called Masked Conditional Residual Coding ( H ( x t m x c | x c ) ). If the mask (m) has a value of 0, the encoder compresses the residual using Conditional Coding. In contrast, if the mask value is 1, the residual is encoded into the bitstream using Conditional Residual Coding. Therefore, they use a soft decision process with the mask to select which methods to use for encoding. With this structure, they achieve significantly better performance than other coding methods.
Meanwhile, Mentzer et al. [31] do not utilize motion information, and instead implicitly remove redundant information along the temporal axis using a transformer [32]-based entropy model. However, this process is performed only at the 1 / 16 scale. Lu et al. [33] use a hierarchical structure to implicitly reduce temporal redundancy. Nevertheless, the performance of their model is still lower than that of other state-of-the-art models because they also do not utilize motion vectors. Jia et al. [34] use implicit temporal modeling with simple convolutional layers and still outperform VVC by 21% in BD-rate. Jiang et al. [35] use both motion vectors and a transformer structure to capture correlations in the temporal context, particularly focusing on non-local correlations among the frames, and achieve a 38% bit saving on the UVG [36] dataset. Recently, bi-directional referencing has also been introduced in neural video compression models [37,38,39] to process the random access mode.

3. Method

In this section, we first describe the overall structure and then present the details of each proposed module.

3.1. Overall Structure

The goal of neural video compression is to compress the target frame by removing temporal redundancy with previous frames and encoding only the residual information. To do this, we estimate the motion vector ( v t ) between the reference frame ( x t 1 ) and the target frame ( x t ). We estimate v t using pre-trained SPyNet [40], which is widely used by existing neural video compression methods [14,15,18,19,20,21]. This motion vector is compressed and reconstructed ( v ˜ t ) using a motion encoder and decoder. This encoder has four convolutional layers with a kernel size of 3 × 3 and a stride of 2 for downsampling, and residual blocks consisting of two 3 × 3 convolutional layers with Leaky ReLU [41]. Moreover, the qp step ( q s ) is multiplied by the intermediate features at the last layer of each resolution for quality control. The qp step has the shape of q s R 1 × C × 1 × 1 , where C represents the channel size. The motion decoder has a symmetric structure with the encoder except that it uses convolutional layers with a kernel size of 1 × 1 followed by a pixel shuffle [42] for upsampling. Then, we employ a Motion Refiner module to enhance the motion vector. The refined motion vector ( v ^ t ) is used with the Multi-scale Context Fusion module to predict the approximation of the target frame in the feature domain at the original resolution, as well as at 1 / 2 and 1 / 4 resolutions ( C 1 × , C 2 × , and C 4 × ). We use a Temporal Context Mining (TCM) module from [18] as the Multi-scale Context Fusion module to predict ( C 1 × , C 2 × , and C 4 × ). The predicted features at the original resolution ( C 1 × ) are transformed and masked using a Mask Prediction module. Thereafter, the residual is obtained by subtracting these masked features from the features extracted from the target frame in the Residual Generator module. We compress this residual by feeding it into the context encoder, using the multi-scale predicted features as conditional input. In the context encoder and decoder, we introduce the SFF module to more effectively subtract information in the conditional input from the features of the target frame.
On the decoder side, we reconstruct the residual features ( r ^ t ) which are combined with f s c in the Combiner module to obtain the reconstructed features ( f ^ t ). Finally, we transform f ^ t into the reconstructed target frame ( x ^ t ) through a Pixel Mapping block composed of a single convolutional layer. f ^ t contains abundant temporal context information to prevent error propagation. Therefore, we store f ^ t and x ^ t in the reference buffer to utilize them when compressing the next frame.
We set the base channel size of the motion encoder and decoder to 64 for all resolutions. In contrast, the base channel sizes for all other modules are set to 48, 64, 96, 96, and 128 at 1 / 1 , 1 / 2 , 1 / 4 , 1 / 8 , and 1 / 16 resolutions, respectively. In addition, to obtain the qp step vector, we interpolate between the learnable parameters of q s max R 1 × C × 1 × 1 and q s min R 1 × C × 1 × 1 using the following equation:
q s = α · q s max + ( 1 α ) · q s min   ,
where α is the constant obtained by log-scale interpolation in the range from 0 to 1.

3.2. Motion Refiner

After estimating the motion vector between the reference frame and the target frame, we obtain the reconstructed motion vector ( v ˜ t ). Many Conditional Coding models [15,18,19,20,21] focus on warping and compensating the reference frame or features to accurately predict the target frame. They predict the features of C 1 × , C 2 × , and C 4 × , which are approximations of the target frame at the original resolution, 1 / 2 resolution, and 1 / 4 resolution, respectively. However, they often overlook the quality of the reconstructed motion vector.
The reconstructed motion vector ( v ˜ t ) often produces blurred edges around moving objects. We refine this motion before performing warping and compensation at the multi-scale level. First, we warp the reference feature using the reconstructed motion vector. Then, we use the motion vector and the warped features to compute Δ v t .
Δ v t = V θ v ( v ˜ t , W ( f ^ t 1 , v ˜ t ) ) ,
where V θ v denotes the neural network with learnable parameters θ v , and W represents the backward warping operation. Subsequently, we obtain the refined motion vector ( v ^ t ) by adding Δ v t to v ˜ t .
v ^ t = v ˜ t + Δ v t
Using this refined motion vector, we warp and compensate the reference features to predict the approximate features of the target frame ( C 1 × , C 2 × , and C 4 × ). We use these approximate features as conditional input for both the context encoder and the decoder.

3.3. Conditional Masked Feature Residual Coding

Masked Conditional Residual Coding [17] ( H ( x t m x c | x c ) ) is processed in the pixel domain. Therefore, it is not possible to apply different masks to the diverse and high-level representations of the image. To address this problem, we introduce CMFR Coding, which extracts features from the image, applies masking, and performs the subtraction.
First, we use a neural network to extract features from the target image ( F t ( x t ) ). Second, we predict the mask using a neural network. To predict the mask, we use the predicted features at the original resolution ( C 1 × ), the refined motion vector, and the warped reference frame.
s = S θ s ( C 1 × , v ^ t , W ( x ^ t 1 , v ^ t ) ) ,
where S θ s denotes the neural network with learnable parameters θ s . Then, we map C 1 × to F c , C 1 × using a mapping function ( F c ), and use a tanh function [43] to mask F c , C 1 × as shown in the following equation:
f s c = 0.5 · [ F c , C 1 × + F c , C 1 × tanh ( s ) ] = 0.5 · ( 1 + tanh ( s ) ) F c , C 1 ×   ,
where ⊙ represents element-wise multiplication. On the context encoder side, the Residual Generator module is used to obtain the feature residual ( r t ) by subtracting f s c from the feature extracted from x t .
r t = F t ( x t ) f s c
Thereafter, we feed r t into the context encoder along with the conditional inputs predicted by the Multi-scale Context Fusion module.
y t = Enc ( r t , C 1 × , C 2 × , C 4 × ) ,
where y t represents the latent vector. We refer to this residual encoding method as CMFR Coding.
H ( F t ( x t ) m F c ( x c ) | x c )
In contrast, on the Context Decoder side, we use the Combiner module to obtain the reconstructed features of the target frame. The Context Decoder predicts the reconstructed feature residual ( r ^ t ). Then, we add f s c to r ^ t and apply a residual block to obtain the reconstructed features.
f ^ t = RB 1 ( r ^ t + f s c ) ,
where RB1 denotes the residual block consisting of two convolutional layers with a kernel size of 1 × 1 and a Leaky ReLU [41] activation function. Subsequently, we utilize a convolutional layer to transform f ^ t into the pixel domain, producing x ^ t . This approach allows us to store the reconstructed features ( f ^ t ), which contain abundant temporal context information, in the reference buffer.

3.4. Scaled Feature Fusion

The context encoder transforms only residual information, which is the difference between the target frame and the conditional input, into the latent vector ( y t ). To achieve this, we downsample the input features ( r t ), and concatenate them with the predicted condition features ( C 1 × , C 2 × , and C 4 × ) at each resolution. To more effectively subtract conditional information from the input features, we introduce the SFF module.
First, we predict the fusion features (u) and scales (a) using the input features and the predicted condition features as follows:
u = U θ u ( r n × , C n × ) a = tanh ( A θ a ( r n ×   , C n × ) ) ,
where U θ u and A θ a represent neural networks to predict the fusion features and the scales, respectively. r n × denotes the first features at 1 / n resolution from the context encoder or decoder. These scales (a) operate through an attention mechanism, allowing it to focus on the regions that need to be subtracted effectively. Subsequently, we apply the scales to the fusion features and feed them into a convolutional layer to match the channel size of r n × . Then, we subtract the output of the convolutional layer from r n × , concatenate the result with C n × , and feed the combined features into a residual block.
o 1 = r n × Conv ( u a ) o 2 = DRB ( Cat ( o 1 , C n × ) ) ) ,
where Cat denotes the concatenation operation. DRB represents the depth-wise residual block, as depicted in Figure 2 (D-ResBlk). This module is used in both the context encoder and decoder at 1 / 2 and 1 / 4 resolutions.

4. Experiments

4.1. Implementation Details

To train our model, we use the Vimeo-90K [44] dataset, which consists of approximately 90,000 sequences, each containing seven frames. Each frame has a resolution of 448 × 256 . During the training stages, we randomly crop frames to a size of 256 × 256 and apply random vertical and horizontal flips. Additionally, we randomly reverse the order of the frames in each sequence. We train our model with the multi-stage training strategy used in DCVC-TCM [18]. Unlike DCVC-TCM, we modify the training process in terms of compression quality control. We find that training our model becomes unstable when variable rate control modules are included from the beginning. Therefore, we first train our model at the highest video quality. After training at the highest quality, we incorporate variable rate control modules to enable training across varying compression levels. Then, to mitigate error propagation, we use a cascaded loss [18] to train the entire model on consecutive multi-frames.
In the training stage with multi-frames, we use six P-frames instead of the five that DCVC-TCM used. For the loss function, we use the Lagrange multiplier λ , which balances the quality of the reconstructed frame and the bit rate in a random range of [ 320 , 2600 ] . In addition, to reduce error propagation and maintain compression performance over long temporal sequences, we fine-tune our model using the BVI-DVC [45] dataset. This dataset consists of approximately 800 sequences at various resolutions. We fine-tune our model using 36 consecutive frames with cascaded loss [18] for 4 epochs at a learning rate of 1 × 10 5 . Then, we decrease the learning rate to 1 × 10 6 for 1 epoch. For the MS-SSIM distortion, we use the 1 MS-SSIM ( x ^ t , x t ) distortion function to train our model with λ values from 10 to 81.

4.2. Evaluation

To evaluate the video compression performance of each model, we use the UVG [36] dataset, which consists of seven sequences at a resolution of 1920 × 1080 . Moreover, we use HEVC Class A to E [46], which have resolutions of 2560 × 1600 , 1920 × 1080 , 832 × 480 , 416 × 240 , and 1280 × 720 , respectively. Additionally, we utilize the MCL-JCV [47] dataset. This dataset consists of 30 sequences with a resolution of 1920 × 1080 . For all test datasets, we apply reflection padding to each input sequence to ensure that the width and height of the frames are multiples of 64. We use an intra period of 32 and a total of 96 frames for compression during the test.
To compare the rate-distortion performance of our model, we use HEVC [2], VP9 [3], and VVC [5] for traditional codec methods. In the case of HEVC, we use the HEVC reference software (HM) [48] version 18.0. For VP9, we use FFmpeg [49] version 7.0 with libvpx-vp9 of version 1.14.0. In the case of VVC, we use the VVC test model (VTM) [50] of version 23.4. For neural network-based P-frame video compression methods, we compare our model with DCVC [15], DCVC-TCM [18], DCVC-HEM [19], and DCVC-DC [20]. To evaluate rate-distortion performance, we measure PSNR and bits per pixel (BPP) and plot these values on a graph. Therefore, the graph on the left is a better model, as it uses fewer bits for the same visual quality. Furthermore, we utilize the BD-rate [51] to confirm the average bit savings achieved at the same visual quality level.

4.3. Rate-Distortion Performance

Figure 3 and Figure 4 show the rate-distortion curves. The x-axis and y-axis represent the bpp and PSNR, respectively. Our model outperforms all traditional video codec methods, DCVC, DCVC-TCM, and DCVC-HEM, across all test sequences. Moreover, our model demonstrates higher compression performance in most sequences compared to DCVC-DC. Table 2 shows the BD-rate using VTM as an anchor with the PSNR distortion function, the total number of parameters, and the encoding and decoding time. To measure inference time, we use sequences with a resolution of 1920 × 1080 and average the results to obtain milliseconds per frame. For GPU measurements, we use the RTX A6000 Ada, while we utilize Ryzen 5 7500F for the CPU. The results indicate that our model demonstrates better compression performance than other models across most of the test sequences. Furthermore, compared to the DCVC-DC model, our model shows higher performance while requiring fewer parameters and less time for decoding. Table 3 represents the BD-rate with the MS-SSIM distortion function. As shown in the table, our model also achieves higher performance than others on most test sequences under the MS-SSIM metric.
Figure 5 presents the qualitative comparison results. As shown in the upper image, our model preserves the pattern of the floor tiles better (blue box) and retains the folding lines of the parasol more clearly (red box). In the lower image, the texture of the girl’s hair is better preserved in our result (blue box). Moreover, the dot below the number 4 on the die is clearer than in the results of the others. In addition, the structure of the wrinkles on the orange bag is better maintained (green box).

5. Discussion

5.1. Conditional Masked Feature Residual Coding

We compare our proposed method, CMFR Coding, with other Conditional Coding approaches. Figure 6 shows the rate-distortion curves of various coding methods. “Base_NoRefFeat” denotes the Conditional Coding model that uses reference frames instead of reference features. Therefore, this model fails to utilize abundant temporal context information. By applying Conditional Residual Coding or Masked Conditional Residual Coding to this base model, we find that the performance improves by 0.92% and 1.99%, respectively. When we apply our CMFR Coding to the base model, we observe a significant performance improvement of 3.46%. “Base” represents the base model, which uses reference features to leverage abundant temporal context information. Conditional Residual Coding and Masked Conditional Residual Coding cannot utilize this base model structure. In contrast, applying our method to both the “Base” and “Base_NoRefFeat” models demonstrates that CMFR Coding consistently improves rate-distortion performance. Table 4 and Figure 7 illustrate the performance improvements achieved by applying our proposed modules, along with the increases in model parameters and encoding/decoding times. Compared to base model A, model B from Table 4 achieves a performance improvement of 3.42% with a relatively small increase in the number of parameters. However, since CMFR Coding is applied at the original resolution, the inference time increases slightly.
Figure 8 shows that the masks of our CMFR Coding are applied differently for each channel. (a) in Figure 8 shows the target frame, while (b) depicts the predicted features, which are F c , C 1 × in Equation (5). (c) illustrates the masks, defined as 0.5 · ( 1 + tanh ( s ) ) in Equation (5). Finally, (d) presents the masked predicted features, denoted as f s c in Equation (5). These results indicate that the mask values change significantly at the boundaries of the moving objects. Moreover, although the spatial location is the same, the masks are applied differently across channels. For the same spatial location, some channels activate with high mask values, while others deactivate with low mask values.

5.2. Motion Refiner

Figure 9 visualizes the results of the Motion Refiner module. This module refines the decoded motion vector using the reference features. (a) shows the decoded motion vector ( v ˜ t ) obtained through the decoder, and (b) depicts the refined motion vector ( v ^ t ). In (a), the edges of the moving object appear blurred. In contrast, the edges of the refined motion vector are clear and sharp. The rate-distortion plot in Figure 7 shows that the Motion Refiner module consistently improves video compression performance across the entire quality range. Furthermore, as shown in Table 4, this module achieves a performance improvement of 2.9%, while it requires only an additional 0.7 M parameters and increases decoding time by approximately 23 ms.
Figure 10 demonstrates how the neural network refines the motion vector using reference features. (a) and (b) represent the frames with index 0 and 60, respectively. Moreover, (c) and (d) depict the refined motion vectors for frame index 61. (c) uses the reference feature from (a), while (d) utilizes the reference feature from (b). Even if the structure of the target motion is not aligned with the reference frame, the edge shapes remain considerably similar. The neural network can use this characteristic to refine the edge or overall shape of the motion vector. As shown in Figure 10, the motion vector of frame index 61 has a similar edge shape with the previous frame (b). Therefore, (d) demonstrates a sharper edge and more accurate structure.
Figure 11 shows the results of predicting Δ v t with reference features ( f ^ t 1 ) or reference frames ( x ^ t 1 ) in the Motion Refiner module. Using f ^ t 1 to predict Δ v t results in significantly better performance compared to x ^ t 1 . Therefore, we use reference features for motion refinement, as they are more effective.

5.3. Scaled Feature Fusion

As shown in Figure 7 and Table 4, the SFF module increases performance by 5.44%, with a parameter increase of about 1.14 M. Furthermore, Table 5 demonstrates that the SFF module effectively improves rate-distortion performance. This table shows the performance with the SFF module (“w/SFF”) and when it is replaced by three consecutive depth-wise residual blocks (“w/3 D-ResBlks”). These results indicate that the three consecutive residual blocks, which use more parameters than the SFF module, result in 1.46% performance increase. In contrast, our SFF module demonstrates 5.87% performance improvement while using fewer parameters.
We visualize the intermediate features from the SFF module. To do this, we select the first SFF module in the context encoder and define two points for visualization: point A (red) and point B (purple), as shown in the SFF module block diagram in Figure 2. Then, we freeze the pre-trained neural network and train only an additional decoder that transforms the features at points A and B into the pixel domain. This decoder consists of three and four residual blocks for points A and B, respectively, followed by a convolutional layer and a pixel shuffle layer. Using this additional decoder, we can visualize what information has been removed by transforming the intermediate features into the pixel domain. Through this process, we confirm whether the conditional information is actually subtracted from the target feature r n × through the feature subtraction operation. Figure 12a shows a PSNR of 22.32 dB at point A of the SFF module in the pixel domain. After the subtraction operation with r n × , the PSNR is reduced to 18.36 dB at point B, as shown in (b). Moreover, as illustrated in Figure 12, the colors are distorted and many details are lost. These results confirm that the SFF module subtracts conditional information from the target feature r n × through the subtraction operation.

5.4. Error Propagation

In this section, we evaluate the effect of error propagation and verify the effectiveness of using reconstructed features as references and fine-tuning the model with 36 consecutive frames. In the case of Residual Coding in the pixel domain, the decoder reconstructs the residual ( r ^ ) and then adds it to the predicted frame ( x c ) to obtain the reconstructed frame ( x ^ t = r ^ + x c ). However, x ^ t , which has only three channels, loses abundant temporal context information. To cope with this problem, DCVC-TCM [18] stores reconstructed features in the buffer, which contain richer temporal information across multiple channels.
To verify this, we adapt the evaluation method for error propagation utilized in DCVC-FM [21]. We use “KristenAndSara” from the HEVC test sequence class E [46] as the test sequence. Only the first frame is set as an intra frame, and the remaining 499 frames are set as P frames to induce error propagation. We evaluate the “Base” and “Base_NoRefFeat” models from Figure 6, which use reconstructed features and frame as reference, respectively. Figure 13a represents the effect of the reference type on the intensity of error propagation. The x- and y-axes represent the frame index and the PSNR value, respectively. Moreover, “Ref_Features” and “Ref_Frame” correspond to the “Base” and “Base_NoRefFeat” models, respectively. The graphs labeled “moving_avg” indicate that a moving average with a window size of 16 is applied to the PSNR values along the frame index axis. As shown in Figure 13a, “Ref_Features” shows a decrease in PSNR until around frame 100, and then it remains stable in visual quality. In contrast, the PSNR of “Ref_Frame” continues to decrease, indicating more severe error propagation.
Moreover, the Vimeo-90K [44] dataset consists of seven frames per sequence, allowing a maximum of six consecutive P-frames to be used for training, as the first frame is used as the I-frame. This leads to a significant gap between training and test settings. Therefore, it is necessary to train the network under similar conditions to the test environment in order to effectively mitigate error propagation in long-term sequences. In Figure 13b, “w/o Finetune” refers to the model trained only on the Vimeo-90k dataset, while “w/Finetune” indicates the result of further fine-tuning this model using 36 consecutive frames from BVI-DVC [45]. As shown in this figure, “w/Finetune” shows stable PSNR from around frame 100, while “w/o Finetune” shows a continuous decrease in PSNR. This observation demonstrates that reducing the gap between the training and test environments effectively mitigates error propagation.

5.5. Limitations and Future Works

Although our proposed methods show performance improvements in terms of rate-distortion, several limitations remain. The first limitation is the memory size of the Decoded Picture Buffer (DPB). In the case of Conditional Residual and Masked Conditional Residual Coding, only 4 × H × W × 3 bytes are required for the size of the DPB when using 32-bit floating point. In contrast, feature-based method, including our CMFR Coding, store both the reconstructed frame and the feature maps. Therefore, when the channel size is 48, the required memory size is H × W × ( 3 + 48 ) × 4 bytes, which is approximately 17 times larger than that of frame buffering methods. In terms of latency, there is no significant difference between storing the reconstructed frame and feature maps in terms of neural network structure. This is because both approaches transform the reconstructed frame or feature map into another feature domain using a convolutional layer and use it for motion compensation. In the case of data transfer over the bus, we measure the time required to transfer frame and feature maps from the CPU to GPU over 100 iterations. The total transfer time for frames is 0.1163 s, while the feature maps required 1.3621 s. This indicates that feature map transfer is approximately 11 times slower. However, if the reconstructed frame or feature is stored in the same device buffer as the neural network, the reference data fetching speed becomes much faster than the values we measured.
The second limitation is the large memory requirement in inference. We analyze how much GPU memory is required for decoding at commonly used resolutions. A resolution of 1280 × 720 , which contains 921 K pixels, requires approximately 8.35 GB of video memory. For 1920 × 1080 (2073 K pixels), the memory usage increases to about 18.24 GB, and for 2560 × 1600 (4096 K pixels), it requires 34.81 GB. In the case of 4 K resolution, which is now widely adopted, approximately 70 GB of GPU memory is required, presenting a challenge for deploying neural video compression models to edge devices. To address this issue, one of the most straightforward approaches is to reduce the bit width of the data. Replacing 32-bit floating-point precision with 16-bit or 8-bit, along with fine-tuning [52], can significantly reduce both memory usage and inference time. Additionally, techniques such as pruning [53] could be utilized in future work to further reduce memory and computational costs.
The third limitation is the long decoding time. Table 2 presents a comparison of encoding and decoding time as well as KMACs between the proposed model, other SoTA neural video compression models, and traditional codecs. For traditional codecs, unlike the neural network models, KMACs vary depending on the content, so we measure only CPU-based inference time. As shown in the Table 2, our model achieves faster encoding speed than VVC when using either GPU or CPU. However, the decoding speed is approximately 10 times slower than VVC, even with GPU acceleration. Moreover, we observe that although DCVC-DC and our model show lower KMACs than DCVC-HEM and DCVC-TCM, their inference times are longer. This is due to the network architecture, which utilize numerous 1 × 1 convolutional layers and depth-wise convolutional layers, resulting in deeper networks. These operations have relatively low computational complexity individually. However, when they are stacked and proceed sequentially, the total inference time can significantly increase, as such a sequential process cannot benefit from parallelism or fast algorithms. As a result, the inference time becomes longer despite the lower KMACs.
This long inference time leads to a significant challenge for using neural video compression models in practical environments, despite their superior performance. Therefore, reducing decoding time is critical for future works. One of the main bottlenecks in decoding is arithmetic coding performed on the CPU and data transfer between the CPU and GPU. They account for approximately 20% (65.34 ms per frame) of the total decoding time. To address this, future work could focus on performing arithmetic coding directly on the GPU or designing new architectures that allow arithmetic coding and neural network computation to be processed in parallel.

6. Conclusions

In this study, we propose Conditional Masked Feature Residual (CMFR) Coding. With this approach, we can minimize the residual information between the target frame and the compensated features. For CMFR Coding, we utilize neural networks to extract the diverse and high-level representations from the target frame. Then, we predict masks and apply them to the predicted features to reduce the residual. Moreover, we introduce the Motion Refiner, which enhances the quality of the decoded optical flow. Additionally, we propose the Scaled Feature Fusion (SFF) module to more effectively remove and add conditional information in the context encoder and decoder, respectively. We demonstrate that our methods effectively minimize residual information and achieve a 36.28% performance improvement compared to VVC on the HEVC test sequences. Furthermore, we analyze limitations of the proposed model in terms of inference time, memory usage, and buffer size.

Author Contributions

Conceptualization, C.S. and Y.K.; Methodology, C.S., Y.K. and K.C.; Software, C.S.; Validation, C.S., Y.K. and K.C.; Formal analysis, K.C.; Investigation, Y.K.; Resources, C.S.; Data curation, C.S. and S.L.; Writing—original draft preparation, C.S.; Writing—review and editing, C.S., Y.K., K.C. and S.L.; Supervision, S.L.; Project administration, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Korea Planning & Evaluation Institute of Industrial Technology(KEIT) grant funded by the Korea government (MOTIE) (No. RS-2024-00442120, Development of AI technology capable of robustly recognizing abnormal and dangerous situations and behaviors during night and bad weather conditions). This work was also supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT)(No. RS-2024-00423362).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicity available datasets were analyzed in this study [36,44,45,46,47].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wiegand, T.; Sullivan, G.J.; Bjontegaard, G.; Luthra, A. Overview of the H.264/AVC video coding standard. IEEE Trans. Circuits Syst. Video Technol. 2003, 13, 560–576. [Google Scholar] [CrossRef]
  2. Sullivan, G.J.; Ohm, J.; Han, W.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  3. Mukherjee, D.; Bankoski, J.; Grange, A.; Han, J.; Koleszar, J.; Wilkins, P.; Xu, Y.; Bultje, R. The latest open-source video codec vp9—An overview and preliminary results. In Proceedings of the 2013 Picture Coding Symposium (PCS), San Jose, CA, USA, 8–11 December 2013; pp. 390–393. [Google Scholar]
  4. Han, J.; Li, B.; Mukherjee, D.; Chiang, C.H.; Grange, A.; Chen, C.; Su, H.; Parker, S.; Deng, S.; Joshi, U.; et al. A Technical Overview of AV1. Proc. IEEE 2021, 109, 1435–1462. [Google Scholar] [CrossRef]
  5. Bross, B.; Wang, Y.K.; Ye, Y.; Liu, S.; Chen, J.; Sullivan, G.J.; Ohm, J.R. Overview of the Versatile Video Coding (VVC) Standard and its Applications. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3736–3764. [Google Scholar] [CrossRef]
  6. Lim, D.; Choi, J. FDI-VSR: Video Super-Resolution Through Frequency-Domain Integration and Dynamic Offset Estimation. Sensors 2025, 25, 2402. [Google Scholar] [CrossRef] [PubMed]
  7. Choi, J.; Oh, T.-H. Joint Video Super-Resolution and Frame Interpolation via Permutation Invariance. Sensors 2023, 23, 2529. [Google Scholar] [CrossRef] [PubMed]
  8. Kumari, P.; Keck, S.; Sohn, E.; Kern, J.; Raedle, M. Advanced Imaging Integration: Multi-Modal Raman Light Sheet Microscopy Combined with Zero-Shot Learning for Denoising and Super-Resolution. Sensors 2024, 24, 7083. [Google Scholar] [CrossRef] [PubMed]
  9. Sun, J.; Yuan, Q.; Shen, H.; Li, J.; Zhang, L. A Single-Frame and Multi-Frame Cascaded Image Super-Resolution Method. Sensors 2024, 24, 5566. [Google Scholar] [CrossRef] [PubMed]
  10. Lu, M.; Zhang, P. Real-World Video Super-Resolution with a Degradation-Adaptive Model. Sensors 2024, 24, 2211. [Google Scholar] [CrossRef] [PubMed]
  11. Tsai, F.J.; Peng, Y.T.; Tsai, C.C.; Lin, Y.Y.; Lin, C.W. Banet: A blur-aware attention network for dynamic scene deblurring. IEEE Trans. Image Process. 2022, 31, 6789–6799. [Google Scholar] [CrossRef] [PubMed]
  12. Wei, P.; Xie, Z.; Li, G.; Lin, L. Taylor neural network for real-world image super-resolution. IEEE Trans. Image Process. 2023, 32, 1942–1951. [Google Scholar] [CrossRef] [PubMed]
  13. Feng, G.; Han, L.; Jia, Y.; Huang, P. Single Image Defocus Deblurring Based on Structural Information Enhancement. IEEE Access 2024, 12, 153471–153480. [Google Scholar] [CrossRef]
  14. Lu, G.; Ouyang, W.; Xu, D.; Zhang, X.; Cai, C.; Gao, Z. DVC: An end-to-end deep video compression framework. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 10998–11007. [Google Scholar]
  15. Li, J.; Li, B.; Lu, Y. Deep contextual video compression. In Proceedings of the 34th Conference on Neural Information Processing Systems (NIPS), New Orleans, LA, USA, 10–16 December 2021; pp. 1–12. [Google Scholar]
  16. Brand, F.; Seiler, J.; Kaup, A. Conditional Residual Coding: A Remedy for Bottleneck Problems in Conditional Inter Frame Coding. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 6445–6459. [Google Scholar] [CrossRef]
  17. Chen, Y.-H.; Xie, H.-S.; Chen, C.-W.; Gao, Z.-L.; Benjak, M.; Peng, W.-H.; Ostermann, J. MaskCRT: Masked Conditional Residual Transformer for Learned Video Compression. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 11980–11992. [Google Scholar] [CrossRef]
  18. Sheng, X.; Li, J.; Li, B.; Li, L.; Liu, D.; Lu, Y. Temporal Context Mining for Learned Video Compression. IEEE Trans. Multimed. 2023, 25, 7311–7322. [Google Scholar] [CrossRef]
  19. Li, J.; Li, B.; Lu, Y. Hybrid Spatial-Temporal Entropy Modelling for Neural Video Compression. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 1503–1511. [Google Scholar]
  20. Li, J.; Li, B.; Lu, Y. Neural video compression with diverse contexts. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 22616–22626. [Google Scholar]
  21. Li, J.; Li, B.; Lu, Y. Neural video compression with feature modulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 26099–26108. [Google Scholar]
  22. Jin, I.; Xia, F.; Ding, F.; Zhang, X.; Liu, M.; Zhao, Y.; Lin, W.; Meng, L. Customizable ROI-Based Deep Image Compression. arXiv 2025, arXiv:2507.00373. [Google Scholar] [CrossRef]
  23. He, D.; Zheng, Y.; Sun, B.; Wang, Y.; Qin, H. Checkerboard context model for efficient learned image compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 14771–14780. [Google Scholar]
  24. Ballé, J.; Minnen, D.; Singh, S.; Hwang, S.J.; Johnston, N. Variational image compression with a scale hyperprior. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–23. [Google Scholar]
  25. x265 HEVC Encoder/h.265 Video Codec. Available online: https://www.x265.org/ (accessed on 1 July 2025).
  26. Agustsson, E.; Minnen, D.; Johnston, N.; Ballé, J.; Hwang, S.J.; Toderici, G. Scale-space flow for end-to-end optimized video compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 8500–8509. [Google Scholar]
  27. Hu, Z.; Lu, G.; Guo, J.; Liu, S.; Jiang, W.; Xu, D. Coarse-to-fine deep video coding with hyperprior-guided mode prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5911–5920. [Google Scholar]
  28. Rippel, O.; Anderson, A.G.; Tatwawadi, K.; Nair, S.; Lytle, C.; Bourdev, L. ELF-VC: Efficient learned flexible-rate video coding. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 14459–14468. [Google Scholar]
  29. Lin, J.; Liu, D.; Li, H.; Wu, F. M-LVC: Multiple Frames Prediction for Learned Video Compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 3543–3551. [Google Scholar]
  30. Brand, F.; Seiler, J.; Kaup, A. On benefits and challenges of conditional interframe video coding in light of information theory. In Proceedings of the 2022 Picture Coding Symposium (PCS), San Jose, CA, USA, 7–9 December 2022; pp. 289–293. [Google Scholar]
  31. Mentzer, F.; Toderici, G.; Minnen, D.; Hwang, S.J.; Caelles, S.; Lucic, M.; Agustsson, E. VCT: A video compression transformer. In Proceedings of the 36th Conference on Neural Information Processing Systems (NIPS), New Orleans, LA, USA, 10–16 December 2023; pp. 13091–13103. [Google Scholar]
  32. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  33. Lu, M.; Duan, Z.; Zhu, F.; Ma, Z. Deep hierarchical video compression. In Proceedings of the 38th Annual AAAI Conference on Artificial Intelligence (AAAI), Vancouver, BC, Canada, 20–27 February 2024; pp. 8859–8867. [Google Scholar]
  34. Jia, Z.; Li, B.; Li, J.; Xie, W.; Qi, L.; Li, H.; Lu, Y. Towards Practical Real-Time Neural Video Compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 10–17 June 2025; pp. 12543–12552. [Google Scholar]
  35. Jiang, W.; Li, J.; Zhang, K.; Zhang, L. Ecvc: Exploiting non-local correlations in multiple frames for contextual video compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 10–17 June 2025; pp. 7331–7341. [Google Scholar]
  36. Mercat, A.; Viitanen, M.; Vanne, J. UVG dataset: 50/120 fps 4K sequences for video codec analysis and development. In Proceedings of the 11th ACM Multimedia Systems Conference, Istanbul, Turkey, 8–11 June 2020; pp. 297–302. [Google Scholar]
  37. Sheng, X.; Li, L.; Liu, D.; Wang, S. Bi-directional deep contextual video compression. IEEE Trans. Multimed. 2025, 1–15. [Google Scholar] [CrossRef]
  38. Zhai, Y.; Tang, L.; Jiang, W.; Yang, J.; Wang, R. L-LBVC: Long-Term Motion Estimation and Prediction for Learned Bi-Directional Video Compression. In Proceedings of the IEEE Data Compression Conference (DCC), Snowbird, UT, USA, 19–21 March 2025; pp. 53–62. [Google Scholar]
  39. Alexandre, D.; Hang, H.M.; Peng, W.H. Hierarchical B-frame video coding using two-layer CANF without motion coding. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 10249–10258. [Google Scholar]
  40. Ranjan, A.; Black, M.J. Optical Flow Estimation Using a Spatial Pyramid Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2720–2729. [Google Scholar]
  41. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. Proc. ICML 2013, 30, 3. [Google Scholar]
  42. Shi, W.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  43. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  44. Xue, T.; Chen, B.; Wu, J.; Wei, D.; Freeman, W.T. Video enhancement with task-oriented flow. Int. J. Comput. Vis. 2019, 127, 1106–1125. [Google Scholar] [CrossRef]
  45. Ma, D.; Zhang, F.; Bull, D.R. BVI-DVC: A Training Database for Deep Video Compression. IEEE Trans. Multimed. 2022, 24, 3847–3858. [Google Scholar] [CrossRef]
  46. Bossen, F. Common test conditions and software reference configurations. In Proceedings of the 3rd. JCT-VC Meeting, Guangzhou, China, 7–15 October 2013; Volume 12. [Google Scholar]
  47. Wang, H.; Gan, W.; Hu, S.; Lin, J.Y.; Jin, L.; Song, L.; Wang, P.; Katsavounidis, I.; Aaron, A.; Kuo, C.C.J. MCL-JCV: A JND-based H.264/AVC video quality assessment dataset. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 1509–1513. [Google Scholar]
  48. Joint Video Expoloration Team (JVET) of ITU-T VCEG and ISO/IEC MPEG. HEVC Reference Software. Version 18.0. Available online: https://vcgit.hhi.fraunhofer.de/jvet/HM/-/tree/HM-18.0?ref_type=tags (accessed on 17 January 2025).
  49. FFmpeg. Version 7.0. Available online: https://github.com/FFmpeg/FFmpeg/tree/release/7.0 (accessed on 4 September 2024).
  50. Joint Video Expoloration Team (JVET) of ITU-T VCEG and ISO/IEC MPEG. VVC Reference Software. Version 23.4. Available online: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftware_VTM/-/tree/VTM-23.4?ref_type=tags (accessed on 21 October 2024).
  51. Bjøntegaard, G. Calculation of average PSNR differences between RDcurves. In Proceedings of the Thirteenth Meeting, Austin, TX, USA, 2–4 April 2001. Tech. Rep., VCEG-M33, ITU-T SG16/Q6. [Google Scholar]
  52. Jacob, B.; Kligys, S.; Chen, B.; Zhu, M.; Tang, M.; Howard, A.; Adam, H.; Kalenichenko, D. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2704–2713. [Google Scholar]
  53. Han, S.; Pool, J.; Tran, J.; Dally, W. Learning both weights and connections for efficient neural network. In Proceedings of the 28th Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
Figure 1. Upper Left: Target and Predict signals. Bottom Left: Decomposing each signal into low and high frequencies. To obtain the low-frequency signals, t 1 and p 1 , we utilize a mean filter, while we subtract the low-frequency signals from the original signals to obtain the high-frequency signals, t 2 and p 2 .
Figure 1. Upper Left: Target and Predict signals. Bottom Left: Decomposing each signal into low and high frequencies. To obtain the low-frequency signals, t 1 and p 1 , we utilize a mean filter, while we subtract the low-frequency signals from the original signals to obtain the high-frequency signals, t 2 and p 2 .
Sensors 25 04460 g001
Figure 2. Overall structure of the network. The module diagrams colored in red, blue, and yellow represent the structures of our proposed Conditional Masked Feature Residual (CMFR) Coding, Scaled Feature Fusion (SFF) module, and Motion Refiner, respectively. “Conv” without any additional mark refers to a 1 × 1 convolutional layer, while “Conv, n c ” represents the 1 × 1 convolutional layer with output channel size multiplied by a factor of n. “DWConv” denotes the depth-wise convolutional layer with a kernel size of 3 × 3 . “ResBlk k × k ” denotes the residual block consisting of two k × k convolutional layers with a LeakyReLU [41].
Figure 2. Overall structure of the network. The module diagrams colored in red, blue, and yellow represent the structures of our proposed Conditional Masked Feature Residual (CMFR) Coding, Scaled Feature Fusion (SFF) module, and Motion Refiner, respectively. “Conv” without any additional mark refers to a 1 × 1 convolutional layer, while “Conv, n c ” represents the 1 × 1 convolutional layer with output channel size multiplied by a factor of n. “DWConv” denotes the depth-wise convolutional layer with a kernel size of 3 × 3 . “ResBlk k × k ” denotes the residual block consisting of two k × k convolutional layers with a LeakyReLU [41].
Sensors 25 04460 g002
Figure 3. Rate-distortion performance comparison for different test sequences against several models.
Figure 3. Rate-distortion performance comparison for different test sequences against several models.
Sensors 25 04460 g003
Figure 4. Rate-distortion performance comparison for different test sequences against several models.
Figure 4. Rate-distortion performance comparison for different test sequences against several models.
Sensors 25 04460 g004
Figure 5. Qualitative comparison results with other SoTA neural video compression models and traditional codec. The values under the images indicate PSNR and BPP, respectively.
Figure 5. Qualitative comparison results with other SoTA neural video compression models and traditional codec. The values under the images indicate PSNR and BPP, respectively.
Sensors 25 04460 g005
Figure 6. Ablation study results of various types of Conditional Coding on all HEVC test sequences. “NoRefFeat” indicates that we use the reference frame instead of features. “CMFR” denotes the proposed Conditional Masked Feature Residual Coding. “MCR” and “CR” represent Masked Conditional Residual Coding and Conditional Residual Coding, respectively.
Figure 6. Ablation study results of various types of Conditional Coding on all HEVC test sequences. “NoRefFeat” indicates that we use the reference frame instead of features. “CMFR” denotes the proposed Conditional Masked Feature Residual Coding. “MCR” and “CR” represent Masked Conditional Residual Coding and Conditional Residual Coding, respectively.
Sensors 25 04460 g006
Figure 7. Ablation study result of each module on all HEVC test sequences. CMFR denotes the Conditional Masked Feature Residual Coding, and SFF represents the Scaled Feature Fusion module.
Figure 7. Ablation study result of each module on all HEVC test sequences. CMFR denotes the Conditional Masked Feature Residual Coding, and SFF represents the Scaled Feature Fusion module.
Sensors 25 04460 g007
Figure 8. Visualization results of the masks and the masked predicted features from the Masked Prediction module from Figure 2 and Equation (5).
Figure 8. Visualization results of the masks and the masked predicted features from the Masked Prediction module from Figure 2 and Equation (5).
Sensors 25 04460 g008
Figure 9. (a) Reconstructed motion vector, which is the output of the motion decoder. (b) Refined motion vector.
Figure 9. (a) Reconstructed motion vector, which is the output of the motion decoder. (b) Refined motion vector.
Sensors 25 04460 g009
Figure 10. Motion refining analysis. (a) and (b) represent the frames with index 0 and 60, respectively. (c,d) depict the refined motion vectors for frame index 61. (c) uses the reference feature from (a), while (d) utilizes the reference feature from (b).
Figure 10. Motion refining analysis. (a) and (b) represent the frames with index 0 and 60, respectively. (c,d) depict the refined motion vectors for frame index 61. (c) uses the reference feature from (a), while (d) utilizes the reference feature from (b).
Sensors 25 04460 g010
Figure 11. Rate-distortion results for motion refinement with reference features (w/Ref_Features) or reference frame (w/Ref_Frame) on all HEVC test sequences.
Figure 11. Rate-distortion results for motion refinement with reference features (w/Ref_Features) or reference frame (w/Ref_Frame) on all HEVC test sequences.
Sensors 25 04460 g011
Figure 12. Visualization of features from the SFF module. We transform the features at point A (red) and B (purple) from the SFF module block diagram in Figure 2 into the pixel domain using additional decoder. (a) shows the feature at point A, and (b) shows the feature at point B, both transformed into the pixel domain.
Figure 12. Visualization of features from the SFF module. We transform the features at point A (red) and B (purple) from the SFF module block diagram in Figure 2 into the pixel domain using additional decoder. (a) shows the feature at point A, and (b) shows the feature at point B, both transformed into the pixel domain.
Sensors 25 04460 g012
Figure 13. Visualization of error propagation effects. (a): Comparison between using reconstructed frames or features as reference. (b): Evaluating the effectiveness of fine-tuning on 36 consecutive frames using the BVI-DVC [45] dataset.
Figure 13. Visualization of error propagation effects. (a): Comparison between using reconstructed frames or features as reference. (b): Evaluating the effectiveness of fine-tuning on 36 consecutive frames using the BVI-DVC [45] dataset.
Sensors 25 04460 g013
Table 1. Summary of notations and terminology used in this paper.
Table 1. Summary of notations and terminology used in this paper.
SymbolFull NameExplanation
x t Target frameThe current frame to be encoded.
x c Compensated frame/PredictionThe frame predicted from previous frames with motion compensation.
mMaskA soft mask that determines how much of the predicted pixel or features to retain.
x ^ t Reconstructed frameThe decoded current frame.
F ^ t Reconstructed featuresThe decoded current features. This can be mapped to x ^ t with a neural network.
C N × Conditional input 1 / N resolution features predicted by the reference frame or features with motion vector.
x ^ t 1 Reference frameThe previously decoded frame used to encode the target frame.
f ^ t 1 Reference featuresThe previously decoded features used to encode the target frame.
Table 2. Rate-distortion performance comparison with other models across various test sequences on PSNR distortion. Inference time is measured using sequences with a resolution of 1920 × 1080 .
Table 2. Rate-distortion performance comparison with other models across various test sequences on PSNR distortion. Inference time is measured using sequences with a resolution of 1920 × 1080 .
ModelDatasetVP9HMVTMDCVCDCVC-TCMDCVC-HEMDCVC-DCOurs
BD-rate (%)UVG54.8140.820.00168.370.27−19.74−35.06−37.90
HEVC-A48.7736.670.0085.81−14.66−27.74−39.67−41.09
HEVC-B65.0644.720.00131.66−3.56−18.49−32.07−33.13
HEVC-C40.9930.760.00162.1633.0411.53−24.07−27.48
HEVC-D39.0829.830.00136.3012.97−1.97−37.01−41.37
HEVC-E67.0545.010.00285.2413.431.15−33.68−30.08
HEVC-avg46.6935.650.00146.6110.21−6.38−34.28−36.28
MCL-JCV57.7442.940.00139.6310.40−13.82−29.35−27.34
Parameters (M)---7.9410.7117.5219.7819.48
GPU Enc Time (ms/frame)---2438.34374.15318.33464.75482.93
GPU Dec Time (ms/frame)---12,345.73230.37192.01359.74334.15
CPU Enc Time (ms/frame)307.898357.6237,905.3117,831.3626,427.8923,015.4835,378.4337,520.83
CPU Dec Time (ms/frame)6.4130.7145.816,446.4223,247.0319,253.7127,824.4626,073.00
KMACs/pixel (Enc)---1180.381750.621700.871360.051490.21
KMACs/pixel (Dec)---776.26933.101270.31930.52918.71
Table 3. Rate-distortion performance comparison with other models across various test sequences on MS-SSIM distortion.
Table 3. Rate-distortion performance comparison with other models across various test sequences on MS-SSIM distortion.
ModelDatasetVP9HMVTMDCVCDCVC-TCMDCVC-HEMDCVC-DCOurs
BD-rate (%)UVG32.2318.680.0057.97−12.22−36.08−40.44−42.07
HEVC-A35.2224.730.0071.61−0.15−31.46−38.18−37.25
HEVC-B33.2423.190.0055.32−15.12−45.74−51.15−52.46
HEVC-C28.3522.070.0040.13−17.88−41.57−53.48−55.61
HEVC-D30.125.160.0018.73−31.02−54.20−61.91−63.79
HEVC-E32.0828.490.00104.56−28.67−54.57−59.83−57.32
HEVC-avg32.8423.410.0062.48−25.4−41.29−53.41−56.41
MCL-JCV31.5727.630.0029.06−11.77−36.61−48.75−46.03
Table 4. Ablation study results of each proposed module on all HEVC test sequences. Inference time is measured using sequences with a resolution of 1920 × 1080 on the RTX A6000 ada.
Table 4. Ablation study results of each proposed module on all HEVC test sequences. Inference time is measured using sequences with a resolution of 1920 × 1080 on the RTX A6000 ada.
ABCD
CMFR
MotionRefiner
SFF
BD-rate (%)0.0−3.42−6.32−11.76
Parameters (M)17.4117.6418.3419.48
GPU Enc Time (ms/frame)332.79398.66411.18482.93
GPU Dec Time (ms/frame)212.48275.73298.03334.15
Table 5. Ablation study result of Scaled Feature Fusion module on all HEVC test sequences. “w/SFF” represents the model with the SFF module, while “w/o SFF” denotes the model without it. In addition, “w/3 D-ResBlks” indicates the model where the SFF module is replaced with three consecutive depth-wise residual blocks.
Table 5. Ablation study result of Scaled Feature Fusion module on all HEVC test sequences. “w/SFF” represents the model with the SFF module, while “w/o SFF” denotes the model without it. In addition, “w/3 D-ResBlks” indicates the model where the SFF module is replaced with three consecutive depth-wise residual blocks.
Modelw/o SFFw/SFFw/3 D-ResBlks
BD-rate (%)0−5.87−1.46
Parameters (M)18.3419.4820.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shin, C.; Kim, Y.; Choi, K.; Lee, S. Masked Feature Residual Coding for Neural Video Compression. Sensors 2025, 25, 4460. https://doi.org/10.3390/s25144460

AMA Style

Shin C, Kim Y, Choi K, Lee S. Masked Feature Residual Coding for Neural Video Compression. Sensors. 2025; 25(14):4460. https://doi.org/10.3390/s25144460

Chicago/Turabian Style

Shin, Chajin, Yonghwan Kim, KwangPyo Choi, and Sangyoun Lee. 2025. "Masked Feature Residual Coding for Neural Video Compression" Sensors 25, no. 14: 4460. https://doi.org/10.3390/s25144460

APA Style

Shin, C., Kim, Y., Choi, K., & Lee, S. (2025). Masked Feature Residual Coding for Neural Video Compression. Sensors, 25(14), 4460. https://doi.org/10.3390/s25144460

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop