Next Article in Journal
Study of Thermodynamic Properties of Ks0, Λ, Ξ, and d/d_ Produced in Symmetric Proton–Proton Collisions at SNN = 0.9 TeV and 7 TeV
Next Article in Special Issue
Strictly Chronological CNN Embeddings with Gradient-Boosted Trees for Next-Day Log-Return Forecasting
Previous Article in Journal
Design Method of Small Recreational Vehicle’s Interior Space Based on User Behavior Data Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Video Frame Interpolation for Extreme Motion Scenes Based on Dual Alignment and Region-Adaptive Interaction

1
Shanghai Film Academy, Shanghai University, Shanghai 200072, China
2
Data and Target Engineering, Information Engineering University, Zhengzhou 450000, China
3
College of Mechanical Engineering, Taiyuan University of Technology, Taiyuan 030024, China
4
Shanghai Engineering Research Center of Motion Picture Special Effects, Shanghai 200072, China
*
Authors to whom correspondence should be addressed.
Symmetry 2025, 17(12), 2097; https://doi.org/10.3390/sym17122097
Submission received: 1 November 2025 / Revised: 28 November 2025 / Accepted: 3 December 2025 / Published: 6 December 2025
(This article belongs to the Special Issue Symmetry in Artificial Intelligence and Applications)

Abstract

Video frame interpolation in ultra-high-definition extreme motion scenes remains highly challenging due to large displacements, nonlinear motion, and occlusions that disrupt spatio-temporal symmetry. To address this issue, this study proposes a frame interpolation method for extreme motion scenes based on dual alignment and region-adaptive interaction from the perspectives of cross-frame localization and adaptive reconstruction. Specifically, we design a two-stage motion information alignment strategy that obtains two types of motion information via optical flow estimation and offset estimation, and it progressively guides reference pixels for accurate long-range cross-frame localization, mitigating structural misalignment caused by limited receptive fields while simultaneously alleviating spatiotemporal asymmetry caused by inconsistent inter-frame motion speed and direction. Based on this, we introduce a region-adaptive interaction module that automatically adapts motion representations for different regions through cross-frame interaction and leverages distinct attention pathways to accurately capture both the global context and local high-frequency motion details. This achieves a dynamic feature fusion tailored to regional characteristics, significantly enhancing the model’s ability to perceive the overall structure and texture details in extreme motion scenarios. In addition, the introduction of a motion compensation module explicitly captures pixel motion relationships by constructing a global correlation matrix that compensates for the positioning errors of the dual alignment module in extreme motion or occlusion areas. The experimental results demonstrate that the proposed method achieves excellent overall performance in ultra-high-definition extreme motion scenes, with a PSNR improvement of 0.05 dB over state-of-the-art methods. In multi-frame interpolation tasks, it achieves an average PSNR gain of 0.31 dB, demonstrating strong cross-scene interpolation capability.

1. Introduction

Video frame interpolation aims to increase the frame rate of a video by synthesizing new frames within consecutive video sequences. This can effectively mitigate common issues in low-quality videos, such as jitter and blur, and it is therefore widely used to balance the video coding rate and quality [1], support view synthesis [2], enhance video quality [3], and allow video deblurring [4,5].
The massive generation of ultra-high-definition (UHD) video (4K or 8K) data poses a significant challenge to current deep-learning-based video frame interpolation methods. In [6], Sim et al. first introduced the concept of extreme motion in the context of video frame interpolation to describe fast, large-magnitude pixel movements present in UHD videos. Extreme motion refers to high-speed dynamic changes in which pixel displacements far exceed the typical distribution found in conventional video datasets. Such motions often extend beyond the receptive field of deep neural networks, making it difficult for models to capture long-range pixel correspondences. Meanwhile, these scenarios are frequently accompanied by complex occlusions and nonlinear lighting variations, which disrupt the temporal symmetry and motion symmetry that should exist between the input frames I 0 and I 1 and the intermediate frame I 0.5 , thereby imposing stricter demands on the spatiotemporal modeling capabilities of existing video frame interpolation methods.
Recently, some works [6,7,8] have proposed methods for frame interpolation in 4K videos and extreme motion scenes. However, these approaches [9,10] typically rely on a single optical flow estimation to infer pixel trajectories. When motion magnitudes exceed the network’s receptive field, the models struggle to maintain spatiotemporal symmetry between frames. In particular, under large-scale motion and severe occlusions, errors in optical flow estimation tend to accumulate, resulting in structural misalignment and texture degradation in the reconstructed intermediate frames. As shown in Figure 1, noticeable artifacts and blurring (red arrows) appear in the results of current SOTA methods, along with information loss at occlusion boundaries (yellow arrows).
To address the limitations of existing methods in handling extreme motion scenes in UHD videos, this paper proposes a video interpolation method based on dual alignment and region-adaptive interaction. The framework consists of two stages. In the first stage, we design a dual alignment module (DAM), which simultaneously estimates optical flow and offsets to obtain two types of motion information. These are applied progressively in a stepwise manner to achieve precise alignment of reference pixels between adjacent frames, effectively reducing the difficulty of motion fitting. Unlike mainstream flow-based methods [9,10,11], the DAM aligns the input frames primarily along the temporal dimension, alleviating spatiotemporal asymmetry caused by nonlinear motion speed variations and inconsistent inter-frame motion directions.
In the second stage, inspired by the frequency-aware strategy, we propose a region-adaptive interaction module (RAIM) to jointly model global semantics and local motion features. As shown in Figure 2, we apply the discrete Fourier transform (DFT) to 4K video frames and compute the RGB frequency-domain difference map by subtracting the spectra of consecutive frames. It can be observed that large-motion regions correspond to local high-frequency components, while the overall semantic structure is mainly represented by low-frequency components. This frequency-domain distinction reveals the distribution pattern of motion and structural features in videos. Based on this observation, the RAIM first employs a window partitioning strategy, dividing adjacent frames into regions of different scales to separate local motion from global structural information. Then, a cross-frame interaction strategy establishes adaptive correlations among motion regions at different scales. Next, a region-separated attention mechanism (RSAM) is introduced to learn from these correlated regions, simultaneously capturing global context and local high-frequency motion information. Unlike existing methods [9,12,13,14,15] that rely solely on spatial feature stacking or a single attention mechanism, the RAIM effectively restores the spatiotemporal symmetry disrupted in extreme motion scenarios through frequency-inspired, region-adaptive interactions, significantly enhancing the model’s holistic perception of motion–structure relationships. Additionally, to compensate for errors in pixel localization from the DAM, we design a motion compensation module (MCM), which explicitly computes the displacement vector field between pixels based on a global correlation matrix, providing additional motion information to further improve interpolation accuracy.
Our contributions can be summarized as follows:
  • We propose a novel video frame interpolation method for extreme motion scenes in UHD videos.
  • We design a DAM to reduce pixel localization difficulty through dual motion information, prevent motion error accumulation, and guide spatiotemporal information aggregation across frames.
  • We propose an RAIM that learns different timestamp information through adaptive interaction between motion regions of varying scales in neighboring frames, effectively enhancing the perception of the overall structure and texture details.
  • Through experiments, our method achieves superior interpolation performance in high-resolution, extreme motion scenes and exhibits stronger robustness and stability in cross-scene generalization.

2. Related Work

Existing video frame interpolation methods are primarily categorized into flow-based [16,17,18] and kernel-based methods [12,13].

2.1. Flow-Based Methods

Flow-based methods estimate bidirectional optical flow between input frames to fit the motion trajectories of pixels in interframes. A representative method is SuperSloMo [16], which employs a predefined linear motion assumption to resample bidirectional optical flow between input frames, thereby obtaining optical flow at intermediate time points. To better model complex real-world motion, several studies improve inter-frame motion trajectories. For example, QVI [17] performs interpolation using quadratic trajectories constructed from acceleration information between video frames. Other works introduce coarse-to-fine optical flow estimation frameworks to adaptively handle motions of varying magnitudes [18,19,20]. Jin et al. [19] propose EBME, which employs enhanced bidirectional motion estimation to achieve a flexible and compact network design with only 3.9 M parameters; building on this, they further propose UPR-Net [20], which unifies motion estimation and frame synthesis within a single pyramid recurrent network, realizing a coarse-to-fine frame synthesis process. To improve computational efficiency, Huang et al. [11] propose RIFE, which uses IFNet, a network based on convolution and deconvolution operations, to estimate intermediate flow in an end-to-end and real-time manner. EMA-VFI [9] employs a hybrid CNN-Transformer architecture and designs an inter-frame attention mechanism to simultaneously capture motion and appearance information, providing an important reference for subsequent work [21]. For large motion scenes, Liu et al. [10] proposed the SGM-VFI, which detects defective regions in intermediate optical flow through sparse global matching and demonstrates excellent performance on large motion benchmark datasets.
Although these methods achieve significant progress, in extreme motion scenes, the large motion magnitude causes single optical flow estimation to easily accumulate errors, making it difficult to accurately model pixel trajectories and, thereby, adversely affecting the frame rendering quality.

2.2. Kernel-Based Methods

The pioneering work in this category is AdaConv, an adaptive convolution kernel proposed by Niklaus et al. [22] in 2017. Its core concept involves performing convolution operations on local pixels of two input frames through an end-to-end DNN model, without requiring additional optical flow input. To reduce model parameters, Niklaus et al. [12] further decompose the 2D convolution kernels into pairs of 1D kernels, resulting in SepConv. However, both AdaConv and SepConv rely on fixed-shape kernels, which limits their ability to handle motions that exceed the predefined kernel size. To address this challenge, Lee et al. [13] propose AdaCoF, a general and flexible transformation module based on Deformable Convolution (DConv) [23], which adaptively estimates multiple offsets for each target pixel. Inspired by DConv, Cheng et al. introduce DSepConv [24] and EDSC [25]. To further expand the sampling range on the input frames, Shi et al. [26] propose GDConvNet based on generalized deformable convolution. To address the issue of large parameter counts in AdaCoF, Ding et al. [27] proposed the compression-driven optimization scheme CDFI. In addition, some works introduce a Vision Transformer (ViT) [28] into video frame interpolation [29,30], effectively addressing the issue of limited receptive fields in CNNs.
Overall, while kernel-based methods can prevent the accumulation of optical flow errors, they still exhibit significant limitations in capturing the global dependencies of inter-frame pixels in extreme motion scenes.

3. UHD4K120FPS-N Dataset

To address the task of frame interpolation for extreme motion scenes in UHD video, we construct a 4K@120fps dataset named UHD4K120FPS-N. Specifically, we capture a total of 300 HDR raw video clips using a SONY Alpha1 camera (Tokyo, Japan), with each clip lasting 5 to 20 s, covering diverse real-world extreme motion scenarios. To ensure data quality, low-quality videos exhibiting blur or defocus are first manually filtered out. Then, scene segmentation is performed on the raw videos using shot detection and background similarity removal methods [31], splitting them into coherent frame sequences. Each sequence is sampled in groups of nine consecutive frames (frame1.png to frame9.png), with each group treated as an independent scene.
Due to GPU memory limitations, the raw 4K videos are temporarily downsampled to a resolution of 1920 × 1080 for optical flow computation using the RAFT-large algorithm [32]. This preprocessing step is solely for motion analysis and scene filtering, without affecting the fidelity of the original 4K data. After calculating the optical flow magnitude between the first and last frames (equivalent to the average optical flow magnitude between input frames in 30 fps), sequences with an average flow magnitude below 3 pixels are removed, resulting in 145 valid samples. Fifteen sequences are randomly selected as test samples based on the percentile distribution, with the remaining sequences being used for training. Training samples are randomly cropped multiple times at a size of 128 × 128 in the same spatial position across 9 consecutive frames. Each sample is required to maintain an average optical flow magnitude of no less than 3 pixels between neighboring frames, with outliers removed. This yielded 50,400 training samples, denoted as UHD-N-Train, with the average optical flow magnitude between neighboring frames ranging from 3 to 164.9 pixels. The test samples, denoted as UHD-N-Test, exhibit an average optical flow magnitude between neighboring frames ranging from 8.5 to 244.6 pixels.
Figure 3 shows some samples from the UHD4K120FPS-N dataset and their corresponding optical flow magnitudes. Table 1 compares the optical flow magnitude and high-frequency feature ratio between the benchmark datasets and our dataset. The high-frequency feature ratio measures the frequency-domain information of regions with large pixel displacements. It is quantified by calculating the proportion of energy in the high-frequency regions of the image power spectrum, where regions with a distance from the spectrum center exceeding 0.2 are defined as high-frequency regions. Higher values indicate that the image contains more details and textures. The results show that the UHD4K120FPS-N dataset exhibits higher optical flow magnitudes and high-frequency feature ratios, demonstrating its clear advantage for training models on extreme motion scenes.

4. Methods

This task aims to synthesize the intermediate frame I t ( t = 0.5 ) from two given RGB frames I 0 , I 1 R H × W × 3 . As illustrated in Figure 4, the overall framework consists of three main components: low-level feature extraction (LFE), inter-frame feature interaction perception (IFIP), and predicted frame reconstruction (PFR). The proposed method adopts a hybrid structure of a CNN and Transformer, consisting of three convolutional layers and two Transformer layers, and the entire process is divided into four stages. In the first stage, the CNN extracts low-level features F l i . In the second stage, F l i is simultaneously fed into the optical flow estimation network (OFENet) and the DAM. OFENet performs bidirectional optical flow estimation to obtain f 0 1 i and f 1 0 i , while the DAM utilizes flow and offset information to achieve precise alignment of long-range reference pixels across frames. In the third stage, the aligned features F i a l i g n ( i = 1 , 2 ) are passed into the interaction perception Transformer block (IPTBlock), which is built on the Transformer architecture and integrates the RAIM and the MCM to enhance cross-frame feature fusion and supplement additional motion information. In the fourth stage, F s t i , F m i , and F l i are fused and input into PFR, where the intermediate frame I 0.5 is generated through motion estimation and a refinement network.

4.1. Dual Alignment Module

To mitigate the spatiotemporal asymmetry caused by large-magnitude nonlinear motion, th eDAM adopts a coarse-to-fine dual alignment strategy, and its overall structure is illustrated in Figure 5a. In the first stage, the optical flow alignment module (OFAM) warps cross-frame features using bidirectional optical flow, providing a temporally consistent coarse alignment that partially alleviates frame misalignment caused by long-range nonlinear displacements. In the second stage, the offset alignment module (OAM) estimates spatial offsets based on local pixel cosine similarity, further refining pixel positions and correcting local inaccuracies introduced by optical-flow warping to achieve fine-grained alignment. This hierarchical design allows the OFAM to handle coarse temporal alignment while the OAM mitigates residual spatial deviations. The dual-alignment mechanism ultimately provides a stable feature alignment foundation for subsequent attention computations, effectively alleviating spatiotemporal asymmetry caused by nonlinear motion and enhancing structural consistency in extreme motion scenes.

4.1.1. Optical Flow Alignment Module

For clarity, this section takes the alignment from input frame I 0 to I 1 as an example. Given two input features F 0 l , F 1 l R C × H × W , we first construct an empty sampling grid g ( i , j ) according to the tensor shape of F 0 l . Adding g ( i , j ) to the optical flow field f 0 1 yields the target sampling coordinates. For a pixel p ( i , j ) on the feature map, its corresponding position in F 1 l is p ( i + Δ x i , j , j + Δ y i , j ) , where ( Δ x , Δ y ) denotes the optical flow magnitude at position ( i , j ) in F 0 l . The corresponding sampling position in F 1 l for any pixel on the entire feature map is:
g ˜ ( i , j ) = g ( i , j ) + f 0 1 ( i , j ) = i + Δ x i , j j + Δ y i , j
After obtaining the sampling locations, g ˜ ( i , j ) is normalized. Finally, the aligned feature F f l o w 0 1 a l i g n is obtained from F 1 l through the sampling mapping function S ( · ) , thereby achieving reference pixel alignment from I 0 to I 1 . The computation is as follows:
F f l o w 0 1 a l i g n = S ( I 1 , g ˜ ( i , j ) )
where S ( · ) denotes a backward-warping sampling operator that extracts feature values from the source feature map according to flow-guided coordinates using nearest-neighbor interpolation.

4.1.2. Offset Alignment Module

To further align pixels in boundary regions and reduce blurring in inconsistent areas within categories, the OAM achieves fine intra-frame pixel alignment by calculating similarity between pixel neighborhoods in the spatial dimension, as shown in Figure 5a. First, the cosine similarity is calculated between pixel p ( i , j ) at a certain position in F f l o w 0 1 a l i g n and the neighboring pixels within the k × k convolution window. Excluding the central pixel, there are k 2 1 pixels in total. The calculation process is as follows:
S i , j ( p , q ) = c = 1 C F c , i , j a l i g n F c , i + p , j + q a l i g n c = 1 C F c , i , j a l i g n 2 · c = 1 C F c , i + p , j + q a l i g n 2
where S R k 2 1 , channel index is c = 1 , , C , the neighborhood offset ( p , q ) k 1 2 , , k 1 2 and ( p , q ) ( 0 , 0 ) . We set k = 3 , indicating the calculation of similarity between this pixel and the 8 pixels in its neighborhood.
To avoid interference from smoothed pixels in low-frequency regions and ensure that the offset estimation module focuses on local high-frequency motion areas, we generate a high-frequency map from F f l o w 0 1 a l i g n using the high-frequency mapping function F h f ( · ) , which is similar to an occlusion mask:
H = F h f ( F f l o w 0 1 a l i g n )
where S R 1 × H × W . During secondary sampling, the high-frequency map H guides offset sampling toward pixels with high intra-category similarity by focusing on regions of significant change within the image.
Through the convolutional layer, the tensor V controlling the sampling direction of the offset is obtained. The tensor M controlling the sampling size of the offset is derived via the Sigmoid function. Finally, multiplying V and M yields the predicted offset Δ O R 2 × H × W . Subsequently, any pixel point in F f l o w 0 1 a l i g n is sampled at its new coordinate position to obtain the final aligned output feature F i a l i g n R C × H × W . The overall computation process is as follows:
V = Conv 3 × 3 Concat ( F f l o w 0 1 a l i g n , S )
M = Sigmoid Conv 3 × 3 Concat ( F f l o w 0 1 a l i g n , S , H )
Δ O = M · V
F i a l i g n = S ( F f l o w 0 1 a l i g n , Δ O )
The OAM achieves more reliable processing of localized high-frequency motion regions through dynamic high-frequency response and low-frequency suppression strategies. It drives the model to locate rapidly changing areas and samples pixels with high intra-category similarity within the frame to achieve secondary fine alignment.

4.2. Region-Adaptive Interaction Module

The overall framework of the RAIM is illustrated in Figure 5b. It first employs cross-frame interaction perception to adaptively associate different regions within the neighborhood, thereby enhancing feature expression across distinct areas. Subsequently, it incorporates both global attention (G-Att) and local attention (L-Att) in the RSAM to capture global dependencies and local high-frequency motion information, respectively. Finally, based on the characteristics of different regions, global and local features are synergistically fused to achieve spatiotemporal information aggregation. The specific implementation process comprises the following four steps.

4.2.1. Assigning the Attention Head

We reshape F 0 , F 1 R C × H × W along the channel dimension into a sequence F R C × N of length N = H × W as input. The RAIM divides the attention heads into two groups based on different regional characteristics, with the following computation process:
h L A + h G A = h
d L A = h L A · d h e a d , d G A = h G A · d h e a d
where h L A and h G A represent the numbers of local attention heads and global attention heads, respectively. h represents the total number of heads in this layer, setting h L A = h G A = h 2 . d h e a d = C h is the dimension of each attention head, d L A and d G A represent the total dimension of the attention heads in L-Att and G-Att, respectively.

4.2.2. L-Att Branch

L-Att employs non-overlapping local window attention to focus on capturing local high-frequency motion information. Within the L-Att branch, cross-frame adaptive interaction perception is achieved through a window partitioning strategy. Specifically, the window is w R r × r , and the number of windows is N w = H × W r 2 , so the window partitioning formula is as follows:
{ F 0 , g } g = 1 N w = W r ( F 0 ) , { F 1 , g } g = 1 N w = W r ( F 1 )
where F 0 , g , F 1 , g R r 2 × C denotes the token sequence in the g-th window of the two frames. W r ( · ) is the window partitioning function. Subsequently, using F 0 as the Query and the local features of F 1 as the Key and Value, their projection matrices are computed via linear mapping. The formula is as follows:
Q g ( m ) = F 0 , g W Q ( m ) , K g ( m ) = F 1 , g W K ( m ) , V g ( m ) = F 1 , g W V ( m )
where W Q ( m ) , W K ( m ) , W V ( m ) R C × d are the learnable weight matrices of the m-th attention head within the window, where m { 1 , , h H F } , and d = d L A h L A denotes the dimensionality of a single attention head in the L-Att branch. The above process enables adaptive cross-frame interaction perception from I 0 to I 1 at the local region level, allowing I 0 to adaptively associate with local regions of I 1 . Subsequently, attention is computed within each window to achieve inter-frame local high-frequency feature aggregation as follows:
A t t g ( m ) = Softmax Q g ( m ) K g ( m ) d V g ( m )
Finally, the attention weights from all heads within each window are concatenated to compute the multi-head attention for that window. W 0 L A R d L A 2 represents the learnable weight matrix, and the output A t t L A R N × d L A for the L-Att branch is obtained through linear mapping, as shown in the following formula:
A t t g L A = Concat A t t g ( 1 ) , A t t g ( 2 ) , , A t t g ( h L A ) W o L A
A t t L A = Reshape A t t 1 L A , A t t 2 L A , , A t t N w L A

4.2.3. G-Att Branch

L-Att is fundamentally implemented based on window attention, perceiving only local high-frequency regions. To extract global semantic information, the cross-frame adaptive interaction perception method in the G-Att branch is achieved through average pooling. Specifically, F 1 is downsampled to F ˜ 1 R N s × C to reduce the scale and expand the receptive field range, with the downsampling factor s set to 2, yielding N s = N s 2 . To establish cross-frame attention, the Query is derived from F 0 , while the Key and Value are obtained from the downsampled feature F ˜ 1 . All three components are generated through linear projection and are computed as follows:
Q = F 0 W Q G A , K = F ˜ 1 W K G A , V = F ˜ 1 W V G A
where Q R N × h G A × d , K , V R N s × h G A × d . W Q G A R C × d G A is the learnable weight matrix for Q, while W K G A , W V G A R C × d G A are the learnable weight matrices for K and V, respectively. d = d G A h G A is the dimension of a single attention head in the G-Att branch.
The above process achieves cross-frame adaptive interaction perception across the global regions I 0 to I 1 . For I 0 , it has now been adaptively associated with the global context of the lower-scale I 1 . Next, the attention for each attention head m is computed using the following formula:
A t t ( m ) = Softmax Q ( m ) K ( m ) d V ( m )
where Q ( m ) R N × d , K ( m ) , V ( m ) R N s × d , A t t ( m ) R N × d . Based on the above formula, G-Att can compute attention between Q, K, and V at different scales. Since the global region changes slowly overall, this approach not only captures cross-scale dependencies but also comprehensively focuses on global semantic information. Finally, concatenate all attention heads and apply a linear mapping to obtain the G-Att output A t t G A R N × d G A , and W 0 G A R d G A 2 is a learnable weight matrix, as shown in the following formula:
A t t G A = Concat A t t 1 , A t t 2 , , A t t h G A W o G A

4.3. Motion Compensation Module

According to [11], for optical flow computation in UHD video sequences, a coarse flow field can first be estimated on low-resolution images and then upsampled to the original resolution. This approach reduces computational complexity while providing an approximate flow field, with the low-resolution flow serving as a motion prior to supplementing the motion information in high-resolution sequences.
To compensate for potential alignment errors from the DAM, the MCM leverages global correlations to provide additional motion information, as illustrated in Figure 5c. This module explicitly models the inter-frame motion field on low-resolution feature maps by computing the global correlation matrix, providing supplementary motion information for the subsequent predicted frame reconstruction stage and, thereby, mitigating alignment errors that could result from depending entirely on the DAM. Specifically, given two feature tensors obtained from the RAIM, they are firstly reshaped into token sequences X 0 , X 1 R C × N , N = H × W . Then, their inner product is computed to derive the global correlation matrix. Finally, the probability matching distribution is obtained via the Softmax function as follows:
C o r r = 1 C · ( X 0 X 1 )
P = Softmax ( C o r r )
where each element of C o r r R N × N represents the correlation between the i-th position in X 0 and the j-th position in X 1 , P R N × N is the probability distribution of the match. The initial coordinate grid G R N × 2 is embedded into coordinate grid G ˜ R N × D through the linear mapping, and D is the motion embedding dimension. Subsequently, P and the embedded positions of the target image are weighted to obtain the coordinate distribution M for each pixel point match. Finally, the difference between the new position coordinates of each pixel and its initial embedded coordinates is calculated to derive the motion feature F m R 2 × H × W .

4.4. Optical Flow Estimation Network

To enable the DAM to achieve efficient and accurate long-range inter-frame pixel alignment, we design a lightweight CNN-based optical flow estimation subnetwork to provide motion information to the DAM. OFENet adopts a classical stacked convolutional architecture, where layer-by-layer feature extraction and nonlinear transformations map the feature maps F l , 0 f , F l , 1 f of two input frames into a pixel-wise optical flow field f 0 1 , f 1 0 R 2 × H × W . The network consists of five convolutional layers: the first four layers have 64 output channels with a stride of 1 and padding of 1, while the final layer outputs 2 channels. The overall architecture is illustrated in Figure 5d.

5. Experiment

5.1. Datasets and Evaluation Metrics

5.1.1. Datasets

We train our models on the UHD-N-Train and Vimeo90K [31] training sets. UHD-N-Train contains 50,400 triplets, while Vimeo90K [31] consists of 51,312 triplets with a spatial resolution of 448 × 256 . During training, the samples are cropped to 256 × 256 . Both datasets are augmented with random flipping, temporal reversal, and rotation.
To fully evaluate the performance of our method in UHD extreme motion scenes and other motion scenes, we conducted evaluations on our self-built dataset and multiple benchmark datasets, as detailed below:
  • UHD4K120FPS-N. This dataset is a 4K dataset covering extreme motion scenes, containing triplets from 15 different scenes, where I 0 and I 1 are the input frames, and I 0.5 is the ground truth.
  • Vimeo90K [31]. This contains 3782 triplets with a resolution of 448 × 256 .
  • UCF101 [33]. This contains 379 video sequences with a resolution of 256 × 256.
  • Middlebury [34]. We evaluate on the Middlebury OTHER, which contains 12 sequences from different scenes with a resolution of approximately 640 × 480 .
  • X4K1000FPS [6]. This is commonly used to evaluate 4K video frame interpolation tasks. XTest contains 15 consecutive 33-frame 4K video sequences. Following the settings in [10], we select the 0th and 32nd frames from each sequence as input and evaluate the quality of the generated 16th frame. This new test set is denoted as XTest-L.
  • Xiph [35]. This dataset contains 8 4K scene sequences, each with 100 consecutive frames. Following the same setup as XTest-L in [6], 192 test instances are obtained and denoted as Xiph-L.

5.1.2. Evaluation Metrics

This paper employs three evaluation metrics: the peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [36], and average interpolation error (IE). The PSNR and SSIM are used for evaluation on Vimeo90K [31], UCF101 [33], X4K1000FPS [6], and Xiph [35]. To ensure fair comparison and reproducibility, the PSNR and SSIM are first computed for each sequence and then averaged across all sequences. The IE is used for evaluation on Middlebury [34]. Additionally, we compare the model parameters and runtime, where runtime is measured as the average time over 100 iterations for all models on the same hardware, NVIDIA GeForce RTX 2080 Ti GPU (manufactured by NVIDIA Corporation, Santa Clara, CA, USA).The test samples are random data with a resolution of 640 × 480 .

5.2. Implementation Details

5.2.1. Training Details

The batch size is set to 32, and training is conducted for 300 epochs. The optimizer is AdamW [37] with β 1 = 0.9 , β 2 = 0.999 , and a weight decay of 1 × 10 4 . Before training, the learning rate is warmed up for 2000 steps to 2 × 10 4 and then gradually decayed to 2 × 10 5 using cosine annealing. Following [9], the loss function consists of warp loss and Laplacian loss. Except for the runtime evaluation, all experiments are performed on an NVIDIA A100 Tensor Core GPU.

5.2.2. Network Architecture

Our model is based on a five-layer architecture with an initial channel width of 32 and a downsampling rate of 2. The first three layers are convolutional layers with a depth of 2, while the last two layers are Transformer layers with a depth of 4. The channels are 32, 64, 128, 256, and 512, respectively. The RAIM has a total of eight attention heads, and the local branch window size for RSAM is set to 2.

5.3. Quantitative Comparison with Previous Methods

5.3.1. Comparison in UHD Extreme Motion Scenes

Table 2 reports the quantitative evaluation results of our method against 19 SOTA methods on the UHD-N-Test dataset, and all methods are compared for the single-frame (2×) interpolation task. These methods include ToFlow [31], SepConv [12], AdaCoF [13], CDFI [27], BMBC [38], XVFIv [6], CAIN [14], DAIN [15], UPR-base [20], UPR-large [20], EMA-small [9], RIFEm [11], EBME-H [19], EBME [19], FILM [39], M2M-PWC [40], FLAVR [41], SGM-small-1/2-points [10], and SGM-local-branch [10]. The PSNR and SSIM are better when higher. The PSNR measures the pixel-wise consistency between the generated and ground-truth frames, and a high PSNR indicates that the model can effectively learn motion information and reconstruct high-quality frames in extreme motion scenes. The results show that our method achieves the best performance in terms of both the PSNR and SSIM, surpassing the second-best result by 0.05 dB in PSNR and 0.0006 in SSIM. Compared with other mainstream methods designed for handling large motion, our method also demonstrates significant advantages, outperforming FILM [39] by 0.57 dB and SGM-local-branch [10] by 1.78 dB in PSNR. These results clearly validate the effectiveness of the proposed method in UHD extreme motion scenes.
Under the same hardware conditions, our method has a runtime of 42 ms. Compared with other 4K video frame interpolation methods, it is faster than XVFIv [6] and FILM [39], but it is still slower than the real-time interpolation method RIFEm [11]. This indicates that the applicability of our method in practical real-time scenes is limited. To further reduce runtime, the most direct approach is to use higher-performance hardware. Additionally, from a model optimization perspective, the RSAM can be further optimized to reduce its computational complexity, which would help lower the overall runtime and improve real-time performance (see Section 5.5.4 for discussion).
In addition, we perform quantitative comparisons with four mainstream methods on the public UHD datasets XTest-L [6] and Xiph-L [35] to further verify the generalization capability of our method, as shown in Table 3. The results show that the proposed method exhibits excellent overall performance in handling extreme motion, and it is not inferior to advanced methods specifically designed for UHD scenarios. It achieved the best results on the XTest-L dataset and demonstrated superior overall performance on the Xiph-L dataset. Compared with EMA-VFI-small [9] and SGM-VFI-local-branch [10], our method achieved improvements of 1.1 dB and 0.45 dB, respectively, on the XTest-L-4K dataset, further validating its adaptability and generalization capability in extreme motion scenes.
We also conduct multi-frame (8×) interpolation experiments on XTest [6] datasets of different resolutions, following the experimental settings of M2M [40] and EMA-VFI [9], as shown in Table 4. Our method achieves the best performance at both resolutions, indicating smaller pixel-level reconstruction errors and interpolation results closer to the ground truth. Combined with inference time analysis, this demonstrates that our method provides efficient and stable multi-frame interpolation in practical applications.

5.3.2. Comparison on Low-Resolution Benchmarks

Table 5 presents the quantitative comparison of our method with 20 SOTA methods on low-resolution datasets, including Vimeo90K [31], UCF101 [33], and Middlebury [34], to evaluate its performance in easy motion scenes. The results show that on the Vimeo90K dataset, our method achieves the best performance in both evaluation metrics, surpassing the second-best method FLAVR [41] by 0.09 dB in PSNR and 0.0004 in SSIM. On the UCF101 dataset, our method also achieves a strong second-best PSNR.
Similarly to our approach, FGDCN-S [43] and VFIFT-Conv [44] also introduce optical flow, but they rely on only a single type of motion information. On the Vimeo90K dataset, our method significantly outperforms VFIFT-Conv [44] and FGDCN-S [43], with PSNR improvements of 0.37 dB and 0.15 dB, respectively. This clearly demonstrates the effectiveness of our approach in leveraging two types of motion information. In summary, our method also shows outstanding performance in low-resolution scenarios with easy motion.
Table 5. Quantitative comparison with 20 SOTA methods on the Vimeo90K [31], UCF101 [33], and Middlebury [34] datasets. The best and second-best results are indicated with bold and underlined text, respectively. “↑” indicate higher is better, "↓" indicate lower is better.
Table 5. Quantitative comparison with 20 SOTA methods on the Vimeo90K [31], UCF101 [33], and Middlebury [34] datasets. The best and second-best results are indicated with bold and underlined text, respectively. “↑” indicate higher is better, "↓" indicate lower is better.
MethodVimeo90KUCF101Middlebury
PSNR ↑SSIM ↑PSNR ↑SSIM ↑IE ↓
Other
CAIN [14]34.650.973034.910.96902.28
FLAVR [41]36.300.975033.330.9710-
Kernel-base
SepConv [12]33.790.970234.780.96692.27
AdaCoF [13]34.470.973034.900.96802.24
CDFI [27]35.170.977035.210.96901.98
EDSC- L C  [25]34.840.975035.130.96802.02
Flow-base
SoftSplat [45]36.100.970035.390.95201.81
ToFlow [31]33.730.968234.580.96672.15
BMBC [38]35.010.976435.150.96892.04
XVFIv [6]35.070.968135.180.9519-
DAIN [15]34.710.975634.990.96832.04
ABME [46]36.180.980535.380.96982.01
IFRNet [42]35.800.979435.290.96931.95
UPR-base [20]36.030.980135.410.9698-
EMA-small [9]36.070.979735.340.96961.94
RIFE [11]35.610.977935.280.96901.96
EBME-H* [19]36.190.980735.410.9697-
FGDCN-S [43]36.240.980635.420.96981.94
VFIFT-Conv [44]36.020.979835.650.9793-
M2M-PWC [40]35.470.977835.280.96942.09
Ours36.390.981135.460.97002.01

5.4. Qualitative Comparison with Previous Methods

5.4.1. Visual Comparison in UHD Extreme Motion Scenes

On the UHD-N-Test dataset, we visually compare the interpolation performance of our method with eight SOTA approaches, all trained under the same settings on UHD-N-Train. As shown in Figure 6, our method produces more reasonable results in extreme motion scenes. In particular, in Example 5, our method reconstructs the volleyball contours clearly compared with other methods. Example 4 shows that, except for our method, EBME [19], and SGM-local-branch [10], the interpolation results of other methods exhibit blur or artifacts, indicating their inability to handle local texture regions effectively. In Example 3 (occluded regions), all methods exhibit some degree of blur along occlusion boundaries, caused by the lack of reference information, which makes accurate pixel displacement estimation difficult. For our method, although the boundaries remain slightly blurred, the RAIM can leverage information from neighboring visible pixels to infer structure and texture more reasonably. This slight blurring is not a flaw of the method but an unavoidable challenge in extreme motion scenes, and our design has substantially mitigated this issue. Moreover, for nonlinear extreme motion (rotational motion in Example 1), all methods fail to generate clear water flow, whereas our interpolation results appear visually more reasonable. In summary, the visualization results of UHD-N-Test fully demonstrate the effectiveness of our method in handling extreme motion scenes while also exhibiting excellent interpolation performance in other complex scenes.
To further validate the performance of the proposed method on UHD datasets, we conduct qualitative comparisons on the XTest-L-4K [6] dataset. We focus on evaluating extreme motion scenes and do not adopt the experimental setup of XVFI [6] on the XTest [6] dataset; instead, we select inputs with the largest temporal intervals for evaluation. As shown in Figure 7, our method achieves interpolation results in subjective visual quality that are comparable to existing SOTA methods, effectively confirming its generalization ability in UHD video frame interpolation tasks.

5.4.2. Visual Comparison on Low-Resolution Benchmarks

Figure 8 presents a visual comparison of our method against six mainstream approaches on the Vimeo90K [31] dataset. This dataset mainly contains low-resolution scenes with easy motion, so we focus on the consistency between the interpolated results and the ground truth (GT). For example, in Example 2, the GT shows the overall contour of the athlete’s legs, but the details are blurry, and there are significant occluded regions. Both our method and EMA-VFI [9] are able to recover these severely occluded areas, producing results with more reasonable overall structures. Across all examples, other mainstream methods perform less favorably. For instance, EBME [19] generates results with noticeable noise, while M2M [40] produces significant artifacts. In contrast, our method demonstrates clear advantages in interpolation quality.

5.5. Ablation Study

5.5.1. Ablation Study of Dual Alignment Module

To validate the effectiveness of the DAM, we conduct ablation experiments on the optical flow alignment module (OFAM) and the offset alignment module (OAM) separately on the UHD-N-Test and Vimeo90K [31] datasets, as shown in Table 6. First, we replace OFENet with several mainstream pre-trained optical flow estimation methods, including SPyNet [47], RAFT [32], and GMFlow [48]. The results show that OFENet achieves the best performance, demonstrating its adaptability and effectiveness for video frame interpolation. Additionally, to analyze the impact of the OAM on overall performance, we remove it and perform alignment using only the OFAM. The results indicate that removing the OAM decreases performance by 0.29 dB, but it still outperforms some pre-trained optical flow methods. These findings confirm the strong alignment capability of the DAM in extreme motion scenes.

5.5.2. Ablation Study of Region-Adaptive Interaction Module

To accelerate the research process, all methods are quantitatively evaluated after 100 training epochs. To validate the effectiveness of the proposed cross-frame adaptive interaction strategy, we replace the RAIM with the classical cross-attention mechanism. To validate the effectiveness of the proposed cross-frame adaptive interaction perception strategy, the RAIM was replaced with a classical cross-attention mechanism. Unlike the RAIM, this attention mechanism does not associate regions at different scales through cross-frame adaptive interaction perception nor aggregate global and local features separately, instead calculating attention only between neighboring frames. As shown in Table 7, on the UHD-N-Test dataset, the model with the RAIM achieves a PSNR 0.29 dB higher than that with cross-attention. This demonstrates the RAIM’s superior performance across diverse motion scenes and the effectiveness of the cross-frame adaptive interaction strategy.
Furthermore, to further analyze the performance of the RAIM and the effectiveness of the RSAM, we compare different window sizes in the cross-frame adaptive interaction, as shown in Table 7. Specifically, the window sizes are set to 3, 2, and 1. The results indicate that a window size of 2 (our model) achieves the best performance. When the window size is 3, since the feature maps in the last two layers of the model ( 32 × 32 and 16 × 16 ) are already small, a larger window increases the computational complexity of attention (as shown in Table 8). When the window size is 1, the model performance drops significantly. In this case, the RSAM degenerates to using only global attention (G-Att), which is equivalent to not introducing the cross-frame adaptive interaction strategy in the RAIM, preventing the differentiation of distinct motion regions and, thereby, limiting the modeling and understanding of extreme motion.

5.5.3. Ablation Study of Motion Compensation Module

We validate the effectiveness of the MCM by removing it from the model. As shown in Table 7, compared with the complete model, the model without the MCM exhibits a performance drop of 0.11 dB. This result indicates that the motion information provided by the MCM enhances model performance, indirectly verifying its ability to effectively suppress alignment errors in the DAM.

5.5.4. Computational Complexity Analysis of the Region-Separated Attention Mechanism

Table 8 presents the performance analysis of the RSAM, G-Att, and L-Att. Specifically, in the last two layers of the model, the feature map sizes are 32 × 32 and 16 × 16, with a depth of 4 per layer. For each layer, we report the total number of parameters and computational complexity, along with the percentage of computation contributed by G-Att and L-Att. The results in the table reveal that, theoretically, the computational complexity of L-Att scales quadratically with the window size, whereas the complexity of G-Att is inversely proportional to the square of the downsampling rate. In practical applications, L-Att accounts for a larger proportion of computation in the last layer, while G-Att dominates in the penultimate layer.

5.5.5. Visual Analysis of the Region-Adaptive Interaction Module

To fully validate the cross-frame adaptive interaction perception capability of the RAIM and its ability to model both global context and local high-frequency motion information, Figure 9 presents the attention weight visualization heatmap results for the final layer of the RAIM within the Transformer. Red and blue regions indicate areas with higher and lower attention weights, respectively. The results show that in the RAIM heatmap, the attention mechanisms focus on areas that highly align with the locally high-frequency motion regions in the original image. For example, the basketball movement area in the first row and the rapid arm movement area in the second row both exhibit high attention weights. In the L-Att heatmap, red indicates high-frequency feature regions, with results closely matching the high-weight areas in the RAIM heatmap, demonstrating that L-Att focuses on learning local high-frequency motion information. In contrast, in the G-Att heatmap, blue indicates low-frequency feature regions, revealing that G-Att prioritizes learning global contextual information. Overall, these visualizations intuitively demonstrate the effectiveness of the proposed RAIM and its superior representation capabilities for distinct spatial features.

6. Conclusions

This paper addresses the challenge of accurately fitting extreme motion in UHD scenes by proposing a high-performance video interpolation method based on dual alignment and region-adaptive interaction. First, we design the dual alignment module, which achieves long-range pixel alignment in stages using both optical flow and offset motion information, effectively reducing the accumulation of motion fitting errors and alleviating spatiotemporal asymmetry between frames. Then, we introduce the region-adaptive interaction module, which employs a cross-frame adaptive interaction strategy to dynamically establish correlations and fuse information across different motion regions in neighboring frames, effectively mitigating blurring at occlusion boundaries. Furthermore, to compensate for alignment errors, we introduce the motion compensation module, which provides additional motion information through explicit motion estimation. Extensive quantitative and qualitative experiments demonstrate that our method exhibits outstanding interpolation performance, strong robustness, and effective restoration of spatiotemporal symmetry in extreme motion scenes on both our self-built UHD dataset and multiple benchmark datasets, validating its applicability and reliability.

Author Contributions

Conceptualization, X.N.; methodology, X.N. and Y.D.; software, X.N.; validation, X.N., J.Q., J.D. and K.Y.; data curation, X.N., J.Q., J.D. and K.Y.; investigation, X.N., J.Q., J.D. and K.Y.; resources, Y.D.; writing—original draft preparation, X.N.; writing—review and editing, J.Q., J.D., K.Y. and Y.D.; visualization, X.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Nos. 61303093 and 61402278) and the Shanghai Natural Science Foundation (No. 19ZR1419100).

Data Availability Statement

The original data presented in the study are openly available in a GitHub repository at https://github.com/nxkobe/ExtremeMotion-VFI, accessed on 28 November 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jiang, W.; Li, J.; Zhang, K.; Zhang, L. Ecvc: Exploiting non-local correlations in multiple frames for contextual video compression. In Proceedings of the Computer Vision and Pattern Recognition Conference, Jammu, India, 16–18 July 2025; pp. 7331–7341. [Google Scholar]
  2. Liao, G.; Li, Q.; Bao, Z.; Qiu, G.; Liu, K. Spc-gs: Gaussian splatting with semantic-prompt consistency for indoor open-world free-view synthesis from sparse inputs. In Proceedings of the Computer Vision and Pattern Recognition Conference, Jammu, India, 16–18 July 2025; pp. 11264–11274. [Google Scholar]
  3. Kim, E.; Kim, H.; Jin, K.H.; Yoo, J. BF-STVSR: B-Splines and Fourier—Best Friends for High Fidelity Spatial-Temporal Video Super-Resolution. In Proceedings of the Computer Vision and Pattern Recognition Conference, Jammu, India, 16–18 July 2025; pp. 28009–28018. [Google Scholar]
  4. Shen, W.; Bao, W.; Zhai, G.; Chen, L.; Min, X.; Gao, Z. Blurry video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 5114–5123. [Google Scholar]
  5. Yan, Z.; Lei, P.; Wang, T.; Fang, F.; Zhang, J.; Huang, Y.; Song, H. Explicit Depth-Aware Blurry Video Frame Interpolation Guided by Differential Curves. In Proceedings of the Computer Vision and Pattern Recognition Conference, Jammu, India, 16–18 July 2025; pp. 1994–2004. [Google Scholar]
  6. Sim, H.; Oh, J.; Kim, M. Xvfi: Extreme video frame interpolation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 14489–14498. [Google Scholar]
  7. Ahn, H.E.; Jeong, J.; Kim, J.W. A fast 4k video frame interpolation using a hybrid task-based convolutional neural network. Symmetry 2019, 11, 619. [Google Scholar] [CrossRef]
  8. Park, J.; Kim, J.; Kim, C.S. Biformer: Learning bilateral motion estimation via bilateral transformer for 4k video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1568–1577. [Google Scholar]
  9. Zhang, G.; Zhu, Y.; Wang, H.; Chen, Y.; Wu, G.; Wang, L. Extracting motion and appearance via inter-frame attention for efficient video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5682–5692. [Google Scholar]
  10. Liu, C.; Zhang, G.; Zhao, R.; Wang, L. Sparse global matching for video frame interpolation with large motion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 19125–19134. [Google Scholar]
  11. Huang, Z.; Zhang, T.; Heng, W.; Shi, B.; Zhou, S. Real-time intermediate flow estimation for video frame interpolation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 624–642. [Google Scholar]
  12. Niklaus, S.; Mai, L.; Liu, F. Video frame interpolation via adaptive separable convolution. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 261–270. [Google Scholar]
  13. Lee, H.; Kim, T.; Chung, T.y.; Pak, D.; Ban, Y.; Lee, S. Adacof: Adaptive collaboration of flows for video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 5316–5325. [Google Scholar]
  14. Choi, M.; Kim, H.; Han, B.; Xu, N.; Lee, K.M. Channel attention is all you need for video frame interpolation. Proc. AAAI Conf. Artif. Intell. 2020, 34, 10663–10671. [Google Scholar] [CrossRef]
  15. Bao, W.; Lai, W.S.; Ma, C.; Zhang, X.; Gao, Z.; Yang, M.H. Depth-aware video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3703–3712. [Google Scholar]
  16. Jiang, H.; Sun, D.; Jampani, V.; Yang, M.H.; Learned-Miller, E.; Kautz, J. Super slomo: High quality estimation of multiple intermediate frames for video interpolation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9000–9008. [Google Scholar]
  17. Xu, X.; Siyao, L.; Sun, W.; Yin, Q.; Yang, M.H. Quadratic video interpolation. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32. [Google Scholar]
  18. Xiao, J.; Xu, K.; Hu, M.; Liao, L.; Wang, Z.; Lin, C.W.; Wang, M.; Satoh, S. Progressive motion boosting for video frame interpolation. IEEE Trans. Multimed. 2022, 25, 8076–8090. [Google Scholar] [CrossRef]
  19. Jin, X.; Wu, L.; Shen, G.; Chen, Y.; Chen, J.; Koo, J.; Hahm, C.h. Enhanced bi-directional motion estimation for video frame interpolation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 5049–5057. [Google Scholar]
  20. Jin, X.; Wu, L.; Chen, J.; Chen, Y.; Koo, J.; Hahm, C.h. A unified pyramid recurrent network for video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1578–1587. [Google Scholar]
  21. Shen, T.; Li, D.; Gao, Z.; Tian, L.; Barsoum, E. Ladder: An efficient framework for video frame interpolation. arXiv 2024, arXiv:2404.11108. [Google Scholar] [CrossRef]
  22. Niklaus, S.; Mai, L.; Liu, F. Video frame interpolation via adaptive convolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 670–679. [Google Scholar]
  23. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  24. Cheng, X.; Chen, Z. Video frame interpolation via deformable separable convolution. Proc. AAAI Conf. Artif. Intell. 2020, 34, 10607–10614. [Google Scholar] [CrossRef]
  25. Cheng, X.; Chen, Z. Multiple video frame interpolation via enhanced deformable separable convolution. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7029–7045. [Google Scholar] [CrossRef] [PubMed]
  26. Shi, Z.; Liu, X.; Shi, K.; Dai, L.; Chen, J. Video frame interpolation via generalized deformable convolution. IEEE Trans. Multimed. 2021, 24, 426–439. [Google Scholar] [CrossRef]
  27. Ding, T.; Liang, L.; Zhu, Z.; Zharkov, I. Cdfi: Compression-driven network design for frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 8001–8011. [Google Scholar]
  28. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef] [PubMed]
  29. Huang, F.; Zhang, X.; Xu, Y.; Wang, X.; Wu, X. Video Frame Interpolation for Polarization via Swin-Transformer. arXiv 2024, arXiv:2406.11371. [Google Scholar] [CrossRef]
  30. Shi, Z.; Xu, X.; Liu, X.; Chen, J.; Yang, M.H. Video frame interpolation transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 17482–17491. [Google Scholar]
  31. Xue, T.; Chen, B.; Wu, J.; Wei, D.; Freeman, W.T. Video enhancement with task-oriented flow. Int. J. Comput. Vis. 2019, 127, 1106–1125. [Google Scholar] [CrossRef]
  32. Teed, Z.; Deng, J. Raft: Recurrent all-pairs field transforms for optical flow. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 402–419. [Google Scholar]
  33. Soomro, K.; Zamir, A.R.; Shah, M. Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv 2012, arXiv:1212.0402. [Google Scholar] [CrossRef]
  34. Baker, S.; Scharstein, D.; Lewis, J.P.; Roth, S.; Black, M.J.; Szeliski, R. A database and evaluation methodology for optical flow. Int. J. Comput. Vis. 2011, 92, 1–31. [Google Scholar] [CrossRef]
  35. Montgomery, C.; Lars, H. Xiph. org video test media (derf’s collection). Online 1994, 6, 1. Available online: https://media.xiph.org/video/derf (accessed on 1 April 2025).
  36. Feijoo, D.; Benito, J.C.; Garcia, A.; Conde, M.V. Darkir: Robust low-light image restoration. In Proceedings of the Computer Vision and Pattern Recognition Conference, Jammu, India, 16–18 July 2025; pp. 10879–10889. [Google Scholar]
  37. Loshchilov, I.; Hutter, F. Fixing weight decay regularization in adam. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  38. Park, J.; Ko, K.; Lee, C.; Kim, C.S. Bmbc: Bilateral motion estimation with bilateral cost volume for video interpolation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 109–125. [Google Scholar]
  39. Reda, F.; Kontkanen, J.; Tabellion, E.; Sun, D.; Pantofaru, C.; Curless, B. Film: Frame interpolation for large motion. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 250–266. [Google Scholar]
  40. Hu, P.; Niklaus, S.; Sclaroff, S.; Saenko, K. Many-to-many splatting for efficient video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 3553–3562. [Google Scholar]
  41. Kalluri, T.; Pathak, D.; Chandraker, M.; Tran, D. Flavr: Flow-agnostic video representations for fast frame interpolation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 2071–2082. [Google Scholar]
  42. Kong, L.; Jiang, B.; Luo, D.; Chu, W.; Huang, X.; Tai, Y.; Wang, C.; Yang, J. Ifrnet: Intermediate feature refine network for efficient frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 1969–1978. [Google Scholar]
  43. Lei, P.; Fang, F.; Zeng, T.; Zhang, G. Flow guidance deformable compensation network for video frame interpolation. IEEE Trans. Multimed. 2023, 26, 1801–1812. [Google Scholar] [CrossRef]
  44. Gao, P.; Tian, H.; Qin, J. Video frame interpolation with flow transformer. arXiv 2023, arXiv:2307.16144. [Google Scholar] [CrossRef]
  45. Niklaus, S.; Liu, F. Softmax splatting for video frame interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 5437–5446. [Google Scholar]
  46. Park, J.; Lee, C.; Kim, C.S. Asymmetric bilateral motion estimation for video frame interpolation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 14539–14548. [Google Scholar]
  47. Ranjan, A.; Black, M.J. Optical flow estimation using a spatial pyramid network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4161–4170. [Google Scholar]
  48. Xu, H.; Zhang, J.; Cai, J.; Rezatofighi, H.; Tao, D. Gmflow: Learning optical flow via global matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 8121–8130. [Google Scholar]
Figure 1. Visual comparison with state-of-the-art (SOTA) methods in extreme motion scenes. GT represents the ground truth.
Figure 1. Visual comparison with state-of-the-art (SOTA) methods in extreme motion scenes. GT represents the ground truth.
Symmetry 17 02097 g001
Figure 2. Visualization results of high-frequency and low-frequency component differences.
Figure 2. Visualization results of high-frequency and low-frequency component differences.
Symmetry 17 02097 g002
Figure 3. Example samples from the UHD4K120FPS-N dataset and their corresponding inter-frame optical flow magnitudes.
Figure 3. Example samples from the UHD4K120FPS-N dataset and their corresponding inter-frame optical flow magnitudes.
Symmetry 17 02097 g003
Figure 4. The overall framework of the model proposed in this paper. The model primarily consists of three components: low-level feature extraction (LFE), inter-frame feature interaction perception (IFIP), and predicted frame reconstruction (PFR). Among these, our main contribution lies in IFIP.
Figure 4. The overall framework of the model proposed in this paper. The model primarily consists of three components: low-level feature extraction (LFE), inter-frame feature interaction perception (IFIP), and predicted frame reconstruction (PFR). Among these, our main contribution lies in IFIP.
Symmetry 17 02097 g004
Figure 5. The structure of each module proposed in this paper, including the dual alignment module, region-adaptive interaction module, motion compensation module, and optical flow estimation network.
Figure 5. The structure of each module proposed in this paper, including the dual alignment module, region-adaptive interaction module, motion compensation module, and optical flow estimation network.
Symmetry 17 02097 g005
Figure 6. Qualitative comparison with SOTA methods on the UHD-N-Test dataset.
Figure 6. Qualitative comparison with SOTA methods on the UHD-N-Test dataset.
Symmetry 17 02097 g006
Figure 7. Qualitative comparison with SOTA methods on the XTest-L-4K [6] dataset.
Figure 7. Qualitative comparison with SOTA methods on the XTest-L-4K [6] dataset.
Symmetry 17 02097 g007
Figure 8. Qualitative comparison with SOTA methods on the Vimeo90K [31] dataset.
Figure 8. Qualitative comparison with SOTA methods on the Vimeo90K [31] dataset.
Symmetry 17 02097 g008
Figure 9. Visualization of heat maps for the RAIM.
Figure 9. Visualization of heat maps for the RAIM.
Symmetry 17 02097 g009
Table 1. Percentile distributions of the optical flow magnitude and high-frequency feature ratio for the UHD4K120FPS-N and benchmark datasets. Higher values indicate larger motion magnitudes and a higher high-frequency feature ratio.
Table 1. Percentile distributions of the optical flow magnitude and high-frequency feature ratio for the UHD4K120FPS-N and benchmark datasets. Higher values indicate larger motion magnitudes and a higher high-frequency feature ratio.
DatasetOptical Flow MagnitudesHigh-Frequency Feature Ratios
25th50th75th25th50th75th
Vimeo90K-Test [31]3.05.07.3---
Vimeo90K-Train [31]3.45.68.0---
SNU-FILM-Hard [14]2.75.617.1---
SNU-FILM-Extreme [14]5.311.934.8---
X-Test [6]19.275.6141.00.00020.00020.0004
X-Train [6]6.719.665.00.00010.00020.0003
UHD-N-Test70.589.7216.60.00010.00020.0004
UHD-N-Train23.730.844.10.00020.00040.0008
Table 2. Quantitative comparison with 19 SOTA methods on the UHD-N-Test dataset. The best and second-best results are indicated with bold and underlined text, respectively. “↑” indicate higher is better.
Table 2. Quantitative comparison with 19 SOTA methods on the UHD-N-Test dataset. The best and second-best results are indicated with bold and underlined text, respectively. “↑” indicate higher is better.
MethodPSNR ↑SSIM ↑Parameters
(Million)
Runtime
(ms)
Other
CAIN [14]26.300.888042.837
FLAVR [41]27.520.890142.437
Kernel-base
SepConv [12]24.970.809721.6200
AdaCoF [13]25.890.887422.930
CDFI [27]26.040.84545.0172
Flow-base
ToFlow [31]24.560.80421.184
BMBC [38]26.080.843211.0822
XVFIv [6]26.210.85595.598
DAIN [15]25.450.844824.0151
UPR-base [20]26.940.89721.742
UPR-large [20]23.600.83313.762
RIFEm [11]28.410.91049.812
EBME-H [19]23.460.78293.940
EBME [19]26.760.89103.920
M2M-PWC [40]23.620.83507.632
FILM- L S  [39]27.890.893734.4101
EMA-small [9]26.880.895114.530
SGM-local-branch [10]26.680.905715.457
SGM-small-1/2-points [10]26.860.890120.856
Ours28.460.911029.842
Table 3. Quantitative comparison of 2× interpolation on the XTest-L [6] and Xiph-L [35] datasets with the SOTA methods. The best and second-best results are indicated with bold and underlined text, respectively. “↑” indicate higher is better.
Table 3. Quantitative comparison of 2× interpolation on the XTest-L [6] and Xiph-L [35] datasets with the SOTA methods. The best and second-best results are indicated with bold and underlined text, respectively. “↑” indicate higher is better.
MethodXTest-L-2KXTest-L-4KXiph-L-2KXiph-L-4K
PSNR ↑SSIM ↑PSNR ↑SSIM ↑PSNR ↑SSIM ↑PSNR ↑SSIM ↑
XVFI [6]29.820.895129.020.886629.170.844928.090.7889
RIFE [11]29.870.880528.980.875630.180.863328.070.7982
EMA-small [9]29.510.877528.600.873330.540.871828.400.8109
SGM-local-branch [10]30.390.894629.250.886130.890.874528.590.8115
Ours30.450.896129.700.890830.860.874628.870.8134
Table 4. Quantitative comparison of 8× interpolation on the XTest [6] dataset with the SOTA methods. “†” denotes training on the XTrain [6] dataset. The best and second-best results are indicated with bold and underlined text, respectively.
Table 4. Quantitative comparison of 8× interpolation on the XTest [6] dataset with the SOTA methods. “†” denotes training on the XTrain [6] dataset. The best and second-best results are indicated with bold and underlined text, respectively.
DAIN [15]IFRNet [42]XVFI † [6]EMA-Small [9]M2M [40]Ours
2K29.3331.5330.8531.8932.1332.41
4K26.7830.4630.1230.8930.8831.23
Table 6. Ablation experiment on the dual alignment module. “w/o” denotes without. The best results are indicated with bold text.
Table 6. Ablation experiment on the dual alignment module. “w/o” denotes without. The best results are indicated with bold text.
MethodUHD-N-TestVimeo90K
PSNR ↑SSIM ↑PSNR ↑SSIM ↑
SPyNet [47]27.940.909036.150.9801
RAFT [32]28.050.909836.180.9806
GMFlow [48]28.210.910436.210.9805
w/o OAM28.170.910736.260.9805
Ours28.460.911036.390.9811
Table 7. Ablation experiment on the RAIM and MCM. The best results are indicated with bold text.
Table 7. Ablation experiment on the RAIM and MCM. The best results are indicated with bold text.
MethodUHD-N-TestVimeo90K
PSNR ↑SSIM ↑PSNR ↑SSIM ↑
3 × 3 Window26.420.853535.700.9780
1 × 1 Window (Global)26.310.860035.660.9780
w/o MCM27.200.862735.750.9788
Cross-Attention27.020.864335.680.9779
Ours27.310.865435.840.9795
Table 8. Complexity and parameter analysis of the RSAM.
Table 8. Complexity and parameter analysis of the RSAM.
MethodTheoreticalActual
FLOPsParameters (M)FLOPs (G)
16 × 1632 × 3216 × 1632 × 32
L-Att O ( H W ( 3 C d L A + 4 h L A r 2 d h e a d + d L A 2 ) ) 1.840.460.47 (61%)0.47 (47%)
G-Att O ( H W ( C d G A ( 1 + 2 s 2 ) + 4 h G A s 2 d h e a d + d G A 2 ) ) 1.840.460.30 (39%)0.54 (53%)
RSAM F L O P L A + F L O P G A 3.680.920.771.01
Total-29.87182
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ning, X.; Qu, J.; Duan, J.; Yang, K.; Ding, Y. Video Frame Interpolation for Extreme Motion Scenes Based on Dual Alignment and Region-Adaptive Interaction. Symmetry 2025, 17, 2097. https://doi.org/10.3390/sym17122097

AMA Style

Ning X, Qu J, Duan J, Yang K, Ding Y. Video Frame Interpolation for Extreme Motion Scenes Based on Dual Alignment and Region-Adaptive Interaction. Symmetry. 2025; 17(12):2097. https://doi.org/10.3390/sym17122097

Chicago/Turabian Style

Ning, Xin, Jiantao Qu, Junyi Duan, Kun Yang, and Youdong Ding. 2025. "Video Frame Interpolation for Extreme Motion Scenes Based on Dual Alignment and Region-Adaptive Interaction" Symmetry 17, no. 12: 2097. https://doi.org/10.3390/sym17122097

APA Style

Ning, X., Qu, J., Duan, J., Yang, K., & Ding, Y. (2025). Video Frame Interpolation for Extreme Motion Scenes Based on Dual Alignment and Region-Adaptive Interaction. Symmetry, 17(12), 2097. https://doi.org/10.3390/sym17122097

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop