Next Article in Journal
GMDNet: Grouped Encoder-Mixer-Decoder Architecture Based on the Role of Modalities for Brain Tumor MRI Image Segmentation
Previous Article in Journal
Enhanced Power Sharing Control of an Islanded DC Microgrid with Unmatched Line Impedances
Previous Article in Special Issue
Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Temporal Map-Based Boundary Refinement Network for Video Moment Localization

1
College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
2
School of Artificial Intelligence, Anhui University, Hefei 230039, China
3
College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China
*
Authors to whom correspondence should be addressed.
Electronics 2025, 14(8), 1657; https://doi.org/10.3390/electronics14081657
Submission received: 29 March 2025 / Revised: 16 April 2025 / Accepted: 17 April 2025 / Published: 19 April 2025
(This article belongs to the Special Issue Big Model Techniques for Image Processing)

Abstract

:
Video moment localization has gradually become a hot research problem in video understanding. Despite much remarkable progress, there are still two limitations in the following aspects. Firstly, previous methods usually directly regarded the moment candidate with the highest confidence score as the final localization result; they overlooked the inevitable deviations between moment candidates and the ground truth. Secondly, past models have not considered the problem that moment candidates with various qualities have different impacts on model training. Therefore, this paper proposes a novel Temporal Map-based Boundary Refinement Network to solve the above problems. Specifically, besides the conventional confidence scores’ prediction network, we introduce a boundary refinement network based on the 2D temporal map, which can fine-tune the temporal boundaries of generated moment candidates to obtain more precise results. Additionally, to discriminately treat diverse moment candidates with different qualities and further boost the localization performance, we devise an innovative weighted refinement loss to guide the model to focus more on refining processes of those moment candidates closer to the ground truth. Finally, we evaluate our model on two publicly available datasets, and the results of extensive experiments show that our technique outperforms the state-of-the-art methods (e.g., for the ActivityNet Captions dataset, a relative improvement of 4.63% on R1@0.7).

1. Introduction

With the rapid development of the Internet, we have entered the era of the Internet of Everything. Recently, massive amounts of multimedia data have poured into people’s daily lives, coming from speech, texts, images, videos, etc. [1,2,3]. Although they bring unprecedented convenience to humans, we face many formidable challenges. How to quickly and accurately retrieve valuable information from plenty of chaotic data, especially video data, has become the mainstream research direction of information retrieval. Video moment localization, one of the most typical topics in multimedia retrieval and the computer vision community [4,5,6], is a critical and challenging problem.
Video moment localization (VML for short) aims to precisely locate the corresponding segment in a long untrimmed video with the given natural language query. For instance, as query 2 shows in Figure 1, “she washes her face and wipes it down with a rag still speaking to the camera” includes “wash face”, “wipe face with a rag”, and “speaking to camera”, three actions. Therefore, we need to obtain the segment that contains at least these three consecutive action scenes from the video. In addition, it is worth mentioning that VML tasks have penetrated many industrial fields [7] (e.g., Intelligent Security and Human–Machine Interaction) and also have beneficial impacts on several related cross-modal retrieval tasks, such as visual dialogue [8], video captioning [9,10], and so on.
To solve this intractable yet meaningful task, we will face two significant challenges. (1) Text queries and videos exhibit significant heterogeneity since they come from different modalities, so there is a substantial semantic gap between them. (2) The background and action scenes presented in videos are complicated and dynamic. Effectively tackling these issues requires both a thorough comprehension of video content and accurate cross-modal semantic alignment.
To overcome these challenges, researchers have developed various strategies to address VML tasks effectively. Until now, mainstream approaches have been generally classified into two categories. (1) Proposal-free methods, which instead of generating abundant moment candidates, immediately predict the start and end time of the target segment or the probability distributions of the two boundaries over all frames within a video. Due to the inherent characteristics of this sort of approach, it is no longer necessary to generate moment candidates, so it is computationally efficient. However, they mainly focus on directly predicting a more precise boundary while ignoring the rich visual content in video clips, leading to mediocre localization results. (2) Proposal-based methods follow the pipeline of Object Detection generally. Specifically, it first generates many moment candidates in a pre-defined manner and then predicts their confidence scores. Subsequently, it sorts these moment candidates according to their confidence scores and regards the candidate with the highest score as the final result. Here, it is worth noting that the familiar way to generate moment candidates is the sliding window method, obtaining abundant moment candidates by dense sampling with multiple windows of different sizes on the time dimension of videos. Compared with proposal-free methods, their final prediction results are more impressive due to the more accurate comprehension of the visual content in the video. Nonetheless, proposal-based methods also have drawbacks, such as the heavy computational burden and over-relying on the heuristic algorithms to generate moment candidates.
Motivated by this, we strive to propose a novel Temporal Map based Boundary Refinement Network (TMBRN) in this paper. Indeed, our solution also follows the pipeline of proposal-based methods because of its better localization performance. Specifically, to solve the problem of relying excessively on heuristic algorithms to generate moment candidates in most previous proposal-based methods, we construct a 2D temporal feature map to obtain moment candidates flexibly. As presented in Figure 1, the value in the horizontal and vertical axis in the 2D coordinate system represents the start’s and end’s clip index, respectively. Notably, each grid in the temporal map represents a moment proposal (candidate). We can obtain abundant moment proposals with extremely fine-grained time spans in such an advanced way. However, in conventional proposal-based methods, the moment candidate with the highest confidence score will be directly regarded as the final localization result. Nonetheless, no matter how fine the granularity of grid partitioning is, it is always inconsistent with the start and end times of the target moment. Therefore, besides the confidence scores’ prediction network, we introduce a boundary refinement network based on such a temporal map to fine-tune the endpoints of the moment candidates further, so that they are as close as possible to the boundaries of the target moment. Meanwhile, to reduce the heavy computation burden caused by the generated massive moment candidates, we develop a multi-step sampling strategy to drop some unnecessary moment proposals.
Meanwhile, we observe that the existing models do not consider the problem that moment candidates with different qualities may have different effects on model training. For instance, as shown in Figure 2, both proposal 1 and proposal 2 are sampled through the constructed temporal map, but proposal 1 is closer to the ground truth. If treating these moment proposals equally in the actual model learning process, we will fail to take full advantage of the superiority of proposal 1 over proposal 2. To solve this problem, we propose a novel weighted refinement loss function to strengthen the boundary refinement process of proposal 1 and weaken that of proposal 2. As far as we know, in the VML task, we are the first to incorporate a weighted factor into the loss function to distinguish the boundary refinement process for moment candidates with different qualities.
In conclusion, our main contributions are summarized below:
  • In this paper, a Temporal Map-based Boundary Refinement Network framework is proposed to handle VML tasks. Firstly, in addition to predicting the confidence scores for moment candidates, we introduce a boundary refinement network based on a temporal map to fine-tune the boundaries of these generated moment candidates. Then, to reduce the heavy computation burden, we leverage a multi-step sampling strategy to drop some redundant moment candidates.
  • We innovatively design a weighted refinement loss function to guide the model to focus more on the boundary refinement process of those moment candidates closer to the ground truth, thereby further improving the model’s performance.
  • Extensive experiments are conducted on two datasets, and the results demonstrate that our proposed model has outperformed the state-of-the-art methods.
Roadmap. Moreover, the remainder of this paper is structured as follows. In Section 2, we introduce the related works. In Section 3, the proposed approach is described exhaustively. In Section 4, we present comprehensive experimental results and in-depth analysis to validate our approach. Finally, in Section 5, we conclude this paper. In addition, all vectors and matrices are denoted in bold.

2. Related Work

In this section, according to the evolution of video retrieval issues, we will briefly describe some related work of text–video retrieval, temporal action localization, and video moment localization in sequence.

2.1. Text–Video Retrieval

The conventional video retrieval task is to find the entire video that includes the scene described by the text query from the video corpus. It is worth noting that this type of task does not need to localize the specific moment. In addition, video retrieval is a bidirectional task in many previous works. Namely, the video can be searched through the given text, and the relevant text can also be found through the given video. Most of the early studies about video retrieval were based on the idea of common space [11,12,13]. That is to say, they project both text and video into a joint common space, where semantically identical video and text are closer. To improve retrieval efficiency, ref. [14] proposed a deep hash framework for a large-scale video similarity search to learn lightweight yet very effective binary hash codes. In [15], aiming to obtain a comprehensive understanding of both text and video content, a dual deep encoding network was proposed to represent video and query into three levels of features, which contain global, temporal-aware, and local encoding. As traditional approaches tend to ignore the deep fusion between text and visual content, Cao et al. [16] proposed a novel co-attention network to explicitly emphasize the cross-modal feature interaction between video and recipe (text). Moreover, conventional common space learning is not enough to represent complex visual and textual details. To solve this problem, Chen et al. [17] introduced a hierarchical graph reasoning (HGR) model, which decomposed the text into a multilevel semantic graph containing three levels of scenario, actions, and entities. Similarly, hierarchical text representations are generated through attention-based graph reasoning. By this, the video representation with different levels can be guided to perform cross-modal matching for capturing global and local interactive details. However, to realize video retrieval of complex queries, a tree-augmented cross-modal encoding method was proposed by jointly learning the semantic structure of text query and the temporal representation of video [18].

2.2. Temporal Action Localization

The purpose of the temporal action localization task is to localize all small segments from a video and predict their classes of action. For this kind of task, Shou et al. [19] proposed a sequential multi-stage CNN network, which includes a proposal identification network, a classification network, and a localization network. These three sub-networks were designed to identify, classify, and regress candidates, respectively, achieving precise action localization. Meanwhile, the reuse of feature units guaranteed fast and efficient computation. Buch et al. [20] introduced the single-stream temporal action proposal network (SST), a new efficient deep architecture to generate candidates. Specifically, it can run continuously in long frame sequences through a single stream and eliminate the necessity of splitting the input video into short overlapping clips or temporal windows. Most of the time, many previous methods handled all action proposals individually and did not explicitly take advantage of their correlations in the model learning phase. However, the relationship between candidates is necessary for temporal action localization. Therefore, Zeng et al. [21] utilized graph convolutional networks (GCNs) to explore the potential relations between candidates.

2.3. Video Moment Localization

Compared to video retrieval and temporal action localization tasks, the video moment localization task is more suitable for the actual situation. Therefore, researchers have paid more and more attention to this open issue in recent years. In short, previous research solutions can be roughly divided into two categories: proposal-based and proposal-free methods.
The proposal-based methods generally follow a proposal-and-rank pipeline. Specifically, all moment candidates are generated first and sorted by their predicted confidence scores, and then the one with the highest score is regarded as the final prediction result. Chen et al. [22] proposed a novel temporal grounding network to capture the evolving fine-grained frame–word interaction between video and natural language in real time. In contrast to the traditional sliding window method, the final localization result was acquired by extracting the video once, which can greatly reduce the prediction time cost. Zhang et al. [23] proposed a cross-modal interaction network encompassing GCNs and the Multi-Head Self-Attention module to learn more fine-grained representation for each modality. Xu et al. [24] introduced a query-guided segment proposal network, which integrated the features of text query in the early phase of moment candidate generation to avoid the production of irrelevant video segments. In [25], the authors proposed a fast video moment retrieval model, which introduced a fine-grained semantic distillation framework to improve the efficiency of the common space.
The proposal-free methods directly predict the two boundaries of the target moment or the probability distribution of the boundary in all video frames. Reference [26] proposed a collaborative attention mechanism to achieve multi-modal interaction effects and enhance the localization precision. Lu et al. [27] proposed a new dense bottom-up grounding framework, which regarded all frames in the ground truth interval as foreground, and each foreground frame carried out a specific distance regression according to the offset between it and the bi-boundary. Mun et al. [28] proposed a hierarchical feature interaction model to extract middle-level features corresponding to significant semantic entities for textual phrases. Li et al. [29] devised a context pyramid network to study multi-scale temporal correlation in the video.
In addition to the above two kinds of mainstream video moment localization methods, many scholars have applied the currently popular reinforcement learning technology to address VML tasks. The conventional proposal-based methods have to generate abundant candidates, which leads to substantial computational overhead. For this problem, He et al. [30] described the task as a sequential decision problem by learning an agent that gradually adjusts the time boundary according to some specific strategies. Specifically, they proposed a reinforcement learning framework based on multi-task learning, which takes additional supervised boundary information into account during the training phase, and finally improves the location performance. To solve the issue that most existing methods are inefficient and deviate from the characteristics of human perception, Wu et al. [31] proposed a new progressive reinforcement learning framework with the tree structure strategy, which adjusts the temporal boundaries sequentially through an iterative process.

3. Proposed Approach

In this section, we first elaborate on the formulation of VML tasks. Then, we will describe every module of our proposed TMBRN in detail. In addition, the overview of the TMBRN is shown in Figure 3.

3.1. Problem Formulation

In the case of a given pair of a long untrimmed video V and a natural language query Q, the VML task aims to exactly localize a short segment S which is semantically the same as the text query. For computational convenience, video is denoted as V = { f i } i = 0 l V 1 , where f i indicates the i-th frame and l V represents the total number of frames in the video. In the same way, a natural language query can be indicated as Q = { q i } i = 0 l Q 1 , in which q i represents the i-th word and l Q stands for the total number of words in the text query. Therefore, the key challenge of VML tasks is how to locate a precise segment S = { f i } t s t e according to the textual query, where t s and t e represent the index of the start and end frame.

3.2. Feature Extraction

For a given text query Q, we first obtain the embedding w i R d q of the word q i by the released Glove Word2vec [32], where d q represents the embedding size. It is well known that LSTM (Long Short-Term Memory) [33] has demonstrated exceptional capability in extracting contextual information in the deep learning community. Therefore, we input the sentence query Q into a three-layer bidirectional LSTM, then integrate all hidden states at each time step. Consequently, we obtain the contextual feature of the query denoted as Q h R l Q × d h , where d h indicates the dimension of the hidden state. Formally, it can be represented as:
h i f = LSTM q f ( w i , h i 1 f ) ,
h i b = LSTM q b ( w i , h i 1 b ) ,
q i h = [ h i f ; h i b ] ,
where LSTM q f and LSTM q b are the forward and backward LSTM network, h i f and h i b represent the hidden states of the i-th unit of the forward and backward LSTM network, and [ ; ] stands for the concatenation operation of tensors.
Due to the nature of video data, it is essential to simultaneously analyze spatial and temporal information. Fortunately, some researchers have proposed several 3D convolutional neural networks (3D-CNNs) [34,35] to solve this problem. For the sequence of the input video frames, we split them into several small clips with the same length T. Subsequently, we perform uniform downsampling with fixed intervals to obtain N small clips. By this, we extract the video feature denoted as V R N × d v by the pre-trained 3D-CNN for every sampled clip. Then, a bidirectional LSTM is utilized to learn the local context information within the video, similar to the text query. In addition, the Transformer [36,37] has been proven to have a strong ability to capture long-distance dependencies in previous work, so we adopt it to obtain more informative video representations with global perspectives. As a result, we derive the contextual representation of the video denoted as V h R N × d h , where d h represents the dimension of the feature.

3.3. Cross-Modal Feature Fusion

Through the previous parts, we have gained the representations of videos and text queries. However, significant modality heterogeneity persists, necessitating effective feature fusion and precise semantic alignment between these distinct representations. In this paper, we use a Context–Query Attention (CQA) module to achieve the purpose of feature interaction [38,39]. Originally, the CQA module was widely used for QA (Question Answering) tasks, where C and Q stand for the context and query [40]. Specifically, we first calculate a correlation matrix M R N × l Q between the video and text, and the calculation formula is denoted as:
f ( v i , q j ) = w 0 [ v i ; q j ; v i q j ] ,
where ⊙ is the symbol of the Hadamard product and w 0 is a trainable vector for linear transformation. In addition, v i and q j indicate the feature vector of the i-th clip in the video and the j-th word in the text, respectively. Therefore, f ( v i , q j ) represents the similarity between v i and q j , which corresponds to the value of the i-th row and the j-th column in the matrix M .
Then, we can further obtain the text-aware video feature A v 2 q and the self-attended video feature A v 2 v by the attention mechanism, while the calculation formula is expressed as:
A v 2 q = M r · Q h ,
A v 2 v = M r · M c T · V h ,
where M r and M c are computed by row and column-wise SoftMax operation on similarity matrix M . Subsequently, we obtain the ultimate cross-modal representations between video and text, denoted as V q R N × d , which can be calculated by the following formula:
V q = MLP ( [ V h ; A v 2 q ; V h A v 2 q ; V h A v 2 v ] ) ,
where MLP is a Multi-Layer Perceptron network, which can linearly transform the dimension of the fused cross-modal feature to d.

3.4. Boundary Refinement Network

In this part, we first describe the concrete construction process of the 2D temporal map in detail, which is leveraged to obtain a mass of moment candidates with extremely fine-granularity time spans. Then, to reduce the heavy computation burden, we leverage a multi-step sampling strategy to drop some redundant moment candidates. In the end, we further introduce a boundary refinement network based on such a temporal map to fine-tune the boundaries of the moment candidates further.
Through the previous subsection, we have obtained the fused cross-modal feature V q , and we will elaborate on the construction process of the temporal feature map in this part. We assume that the feature of a moment proposal can be represented as f a b = m a x p o o l ( V a q , V a + 1 q , · · · , V b q ) , where V a q indicates the fused feature of the a-th clip in video and 0 a b N . Therefore, we can obtain diverse moment candidates with many different time spans. In this manner, we can construct a whole 2D temporal feature map, where each grid corresponds to a video moment proposal, and their features are obtained by consecutive max-pooling. Formally, this temporal feature map can be denoted as F R N × N × d . Furthermore, it is worth noting that the start clip index of each moment candidate in the feature map must not be larger than the end clip index. Therefore, moment candidates below the diagonal are invalid in the temporal feature map. In our experimental implementations, we pad these moment candidates with zero vectors.
However, the total number of moment candidates in such a temporal feature map is still substantial, which will bring a heavy computation burden. To solve this problem, we use a multi-step sampling strategy to remove some redundant moment candidates. Namely, we perform dense sampling when the length of the moment candidate is short. On the contrary, as the length of the moment proposal becomes longer, the sampling probability becomes lower and lower. Specifically, when the size of the temporal map N 16 , we sample all moment candidates in the map. For simplicity, we adopt N = 64 as an example to illustrate the proposed strategy. As shown in Figure 4, we divide all moment candidates into four regions based on the length. In addition, indices of the moment candidate’s start and end clip can be denoted as i and j. When 0 j i 15 , i.e., the lengths of moment candidates range from 1 to 16 clips, we fetch all moment proposals. If 16 j i 31 , i.e., the time span of moment candidates ranges from 17 to 32 clips, we obtain moment proposals with a two-clip stride at every specific moment candidate length. In detail, we take out proposals from the set of moment candidate length L { 18 , 20 , , 32 } . Similarly, when the moment candidate lengths is from 33 to 48 (i.e., 32 j i 47 ), we obtain moment proposals with a three-clip stride at every specific moment candidate length. The corresponding set of moment candidate length is L { 35 , 38 , , 47 } . Finally, as the moment candidate length is from 49 to 64 (i.e., 48 j i 63 ), we obtain moment proposals with a four-clip stride at every specific moment candidate length. The corresponding set of moment candidate lengths is L { 52 , 56 , 60 , 64 } . In addition, it is worth explaining that we adopt such a strategy because of a common phenomenon that the video segment needed to locate generally takes up less than one-quarter of the length of the whole video in most cases, and the longer the target segment, the lower the probability. Therefore, the longer the duration of moment candidates, the sparser the sampling. Furthermore, we can use the above rule to infer when N > 64 .
Most of the previous proposal-based methods predominantly relied on ingenious strategies to generate high-quality moment candidates, then predicted their confidence scores and directly selected the moment proposal with the highest score as the final localization result. However, as is depicted in Figure 5, although these moment candidates are segmented so exquisitely through the temporal map in this paper, there are still some deviations compared with the ground truth. Motivated by this, in addition to the usual confidence score prediction network, we innovatively introduce a boundary refinement network based on the constructed temporal feature map to continuously fine-tune the boundaries of generated moment candidates, further improving localization performance.
As we know, continuous convolution can gradually expand the receptive field and constantly fuse the adjacent context information [6,41]. Inspired by this, the boundary refinement network encompasses L continuous convolutional layers, the first L 1 convolutional layers are leveraged to fuse context information continuously, and the last convolutional layer followed with an activation function is used for scoring and boundary refining. To guarantee that the shape of the output is equal to the input, we perform proper padding before each convolution operation. In addition, it is worth noting that moment candidates padded with zero in the temporal feature map are not taken into computation since they are invalid.
Finally, we obtained a confidence score map and an offsets map based on the temporal feature map. Here, the offsets predicted by the boundary refinement network refer to the gap between the two boundaries of each moment candidate in the temporal map and the ground truth. Then, we collect the confidence scores and offsets of all valid moment candidates in the temporal map denoted as P ^ = { p ^ i } i = 0 C , O ^ = { o ^ i s , o ^ i e } i = 0 C , respectively, where C stands for the number of valid moment candidates. Therefore, the one with the highest confidence score is regarded as the correct moment candidate, and the final prediction result is the sum of the original timestamp of this candidate and its corresponding predicted offsets.

3.5. Loss Function

As shown in Figure 1, the start and end boundaries of each moment candidate in the temporal feature map can be expressed as ( i δ , ( j + 1 ) δ ) , where i and j stand for the indexes of start and end clips in the temporal map, respectively. Moreover, δ indicates a unit of time in the temporal map, which refers to the length of each grid (converted to absolute time). Specifically, it can be calculated with δ = D N , where D stands for the duration of the entire video. Based on the above description, we can calculate the IoU (Intersection of Union) between each moment candidate and ground truth in the temporal map. As shown in Figure 6, the IoU can be computed by the following formula:
IoU = Area of overlap Area   of   union = min ( t e ^ , t e ) max ( t s ^ , t s ) max ( t e ^ , t e ) min ( t s ^ , t s ) ,
where t s ^ and t e ^ represent the start and end time of the predicted video segment by the boundary refinement network, while t s and t e indicate the start and end time of the ground truth.

3.5.1. Alignment Loss

Based on this, the IoU of the i-th moment candidate in the temporal map can be readily calculated and denoted as p i . Meanwhile, we can also obtain the label of offsets between every moment candidate in the temporal map and ground truth, which can be denoted as o i = ( o i s , o i e ) . In the process of TMBRN training, we utilize a scaled IoU as a label to replace the traditional hard binary label [25], which can be calculated by the following formula:
y i = 0 p i t m i n , p i t m i n t m a x t m i n t m i n < p i < t m a x , 1 p i t m a x ,
where t m i n and t m a x represent two thresholds used to scale the IoU label. Consequently, y i can serve as a supervised signal in the model training phase.
To bridge the semantic gap between the text and video, we introduce an alignment loss function based on binary cross-entropy loss, which can be expressed by the following formula:
L a l i g n = 1 C i = 1 C y i log p ^ i + ( 1 y i ) log ( 1 p ^ i ) ,
where C stands for the number of valid sampled moment candidates in the temporal map and p ^ i represents the predicted confidence score of the i-th moment proposal.

3.5.2. Weighted Refinement Loss

In addition to the alignment loss, it is essential to incorporate the loss of offsets predicted by the boundary refinement network. Because the boundaries of the moment candidates generated by the temporal map have been fixed, the model’s performance will be limited if we train the TMBRN only with the alignment loss. Motivated by this, we introduce a refinement loss to train the model jointly. Because many moment candidates significantly deviate from the target video segment (i.e., the IoUs between them with ground truth are very small), it is unprofitable to refine these moment proposals. To handle this dilemma, we devise an unprecedented weighted refinement loss function. Specifically, we inject the IoUs between the original boundaries of moment candidates and the ground truth into the naive refinement loss as a weighted factor. By this, we can guide the network to pay more attention to fine-tuning those moment candidates closer to the ground truth, and then improve the localization performance of the model. Formally, the weighted refinement loss function can be expressed as:
L w r e f = 1 C i = 0 C y i [ R ( o i s o ^ i s ) + R ( o i e o ^ i e ) ] ,
where R ( · ) is the smooth L 1 function. It is worth noting that we adopt the scaled IoU y i as a weighted factor here. Furthermore, we can reduce training computation costs to some extent in this way. The reason is that the weighted factor of those moment candidates having no overlap with the ground truth will be assigned 0.
Combined with that mentioned above, the total loss is expressed as:
L = L a l i g n + λ × L w r e f ,
where λ is a hyperparameter used to balance two loss functions.
During inference, we sort all valid moment candidates by the predicted confidence scores and the one with the highest score is our need. It is worth mentioning that the final result should be the sum of the original boundaries of each moment candidate (converted to absolute time) in the temporal map and its corresponding refined offsets. In addition, we use the NMS (Non-Maximum Suppression) to drop some redundant moment candidates.

4. Experiments

We conduct extensive experiments on two popular and challenging datasets, i.e., Charades-STA [42] and ActivityNet Captions [43], to evaluate the efficiency of our proposed TMBRN model. In this section, we first briefly introduce these two datasets, the evaluation metrics in the experiments, and the relevant details of the experimental implementation. Then, we compare the experimental results of the proposed model with the state-of-the-art methods in recent years and analyze them legitimately. At last, we conduct several ablation experiments to quantify the impact of various components in the TMBRN model on performance and present the visualized video moment localization result.

4.1. Datasets

Charades-STA: It is built on the Charades dataset [44], which was originally used for the temporal action localization task. Charades contains 9848 videos of daily indoor activity, with an average duration of about 30 s. Moreover, it includes 15 activity scenes and 157 action categories. On this basis, Gao et al. split these videos in Charades into several small segments according to the initial annotations, and constructed the Charades-STA dataset by annotating each segment with a start time, an end time, and a sentence. Following the previous work [42], we utilize 12,408 video–text pairs to constitute the training set and 3720 pairs to compose the test set in our experiments. ActivityNet Captions: It is built on the ActivityNet v1.3 dataset [45], which contains 19,994 untrimmed YouTube videos. This dataset includes approximately 100 K descriptive sentences and the average duration of videos is about 120 s. Each video contains at least two annotations, the timestamp label (i.e., the start and end time), and the hand-written sentence description. Therefore, according to the official split principle, we used 37,421 moment–text pairs to construct the training set, 17,505 pairs to form the validation set, and 17,031 pairs to form the test set.

4.2. Evaluation Metrics

Similar to the evaluation metric of previous work, we also utilize R n @ m as the key indicator to evaluate the experiment results of our proposed approach. m is a threshold, which means that if the IoU value between the predicted result and the ground truth is greater than it, we regard this result as a correct prediction. n stands for the top-n results sorted by confidence scores. Therefore, R n @ m is the percentage of samples having at least one correct prediction in the top-n results. In addition, for these two datasets, we adopt n { 1 , 5 } and m { 0.5 , 0.7 } for Charades-STA, and n { 1 , 5 } and m { 0.3 , 0.5 , 0.7 } for ActivityNet Captions.

4.3. Implementation Details

To make a fair comparison with some previous work, we maintain identical experimental setups. When extracting the embedding of each word in sentences, we use the pre-trained Glove-300 word2vec model [32]; hence, the length of all word vectors is 300, i.e., d q = 300 . We statistically analyzed the length of all sentence queries (i.e., the number of words) in the dataset, and to make the code implementation more convenient, we set the length of all sentences to a uniform value L Q . Specifically, when the length of the sentence is greater than L Q , we truncate the subsequent content, whereas we pad with zero vectors for shorter sentences. We set the L Q to 10 for Charades-STA and 25 for ActivityNet Captions. As for the video feature, we utilize several pre-trained 3D ConvNets to extract. Here, we adopt the C3D [34] and the VGG [46] features for Charades-STA and the C3D feature for ActivityNet Captions. The length of small clips T is set to 16 frames and the number of downsampled clips N is set to 16 for Charades-STA and 64 for ActivityNet Captions. In the Transformer module, which is leveraged to capture long-distance dependencies in the video, the number of attention heads is set to 8. In this paper, the dimension d h of all hidden states is uniformly set to 512. In the boundary refinement network, we leverage continuous L convolution layers with kernel size K, and we set L = 8 and K = 5 for Charades-STA and L = 4 and K = 9 for ActivityNet Captions. The threshold t m a x and t m i n are set to 1.0 and 0.5 for all datasets. The hyperparameter λ used to balance alignment and refinement loss is set to 10.0 for Charades-STA and 0.1 for ActivityNet Captions. Moreover, the NMS threshold is set to 0.4 for Charades-STA and 0.5 for ActivityNet Captions. The learning rate and batch size are set to 1 × 10 4 and 32, respectively. We use both these two datasets to train the TMBRN with 100 epochs. The PyTorch framework (torch 1.10.0 version) is used in our experiments, and Adam is selected as the optimizer; then, the model is trained on a computer equipped with two NVIDIA RTX3090 GPUs.

4.4. Performance Comparison

We evaluate the proposed model through extensive experiments on two extremely challenging benchmarks: the Charades-STA and ActivityNet Captions datasets. To demonstrate the superiority of the TMBRN, we compared it with the state-of-the-art approaches in recent years, including:
  • Proposal-based methods: CTRL [42], MCN [47], ACRN [48], ROLE [49], TGN [22], MAC [50], QSPN [24], SAP [51], SCDM [52], 2D-TAN [53], CBP [54], BPNet [55], FVMR [25], TLL [56], MIM [57].
  • Proposal-free methods: DEBUG [27], ABLR [26], GDP [58], LGI [28], PMI [59], CPNet [29], PLPNet [60], TCFN [61], STCNet [62], FVC [63].
  • Reinforcement learning methods: RWM [30], SM-RL [64], TSP-PRL [31], TripNet [65].
The results on the Charades-STA dataset are presented in Table 1, and the results on the ActivityNet Captions dataset are shown in Table 2. The values of best performance in the table are highlighted in bold font. It is worth noting that we compare the experimental results on both C3D and VGG features in the Charades-STA dataset.
Table 1 shows the experimental results of the TMBRN on the Charades-STA dataset. As demonstrated in the table, our model achieves almost the best results with C3D and VGG visual features. When using C3D features, the TMBRN has an improvement of 5.64% over the second-best FVMR on the R5@0.5 evaluation metric. For VGG features, our TMBRN outperforms state-of-the-art methods on all metrics. In addition, our proposed model has a more significant increase on the R5 metrics than the R1 metrics compared to the other models. This may be because moment candidates provided by the temporal feature map in the TMBRN are more flexible than those conventional proposal-based methods, which can achieve the goal of generating more reasonable moment candidate segments with different time spans.
Similarly, the experimental results of the TMBRN on the ActivityNet Captions dataset are presented in Table 2. The data presented in Table 2 clearly demonstrate that our proposed model achieves the best results on all metrics. Therefore, it confirms that the TMBRN predicts the score of moment candidates more accurately and the refined offset is also precise in the prediction process. In addition, it benefits from the weighted factor of the refinement loss function. Because of this, when training the TMBRN, we can urge the network to concentrate more on fine-tuning those moment candidates closer to the ground truth. As a result, the localization performance can be further improved in this way.

4.5. Ablation Studies

In this part, we conducted ablation experiments to elaborate on the influence of some components of the proposed model and parameter settings on the localization precision. These include the number of stacked convolution layers and the size of the convolution kernel in the boundary refinement network, the setting of upper and lower thresholds of the scaled IoU, the weighted factor of the refinement loss, the feature fusion module CQA, and the hyperparameter λ used to balance alignment loss and refinement loss.
Figure 7 shows the impact of the number of convolution layers and the size of the convolution kernel on localization performance in the boundary refinement network. When studying the number of convolution layers, we first empirically fixed the convolution kernel size K = 9 , and we can see that the best performance occurs when L = 4 from the left chart. However, the performance declines when the number of convolution layers increases continuously. This phenomenon can be attributed to the potential loss of crucially original features when stacking excessive convolutional layers. Similarly, we fixed the number of convolution layers L = 4 , and we can intuitively observe from the right subgraph that the localization precision reaches the maximum when K = 9 . When K < 9 , the larger the convolution kernel, the better the performance, which means that a small convolution kernel cannot effectively aggregate the context information of adjacent moment candidates. When K > 9 , as the size of the convolution kernel increases, the performance begins to decline. This may be because if the convolution kernel is too large, information fusion is not smooth, degrading the model’s localization performance. Therefore, the more convolution layers and the larger convolution kernel sizes do not denote better performance. Moreover, these two experimental results convincingly verify that the values of K and L in this paper are set reasonably.
Figure 8 shows the impact of different threshold settings of the scaled IoU. The horizontal axis represents the difference value between the threshold t m a x and the threshold t m i n , while the three bars in each group denote the t m i n = 0.3 / 0.4 / 0.5 , respectively. As can be seen from the figure, the smaller the difference value between the upper and lower thresholds, the lower the overall localization accuracy of the model. This may be because when all the IoU values are converted into a smaller interval, the diversity of the original IoU values is lost, which conversely makes the training effect of alignment loss worse. In addition, the larger the t m i n , the better the model performance. We think the potential reason is that the scaled IoU solution directly regards moment proposals with a small IoU value as invalid moment candidates, which can reduce the complexity of the prediction.
Table 3 and Table 4 present the importance of the weighted factor of the refinement loss on the Charades-STA and the ActivityNet Captions dataset. These two tables reveal that the absence of the weighting factor in the refinement loss (i.e., the first row) leads to a substantial degradation in model performance across all evaluation metrics. We preliminarily consider the reason to be that in the training phase, where we added the scaled IoUs as the weighted factor, the model can pay more attention to refining those moment proposals closer to the ground truth, and weaken other moment candidates that are far away from the ground truth, thereby making the model training more efficient. In addition, to further demonstrate this thought, we utilize the original IoU values to replace the scaled IoUs as the weighted factor of (11) (i.e., the second row). Finally, we can conclude that while its performance is inferior to the complete TMBRN, it is much better than the refinement loss function without any weighted factor.
Table 5 and Table 6 demonstrate the importance of the scaled IoU on the Charades-STA and ActivityNet Captions datasets, respectively. As shown in these two tables, the higher the confidence threshold, the more significant the improvement in the scaled IoU. This is also highly consistent with the goal of our VML tasks, i.e., as much overlap as possible between the predicted fragment and the ground truth.
Table 7 shows the results of the TMBRN with different multi-modal feature fusion strategies on the ActivityNet Captions dataset. In this paper, we use the feature fusion scheme of CTRL and the cross-attention mechanism to conduct ablation experiments. It can be intuitively seen from the table that replacing the CQA module of the TMBRN with the fusion module of CTRL leads to the worst performance. Because the feature fusion module of CTRL only comprises some plain tensor operations, e.g., element-wise multiplication and concatenation, it fails to align multi-modal semantics well. The model’s performance improved significantly when using the cross-attention mechanism to integrate features, but it is still worse than utilizing the CQA module. However, even if we fuse cross-modal features without the CQA module, the TMBRN surpasses some recent methods, proving its superiority and robustness.
Figure 9 shows the impact of the hyperparameter λ , which is used to balance the alignment loss and the weighted refinement loss on the performance of the TMBRN. It can be seen from the figure that both a larger and a smaller λ will degrade the model performance. Across all metrics, the best performance is achieved when λ = 10 in this paper. Through such an experiment, we find that the two terms of (12) are of the same magnitude order only when λ = 10 . This indicates that in the actual model training phase, we need to set an appropriate λ to make the model treat the two kinds of loss equally, which is beneficial to improving the localization precision of the model.

4.6. Visualization Results

Figure 10 demonstrates some visualized results, which can intuitively verify the reliability of our devised TMBRN. As shown in the top subfigure, for the given query, “A person takes a book out from the entertainment center”, the TMBRN generates a video segment whose IoU is about 91%. However, this is the result of the Charades-STA dataset, in which the duration of the video is relatively shorter. As a comparison, in the bottom subfigure, we also show the visualized results on the ActivityNet Captions dataset in which the video has a 2 min average duration. For the more complex query, “The man takes out the car, and other men start fighting for the car, and the three are riding it”, which contains three different actions, the TMBRN also outputs a video segment with 90% localization precision. This further confirms that our proposed TMBRN model has strong stability and high reliability.

5. Conclusions

In this paper, we develop a Temporal Map-based Boundary Refinement Network (TMBRN) for video moment localization tasks. We obtain abundant moment candidates with extremely fine-granularity time spans by a constructed 2D temporal map. Based on such a temporal map, in addition to predicting the confidence scores, we introduce a boundary refinement network to fine-tune the boundaries of moment candidates to obtain more precise localization results. Meanwhile, we adopt a multi-step sampling strategy to remove some redundant moment candidates to reduce the heavy computation burden. Moreover, we design an innovative weighted refinement loss function to enable the model to focus more on refining the boundaries of those moment candidates closer to the ground truth, greatly heightening the model’s performance. Finally, by extensive experiments on two publicly available datasets, the results suggest that our proposed method is highly effective for handling VML tasks and outperforms the state-of-the-art methods.
Meanwhile, our model also has some weaknesses; for example, for complex textual queries (including multiple actions), it may not be able to accurately model its semantics, resulting in poor multi-modal semantic alignment and inferior localization performance. Our future work may consider the development of more effective text encoders, obtaining informative contextual features to enhance the capability of language modeling.

Author Contributions

L.L. implemented the model and wrote the original draft. D.L. reviewed and edited the manuscript. C.Z. revised the paper. Y.Z. and H.R. analyzed the experimental results and prepared figures. L.Z. acquired funding. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grants 62472161, 61836016, and 62202163, and the Natural Science Foundation of Hunan Province under Grants 2023JJ30169 and 2022JJ40190.

Data Availability Statement

All datasets can be accessed through their respective official websites or mirrors. The Charades-STA dataset can be accessed at https://prior.allenai.org/projects/charades. The ActivityNet Captions dataset can be accessed at http://activity-net.org/challenges/2016/download.html.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tian, C.; Zheng, M.; Li, B.; Zhang, Y.; Zhang, S.; Zhang, D. Perceptive self-supervised learning network for noisy image watermark removal. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7069–7079. [Google Scholar] [CrossRef]
  2. Zhang, J.; Zhao, B.; Zhang, W.; Miao, Q. Knowledge Translator: Cross-Lingual Course Video Text Style Transform via Imposed Sequential Attention Networks. Electronics 2025, 14, 1213. [Google Scholar] [CrossRef]
  3. Tian, C.; Zheng, M.; Zuo, W.; Zhang, S.; Zhang, Y.; Lin, C.W. A cross transformer for image denoising. Inf. Fusion 2024, 102, 102043. [Google Scholar] [CrossRef]
  4. Liu, J.; Liu, W.; Han, K. MNv3-MFAE: A Lightweight Network for Video Action Recognition. Electronics 2025, 14, 981. [Google Scholar] [CrossRef]
  5. Zhang, H.; Li, Z.; Yang, J.; Wang, X.; Guo, C.; Feng, C. Revisiting Hard Negative Mining in Contrastive Learning for Visual Understanding. Electronics 2023, 12, 4884. [Google Scholar] [CrossRef]
  6. Tian, C.; Zhang, X.; Liang, X.; Li, B.; Sun, Y.; Zhang, S. Knowledge distillation with fast CNN for license plate detection. IEEE Trans. Intell. Veh. 2023, 1–7. [Google Scholar] [CrossRef]
  7. Tian, C.; Xiao, J.; Zhang, B.; Zuo, W.; Zhang, Y.; Lin, C.W. A self-supervised network for image denoising and watermark removal. Neural Netw. 2024, 174, 106218. [Google Scholar] [CrossRef] [PubMed]
  8. Matsumori, S.; Shingyouchi, K.; Abe, Y.; Fukuchi, Y.; Sugiura, K.; Imai, M. Unified questioner transformer for descriptive question generation in goal-oriented visual dialogue. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1898–1907. [Google Scholar]
  9. Song, J.; Guo, Y.; Gao, L.; Li, X.; Hanjalic, A.; Shen, H.T. From deterministic to generative: Multimodal stochastic RNNs for video captioning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 3047–3058. [Google Scholar] [CrossRef] [PubMed]
  10. Fu, Y.; Yang, Y.; Ye, O. Video Captioning Method Based on Semantic Topic Association. Electronics 2025, 14, 905. [Google Scholar] [CrossRef]
  11. Wang, B.; Yang, Y.; Xu, X.; Hanjalic, A.; Shen, H.T. Adversarial cross-modal retrieval. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 154–162. [Google Scholar]
  12. Xu, X.; Wang, T.; Yang, Y.; Zuo, L.; Shen, F.; Shen, H.T. Cross-modal attention with semantic consistence for image–text matching. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5412–5425. [Google Scholar] [CrossRef] [PubMed]
  13. Shen, H.T.; Liu, L.; Yang, Y.; Xu, X.; Huang, Z.; Shen, F.; Hong, R. Exploiting subspace relation in semantic labels for cross-modal hashing. IEEE Trans. Knowl. Data Eng. 2020, 33, 3351–3365. [Google Scholar] [CrossRef]
  14. Wu, G.; Han, J.; Guo, Y.; Liu, L.; Ding, G.; Ni, Q.; Shao, L. Unsupervised deep video hashing via balanced code for large-scale video retrieval. IEEE Trans. Image Process. 2018, 28, 1993–2007. [Google Scholar] [CrossRef] [PubMed]
  15. Dong, J.; Li, X.; Xu, C.; Ji, S.; He, Y.; Yang, G.; Wang, X. Dual encoding for zero-example video retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9346–9355. [Google Scholar]
  16. Cao, D.; Yu, Z.; Zhang, H.; Fang, J.; Nie, L.; Tian, Q. Video-based cross-modal recipe retrieval. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1685–1693. [Google Scholar]
  17. Chen, S.; Zhao, Y.; Jin, Q.; Wu, Q. Fine-grained video-text retrieval with hierarchical graph reasoning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10638–10647. [Google Scholar]
  18. Yang, X.; Dong, J.; Cao, Y.; Wang, X.; Wang, M.; Chua, T.S. Tree-augmented cross-modal encoding for complex-query video retrieval. In Proceedings of the 43rd international ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 25–30 July 2020; pp. 1339–1348. [Google Scholar]
  19. Shou, Z.; Wang, D.; Chang, S.F. Temporal action localization in untrimmed videos via multi-stage cnns. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1049–1058. [Google Scholar]
  20. Buch, S.; Escorcia, V.; Shen, C.; Ghanem, B.; Carlos Niebles, J. Sst: Single-stream temporal action proposals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2911–2920. [Google Scholar]
  21. Zeng, R.; Huang, W.; Tan, M.; Rong, Y.; Zhao, P.; Huang, J.; Gan, C. Graph convolutional networks for temporal action localization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7094–7103. [Google Scholar]
  22. Chen, J.; Chen, X.; Ma, L.; Jie, Z.; Chua, T.S. Temporally grounding natural sentence in video. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 162–171. [Google Scholar]
  23. Zhang, Z.; Lin, Z.; Zhao, Z.; Xiao, Z. Cross-modal interaction networks for query-based moment retrieval in videos. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 655–664. [Google Scholar]
  24. Xu, H.; He, K.; Plummer, B.A.; Sigal, L.; Sclaroff, S.; Saenko, K. Multilevel language and vision integration for text-to-clip retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9062–9069. [Google Scholar]
  25. Gao, J.; Xu, C. Fast video moment retrieval. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1523–1532. [Google Scholar]
  26. Yuan, Y.; Mei, T.; Zhu, W. To find where you talk: Temporal sentence localization in video with attention based location regression. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9159–9166. [Google Scholar]
  27. Lu, C.; Chen, L.; Tan, C.; Li, X.; Xiao, J. Debug: A dense bottom-up grounding approach for natural language video localization. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 5144–5153. [Google Scholar]
  28. Mun, J.; Cho, M.; Han, B. Local-global video-text interactions for temporal grounding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10810–10819. [Google Scholar]
  29. Li, K.; Guo, D.; Wang, M. Proposal-free video grounding with contextual pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 19–21 May 2021; Volume 35, pp. 1902–1910. [Google Scholar]
  30. He, D.; Zhao, X.; Huang, J.; Li, F.; Liu, X.; Wen, S. Read, watch, and move: Reinforcement learning for temporally grounding natural language descriptions in videos. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8393–8400. [Google Scholar]
  31. Wu, J.; Li, G.; Liu, S.; Lin, L. Tree-structured policy based progressive reinforcement learning for temporally language grounding in video. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12386–12393. [Google Scholar]
  32. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  33. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  34. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar]
  35. Carreira, J.; Zisserman, A. Quo vadis, action recognition? a new model and the kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6299–6308. [Google Scholar]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  37. Tian, C.; Zheng, M.; Lin, C.W.; Li, Z.; Zhang, D. Heterogeneous window transformer for image denoising. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 6621–6632. [Google Scholar] [CrossRef]
  38. Xiong, C.; Zhong, V.; Socher, R. Dynamic Coattention Networks For Question Answering. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  39. Yu, A.W.; Dohan, D.; Le, Q.; Luong, T.; Zhao, R.; Chen, K. Fast and accurate reading comprehension by combining self-attention and convolution. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; Volume 2. [Google Scholar]
  40. Seo, M.; Kembhavi, A.; Farhadi, A.; Hajishirzi, H. Bidirectional Attention Flow for Machine Comprehension. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  41. Tian, C.; Zheng, M.; Jiao, T.; Zuo, W.; Zhang, Y.; Lin, C.W. A self-supervised CNN for image watermark removal. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7566–7576. [Google Scholar] [CrossRef]
  42. Gao, J.; Sun, C.; Yang, Z.; Nevatia, R. Tall: Temporal activity localization via language query. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5267–5275. [Google Scholar]
  43. Krishna, R.; Hata, K.; Ren, F.; Fei-Fei, L.; Carlos Niebles, J. Dense-captioning events in videos. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 706–715. [Google Scholar]
  44. Sigurdsson, G.A.; Varol, G.; Wang, X.; Farhadi, A.; Laptev, I.; Gupta, A. Hollywood in homes: Crowdsourcing data collection for activity understanding. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 510–526. [Google Scholar]
  45. Caba Heilbron, F.; Escorcia, V.; Ghanem, B.; Carlos Niebles, J. ActivityNet: A Large-Scale Video Benchmark for Human Activity Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 961–970. [Google Scholar]
  46. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  47. Anne Hendricks, L.; Wang, O.; Shechtman, E.; Sivic, J.; Darrell, T.; Russell, B. Localizing moments in video with natural language. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5803–5812. [Google Scholar]
  48. Liu, M.; Wang, X.; Nie, L.; He, X.; Chen, B.; Chua, T.S. Attentive moment retrieval in videos. In Proceedings of the 41st international ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 15–24. [Google Scholar]
  49. Liu, M.; Wang, X.; Nie, L.; Tian, Q.; Chen, B.; Chua, T.S. Cross-modal moment localization in videos. In Proceedings of the 26th ACM international Conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018; pp. 843–851. [Google Scholar]
  50. Ge, R.; Gao, J.; Chen, K.; Nevatia, R. Mac: Mining activity concepts for language-based temporal localization. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 245–253. [Google Scholar]
  51. Chen, S.; Jiang, Y.G. Semantic proposal for activity localization in videos via sentence query. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8199–8206. [Google Scholar]
  52. Yuan, Y.; Ma, L.; Wang, J.; Liu, W.; Zhu, W. Semantic conditioned dynamic modulation for temporal sentence grounding in videos. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar] [CrossRef] [PubMed]
  53. Zhang, S.; Peng, H.; Fu, J.; Luo, J. Learning 2d temporal adjacent networks for moment localization with natural language. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12870–12877. [Google Scholar]
  54. Wang, J.; Ma, L.; Jiang, W. Temporally grounding language queries in videos by contextual boundary-aware prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12168–12175. [Google Scholar]
  55. Xiao, S.; Chen, L.; Zhang, S.; Ji, W.; Shao, J.; Ye, L.; Xiao, J. Boundary proposal network for two-stage natural language video localization. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 19–21 May 2021; Volume 35, pp. 2986–2994. [Google Scholar]
  56. Bao, P.; Mu, Y. Learning Sample Importance for Cross-Scenario Video Temporal Grounding. In Proceedings of the 2022 International Conference on Multimedia Retrieval, Newark, NJ, USA, 27–30 June 2022; pp. 322–329. [Google Scholar]
  57. Li, J.; Zhang, F.; Lin, S.; Zhou, F.; Wang, R. Mim: Lightweight multi-modal interaction model for joint video moment retrieval and highlight detection. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 1961–1966. [Google Scholar]
  58. Chen, L.; Lu, C.; Tang, S.; Xiao, J.; Zhang, D.; Tan, C.; Li, X. Rethinking the bottom-up framework for query-based video localization. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10551–10558. [Google Scholar]
  59. Chen, S.; Jiang, W.; Liu, W.; Jiang, Y.G. Learning modality interaction for temporal sentence localization and event captioning in videos. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 333–351. [Google Scholar]
  60. Li, S.; Li, C.; Zheng, M.; Liu, Y. Phrase-level Prediction for Video Temporal Localization. In Proceedings of the 2022 International Conference on Multimedia Retrieval, Newark, NJ, USA, 27–30 June 2022; pp. 360–368. [Google Scholar]
  61. Lv, Z.; Su, B. Temporal-enhanced Cross-modality Fusion Network for Video Sentence Grounding. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 1487–1492. [Google Scholar]
  62. Wang, Y.; Li, K.; Chen, G.; Zhang, Y.; Guo, D.; Wang, M. Spatiotemporal contrastive modeling for video moment retrieval. World Wide Web 2023, 26, 1525–1544. [Google Scholar] [CrossRef]
  63. Shi, M.; Su, Y.; Lin, X.; Zao, B.; Kong, S.; Tan, M. Frame as Video Clip: Proposal-Free Moment Retrieval by Semantic Aligned Frames. IEEE Trans. Ind. Inform. 2024, 20, 13158–13168. [Google Scholar] [CrossRef]
  64. Wang, W.; Huang, Y.; Wang, L. Language-driven temporal activity localization: A semantic matching reinforcement learning model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 334–343. [Google Scholar]
  65. Hahn, M.; Kadav, A.; Rehg, J.M.; Graf, H.P. Tripping through time: Efficient localization of activities in videos. arXiv 2019, arXiv:1904.09936. [Google Scholar]
Figure 1. A diagram of VML tasks and the 2D temporal map.
Figure 1. A diagram of VML tasks and the 2D temporal map.
Electronics 14 01657 g001
Figure 2. A demonstration of the weighted refinement loss.
Figure 2. A demonstration of the weighted refinement loss.
Electronics 14 01657 g002
Figure 3. The architecture of the Temporal Map-based Boundary Refinement Network. The TMBRN primarily comprises three modules: feature extraction, feature fusion, and the boundary refinement network. The two blue dotted boxes of the figure represent the video and text encoding modules; the box filled with green is the cross-modal fusion module, and the red dotted box indicates the boundary refinement network based on the 2D temporal map.
Figure 3. The architecture of the Temporal Map-based Boundary Refinement Network. The TMBRN primarily comprises three modules: feature extraction, feature fusion, and the boundary refinement network. The two blue dotted boxes of the figure represent the video and text encoding modules; the box filled with green is the cross-modal fusion module, and the red dotted box indicates the boundary refinement network based on the 2D temporal map.
Electronics 14 01657 g003
Figure 4. A demonstration of the multi-step sampling on the 2D temporal map. In the Figure, the grids marked blue, orange, purple, and green represent moment proposals sampled with four different interval segments, respectively. The blue area denotes the moment candidates whose periods range from 1 to 16 clips, where we sampled all moment proposals. The orange region represents the moment candidates whose lengths range from 17 to 32 clips, sampled with a two-clip stride on both the horizontal and vertical axis in the 2D coordinate system. The purple area stands for the moment candidates whose durations range from 33 to 48 clips, sampled with a three-clip stride. Similarly, the green region indicates moment candidates whose lengths range from 49 to 64 clips, sampled with a four-clip stride.
Figure 4. A demonstration of the multi-step sampling on the 2D temporal map. In the Figure, the grids marked blue, orange, purple, and green represent moment proposals sampled with four different interval segments, respectively. The blue area denotes the moment candidates whose periods range from 1 to 16 clips, where we sampled all moment proposals. The orange region represents the moment candidates whose lengths range from 17 to 32 clips, sampled with a two-clip stride on both the horizontal and vertical axis in the 2D coordinate system. The purple area stands for the moment candidates whose durations range from 33 to 48 clips, sampled with a three-clip stride. Similarly, the green region indicates moment candidates whose lengths range from 49 to 64 clips, sampled with a four-clip stride.
Electronics 14 01657 g004
Figure 5. The diagram of the boundary refinement network.
Figure 5. The diagram of the boundary refinement network.
Electronics 14 01657 g005
Figure 6. An illustration of Intersection of Union (IoU).
Figure 6. An illustration of Intersection of Union (IoU).
Electronics 14 01657 g006
Figure 7. The ablation study of the number of stacked convolution layers and the convolution kernel size on the ActivityNet Captions dataset. The left chart shows the result of the R1@0.7 metric, and the right one represents the result of R5@0.7.
Figure 7. The ablation study of the number of stacked convolution layers and the convolution kernel size on the ActivityNet Captions dataset. The left chart shows the result of the R1@0.7 metric, and the right one represents the result of R5@0.7.
Electronics 14 01657 g007
Figure 8. The ablation study of the various value of t m i n and t m a x on the Charades-STA dataset at the R5@0.5 metric.
Figure 8. The ablation study of the various value of t m i n and t m a x on the Charades-STA dataset at the R5@0.5 metric.
Electronics 14 01657 g008
Figure 9. The ablation study of the hyperparameter λ on the Charades-STA dataset. From left to right, each subgraph represents the result of R1@0.5, R1@0.7, R5@0.5, and R5@0.7, respectively.
Figure 9. The ablation study of the hyperparameter λ on the Charades-STA dataset. From left to right, each subgraph represents the result of R1@0.5, R1@0.7, R5@0.5, and R5@0.7, respectively.
Electronics 14 01657 g009
Figure 10. The top and bottom subfigures refer to the visualization results of the Charades-STA and ActivityNet Captions, respectively.
Figure 10. The top and bottom subfigures refer to the visualization results of the Charades-STA and ActivityNet Captions, respectively.
Electronics 14 01657 g010
Table 1. Performance comparison with other methods on Charades-STA dataset.
Table 1. Performance comparison with other methods on Charades-STA dataset.
Method TypesMethodsFeaturer1@0.5r1@0.7r5@0.5r5@0.7
Reinforcement learningRWMC3D34.1213.74--
TripNet38.2916.07--
TSP-PRL37.3917.69--
Proposal-freeDEBUGC3D37.3917.69--
GDP39.4718.49--
PMI39.7319.27--
STCNet38.4420.73--
FVC41.5622.61--
Proposal-basedCTRLC3D23.638.8958.9229.52
ACRN20.267.6471.9927.79
ROLE12.12-40.59-
MAC30.4812.264.8435.13
QSPN35.615.879.445.4
CBP36.818.8770.9450.19
BPNet38.2520.51--
FVMR38.1618.2282.1844.96
TMBRN42.3323.2887.8252.2
Reinforcement learningSM-RLVGG24.3611.1761.2532.08
Proposal-freeABLRVGG24.369.01--
PLPNet41.8820.56--
Proposal-basedMCNVGG17.468.0148.2226.73
SAP27.4213.3666.3738.15
2D-TAN39.7023.3180.3251.26
MIM43.9225.8987.0752.26
TMBRN45.1427.0388.4953.31
Table 2. Performance comparison with other methods on ActivityNet Captions dataset.
Table 2. Performance comparison with other methods on ActivityNet Captions dataset.
Method TypesMethodsFeatureR1@0.3R1@0.5R1@0.7R5@0.3R5@0.5R5@0.7
Reinforcement learningRWMC3D-36.9----
TripNet48.4232.1913.93---
TSP-PRL56.0838.76----
Proposal-freeDEBUGC3D55.9139.72----
ABLR55.6736.79----
GDP56.1739.27----
LGI58.5241.5123.07---
PMI59.6938.2817.83---
CPNet-40.5621.63---
PLPNet56.9239.2020.91---
TCFN56.8140.5824.7367.8755.5139.42
STCNet-40.1521.85---
FVC63.5946.3226.01---
Proposal-basedCTRLC3D47.4329.0110.3475.3259.1737.54
MCN39.3521.366.4368.1253.2329.70
ACRN49.7031.6711.2576.5060.3438.57
TGN43.8127.93-54.5644.20-
SCDM54.8036.7519.8677.2964.9941.53
QSPN52.1333.2613.4377.7262.3940.78
2D-TAN59.4544.5126.5485.5377.1361.96
CBP54.3035.7617.8077.6365.8946.20
BPNet58.9842.0724.69---
TLL-44.2427.01-75.2260.23
TMBRN64.5147.0228.2685.8977.2262.32
Table 3. Results of ablation experiments of the weighted factor (WF) of the refinement loss on the Charades-STA dataset. The first row stands for the TMBRN without any weighted factors, and the second row indicates that the TMBRN regards the original IoU (OIoU) as the weighted factor.
Table 3. Results of ablation experiments of the weighted factor (WF) of the refinement loss on the Charades-STA dataset. The first row stands for the TMBRN without any weighted factors, and the second row indicates that the TMBRN regards the original IoU (OIoU) as the weighted factor.
MethodR1@0.5R1@0.7R5@0.5R5@0.7
TMBRN w/o WF32.1215.4638.6818.92
TMBRN w/ WF(OIoU)36.3419.1182.5044.33
TMBRN42.3323.2887.8252.2
Table 4. Results of ablation experiments of the weighted factor (WF) of the refinement loss on the ActivityNet Captions dataset. The first row stands for the TMBRN without any weighted factors, and the second row indicates that the TMBRN regards the original IoU (OIoU) as the weighted factor.
Table 4. Results of ablation experiments of the weighted factor (WF) of the refinement loss on the ActivityNet Captions dataset. The first row stands for the TMBRN without any weighted factors, and the second row indicates that the TMBRN regards the original IoU (OIoU) as the weighted factor.
MethodR1@0.3R1@0.5R1@0.7R5@0.3R5@0.5R5@0.7
TMBRN w/o WF52.6035.5417.6365.0749.6429.39
TMBRN w/ WF(OIoU)54.6840.1123.2182.9072.1753.36
TMBRN64.5147.0228.2685.8977.2262.32
Table 5. Results of ablation experiments of scaled IoU on the Charades-STA dataset.
Table 5. Results of ablation experiments of scaled IoU on the Charades-STA dataset.
MethodR1@0.5R1@0.7R5@0.5R5@0.7
TMBRN w/o scaled IoU34.7017.3975.2240.51
TMBRN42.3323.2887.8252.2
Table 6. Results of ablation experiments of scaled IoU on the ActivityNet Captions dataset.
Table 6. Results of ablation experiments of scaled IoU on the ActivityNet Captions dataset.
MethodR1@0.3R1@0.5R1@0.7R5@0.3R5@0.5R5@0.7
TMBRN w/o scaled IoU53.8338.5820.5178.1867.2449.95
TMBRN64.5147.0228.2685.8977.2262.32
Table 7. Results of ablation experiments of the feature fusion module on the ActivityNet Captions dataset. The first row indicates that the feature fusion module of the TMBRN is replaced with that of CTRL [42], and the second row represents that it is replaced with the cross-attention (CA) module.
Table 7. Results of ablation experiments of the feature fusion module on the ActivityNet Captions dataset. The first row indicates that the feature fusion module of the TMBRN is replaced with that of CTRL [42], and the second row represents that it is replaced with the cross-attention (CA) module.
MethodR1@0.3R1@0.5R1@0.7R5@0.3R5@0.5R5@0.7
TMBRN w/ CTRL56.3041.0724.9183.1972.4853.91
TMBRN w/ CA56.8942.4726.1383.4074.0257.69
TMBRN64.5147.0228.2685.8977.2262.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lyu, L.; Liu, D.; Zhang, C.; Zhang, Y.; Ruan, H.; Zhu, L. Temporal Map-Based Boundary Refinement Network for Video Moment Localization. Electronics 2025, 14, 1657. https://doi.org/10.3390/electronics14081657

AMA Style

Lyu L, Liu D, Zhang C, Zhang Y, Ruan H, Zhu L. Temporal Map-Based Boundary Refinement Network for Video Moment Localization. Electronics. 2025; 14(8):1657. https://doi.org/10.3390/electronics14081657

Chicago/Turabian Style

Lyu, Liang, Deyin Liu, Chengyuan Zhang, Yilin Zhang, Haoyu Ruan, and Lei Zhu. 2025. "Temporal Map-Based Boundary Refinement Network for Video Moment Localization" Electronics 14, no. 8: 1657. https://doi.org/10.3390/electronics14081657

APA Style

Lyu, L., Liu, D., Zhang, C., Zhang, Y., Ruan, H., & Zhu, L. (2025). Temporal Map-Based Boundary Refinement Network for Video Moment Localization. Electronics, 14(8), 1657. https://doi.org/10.3390/electronics14081657

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop