Next Article in Journal
Rethinking Liquid Crystal Tunable Phase Shifter Design with Inverted Microstrip Lines at 1–67 GHz by Dissipative Loss Analysis
Next Article in Special Issue
Time Synchronization and Space Registration of Roadside LiDAR and Camera
Previous Article in Journal
Incorporating Derivative-Free Convexity with Trigonometric Simplex Designs for Learning-Rate Estimation of Stochastic Gradient-Descent Method
Previous Article in Special Issue
Few-Shot Learning Based on Double Pooling Squeeze and Excitation Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

VMLH: Efficient Video Moment Location via Hashing

1
School of Computer Science and Technology, Shandong Jianzhu University, Jinan 250101, China
2
School of Journalism and Communication, Shandong Normal University, Jinan 250014, China
3
School of Software, Shandong University, Jinan 250101, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(2), 420; https://doi.org/10.3390/electronics12020420
Submission received: 1 November 2022 / Revised: 24 December 2022 / Accepted: 9 January 2023 / Published: 13 January 2023
(This article belongs to the Special Issue Deep Perception in Autonomous Driving)

Abstract

:
Video-moment location by query is a hot topic in video understanding. However, most of the existing methods ignore the importance of location efficiency in practical application scenarios; video and query sentences have to be fed into the network at the same time during the retrieval, which leads to low efficiency. To address this issue, in this study, we propose an efficient video moment location via hashing (VMLH). In the proposed method, query sentences and video clips are, respectively, converted into hash codes and hash code sets, in which the semantic similarity between query sentences and video clips is preserved. The location prediction network is designed to predict the corresponding timestamp according to the similarity among hash codes, and the videos do not need to be fed into the network during the process of retrieval and location. Furthermore, different from the existing methods, which require complex interactions and fusion between video and query sentences, the proposed VMLH method only needs a simple XOR operation among codes to locate the video moment with high efficiency. This paper lays the foundation for fast video clip positioning and makes it possible to apply large-scale video clip positioning in practice. The experimental results on two public datasets demonstrate the effectiveness of the method.

1. Introduction

Given a video and a query sentence, the moment location task needs to find the start and end timestamps of the video clip that best match the query sentence. In the video-moment location task, “moment” indicates the start and end times of the query text corresponding to the video; that is, we need to obtain two time stamps of the video. Figure 1 illustrates the objective of this task through a simple example. Generally, video-moment location focuses on improving accuracy by using the fine-grained matching relationship among modes or reducing the computational cost by avoiding multiple candidate windows.
Recently, there has been much research into video-moment location. Hendricks et al. [1] proposed a moment context network, which roughly located segments by calculating the similarity between query sentences and different parts of different scales of the video. However, this method was not accurate enough and required a lot of calculation. In contrast, Gao et al. [2] proposed a model based on a sliding window, which could finetune the boundary in the window to achieve more fine-grained positioning. In order to further improve the accuracy, the cross-modal attention mechanism [3] has been widely used. For example, Xu et al. [4] had these two patterns interact at an early stage to produce semantically richer suggestions. Zhang et al. [5] proposed a network of simultaneous proposal and reasoning. In order to further reduce the computational workload related to candidate or sliding windows, Wang et al. [6] used deep reinforcement learning to automatically obtain the best window. In addition, Ghosh et al. [7] designed a method for sensing boundaries. In addition, some works in other fields have played a positive role in the video-moment location task. For example, the video object segmentation can accurately extract object information from a video [8,9]. Lu et al. [10] proposed a CO-attention Siamese Network (COSNet), to address the unsupervised video object segmentation task from a holistic view. Object tracking tasks can assist in obtaining key object information in video [11,12]. In addition, the Modified Lazy Video Transmission Algorithm (MLVTA) proposed by Sodhro et al. [13] has laid a foundation for online video-moment location.
Generally speaking, most of the existing video-moment location methods only focus on accuracy. However, with the explosive growth of multimedia data, especially video data, it becomes important to locate and retrieve large databases quickly.
In essence, video-moment location using natural language query belongs to the field of cross-modal retrieval, in which we use the query in text modality to retrieve one or more video clips. For the task of cross-modal retrieval, a common space is always needed to evaluate similarity. As is known, hashing maps data points to low-dimensional binary codes in Hamming space, which can be considered a common space for different modalities. Furthermore, the binary representation of hash codes is a widely studied solution for the fast retrieval of large-scale data. Due to the low storage requirements of binary hash codes, hashing can improve the retrieval speed and reduce memory storage, and the similarity among hash codes can be effectively calculated through fast XOR operation in Hamming space. Therefore, in this study, we use hashing to perform the moment location task.
The advantages of video-moment location using hashing are twofold. Firstly, using hash codes to store video semantic information can greatly reduce the space occupation, which has been effective for the growth of massive data in recent years. Secondly, the semantic matching between video and query sentences can be completed by bit operation of hash code, which can greatly improve the query speed.
In this study, we propose a video-moment location method with hashing (VMLH), which consists of three parts: a video-hashing network, a sentence-hashing network, and a location-prediction network. VMLH can carry out end-to-end training. In the positioning process, we only need a language query, which is flexible and can avoid the consumption of space and time.
The contributions of this study are summarized as follows:
  • An efficient video-moment location method with hashing is proposed, which makes full use of hash retrieval and greatly improves the efficiency of the task.
  • There is no complex interaction and fusion process in the proposed method, and videos do not need to be fed into the network during the location, which leads to higher efficiency and better scalability compared with the existing methods.

2. Related Work

Given that moment localization, activity localization and video retrieval with text queries in are two related tasks, we provide a brief description of them.

2.1. Activity Localization

Activity localization is the process of locating the start and end times of certain actions in a video. The purpose of the activity localization task is to enable the machine to recognize the actions occurring and predict when they transpire in the video automatically. There are only a few common types of actions that can be localized, such as running, jumping, throwing, etc. In contrast to video-moment location, activity location cannot be queried using natural language, and the categories of actions are limited, but its mature models are often used as backbone networks for other video tasks. Earlier work [14] localized by performing frame-level or window-level classification, followed by manual aggregation. Later, a two-stage approach [15] of proposal generation and boundary fine-tuning was used. Some models [16] now combine proposal generation and boundary fine-tuning for end-to-end training.

2.2. Video Retrieval with Sentence Queries

Video retrieval with text queries can obtain the entire video from a collection of videos associated with a text description. In contrast to the moment localization in this study, it does not need to predict the start and end timestamps of the moment, and its main difficulty is in learning to distinguish between different videos, rather than different parts of the same video. Currently, the dominant approach to this cross-modal retrieval is to encode different modal features into a joint embedding space to measure semantic similarity. Mithun et al. [17] encoded the video and text into global vectors. Although this global representation was efficient, it may lead to the loss of some critical details. To avoid these problems, Yu et al. [18] computed the matching relationship between words and frames and further computed the matching relationship between the entire video and the query. However, the natural language usually containing the logical structure is complex, and sometimes partial matches are not representative of the overall match relationship. For example, there is a piece of natural language with logical structure, “A man jumps from the ground for the second time”, which may locate two jump clips in the video. To this end, a number of studies, such as the method in [19], have been working on this.

3. Proposed Method

This section will first provide a description of the moment location task and then describe the proposed VMLH model and its training and reasoning process. Figure 2 shows the framework of the VMLH, which includes three components: a video-hashing network, a sentence-hashing network, and a location-prediction network. Each part will be detailed in the next section. The main framework of the model is to extract video features through the I3D network and extract text features through GloVe; then, it uses GRU to extract the advanced features of these two modes and map them to a hash matrix; next, the similarity between the two hash matrices is calculated and fed into the MLP to locate the final timestamp. The idea of the VMLH is to train a model so that it can generate a set of hash codes H v = { h t v } t = 1 T for video v ( h t v is the hash code of video clip c t , and T is the total number of video clips) and generate hash code h s for a query sentence s. Generally, the task of hashing is to convert the original video, image, or text into binary codes. The values −1 and 1 are obtained with the proposed method, and we also can convert them into 0 and 1 for XOR operations. Then, using the location prediction network, the start and end timestamps are predicted by calculating the similarity among the hash codes.

3.1. Video Hashing Network

Given a raw video, a video encoder is used to transform it into a sequence of features V = { c t } t = 1 T . These features are then passed through a bidirectional gated recurrent unit (GRU) [20] network to mine the timing information, and a fully connected layer (FC) with activation functions is used to generate a real vector at each moment. Finally, the hash code h t v is obtained as follows:
h t g 1 = G R U ( c t , h t 1 g 1 ) ,
r t v = T a n h ( W α h t g 1 + b α ) ,
h t v = s i g n ( r t v ) ,
where h t g 1 is the concatenation from the bidirectional output of the GRU at time step t, and r t v refers to an undiversified vector of real numbers. The symbols W α and b α are the learnable matrix and bias, respectively.

3.2. Sentence-Hashing Network

The structure of the sentence hashing network is almost identical to that of the video hashing network, except that only the output of the last time step of the GRU is used as the overall semantics of the sentence. A query sentence with N words can be represented as S = { w n } n = 1 N , after extracting the features using GloVe [21], and the hash code h s of the query sentence is obtained as follows:
h n g 2 = G R U ( w n , h t 1 g 2 ) ,
r s = T a n h ( W β h N g 2 + b β ) ,
h s = s i g n ( r s ) ,
where h n g 2 is the concatenation from the bidirectional output of the GRU at time step n, and r s represents the real vectors of the output of the last time step N. The symbols W β and b β are the learnable matrix and bias, respectively.

3.3. Location Prediction Network

The role of the location prediction network is to calculate the corresponding start and end times of the moment based on the distribution of the similarity scores. Specifically, the similarity score s t h at each time step is calculated by the following formula:
s t h = S i g m o i d ( μ h t v · h s ) ,
where S i g m o i d ( · ) is a sigmoid function. The symbol μ is the deflation factor to prevent the similarity values from straying too far from the origin and causing the gradient to disappear. Subsequently, the similarity scores s t h for each time step are collapsed into a vector s h and fed into a multilayer perceptron. Then, using the multilayer perceptron, we obtain a start and end timestamp as follows:
h f = T a n h ( W γ s h + b γ ) ,
l = W ζ h f + b ζ ,
where W γ , W ζ , b γ , and b ζ are learnable parameters. The vector l represents the predicted normative moment, and it consists of two items l s and l e , which represent the predicted start and end times, respectively.

3.4. Training and Inference

Given a video and a query sentence whose corresponding start and end times are t s and t e , respectively, the ground truth score s t g at time t for this duration corresponds to 1, and the remainder corresponds to 0. s t r is the similarity score of time t obtained by our method. We move the semantically similar fragments of sentence and video closer together in Hamming space. The similarity loss L s is:
L s = 1 T t = 1 T s t g l o g ( s t r ) + ( 1 s t g ) l o g ( 1 s t r ) .
A smoothed L1 loss function R ( · ) [22] is used to calculate the location loss L l :
L l = R ( l s t s n ) + R ( l e t e n ) ,
where t s n and t e n are the start and end times after normalization, respectively. Finally, the entire loss function is described as follows:
L = L s + λ L l ,
where λ is a parameter.
The inference process can be completed end-to-end, or it can first generate the hash code for storage through the corresponding network and then use the location-prediction network for matching when necessary.
After obtaining l s and l e , the predicted start and end timestamps t s * and t e * can be obtained by calculating the following:
t s * = l s × d u r a t i o n ,
t e * = l e × d u r a t i o n ,
where d u r a t i o n indicates the time duration of the entire video.
To better represent the complete training process, the algorithm for training the VMLH is shown in Algorithm 1. The overall flow chart of this method is shown in Figure 3.
Algorithm 1: Learning algorithm for VMLH
Input: Training video V = { c t } t = 1 T and the query sentence S = { w n } n = 1 N . The ground truth timestamps t s n and t e n .
Output: Predicted video start and end timestamps t s * and t e * .
Initialization: Initialize the model with the pretrained I3D and GloVe parameter file.
Repeat: Randomly sample a minibatch of video from V , obtain the corresponding sentence S according to the video ID, and perform the following operations:
1 
Obtain the video feature matrix c t and the corresponding text feature matrix w n .
2 
For each clip feature c t , obtain the video hash codes h t v . Obtain the hash codes h s of sentence feature w n .
3 
XOR the hash codes of each video clip h t v with the sentence hash codes h s to obtain the similarity score s t h . Feed the similarity score into the multilayer perceptron to obtain the start and end time stamps.
4 
Calculate L s and L l , according to the similarity score s t h and t s n and t e n in the INPUT.
5 
Update the parameters by utilizing backpropagation.
Until: a fixed number of iterations

4. Experiments

4.1. Datasets

We evaluated the proposed VMLH on two widely used benchmark video datasets, Charades-STA [2] and ActivityNet Captions [23].

4.1.1. Charades-STA

The Charades-STA dataset [2] is annotated by the semi-automatic method. The average length of sentences is 8.6 words, and there is no complex logical structure. As a result, the sentences are simpler in form and usually not very long. This dataset includes 9848 tagged videos, each lasting approximately 30 s, showing the behavior of 267 different people on three continents.

4.1.2. ActivityNet Captions

In this dataset, each sentence corresponds to a moment of the video, which can be anywhere from a few seconds to over a hundred seconds in length. The sentences themselves are also highly complex, can be very long, and can contain multiple consecutive actions. On average, each of the 20 k videos in an ActivityNet caption contains 3.65 sentences.

4.2. Experimental Settings

Except as specifically mentioned, the hyperparameter settings were identical for both datasets. All the experiments of the model efficiency were run in an Nvidia TITAN Xp GPU on Ubuntu 16.04 with 256 GB memory. The video and sentence hash codes were both 64-bit. The 500-dimensional C3D [24] features and 1024-dimensional I3D [25] features were used in the datasets ActivityNet and Charades-STA, respectively. The output dimension of the sentence encoder GloVe was 300. The hidden layer sizes of the LSTMs and FCs were 256 and 128, respectively, regardless of whether they were in a video or a sentence hash network. In addition, λ and μ were set to 0.01 and 1/6, respectively. Moreover, we sampled the videos on the Charades-STA and ActivityNet evenly to 64 clips and 128 clips, respectively. All experiments were conducted using the Adam optimizer with a learning rate of 0.001 and a batch size of 64 for 50 epochs in PyTorch. In building the project code, we used the framework of Pytorch and Torch Lightning, and we used Wisdom to visualize the training process.

4.3. Evaluation Metrics

The IoU metric denotes the intersection of the predicted and ground truth moment over their union. For a fair comparison, we adopted “R@n, IoU = m” as the evaluation metric for our study. Specifically, “R@n, IoU = m” is defined as the percentage of queries having at least one result satisfying IoU ≤ m in the top n results. Specifically, R@1 means that only one video segment was predicted. Note that our VMLH provided only one pair of timestamps, so that n = 1 in the experiment we reported.

4.4. Accuracy Performance

Table 1 and Table 2 show the comparison between the VMLH and other state-of-the-art methods. The experimental results using real features were additionally provided for both datasets. The experimental results demonstrated that our proposed model had higher accuracy compared to other state-of-the-art methods even in the absence of earlier feature interaction. Moreover, using discrete hash codes in training, the location prediction network avoided inconsistencies between the training and test data types without significant loss of accuracy.

4.5. Model Efficiency

Table 3 shows the efficiency comparison between our model and the other models in a single run time. The single run time is the average time taken to locate one moment in a video. VMLH-full means that neither the sentence nor the video was pre-stored as hash codes. VMLH-vh represents the video was pre-stored as hash codes, while the sentences had to move through the sentence hashing network. VMLH-h means both used hash for retrieval. Given that using hash for retrieval only requires going through the location prediction network, the model’s computation was reduced by a factor of 10. When large batches were considered, the average time per retrieval required was reduced to the level of microseconds.

4.6. Ablation Experiment

We performed ablation experiments on two datasets. Figure 4 shows the results of the location efficiency using hash codes of different lengths and without hash generators. We also used a single run time to evaluate the positioning efficiency. According to the experimental results, the hash generator greatly improved the localization efficiency. We conducted experiments on the influence of the hash code length on the precision, and the experimental results are shown in Table 4. The length of the hash code affected the representation ability of the features. We conducted precision experiments on the settings of 32-bit, 64-bit, and 128-bit hash codes. We chose the 64-bit hash code. Although the accuracy was slightly lower, the location efficiency was greatly improved.

4.7. Convergence Analysis

Experiments were performed on the two datasets to evaluate the convergence performance of the proposed VMLH. In the experiments described in this section, we used the relative losses to evaluate the convergence of the VMLH. Figure 5 shows that as the iterations increased, the relative loss became fairly small and stable. The convergence experiments showed that the VMLH reached convergence quickly during training, which greatly reduced the training time.

4.8. Qualitative Results

Figure 6 shows three visual predictions on the Charades-STA dataset. The area represented by the blue line is the ground truth; the green line is the predicted segment. We expect to estimate the start and end timestamps closer to the ground truth. The three examples in Figure 6 show that the prediction result has a high IoU on the Charades-STA dataset. However, in the first example, the start and end timestamps of the prediction have errors compared to the ground truth. Therefore, the first example results are unsatisfactory and may lead to considerable errors in the prediction results on long videos.

5. Conclusions

In this study, we proposed hashing to solve the moment location problem. The proposed VMLH model is efficient, reducing the storage space with considerable accuracy. Our proposed method still contains limitations: 1. storing hash codes in advance video and sentence hashing networks may cause additional memory consumption. 2. The XOR operation using hash codes may fail to fuse the modes, which will cause the loss of semantic information between the two models. Therefore, further improving the accuracy without early feature interaction requires more research. In addition, considering the short video duration used for training in these two datasets, the performance with long videos requires investigation in future studies. Through the ablation experiment, we concluded that the hash code as the video feature affects the accuracy. In future works, we will conduct in-depth research on this; we will enable the hash codes to learn better feature representation. This aspect will immensely improve the accuracy of our method.

Author Contributions

Conceptualization, Z.T.; Data curation, X.L.; Formal analysis, X.N.; Funding acquisition, X.N.; Investigation, F.D.; Methodology, X.L.; Project administration, X.N.; Resources, C.L.; Software, F.D. and C.L.; Supervision, X.N.; Validation, X.N.; Writing—original draft, Z.T.; Writing—review and editing, X.L. and X.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (62176141, 62102235), the Shandong Provincial Natural Science Foundation for Distinguished Young Scholars (ZR2021JQ26), the Shandong Provincial Natural Science Foundation (ZR2020QF029), and the Taishan Scholar Project of Shandong Province (tsqn202103088).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hendricks, L.A.; Wang, O.; Shechtman, E.; Sivic, J.; Darrell, T.; Russell, B.C. Localizing Moments in Video with Natural Language. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5804–5813. [Google Scholar] [CrossRef] [Green Version]
  2. Gao, J.; Sun, C.; Yang, Z.; Nevatia, R. TALL: Temporal Activity Localization via Language Query. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5277–5285. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, J.; Chen, X.; Ma, L.; Jie, Z.; Chua, T. Temporally Grounding Natural Sentence in Video. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 162–171. [Google Scholar] [CrossRef] [Green Version]
  4. Xu, H.; He, K.; Plummer, B.A.; Sigal, L.; Sclaroff, S.; Saenko, K. Multilevel Language and Vision Integration for Text-to-Clip Retrieval. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 9062–9069. [Google Scholar] [CrossRef] [Green Version]
  5. Zhang, D.; Dai, X.; Wang, X.; Wang, Y.; Davis, L.S. MAN: Moment Alignment Network for Natural Language Moment Retrieval via Iterative Graph Adjustment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1247–1257. [Google Scholar] [CrossRef] [Green Version]
  6. Wang, W.; Huang, Y.; Wang, L. Language-Driven Temporal Activity Localization: A Semantic Matching Reinforcement Learning Model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Long Beach, CA, USA, 16–20 June 2019; pp. 334–343. [Google Scholar] [CrossRef]
  7. Ghosh, S.; Agarwal, A.; Parekh, Z.; Hauptmann, A.G. ExCL: Extractive Clip Localization Using Natural Language Descriptions. arXiv 2019, arXiv:1904.02755. [Google Scholar] [CrossRef]
  8. Lu, X.; Wang, W.; Shen, J.; Crandall, D.; Luo, J. Zero-shot video object segmentation with co-attention siamese networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2228–2242. [Google Scholar] [CrossRef] [PubMed]
  9. Lu, X.; Wang, W.; Shen, J.; Crandall, D.J.; Van Gool, L. Segmenting objects from relational visual data. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7885–7897. [Google Scholar] [CrossRef] [PubMed]
  10. Lu, X.; Wang, W.; Ma, C.; Shen, J.; Shao, L.; Porikli, F. See more, know more: Unsupervised video object segmentation with co-attention siamese networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3623–3632. [Google Scholar]
  11. Shen, J.; Liu, Y.; Dong, X.; Lu, X.; Khan, F.S.; Hoi, S. Distilled Siamese networks for visual tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 8896–8909. [Google Scholar] [CrossRef] [PubMed]
  12. Lu, X.; Ma, C.; Ni, B.; Yang, X. Adaptive region proposal with channel regularization for robust object tracking. IEEE Trans. Circuits Syst. Video Technol. 2019, 31, 1268–1282. [Google Scholar] [CrossRef]
  13. Sodhro, A.H.; Sangaiah, A.K.; Sodhro, G.H.; Lodro, M.M.; Sekhari, A.; Ouzrout, Y.; Pirbhulal, S.; Fatima, K. Medical quality of service optimization over internet of multimedia things. In Computational Intelligence for Multimedia Big Data on the Cloud with Engineering Applications; Elsevier: Amsterdam, The Netherlands, 2018; pp. 271–295. [Google Scholar]
  14. Shou, Z.; Chan, J.; Zareian, A.; Miyazawa, K.; Chang, S. CDC: Convolutional-De-Convolutional Networks for Precise Temporal Action Localization in Untrimmed Videos. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1417–1426. [Google Scholar] [CrossRef] [Green Version]
  15. Shou, Z.; Wang, D.; Chang, S. Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1049–1058. [Google Scholar] [CrossRef] [Green Version]
  16. Lin, T.; Zhao, X.; Shou, Z. Single Shot Temporal Action Detection. In Proceedings of the 2017 ACM on Multimedia Conference, Mountain View, CA, USA, 23–27 October 2017; pp. 988–996. [Google Scholar] [CrossRef] [Green Version]
  17. Mithun, N.C.; Li, J.; Metze, F.; Roy-Chowdhury, A.K. Learning Joint Embedding with Multimodal Cues for Cross-Modal Video-Text Retrieval. In Proceedings of the 2018 ACM on International Conference on Multimedia Retrieval, Yokohama, Japan, 11–14 June 2018; pp. 19–27. [Google Scholar] [CrossRef] [Green Version]
  18. Yu, Y.; Kim, J.; Kim, G. A Joint Sequence Fusion Model for Video Question Answering and Retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Volume 11211, pp. 487–503. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, S.; Zhao, Y.; Jin, Q.; Wu, Q. Fine-Grained Video-Text Retrieval with Hierarchical Graph Reasoning. In Proceedings of the 2020 IEEE/CVF Conference on CVPR, Seattle, WA, USA, 13–19 June 2020; pp. 10635–10644. [Google Scholar] [CrossRef]
  20. Chung, J.; Gülçehre, Ç.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  21. Pennington, J.; Socher, R.; Manning, C. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; Association for Computational Linguistics: Doha, Qatar, 2014; pp. 1532–1543. [Google Scholar] [CrossRef]
  22. Girshick, R.B. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  23. Krishna, R.; Hata, K.; Ren, F.; Fei-Fei, L.; Niebles, J.C. Dense-Captioning Events in Videos. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 706–715. [Google Scholar] [CrossRef]
  24. Tran, D.; Bourdev, L.D.; Fergus, R.; Torresani, L.; Paluri, M. Learning Spatiotemporal Features with 3D Convolutional Networks. In Proceedings of the 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar] [CrossRef] [Green Version]
  25. Carreira, J.; Zisserman, A. Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 4724–4733. [Google Scholar] [CrossRef] [Green Version]
  26. Liu, L.; Shao, L. Sequential Compact Code Learning for Unsupervised Image Hashing. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 2526–2536. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Hahn, M.; Kadav, A.; Rehg, J.M.; Graf, H.P. Tripping through time: Efficient Localization of Activities in Videos. In Proceedings of the 31st British Machine Vision Conference 2020, Virtual, 7–10 September 2020. [Google Scholar]
  28. Liu, M.; Wang, X.; Nie, L.; He, X.; Chen, B.; Chua, T. Attentive Moment Retrieval in Videos. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 15–24. [Google Scholar] [CrossRef]
  29. Ge, R.; Gao, J.; Chen, K.; Nevatia, R. MAC: Mining Activity Concepts for Language-Based Temporal Localization. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 245–253. [Google Scholar] [CrossRef] [Green Version]
  30. Yuan, Y.; Mei, T.; Zhu, W. To Find Where You Talk: Temporal Sentence Localization in Video with Attention Based Location Regression. In Proceedings of the The Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 9159–9166. [Google Scholar] [CrossRef] [Green Version]
  31. Hu, Y.; Liu, M.; Su, X.; Gao, Z.; Nie, L. Video Moment Localization via Deep Cross-Modal Hashing. IEEE Trans. Image Process. 2021, 30, 4667–4677. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The goal of moment location is to identify a moment from a video that is most relevant to the description of the query sentence. For example, if one wants to find a clip of a man peeling potatoes, then the query can be “a man is peeling potatoes”. The start and end timestamps are 13.7 s and 21.0 s, respectively, and the moment location task seeks to find the most relevant timestamps matching the query sentence.
Figure 1. The goal of moment location is to identify a moment from a video that is most relevant to the description of the query sentence. For example, if one wants to find a clip of a man peeling potatoes, then the query can be “a man is peeling potatoes”. The start and end timestamps are 13.7 s and 21.0 s, respectively, and the moment location task seeks to find the most relevant timestamps matching the query sentence.
Electronics 12 00420 g001
Figure 2. The framework of VMLH. VMLH consists of three parts: a video hashing network, a sentence hashing network, and a location prediction network.
Figure 2. The framework of VMLH. VMLH consists of three parts: a video hashing network, a sentence hashing network, and a location prediction network.
Electronics 12 00420 g002
Figure 3. VMLH Training Flow Chart.
Figure 3. VMLH Training Flow Chart.
Electronics 12 00420 g003
Figure 4. Effects of different lengths of hash codes and of not using hash generators on the location efficiency.
Figure 4. Effects of different lengths of hash codes and of not using hash generators on the location efficiency.
Electronics 12 00420 g004
Figure 5. Convergence curve. Relative loss results of the VMLH on the two datasets after 1000 iterations.
Figure 5. Convergence curve. Relative loss results of the VMLH on the two datasets after 1000 iterations.
Electronics 12 00420 g005
Figure 6. We randomly selected three location results for visual output on the Charades-STA dataset. In addition, there is a high IoU result on this dataset.
Figure 6. We randomly selected three location results for visual output on the Charades-STA dataset. In addition, there is a high IoU result on this dataset.
Electronics 12 00420 g006
Table 1. R@1 performance comparison for the ActivityNet Captions dataset (%).
Table 1. R@1 performance comparison for the ActivityNet Captions dataset (%).
ComparisonCross–Merge Ratio
MethodIoU = 0.3IoU = 0.5IoU = 0.7
MCN [1]39.3521.366.43
CTRL [26]47.4329.0110.34
TGN [3]45.5128.47-
TripNet [27]48.4232.1913.93
ACRN [28]49.7031.6711.25
VMLH52.1534.5017.16
Table 2. R@1 performance comparison for the Charades-STA dataset (%).
Table 2. R@1 performance comparison for the Charades-STA dataset (%).
ComparisonCross–Merge Ratio
MethodIoU = 0.5IoU = 0.7
CTRL [2]23.638.89
MLVI [4]35.6015.80
ACL-K [29]30.4812.20
ACRN [28]20.267.64
SM-RL [6]24.3611.17
QSPN [4]35.6015.80
TripNet [27]36.6114.50
VMLH43.8020.32
Table 3. Comparison of the models’ efficiency for a single run time (s).
Table 3. Comparison of the models’ efficiency for a single run time (s).
MethodSingle Run Time (s)
CTRL [2]3.41
ACRN [28]4.42
ABLR [30]0.06
CMHN [31]0.0076
VMLH-full0.0093
VMLH-vh0.0036
VMLH-h0.0007
Table 4. Hash code length ablation experiment.
Table 4. Hash code length ablation experiment.
Hash CodeCharades-STAActivityNet Captions
LengthIoU = 0.5IoU= 0.7IoU = 0.3IoU = 0.5IoU = 0.7
32 bit42.2920.1651.0033.7416.63
64 bit43.8020.3252.1534.5017.16
128 bit43.9920.7652.8734.3617.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, Z.; Dong, F.; Liu, X.; Li, C.; Nie, X. VMLH: Efficient Video Moment Location via Hashing. Electronics 2023, 12, 420. https://doi.org/10.3390/electronics12020420

AMA Style

Tan Z, Dong F, Liu X, Li C, Nie X. VMLH: Efficient Video Moment Location via Hashing. Electronics. 2023; 12(2):420. https://doi.org/10.3390/electronics12020420

Chicago/Turabian Style

Tan, Zhifang, Fei Dong, Xinfang Liu, Chenglong Li, and Xiushan Nie. 2023. "VMLH: Efficient Video Moment Location via Hashing" Electronics 12, no. 2: 420. https://doi.org/10.3390/electronics12020420

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop