Next Article in Journal
Mathematical Modeling and the Use of Network Models as Epidemiological Tools
Next Article in Special Issue
Deep Spatial-Temporal Neural Network for Dense Non-Rigid Structure from Motion
Previous Article in Journal
Positive-Unlabeled Learning for Network Link Prediction
Previous Article in Special Issue
Boosting Unsupervised Dorsal Hand Vein Segmentation with U-Net Variants
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Level Cross-Modal Semantic Alignment Network for Video–Text Retrieval

1
School of Advanced Manufacturing Engineering, Hefei University, Hefei 230601, China
2
Anhui International Joint Research Center for Ancient Architecture Intellisencing and Multi-Dimensional Modeling, Anhui Jianzhu University, Hefei 230601, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(18), 3346; https://doi.org/10.3390/math10183346
Submission received: 22 July 2022 / Revised: 1 September 2022 / Accepted: 10 September 2022 / Published: 15 September 2022
(This article belongs to the Special Issue Computer Vision and Pattern Recognition with Applications)

Abstract

:
This paper strives to improve the performance of video–text retrieval. To date, many algorithms have been proposed to facilitate the similarity measure of video–text retrieval from the single global semantic to multi-level semantics. However, these methods may suffer from the following limitations: (1) largely ignore the relationship semantic which results in semantic levels are insufficient; (2) it is incomplete to constrain the real-valued features of different modalities to be in the same space only through the feature distance measurement; (3) fail to handle the problem that the distributions of attribute labels in different semantic levels are heavily imbalanced. To overcome the above limitations, this paper proposes a novel multi-level cross-modal semantic alignment network (MCSAN) for video–text retrieval by jointly modeling video–text similarity on global, entity, action and relationship semantic levels in a unified deep model. Specifically, both video and text are first decomposed into global, entity, action and relationship semantic levels by carefully designing spatial–temporal semantic learning structures. Then, we utilize KLDivLoss and a cross-modal parameter-share attribute projection layer as statistical constraints to ensure that representations from different modalities in different semantic levels are projected into a common semantic space. In addition, a novel focal binary cross-entropy (FBCE) loss function is presented, which is the first effort to model the unbalanced attribute distribution problem for video–text retrieval. MCSAN is practically effective to take the advantage of the complementary information among four semantic levels. Extensive experiments on two challenging video–text retrieval datasets, namely, MSR-VTT and VATEX, show the viability of our method.

1. Introduction

With the popularity of online short video sharing platforms (e.g., TikTok), video–text retrieval (using a video clip as query to find similar sentences and vice versa) becomes a fundamental and crucial information retrieval problem for both users and platforms. As a specific cross-modal retrieval [1] task, many research papers about video–text retrieval have been presented and achieved remarkable progress from 2018 to the present [2,3,4,5,6].
Measuring the semantic similarity between any pair of video–text is very important for video–text retrieval. Basically, existing studies on video–text retrieval can be summarized into two categories according to the roughness of video and text semantic representations: (1) The first one is single-level encoding methods [2,6,7,8], which leverage convolutional neural networks (CNN), recurrent neural networks (RNN) or Transformer to learn global representation (i.e., single level) of video/text. However, compacting video/text into a single global level representation is insufficient due to neglecting the local and fine-grained static/dynamic information that widely existed in video/text. (2) The second one is multi-level encoding methods [3,4,9,10,11], which convert video/text into hybrid attribute spaces (e.g., global, entity and action). The total similarity of the video–text pair is the fusion of the cross-modal similarities in different semantic levels.
Recently, large-scale visual-and-language pretraining (VLP) models (e.g., DALL-E [12], CLIP [13] and MVP [14]), which can learn representations grounded in both visual textual contexts via a self-supervised learning manner, have been successfully applied in various downstream multi-modal tasks such as visual question answering [15], video captioning [16] and visual reasoning [17]. Some video–text retrieval methods utilize VLP models to obtain more representative video and text features [18,19,20,21]. Although these approaches show promising performance on video–text retrieval tasks, training VLP means leveraging lots of extra multi-modal training data and a video–text retrieval framework with the VLP model also means adding a number of parameters.
Despite the significance and value of the methods in the above algorithms, they still suffer from three critical shortcomings: (1) Semantic levels are insufficient. Though action and entity semantic levels are considered in [3,9,11] and get better performance, the semantic similarity in relationship level is ignored. However, several works [22,23,24] demonstrate that modeling relationships can improve the performance of many multi-modal understanding tasks significantly. Similarly, as shown in Figure 1, relationships are also common clues in video–text retrieval scenarios. (2) Incomplete joint space representation. To compare the cross-modal representations in different semantic levels, most existing works [25,26] utilize ranking or classification losses. Although it has achieved certain results, there is no statistical guarantee that real-valued vectors encoded from different modalities are in a common space. (3) Unbalanced data distribution problem is neglected. Figure 2 shows the number of five entities in MSR-VTT [27] dataset, from it we can see that attributes distribution has a serious long tail effect. Unfortunately, to the best of our knowledge, no video–text retrieval frameworks are handling this problem explicitly.
To tackle these limitations, in this paper, we present a novel Multi-Level Cross-Modal Semantic Alignment Network (MCSAN) for video–text retrieval by jointly modeling the video/text context information and unbalanced attributes distribution challenges in a unified deep model. To achieve a robust video–text retrieval, as shown in Figure 1, besides the global representation, we first construct multi-level semantic dictionaries (entity, action and relationship) from the corpus. Then, we design several practical semantic encoding modules including CNN, RNN, GNN [28] and Multi-head self-attention [29], which are utilized to encode both video and text into multi-level semantic spaces (i.e., global, entity, action and relationship). Meanwhile, not only using feature distance, we also utilize KLDivLoss and a cross-modal parameter-share attribute projection layer as statistical constraints to ensure that representations from different modalities in different semantic levels are projected into a common semantic space. In addition, we design a novel multi-label regression loss to handle the unbalanced attributes distribution problem. Extensive experimental results on two widely-used video–text retrieval benchmark datasets, i.e., MSR-VTT [27] and VATEX [30], to validate the effectiveness and superiority of our MCSAN.
In summary, this paper makes the following contributions.
  • A novel Multi-Level Cross-Modal Semantic Alignment Network (MCSAN) is developed for video–text retrieval tasks. The semantic similarity between video and text is fully considered at four semantic levels (global, action, entity and relationship) by carefully designing spatial–temporal semantic learning structures and cross-modal representation distribution constraints in an end-to-end framework, which offers an alternative orientation for video–text retrieval.
  • Compared with existing works, we are the first to measure the similarity of video–text pairs using relationship semantic level. We demonstrate that relationship, which is also a complementary and crucial clue for video–text retrieval.
  • We introduce a novel focal binary cross-entropy (FBCE) loss function for multi-label attributes regression. Though several papers adopt attributes to improve the performance of video–text retrieval, to the best of our knowledge, this is the first effort to consider and handle the unbalanced attributes distribution problem in video–text retrieval scenarios.
  • By carefully incorporating spatial–temporal semantic learning structures, cross-modal representation distribution constraints, novel multi-label classification loss function, adaptive graph learning, and cross-modal contextual attention module, our proposed method can jointly enjoy the merits of alleviate semantic gap and long-tailed multi-level multi-modal attributes learning for video–text retrieval. Moreover, we conduct extensive experiments on two popular benchmarks, i.e., MSR-VTT [27] and VATEX [30], demonstrating the rationality and effectiveness of the proposed multi-level cross-modal semantic alignment network.
The remainder of this paper is organized as follows: Section 2 reviews the related work. We present the detailed architecture of the multi-level cross-modal semantic alignment network in Section 3, while Section 4 describes the datasets, settings, implementation details, experimental results, and limitations. Finally, we conclude the paper with future work in Section 5.

2. Related Work

The general process of the current video–text retrieval framework can be roughly divided into two blocks, namely, modality encoding (embedding video and text into vectors), and similarity calculation (predicting the similarity score of two modalities). Therefore, in this section, we briefly review typical video–text retrieval algorithms in the above two views.

2.1. Modality Encoding

Existing video and text encoding methods can be roughly divided into two categories. The first one is single-level encoding methods [2,6,7,8,31,32], which leverage convolutional neural networks (CNN), recurrent neural networks (RNN) or Transformer to learn global representation (i.e., single level) of video/text. However, compacting video/text into a single global level representation is insufficient due to neglecting the local and fine-grained static/dynamic information that widely existed in video/text. The second one is multi-level encoding methods [3,4,9,10,11], which convert video/text into hybrid attribute spaces (e.g., global, entity and action). The total similarity of the video–text pair is the fusion of the cross-modal similarities in different semantic levels. The popular pipeline for video/text encoding is that: (1) For video encoding, using the pre-trained image classification or video clip classification model to extract the frame-wise feature. For text encoding, using Word2Vec, Glove, or Bert to extract the word-wise feature. (2) Utilizing CNN, RNN, or Transformer to handle temporal information of both video and text.
To enhance the representation ability, Yang et al. [33] constructed a tree structure to model text. Dong et al. [34] proposed a reading-strategy-inspired visual representation learning framework to represent videos. Some works [35,36,37,38] made full use of multi-modal cues, e.g., appearance, audio, motion, OCR and face, for video encoding. Although these methods have made some progress, the training pipeline of these methods are very complex, and even a large amount of additional annotation information is required.
Recently, thanks to the rapid development of VLP models [12,13,14], there are some works [18,19,20,21] tend to utilize the VLP model encoding the video and text directly. Although these approaches show promising performance on video–text retrieval tasks since they have better feature representation ability, training VLP means leveraging lots of extra multi-modal training data and the video–text retrieval framework with the VLP model also means adding a mount of parameters. Moreover, due to the training codes of these VLP based video–text retrieval algorithms are not publicly available, we have not compared these methods in this paper.

2.2. Similarity Calculation

The most common method of calculating cross-modal similarity in video–text retrieval task is that first project video and text into a common semantic space and then compute the feature distance in the common space [34,39,40,41,42]. Additionally, Lei et al. [39] concatenated video and text features, and fed them into a transformer for multi-modal fusion. Directly predicting the cross-modal similarity by fusing multi-modal representations without learning the common semantic space explicitly in other cross-modal understanding tasks [43,44] is not uncommon. Yu et al. [43] proposed a guided attention module to fuse text and image representations for visual question answering. Yu et al. [44] presented a multi-modal Transformer to cascade the image and text features in depth. for image captioning. Though the recent image-text retrieval literature [45] has demonstrated that multi-modal fusion can make the model more stable, extending it for video retrieval is unwise because it is at the expense of efficiency. It should be noted that all existing methods (whether it is common space base framework) calculate the similarity score either in global or entity and action semantic levels, which neglect the relationship information widely existing in multimedia data. Instead, we are the first to measure the similarity of video–text pairs using relationship semantic level, which is also a complementary and crucial clue for video–text retrieval.

3. Methodology

3.1. Problem Statement

Video–text retrieval includes two aspects: (1) Providing a sentence as a query and return the corresponding videos. (2) Providing a video as a query and return the corresponding sentences. The major challenge to achieve these goals is modeling the similarity between any video–text pairs. Formally, given a video V and a sentence T, the video–text retrieval model will output S ( V , T ) = ( 0 , 1 ) to indicate the similarity of this video–text pair, where the larger the value of S ( V , T ) , the more similar the semantics of video V and text T.

3.2. Overall Framework

We introduce a novel multi-level cross-modal semantic alignment network (MCSAN) to improve the performance of video–text retrieval tasks. In order to precisely handle both the semantics of video and sentence, we employ several carefully designed spatial–temporal representation structures to encode both video and sentence into four semantic levels (global, action, entity and relationship). To calculate the similarity of video and sentence, we utilize metric learning, statistical distribution constraints and cross-modal parameter-share attribute projection layers to ensure multi-level representations extracted from different modalities are in multi-level common spaces. As the semantic distribution is seriously uneven, we propose a novel multi-label regression loss that can automatical focus on samples that are hard to be identified. The overall architecture of MCSAN is illustrated in Figure 3; in that which follows, we depict it in detail.
The proposed MCSAN is reasonable: (1) Obviously, as shown in Figure 3, a query sentence (e.g., “Cars racing on a road surrounded by lots of people.”) corresponding to a specific video clip has multiple semantic levels including global-level (the whole sentence), verb-level (actions), and noun-level (entities). The multiple semantic levels involves complicated interactions between them, which are actually organized as a relationship structure, indicating that a textual query can be effectively grounded onto the video by properly aligning different semantic levels with the corresponding video parts. Although some previous methods attempt to utilize the multi-level semantic structure, they fail to capture the fine-grained and explicit relationship information. (2) Intuitively, in video retrieval or video captioning tasks, humans usually takes the relation-aware context of the visual structure into consideration and select the correct temporal moment.

3.3. Multi-Level Attributes Vocabulary Construction

As mentioned above, besides global representation, we also represent both video and text into three attribute spaces, i.e., action, entity and relationship. To achieve this goal, attribute vocabulary is necessary for training the encoding modules. In this work, we employ the Stanford CoreNLP Natural Language Processing Toolkit [46] to build action, entity and relationship vocabularies. Specifically, for a video–text retrieval dataset, we collect all sentences as original corpus, and extract nouns, verbs and subject–verb–object (SVO) triplets, which are deemed as entity, action and relationship, respectively. Then, we lemmatize all words to eliminate duplication of attributes. Finally, based on the frequency statistics, we select the top K e nouns, K a verbs and K r SVO triplets as the final entity, action and relationship vocabularies, respectively.

3.4. Multi-Level Text and Video Semantics Encoding

In this section, we detail how to encode both video/text into global, entity, action and relationship semantic spaces.

3.4.1. Multi-Level Text Semantics Encoding

(1) Text parsing and initial embedding.
For a given sentence T, we first utilize the pre-trained word embedding model as the textual encoder to extract word initial embeddings E = [ e 1 , e 2 , , e M ] , where M denotes the number of words in sentence T.
(2) Text global semantic encoding.
The text global semantic encoding F t g is calculated by pre-trained S e n t e n c e _ B E R T [47] since it sets a new state-of-the-art performance on semantic textual similarity task:
F t g = S e n t e n c e _ B E R T ( T ) ,
(3) Text entity, action semantic encoding.
Since prior works [10,11] demonstrate that relational GCN [48] performs well on local-level text representations, we adopt relational GCN to encode the semantics of text entity and action, which are denoted as F t e and F t a , respectively. The structure of relational GCN is the same as [11] except not handle global feature, please refer to [11] for details.
(4) Text relationship semantic encoding.
Compared with entity or action, the relationship in a sentence is complex, e.g., “A dog is chasing a frisbee on the grass” includes two relationships (“dog chasing frisbee” and “dog on grass”) and words in these relationships are nonadjacent. Addressing it, we utilize a Transformer with multi-head self-attention block (as shown in Figure 4) to encode the text relationship semantic.
Specifically, the self-attention layer produces the attention output as follows:
S t r = s o f t m a x ( Q t K t d ) V t ,
where d indicates the feature dimension. Q t , K t and V t are Query, Key and Value matrixes of the self-attention layer, which are based on the original word initial embedding with positing encoding E ^ :
Q t = W t q E ^ , K = W t k E ^ , V = W t v E ^ ,
where W t q , W t k , and W t v are the embedding weights of the self-attention layer.
Based on several affinity matrixes, then compute the context representation H t r :
H ^ t r = L a y e r N o r m ( S t r + E ^ ) ,
H t r = L a y e r N o r m ( H ^ t r + L i n e a r ( R e L U ( L i n e a r ( H ^ t r ) ) ) ) ,
where LayerNorm is the layer normalization operation and Linear(ReLU()) is a fully connected position-wise FFN layer.
Finally, the text relationship semantic encoding F t r is obtained by applying mean pooling on H t r , that is:
F t r = 1 N i = 1 N H t i r ,
where H t i r is the context representation of i-th word.

3.4.2. Multi-Level Video Semantics Encoding

Parallel to multi-level text semantics encoding, we also encode video into global, action, entity and relationship spaces.
(1) Video parsing and initial embedding.
Given a video V, we first uniformly sample N frames, then use the pre-trained feature extraction model to extract general frame-wise features, i.e., F = [ f 1 , f 2 , , f N ] , here, f i represents the i-th frame general feature of video V. We then leverage four residual sub-networks to convert the general frame-level features to level-specific frame initial embeddings as follows:
F g = R e L U ( F + F C g ( F ) ) ,
F e = R e L U ( F + F C e ( F ) ) ,
F a = R e L U ( F + F C a ( F ) ) ,
F r = R e L U ( F + F C r ( F ) ) ,
where F g = [ f 1 g , f 2 g , , f N g ] R N × D v , F e = [ f 1 e , f 2 e , , f N e ] R N × D v , F a = [ f 1 a , f 2 a , , f N a ] R N × D v and F r = [ f 1 r , f 2 r , , f N r ] R N × D v are frame-wise global, entity, action and relationship initial embeddings, respectively. Here, D v is the dimension of the extracted visual feature. F C g , F C e , F C a and F C r are four different fully connected layers.
(2) Video global semantic encoding.
Bi-directional recurrent neural network [49] is an effective structure to handle both past and future contextual information of a given video sequence. We adopt a bidirectional GRU with soft attention (BiGRU-SA) structure to learn video global semantic encoding from frame-wise global initial embedding F g .
As shown in Figure 5, to extract more salient global semantic and suppress unimportant frames automatically, we first utilize the attention mechanism to refine the F g , i.e.,
F ^ g = α F g ,
where F ^ g is frame-wise global initial embedding after soft attention, and α is the attention score which is computed as follows:
α = s o f t m a x ( F g W ) ,
where W is the learnable weight matrix.
Let G R U and G R U indicate the forward and backward GRUs, h and h are corresponding hidden states generated as follows:
h = G R U ( F ^ g ) ,
h = G R U ( F ^ g ) ,
Finally, we concatenate the last hidden states of h and h to obtain the final video global semantic encoding:
F v g = c o n c a t ( h N , h 1 ) ,
where concat denotes the concatenate operation, h N represents the hidden vector at the N-th time step of the forward GRU layer, and h 1 represents the hidden vector at the 1-th time step of the backward GRU layer.
(3) Video entity semantic encoding.
Intuitively, the video entity depends on each frame. To be specific, on the one hand, if any frame contains a certain entity means that the entire video also has this entity concept. On the other hand, if the video does not contain a certain entity concept means that all frames do not contain this entity concept. We adapt 1D CNN with kernel size 1 to model this characteristic, as illustrated in Figure 6a, the final video entity semantic encoding F v e can be represented as follows:
F v e = R e L U ( C o n v 1 d k = 1 ( F e ) ) ,
(4) Video action semantic encoding.
We also adapt 1D CNN to encode video action semantic. However, different from entity, action is motion information, which depends on several continuous video frames. Therefore, we integrate three 1D CNN with different kernel sizes to handle this challenge, the final video action semantic encoding F v a can be represented as follows:
F v a = β 1 R e L U ( C o n v 1 d k = 2 ( F e ) ) + β 2 R e L U ( C o n v 1 d k = 3 ( F e ) ) + β 3 R e L U ( C o n v 1 d k = 5 ( F e ) ) ,
where β 1 , β 2 , and β 3 serve as the balance factor.
(5) Video relationship semantic encoding.
Different from entity and action, the relationship semantics associated two objects may appear in discontinuous frames. To address this challenge, consistent with text relation semantic extraction, we also present a self-attention based transformer encoder (as shown in Figure 7) to compute the final video relationship semantic encoding F v r from the frame-wise relationship initial embedding F r .
Specifically, the self-attention layer produces the attention output S v r as follows:
S v r = s o f t m a x ( Q v K v d ) V v ,
where d indicates the feature dimension. Q v , K v and V v are query, key and value matrixes of the self-attention layer, which are based on the original frame-wise relationship initial embedding with positing encoding F ^ r :
Q v = W v q F ^ r , K = W v k F ^ r , V = W v v F ^ r ,
where W v q , W v k , and W v v are the embedding weights of the self-attention layer.
Based on several affinity matrixes, then compute the context representation H v r :
H ^ v r = L a y e r N o r m ( S v r + F ^ r ) ,
H v r = L a y e r N o r m ( H ^ v r + L i n e a r ( R e L U ( L i n e a r ( H ^ v r ) ) ) ) ,
where LayerNorm is the layer normalization operation and Linear(ReLU()) is a fully connected position-wise FFN layer.
Finally, the video relationship semantic encoding F v r is obtained by applying mean pooling on H v r , that is:
F v r = 1 N i = 1 N H v i r ,
where H v i r is the context representation of ith frame.

3.5. Multi-Level Text and Video Semantics Alignment

In this section, we detail how to project the real-valued semantic vectors extracted from different modalities into a common space and calculate their similarity at different semantic levels.

3.5.1. Global Alignment

In order to make different dimensions in F v g and F t g fully interactive, we first use Holographic Embeddings [50] to fusion F v g and F t g , then leverage a fully connected layer (FC) to calculate the similarity of F v g and F t g , formally:
s i m ( F v g , F t g ) = σ ( F C ( F v g F t g ) ) ,
where s i m ( F v g , F t g ) denotes the global semantic similarity, σ is Sigmoid non-linear function which projects the similarity score into ( 0 , 1 ) , ★ represents circular correlation operation, i.e., [ a b ] k = i = 0 d f 1 a i b k + i % d f , here d f is feature dimension.

3.5.2. Entity, Action and Relationship Alignment

(1) Statistical distribution constraints.
In order to ensure entity, action and relationship features extracted from different modalities that are projected into a common space, we leverage cross-modal parameter-share attribute projection layer (i.e., F C * s h a r e ) in different semantic levels. That is,
p v e = R e L U ( F C e s h a r e ( F v e ) ) ,   p t e = R e L U ( F C e s h a r e ( F t e ) ) ,
p v a = R e L U ( F C a s h a r e ( F v a ) ) ,   p t a = R e L U ( F C a s h a r e ( F t a ) ) ,
p v r = R e L U ( F C r s h a r e ( F v r ) ) ,   p t r = R e L U ( F C r s h a r e ( F t r ) ) ,
where p v e , p v a and p v r are the predicted entity, action and relationship of the video. p t e , p t a and p t r are the predicted entity, action and relationship of the text.
In addition, we also consider minimizing the cross-modal feature distribution in different semantic levels as follows:
m i n ( d i s t r i b u t i o n ( F v e ) ,   d i s t r i b u t i o n ( F t e ) ) ,
m i n ( d i s t r i b u t i o n ( F v a ) ,   d i s t r i b u t i o n ( F t a ) ) ,
m i n ( d i s t r i b u t i o n ( F v r ) ,   d i s t r i b u t i o n ( F t r ) ) ,
where F denotes semantic representation matrix of a mini-batch.
(2) Similarity calculation.
We use the cosine similarity to measure cross-modal similarity in entity, action and relationship levels, that is:
s i m ( F v e , F t e ) = c o s ( F v e , F t e ) ,
s i m ( F v a , F t a ) = c o s ( F v a , F t a ) ,
s i m ( F v r , F t r ) = c o s ( F v r , F t r ) ,
where s i m ( F v e , F t e ) , s i m ( F v a , F t a ) and s i m ( F v r , F t r ) denote the entity, action and relationship semantic similarities of the video V and the sentence T, respectively.

3.6. Training and Inference

3.6.1. Training

Once all similarity scores in four semantic levels are computed, the similarity between the video V and the sentence T is calculated by:
s i m ( V , T ) = ( s i m ( F v g , F t g ) + s i m ( F v e , F t e ) + s i m ( F v a , F t a ) + s i m ( F v r , F t r ) ) / 4 ,
where s i m ( V , T ) denotes the overall similarity of the video V and the sentence T.
We use the triplet ranking loss with hard negative sampling strategy (i.e., L r a n k ) to optimize MCSAN, that is:
L r a n k = m a x ( 0 , m + s i m ( V , T ) s i m ( V , T ) ) + m a x ( 0 , m + s i m ( V , T ) s i m ( V , T ) ) ,
where ( V , T ) are the positive pair, T and V are the hardest negatives of text and video in a mini-batch, and m is a positive constant, which we term the margin.
In addition, we also use a novel multi-label regression loss for entity, action and relationship semantic learning. The original video–text retrieval dataset has no attribute annotations. Therefore, annotating entity, action and relationship for each video/text is crucial for learning multi-level representations. To ensure video and text share a common attribute space, we annotate video/text with the same vocabulary (Section 3.3). To be specific, for each sentence, we also use Standford NLP ToolKit to parse its entity, action and relationship as annotations. For each video, we parse all corresponding sentences and take their union as the annotations. Finally, all annotations are labeled in one-hot format based on the vocabulary (that is, entity, action and relationship can be predicted by a multi-label regression manner).
As illustrated in Section 1, the distribution of attributes (i.e., entity, action and relationship) is very imbalanced. Thus, the widely-used binary cross-entropy (BCE) loss may affect the generalization performance of the video–text retrieval model. To address it, inspired by the Focal loss [51], which performed successfully on unbalanced object detection tasks, we propose a novel focal binary cross-entropy (FBCE) loss for entity, action and relationship learning as:
L e = 1 K i = 1 K ( y v , i e α e ( 1 p v , i e ) γ e l o g ( p v , i e ) + ( 1 y v , i e ) α e ( 1 p v , i e ) γ e l o g ( 1 p v , i e ) ) 1 K i = 1 K ( y t , i e α e ( 1 p t , i e ) γ e l o g ( p t , i e ) + ( 1 y t , i e ) α e ( 1 p t , i e ) γ e l o g ( 1 p t , i e ) ) ,
L a = 1 K i = 1 K ( y v , i a α a ( 1 p v , i a ) γ a l o g ( p v , i a ) + ( 1 y v , i a ) α a ( 1 p v , i a ) γ a l o g ( 1 p v , i a ) ) 1 K i = 1 K ( y t , i a α a ( 1 p t , i a ) γ a l o g ( p t , i a ) + ( 1 y t , i a ) α a ( 1 p t , i a ) γ a l o g ( 1 p t , i a ) ) ,
L r = 1 K i = 1 K ( y v , i r α r ( 1 p v , i r ) γ r l o g ( p v , i r ) + ( 1 y v , i r ) α r ( 1 p v , i r ) γ r l o g ( 1 p v , i r ) ) 1 K i = 1 K ( y t , i r α r ( 1 p t , i r ) γ r l o g ( p t , i r ) + ( 1 y t , i r ) α r ( 1 p t , i r ) γ r l o g ( 1 p t , i r ) ) ,
where K is the batch size. y v , i e , y v , i a and y v , i r are the ground-truth entity, action and relationship of the i-th video. y t , i e , y t , i a and y t , i r are the ground-truth entity, action and relationship of the i-th text. Please refer to [51] for the details of factors α and γ .
In addition, we leverage Kullback–Leibler divergence loss (KLDivLoss) to constrain that features from different modalities have the same distribution. That is:
L e _ k l = K L D i v L o s s ( F v e , F t e ) ,
L a _ k l = K L D i v L o s s ( F v a , F t a ) ,
L r _ k l = K L D i v L o s s ( F v r , F t r ) ,
Finally, we combine the above losses to obtain the total objective function L t o t a l as follows:
L t o t a l = L r a n k + λ 1 ( L e + L a + L r ) + λ 2 ( L e _ k l + L a _ k l + L r _ k l ) ,
where λ 1 , λ 2 are trade-off hyper-parameters.

3.6.2. Inference

We simply sort the similarity scores of all video–text pairs for video–text retrieval.

4. Experiments and Results

To justify the effectiveness of our MCSAN model, we carried out extensive experiments under the bidirectional retrieval scenario including (1) text-to-video retrieval, i.e., retrieving videos that are semantically consistent with a given sentence query; and (2) video-to-text retrieval, i.e., retrieving sentences that are semantically consistent with a given video query. In this section, we first give the details about experimental settings, including datasets, evaluation metrics and implementation details. Then, we compare our MCSAN with several state-of-the-art baselines. Next, we explore the effectiveness of each component in our model via ablation studies. Finally, several qualitative and visualization experimental results are reported.

4.1. Experimental Settings

4.1.1. Datasets

We conduct extensive experiments on two most widely-used video–text retrieval datasets, namely, MSR-VTT [27] and VATEX [30], to evaluate our proposed method and several state-of-the-art baselines.
MSR-VTT consists of 10,000 video clips, where each video is described by 20 different sentences. Following the official setting, this dataset is split into 6573 training videos, 497 validation videos, and 2990 testing videos.
VATEX is a large-scale dataset including more than 30,000 video clips, where each video is associated with 10 English and 10 Chinese sentences. Here, only English sentences are used in our experiments. Following the settings in previous work [3,9,11], where 25,991, 1500, 1500 video clips are utilized for training, validation and testing, respectively.

4.1.2. Evaluation Metrics

Following prior cross-modal retrieval works [11,52], rank-based metric, i.e., Recall at K (R@K for short), Sum of all Recalls (SumR for short), are adopted to measure the overall video–text retrieval performance. The higher R@K and SumR indicate that the model has better retrieval performance.

4.1.3. Implementation Details

For a fair comparison with all the compared methods, we apply the same feature as HANet [11]. To be specific, for visual embedding, we use the pre-trained ResNet152 [53] and I3D [54] to extract frame features of MSR-VTT and VATEX, respectively. For text embedding, we use the pre-trained Glove [55] to extract word features of both MSR-VTT and VATEX. Here, the dimension of text embedding is 300 (extracted from Glove), and the dimension of visual embedding are 2048 (extracted from ResNet152) and 1024 (extracted from I3D) for MSR-VTT and VATEX, respectively. We optimized our MCSAN using PyTorch on 1 GeForce RTX 2080Ti GPU. Following HANet [11], the Adam [56] optimizer is employed with learning rate of 1 × 10 4 and the model is trained for 50 epochs with the batch size of 64. The trade-off hyper-parameters λ 1 and λ 2 are empirically set to 0.1 and 0.001, respectively. We estimate the hyperparameter margin m by running a grid search on the corresponding dataset, and the optimal value of m is empirically set to 0.25. In our FBCE loss, following the original Focal loss [51], α e , α a and α r are empirically set to 0.25, γ e , γ a and γ r are empirically set to 2. Note that, the model with the highest SumR value on validation set is deemed as the best model to report the performance on the test set.

4.2. Comparison with State-of-the-Art Methods

We compare our model with two types of baselines: compute the similarity between video and text only using the global features, i.e., VSE [57], VSE++ [58], Mithum et al. [2], W2VV [41], W2VV++ [8], Dual Encoding [3], CE [36], TCE [33], Zhao et al. [59], and compute the similarity between video and text using several level features, i.e., HGR [9], HANet [11], HSL [60]. The experimental results of our proposed MCSAN and all baseline approaches are reported in Table 1 and Table 2. Note that all models are trained using the same initial video/text features on the same dataset. From Table 1 and Table 2, we can have the following observations:
(1) In the two datasets, methods utilizing multi-level features perform better than these algorithms only use the global feature significantly, indicating that representations extracted from different semantic levels are complementary for video–text retrieval task.
(2) For all video–text retrieval methods, video-to-text has better performance than text-to-video. We think the reason is that on existing video–text retrieval dataset, one video corresponds to multiple sentences; however, one text can only correspond to one video. The results indicate that collecting a new video–text retrieval dataset with proportional numbers of video and text may be a feasible research direction.
(3) The proposed MCSAN outperforms all the baselines on MSR-VTT and VATEX datasets. The results demonstrate that the proposed model can jointly model multi-modal context information and multi-level semantics of both video and text in a unified deep model, which can better capture the underlying complementary representation of video/text, so as to improve the performance of video–text retrieval.
Table 1. Comparisons with the state-of-the-art video–text retrieval methods on the MSR-VTT dataset. The best performance are marked in bold. Our proposed model performs the best.
Table 1. Comparisons with the state-of-the-art video–text retrieval methods on the MSR-VTT dataset. The best performance are marked in bold. Our proposed model performs the best.
MethodText-to-VideoVideo-to-TextSumR
R@1R@5R@10R@1R@5R@10
VSE [57]5.016.424.67.720.331.2105.2
VSE++ [58]5.717.124.810.225.435.1118.3
Mithum et al. [2]5.817.625.210.526.735.9121.7
W2VV [41]6.118.727.511.828.939.1132.1
Dual Encoding [3]7.722.031.813.030.843.3148.6
TCE [33]7.722.532.1----
Zhao et al. [59]8.825.536.514.033.144.9162.8
HGR [9]9.226.236.515.036.748.8172.4
HANet [11]9.327.038.116.139.252.1181.8
HSL [60]10.829.238.520.345.155.9201.7
MCSAN10.929.640.121.146.857.4205.3
Table 2. Comparisons with the state-of-the-art video–text retrieval methods on the VATEX dataset. The best performance are marked in bold. The overall performance are indicated by Sum of Recalls (SumR). MCSAN is the best.
Table 2. Comparisons with the state-of-the-art video–text retrieval methods on the VATEX dataset. The best performance are marked in bold. The overall performance are indicated by Sum of Recalls (SumR). MCSAN is the best.
MethodText-to-VideoVideo-to-TextSumR
R@1R@5R@10R@1R@5R@10
W2VV [41]14.636.346.139.669.579.4285.5
VSE++ [58]31.365.876.442.973.983.6373.9
CE [36]31.168.780.241.371.082.3374.6
W2VV++ [8]32.068.278.841.875.184.3380.2
Dual Encoding [3]31.167.478.9----
HGR [9]35.173.583.5----
HSL [60]36.873.683.746.875.785.1401.7
HANet [11]36.474.184.149.179.586.2409.4
MCSAN36.674.284.151.481.386.8411.8

4.3. Ablation Studies

Because the proposed MCSAN contains multiple key components, we additional compare variants of MCSAN with respect to the following perspectives to verify the effectiveness of each component in MCSAN: (1) the effect of the relationship information, (2) the impact of the statistical distribution constraints, and (3) the effect of the proposed FBCE loss for handling unbalanced attributes. The following MCSAN variants are designed for comparison. The following MCSAN variants are designed for comparison.
(1) MCSAN¬R: A variant of MCSAN with the relationship encoding level being removed.
(2) MCSAN¬S: A variant of MCSAN with the statistical distribution constraints being removed.
(3) MCSAN¬L: A variant of MCSAN with the FBCE loss being removed, and only using the conventional BCE loss for entity, action and relationship learning.
The ablation study results are shown in Table 3. We can observe that the combination of all three components has the best performance and the performance will degrade if remove any components. The results demonstrate that all components in our MCSAN are effective. Moreover, we can also observe that if the relationship encoding is removed, the performance will degrade the most, which indicates that explicitly relationship understanding is crucial for video–text retrieval.

4.4. Qualitative Analyses

We visualize three text-to-video retrieval examples on the MSR-VTT test set in Figure 8. For each query sentence, Figure 8 shows the top-3 scored videos evaluated by our MCSAN. We find that the proposed model is able to recognize the relationship “man at top speed” since comprehensive relationship between concepts is explicitly modeled in our framework. The results demonstrate that the proposed method can handle different types of video/text semantics well.

4.5. Limitations

Although we achieve good results in the previous experiments, there are still some inevitable problems. Note that Rao et al. [45] demonstrated that modality embedding is crucial for the cross modal retrieval performance improvement. However, only simple CNN/Transformer structures are adopted in this paper. Thus, we will explore a more effective way to extract text/video semantic features in the future. Moreover, we will try to leverage additional knowledge information (e.g., knowledge embedding learnt from the external sentence corpus) since it can provide useful complementary information for video–text retrieval.

5. Conclusions

In this paper, we propose a novel Multi-Level Cross-Modal Semantic Alignment Network (MCSAN) for video–text retrieval tasks. The semantic similarity between video and text is fully considered at four semantic levels (global, action, entity and relationship) by carefully designed spatial–temporal semantic learning structures and cross-modal representation distribution constraints in an end-to-end framework. We make the first attempt to handle the cross-modal relationship semantic consistency without relying on any extern corpus in video–text retrieval task. Moreover, we introduce a novel focal binary cross-entropy (FBCE) loss function, which also is the first effort to model the unbalanced attribute distribution problem for video–text retrieval. We conduct extensive experiments on two video–text retrieval benchmarks, i.e., MSR-VTT and VATEX. The quantitative and qualitative experimental results demonstrate the effectiveness of the proposed framework. We achieve state-of-the-art performance against existing non-VLP-based video–text retrieval frameworks. Due to that most recent VLP based video–text retrieval algorithms are not publicly available, we hope that future work in the field of video–text retrieval should pay more attention to the reproducibilities and capabilities of the model, rather than use all the tricks to report a hard-to-reproduce better performance.

Author Contributions

Conceptualization, methodology, and writing—original draft preparation, F.N.; writing—review and editing, Y.G.; visualization, L.D.; funding acquisition, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation (NSF) of China (No. 61902104), the Anhui Provincial Key Research and Development Program (No. 2022a05020042, No. 2022a05020043), the Anhui Provincial Natural Science Foundation (No. 2008085QF295, No. 2008085QF316), the University Natural Sciences Research Project of Anhui Province (No. KJ2020A0651), New Energy Vehicle and Intelligent Networked Vehicle Innovation Project (No. wfgcyh2020477), and Anhui International Joint Research Center for Ancient Architecture Intellisencing and Multi-Dimensional Modeling (No. GJZZX2021KF01).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository. The data presented in this study are openly available at https://github.com/Roc-Ng/HANet, accessed on 26 July 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kaur, P.; Pannu, H.S.; Malhi, A.K. Comparative analysis on cross-modal information retrieval: A review. Comput. Sci. Rev. 2021, 39, 100336. [Google Scholar] [CrossRef]
  2. Mithun, N.C.; Li, J.; Metze, F.; Roy-Chowdhury, A.K. Learning joint embedding with multimodal cues for cross-modal video–text retrieval. In Proceedings of the 2018 ACM on International Conference on Multimedia Retrieval, Yokohama, Japan, 11–14 June 2018; pp. 19–27. [Google Scholar]
  3. Dong, J.; Li, X.; Xu, C.; Ji, S.; He, Y.; Yang, G.; Wang, X. Dual encoding for zero-example video retrieval. In Proceedings of the IEEE/CVF Conference On Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9346–9355. [Google Scholar]
  4. Wang, W.; Gao, J.; Yang, X.; Xu, C. Learning coarse-to-fine graph neural networks for video–text retrieval. IEEE Trans. Multimed. 2020, 23, 2386–2397. [Google Scholar] [CrossRef]
  5. Jin, W.; Zhao, Z.; Zhang, P.; Zhu, J.; He, X.; Zhuang, Y. Hierarchical cross-modal graph consistency learning for video–text retrieval. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021; pp. 1114–1124. [Google Scholar]
  6. Gorti, S.K.; Vouitsis, N.; Ma, J.; Golestan, K.; Volkovs, M.; Garg, A.; Yu, G. X-Pool: Cross-Modal Language-Video Attention for Text-Video Retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 5006–5015. [Google Scholar]
  7. Feng, Z.; Zeng, Z.; Guo, C.; Li, Z. Exploiting visual semantic reasoning for video–text retrieval. arXiv 2020, arXiv:2006.08889. [Google Scholar]
  8. Li, X.; Xu, C.; Yang, G.; Chen, Z.; Dong, J. W2vv++ fully deep learning for ad-hoc video search. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1786–1794. [Google Scholar]
  9. Chen, S.; Zhao, Y.; Jin, Q.; Wu, Q. Fine-grained video–text retrieval with hierarchical graph reasoning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10638–10647. [Google Scholar]
  10. Wray, M.; Larlus, D.; Csurka, G.; Damen, D. Fine-grained action retrieval through multiple parts-of-speech embeddings. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–3 November 2019; pp. 450–459. [Google Scholar]
  11. Wu, P.; He, X.; Tang, M.; Lv, Y.; Liu, J. HANet: Hierarchical Alignment Networks for Video–text Retrieval. In Proceedings of the 29th ACM international conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 3518–3527. [Google Scholar]
  12. Reddy, M.D.M.; Basha, M.S.M.; Hari, M.M.C.; Penchalaiah, M.N. Dall-e: Creating images from text. UGC Care Group I J. 2021, 8, 71–75. [Google Scholar]
  13. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning (PMLR), Virtual Event, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  14. Wei, L.; Xie, L.; Zhou, W.; Li, H.; Tian, Q. MVP: Multimodality-guided Visual Pre-training. arXiv 2022, arXiv:2203.05175. [Google Scholar]
  15. Yang, Z.; Garcia, N.; Chu, C.; Otani, M.; Nakashima, Y.; Takemura, H. Bert representations for video question answering. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1556–1565. [Google Scholar]
  16. Tang, M.; Wang, Z.; Liu, Z.; Rao, F.; Li, D.; Li, X. Clip4caption: Clip for video caption. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 4858–4862. [Google Scholar]
  17. Wang, Z.; Codella, N.; Chen, Y.C.; Zhou, L.; Yang, J.; Dai, X.; Xiao, B.; You, H.; Chang, S.F.; Yuan, L. CLIP-TD: CLIP Targeted Distillation for Vision-Language Tasks. arXiv 2022, arXiv:2201.05729. [Google Scholar]
  18. Luo, J.; Li, Y.; Pan, Y.; Yao, T.; Chao, H.; Mei, T. CoCo-BERT: Improving video-language pre-training with contrastive cross-modal matching and denoising. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 5600–5608. [Google Scholar]
  19. Fang, H.; Xiong, P.; Xu, L.; Chen, Y. Clip2video: Mastering video–text retrieval via image clip. arXiv 2021, arXiv:2106.11097. [Google Scholar]
  20. Luo, H.; Ji, L.; Zhong, M.; Chen, Y.; Lei, W.; Duan, N.; Li, T. Clip4clip: An empirical study of clip for end to end video clip retrieval. arXiv 2021, arXiv:2104.08860. [Google Scholar] [CrossRef]
  21. Gao, Z.; Liu, J.; Chen, S.; Chang, D.; Zhang, H.; Yuan, J. Clip2tv: An empirical study on transformer-based methods for video–text retrieval. arXiv 2021, arXiv:2111.05610. [Google Scholar]
  22. Nian, F.; Bao, B.K.; Li, T.; Xu, C. Multi-modal knowledge representation learning via webly-supervised relationships mining. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 411–419. [Google Scholar]
  23. Yao, T.; Pan, Y.; Li, Y.; Mei, T. Exploring visual relationship for image captioning. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 684–699. [Google Scholar]
  24. Wang, S.; Wang, R.; Yao, Z.; Shan, S.; Chen, X. Cross-modal scene graph matching for relationship-aware image-text retrieval. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1508–1517. [Google Scholar]
  25. Shvetsova, N.; Chen, B.; Rouditchenko, A.; Thomas, S.; Kingsbury, B.; Feris, R.S.; Harwath, D.; Glass, J.; Kuehne, H. Everything at Once-Multi-Modal Fusion Transformer for Video Retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 20020–20029. [Google Scholar]
  26. Wray, M.; Doughty, H.; Damen, D. On semantic similarity in video retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 3650–3660. [Google Scholar]
  27. Xu, J.; Mei, T.; Yao, T.; Rui, Y. Msr-vtt: A large video description dataset for bridging video and language. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5288–5296. [Google Scholar]
  28. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  30. Wang, X.; Wu, J.; Chen, J.; Li, L.; Wang, Y.F.; Wang, W.Y. Vatex: A large-scale, high-quality multilingual dataset for video-and-language research. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–3 November 2019; pp. 4581–4591. [Google Scholar]
  31. Li, T.; Ni, B.; Xu, M.; Wang, M.; Gao, Q.; Yan, S. Data-driven affective filtering for images and videos. IEEE Trans. Cybern. 2015, 45, 2336–2349. [Google Scholar] [CrossRef]
  32. Nian, F.; Li, T.; Wu, X.; Gao, Q.; Li, F. Efficient near-duplicate image detection with a local-based binary representation. Multimed. Tools Appl. 2016, 75, 2435–2452. [Google Scholar] [CrossRef]
  33. Yang, X.; Dong, J.; Cao, Y.; Wang, X.; Wang, M.; Chua, T.S. Tree-augmented cross-modal encoding for complex-query video retrieval. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 1339–1348. [Google Scholar]
  34. Dong, J.; Wang, Y.; Chen, X.; Qu, X.; Li, X.; He, Y.; Wang, X. Reading-strategy inspired visual representation learning for text-to-video retrieval. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5680–5694. [Google Scholar] [CrossRef]
  35. Gabeur, V.; Sun, C.; Alahari, K.; Schmid, C. Multi-modal transformer for video retrieval. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 214–229. [Google Scholar]
  36. Liu, Y.; Albanie, S.; Nagrani, A.; Zisserman, A. Use what you have: Video retrieval using representations from collaborative experts. arXiv 2019, arXiv:1907.13487. [Google Scholar]
  37. Miech, A.; Laptev, I.; Sivic, J. Learning a text-video embedding from incomplete and heterogeneous data. arXiv 2018, arXiv:1804.02516. [Google Scholar]
  38. Gabeur, V.; Nagrani, A.; Sun, C.; Alahari, K.; Schmid, C. Masking modalities for cross-modal video retrieval. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 1766–1775. [Google Scholar]
  39. Song, X.; Chen, J.; Wu, Z.; Jiang, Y.G. Spatial–temporal graphs for cross-modal text2video retrieval. IEEE Trans. Multimed. 2021, 14, 2914–2923. [Google Scholar] [CrossRef]
  40. Miech, A.; Zhukov, D.; Alayrac, J.B.; Tapaswi, M.; Laptev, I.; Sivic, J. Howto100m: Learning a text-video embedding by watching hundred million narrated video clips. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–3 November 2019; pp. 2630–2640. [Google Scholar]
  41. Dong, J.; Li, X.; Snoek, C.G. Predicting visual features from text for image and video caption retrieval. IEEE Trans. Multimed. 2018, 20, 3377–3388. [Google Scholar] [CrossRef]
  42. Dong, J.; Long, Z.; Mao, X.; Lin, C.; He, Y.; Ji, S. Multi-level alignment network for domain adaptive cross-modal retrieval. Neurocomputing 2021, 440, 207–219. [Google Scholar] [CrossRef]
  43. Yu, Z.; Yu, J.; Cui, Y.; Tao, D.; Tian, Q. Deep modular co-attention networks for visual question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 6281–6290. [Google Scholar]
  44. Yu, J.; Li, J.; Yu, Z.; Huang, Q. Multimodal transformer with multi-view visual representation for image captioning. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 4467–4480. [Google Scholar] [CrossRef]
  45. Rao, J.; Wang, F.; Ding, L.; Qi, S.; Zhan, Y.; Liu, W.; Tao, D. Where Does the Performance Improvement Come From?—A Reproducibility Concern about Image-Text Retrieval. arXiv 2022, arXiv:2203.03853. [Google Scholar]
  46. Manning, C.D.; Surdeanu, M.; Bauer, J.; Finkel, J.R.; Bethard, S.; McClosky, D. The Stanford CoreNLP natural language processing toolkit. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Baltimore, MD, USA, 23–24 June 2014; pp. 55–60. [Google Scholar]
  47. Reimers, N.; Gurevych, I. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv 2019, arXiv:1908.10084. [Google Scholar]
  48. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Berg, R.v.d.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the European Semantic Web Conference, Heraklion, Greece, 3–7 June 2018; pp. 593–607. [Google Scholar]
  49. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
  50. Nickel, M.; Rosasco, L.; Poggio, T. Holographic embeddings of knowledge graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1955–1961. [Google Scholar]
  51. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  52. Qu, L.; Liu, M.; Wu, J.; Gao, Z.; Nie, L. Dynamic modality interaction modeling for image-text retrieval. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 11–15 July 2021; pp. 1104–1113. [Google Scholar]
  53. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  54. Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6299–6308. [Google Scholar]
  55. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  56. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  57. Kiros, R.; Salakhutdinov, R.; Zemel, R.S. Unifying visual-semantic embeddings with multimodal neural language models. arXiv 2014, arXiv:1411.2539. [Google Scholar]
  58. Faghri, F.; Fleet, D.J.; Kiros, J.R.; Fidler, S. Vse++: Improving visual-semantic embeddings with hard negatives. arXiv 2017, arXiv:1707.05612. [Google Scholar]
  59. Zhao, R.; Zheng, K.; Zha, Z.J. Stacked convolutional deep encoding network for video–text retrieval. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), Virtual, 6–10 July 2020; pp. 1–6. [Google Scholar]
  60. Dong, J.; Li, X.; Xu, C.; Yang, X.; Yang, G.; Wang, X.; Wang, M. Dual encoding for video retrieval by text. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4065–4080. [Google Scholar] [CrossRef]
Figure 1. An illustration of both video and text are decomposed into global, entity, action and relationship semantic levels.
Figure 1. An illustration of both video and text are decomposed into global, entity, action and relationship semantic levels.
Mathematics 10 03346 g001
Figure 2. The number of 5 entities in MSR-VTT corpus, we can see that attributes distribution is very uneven.
Figure 2. The number of 5 entities in MSR-VTT corpus, we can see that attributes distribution is very uneven.
Mathematics 10 03346 g002
Figure 3. Framework of the proposed multi-level cross-modal semantic alignment network (MCSAN).
Figure 3. Framework of the proposed multi-level cross-modal semantic alignment network (MCSAN).
Mathematics 10 03346 g003
Figure 4. Text relationship semantic encoding architecture.
Figure 4. Text relationship semantic encoding architecture.
Mathematics 10 03346 g004
Figure 5. Video global semantic encoding architecture.
Figure 5. Video global semantic encoding architecture.
Mathematics 10 03346 g005
Figure 6. The structures of video entity and action semantic encoding using different 1D CNN.
Figure 6. The structures of video entity and action semantic encoding using different 1D CNN.
Mathematics 10 03346 g006
Figure 7. Video relationship semantic encoding architecture.
Figure 7. Video relationship semantic encoding architecture.
Mathematics 10 03346 g007
Figure 8. Top 3 text-to-video retrieval examples on the MSR-VTT test set.
Figure 8. Top 3 text-to-video retrieval examples on the MSR-VTT test set.
Mathematics 10 03346 g008
Table 3. Results of comparison among different variants in MCSAN on MSR-VTT dataset. The best performance are marked in bold. MCSAN exploiting all the four semantic levels (global, entity, action and realtionship) is the best.
Table 3. Results of comparison among different variants in MCSAN on MSR-VTT dataset. The best performance are marked in bold. MCSAN exploiting all the four semantic levels (global, entity, action and realtionship) is the best.
MethodText-to-VideoVideo-to-TextSumR
R@1R@5R@10R@1R@5R@10
MCSAN¬R9.427.137.917.242.354.7195.8
MCSAN¬S10.127.839.819.945.957.0202.6
MCSAN¬L10.628.839.220.846.456.6203.7
MCSAN10.929.640.121.146.857.4205.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nian, F.; Ding, L.; Hu, Y.; Gu, Y. Multi-Level Cross-Modal Semantic Alignment Network for Video–Text Retrieval. Mathematics 2022, 10, 3346. https://doi.org/10.3390/math10183346

AMA Style

Nian F, Ding L, Hu Y, Gu Y. Multi-Level Cross-Modal Semantic Alignment Network for Video–Text Retrieval. Mathematics. 2022; 10(18):3346. https://doi.org/10.3390/math10183346

Chicago/Turabian Style

Nian, Fudong, Ling Ding, Yuxia Hu, and Yanhong Gu. 2022. "Multi-Level Cross-Modal Semantic Alignment Network for Video–Text Retrieval" Mathematics 10, no. 18: 3346. https://doi.org/10.3390/math10183346

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop