Abstract
Dense Video Captioning (DVC) represents the cutting edge of advanced multimedia tasks, focusing on generating a series of temporally precise descriptions for events unfolding within a video. In contrast to traditional video captioning, which usually offers a singular summary or caption for an entire video, DVC demands the identification of multiple events within a video, the determination of their exact temporal boundaries, and the production of natural language descriptions for each event. This review paper presents a thorough examination of the latest techniques, datasets, and evaluation protocols in the field of DVC. We categorize and assess existing methodologies, delve into the characteristics, strengths, and limitations of widely utilized datasets, and underscore the challenges and opportunities associated with evaluating DVC models. Furthermore, we pinpoint current research trends, open challenges, and potential avenues for future exploration in this domain. The primary contributions of this review encompass: (1) a comprehensive survey of state-of-the-art DVC techniques, (2) an extensive review of commonly employed datasets, (3) a discussion on evaluation metrics and protocols, and (4) the identification of emerging trends and future directions.
1. Introduction
In the rapidly evolving landscape of computer vision and natural language processing (NLP), video understanding has emerged as a pivotal area of research. With the proliferation of video content across various platforms [1], the need for automated systems that can effectively analyze, interpret, and summarize video data has become increasingly urgent. Dense Video Captioning (DVC) stands at the forefront of this advancement [2], representing a sophisticated task that combines video analysis with language generation.
DVC aims to localize and describe multiple events within long, untrimmed videos [3]. Unlike traditional video captioning, which typically generates a single summary sentence or a sequence of sentences for the entire video, DVC requires the model to detect and caption multiple events occurring at different time intervals within the video. This task is highly challenging due to the complexity and diversity of video content [4], the variable length and overlap of events, and the need for accurate temporal localization and coherent natural language descriptions.
The motivation behind DVC stems from its potential applications in various domains. For instance, in video retrieval systems, DVC can enable more precise and intuitive search capabilities by allowing users to query based on event descriptions [5]. In video accessibility, DVC can provide detailed captions for deaf and hard-of-hearing individuals, enhancing their understanding and enjoyment of video content. Additionally, DVC has applications in content analysis [6], surveillance, and sports video summarization, where automatic event detection and description can save time and resources.
This comprehensive review paper aims to provide a detailed analysis of the state-of-the-art techniques, datasets, and evaluation protocols [7] for Dense Video Captioning (DVC). The paper is structured to cover the following key areas:
Fundamentals of Dense Video Captioning: This section introduces the fundamental concepts and challenges associated with DVC, including the subprocesses of video feature extraction, temporal event localization, and dense caption generation.
Techniques for Dense Video Captioning: In this section, we survey the existing methodologies for DVC, categorizing them into key subprocesses and discussing the strengths and limitations of each approach. We will cover a wide range of techniques, from traditional computer vision methods to deep learning-based architectures, including encoder-decoder models, attention mechanisms, and multimodal fusion techniques.
Datasets for Dense Video Captioning: This section presents an overview of the most widely used datasets for training and evaluating DVC models. We will discuss the characteristics, strengths, and limitations of each dataset, highlighting the challenges and opportunities they present for researchers in the field.
Evaluation Protocols and Metrics: In this section, we review the evaluation metrics commonly employed in DVC research, including BLEU [8], METEOR [9], CIDEr [10], ROUGE-L [11], SODA [12], SPICE [13], and WMD [14]. We will discuss the challenges in evaluating DVC models and the importance of developing more comprehensive and interpretable evaluation protocols.
Emerging Trends and Future Directions: This section summarizes the current research trends in DVC and identifies open challenges and limitations. We will propose potential future directions for DVC research, including improved event recognition and localization algorithms, enhanced multimodal integration, and more efficient and interpretable models. The contributions of this review paper are as follows:
- (1)
- A comprehensive survey of the state-of-the-art techniques for Dense Video Captioning, categorizing and analyzing the existing methodologies.
- (2)
- An in-depth review of the most widely used datasets for DVC, highlighting their strengths and limitations.
- (3)
- A detailed discussion of the evaluation metrics and protocols commonly employed in DVC research, including challenges and opportunities for future developments.
- (4)
- Identification of emerging trends and future directions in DVC research, providing valuable insights for researchers and practitioners in the field.
2. Fundamentals of Dense Video Captioning
2.1. Definition and Objectives of DVC
Dense Video Captioning (DVC) represents an advanced multimedia task that strives to automatically produce a series of temporally localized descriptions for events unfolding within a video [15]. Unlike conventional video captioning, which usually offers a solitary summary or caption per video, DVC demands the identification of multiple events within a video, the determination of their exact temporal boundaries, and the generation of a natural language description for each event. The core objective of DVC is to furnish a thorough, detailed, and human-readable comprehension of video content, thereby enhancing video retrieval, indexing, and browsing experiences.
2.2. Key Subprocesses
Video Feature Extraction (VFE): VFE constitutes the initial critical subprocess in DVC [16,17]. It entails extracting meaningful representations from raw video data for subsequent tasks like event localization and caption generation. This typically involves employing deep learning models, such as Convolutional Neural Networks (CNNs) for spatial feature extraction and Recurrent Neural Networks (RNNs) or Transformers for temporal feature encoding. The aim of VFE is to capture both the visual content and the temporal dynamics of the video.
Temporal Event Localization (TEL): TEL pertains to identifying the start and end times of events within a video. This subprocess necessitates precise temporal segmentation of the video into meaningful event segments. TEL can be tackled using diverse techniques, including sliding window methods, proposal-based methods, and end-to-end detection frameworks. The challenge resides in accurately detecting events, particularly when they overlap or occur in rapid succession.
Dense Caption Generation (DCG): DCG is the culminating subprocess in DVC, where natural language descriptions are crafted for each localized event. This involves encoding the visual and temporal features of the event into a format that can be decoded into a coherent sentence or phrase. DCG often harnesses encoder-decoder architectures, such as sequence-to-sequence models or Transformer-based models, trained on extensive datasets of video-caption pairs. The objective is to generate captions that are both precise and fluent, encapsulating the essential details of the event.
2.3. Challenges in DVC
Event Recognition and Overlapping Events: A primary challenge in DVC is accurately recognizing and differentiating between distinct events, especially when they occur simultaneously or overlap in time. This necessitates models with a robust ability to distinguish between visually similar events and manage temporal ambiguities.
Event Localization Accuracy: Precise event localization is vital for generating meaningful captions. However, achieving high localization accuracy is challenging due to the complexity and variability of video content. Factors like camera motion, occlusions, and varying event durations can all contribute to localization errors.
Multimodal Integration and Natural Language Generation: DVC requires the integration of multiple modalities, encompassing visual, auditory, and temporal information. Effective multimodal fusion is crucial for generating accurate and informative captions. Furthermore, the generated captions must be natural and fluent, which mandates models with a profound understanding of language structure and semantics.
3. Techniques for Dense Video Captioning
Recent years have seen remarkable progress in video captioning: the automated generation of descriptive text for video content. Central to these advancements are attention-based methodologies, which enable models to dynamically focus on the most relevant spatiotemporal features within videos. This review provides a structured analysis of cutting-edge techniques in dense video captioning, emphasizing three methodological pillars: attention mechanisms, transformer architectures, and encoder-decoder frameworks. This study highlights innovations in modeling long-range temporal dependencies and integrating multimodal cues (visual, auditory, textual) to enhance caption relevance and fidelity. These technical breakthroughs are particularly valuable given the growing demand for video captioning in human–computer interaction systems, content-based retrieval platforms, and automated video indexing pipelines. As visualized in Figure 1, our analysis systematically dissects current approaches through three complementary lenses-attention-driven strategies, transformer-based paradigms, and encoder-decoder architectures-to offer a comprehensive roadmap for researchers and practitioners in this dynamic field.
Figure 1.
Techniques for Dense Video Captioning.
3.1. Encoder-Decoder Based Approaches
3.1.1. Basic Encoder-Decoder Models
Video captioning aims to generate natural language descriptions of video content, and the basic encoder-decoder framework has been widely adopted for this task. Early methods employed models such as Recurrent Neural Networks (RNNs) and Convolutional Neural Networks (CNNs) to encode video features and decode them into captions. For instance, Venugopalan et al. [18] introduced the first encoder-decoder model for video captioning. As illustrated in Figure 2, features from the layer are extracted from individual frames. Here, refers to the fully connected layer features obtained from each frame using a pre-trained convolutional neural network (CNN), typically extracted from the layer of models like VGGNet. These features are then mean-pooled over the entire duration of the video and subsequently fed into an LSTM (Long Short-Term Memory) network at each time step. The LSTM generates a word per time step, considering both video features and the preceding word, until the end-of-sentence tag is chosen. Yao et al. [19] further improved this approach by incorporating temporal attention mechanisms.
Figure 2.
Redrawn schematic of the pioneering video caption encoder-decoder network, adapted from [19].
However, these basic models often suffer from limited performance. They struggle to capture complex temporal and spatial dependencies in videos, which leads to generated captions that lack specificity and diversity. This limitation underscores the need for more advanced techniques to enhance the quality and accuracy of video captioning.
3.1.2. Multimodal Fusion Encoder-Decoder Models
Multimodal fusion approaches aim to integrate information from video, audio, and text to generate more accurate captions. Hori et al. [20] introduced an attention-based method for multimodal fusion in video description, as shown in Figure 3. This method utilizes temporal attention mechanisms to selectively focus on encoded features from specific time frames. By combining visual, motion, and audio features, it captures a more comprehensive representation of the video content, enhancing the quality of the generated captions.
Figure 3.
Redrawn architecture of multimodal fusion encoder-decoder models (adapted from [20]) for video caption generation.
Building upon the idea that videos contain both visual and auditory information, Huang et al. [21] proposed a model that leverages both visual and audio features. Their model employs separate encoders for each modality and generates captions based on a fused multimodal representation. Similarly, Aafaq et al. [22] introduced a model that incorporates early linguistic information fusion through a visual-semantic embedding framework. However, these methods often require costly additional annotations for non-video modalities and often rely on simple fusion techniques, such as concatenation or element-wise summation. These techniques may not fully exploit the complementary nature of multimodal information, highlighting the need for more sophisticated fusion strategies.
3.1.3. Attention-Based Encoder-Decoder Models
To overcome the limitations of basic encoder-decoder models, attention mechanisms have been introduced. These mechanisms enable the model to focus on salient parts of the video while generating captions by dynamically adjusting the weight assigned to different video frames based on their relevance to the currently generated caption. Yu et al. [23] proposed a hierarchical RNN with an attention mechanism to encode videos at multiple temporal scales. The overall structure of this hierarchical RNN-based attention encoder-decoder framework is illustrated in Figure 4.
Figure 4.
Redrawn hierarchical RNN-based attention encoder-decoder framework for video captioning (adapted from [23]): Input in green, output in blue, recurrent components in red, and sentence generator reinitialization indicated by orange arrow.
Building on this, Pan et al. [24] introduced a spatio-temporal attention model that captures both spatial and temporal attention, further enhancing the model’s ability to focus on important video features. Despite these advancements, attention-based models still face challenges in capturing long-range dependencies and generating diverse captions. Ongoing research continues to explore solutions to these remaining issues.
3.1.4. Hierarchical Encoder-Decoder Models
Hierarchical encoder-decoder models aim to capture both local and global temporal structures in videos by employing multiple encoders and decoders that operate at different temporal scales. Wang et al. [25] proposed a hierarchical recurrent neural encoder that effectively captures both short-term and long-term dependencies in videos, which are then decoded into captions using an LSTM. This approach highlights the potential of hierarchical models in understanding complex video structures.
Building on this concept, Gao et al. [26] introduced the Hierarchical Representation Network with Auxiliary Tasks (HRNAT), as illustrated in Figure 5. This framework incorporates auxiliary tasks, such as cross-modality matching and syntax-guiding, to enhance caption generation. The novelty of their approach lies in the shared knowledge and parameters between auxiliary tasks and the main stream, including the Hierarchical Encoder Module, Hierarchical Attention Mechanism, and Language Generation Mechanism. While these hierarchical approaches offer a powerful way to model the complex structure of videos and captions at multiple levels of granularity, they can also increase model complexity and computational cost due to the multiple stages of processing involved.
Figure 5.
Redrawn hierarchical encoder-decoder model (adapted from [26]): HRNAT overview for video captioning (vc-HRNAT) with Event-level (orange), Action-level (light red), and Object-level (green) detail.
3.1.5. Memory-Augmented Encoder-Decoder Models
Memory-augmented encoder-decoder models have emerged as a powerful approach to enrich video captioning with external knowledge. These models integrate external memory to store and retrieve pertinent information, enabling them to capture implicit knowledge and produce more diverse captions. Wang et al. [27] proposed a memory-attended recurrent network that leverages an external memory to store visual and textual context. This context is then accessed during caption generation to enhance the overall context of the captions.
Building on this concept, Jing et al. [28] proposed the Memory-based Augmentation Network (MAN), as shown in Figure 6. The MAN consists of two key components: a memory-augmented encoder and a decoder. The encoder integrates a visual memory module that stores diverse visual contexts from the dataset and employs contextual-aware attention to extract essential video-related features. Conversely, the decoder incorporates a symmetric textual memory to capture external linguistic cues from dataset descriptions, using linguistic-aware cross-attention to gather relevant language features for each token at every time step. By leveraging visual and textual memories learned from the video-language dataset, the MAN enhances contextual understanding, thereby improving the quality and diversity of generated captions. While the performance of such memory-augmented models depends on the quality and diversity of the stored information, and maintaining the memory can be computationally intensive, their ability to produce high-quality and diverse captions makes them a promising direction in video captioning research.
Figure 6.
Overview of the proposed MAN architecture for video captioning (redrawn by the authors, adapted from [28]).
3.1.6. Transformer-Based Encoder-Decoder Models
Transformer models, initially designed for natural language processing, have recently gained traction in the field of video captioning due to their exceptional capacity to capture long-range dependencies. Yang et al. [29] pioneered the use of a Transformer-based model for video captioning, which outperformed traditional RNN-based approaches. Building on this success, Zhou et al. [30] further advanced the field by presenting an end-to-end Transformer model for dense video captioning. This model utilizes an encoder for video feature extraction and a decoder for caption generation, significantly enhancing efficiency through parallel processing.
More recently, the introduction of Vid2Seq by Yang et al. [31] has marked a new milestone in video captioning. As depicted in Figure 7, Vid2Seq is a pretrained Transformer-based encoder-decoder model specifically designed for narrated videos. It takes video frames and transcribed speech sequence as inputs, and outputs an event sequence , where each event includes a textual description and corresponding timestamps in the video. Despite its top performance across multiple benchmarks, it is important to note that Transformer models require substantial data for training, are computationally intensive, and often exhibit less interpretability than RNN-based models due to their reliance on self-attention mechanisms.
Figure 7.
Transformer-based Encoder-Decoder Models for video captioning (redrawn by the authors, adapted from [31]): This approach frames dense event captioning as a sequence-to-sequence task, utilizing special time tokens to enable the model to comprehend and produce token sequences containing both textual semantics and temporal localization details, thereby anchoring each sentence to the corresponding video segment.
3.2. Discussion-Encoder-Decoder Based APPROACHES
Encoder-decoder-based approaches for video captioning have significantly evolved, transitioning from simple CNN-RNN models to complex Transformer-based and modular frameworks. Attention mechanisms have been crucial for improving visual-textual alignment. Despite progress, challenges persist in generating diverse and accurate captions, capturing long-range dependencies, and incorporating external knowledge. Video captioning, vital for applications like video retrieval and human–computer interaction, has garnered attention.
3.3. Transformer-Based Approaches
3.3.1. Knowledge-Enhanced Transformer Models
Knowledge-enhanced transformer models are designed to integrate external knowledge into the video captioning process, aiming to enhance the diversity and accuracy of the generated captions. Gu et al. [32] introduced the Text with Knowledge Graph Augmented Transformer (TextKG) for this purpose. In TextKG, both the external and internal streams utilize the self-attention module to model interactions among multi-modality information. Specifically, the external stream focuses on modeling interactions between knowledge graphs and video content, while the internal stream exploits multi-modal information within videos. This approach allows the model to leverage external knowledge effectively.
Similarly, Zhang et al. [33] proposed a Center-enhanced Video Captioning model with Multimodal Semantic Alignment. This model integrates feature extraction and caption generation into a unified framework and introduces a cluster center enhancement module to capture key information within the multimodal representation. This facilitates more accurate caption generation. Both of these models, TextKG and the Center-enhanced Video Captioning model, effectively utilize external knowledge to overcome the limitations of relying solely on video-description pairs. By incorporating knowledge graphs and memory networks, they capture richer context, thereby enhancing the cognitive capability of the generated captions. However, the performance of these models is contingent upon the quality and coverage of the external knowledge sources they utilize.
3.3.2. Sparse Attention-Based Transformer Models
Sparse attention-based transformer models have proven effective for video captioning by focusing on informative regions, reducing redundancy and computational complexity. These models enhance efficiency and caption quality via sparse attention mechanisms. Notable examples include the Sparse Boundary-Aware Transformer (SBAT) by Jin et al. [34], which uses boundary-aware pooling and local correlation to capture features and outperform state-of-the-art methods, and the Universal Attention Transformer (UAT) by Im et al. [35], which employs a full transformer structure with ViT, FEGs, and UEA for competitive performance. As illustrated in Figure 8, the encoder-decoder structure with sparse boundary-aware attention is used for video captioning.
Figure 8.
The SBAT architecture (redrawn by the authors, adapted from [34]) integrates a sparse boundary-aware strategy (Sp) into both encoder and decoder multihead attention blocks.
However, these models have limitations. Performance hinges on the accuracy of sparse attention, and misidentification or omission of key details can degrade captions. Designing these mechanisms requires careful hyperparameter tuning, which is time-consuming and dataset-specific. Moreover, fast-moving objects or complex scenes pose challenges for selecting sparse, informative regions. Efficient models like SBAT and SnapCap aim to reduce complexity and improve speed, with SnapCap generating captions directly from the compressed video to avoid costly processing.
3.3.3. Hierarchical Transformer Models
Hierarchical transformer models are designed to capture the intricate hierarchical structure of videos, which encompasses scenes, actions, and objects at multiple levels of granularity. Ye et al. [36] proposed a Hierarchical Modular Network (HMN) specifically for video captioning. This network consists of three key modules: the entity module, which identifies principal objects within the video; the predicate (or action) module, which learns action representations; and the sentence module, which generates the final captions. As illustrated in Figure 9, the HMN adheres to the conventional encoder-decoder paradigm, with the HMN itself serving as the encoder.
Figure 9.
The proposed Hierarchical Modular Network (redrawn by the authors, adapted from [36]) functions as a robust video encoder, linking video representations to linguistic semantics through three levels using entity, predicate, and sentence modules. Each sentence module has its own input, which is extracted from captions, and linguistic guidance.
Similarly, Ging et al. [37] introduced COOT, another hierarchical model for video captioning. COOT includes components for capturing relationships between frames and words, producing clip and sentence features, and generating final video and paragraph embeddings. Both HMN and COOT decompose the video captioning task into sub-tasks, such as scene recognition, action recognition, and caption generation, and then combine the results of these sub-tasks to produce the final caption. By leveraging the hierarchical structure of videos, these models are able to generate captions that are more semantically meaningful and grammatically correct. However, the effective design of these modular architectures requires careful consideration to ensure seamless information flow across the different levels of hierarchy.
3.3.4. Single-Stream Transformer Models
To tackle the challenges of video captioning, single-stream transformer models have emerged, harnessing the potent self-attention mechanism inherent in transformers. These models strive to integrate both visual and textual information within a cohesive framework. Chen et al. [38] presented the Two-View Transformer (TVT) network, which employs a transformer encoder to capture temporal information and a fusion decoder to merge visual and textual modalities. TVT exhibits enhanced performance over RNN-based approaches, underscoring the efficiency and efficacy of transformers in the realm of video captioning.
Building on this, Lei et al. [39] introduced MART, which enhances the transformer architecture by incorporating an external memory module to facilitate sentence-level recurrence. This memory module aids in modeling the context of preceding video segments and sentences, thereby producing more coherent paragraph captions. In MART, notably, the encoder and decoder are shared, as illustrated in Figure 10 (left). Initially, video and text inputs are encoded and normalized separately. Figure 10 (right) showcases a modified version of Transformer-XL tailored for video paragraph captioning. Nonetheless, these models encounter difficulties with complex temporal dynamics and multi-modal interactions, and they demand substantial amounts of paired data, which is both costly and time-consuming to acquire.
Figure 10.
The figure illustrates the proposed Memory-Augmented Recurrent Transformer (MART) for video paragraph captioning (redrawn by the authors, adapted from [39]). On the right, the Transformer-XL model for the same task is depicted. Relative PE stands for Relative Positional Encoding, and SG(·) denotes stop-gradient [40].
3.3.5. Multi-Stream Transformer Models
To capture the diverse information in videos more effectively, multi-stream transformer models have been developed. Wang et al. introduced the Collaborative Three-Stream Transformers (COST) framework [41], which is illustrated in Figure 11. This framework comprises Video-Text, Detection-Text, and Action-Text branches, each tailored to model interactions between global video appearances, detected objects, and actions, respectively. By fusing these branches, COST generates comprehensive captions, and a cross-granularity attention module ensures feature alignment across the different streams. This design allows COST to capture intricate interactions between objects and their actions/relations across various modalities.
Figure 11.
The proposed COST method’s network architecture (redrawn by the authors, adapted from [41]) comprises three transformer branches: Action-Text, Video-Text, and Detection-Text. A cross-granularity attention module aligns the interactions modeled by these branches, with Y, H, and X representing their respective interactions.
In a different approach, Sun et al. [42] presented SnapCap, a method that bypasses the reconstruction process and generates captions directly from compressed video measurements. This method is guided by a teacher-student knowledge distillation framework using CLIP. While multi-stream transformer models, such as COST, significantly enhance video captioning performance by leveraging complementary semantic cues, they do come with higher computational costs due to their multiple streams. SnapCap, on the other hand, offers an alternative that may address some of these computational challenges while still aiming for accurate caption generation.
3.4. Discussion-Transformer Based APPROACHES
Transformer-based approaches have shown significant promise in video captioning. Basic Transformer models have demonstrated improved performance over RNN-based methods due to their ability to capture long-range dependencies. Enhanced Transformer models, such as VideoBERT, MART, and SBAT, further improve performance by incorporating additional mechanisms like memory augmentation, sparse attention, and boundary-aware pooling. Collaborative frameworks like COST and knowledge graph-augmented models provide additional context to generate more informative captions. Finally, efficient Transformer models like SnapCap and factorized autoregressive decoding enable video captioning in resource-constrained environments.
3.5. Attention-Based Approaches
3.5.1. Spatial Attention Models
Spatial attention models have become crucial in enhancing video caption generation by highlighting key regions within frames that are essential for understanding content. Their primary goal is to capture the most prominent areas, extracting high-resolution visual features necessary for accurate captioning. By complementing temporal attention frameworks, spatial attention mechanisms delve into the spatial details of each frame, expertly identifying significant objects and activities. This combination of spatial and temporal cues improves the overall comprehension of video content.
An early notable example is the Attention-Based Multimodal Fusion for Video Description (ABFVD) model by Hori et al. [20], which integrates image and motion attributes through a multimodal attention system. However, ABFVD’s reliance on basic feature combinations may limit its ability to adaptively assess feature relevance across different descriptive contexts. In contrast, Tu et al. [43] introduced the spatial-temporal attention (STAT) model, skillfully combining both spatial and temporal dimensions to focus on relevant regions and frames. As illustrated in Figure 12, the STAT model takes global motion features, local features, and model status information as input, producing dynamic visual representations that are fed into each LSTM decoder iteration. Nevertheless, a significant limitation of spatial attention models is their potential over-reliance on predefined object detection methods or region proposal networks, which may not consistently identify contextually relevant areas. Additionally, these models can be computationally intensive, especially with large and complex scenes, posing challenges for real-time implementation and resource management.
Figure 12.
The STAT unit (redrawn by the authors, adapted from [43]) is depicted, receiving local features, global-motion features, and model status as inputs. It generates dynamic visual features for LSTM decoder iterations. Layer 1 applies spatial attention to local features. Layer 2 focuses on temporal attention for both feature types. Layer 3 fuses two temporal representations.
3.5.2. Temporal Attention Models
Temporal attention models have emerged as a prominent approach in the field of video captioning, enabling dynamic focus on the most relevant frames or temporal segments within a video. Yao et al. [19] were pioneers in this domain, introducing a model that utilizes a soft attention mechanism. This mechanism assigns weights to frames based on their relevance to the current word being generated, allowing the model to selectively attend to various temporal regions for improved caption prediction.
Building on this foundation, Yan et al. [16] proposed the Spatial-Temporal Attention Mechanism (STAT) for video captioning. As illustrated in Figure 13, their framework integrates three types of features (CNN, C3D, and R-CNN) through a two-stage attention mechanism. Initially, given the current semantic context, a spatial attention mechanism guides the decoder to select local features with higher spatial attention weights, representing significant regions within frames. Subsequently, a temporal attention mechanism enables the decoder to select global and motion features, along with the previously identified local features. Ultimately, these three types of features are fused to represent the information of keyframes, enhancing the model’s ability to generate more detailed descriptions.
Figure 13.
The STAT video caption framework (redrawn by the authors, adapted from [16]) utilizes a spatial-temporal attention mechanism and comprises three main modules: (1) Feature extraction, (2) LSTM with spatial-temporal attention, and (3) caption generation.
3.5.3. Semantic Attention Models
Semantic attention models aim to integrate high-level semantic information into the captioning process, thereby generating more coherent and meaningful captions by understanding the video’s content. Ye et al. [36] proposed a hierarchical network for video captioning that uses three levels to link video representations with linguistic semantics, extracting semantic features at different levels to guide caption generation. Similarly, Gao et al. [44] introduced a unified framework (Figure 14, named aLSTMs, an attention-based Long Short-Term Memory model with semantic consistency. This framework takes the dynamic weighted sum of local spatio-temporal 2D CNN feature vectors and 3D CNN feature vectors as input for the LSTM decoder, integrating multi-word embedding and cross-view methodology to project generated words and visual features into a common space, bridging the semantic gap between videos and corresponding sentences.
Figure 14.
The proposed method, aLSTMs (redrawn by the authors, adapted from [44]), includes: (1) an LSTM visual encoder for video frame processing, (2) an attention-based LSTM for word generation with key feature focus, and (3) a cross-view module aligning sentences with visual content. For demonstration, words are color-coded to indicate their most relevant video frames.
Yousif et al. [45] presented a Semantic-Based Temporal Attention Network (STAN) for Arabic video captioning, utilizing a Semantic Representation Network (SRN) to extract semantic features and employing temporal attention to align keyframes with relevant semantic tags. Guo et al. [46] introduced an attention-based LSTM method focused on semantic consistency, combining attention with LSTM to capture significant video structures and explore multimodal representation correlations. Despite these advancements in improving semantic richness in captions, semantic attention models face challenges related to the availability and quality of semantic features and require sophisticated architectures and training strategies, increasing computational complexity and training time.
3.5.4. Multi-Modal Attention Models
To generate more comprehensive and accurate captions, multi-modal attention models aim to fuse information from various sources, such as visual, auditory, and textual data. Hori et al. [20] introduced an attention-based multimodal fusion model that incorporates visual, motion, and audio features for descriptive video content generation. Similarly, Ji et al. [47] proposed an Attention-based Dual Learning Approach for Video Captioning, which includes a caption generation and a video reconstruction module to bridge the semantic gap between videos and captions by jointly learning both modalities.
Lin et al. [48] presented a multi-modal feature fusion method with feature attention specifically for the VATEX captioning challenge. This method extracts motion, appearance, semantic, and audio features, utilizing a feature attention module to focus on different features during the decoding process. However, these multi-modal attention models often have higher computational complexity due to the integration of multiple modalities. Furthermore, acquiring high-quality multi-modal features, particularly for audio, can be challenging and may require specialized pre-processing and feature extraction techniques.
3.6. Discussion-Attention Based Models
Attention-based approaches have proven to be highly effective in Video Captioning, enabling models to focus on the most relevant visual and temporal information within videos. From traditional Encoder-Decoder frameworks with single-head attention to advanced Transformer models with multi-head attention, the field has witnessed significant progress. Furthermore, the integration of semantic information and object relationships has further enriched the generated captions, pushing the state-of-the-art in this exciting domain. As we continue to explore more sophisticated attention mechanisms and their applications, the potential for creating truly intelligent video description systems remains vast and promising.
4. Datasets for Dense Video Captioning
Dense video captioning is a task that involves generating natural language descriptions for multiple events occurring in a video, and it heavily relies on the availability of well-annotated datasets. This section provides an overview of commonly used datasets for dense video captioning, followed by detailed descriptions, characteristics, limitations, and the data preprocessing and augmentation techniques associated with each dataset. Table 1 presents a comparison of key attributes among various datasets, highlighting benchmark ones that have been utilized in recent video description research [1]. Additionally, it summarizes comparisons involving video caption datasets and 3D gaze-related datasets.
Table 1.
Comparison of video caption datasets across different domains.
4.1. Detailed Descriptions
- (1)
- MSVDMSVD [49] (Microsoft Video Description) is a pioneering and widely adopted dataset for video captioning tasks. It features short video clips accompanied by descriptive sentences. Each video clip in MSVD is annotated with a single sentence description, providing a concise summary of the clip’s content. The MSVD dataset comprises over 1970 video clips, offering a substantial collection for research and development in video captioning.
- (2)
- TACoSTACoS [50] (Textually Annotated Cooking Scenes) is a multimodal corpus containing high-quality videos of cooking activities aligned with multiple natural language descriptions. This facilitates grounding action descriptions in visual information. The corpus provides sentence-level alignments between textual descriptions and video segments, alongside annotations indicating similarity between action descriptions. TACoS comprises 127 videos with 2540 textual descriptions, resulting in 2206 aligned descriptions after filtering. It contains 146,771 words and 11,796 unique sentences.
- (3)
- MPII-MDMPII-MD [51] is a dataset sourced from movies for video description. It aims to facilitate research on generating multi-sentence descriptions for videos. Each video is annotated with multiple temporally localized sentences that cover key events and details. The dataset contains 94 movies, amounting to 68,000 clips with 68.300 multi-sentence descriptions.
- (4)
- TGIFTGIF [52] (Tumblr GIF) is a large-scale dataset designed for animated GIF description, containing 100,000 GIFs collected from Tumblr and 120,000 natural language descriptions sourced via crowdsourcing. Each GIF is accompanied by multiple sentences to capture its visual content. High-quality annotations are ensured through syntactic and semantic validation, ensuring strong visual-textual associations. Workers were selected based on language proficiency and performance. The dataset is split into 90K training, 10K validation, and 10K test GIFs, providing ample data for model training and evaluation.
- (5)
- MSR-VTTMSR-VTT [7] (Microsoft Research Video to Text) is a diverse video dataset tailored for various video-to-text generation tasks, including video captioning and dense video captioning. It spans 20 broad categories, offering a rich variety of video content. Annotations for MSR-VTT are provided by 1327 AMT workers, ensuring each video clip is accompanied by multiple captions that offer comprehensive coverage of various aspects of the video content, enhancing the understanding of video semantics. MSR-VTT comprises 10,000 video clips sourced from 7180 videos, with an average of 20 captions per video.
- (6)
- CharadesCharades is a large-scale video dataset designed for activity understanding and video description tasks [53]. It contains videos of people performing daily activities in various indoor scenes. The annotations for Charades include action labels, temporal intervals, and natural language descriptions, collected through Amazon Mechanical Turk to provide comprehensive coverage of video content. Charades comprises 9848 videos, recorded by 267 people in 15 different indoor scenes. Each video is annotated with multiple action classes and descriptions.
- (7)
- VTWVTW (Video Titles in the Wild) is a large-scale dataset of user-generated videos automatically crawled from online communities [54]. It contains 18,100 videos with an average duration of 1.5 min, each associated with a concise title produced by an editor. Each video is annotated with a single title sentence describing the most salient event, accompanied by 1–3 longer description sentences. Highlight moments corresponding to the titles are manually labeled in a subset. The dataset spans 213.2 h in total, making it a challenging benchmark for video title generation tasks.
- (8)
- ActivityNet CaptionsActivityNet Captions [2] is a large-scale dataset for dense video captioning, featuring over 20,000 videos segmented into multiple temporal intervals. With 100K descriptions totaling 849 h, each averaging 13.48 words and capturing 36 s of video content, it comprehensively annotates activities within each interval. Annotations in the dataset include multiple temporally localized sentences per video, averaging 3.65 per clip, which collectively cover 94.6% of video content with 10% overlap, ideal for studying concurrent events. Covering over 849 video hours, the dataset encompasses 20K videos and generates 100k sentences. Both sentence length and count per video follow normal distributions, indicative of varying video durations.
- (9)
- YouCook2YouCook2 [4] is a comprehensive cooking video dataset, featuring 2000 YouTube videos across 89 recipes from diverse cuisines. It covers a wide range of cooking styles, ingredients, and utensils, primarily used for video summarization, recipe generation, and dense video captioning tasks. YouCook2 includes temporal boundaries of cooking steps and ingredients, along with textual descriptions for each step. These annotations provide detailed, step-by-step insights into the cooking process. Comprising 2000 cooking videos, YouCook2 offers a substantial dataset for research and development in cooking-related video analysis.
- (10)
- BDD-XThe BDD-X [55] dataset is a large-scale video dataset designed for explaining self-driving vehicle behaviors. It contains over 77 h of driving videos with human-annotated descriptions and explanations of vehicle actions. Each video clip is annotated with action descriptions and justifications. Annotators view the videos from a driving instructor’s perspective, describing what the driver is doing and why. The dataset comprises 6984 videos, split into training (5588), validation (698), and test sets (698). Over 26,000 actions are annotated across 8.4 million frames.
- (11)
- VideoStoryVideoStory [56] is a novel dataset tailored for multi-sentence video description. It comprises 20k videos sourced from social media, spanning diverse topics and engaging narratives. Each video is annotated with multiple paragraphs, containing 123k sentences temporally aligned to video segments. On average, each video has 4.67 annotated sentences. The dataset encompasses 396 h of video content, divided into training (17,098 videos), validation (999 videos), test (1011 videos), and blind test sets (1039 videos).
- (12)
- M-VADM-VAD [57] (Montreal Video Annotation Dataset) is a large-scale movie description dataset consisting of video clips from various movies. It aims to support research on video captioning and related tasks. Each video clip is paired with a textual description. Annotations include face tracks associated with characters mentioned in the captions. The dataset contains over 92 movies, resulting in a substantial number of video clips and textual descriptions, providing rich material for video captioning research.
- (13)
- VATEXVATEX [58] (Video and Text) is a large-scale dataset comprising 41,269 video clips, each approximately 10 s long, sourced from the Kinetics-600 dataset. Each video clip in VATEX is annotated with 10 English and 10 Chinese descriptions, providing rich linguistic information across two languages. VATEX boasts a total of 41,269 video clips, making it a substantial resource for video description tasks.
- (14)
- TVCTVC [59] (TV Show Caption) is a large-scale multimodal video caption dataset built upon the TVR dataset. It requires systems to gather information from both video and subtitles to generate relevant descriptions. Each annotated moment in TVR has additional descriptions collected to form the TVC dataset, totaling 262,000 descriptions for 108 [53] moments. Descriptions may focus on video, subtitles, or both. The dataset includes 174,350 training descriptions, 43,580 validation descriptions, and 21,780 public test descriptions, split across numerous videos and moments.
- (15)
- ViTTThe ViTT [21] (Video Timeline Tags) dataset comprises 8000 untrimmed instructional videos, tailored for tasks that involve video content analysis and tagging. It aims to address the uniformity issue in YouCook2 videos by sampling videos with cooking/recipe labels from YouTube-8M. Each video in ViTT is annotated with 7.1 temporally-localized short tags, offering detailed insights into various aspects of the video content. Annotators have identified each step and assigned descriptive yet concise tags. The ViTT dataset boasts 8000 videos, with an average duration of 250 s per video. It contains a vast array of unique tags and token types, making it a comprehensive resource for video analysis tasks.
- (16)
- VC_NBA_2022VC_NBA_2022 [60] is a specialized basketball video dataset designed for knowledge-guided entity-aware video captioning tasks. This dataset is intended to support the generation of text descriptions that include specific entity names and fine-grained actions, particularly tailored for basketball live text broadcasts. VC_NBA_2022 provides annotations beyond conventional video captions. It leverages a multimodal basketball game knowledge graph (KG_NBA_2022) to offer additional context and knowledge, such as player images and names. Each video segment is annotated with detailed captions covering 9 types of fine-grained shooting events and incorporating player-related information from the knowledge graph. The dataset contains 3977 basketball game videos, with 3162 clips used for training and 786 clips for testing. Each video has one text description and associated candidate player information (images and names). The dataset is constructed to focus on shot and rebound events, which are the most prevalent in basketball games.
- (17)
- WTSWTS [61] (Woven Traffic Safety Dataset) is a specialized video dataset designed for fine-grained spatial-temporal understanding tasks in traffic scenarios, particularly focusing on pedestrian-centric events. Comprehensive textual descriptions and unique 3D Gaze data are provided for each video event, capturing detailed behaviors of both vehicles and pedestrians across diverse traffic situations. The dataset contains over 1.2k video events, with rich annotations that enable in-depth analysis and understanding of traffic dynamics.
4.2. Characteristics and Limitations of Each Dataset
- (1)
- MSVDMSVD is an early and well-established dataset that features short video clips accompanied by concise descriptions. However, its size and diversity are somewhat limited, and the single sentence descriptions may not fully capture all aspects of the video content.
- (2)
- TACoSTACoS is a multimodal corpus that offers video-text alignments along with diverse cooking actions and fine-grained sentence-level annotations. Nevertheless, it is restricted to the cooking domain and thus may lack the diversity needed for broader actions.
- (3)
- MPII-MDMPII-MD centers on movie descriptions and contains 94 videos paired with 68k sentences, providing multi-sentence annotations aligned with video clips. But this dataset is confined to movie-related content and lacks the diversity of topics typically found in social media videos.
- (4)
- TGIFTGIF consists of 100K animated GIFs coupled with 120K natural language descriptions, showcasing rich motion information and cohesive visual stories. Yet, due to the nature of animated GIFs, it lacks audio and long-form narratives, which limits its complexity compared to full videos.
- (5)
- MSR-VTTMSR-VTT boasts diverse video content spanning various topics, with multiple captions per video segment to offer a comprehensive understanding. However, the annotations may occasionally be redundant or overlapping, and the quality of captions can vary, potentially affecting model training.
- (6)
- CharadesCharades, comprising 9,848 annotated videos of daily activities involving 267 people, covers 157 action classes and facilitates research in action recognition and video description. Nonetheless, the dataset’s focus on home environments restricts diversity in backgrounds and activities, possibly impacting generalization.
- (7)
- VTWThe VTW dataset holds 18,100 automatically crawled user-generated videos with diverse titles, presenting a challenging benchmark for video title generation. However, the temporal location and extent of highlights in most videos are unknown, posing challenges for highlight-sensitive methods.
- (8)
- ActivityNet CaptionsActivityNet Captions is a large-scale dataset characterized by diverse video content, fine-grained temporal annotations, and rich textual descriptions. The complexity of the videos might make it difficult for models to accurately capture all details, and some annotations may exhibit varying levels of granularity.
- (9)
- YouCook2YouCook2 zeroes in on cooking activities, making it highly domain-specific with detailed step-by-step annotations. Compared to general-purpose datasets, it has limited diversity in video content, and its annotations are biased towards cooking terminology.
- (10)
- BDD-XBDD-X contains over 77 h of driving videos with 26k human-annotated descriptions and explanations, covering diverse driving scenarios and conditions. However, it is important to note that the annotations are rationalizations from observers rather than capturing drivers’ internal thought processes.
- (11)
- VideoStoryVideoStory comprises 20k social media videos paired with 123k multi-sentence descriptions, spanning a variety of topics and demonstrating temporal consistency. But there’s a possibility that its representativeness might be limited as it tends to focus on engaging content when it comes to all social media videos.
- (12)
- M-VADThe M-VAD dataset holds 92 movies with detailed descriptions that cover different genres. It zeroes in on captions for movie clips, which makes it quite suitable for video captioning research. Nevertheless, it lacks character visual annotations as well as proper names in the captions, thereby restricting its application in tasks that require naming capabilities.
- (13)
- VATEXVATEX shines in its focus on a broad range of video activities, rendering it versatile for various video understanding tasks. It offers detailed and diverse annotations that touch upon multiple aspects of the video content. Yet, despite being extensive, it may not be deeply specialized in one particular domain like some other datasets. Moreover, while comprehensive, its annotations might not be finely tuned to the specific terminology or nuances of highly specialized fields.
- (14)
- TVCTVC stands out as a large-scale dataset equipped with multimodal descriptions and embraces diverse context types such as video, subtitle, or both, along with rich human interactions. However, the annotations within this dataset may vary in their level of detail, and some descriptions could be more centered around one modality.
- (15)
- ViTTViTT centers on text-to-video retrieval tasks, making it immensely valuable for video-text understanding and alignment. It provides detailed annotations that link textual descriptions to corresponding video segments. Still, when compared to general-purpose video datasets, its focus is narrower, concentrating on specific types of videos pertinent to the text-to-video retrieval task. Additionally, the annotations might lean towards the terminology and context commonly seen in text-video pairing scenarios.
- (16)
- VC_NBA_2022The VC_NBA_2022 dataset presents a specialized collection of basketball game videos, covering a multitude of in-game events. It delivers detailed and specific annotations for shooting incidents along with player-centric details, offering an exhaustive comprehension of basketball game situations. Nonetheless, the dataset’s annotations could occasionally be skewed towards certain aspects of the game, perhaps overlooking other vital elements. Furthermore, the quality and consistency of these annotations can fluctuate, which might influence the efficacy of model training.
- (17)
- WTSWTS has a distinct focus on traffic safety scenarios, making it highly pertinent to studies involving pedestrian and vehicle interactions. It includes thorough annotations comprising textual descriptions and 3D gaze data for in-depth analysis. Although, in contrast to general-purpose video datasets, it has a more confined scope concerning video content as it concentrates solely on traffic safety circumstances. Furthermore, its annotations may exhibit bias towards traffic-related terminology and specific facets of pedestrian and vehicle behavior.
4.3. Data Preprocessing and Augmentation Techniques
Effective data preprocessing and augmentation play a pivotal role in enhancing the performance of dense video captioning models. When dealing with the datasets mentioned, several common techniques are employed to optimize the data for training these models.
Firstly, temporal segmentation is applied to videos. This involves dividing them into meaningful clips based on temporal annotations [62,63,64]. By focusing on relevant parts of the video, this technique helps in improving the accuracy and relevance of the captions generated by the model [65,66,67].
Next, frame sampling is utilized to extract keyframes from each video segment. This step is crucial for reducing computational load while capturing essential visual information [4,7,68]. By selecting representative frames, the model can process the video more efficiently without losing important details [66,69,70,71].
In addition to visual data preprocessing, text preprocessing is also essential. Captions are cleaned by removing punctuation and stop words, and undergoing tokenization and lemmatization [64,67]. This process ensures that the captions are in a standardized format, making it easier for the model to understand and generate accurate descriptions [64,67].
To further enhance the diversity and robustness of the training data, data augmentation techniques are employed. One such technique is temporal shifting, where the boundaries of temporal segments are slightly shifted [72]. This helps create more robust models by exposing them to variations in the temporal boundaries of events.
Another augmentation technique is caption paraphrasing. By generating paraphrased versions of captions, the diversity of the training data is increased [72,73]. This encourages the model to learn different ways of describing the same event, improving its ability to generate varied and accurate captions [74].
Lastly, synthetic noise is added to video frames or captions. This technique enhances the model’s robustness by exposing it to noisy data during training [71]. By learning to handle synthetic noise, the model becomes better equipped to deal with real-world noise and variations, leading to improved performance in dense video captioning tasks.
5. Evaluation Protocols and Metrics
Assessing video descriptions, whether automatically or manually generated, remains challenging due to the absence of a definitive “correct answer” or ground truth for evaluating accuracy. A single video can be accurately depicted using a range of sentences that display both syntactic and semantic variety. For instance, consider the various captions for vid1447 from the MSVD dataset, as illustrated in Figure 15. One caption states, “Two men are rolling huge tires sideways down a street while spectators watch”, offering one viewpoint. Another says, “Men are flipping large tires in a race”, presenting a different angle. Yet another description notes, “Two men are pushing large tires”, emphasizing a distinct aspect of the action. Variations also emerge in how strength and activity are portrayed, such as “A strong person pulling a big tire”, or the more concise version, “Two men are rolling tires”. Additionally, some captions incorporate extra context, like “A short clip about Jeff Holder, the strongest man in Southern California, who is moving the tire and proving his strength”. Each of these captions captures the essence of the video in a unique yet equally valid way, with differing focuses, levels of detail, and depictions of the actions and participants.
Figure 15.
An instance from the MSVD dataset is provided along with its corresponding ground truth captions. Observe the varied descriptions of the same video clip. Each caption presents the activity either wholly or partially in a distinct manner.
5.1. Key Metrics
Video description merges computer vision (CV) with natural language processing (NLP). BLEU [8], METEOR [9], CIDEr [10], ROUGE-L [11], SODA [12], SPICE [13], and WMD [14] serve as crucial metrics for evaluating auto-generated captions for videos and images. Derived from NLP, these metrics assess various aspects like n-gram overlap, semantic similarity, and uniqueness, providing rapid and cost-effective evaluations that complement human assessments of adequacy, fidelity, and fluency. Each metric has unique strengths and aims to align with professional translation quality. Table 2 summarizes the metric names, purposes, methodologies, and limitations, with detailed descriptions following below.
Table 2.
Summary of Metrics Used for Video Description Evaluation.
5.1.1. BLEU
BLEU (Bilingual Evaluation Understudy) is a metric commonly used in machine translation and has been adapted for video captioning evaluation [8]. It measures the n-gram precision between the generated caption and the reference captions. BLEU scores range from 0 to 1, with higher scores indicating better performance. However, BLEU has limitations in capturing the semantic similarity between captions and may penalize models for generating grammatically correct but semantically different captions from the references.
The core idea behind BLEU is to compute the precision of n-grams (sequences of n consecutive words) between the candidate text and the reference texts. BLEU calculates a weighted geometric mean of the n-gram precisions for different values of n, typically up to 4. Additionally, BLEU incorporates a brevity penalty to penalize candidate texts that are too short. The BLEU score is calculated using the following formula [8]:
where is the brevity penalty, N is the maximum n-gram length (typically 4), is the weight for n-gram precision, typically set to for uniform weighting; is the modified n-gram precision for length n.
The brevity penalty (BP) is defined as:
where c is the length of the candidate text, r is the effective reference length, calculated as the closest reference length.
5.1.2. METEOR
METEOR (Metric for Evaluation of Translation with Explicit Ordering) is another metric that combines Precision (P) and Recall (R), taking into account both exact word matches and paraphrases [9]. It uses a synonym dictionary and stems words to improve the evaluation of semantic similarity. METEOR is considered more robust than BLEU in handling synonyms and different word orderings. Moreover, it explicitly takes into account the word order in the translated sentence, which is vital for capturing the fluency and grammatical correctness of the translation.
The computation of the METEOR score involves five steps:
- (1)
- Word AlignmentThe initial step involves aligning words between the machine-generated and reference translations using exact matching, stemming (Porter stemmer), and synonymy matching (WordNet). This alignment is carried out in stages, each focusing on a different matching criterion.
- (2)
- Precision and Recall CalculationOnce aligned METEOR calculates Precision (P) as the ratio of correctly matched unigrams in the translation to the total unigrams, and Recall (R) as the ratio of correctly matched unigrams to the total in the reference.
- (3)
- F-measure CalculationMETEOR then computes the harmonic mean (F-measure) of Precision (P) and Recall (R), with more weight given to Recall (R).
- (4)
- Penalty for FragmentationTo penalize translations with poor ordering or gaps, METEOR calculates a fragmentation Penalty based on the number of chunks formed by adjacent matched unigrams in the translation and reference.where and are parameters that control the Penalty.
- (5)
- Final METEOR Score
5.1.3. CIDEr
CIDEr (Consensus-based Image Description Evaluation) is tailored for assessing image and video caption [10]. It gauges the level of agreement among multiple reference captions by calculating the cosine similarity between the TF-IDF vectors of the candidate caption and those of the references. This metric excels at capturing the relevance and salience of the generated captions, embodying the idea that a high-quality caption should resemble how most humans would describe the image.
CIDEr evaluates captions by: matching n-grams (1−4 words) for grammatical and semantic richness, applying TF-IDF weighting to downplay common n-grams, computing average cosine similarity for each n-gram length to measure overlap, and combining n-gram scores with uniform weights to obtain the final score. The metric’s formula is as follows:
- (1)
- TF-IDF Weightingwhere is the frequency of n-gram in reference caption , is the vocabulary of all n-grams, is the total number of images, and q is a constant typically set to the average length of reference captions.
- (2)
- Cosine Similarity for n-grams of Length nwhere is the candidate caption, is the set of reference captions, and and are the TF-IDF vectors for the n-grams of length n in the candidate and reference captions, respectively.
- (3)
- Combined CIDEr Scorewhere N is the maximum n-gram length (typically 4), and are the weights for different n-gram lengths (typically set to for uniform weights).
5.1.4. ROUGE-L
ROUGE (Recall-Oriented Understudy for Gisting Evaluation) is a collection of metrics designed to assess automatic summarization, and it has been adapted for evaluating video captioning as well [11]. Specifically, ROUGE-L measures the longest common subsequence (LCS) between a generated caption and reference captions, with a particular focus on the recall of content. This metric is valuable for gauging the completeness of the generated captions. ROUGE-L offers both Recall () and Precision () scores, and the overall ROUGE-L score is typically reported as their F-measure ().
The computation of ROUGE-L involves three steps:
- (1)
- LCS CalculationFor two sequences X and Y, let denote the length of their longest common subsequence.
- (2)
- Recall and PrecisionRecall () is determined by the ratio of the length of the longest common subsequence (LCS) to the length of the reference sequence. Precision (), on the other hand, is calculated as the ratio of the length of the LCS to the length of the candidate sequence.
- (3)
- F-MeasureThe F-measure () is the harmonic mean of Recall () and Precision ().where is a parameter that controls the relative importance of Recall and Precision. In ROUGE evaluations, is often set to a very large value (e.g., ), effectively making equal to .
5.1.5. SODA
SODA (Story Oriented Dense Video Captioning Evaluation Framework) is a metric specifically designed to evaluate video story descriptions [12]. It takes into account the temporal ordering of events and penalizes redundant captions to ensure accuracy. SODA assesses the quality and relevance of generated captions by maximizing the sum of Intersection over Union (IoU) scores and computing an F-measure based on METEOR scores. The framework comprises two main components: optimal matching using dynamic programming, and an F-measure specifically tailored for evaluating video story descriptions.
- (1)
- Optimal Matching Using Dynamic ProgrammingSODA uses dynamic programming to find the best match between generated and reference captions, maximizing their IoU score sum while considering temporal order. The IoU between a generated caption p and a reference caption g is defined as:where and represent the start and end times of the event proposals corresponding to the captions.Given a set of reference captions G and generated captions P, SODA finds the optimal matching by solving:where S holds the maximum score of optimal matching between the first i generated captions and the first j reference captions, and C is the cost matrix based on IoU scores.
- (2)
- F-measure for Evaluating Video Story DescriptionTo penalize redundant captions and ensure the generated captions cover all relevant events, SODA computes Precision and Recall based on the optimal matching and derives F-measure scores:where represents an evaluation metric like METEOR.
5.1.6. SPICE
SPICE (Semantic Propositional Image Caption Evaluation) is a metric for assessing image caption quality by analyzing semantic content, utilizing a graph-based semantic representation called a scene graph to encode objects, attributes, and relationships present in captions, thus abstracting away lexical and syntactic idiosyncrasies of natural language. It calculates an F-score over logical tuples to reflect semantic propositional similarity, providing a more human-like evaluation.
To generate a scene graph, SPICE employs a two-stage approach: first, a dependency parser establishes syntactic dependencies between words in the caption, resulting in a dependency syntax tree; second, this tree is mapped to a scene graph using a rule-based system, consisting of object nodes, attribute nodes, and relationship edges [13].
SPICE computes a similarity score between a candidate caption c and reference captions by extracting logical tuples from their parsed scene graphs using a function T:
where is the scene graph for caption c, is the set of object nodes, is the set of relationship edges, and is the set of attribute nodes. Each tuple contains one, two, or three elements, representing objects, relationships, and attributes, respectively.
SPICE then computes precision P(c, S), recall R(c, S), and F-score based on the matching tuples between the candidate scene graph and the reference scene graphs:
where ⊗ denotes the tuple matching operator, which returns the set of matching tuples between two scene graphs. For matching tuples, SPICE uses WordNet synonym matching, similar to METEOR.
5.1.7. WMD
WMD (Word Mover’s Distance) is a metric for evaluating text document similarity [14], based on the Earth Mover’s Distance. It measures the minimum distance to move embedded words from one document to match those of another using word2vec, capturing semantic similarity for tasks like document retrieval and classification. Given two documents represented as normalized bag-of-words (nBOW) histograms d and , WMD calculates the minimum travel distance between their embedded words.
Formally, let X and be the sets of embedded word vectors for documents d and , respectively. The WMD is defined as the solution to the following optimization problem:
subject to:
here, T is a flow matrix where represents the amount of “mass” moved from word i in document d to word j in document . The cost is the Euclidean distance between the embedded word vectors and :
The constraints ensure that the entire “mass” of each word in the source document is moved to the target document, and vice versa.
5.2. Challenges in Evaluation and Benchmarking
Despite the advancements in evaluation metrics, there are still challenges in evaluating and benchmarking DVC models.
- (1)
- Diversity of Video ContentVideos contain a wide range of events, contexts, and interactions, making it difficult to create a comprehensive set of reference captions that capture all possible descriptions. This diversity can lead to inconsistencies in evaluation results.
- (2)
- Semantic SimilarityCurrent metrics rely heavily on n-gram matching and may not accurately capture the semantic similarity between captions. This can result in models being penalized for generating captions that are semantically correct but differ in wording from the references.
- (3)
- Temporal LocalizationDVC involves not only generating accurate captions but also locating the events in the video. Evaluating the temporal localization of events is challenging, as it requires precise annotations and may not be fully captured by existing metrics.
5.3. Discussion on the Interpretability and Fairness of Evaluation Metrics
The interpretability and fairness of evaluation metrics are crucial for the development of DVC models.
- (1)
- InterpretabilityMetrics should provide clear insights into the strengths and weaknesses of the models. For example, a low BLEU score may indicate a lack of n-gram overlap, while a high CIDEr score may suggest good consensus with reference captions. However, it is important to interpret these scores in the context of the specific task and dataset.
- (2)
- FairnessEvaluation metrics should be fair and unbiased, reflecting the true performance of the models. This requires careful selection of reference captions and consideration of the diversity of video content. Metrics should also be robust to variations in annotation quality and consistency.
To address these challenges, future work should focus on developing more sophisticated evaluation metrics that can capture the semantic similarity between captions and account for the diversity of video content. Additionally, there is a need for larger and more diverse datasets with high-quality annotations to support the evaluation and benchmarking of DVC models.
5.4. Benchmark Results
We present an overview of the benchmark results achieved by various methods across multiple video description datasets, including MSVD [49], TACoS [50], MPII-MD [51], MPII-MD [51], TGIF [52], Youtube2text [75], MSR-VTT [7], Charades [53], VTW [54], ActivityNet Captions [2], YouCook2 [4], BDD-X [55], VideoStory [56], M-VAD [57], VATEX [58], TVC [59], ViTT [21], VC_NBA_2022 [60], and WTS [61]. The methods are evaluated using a range of automatic metrics: BLEU [8], METEOR [9], CIDEr [10], ROUGE-L [11], SODA [12], SPICE [13], and WMD [14].
5.4.1. MSVD Dataset: Method Performance Overview
On the MSVD Dataset, various methods have been evaluated using the BLEU, METEOR, CIDEr, and ROUGE-L metrics. As shown in Table 3, several methods have reported results for this dataset, with specific models achieving notable performances. Here, we summarize the best-performing methods, along with their published scores.
Encoder-decoder based approaches: Encoder-decoder architectures have shown promising results on the MSVD dataset. Among the recent methods, Howtocaption [76] stands out with the highest BLEU (70.4), METEOR (46.4), CIDEr (154.2), and ROUGE-L (83.2) scores. This method leverages a combination of advanced encoding and decoding techniques to generate high-quality captions. Other notable methods include KG-VCN [77] and ATMNet [78], which also report competitive scores across all metrics. Earlier works, such as VNS-GRU [79] and RMN [80], also employed encoder-decoder frameworks but achieved lower scores compared to more recent methods. This highlights the progress made in improving video description techniques over the years.
Transformer-based approaches: Transformer-based methods have gained popularity in recent years due to their ability to capture long-range dependencies in sequences. CEMSA [33] is a notable transformer-based approach that reports competitive scores on the MSVD dataset, with BLEU (60.9), METEOR (40.5), CIDEr (117.9), and ROUGE-L (77.9). Other transformer-based methods, such as COST [41] and TextKG [32], also show promising results. However, some transformer-based methods, like SnapCap [42] and UAT-FEGs [35], report lower scores compared to encoder-decoder based approaches. This suggests that while transformers are powerful architectures, their effectiveness in video description tasks may depend on the specific implementation and design choices.
Table 3.
Comparison of video captioning methods on the MSVD dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
Table 3.
Comparison of video captioning methods on the MSVD dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
| NO | Method | Method’s Category | BLEU-4 | METEOR | CIDEr | ROUGE-L | Year |
|---|---|---|---|---|---|---|---|
| 1 | ASGNet [81] | Encoder-decoder | 55.2 | 36.6 | 101.8 | 74.3 | 2025 |
| 2 | CroCaps [82] | Encoder-decoder | 58.2 | 39.79 | 112.29 | 77.41 | 2025 |
| 3 | KG-VCN [77] | Encoder-decoder | 64.9 | 39.7 | 107.1 | 77.2 | 2025 |
| 4 | PKG [83] | Encoder-decoder | 60.1 | 39.3 | 107.2 | 76.2 | 2025 |
| 5 | ATMNet [78] | Encoder-decoder | 58.8 | 41.1 | 121.9 | 78.2 | 2025 |
| 6 | Howtocaption [76] | Encoder-decoder | 70.4 | 46.4 | 154.2 | 83.2 | 2024 |
| 7 | MAN [28] | Encoder-decoder | 59.7 | 37.3 | 101.5 | 74.3 | 2024 |
| 8 | GSEN [84] | Encoder-decoder | 58.8 | 37.6 | 102.5 | 75.2 | 2024 |
| 9 | EDS [85] | Encoder-decoder | 59.6 | 39.5 | 110.2 | 76.2 | 2024 |
| 10 | SATLF [86] | Encoder-decoder | 60.9 | 40.8 | 110.9 | 77.5 | 2024 |
| 11 | GLG [87] | Encoder-decoder | 63.6 | 42.4 | 107.8 | 80.8 | 2024 |
| 12 | CMGNet [88] | Encoder-decoder | 54.2 | 36.9 | 96.2 | 74.5 | 2024 |
| 13 | ViT/L14 [89] | Encoder-decoder | 60.1 | 41.4 | 121.5 | 78.2 | 2023 |
| 14 | CARE [90] | Encoder-decoder | 56.3 | 39.11 | 106.9 | 75.6 | 2023 |
| 15 | VCRN [91] | Encoder-decoder | 59.1 | 37.4 | 100.8 | 74.6 | 2023 |
| 16 | RSFD [92] | Encoder-decoder | 51.2 | 35.7 | 96.7 | 72.9 | 2023 |
| 17 | vc-HRNAT [26] | Encoder-decoder | 57.7 | 36.8 | 98.1 | 74.1 | 2022 |
| 18 | SemSynAN [93] | Encoder-decoder | 64.4 | 41.9 | 111.5 | 79.5 | 2021 |
| 19 | TTA [94] | Encoder-decoder | 51.8 | 35.5 | 87.7 | 72.4 | 2021 |
| 20 | AVSSN [93] | Encoder-decoder | 62.3 | 39.2 | 107.7 | 76.8 | 2021 |
| 21 | RMN [80] | Encoder-decoder | 54.6 | 36.5 | 94.4 | 73.4 | 2020 |
| 22 | SAAT [95] | Encoder-decoder | 46.5 | 33.5 | 81.0 | 69.4 | 2020 |
| 23 | VNS-GRU [79] | Encoder-decoder | 66.5 | 42.1 | 121.5 | 79.7 | 2020 |
| 24 | JSRL-VCT [96] | Encoder-decoder | 52.8 | 36.1 | 87.8 | 71.8 | 2019 |
| 25 | GFN-POS [27] | Encoder-decoder | 53.9 | 34.9 | 91.0 | 72.1 | 2019 |
| 26 | DCM [97] | Encoder-decoder | 53.3 | 35.6 | 83.1 | 71.2 | 2019 |
| 27 | CEMSA [33] | Transformer | 60.9 | 40.5 | 117.9 | 77.9 | 2024 |
| 28 | SnapCap [42] | Transformer | 51.7 | 36.5 | 94.7 | 73.5 | 2024 |
| 29 | COST [41] | Transformer | 56.3 | 37.2 | 99.2 | 74.3 | 2023 |
| 30 | TextKG [32] | Transformer | 60.8 | 38.5 | 105.2 | 75.1 | 2023 |
| 31 | UAT-FEGs [35] | Transformer | 56.5 | 36.4 | 92.8 | 72.8 | 2022 |
| 32 | HMN [36] | Transformer | 59.2 | 37.7 | 104.0 | 75.1 | 2022 |
| 33 | O2NA [98] | Transformer | 55.4 | 37.4 | 96.4 | 74.5 | 2021 |
| 34 | STGN [99] | Transformer | 52.2 | 36.9 | 93.0 | 73.9 | 2020 |
| 35 | SBAT [34] | Transformer | 53.1 | 35.3 | 89.5 | 72.3 | 2020 |
| 36 | TVT [38] | Transformer | 53.21 | 35.23 | 86.76 | 72.01 | 2018 |
| 37 | STAN [45] | Attention | 40.0 | 40.2 | 72.1 | 55.2 | 2025 |
| 38 | IVRC [100] | Attention | 58.8 | 40.3 | 116.0 | 77.4 | 2024 |
| 39 | MesNet [101] | Attention | 54.0 | 36.3 | 100.1 | 73.6 | 2023 |
| 40 | ADL [47] | Attention | 54.1 | 35.7 | 81.6 | 70.4 | 2022 |
| 41 | STAG [102] | Attention | 58.6 | 37.1 | 91.5 | 73.0 | 2022 |
| 42 | SemSynAN [103] | Attention | 64.4 | 41.9 | 111.5 | 79.5 | 2021 |
| 43 | T-DL [104] | Attention | 55.1 | 36.4 | 85.7 | 72.2 | 2021 |
| 44 | SGN [105] | Attention | 52.8 | 35.5 | 94.3 | 72.9 | 2021 |
| 45 | ORG-TRL [106] | Attention | 54.3 | 36.4 | 95.2 | 73.9 | 2020 |
| 46 | MAM-RNN [107] | Attention | 41.3 | 32.2 | 53.9 | 68.8 | 2017 |
Attention-based approaches: Attention-based methods have been widely used in video description to focus on relevant parts of the video when generating captions. However, the performance of attention-based methods on the MSVD dataset varies widely. SemSynAN [103] reports relatively high scores, with BLEU (64.4), METEOR (41.9), CIDEr (111.5), and ROUGE-L (79.5). This method leverages semantic and syntactic information to generate captions. On the other hand, some attention-based methods, such as ADL [47] and T-DL [104], report lower scores. This may be due to differences in the attention mechanisms used or the overall design of the methods.
5.4.2. MSR-VTT Dataset: Method Performance Overview
On the MSR-VTT Dataset, numerous methods have been assessed using the BLEU, METEOR, CIDEr, and ROUGE-L metrics. As presented in Table 4, various models have reported results for this dataset, with certain approaches achieving remarkable performances. Here, we provide an overview of the top-performing methods, along with their published scores.
Table 4.
Comparison of video captioning methods on the MSR-VTT dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
Encoder-decoder based approaches: Encoder-decoder architectures have demonstrated impressive results on the MSR-VTT dataset. Among the latest methods, GLG [87] emerges as a standout with high scores across all metrics, including BLEU (49.1), METEOR (31.9), CIDEr (63.6), and ROUGE-L (66.1). This method likely benefits from its innovative encoding and decoding strategies to produce accurate captions. Other noteworthy methods include Howtocaption [76] and ATMNet [78], both of which also report competitive scores. Earlier works, such as AVSSN [93] and RMN [80], also adopted encoder-decoder frameworks but achieved lower scores compared to the more recent methods. This underscores the significant advancements made in video description techniques over time.
Transformer-based approaches: Transformer-based approaches have surged in popularity recently, driven by their adeptness at modeling long-range dependencies within sequences. TextKG [32], a notable transformer-based methodology, has achieved impressive scores on the MSR-VTT dataset, boasting a BLEU of 46.6, METEOR of 30.5, CIDEr of 60.8, and ROUGE-L of 64.8. Similarly, HMN [36] is another transformer-based method that has demonstrated promising outcomes. Nevertheless, certain transformer-based techniques, such as SnapCap [42] and UAT-FEGs [35], have yielded relatively lower scores when juxtaposed with encoder-decoder-based counterparts. This disparity underscores that while transformers undeniably possess potent architectural strengths, their efficacy in video description tasks, specifically on the MSR-VTT dataset, may be contingent upon nuanced implementation specifics and design decisions.
Attention-based approaches: Attention-based methods have been extensively used in video descriptions to focus on pertinent parts of the video when generating captions. Nevertheless, the performance of attention-based methods on the MSR-VTT dataset exhibits considerable variation. IVRC [100] reports relatively high scores, with BLEU (43.7), METEOR (30.2), CIDEr (57.1), and ROUGE-L (63.0). This method may leverage sophisticated attention mechanisms to enhance caption generation. Conversely, some attention-based methods, like ADL [47] and T-DL [104], report lower scores. This discrepancy could be attributed to differences in the attention mechanisms employed or the overall architecture of the methods. The varied performance of attention-based approaches highlights the importance of careful design and implementation in achieving optimal results on the MSR-VTT dataset.
5.4.3. VATEX Dataset: Method Performance Overview
On the VATEX Dataset, a variety of methods have been assessed using the BLEU, METEOR, CIDEr, and ROUGE-L metrics. As presented in Table 5, multiple approaches have reported results for this dataset, with certain models demonstrating notable performances. Here, we provide an overview of the best-performing methods along with their published scores.
Table 5.
Comparison of video captioning methods on the VATEX dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
Encoder-decoder based approaches: Encoder-decoder architectures have exhibited promising results on the VATEX dataset. Among the listed methods, CARE [90] stands out with a BLEU score of 37.5, METEOR score of 25.1, CIDEr score of 63.1, and ROUGE-L score of 52.4. This method incorporates concept-aware reasoning to enhance video captioning. Other competitive methods include ViT/L14 [89] with notable scores across all metrics, highlighting the effectiveness of vision transformers in this task. Earlier works, such as vc-HRNAT [26] and RCG [111], also utilized encoder-decoder frameworks but achieved lower scores compared to more recent methods. This underscores the advancements made in video description techniques over time.
Transformer-based approaches: Transformer-based methods have demonstrated their prowess in capturing long-range dependencies, making them suitable for video captioning tasks. X-Lin+Tr(VATEX-En) [112] is a prominent transformer-based approach that reports impressive scores, with a BLEU score of 40.7, METEOR score of 25.8, CIDEr score of 81.4, and ROUGE-L score of 53.7. Its counterpart, X-Lin+Tr(VATEX-Ch), also performs well, particularly in METEOR and ROUGE-L metrics, indicating its adaptability to different language settings. It is worth noting that while transformer-based methods have shown great potential, their performance can vary depending on the specific implementation and design choices, as evidenced by the difference in scores between the English and Chinese versions of X-Lin+Tr.
Attention-based approaches: Attention-based methods have been extensively used in video description to focus on relevant video frames and features when generating captions.On the VATEX dataset, the performance of attention-based methods varies. FAtt(VATEX-En) [48] achieves competitive results, with a BLEU-4 score of 39.2, a METEOR score of 25.0, a CIDEr score of 76.0, and a ROUGE-L score of 52.7. This method multi-level features and attention mechanisms to improve captioning accuracy. However, other attention-based methods, such as IVRC [100] and ORG-TRL [106], report lower scores. This variability may be attributed to differences in the attention mechanisms employed, the overall architecture design, or the specific features used for caption generation.
5.4.4. ActivityNet Captions Dataset: Method Performance Overview
On the ActivityNet Captions Dataset, numerous methods have been assessed using the BLEU, METEOR, CIDEr, and ROUGE-L metrics. As presented in Table 6, several approaches have reported results for this dataset, with certain models demonstrating notable performances. Here, we provide an overview of the best-performing methods alongside their published scores.
Encoder-decoder based approaches: Encoder-decoder frameworks have exhibited promising outcomes on the ActivityNet Captions Dataset. Among the methods listed, COST [41] stands out with a BLEU score of 11.88, METEOR score of 15.70, CIDEr score of 29.64, and ROUGE-L score of 6.11. This approach leverages collaborative mechanisms for improved video captioning. Another noteworthy method is VSJM-Net [22], which achieves competitive scores, particularly in the METEOR and ROUGE-L metrics. Earlier works, such as VC-FF [113] and JSRL-VCT [96], also employed encoder-decoder architectures but obtained lower scores compared to more recent methods. This underscores the advancements made in video description techniques over time.
Transformer-based approaches: Transformer-based methods have gained traction due to their capacity to capture long-range dependencies in video sequences. Among the transformer-based approaches listed, COOT [37] performs notably well, with a BLEU score of 17.43, METEOR score of 15.99, CIDEr score of 28.19, and ROUGE-L score of 31.45. This method effectively utilizes transformers for video captioning tasks. Other transformer-based methods, such as SART [114] and TextKG [32], also show promising results. However, it is worth noting that some transformer-based approaches, like BMT-V+sm [115], report lower scores. This suggests that the effectiveness of transformers in video description tasks on the ActivityNet Captions Dataset may be influenced by the specific architectural choices and implementation details.
Attention-based approaches: Attention-based methods have been extensively used in video descriptions to focus on relevant video frames and features when generating captions. However, the performance of these methods on the ActivityNet Captions Dataset varies. EC-SL [116] reports relatively modest scores, with a BLEU score of 1.33, METEOR score of 7.49, CIDEr score of 21.21, and ROUGE-L score of 13.02. This method emphasizes the use of attention mechanisms for improved caption generation. Other attention-based methods, such as WS-DEC [117] and Bi-SST [118], also report lower scores compared to encoder-decoder and transformer-based approaches. This variability in performance may be attributed to differences in the attention mechanisms employed or the overall design of the methods. Overall, while attention-based methods have shown potential, further improvements are needed to achieve competitive results on the ActivityNet Dataset.
Table 6.
Comparison of video captioning methods on the ActivityNet Captions dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
Table 6.
Comparison of video captioning methods on the ActivityNet Captions dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
| NO | Method | Method’s Category | BLEU-4 | METEOR | CIDEr | ROUGE-L | Year |
|---|---|---|---|---|---|---|---|
| 1 | COST [41] | Encoder-decoder | 11.88 | 15.70 | 29.64 | - | 2023 |
| 2 | MPP-net [119] | Encoder-decoder | 12.75 | 16.01 | 29.35 | - | 2023 |
| 3 | VSJM-Net [22] | Encoder-decoder | 3.97 | 12.89 | 26.52 | 25.37 | 2022 |
| 4 | VC-FF [113] | Encoder-decoder | 2.76 | 7.02 | 26.55 | 18.16 | 2021 |
| 5 | JSRL-VCT [96] | Encoder-decoder | 1.9 | 11.30 | 44.20 | 22.40 | 2019 |
| 6 | TextKG [32] | Transformer | 11.3 | 16.5 | 26.6 | 6.3 | 2023 |
| 7 | Vltint [120] | Transformer | 14.93 | 18.16 | 33.07 | 36.86 | 2023 |
| 8 | SART [114] | Transformer | 11.35 | 16.21 | 28.35 | 7.18 | 2022 |
| 9 | BMT-V+sm [115] | Transformer | 2.55 | 8.65 | 13.48 | 13.62 | 2021 |
| 10 | COOT [37] | Transformer | 17.43 | 15.99 | 28.19 | 31.45 | 2020 |
| 11 | MART [39] | Transformer | 10.85 | 15.99 | 28.19 | 6.64 | 2020 |
| 12 | EC-SL [116] | Attention | 1.33 | 7.49 | 21.21 | 13.02 | 2021 |
| 13 | WS-DEC [117] | Attention | 1.27 | 6.30 | 18.77 | 12.55 | 2018 |
| 14 | Bi-SST [118] | Attention | 2.30 | 9.60 | 12.68 | 19.10 | 2018 |
5.4.5. YouCook2 Dataset: Method Performance Overview
On the YouCook2 Dataset, various methods have been evaluated using the BLEU, METEOR, CIDEr, and ROUGE-L metrics. As shown in Table 7, several methods have reported results for this dataset, with specific models achieving notable performances. Here, we summarize the best-performing methods along with their published scores.
Encoder-decoder based approaches: Encoder-decoder architectures have demonstrated strong performance on the YouCook2 Dataset. Among the listed methods, UniVL [121] stands out with a BLEU score of 17.35, METEOR score of 22.35, CIDEr score of 1.81, and ROUGE-L score of 46.52. This method effectively combines video and text representations, leveraging a unified encoder-decoder framework to generate accurate captions. Other competitive methods in this category include COTE [122] and E2vidD6-MASSvid-BiD [21], which also report high scores across multiple metrics. Earlier works, such as VSJM-Net [22], achieved lower scores compared to more recent methods. This reflects the ongoing progress in developing advanced encoder-decoder frameworks for video description tasks.
Transformer-based approaches: Transformer-based methods have shown promise on the YouCook2 Dataset, particularly in capturing the temporal dependencies and contextual information in videos. COOT [37] is a notable transformer-based approach that reports a BLEU score of 17.97, METEOR score of 19.85, CIDEr score of 57.24, and ROUGE-L score of 37.94. This method utilizes a cross-modal transformer to fuse video and text features effectively. Other transformer-based methods, such as TextKG [32] and SART [114], also exhibit competitive performance. However, it is worth noting that some transformer-based methods, like MART [39], report lower scores compared to encoder-decoder-based approaches. This suggests that the effectiveness of transformer architectures in video description tasks may be influenced by the specific design and implementation choices.
Attention-based approaches: Attention-based methods have been employed in video description to focus on relevant parts of the video frames and temporal sequences when generating captions. However, the performance of these methods on the YouCook2 Dataset is varied. ActBERT [123], for instance, reports a BLEU score of 5.41, METEOR score of 13.30, CIDEr score of 0.65, and ROUGE-L score of 30.56. This method incorporates both local and global attention mechanisms to capture fine-grained visual and temporal information. In contrast, Bi-SST [118] reports lower scores across all metrics. This variation in performance may be attributed to differences in the attention mechanisms used, the overall architecture design, or the ability to effectively capture and utilize contextual information in the video description process. Further research and development in attention-based methods are needed to improve their performance on the YouCook2 Dataset and other video description tasks.
Table 7.
Comparison of video captioning methods on the YouCook2 dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
Table 7.
Comparison of video captioning methods on the YouCook2 dataset, showing BLEU-4 (in percent), METEOR, CIDEr, and ROUGE-L scores per sentence for short descriptions.
| NO | Method | Method’s Category | BLEU-4 | METEOR | CIDEr | ROUGE-L | Year |
|---|---|---|---|---|---|---|---|
| 1 | Howtocaption [76] | Encoder-decoder | 8.8 | 15.9 | 116.4 | 37.3 | 2024 |
| 2 | COTE [122] | Encoder-decoder | 12.14 | 19.77 | 1.22 | 43.26 | 2024 |
| 3 | COST [41] | Encoder-decoder | 8.65 | 15.62 | 1.05 | 36.5 | 2023 |
| 4 | VSJM-Net [22] | Encoder-decoder | 1.09 | 4.31 | 9.07 | 10.51 | 2022 |
| 5 | MV-GPT [108] | Encoder-decoder | 21.26 | 26.36 | 2.14 | 48.58 | 2022 |
| 6 | E2vidD6-MASSvid-BiD [21] | Encoder-decoder | 12.04 | 18.32 | 1.23 | 39.03 | 2020 |
| 7 | UniVL [121] | Encoder-decoder | 17.35 | 22.35 | 1.81 | 46.52 | 2020 |
| 8 | TextKG [32] | Transformer | 14.0 | 22.1 | 75.9 | - | 2023 |
| 9 | SART [114] | Transformer | 11.43 | 19.91 | 57.66 | 8.58 | 2022 |
| 10 | MART [39] | Transformer | 11.30 | 19.85 | 57.24 | 6.69 | 2020 |
| 11 | COOT [37] | Transformer | 17.97 | 19.85 | 57.24 | 37.94 | 2020 |
| 12 | Videobert+S3D [124] | Transformer | 4.33 | 11.94 | 0.55 | 28.8 | 2019 |
| 13 | ActBERT [123] | Attention | 5.41 | 13.30 | 0.65 | 30.56 | 2020 |
| 14 | Bi-SST [118] | Attention | 2.30 | 9.60 | 12.68 | 19.10 | 2018 |
5.4.6. VITT Dataset: Method Performance Overview
On the VITT Dataset, a variety of methods have been assessed using the BLEU, METEOR, CIDEr, ROUGE-L, and SODA metrics. As presented in Table 8, multiple approaches have reported results for this dataset, with certain models demonstrating notable performances. Here, we provide an overview of the top-performing methods along with their published scores.
Encoder-decoder based approaches: Encoder-decoder architectures have exhibited promising outcomes on the VITT dataset. Among the listed methods, VidChapters-7M [125] stands out with a high METEOR score of 9.6, CIDEr score of 50.9, and SODA score of 15.1. This approach leverages advanced techniques to effectively generate descriptive captions. Other significant methods include Vid2Seq [31] and AIEM [126], which also report competitive scores, particularly in the CIDEr and SODA metrics, respectively.
Earlier works, such as SimpleConcat_WikiHow_T5 [127], although not reporting scores for all metrics, show a BLEU score of 1.34 and a ROUGE-L score of 9.21. This indicates progress in video description techniques, though more recent methods have surpassed these initial efforts.
Transformer-based approaches: Transformer-based methods have become increasingly popular due to their capacity to capture long-range dependencies. FAD [128] is a notable transformer-based approach on the VITT dataset, reporting a METEOR score of 18.9, a CIDEr score of 37.2, and a SODA score of 10.2. While this method demonstrates competitive performance, it is worth noting that other transformer-based approaches might exist that could further push the boundaries of performance on this dataset.
However, it is evident that transformer-based methods, like FAD, may not always outperform encoder-decoder-based approaches across all metrics. This underscores the importance of tailored implementations and design choices in maximizing the effectiveness of these architectures for video description tasks.
Table 8.
Comparison of video captioning methods on the ViTT dataset, showing BLEU-4 (in percent), METEOR, CIDEr, ROUGE-L, and SODA scores per sentence for short descriptions.
Table 8.
Comparison of video captioning methods on the ViTT dataset, showing BLEU-4 (in percent), METEOR, CIDEr, ROUGE-L, and SODA scores per sentence for short descriptions.
| NO | Method | Method’s Category | BLEU-4 | METEOR | CIDEr | ROUGE-L | SODA | Year |
|---|---|---|---|---|---|---|---|---|
| 1 | AIEM [126] | Encoder-decoder | - | 44.9 | 11.3 | - | 11.8 | 2025 |
| 2 | Streaming_Vid2Seq [129] | Encoder-decoder | - | 5.8 | 25.2 | - | 10.0 | 2024 |
| 3 | Vid2Seq [31] | Encoder-decoder | - | 8.5 | 43.5 | - | 13.5 | 2023 |
| 4 | VidChapters-7M [125] | Encoder-decoder | - | 9.6 | 50.9 | - | 15.1 | 2023 |
| 5 | SimpleConcat_WikiHow_T5 [127] | Encoder-decoder | 1.34 | 7.97 | 0.25 | 9.21 | - | 2022 |
| 6 | FAD [128] | Transformer | - | 18.9 | 37.2 | - | 10.2 | 2024 |
5.4.7. Eight Datasets: Method Performance Overview
Various methods have been evaluated on eight datasets using BLEU, METEOR, CIDEr, and ROUGE-L metrics, as presented in Table 9. Below, we provide an overview of the performance of these methods on each dataset.
VC_NBA_2022 Dataset: On the VC_NBA_2022 dataset, KEANetT [130] stands out with the highest scores across all metrics: BLEU (32.4), METEOR (28.0), CIDEr (138.5), and ROUGE-L (54.9). This dominance is likely due to its advanced multimodal fusion and context-aware techniques for generating high-quality video descriptions. Other methods listed for this dataset are not directly comparable given KEANetT’s superior performance.
WTS Dataset: The WTS dataset has witnessed a variety of methods being evaluated, with significant performance differences. Among recent methods, TrafficVLM(VC) [131] achieves the highest scores: BLEU (0.419), METEOR (0.493), CIDEr (0.609), and ROUGE-L (0.589). This method’s video-centric design and effective use of multimodal information likely contribute to its success. Competitive methods include CityLLaVA [132] and TrafficVLM(PC) [131], while earlier works like Video-LLaMA [133] and Video-ChatGPT [134] show lower performance, highlighting recent advancements in the field.
TGIF Dataset: On the TGIF dataset, CMGNet [88] excels with the highest scores: BLEU (16.2), METEOR (18.4), CIDEr (53.7), and ROUGE-L (43.7). This method likely employs sophisticated temporal modeling and feature extraction techniques to capture video event dynamics. Other notable methods, such as SAAT [95] and RMN [80], also report competitive scores, demonstrating the effectiveness of temporal reasoning and syntactic awareness in video description tasks.
TVC Dataset: The TVC dataset has seen a range of methods evaluated, with HERO [135] achieving the highest scores: BLEU (12.35), METEOR (17.64), CIDEr (49.98), and ROUGE-L (34.16). HERO’s hierarchical encoding and decoding framework likely contribute to its success in capturing video content structure. Competitive methods include COTE [122] and MMT [59], suggesting that contrastive learning and multimodal fusion are particularly effective on this dataset.
Charades Dataset: On the Charades dataset, HRL [25] stands out with the highest scores: BLEU (18.8), METEOR (19.5), CIDEr (23.2), and ROUGE-L (41.4). This method’s hierarchical reinforcement learning approach likely generates coherent and relevant video descriptions. Given the dataset’s focus on complex activities, methods that effectively model temporal dependencies and activity hierarchies tend to perform better.
TACoS Dataset: The TACoS Dataset has seen evaluations from methods with varying performances. LSCEC [136] achieves notably high BLEU (47.6) but lower scores on other metrics, such as METEOR (0.352), CIDEr (1.774), and ROUGE-L (0.725). This discrepancy may suggest a focus on specific aspects of caption generation at the expense of others. JEDDi-Net [137] reports more balanced scores, indicating its effectiveness in capturing both visual and textual information. These results highlight the need for holistic approaches that can perform well across all evaluation metrics.
VideoStory Dataset: ResBRNN [138] achieves the highest scores on the VideoStory dataset: BLEU (15.6), METEOR (20.1), CIDEr (103.6), and ROUGE-L (29.9). This method’s recurrent neural network with residual connections likely effectively captures long-term dependencies in video narratives. Earlier works, such as m-RNN [139], show lower performance, highlighting advancements in modeling video stories over time.
MPII-MD Dataset: On the MPII-MD dataset, Boundary-aware [140] achieves the highest scores: BLEU (0.8), METEOR (7.0), CIDEr (10.8), and ROUGE-L (16.7). This method’s hierarchical boundary-aware framework likely effectively segments and describes video content. Other methods, such as VL-LSTM [141] and SMT [51], report lower scores, indicating the challenges of describing complex and diverse video content in this dataset. These findings suggest that methods that effectively segment and align video content with textual descriptions perform better on MPII-MD.
Table 9.
Comparison of video captioning methods on nine other benchmark datasets, showing BLEU-4, METEOR, CIDEr, and ROUGE-L scores for each sentence in short descriptions.
Table 9.
Comparison of video captioning methods on nine other benchmark datasets, showing BLEU-4, METEOR, CIDEr, and ROUGE-L scores for each sentence in short descriptions.
| NO | Method | Method’s Category | Dataset | BLEU-4 | METEOR | CIDEr | ROUGE-L | Year |
|---|---|---|---|---|---|---|---|---|
| 1 | KEANetT [130] | Encoder-decoder | VC_NBA_2022 | 32.4 | 28.0 | 138.5 | 54.9 | 2025 |
| 2 | Instance-VideoLLM [61] | Attention | WTS | 0.121 | 0.409 | 0.389 | 0.417 | 2025 |
| 3 | UCF-SST-NLP [131] | Attention | WTS | 0.2005 | 0.4115 | 0.5573 | 0.4416 | 2024 |
| 4 | TrafficVLM(VC) [142] | Encoder-decoder | WTS | 0.419 | 0.493 | 0.609 | 0.589 | 2024 |
| 5 | TrafficVLM(PC) [142] | Encoder-decoder | WTS | 0.316 | 0.378 | 0.61 | 0.438 | 2024 |
| 6 | CityLLaVA [132] | Encoder-decoder | WTS | 0.289 | 0.484 | 1.044 | 0.481 | 2024 |
| 7 | Video-LLaMA [133] | Encoder-decoder | WTS | 0.045 | 0.247 | 0.210 | 0.226 | 2023 |
| 8 | Video-ChatGPT [134] | Encoder-decoder | WTS | 0.072 | 0.267 | 0.282 | 0.266 | 2023 |
| 9 | +VSE [143] | Transformer | BDD-X | 21.82 | 25.42 | 162.12 | 42.47 | 2024 |
| 10 | MVVC [144] | Transformer | BDD-X | 17.5 | 20.7 | 107.0 | 44.4 | 2024 |
| 11 | ADAPT [145] | Transformer | BDD-X | 11.4 | 15.2 | 102.6 | 32.0 | 2023 |
| 12 | CMGNet [88] | Encoder-decoder | TGIF | 16.2 | 18.4 | 53.7 | 43.7 | 2024 |
| 13 | SAAT [95] | Encoder-decoder | TGIF | 15.3 | 18.1 | 49.6 | 42.3 | 2020 |
| 14 | RMN [80] | Encoder-decoder | TGIF | 15.6 | 18.1 | 50.1 | 42.6 | 2020 |
| 15 | COTE [122] | Encoder-decoder | TVC | 12.20 | 17.56 | 45.81 | 33.24 | 2024 |
| 16 | MMT [59] | Transformer | TVC | 10.87 | 16.91 | 45.38 | 32.81 | 2020 |
| 17 | HERO [135] | Transformer | TVC | 12.35 | 17.64 | 49.98 | 34.16 | 2020 |
| 18 | HRL [25] | Encoder-decoder | Charades | 18.8 | 19.5 | 23.2 | 41.4 | 2018 |
| 19 | LSCEC [136] | Encoder-decoder | TACoS | 47.6 | 0.352 | 1.774 | 0.725 | 2021 |
| 20 | JEDDi-Net [137] | Encoder-decoder | TACoS | 18.1 | 23.85 | 103.98 | 50.85 | 2018 |
| 21 | ResBRNN [138] | Encoder-decoder | VideoStory | 15.6 | 20.1 | 103.6 | 29.9 | 2019 |
| 22 | m-RNN [139] | Encoder-decoder | VideoStory | 11.8 | 18.0 | 81.3 | 28.5 | 2015 |
| 23 | Boundary-aware [140] | Encoder-decoder | MPII-MD | 0.8 | 7.0 | 10.8 | 16.7 | 2017 |
| 24 | VL-LSTM [141] | Encoder-decoder | MPII-MD | 0.80 | 7.03 | 9.98 | 16.02 | 2015 |
| 25 | SMT [51] | Encoder-decoder | MPII-MD | 0.47 | 5.59 | 13.21 | 8.14 | 2015 |
6. Emerging Trends and Future Directions
Dense Video Captioning (DVC) has seen significant advancements in recent years, driven by the integration of deep learning and multimodal analysis. This section delves into the current research trends, open challenges, limitations, and potential future directions in DVC.
6.1. Current Research Trends in DVC
The field of DVC is rapidly evolving, with researchers exploring various techniques to improve event recognition, localization, and description generation. One notable trend is the increasing use of Transformer-based architectures. Unlike traditional LSTM or encoder-decoder models, Transformers can capture long-range dependencies in video data, leading to more accurate and coherent captions. For instance, the PDVC framework [17] employs a deformable Transformer with parallel decoding heads for localization and captioning, achieving state-of-the-art performance.
Another trend is the integration of multimodal information. DVC systems are no longer limited to visual features but also incorporate audio, language, and other modalities to enhance the robustness and descriptive capabilities of the system. Multimodal fusion techniques, such as those proposed in Aafaq et al. [22], leverage the natural kinship between events and their human-provided descriptions to generate semantically meaningful event proposals.
6.2. Open Challenges and Limitations
Despite these advancements, DVC still faces several challenges and limitations. One major challenge is the complexity of video data, which contains a vast amount of spatiotemporal information. Effectively extracting and utilizing this information remains a core challenge in video feature extraction. Moreover, the high-dimensional feature vectors of video data can lead to high computational complexity, necessitating efficient feature extraction, and dimensionality reduction techniques.
Another limitation is the performance of event recognition and localization algorithms. While significant progress has been made, these algorithms still struggle with overlapping events, varying event durations, and subtle visual cues. Improving the accuracy and robustness of these algorithms is crucial for generating high-quality captions.
6.3. Potential Future Directions
- (1)
- Improved Event Recognition and Localization AlgorithmsTo address the limitations of current event recognition and localization algorithms, researchers are exploring novel approaches such as contrastive learning and self-supervised learning. These methods can leverage unlabeled video data to learn more discriminative features, improving the performance of event detection and localization. Additionally, integrating temporal context and semantic information can further enhance the accuracy of event recognition.
- (2)
- Enhanced Multimodal Integration and Cross-Media UnderstandingMultimodal integration is key to improving the descriptive capabilities of DVC systems. Future research should focus on developing more sophisticated multimodal fusion techniques that can effectively combine visual, audio, and language features. Moreover, cross-media understanding, which involves linking video content to related text, audio, and other media, holds promise for advancing DVC applications in various domains.
- (3)
- More Efficient and Interpretable ModelsAs DVC systems become more complex, the need for efficient and interpretable models becomes increasingly important. Researchers are exploring various techniques to reduce the computational complexity of DVC models, such as model pruning, quantization, and knowledge distillation. Additionally, developing interpretable models that can provide insights into the decision-making process of DVC systems will enhance their trustworthiness and usability.
- (4)
- Applications in Video Retrieval, Accessibility, and Content AnalysisDVC has a wide range of applications, including video retrieval, accessibility, and content analysis. In video retrieval, DVC can improve the accuracy and efficiency of search results by generating descriptive text for each video segment. For accessibility, DVC can help visually impaired individuals better understand video content by providing detailed descriptions of events and actions. In content analysis, DVC can provide valuable insights into video data, facilitating tasks such as automatic annotation, classification, and summarization.
7. Conclusions
7.1. Summary of the Review’s Key Findings
This comprehensive review of Dense Video Captioning (DVC) has unveiled several pivotal findings that encapsulate the present state of the field. Firstly, DVC marks a substantial leap in video understanding, effectively bridging the gap between video recognition and natural language generation. The evolution of DVC techniques has been remarkable, transitioning from early methods that heavily relied on hand-crafted features and template-based generation to recent approaches that harness the power of deep learning, particularly transformer-based architectures, resulting in significant performance enhancements.
Secondly, the availability of diverse and large-scale datasets has played a crucial role in propelling advancements in DVC. Datasets like ActivityNet Captions, YouTube2Text, and Charades have provided essential annotations for training and evaluating DVC models, enabling researchers to benchmark their methods against state-of-the-art performances. These datasets encompass a broad spectrum of video content and include temporal annotations for precise event localization within videos, which is vital for generating accurate captions.
Thirdly, the integration of multimodal information has emerged as a prominent trend in DVC research. By amalgamating visual, auditory, and textual cues, models can gain a deeper understanding of video content and generate more informative and contextually relevant captions. This multimodal fusion can occur at various stages, including early fusion (combining features prior to encoding), late fusion (combining predictions after decoding), and hybrid approaches that utilize both strategies.
Lastly, the review underscores the challenges and limitations inherent in current DVC techniques. These encompass the complexity of video content, which may feature multiple overlapping events and background noise, as well as the inherent ambiguity in natural language, leading to multiple plausible captions for the same video segment. Furthermore, the computational demands of deep learning models and the necessity for substantial amounts of labeled data remain significant obstacles to the widespread adoption of DVC technology.
7.2. Contributions and Implications for the Field of DVC
This review contributes to the DVC field by offering a comprehensive overview of existing techniques, datasets, and challenges. By synthesizing information from a multitude of sources, this review presents a clear picture of the current DVC research landscape and identifies key trends and avenues for future exploration.
The implications of this review for the DVC field are multifaceted. Firstly, it emphasizes the importance of ongoing innovation in model architectures and training strategies to tackle the challenges posed by video complexity and natural language ambiguity. Researchers should delve into new methods for more effectively integrating multimodal information and developing robust and scalable models capable of handling large-scale video data.
Secondly, the review underscores the need for diverse and high-quality datasets to support the development and evaluation of DVC models. As the field advances, there will be a growing demand for datasets that encompass a wider array of video genres, languages, and cultural contexts. Additionally, there is a need for datasets that incorporate more challenging and realistic scenarios, such as videos captured in unconstrained environments or with low-quality cameras.
Finally, this review draws attention to the broader implications of DVC technology for applications like video retrieval, video summarization, and assistive technologies for the visually impaired. As DVC models become more accurate and efficient, they have the potential to revolutionize these and other domains by facilitating more intuitive and natural interactions with video content.
7.3. Final Thoughts and Recommendations for Future Research
In conclusion, Dense Video Captioning represents a promising avenue in video understanding and natural language generation. While considerable progress has been made in recent years, numerous challenges and opportunities for future research remain. To propel the field of DVC forward, we recommend the following research directions:
- (1)
- Exploration of Novel Model ArchitecturesResearchers should continue to investigate new model architectures that can more effectively capture the temporal and spatial dynamics of video content. This may involve the development of more advanced attention mechanisms, the integration of graph neural networks for modeling relationships between objects and events, or the utilization of reinforcement learning to optimize caption generation.
- (2)
- Enhancement of Multimodal Fusion TechniquesAs multimodal information becomes increasingly crucial for DVC, researchers should focus on refining fusion techniques that can harness the complementary strengths of visual, auditory, and textual cues. This may entail employing more sophisticated feature extraction methods, developing better alignment strategies for multimodal features, or exploring innovative ways to combine predictions from different modalities.
- (3)
- Creation of Diverse and High-Quality DatasetsTo bolster the development and evaluation of DVC models, researchers should endeavor to create diverse and high-quality datasets that span a wide range of video genres, languages, and cultural contexts. Additionally, there is a need for datasets that incorporate more challenging and realistic scenarios to test the robustness and scalability of DVC models.
- (4)
- Development of Temporal-Specific Evaluation MetricsGiven the limitations of current NLP-based metrics in assessing temporal event localization, future research should prioritize the creation of evaluation metrics specifically designed for dense video captioning. These metrics should focus on temporal Precision (P) and Recall (R), event alignment accuracy, and the ability to capture temporal dependencies. Potential approaches may include developing metrics that evaluate the temporal coherence of generated captions, measuring the alignment between predicted and ground-truth event boundaries, or incorporating temporal consistency as a key criterion. Such advancements would provide a more comprehensive and relevant framework for evaluating the performance of dense video captioning systems in real-world applications.
- (5)
- Exploration of Applications and Use CasesLastly, researchers should explore the potential applications and use cases of DVC technology to demonstrate its value and drive adoption in real-world scenarios. This may involve collaborating with industry partners to develop prototypes and pilot projects, conducting user studies to grasp the needs and preferences of end-users, or exploring new ways to integrate DVC technology into existing systems and platforms.
Author Contributions
Conceptualization, Z.L.; methodology, Z.L.; validation, Z.L.; formal analysis, Z.L.; investigation, Z.L.; writing—original draft preparation, Z.L.; writing—review and editing, R.S.; visualization, R.S.; supervision, Z.L.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.
Funding
This work was mainly supported by the National Natural Science Foundation of China (Grant No. 62162061 and Grant No. 62262066). This study was also supported by the Doctoral Research Foundation of Xinjiang Normal University (Grant No. XJNUBS2115), and the Xinjiang Normal University Youth Top Talents Project (Grant No. XJNUQB2022-21). In addition, this study was also supported by the Xinjiang Key Research and Development Program (2022B01007-1).
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Rafiq, G.; Rafiq, M.; Choi, G.S. Video description: A comprehensive survey of deep learning approaches. Artif. Intell. Rev. 2023, 56, 13293–13372. [Google Scholar] [CrossRef]
- Krishna, R.; Hata, K.; Ren, F.; Fei-Fei, L.; Niebles, J.C. Dense-captioning events in videos. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 706–715. [Google Scholar] [CrossRef]
- Li, Y.; Yao, T.; Pan, Y.; Chao, H.; Mei, T. Jointly localizing and describing events for dense video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition alt, Lake City, UT, USA, 18–23 June 2018; pp. 7492–7500. [Google Scholar] [CrossRef]
- Zhou, L.; Xu, C.; Corso, J. Towards automatic learning of procedures from web instructional videos. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar] [CrossRef]
- Rafiq, M.; Rafiq, G.; Choi, G.S. Video description: Datasets & evaluation metrics. IEEE Access 2021, 9, 121665–121685. [Google Scholar] [CrossRef]
- Hou, R.; Chen, C.; Shah, M. Tube convolutional neural network (t-cnn) for action detection in videos. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5822–5831. [Google Scholar] [CrossRef]
- Xu, J.; Mei, T.; Yao, T.; Rui, Y. Msr-vtt: A large video description dataset for bridging video and language. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5288–5296. [Google Scholar] [CrossRef]
- Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.-J. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 7–12 July 2002; pp. 311–318. [Google Scholar] [CrossRef]
- Lavie, A.; Agarwal, A. Meteor: An automatic metric for MT evaluation with high levels of correlation with human judgments. In Proceedings of the Second Workshop on Statistical Machine Translation, Prague, Czech Republic, 23 June 2007; pp. 228–231. [Google Scholar]
- Vedantam, R.; Lawrence Zitnick, C.; Parikh, D. Cider: Consensus-based image description evaluation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4566–4575. [Google Scholar] [CrossRef]
- Chin-Yew, L. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Workshop on Text Summarization Branches Out, Barcelona, Spain, 25–26 July 2004; pp. 74–81. Available online: https://aclanthology.org/W04-1013/ (accessed on 21 March 2025).
- Fujita, S.; Hirao, T.; Kamigaito, H.; Okumura, M.; Nagata, M. SODA: Story oriented dense video captioning evaluation framework. In Proceedings of the European Conference on Computer Vision, Virtual, Online, UK, 23–28 August 2020; pp. 517–531. [Google Scholar] [CrossRef]
- Anderson, P.; Fernando, B.; Johnson, M.; Gould, S. Spice: Semantic propositional image caption evaluation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 382–398. [Google Scholar] [CrossRef]
- Kusner, M.; Sun, Y.; Kolkin, N.; Weinberger, K. From word embeddings to document distances. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 957–966. Available online: https://proceedings.mlr.press/v37/kusnerb15 (accessed on 21 March 2025).
- Aafaq, N.; Mian, A.; Liu, W.; Gilani, S.Z.; Shah, M. Video description: A survey of methods, datasets, and evaluation metrics. ACM Comput. Surv. 2019, 52, 1–37. [Google Scholar] [CrossRef]
- Yan, C.; Tu, Y.; Wang, X.; Zhang, Y.; Hao, X.; Zhang, Y.; Dai, Q. Spatial-temporal attention mechanism for video captioning. IEEE Trans. Multimed. 2019, 22, 229–241. [Google Scholar] [CrossRef]
- Wang, T.; Zhang, R.; Lu, Z.; Zheng, F.; Cheng, R.; Luo, P. End-to-end dense video captioning with parallel decoding. In Proceedings of the IEEE International Conference on Computer Vision, Online, Canada, 11–17 October 2021; pp. 6847–6857. [Google Scholar] [CrossRef]
- Venugopalan, S.; Xu, H.; Donahue, J.; Rohrbach, M.; Mooney, R.; Saenko, K. Translating videos to natural language using deep recurrent neural networks. arXiv 2014, arXiv:1412.4729. [Google Scholar] [CrossRef]
- Yao, L.; Torabi, A.; Cho, K.; Ballas, N.; Pal, C.; Larochelle, H.; Courville, A. Describing videos by exploiting temporal structure. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4507–4515. [Google Scholar] [CrossRef]
- Hori, C.; Hori, T.; Lee, T.-Y.; Zhang, Z.; Harsham, B.; Hershey, J.R.; Marks, T.K.; Sumi, K. Attention-based multimodal fusion for video description. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4193–4202. [Google Scholar] [CrossRef]
- Huang, G.; Pang, B.; Zhu, Z.; Rivera, C.; Soricut, R. Multimodal pretraining for dense video captioning. arXiv 2020, arXiv:2011.11760. [Google Scholar] [CrossRef]
- Aafaq, N.; Mian, A.; Akhtar, N.; Liu, W.; Shah, M. Dense video captioning with early linguistic information fusion. IEEE Trans. Multimed. 2022, 25, 2309–2322. [Google Scholar] [CrossRef]
- Yu, H.; Wang, J.; Huang, Z.; Yang, Y.; Xu, W. Video paragraph captioning using hierarchical recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26–30 June 2016; pp. 4584–4593. [Google Scholar] [CrossRef]
- Pan, Y.; Mei, T.; Yao, T.; Li, H.; Rui, Y. Jointly modeling embedding and translation to bridge video and language. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, USA, 26–30 June 2016; pp. 4594–4602. [Google Scholar] [CrossRef]
- Wang, X.; Chen, W.; Wu, J.; Wang, Y.-F.; Wang, W.Y. Video captioning via hierarchical reinforcement learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4213–4222. [Google Scholar] [CrossRef]
- Gao, L.; Lei, Y.; Zeng, P.; Song, J.; Wang, M.; Shen, H.T. Hierarchical representation network with auxiliary tasks for video captioning and video question answering. IEEE Trans. Image Process. 2021, 31, 202–215. [Google Scholar] [CrossRef]
- Wang, B.; Ma, L.; Zhang, W.; Jiang, W.; Wang, J.; Liu, W. Controllable video captioning with POS sequence guidance based on gated fusion network. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27–2 November 2019; pp. 2641–2650. [Google Scholar] [CrossRef]
- Jing, S.; Zhang, H.; Zeng, P.; Gao, L.; Song, J.; Shen, H.T. Memory-based augmentation network for video captioning. IEEE Trans. Multimed. 2023, 26, 2367–2379. [Google Scholar] [CrossRef]
- Yang, X.; Tang, K.; Zhang, H.; Cai, J. Auto-encoding scene graphs for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10685–10694. [Google Scholar] [CrossRef]
- Zhou, L.; Zhou, Y.; Corso, J.J.; Socher, R.; Xiong, C. End-to-end dense video captioning with masked transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8739–8748. [Google Scholar] [CrossRef]
- Yang, A.; Nagrani, A.; Seo, P.H.; Miech, A.; Pont-Tuset, J.; Laptev, I.; Sivic, J.; Schmid, C. Vid2seq: Large-scale pretraining of a visual language model for dense video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10714–10726. [Google Scholar] [CrossRef]
- Gu, X.; Chen, G.; Wang, Y.; Zhang, L.; Luo, T.; Wen, L. Text with knowledge graph augmented transformer for video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 18941–18951. [Google Scholar] [CrossRef]
- Zhang, B.; Gao, J.; Yuan, Y. Center-enhanced video captioning model with multimodal semantic alignment. Neural Netw. 2024, 180, 106744. [Google Scholar] [CrossRef]
- Jin, T.; Huang, S.; Chen, M.; Li, Y.; Zhang, Z. SBAT: Video captioning with sparse boundary-aware transformer. arXiv 2020, arXiv:2007.11888. [Google Scholar] [CrossRef]
- Im, H.; Choi, Y.-S. UAT: Universal attention transformer for video captioning. Sensors 2022, 22, 4817. [Google Scholar] [CrossRef]
- Ye, H.; Li, G.; Qi, Y.; Wang, S.; Huang, Q.; Yang, M.-H. Hierarchical modular network for video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17939–17948. [Google Scholar] [CrossRef]
- Ging, S.; Zolfaghari, M.; Pirsiavash, H.; Brox, T. Coot: Cooperative hierarchical transformer for video-text representation learning. Adv. Neural Inf. Process. Syst. 2020, 33, 22605–22618. [Google Scholar] [CrossRef]
- Chen, M.; Li, Y.; Zhang, Z.; Huang, S. TVT: Two-view transformer network for video captioning. In Proceedings of the Asian Conference on Machine Learning, Beijing, China, 14–16 November 2018; pp. 847–862. Available online: https://proceedings.mlr.press/v95/chen18b (accessed on 21 March 2025).
- Lei, J.; Wang, L.; Shen, Y.; Yu, D.; Berg, T.L.; Bansal, M. Mart: Memory-augmented recurrent transformer for coherent video paragraph captioning. arXiv 2020, arXiv:2005.05402. [Google Scholar] [CrossRef]
- Dai, Z.; Yang, Z.; Yang, Y.; Carbonell, J.; Le, Q.V.; Salakhutdinov, R. Transformer-xl: Attentive language models beyond a fixed-length context. arXiv 2019, arXiv:1901.02860. [Google Scholar] [CrossRef]
- Wang, H.; Zhang, L.; Fan, H.; Luo, T. Collaborative three-stream transformers for video captioning. Comput. Vis. Image Underst. 2023, 235, 103799. [Google Scholar] [CrossRef]
- Sun, J.; Su, Y.; Zhang, H.; Cheng, Z.; Zeng, Z.; Wang, Z.; Chen, B.; Yuan, X. SnapCap: Efficient Snapshot Compressive Video Captioning. arXiv 2024, arXiv:2401.04903. [Google Scholar] [CrossRef]
- Tu, Y.; Zhang, X.; Liu, B.; Yan, C. Video description with spatial-temporal attention. In Proceedings of the ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1014–1022. [Google Scholar] [CrossRef]
- Gao, L.; Guo, Z.; Zhang, H.; Xu, X.; Shen, H.T. Video captioning with attention-based LSTM and semantic consistency. IEEE Trans. Multimed. 2017, 19, 2045–2055. [Google Scholar] [CrossRef]
- Yousif, A.J.; Al-Jammas, M.H. Semantic-based temporal attention network for Arabic Video Captioning. Nat. Lang. Process. J. 2025, 10, 100122. [Google Scholar] [CrossRef]
- Guo, Z.; Gao, L.; Song, J.; Xu, X.; Shao, J.; Shen, H.T. Attention-based LSTM with semantic consistency for videos captioning. In Proceedings of the ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 357–361. [Google Scholar] [CrossRef]
- Ji, W.; Wang, R.; Tian, Y.; Wang, X. An attention based dual learning approach for video captioning. Appl. Soft Comput. 2022, 117, 108332. [Google Scholar] [CrossRef]
- Lin, K.; Gan, Z.; Wang, L. Multi-modal feature fusion with feature attention for VATEX captioning challenge 2020. arXiv 2020, arXiv:2006.03315. [Google Scholar] [CrossRef]
- Chen, D.; Dolan, W.B. Collecting highly parallel data for paraphrase evaluation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; pp. 190–200. Available online: https://aclanthology.org/P11-1020/ (accessed on 21 March 2025).
- Regneri, M.; Rohrbach, M.; Wetzel, D.; Thater, S.; Schiele, B.; Pinkal, M. Grounding action descriptions in videos. Trans. Assoc. Comput. Linguist. 2013, 1, 25–36. [Google Scholar] [CrossRef]
- Rohrbach, A.; Rohrbach, M.; Tandon, N.; Schiele, B. A dataset for movie description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3202–3212. [Google Scholar] [CrossRef]
- Li, Y.; Song, Y.; Cao, L.; Tetreault, J.; Goldberg, L.; Jaimes, A.; Luo, J. TGIF: A new dataset and benchmark on animated GIF description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4641–4650. [Google Scholar] [CrossRef]
- Sigurdsson, G.A.; Varol, G.; Wang, X.; Farhadi, A.; Laptev, I.; Gupta, A. Hollywood in homes: Crowdsourcing data collection for activity understanding. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 510–526. [Google Scholar] [CrossRef]
- Zeng, K.-H.; Chen, T.-H.; Niebles, J.C.; Sun, M. Title generation for user generated videos. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 609–625. [Google Scholar] [CrossRef]
- Kim, J.; Rohrbach, A.; Darrell, T.; Canny, J.; Akata, Z. Textual explanations for self-driving vehicles. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 563–578. [Google Scholar] [CrossRef]
- Gella, S.; Lewis, M.; Rohrbach, M. A dataset for telling the stories of social media videos. In Proceedings of the Conference on Empirical Methods in Natural Language, Brussels, Belgium, 31 October–4 November 2018; pp. 968–974. [Google Scholar] [CrossRef]
- Pini, S.; Cornia, M.; Bolelli, F.; Baraldi, L.; Cucchiara, R. M-VAD names: A dataset for video captioning with naming. Multimed. Tools Appl. 2019, 78, 14007–14027. [Google Scholar] [CrossRef]
- Wang, X.; Wu, J.; Chen, J.; Li, L.; Wang, Y.-F.; Wang, W.Y. Vatex: A large-scale, high-quality multilingual dataset for video-and-language research. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4581–4591. [Google Scholar] [CrossRef]
- Lei, J.; Yu, L.; Berg, T.L.; Bansal, M. TVR: A large-scale dataset for video-subtitle moment retrieval. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 447–463. [Google Scholar] [CrossRef]
- Xi, Z.; Shi, G.; Wu, L.; Li, X.; Yan, J.; Wang, L.; Liu, Z. Knowledge Graph Supported Benchmark and Video Captioning for Basketball. arXiv 2024, arXiv:2401.13888. [Google Scholar] [CrossRef]
- Kong, Q.; Kawana, Y.; Saini, R.; Kumar, A.; Pan, J.; Gu, T.; Ozao, Y.; Opra, B.; Sato, Y.; Kobori, N. Wts: A pedestrian-centric traffic video dataset for fine-grained spatial-temporal understanding. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 1–18. [Google Scholar] [CrossRef]
- Yang, T.; Jia, J.; Wang, B.; Cheng, Y.; Li, Y.; Hao, D.; Cao, X.; Chen, Q.; Li, H.; Jiang, P.; et al. Spatiotemporal Fine-grained Video Description for Short Videos. In Proceedings of the ACM International Conference on Multimedia, Melbourne, Australia, 28–31 October 2024; pp. 3945–3954. [Google Scholar] [CrossRef]
- Zhang, G.; Ren, J.; Gu, J.; Tresp, V. Multi-event video-text retrieval. In Proceedings of the IEEE International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 22113–22123. [Google Scholar] [CrossRef]
- Shen, X.; Li, D.; Zhou, J.; Qin, Z.; He, B.; Han, X.; Li, A.; Dai, Y.; Kong, L.; Wang, M.; et al. Fine-grained audible video description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10585–10596. [Google Scholar] [CrossRef]
- Islam, M.M.; Ho, N.; Yang, X.; Nagarajan, T.; Torresani, L.; Bertasius, G. Video ReCap: Recursive Captioning of Hour-Long Videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 18198–18208. [Google Scholar] [CrossRef]
- Kim, M.; Kim, H.B.; Moon, J.; Choi, J.; Kim, S.T. Do You Remember? Dense Video Captioning with Cross-Modal Memory Retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 13894–13904. [Google Scholar] [CrossRef]
- Ohkawa, T.; Yagi, T.; Nishimura, T.; Furuta, R.; Hashimoto, A.; Ushiku, Y.; Sato, Y. Exo2EgoDVC: Dense Video Captioning of Egocentric Procedural Activities Using Web Instructional Videos. arXiv 2023, arXiv:2311.16444. [Google Scholar] [CrossRef]
- Wu, J.; Wang, J.; Yang, Z.; Gan, Z.; Liu, Z.; Yuan, J.; Wang, L. Grit: A generative region-to-text transformer for object understanding. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 207–224. [Google Scholar] [CrossRef]
- Liu, C.; Hu, C.; Liu, Q.; Aggarwal, J.K. Video event description in scene context. Neurocomputing 2013, 119, 82–93. [Google Scholar] [CrossRef]
- Zhou, X.; Arnab, A.; Sun, C.; Schmid, C. Dense Video Object Captioning from Disjoint Supervision. arXiv 2023, arXiv:2306.11729. [Google Scholar] [CrossRef]
- Xiao, J.; Wu, Y.; Chen, Y.; Wang, S.; Wang, Z.; Ma, J. LSTFE-net: Long short-term feature enhancement network for video small object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 14613–14622. [Google Scholar] [CrossRef]
- Gao, H.; Xu, H.; Cai, Q.-Z.; Wang, R.; Yu, F.; Darrell, T. Disentangling propagation and generation for video prediction. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9006–9015. [Google Scholar] [CrossRef]
- Ramanathan, V.; Liang, P.; Fei-Fei, L. Video event understanding using natural language descriptions. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 905–912. [Google Scholar] [CrossRef]
- Whitehead, S.; Ji, H.; Bansal, M.; Chang, S.-F.; Voss, C. Incorporating background knowledge into video description generation. In Proceedings of the Conference on Empirical Methods in Natural Language, Brussels, Belgium, 31 October–4 November 2018; pp. 3992–4001. [Google Scholar] [CrossRef]
- Guadarrama, S.; Krishnamoorthy, N.; Malkarnenkar, G.; Venugopalan, S.; Mooney, R.; Darrell, T.; Saenko, K. Youtube2text: Recognizing and describing arbitrary activities using semantic hierarchies and zero-shot recognition. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2712–2719. [Google Scholar] [CrossRef]
- Shvetsova, N.; Kukleva, A.; Hong, X.; Rupprecht, C.; Schiele, B.; Kuehne, H. Howtocaption: Prompting llms to transform video annotations at scale. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 1–18. [Google Scholar] [CrossRef]
- Yuan, F.; Gu, S.; Zhang, X.; Fang, Z. Fully exploring object relation interaction and hidden state attention for video captioning. Pattern Recognit. 2025, 159, 111138. [Google Scholar] [CrossRef]
- Xu, T.; Zhang, Y.; Song, X.; Feng, Z.-H.; Wu, X.-J. ATMNet: Adaptive Two-Stage Modular Network for Accurate Video Captioning. IEEE Trans. Syst. Man, Cybern. Syst. 2025, 55, 2821–2833. [Google Scholar] [CrossRef]
- Chen, H.; Li, J.; Hu, X. Delving deeper into the decoder for video captioning. In ECAI 2020; IOS Press: Amsterdam, The Netherlands, 2020; pp. 1079–1086. [Google Scholar] [CrossRef]
- Tan, G.; Liu, D.; Wang, M.; Zha, Z.-J. Learning to discretely compose reasoning module networks for video captioning. arXiv 2020, arXiv:2007.09049. [Google Scholar] [CrossRef]
- Liu, Y.; Zhu, H.; Wu, Z.; Du, S.; Wu, S.; Shi, J. Adaptive semantic guidance network for video captioning. Comput. Vis. Image Underst. 2025, 251, 104255. [Google Scholar] [CrossRef]
- Xu, W.; Xu, Y.; Miao, Z.; Cen, Y.-G.; Wan, L.; Ma, X. CroCaps: A CLIP-assisted cross-domain video captioner. Expert Syst. Appl. 2024, 268, 126296. [Google Scholar] [CrossRef]
- Li, P.; Wang, T.; Zhao, X.; Xu, X.; Song, M. Pseudo-labeling with keyword refining for few-supervised video captioning. Pattern Recognit. 2025, 159, 111176. [Google Scholar] [CrossRef]
- Luo, X.; Luo, X.; Wang, D.; Liu, J.; Wan, B.; Zhao, L. Global semantic enhancement network for video captioning. Pattern Recognit. 2024, 145, 109906. [Google Scholar] [CrossRef]
- Lou, Y.; Zhang, W.; Song, X.; Hua, Y.; Wu, X.-J. EDS: Exploring deeper into semantics for video captioning. Pattern Recognit. Lett. 2024, 186, 133–140. [Google Scholar] [CrossRef]
- Ma, Y.; Zhu, Z.; Qi, Y.; Beheshti, A.; Li, Y.; Qing, L.; Li, G. Style-aware two-stage learning framework for video captioning. Knowl.-Based Syst. 2024, 301, 112258. [Google Scholar] [CrossRef]
- Zhao, H.; Chen, Z.; Yang, Y. Multi-scale features with temporal information guidance for video captioning. Eng. Appl. Artif. Intell. 2024, 137, 109102. [Google Scholar] [CrossRef]
- Rao, Q.; Yu, X.; Li, G.; Zhu, L. CMGNet: Collaborative multi-modal graph network for video captioning. Comput. Vis. Image Underst. 2024, 238, 103864. [Google Scholar] [CrossRef]
- Shen, Y.; Gu, X.; Xu, K.; Fan, H.; Wen, L.; Zhang, L. Accurate and fast compressed video captioning. In Proceedings of the IEEE International Conference on Computer Vision, Paris, France, 23–28 October 2023; pp. 15558–15567. [Google Scholar] [CrossRef]
- Yang, B.; Cao, M.; Zou, Y. Concept-aware video captioning: Describing videos with effective prior information. IEEE Trans. Image Process. 2023, 32, 5366–5378. [Google Scholar] [CrossRef]
- Zeng, P.; Zhang, H.; Gao, L.; Li, X.; Qian, J.; Shen, H.T. Visual commonsense-aware representation network for video captioning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 36, 1092–1103. [Google Scholar] [CrossRef]
- Zhong, X.; Li, Z.; Chen, S.; Jiang, K.; Chen, C.; Ye, M. Refined semantic enhancement towards frequency diffusion for video captioning. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington DC, USA, 7–14 February 2023; Volume 37, pp. 3724–3732. [Google Scholar] [CrossRef]
- Perez-Martin, J.; Bustos, B.; P’erez, J. Improving video captioning with temporal composition of a visual-syntactic embedding. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3039–3049. [Google Scholar] [CrossRef]
- Tu, Y.; Zhou, C.; Guo, J.; Gao, S.; Yu, Z. Enhancing the alignment between target words and corresponding frames for video captioning. Pattern Recognit. 2021, 111, 107702. [Google Scholar] [CrossRef]
- Zheng, Q.; Wang, C.; Tao, D. Syntax-aware action targeting for video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13096–13105. [Google Scholar] [CrossRef]
- Hou, J.; Wu, X.; Zhao, W.; Luo, J.; Jia, Y. Joint syntax representation learning and visual cue translation for video captioning. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8918–8927. [Google Scholar] [CrossRef]
- Xiao, H.; Shi, J. Diverse video captioning through latent variable expansion. Pattern Recognit. Lett. 2022, 160, 19–25. [Google Scholar] [CrossRef]
- Liu, F.; Ren, X.; Wu, X.; Yang, B.; Ge, S.; Zou, Y.; Sun, X. O2NA: An object-oriented non-autoregressive approach for controllable video captioning. arXiv 2021, arXiv:2108.02359. [Google Scholar] [CrossRef]
- Pan, B.; Cai, H.; Huang, D.-A.; Lee, K.-H.; Gaidon, A.; Adeli, E.; Niebles, J.C. Spatio-temporal graph for video captioning with knowledge distillation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10870–10879. [Google Scholar] [CrossRef]
- Tian, M.; Li, G.; Qi, Y.; Wang, S.; Sheng, Q.Z.; Huang, Q. Rethink video retrieval representation for video captioning. Pattern Recognit. 2024, 156, 110744. [Google Scholar] [CrossRef]
- Niu, T.-Z.; Dong, S.-S.; Chen, Z.-D.; Luo, X.; Huang, Z.; Guo, S.; Xu, X.-S. A multi-layer memory sharing network for video captioning. Pattern Recognit. 2023, 136, 109202. [Google Scholar] [CrossRef]
- Xue, P.; Zhou, B. Exploring the Spatio-Temporal Aware Graph for video captioning. IET Comput. Vis. 2022, 16, 456–467. [Google Scholar] [CrossRef]
- Perez-Martin, J.; Bustos, B.; P’erez, J. Attentive visual semantic specialized network for video captioning. In Proceedings of the 2020 25th International Conference on Pattern Recognition, Milan, Italy, 10–15 January 2021; pp. 5767–5774. [Google Scholar] [CrossRef]
- Peng, Y.; Wang, C.; Pei, Y.; Li, Y. Video captioning with global and local text attention. Vis. Comput. 2022, 38, 4267–4278. [Google Scholar] [CrossRef]
- Ryu, H.; Kang, S.; Kang, H.; Yoo, C.D. Semantic grouping network for video captioning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 2514–2522. [Google Scholar] [CrossRef]
- Zhang, Z.; Shi, Y.; Yuan, C.; Li, B.; Wang, P.; Hu, W.; Zha, Z.-J. Object relational graph with teacher-recommended learning for video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13278–13288. [Google Scholar] [CrossRef]
- Li, X.; Zhao, B.; Lu, X. MAM-RNN: Multi-level attention model based RNN for video captioning. In Proceedings of the International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; Volume 2017, pp. 2208–2214. [Google Scholar] [CrossRef]
- Seo, P.H.; Nagrani, A.; Arnab, A.; Schmid, C. End-to-end generative pretraining for multimodal video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17959–17968. [Google Scholar] [CrossRef]
- Tang, M.; Wang, Z.; Liu, Z.; Rao, F.; Li, D.; Li, X. Clip4caption: Clip for video caption. In Proceedings of the ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 4858–4862. [Google Scholar] [CrossRef]
- Chen, H.; Lin, K.; Maye, A.; Li, J.; Hu, X. A semantics-assisted video captioning model trained with scheduled sampling. Front. Robot. AI 2020, 7, 475767. [Google Scholar] [CrossRef]
- Zhang, Z.; Qi, Z.; Yuan, C.; Shan, Y.; Li, B.; Deng, Y.; Hu, W. Open-book video captioning with retrieve-copy-generate network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 9837–9846. [Google Scholar] [CrossRef]
- Zhu, X.; Guo, L.; Yao, P.; Lu, S.; Liu, W.; Liu, J. Vatex video captioning challenge 2020: Multi-view features and hybrid reward strategies for video captioning. arXiv 2019, arXiv:1910.11102. [Google Scholar] [CrossRef]
- Hosseinzadeh, M.; Wang, Y. Video captioning of future frames. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 980–989. [Google Scholar] [CrossRef]
- Man, X.; Ouyang, D.; Li, X.; Song, J.; Shao, J. Scenario-aware recurrent transformer for goal-directed video captioning. ACM Trans. Multimed. Comput. Commun. Appl. 2022, 18, 1–17. [Google Scholar] [CrossRef]
- Estevam, V.; Laroca, R.; Pedrini, H.; Menotti, D. Dense video captioning using unsupervised semantic information. J. Vis. Commun. Image Represent. 2024, 107, 104385. [Google Scholar] [CrossRef]
- Chen, S.; Jiang, Y.-G. Towards bridging event captioner and sentence localizer for weakly supervised dense event captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 8425–8435. [Google Scholar] [CrossRef]
- Duan, X.; Huang, W.; Gan, C.; Wang, J.; Zhu, W.; Huang, J. Weakly supervised dense event captioning in videos. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 3029–3039. [Google Scholar] [CrossRef]
- Wang, J.; Jiang, W.; Ma, L.; Liu, W.; Xu, Y. Bidirectional attentive fusion with context gating for dense video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7190–7198. [Google Scholar] [CrossRef]
- Wei, Y.; Yuan, S.; Chen, M.; Shen, X.; Wang, L.; Shen, L.; Yan, Z. MPP-net: Multi-perspective perception network for dense video captioning. Neurocomputing 2023, 552, 126523. [Google Scholar] [CrossRef]
- Yamazaki, K.; Vo, K.; Truong, Q.S.; Raj, B.; Le, N. Vltint: Visual-linguistic transformer-in-transformer for coherent video paragraph captioning. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington DC, USA, 7–14 February 2023; Volume 37, pp. 3081–3090. [Google Scholar] [CrossRef]
- Luo, H.; Ji, L.; Shi, B.; Huang, H.; Duan, N.; Li, T.; Li, J.; Bharti, T.; Zhou, M. Univl: A unified video and language pre-training model for multimodal understanding and generation. arXiv 2020, arXiv:2002.06353. [Google Scholar] [CrossRef]
- Zeng, Y.; Wang, Y.; Liao, D.; Li, G.; Xu, J.; Man, H.; Liu, B.; Xu, X. Contrastive topic-enhanced network for video captioning. Expert Syst. Appl. 2024, 237, 121601. [Google Scholar] [CrossRef]
- Zhu, L.; Yang, Y. Actbert: Learning global-local video-text representations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8746–8755. [Google Scholar] [CrossRef]
- Sun, C.; Myers, A.; Vondrick, C.; Murphy, K.; Schmid, C. Videobert: A joint model for video and language representation learning. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7464–7473. [Google Scholar] [CrossRef]
- Yang, A.; Nagrani, A.; Laptev, I.; Sivic, J.; Schmid, C. Vidchapters-7m: Video chapters at scale. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 11–17 December 2023; pp. 49428–49444. [Google Scholar] [CrossRef]
- Kim, D.; Piergiovanni, A.J.; Angelova, A. Actions Inspire Every Moment: Online Action-Augmented Dense Video Captioning. 2024. Available online: https://openreview.net/forum?id=oO3oXJ19Pb (accessed on 21 March 2025).
- Zhu, W.; Pang, B.; Thapliyal, A.V.; Wang, W.Y.; Soricut, R. End-to-end dense video captioning as sequence generation. arXiv 2022, arXiv:2204.08121. [Google Scholar] [CrossRef]
- Piergiovanni, A.J.; Kim, D.; Ryoo, M.S.; Noble, I.; Angelova, A. Whats in a Video: Factorized Autoregressive Decoding for Online Dense Video Captioning. arXiv 2024, arXiv:2411.14688. [Google Scholar] [CrossRef]
- Zhou, X.; Arnab, A.; Buch, S.; Yan, S.; Myers, A.; Xiong, X.; Nagrani, A.; Schmid, C. Streaming dense video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–20 June 2024; pp. 18243–18252. [Google Scholar] [CrossRef]
- Xi, Z.; Shi, G.; Li, X.; Yan, J.; Li, Z.; Wu, L.; Liu, Z.; Wang, L. A simple yet effective knowledge guided method for entity-aware video captioning on a basketball benchmark. Neurocomputing 2025, 619, 129177. [Google Scholar] [CrossRef]
- Shoman, M.; Wang, D.; Aboah, A.; Abdel-Aty, M. Enhancing traffic safety with parallel dense video captioning for end-to-end event analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 16–20 June 2024; pp. 7125–7133. [Google Scholar] [CrossRef]
- Duan, Z.; Cheng, H.; Xu, D.; Wu, X.; Zhang, X.; Ye, X.; Xie, Z. Cityllava: Efficient fine-tuning for vlms in city scenario. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 16–20 June 2024; pp. 7180–7189. [Google Scholar] [CrossRef]
- Zhang, H.; Li, X.; Bing, L. Video-llama: An instruction-tuned audio-visual language model for video understanding. arXiv 2023, arXiv:2306.02858. [Google Scholar] [CrossRef]
- Maaz, M.; Rasheed, H.; Khan, S.; Khan, F.S. Video-chatgpt: Towards detailed video understanding via large vision and language models. arXiv 2023, arXiv:2306.05424. [Google Scholar] [CrossRef]
- Li, L.; Chen, Y.-C.; Cheng, Y.; Gan, Z.; Yu, L.; Liu, J. Hero: Hierarchical encoder for video+ language omni-representation pre-training. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Online, 16–20 November 2020; pp. 2046–2065. [Google Scholar] [CrossRef]
- Mahmud, T.; Billah, M.; Hasan, M.; Roy-Chowdhury, A.K. Prediction and description of near-future activities in video. Comput. Vis. Image Underst. 2021, 210, 103230. [Google Scholar] [CrossRef]
- Xu, H.; Li, B.; Ramanishka, V.; Sigal, L.; Saenko, K. Joint event detection and description in continuous video streams. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision, Honolulu, HI, USA, 7–11 January 2019; pp. 396–405. [Google Scholar] [CrossRef]
- Li, J.; Wong, Y.; Zhao, Q.; Kankanhalli, M.S. Video storytelling: Textual summaries for events. IEEE Trans. Multimed. 2019, 22, 554–565. [Google Scholar] [CrossRef]
- Mao, J. Deep captioning with multimodal recurrent neural networks (m-rnn). arXiv 2014, arXiv:1412.6632. [Google Scholar] [CrossRef]
- Baraldi, L.; Grana, C.; Cucchiara, R. Hierarchical boundary-aware neural encoder for video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1657–1666. [Google Scholar] [CrossRef]
- Rohrbach, A.; Rohrbach, M.; Schiele, B. The Long-Short Story of Movie Description. arXiv 2015, arXiv:1506.01698. [Google Scholar] [CrossRef]
- Dinh, Q.M.; Ho, M.K.; Dang, A.Q.; Tran, H.P. Trafficvlm: A controllable visual language model for traffic video captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 17–18 June 2024; pp. 7134–7143. [Google Scholar] [CrossRef]
- Chen, F.; Xu, C.; Jia, Q.; Wang, Y.; Liu, Y.; Zhang, H.; Wang, E. Egocentric Vehicle Dense Video Captioning. In Proceedings of the ACM International Conference on Multimedia, Melbourne, Australia, 28 October–1 November 2024; pp. 137–146. [Google Scholar] [CrossRef]
- Wu, S.; Gao, Y.; Yang, W.; Li, H.; Zhu, G. End-to-End Video Captioning Based on Multiview Semantic Alignment for Human–Machine Fusion. IEEE Trans. Autom. Sci. Eng. 2024, 22, 4682–4690. [Google Scholar] [CrossRef]
- Jin, B.; Liu, X.; Zheng, Y.; Li, P.; Zhao, H.; Zhang, T.; Zheng, Y.; Zhou, G.; Liu, J. Adapt: Action-aware driving caption transformer. In Proceedings of the IEEE International Conference on Robotics and Automation, Yokohama, Japan, 29 May–2 June 2023; pp. 7554–7561. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).