Next Article in Journal
Federated Learning and EEL-Levy Optimization in CPS ShieldNet Fusion: A New Paradigm for Cyber–Physical Security
Previous Article in Journal
Covert Communications via Full-Duplex User Relaying
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Advances in Video Emotion Recognition: Challenges and Trends

1
School of Mathematics and Computer Science, Gannan Normal University, Ganzhou 341000, China
2
Key Laboratory of Data Science and Artificial Intelligence of Jiangxi Education Institutes, Gannan Normal University, Ganzhou 341000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3615; https://doi.org/10.3390/s25123615
Submission received: 30 April 2025 / Revised: 1 June 2025 / Accepted: 6 June 2025 / Published: 9 June 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Video emotion recognition (VER), situated at the convergence of affective computing and computer vision, aims to predict the primary emotion evoked in most viewers through video content, with extensive applications in video recommendation, human–computer interaction, and intelligent education. This paper commences with an analysis of the psychological models that constitute the foundation of VER theory. The paper further elaborates on datasets and evaluation metrics commonly utilized in VER. Then, the paper reviews VER algorithms according to their categories, and compares and analyzes the experimental results of classic methods on four datasets. Based on a comprehensive analysis and investigations, the paper identifies the prevailing challenges currently faced in the VER field, including gaps between emotional representations and labels, large-scale and high-quality VER datasets, and the efficient integration of multiple modalities. Furthermore, this study proposes potential research directions to address these challenges, e.g., advanced neural network architectures, efficient multimodal fusion strategies, high-quality emotional representation, and robust active learning strategies.

1. Introduction

Emotion is a psychophysiological phenomenon [1,2,3]. Video emotion recognition (VER), also known as affective video content analysis, is an interdisciplinary domain where affective computing and computer vision intersect. The objective of VER is to predict the principal emotion of most audiences after they watch a video. VER has received extensive attention on account of its wide-ranging applications [4,5,6], including video recommendation, human–computer interaction, intelligent education, and advertising effectiveness evaluation. By analyzing audiovisual cues, VER systems infer the emotional responses of viewers, enabling more personalized and emotionally intelligent technologies.
The origins of VER are rooted in the pioneering studies of affective computing. Initially focused on recognizing emotions from static images or audio signals, research gradually extended to video data as computational power and machine learning techniques advanced [7]. Videos, unlike static images, provide temporal dynamics and multimodal information, making them richer but also more complex for emotion analysis. In 2005, Hanjalic and Xu [8] proposed a framework to represent and learn the emotional content. Afterward, many researchers began to study the problems in VER. Early studies in the VER field primarily relied on handcrafted features extracted from visual and auditory modalities [9,10]. Because emotions are simultaneously influenced by visual and auditory information, some studies [11,12] focused on fusing audiovisual features to learn complex emotional representations in videos. Recent research in VER has achieved end-to-end learning of emotional representations [13,14] and has explored the utilization of weakly supervised learning techniques [15].
Generally, algorithms for VER can be classified into two distinct groups: (1) handcrafted-feature-based algorithms, which utilize handcrafted emotion features to describe video content; (2) neural-network-based algorithms, which employ neural networks to learn the emotional information of videos from massive data and build emotion models. Despite the rapid progress, VER continues to pose challenges, which are primarily attributed to the subtleties and context-sensitive nature inherent in human emotional expression. The differences in cultural expressions, individual differences in emotional intensity, and the presence of ambiguous or mixed emotions further complicate the recognition process. Furthermore, the temporal dynamics of videos demand that models effectively capture both spatial and temporal dependencies.
This study conducted an in-depth review of recent developments in VER, with a focus on the psychological foundations, datasets, evaluation metrics, and state-of-the-art algorithms. It also identified key challenges and proposed future research directions aimed at addressing these challenges, thereby laying the groundwork for more robust VER systems. The major contributions are summarized below:
  • Comparative analysis of classic algorithms: This study systematically evaluated classic algorithms in the VER field, including the comparison and analysis of their performance on four datasets. This evaluation provides a benchmark reference for researchers to select or refine algorithms for specific VER tasks, while highlighting unresolved challenges that require further innovation.
  • Comprehensive analysis of challenges: This study comprehensively analyzed the pivotal challenges hindering the development of VER, categorizing them into three interdependent dimensions. Through this analysis, this study identified the root causes of current limitations, established a foundation for future research, and bridged theoretical insights with actionable strategies to advance the field.
  • Actionable future research directions: This paper proposes actionable research directions to address the key challenges, thereby promoting the development of the VER field. By systematically linking these directions to the challenges, the paper provides a clear roadmap for future innovations, thereby guiding researchers in designing precise and efficient VER systems.
The subsequent sections are organized as follows: Section 2 and Section 3 present psychological models and benchmark datasets in VER, respectively. Then, Section 4 and Section 5 illustrate methods and results in VER. Moreover, a discussion and conclusion are introduced in Section 6 and Section 7, respectively. Acronyms are reported in Table 1.

2. Psychological Models for VER

In the field of psychology, there are two main types of models used to measure emotions, namely, the discrete emotion model and the dimensional emotion model. The discrete emotion model divides emotions into several relatively independent emotion categories, which has the advantages of simplicity. In 1980, Plutchik [48] proposed that humans have eight basic emotion categories, i.e., anger, anticipation, disgust, fear, joy, sadness, surprise, and trust. In 1992, Ekman [49] introduced six basic emotion categories. Moreover, other emotion categories, such as boredom and excitement, have also been used to describe video emotions [50,51]. Recently, more emotion categories have been reported, e.g., 27 categories [52] and 80 categories [53]. Limited by the number of emotion categories, the discrete emotion model has the following drawbacks. First, the discrete emotion model merely represents a restricted range of emotions. Second, there is a certain correlation between some emotion categories, but it is difficult for the discrete emotion model to measure and represent this correlation. Finally, the discrete emotion model cannot describe the process of emotion generation, development, and disappearance.
Some researchers proposed the dimensional emotion models [54,55], where a multidimensional space is utilized to describe continuous emotions. Specifically, the dimensional emotion model employs a dimension in the multidimensional space to describe a certain aspect of emotion and a coordinate point in the space to represent a certain emotional state. Three dimensions are utilized to describe the emotion model, i.e., valence, arousal, and dominance. Among them, the valence dimension measures the degree of happiness of emotions, ranging from depression to ecstasy. The arousal dimension evaluates the intensity of activation and emotional stimulation, ranging from passive to active. The dominance dimension captures the extent of emotional authority and control, and measures the degree of affecting others or being affected by others, ranging from submission to dominance. As the dominance dimension is hard to measure, the two-dimensional emotion model is generally used in the emotion analysis field, which is called the valence–arousal emotion model. Moreover, Russell [54] demonstrated that emotion categories can be mapped to the dimensional emotion model.

3. VER Datasets

To advance research in VER, some datasets containing diverse and rich emotional content have been created to train and test emotion recognition methods. These datasets include video clips extracted from movies and social media platforms, ensuring a variety of contexts and cultural backgrounds. The diversity in these datasets is crucial for training robust methods that can generalize across different scenarios. Furthermore, they are annotated with emotional labels or continuous emotion dimensions, providing valuable ground truth for supervised learning tasks. Table 2 reports the datasets’ information for VER, where F1, MSE, and PCC denote the F1 score, mean-square error, and Pearson correlation coefficient, respectively.
DEAP [56] is a multimodal emotion dataset that contains testing videos and audience response data. To collect emotional responses, 32 participants were engaged in viewing 40 music video clips, each lasting one minute. Additionally, facial videos were collected from 22 participants. Due to copyright licensing restrictions, DEAP does not provide the music videos. Links for downloading the videos are public.
MAHNOB-HCI [57] is a multimodal dataset for emotion analysis. This dataset was developed through 27 participants from diverse gender and cultural backgrounds, ensuring extensive emotional responses. MAHNOB-HCI comprises 20 videos, each annotated according to the valence–arousal–dominance model to capture emotional states comprehensively. Moreover, the dataset includes various physiological signals, respiratory amplitude, and skin temperature measurements, providing rich data for analyzing human emotional reactions in depth.
VideoEmotion-8 [9] is structured around the Plutchik emotion model, which categorizes human emotions into eight distinct types. The dataset comprises 1101 videos, which exhibit a wide range of content and contextual variations, guaranteeing robust emotional expression coverage. It provides 10 unique splits of videos, each categorized into training and testing subsets. Roughly two-thirds of the samples within each split are designated for model training, with the residual portions allocated for evaluation purposes. To assess the model performance, VideoEmotion employs the mean accuracy across all 10 testing splits as its primary metric. Overall, VideoEmotion offers a well-structured dataset for investigating emotion recognition in video content.
LIRIS-ACCEDE [29,58] employs the valence–arousal dimensional model to describe emotions, and is a classic dataset for evaluating VER algorithms. From 2015 to 2018, in the movie emotional impact competition organized by MediaEval, the competition organizers used LIRIS-ACCEDE as the competition dataset [60,61,62,63]. Discrete LIRIS-ACCEDE contains 9800 videos from 160 movies. The types of these films are diverse. The cumulative runtime of these clips approximates 73 h, with individual clips varying in length between 8 and 12 s. Each clip is given an emotional label on the valence and arousal dimensions. Continuous LIRIS-ACCEDE [64] selects 30 movies from the 160 movies, which are annotated once per second. In the valence and arousal dimensions, MSE and PCC were utilized as evaluation indicators, respectively.
Ekman-6 [26] is built upon the Ekman model. It includes 1637 videos, which have been carefully annotated to represent the six basic emotions outlined by Ekman’s model. The diversity of sources ensures that the dataset captures a wide range of emotional expressions, cultural contexts, and real-world scenarios. Regarding evaluation, it employs accuracy as a metric for assessing the performance. Ekman-6 constitutes a valuable dataset for promoting VER research.
CAER [16] has an extensive collection of videos, comprising 13,201 clips from 79 TV shows. Each video clip within CAER has been annotated into seven distinct emotion categories. CAER provides a sufficient number of large-scale videos to learn models for context-aware VER. Accuracy is the metric for CAER. In summary, CAER is a valuable dataset for advancing research in VER. Its extensive size and emphasis on context-aware analysis make it an essential tool for developing more sophisticated emotion intelligence technologies.
Video-Danmu [59] contains 4056 clips and 371,177 danmus (i.e., synchronized comments), which are collected from an online video platform. The dataset categorizes each video into one of seven affective categories. Precision and accuracy are utilized as key performance indicators for Video-Danmu. Within the realm of VER, Video-Danmu was the first one to provide the text modality, i.e., synchronized comments.
EEV [20] is the most extensive dataset in VER, including about 1700 h across 23,574 individual videos. The proportions of samples in the training, validation, and test sets are approximately 60%, 15%, and 25%, respectively. In 2021, a challenge was held using EEV as the competition dataset, with PCC as the metric. This vast dataset provides researchers with an unparalleled resource for studying human emotions in depth.
MM-AU [5] contains 8399 video samples, covering a wide range of topics and languages. This dataset offers annotations for 18 topic categories, 3 tone categories, and 2 social message categories, rendering it applicable for three tasks, including VER. These annotations were generated utilizing a semi-automatic approach. F1 and accuracy are the metrics for the MM-AU dataset.
VAD [45] is a comprehensive dataset designed for VER. It sources videos from popular Chinese online platforms, e.g., Bilibili. The dataset contains 19,267 clips, each accompanied by synchronized danmu comments that reflect the viewers’ emotional responses to specific moments in videos. These clips are annotated with emotional labels, enabling research on multi-task learning across different emotion models. The evaluation metrics for VAD are accuracy and F1.

4. VER Algorithms

Algorithms for VER utilize visual and audio features to infer the emotional reactions of the predominant audiences. Generally, VER algorithms are categorized into two classifications, i.e., handcrafted-feature-based algorithms and neural-network-based algorithms. As shown in Figure 1, before 2015, most VER algorithms predominantly relied on handcrafted features. However, since 2015, neural networks have been progressively adopted in the VER field, and the use of handcrafted features has gradually decreased in this field. Notably, since 2020, end-to-end neural networks have been designed to address the challenges in the VER domain. In recent years, neural networks have become the mainstream and hot technology in this field.
Although the use of handcrafted features is gradually decreasing in this field, they remain effective in scenarios with limited data, constrained computational resources, or high interpretability requirements. With the advent of big data and powerful GPUs, neural networks have demonstrated superior generalization in high-dimensional data (e.g., images, videos) through end-to-end feature learning. Deep neural networks can automatically extract hierarchical features while minimizing manual intervention. Therefore, handcrafted features retain competitiveness in small-sample tasks or lightweight applications, whereas deep learning excels in large-scale and complex tasks.

4.1. Handcrafted-Feature-Based Algorithms

In VER, handcrafted-feature-based algorithms utilize handcrafted design features to describe videos, and their core part is video feature extraction and emotion learning. Figure 2 shows a diagram of a classic algorithm [10] based on handcrafted features. Generally, videos contain visual and audio data, from which visual features and audio features can be extracted, respectively. The following will introduce the relevant algorithms in visual emotion features, audio emotion features, and emotion learning.

4.1.1. Visual Emotion Features

Visual features are essential for describing emotions. Based on their capacity to characterize semantic content, visual elements are broadly categorized into LVFs and HVFs. LVFs primarily capture fundamental attributes of video data, such as color, texture, and shape. For VER, commonly used LVFs include GIST [67], LBP [28], HSH [25], DSIFT [19], HOG [24], and self-similarity [68].
HVFs distinguish themselves from LVFs by encoding meaning rather than basic data. While LVFs focus on rudimentary traits, HVFs represent abstract concepts, such as sentiment information, scene information, object information, and motion information. For VER, HVFs mainly encompass Classemes [69], ObjectBank [70], SentiBank [65], MKT [71], and TSN [44].
Generally, different visual features possess distinct capabilities for describing emotions. To determine which visual features can describe video emotions more effectively, comparative experiments were performed on several datasets. On VideoEmotion-8, Jiang et al. [9] compared four LVFs, i.e., GIST, LBP, HOG, and DSIFT. Among them, they discovered that DSIFT obtained relatively better experimental outcomes. On LIRIS-ACCEDE, Baecchi et al. [72] evaluated five HVFs, namely, VGG-FC7 [46], VGG-FC8 [46], Sent-FC7 [73], Sent-FC8 [73], and deep face [74], and found that Sent-FC8 better characterized the emotions in videos. Moreover, Guo et al. [75] also assessed two features based on the VGG [46] network, i.e., OVGG and OFVGG [75], and discovered that OFVGG achieved relatively better experimental results. Furthermore, Yi et al. [10] appraised four visual features, namely, DSIFT, HSH, MKT, and TSN, and found that TSN and MKT performed better in the two dimensions separately.
According to the above findings, several key conclusions were drawn. First, HVFs generally demonstrate a superior ability to describe emotions compared with LVFs. Second, the integration of diverse feature types tends to enhance performance. While expanding the quantity of features alone may not guarantee enhanced performance, the integration of complementary features (e.g., combining low-level texture details with high-level object semantics) promotes predictive accuracy. By leveraging diverse but complementary information, fusion mitigates individual feature limitations, captures multidimensional correlations, and reduces the influence of any single feature.

4.1.2. Audio Emotion Features

The sound in videos can affect the emotions of audiences. In VER, audio features are also crucial. Audio data contains both semantic information and non-semantic information, each of which contributes to emotional perception. The semantic information of audio generally refers to the semantic content conveyed by the sound, while the non-semantic information of audio mainly includes the tone, melody, rhythm, timbre, and speech rate.
Generally, audio features are classified into two categories, namely, LAFs and HAFs. In previous VER studies, LAFs mainly included MFCC [31] and ZCR [47]. HAFs are generally based on LAFs and describe audio data at a higher level. For VER, HAFs mainly include EmoLarge [76], EmoBase10 [77], IS13 [78], GeMAPS [23], SoundNet [79], and VGGish [80].
Different audio features have varying abilities to describe emotions. To find out which audio features are more suitable for emotional description, LIRIS-ACCEDE was employed by the paper [10] to evaluate four audio features, namely, MFCC, EmoLarge, EmoBase10, and IS13. Through the comparative experiments, EmoBase10 attained relatively superior performance. Moreover, Guo et al. [75] compared seven audio features, namely, MFCC, IS13, EmoBase10, EmoLarge, GeMAPS, SoundNet, and VGGish, and discovered that SoundNet and VGGish obtained superior performance in the arousal and valence dimensions, respectively.
The following inferences can be drawn. First, HAFs generally tend to yield better performance than LAFs. This is because HAFs are generally constructed based on LAFs, which improves their generalization capability. Second, exploring the extraction of HAFs from large datasets is a valuable research direction.

4.1.3. Emotion Learning

Handcrafted-feature-based algorithms can model emotional content through classic machine learning algorithms. The learning algorithms employed in previous relevant studies mainly include HMM, SVM [41], RBM [38], MLP [33], LAR [27], and passive aggressive [81].
According to the valence–arousal model, Xu et al. [82] designed a hierarchical method and learned emotional information by utilizing HMM. Baecchi et al. [72] adopted emotion-related features to depict videos and used SVM to learn emotional content. On LIRIS-ACCEDE, Yi et al. [10] evaluated four learning algorithms and discovered that the method using SVM obtained relatively better performance under the same experimental conditions. In the arousal and valence dimensions, Guo et al. [75] utilized the passive-aggressive and least-angle regression algorithms to predict emotion, respectively. Because SVM possesses strong generalization capabilities, many VER algorithms utilize SVM to learn emotion models.
In summary, approaches utilizing handcrafted features and traditional machine learning algorithms have achieved some success in the domain of VER. According to visual theory, researchers designed features and customized them for the characteristics of video data. While handcrafted features offer high customizability, these features rely on shallow linear structures and have difficulty in effectively modeling large-scale and complex data. On large-scale datasets, such algorithms often attain less-than-satisfactory performance. Therefore, neural networks have been increasingly adopted in this field.

4.2. Neural-Network-Based Algorithms

In VER, neural-network-based algorithms describe video content through neural networks, with the core being the design and training of neural networks. Depending on whether there exists an independent feature extraction module, these algorithms are typically classified into two major categories, i.e., two-stage VER algorithms and end-to-end VER algorithms.

4.2.1. Two-Stage VER Algorithms

According to the order of video data being modeled, the two-stage VER algorithms primarily include two steps. Initially, feature vectors are derived from video data. Subsequently, a neural network architecture is constructed, and the model undergoes training utilizing these vectors as input. Some classic algorithms are introduced below.
Generally, there exists a nonlinear relationship between different modes. E-MDBM [21] was designed to model nonlinear relationships between modalities. E-MDBM contains three separate paths (i.e., vision, auditory, and text), with each composed of stacked restricted Boltzmann machines to learn modality-specific representations. By combining these paths, the model captures the nonlinear relationship between modalities. The final representation is a shared embedding space where statistical properties from different modalities are unified. In this representation, high-level semantic correlations between the three paths are leveraged to represent videos.
To effectively fuse event, object, and scene features, Chen et al. [18] designed CFN. Pre-trained models from external datasets are leveraged to extract vectors of the three features. Following feature extraction, these vectors are processed by using L1 and RootSift [83] normalizations. These normalized features are then integrated through CFN to produce accurate predictions. CFN mainly contains linear and softmax layers.
To mine the correlation between visual and audio modalities, MMDRBN [36] was designed. By minimizing the KL divergence, MMDRBN is transformed into a multimodal inference deep network. During training, the back propagation strategy is utilized to optimize MMDRBN. During testing, the trained model is employed to predict emotional labels of videos. To address the limitations of MMDRBN, Wang et al. [84] proposed an improved version called knowledge-augmented MMDRBN. By constructing a representation that bridges visual and auditory elements with semantic attributes, the model achieves a more comprehensive understanding of emotional expression in videos.
The labels of VER datasets have inherent noise, which poses a significant challenge for obtaining reliable supervised data to train emotion models effectively. MMDDN [34] was devised to solve this problem. By leveraging the embedding network, MMDDN fuses rich multimodal information. LSTM [30] is used to predict the label of a sample. Moreover, MMDQEN [35] further improved MMDDN. MMDQEN adopts a non-parametric approach instead of relying on LSTM, which significantly reduces the computational burden associated with fusing temporal information. By inferring potential labels from noisy training samples, MMDQEN provides more accurate annotations for emotional classifiers.
Humans perceive the world through multimodal information, and their emotional states are influenced by previously encountered visual scenes and auditory cues. Therefore, understanding the relative importance of multiple modalities and the temporal relationships within inputs is crucial for accurately predicting emotions in videos. To leverage these relationships effectively, AFRN was proposed by [11]. As illustrated in Figure 3, three layers were designed for extracting robust input features, integrating temporal information, and combining information across different modalities.
To address the challenge of multimodal fusion in VER, MMLGAN [37] was introduced. This approach extends traditional attention mechanisms to facilitate multi-level data integration and enhances the representation of videos by designing a multimodal fusion unit. This unit operates in two stages: local and global attention. The local attention stage selects key components from various streams, ensuring that the most relevant features are highlighted. During the global attention stage, it captures the temporal distribution of information, providing a comprehensive understanding of the video content over time. MMLGAN selectively emphasizes crucial emotional elements, thereby improving the overall representation of emotions in videos.
By leveraging self-attention to capture relationships across modalities and time, AttendAffectNet [66] was designed to systematically analyze viewers’ emotional reactions. Specifically, three variants were designed. The feature AttendAffectNet, which utilizes self-attention on multimodal features to capture inter-modal relationships, achieves superior performance between the three variants. Through the innovative usage of self-attention mechanisms, AttendAffectNet learns the complex interactions between different modalities, thereby enhancing the accuracy of emotional response prediction.
To address the issue of ignoring contextual cues in videos, CAF [17] was proposed. Key regions are extracted to capture the emotional cues contained in videos. RPN is utilized to derive features from key regions, facilitating the construction of an emotional similarity graph. To enhance the effectiveness, FNN is used to allocate weights to diverse regions based on their emotional significance, followed by a GCN, which elucidates the interconnections between these key regions. Additionally, MFCC is extracted from the auditory modality to complement the visual modality.
In the EEV challenge held in 2021, methods [85,86,87] won first, second, and third places, respectively. Among them, Huynh et al. [85] used EfficientNet [88] and TRILL [43] to calculate visual and auditory features. The temporal convolutional network was designed to learn temporal relationships. Lin et al. [86] utilized Swin-L [89] and VGGish to extract visual and auditory features. Meanwhile, two layers of bidirectional GRUs are proposed to build the emotion model. Yan et al. [87] used Inception-Resnet-v2 [90] and S3D [39] to calculate visual characteristics, employed VGGish to extract auditory features, and proposed a fusion block to combine the above characteristics.
By integrating visual–audio representations, deep graph fusion [91] was proposed for estimating the evoked expressions of viewers. First, salient feature vectors are extracted from videos via pre-trained models. Second, the vectors are fed to the graph structure and processed through a GCN to generate node embeddings. Third, integration mechanisms are applied to synthesize the depictions from the audiovisual pathways. Ultimately, the resultant embeddings are leveraged to estimate scores of samples. Moreover, a semantic embedding loss is incorporated to enhance the performance.
By fusing multimodal features from multiple stages, UMFN [12] was devised to improve AFRN. As shown in Figure 4, three modalities of video are utilized as the input into UMFN, and the unified fusion layer was designed to fuse the output information. Multiple modalities are integrated by UMFN, thereby enhancing the model’s capability to characterize video-based emotional content.
Based on weakly supervised learning, CTEN [15] was proposed to equip the network with contextual awareness for interpreting emotional expressions. The 3D Resnet-101 and 2D Resnet-50 are utilized to calculate characteristics for audiovisual streams. Then, a module is introduced to learn the temporal relationship of two modalities. By choosing keyframes, the model concentrates on contexts containing complementary information.
To overcome the limitations of MLLMs, which predominantly focus on semantic video content, an approach named StimuVAR [40] has been introduced. This approach includes two-level awareness, i.e., frame-level awareness, which involves sampling video frames likely to elicit emotional responses, and token-level awareness, which executes tube selection to focus on emotion-triggered spatiotemporal regions. Additionally, instruction data has been created to conduct affective training, guiding MLLMs’ reasoning capabilities toward emotional focus.
To reduce the noise in the temporal domain, TE [42] is introduced. This method takes motion, semantic, and audio modalities as the input. A module is designed to enhance the temporal information of the modalities. Through the interaction with these modalities, this module enhances the key temporal information and suppresses the irrelevant information.
In summary, the above algorithms use the two-stage framework to build a video emotion model. The primary research directions in this domain concentrate on three main aspects: first, exploring or identifying features that are more suitable for describing emotions in videos; second, researching more efficient temporal modeling algorithms; and third, investigating improved methods for effectively fusing multiple features.

4.2.2. End-to-End VER Algorithms

Unlike two-stage VER algorithms, end-to-end VER algorithms implement the extraction of video features and the construction of emotion models in a network and ultimately achieve the prediction of video labels. Due to limitations such as GPU memory and sample data size, it is challenging to jointly train long-range visual and audio models in a single network.
VAANet [13] was designed to implement an end-to-end scheme for VER. In VAANet, three attention strategies are integrated into the visual 3D CNN. Concurrently, temporal attentions are added to the audio 2D CNN. Under the polarity–emotion hierarchy constraint, a loss function is formulated to steer the attention mechanism. This network constitutes the first end-to-end model within the domain of VER.
A network [22] was designed to learn frame-level emotional information. Regarding the architecture, ResNet-50 [92] was selected as the backbone. Frame-level features are fused via the pooling method to create video-level representations. Through the alignment of cross-domain features, the network can obtain knowledge from the source dataset while maintaining adaptability to the target video frames.
To address the challenge of encoding long-range contextual correlations in videos, LRCANet [4] was proposed. A diagram of LRCANet is visualized in Figure 5. A spatio-temporal correlation-aware block was devised to capture long-range relations between input tokens, where local correlations are learned through convolutional layers, and spatio-temporal relationships are learned by the inter-image cross-attention. To enhance sample diversity, a dual-augmentation fusion layer is introduced, which integrates each frame with its corresponding temporal counterpart. Furthermore, a long-range sampling layer is devised to create samples spanning extensive spatial and temporal domains, ensuring rich and varied representations.
To address the long-standing issue of limited training data for VER, a masked learning framework named MART [14] was proposed. The core idea is to learn robust VER representations based on the MAE paradigm. First, emotional cues are extracted, and the reliability of emotional cues is verified by calculating the matching degree between emotional dictionaries and videos. Second, a masked strategy is proposed to reconstruct the temporal distribution of the masked segments. Finally, cross-modal attention is utilized to construct a complementary learning block. Without requiring the incorporation of additional large-scale datasets, MART learns the emotional cues from video content.
Several challenges in the field of VER have been addressed through various methods, which have yielded significant advancements and established foundational approaches. However, certain limitations and unresolved issues still persist within this domain, requiring further exploration and refinement. This work undertakes an examination of these existing problems, with a particular focus on identifying potential research directions for future investigations.

5. Results

In the field of VER, the datasets frequently employed to assess algorithm performance are the MediaEval 2015 task of LIRIS-ACCEDE, EEV, VideoEmotion-8, and Ekman-6. On these four datasets, this work compared the experimental results of a series of classic methods and drew relevant inferences. Specifically, Table 3 shows the results of classic methods on the MediaEval 2015 task of LIRIS-ACCEDE, where ACC denotes accuracy. Table 4 compares the PCC values of related algorithms on EEV, Table 5 reports the accuracy values of relevant methods on VideoEmotion-8, and Table 6 summarizes the accuracy values of classic methods on Ekman-6. Moreover, the experimental results presented in these tables were sourced from the corresponding papers of these methods.
As summarized in Table 3, experimental outcomes indicate diverse performance across different approaches for VER. The majority of methods leverage both visual and audio modalities, with a few relying solely on visual features. In terms of the arousal dimension, the accuracy ranged from 45.04% to 60.88%, while the valence prediction presented a range of 36.12% to 48.61%. Notably, methods such as MLG-S and AFRN achieved superior performances, where AFRN reached 58.22% for arousal and 48.61% for valence. This indicates that multimodal approaches, particularly those integrating advanced deep learning architectures, tended to outperform the unimodal methods. The results also demonstrate a gradual improvement in accuracy over time, likely due to advancements in model design and feature extraction techniques. However, compared with arousal, the performance of valence prediction was relatively lower, which might suggest inherent challenges or dataset biases in this dimension. Overall, these findings highlight the effectiveness of multimodal fusion strategies and the potential for further improvements in VER models.
Table 4 presents the results of classic methods on EEV, with the results measured using PCC. All methods utilize both visual and audio modalities. The PCC values ranged from 0.00819 to 0.05477 for TCN, indicating that while all the methods showed some ability to capture emotional correlations, their effectiveness is limited. Specifically, TCN achieved the highest PCC scores, indicating that the architecture of TCN better models the temporal or multimodal information of emotions on this dataset. Despite employing both visual and audio modalities, these methods exhibited modest performances on EEV, highlighting the complexity of VER in this context.
Table 5 reports the results of classic methods on the VideoEmotion-8 dataset, with accuracy values that ranged from 40.50% to 59.39%. Notably, TE [42] achieved the highest accuracy at 59.39%, closely followed by FAEIL [22], LRCANet [4], and CTEN [15] at 57.63%, 57.40%, and 57.30%, respectively. Most methods concurrently utilized audiovisual modalities, underscoring the critical role of cross-modal fusion within the VER framework. Methods relying only on the visual modality, e.g., KeyFrame [106] and DFAN [107], generally performed less effectively, where they achieved performances of 52.85% and 53.34%, respectively. This underscores the benefit of combining multiple modalities to enhance the recognition accuracy. In conclusion, technologies focusing on multimodal fusion, temporal modeling, and advanced network architectures play a crucial role in achieving outstanding results.
As reported in Table 6, the performance reveals significant variations across different approaches, where FAEIL [22] achieved the highest accuracy at 60.44%, followed closely by LRCANet [4] and KeyFrame [106] at 59.78% and 59.51%, respectively. Notably, methods such as CFN [18], DFAN [107], and KeyFrame [106], which rely solely on the visual modality, still achieved relatively high performances, suggesting that visual features alone can be highly effective on this dataset. Overall, these discoveries accentuate the significance of model design and feature extraction strategies, along with the potential advantages of concentrating on visual information in some circumstances.
The aforementioned methods mainly adopted three types of network architectures, i.e., LSTM, a CNN, and Transformer. In the field of VER, the comparison of the three types of architectures is presented below:
  • LSTMs leverage a recurrent structure to process data sequentially, capturing long-term temporal dependencies. In the VER field, LSTMs learn contextual information from previous frames to infer emotions. However, long videos may still suffer from partial forgetting. Moreover, LSTMs are limited in spatial modeling, and their sequential processing leads to slow training.
  • CNNs use kernels to extract local features and build hierarchical abstractions via stacked layers, which enables them to learn video content efficiently. Their limitations include limited receptive fields and requiring deep stacking for long-range dependencies. In practice, the kernel size and depth must be carefully designed, and 3D convolutions are essential for video data.
  • Transformers employ self-attention to learn dependencies across elements in a video, thereby establishing cross-temporal relationships. Their primary strength lies in building long-range dependency models. However, their quadratic complexity in attention matrices limits long-sequence processing. In practice, local or sparse attention variants can reduce costs, and positional encodings are critical for videos.
Attention mechanisms have achieved significant progress in affective computing, and various related techniques have been customized for specific task requirements. For example, MMLGAN [37] captures the correlations between multiple modalities through a fusion block and an improved attention mechanism. AttendAffectNet [66] explores feature-level and temporal-level interactions of multimodal inputs via self-attention. Moreover, the hierarchical attention [108] is devised to improve discriminative features, thereby enabling the effective fusion of multiple features. This diversification of these techniques highlights the flexibility of attention mechanisms and emphasizes the necessity of domain-specific optimization in the context of affective computing.
From these tables, the following conclusions can be inferred. First, methods that utilize visual and audio modalities tend to achieve higher performance compared with those relying solely on a single modality. Second, the design of the model architecture plays a vital role in determining the performance. Advanced architectures, such as those integrating temporal modeling and multimodal fusion techniques, surpass simpler models. Third, although many studies have made a series of advances in this field, the accuracy of VER remains relatively low.

6. Discussions

6.1. Challenges

The domain of VER has witnessed notable advancements over the past decade. However, there are still several key scientific challenges that require further exploration and research, as follows.

6.1.1. Gaps Between Emotional Representations and Labels

In terms of theoretical research, the direct connection between video emotional representations and labels remains unclear, which makes it difficult to attain better experimental results on benchmark datasets. Human emotions are often complex and context-dependent. VER requires capturing subtle emotional cues from both visual and auditory modalities, but existing emotional representations may overlook complex emotional cues.
Moreover, current models may not be able to fully capture context. Similar visual or auditory cues can evoke different emotions in different contexts, leading to prediction errors when the context is ignored. Therefore, one of the critical challenges in VER is the emotional gaps between representations and labels. How to design an excellent emotional network model to establish a direct connection between representations and labels is a problem worthy of further research.

6.1.2. Large-Scale and High-Quality VER Datasets

Compared with objective tasks, such as image classification, building large-scale, high-quality VER datasets poses significant challenges. The subjectivity of emotions introduces great complexity in data annotation, as different annotators may have different perceptions of the same video due to psychological or cultural differences. This inherent subjectivity can lead to potential label inconsistencies, which can reduce the generalization ability if left unaddressed.
The main research significance lies in developing innovative methods to effectively annotate large-scale VER datasets while minimizing label noise. Given the limitations of human and financial resources, traditional manual annotation methods are impractical for constructing large-scale datasets. Therefore, exploring intelligent algorithms to assist the annotation process becomes crucial. Such algorithms may include active learning strategies and weakly supervised strategies.
Furthermore, addressing the label noise problem is essential for maintaining data quality. Techniques such as multi-annotator consensus mechanisms or statistical methods for identifying and correcting noisy labels can significantly improve the reliability of VER datasets. These advancements not only benefit VER but also extend to other domains involving subjective evaluations, making this research direction highly valuable in the broader application of machine learning and affective computing. In conclusion, developing efficient annotation methods and noise reduction strategies for large-scale VER datasets is crucial for advancing VER technologies.

6.1.3. Efficient Integration of Multiple Modalities

Videos typically contain both visual and auditory modalities, which jointly influence the emotions of viewers. Current research has not fully explored the intricate correlations between multiple modalities, particularly in three aspects, namely, temporal alignment, dynamic modality weighting, and noise control.
First, temporal alignment between modalities presents significant complexities. For instance, video data is typically sampled at 30 FPS, while audio is captured at 16 kHz, leading to discrepancies in time-sensitive emotional cues. This mismatch can degrade the recognition accuracy. To address this, dynamic time warping mechanisms can be employed to align multimodal temporal sequences and ensure consistency at critical time points, mitigating errors caused by sampling rate differences.
Second, the contribution of modalities to emotional expression varies. For example, visual modality is more informative for detecting anger or happiness, while audio modality is critical for identifying sadness or anxiety. Assigning equal weights to all modalities may lead to suboptimal performance by overlooking dominant features. Therefore, dynamic weighting strategies, such as attention mechanisms or uncertainty modeling, may be essential to adaptively adjust the modality importance.
Third, multimodal fusion often introduces redundant or conflicting information. For example, background noise in videos may confuse visual feature extraction, irrelevant text could mislead emotional judgments, and imbalances in data volume across modalities may cause overfitting to noise. To alleviate these issues, techniques like sparse representations can filter redundant features, and cross-modal filtering mechanisms can eliminate contradictions, thus enhancing the model’s robustness.
Through the above strategies, multiple modalities can be efficiently integrated, thereby building a more accurate VER system. The significance of the research lies in developing an efficient multimodal fusion architecture to handle the aforementioned issues. These advancements will not only improve the performance of VER models but also contribute to a deeper understanding of multimedia content. By addressing this challenge, researchers can create more powerful VER systems capable of capturing subtle emotional cues in real-life situations. This will significantly promote progress in fields such as emotion-centric technologies and personalized content recommendation.

6.2. Future Work

Future orientations in VER focus on addressing the key challenges, including the gaps between emotional representations and labels, the efficient integration of multiple modalities, and the establishment of high-quality datasets. These challenges are critical to advancing VER technologies and improving their real-world applications.
First, bridging the gap between emotional representations and labels requires innovative approaches to better capture the nature of human emotions. Future research could explore advanced neural network architectures, such as graph-based models or hierarchical attention [108], to model the subtle relationships between visual and auditory cues and their corresponding labels. Furthermore, integrating emotional psychological theories into the design of representation learning frameworks could help create more interpretable and effective VER systems.
Second, building large-scale and high-quality VER datasets is essential for training robust models. To address the subjectivity of emotions and reduce label noise, researchers should explore active learning strategies to selectively annotate samples with high uncertainty and utilize weakly supervised techniques to generate reliable labels on a large scale. Additionally, exploring automated annotation methods, such as using pre-trained models or crowdsourcing frameworks with quality control mechanisms, could significantly alleviate the burden of manual annotation. Methods for identifying and mitigating noisy labels, such as multi-annotator consensus approaches or statistical noise reduction approaches, should also be essential to ensure the reliability of datasets.
Third, enhancing the integration of multiple modalities is crucial for promoting the performance of VER systems. Future research should focus on designing multimodal fusion architectures that can effectively align and integrate visual and auditory modalities in both spatial and temporal domains. For example, cross-modal Transformers or attention-based mechanisms could be developed to capture the complementarity of multimodalities and model their temporal relationships more accurately. Moreover, exploring modality-specific processing techniques, such as separate encoders for visual and auditory inputs followed by adaptive fusion layers, could help maximize the contribution of each modality to VER.
In conclusion, addressing these challenges will not only advance VER technologies but also have a broader influence on emotion-centric applications and multimedia analysis. By exploring more sophisticated models, datasets, and multimodal integration strategies, researchers can create VER systems that are more accurate, reliable, and capable of capturing emotions expressed in videos.

7. Conclusions

VER is a crucial research domain in affective computing, possessing considerable social significance and vast application potential. Many scholars have attained research outcomes in the aspects of emotional psychological models, VER datasets, and VER algorithms. Regarding the aforementioned three aspects, this paper elaborately expounds on the relevant research advancements in the field of VER, filling the gap in the survey of this domain.
On the basis of investigations and analysis, this paper puts forward the current challenges confronted in this field, including gaps between emotional representations and labels, large-scale and high-quality VER datasets, and the efficient integration of multiple modalities. Moreover, this paper predicts the potential research directions, such as advanced neural network architectures, efficient multimodal fusion strategies, high-quality emotional representation, and robust active learning strategies.

Author Contributions

Conceptualization, Y.Y.; investigation, Y.Z.; resources, Y.Y. and T.W.; writing—original draft preparation, Y.Z. and J.Z.; writing—review and editing, Y.Z., Y.Y., T.W. and J.Z.; administration, Y.Y.; funding acquisition, Y.Y. and T.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant No. 62362003), and the Natural Science Foundation of Jiangxi Province (Grant No. 20232BAB202017, 20242BAB26024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Goel, S.; Jara-Ettinger, J.; Ong, D.C.; Gendron, M. Face and context integration in emotion inference is limited and variable across categories and individuals. Nat. Commun. 2024, 15, 2443. [Google Scholar] [CrossRef] [PubMed]
  2. Nomiya, H.; Shimokawa, K.; Namba, S.; Osumi, M.; Sato, W. An artificial intelligence model for sensing affective valence and arousal from facial images. Sensors 2025, 25, 1188. [Google Scholar] [CrossRef] [PubMed]
  3. Sun, W.; Yan, X.; Su, Y.; Wang, G.; Zhang, Y. MSDSANet: Multimodal emotion recognition based on multi-stream network and dual-scale attention network feature representation. Sensors 2025, 25, 2029. [Google Scholar] [CrossRef] [PubMed]
  4. Yi, Y.; Zhou, J.; Wang, H.; Tang, P.; Wang, M. Emotion recognition in user-generated videos with long-range correlation-aware network. IET Image Process. 2024, 18, 3288–3301. [Google Scholar] [CrossRef]
  5. Bose, D.; Hebbar, R.; Feng, T.; Somandepalli, K.; Xu, A.; Narayanan, S. MM-AU: Towards multimodal understanding of advertisement videos. In Proceedings of the ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 86–95. [Google Scholar]
  6. Antonov, A.; Kumar, S.S.; Wei, J.; Headley, W.; Wood, O.; Montana, G. Decoding viewer emotions in video ads. Sci. Rep. 2024, 14, 26382. [Google Scholar] [CrossRef]
  7. Khare, S.K.; Blanes-Vidal, V.; Nadimi, E.S.; Acharya, U.R. Emotion recognition and artificial intelligence: A systematic review (2014–2023) and research recommendations. Inf. Fusion 2024, 102, 102019. [Google Scholar] [CrossRef]
  8. Hanjalic, A.; Xu, L.Q. Affective video content representation and modeling. IEEE Trans. Multimed. 2005, 7, 143–154. [Google Scholar] [CrossRef]
  9. Jiang, Y.G.; Xu, B.; Xue, X. Predicting emotions in user-generated videos. In Proceedings of the Association for the Advancement of Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; pp. 73–79. [Google Scholar]
  10. Yi, Y.; Wang, H. Multi-modal learning for affective content analysis in movies. Multimed. Tools Appl. 2019, 78, 13331–13350. [Google Scholar] [CrossRef]
  11. Yi, Y.; Wang, H.; Li, Q. Affective video content analysis with adaptive fusion recurrent network. IEEE Trans. Multimed. 2020, 22, 2454–2466. [Google Scholar] [CrossRef]
  12. Yi, Y.; Wang, H.; Tang, P. Unified multi-stage fusion network for affective video content analysis. Electron. Lett. 2022, 58, 795–797. [Google Scholar] [CrossRef]
  13. Zhao, S.; Ma, Y.; Gu, Y.; Yang, J.; Xing, T.; Xu, P.; Hu, R.; Chai, H.; Keutzer, K. An End-to-End visual-audio attention network for emotion recognition in user-generated videos. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 303–311. [Google Scholar]
  14. Zhang, Z.; Zhao, P.; Park, E.; Yang, J. MART: Masked affective representation learning via masked temporal distribution distillation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–18 June 2024; pp. 12830–12840. [Google Scholar]
  15. Zhang, Z.; Wang, L.; Yang, J. Weakly supervised video emotion detection and prediction via cross-modal temporal erasing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 18888–18897. [Google Scholar]
  16. Lee, J.; Kim, S.; Kim, S.; Park, J.; Sohn, K. Context-aware emotion recognition networks. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10143–10152. [Google Scholar]
  17. Cheng, H.; Tie, Y.; Qi, L.; Jin, C. Context-aware based visual-audio feature fusion for emotion recognition. In Proceedings of the International Joint Conference on Neural Networks, Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  18. Chen, C.; Wu, Z.; Jiang, Y.G. Emotion in context: Deep semantic feature fusion for video emotion recognition. In Proceedings of the ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 127–131. [Google Scholar]
  19. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  20. Sun, J.J.; Liu, T.; Cowen, A.S.; Schroff, F.; Adam, H.; Prasad, G. EEV: A large-scale dataset for studying evoked expressions from video. arXiv 2021, arXiv:2001.05488. [Google Scholar]
  21. Pang, L.; Zhu, S.; Ngo, C.W. Deep multimodal learning for affective analysis and retrieval. IEEE Trans. Multimed. 2015, 17, 2008–2020. [Google Scholar] [CrossRef]
  22. Zhang, H.; Xu, M. Recognition of emotions in user-generated videos through frame-level adaptation and emotion intensity learning. IEEE Trans. Multimed. 2023, 25, 881–891. [Google Scholar] [CrossRef]
  23. Eyben, F.; Scherer, K.R.; Schuller, B.W.; Sundberg, J.; André, E.; Busso, C.; Devillers, L.Y.; Epps, J.; Laukka, P.; Narayanan, S.S.; et al. The Geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing. IEEE Trans. Affect. Comput. 2015, 7, 190–202. [Google Scholar] [CrossRef]
  24. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–26 June 2005; pp. 886–893. [Google Scholar]
  25. Sural, S.; Qian, G.; Pramanik, S. Segmentation and histogram generation using the HSV color space for image retrieval. In Proceedings of the IEEE International Conference on Image Processing, Rochester, NY, USA, 22–25 September 2002; pp. 589–592. [Google Scholar]
  26. Xu, B.; Fu, Y.; Jiang, Y.; Li, B.; Sigal, L. Heterogeneous knowledge transfer in video emotion recognition, attribution and summarization. IEEE Trans. Affect. Comput. 2018, 9, 255–270. [Google Scholar] [CrossRef]
  27. Efron, B.; Hastie, T.; Johnstone, I.; Tibshirani, R. Least angle regression. Ann. Stat. 2004, 32, 407–499. [Google Scholar] [CrossRef]
  28. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  29. Baveye, Y.; Dellandrea, E.; Chamaret, C.; Chen, L. LIRIS-ACCEDE: A video database for affective content analysis. IEEE Trans. Affect. Comput. 2015, 6, 43–55. [Google Scholar] [CrossRef]
  30. Schmidhuber, J.; Hochreiter, S. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar]
  31. Davis, S.B.; Mermelstein, P. Evaluation of acoustic parameters for monosyllabic word identification. J. Acoust. Soc. Am. 1978, 64, S180–S181. [Google Scholar] [CrossRef]
  32. Ji, X.; Dong, Z.; Zhou, G.; Lai, C.S.; Qi, D. MLG-NCS: Multimodal local–global neuromorphic computing system for affective video content analysis. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 5137–5149. [Google Scholar] [CrossRef]
  33. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  34. Zhu, Y.; Chen, Z.; Wu, F. Multimodal deep denoise framework for affective video content analysis. In Proceedings of the ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 130–138. [Google Scholar]
  35. Zhu, Y.; Chen, Z.; Wu, F. Affective Video Content Analysis via Multimodal Deep Quality Embedding Network. IEEE Trans. Affect. Comput. 2022, 13, 1401–1415. [Google Scholar] [CrossRef]
  36. Gan, Q.; Wang, S.; Hao, L.; Ji, Q. A multimodal deep regression bayesian network for affective video content analyses. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5113–5122. [Google Scholar]
  37. Ou, Y.; Chen, Z.; Wu, F. Multimodal local-global attention network for affective video content analysis. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1901–1914. [Google Scholar] [CrossRef]
  38. Tieleman, T. Training restricted Boltzmann machines using approximations to the likelihood gradient. In Proceedings of the International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1064–1071. [Google Scholar]
  39. Xie, S.; Sun, C.; Huang, J.; Tu, Z.; Murphy, K. Rethinking spatiotemporal feature learning: Speed-accuracy trade-offs in video classification. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 305–321. [Google Scholar]
  40. Guo, Y.; Siddiqui, F.; Zhao, Y.; Chellappa, R.; Lo, S.Y. StimuVAR: Spatiotemporal stimuli-aware video affective reasoning with multimodal large language models. arXiv 2024, arXiv:2409.00304. [Google Scholar]
  41. Yu, H.F.; Huang, F.L.; Lin, C.J. Dual coordinate descent methods for logistic regression and maximum entropy models. Mach. Learn. 2011, 85, 41–75. [Google Scholar] [CrossRef]
  42. Li, X.; Wang, S.; Huang, X. Temporal enhancement for video affective content analysis. In Proceedings of the ACM International Conference on Multimedia, Melbourne, VIC, Australia, 28 October–1 November 2024; pp. 642–650. [Google Scholar]
  43. Shor, J.; Jansen, A.; Maor, R.; Lang, O.; Tuval, O.; Quitry, F.d.C.; Tagliasacchi, M.; Shavitt, I.; Emanuel, D.; Haviv, Y. Towards learning a universal non-semantic representation of speech. arXiv 2020, arXiv:2002.12764. [Google Scholar]
  44. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Gool, L.V. Temporal segment networks: Towards good practices for deep action recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 20–36. [Google Scholar]
  45. Wang, S.; Li, X.; Zheng, F.; Pan, J.; Li, X.; Chang, Y.; Zhu, Z.; Li, Q.; Wang, J.; Xiao, Y. VAD: A video affective dataset with danmu. IEEE Trans. Affect. Comput. 2024, 15, 1889–1905. [Google Scholar] [CrossRef]
  46. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  47. Gouyon, F.; Pachet, F.; Delerue, O. On the use of zero-crossing rate for an application of classification of percussive sounds. In Proceedings of the COST G-6 Conference on Digital Audio Effects, Verona, Italy, 7–9 December 2000; p. 16. [Google Scholar]
  48. Plutchik, R. Emotion, Theory, Research, and Experience; Academic Press: Cambridge, MA, USA, 1980. [Google Scholar]
  49. Ekman, P. An Argument for Basic Emotions. Cogn. Emot. 1992, 6, 169–200. [Google Scholar] [CrossRef]
  50. Arifin, S.; Cheung, P.Y. Affective level video segmentation by utilizing the pleasure-arousal-dominance information. IEEE Trans. Multimed. 2008, 10, 1325–1341. [Google Scholar] [CrossRef]
  51. Zhao, S.; Yao, H.; Sun, X.; Xu, P.; Liu, X.; Ji, R. Video indexing and recommendation based on affective analysis of viewers. In Proceedings of the ACM International Conference on Multimedia, Scottsdale, AZ, USA, 28 November–1 December 2011; pp. 1473–1476. [Google Scholar]
  52. Cowen, A.S.; Keltner, D. Self-report captures 27 distinct categories of emotion bridged by continuous gradients. Proc. Natl. Acad. Sci. USA 2017, 114, E7900–E7909. [Google Scholar] [CrossRef] [PubMed]
  53. Koide-Majima, N.; Nakai, T.; Nishimoto, S. Distinct dimensions of emotion in the human brain and their representation on the cortical surface. NeuroImage 2020, 222, 117258. [Google Scholar] [CrossRef]
  54. Russell, J.A. A circumplex model of affect. J. Personal. Soc. Psychol. 1980, 39, 1161–1178. [Google Scholar] [CrossRef]
  55. Posner, J.; Russell, J.A.; Peterson, B.S. The circumplex model of affect: An integrative approach to affective neuroscience, cognitive development, and psychopathology. Dev. Psychopathol. 2005, 17, 715–734. [Google Scholar] [CrossRef] [PubMed]
  56. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. Deap: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef]
  57. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef]
  58. Dellandréa, E.; Huigsloot, M.; Chen, L.; Baveye, Y.; Xiao, Z.; Sjöberg, M. Datasets column: Predicting the emotional impact of movies. ACM SIGMultimedia Rec. 2019, 10, 6. [Google Scholar] [CrossRef]
  59. Li, C.; Wang, J.; Wang, H.; Zhao, M.; Li, W.; Deng, X. Visual-texual emotion analysis with deep coupled video and danmu neural networks. IEEE Trans. Multimed. 2020, 22, 1634–1646. [Google Scholar] [CrossRef]
  60. Sjöberg, M.; Baveye, Y.; Wang, H.; Quang, V.L.; Ionescu, B.; Dellandréa, E.; Schedl, M.; Demarty, C.H.; Chen, L. The MediaEval 2015 affective impact of movies task. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  61. Dellandréa, E.; Chen, L.; Baveye, Y.; Sjöberg, M.V.; Chamaret, C. The MediaEval 2016 emotional impact of movies task. In Proceedings of the MediaEval Workshop, Hilversum, Netherlands, 20-21 October 2016. [Google Scholar]
  62. Dellandréa, E.; Huigsloot, M.; Chen, L.; Baveye, Y.; Sjöberg, M. The MediaEval 2017 emotional impact of movies task. In Proceedings of the MediaEval Workshop, Dublin, Ireland, 13–15 September 2017. [Google Scholar]
  63. Dellandréa, E.; Huigsloot, M.; Chen, L.; Baveye, Y.; Xiao, Z.; Sjöberg, M. The MediaEval 2018 emotional impact of movies task. In Proceedings of the MediaEval Workshop, Sophia Antipolis, France, 29-31 October 2018. [Google Scholar]
  64. Baveye, Y.; Dellandrea, E.; Chamaret, C.; Chen, L. Deep learning vs. kernel methods: Performance for emotion prediction in videos. In Proceedings of the International Conference on Affective Computing and Intelligent Interaction, Xi’an, China, 21–24 September 2015; pp. 77–83. [Google Scholar]
  65. Borth, D.; Chen, T.; Ji, R.; Chang, S.F. Sentibank: Large-scale ontology and classifiers for detecting sentiment and emotions in visual content. In Proceedings of the ACM International Conference on Multimedia, Barcelona, Spain, 21–25 October 2013; pp. 459–460. [Google Scholar]
  66. Thao, H.T.P.; Balamurali, B.; Roig, G.; Herremans, D. Attendaffectnet–emotion prediction of movie viewers using multimodal fusion with self-attention. Sensors 2021, 21, 8356. [Google Scholar] [CrossRef]
  67. Oliva, A.; Torralba, A. Modeling the shape of the scene: A holistic representation of the spatial envelope. Int. J. Comput. Vis. 2001, 42, 145–175. [Google Scholar] [CrossRef]
  68. Irani, E.S.M. Matching local self-similarities across images and videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 18–23 June 2007; pp. 1–8. [Google Scholar]
  69. Torresani, L.; Szummer, M.; Fitzgibbon, A. Efficient object category recognition using classemes. In Proceedings of the European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; pp. 776–789. [Google Scholar]
  70. Li, L.J.; Su, H.; Fei-Fei, L.; Xing, E.P. Object bank: A high-level image representation for scene classification & semantic feature sparsification. In Proceedings of the Neural Information Processing Systems, Hyatt Regency, Vancouver, BC, Canada, 6–11 December 2010; pp. 1378–1386. [Google Scholar]
  71. Yi, Y.; Wang, H. Motion keypoint trajectory and covariance descriptor for human action recognition. Vis. Comput. 2018, 34, 391–403. [Google Scholar] [CrossRef]
  72. Baecchi, C.; Uricchio, T.; Bertini, M.; Del Bimbo, A. Deep sentiment features of context and faces for affective video analysis. In Proceedings of the ACM on International Conference on Multimedia Retrieval, Bucharest, Romania, 6–9 June 2017; pp. 72–77. [Google Scholar]
  73. Chen, T.; Borth, D.; Darrell, T.; Chang, S.F. Deepsentibank: Visual sentiment concept classification with deep convolutional neural networks. arXiv 2014, arXiv:1410.8586. [Google Scholar]
  74. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep face recognition. In Proceedings of the British Machine Vision Conference, Swansea, UK, 7–10 September 2015. [Google Scholar]
  75. Guo, X.; Zhong, W.; Ye, L.; Fang, L.; Heng, Y.; Zhang, Q. Global Affective Video Content Regression Based on Complementary Audio-Visual Features. In Proceedings of the International Conference on Multimedia Modeling, Daejeon, Republic of Korea, 5–8 January 2020; pp. 540–550. [Google Scholar]
  76. Eyben, F.; Weninger, F.; Gross, F.; Schuller, B. Recent developments in opensmile, the munich open-source multimedia feature extractor. In Proceedings of the ACM International Conference on Multimedia, Barcelona, Spain, 21–25 October 2013; pp. 835–838. [Google Scholar]
  77. Schuller, B.; Steidl, S.; Batliner, A.; Burkhardt, F.; Devillers, L.; Müller, C.A.; Narayanan, S.S. The INTERSPEECH 2010 paralinguistic challenge. In Proceedings of the INTERSPEECH, Makuhari, Chiba, Japan 26–30 September 2010. [Google Scholar]
  78. Schuller, B.; Steidl, S.; Batliner, A.; Vinciarelli, A.; Scherer, K.; Ringeval, F.; Chetouani, M.; Weninger, F.; Eyben, F.; Marchi, E.; et al. The INTERSPEECH 2013 computational paralinguistics challenge: Social signals, conflict, emotion, autism. In Proceedings of the Annual Conference of the International Speech Communication Association, Lyon, France, 25–29 August 2013. [Google Scholar]
  79. Aytar, Y.; Vondrick, C.; Torralba, A. Soundnet: Learning sound representations from unlabeled video. In Proceedings of the Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 892–900. [Google Scholar]
  80. Hershey, S.; Chaudhuri, S.; Ellis, D.P.; Gemmeke, J.F.; Jansen, A.; Moore, R.C.; Plakal, M.; Platt, D.; Saurous, R.A.; Seybold, B.; et al. CNN architectures for large-scale audio classification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, New Orleans, LA, USA, 5–9 March 2017; pp. 131–135. [Google Scholar]
  81. Crammer, K.; Dekel, O.; Keshet, J.; Shalev-Shwartz, S.; Singer, Y. Online passive aggressive algorithms. J. Mach. Learn. Res. 2006, 7, 551–585. [Google Scholar]
  82. Xu, M.; Jin, J.S.; Luo, S.; Duan, L. Hierarchical movie affective content analysis based on arousal and valence features. In Proceedings of the ACM International Conference on Multimedia, Vancouver, BC, Canada, 26–31 October 2008; pp. 677–680. [Google Scholar]
  83. Arandjelović, R.; Zisserman, A. Three things everyone should know to improve object retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2911–2918. [Google Scholar]
  84. Wang, S.; Hao, L.; Ji, Q. Knowledge-augmented multimodal deep regression bayesian networks for emotion video tagging. IEEE Trans. Multimed. 2020, 22, 1084–1097. [Google Scholar] [CrossRef]
  85. Huynh, V.; Lee, G.S.; Yang, H.J.; Kim, S.H. Temporal convolution networks with positional encoding for evoked expression estimation. arXiv 2021, arXiv:2106.08596. [Google Scholar]
  86. Lin, K.; Wang, X.; Zheng, Z.; Zhu, L.; Yang, Y. Less is more: Sparse sampling for dense reaction predictions. arXiv 2021, arXiv:2106.01764. [Google Scholar]
  87. Yan, B.; Wang, L.; Gao, K.; Gao, B.; Liu, X.; Ban, C.; Yang, J.; Li, X. Multi-Granularity Network with Modal Attention for Dense Affective Understanding. arXiv 2021, arXiv:2106.09964. [Google Scholar]
  88. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9-15 June 2019; pp. 6105–6114. [Google Scholar]
  89. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  90. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  91. Ho, N.H.; Yang, H.J.; Kim, S.H.; Lee, G.; Yoo, S.B. Deep graph fusion based multimodal evoked expressions from large-scale videos. IEEE Access 2021, 9, 127068–127080. [Google Scholar] [CrossRef]
  92. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27-30 June 2016; pp. 770–778. [Google Scholar]
  93. Mironica, I.; Ionescu, B.; Sjöberg, M.; Schedl, M.; Skowron, M. RFA at MediaEval 2015 affective impact of movies task: A multimodal approach. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  94. Thomas, T.; Domínguez, M.; Ptucha, R. Deep independent audio-visual affect analysis. In Proceedings of the IEEE Global Conference on Signal and Information Processing, Montreal, QC, Canada, 14–16 November 2017; pp. 1417–1421. [Google Scholar]
  95. Dai, Q.; Zhao, R.W.; Wu, Z.; Wang, X.; Gu, Z.; Wu, W.; Jiang, Y.G. Fudan-Huawei at MediaEval 2015: Detecting violent scenes and affective impact in movies with deep learning. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  96. Chakraborty, R.; Maurya, A.K.; Pandharipande, M.; Hassan, E.; Ghosh, H.; Kopparapu, S.K. TCS-ILAB-MediaEval 2015: Affective impact of movies and violent scene detection. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  97. Marin Vlastelica, P.; Hayrapetyan, S.; Tapaswi, M.; Stiefelhagen, R. KIT at MediaEval 2015–Evaluating visual cues for affective impact of movies task. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  98. Seddati, O.; Kulah, E.; Pironkov, G.; Dupont, S.; Mahmoudi, S.; Dutoit, T. UMons at MediaEval 2015 affective impact of movies task including violent scenes detection. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  99. Trigeorgis, G.; Coutinho, E.; Ringeval, F.; Marchi, E.; Zafeiriou, S.; Schuller, B. The ICL-TUM-PASSAU approach for the MediaEval 2015 affective impact of movies task. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  100. Lam, V.; Phan, S.; Le, D.D.; Satoh, S.; Duong, D.A. NII-UIT at MediaEval 2015 affective impact of movies task. In Proceedings of the MediaEval Workshop, Wurzen, Germany, 14–15 September 2015. [Google Scholar]
  101. Guo, J.; Song, B.; Zhang, P.; Ma, M.; Luo, W. Affective video content analysis based on multimodal data fusion in heterogeneous networks. Inf. Fusion 2019, 51, 224–232. [Google Scholar] [CrossRef]
  102. Zhang, J.; Zhao, Y.; Cai, L.; Tu, C.; Wei, W. Video affective effects prediction with multi-modal fusion and shot-long temporal context. arXiv 2019, arXiv:1909.01763. [Google Scholar]
  103. Wang, S.; Wang, C.; Chen, T.; Wang, Y.; Shu, Y.; Ji, Q. Video affective content analysis by exploring domain knowledge. IEEE Trans. Affect. Comput. 2019, 12, 1002–1017. [Google Scholar] [CrossRef]
  104. Huynh, V.T.; Yang, H.J.; Lee, G.S.; Kim, S.H. Prediction of evoked expression from videos with temporal position fusion. Pattern Recognit. Lett. 2023, 172, 245–251. [Google Scholar] [CrossRef]
  105. Peng, X.; Li, K.; Li, J.; Chen, G.; Guo, D. Multi-modality fusion for emotion recognition in videos. In Proceedings of the IJCAI Workshop on Micro-gesture Analysis for Hidden Emotion Understanding, Macau, China, 21–22 August 2023. [Google Scholar]
  106. Wei, J.; Yang, X.; Dong, Y. User-generated video emotion recognition based on key frames. Multimed. Tools Appl. 2021, 80, 14343–14361. [Google Scholar] [CrossRef]
  107. Qiu, H.; He, L.; Wang, F. Dual focus attention network for video emotion recognition. In Proceedings of the IEEE International Conference on Multimedia and Expo, London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar]
  108. Tao, H.; Duan, Q. Hierarchical attention network with progressive feature fusion for facial expression recognition. Neural Netw. 2024, 170, 337–348. [Google Scholar] [CrossRef]
Figure 1. A timeline of classic VER algorithms, namely, SentiBank [65], VideoEmotion [9], E-MDBM [21], CFN [18], MMDRBN [36], ITE [26], MMDDN [34], AFRN [11], MMDQEN [35], AttendAffectNet [66], MMLGAN [37], UMFN [12], CTEN [15], TE [42], StimuVAR [40], VAANet [13], FAEIL [22], LRCANet [4], and MART [14].
Figure 1. A timeline of classic VER algorithms, namely, SentiBank [65], VideoEmotion [9], E-MDBM [21], CFN [18], MMDRBN [36], ITE [26], MMDDN [34], AFRN [11], MMDQEN [35], AttendAffectNet [66], MMLGAN [37], UMFN [12], CTEN [15], TE [42], StimuVAR [40], VAANet [13], FAEIL [22], LRCANet [4], and MART [14].
Sensors 25 03615 g001
Figure 2. A diagram of MML [10].
Figure 2. A diagram of MML [10].
Sensors 25 03615 g002
Figure 3. A diagram of AFRN [11].
Figure 3. A diagram of AFRN [11].
Sensors 25 03615 g003
Figure 4. A diagram of UMFN [12].
Figure 4. A diagram of UMFN [12].
Sensors 25 03615 g004
Figure 5. A diagram of LRCANet [4].
Figure 5. A diagram of LRCANet [4].
Sensors 25 03615 g005
Table 1. Acronyms and descriptions.
Table 1. Acronyms and descriptions.
AcronymDescription
AFRNAdaptive fusion recurrent network [11]
CAERContext-aware emotion recognition [16]
CAFContext-aware framework [17]
CFNContext fusion network [18]
CNNConvolutional Neural Network
CTENCross-modal temporal erasing network [15]
DSIFTDense scale invariant feature transform [19]
EEVEvoked Expression in Video [20]
E-MDBMEnhanced multimodal deep Boltzmann machine [21]
FAEILFrame-level adaptation and emotion intensity learning [22]
FNNFeedforward neural network
GCNGraph convolutional network
GeMAPSGeneva minimalistic acoustic parameter set [23]
GRUsGated recurrent units
HAFsHigh-level audio features
HMMHidden Markov model
HOGHistogram of oriented gradient [24]
HSHHue–saturation histogram [25]
HVFsHigh-level visual features
ITEImage Transfer Encoding [26]
LAFsLow-level audio features
LARLeast-angle regression [27]
LBPLocal binary pattern [28]
LIRIS-ACCEDELIRIS annotated creative commons emotional database [29]
LRCANetLong-range correlation-aware network [4]
LSTMLong short-term memory [30]
LVFsLow-level visual features
MARTMasked affective representation learning [14]
MFCCMel-frequency cepstrum coefficients [31]
MKTMotion keypoint trajectory
MLG-SMultimodal local–global system [32]
MLLMsMultimodal large language models
MLPMulti-layer perceptron [33]
MM-AUMultimodal ads understanding [5]
MMDDNMultimodal deep denoise network [34]
MMDQENMultimodal deep quality embedding network [35]
MMDRBNMultimodal deep regression Bayesian network [36]
MMLMultimodal learning [10]
MMLGANMultimodal local–global attention network [37]
MSEMean-square error
PCCPearson correlation coefficient
RBMRestricted Boltzmann machine [38]
RPNRegion proposal network
S3DSeparable 3D CNN [39]
StimuVARStimuli-aware video affective reasoning [40]
SVMSupport vector machine [41]
TETemporal enhancement [42]
TRILLTriplet loss network [43]
TSNTemporal segment network [44]
UMFNUnified multi-stage fusion network [12]
VAANetVisual audio attention network [13]
VADVideo affective dataset [45]
VERVideo emotion recognition
VGGVisual geometry group [46]
ZCRZero-crossing rate [47]
Table 2. Datasets for VER.
Table 2. Datasets for VER.
DatasetYearEmotion ModelMain Evaluation Metrics
DEAP [56]2012Valence–arousal–dominanceAccuracy
MAHNOB-HCI [57]2012Valence–arousal–dominanceF1
VideoEmotion-8 [9]2014PlutchikAccuracy
LIRIS-ACCEDE [29,58]2015Valence–arousalMSE, PCC, and accuracy
Ekman-6 [26]2018EkmanAccuracy
CAER [16]20196 emotionsAccuracy
Video-Danmu [59]20207 emotionsPrecision and accuracy
EEV [20]20215 emotionsPCC
MM-AU [5]20233 emotionsF1 and accuracy
VAD [45]2024Valence–arousal and 13 emotionsF1 and accuracy
Table 3. Results of classic methods on the MediaEval 2015 task of LIRIS-ACCEDE.
Table 3. Results of classic methods on the MediaEval 2015 task of LIRIS-ACCEDE.
MethodVisualAudioArousal (%)Valence (%)
Mironica et al. [93]45.0436.12
Thomas et al. [94]48.2044.64
Dai et al. [95]48.8441.78
Chakraborty et al. [96]48.9535.66
Marin et al. [97] 51.8938.54
Seddati et al. [98]52.4437.28
Trigeorgis et al. [99]55.7241.48
Lam et al. [100]55.9142.96
Baecchi et al. [72] 55.9845.31
MMDDN + MMCLF [34]56.7545.03
MMDQEN [35]56.7545.03
OFVGG + GeMAPS [101]57.0040.83
MML [10]57.4046.22
Zhang et al. [102]57.5045.90
MLG-S [32]57.9048.20
AFRN [11]58.2248.61
Wang et al. [103]60.8843.74
Table 4. Results of classic methods on EEV.
Table 4. Results of classic methods on EEV.
MethodVisualAudioPCC
Ho et al. [91]0.00819
MGN-MA [87]0.02292
Lin et al. [86]0.04430
TPF [104]0.05400
TCN [85]0.05477
Table 5. Results of classic methods on VideoEmotion-8.
Table 5. Results of classic methods on VideoEmotion-8.
MethodVisualAudioACC(%)
StimuVAR [40] 40.50
Peng et al. [105]43.20
MART [14]50.83
MMLGAN [37]51.14
ITE [26]52.60
CAF [17]52.70
KeyFrame [106] 52.85
DFAN [107] 53.34
VAANet [13]54.50
UMFN [12]55.80
CTEN [15]57.30
LRCANet [4]57.40
FAEIL [22]57.63
TE [42]59.39
Table 6. Results of classic methods on Ekman-6.
Table 6. Results of classic methods on Ekman-6.
MethodVisualAudioACC(%)
ITE [26]51.20
CFN [18] 51.80
MART [14]52.66
CAF [17]53.60
VAANet [13]55.30
DFAN [107] 57.37
CTEN [15]58.20
KeyFrame [106] 59.51
LRCANet [4]59.78
FAEIL [22]60.44
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yi, Y.; Zhou, Y.; Wang, T.; Zhou, J. Advances in Video Emotion Recognition: Challenges and Trends. Sensors 2025, 25, 3615. https://doi.org/10.3390/s25123615

AMA Style

Yi Y, Zhou Y, Wang T, Zhou J. Advances in Video Emotion Recognition: Challenges and Trends. Sensors. 2025; 25(12):3615. https://doi.org/10.3390/s25123615

Chicago/Turabian Style

Yi, Yun, Yunkang Zhou, Tinghua Wang, and Jin Zhou. 2025. "Advances in Video Emotion Recognition: Challenges and Trends" Sensors 25, no. 12: 3615. https://doi.org/10.3390/s25123615

APA Style

Yi, Y., Zhou, Y., Wang, T., & Zhou, J. (2025). Advances in Video Emotion Recognition: Challenges and Trends. Sensors, 25(12), 3615. https://doi.org/10.3390/s25123615

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop