Next Article in Journal
Machine Learning in Thermography Non-Destructive Testing: A Systematic Review
Previous Article in Journal
Evaluation of a Lightweight IoT Protocol for Intelligent Parking Management in Urban Environments
Previous Article in Special Issue
SceEmoNet: A Sentiment Analysis Model with Scene Construction Capability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal Emotion Recognition Based on Graph Neural Networks

1
Educational Service Center, Communication University of China, Beijing 100024, China
2
School of Information and Engineering, Communication University of China, Beijing 100024, China
3
School of Data and Intelligence, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(17), 9622; https://doi.org/10.3390/app15179622
Submission received: 30 July 2025 / Revised: 24 August 2025 / Accepted: 28 August 2025 / Published: 1 September 2025
(This article belongs to the Special Issue Advanced Technologies and Applications of Emotion Recognition)

Abstract

Emotion recognition remains a challenging task in human–computer interaction. With advancements in multimodal computing, multimodal emotion recognition has become increasingly important and significant. To address the existing limitations in multimodal fusion efficiency, emotional–semantic association mining, and long-range context modeling, we propose an innovative graph neural network (GNN)-based framework. Our methodology integrates three key components: (1) a hierarchical sequential fusion (HSF) multimodal integration approach, (2) a sentiment–emotion enhanced joint learning framework, and (3) a context-similarity dual-layer graph architecture (CS-BiGraph). The experimental results demonstrate that our method achieves 69.1% accuracy on the IEMOCAP dataset, establishing a new state-of-the-art performance. For future work, we will explore robust extensions of our framework under real-world scenarios with higher noise levels and investigate the integration of emerging modalities for broader applicability.

1. Introduction

Multimodal dialogue emotion recognition, which integrates textual, acoustic, and visual cues, is a pivotal research area with significant applications in human–computer interaction, clinical diagnosis, and cognitive science [1]. By leveraging multiple modalities, machines can achieve a more comprehensive and accurate understanding of human emotional states. However, effectively modeling the complex interplay between modalities, speakers, and contextual dependencies within a conversation remains a formidable challenge.
The existing methods face three primary limitations. Firstly, many approaches struggle to capture the rich dynamic interactions between modalities, often resorting to simple fusion techniques that neglect semantic alignment and complementary information [2,3]. Secondly, while sentiment analysis is closely related to emotion recognition, most models fail to effectively leverage fine-grained sentiment features to enhance emotion classification, or lack mechanisms for dynamic emotional weight optimization [4]. Lastly, capturing long-range contextual dependencies is crucial, yet traditional RNN-based models suffer from vanishing gradient issues [5,6], while recent GNN-based approaches are often constrained by fixed dialogue order and predefined graph structures [7,8]. These challenges collectively hinder the performance and robustness of current emotion recognition systems.
To overcome these limitations, we propose a novel integrated framework centered around the context-similarity bilevel graph (CS-BiGraph) model. Our framework introduces a sophisticated multimodal fusion method (HSF) and a sentiment-enhanced learning framework (SE) to provide high-quality sentiment-aware features. The core of our work, the CS-BiGraph model, innovatively constructs a bilevel graph to capture both global context similarity and local conversational relationships, leveraging a graph autoencoder for adaptive structure learning. This allows our model to overcome the constraints of fixed graph structures and effectively model long-range dependencies.
The main contributions of this paper are summarized as follows:
  • We propose a novel sentiment-enhanced learning framework (SE) that integrates a RoBERTa-based fine-grained feature extractor with a dynamic emotional weight optimization module, effectively sharpening emotion classification boundaries.
  • We introduce the context-similarity bilevel graph (CS-BiGraph) model, which decouples global and local contextual modeling. Its structure self-learning mechanism, driven by a graph autoencoder, adaptively captures long-range dependencies, overcoming the limitations of conventional GNNs.
  • We conduct extensive experiments on the IEMOCAP and MOSEI datasets. Our integrated framework achieves state-of-the-art performance, demonstrating the effectiveness of our proposed methods, particularly in recognizing neutral and happy emotions.

2. Related Work

2.1. Multimodal Fusion Strategies

Effective fusion of information from different modalities is a cornerstone of multimodal emotion recognition. Early works often employed feature-level or decision-level fusion. For instance, Poria et al. [3] developed dedicated models for each modality before fusing their outputs. Zhou et al. [2] proposed a semi-supervised multipath generative network for fusion. A significant breakthrough came with attention-based mechanisms and transformers, which allow for more dynamic inter-modal interactions. Tsai et al. [9] introduced the cross-modal transformer, which effectively aligns features across modalities. Shou et al. [10] proposed a Low-Rank Matching Attention Method (LMAM), which utilizes low-rank decomposition to significantly reduce the number of parameters in the cross-modal attention mechanism. However, many of these methods still process modalities in a parallel fashion, which may not fully capture the temporal and hierarchical dependencies inherent in human communication. Our HSF fusion method builds upon this by introducing a sequential weighting mechanism to achieve progressive temporally synchronized fusion.

2.2. Sentiment-Enhanced Emotion Recognition

The intrinsic link between sentiment (positive, negative, and neutral) and discrete emotions (happy, sad, and angry) has inspired research into joint learning frameworks. Wu et al. [11] treated emotion classification as the primary task and sentiment classification as an auxiliary task, using shared information to boost performance. More recently, UniMSE [4] proposed a comprehensive knowledge sharing framework across features, labels, and models. While these methods validate the benefit of leveraging sentiment information, they often lack the capacity for fine-grained sentiment analysis or mechanisms to dynamically optimize how much sentiment information should influence the final emotion prediction. Our SE framework addresses this gap by combining a powerful pre-trained language model (RoBERTa) for fine-grained feature extraction with a dynamic weight optimization module.

2.3. Contextual Modeling in Dialogues

Modeling the conversational context is critical for understanding emotional dynamics. Early approaches relied on Recurrent Neural Networks (RNNs) and their variants. ICON [5] and CMN [12] used RNNs to capture inter-utterance temporal relationships. Subsequent models like DialogueRNN [6], SGED [13], and RTER [14] introduced more complex architectures to handle speaker-specific information and long-range dependencies, but the inherent limitations of sequential modeling remained.
To better capture non-sequential relationships, graph neural networks (GNNs) have become a popular choice. Models like DialogueGCN [7], RGAT [15], and ConGCN [8] structure dialogues as graphs, with utterances as nodes and relationships as edges. This allows for more flexible modeling of contextual influences. Nevertheless, these GNN-based methods are typically constrained by fixed graph construction rules (e.g., relying only on speaker order) and limited receptive fields (e.g., fixed window sizes), which hinders their ability to capture complex long-range dependencies adaptively. Our CS-BiGraph model is specifically designed to overcome these constraints by learning the graph structure itself.

2.4. LLM-Based and Prompt-Driven Methods

Another promising direction is the application of prompt learning to emotion recognition. For instance, Wu et al. [16] proposed the MERC-PLTAF framework, which ingeniously constructs both textual and ‘acoustic prompts’ to guide a pre-trained language model. Their work demonstrates the power of leveraging the structured knowledge within LLMs through explicit prompting.

3. Method

The focus of this study is to perform emotion recognition by integrating multimodal data. First, the hierarchical sequential fusion (HSF) method is employed to fuse different modalities. Simultaneously, textual information is input into the sentiment-enhancement (SE) framework to predict sentiment labels for utterances, which assists emotion recognition. The emotion labels and fused multimodal data are then processed through a contextual extractor and input into the context-similarity bilevel graph (CS-BiGraph) model to generate graph structures. Subsequently, the outputs from CS-BiGraph are processed via relational graph convolutional network (RGCN) and graph transformer, with the final emotion recognition results obtained through a classifier. The overall process is shown in Figure 1.

3.1. Feature Extraction

(1) Textual Modality Features: The sBert (sentence-BERT) [17] model is employed to extract key textual features. After processing through sBert, each sentence yields a 768-dimensional feature vector u t , which we denote as u t R L t × d t , where L t is the sequence length and d t = 768.
(2) Audio Modality Features: The open-source toolkit OpenSmile [18] is utilized to extract acoustic features. Acoustic descriptors such as Mel-Frequency Cepstral Coefficients (MFCCs), Fundamental Frequency Contour (F0), and over 64 other acoustic features are extracted to establish mappings between speech signals and emotional states. A total of 6373 acoustic descriptors are used to construct speech emotion features. The feature matrix undergoes standardization preprocessing and is subsequently processed through a fully connected neural network layer for dimensionality reduction, resulting in a 100-dimensional audio feature vector denoted as u a R L a × d a , where d a = 100.
(3) Video Modality Features: The OpenFace [19] toolkit is used to extract facial keypoints and derive rich feature representations. After processing with OpenFace, a 512-dimensional sequential feature vector u v is obtained as the final video modality representation, denoted as u v R L v × d v , where d v = 512.

3.2. Hierarchical Sequential Fusion

HSF is the multimodal fusion module designed in this study. As illustrated in Figure 2, its architecture comprises three parallel cross-fusion blocks designed to enrich each modality’s features by incorporating information from the others.

3.2.1. Cross-Attention Mechanism

At the core of our fusion block is the cross-attention mechanism, which enables a target modality to selectively integrate information from a source modality. To facilitate this, the input feature sequences are linearly projected into three distinct representations: the query ( Q ), the key ( K ), and the value ( V ). The Q representation is derived from the target modality’s features, while the K and V representations are derived from the source modality’s features. This projection is governed by three learnable weight matrices: W Q , W K , and W V . The attention scores are then computed based on the similarity between queries and keys, which in turn weight the values to produce the final fused output.

3.2.2. Fusion Process

The fusion process within HSF unfolds in two stages.
First, the audio and video modalities are enhanced in parallel. The audio feature sequence u a is enriched by attending to both the text features u t and the video features u v . This results in two intermediate representations, A a t and A a v , computed as shown in Equation (1) and Equation (2), respectively.
A a t = softmax ( u a W Q ) ( u t W K ) T d k ( u t W V )
A a v = softmax ( u a W Q ) ( u v W K ) T d k ( W V u v )
where d k is the dimension of the key vectors for normalization. These two enhanced representations are then fused using the Hadamard product (⊙), which performs element-wise multiplication. We chose this operation because it acts as a non-linear feature-gating mechanism, amplifying features that receive strong signals from both the text and video sources while suppressing those supported by only one. The result is then passed through a normalization layer to produce the final enhanced audio features, u a . The video features are similarly enhanced in parallel to produce u v .
Second, in our sequential strategy, the text modality serves as the final integration point. It attends to the already enhanced audio ( u a ) and video ( u v ) features to produce its final representation, u t . For instance, text attending to the enhanced audio features is formulated as
A t a = softmax ( u t W Q ) ( u a W K ) T d k ( u a W V )
All normalization steps employ a residual connection with a learnable coefficient α , as shown in Equation (4). Finally, the three fully enhanced feature sequences are concatenated to form the final multimodal representation X atv (Equation (5)).
u out = LayerNorm ( u in + α · F )
X atv = concat ( u a , u t , u v )

3.3. Sentiment-Enhancement Framework

To predict sentiment labels for utterances and enhance emotion recognition, we design the SE module with the architecture shown in Figure 3.
(1) Embedding Layer: For sentence-level sentiment analysis, inputs are unified as word sequences S = [ w 1 , w 2 , …, w n ]. The pre-trained RoBERTa model converts inputs into continuous vector representations. Each word w i ∈ S is mapped to a continuous vector x i R d (where d denotes the embedding dimension). By stacking these word vectors, we obtain the word embedding matrix X ∈ R n × d .
(2) Global Locator: The architecture of global locator is shown on the left in Figure 4. The word embedding matrix X is fed into the global locator to capture sentence-level contextual information. LSTM Processing: X is input to an LSTM to generate hidden states Ht. Attention Mechanism: Ht is linearly transformed into query (Q), key (K), and value (V) representations, as computed in Equation (6).
A L = softmax ( Q K T d ) V
Residual Processing: The attention output is combined with the original input through residual connections to produce the final global locator output.
(3) Local Locator: The architecture of global locator is shown on the left in Figure 4. This component employs an Encoder–Locator Combination (ELC) mechanism. Encoder Guidance: For the k-th ELC (k > 1), the encoder derives hidden states h k R d by weighted summation of rows in C k 1 (masked context matrix). The weights are determined by l k 1 R n (projected weights from the previous ELC).
Masking Strategy: Before core computations in the k-th ELC, C k 1 undergoes Bernoulli masking with predefined probability Pmask. Each row in C k 1 is independently masked (set to zero) based on weights in l k 1 . This operation is only activated during training (demonstrated in Figure 4).
(4) Linear Layer: A single-layer feedforward neural network constructs the linear layer Equation (7).
D = ( 1 k k h k ) W dis + b dis
The output D represents sentiment features extracted by the SE module, which are concatenated with multimodal fusion features from HSF for downstream processing.

3.4. Contextual Extractor

The fused features X a t v generated by HSF and sentiment features D obtained from the SE module are concatenated and input into the contextual extractor. This module takes the fused features as inputs for each dialogue utterance ( u i ; i = 1, …, n) and employs a transformer encoder to capture contextual dependencies. The architecture is illustrated in Figure 1.

3.5. CS-BiGraph

Our CS-BiGraph model addresses a fundamental limitation in existing GNN-based methods for conversational emotion recognition, such as DialogueGCN and COGMEN [20]. While these models have successfully introduced graphs to model conversational context, they typically construct a single monolithic graph based on rigid predefined rules (e.g., speaker turn order and fixed-size temporal windows). This approach suffers from two key drawbacks: (1) it conflates local turn-by-turn conversational flow with global long-range semantic dependencies, treating them within the same structural representation; and (2) its fixed structure is unable to adapt to the highly dynamic and non-sequential nature of human dialogues, where an utterance might be semantically closer to a much earlier turn than its immediate predecessor.
To overcome these challenges, our CS-BiGraph introduces a novel bilevel architectural paradigm that explicitly decouples these two distinct types of contextual relationships. While the CS-BiGraph paradigm is constructed using established techniques such as K-nearest neighbor (KNN) and graph autoencoder (GAE), its fundamental novelty lies in their architectural orchestration. This specific integration is engineered to create a disentangled bilevel structure that directly addresses a core limitation in prior models: the conflation of local turn-by-turn conversational flow with global long-range semantic dependencies. Therefore, the primary contribution is the design of this specialized architecture, which systematically separates these two contextual facets and models them with dedicated graph structures to achieve a more robust understanding of the dialogue’s emotional landscape.
A Global Similarity Graph: This layer is specifically designed to capture the non-sequential topic-driven semantic relationships across the entire conversation. By connecting nodes based on feature similarity (via KNN and cosine distance), it builds a “shortcut” network that allows information to flow between semantically related but temporally distant utterances, overcoming the limitations of fixed window sizes.
A Local Speaker Relationship Graph: This layer focuses on modeling the turn-by-turn speaker-dependent conversational dynamics. Crucially, instead of using fixed rules, we employ a graph autoencoder (GAE) to adaptively learn the optimal structure of these local interactions directly from the data. This allows the model to dynamically adjust the graph’s connectivity based on the specific dialogue rather than relying on a one-size-fits-all heuristic.
This explicit decoupling of global semantic context from local conversational structure is the fundamental architectural difference between CS-BiGraph and prior GNN-based models. By modeling these two facets of context with specialized mechanisms and then integrating them, our framework achieves a more comprehensive and robust understanding of the dialogue’s emotional landscape. The following describes the implementation details of each graph component.
The CS-BiGraph model consists of a global similarity-based neighbor graph and a speaker relationship graph based on graph autoencoders, as illustrated in Figure 5.
The dialogue is defined as D = U s 1 , U s 2 , …, U s M , where U s 1 = u 1 s 1 , u 2 s 1 , …, u n s 1 represents the utterances from speaker 1, as illustrated in Figure 6.
In this study: Intra-speaker relationships R intra = U i S U i S are relationships between utterances from the same speaker. Inter-speaker relationships R inter U S i U S j i j are relationships between utterances from different speakers. (1) Global Similarity-Based Neighbor Graph: After feature extraction, the K-nearest neighbor (KNN) algorithm calculates node similarities using Equation (8):
dis = i , j ( u i u j ) 2
where d denotes the feature dimension. Smaller distances indicate higher similarity. Subsequently, cosine distances between nodes are computed, and edges are removed if distances exceed a threshold τ (hyperparameter), reducing redundant information and enhancing discriminative features.
(2) Speaker Relationship Graph via Graph Autoencoder (GAE): The graph autoencoder (GAE), an unsupervised graph neural network (GNN) method, optimizes graph structures by reconstructing adjacency matrices through node representation learning. It comprises the following:
Encoder: Maps node features to a low-dimensional latent space via graph convolutional networks (GCNs), learning node embeddings Z. Decoder: Reconstructs the adjacency matrix A ^ using Z, as defined in Equation (9):
A ^ = σ ( Z Z T )
where σ denotes the sigmoid function, mapping similarities to the range [0, 1].

3.6. Prediction Result Generation

After graph construction, the relational graph convolutional network (R-GCN) is employed to train the graph structure, effectively capturing dependencies between inter-speaker and intra-speaker connected utterances, as defined in Equation (10).
x i = θ root · z i + r R j N r ( i ) 1 | N r ( i ) | θ r · z j
where N r ( i ) indicates neighbor indices of node i under relation r R ; θ r o o t , θ r represent learnable parameters of R-GCN; | N r ( i ) | denotes normalization constant; and z indicates utterance-level features from the contextual extractor. The processed graph representations are then fed into a graph transformer layer with ReLU activation and softmax normalization, followed by a classifier to generate final emotion label predictions. The complete pipeline is illustrated in Figure 7.

4. Experiment

4.1. Experimental Environment

Our experimental environment is shown in Table 1.

4.2. Test Dataset Description

IEMOCAP Dataset [21]: Comprises 151 dyadic dialogues with 7433 utterances, each annotated by experts with six emotion labels: happiness, sadness, calmness, anger, excitement, and depression.
CMU-MOSEI Dataset [22]: Contains over 23,000 monologue video clips from YouTube, annotated for sentiment and emotion. Sentiment labels range from −3 (strongly negative) to +3 (strongly positive), while emotion labels include happiness, sadness, disgust, fear, surprise, and anger. Due to multi-label emotion annotations, sentiment labels are prioritized for evaluation. The CMU-MOSEI dataset features multi-label emotion annotations, where a single utterance can be associated with multiple emotion categories. To ensure a fair and consistent comparison with prior works, we follow a standard evaluation protocol for this dataset. Specifically, we transform the multi-label problem into a multi-class classification problem by prioritizing the sentiment labels, which are single-label annotations ranging from −3 (strongly negative) to +3 (strongly positive), resulting in a 7-class classification task. For binary sentiment classification (positive/negative), we treat sentiment scores > 0 as the positive class and scores < 0 as the negative class, excluding neutral samples. The reported accuracy (Acc) metric in our experiments corresponds to the standard classification accuracy on this primary sentiment task.

4.3. Evaluation Metrics

To assess model performance across scenarios, predictions are evaluated on the IEMOCAP test set (1623 utterances) and CMU-MOSEI test set (4662 utterances). Performance is measured using standard metrics: precision, recall, and F1-score, which comprehensively evaluate the model’s ability to accurately recognize diverse emotion categories.

4.4. Computational Complexity Analysis

To provide a formal assessment of our framework’s theoretical underpinnings, we analyze its computational complexity. The overall complexity is primarily determined by its three main components: the hierarchical sequential fusion (HSF), sentiment-enhancement (SE), and CS-BiGraph modules.
Let N be the number of utterances in a dialogue and L be the sequence length of each utterance. The complexity of the feature extractors (e.g., sBERT) is typically O ( L 2 ) per utterance. For our main components, the complexities are as follows:
  • HSF Module: The complexity is dominated by the cross-attention mechanisms. For three modalities, this results in a complexity of O ( N · L 2 ) .
  • SE Module: The RoBERTa-based embedding, along with the subsequent LSTM and attention layers, contribute a complexity of O ( N · L 2 ) .
  • CS-BiGraph Module: This module’s complexity is driven by two parts: the Global Similarity Graph construction using KNN, which is O ( N 2 · d ) , and the GNN operations (GAE and R-GCN), which are approximately O ( E · d ) , where d is the feature dimension and E is the number of edges.
Therefore, the total computational complexity of our framework is approximately O ( N · L 2 + N 2 · d + E · d ) . The quadratic term O ( N 2 · d ) from the KNN graph construction is the dominant factor for long dialogues. This analysis highlights a trade-off between our model’s ability to capture long-range dependencies and its scalability.

5. Results

5.1. Ablation Study

To validate the effectiveness of each proposed component, ablation experiments are conducted by removing individual modules and substituting alternatives: HSF module removal replaced with simple feature concatenation, SE module removal replaced with RNN, and CS-BiGraph removal replaced with conventional GCN. The key findings are presented below (Table 2 and Table 3):
Module Complementarity:
HSF + CS-BiGraph achieves the highest F1-score (56.1%) for happy emotion, demonstrating the synergy between hierarchical sequential fusion and bilevel graph optimization in positive emotion recognition.
Sentiment-Topology Synergy: SE + CS-BiGraph improves the excited emotion F1-score to 76.6%, verifying the benefits of combining sentiment enhancement with global graph modeling.
To further analyze the effectiveness of each component in the bilevel graph structure, we conducted a decomposition analysis of the CS-BiGraph module, with the results shown in Table 4. The removal of the Global Similarity Graph (w/o Similarity Graph) and graph autoencoder (w/o GAE) led to significant performance drops of 3.2% and 1.0%, respectively, on the IEMOCAP dataset, demonstrating the foundational role of graph structure optimization in complex emotion recognition. Notably, when CS-BiGraph remains intact, the synergistic gains between HSF and SE reach their maximum (+1.7%). This enhancement effect is particularly pronounced in long-tail categories such as neutral (69.9%) and angry (68.3%). Compared to the baseline model COGMEN (w/o Similarity Graph, GAE), the triple-module integrated model (HSF + SE + CS-BiGraph) achieves a 1.9× greater improvement (+3.4%) in complex emotion categories than in ordinary ones.
The optimized graph structure enables efficient fusion of HSF’s local temporal features and SE’s global semantic information through its hierarchical information propagation channels. This triple synergistic effect ultimately yields an average F1-score of 70.0%, significantly outperforming single-module (max 68.7%) and dual-module combinations (max 69.3%).

Impact of Pre-Trained Feature Extractors

To address the framework’s dependency on large pre-trained models, we conducted a quantitative analysis by replacing the sBERT textual feature extractor with a lighter non-contextual GloVe model [23]. All the other framework components remained unchanged. The comparative results on the IEMOCAP dataset are presented in Table 5.
The results show a significant performance drop of 5.2%, empirically confirming the importance of high-quality contextual embeddings from models like sBERT. Notably, even with a simpler GloVe extractor, our framework achieves a respectable F1-score of 64.8%. This demonstrates that the architectural design of our framework contributes significantly to its effectiveness, independent of the feature extractor’s power.

5.2. Hyperparameter Analysis

Hyperparameter Settings For the IEMOCAP dataset: Dropout = 0.1, GNNhead = 7, SeqContext = 4, Learning Rate = 1 × 10−4. For the CMU-MOSEI dataset: Dropout = 0.3, GNNhead = 2, SeqContext = 1, Learning Rate = 1 × 10−3.

5.2.1. Number of Attention Layers

To investigate the impact of attention layers in the SE module’s global locator, we fix other hyperparameters and vary the number of attention layers h ∈ {1, 2, 3, 4, 5}. The results (Figure 8) are presented as follows:
Initial performance gain: Increasing h from 1 to 2 enhances key sentiment feature capture, improving accuracy. Diminishing returns: Beyond h = 5, model complexity surges, causing overfitting and reduced generalization.

5.2.2. Number of Neighbor Nodes

To analyze the impact of neighbor quantity in k-NN graph construction, we fix other hyperparameters and vary n ∈ {1, 2, 3, 4, 5, 6, 7, 8, 9, 10}. The results (Figure 9) reveal the following:
Sparse regime ( n 3 ): Insufficient neighbors restrict feature propagation due to sparse graph connectivity. Optimal range ( 4 n 6 ): Balanced local associations and mitigated sparsity yield peak performance. Overdense regime ( n 7 ): Excessive edges introduce low-similarity noise, degrading model robustness.

5.2.3. Similarity Threshold τ

Similarity Threshold τ : To analyze the impact of similarity threshold τ on model performance during graph construction, we fix other hyperparameters and vary τ {0, 0.1, 0.2, …, 1.0}. The results (Figure 10) demonstrate the following:
Low τ regime ( τ 0.3 ): Excessive low-similarity edges introduce significant noise, diluting discriminative emotional associations despite expanded feature propagation.
Optimal range ( 0.4 τ 0.6 ): Balanced edge density and noise suppression achieve peak performance, retaining high-confidence emotional connections while avoiding isolated nodes.
High τ regime ( τ 0.7 ): Over-sparse graphs impair context aggregation due to insufficient neighbor connections, causing severe performance degradation.

5.3. Statistical Significance Analysis

To validate that the performance improvements of our proposed framework are statistically significant and not due to random chance, we conducted a statistical analysis on the IEMOCAP dataset, which serves as our primary benchmark. We specifically compared our full model against a strong baseline, COGMEN.
Following common practice, we performed multiple independent training runs (five runs) for both our model and the baseline, each with different random seeds. We then collected the F1-score for each run and performed a two-sample t-test to assess the significance of the difference between the two sets of results.
The mean F1-score for our model was 70.2 with a standard deviation of 0.53. In comparison, the mean F1-score for the baseline was 67.6 with a standard deviation of 0.72. The t-test yielded a p-value 0.00084. Since p < 0.05 , we can conclude that the performance gain achieved by our model is statistically significant. This analysis provides strong evidence that the architectural innovations in our framework, particularly the CS-BiGraph module, lead to a genuine and reliable improvement in performance.

5.4. Comparison with Existing Works

In this subsection, we further compare the proposed method with state-of-the-art multimodal emotion classification models on the IEMOCAP and CMU-MOSEI datasets.
To provide a comprehensive and up-to-date comparison, we have included several strong baselines as well as two very recent state-of-the-art methods from 2025, LMAM and MERC-PLTAF [16]. The overall performance comparison on the IEMOCAP dataset, which features rich dyadic dialogues and serves as our primary benchmark, is visualized in Figure 11.
As illustrated in Figure 11, our integrated framework outperforms established baselines such as GraphMFT, BiGraph [24], TelME [25], and COGMEN. We also observe that the recently proposed LMAM and MERC-PLTAF achieve higher overall scores. This warrants a deeper analysis into the architectural philosophies of these methods.
The strong performance of the LMAM can be attributed to its highly specialized and parameter-efficient cross-modal fusion module, which excels at its specific subtask. MERC-PLTAF’s success lies in its innovative use of a prompt-based learning paradigm, effectively converting the multimodal problem into a format that maximally leverages the power of large pre-trained language models.
In contrast, our work’s primary contribution is a holistic three-component architecture that addresses multiple distinct challenges in the CER pipeline simultaneously. Our key innovation, the CS-BiGraph module, provides a unique capability in adaptively learning complex non-sequential conversational structures—a mechanism not present in the aforementioned models. As demonstrated in our extensive ablation studies (Table 1), this graph-based contextual modeling is the cornerstone of our framework’s performance. Therefore, we posit that our work offers a complementary and architecturally novel approach, focusing on robust conversational structure modeling, while methods like LMAM and MERC-PLTAF provide powerful solutions for modality fusion and alignment, respectively. Future work could explore integrating these specialized modules into our comprehensive framework.
To provide a more fine-grained analysis of our model’s capabilities, Table 6 details the F1-scores for each of the six emotion categories on the IEMOCAP dataset, comparing our model with several classic and strong baselines for which this detailed data is available, such as BC-LSTM [26], CMN, ICON, DialogueRNN, DialogueGCN, MMGCN [27], TBJE [28], and Multilogue-Net [29]. For the CMU-MOSEI dataset, which consists primarily of monologues, the overall accuracy scores are presented in Table 7.
As noted in our dataset description, CMU-MOSEI primarily consists of monologues. It serves as a crucial benchmark for multi-label classification on single-speaker data streams. Our framework achieves a competitive accuracy score, as shown in Table 7, where we also include results from recent state-of-the-art models for a comprehensive comparison.
As shown in Table 7, our model’s accuracy (48.6%) is competitive, although slightly below recent SOTA models like TAILOR (48.8%) [30] and CARAT (49.4%) [31]. We attribute this difference to a fundamental divergence in research paradigms. These state-of-the-art methods achieve high performance by developing sophisticated intra-task mechanisms, such as TAILOR’s label-specific feature generation or CARAT’s reconstruction-based fusion.
In contrast, our work explores a complementary inter-task approach. The novelty of our SE module lies in leveraging an auxiliary task—sentiment polarity analysis—to provide external knowledge that constrains and guides the primary emotion recognition task. This strategy aims to reduce emotional ambiguity by using broader sentiment cues, a method that proves particularly effective on conversational datasets like IEMOCAP (see Table 6).
Thus, while SOTA models advance intra-task representation learning, our framework contributes an orthogonal perspective on inter-task knowledge integration. We believe these approaches are not mutually exclusive; future work could explore combining our sentiment-enhancement strategy with the advanced fusion mechanisms of models like TAILOR and CARAT.

5.5. Discussion

Our experimental results, presented in Table 6 and Table 7, demonstrate that our proposed framework achieves competitive or state-of-the-art performance compared to a wide range of baseline models. A key strength of our approach is evident on the IEMOCAP dataset, where our model shows a significant 7.1% F1-score improvement on the “happy” emotion over the second-best model from our baseline set—a category where many models traditionally struggle. This highlights our framework’s enhanced discriminative power in handling challenging non-canonical emotional expressions.
However, a critical analysis of our results also reveals important limitations and areas for future improvement, which we discuss in detail below.

5.5.1. Challenges in Distinguishing Semantically Similar Emotions

A primary challenge, reflected in the fine-grained results of Table 6, is the model’s difficulty in distinguishing between semantically similar emotion categories. For instance, while achieving strong performance in most categories, the F1-scores for “angry” (68.3%) and “frustrated” (68.4%) are lower than for “sad” (79.3%) or “excited” (72.9%). This suggests a degree of confusion between emotions that share similar valence and arousal levels, such as the overlap between anger and frustration. This issue stems from two factors. First, the inherent ambiguity in human expression means that these emotions often share highly similar acoustic features (e.g., increased vocal intensity) and lexical cues. Second, capturing the subtle contextual shifts that differentiate them (e.g., frustration often implies an unfulfilled goal, while anger can be more direct) remains a non-trivial task.
Our framework is architecturally designed to mitigate this. The SE module provides a foundational layer of sentiment polarity, helping to separate positive from negative emotions. More critically, our CS-BiGraph module aims to resolve ambiguity by modeling long-range context, learning, for example, that an utterance following a series of failed attempts is more likely to be “frustrated” than “angry”. However, our results indicate that, when local immediate multimodal signals are particularly strong and ambiguous, they can occasionally override the model’s broader contextual understanding. This points to a need for future work on more robust arbitration mechanisms that can better balance local evidence with global context.

5.5.2. The Impact of Data Imbalance and Future Directions

A second factor influencing performance is the well-known class imbalance in the IEMOCAP dataset. As shown in our ablation studies (Table 1), our model achieves its highest F1-score in the “neutral” category (69.9%), which is also the most frequent class. This high performance, while positive, may also indicate a slight bias, where the model learns to default to “neutral” in cases of high uncertainty. This is a common issue that can lead to suboptimal performance on less frequent minority emotion classes.
While the focus of this paper was on architectural innovation, we acknowledge that addressing this data imbalance is a critical next step for improving generalizability. We did not employ explicit mitigation strategies in this work, but future iterations of our framework could readily incorporate them. Promising directions include the following:
(1) class-balanced loss functions, such as focal loss or weighted cross-entropy, to assign higher penalties for misclassifying minority classes;
(2) advanced data augmentation, particularly for synthesizing realistic multimodal data for underrepresented emotions; and
(3) balanced sampling strategies, such as oversampling minority classes during batch creation. We hypothesize that integrating these data-centric strategies with our model-centric innovations will lead to more robust and equitable performance across all emotion categories.
Despite these limitations, by integrating results from both datasets, our model consistently demonstrates strong performance, proving the effectiveness and reliability of its core architectural design in practical applications.

6. Limitations and Future Work Directions

(1) The current multimodal fusion method has only validated the effectiveness of text/audio/video modalities. Future work should extend to additional modalities (e.g., gestures) to verify generalizability.
(2) The emotion modeling focuses solely on polarity dimensions without fully exploring fine-grained features like intensity. Subsequent research could integrate multi-dimensional emotion-labeled datasets and incorporate external knowledge (e.g., gestures and micro-expressions) to enhance model performance.
(3) The dual-graph structure (speaker relationship graph and similarity graph) lacks interactive mechanisms. Future improvements may involve dynamic collaborative graph networks through cross-graph information propagation (e.g., similarity-weighted speaker relationship modeling and structural feature sharing).
(4) The current model has only validated the effectiveness of this method under ideal data conditions. Further research is required to determine whether the model performance remains consistent when one or more modalities are partially missing or contain noise.
(5) Our framework’s performance is inherently linked to the capabilities of the large pre-trained models used for feature extraction (e.g., sBERT and RoBERTa). While leveraging these models provides a strong foundation, the stability and reproducibility of our results are dependent on them. Future work should investigate the framework’s robustness with different, potentially lighter, feature extractors to better assess its architectural modularity and efficiency.
These advancements will enhance the model’s ability to deeply mine multimodal emotional features and further break through the performance bottleneck in emotion recognition.

7. Conclusions

Emotion recognition is a prominent research direction in machine learning and plays a critical role in affective computing. However, challenges such as insufficient multimodal fusion interaction, lack of dynamic emotional weight optimization, and long-range dependency modeling in dialogue history persist.
To address these issues, we propose three innovations: HSF, a hierarchical sequential fusion method; SE, a sentiment-enhanced joint learning framework; and CS-BiGraph, a context-similarity bilevel graph model for dialogue modeling. The integrated framework achieves 69.9% accuracy on the IEMOCAP dataset, setting new benchmarks in “Neutral” and “Happy” emotion recognition. This work demonstrates significant performance improvements and paves the way for future research in emotion recognition applications. Specifically, our framework lays the groundwork for real-world applications such as developing more empathetic chatbots and advanced tools for mental health monitoring.

Author Contributions

Conceptualization, Z.T. and R.Y.; methodology, Z.T. and R.Y.; software, Z.T.; validation, Z.T., R.Y., S.W., J.L. and W.Z.; formal analysis, Z.T.; investigation, Z.T. and R.Y.; writing—original draft preparation, S.W.; writing—review and editing, S.W., Z.T., and R.Y.; visualization, S.W.; supervision, Z.T.; project administration, Z.T.; funding acquisition, Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The public datasets analyzed in this study can be found at their respective sources. The source code and models generated during the study are not publicly available due to institutional policy restrictions.

Acknowledgments

The authors thank all the anonymous reviewers for their insightful comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Picard, R.W. Affective computing: Challenges. Int. J. Hum. Comput. Stud. 2003, 59, 55–64. [Google Scholar] [CrossRef]
  2. Zhou, S.; Jia, J.; Wang, Q.; Dong, Y.; Yin, Y.; Lei, K. Inferring Emotion from Conversational Voice Data: A Semi-Supervised Multi-Path Generative Neural Network Approach. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. Number 1. [Google Scholar] [CrossRef]
  3. Poria, S.; Cambria, E.; Howard, N.; Huang, G.B.; Hussain, A. Fusing audio, visual and textual clues for sentiment analysis from multimodal content. Neurocomputing 2016, 174, 50–59. [Google Scholar] [CrossRef]
  4. Hu, G.; Lin, T.E.; Zhao, Y.; Lu, G.; Wu, Y.; Li, Y. UniMSE: Towards Unified Multimodal Sentiment Analysis and Emotion Recognition. arXiv 2022, arXiv:2211.11256. [Google Scholar] [CrossRef]
  5. Hazarika, D.; Poria, S.; Mihalcea, R.; Cambria, E.; Zimmermann, R. ICON: Interactive Conversational Memory Network for Multimodal Emotion Detection. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 2594–2604. [Google Scholar] [CrossRef]
  6. Majumder, N.; Poria, S.; Hazarika, D.; Mihalcea, R.; Gelbukh, A.; Cambria, E. DialogueRNN: An Attentive RNN for Emotion Detection in Conversations. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 6818–6825, Number 1. [Google Scholar] [CrossRef]
  7. Ghosal, D.; Majumder, N.; Poria, S.; Chhaya, N.; Gelbukh, A. DialogueGCN: A Graph Convolutional Neural Network for Emotion Recognition in Conversation. arXiv 2019, arXiv:1908.11540. [Google Scholar] [CrossRef]
  8. Zhang, D.; Wu, L.; Sun, C.; Li, S.; Zhu, Q.; Zhou, G. Modeling both Context- and Speaker-Sensitive Dependence for Emotion Detection in Multi-speaker Conversations. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 5415–5421. [Google Scholar] [CrossRef]
  9. Tsai, Y.H.H.; Bai, S.; Pu Liang, P.; Kolter, J.Z.; Morency, L.P.; Salakhutdinov, R. Multimodal Transformer for Unaligned Multimodal Language Sequences. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; Volume 2019, pp. 6558–6569. [Google Scholar] [CrossRef]
  10. Shou, Y.; Liu, H.; Cao, X.; Meng, D.; Dong, B. A Low-Rank Matching Attention Based Cross-Modal Feature Fusion Method for Conversational Emotion Recognition. IEEE Trans. Affect. Comput. 2025, 16, 1177–1189. [Google Scholar] [CrossRef]
  11. Wu, L.; Liu, Q.; Zhang, D.; Wang, J.; Li, S.; Zhou, G. Multimodal Emotion Recognition with Auxiliary Sentiment Information. Acta Sci. Nat. Univ. Pekin. 2020, 56, 75–81. [Google Scholar]
  12. Hazarika, D.; Poria, S.; Zadeh, A.; Cambria, E.; Morency, L.P.; Zimmermann, R. Conversational Memory Network for Emotion Recognition in Dyadic Dialogue Videos. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics, New Orleans, LA, USA, 1–6 June 2018; Volume 2018, pp. 2122–2132. [Google Scholar] [CrossRef]
  13. Bao, Y.; Ma, Q.; Wei, L.; Zhou, W.; Hu, S. Speaker-Guided Encoder-Decoder Framework for Emotion Recognition in Conversation. arXiv 2022, arXiv:2206.03173. [Google Scholar] [CrossRef]
  14. Jiao, W.; Lyu, M.; King, I. Real-Time Emotion Recognition via Attention Gated Hierarchical Memory Network. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8002–8009, Number 5. [Google Scholar] [CrossRef]
  15. Ishiwatari, T.; Yasuda, Y.; Miyazaki, T.; Goto, J. Relation-aware Graph Attention Networks with Relational Position Encodings for Emotion Recognition in Conversations. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; Webber, B., Cohn, T., He, Y., Liu, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 7360–7370. [Google Scholar] [CrossRef]
  16. Wu, Y.; Zhang, S.; Li, P. Multi-modal emotion recognition in conversation based on prompt learning with text-audio fusion features. Sci. Rep. 2025, 15, 8855. [Google Scholar] [CrossRef] [PubMed]
  17. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv 2019, arXiv:1908.10084. [Google Scholar] [CrossRef]
  18. Eyben, F.; Wöllmer, M.; Schuller, B. Opensmile: The munich versatile and fast open-source audio feature extractor. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 1459–1462. [Google Scholar] [CrossRef]
  19. Baltrušaitis, T.; Robinson, P.; Morency, L.P. OpenFace: An open source facial behavior analysis toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–9 March 2016; pp. 1–10. [Google Scholar] [CrossRef]
  20. Joshi, A.; Bhat, A.; Jain, A.; Singh, A.V.; Modi, A. COGMEN: COntextualized GNN based Multimodal Emotion recognitioN. arXiv 2022, arXiv:2205.02455. [Google Scholar] [CrossRef]
  21. Busso, C.; Bulut, M.; Lee, C.C.; Kazemzadeh, A.; Mower, E.; Kim, S.; Chang, J.N.; Lee, S.; Narayanan, S.S. IEMOCAP: Interactive emotional dyadic motion capture database. Lang. Resour. Eval. 2008, 42, 335–359. [Google Scholar] [CrossRef]
  22. Bagher Zadeh, A.; Liang, P.P.; Poria, S.; Cambria, E.; Morency, L.P. Multimodal Language Analysis in the Wild: CMU-MOSEI Dataset and Interpretable Dynamic Fusion Graph. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; Gurevych, I., Miyao, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 2236–2246. [Google Scholar] [CrossRef]
  23. Pennington, J.; Socher, R.; Manning, C. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; Moschitti, A., Pang, B., Daelemans, W., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2014; pp. 1532–1543. [Google Scholar] [CrossRef]
  24. Meng, T.; Shou, Y.; Ai, W.; Yin, N.; Li, K. Deep Imbalanced Learning for Multimodal Emotion Recognition in Conversations. IEEE Trans. Artif. Intell. 2024, 5, 6472–6487. [Google Scholar] [CrossRef]
  25. Yun, T.; Lim, H.; Lee, J.; Song, M. TelME: Teacher-leading Multimodal Fusion Network for Emotion Recognition in Conversation. arXiv 2024, arXiv:2401.12987. [Google Scholar] [CrossRef]
  26. Poria, S.; Cambria, E.; Hazarika, D.; Majumder, N.; Zadeh, A.; Morency, L.P. Context-Dependent Sentiment Analysis in User-Generated Videos. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; Barzilay, R., Kan, M.Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; pp. 873–883. [Google Scholar] [CrossRef]
  27. Hu, J.; Liu, Y.; Zhao, J.; Jin, Q. MMGCN: Multimodal Fusion via Deep Graph Convolution Network for Emotion Recognition in Conversation. arXiv 2021, arXiv:2107.06779. [Google Scholar] [CrossRef]
  28. Delbrouck, J.B.; Tits, N.; Brousmiche, M.; Dupont, S. A Transformer-based joint-encoding for Emotion Recognition and Sentiment Analysis. In Proceedings of the Second Grand-Challenge and Workshop on Multimodal Language (Challenge-HML), Online, 10 July 2020; pp. 1–7. [Google Scholar] [CrossRef]
  29. Shenoy, A.; Sardana, A. Multilogue-Net: A Context Aware RNN for Multi-modal Emotion Detection and Sentiment Analysis in Conversation. In Proceedings of the Second Grand-Challenge and Workshop on Multimodal Language (Challenge-HML), Online, 10 July 2020; pp. 19–28. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Chen, M.; Shen, J.; Wang, C. Tailor Versatile Multi-Modal Learning for Multi-Label Emotion Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 9100–9108. [Google Scholar] [CrossRef]
  31. Peng, C.; Chen, K.; Shou, L.; Chen, G. CARAT: Contrastive Feature Reconstruction and Aggregation for Multi-Modal Multi-Label Emotion Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 14581–14589. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed multimodal emotion recognition framework. The framework processes unimodal features (video, audio, and text) through three sequential core modules: (1) multimodal fusion (HSF), which integrates features across modalities; (2) emotional enhancement (SE), which leverages sentiment cues; and (3) graph construction and reasoning (CS-BiGraph), which models contextual dependencies. The final feature representations are fed into a prediction head for classification. Detailed architectures of the HSF, SE, and CS-BiGraph modules are presented in subsequent sections.
Figure 1. Overview of the proposed multimodal emotion recognition framework. The framework processes unimodal features (video, audio, and text) through three sequential core modules: (1) multimodal fusion (HSF), which integrates features across modalities; (2) emotional enhancement (SE), which leverages sentiment cues; and (3) graph construction and reasoning (CS-BiGraph), which models contextual dependencies. The final feature representations are fed into a prediction head for classification. Detailed architectures of the HSF, SE, and CS-BiGraph modules are presented in subsequent sections.
Applsci 15 09622 g001
Figure 2. Architecture of the hierarchical sequential fusion (HSF) module. The inputs x v i d e o , x t e x t , and x a u d i o represent the initial unimodal feature sequences. The module consists of three parallel cross-attention blocks, where each modality is enriched by attending to the other two.
Figure 2. Architecture of the hierarchical sequential fusion (HSF) module. The inputs x v i d e o , x t e x t , and x a u d i o represent the initial unimodal feature sequences. The module consists of three parallel cross-attention blocks, where each modality is enriched by attending to the other two.
Applsci 15 09622 g002
Figure 3. Overall architecture of the sentiment-enhancement (SE) framework. The detailed structures of the global and local locators are presented in Figure 4.
Figure 3. Overall architecture of the sentiment-enhancement (SE) framework. The detailed structures of the global and local locators are presented in Figure 4.
Applsci 15 09622 g003
Figure 4. Detailed architecture of the locator components within the SE framework. (a) Global locator structure; (b) K-th local locator mechanism.
Figure 4. Detailed architecture of the locator components within the SE framework. (a) Global locator structure; (b) K-th local locator mechanism.
Applsci 15 09622 g004
Figure 5. Structure of the context-similarity bilevel graph (CS-BiGraph) module. The model constructs two distinct graphs to capture conversational context: (1) a global similarity-based neighbor graph, where nodes (utterances) are connected based on feature similarity (calculated via cosine similarity and KNN), capturing long-range semantic relationships; (2) a speaker relationship graph, where the structure is shown in Figure 6.
Figure 5. Structure of the context-similarity bilevel graph (CS-BiGraph) module. The model constructs two distinct graphs to capture conversational context: (1) a global similarity-based neighbor graph, where nodes (utterances) are connected based on feature similarity (calculated via cosine similarity and KNN), capturing long-range semantic relationships; (2) a speaker relationship graph, where the structure is shown in Figure 6.
Applsci 15 09622 g005
Figure 6. Illustration of the speaker-to-speaker relational graph in dialogue.
Figure 6. Illustration of the speaker-to-speaker relational graph in dialogue.
Applsci 15 09622 g006
Figure 7. Architecture of the final prediction pipeline. The input consists of the context-aware utterance features z (output from the contextual extractor) and the graph structure generated by the CS-BiGraph module. These are processed by a relational graph convolutional network (R-GCN) to model inter-speaker and intra-speaker dependencies, followed by a graph transformer layer to capture higher-level interactions. A final classifier with a softmax activation function then generates the emotion label predictions (e.g., happy, neutral, etc.).
Figure 7. Architecture of the final prediction pipeline. The input consists of the context-aware utterance features z (output from the contextual extractor) and the graph structure generated by the CS-BiGraph module. These are processed by a relational graph convolutional network (R-GCN) to model inter-speaker and intra-speaker dependencies, followed by a graph transformer layer to capture higher-level interactions. A final classifier with a softmax activation function then generates the emotion label predictions (e.g., happy, neutral, etc.).
Applsci 15 09622 g007
Figure 8. Attention level.
Figure 8. Attention level.
Applsci 15 09622 g008
Figure 9. K-nearest neighbors.
Figure 9. K-nearest neighbors.
Applsci 15 09622 g009
Figure 10. Similarity threshold.
Figure 10. Similarity threshold.
Applsci 15 09622 g010
Figure 11. Comparison with existing works.
Figure 11. Comparison with existing works.
Applsci 15 09622 g011
Table 1. Computational resources and model statistics.
Table 1. Computational resources and model statistics.
ComponentSpecification
Hardware and System
    GPUNVIDIA RTX 3070 Ti (NVIDIA Corporation, Santa Clara, CA, USA) (8 GB VRAM)
    Operating SystemUbuntu 22.04
Software
    Python Version3.10
    PyTorch Version1.12.1
Model Statistics and Training Time
    Total Trainable Parameters7.5 M
    Training Time per Epoch (IEMOCAP)Approx. 7 min
    Total Training Time (IEMOCAP)Approx. 9 h
Table 2. Overall ablation study results on the IEMOCAP dataset.
Table 2. Overall ablation study results on the IEMOCAP dataset.
HappySadNeutralAngryExcitedFrustratedAvg
ModelF1 (%)F1 (%)F1 (%)F1 (%)F1 (%)F1 (%)F1 (%)
Ours54.279.369.968.372.968.470.0
w/o HSF57.978.165.466.876.664.768.7
w/o SE56.181.166.967.473.466.669.3
w/o CS-BiGraph49.878.565.368.776.867.168.8
w/o HSF, SE47.175.569.166.773.369.468.7
w/o HSF, CS-BiGraph55.481.764.168.674.064.468.4
w/o SE, CS-BiGraph51.181.165.166.274.465.368.2
Table 3. Accuracy (Acc) scores (%) of the proposed model in ablation studies on the MOSEI dataset.
Table 3. Accuracy (Acc) scores (%) of the proposed model in ablation studies on the MOSEI dataset.
MethodCMU-MOSEI
Ours48.6
w/o HSF48.5
w/o SE47.9
w/o CS-BiGraph48.2
w/o HSF, SE47.8
w/o HSF, CS-BiGraph48.3
w/o SE, CS-BiGraph47.6
Table 4. Accuracy (Acc) scores (%) of Cs-BiGraph in ablation studies on two datasets.
Table 4. Accuracy (Acc) scores (%) of Cs-BiGraph in ablation studies on two datasets.
MethodIEMOCAPMOSEI
Cs-BiGraph68.947.8
w/o Cosine68.246.5
w/o Similarity Graph65.744.9
w/o GAE67.946.3
w/o Cosine, GAE66.345.6
w/o Similarity Graph, GAE64.543.9
Table 5. Impact of replacing the textual feature extractor on IEMOCAP (Avg F1-%).
Table 5. Impact of replacing the textual feature extractor on IEMOCAP (Avg F1-%).
Textual Feature ExtractorAverage F1-Score (%)
sBERT (our main model)70.0
GloVe (300d vectors)64.8
Table 6. F1-scores of existing models for each emotion category on the IEMOCAP dataset.
Table 6. F1-scores of existing models for each emotion category on the IEMOCAP dataset.
ModelHappySadNeutralAngryExcitedFrustrated
F1 (%)F1 (%)F1 (%)F1 (%)F1 (%)F1 (%)
BC-LSTM35.669.253.566.361.162.4
CMN32.672.956.264.667.963.1
ICON32.874.460.668.268.466.2
DialogueRNN32.878.059.163.373.659.4
DialogueGCN42.884.563.564.263.167.0
MMGCN42.378.761.769.074.362.3
COGMEN44.477.563.162.272.158.6
GraphMFT46.083.163.170.076.963.8
BiGraph47.179.466.767.074.764.0
Ours54.279.369.968.372.968.4
Table 7. Accuracy (Acc) scores (%) of existing models on the CMU-MOSEI dataset.
Table 7. Accuracy (Acc) scores (%) of existing models on the CMU-MOSEI dataset.
ModelAcc (%)
TBJE44.4
Multilogue-Net44.83
Graph-MFN45.0
COGMEN43.9
BiGraph47.2
Ours48.6
TAILOR48.8
CARAT49.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tu, Z.; Yan, R.; Weng, S.; Li, J.; Zhao, W. Multimodal Emotion Recognition Based on Graph Neural Networks. Appl. Sci. 2025, 15, 9622. https://doi.org/10.3390/app15179622

AMA Style

Tu Z, Yan R, Weng S, Li J, Zhao W. Multimodal Emotion Recognition Based on Graph Neural Networks. Applied Sciences. 2025; 15(17):9622. https://doi.org/10.3390/app15179622

Chicago/Turabian Style

Tu, Zhongwen, Raoxin Yan, Sihan Weng, Jiatong Li, and Wei Zhao. 2025. "Multimodal Emotion Recognition Based on Graph Neural Networks" Applied Sciences 15, no. 17: 9622. https://doi.org/10.3390/app15179622

APA Style

Tu, Z., Yan, R., Weng, S., Li, J., & Zhao, W. (2025). Multimodal Emotion Recognition Based on Graph Neural Networks. Applied Sciences, 15(17), 9622. https://doi.org/10.3390/app15179622

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop