Next Article in Journal
Damage Resistance of an fMRI-Spiking Neural Network Based on Speech Recognition Against Stochastic Attack
Previous Article in Journal
Adhesive Performance of Zirconia and Lithium Disilicate Maryland Cantilever Restorations on Prepared and Non-Prepared Abutment Teeth: An In Vitro Comparative Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Graph Learning with Multimodal Fusion for Emotion Recognition in Conversation

1
Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai 200093, China
2
Department of Information Technology, National Technical University Kharkiv Polytechnic Institute, 61000 Kharkiv, Ukraine
3
School of Automation and Electronic Information, Xiangtan University, Xiangtan 411105, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(7), 414; https://doi.org/10.3390/biomimetics10070414
Submission received: 28 April 2025 / Revised: 6 June 2025 / Accepted: 23 June 2025 / Published: 25 June 2025
(This article belongs to the Section Bioinspired Sensorics, Information Processing and Control)

Abstract

Robust emotion recognition is a prerequisite for natural, fluid human–computer interaction, yet conversational settings remain challenging because emotions are shaped simultaneously by global topic flow and local speaker-to-speaker dependencies. Here, we introduce GASMER—Graph-Adaptive Structure for Multimodal Emotion Recognition—a unified architecture that tackles both issues. It uses the correlation structure based on graph neural networks (GNNs) to model the complex dependencies in the conversation, while adaptively learning the graph structure for GNNs. The experiments indicate that our model has strong performance that outperforms all existing graph-based approaches, and remains competitive when compared to recent multimodal fusion models, underscoring the importance of combining fine-grained multimodal fusion with adaptive graph learning for conversational emotion recognition. On the IEMOCAP dataset, GASMER improves accuracy by 2.7% and the weighted F1-score by 3.6% compared to the best baseline. On the MOSEI dataset, it achieves a 1.2% gain in binary classification accuracy (ACC-2).

1. Introduction

Emotions expressed in conversation seldom emerge in isolation [1]; they are tightly coupled to both the thematic context and the affective cues of other speakers [2]. For example, during a discussion about upcoming holidays, the prevailing topic naturally primes participants toward joyful anticipation, making positive affect more likely. Conversely, in a heated debate, one speaker’s aggressive tone can quickly propagate frustration or defensiveness throughout the dialog [3]. Such cases highlight two key factors that any emotion-recognition model must capture [4]: contextual information (topic flow and discourse setting) and inter-speaker dependencies (how one utterance influences another) (as shown in Figure 1).
Treating a dialog as a graph naturally captures these dependencies, and graph neural networks (GNNs) have therefore become a popular choice for conversation modeling. Yet, practical obstacles remain. First, the graph structure is rarely given a priori; it must be inferred from the conversation itself. Second, naïvely constructing a fully connected graph presumes that every utterance influences every other one, incurring O n 2 edges and quickly becoming infeasible for long exchanges. Some studies sidestep this cost with a fixed-width context window, but a static neighborhood is seldom optimal because conversational influence is highly dynamic.
To overcome these limitations, we introduce GASMER—a GNN with Adaptive Structure for Multimodal Emotion Recognition. GASMER rests on two key ideas: Firstly, we introduce the modality fusion adapter (MFA), which can be embedded into each layer of the Transformer, enhancing the efficacy of visual and audio modalities to extract richer and more informative data. Secondly, we propose an adaptive graph structure learning approach that incorporates self-supervised tasks. By dynamically learning the graph structure based on the specific context, our method aims to provide a more efficient and effective solution for modeling conversations.
Recent advances in graph neural networks and multimodal learning have opened up new opportunities for modeling complex social signals in conversation. However, most existing approaches rely on fixed graph structures or treat modality fusion and structural modeling as separate stages. Our motivation in proposing GASMER is to unify these aspects through a dynamic and adaptive framework that is both context-aware and speaker-sensitive. This integrated perspective aligns with the growing interest in embodied AI and emotionally intelligent dialog systems.
In conclusion, we make the following contributions:
  • GASMER framework: We present a multimodal emotion-recognition model that simultaneously exploits contextual history and speaker-to-speaker dependencies through an adaptively learned graph;
  • Self-supervised graph module: We design a lightweight objective that infers dialog structure on-the-fly, eliminating the brittleness of preset graphs;
  • Layer-wise multimodal fusion: The proposed MFA integrates audio-visual signals inside the Transformer, yielding richer, more discriminative representations;
  • Outstanding results: GASMER achieves state-of-the-art performance among graph-based methods, and shows competitive results on the IEMOCAP and MOSEI benchmarks for multimodal emotion recognition.

2. Related Work

Building social chatbots and intelligent dialog systems that can engage in empathetic conversations has long been a central ambition of artificial intelligence [5]. Emotion recognition in conversation (ERC) is critical to this objective, and recent progress has been driven by the increasing availability of high-quality multimodal corpora such as CMU-MOSEI [6] and IEMOCAP [7].

2.1. Multimodal Emotion Recognition

Because of the high correlation observed between emotions and facial expressions [8], the methods using multimodal fusion to improve the effect of emotion recognition tasks are also being increasingly widely used [9]. Datcu et al. [10] integrate audio modality and visual modality for emotion recognition. Joshi et al. [11] adopt a graph neural network to model the inter/intra dependencies of speakers in conversation. Zadeh et al. [12] proposed the Tensor Fusion Network to model the intra- and inter-modality dynamics underlying multimodal sentiment analysis. Wollmer et al. [13] add contextual information to the multimodal task for emotion recognition. Sun et al. [14]; Li et al. [15] treat the conversation as a graph structure and model the contextual information via a graph neural network. Mao et al. [16] propose the concept of emotion dynamics to capture contextual information and use a multi-granular intertalk fusion approach across modalities to model the cross-modal emotion dynamics. Huang et al. [17] proposed to use Transformer to integrate visual and speech information at the model level. Liu et al. [18] proposed CapsGCN to represent the fusion of the multimodal model. Siriwardhana et al. [19] proposed a multimodal emotion recognition framework based on Transformer self-supervised feature fusion, using the trained self-supervised network to extract the features of multimodal information, including textual, audio, and visual features, and at the same time, using a method based on Transformer and the attention mechanism to capture the context of semantic connection between and within modalities. Tan et al. [20] utilized human facial images and EEG signals for multimodal recognition, used CNN to extract face features for classification and SVM-classified EEG signals, and finally obtained the final multimodal emotion classification results through multiple voting. Pandeya et al. [21] extracted face features using 3D-CNN and finally classified them via the late fusion method. Huang et al. [22] proposed the deep multimodal attention fusion method (DMAF), which used the difference and internal correlation between visual and semantic content to identify emotion through a hybrid fusion framework. Lian et al. [23] proposed a Transformer-based session-level multimodal emotion recognition framework containing two steps of context-independent discourse-level feature extraction and context-dependent multimodal feature extraction. One Transformer in the architecture is used to capture the temporal features of unimodal features, and the other cross-modal Transformer is used to learn the cross-modal interaction information on non-aligned multimodal features, and to perform multimodal feature fusion through the attention mechanism.

2.2. Graph Structure Learning

Zhao et al. [24] proposed GAUG-M, which models the weight of edges by calculating the inner product of the features of two nodes. Lim et al. [25] proposed AdaCAD, which considers both node features and graph structure while designing the transformation matrix. To enhance expression ability, Yuan et al. [26] used the cosine kernel to learn node feature representation and encode the local and global graph structure using the additional diffusion kernel. Gidaris et al. [27] created a KNN similarity graph using the cosine similarity of node features as a similarity measure. Wang et al. [28] extended Gidaris’ approach to create a new graph structure based on the cosine similarity of the node features in each layer of the GNN. Halcrow et al. [29] incorporated multiple similarity measures to create KNN graph structures. Zhang et al. [30] proposed to reduce the complexity by defining for each node the local neighborhood and assuming that these local neighborhoods are fully connected. Li et al. [31] generated a fully connected graph based on a bilinear similarity function with learnable parameters. Franceschi et al. [32] learned the Bernoulli distribution for each edge and used these distributions to create the graph structure. Yang et al. [33] updated the initial input structure according to the label and model predictions to increase homogeneity. Chen et al. [34] proposed an iterative approach that projects nodes onto the latent space and constructs an adjacency matrix with a latent representation. Qasim et al. [35] used an MLP as a projection layer to learn the graph structure. Kazi et al. [36] used a GNN as the projection layer and created graph structures separately using features from different layers. To solve the supervised hunger problem, Fatemi et al. [37] introduced a self-supervised task to guide graph generator learning.
Collectively, these studies underscore two trends crucial in our work: (i) the benefit of adaptive graph learning for capturing dynamic conversational structure and (ii) the effectiveness of fine-grained multimodal fusion for robust emotion recognition. GASMER builds upon and unifies these directions through a self-supervised, layer-wise multimodal GNN with an adaptively learned topology.

3. Proposed Method

In the ERC task, an important emotion recognition strategy involves modeling conversational scenarios. In a conversation involving multiple speakers, each speaker’s emotions are influenced by the context and the responses of other speakers. Inspired by this insight, we have primarily focused on two types of situational information to model the conversational scenario: contextual information and speaker-level dependency information.
In our model, we represent these two types of scenario information as follows: Global information—this aspect captures the impact of the overall context on the emotional state of the utterance. Local information—this aspect captures the influence of inter-speaker dependencies and intra-speaker dependencies on the emotional state of the utterance.
Global Information: Recognizing the importance of extracting contextual information and handling multimodal data, we advocate for the adoption of a unified model. In this study, we employ the Transformer encoder [38] to extract global information. To capture the full influence of context on each utterance, we depart from traditional encoding methods that involve adding positional encoding to input features, opting instead for a vanilla Transformer encoder. However, in preparation for the subsequent extraction of local information and the fusion of multimodal data, we introduce the modality fusion adapter (MFA) to be embedded within the Transformer encoder, enabling the learning of temporal information within the MFA.
Local Information: In binary or multi-party dialog systems, the information in adjacent utterances often exerts the greatest influence on emotions. We model the relationship between adjacent utterances through speaker-level dependent information, which is further categorized into inter-speaker and intra-speaker dependencies. Inter-speaker dependence refers to the emotional impact of one person on another in the conversation, while intra-speaker dependence represents the individual’s emotional influence within the conversation, also referred to as emotional inertia. Our approach shares similarities with DialogueGCN [39] and COGMEN [11]; we model the conversation as a graph, with each utterance representing a node and the relationships between nodes being represented by directed edges. We classify the edge relationships into two categories: the relationship between utterances spoken by the same speaker and the relationship between utterances spoken by different speakers. To learn the relationships between the nodes, we utilize the Relational Graph Convolutional Network [40] and Graph Transformer [41].
Additionally, to address challenges related to the learning difficulties caused by preset graph structure defects in graph neural networks, we have designed a self-supervised graph structure learning module to overcome this issue.

3.1. Overall Architecture

As illustrated in Figure 2, GASMER comprises the context extractor, graph generator, local feature extractor, and emotion classifier. Initially, the context feature extractor is employed to extract text, audio, and video features from the input utterance to acquire global context information. Diverging from prior work, we have integrated the modality fusion adapter into the Transformer encoder layer to combine text, audio, and video features, thereby enhancing the temporal information within the Transformer encoder. The features extracted by the context feature extractor serve as inputs for the graph generator to derive the graph structure, which, along with the graph node features, is fed into the graph neural network (GNN). The local feature extractor, leveraging GNN components like the Relational Graph Convolutional Network and Graph Transformer, captures inter-speaker and intra-speaker dependencies from the graph structure. Ultimately, the output features of the local feature extractor are utilized for emotion prediction through the emotion classifier, implemented as a linear layer.

3.2. Context Extractor

The context extractor inputs multimodal utterance features to represent each conversation utterance. The specific input form can be expressed as follows:
U = u 1 , u 2 , , u i   R n
u i = u i a , u i t , u i v
where n is the number of utterances in a conversation, u i a   R d a is the audio feature of utterances, u i t   R d t is the text feature of utterances, and u i v   R d v is the video feature of utterances. Previous work has proven that the text modality is more indicative than other modalities, so we adopt the strategy of using the text modality as the input to the Transformer encoder layer and fusing other modalities in the modality fusion adapter. According to the vanilla Transformer encoding layer, we define a Query, a Key, and a Value vector to encode the input features in the following form:
X = x t = u 1 t , u 2 t , , u i t
Q i = X W i , q
K i = X W i , k
V i = X W i , v
where W i , q ,   W i , k ,   W i , v   R d t × k . Scaled dot product attention captures the interaction relationship between Key and Query and outputs an attention coefficient graph α h as follows:
α i = σ j Q i K i k
where α h   R n × n represents the attention coefficient of the individual attention head for each discourse in the following form:
h e a d i = α i V i   R n × k
H = h e a d 1 h e a d 2   h e a d h W
where W   R k h × d t , H represents the number of attention heads in the multi-head attention mechanism. The obtained H   R h × d t performs the residual connection and enters the feedforward layer in the following form:
H = L a y e r N o r m X + H
Z = R e L U H W 1 W 2
Z = L a y e r N o r m H + Z
where W 1   R d t × m , W 2   R m × d t . The textual features provided by the Transformer encoder that contain the conversation context information are z 1 , z 2 , , z n T = Z   R n × d t .
Different from the previous feature fusion method of simple concatenation, we embedded the modality fusion adapter into the Transformer encoder, fused text features, audio features, and video features, and learned the temporal information that is not included in the Transformer encoder. We refer to the traditional adapter [42] to connect the modality fusion adapter after each Transformer encoder layer.
We take the output text features Z of the Transformer encoder, the audio features of utterance, and the video features of utterance together as the input of the modality fusion adapter in the following form:
x i a v t = u i a z i u i v     R d
X = x a v t = x 1 a v t , x 2 a v t , , x n a v t T
where d = d a + d t + d v . The obtained multimodal fusion feature X enters the modality fusion adapter for downward projection and upward projection to optimize the specific parameters of multimodal fusion, and finally, the feature dimension of X is mapped back to the dimension of text features. For the context extractor of layer j , the multimodal fusion is expressed as follows:
X u = σ W u X j 1 + b u
X v = W v X u + b v
X j = W X v + X j 1
X j = L a y e r N o r m X j
where σ Sigmoid function, W u , W v , W , b u , b v are learnable parameters, and X j 1 denotes the multimodal fusion adapter after   j 1 Transformer layers. At the same time, in order to better capture the temporal information, we follow the practice of AIM [43] and define the input of the multimodal fusion adapter as X   R b × l × d , where b is the size of the batchsize and l is the length of the conversation, so as to learn the temporal relationship between each utterance in the conversation. This explicit operation helps the model to enhance the temporal modeling and keep the number of parameters stable.

3.3. Graph Generator

By considering the significant influence of adjacent utterances on emotions during conversations and taking into account the typical conversation length characteristics, we have adopted a strategy to discretize the adjacency radius. This approach simplifies the conventional task of generating graph structures by focusing on classifying the adjacency radius of graph nodes. Through this simplification, we are able to generate adjacency vectors that effectively represent the graph structure. This method allows us to capture the emotional impact of adjacent utterances in conversations more accurately, leading to improved emotion modeling within dialog systems. We use the MLP as a graph-generator function:
G M L P X ; θ G = Z
P = s o f t m a x Z
N = a r g m a x P
where X   R n × d t is the output of the context extractor, θ is the learnable parameters of MLP, Z   R n × w is the result vector of the graph node neighbor radius,   w is the number of categories after the discretization of the adjacent radius, and N   R n is the last predicted node adjacency radius. Finally, the adjacency vector is generated through N , which considers the relationship between past N utterances and future N utterances.

3.4. Graph Neural Network Components

Considering the significance of speaker-level dependence information, we have leveraged graph-based models to represent these relationships. Within this framework, each utterance is associated with speaker-level and temporal dependencies, which are systematically interconnected through various relationship types. Specifically, if there are M speakers in a conversation, then there are R = M s p e a k e r   o f   u i M s p e a k e r   o f   u j 2 u i   o c c u r s   b e f o r e   o r   a f t e r   u i = 2 M 2 different relationships in the graph. This approach allows us to comprehensively capture the interplay between speaker-level dependence and temporal dynamics within conversations, enabling a more nuanced understanding of the underlying dependencies and facilitating the improved modeling of dialog interactions.
In our framework, we recognize that each speaker in a conversation is influenced by the utterances of other speakers. To model these dependencies effectively, we define clear relationships and employ the Relational Graph Convolution Network (R-GCN) to capture inter-speaker and intra-speaker dependencies, enabling the extraction of local information within the dialog context. The R-GCN operates based on the types of edges present in the graph structure, allowing it to leverage the information from adjacent nodes to update the representation of each node. This tailored approach enables us to capture the intricate relationships and dependencies between speakers in conversations, enhancing the overall understanding of speaker interactions and facilitating more accurate information extraction at the local level.
x i = W 0 x i + r   R j   N r i 1 N r i W r x j
where N r i is the set of adjacent node indices of node i under the relation r   R , W 0 and W r are learnable parameters, N r i are the normalized constants, and x j are the utterance level features from the context extractor.
To enhance the information extracted from the node features, we used GraphTransformer to process the node features obtained from the R-GCN in the following form:
h i = W 1 x i + j   N r i α i , j W 2 x j
where W 1 and W 2 are learnable parameters and α i , j is the attention coefficient calculated via scaled multi-head dot product attention:
α i , j = s o f t m a x W 3 x i T W 4 x j d
where W 3 and W 4 are learnable parameters and d is the scale coefficient.
For a concrete example illustrating how speaker-level relationships are represented in our constructed graph, please refer to the Appendix A.

3.5. Emotion Classifier

In the final stage of our framework, we employ an emotion classifier that utilizes the features extracted by the Graph Transformer. The obtained features are fed through a linear layer to predict the corresponding emotions associated with each node in the graph. The classification task loss is calculated using a SoftMax function to obtain the emotion probability distribution for each node. Subsequently, the cross-entropy loss is computed to measure the disparity between the predicted probability distribution and the actual emotions, enabling the model to learn and optimize its parameters for accurate emotion classification.
h i = R e L U W 1 h i + b 1
P i = s o f t m a x W 2 h i + b 2
y i ^ = a r g m a x P i
where y i ^ is the emotional label corresponding to the predicted utterance.

3.6. Self-Supervision Task

In the existing framework, it is feasible to establish a comprehensive pipeline for conversation emotion recognition utilizing the outlined scheme. However, relying solely on the loss of the classification task for supervising graph structure learning may not yield optimal results. Moreover, parameter fluctuations in the graph generator can significantly impact the accuracy of the overall conversation emotion recognition task. To mitigate these challenges, we introduce a self-supervised task and adopt a differentiated learning rate strategy to separate the training of the graph generator from that of other network structures.
To enhance the efficacy of graph structure learning, incorporating prior information into the graph generation task can be beneficial. The selection of appropriate prior information is crucial as it influences the inductive bias of the model learning process. In this context, we establish a prior graph structure based on the hypothesis: a graph structure suitable for predicting the characteristics of the node is also suitable for predicting the node label [37].
Firstly, considering an extreme case where the node features are the same as its node labels. In the scenario, the graph structure that is effective for predicting the features exhibits homogeneity. Due to the feature being equivalent to the label, the graph structure also demonstrates homogeneity in predicting the labels. As a result, the same graph structure that accurately predicts the features is inherently suitable for predicting the node labels as well.
In general, the node features may not be directly equivalent to the node labels. However, it is possible that a subset of the node features exhibits strong predictability for the labels. In such cases, the graph structure that is effective at predicting this subset of features is likely to be highly predictive for the labels as well. When a specific subset of features holds significant predictive power for the labels, the graph structure aligned with predicting these features also demonstrates a high level of homogeneity in relation to the labels. Consequently, this graph structure is also well-suited for predicting the node labels, due to the strong predictive relationship between the subset features and the labels.
Introducing prior information through a self-supervised task is a valuable approach for enhancing model learning. In this context, we employ a denoising autoencoder (DAE) as the self-supervised task. The denoising autoencoder shares the same structure as the GNN components utilized earlier in the model. It takes as input the noisy version of the node features, derived from the context feature extractor, and the adjacency vector generated by the graph generator. The autoencoder’s objective is to reconstruct the original, denoised node features by removing the added noise elements. We set i d x as the corresponding index of the noise added elements and X i d x as the index value, and set the loss function for the denoising task as follows:
L D A E = L X i d x , G N N D A E X ~ , A ; θ G N N D A E i d x
where A is the adjacency vector generated by the graph generator, L is the mean square error loss, X ~ is the node feature of the noisy version, and θ G N N D A E is a learnable parameter of the denoising autoencoder. Finally, our entire task loss function is as follows:
L = L C + λ L D A E
where L C is the loss of emotion classification, L D A E is the denoising autoencoder loss, and λ is the importance coefficient.

4. Experiments

We compared the GASMER and baseline models on two widely used datasets to evaluate the effect of our proposed model, and performed sufficient ablation experiments to demonstrate the effectiveness of our proposed module.

4.1. Datasets

We selected two widely used datasets for emotion recognition: the Interactive Emotional Dyadic Motion Capture (IEMOCAP) dataset [7] and the Multimodal Opinion Sentiment and Emotion Intensity (MOSEI) dataset [6]. The detailed statistics of the two datasets are shown in Table 1.
MOSEI is a multimodal emotion recognition dataset using 22,860 movie clips from YouTube. This dataset is annotated with seven sentiment annotations (−3(highly negative) to +3(highly positive)) and six emotional labels—happiness, sadness, disgust, fear, surprise, and anger—for each sample. It is important to note that the emotional labels provided in different datasets are often different. IEMOCAP is a binary multimodal emotion recognition dataset, consisting of 7433 samples. We selected two versions of six emotion classifications (happiness, sadness, neutral, anger, excitement, and frustration) and four emotion classifications (anger, sadness, happiness, and neutral) for testing.
For CMU-MOSEI, in addition to the 7-class emotion classification task (ACC-7), we also conduct a binary classification task (ACC-2). In this setting, all samples with positive sentiment scores (greater than 0) are labeled as “positive”, and those with negative sentiment scores (less than 0) are labeled as “negative”. This binary setup is widely adopted in prior works to evaluate polarity-level emotion recognition.

4.2. Data Preprocessing

Like some previous works [44], for IEMOCAP, we used OpenSmile [45] to extract audio features, video features using the method of Baltrusaitis et al. [46], and text features using sBERT [47]. For MOSEI, we extracted audio features using the method of Delbrouck et al. [48] and used Librosa [49] to extract a specific number of filter sets. Video features were extracted using the method of Baltrusaitis et al. [46] and text features were extracted using sBERT.

4.3. Evaluating Indicator

For MOSEI, we use two evaluation settings: ACC-7 for 7-class emotion classification, and ACC-2 for the binary classification of sentiment polarity. The binary task distinguishes between positive and negative sentiments, as described in Section 4.1.

4.4. Baselines

We compared our method to baseline models in many emotion recognition tasks to comprehensively evaluate its performance. For the IEMOCAP, the specific evaluation results are shown in Table 2. The baseline models used include the Tensor Fusion Network (TFN) [12], MMGCN [50], Multiview Sequential Learning Memory Fusion Network (MFN) [51], DialogueRNN [52], DialogueGCN [39], Multimodal Dynamic Fusion Network (MM-DFN) [53], COGMEN [53], and Interactive Conversational Memory Network ICON [54]. For the MOSEI, the specific assessment results are shown in Table 3. The baseline models include the Tensor Fusion Network (TFN), COGMEN, Multimodal Factorization Model (MFM) [55], Low-rank Multimodal Fusion (LMF) [56], Interaction Canonical Correlation Network (ICCN) [57], Self-MM [58], and MMIM [59].

4.5. Experimental Settings

We use PyTorch (2.0.1) [60] as the training framework, PyG [61] to build the GNN components in our model, and Bayesian optimization for hyperparameter tuning. Our trained batch size is set to 64, the graph generator learning rate is set to 3 × 10−5, and the remaining components learning rate is set to 1 × 10−4. The audio feature size is 100, video feature size is 512, text feature size is 768, and fusion vector size is 768. In the loss function, we set the importance coefficient = 10.

4.6. Results

GASMER is significantly better than the SOTA among graph-based methods in all indicators. Compared to the previous SOTA among graph-based methods, GASMER increased the ACC-2 by 1.2% in MOSEI (vs. MMIM), ACC by 2.7% in IEMOCAP (vs. MM-DFN), and 3.6% for the weighted F1-score (vs. COGMEN). Specifically, Table 2 shows the comparison results on IEMOCAP(6-way), showing that GASMER significantly outperforms previous graph-based baseline models in the ACC and weighted F1-score. It is evident that the classification scores for happiness, sadness, and excitement emotions have significantly improved compared to the previous baseline model. This improvement can be attributed to the incorporation of the graph learning module in our approach. Unlike the previous baseline models, our model considers the impact of the graph structure, which contributes to the enhanced performance.
We acknowledge that more recent multimodal fusion approaches such as the work by Ryumina et al. [62] have achieved higher overall results on both IEMOCAP and MOSEI. However, their approach focuses primarily on gated attention mechanisms across modalities, while our method emphasizes the dynamic learning of speaker relationships and conversational structure via self-supervised adaptive graph construction. As such, GASMER continues to offer complementary benefits within graph-based modeling paradigms.
Table 3 presents the comparison results on MOSEI, indicating that GASMER outperforms the previous graph-based baseline models in terms of the ACC-2 and demonstrates comparable performance to the current SOTA in terms of the ACC-7. To better handle class imbalance, we report the Unweighted Average Recall (UAR) or IEMOCAP(6-way), which computes the average recall across all classes, giving equal weight to each class regardless of its sample size. The results show that GASMER consistently achieves high UAR scores, demonstrating its robustness across uneven class distributions. Overall, GASMER consistently outperforms prior graph-based models in most instances, highlighting its superiority in emotion recognition tasks and validating the effectiveness of the GASMER framework.

4.7. Analysis

Upon analyzing the predictions made by our model, we observed a considerable improvement in classifying similar emotions compared to previous baseline models. However, there are still certain emotions, such as anger and frustration, that our model struggles to classify accurately. Additionally, due to the substantial proportion of neutral samples, several instances are wrongly classified as neutral. To illustrate this, we have included the model prediction results for IEMOCAP(6-way) in Figure 3 and calculate the recall as shown in Table 4.
To confirm the impact of GASMER’s graph learning module on the node feature representation, we utilized UMAP [63] to visualize the multimodal fusion representation output by the context extractor and the feature representation after the graph learning module. Figure 4 clearly demonstrates that the features following the graph learning module exhibit improved emotion clusters, highlighting the significance of capturing conversational local dependence through the graph learning module for emotion recognition.

4.8. Ablation Study

We conducted a series of ablation experiments on GASMER, and the specific results are presented in Table 5. In our ablation study, we treat the text modality (T) as the base configuration. This is because prior research and our own experiments show that text features typically carry the most semantically discriminative information for emotion recognition tasks. As such, we do not perform ablation on audio-only (A), video-only (V), or audio-plus-video (A + V) configurations without text, since our full model architecture is built upon the text backbone. Instead, we evaluate the impact of adding audio (A) and/or video (V) on top of text to examine multimodal complementarity. By systematically removing one or several modalities, we aimed to assess the impact of multimodal information on model performance. Our findings indicate that the exclusion of modality information leads to a decline in model performance, underscoring the necessity of leveraging multimodal information for effective emotion recognition tasks. Moreover, the results demonstrate the complementary nature of audio, text, and video modality information.
Interestingly, we observed that audio modality information plays a more significant role than video modality information, with instances where the addition of video modality information even negatively affects the results. This could be attributed to noise in the video modality and its lack of alignment with other modalities. However, GASMER consistently achieves notable performance when utilizing all modalities, showcasing its capability to capture complex interrelationships between modalities.
Furthermore, upon testing the removal of the MFA layer and graph learning (GL) module from GASMER, we observed a decrease in performance, highlighting the effectiveness of these components in multimodal emotion recognition tasks. While we primarily describe the experimental results on IEMOCAP due to space constraints, similar trends were also observed on MOSEI.

5. Discussion

In this study, we proposed GASMER, a graph-adaptive multimodal emotion recognition framework that integrates audio, video, and text modalities through self-supervised graph structure learning. Our experiments on IEMOCAP and MOSEI demonstrate that GASMER achieves superior performance compared to existing methods. Ablation experiments confirm the complementary role of multimodal inputs and the effectiveness of the proposed modality fusion adapter (MFA) and adaptive graph learning (GL) modules. Further visualization using UMAP highlights the improved emotion clustering enabled by the graph structure, demonstrating its contribution to contextual understanding.
We also performed a detailed error analysis. The model shows notable improvement in classifying emotions such as happiness and excitement, but still struggles with confusion between similar emotions like anger and frustration. Additionally, the overrepresentation of neutral samples contributes to prediction bias. These observations emphasize the importance of balanced datasets and more discriminative modeling for fine-grained emotions.
Despite these advancements, the current system operates in an offline mode and relies heavily on global contextual information. For real-world deployment, developing real-time emotion recognition capabilities remains a key challenge. Moreover, the reliance on accurate modality alignment and the performance sensitivity to video noise are areas for future enhancement.

6. Conclusions

We presented GASMER, a novel graph-adaptive multimodal emotion recognition framework designed for conversational settings. The model unifies multimodal feature fusion and speaker-level dependency modeling through a self-supervised graph learning mechanism. GASMER demonstrates competitive or superior performance on benchmark datasets, supported by extensive experiments and analyses.
In future work, we plan to (i) improve the model’s ability to distinguish subtle emotional variations through fine-grained modeling and (ii) develop methods to support real-time contextual emotion recognition for interactive and intelligent systems.

Author Contributions

J.L. (Jian Liu) conceived the project and designed the research process. J.L. (Jian Li) analyzed the experiment data. Z.M. and N.L. analyzed the computational data and created the figures. Q.L. and Y.Y. jointly supervised this work. All authors contributed to the writing of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National key Research & Development plan of Ministry of Science and Technology of China (Grant No. 2023YFC3605800, 2023YFC3605803) and the Young Scientists Fund of the National Natural Science Foundation of China (Grant No. 62206175).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Acknowledgments

The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

To better demonstrate how we model a conversation into a graph and represent speaker-level relationships in the graph, we provide an example to describe the relationship between utterances in a conversation. In this example, we assume that there are two speakers, speaker1 and speaker2, who engage in a conversation consisting of five utterances. The adjacency radius N for each utterance is set to 2. u i 2 , u i , and u i + 2 represent the utterances of speaker1, while u i 1 and u i + 1 represent the utterances of speaker2. After modeling this conversation into a graph with relationships, its specific form is presented in Table A1.
Table A1. The relationships between each node are represented by (→) for future utterances and (←) for past utterances.
Table A1. The relationships between each node are represented by (→) for future utterances and (←) for past utterances.
NodeIntra-Speaker RelationshipInter-Speaker Relationship
u i 2 u i 2 u i 2 ,   u i 2 u i ,   u i 2 u i + 2 u i 2 u i 1 ,   u i 2 u i + 1
u i 1 u i 1 u i 1 ,   u i 1 u i + 1 u i 1 u i 2 ,   u i 1 u i ,   u i 1 u i + 2
u i u i u i 2 ,   u i u i ,   u i u i + 2 u i 2 u i 1 ,   u i 2 u i + 1
u i + 1 u i + 1 u i 1 ,   u i + 1 u i + 1 u i + 1 u i 2 ,   u i + 1 u i ,   u i + 1 u i + 2
u i + 2 u i + 2 u i 2 ,   u i + 2 u i ,   u i + 2 u i + 2 u i + 2 u i 1 ,   u i + 2 u i + 1
Table A2 shows the types of relationships existing in this conversation.
Table A2. The types of relationships between the two speakers in the conversation.
Table A2. The types of relationships between the two speakers in the conversation.
Relationship TypeTemporal RelationshipRelationship
1Past u s p e a k e r 1 u s p e a k e r 1
2Past u s p e a k e r 1 u s p e a k e r 2
3Past u s p e a k e r 2 u s p e a k e r 1
4Past u s p e a k e r 2 u s p e a k e r 2
5Future u s p e a k e r 1 u s p e a k e r 1
6Future u s p e a k e r 1 u s p e a k e r 2
7Future u s p e a k e r 2 u s p e a k e r 1
8Future u s p e a k e r 2 u s p e a k e r 2

References

  1. Pereira, P.; Moniz, H.; Carvalho, J.P. Deep emotion recognition in textual conversations: A survey. Artif. Intell. Rev. 2025, 58, 10. [Google Scholar] [CrossRef]
  2. Alhussein, G.; Alkhodari, M.; Khandoker, A.H.; Hadjileontiadis, L.J. Novel speech-based emotion climate recognition in peers’ conversations incorporating affect dynamics and temporal convolutional neural networks. IEEE Access 2025, 13, 16752–16769. [Google Scholar] [CrossRef]
  3. Xie, Y.; Mao, R. PGIF: A Personality-Guided Iterative Feedback Graph Network for Multimodal Conversational Emotion Recognition. In IEEE Transactions on Computational Social Systems; IEEE: Piscataway, NJ, USA, 2025. [Google Scholar]
  4. Van, C.T.; Tran, T.V.; Nguyen, V.; Hy, T.S. Effective Context Modeling Framework for Emotion Recognition in Conversations. In Proceedings of the ICASSP 2025—2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025; pp. 1–5. [Google Scholar]
  5. Zhou, L.; Gao, J.; Li, D.; Shum, H.-Y. The design and implementation of xiaoice, an empathetic social chatbot. Comput. Linguist. 2020, 46, 53–93. [Google Scholar] [CrossRef]
  6. Zadeh, A.B.; Liang, P.P.; Poria, S.; Cambria, E.; Morency, L.-P. Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 2236–2246. [Google Scholar]
  7. Busso, C.; Bulut, M.; Lee, C.-C.; Kazemzadeh, A.; Mower, E.; Kim, S.; Chang, J.N.; Lee, S.; Narayanan, S.S. IEMOCAP: Interactive emotional dyadic motion capture database. Lang. Resour. Eval. 2008, 42, 335–359. [Google Scholar] [CrossRef]
  8. Ekman, P. Facial expression and emotion. Am. Psychol. 1993, 48, 384. [Google Scholar] [CrossRef]
  9. Sebe, N.; Cohen, I.; Gevers, T.; Huang, T.S. Multimodal approaches for emotion recognition: A survey. In Internet Imaging VI; SPIE: Bellingham, WA, USA, 2005; pp. 56–67. [Google Scholar]
  10. Datcu, D.; Rothkrantz, L.J. Semantic audiovisual data fusion for automatic emotion recognition. In Emotion Recognition: A Pattern Analysis Approach; John Wiley and Sons: Hoboken, NJ, USA, 2015; pp. 411–435. [Google Scholar]
  11. Joshi, A.; Bhat, A.; Jain, A.; Singh, A.V.; Modi, A. COGMEN: COntextualized GNN based Multimodal Emotion recognitioN. arXiv 2022, arXiv:2205.02455. [Google Scholar] [CrossRef]
  12. Zadeh, A.; Chen, M.; Poria, S.; Cambria, E.; Morency, L.P. Tensor Fusion Network for Multimodal Sentiment Analysis. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017. [Google Scholar]
  13. Wöllmer, M.; Metallinou, A.; Eyben, F.; Schuller, B.; Narayanan, S. Context-Sensitive Multimodal Emotion Recognition from Speech and Facial Expression Using Bidirectional Lstm Modeling: Proceedings of the Interspeech 2010, Chiba, Japan, 26–30 September 2010; International Speech Communication Association (ISCA): Baixas, France, 2010. [Google Scholar]
  14. Sun, Y.; Yu, N.; Fu, G. A discourse-aware graph neural network for emotion recognition in multi-party conversation. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, Punta Cana, Dominican Republic, 16–20 November 2021; pp. 2949–2958. [Google Scholar]
  15. Li, J.; Lin, Z.; Fu, P.; Wang, W. Past, present, and future: Conversational emotion recognition through structural modeling of psychological knowledge. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, Punta Cana, Dominican Republic, 16–20 November 2021; pp. 1204–1214. [Google Scholar]
  16. Mao, Y.; Sun, Q.; Liu, G.; Wang, X.; Shen, J. DialogueTRM: Exploring the Intra- and Inter-Modal Emotional Behaviors in the Conversation. arXiv 2020, arXiv:2010.07637. [Google Scholar]
  17. Huang, J.; Tao, J.; Liu, B.; Lian, Z.; Niu, M. Multimodal transformer fusion for continuous emotion recognition. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3507–3511. [Google Scholar]
  18. Liu, J.; Chen, S.; Wang, L.; Liu, Z.; Fu, Y.; Guo, L.; Dang, J. Multimodal emotion recognition with capsule graph convolutional based representation fusion. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 6339–6343. [Google Scholar]
  19. Siriwardhana, S.; Kaluarachchi, T.; Billinghurst, M.; Nanayakkara, S. Multimodal emotion recognition with transformer-based self supervised feature fusion. IEEE Access 2020, 8, 176274–176285. [Google Scholar] [CrossRef]
  20. Tan, Y.; Sun, Z.; Duan, F.; Solé-Casals, J.; Caiafa, C.F. A multimodal emotion recognition method based on facial expressions and electroencephalography. Biomed. Signal Process. Control 2021, 70, 103029. [Google Scholar] [CrossRef]
  21. Pandeya, Y.R.; Lee, J. Deep learning-based late fusion of multimodal information for emotion classification of music video. Multimed. Tools Appl. 2021, 80, 2887–2905. [Google Scholar] [CrossRef]
  22. Huang, F.; Zhang, X.; Zhao, Z.; Xu, J.; Li, Z. Image–text sentiment analysis via deep multimodal attentive fusion. Knowl.-Based Syst. 2019, 167, 26–37. [Google Scholar] [CrossRef]
  23. Lian, Z.; Liu, B.; Tao, J. CTNet: Conversational transformer network for emotion recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 985–1000. [Google Scholar] [CrossRef]
  24. Zhao, T.; Liu, Y.; Neves, L.; Woodford, O.; Jiang, M.; Shah, N. Data augmentation for graph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; pp. 11015–11023. [Google Scholar]
  25. Lim, J.; Um, D.; Chang, H.J.; Jo, D.U.; Choi, J.Y. Class-attentive diffusion network for semi-supervised classification. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; pp. 8601–8609. [Google Scholar]
  26. Yuan, J.; Cao, M.; Cheng, H.; Yu, H.; Xie, J.; Wang, C. A unified structure learning framework for graph attention networks. Neurocomputing 2022, 495, 194–204. [Google Scholar] [CrossRef]
  27. Gidaris, S.; Komodakis, N. Generating classification weights with gnn denoising autoencoders for few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 21–30. [Google Scholar]
  28. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. (Tog) 2019, 38, 1–12. [Google Scholar] [CrossRef]
  29. Halcrow, J.; Mosoi, A.; Ruth, S.; Perozzi, B. Grale: Designing networks for graph learning. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, San Diego, CA, USA, 23–27 August 2020; pp. 2523–2532. [Google Scholar]
  30. Zhang, J.; Zhang, H.; Sun, L.; Xia, C. Only attention is needed for learning graph representations. arXiv 2020, arXiv:2001.05140 v2. [Google Scholar]
  31. Li, R.; Wang, S.; Zhu, F.; Huang, J. Adaptive graph convolutional neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  32. Franceschi, L.; Niepert, M.; Pontil, M.; He, X. Learning discrete structures for graph neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 1972–1982. [Google Scholar]
  33. Yang, L.; Kang, Z.; Cao, X.; Jin, D.; Yang, B.; Guo, Y. Topology Optimization based Graph Convolutional Network. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 4054–4061. [Google Scholar]
  34. Chen, Y.; Wu, L.; Zaki, M.J. Deep Iterative and Adaptive Learning for Graph Neural Networks. arXiv 2019, arXiv:1912.07832. [Google Scholar]
  35. Qasim, S.R.; Kieseler, J.; Iiyama, Y.; Pierini, M. Learning representations of irregular particle-detector geometry with distance-weighted graph networks. Eur. Phys. J. C 2019, 79, 608. [Google Scholar] [CrossRef]
  36. Kazi, A.; Cosmo, L.; Ahmadi, S.-A.; Navab, N.; Bronstein, M.M. Differentiable graph module (dgm) for graph convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1606–1617. [Google Scholar] [CrossRef]
  37. Fatemi, B.; El Asri, L.; Kazemi, S.M. SLAPS: Self-supervision improves structure learning for graph neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 22667–22681. [Google Scholar]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems: Proceedings of the 30th Annual Conference on Neural Information Processing Systems (NeurIPS 2017), San Diego CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  39. Ghosal, D.; Majumder, N.; Poria, S.; Chhaya, N.; Gelbukh, A. Dialoguegcn: A graph convolutional neural network for emotion recognition in conversation. arXiv 2019, arXiv:1908.11540. [Google Scholar] [CrossRef]
  40. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the Semantic Web: 15th International Conference, ESWC 2018, Heraklion, Greece, 3–7 June 2018; pp. 593–607. [Google Scholar]
  41. Shi, Y.; Huang, Z.; Feng, S.; Zhong, H.; Wang, W.; Sun, Y. Masked label prediction: Unified message passing model for semi-supervised classification. arXiv 2020, arXiv:2009.03509. [Google Scholar] [CrossRef]
  42. Houlsby, N.; Giurgiu, A.; Jastrzebski, S.; Morrone, B.; De Laroussilhe, Q.; Gesmundo, A.; Attariyan, M.; Gelly, S. Parameter-efficient transfer learning for NLP. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 2790–2799. [Google Scholar]
  43. Yang, T.; Zhu, Y.; Xie, Y.; Zhang, A.; Chen, C.; Li, M. Aim: Adapting image models for efficient video action recognition. arXiv 2023, arXiv:2302.03024. [Google Scholar]
  44. Fan, C.; Lin, J.; Mao, R.; Cambria, E. Fusing pairwise modalities for emotion recognition in conversations. Inf. Fusion 2024, 106, 102306. [Google Scholar] [CrossRef]
  45. Eyben, F.; Wöllmer, M.; Schuller, B. Opensmile: The munich versatile and fast open-source audio feature extractor. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 1459–1462. [Google Scholar]
  46. Baltrušaitis, T.; Robinson, P.; Morency, L.-P. Openface: An open source facial behavior analysis toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar]
  47. Reimers, N.; Gurevych, I. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv 2019, arXiv:1908.10084. [Google Scholar] [CrossRef]
  48. Delbrouck, J.-B.; Tits, N.; Brousmiche, M.; Dupont, S. A transformer-based joint-encoding for emotion recognition and sentiment analysis. arXiv 2020, arXiv:2006.15955. [Google Scholar] [CrossRef]
  49. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. Librosa: Audio and music signal analysis in python. SciPy 2015, 2015, 18–24. [Google Scholar]
  50. Hu, J.; Liu, Y.; Zhao, J.; Jin, Q. MMGCN: Multimodal fusion via deep graph convolution network for emotion recognition in conversation. arXiv 2021, arXiv:2107.06779. [Google Scholar] [CrossRef]
  51. Zadeh, A.; Liang, P.P.; Mazumder, N.; Poria, S.; Cambria, E.; Morency, L.-P. Memory fusion network for multi-view sequential learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  52. Majumder, N.; Poria, S.; Hazarika, D.; Mihalcea, R.; Gelbukh, A.; Cambria, E. Dialoguernn: An attentive rnn for emotion detection in conversations. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HA, USA, 27 January–1 February 2019; pp. 6818–6825. [Google Scholar]
  53. Hu, D.; Hou, X.; Wei, L.; Jiang, L.; Mo, Y. MM-DFN: Multimodal dynamic fusion network for emotion recognition in conversations. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 7–13 May 2022; pp. 7037–7041. [Google Scholar]
  54. Hazarika, D.; Poria, S.; Mihalcea, R.; Cambria, E.; Zimmermann, R. Icon: Interactive conversational memory network for multimodal emotion detection. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2594–2604. [Google Scholar]
  55. Tsai, Y.-H.H.; Liang, P.P.; Zadeh, A.; Morency, L.-P.; Salakhutdinov, R. Learning factorized multimodal representations. arXiv 2018, arXiv:1806.06176. [Google Scholar] [CrossRef]
  56. Liu, Z.; Shen, Y.; Lakshminarasimhan, V.B.; Liang, P.P.; Zadeh, A.; Morency, L.-P. Efficient low-rank multimodal fusion with modality-specific factors. arXiv 2018, arXiv:1806.00064. [Google Scholar] [CrossRef]
  57. Sun, Z.; Sarma, P.; Sethares, W.; Liang, Y. Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 8992–8999. [Google Scholar]
  58. Yu, W.; Xu, H.; Yuan, Z.; Wu, J. Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; pp. 10790–10797. [Google Scholar]
  59. Han, W.; Chen, H.; Poria, S. Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis. arXiv 2021, arXiv:2109.00412. [Google Scholar] [CrossRef]
  60. Paszke, A. Pytorch: An imperative style, high-performance deep learning library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
  61. Fey, M.; Lenssen, J.E. Fast graph representation learning with PyTorch Geometric. arXiv 2019, arXiv:1903.02428. [Google Scholar] [CrossRef]
  62. Ryumina, E.; Ryumin, D.; Axyonov, A.; Ivanko, D.; Karpov, A. Multi-corpus emotion recognition method based on cross-modal gated attention fusion. Pattern Recognit. Lett. 2025, 190, 192–200. [Google Scholar] [CrossRef]
  63. Becht, E.; McInnes, L.; Healy, J.; Dutertre, C.-A.; Kwok, I.W.; Ng, L.G.; Ginhoux, F.; Newell, E.W. Dimensionality reduction for visualizing single-cell data using UMAP. Nat. Biotechnol. 2019, 37, 38–44. [Google Scholar] [CrossRef]
Figure 1. An example conversation that reveals the influence of the context and the utterance emotion of speaker-1 on the utterance of speaker-2.
Figure 1. An example conversation that reveals the influence of the context and the utterance emotion of speaker-1 on the utterance of speaker-2.
Biomimetics 10 00414 g001
Figure 2. The overview of GASMER.
Figure 2. The overview of GASMER.
Biomimetics 10 00414 g002
Figure 3. Confusion matrix for IEMOCAP(6-way).
Figure 3. Confusion matrix for IEMOCAP(6-way).
Biomimetics 10 00414 g003
Figure 4. Take IEMOCAP(4-way) as an example, UMAP visual comparison, (a) is the feature before the GNN component, (b) is the feature after the graph learning component. It can be seen that the characteristics after the graph learning component form a better emotion cluster, showing the importance of GNN component and graph learning module.
Figure 4. Take IEMOCAP(4-way) as an example, UMAP visual comparison, (a) is the feature before the GNN component, (b) is the feature after the graph learning component. It can be seen that the characteristics after the graph learning component form a better emotion cluster, showing the importance of GNN component and graph learning module.
Biomimetics 10 00414 g004
Table 1. Detailed segmentation information for IEMOCAP and MOSEI.
Table 1. Detailed segmentation information for IEMOCAP and MOSEI.
DatasetTrainValidTestAll
IEMOCAP514666416237433
MOSEI16,3271871466222,860
Table 2. Results on IEMOCAP(6-way). WF1: Weighted F1-score, computed with class support as weights.
Table 2. Results on IEMOCAP(6-way). WF1: Weighted F1-score, computed with class support as weights.
ModelIEMOCAP
HappySadNeutralAngryExcitedFrustratedAvg
F1 ↑F1 ↑F1 ↑F1 ↑F1 ↑F1 ↑ACC ↑WF1 ↑
TFN33.768.655.164.262.461.258.858.5
MMGCN42.378.661.769.074.362.3-66.2
MFN34.170.552.166.862.162.560.159.9
DialogueRNN32.878.059.163.373.659.463.362.8
DialogueGCN42.787.563.564.163.166.965.264.2
ICON32.874.460.668.268.466.264.063.5
COGMEN51.981.768.666.075.358.268.267.6
MM-DFN42.278.966.469.775.566.368.268.1
GASMER64.285.467.966.580.362.770.971.2
Table 3. Results on MOSEI.
Table 3. Results on MOSEI.
MOSEI
TFNCOGMENMFMLMFICCNSelf-MMMMIMGASMER
ACC-7 ↑50.243.951.348.051.6-54.254.3
ACC-2 ↑82.584.384.482.084.285.185.987.1
Table 4. Recall and UAR or IEMOCAP(6-way).
Table 4. Recall and UAR or IEMOCAP(6-way).
RecallUAR
HappySadNeutralAngryExcitedFrustrated
70.83%85.31%66.15%78.24%70.90%58.01%71.57%
Table 5. Ablation experiments on the IEMOCAP dataset, all evaluation indicators use F1-score (%). The results illustrate the effectiveness of the MFA layer and graph learning module.
Table 5. Ablation experiments on the IEMOCAP dataset, all evaluation indicators use F1-score (%). The results illustrate the effectiveness of the MFA layer and graph learning module.
ModalitiesTA + TT + VA + T + V
IEMOCAP
6-way
Actual65.668.464.771.2
w/o MFA63.664.763.264.8
w/o GL63.967.264.968.2
IEMOCAP
4-way
Actual81.483.279.985.2
w/o MFA80.081.381.283.4
w/o GL80.783.380.683.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Li, J.; Dong, J.; Mo, Z.; Liu, N.; Li, Q.; Yuan, Y. Adaptive Graph Learning with Multimodal Fusion for Emotion Recognition in Conversation. Biomimetics 2025, 10, 414. https://doi.org/10.3390/biomimetics10070414

AMA Style

Liu J, Li J, Dong J, Mo Z, Liu N, Li Q, Yuan Y. Adaptive Graph Learning with Multimodal Fusion for Emotion Recognition in Conversation. Biomimetics. 2025; 10(7):414. https://doi.org/10.3390/biomimetics10070414

Chicago/Turabian Style

Liu, Jian, Jian Li, Jiawei Dong, Zifan Mo, Na Liu, Qingdu Li, and Ye Yuan. 2025. "Adaptive Graph Learning with Multimodal Fusion for Emotion Recognition in Conversation" Biomimetics 10, no. 7: 414. https://doi.org/10.3390/biomimetics10070414

APA Style

Liu, J., Li, J., Dong, J., Mo, Z., Liu, N., Li, Q., & Yuan, Y. (2025). Adaptive Graph Learning with Multimodal Fusion for Emotion Recognition in Conversation. Biomimetics, 10(7), 414. https://doi.org/10.3390/biomimetics10070414

Article Metrics

Back to TopTop