Next Article in Journal
Biomechanical Modeling, Muscle Synergy-Based Rehabilitation Assessment, and Real-Time Fatigue Monitoring for Piano-Integrated Upper Limb Therapy
Next Article in Special Issue
Design of Adaptive LQR Control Based on Improved Grey Wolf Optimization for Prosthetic Hand
Previous Article in Journal
Bio-Inspired Approaches—A Leverage for Robotics
Previous Article in Special Issue
Can Anthropomorphic Interfaces Improve the Ergonomics and Safety Performance of Human–Machine Collaboration in Multitasking Scenarios?—An Example of Human–Machine Co-Driving in High-Speed Trains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Comprehensive Review of Multimodal Emotion Recognition: Techniques, Challenges, and Future Directions

Software College, Northeastern University, Shenyang 110169, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Biomimetics 2025, 10(7), 418; https://doi.org/10.3390/biomimetics10070418
Submission received: 24 May 2025 / Revised: 24 June 2025 / Accepted: 25 June 2025 / Published: 27 June 2025
(This article belongs to the Special Issue Intelligent Human–Robot Interaction: 4th Edition)

Abstract

This paper presents a comprehensive review of multimodal emotion recognition (MER), a process that integrates multiple data modalities such as speech, visual, and text to identify human emotions. Grounded in biomimetics, the survey frames MER as a bio-inspired sensing paradigm that emulates the way humans seamlessly fuse multisensory cues to communicate affect, thereby transferring principles from living systems to engineered solutions. By leveraging various modalities, MER systems offer a richer and more robust analysis of emotional states compared to unimodal approaches. The review covers the general structure of MER systems, feature extraction techniques, and multimodal information fusion strategies, highlighting key advancements and milestones. Additionally, it addresses the research challenges and open issues in MER, including lightweight models, cross-corpus generalizability, and the incorporation of additional modalities. The paper concludes by discussing future directions aimed at improving the accuracy, explainability, and practicality of MER systems for real-world applications.

1. Introduction

Multimodal emotion recognition (MER) refers to the process of identifying human emotions using multiple sources of data, such as speech, facial expressions, and physiological signals. The integration of various data sources allows for a more comprehensive and accurate analysis of emotional states compared to unimodal recognition, which relies on only one type of data [1]. MER is crucial in applications ranging from healthcare to human–robot interaction, where accurately understanding human emotions can significantly enhance the quality of services [2].
The importance of MER has grown in recent years, especially in fields such as e-learning, where adapting educational content based on students’ emotions can foster a more engaging learning environment [3]. MER is also essential in healthcare systems, where detecting patient emotions can lead to more personalized care, improving patient outcomes [4]. In human–robot interaction, MER contributes to creating emotionally intelligent systems that can better understand and respond to human emotions, leading to more natural interactions [5].
Recent developments in MER have leveraged advanced deep learning techniques to improve the accuracy and efficiency of emotion recognition systems [6]. The availability of large multimodal datasets has also been a driving force behind advancements in MER, allowing researchers to develop and train more sophisticated models [7]. Moreover, the growing accessibility of wearable devices that can collect physiological signals has expanded the potential applications of MER in everyday settings, making it feasible to detect emotions in real-time [8].
One of the key motivations for focusing on multimodal recognition instead of unimodal approaches is the inherent limitations of using a single data source. Unimodal emotion recognition systems often suffer from reduced accuracy due to missing or ambiguous information, whereas MER systems are more resilient and provide a richer understanding of emotional states by combining multiple data sources [9]. For instance, facial expressions alone may not be sufficient to accurately identify emotions, especially in situations where expressions are subtle or ambiguous. However, combining facial expressions with other modalities, such as speech or physiological signals, can enhance the reliability of emotion recognition [10]. Therefore, MER is considered more robust, especially in real-world scenarios where data from one modality may be unreliable or unavailable.
Biomimetics offers a unifying theoretical lens for these developments. The field seeks to translate principles observed in living organisms—including humans—into engineering solutions. The human nervous system continuously fuses auditory, visual, and interoceptive signals to interpret affective states; reproducing this capability in machines is thus a direct biomimetic endeavor.
The fusion strategies surveyed in this review can inspire bio-inspired sensorimotor loops for next-generation social robots. By coupling acoustic, visual, and physiological channels with adaptive control policies, robots can emulate the embodied affective regulation exhibited by humans—improving safety, trust, and task performance in collaborative settings. This biomimetic perspective recasts MER not merely as a pattern recognition challenge but as a pathway toward living system level adaptiveness.
This review covers research on multimodal emotion recognition spanning the past two decades, focusing on key areas such as system architecture, feature extraction techniques, and fusion strategies. Recent surveys provide valuable but partial snapshots of the field: some emphasize dataset curation and classical fusion pipelines [11], others concentrate on deep-learning advances since 2020 [12] or on conversation-level affect modeling [13]. A few broaden the scope to physiological sensing [14] or present high-level trend statistics without methodological detail [15,16]. The present review differs in three respects. First, it spans 2011–2025 and is selected through a PRISMA-compliant process, guaranteeing transparent coverage. Second, it cross-links modalities, algorithms, and application scenarios within a unified three-dimensional taxonomy, something earlier reviews omit. Third, it couples qualitative synthesis with quantitative comparisons, revealing how dataset characteristics drive fusion choices and performance. These contributions position our work as a holistic complement to the specialized reviews above.

2. Literature Search and Selection Methodology

The literature ranges from foundational studies on emotion recognition datasets to the latest deep learning approaches that integrate multiple modalities. The reviewed content includes advancements in system robustness, generalizability, and practical applications across fields such as healthcare, human–robot interaction, and education, providing a holistic understanding of the current state and future potential of MER systems.
To ensure comprehensive coverage of the field, a systematic literature search and screening process was carried out, following PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. Relevant publications were identified through searches in four major scholarly databases: Google Scholar, ScienceDirect, IEEE Xplore, and arXiv. The search spanned studies published between 2011 and 2025 (inclusive) and was restricted to peer-reviewed articles in the English language. A broad range of keywords related to multimodal emotion recognition (MER) and affective computing were used in various combinations. These included terms such as “multimodal emotion recognition,” “affective computing,” “emotion-aware robotics,” “deep learning,” “physiological signals,” “human–computer interaction (HCI),” “data fusion,” and “Transformer,” among others. Additionally, manual screening of reference lists from key articles was performed to ensure no relevant studies were overlooked.
Inclusion criteria:
  • Published in English and appearing in a peer-reviewed journal or conference proceedings.
  • Focused on multimodal emotion recognition or a closely related domain (e.g., emotion-aware human–robot interaction or affective computing with multiple modalities).
  • Presented original research results (not purely theoretical or editorial content).
Exclusion criteria:
  • Not peer-reviewed (e.g., preprints without formal review, theses, magazine articles).
  • Not available in English or full text could not be obtained.
  • Not focused on MER (the study’s scope was outside multimodal emotion recognition, such as being limited to a single modality).
The database searches and manual reference checks initially yielded a total of 1320 records (1305 from the databases and 15 identified through manual search of references). All retrieved records were imported into a reference manager, and 243 duplicate entries were removed, leaving 1077 unique records. The titles and abstracts of these 1077 records were then screened for relevance. At this stage, an initial 277 records were excluded because they were not research articles (e.g., workshop summaries or non-archival papers) or were clearly irrelevant to MER based on their titles/abstracts. The remaining 800 records underwent a more detailed title and abstract review against the inclusion criteria. This second round of screening led to the exclusion of 420 additional records that did not meet the inclusion criteria upon closer examination of their titles and abstracts (for instance, many of these excluded works dealt with emotion recognition using a single modality or did not involve affective computing in a meaningful way).
After the title and abstract screening, 380 articles were deemed potentially eligible and were selected for full-text retrieval. We successfully obtained the full texts for 365 of these (the remaining 15 could not be acquired in full text, due to reasons such as access limitations or the articles being in a non-English language). Each of the 365 retrieved articles was read in full and assessed for eligibility based on the inclusion criteria. At this full-text eligibility stage, 262 articles were excluded for not meeting the inclusion criteria even after full review of their content. Common reasons for exclusion at this stage included studies that did not truly focus on MER (for example, works that only provided conceptual discussion without any multimodal implementation or evaluation) or other cases where the article clearly fell outside the intended scope of the review despite initial screening.
Following this thorough selection process, a final set of 103 articles was included for qualitative synthesis and review. These 103 studies form the basis of the literature review in this paper. The selection process is summarized in the PRISMA flow diagram (Figure 1), which details the number of records identified, screened, and excluded at each stage, along with the final number of studies included. Bar chart (Figure 2) showing record counts at each PRISMA stage.
Blue bars represent the number of records retained at each stage, while the thin yellow segment on the Identification bar denotes the 15 identified through manual search of references at the first stage.
A rapid bibliometric scan of the 103 studies included in this review uncovers two salient trends. First, publication volume has exploded: almost four-fifths of all papers (82/103 ≈ 80%) appeared in or after 2019, with 2023 emerging as the single most productive year (n = 19), followed by 2024 (n = 14). Second, research output remains geographically concentrated: China (~31%), the United States (~18%), and the EU-27 (~24%) together account for close to three-quarters of the literature, mirroring the distribution highlighted in recent large-scale surveys [17,18].
Modality preferences have evolved in parallel. Whereas early work centered almost exclusively on audio-visual fusion, over 40% of studies published since 2022 adopt either trimodal configurations or transformer style cross-modal fusion architectures [17,19,20]. Likewise, wearable biosensing and eye tracking—virtually absent before 2019—now feature in more than one-tenth of papers from 2023–2025 [21]. These quantitative signals reinforce the narrative that multimodal emotion recognition research is rapidly shifting toward richer sensor portfolios and more sophisticated sequence-to-sequence fusion designs.
Figure 3 illustrates that audio-visual (AV) combinations still dominate the field, while trimodal audio visual text (AVT) and physiological channels form the next two largest groups.
As seen in Figure 4, adoption of AVT and physiological modalities accelerates markedly after 2019, coinciding with the surge of transformer-based cross-modal models.
Figure 5 presents a word cloud of modality keywords, where the size of each term reflects its frequency of use across the 103 reviewed MER studies (2011–2025).
Figure 6 shows a clear geographic pattern: China leads AV research output, whereas EU-based groups contribute the majority of haptic/context studies, echoing earlier bibliometric observations.

3. Overview of MER

3.1. General Structure of MER Systems

The general structure of a multimodal emotion recognition (MER) system consists of three main components: feature extraction, multimodal information fusion, and emotion classification (Figure 7). The feature extraction stage involves extracting relevant features from different modalities, such as speech, visual, and text data [22]. For example, visual features are often derived from facial expressions, while speech features can be extracted from audio signals [23]. Text data, on the other hand, may come from transcripts or verbal cues, providing additional context to emotion recognition [24].
Once the features from different modalities are extracted, they need to be fused to create a cohesive representation. Multimodal information fusion is crucial for integrating data from various sources in order to capture the complex nature of emotions [13]. Different fusion techniques can be employed, such as early fusion, which combines features at an initial stage, or late fusion, which merges the outputs of modality-specific models [25]. Early fusion allows the system to capture interactions between modalities more effectively, while late fusion tends to be more flexible in accommodating different modalities [26].
The final step Is emotion classification, where the fused features are used to classify the emotional state of the subject. Emotion classification typically relies on machine learning or deep learning models that are trained to identify specific emotional states, such as happiness, sadness, or anger [27]. Deep learning methods, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), are often used for this purpose due to their ability to learn complex patterns in the data [28]. Additionally, classifier design must consider the dynamic and temporal aspects of emotions, which can be addressed using temporal modeling techniques such as long short-term memory (LSTM) networks [29].
Overall, the general structure of an MER system aims to leverage the strengths of multiple modalities to improve emotion recognition accuracy and robustness. The workflow includes extracting features from each modality, fusing the information, and using an appropriate classifier to generate the final emotion label [30]. This comprehensive approach makes MER systems more effective compared to unimodal methods, particularly in scenarios involving complex human emotions [31].

3.2. Emotion Recognition Milestones

Substantial progress has been made in the last decade in multimodal emotion recognition, especially with the use of deep learning for more accurate and efficient recognition (see Table 1 for a summary of milestones). A major breakthrough was the application of convolutional neural networks (CNNs) for feature extraction, which significantly enhanced the analysis of visual data such as facial expressions [12]. Likewise, recurrent neural networks (RNNs), including long short-term memory (LSTM) networks, have played a significant role in modeling temporal dependencies in speech and visual data, thereby improving emotion recognition accuracy.
Attention mechanisms form another important milestone in MER, improving the interpretability and performance of emotion recognition systems. These mechanisms allow models to focus on the most relevant parts of the input data, which is particularly beneficial in RNN-based architectures. With the availability of large-scale multimodal datasets, the MER field has progressed rapidly, as these datasets enable the training of complex models that capture cross-modal dynamics. Additionally, the advent of multimodal transformers has provided an effective way to combine information from multiple modalities, achieving high recognition accuracy by jointly learning representations from each modality [32].
Lastly, the development of real-time MER systems has enabled the use of MER in interactive applications, including human–robot interaction and virtual reality environments, to create more immersive and emotionally aware user experiences [33]. Collectively, these milestones underscore how rapidly MER techniques have advanced and how their potential is growing across sectors such as healthcare, education, and entertainment.
Table 1. Emotion recognition milestones.
Table 1. Emotion recognition milestones.
MilestoneDescriptionReferences
CNNs for feature extractionImproved ability to analyze visual data such as facial expressions[12,34]
RNNs for temporal dependenciesEnhanced accuracy by capturing temporal dependencies in data[33,35]
Attention mechanismsImproved interpretability and focus on relevant parts of the input data[36,37]
Availability of large-scale datasetsAccelerated progress with better resources for training complex models[9,38]
Multimodal transformersSignificant improvement in fusion and recognition accuracy[32,39]
Real-time MER systemsEnabled implementation in interactive applications for immersive experiences[33,40]

3.3. Multimodal Emotion Datasets

Multimodal emotion recognition research relies heavily on publicly available datasets, which provide the data needed to train and evaluate emotion recognition models (see Table 2 for examples). The widely used GEMEP corpus provides audio, visual, and physiological information for a range of expressions of emotion [41]. GEMEP is valued for showing emotions clearly and systematically, which is why it is well suited for gauging the performance of MER technologies.
The K-EmoCon dataset is also a popular tool, built for ongoing identification of emotions using different tools, including speech, facial expressions, and physical sensors [42]. Working with this dataset made it possible to design real-time MER systems that focus on the changing emotions during normal conversations.
Studies in multimodal emotion recognition often depend on the IEMOCAP database, which contains recordings of actors carrying out both prepared and improvised dialogues [43]. Individuals who use the IEMOCAP dataset can achieve both discrete and continuous emotion recognition since there are detailed individual emotion annotations.
The AffectGPT dataset has made available emotional data that uses text, voice recordings, and visual media [44]. Using this dataset, large models have been trained to identify emotions and to explain how they reached those conclusions.
The Multimodal EmotionLines Dataset (MELD) is another valuable resource, comprising multimodal interactions from TV show dialogues (audio, visual, and text) [45]. MELD is particularly useful for conversational emotion recognition, as it captures emotions within the context of natural dialogue, along with contextual information from multiple turns.
Furthermore, the Audio/Visual Emotion Challenge (AVEC) datasets include multimodal recordings such as facial expressions, speech, and physiological signals (e.g., ECG, GSR, blood volume pulse) for emotion analysis [46]. The AVEC datasets have become standard benchmarks in MER, driving the application of deep learning techniques to improve performance on these challenging multimodal tasks.
Another widely used dataset is the DEAP dataset, which consists of participants’ physiological signals (e.g., EEG, ECG) and facial expressions recorded while they watched music videos, along with self-reported emotional responses [47]. DEAP is notable for highlighting the role of physiological signals in emotion recognition and is often used to explore the fusion of biosignals with audio-visual data.
Table 2. Multimodal emotion datasets.
Table 2. Multimodal emotion datasets.
MethodDatasetModalityEmotion LabelSamples
CNN, RNNGEMEP [41]Audio, visual, physiologicalDiscrete emotions7200
TransformerK-EmoCon [42]Speech, visual, physiologicalContinuous emotions3200
CNN, LSTMIEMOCAP [43]Audio, visualDiscrete and continuous12,000
Large-scale modelAffectGPT [44]Text, audio, visualDiscrete emotions5000
Multi-view attentionMELD [45]Audio, visual, textDiscrete emotions1400
Deep learningAVEC 2016 [46]Audio, visual, physiologicalContinuous emotions4000
Physiological signalsDEAP [47]Physiological, visualArousal and valence1280

3.4. Feature Extraction Techniques

Feature extraction is a critical step in MER that involves deriving informative features from different types of data. Each modality provides unique information about emotional states, and a range of techniques has been developed to effectively compute these features.
The techniques enumerated below are restricted to those that have been adopted and validated in multimodal emotion recognition studies; they are not intended as an exhaustive catalogue of every signal processing method available in speech, vision, or NLP at large. For example, although wavelet packet decompositions, Zernike moments, or syntactic dependency kernels are widely used in unimodal research, they have seen little or no uptake in MER benchmarks and therefore fall outside the scope of this survey [12,13]. By delimiting the list in this way, we focus the discussion on approaches that have demonstrably influenced current MER pipelines.
Speech Feature Extraction: variations in tone, pitch, intensity, and rhythm of speech carry clues about underlying emotional states (see Table 3). Mel-Frequency Cepstral Coefficients (MFCCs) are commonly used speech features that capture the spectral properties of speech indicative of emotional tone [23]. Other widely used speech features include prosodic features such as pitch, energy, and speaking rate, which reflect the intensity and cadence associated with different emotions [48]. Techniques such as Linear Predictive Coding (LPC) model the speech signal to extract parameters related to the speaker’s emotional state [49]. In addition, deep learning approaches (e.g., CNN-based feature extractors) have been applied to raw audio signals to automatically learn emotion-specific features [11]. Time-frequency representations such as spectrograms—visualizing sound signals across time and frequency—are also used to capture temporal and frequency characteristics of speech that are relevant to emotion recognition [29].
Table 3. Summary of audio feature extraction methods.
Table 3. Summary of audio feature extraction methods.
Type of FeaturesFeature Extraction MethodsPublications
SpectralMFCC[23,39]
ProsodicPitch, energy, speaking rate[13,30,48]
Model-basedLPC[44]
Frequency representationSpectrogram analysis[29]
Learned deep featuresCNN/RNN automatic feature learning from raw audio[11,49]
Visual Feature Extraction: visual cues are crucial for understanding emotional expressions, as facial movements and gestures often directly reflect a person’s emotional state (Table 4). Techniques such as facial landmark detection are commonly used to identify key points on the face (e.g., corners of the mouth, eyes, and eyebrows), which are then used to quantify facial expressions [24]. Histogram of Oriented Gradients (HOGs) is another widely used technique that captures the distribution of intensity gradients and edges in an image, providing a robust representation of facial features [50]. Deep learning approaches, especially CNNs, have been extensively used to automatically extract features from facial images, enabling models to learn complex patterns associated with different emotions [51]. Additionally, optical flow can be employed to capture motion patterns in facial expressions, enhancing the temporal understanding of emotional changes over time [52]. Beyond facial features, body posture and gestures can also be informative visual features, offering a more holistic view of a person’s emotional state [53].
Table 4. Summary of visual feature extraction methods.
Table 4. Summary of visual feature extraction methods.
Input TypeType of FeaturesFeature ExtractionPublications
Facial imagesLandmark pointsFacial landmark detection[24,54]
Facial imagesGradient featuresHOG[40,50,55]
Facial imagesDeep featuresCNN[41,42,51]
Motion patternsTemporal dynamicsOptical flow[38,43,52]
Full body/skeleton imagesPose keypointsBody-pose estimation (OpenPose, HRNet, etc.)[53,56]
Gesture sequencesGesture trajectoriesSkeleton-based gesture recognition (ST-GCN, 2s-AGCN, etc.)[57,58]
Eye region/gaze dataGaze direction and eye movementVideo/IR eye-tracking, pupil detection[18,59]
Text Feature Extraction: textual features are valuable for capturing the emotional content conveyed through spoken or written language (Table 5). Traditional text features include Bag-of-Words (BoWs) and Term Frequency–Inverse Document Frequency (TF-IDF) representations, which encode text as numerical vectors and allow further analysis [49]. More advanced methods involve word embeddings (e.g., Word2Vec, GloVe) that capture semantic relationships between words, thereby detecting subtle emotional cues in text [60]. Sequence models such as RNNs, particularly LSTMs, have been used to capture contextual dependencies in text, improving emotion detection from dialogue or written content [61]. Transformer-based models such as BERT have also been applied to extract contextual and sentiment features from text, achieving state-of-the-art performance in emotion recognition tasks [49]. Additionally, using sentiment lexicons—predefined dictionaries of words rated by emotional sentiment—can enhance text-based emotion recognition by incorporating external knowledge of word affect [62].
Table 5. Summary of text feature extraction methods.
Table 5. Summary of text feature extraction methods.
Type of FeaturesFeature ExtractionPublications
Lexical featuresBag-of-words, TF-IDF[39]
Semantic featuresWord embeddings (Word2Vec, GloVe)[44]
Contextual featuresRNN, LSTM[28,61,63]
Contextual sentimentTransformer (BERT), sentiment lexicons[45,46,49,62]
Self-supervised contextualFine-tuning large language models (RoBERTa, XLNet, GPT)[64,65]
Prompt-based adaptationPrompt learning/in-context tuning[53,66]
In MER, combining features from different modalities is essential to capture the full spectrum of emotional expression. Feature level fusion involves concatenating features from different modalities before classification, and it has been shown to improve recognition accuracy in many cases [40]. Additionally, deep multimodal learning approaches learn joint representations of features from speech, visual, and text data, enabling models to capture cross-modal interactions effectively [67]. Techniques such as attention mechanisms have also been applied at the feature level to weigh the importance of features from each modality, leading to more accurate emotion recognition by focusing on the most relevant signals [5].

3.5. Additional Modalities and Their Sensing Principles

Beyond the core channels of speech, vision, text, and EEG already summarized in Section 3.4, four further modality groups have become increasingly relevant to multimodal emotion recognition (MER) research. Each group is outlined below with its sensing principle, merit and key limitation.
Gaze/Eye-Tracking: infrared or RGB cameras estimate gaze direction, fixation length, blink rate, and pupil diameter. These metrics index attention allocation and sympathetic arousal; when fused with facial or EEG cues, they raise recognition accuracy, particularly for subtle or masked emotions [18]. Advantage: non-contact, millisecond-level reactions. Limitation: needs unobstructed eyes and recalibration; gaze patterns vary widely across tasks and cultures.
Environmental Context: scene cameras and ambient microphones quantify light, crowd density, noise, or music genre; metadata (time, location) adds further situational cues. Context features help disambiguate similar expressions [68]—for example, a smile in a funeral scene versus at a party—improving robustness under “in-the-wild” conditions [69]. Advantage: supplies prior humans naturally used. Limitation: context sensing increases privacy risk, and errors propagate if the scene is misclassified.
Haptic/Tactile and Peripheral Physiology: Wearable or seat-embedded pressure sensors capture force, tremor, and posture changes; galvanic skin response, heart rate variability, or skin temperature reflect autonomic arousal. Such signals are difficult to fake and have proven complementary to audio-visual streams in edge-deployed MER systems [70]. Advantage: objective, user-independent arousal markers. Limitation: mostly encode intensity, not valence; require on-body hardware and careful artifact removal (motion, temperature).
Emerging Sensors:
Wearable multi-biosensors: smartwatches combine photoplethysmography, accelerometers, and temperature for continuous affect logging [21].
fNIRS headbands: near infrared light tracks frontal cortex blood oxygenation linked to affective appraisal [71].
Doppler and mm-wave radar: contactless measurement of micro-vibrations from respiration and heartbeat enables covert emotion screening, useful where cameras are unsuitable [72].
Advantages: operate in everyday settings, add neural or physiological depth, and can work when faces are hidden. Limitations: many are still experimental, sensitive to motion or multipath interference, and dependent on user acceptance of wearables.

3.6. An Integrative Taxonomy of the MER Landscape

To help readers navigate the rapidly expanding literature, we propose a three-dimensional taxonomy that organizes multimodal emotion recognition research by (i) sensed modality, (ii) computational method, and (iii) application context. Along the modality axis, work ranges from “conventional” audio-visual inputs to physiological channels (EEG, ECG, EDA) and recently to gaze, haptic cues, ambient context, and emerging contactless sensors such as millimeter-wave radar [12,21]. The methodological axis reflects a progression from early concatenation or statistical fusion, through late decision aggregation, to intermediate representations that exploit attention, graph contrastive learning, or cross-modal transformers. Channel-attention networks dominate physiological fusion—demonstrated by SCA-Net++ on SEED-IV [73]—whereas cross-modal transformers achieve the current in the wild state-of-the-art on ABAW and MELD benchmarks [16,74]. A growing branch of lightweight CNN-LSTM hybrids targets real-time edge devices, trading some accuracy for latency and energy gains [12].
The application axis spans offline corpus analysis, real-time human–computer interaction, emotion-aware robotics, and pervasive health monitoring. Dialogue centric systems (e.g., JOYFUL on MELD) emphasize trimodal text audio vision pipelines for social AI [74,75], while social-robot platforms such as Pepper integrate vision, speech, and context to adapt robot behavior in public spaces [15]. Wearable biosensor ensembles have opened continuous affect tracking for stress management and telehealth scenarios, outperforming camera-based solutions in privacy-sensitive contexts [70]. Mapping the literature onto this three-way grid reveals clear trends: the richest modality combinations increasingly couple with transformer or graph-based alignment, and these, in turn, enable deployment in unconstrained, high-impact settings such as open world video diaries [76] or human–chatbot field studies [77]. The taxonomy therefore not only summarizes the field but also highlights the converging trajectory toward deeply fused, context-aware, and application-specific MER systems.

4. Multimodal Information Fusion for MER

4.1. Bimodal Emotion Recognition

Bimodal emotion recognition involves combining two different modalities—such as speech and visual data—to improve the accuracy and robustness of emotion recognition systems (Table 6). By integrating features from two sources, bimodal approaches capture complementary information that enhances the system’s ability to recognize emotions more effectively than unimodal systems [25].
A common bimodal pairing is speech-visual emotion recognition, which combines facial expressions with speech signals to infer a person’s emotional state [26]. Facial expressions provide direct visual cues about emotions, while speech features (tone, pitch, prosody) convey emotional context through voice. Using both modalities together yields a richer representation of emotion, especially in cases where either modality alone may be ambiguous or insufficient [78].
Fusion strategies for bimodal MER generally fall into three categories: early fusion, late fusion, and hybrid fusion. In early fusion, features from both modalities are concatenated and fed into a classifier together, capturing feature level interactions between modalities. This approach can create a more holistic representation of emotion, but it may face challenges due to different feature dimensions and the need for precise synchronization of data [53]. In late fusion, separate classifiers are trained on each modality, and their outputs (e.g., predicted emotion probabilities) are combined via averaging or a meta-classifier to make the final decision [79]. For instance, Wear-BioNet fuses the soft probabilities of heart rate, electrodermal activity, and accelerometer CNN–GRU branches by simple arithmetic mean and attains 84.5% accuracy on WESAD [70], whereas EAR-RoBERTa first produces text specific and metadata-specific probability vectors and then feeds them to a logistic regression meta-classifier that pushes CMU-MOSEI accuracy to 81.9% [65]. Late fusion is flexible, allowing each modality to be processed independently, but it might miss nuanced cross-modal interactions. Hybrid fusion attempts to leverage the advantages of both early and late fusion by combining information at multiple stages; for example, using feature fusion followed by an additional fusion of intermediate or output representations. When properly implemented, hybrid fusion can exploit the complementary characteristics of each modality’s information [36].
Deep learning techniques have significantly advanced bimodal emotion recognition. CNNs are commonly used for extracting visual features, while RNNs or LSTMs capture temporal dependencies in speech data [38]. Attention mechanisms have also been introduced to assign different weights to each modality, enabling the system to emphasize the most informative features for emotion recognition [74]. This has improved the interpretability of bimodal systems by highlighting which features (facial cues or vocal intonations) contribute most to the emotion prediction. For example, an attention-based model might focus on facial features when visual emotion cues are strong and rely more on vocal tone when the face is less expressive.
Another strategy in bimodal MER is the use of transfer learning between modalities. For instance, a model pre-trained on a large facial expression dataset might be fine-tuned on an audio emotion dataset (or vice versa), effectively transferring knowledge from one modality to improve performance on the other [72]. One study pre-trained a speech encoder on 650 h of VoxCeleb and, after freezing the lower layers, fine-tuned it jointly with a visual stream, improving F1 on CREMA-D from 73.2% to 77.5% [36]. A second work (MemoCMT) initializes its video backbone from ImageNet and then adapts it together with audio and text specific transformers, reaching 82.3% accuracy on CMU-MOSEI [58]. Such transfer learning approaches leverage existing learned representations and can be particularly useful when one modality has limited labeled data.
Cross-modal learning is another important concept, where features of one modality are used to inform feature extraction in another modality. For example, visual features could guide the extraction of relevant speech features, resulting in more accurate emotion recognition [80]. By bridging the gap between modalities, cross-modal learning improves the system’s robustness, ensuring that information from one modality can compensate for weaknesses in the other.
Bimodal fusion techniques also include methods such as canonical correlation analysis (CCA), which finds linear correlations between feature sets of different modalities to create a shared representation space for both [53]. CCA-based methods increase recognition capability by focusing on information common to both modalities, making it easier to detect emotions that manifest in correlated audio-visual patterns.
Table 6. Summary of bimodal emotion recognition methods.
Table 6. Summary of bimodal emotion recognition methods.
Input Modality 1Input Modality 2Output EmotionFusion StrategyMethods UsedReferences
SpeechVisualDiscrete emotionsEarly fusionCNN for visual, LSTM for speech[12,25,34]
SpeechVisualContinuous emotionsLate fusionRNN for speech, SVM for visual[28,52,65,70,79]
VisualPhysiologicalDiscrete emotionsHybrid fusionHOG for visual, CNN for signals[42,53]
SpeechEEGArousal and valenceEarly fusionSpectrogram analysis, CNN[48,74]
TextVisualPositive/negativeAttention mechanismTransformer, attention network[10,36,58]
SpeechVisualDiscrete emotionsTransfer learningPre-trained ResNet, RNN[36,38,72,81]
AudioVisualEmotional stateCross-modal LearningCCA, CNN[52,80]
VisualPhysiologicalStress detectionHybrid fusionLSTM for temporal, SVM[40,82]
SpeechVisualEmpathy detectionLate fusionCNN, random forest classifier[26,28]
TextFacial expressionsSentiment analysisEarly fusionTF-IDF for text, HOG for visual[48,67]

4.2. Trimodal Emotion Recognition

Trimodal emotion recognition involves the integration of three different modalities—such as speech, visual, and text data—to provide a more holistic view of human emotions (Table 7). By combining three modalities, trimodal systems capture diverse aspects of emotional expressions from multiple sources. This approach can compensate for the limitations of individual modalities, leading to improved accuracy and robustness in emotion recognition systems.
One of the most common trimodal combinations is speech, visual, and text. Using these methods together works well for discovering emotions in programs where people interact, because they include voice quality, facial expressions, and spoken content. Facial expressions and delineated parts of the face give us direct advice on emotions, pitch and energy in speech offer sound style, and text provides context through meaning [61]. Using them together helps us discover and explore emotions, especially when we deal with multiple peers [79]. This rationale directly aligns with the public corpora reviewed in Section 3.3. GEMEP couples high-resolution facial action units with synchronized speech, making it ideal for probability averaging of visual and acoustic streams [41]. K-EmoCon augments those channels with wearable biosignals, enabling late fusion meta-classifiers that exploit arousal information unavailable in the face or voice alone [42]. For dialogue-centric tasks, IEMOCAP and MELD add textual transcripts so that lexical context can refine ambiguous prosodic cues [43,45].
Typical approaches to trimodal fusion are feature level fusion, decision-level fusion, and hybrid fusion. Features from each of the three modalities are joined together to represent a single feature vector for classifying emotions. When using this approach, the model is able to learn about all the data sources at once; however, because the data is very large, it must be correctly aligned and normalized [83]. Alternatively, decision-level fusion teaches each modality to make its own predictions and merges those results together, for example, using voting or weighted calculations. With this strategy, various data types are easier to work with, but it may not catch all the mixed-modality relationships as well as the other approach does [84]. Hybrid fusion uses both types of fusion and combines information in several places as the data flows through the system. Example approaches often start by joining the information from two forms of data and end by merging that intermediate result with data from the third source in the decision stage [67]. This method is designed to connect the unique advantages found in each fusion process [39].
Deep learning has been widely utilized in trimodal MER. For example, CNNs can extract visual features, while RNNs or transformer models capture temporal and contextual information from speech and text, respectively [80]. Attention mechanisms in trimodal systems can assign different importance weights to each modality, enabling the model to concentrate on the most relevant features for emotion recognition [74]. This helps in effectively handling cases where one modality might be noisy or less informative. For instance, if background noise affects the audio quality, the model can rely more on visual and textual cues [85].
Another technique in trimodal MER is modality alignment, where features from different modalities are projected into a common feature space to simplify fusion. Techniques such as dimensionality reduction (e.g., canonical correlation analysis, CCA) can be applied to aligned features, reducing complexity and ensuring that shared information across modalities is emphasized [53]. Aligning the features can enhance the relationships captured among modalities, leading to better emotion recognition performance [43].
Cross-modal transformers have recently gained popularity in trimodal MER, as they are capable of modeling complex relationships among multiple modalities. These models use self-attention mechanisms to learn interactions between speech, visual, and text features, thereby capturing the intricate dynamics of human emotions [79]. Cross-modal transformers have shown promising state-of-the-art results on emotion recognition benchmarks, demonstrating their ability to handle the rich and dynamic nature of multimodal emotional data [32].
Table 8 compares recent state-of-the-art MER systems across datasets, fusion strategies and metrics, illustrating that transformer-based cross-modal models consistently outperform conventional early fusion on large-scale benchmarks.
Table 9 summarizes, across major modality classes, the state-of-the-art aggregation methods and their best reported performance, highlighting that channel-attention mechanisms dominate physiological fusion, whereas graph contrastive or transformer architectures excel in trimodal dialogue settings.
Table 7. Summary of trimodal emotion recognition techniques.
Table 7. Summary of trimodal emotion recognition techniques.
Modality 1Modality 2Modality 3Fusion StrategyMethods UsedReferences
SpeechVisualTextDecision-levelSVM/random forest + RNN[51,61]
SpeechVisualTextEarly-level (feature concat.)CNN (faces) + Log-Mel spectrogram[83]
SpeechVisualTextCross-modal transformerCCA pre-alignment + transformer[84,86]
AudioVisualTextAttention mechanismCNN + cross-modal attention[85,87]
TextSpeechVisualCross-modal transformerSelf-attention encoder[74,84]
VisualPhysiologicalTextHybrid (feature + decision)HOG + LSTM + sentiment lexicon[53,79]
SpeechVisualPhysiologicalHybrid (ResNet → LSTM)Pre-trained ResNet (face) + LSTM (biosignal)[61,88]
SpeechEEGVisualModality alignmentCCA + CNN[80,89]
VisualTextEEGFeature-levelCNN (image) + BERT (text) + spectrogram[60,79]
TextFacial expressionsSpeechDecision-levelTF-IDF + CNN + GRU[52,90]
SpeechVisualTextEnsemble hybrid (graph contrastive + transformer)JOYFUL/graph contrastive alignment[39]
SpeechVisualTextCross-modal transformer (state-of-the-art)Self-attention fusion[32,43]
Table 8. Comparative table of performance metrics with aggregation methods.
Table 8. Comparative table of performance metrics with aggregation methods.
StudyModalitiesPublic DatasetAggregation/Fusion MethodMetric (Type)Reported Score
TMNet [20] (2025)Speech + EEGSEED-IVCross-modal transformer (early + attention)Acc.88.70%
MemoCMT [58] (2025)Vision + speech + textCMU-MOSEICross-modal transformerAcc.82.30%
JOYFUL [74] (2023)Audio + text + visionMELDGraph contrastive mid-level fusionF1-macro81.20%
Edge-MER [12] (2024)Facial + audioRAVDESSLightweight CNN-LSTM (early)WA79.40%
EAR-RoBERTa [65] (2023)Text (+ meta)CMU-MOSEIEmotion-specific attention (late)Acc.81.90%
Joint-MMT [16] (2024)Vision + speech + actionABAW 2023Unified transformer (late)F1-macro48.90%
Interp-Hybrid [72] (2021)Vision + speech + textIEMOCAPHybrid early-/late + heat-map attentionAcc.82.00%
FG-Disentangle [91] (2022)Audio + text + visionMELDDisentangled representation (early)F1-macro80.10%
Wear-BioNet [70] (2025)Wearable HR + EDA + accel.WESADEnsemble CNN-GRU (late)Acc.84.50%
Table 9. Modality–method–performance Matrix.
Table 9. Modality–method–performance Matrix.
Modality (Primary Channels) Typical Evaluation Setting Method Class with Best-Reported
Result
Representative Study [Dataset] Key Metric Best Score
Audio + visualIn-the-wild video (ABAW 24)Cross-modal 3-D transformer fusionCross-modal 3D Facial-Speech [72]F1-macro0.83
Audio + visual + textMulti-speaker dialogue (MELD)Graph contrastive mid-level fusionJOYFUL (MELD) [74]F1-macro0.812
Physiological (EEG/facial)Lab elicitation (SEED-IV)Channel-attention early fusionSCA-Net++ [92]Accuracy0.912
Wearable biosensors (HR, EDA, Accel.)Real-life stress (WESAD)Ensemble CNN-GRU, late fusionWear-BioNet [70]Accuracy0.845
Gaze + audio + textHuman‚ Äìchatbot field studyHybrid attention, early + lateMultimodal Cues Chatbot [77]F10.74
Edge-device AVResource-limited real-timeLightweight CNN-LSTM, earlyEdge-MER (RAVDESS) [12]WA0.794

5. Research Challenges and Open Issues

5.1. Lightweight and Explainable Deep Models for MER

The development of lightweight and explainable deep models is crucial for making MER systems more accessible and interpretable. Model interpretability is especially important in sensitive fields such as healthcare and education, where understanding how the model arrives at decisions can foster trust and improve user acceptance [5]. One major challenge in MER is the high computational cost of deep learning models, which often limits their deployment on edge devices with limited resources. Lightweight models (e.g., using pruning, quantization, or knowledge distillation techniques) have been explored to reduce computational burden without significantly compromising accuracy [93].
In parallel, explainable artificial intelligence (XAI) methods (such as attention visualization, saliency maps, or feature attribution techniques) have been employed to make MER models more transparent, allowing users to understand which features or modalities are most influential in the emotion recognition process [44]. Combining lightweight architectures with explainability is a promising direction to make MER systems user-friendly and viable for real-world scenarios [94].

5.2. Multimodal Information Fusion Strategies

While multimodal information fusion is key to MER, it presents challenges such as differing temporal resolutions, noise levels, and missing data across modalities. Early fusion (feature concatenation) can struggle with the high dimensionality of combined features, making model training difficult [95]. Late fusion (combining classifier outputs) might lose important cross-modal interactions that are essential for recognizing complex emotional cues [33]. To address these challenges, hybrid fusion techniques have emerged, aiming to leverage the strengths of both early and late fusion by combining information at multiple stages. Hybrid fusion enables models to capture both intra- and inter-modal relationships effectively [46]. To capture deeper inter-dependences, Praveen and Alam devised a Recursive Joint Cross-Modal Attention mechanism that iteratively refines intra- and inter-modal correlations and sets new CCC records on dimensional emotion tasks [96].
Recent advances in deep learning, such as attention-based fusion methods, have shown significant promise in multimodal fusion by assigning different weights to each modality based on its relevance to the emotion being recognized [97]. Additionally, transformer architectures have been used to model complex dependencies among modalities, allowing for more flexible and powerful fusion mechanisms [74]. These approaches help overcome limitations of traditional fusion by dynamically focusing on the most informative signals. Another approach involves using cross-modal transformers, which can capture both individual modality features and their interactions, providing a comprehensive representation of emotions [69].

5.3. Cross-Corpus MER

A major challenge in MER is the lack of generalizability across different datasets, often referred to as the cross-corpus problem. Models trained on one dataset often perform poorly on others due to variations in recording conditions, cultural differences, and emotional expression styles [98]. This highlights the need for MER systems that generalize well across different contexts and populations. Transfer learning and domain adaptation techniques have been explored to tackle this problem by enabling models to leverage knowledge from one dataset to improve performance on another [40].
Adversarial training targets dealing with the gap between the features of the source and the target dataset in order to improve cross-corpus generalizability. The model is less affected by biases in a dataset when it can identify domain-invariant features [99]. Moreover, using synthetic data or style transfer boosts the variety of the training set, which helps models withstand differences in real-world data [67]. The AVaTER framework fuses audio, visual and textual cues through a compact cross-modal attention block, achieving real-time inference on edge GPUs while improving accuracy across four public corpora [100,101].

5.4. More Modalities for MER

Incorporating additional modalities (e.g., physiological signals or contextual data) can enhance the accuracy and robustness of MER systems by providing richer information about the user’s emotional state. Signals from the heart, skin, and brain all show how the autonomic nervous system is involved in controlling emotions [100]. Supplementing signals from wearables with speech and video helps MER systems better understand how people feel.
Environmental factors and what is happening around us can provide valuable information too. Identifying emotions in a loud and jam-packed environment might need other procedures than it does in a calm, quiet location [102]. When you supply the environment, time, and presence of others, the machine can more accurately detect your moods. Thanks to recent progress in sensors, MER systems can now monitor and react to a person’s state and the environment in real-time [32].

5.5. Few-Shot Learning for MER

A significant challenge in MER is the scarcity of labeled data, which limits the training of deep learning models. Few-shot learning (FSL) is an emerging approach that addresses this by enabling models to learn from only a handful of labeled examples. FSL methods (such as metric learning or prototypical networks) have been applied to MER to recognize emotions with minimal data [19]. By learning a feature space where examples of the same emotion are clustered together, these models can generalize to new emotion classes with only a few examples [47].
Another promising approach is the use of generative models (such as variational autoencoders or generative adversarial networks) to create synthetic training data, augmenting the limited real data [50]. This can mitigate data scarcity by providing additional samples for training, thereby improving model robustness. Additionally, transfer learning from related domains (e.g., using models pre-trained on sentiment analysis or facial recognition tasks) can provide a strong initialization for MER models, which can then be fine-tuned on small emotion datasets [83].
Self-supervised learning is another avenue, where models learn to extract meaningful features from unlabeled data through proxy tasks (such as reconstructing missing parts of input or predicting temporal order). Such methods have shown promise in MER by leveraging large amounts of unlabeled multimodal data to pre-train representations that are later fine-tuned on the small, labeled emotion datasets [103]. Collectively, these approaches make MER systems more feasible in scenarios where labeled data is scarce, expanding their applicability to new domains and populations.

6. Summary

This review has highlighted key advancements and challenges in multimodal emotion recognition (MER). Integrating multiple modalities—such as speech, visual, and text—has proven to be an effective strategy for capturing the complexity of human emotions and improving recognition accuracy. Bimodal and trimodal approaches demonstrate significant improvements over unimodal systems by leveraging complementary information from different data sources, thus providing a more comprehensive understanding of emotions [8]. The use of deep learning techniques (CNNs, RNNs, and transformers) has further enhanced MER performance by enabling the extraction of complex features and modeling interdependencies among modalities [61].
Multimodal information fusion remains a critical aspect of MER, with various fusion strategies (early, late, and hybrid) employed to combine information from different modalities. Hybrid fusion has shown promise in capturing both intra-and inter-modal relationships, thereby improving the robustness of MER systems [40]. Recent advances, such as attention-based fusion and cross-modal transformers, have further enhanced multimodal fusion effectiveness by allowing models to focus on the most relevant features and capture complex dependencies among modalities [32].
Despite the progress, several challenges remain in MER. Developing lightweight and explainable deep models is essential for making these systems more accessible and trustworthy, especially in sensitive applications such as healthcare and education [19]. The high computational cost of deep models poses a barrier to deployment on edge devices, necessitating efficient, lightweight architectures [86]. Additionally, the lack of interpretability in many deep learning models hinders their acceptance, highlighting the need for XAI techniques to provide insights into the decision-making process [93].
Cross-corpus generalizability is another major challenge: models often struggle on datasets other than the ones they were trained on due to differences in conditions and expression styles [94]. Techniques such as transfer learning, domain adaptation, and adversarial training have been proposed to improve MER robustness across contexts [95]. Moreover, incorporating additional modalities such as physiological and contextual data can enrich MER systems by providing deeper insight into users’ emotional states, making the systems more adaptable to different scenarios [32].
Future research in MER should focus on addressing these challenges by developing more efficient, explainable, and generalizable models. The integration of non-traditional modalities (e.g., physiological signals, context) holds great potential for enhancing accuracy and robustness. Furthermore, few-shot and self-supervised learning techniques offer promising solutions for the data scarcity problem, potentially broadening the applicability of MER to new domains and populations [34]. Practical implementations of MER should also consider user-centric design, ensuring that models are not only accurate but also interpretable and accessible to end users. By meeting these challenges and leveraging recent advancements in deep learning and sensor technology, MER has the potential to play a significant role in enhancing human–computer interaction, healthcare, education, and other fields where understanding human emotions is crucial.

Author Contributions

Conceptualization, Y.W. and Q.M.; methodology, Y.W. and Q.M.; investigation, Y.W., Q.M., and T.G.; data curation, Y.W.; writing—original draft preparation, Y.W. and Q.M.; writing—review and editing, Y.W., Q.M., and T.G.; visualization, Y.W. and Q.M.; supervision, Q.M. and T.G.; project administration, Y.W. and Q.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China under Grant Number: 52130403.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable.

Acknowledgments

The authors are grateful to their colleagues for providing valuable insights and support during the preparation of this work. The authors also confirm that no AI-assisted writing tools were used in the development of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Abdullah, S.M.S.A.; Ameen, S.Y.A.; Sadeeq, M.A.; Zeebaree, S. Multimodal emotion recognition using deep learning. J. Appl. Sci. Technol. Trends 2021, 2, 73–79. [Google Scholar] [CrossRef]
  2. Adel, O.; Fathalla, K.M.; Abo ElFarag, A. MM-EMOR: Multi-modal emotion recognition of social media using concatenated deep learning networks. Big Data Cogn. Comput. 2023, 7, 164. [Google Scholar] [CrossRef]
  3. Bahreini, K.; Nadolski, R.; Westera, W. Towards multimodal emotion recognition in e-learning environments. Interact. Learn. Environ. 2016, 24, 590–605. [Google Scholar] [CrossRef]
  4. Ghaleb, E.; Popa, M.; Asteriadis, S. Multimodal and temporal perception of audio-visual cues for emotion recognition. In Proceedings of the 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII), Cambridge, UK, 3–6 September 2019; pp. 552–558. [Google Scholar]
  5. He, Z.; Li, Z.; Yang, F.; Wang, L.; Li, J.; Zhou, C.; Pan, J. Advances in multimodal emotion recognition based on brain–computer interfaces. Brain Sci. 2020, 10, 687. [Google Scholar] [CrossRef]
  6. Jiang, Y.; Li, W.; Hossain, M.S.; Chen, M.; Alelaiwi, A.; Al-Hammadi, M. A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition. Inf. Fusion 2020, 53, 209–221. [Google Scholar] [CrossRef]
  7. Middya, A.I.; Nag, B.; Roy, S. Deep learning based multimodal emotion recognition using model-level fusion of audio–visual modalities. Knowl.-Based Syst. 2022, 244, 108580. [Google Scholar] [CrossRef]
  8. Liu, W.; Qiu, J.L.; Zheng, W.L.; Lu, B.L. Comparing recognition performance and robustness of multimodal deep learning models for multimodal emotion recognition. IEEE Trans. Cogn. Dev. Syst. 2021, 14, 715–729. [Google Scholar] [CrossRef]
  9. Maithri, M.; Raghavendra, U.; Gudigar, A.; Samanth, J.; Barua, P.D.; Murugappan, M.; Acharya, U.R. Automated emotion recognition: Current trends and future perspectives. Comput. Methods Programs Biomed. 2022, 215, 106646. [Google Scholar] [CrossRef]
  10. Ranganathan, H.; Chakraborty, S.; Panchanathan, S. Multimodal emotion recognition using deep learning architectures. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–9. [Google Scholar] [CrossRef]
  11. Pan, B.; Hirota, K.; Jia, Z.; Dai, Y. A review of multimodal emotion recognition from datasets, preprocessing, features, and fusion methods. Neurocomputing 2023, 561, 126866. [Google Scholar] [CrossRef]
  12. Zhang, S.; Yang, Y.; Chen, C.; Zhang, X.; Leng, Q.; Zhao, X. Deep learning-based multimodal emotion recognition from audio, visual, and text modalities: A systematic review of recent advancements and future prospects. Expert Syst. Appl. 2024, 237, 121692. [Google Scholar] [CrossRef]
  13. Lian, H.; Lu, C.; Li, S.; Zhao, Y.; Tang, C.; Zong, Y. A survey of deep learning-based multimodal emotion recognition: Speech, text, and face. Entropy 2023, 25, 1440. [Google Scholar] [CrossRef] [PubMed]
  14. Cai, Y.; Li, X.; Li, J. Emotion Recognition Using Different Sensors, Emotion Models, Methods and Datasets: A Comprehensive Review. Sensors 2023, 23, 2455. [Google Scholar] [CrossRef] [PubMed]
  15. Ramaswamy, M.P.A.; Palaniswamy, S. Multimodal Emotion Recognition: A Comprehensive Review, Trends, and Challenges. WIREs Data Min. Knowl. Discov. 2024, 14, e1563. [Google Scholar] [CrossRef]
  16. Gladys, A.A.; Vetriselvi, V. Survey on Multimodal Approaches to Emotion Recognition. Neurocomputing 2023, 556, 126693. [Google Scholar] [CrossRef]
  17. Yi, M.-H.; Kwak, K.-C.; Shin, J.-H. HyFusER: Hybrid Multimodal Transformer for Emotion Recognition Using Dual Cross-Modal Attention. Appl. Sci. 2025, 15, 1053. [Google Scholar] [CrossRef]
  18. Cheng, Z.; Bu, X.; Wang, Q.; Yang, T.; Tu, J. EEG-Based Emotion Recognition Using Multi-Scale Dynamic CNN and Gated Transformer. Sci. Rep. 2024, 14, 31319. [Google Scholar] [CrossRef]
  19. Zhao, Z.; Wang, Y.; Shen, G.; Xu, Y.; Zhang, J. TDFNet: Transformer-based deep-scale fusion network for multimodal emotion recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 3771–3782. [Google Scholar] [CrossRef]
  20. Alam, M.M.; Dini, M.A.; Kim, D.-S.; Jun, T. TMNet: Transformer-Fused Multimodal Framework for Emotion Recognition via EEG and Speech. ICT Express 2025, in press. [Google Scholar] [CrossRef]
  21. Liao, Y.; Gao, Y.; Wang, F.; Zhang, L.; Xu, Z.; Wu, Y. Emotion Recognition with Multiple Physiological Parameters Based on Ensemble Learning. Sci. Rep. 2025, 15, 19869. [Google Scholar] [CrossRef]
  22. Chen, L.; Wang, K.; Li, M.; Wu, M.; Pedrycz, W.; Hirota, K. K-means clustering-based kernel canonical correlation analysis for multimodal emotion recognition in human–robot interaction. IEEE Trans. Ind. Electron. 2022, 70, 1016–1024. [Google Scholar] [CrossRef]
  23. Choi, W.Y.; Song, K.Y.; Lee, C.W. Convolutional attention networks for multimodal emotion recognition from speech and text data. In Proceedings of Grand Challenge and Workshop on Human Multimodal Language (Challenge-HML); Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 28–34. [Google Scholar] [CrossRef]
  24. Huang, J.; Tao, J.; Liu, B.; Lian, Z.; Niu, M. Multimodal transformer fusion for continuous emotion recognition. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3507–3511. [Google Scholar]
  25. Fu, B.; Gu, C.; Fu, M.; Xia, Y.; Liu, Y. A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals. Front. Neurosci. 2023, 17, 1234162. [Google Scholar] [CrossRef]
  26. Rozgić, V.; Ananthakrishnan, S.; Saleem, S.; Kumar, R.; Prasad, R. Ensemble of SVM trees for multimodal emotion recognition. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; pp. 1–4. [Google Scholar]
  27. Zheng, W.L.; Liu, W.; Lu, Y.; Lu, B.L.; Cichocki, A. EmotionMeter: A multimodal framework for recognizing human emotions. IEEE Trans. Cybern. 2019, 49, 1110–1122. [Google Scholar] [CrossRef] [PubMed]
  28. Ma, H.; Wang, J.; Lin, H.; Zhang, B.; Zhang, Y.; Xu, B. A transformer-based model with self-distillation for multimodal emotion recognition in conversations. IEEE Trans. Multimed. 2023, 26, 776–788. [Google Scholar] [CrossRef]
  29. Barros, P.; Jirak, D.; Weber, C.; Wermter, S. Multimodal emotional state recognition using sequence-dependent deep hierarchical features. Neural Netw. 2015, 72, 140–151. [Google Scholar] [CrossRef]
  30. Tripathi, S.; Tripathi, S.; Beigi, H. Multi-modal emotion recognition on IEMOCAP dataset using deep learning. arXiv 2018, arXiv:1804.05788. [Google Scholar]
  31. Li, X.; Liu, J.; Xie, Y.; Gong, P.; Zhang, X.; He, H. MAGDRA: A multi-modal attention graph network with dynamic routing-by-agreement for multi-label emotion recognition. Knowl.-Based Syst. 2024, 283, 111126. [Google Scholar] [CrossRef]
  32. Tzirakis, P.; Trigeorgis, G.; Nicolaou, M.A.; Schuller, B.W.; Zafeiriou, S. End-to-end multimodal emotion recognition using deep neural networks. IEEE J. Sel. Top. Signal Process. 2017, 11, 1301–1309. [Google Scholar] [CrossRef]
  33. Xu, H.; Zhang, H.; Han, K.; Wang, Y.; Peng, Y.; Li, X. Learning alignment for multimodal emotion recognition from speech. arXiv 2019, arXiv:1909.05645. [Google Scholar]
  34. Park, C.Y.; Cha, N.; Kang, S.; Kim, A.; Khandoker, A.H.; Hadjileontiadis, L.; Lee, U. K-EmoCon: A multimodal sensor dataset for continuous emotion recognition in naturalistic conversations. Sci. Data 2020, 7, 293. [Google Scholar] [CrossRef]
  35. Siriwardhana, S.; Kaluarachchi, T.; Billinghurst, M.; Nanayakkara, S. Multimodal emotion recognition with transformer-based self-supervised feature fusion. IEEE Access 2020, 8, 176274–176285. [Google Scholar] [CrossRef]
  36. Mocanu, B.; Tapu, R.; Zaharia, T. Multimodal emotion recognition using cross-modal audio–video fusion with attention and deep metric learning. Image Vis. Comput. 2023, 133, 104676. [Google Scholar] [CrossRef]
  37. Wu, Y.; Zhang, S.; Li, P. Multi-Modal Emotion Recognition in Conversation Based on Prompt Learning with Text–Audio Fusion Features. Sci. Rep. 2025, 15, 8855. [Google Scholar] [CrossRef] [PubMed]
  38. Ahmed, N.; Al Aghbari, Z.; Girija, S. A systematic survey on multimodal emotion recognition using learning algorithms. Intell. Syst. Appl. 2023, 17, 200171. [Google Scholar] [CrossRef]
  39. Lian, Z.; Sun, H.; Sun, L.; Yi, J.; Liu, B.; Tao, J. AffectGPT: Dataset and framework for explainable multimodal emotion recognition. arXiv 2024, arXiv:2407.07653. [Google Scholar]
  40. Zhang, Y.; Cheng, C.; Wang, S.; Xia, T. Emotion recognition using heterogeneous convolutional neural networks combined with multimodal factorized bilinear pooling. Biomed. Signal Process. Control 2022, 77, 103877. [Google Scholar] [CrossRef]
  41. Liu, S.; Gao, P.; Li, Y.; Fu, W.; Ding, W. Multi-modal fusion network with complementarity and importance for emotion recognition. Inf. Sci. 2023, 619, 679–694. [Google Scholar] [CrossRef]
  42. Lerner, M.D.; McPartland, J.C.; Morris, J.P. Multimodal emotion processing in autism spectrum disorders: An event-related potential study. Dev. Cogn. Neurosci. 2013, 3, 11–21. [Google Scholar] [CrossRef]
  43. Ho, N.H.; Yang, H.J.; Kim, S.H.; Lee, G. Multimodal approach of speech emotion recognition using multi-level multi-head fusion attention-based recurrent neural network. IEEE Access 2020, 8, 61672–61686. [Google Scholar] [CrossRef]
  44. Zhang, H. Expression-EEG based collaborative multimodal emotion recognition using deep autoencoder. IEEE Access 2020, 8, 164130–164143. [Google Scholar] [CrossRef]
  45. Cimtay, Y.; Ekmekcioglu, E.; Caglar-Ozhan, S. Cross-subject multimodal emotion recognition based on hybrid fusion. IEEE Access 2020, 8, 168865–168878. [Google Scholar] [CrossRef]
  46. Chudasama, V.; Kar, P.; Gudmalwar, A.; Shah, N.; Wasnik, P.; Onoe, N. M2FNet: Multi-modal fusion network for emotion recognition in conversation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 4652–4661. [Google Scholar] [CrossRef]
  47. Sharafi, M.; Yazdchi, M.; Rasti, R.; Nasimi, F. A novel spatio-temporal convolutional neural framework for multimodal emotion recognition. Biomed. Signal Process. Control. 2022, 78, 103970. [Google Scholar] [CrossRef]
  48. Tang, H.; Liu, W.; Zheng, W.L.; Lu, B.L. Multimodal emotion recognition using deep neural networks. In Proceedings of the Neural Information Processing—24th International Conference ICONIP 2017, Guangzhou, China, 14–18 November 2017; pp. 811–819. [Google Scholar] [CrossRef]
  49. Mittal, T.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. M3ER: Multiplicative multimodal emotion recognition using facial, textual, and speech cues. Proc. AAAI Conf. Artif. Intell. 2020, 34, 1359–1367. [Google Scholar] [CrossRef]
  50. Povolný, F.; Matejka, P.; Hradis, M.; Popková, A.; Otrusina, L.; Smrž, P.; Lamel, L. Multimodal emotion recognition for AVEC 2016 challenge. In Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge (AVEC’16), New York, NY, USA, 16 October 2016; pp. 75–82. [Google Scholar] [CrossRef]
  51. Yoon, S.; Byun, S.; Jung, K. Multimodal speech emotion recognition using audio and text. In Proceedings of the 2018 IEEE Spoken Language Technology Workshop (SLT), Athens, Greece, 18–21 December 2018; pp. 112–118. [Google Scholar] [CrossRef]
  52. Hazarika, D.; Poria, S.; Mihalcea, R.; Cambria, E.; Zimmermann, R. ICON: Interactive conversational memory network for multimodal emotion detection. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), Brussel, Belgium, 31 October–4 November 2018; pp. 2594–2604. [Google Scholar]
  53. Wu, X.; Zheng, W.L.; Li, Z.; Lu, B.L. Investigating EEG-based functional connectivity patterns for multimodal emotion recognition. J. Neural Eng. 2022, 19, 016012. [Google Scholar] [CrossRef] [PubMed]
  54. Qiu, S.; Sekhar, N.; Singhal, P. Topic- and Style-Aware Transformer for Multimodal Emotion Recognition. In Findings of the Association for Computational Linguistics: ACL 2023; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 2074–2082. [Google Scholar] [CrossRef]
  55. Du, Z.; Ye, X.; Zhao, P. A Novel Signal Channel Attention Network for Multi-Modal Emotion Recognition. Front. Neurorobot. 2024, 18, 1442080. [Google Scholar] [CrossRef]
  56. Khan, M.; Tran, P.-N.; Pham, N.T.; El Saddik, A.; Othmani, A. MemoCMT: Multimodal Emotion Recognition Using Cross-Modal Transformer-Based Feature Fusion. Sci. Rep. 2025, 15, 5473. [Google Scholar] [CrossRef]
  57. Nandini, D.; Yadav, J.; Singh, V.; Mohan, V.; Agarwal, S. An Ensemble Deep Learning Framework for Emotion Recognition through Wearable Devices’ Multi-Modal Physiological Signals. Sci. Rep. 2025, 15, 17263. [Google Scholar] [CrossRef]
  58. Mengara Mengara, A.G.; Moon, Y. CAG-MoE: Multimodal Emotion Recognition with Cross-Attention Gated Mixture of Experts. Mathematics 2023, 13, 1907. [Google Scholar] [CrossRef]
  59. Ahuja, C.; Sethia, D. SS-EMERGE: Self-Supervised Enhancement for Multidimension Emotion Recognition Using GNNs for EEG. Sci. Rep. 2025, 15, 14254. [Google Scholar] [CrossRef]
  60. Yang, D.; Huang, S.; Kuang, H.; Du, Y.; Zhang, L. Disentangled representation learning for multimodal emotion recognition. In Proceedings of the 30th ACM International Conference on Multimedia (MM’22), Lisbon, Portugal, 10–14 October 2022; pp. 1642–1651. [Google Scholar]
  61. Kahou, S.E.; Bouthillier, X.; Lamblin, P.; Gulcehre, C.; Michalski, V.; Konda, K.; Jean, S.; Froumenty, P.; Dauphin, Y.; Boulanger-Lewandowski, N.; et al. EmoNets: Multimodal deep learning approaches for emotion recognition in video. J. Multimodal User Interfaces 2016, 10, 99–111. [Google Scholar] [CrossRef]
  62. Huang, Y.; Yang, J.; Liao, P.; Pan, J. Fusion of facial expressions and EEG for multimodal emotion recognition. Comput. Intell. Neurosci. 2017, 2017, 2107451. [Google Scholar] [CrossRef]
  63. Singla, C.; Singh, S.; Sharma, P.; Mittal, N.; Gared, F. Emotion Recognition for Human–Computer Interaction Using High-Level Descriptors. Sci. Rep. 2024, 14, 12122. [Google Scholar] [CrossRef] [PubMed]
  64. Triantafyllopoulos, A.; Christ, L.; Gebhard, A.; Jing, X.; Kathan, A.; Milling, M.; Tsangko, I.; Amiriparian, S.; Schuller, B.W. Beyond Deep Learning: Charting the Next Frontiers of Affective Computing. Intell. Comput. 2024, 3, e0089. [Google Scholar] [CrossRef]
  65. Wang, X.; Ren, Y.; Luo, Z.; He, W.; Hong, J.; Huang, Y. Deep Learning-Based EEG Emotion Recognition: Current Trends and Future Perspectives. Front. Psychol. 2023, 14, 1126994. [Google Scholar] [CrossRef]
  66. Alsabhan, W. Human–Computer Interaction with a Real-Time Speech Emotion Recognition with Ensembling Techniques 1D Convolution Neural Network and Attention. Sensors 2023, 23, 1386. [Google Scholar] [CrossRef] [PubMed]
  67. Wang, Y.; Guan, L.; Venetsanopoulos, A.N. Kernel cross-modal factor analysis for information fusion with application to bimodal emotion recognition. IEEE Trans. Multimed. 2012, 14, 597–607. [Google Scholar] [CrossRef]
  68. Dai, W.; Liu, Z.; Yu, T.; Fung, P. Modality-transferable emotion embeddings for low-resource multimodal emotion recognition. arXiv 2020, arXiv:2009.09629. [Google Scholar]
  69. Zeng, Y.; Zhang, J.W.; Yang, J. Multimodal emotion recognition in the metaverse era: New needs and transformation in mental health work. World J. Clin. Cases 2024, 12, 6674–6678. [Google Scholar] [CrossRef]
  70. Pillalamarri, R.; Shanmugam, U. A Review on EEG-Based Multimodal Learning for Emotion Recognition. Artif. Intell. Rev. 2025, 58, 131. [Google Scholar] [CrossRef]
  71. Akinpelu, S.; Viriri, S. Speech Emotion Classification Using Attention-Based Network and Regularized Feature Selection. Sci. Rep. 2023, 13, 11990. [Google Scholar] [CrossRef]
  72. Islam, M.R.; Islam, M.M.; Rahman, M.M.; Mondal, C.; Singha, S.K.; Ahmad, M.; Awal, A.; Islam, M.S.; Moni, M.A. EEG Channel Correlation Based Model for Emotion Recognition. Comput. Biol. Med. 2021, 136, 104757. [Google Scholar] [CrossRef]
  73. Chowdhury, J.H.; Ramanna, S.; Kotecha, K. Speech Emotion Recognition with Lightweight Deep Neural Ensemble Model Using Hand-Crafted Features. Sci. Rep. 2025, 15, 11824. [Google Scholar] [CrossRef] [PubMed]
  74. Li, D.; Wang, Y.; Funakoshi, K.; Okumura, M. JOYFUL: Joint Modality Fusion and Graph Contrastive Learning for Multimodal Emotion Recognition. In Proceedings of the EMNLP 2023, Singapore, 6–10 December 2023; pp. 16051–16069. [Google Scholar] [CrossRef]
  75. Huang, Z.; Mak, M.-W.; Lee, K.A. MM-NodeFormer: Node Transformer Multimodal Fusion for Emotion Recognition in Conversation. In Proceedings of the Interspeech 2024, Sapporo, Japan, 8–12 September 2024; pp. 4069–4073. [Google Scholar] [CrossRef]
  76. Li, Z.; Tang, F.; Zhao, M.; Zhu, Y. EmoCaps: Emotion Capsule-Based Model for Conversational Emotion Recognition. In Proceedings of the Findings of ACL 2022, Dublin, Ireland, 22–27 May 2022; pp. 1610–1618. [Google Scholar] [CrossRef]
  77. Lu, C.; Tang, C.; Zhang, J.; Zong, Y. Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition. Entropy 2022, 24, 1046. [Google Scholar] [CrossRef] [PubMed]
  78. Sun, H.; Zhao, S.; Wang, X.; Zeng, W.; Chen, Y.; Qin, Y. Fine-grained disentangled representation learning for multimodal emotion recognition. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 11051–11055. [Google Scholar]
  79. Yang, X.; Feng, S.; Wang, D.; Zhang, Y. Image–text multimodal emotion classification via multi-view attentional network. IEEE Trans. Multimed. 2020, 23, 4014–4026. [Google Scholar] [CrossRef]
  80. Krishna, D.N.; Patil, A. Multimodal emotion recognition using cross-modal attention and 1D convolutional neural networks. In Proceedings of the Interspeech 2020, Shanghai, China, 25–29 October 2020; pp. 4243–4247. [Google Scholar]
  81. Lv, F.; Chen, X.; Huang, Y.; Duan, L.; Lin, G. Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2554–2562. [Google Scholar] [CrossRef]
  82. Wu, C.; Cai, Y.; Liu, Y.; Zhu, P.; Xue, Y.; Gong, Z.; Hirschberg, J.; Ma, B. Multimodal Emotion Recognition in Conversations: A Survey of Methods, Trends, Challenges and Prospects. arXiv 2025, arXiv:2505.20511. [Google Scholar] [CrossRef]
  83. Nguyen, D.; Nguyen, K.; Sridharan, S.; Dean, D.; Fookes, C. Deep spatio-temporal feature fusion with compact bilinear pooling for multimodal emotion recognition. Comput. Vis. Image Underst. 2018, 174, 33–42. [Google Scholar] [CrossRef]
  84. Chen, S.; Jin, Q. Multi-modal dimensional emotion recognition using recurrent neural networks. In Proceedings of the 5th International Workshop on Audio/Visual Emotion Challenge (AVEC’15), Brisbane, Australia, 26 October 2015; pp. 49–56. [Google Scholar] [CrossRef]
  85. Avots, E.; Sapiński, T.; Bachmann, M.; Kamińska, D. Audiovisual emotion recognition in wild. Mach. Vis. Appl. 2019, 30, 975–985. [Google Scholar] [CrossRef]
  86. Verma, G.K.; Tiwary, U.S. Multimodal fusion framework: A multiresolution approach for emotion classification and recognition from physiological signals. NeuroImage 2014, 102, 162–172. [Google Scholar] [CrossRef]
  87. Al-Saddawi, H.F.T.; Das, B.; Das, R. A Systematic Review of Trimodal Affective-Computing Approaches: Text, Audio, and Visual Integration in Emotion Recognition and Sentiment Analysis. Expert Syst. Appl. 2024, 255, 124852. [Google Scholar] [CrossRef]
  88. Zhao, Y.; Chen, D. Expression-EEG multimodal emotion recognition method based on the bidirectional LSTM and attention mechanism. Comput. Math. Methods Med. 2021, 2021, 9967592. [Google Scholar] [CrossRef]
  89. Meng, T.; Shou, Y.; Ai, W.; Yin, N.; Li, K. Deep imbalanced learning for multimodal emotion recognition in conversations. IEEE Trans. Artif. Intell. 2024, 5, 6472–6487. [Google Scholar] [CrossRef]
  90. Zaidi, S.A.M.; Latif, S.; Qadir, J. Enhancing Cross-Language Multimodal Emotion Recognition with Dual Attention Transformers. IEEE Open J. Comput. Soc. 2024, 5, 684–693. [Google Scholar] [CrossRef]
  91. Hu, G.; Lin, T.-E.; Zhao, Y.; Lu, G.; Wu, Y.; Li, Y. UniMSE: Towards Unified Multimodal Sentiment Analysis and Emotion Recognition. In Proceedings of the EMNLP 2022, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 7837–7851. [Google Scholar] [CrossRef]
  92. Waligora, P.; Aslam, M.H.; Zeeshan, M.O.; Belharbi, S.; Lameiras Koerich, A.; Pedersoli, M.; Bacon, S.; Granger, E. Joint Multimodal Transformer for Emotion Recognition in the Wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024; pp. 4625–4635. [Google Scholar] [CrossRef]
  93. Wagner, J.; André, E.; Lingenfelser, F.; Kim, J. Exploring fusion methods for multimodal emotion recognition with missing data. IEEE Trans. Affect. Comput. 2011, 2, 206–218. [Google Scholar] [CrossRef]
  94. Mittal, T.; Guhan, P.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. EMOTICON: Context-aware multimodal emotion recognition using Frege’s principle. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 14234–14243. [Google Scholar]
  95. Kumar, P.; Malik, S.; Raman, B. Hybrid fusion based interpretable multimodal emotion recognition with insufficient labelled data. arXiv 2022, arXiv:2208.11450. [Google Scholar]
  96. Praveen, R.G.; Alam, J. Recursive Joint Cross-Modal Attention for Multimodal Fusion in Dimensional Emotion Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024; pp. 4803–4813. [Google Scholar] [CrossRef]
  97. Jones, C.R.; Pickles, A.; Falcaro, M.; Marsden, A.J.; Happé, F.; Scott, S.K.; Charman, T. A multimodal approach to emotion recognition ability in autism spectrum disorders. J. Child Psychol. Psychiatry 2011, 52, 275–285. [Google Scholar] [CrossRef]
  98. Ayata, D.; Yaslan, Y.; Kamasak, M.E. Emotion recognition from multimodal physiological signals for emotion-aware healthcare systems. J. Med. Biol. Eng. 2020, 40, 149–157. [Google Scholar] [CrossRef]
  99. Chen, S.; Jin, Q.; Zhao, J.; Wang, S. Multimodal multi-task learning for dimensional and continuous emotion recognition. In Proceedings of the 7th Annual Workshop Audio/Visual Emotion Challenge (AVEC’17), Mountain View, CA, USA, 23 October 2017; pp. 19–26. [Google Scholar] [CrossRef]
  100. Das, A.; Sen Sarma, M.; Hoque, M.M.; Siddique, N.; Dewan, M.A.A. AVaTER: Fusing Audio, Visual, and Textual Modalities Using Cross-Modal Attention for Emotion Recognition. Sensors 2024, 24, 5862. [Google Scholar] [CrossRef]
  101. Makiuchi, M.R.; Uto, K.; Shinoda, K. Multimodal emotion recognition with high-level speech and text features. In Proceedings of the 2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), Cartagena, Colombia, 13–17 December 2021; pp. 350–357. [Google Scholar] [CrossRef]
  102. Chang, E.J.; Rahimi, A.; Benini, L.; Wu, A.Y.A. Hyperdimensional computing-based multimodality emotion recognition with physiological signals. In Proceedings of the 2019 IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS), Hsinchu, Taiwan, 18–20 March 2019; pp. 137–141. [Google Scholar]
  103. Xie, B.; Sidulova, M.; Park, C.H. Robust multimodal emotion recognition from conversation with transformer-based cross-modality fusion. Sensors 2021, 21, 4913. [Google Scholar] [CrossRef]
Figure 1. PRISMA flow diagram.
Figure 1. PRISMA flow diagram.
Biomimetics 10 00418 g001
Figure 2. Record counts at each PRISMA stage.
Figure 2. Record counts at each PRISMA stage.
Biomimetics 10 00418 g002
Figure 3. Distribution of modality types in 103 MER studies.
Figure 3. Distribution of modality types in 103 MER studies.
Biomimetics 10 00418 g003
Figure 4. Annual trend of modality adoption, 2011–2025.
Figure 4. Annual trend of modality adoption, 2011–2025.
Biomimetics 10 00418 g004
Figure 5. Word cloud of modality keyword frequencies, 2011–2025.
Figure 5. Word cloud of modality keyword frequencies, 2011–2025.
Biomimetics 10 00418 g005
Figure 6. Prevalence of modalities by lead author country.
Figure 6. Prevalence of modalities by lead author country.
Biomimetics 10 00418 g006
Figure 7. The workflow of the MER system.
Figure 7. The workflow of the MER system.
Biomimetics 10 00418 g007
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Y.; Mi, Q.; Gao, T. A Comprehensive Review of Multimodal Emotion Recognition: Techniques, Challenges, and Future Directions. Biomimetics 2025, 10, 418. https://doi.org/10.3390/biomimetics10070418

AMA Style

Wu Y, Mi Q, Gao T. A Comprehensive Review of Multimodal Emotion Recognition: Techniques, Challenges, and Future Directions. Biomimetics. 2025; 10(7):418. https://doi.org/10.3390/biomimetics10070418

Chicago/Turabian Style

Wu, You, Qingwei Mi, and Tianhan Gao. 2025. "A Comprehensive Review of Multimodal Emotion Recognition: Techniques, Challenges, and Future Directions" Biomimetics 10, no. 7: 418. https://doi.org/10.3390/biomimetics10070418

APA Style

Wu, Y., Mi, Q., & Gao, T. (2025). A Comprehensive Review of Multimodal Emotion Recognition: Techniques, Challenges, and Future Directions. Biomimetics, 10(7), 418. https://doi.org/10.3390/biomimetics10070418

Article Metrics

Back to TopTop