Next Article in Journal
IDG-ViolenceNet: A Video Violence Detection Model Integrating Identity-Aware Graphs and 3D-CNN
Previous Article in Journal
A Novel Pavement Abrasion Test for Assessing Injury Risk to Vulnerable Road Users
Previous Article in Special Issue
Robust Industrial Surface Defect Detection Using Statistical Feature Extraction and Capsule Network Architectures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

M3ENet: A Multi-Modal Fusion Network for Efficient Micro-Expression Recognition

by
Ke Zhao
1,2,*,
Xuanyu Liu
3 and
Guangqian Yang
4
1
Electronic Science and Technology Museum, University of Electronic Science and Technology of China, Chengdu 611731, China
2
School of Public Administration, University of Electronic Science and Technology of China, Chengdu 611731, China
3
State Key Laboratory of Photonics and Communications, Peking University, Beijing 100871, China
4
Department of Biomedical Engineering, Hong Kong Polytechnic University, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(20), 6276; https://doi.org/10.3390/s25206276 (registering DOI)
Submission received: 1 September 2025 / Revised: 7 October 2025 / Accepted: 8 October 2025 / Published: 10 October 2025
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems—2nd Edition)

Abstract

Micro-expression recognition (MER) aims to detect brief and subtle facial movements that reveal suppressed emotions, discerning authentic emotional responses in scenarios such as visitor experience analysis in museum settings. However, it remains a highly challenging task due to the fleeting duration, low intensity, and limited availability of annotated data. Most existing approaches rely solely on either appearance or motion cues, thereby restricting their ability to capture expressive information fully. To overcome these limitations, we propose a lightweight multi-modal fusion network, termed M3ENet, which integrates both motion and appearance cues through early-stage feature fusion. Specifically, our model extracts horizontal, vertical, and strain-based optical flow between the onset and apex frames, alongside RGB images from the onset, apex, and offset frames. These inputs are processed by two modality-specific subnetworks, whose features are fused to exploit complementary information for robust classification. To improve generalization in low data regimes, we employ targeted data augmentation and adopt focal loss to mitigate class imbalance. Extensive experiments on five benchmark datasets, including CASME I, CASME II, CAS(ME)2, SAMM, and MMEW, demonstrate that M3ENet achieves state-of-the-art performance with high efficiency. Ablation studies and Grad-CAM visualizations further confirm the effectiveness and interpretability of the proposed architecture.

1. Introduction

Facial expression recognition (FER) plays a crucial role in understanding human emotions, with applications across various domains such as healthcare, education, surveillance, and biomedical engineering. By analyzing facial muscle movements, FER enables machines to detect emotional cues, facilitating natural human–computer interaction. In biomedical engineering, FER aids in monitoring patient emotions, detecting psychological conditions, and supporting therapeutic interventions for neurological disorders. Additionally, FER is increasingly used to analyze visitor experience in museums [1], providing insights into emotional engagement and satisfaction, which can help improve exhibits and overall visitor interaction. Among various facial expressions, micro-expressions (MEs) are particularly significant due to their involuntary, subtle, and fleeting nature, typically lasting between 1/25 and 1/5 s [2,3]. Unlike macro-expressions, MEs are automatic and difficult to control, making them reliable indicators of a person’s genuine emotional state. These brief facial movements often emerge when individuals attempt to suppress or conceal their true feelings, particularly in high-stakes or emotionally intense situations. As a result, MEs are crucial for applications in lie detection, criminal interrogation, and psychological assessments, where understanding concealed emotions is essential. In biomedical contexts, recognizing MEs holds great promise for monitoring the emotional and psychological health of patients, particularly those with conditions such as Parkinson’s disease, autism spectrum disorders, or facial paralysis. By detecting these subtle emotional cues, FER systems can provide real-time insights into a patient’s emotional well-being, supporting more effective therapies and interventions. Despite their potential, the recognition of MEs remains a challenging task due to their short duration, subtle intensity, and the limited availability of annotated datasets. To address these challenges, biomedical engineering approaches combining image processing, machine learning, and computer vision are being developed to create more efficient and accurate recognition systems, making MEs a valuable tool for both research and clinical applications.
Early research in micro-expression recognition predominantly relied on handcrafted features, which aimed to extract discriminative patterns from shallow visual cues. Among these, Local Binary Patterns on Three Orthogonal Planes (LBP-TOP) [4] and its variants have been widely adopted due to their compact representation, low computational cost, and robustness to illumination changes. Several extensions have been proposed to enhance LBP’s spatiotemporal modeling ability, such as Local Binary Pattern with Six Intersection Points (LBP-SIP) [5], Local Binary Pattern with Mean Orthogonal Planes (LBP-MOP) [6], and Spatiotemporal Completed Local Quantization Patterns (STCLQP) [7]. These methods extract local texture descriptors across temporal frames to represent subtle facial movements. In parallel, optical flow-based techniques have also been explored to capture the subtle motions between consecutive frames. Optical flow estimates pixel-wise movement by computing brightness changes, making it suitable for detecting subtle facial muscle dynamics. Representative approaches include Optical Strain Features (OSFs) [8], Optical Strain Weights (OSWs) [9], Fuzzy Histogram of Oriented Optical Flow (FHOOF) [10], Main Directional Mean Optical Flow (MDMO) [11], and the more recent Bi-Weighted Oriented Optical Flow (Bi-WOOF) [12]. While these handcrafted approaches offer interpretability and require relatively fewer training samples, they often suffer from limited discriminative power and generalization capability, especially under uncontrolled conditions involving variations in head pose, illumination, and individual differences. Moreover, the reliance on carefully designed descriptors and prior knowledge limits their scalability and adaptability in practical applications.
With the advent of deep learning, particularly convolutional neural networks (CNNs), researchers have increasingly explored data-driven methods for micro-expression recognition. Early attempts, such as the work by Patel et al. [13], adopted standard CNN architectures but failed to outperform traditional handcrafted descriptors due to overfitting and insufficient data. Subsequent works addressed these limitations by integrating domain-specific priors. For instance, Peng et al. [14] introduced a dual-temporal-scale CNN using optical flow inputs to capture multi-scale motion features.
To further exploit temporal dynamics, some methods combined CNNs with recurrent structures such as LSTM [15] or introduced frame alignment techniques (e.g., TIM [16]) to standardize input lengths. Liong et al. [17] proposed a Shallow Triple-Stream 3D CNN (STSTNet) to process horizontal, vertical, and optical strain components of apex-frame optical flow, achieving promising performance with low model complexity. More recently, inspired by the success of Transformers in modeling long-range dependencies, Wang et al. [18] adopted a Transformer encoder to extract global spatial representations from long-term optical flow sequences, further combined with LSTM for temporal modeling.
Despite these advances, most existing deep learning methods remain confined to a single modality, typically either RGB sequences or optical flow. As a result, they fail to exploit the complementary nature of appearanceand motion information. Empirical studies have shown that while optical flow captures subtle motion cues, RGB frames retain rich appearance and texture details, both of which are crucial for accurately identifying fleeting facial expressions. Relying solely on one modality often limits the representational power of the network. It leads to suboptimal performance, especially under challenging scenarios such as low intensity, ambiguous expressions, or inter-subject variation.
These limitations highlight the necessity of multi-modal fusion, which aims to integrate appearance and motion features in a unified framework. By jointly modeling RGB and optical flow representations, a well-designed multi-modal system can effectively leverage their complementary strengths, improve robustness to visual variability, and enhance overall recognition accuracy. This motivates us to develop a lightweight, multi-stream architecture that fuses features at an early stage, enabling more effective and efficient micro-expression recognition.
To address these limitations, we propose a novel five-stream CNN framework for unified feature extraction from both RGB frames and optical flow data. Specifically, we extract three types of optical flow components—horizontal, vertical, and optical strain—between the onset and apex frames to capture subtle motion dynamics. In parallel, we use three key RGB frames—the onset, apex, and offset frames—as inputs to extract appearance-based features. This design not only reduces the computational burden compared to processing full video clips but also eliminates redundant frames while preserving crucial temporal dynamics of micro-expressions. We employ two independent subnetworks to learn modality-specific representations for optical flow and RGB inputs, respectively. These features are subsequently fused at the feature level to leverage their complementary information. The final classification is achieved through a fully connected layer followed by a softmax function. Furthermore, to mitigate the limited data availability in current micro-expression datasets, we design a comprehensive data augmentation strategy, including image flipping, rotation, and additive noise, which significantly boosts performance and enhances generalization across different datasets. Our contributions are summarized below:
  • We propose a unified five-stream CNN architecture that effectively integrates both optical flow and RGB modalities, enabling joint modeling of motion and appearance cues critical for micro-expression recognition.
  • To tackle sample scarcity and label imbalance in micro-expression datasets, we introduce a customized data augmentation pipeline and incorporate focal loss, leading to consistent performance gains across standard benchmarks.
  • We conduct extensive experiments on publicly available datasets and demonstrate that our approach achieves state-of-the-art (SOTA) performance while maintaining high computational efficiency, offering a strong baseline for robust and interpretable multi-modal micro-expression analysis.

2. Related Work

2.1. ME Datasets

The construction of high-quality ME datasets is crucial for accurate ME recognition. Unlike other computer vision tasks’ datasets, such as ImageNet [19] or CIFAR-10 [20], where data collection is relatively straightforward, MEs are characterized by their spontaneous, involuntary, and fleeting nature, often triggered under high-stakes or emotionally intense conditions. This makes both data acquisition and annotation significantly more challenging.
In recent years, several spontaneous ME datasets have been introduced, including SMIC [21] and its extended version SMIC-E, CASME [22], CASME II [23], CAS(ME)2 [24], SAMM [25], and the Macro- and Micro-Expression Warehouse (MMEW) [26]. This paper focuses on these spontaneous datasets.
The following summarizes the key characteristics of the most widely used spontaneous ME datasets, highlighting their data collection protocols, annotation strategies, and contributions to the field. These datasets provide the foundation for training and evaluating ME recognition models under various settings.
  • SMIC [21]: SMIC (Spontaneous Micro-Expression Corpus) was the first public spontaneous ME dataset, containing 71 video clips from 20 participants. It includes three modalities: high speed (HS, 100 fps), visible light (VIS, 25 fps), and near infrared (NIR), captured while participants watched emotion-inducing videos under instructions to maintain a neutral expression. Each clip is labeled via participant self-report as “positive,” “negative,” or “surprise,” and FACS-trained coders provide frame-level annotations to ensure reliability.
  • CASME [22]: CASME (Chinese Academy of Sciences Micro-Expression) contains 195 spontaneous MEs elicited from 35 participants. Videos were captured at 60 fps in a controlled setting, with coding of onset, apex, and offset frames. Emotion labels stem from a combination of participant self-reports and AU-based FACS.
  • CASME II [23]: CASME II builds on CASME with 247 spontaneous micro-expression samples from 26 participants. Videos were recorded at an improved 200 fps and 280 × 340 px resolution. Labeling combines AUs, subjective reports, and contextual video information, with ambiguous cases tagged as “others”. It is the most widely used dataset for MER.
  • CAS(ME)2 [24]: CAS(ME)2 extends CASME II by capturing both spontaneous micro- and macro-expressions from the same 22 participants. It contains 57 micro-expression samples and 300 cropped macro-expression samples, all recorded at 30 fps. In addition, 87 long video sequences are provided for joint evaluation of expression spotting and recognition. Emotion labels are assigned based on a combination of facial action units (AUs), the nature of the emotion-inducing videos, and participant self-reports, categorized into “positive,” “negative,” “surprise,” and “others.” The dataset facilitates research into cross-scale expression dynamics and supports both recognition and temporal spotting tasks.
  • SAMM [25]: SAMM (Spontaneous Actions and Micro-Movements) is a high-resolution spontaneous ME dataset comprising 159 micro-expression samples recorded at 200 or 300 fps under controlled lighting conditions, with facial regions captured at 2040 × 1088 resolution. The dataset emphasizes diversity by including 32 participants from 13 different ethnicities. Each ME is annotated with onset, apex, and offset frames, as well as corresponding facial action units (AUs). Emotion categories include contempt, disgust, fear, anger, sadness, happiness, and surprise. To induce genuine emotional responses, participant-specific video stimuli and incentive-based protocols were employed. An extended version, SAMM Long Videos [27], provides 147 long videos containing both micro- and macro-expressions, further supporting research in expression spotting across temporal scales.
  • MMEW [26]: MMEW (Macro- and Micro-Expression Warehouse) is a multi-modal database containing 300 micro- and 900 macro-expression clips, covering six basic emotions (happiness, surprise, anger, disgust, fear, sadness). Data are available in RGB, depth, and infrared formats, with naming conventions that simplify emotion-specific retrieval. It supports deep learning research due to its scale and richness.
  • Composite Dataset [28] and CMED [29]: The Composite Dataset (from MEGC2019) merges CASME II, SAMM, and SMIC-HS, unifying emotion labels into positive, negative, and surprise to standardize evaluation. The Compound Micro-Expression Dataset (CMED) aggregates MEs from CASME, CASME II, CAS(ME)2, SMIC-HS, and SAMM, categorizing expressions into basic and compound emotions, reflecting the psychological realism of naturally co-occurring affective states.
Table 1 provides a detailed comparison of existing spontaneous ME datasets. Despite their contributions, these corpora (e.g., SMIC, CASME, CASME II, CAS(ME)2, SAMM, and MMEW) still face critical limitations that hinder the external validity and scalability of MER systems. First, sample sizes are small and highly imbalanced: CAS(ME)2 contains only 57 clips, CASME II has 247, SAMM 159, and even MMEW provides just 300 micro-expression clips, with minority categories such as fear or contempt especially underrepresented. This scarcity often forces studies to focus only on categories with more than ten samples, leading to coarse-grained recognition and overfitting. Second, ecological validity is limited because most recordings were collected under controlled laboratory conditions with uniform illumination and frontal poses, far from in-the-wild scenarios with head movements, occlusions, or compression artifacts. Third, population diversity remains narrow: apart from SAMM’s effort to include multiple ethnicities, demographic coverage is restricted, and heterogeneous frame rates, resolutions, and label taxonomies complicate cross-dataset learning and fair comparison.
Looking forward, several directions are vital for the community: (i) building larger, demographically diverse, and ecologically valid corpora across naturalistic contexts (e.g., social interaction, telehealth, surveillance); (ii) standardizing annotation protocols by harmonizing AU coding, emotion taxonomies, and onset–apex–offset timestamps, while establishing cross-dataset benchmarks to assess domain shift; and (iii) leveraging synthetic augmentation (e.g., GANs, diffusion models) to mitigate class imbalance without distorting micro-expression dynamics.

2.2. Deep Learning-Based Micro-Expression Recognition Pipeline

Deep learning has significantly advanced micro-expression recognition (MER) by enabling models to automatically extract discriminative features from raw visual data. A typical deep MER pipeline consists of three key stages: advanced preprocessing, specialized network architectures, and training strategies tailored to the challenges of MER.
Preprocessing. The pipeline typically begins with face detection and alignment, using either traditional methods (e.g., Viola-Jones [30]) or modern CNN-based detectors [31]. To amplify the subtle facial movements of micro-expressions, motion magnification techniques such as Eulerian Video Magnification [32] are widely applied. Temporal normalization ensures that all video clips are resampled to a consistent length, often via the Temporal Interpolation Model (TIM) or CNN-based frame interpolation [33]. To focus on informative facial areas, researchers frequently extract task-relevant regions of interest (ROIs), such as those defined by the Facial Action Coding System (FACS) or anatomical priors [34]. Given the limited size of existing MER datasets, data augmentation is essential, extending from basic techniques (e.g., flipping, cropping) to more sophisticated strategies such as multi-scale motion magnification or GAN-based data synthesis [35,36].
Network Architectures. Early MER models relied on shallow 2D CNNs applied to key frames (e.g., apex frames) to reduce overfitting. More recent works adopt deeper architectures with pretrained backbones (e.g., VGG, ResNet), and incorporate spatiotemporal modeling via 3D CNNs or recurrent layers like LSTM and GRU [37]. Many designs adopt CNN-RNN cascades to model frame-wise evolution. Beyond these, Graph Convolutional Networks (GCNs) have been introduced to capture spatial or temporal dependencies among facial landmarks or action units [38]. Other variants incorporate capsule networks or attention modules (e.g., CBAM) to enhance sensitivity to subtle spatial changes. Transformer-based models, such as HTNet [18], divide the face into semantically meaningful regions (e.g., eyes, mouth) and use hierarchical self-attention to learn both local dynamics and global interactions. Overall, MER models span single-stream 2D/3D CNNs, CNN-RNN hybrids, multi-stream fusion networks, GCNs, and Transformers.
Training Strategies. Most MER models employ the standard softmax (cross-entropy) loss for classification, but often combine it with auxiliary objectives to address the high intra-class similarity and severe class imbalance in MER datasets. Metric learning-based losses, such as triplet loss [39] and center loss, are used to improve feature space discriminability. Focal loss [37] is frequently adopted to emphasize hard or under-represented samples. These combined objectives have been shown to improve model robustness and generalization under limited supervision.
In summary, recent deep learning-based MER systems follow a consistent yet evolving pipeline: comprehensive preprocessing to enhance subtle motions and temporal consistency, specialized architectures for fine-grained spatiotemporal representation, and training strategies designed to cope with limited and imbalanced data. These innovations collectively address the fundamental challenges of micro-expression recognition, including low motion intensity, short duration, and scarce annotations.

2.3. Multi-Modal Approaches for Micro-Expression Recognition

To better exploit the subtle and brief nature of micro-expressions (MEs), recent studies have explored multi-modal inputs that capture complementary aspects of facial dynamics. Instead of relying solely on RGB video frames, many works integrate various visual cues such as apex frames, optical flow, and facial landmarks to enhance representation quality. For example, Liu et al. [40] utilize apex frames to capture peak expression details while leveraging optical flow to model temporal motion, effectively extracting both spatial and temporal features. Expanding on this idea, Song et al. [41] introduce local facial regions from the apex frame to model the relationships among facial sub-areas, improving robustness to expression variability. Sun et al. [42] combine optical flow and full video sequences to fully mine temporal dynamics. Furthermore, inspired by the success of facial landmarks in expression analysis, Kumar et al. [43] propose fusing landmark graphs with optical flow to enhance the discriminative capacity of the network.
These multi-modal designs have led to state-of-the-art performance by maximizing the use of limited ME data. However, several limitations remain. First, the input modalities used by existing methods often fall into two extremes: either full video sequences, which are computationally expensive and may contain redundant or irrelevant frames, or single RGB frames (typically the apex frame), which may overlook important temporal cues and fail to fully exploit the visual modality, leading to suboptimal performance. Second, despite their empirical success, most current networks lack interpretability. The fusion of heterogeneous inputs is typically implemented through black-box neural modules, offering little insight into which features or modalities contribute most to the final prediction. This limits the transparency and trustworthiness of such systems, especially in sensitive applications like deception detection or clinical diagnosis.
Overall, while multi-modal approaches effectively enhance MER performance by integrating complementary visual cues, there remains a pressing need for more efficient input selection strategies and interpretability analysis that can offer both high accuracy and transparency.

3. Methods

Our method is designed to accurately recognize micro-expressions by leveraging both appearance and motion cues. As illustrated in Figure 1, the overall framework comprises four key components: a preprocessing module, RGB and optical flow feature encoders, and an output module. Each module collaboratively extracts and fuses multi-modal information.

3.1. Preprocessing Module

The preprocessing module is designed to prepare the input data for feature extraction, comprising four essential steps: face detection, alignment, apex frame spotting, and optical flow computation.
We adopt the OpenCV library with the Local Binary Features (LBF) model to perform face detection and landmark tracking across video frames. Subsequently, all frames are resized to H × W pixels, resulting in standardized inputs represented as I i R 3 × H × W .
To reduce computational redundancy while preserving the most informative motion cues, we perform apex frame spotting. Specifically, for each video clip, domain-expert annotators label three key frames: the onset frame (beginning of the expression), the apex frame (moment of peak expression), and the offset frame (end of the expression). These frames are respectively denoted as I onset , I apex , and I offset . This sparse frame selection strategy significantly reduces data volume and computational cost while retaining the critical temporal information necessary for micro-expression analysis. The selected RGB frames { I onset , I apex , I offset } are subsequently fed into the RGB feature extraction module to encode static spatial features corresponding to different temporal stages of expression development.
In parallel, to capture the subtle dynamic changes between frames, we compute optical flow between the onset and apex frames using the Dual TV-L1 algorithm [44], which is known for its robustness to illumination variation and preservation of motion discontinuities. This yields a dense motion field O that encodes pixel-wise displacements in both horizontal and vertical directions:
O = { ( u ( x , y ) , v ( x , y ) ) x = 1 , 2 , . . . , H ; y = 1 , 2 , . . . , W }
where u ( x , y ) and v ( x , y ) represent the horizontal and vertical motion vectors at pixel ( x , y ) , respectively.
To further characterize localized facial muscle deformation, we derive the optical strain map ϵ ( x , y ) from the flow field. This is achieved by computing the spatial gradients of u and v, forming the strain tensor:
ϵ = ϵ x x = u x ϵ x y = 1 2 u y + v x ϵ y x = 1 2 v x + u y ϵ y y = v y
where ϵ x x and ϵ y y denote normal strain, and ϵ x y and ϵ y x represent shear strain components.
The overall magnitude of strain at each pixel is quantified using the Frobenius norm of the strain tensor:
ϵ ( x , y ) = ϵ x x 2 + ϵ y y 2 + ϵ x y 2 + ϵ y x 2 .
Finally, the computed optical flow maps are organized as three separate single-channel feature maps: the horizontal component O u = u ( x , y ) R 1 × H × W , the vertical component O v = v ( x , y ) R 1 × H × W , and the optical strain map O ϵ = ϵ ( x , y ) R 1 × H × W . These are fed individually into the optical flow feature extraction branch for spatiotemporal representation learning.

3.2. Multi-Modal Feature Extraction

The multi-modal feature extraction stage is designed to learn complementary representations from static RGB appearance and dynamic motion features. To this end, we construct two separate encoding branches: one for RGB frames and one for optical flow maps.

3.2.1. RGB Feature Encoder

To encode the appearance information of facial expressions at different time points, we extract features from the onset, apex, and offset frames using a shared shallow convolutional network. Each frame I R 3 × H × W is passed through the following convolutional block:
F = ReLU ( BN ( Conv 7 × 7 ( I ) ) ) .
Conv RGB ( I ) = ReLU ( BN ( Conv 3 × 3 ( F ) ) ) .
This results in three intermediate feature maps: F onset , F apex , F offset R D r g b × H × W , where D r g b denotes the feature dimension.
To enhance the model’s ability to capture temporal contrast and subtle visual changes in facial appearance, we perform feature concatenation between the apex frame and its adjacent frames:
F rgb , 1 = Concat ( F onset , F apex ) .
F rgb , 2 = Concat ( F offset , F apex ) .
This design allows the network to explicitly model the relative differences in appearance between the apex frame and both the start and end of the expression. Compared to directly processing each frame individually, this strategy emphasizes the local variations that are critical for distinguishing subtle micro-expressions.
Each concatenated feature map is then passed through an independent residual block:
F rgb , i = ResBlock RGB ( F rgb , i ) , i = 1 , 2
Finally, the outputs are flattened into two 1D feature vectors:
f rgb , i = Flatten ( F rgb , i ) R d , i = 1 , 2
These two vectors jointly represent the RGB-based spatial appearance features for subsequent fusion and classification.

3.2.2. Optical Flow Feature Encoder

The optical flow encoder processes the horizontal flow map O u , vertical flow map O v , and optical strain map O ϵ , each treated as a single-channel input. For each flow map O R 1 × H × W , we first apply a convolutional layer to project it into a high-dimensional feature space:
F O ( 0 ) = ReLU ( BN ( Conv 3 × 3 ( O ) ) ) R D o f × H × W .
Then, N stacked residual blocks are applied to progressively extract deeper motion representations:
F O ( N ) = ResBlock ( N ) ( ResBlock ( 1 ) ( F O ( 0 ) ) ) .
Finally, each output feature map is flattened into a 1D vector:
f O = Flatten ( F O ( N ) ) R d .
where O { O u , O v , O ϵ } , and the resulting vectors f O u , f O v , f O ϵ form the optical flow-based motion representation.

3.3. Output Module

The output module integrates the appearance and motion representations for final classification. Specifically, the two RGB-based feature vectors f rgb , 1 and f rgb , 2 , and the three flow-based vectors f O u , f O v , f O ϵ are concatenated into a unified representation:
f concat = Concat ( f rgb , 1 , f rgb , 2 , f O u , f O v , f O ϵ ) .
The fused feature vector f concat R 5 d is then passed through two fully connected layers with ReLU activation:
h 1 = ReLU ( FC 1 ( f concat ) ) .
h 2 = ReLU ( FC 2 ( h 1 ) ) .
Finally, the classification logits are computed and normalized using the softmax function to obtain the predicted probability distribution:
y ^ = Softmax ( FC out ( h 2 ) ) R C d .
Here, C d denotes the number of emotion categories in dataset d.

4. Experiments and Results

In this section, we first describe the experimental settings, including the used datasets, evaluation metrics, baseline methods for comparison, and implementation details. We then evaluate the performance of the proposed M3ENet method across multiple ME datasets, followed by interpretability analysis and ablation studies to validate the reliability and effectiveness of the proposed framework.

4.1. Experimental Setup

4.1.1. Datasets

To demonstrate the effectiveness of the proposed framework, we conduct extensive MER experiments on five widely used public datasets: CASME, CASME II, SAMM, CAS(ME)2, and MMEW. Detailed descriptions of these datasets are provided in Section 2. An illustration of the data augmentation strategies used in our framework is shown in Figure 2.
To address the issue of limited training data, we apply data augmentation techniques, including image flipping, rotation, and additive noise. Each original sequence is augmented into five variants, effectively increasing the dataset size fivefold. These augmentations enhance data diversity, mitigate overfitting, and improve generalization. Representative examples of the augmented data are shown to illustrate the effectiveness of these transformations, as illustrated in Figure 3.
For all five datasets, we adopt the leave-one-subject-out (LOSO) evaluation protocol to ensure a fair and reliable assessment. Specifically, in each fold, the samples of one subject are used as the test set, while the remaining samples are used for training. This protocol prevents identity overlap between training and testing sets, thus avoiding subject-dependent bias and enhancing the reliability of evaluation results.

4.1.2. Evaluation Metrics

To comprehensively evaluate the performance of MER models, we report the following metrics: precision, recall, F1-score, accuracy, and the area under the ROC curve (AUC). Their definitions are given below:
Precision = T P T P + F P ,
Recall = T P T P + F N ,
F 1 - score = 2 · Precision · Recall Precision + Recall ,
Accuracy = T P + T N T P + T N + F P + F N ,
AUC = i = 1 N 1 F P R i + 1 F P R i · T P R i + 1 + T P R i 2 ,
where T P , T N , F P , and F N denote the numbers of true positives, true negatives, false positives, and false negatives, respectively. In the AUC formula, T P R i = T P i T P i + F N i and F P R i = F P i F P i + T N i represent the true positive rate and false positive rate at the i-th threshold, and N denotes the number of evaluation thresholds.

4.1.3. Baselines

To validate the effectiveness of the proposed method, we conduct comprehensive comparisons with a diverse set of baseline approaches, including generic deep learning models, ME-specific single-modal methods, and ME-specific multi-modal methods.
Generic Deep Learning Methods: These methods were originally designed for general image recognition tasks and are widely used as backbones or baselines in MER studies. Despite not being tailored for micro-expression recognition, they offer a strong benchmark for evaluating feature extraction capabilities.
  • AlexNet [45]: A classic CNN architecture consisting of five convolutional layers and three fully connected layers. It is used here as a shallow baseline for visual feature learning.
  • VGG-16 [46]: A deeper architecture with small convolution kernels (3 × 3) and a uniform design. VGG-16 is known for its strong representation ability and is commonly used for fine-grained recognition tasks.
  • ResNet-18 [47]: Incorporates residual connections to mitigate the vanishing gradient problem in deep networks. Its lightweight version (ResNet-18) is suitable for small-scale datasets like MER.
  • GoogLeNet (Inception v1) [48]: Utilizes multi-scale convolution within inception modules, allowing more expressive features with fewer parameters.
Single-Modal ME-Specific Methods: These approaches are explicitly designed for MER and typically leverage temporal modeling techniques (e.g., LSTM or 3D CNNs) or incorporate domain-specific priors to better capture the subtle and transient nature of micro-expressions. They usually focus on a single modality, such as RGB or optical flow.
  • Off-ApexNet [49]: A dual-stream CNN model that captures subtle facial motion by computing optical flow between the onset and apex frames. The network separately processes the horizontal and vertical components of flow to learn discriminative motion features. It is one of the earliest deep learning frameworks tailored for MER tasks.
  • STSTNet [17]: A lightweight three-stream 3D CNN that processes horizontal flow, vertical flow, and optical strain simultaneously. By employing shallow temporal convolution, it effectively models the short-term dynamics of micro-expressions. The architecture is highly efficient and well suited for small-scale MER datasets.
  • HTNet [18]: A hierarchical Transformer designed for micro-expression recognition. It divides the face into four regions and uses local self-attention to capture subtle muscle movements, while an aggregation layer models interactions between eye and lip areas.
  • LAENet [50]: A local self-attention encoding network that focuses on critical facial regions with subtle muscle movements. It incorporates spatial attention to selectively enhance local features relevant to micro-expression.
  • SSRLTS-ViT [51]: SSRLTS-ViT is a three-stream Vision Transformer baseline for micro-expression recognition that learns from three optical flow components to capture subtle facial motion with global context.
Multi-Modal ME-Specific Methods: These methods extend single-modal approaches by combining multiple modalities (e.g., RGB, optical flow) to improve discrimination. They often use fusion strategies or cross-modal learning to exploit inter-modal relationships.
  • CNNCapsNet [40]: This method integrates five input streams, including vertical and horizontal optical flow between onset–apex and apex–offset, as well as the apex grayscale image. A multi-stream CNN is used for feature extraction, followed by a Capsule Network for final classification.

4.1.4. Implementation Details

The experiments were conducted on an Ubuntu 20.10 system equipped with 4 Intel Xeon Platinum 8375C CPUs (Intel Corporation, Santa Clara, CA, USA), a single NVIDIA GeForce RTX 4090 GPU (NVIDIA Corporation, Santa Clara, CA, USA), and 256 GB of RAM. The model was trained for 50 epochs with a batch size of 128 using the AdamW optimizer, a cosine annealing learning rate schedule, and a weight decay of 0.05. The initial learning rate was set to 0.001, and the momentum coefficient β was fixed at 0.999. The input images were uniformly resized to 48 × 48 .
To address the issue of class imbalance, where certain expression categories (e.g., “disgust”) are heavily underrepresented, we employ the focal loss [52], which dynamically adjusts the contribution of each sample to the loss based on its classification difficulty.
The focal loss is formulated as
L F L = α t ( 1 p t ) γ log ( p t )
where p t is the model’s predicted probability for the true class label, α t ( 0 , 1 ) is a class-specific weighting factor to handle imbalance, and γ 0 is the focusing parameter that down-weights well-classified examples (i.e., those with high p t ) and focuses learning on hard, misclassified samples.
When γ = 0 , the focal loss reduces to standard cross-entropy. As γ increases, the relative loss assigned to easier examples decreases, thus encouraging the model to pay more attention to underrepresented or ambiguous samples. This property makes focal loss particularly suitable for micro-expression datasets where some emotion categories may be sparsely distributed.
In our experiments, we set γ = 2 to effectively focus learning on hard samples. The class-wise weighting factor α t is computed based on the inverse frequency of each category as
α t = 1 n t N ,
where n t denotes the number of samples in class t, ensuring higher weights for underrepresented categories.

4.2. Performance Evaluation

4.2.1. Overall Performance

Table 2 demonstrates that M3ENet consistently outperforms all baseline methods across all five datasets, achieving state-of-the-art (SOTA) performance. Specifically, M3ENet attains an average accuracy of 81.5%, an F1-score of 0.736, and an AUC of 0.953. These results highlight the superiority of our network architecture in effectively extracting and integrating features from both RGB and optical flow modalities, leading to significant improvements over single-modal methods. In contrast, existing multi-modal approaches often suffer from suboptimal modality selection and inadequate fusion strategies, resulting in limited performance gains. By carefully designing and aligning complementary representations across modalities, M3ENet enables more accurate and robust micro-expression recognition. Additionally, the confusion matrices of M3ENet on all five datasets (Figure 4) provide detailed insights into the model’s performance across different micro-expression categories.

4.2.2. Interpretability Analysis

This interpretability is of paramount importance for enhancing the trustworthiness and user acceptance of the model, particularly in applications requiring high precision and reliability, such as mental health assessments and interpersonal interaction analysis. In our study, we employed Grad-Cam [53] visualization techniques to elucidate the decision-making processes of our micro-expression recognition network. The visualizations depicted in Figure 5 demonstrate the model’s capability to accurately identify regions of the face that are crucial for micro-expression recognition, often corresponding to action units (AUs). The Grad-Cam visualizations not only validate the model’s precision in recognizing micro-expressions but also provide transparency into the model’s decision-making process, thereby enhancing its interpretability.

4.2.3. Ablation Analysis

To validate the effectiveness of the proposed approach, we conduct ablation studies by altering the network architecture and training strategies. Specifically, the training strategy variant involves removing data augmentation (w/o Data augmentation), while architectural variants include removing the RGB input stream (w/o RGB input), removing the optical flow input stream (w/o Optical flow input), using only a single RGB image as input (Single RGB input), and using the shuffled RGB image as input (Shuffled RGB input).
The ablation results in Table 3 clearly demonstrate the importance of each component in our framework. First, removing data augmentation (w/o Data augmentation) leads to a dramatic drop in all evaluation metrics, confirming the critical role of data diversity in preventing overfitting and enhancing generalization. Second, when comparing the two input modalities, removing the optical flow stream (w/o Optical flow input) causes a more severe performance degradation than removing the RGB stream (w/o RGB input). This highlights the indispensable contribution of motion information in capturing subtle facial dynamics for micro-expression recognition. Furthermore, the results of the Single RGB input variant indicate that a single static frame provides only limited discriminative cues, while the poor performance with Shuffled RGB input demonstrates that it is the texture information within RGB frames, rather than the RGB modality itself, that effectively guides the model’s decisions together, these findings validate that both spatial and temporal information, along with effective training strategies, are indispensable for the superior performance of our proposed M3ENet.

4.2.4. Efficiency and Complexity Analysis

By extracting key frames from the video stream, we substantially reduce the dimensionality of the network input, which simplifies deployment in practical systems. To quantify computational budget, we report three standard metrics: floating-point operations (FLOPs), parameter count, and per-frame inference latency. As summarized in Table 4, Transformer-based baselines (e.g., SSRLTS-ViT and HTNet) exhibit prohibitive latency and computational cost, which undermines real-time deployment. Conventional unimodal models are lighter, yet their accuracy is often insufficient for production use. Our method, M3ENet, attains a more favorable accuracy–efficiency trade-off: by leveraging multi-modal feature fusion, it markedly improves performance while maintaining a lightweight computational footprint, facilitating deployment in resource-constrained, real-time settings.

5. Conclusions

In this paper, we propose a novel network architecture for micro-expression recognition using multi-modal inputs, which is applicable to a variety of real-world scenarios such as visitor experience analysis in museum environments. We demonstrate its superior performance on five publicly available datasets (CASME I, CASME II, CAS(ME)2, SAMM, and MMEW). Specifically, the network takes optical flow data and RGB images from the onset, apex, and offset frames as inputs, with dedicated encoders designed for each modality to extract complementary features. Data augmentation and focal loss are employed to address data scarcity and class imbalance, significantly improving recognition accuracy across datasets. Experimental results show that, thanks to the multi-modal input design and the carefully crafted architecture, M3ENet significantly outperforms both conventional general-purpose models and existing single-modal and multi-modal MER methods. Furthermore, we employ Grad-CAM to visualize the model’s attention, demonstrating the interpretability and rationality of its decision-making. Comprehensive ablation studies further validate the importance of each proposed component to the overall model performance.

Author Contributions

K.Z. designed the study framework, supervised the research, and drafted the manuscript. X.L. was responsible for network architecture design and conducted numerical analysis experiments. G.Y. selected the public datasets and supervised the experimental evaluation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Ministry of Education of China Industry-University Cooperative Education Program (231006551161619), the Sichuan Science and Technology Program (24RKX0776), the AI Technology-Enabled Undergraduate Teaching Reform Projects at the University of Electronic Science and Technology of China (UESTC) (2024AIXM100), the Sichuan Provincial Science Popularization Base Assessment Project (2023JDKPJD0036), and the Graduate Education and Teaching Reform Program at UESTC (JYJG2024059).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Benchmark datasets (CASME I, CASME II, CAS(ME)2, SAMM, and MMEW) are publicly available from the respective providers (see Section 4.1). Source code is available at https://github.com/xuanyv/M3ENet (accessed on 1 September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, K.; Zhang, S. Research on Museum Visitor Experience Based on Micro-Expression Recognition. Int. J. Psychophysiol. (IOP) 2021, 168, S147. [Google Scholar] [CrossRef]
  2. Ekman, P.; Friesen, W.V. Constants across cultures in the face and emotion. J. Personal. Soc. Psychol. 1971, 17, 124. [Google Scholar] [CrossRef]
  3. Ekman, P. Lie catching and microexpressions. Philos. Decept. 2009, 1, 5. [Google Scholar]
  4. Zhao, G.; Pietikainen, M. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 915–928. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, Y.; See, J.; Phan, R.C.W.; Oh, Y.H. Lbp with six intersection points: Reducing redundant information in lbp-top for micro-expression recognition. In Proceedings of the Asian Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 525–537. [Google Scholar]
  6. Wang, Y.; See, J.; Phan, R.C.W.; Oh, Y.H. Efficient spatio-temporal local binary patterns for spontaneous facial micro-expression recognition. PLoS ONE 2015, 10, e0124674. [Google Scholar] [CrossRef]
  7. Huang, X.; Zhao, G.; Hong, X.; Zheng, W.; Pietikäinen, M. Spontaneous facial micro-expression analysis using spatiotemporal completed local quantized patterns. Neurocomputing 2016, 175, 564–578. [Google Scholar] [CrossRef]
  8. Liong, S.T.; Phan, R.C.W.; See, J.; Oh, Y.H.; Wong, K. Optical strain based recognition of subtle emotions. In Proceedings of the 2014 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Kuching, Malaysia, 1–4 December 2014; pp. 180–184. [Google Scholar]
  9. Liong, S.T.; See, J.; Phan, R.C.W.; Le Ngo, A.C.; Oh, Y.H.; Wong, K. Subtle expression recognition using optical strain weighted features. In Proceedings of the Asian Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 644–657. [Google Scholar]
  10. Happy, S.; Routray, A. Fuzzy histogram of optical flow orientations for micro-expression recognition. IEEE Trans. Affect. Comput. 2017, 10, 394–406. [Google Scholar] [CrossRef]
  11. Liu, Y.J.; Zhang, J.K.; Yan, W.J.; Wang, S.J.; Zhao, G.; Fu, X. A main directional mean optical flow feature for spontaneous micro-expression recognition. IEEE Trans. Affect. Comput. 2015, 7, 299–310. [Google Scholar] [CrossRef]
  12. Liong, S.T.; See, J.; Wong, K.; Phan, R.C.W. Less is more: Micro-expression recognition from video using apex frame. Signal Process. Image Commun. 2018, 62, 82–92. [Google Scholar] [CrossRef]
  13. Patel, D.; Hong, X.; Zhao, G. Selective deep features for micro-expression recognition. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2258–2263. [Google Scholar]
  14. Peng, M.; Wang, C.; Chen, T.; Liu, G.; Fu, X. Dual temporal scale convolutional neural network for micro-expression recognition. Front. Psychol. 2017, 8, 1745. [Google Scholar] [CrossRef]
  15. Bai, M.; Goecke, R. Investigating LSTM for micro-expression recognition. In Proceedings of the Companion Publication of the 2020 International Conference on Multimodal Interaction, Utrecht, The Netherlands, 25–29 October 2020; pp. 7–11. [Google Scholar]
  16. Zhou, Z.; Zhao, G.; Pietikäinen, M. Towards a practical lipreading system. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 137–144. [Google Scholar]
  17. Liong, S.; See, J.; Wong, K.; Phan, R. Shallow Triple Stream Three-dimensional CNN (STSTNet) for Micro-expression Recognition. In Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition (FG), Lille, France, 14–18 May 2019. [Google Scholar]
  18. Wang, Z.; Zhang, K.; Luo, W.; Sankaranarayana, R. Htnet for micro-expression recognition. Neurocomputing 2024, 602, 128196. [Google Scholar] [CrossRef]
  19. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  20. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; Technical Report; University of Toronto, Department of Computer Science: Toronto, ON, Canada, 2009. [Google Scholar]
  21. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement, collection and baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar]
  22. Yan, W.J.; Wu, Q.; Liu, Y.J.; Wang, S.J.; Fu, X. CASME database: A dataset of spontaneous micro-expressions collected from neutralized faces. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–7. [Google Scholar]
  23. Yan, W.J.; Li, X.; Wang, S.J.; Zhao, G.; Liu, Y.J.; Chen, Y.H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef]
  24. Qu, F.; Wang, S.J.; Yan, W.J.; Li, H.; Wu, S.; Fu, X. CAS (ME)2: A database for spontaneous macro-expression and micro-expression spotting and recognition. IEEE Trans. Affect. Comput. 2017, 9, 424–436. [Google Scholar] [CrossRef]
  25. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. Samm: A spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 2016, 9, 116–129. [Google Scholar] [CrossRef]
  26. Ben, X.; Ren, Y.; Zhang, J.; Wang, S.J.; Kpalma, K.; Meng, W.; Liu, Y.J. Video-based facial micro-expression analysis: A survey of datasets, features and algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5826–5846. [Google Scholar] [CrossRef]
  27. Yap, C.H.; Kendrick, C.; Yap, M.H. Samm long videos: A spontaneous facial micro-and macro-expressions dataset. In Proceedings of the 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 16–20 November 2020; pp. 771–776. [Google Scholar]
  28. See, J.; Yap, M.H.; Li, J.; Hong, X.; Wang, S.J. Megc 2019—The second facial micro-expressions grand challenge. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–5. [Google Scholar]
  29. Zhao, Y.; Xu, J. A convolutional neural network for compound micro-expression recognition. Sensors 2019, 19, 5553. [Google Scholar] [CrossRef]
  30. Viola, P.; Jones, M.J. Robust real-time face detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
  31. Matsugu, M.; Mori, K.; Mitari, Y.; Kaneda, Y. Subject independent facial expression recognition with robust face detection using a convolutional neural network. Neural Netw. 2003, 16, 555–559. [Google Scholar] [CrossRef] [PubMed]
  32. Wu, H.Y.; Rubinstein, M.; Shih, E.; Guttag, J.; Durand, F.; Freeman, W. Eulerian video magnification for revealing subtle changes in the world. ACM Trans. Graph. (TOG) 2012, 31, 1–8. [Google Scholar] [CrossRef]
  33. Niklaus, S.; Liu, F. Context-aware synthesis for video frame interpolation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1701–1710. [Google Scholar]
  34. Li, X.; Hong, X.; Moilanen, A.; Huang, X.; Pfister, T.; Zhao, G.; Pietikäinen, M. Towards reading hidden emotions: A comparative study of spontaneous micro-expression spotting and recognition methods. IEEE Trans. Affect. Comput. 2017, 9, 563–577. [Google Scholar] [CrossRef]
  35. Xie, H.X.; Lo, L.; Shuai, H.H.; Cheng, W.H. Au-assisted graph attention convolutional network for micro-expression recognition. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2871–2880. [Google Scholar]
  36. Yu, J.; Zhang, C.; Song, Y.; Cai, W. ICE-GAN: Identity-aware and capsule-enhanced GAN with graph-based reasoning for micro-expression recognition and synthesis. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  37. Xia, Z.; Hong, X.; Gao, X.; Feng, X.; Zhao, G. Spatiotemporal recurrent convolutional networks for recognizing spontaneous micro-expressions. IEEE Trans. Multimed. 2019, 22, 626–640. [Google Scholar] [CrossRef]
  38. Lei, L.; Li, J.; Chen, T.; Li, S. A novel graph-tcn with a graph structured representation for micro-expression recognition. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2237–2245. [Google Scholar]
  39. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  40. Liu, N.; Liu, X.; Zhang, Z.; Xu, X.; Chen, T. Offset or onset frame: A multi-stream convolutional neural network with capsulenet module for micro-expression recognition. In Proceedings of the 2020 5th International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), Okinawa, Japan, 18–20 November 2020; pp. 236–240. [Google Scholar]
  41. Xie, Z.; Zhao, C. Dual-branch cross-attention network for micro-expression recognition with transformer variants. Electronics 2024, 13, 461. [Google Scholar] [CrossRef]
  42. Wang, S.; Guan, S.; Lin, H.; Huang, J.; Long, F.; Yao, J. Micro-expression recognition based on optical flow and pcanet+. Sensors 2022, 22, 4296. [Google Scholar] [CrossRef]
  43. Kumar, A.J.R.; Bhanu, B. Micro-expression classification based on landmark relations with graph attention convolutional network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 1511–1520. [Google Scholar]
  44. Zach, C.; Pock, T.; Bischof, H. A duality based approach for realtime tv-l 1 optical flow. In Proceedings of the Joint Pattern Recognition Symposium; Springer: Berlin/Heidelberg, Germany, 2007; pp. 214–223. [Google Scholar]
  45. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 25, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  46. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  47. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  48. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  49. Gan, Y.S.; Liong, S.T.; Yau, W.C.; Huang, Y.C.; Tan, L.K. OFF-ApexNet on micro-expression recognition system. Signal Process. Image Commun. 2019, 74, 129–139. [Google Scholar] [CrossRef]
  50. Gan, Y.S.; Lien, S.E.; Chiang, Y.C.; Liong, S.T. LAENet for micro-expression recognition. Vis. Comput. 2024, 40, 585–599. [Google Scholar] [CrossRef]
  51. Zhang, H.; Yin, L.; Zhang, H.; Wu, X. Facial micro-expression recognition using three-stream vision transformer network with sparse sampling and relabeling. Signal Image Video Process. 2024, 18, 3761–3771. [Google Scholar] [CrossRef]
  52. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  53. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. Illustration of the proposed five-stream architecture for micro-expression analysis. The network consists of a preprocessing module, an RGB feature encoder, an optical flow feature encoder, and an output module.
Figure 1. Illustration of the proposed five-stream architecture for micro-expression analysis. The network consists of a preprocessing module, an RGB feature encoder, an optical flow feature encoder, and an output module.
Sensors 25 06276 g001
Figure 2. Illustration of representative micro-expression samples from multiple datasets used in this work.
Figure 2. Illustration of representative micro-expression samples from multiple datasets used in this work.
Sensors 25 06276 g002
Figure 3. Illustration of the proposed data augmentation strategy with random rotation angles and noise levels across samples.
Figure 3. Illustration of the proposed data augmentation strategy with random rotation angles and noise levels across samples.
Sensors 25 06276 g003
Figure 4. Confusion matrices of model predictions on the five ME datasets.
Figure 4. Confusion matrices of model predictions on the five ME datasets.
Sensors 25 06276 g004
Figure 5. Illustration of Grad-CAM visualizations highlighting model attention on key facial regions across micro-expression types and datasets.
Figure 5. Illustration of Grad-CAM visualizations highlighting model attention on key facial regions across micro-expression types and datasets.
Sensors 25 06276 g005
Table 1. Summary of previous spontaneous ME datasets.
Table 1. Summary of previous spontaneous ME datasets.
DatasetSMIC
 [21]
CASME
 [22]
CASME II
 [23]
CAS(ME)2
 [24]
SAMM
 [25]
SAMM long
 [25]
MMEW
 [26]
Subjects16/8/8353522323236
ME samples164/71/7119524757159159300
Resolution640 × 480640 × 480640 × 480640 × 4802040 × 10882040 × 10881920 × 1080
Facial size190 × 230150 × 90250 × 340-400 × 400-400 × 400
Frame Rate100/25/2560603020020090
Expressions3854777
EnvironLabLabLabLabLabLabLab
Ethnicity311113131
AU
Apex
Note: ✓ indicates that the dataset contains the corresponding data, while ✗ indicates that the dataset does not include the corresponding data.
Table 2. Performance comparison between M3ENet and baselines on five datasets. Bold indicates the best, and underline indicates the second best.
Table 2. Performance comparison between M3ENet and baselines on five datasets. Bold indicates the best, and underline indicates the second best.
DatasetMetricANV16R18GNOANSSNHNLNViTCCNM3ENet
CASME IPrecision0.3380.1510.1770.1230.3400.6400.0510.3790.8090.2510.780
Recall0.3250.1420.1780.1490.3490.4170.1430.2630.6290.2860.746
F1_score0.3230.0920.1560.0890.3380.4460.0750.2580.6910.2650.760
Accuracy0.5990.3460.3770.3660.5790.6230.3560.5410.7260.5860.856
AUC0.8370.3230.5260.6220.8430.8930.1120.6940.9530.7460.974
CASME IIPrecision0.3490.3450.2880.1300.4900.5070.0780.5460.7370.3550.718
Recall0.3360.2000.2720.1750.4050.4120.1670.3080.5300.3230.715
F1_score0.3340.1730.2270.1250.4260.4380.1060.2980.5840.3270.715
Accuracy0.5670.4860.3720.4680.5710.6330.4680.5730.6860.5230.803
AUC0.8070.6530.6980.7130.8260.8810.2590.8100.9130.8160.948
CAS(ME)2Precision0.1600.0760.0890.0370.1620.2970.0370.2040.4050.0900.621
Recall0.1580.1230.1310.1250.1840.3200.1250.1810.3710.1360.650
F1_score0.1260.0840.0690.0570.1670.3030.0570.1600.3620.1000.631
Accuracy0.3190.2780.2360.2960.3430.4310.2960.3520.4950.2870.731
AUC0.6550.4330.5980.3880.6900.7500.1220.6160.8150.6120.921
SAMMPrecision0.3810.2230.4590.0510.5210.4880.0510.2640.7120.3060.757
Recall0.2950.1500.4030.1420.2900.3520.1430.2010.4720.2790.740
F1_score0.2880.0920.4160.0750.2930.3710.0750.1730.5220.2770.747
Accuracy0.5020.3600.5480.3540.4970.5610.3570.4280.6290.4670.806
AUC0.7480.3760.7770.5400.7690.8080.1220.7330.8630.7400.944
MMEWPrecision0.5300.4350.7350.3070.7330.6920.2450.6780.7340.6030.822
Recall0.5180.4090.7070.3580.5340.5850.1460.4950.6560.4870.839
F1_score0.5130.3910.7150.3230.5540.6090.0730.5080.6820.4950.829
Accuracy0.6880.6080.7830.5690.7430.7300.3020.6830.7580.6900.878
AUC0.8970.8580.9390.8190.9430.9290.3870.8990.9410.9090.976
AveragePrecision0.3520.2460.3500.1300.4490.5250.0920.4140.6800.3210.740
Recall0.3270.2050.3380.1900.3520.4170.1450.2900.5320.3020.738
F1_score0.3170.1660.3170.1340.3550.4330.0770.2800.5680.2940.736
Accuracy0.5350.4150.4630.4110.5460.5950.3560.5150.6590.5100.815
AUC0.7890.5280.7080.6160.8140.8520.2000.7500.8970.7650.953
Abbreviations: AN = AlexNet [45], V16 = VGG-16 [46], R18 = ResNet-18 [47], GN = GoogLeNet [48], OAN = OFF-ApexNet [49], SSN = STSTNet [17], HN = HTNet [18], LN = LAENet [50], ViT = SSRLTS-ViT [51], CCN = CNNCapsNet [40].
Table 3. Ablation results of M3ENet under different settings. Bold indicates the best, and underline indicates the second best.
Table 3. Ablation results of M3ENet under different settings. Bold indicates the best, and underline indicates the second best.
Ablation VariantPrecisionRecallF1_ScoreAccuracyAUC
Standard0.7400.7380.7360.8150.953
w/o Data augmentation0.3060.2670.2680.4370.637
w/o Optical flow input0.6400.6220.6250.7050.896
w/o RGB input0.7420.7150.7230.7760.939
Single RGB input0.7310.7260.7220.7930.943
Shuffled RGB input0.6210.5710.5840.6780.891
Table 4. Efficiency and complexity across baselines. Latency is per frame in milliseconds (ms).
Table 4. Efficiency and complexity across baselines. Latency is per frame in milliseconds (ms).
ModelParameters (M)FLOPs (M)Latency (ms)
M3ENet6.0142.875.08
SSRLTS-ViT66.42643.0522.94
ResNet-1811.1798.443.23
VGG-1633.61723.581.45
GoogLeNet5.9776.405.47
EfficientNet7.980.163.46
HTNet139.65215.4513.59
CNNCapsNet12.51477.102.47
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, K.; Liu, X.; Yang, G. M3ENet: A Multi-Modal Fusion Network for Efficient Micro-Expression Recognition. Sensors 2025, 25, 6276. https://doi.org/10.3390/s25206276

AMA Style

Zhao K, Liu X, Yang G. M3ENet: A Multi-Modal Fusion Network for Efficient Micro-Expression Recognition. Sensors. 2025; 25(20):6276. https://doi.org/10.3390/s25206276

Chicago/Turabian Style

Zhao, Ke, Xuanyu Liu, and Guangqian Yang. 2025. "M3ENet: A Multi-Modal Fusion Network for Efficient Micro-Expression Recognition" Sensors 25, no. 20: 6276. https://doi.org/10.3390/s25206276

APA Style

Zhao, K., Liu, X., & Yang, G. (2025). M3ENet: A Multi-Modal Fusion Network for Efficient Micro-Expression Recognition. Sensors, 25(20), 6276. https://doi.org/10.3390/s25206276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop