Skip Content
You are currently on the new version of our website. Access the old version .
EntropyEntropy
  • Article
  • Open Access

29 January 2026

Cross-Dataset Facial Micro-Expression Recognition with Regularization Learning and Action Unit-Guided Data Augmentation

,
,
,
and
1
Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
Tech X Academy, Shenzhen Polytechnic University, Shenzhen 518055, China
3
School of Computer Science and Artificial Intelligence, Guangdong University of Education, Guangzhou 510303, China
4
Shenzhen Lower Limb Intelligent Rehabilitation Engineering Research Center, Shenzhen 518055, China

Abstract

With the growing development of facial micro-expression recognition technology, its practical application value has attracted increasing attention. In real-world scenarios, facial micro-expression recognition typically involves cross-dataset evaluation, where training and testing samples come from different datasets. Specifically, cross-dataset micro-expression recognition employs multi-dataset composite training and unseen single-dataset testing. This setup introduces two major challenges: inconsistent feature distributions across training sets and data imbalance. To address the distribution discrepancy of the same category across different training datasets, we propose a plug-and-play batch regularization learning module that constrains weight discrepancies across datasets through information-theoretic regularization, facilitating the learning of domain-invariant representations while preventing overfitting to specific source domains. To mitigate the data imbalance issue, we propose an Action Unit (AU)-guided generative adversarial network (GAN) for synthesizing micro-expression samples. This approach uses K-means clustering to obtain cluster centers of AU intensities for each category, which are then used to guide the GAN in generating balanced micro-expression samples. To validate the effectiveness of the proposed methods, extensive experiments are conducted on CNN, ResNet, and PoolFormer architectures. The results demonstrate that our approach achieves superior performance in cross-dataset recognition compared to state-of-the-art methods.

1. Introduction

Micro-expressions refer to involuntary facial movements that reveal concealed emotions, characterized by their brief duration and low intensity. Due to their spontaneous nature, micro-expressions serve as a critical cue for inferring true emotions [1,2,3]. Since their introduction, they have been extensively studied across disciplines including psychology and sociology. However, accurately detecting micro-expressions with the naked eye is challenging for non-experts, which has motivated the development of computer-based automated micro-expression analysis [4,5,6]. As a key technology in human affective computing, micro-expression recognition holds significant promise in critical fields such as suicide prevention [7], criminal interrogation [8], and national security [9].
The advancement of micro-expression recognition research relies fundamentally on the support of well-annotated datasets. Despite the considerable challenges given the spontaneous and subtle nature of micro-expressions, persistent research efforts have yielded several valuable open-source datasets, including SMIC [10], CASME [11], CASME II [12], SAMM [13], and MMEW [14]. These datasets provide an essential foundation for model training and evaluation. However, existing datasets still face critical limitations: the total number of publicly available samples is relatively small (less than 2000 in total), which restricts model complexity and generalizability, and significant class imbalance across emotion categories further complicates the development of robust recognition. These issues pose challenges to building reliable micro-expression recognition models.
Although many effective micro-expression recognition models have been developed using existing datasets, the prevailing methodology remains largely dependent on an intra-dataset evaluation protocol, wherein models are trained and tested on data from the same source [5,8]. This approach ensures highly consistent feature distributions between training and testing splits, which facilitates learning but also causes models to overfit to dataset-specific statistical biases. In real-world applications, however, models are required to generalize to data collected under different conditions with divergent quality and feature distributions. Under such circumstances, conventional models developed under the intra-dataset paradigm typically suffer a significant performance degradation, revealing a critical lack of generalization capability [15].
To address the generalization challenge in micro-expression recognition, the Micro-Expression Grand Challenge 2018 (MEGC2018) and 2019 (MEGC2019) introduced two protocols: Composite Dataset Evaluation (CDE) [16] that merges multiple datasets for training and testing to learn general features, and Holdout Dataset Evaluation (HDE) [17] that trains on one dataset and tests on another to simulate cross-dataset scenarios. For CDE, Liu et al. integrate optical flow motion analysis, structural feature pooling, and cross-domain knowledge transfer to align feature distributions across source datasets [18]. For HDE, Zong et al. propose a Target Sample Re-Generator on the CASME II and SMIC datasets. This method generates target samples with a feature distribution similar to the source domain to minimize distribution discrepancies [19]. Although these methods have achieved promising results, as shown in the left part of Figure 1, existing evaluation protocols suffer from notable limitations in assessing model generalizability. The CDE protocol, despite its ability to learn distribution discrepancies across multiple source datasets, does not evaluate models on truly unseen data. Meanwhile, conventional HDE setups typically utilize only two datasets, which prevents models from being exposed to sufficiently diverse sample distributions. Consequently, the generalizability of models trained under these paradigms is often limited [20].
Figure 1. Schematic illustration of different micro-expression recognition evaluation protocols.
To overcome these limitations and better simulate real-world conditions, we propose a comprehensive cross-dataset micro-expression recognition framework based on a leave-one-dataset-out strategy across five datasets. For instance, Datasets 2–5 are combined for training, while Dataset 1 is held out for testing. As shown in the right part of Figure 1, this approach maximizes data utilization while ensuring rigorous evaluation on completely unseen data distributions, thereby closely mimicking real-world scenarios where models are required to generalize to data from entirely new sources.
While this framework significantly enhances evaluation rigor, it simultaneously amplifies inherent challenges in cross-dataset recognition. The first challenge stems from feature distribution inconsistency. Disparities in acquisition hardware, illumination environments, and subject populations create substantial distribution shifts across datasets. When multiple datasets are merged for training, these discrepancies not only intensify internal feature variations within the training set but also magnify the distribution gap with the held-out test set. The second challenge is dataset imbalance. Each dataset exhibits unevenness in both quantity and class distribution, and combining multiple datasets further intensifies this imbalance. This composite imbalance may lead the model to over-rely on datasets or classes with larger sample sizes during training, thereby reducing its generalization capability across different data sources. We provide a detailed visual analysis of these cross-dataset challenges in Section 2.
To address feature distribution inconsistency in cross-dataset recognition, we propose a distribution-balanced batch regularization learning (BRL) approach. Implemented as a specialized loss component through a self-attention mechanism, this method establishes dataset grouping constraints in the feature space and minimizes inter-group attention weight differences to force balanced feature learning from all source domains. The BRL module acts as an information flow regularizer, minimizing the entropy disparity in cross-domain attention distribution and encouraging the extraction of domain-invariant features with higher mutual information to emotion labels. Experiments demonstrate that BRL module effectively prevents overfitting to individual datasets and significantly improves cross-dataset generalization. To address data imbalance, we propose a data augmentation method based on Action Unit (AU) intensity clustering. By analyzing the AU intensity distribution of the same emotion category across different datasets, we extract representative cluster centroids as average AU weights to guide the ULME-GAN network [21] in generating semantically consistent samples that align with the real data distribution. By effectively expanding samples, this strategy enhances the information diversity of the training set while preserving the essential statistical properties and semantic plausibility of micro-expressions. Comprehensive experiments using three mainstream networks (CNN, ResNet, and PoolFormer) on five spontaneous datasets (SMIC, CASME, CASME II, SAMM, and MMEW) demonstrate substantial performance improvements of our approach over state-of-the-art methods under the rigorous cross-dataset evaluation protocol.
The main contributions of this study are summarized as follows:
(1) A rigorous cross-dataset evaluation protocol is established for micro-expression recognition, systematically addressing key challenges through targeted solutions while achieving state-of-the-art performance under this realistic setting.
(2) A batch regularization learning strategy is proposed to address feature distribution inconsistency across datasets, implemented as a loss term that explicitly balances the model’s attention to different datasets at the representation learning level.
(3) An AU-guided data generation strategy is developed, which not only effectively mitigates class imbalance but also maintains the authenticity of data distributions, contributing to improved model performance.
The rest of this paper is organized as follows: Section 2 reviews related work. Section 3 provides a systematic analysis of cross-dataset challenges through visualizations and elaborates the proposed targeted solutions. Section 4 presents the experimental results and discussions. Section 5 concludes this study and suggests future research directions.

3. Analysis of Cross-Dataset Challenges and Proposed Methods

As illustrated in Figure 1, this study investigates a more challenging cross-dataset recognition paradigm that integrates characteristics of both CDE and HDE protocols. We begin by conducting comprehensive analysis to identify and characterize two critical challenges in this rigorous evaluation protocol: feature distribution inconsistency across source domains and inherent data imbalance. Building on these analysis, we subsequently present a targeted solution framework incorporating Batch Group Regularization for domain-invariant representation learning and AU-guided data augmentation for dataset rebalancing.

3.1. Analysis of Cross-Dataset Challenges

The adoption of CDE and HDE evaluation protocols has improved model generalization to some extent. However, in the context of rigorous cross-dataset recognition involving large-scale multi-dataset scenarios that better reflect real-world conditions, this paradigm amplifies two inherent challenges: feature distribution inconsistency and data imbalance.

3.1.1. Feature Distribution Inconsistency

Feature distribution differences are a core challenge in cross-dataset micro-expression recognition. These differences come from variations in how datasets are collected, including the capture devices, experimental conditions, and participant groups. From an information-theoretic perspective, these variations introduce domain-specific biases that increase the uncertainty (or entropy) of the composite feature space, making it difficult for models to extract stable, discriminative signals for emotion recognition.
(1) Data Acquisition Differences
As summarized in Table 1, the five benchmark datasets employed in this study exhibit significant differences in acquisition specifications. Camera frame rates range from 60 fps to 200 fps, while facial region resolutions vary from 150 × 190 to 400 × 400 pixels. These technical discrepancies directly introduce heterogeneity in both spatiotemporal information density and image detail across datasets. This heterogeneity can be viewed as differing informational content per sample across domains. For instance, a higher frame rate (e.g., 200 fps) provides a denser temporal signal, potentially carrying more information about the dynamics of a micro-expression, whereas variations in resolution affect the spatial information entropy of the facial details.
Table 1. Comparison of the five spontaneous micro-expression datasets used in this study.
Table 1. Comparison of the five spontaneous micro-expression datasets used in this study.
DatasetsFrame Rate (fps)Facial ResolutionSample Size Per CategoryTotal
PositiveNegativeSurprise
SMIC [10]100190 × 230517043164
CASME [11]60150 × 1909532183
CASME II [12]200280 × 340327225129
SAMM [13]200400 × 400268015121
MMEW [14]90400 × 4003610989234
(2) Experimental Environment and Ethnicity
Besides equipment parameters, differences in experimental environments introduce systematic biases in image characteristics across datasets. These environmental factors include lighting conditions, experimental settings, and participant ethnicity. As shown in Figure 2, “happiness” micro-expression samples from five different datasets exhibit noticeable differences in illumination characteristics, lighting intensity, and skin tone. These visual variations are unrelated to micro-expression muscle movements, yet deep neural networks may learn them as shortcut features during training. Such shortcut learning interferes with the acquisition of generalizable micro-expression features for cross-dataset transfer, ultimately limiting performance in practical applications.
Figure 2. Sample images of the “happiness” category from different datasets.
(3) Evidence from Feature Visualization
To evaluate feature distribution inconsistency in cross-dataset scenarios, we train a convolutional neural network using all “happiness” category samples from these five datasets and visualize their distributions in the feature space. Specifically, we project the high-dimensional feature representations into a two-dimensional space for visualization using t-SNE [36]. As shown in Figure 3, the results clearly demonstrate that samples from the same dataset form tight clusters, while clear distribution gaps exist between different datasets. This finding indicates that as more datasets are combined for training, the feature distribution differences across source domains become more pronounced. In this cross-dataset paradigm, models struggle to learn unified semantic representations, and the intensified distribution inconsistency leads to isolated feature subspaces, thereby limiting generalization performance on unseen datasets.
Figure 3. Visualization of the feature space distribution for “happiness” samples from different datasets.
The visualization in Figure 3 also reveals a critical information structure problem. The strong intra-dataset clustering and inter-dataset separation imply that the learned features are highly predictive of their source dataset domain, sharing substantial mutual information with the dataset identity, but are less informative about the micro-expression emotion category. Ideally, for robust cross-dataset recognition, the features should be highly informative of the emotion label while being invariant to the dataset source. This misalignment increases the uncertainty in predicting the emotion from the features, making accurate classification across domains challenging. Therefore, our proposed method aims to rectify this by guiding the model to learn features that are more discriminative for emotions and less dependent on the dataset origin.

3.1.2. Data Imbalance

Beyond feature distribution differences, the imbalance in data size across datasets presents another important challenge. This issue mainly appears in two aspects: overall data quantity differences and uneven distribution at the category level.
(1) Overall Data Quantity Differences
As shown in Table 1, the five datasets comprise a total of 731 samples, but their distribution is highly uneven. For example, the MMEW dataset contains 234 samples, while CASME has only 83 samples. This quantity difference causes models to over-rely on larger datasets during composite training, leading to overfitting. Consequently, the model performs well on certain datasets but degrades significantly on others. This can be interpreted as the model allocating disproportionate attention entropy or computational resources to domains with higher sample counts, essentially forming an inefficient information acquisition strategy that fails to maximize knowledge from all available sources.
(2) Class Distribution Imbalance
A more serious problem is the imbalance at the class level. The number of samples from different datasets varies greatly across emotion categories. For instance, in the “Positive” category, SMIC has 51 samples while CASME has only 9 samples; in the “Surprise” category, MMEW has 89 samples while SAMM has only 15 samples. This class imbalance results in certain categories being predominantly represented by limited datasets during training. Consequently, learned features reflect characteristics of these dominant datasets rather than essential micro-expression attributes, which significantly undermines the model generalization across datasets.

3.2. Proposed Methods

Figure 4 presents the overall architecture of the proposed cross-dataset micro-expression recognition framework. The method employs five spontaneous micro-expression datasets in a leave-one-dataset-out protocol: four datasets form a composite training set, while the remaining unseen dataset serves as the test set to evaluate model generalization. Taking input apex frames as an example, the framework first performs AU detection and clustering analysis to construct category-specific AU intensity matrices. These matrices subsequently guide the ULME-GAN [21] in generating augmented micro-expression samples, as illustrated in the blue dashed box of Figure 4 and detailed in Section 3.2.2. Both original and augmented samples are processed through a backbone network for feature extraction. During training, a batch regularization learning module optimizes the process by computing a regularization loss via self-attention mechanism. This BRL loss is combined with cross-entropy loss through weighted summation to form the final optimization objective. The trained model is ultimately evaluated on the completely unseen test dataset to assess cross-dataset recognition performance.
Figure 4. The overall architecture of the proposed method.

3.2.1. Batch Regularization Learning (BRL)

In cross-dataset micro-expression recognition, models trained on composite datasets often develop biased feature representations due to distribution discrepancies among source domains. This bias manifests as overfitting to certain datasets while underfitting others, ultimately compromising generalization performance. From an information-theoretic perspective, this bias represents an inefficient allocation of the model’s information acquisition capacity—it extracts excessive information from some domains while neglecting others, leading to suboptimal information integration across sources. To address this challenge, we propose the BRL module that explicitly balances the model’s attention distribution across different datasets during training, ensuring a more equitable information flow from all source domains.
The BRL module consists of two key components: a self-attention weighting mechanism and group-wise regularization learning. Figure 5 illustrates the computational workflow of the proposed BRL module.
Figure 5. The computational workflow of the BRL module.
First, the self-attention mechanism [37] computes sample-wise importance weights based on the extracted features. Let x i j R D denote the feature vector of the j-th sample from the i-th dataset, where D is the feature dimension. This feature vector is projected through a fully connesxcted layer to obtain a self-attention weight:
a i j = σ W x i j + b
where a i j is the computed self-attention weight, W and b denote the weight matrix and bias term of the fully connected layer, and σ is the sigmoid activation function. The magnitude of a i j indicates the sample’s relative contribution to model learning, where higher values correspond to greater influence. Conceptually, a i j quantifies the information gain the model attributes to each sample during the learning process.
The regularization component utilizes these attention weights to promote balanced learning across different source datasets. During training, samples in each batch are grouped according to their source dataset, and the average attention weight m i is computed for each group as follows:
m i = 1 n i j = 1 n i a i j , i 1 , 2 , 3 , 4
where n i denotes the number of samples from the i-th dataset in the current batch, and m i represents the mean attention weight for that dataset. Here, m i serves as a proxy for the average information contribution from domain i to the current learning step. A balanced set of m i values indicates that the model is extracting knowledge uniformly from all available datasets, thereby encouraging the learning of domain-invariant representations. Conversely, the imbalanced set of m i values suggest that the model’s learned features are likely biased toward specific domains, containing dataset-specific information that hinders cross-dataset generalization.
The BRL module functions through a loss term L B R L , which is defined as the range between the maximum and minimum of these dataset-wise average attention weights:
L B R L = H L
where H = max ( m 1 , m 2 , m 3 , m 4 ) and L = min ( m 1 , m 2 , m 3 , m 4 ) represent the highest and lowest average attention weights among the four dataset groups, respectively. By minimizing this range-based loss, the module explicitly penalizes attention bias to any particular dataset, thereby enforcing more balanced feature learning across all source domains and encouraging the discovery of domain-invariant representations. Minimizing H L is equivalent to reducing the entropy (or variance) in the distribution of information contributions { m i } across domains. This regularization encourages the model to become a fair information processor that extracts knowledge uniformly from all sources. In cross-dataset micro-expression recognition, balanced attention compels the model to focus on shared emotion-related semantics across datasets, thereby avoiding overfitting to non-transferable, dataset-specific artifacts such as illumination, resolution, or demographic variations.
The overall optimization objective integrates the proposed BRL loss L B R L with the standard multi-class cross-entropy loss L C E to jointly optimize classification performance and cross-dataset generalization. The complete loss function L t o t a l combines both objectives through setting weights:
L t o t a l = ( 1 λ ) L C E + λ L B R L
where λ [ 0 , 1 ) is a balancing coefficient that controls the relative contribution of each loss component. This formulation ensures that the cross-entropy loss L C E always maintains a non-zero weight to preserve classification capability, while the batch regularization loss L B R L provides adjustable constraint strength for feature distribution alignment. This dual-objective framework can be interpreted as simultaneously minimizing the prediction uncertainty (via cross-entropy) and regularizing the information acquisition strategy (via BRL) to promote domain invariance.
Through the optimization of this composite objective, the model simultaneously maintains discriminative power for accurate classification while learning to allocate attention resources equitably across all source domains. This dual optimization strategy promotes the learning of domain-invariant feature representations that capture essential micro-expression characteristics while minimizing dataset-specific biases, ultimately enhancing generalization capability in cross-dataset scenarios. The resulting features exhibit higher mutual information with the true emotion labels and lower dependence on dataset identifiers, aligning with the information-theoretic principle of learning maximally informative yet minimally domain-specific representations.
Designed as a plug-and-play component, the BRL module operates at the mini-batch level, enabling application to diverse cross-domain learning tasks beyond micro-expression recognition. Its batch-wise processing mechanism allows seamless integration into standard deep learning pipelines without requiring modifications to the host network architecture. This modular design establishes a portable framework for enhancing model robustness across various cross-dataset applications. More broadly, it offers a principled, information-theoretic approach to mitigating source bias in any multi-source learning scenario by explicitly regularizing the information flow during training.

3.2.2. AU-Guided Micro-Expression Augmentation

As shown in Table 1, the five spontaneous micro-expression datasets employed in this study exhibit significant data imbalance, which can bias the model to majority classes and impair its generalization capability. To address this issue while ensuring the authenticity of synthesized samples, we propose an AU-guided data augmentation strategy that generates semantically consistent micro-expression samples aligned with the real data distribution.
The proposed method builds upon the ULME-GAN framework [21], which leverages a combination of adversarial loss, attention loss, conditional expression loss, and identity loss to ensure the generation of realistic and identity-consistent micro-expression sequences. Our approach extends this framework by incorporating AU intensity clustering to guide the generation process.
Specifically, for each emotion category in the composite training set, we first extract AU intensity values from all available samples using OpenFace [38]. The intensity vector for the p-th sample is defined as:
AU p = [ I AU 1 ( p ) , I AU 2 ( p ) , , I AU 45 ( p ) ] , p [ 1 , M ]
where I AU k ( p ) denotes the intensity value of the AUk for sample p, with k representing the specific AU identifier (1, 2, 4, 5, 6, 7, 9, 10, 12, 14, 15, 17, 20, 23, 25, 26, and 45), and M represents the total number of samples in the current emotion category. The complete AU intensity set for the category is then:
A = { AU 1 , AU 2 , , AU M }
To obtain a representative AU pattern for each emotion category, we apply K-means clustering to the AU intensity set A and identify the optimal cluster centroid. The centroid vector AU c is computed as the mean of all AU intensity vectors in the dominant cluster, representing the prototypical AU activation pattern for that emotion category. This centroid serves as the conditional input to ULME-GAN, guiding the generation of synthetic samples that embody the characteristic AU configuration of the category while maintaining semantic consistency with real micro-expressions.
The generation process follows the adversarial training framework, where the generator G learns to transform neutral face images into expression-carrying faces conditioned on the centroid vector AU c . The training objective incorporates multiple loss components: adversarial loss ( L adv ) to ensure visual realism, attention loss ( L att ) to preserve spatial coherence, conditional expression loss ( L exp ) to maintain expression fidelity, and identity preservation loss ( L id ) to retain subject characteristics.
To establish a balanced dataset across all emotion categories, we use the MMEW dataset as the reference due to its largest sample size of 234 samples. Each dataset undergoes augmentation to reach this target, with generated samples distributed evenly across all emotion categories. Table 2 provides the detailed numbers of augmented samples required for each dataset. It should be noted that the MMEW dataset serves as the benchmark in this process and therefore does not require additional augmentation.
Table 2. Number of augmented samples per category for each dataset.
This AU-guided augmentation strategy offers significant advantages by directly utilizing the clustered AU intensity vectors as the standard for data generation. This approach ensures that the synthesized samples accurately reflect the most representative and prototypical facial muscle movement patterns for each emotion category, effectively capturing the essential characteristics of genuine micro-expressions. By employing these statistically meaningful AU patterns derived from real data distributions, our method not only effectively mitigates data imbalance across datasets but also maintains high semantic authenticity in the generated samples. Consequently, the augmentation process enhances the model’s capacity to learn robust and domain-invariant feature representations that generalize well across all source domains.

3.2.3. Backbone Networks

To validate the effectiveness of our proposed methods independently of network architecture complexity, we employ three representative backbones: CNN, ResNet, and PoolFormer. This selection ensures that performance improvements stem from our proposed BRL and AU-guided augmentation approaches rather than sophisticated network designs. As shown in Figure 6, these widely-used backbone networks provide diverse architectural paradigms for micro-expression recognition.
Figure 6. The backbone network architectures of the (a) CNN, (b) ResNet, and (c) PoolFormer.
(1) CNN
The CNN backbone, illustrated in Figure 6a, consists of a sequence of convolutional and pooling layers for hierarchical feature extraction. Input images are processed using convolutional kernels of sizes 5 × 5 and 3 × 3 , each followed by a 2 × 2 pooling layer to progressively reduce spatial resolution and enhance translational invariance. The extracted features are flattened and passed through a fully connected layer to produce predictions across the three target emotion categories. This simple yet effective architecture serves as a baseline to evaluate the generalizability of our proposed methods.
(2) ResNet
The ResNet backbone is originally proposed to address the issues of vanishing gradients and network degradation in very deep convolutional networks [39]. As illustrated in Figure 6b, we employ ResNet-18 in this study, a lightweight variant of the architecture. The core residual building block utilizes skip connections that bypass one or more convolutional layers, enabling a direct, element-wise addition between the input and the transformed feature. This design establishes a residual learning framework that effectively alleviates the vanishing gradient problem in deep networks and has demonstrated strong performance across various recognition tasks.
(3) PoolFormer
PoolFormer represents a transformer-type architecture that substitutes the standard self-attention mechanism with spatial pooling operations [40]. This design significantly lowers computational cost while maintaining competitive performance. As illustrated in Figure 6c, we adopt the PoolFormer-S12 configuration in this study, which stacks 12 layers of identical PoolFormer blocks. Each block applies a 3 × 3 average pooling layer to mix spatial information, followed by channel-wise MLP modules and residual connections. The network progressively reduces spatial resolution while increasing channel dimensions across four stages. This efficient architecture provides a practical balance between accuracy and speed, making it suitable for micro-expression recognition tasks where computational efficiency and cross-dataset generalization are both important.

4. Experimental Results and Analysis

In this section, we present a comprehensive evaluation of the proposed method to validate its effectiveness. We begin by describing the experimental setup, including descriptions of the datasets employed, the data preprocessing pipeline, and the implementation details. Subsequently, the evaluation metrics used for quantitative analysis are defined. A thorough analysis of the results is then provided, encompassing an ablation study to dissect the contribution of each proposed component, followed by a comparative analysis with the state-of-the-art methods.

4.1. Datasets

This study employs five publicly available spontaneous micro-expression datasets (SMIC [10], CASME [11], CASME II [12], SAMM [13], and MMEW [14]) to establish a comprehensive benchmark for cross-dataset evaluation. Except for SMIC, all other datasets provide apex frame annotations and corresponding AU annotations. For SMIC, we followed established research conventions by using the frame with the largest detected motion magnitude as the apex frame. Each micro-expression dataset contains 3 to 8 emotion categories. To maintain categorical consistency across datasets with different original emotion labels, all samples are mapped into three emotion categories: Positive, Negative, and Surprise.
As illustrated in Figure 1 and detailed in Table 3, the cross-dataset evaluation follows a leave-one-dataset-out protocol. For each test round, one dataset is held out as the test set, while the remaining four are combined to form a composite training set. This process is repeated five times, ensuring each dataset serves as the test set. It is important to note that all augmented samples are used exclusively for training and do not participate in the testing phase. The original sample sizes for each test dataset are provided in Table 1.
Table 3. Comparison of training sample sizes with and without data augmentation for cross-dataset evaluation.
Table 3 details the sample composition of the composite training sets corresponding to each testing scenario, comparing the original and augmented sample sizes. The data augmentation strategy effectively balances the training distribution, with each composite set reaching 936 samples after augmentation, significantly enhancing the model’s exposure to diverse sample distributions.

4.2. Data Preprocessing and Implementation Details

All samples from these five micro-expression datasets undergo the same data preprocessing procedure, with facial regions cropped to 224 × 224 pixels. In this study, we employ two different input modalities to comprehensively evaluate model performance: RGB apex frames and optical flow images. The RGB apex frames capture spatial texture features at the peak expression intensity, while the optical flow images, computed between onset and apex frames using the Recurrent All-Pairs Field Transforms (RAFT) algorithm [41], characterize temporal motion patterns of micro-expressions. Since the AU-guided augmentation strategy specifically operates on single-frame AU features, data augmentation is exclusively applied to RGB apex frame inputs.
All experiments are implemented using the PyTorch framework (Version: 2.3.1) and evaluated on three backbone networks: CNN, ResNet18, and PoolFormer-S12. The models are trained for 100 epochs with an initial learning rate of 0.0002 and weight decay of 5 × 10 4 . To ensure stable optimization, the BRL loss is incorporated after the 50th epoch, and the best performance after this point is recorded. To systematically evaluate the effect of the proposed BRL module, we compare five different weighting coefficients λ = [ 0 , 0.3 , 0.5 , 0.7 , 0.9 ] in the total loss function L total = ( 1 λ ) L C E + λ L B R L , where λ = 0 serves as the baseline without BRL regularization. All hyperparameters are tuned to achieve optimal performance across different experimental settings.

4.3. Evaluation Metrics

To comprehensively evaluate model performance in cross-dataset micro-expression recognition, we employ accuracy (Acc) alongside two additional metrics that account for class imbalance: unweighted average recall (UAR) and unweighted F1-score (UF1). These metrics are defined as follows:
Acc = c = 1 C T P c N UAR = 1 C c = 1 C T P c N c UF 1 = 1 C c = 1 C 2 × T P c 2 × T P c + F P c + F N c
where T P c , F P c , and F N c represent true positives, false positives, and false negatives for class c, respectively; N c denotes the total number of samples in class c; N is the total number of samples; and C is the number of classes. While Acc provides an overall performance measure, UAR and UF1 offer more balanced assessments by giving equal weight to each emotion category, thus mitigating the bias to majority classes that commonly exists in imbalanced micro-expression datasets.

4.4. Ablation Experiments

4.4.1. Ablation of λ

An ablation study is conducted to analyze the influence of the weighting coefficient λ in the combined loss function L total = ( 1 λ ) L CE + λ L BRL . The experiment uses apex frame inputs on the CNN backbone and is evaluated on the CASME II dataset. The parameter λ is tested with values of [ 0 , 0.3 , 0.5 , 0.7 , 0.9 ] , where larger values indicate greater emphasis on the batch regularization loss component.
As shown in Table 4, the optimal performance is achieved at λ = 0.5 , where the model attains 56.59% Acc, 43.07% UAR, and 41.89% UF1. Compared to the baseline without BRL module ( λ = 0 ), this configuration yields improvements of 8.53% in Acc, 2.73% in UAR, and 2.70% in UF1. This balanced weighting allows the model to maintain strong classification capability while effectively utilizing the dataset-balancing regularization provided by the BRL module. When λ decreases to 0.3, the regularization effect is insufficient to adequately address feature distribution discrepancies across source domains, resulting in suboptimal performance. Conversely, when λ increases to 0.7 or 0.9, the excessive emphasis on dataset alignment compromises the model’s discriminative power for micro-expression classification, leading to significant performance degradation, with λ = 0.9 performing even worse than the baseline case.
Table 4. Performance comparison with different λ values on CASME II dataset using apex frames and CNN backbone.
The performance trend across different λ values—rising to an optimum at λ = 0.5 and then declining sharply at λ = 0.7 and 0.9 —reveals a critical trade-off in domain generalization. A low λ provides insufficient regularization to align features across domains, limiting generalization, while a high λ overly suppresses discriminative power for emotion classification, causing performance to fall below the baseline. These results confirm that an appropriate balance between classification learning and domain-invariant feature learning is essential for effective cross-dataset micro-expression recognition. The λ = 0.5 configuration achieves this optimal trade-off, and its equal weighting (1:1) between the cross-entropy loss and the BRL loss is adopted in all subsequent experiments.

4.4.2. Analysis with Apex Frame Inputs

This section presents the experimental results under the apex frame input setting, evaluating cross-dataset recognition performance across different datasets, backbone architectures, and the effect of the BRL module. Table 5 summarizes the comprehensive comparison, where bold values indicate the best performance for each dataset under each evaluation metric.
Table 5. Classification results of apex frame inputs across datasets with different backbones and BRL configurations. Bold values indicate the best performance for each metric within each dataset.
At the dataset level, CASME achieves the highest recognition Acc among all test configurations, reaching 69.88% with both CNN+BRL and ResNet+BRL. This superior performance can be attributed to the exclusion of CASME during training, which results in the largest composite training set among all test scenarios due to CASME’s relatively small sample size. In contrast, SMIC consistently exhibits the lowest recognition accuracy across all configurations, a trend that persists throughout the subsequent experimental results. This performance gap stems from two primary factors. First, SMIC has a relatively small training sample size, which limits the amount of representative data available for model optimization. Second, and more critically, SMIC does not provide explicit apex frame annotations. The use of detected apex frames—defined as the frame with the largest motion magnitude within the sequence—introduces potential temporal misalignment, particularly affecting optical flow features that rely on precise onset–apex timing. These limitations collectively hinder the model’s ability to learn robust and well-aligned features from this dataset.
Regarding backbone architectures, CNN demonstrates the most substantial improvements when integrated with the BRL module, with average performance increases of 6.95% in Acc, 6.62% in UAR, and 8.70% in UF1. Both CNN and ResNet significantly outperform PoolFormer across most evaluation metrics. Overall, the CNN architecture combined with the BRL module delivers the best performance across most datasets (except CASME), demonstrating its superior capability in handling cross-dataset micro-expression recognition tasks when using apex frames as input.
The incorporation of the BRL module effectively enhances model performance across all backbone networks. The most notable improvement is observed in the CNN architecture, where the addition of BRL boosts the average accuracy from 52.28% to 59.23%. This demonstrates the effectiveness of the proposed batch regularization learning in addressing feature distribution discrepancies across datasets, particularly for architectures with moderate complexity that are well-suited for the scale of available micro-expression data.

4.4.3. Analysis with Optical Flow Inputs

This section evaluates the effectiveness of temporal motion features for cross-dataset micro-expression recognition, using optical flow sequences computed between onset and apex frames via the RAFT algorithm [41]. Table 6 presents the comprehensive results, with bold values indicating the best performance for each dataset under each evaluation metric.
Table 6. Classification results of optical flow inputs across datasets with different backbones and BRL configurations. Bold values indicate the best performance for each metric within each dataset.
Compared to the apex frame inputs analyzed in Section 4.4.2, optical flow inputs significantly outperform apex frame inputs (Table 5) across all evaluation metrics and backbone architectures. This performance advantage demonstrates that temporal motion patterns capture more robust and transferable characteristics for cross-dataset recognition compared to spatial appearance features from single frames. Among the architectures, the PoolFormer architecture exhibits the most substantial improvement with optical flow inputs, achieving performance increases of over 10% across respective metrics. The incorporation of the BRL module further enhances model performance with optical flow inputs across all backbone networks. The CNN architecture achieves the most consistent improvements, with performance gains of 3.76% in Acc, 4.26% in UAR, and 3.59% in UF1 after BRL integration. While the PoolFormer+BRL configuration delivers particularly competitive results on SMIC and SAMM datasets, CNN+BRL achieves superior performance in the remaining cases.
These findings collectively demonstrate that optical flow features provide more discriminative temporal representations than single apex frames for cross-dataset micro-expression recognition. The superior generalizability of optical flow stems from its inherent robustness to cross-domain variations. Unlike appearance features from single apex frames, which are sensitive to illumination and camera differences, optical flow encodes relative motion, thereby normalizing dataset-specific biases. Micro-expressions are transient motion events, and optical flow directly captures the dynamics of facial muscle activations over time, which are more consistent across data collection setups. Consequently, models learn transferable motion patterns rather than appearance artifacts, leading to stronger generalization to unseen domains. Moreover, the BRL module effectively enhances model performance with different input types and architectural designs, confirming its robustness and general applicability across varied experimental conditions.

4.4.4. Ablation of Data Augmentation

This section evaluates the proposed AU-guided data augmentation strategy, which generates synthetic samples based on clustered AU intensity patterns for each micro-expression category. As the augmentation operates on single-frame AU characteristics, we validate its effectiveness using apex frame inputs across the three backbone architectures with BRL module integration. Table 7 presents the results with data augmentation, where bold values indicate the best performance for each dataset under each evaluation metric.
Table 7. Classification results using apex frames as input with BRL module and AU-guided data augmentation on different backbones. Bold values indicate the best performance for each metric within each dataset.
When comparing with the non-augmented results in Table 5, the integration of AU-guided data augmentation yields significant performance gains, with particularly notable improvements in UAR and UF1 metrics. This trend appears across all three backbone architectures. The CNN+BRL configuration with augmentation demonstrates remarkable gains on the MMEW dataset, achieving 6.40% higher UAR and 11.02% better UF1 compared to the non-augmented case. At the architecture level, ResNet+BRL+Augmentation shows the most substantial average improvements with 3.95% higher UAR and 5.28% better UF1, followed by PoolFormer and CNN. The pronounced enhancement in UAR and UF1, which are metrics specifically designed to evaluate performance on imbalanced datasets, confirms that the augmentation strategy effectively mitigates inter-dataset sample quantity disparity by generating representative samples that better capture characteristic AU patterns for each emotion category.
These results demonstrate that the AU-guided data augmentation effectively addresses dataset imbalance in cross-dataset micro-expression recognition. The method generates semantically consistent samples by leveraging statistically derived AU intensity centroids from real data distributions. This approach maintains computational efficiency through single-pass generation via the ULME-GAN network while specifically targeting class imbalance at the AU level. By preserving essential facial action characteristics of each emotion category, the augmentation diversifies training data while ensuring feature authenticity.

4.4.5. Ablation Comparison and Visualization

To comprehensively evaluate the overall effectiveness of the proposed modules, we analyze the average cross-dataset recognition performance across all five datasets using the CNN backbone, which has demonstrated superior performance in previous experiments. Table 8 presents the ablation results of different module combinations, revealing that both the BRL module and data augmentation contribute significantly to performance improvement.
Table 8. Cross-dataset micro-expression recognition results (averaged over five datasets) with CNN backbone under different configurations.
The results in Table 8 indicate a clear performance hierarchy among different configurations. The baseline model using apex frames without any proposed modules achieves 52.28% Acc, 40.13% UAR, and 36.46% UF1. Incorporating the BRL module brings substantial improvements, increasing these metrics to 59.23%, 46.75%, and 45.16% respectively. The addition of data augmentation further enhances performance, particularly on UAR and UF1, which rise to 49.10% and 47.83%. Notably, optical flow features outperform apex frames across all configurations, with the combination of optical flow inputs and BRL module achieving the best overall performance at 63.50% Acc, 53.63% UAR, and 53.07% UF1.
We further investigate the impact of BRL through feature visualization. For each trained model, test samples are processed to extract features, which are then reduced to 2D space using Principal Component Analysis (PCA). A logistic regression classifier is trained on these reduced-dimension features to identify and plot decision boundaries. Figure 7 presents the visualization results under four different configurations.
Figure 7. Visualization using logistic regression for features extracted by a CNN trained on four datasets and tested on CASME II. The red circles highlight the areas where samples are clustered. (a) Classification using apex frames; (b) Classification using apex frames with BRL Loss; (c) Classification using optical flow features; (d) Classification using optical flow features with BRL module.
The visualization reveals distinct patterns in feature learning. For apex frame inputs (Figure 7a), the feature distribution appears chaotic without BRL, making it difficult to establish clear decision boundaries. Although BRL integration (Figure 7b) brings moderate improvement in feature clustering, significant inter-class mixing persists, indicating the limited discriminative capacity of spatial features alone in cross-dataset scenarios. In contrast, optical flow inputs (Figure 7c,d) demonstrate better inherent separability, with the BRL module further enhancing class discrimination. Particularly in Figure 7d, the feature clusters become more compact and well-separated, as highlighted by the red circles, indicating that the combination of optical flow features and BRL regularization effectively learns discriminative and dataset-invariant representations. This advantage stems from the fundamental characteristic that optical flow features directly encode temporal dynamics of facial muscle movements, which captures the essential nature of micro-expressions more effectively than static texture features from apex frames. Furthermore, motion patterns exhibit greater robustness to inter-dataset appearance variations such as illumination and subject demographics.
These experimental results demonstrate that the proposed methods effectively address feature distribution inconsistency and data imbalance in cross-dataset micro-expression recognition, with optical flow features providing superior temporal representations and the BRL module enhancing feature discriminability across diverse dataset distributions.

4.5. Comparison with Other Methods

Due to the relatively limited research on cross-dataset micro-expression recognition and the inconsistent use of evaluation datasets across existing studies, we compare our method with available state-of-the-art approaches under the same tested dataset. Since most existing methods primarily report Acc and UAR with one decimal place precision, Table 9 presents our best results rounded to one decimal place for comparison. Bold values indicate the best performance for each metric across different datasets.
As shown in Table 9, our method achieves superior performance across most datasets and evaluation metrics. On the CASME II dataset, our approach attains 73.6% Acc and 68.9% UAR, surpassing the best existing methods by 7.4% in Acc compared to DR [34] and by 4.6% in UAR relative to ATNet [31]. Similarly, on the SAMM dataset, our method reaches 67.8% Acc and 53.2% UAR, representing an 11.9% improvement in Acc over RNMA [42] and a 7.4% gain in UAR compared to ATNet [31]. However, on the SMIC dataset, our method does not achieve the best performance, obtaining 51.2% accuracy and 46.6% UAR. This performance limitation can be attributed to the absence of precise apex frame annotations in SMIC, where optical flow features are extracted using the middle frame of the sequence rather than accurately identified apex frames, potentially introducing temporal misalignment that compromises feature quality.
Table 9. Comparison with state-of-the-art methods (Acc and UAR). Bold values indicate the best results.
Table 9. Comparison with state-of-the-art methods (Acc and UAR). Bold values indicate the best results.
MethodSMICCASME IISAMM
AccUARAccUARAccUAR
LBP-TOP [17]--23.231.633.832.7
3DHOG [17]--37.318.735.326.9
HOOF [17]--26.534.644.434.9
DR [34]54.954.766.249.6--
D3DCNN [43]--44.7-36.9-
TFMVN [44]--45.536.7--
ELRCN [45]--38.432.248.538.2
RN [26]--57.833.754.444.0
RNMA [42]--58.434.155.942.7
RSTR [33]45.1-56.2---
ATNet [31]-52.3-64.3-45.8
Ours51.246.673.668.967.853.2
These results demonstrate that the proposed BRL module and AU-guided data augmentation effectively address the core challenges in cross-dataset micro-expression recognition: feature distribution inconsistency and data imbalance. Notably, these improvements are achieved through targeted methodological innovations rather than increased architectural complexity. The performance superiority across datasets confirms that our approach learns discriminative and dataset-invariant feature representations, demonstrating both effectiveness and strong generalizability for micro-expression recognition.

5. Conclusions and Discussion

Cross-dataset micro-expression recognition presents a critical challenge for real-world applications, where models must maintain performance when encountering data from previously unseen sources. To address this challenge systematically, this study establishes a rigorous evaluation paradigm using five publicly available spontaneous micro-expression datasets under a leave-one-dataset-out protocol. This framework reveals two fundamental difficulties in cross-dataset scenarios: feature distribution inconsistency across source domains and inherent dataset imbalance. We propose targeted solutions for each challenge and demonstrate state-of-the-art performance across multiple evaluation metrics.
Through comprehensive experimental analysis and visualization, we identify that feature distribution shifts across datasets cause models to learn dataset-specific biases rather than generalizable micro-expression characteristics. Additionally, the natural imbalance in sample quantities among different datasets introduces training biases that further degrade cross-dataset generalization capability. These interrelated issues significantly impact model performance in practical cross-dataset applications. The core of the problem lies in the model’s inefficient information acquisition strategy. It tends to overfit to high-entropy but domain-specific signals, while underutilizing the essential, low-entropy stable emotional cues that are consistent across domains.
To address feature distribution inconsistency, we propose BRL as a plug-and-play learning strategy that enhances model generalization by explicitly balancing its attention across multiple source domains. This approach adaptively adjusts feature importance during training to encourage domain-invariant representation learning. The BRL module functions as an information flow regularizer, minimizing the entropy in attention distribution across domains and promoting a more equitable extraction of information from all available sources. The modular design enables integration into various backbone architectures and facilitates straightforward transfer to similar cross-domain recognition tasks. For the dataset imbalance problem, we introduce an AU-guided data augmentation strategy that generates semantically consistent samples based on clustered AU intensity patterns. Importantly, these solutions achieve significant performance improvements on three conventional backbone networks (CNN, ResNet, and PoolFormer), demonstrating that our methods effectively address the core challenges without relying on architectural complexity.
Despite the promising results, several limitations remain. Performance variations across different test datasets indicate that more universal micro-expression representations need to be developed. Future work will focus on advanced temporal feature learning techniques and domain generalization methods to create more robust cross-dataset recognition systems. The exploration of self-supervised learning paradigms and the integration of physiological prior knowledge also present promising directions for enhancing cross-dataset micro-expression recognition.

Author Contributions

Data curation, H.X.; Formal analysis, J.Z.; Investigation, J.Z. and X.L.; Methodology, J.Z. and L.W.; Visualization, H.X.; Writing—original draft, J.Z. and X.L.; Writing—review & editing, T.W. All authors have read and agreed to the published version of the manuscript.

Funding

We gratefully acknowledge the support from the Postdoctoral Research Startup Fund Project of Shenzhen Polytechnic University (6024331018K). This work is also purported in part by the National Key Research and Development Program of China (2024YFE0216500), the National Natural Science Foundation of China (62573402), the Shenzhen Strategic Emerging Industry Support Plans (XMHT20230115002), and the Shenzhen Sustainable Development Sci-Tech project (KCXFZ20230731093501003).

Data Availability Statement

Data sharing is not applicable.

Acknowledgments

We are grateful to the researchers who built the public micro-expression datasets, as their work laid the essential foundation for this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pfister, T.; Li, X.; Zhao, G.; Pietikäinen, M. Recognising spontaneous facial micro-expressions. In Proceedings of the 2011 International Conference On Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1449–1456. [Google Scholar] [CrossRef]
  2. Li, X.; Hong, X.; Moilanen, A.; Huang, X.; Pfister, T.; Zhao, G.; Pietikäinen, M. Towards reading hidden emotions: A comparative study of spontaneous micro-expression spotting and recognition methods. IEEE Trans. Affect. Comput. 2017, 9, 563–577. [Google Scholar] [CrossRef]
  3. Malik, P.; Singh, J.; Ali, F.; Sehra, S.S.; Kwak, D. Action unit based micro-expression recognition framework for driver emotional state detection. Sci. Rep. 2025, 15, 27824. [Google Scholar] [CrossRef]
  4. Shuai, T.; Beng, S.; Khalid, F.B.; Rahmat, R.W.B.O. Advances in Facial Micro-Expression Detection and Recognition: A Comprehensive Review. Information 2025, 16, 876. [Google Scholar] [CrossRef]
  5. Li, Y.; Wei, J.; Liu, Y.; Kauttonen, J.; Zhao, G. Deep learning for micro-expression recognition: A survey. IEEE Trans. Affect. Comput. 2022, 13, 2028–2046. [Google Scholar] [CrossRef]
  6. Yang, H.; Xie, L.; Pan, H.; Li, C.; Wang, Z.; Zhong, J. Multimodal Attention Dynamic Fusion Network for Facial Micro-Expression Recognition. Entropy 2023, 25, 1246. [Google Scholar] [CrossRef]
  7. Gu, S.; Sun, X.; Chen, B.; Tao, W. Depression Micro-Expression Recognition Technology Based on Multimodal Knowledge Graphs. Trait. Signal 2024, 41, 2047. [Google Scholar] [CrossRef]
  8. Xie, H.X.; Lo, L.; Shuai, H.H.; Cheng, W.H. An overview of facial micro-expression analysis: Data, methodology and challenge. IEEE Trans. Affect. Comput. 2022, 14, 1857–1875. [Google Scholar] [CrossRef]
  9. Maryala, S.; Kommula, V.; Kumarapu, S.; A, K. Micro-expression Analysis for Security and Law Enforcment using Deep Learning. In Proceedings of the 2025 International Conference on Emerging Smart Computing and Informatics (ESCI), Pune, India, 5–7 March 2025; pp. 1–6. [Google Scholar] [CrossRef]
  10. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement, collection and baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar] [CrossRef]
  11. Yan, W.J.; Wu, Q.; Liu, Y.J.; Wang, S.J.; Fu, X. CASME database: A dataset of spontaneous micro-expressions collected from neutralized faces. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–7. [Google Scholar] [CrossRef]
  12. Yan, W.J.; Li, X.; Wang, S.J.; Zhao, G.; Liu, Y.J.; Chen, Y.H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef] [PubMed]
  13. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. SAMM: A spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 2016, 9, 116–129. [Google Scholar] [CrossRef]
  14. Ben, X.; Ren, Y.; Zhang, J.; Wang, S.J.; Kpalma, K.; Meng, W.; Liu, Y.J. Video-based facial micro-expression analysis: A survey of datasets, features and algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5826–5846. [Google Scholar] [CrossRef]
  15. Zhou, L.; Mao, Q.; Xue, L. Cross-database micro-expression recognition: A style aggregated and attention transfer approach. In Proceedings of the 2019 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Shanghai, China, 8–12 July 2019; pp. 102–107. [Google Scholar] [CrossRef]
  16. See, J.; Yap, M.H.; Li, J.; Hong, X.; Wang, S.J. Megc 2019—The second facial micro-expressions grand challenge. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–5. [Google Scholar] [CrossRef]
  17. Yap, M.H.; See, J.; Hong, X.; Wang, S.J. Facial micro-expressions grand challenge 2018 summary. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 675–678. [Google Scholar] [CrossRef]
  18. Liu, Y.; Du, H.; Zheng, L.; Gedeon, T. A neural micro-expression recognizer. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–4. [Google Scholar] [CrossRef]
  19. Zong, Y.; Huang, X.; Zheng, W.; Cui, Z.; Zhao, G. Learning a target sample re-generator for cross-database micro-expression recognition. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 872–880. [Google Scholar] [CrossRef]
  20. Varanka, T.; Li, Y.; Peng, W.; Zhao, G. Data leakage and evaluation issues in micro-expression analysis. IEEE Trans. Affect. Comput. 2023, 15, 186–197. [Google Scholar] [CrossRef]
  21. Zhou, J.; Sun, S.; Xia, H.; Liu, X.; Wang, H.; Chen, T. ULME-GAN: A generative adversarial network for micro-expression sequence generation. Appl. Intell. 2024, 54, 490–502. [Google Scholar] [CrossRef]
  22. Verma, M.; Vipparthi, S.K.; Singh, G. Deep insights of learning-based micro expression recognition: A perspective on promises, challenges, and research needs. IEEE Trans. Cogn. Dev. Syst. 2022, 15, 1051–1069. [Google Scholar] [CrossRef]
  23. Goh, K.M.; Ng, C.H.; Lim, L.L.; Sheikh, U.U. Micro-expression recognition: An updated review of current trends, challenges and solutions. Vis. Comput. 2020, 36, 445–468. [Google Scholar] [CrossRef]
  24. Zhou, L.; Mao, Q.; Xue, L. Dual-inception network for cross-database micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–5. [Google Scholar] [CrossRef]
  25. Wei, M.; Zong, Y.; Jiang, X.; Lu, C.; Liu, J. Micro-Expression Recognition Using Uncertainty-Aware Magnification-Robust Networks. Entropy 2022, 24, 1271. [Google Scholar] [CrossRef]
  26. Peng, M.; Wu, Z.; Zhang, Z.; Chen, T. From macro to micro expression recognition: Deep learning on small datasets using transfer learning. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 657–661. [Google Scholar] [CrossRef]
  27. Yu, J.; Zhang, C.; Song, Y.; Cai, W. ICE-GAN: Identity-aware and capsule-enhanced GAN with graph-based reasoning for micro-expression recognition and synthesis. In Proceedings of the 2021 International joint conference on neural networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar] [CrossRef]
  28. Zhang, M.; Yao, L.; Wenzhong, Y.; Yin, Y. GLFNet: Attention Mechanism-Based Global–Local Feature Fusion Network for Micro-Expression Recognition. Entropy 2025, 27, 1023. [Google Scholar] [CrossRef]
  29. Gan, C.; Xiao, J.; Zhu, Q.; Zhu, Y. Macro-expression-guided micro-expression recognition: A motion similarity perspective. Pattern Recognit. 2025, 171, 112237. [Google Scholar] [CrossRef]
  30. Zhang, M.; Yang, W.; Wang, L.; Wu, Z.; Chen, D. HFA-Net: Hierarchical feature aggregation network for micro-expression recognition. Complex Intell. Syst. 2025, 11, 169. [Google Scholar] [CrossRef]
  31. Peng, M.; Wang, C.; Bi, T.; Shi, Y.; Zhou, X.; Chen, T. A novel apex-time network for cross-dataset micro-expression recognition. In Proceedings of the 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII), Cambridge, UK, 3–6 September 2019; pp. 1–6. [Google Scholar] [CrossRef]
  32. Mao, Q.; Zhou, L.; Zheng, W.; Shao, X.; Huang, X. Objective class-based micro-expression recognition under partial occlusion via region-inspired relation reasoning network. IEEE Trans. Affect. Comput. 2022, 13, 1998–2016. [Google Scholar] [CrossRef]
  33. Zhang, T.; Zong, Y.; Zheng, W.; Chen, C.L.P.; Hong, X.; Tang, C.; Cui, Z.; Zhao, G. Cross-Database Micro-Expression Recognition: A Benchmark. IEEE Trans. Knowl. Data Eng. 2022, 34, 544–559. [Google Scholar] [CrossRef]
  34. Zong, Y.; Zheng, W.; Huang, X.; Shi, J.; Cui, Z.; Zhao, G. Domain regeneration for cross-database micro-expression recognition. IEEE Trans. Image Process. 2018, 27, 2484–2498. [Google Scholar] [CrossRef]
  35. Wang, Z.; Fu, W.; Zhang, Y.; Li, J.; Gong, W.; Gonzàlez, J. MCNet: Meta-clustering learning network for micro-expression recognition. J. Electron. Imaging 2024, 33, 023014. [Google Scholar] [CrossRef]
  36. van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  37. Wang, K.; Peng, X.; Yang, J.; Lu, S.; Qiao, Y. Suppressing uncertainties for large-scale facial expression recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6897–6906. [Google Scholar] [CrossRef]
  38. Baltrušaitis, T.; Mahmoud, M.; Robinson, P. Cross-dataset learning and person-specific normalisation for automatic action unit detection. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; Volume 6, pp. 1–6. [Google Scholar] [CrossRef]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  40. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. MetaFormer is Actually What You Need for Vision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10819–10829. [Google Scholar] [CrossRef]
  41. Teed, Z.; Deng, J. RAFT: Recurrent All-Pairs Field Transforms for Optical Flow. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 402–419. [Google Scholar] [CrossRef]
  42. Wang, C.; Peng, M.; Bi, T.; Chen, T. Micro-Attention for Micro-Expression recognition. Neurocomputing 2020, 410, 354–362. [Google Scholar] [CrossRef]
  43. Talluri, K.K.; Fiedler, M.A.; Al-Hamadi, A. Deep 3d Convolutional Neural Network for Facial Micro-Expression Analysis from Video Images. Appl. Sci. 2022, 12, 11078. [Google Scholar] [CrossRef]
  44. Zhang, M.; Huan, Z.; Shang, L. Micro-Expression Recognition Using Micro-Variation Boosted Heat Areas. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision, Nanjing, China, 16–18 October 2020; pp. 531–543. [Google Scholar] [CrossRef]
  45. Khor, H.Q.; See, J.; Phan, R.C.W.; Lin, W. Enriched Long-term Recurrent Convolutional Network for Facial Micro-Expression Recognition. In Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition, Xi’an, China, 15–19 May 2018; pp. 667–674. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.