You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

31 December 2025

ARTNet for Micro-Expression Recognition

,
,
,
and
1
Institute of Energy, Hefei Comprehensive National Science Center (Anhui Energy Laboratory), Hefei 230031, China
2
School of Aeronautics and Astronautics, Sichuan University, Chengdu 610207, China
3
National Key Laboratory of Fundamental Science on Synthetic Vision, Sichuan University, Chengdu 610065, China
*
Authors to whom correspondence should be addressed.
Sensors2026, 26(1), 247;https://doi.org/10.3390/s26010247 
(registering DOI)
This article belongs to the Section Intelligent Sensors

Abstract

The field of micro-expression recognition (MER) has garnered considerable attention for its potential to reveal an individual’s genuine emotional state. However, MER remains a formidable challenge, primarily due to the subtle nature and brief duration of micro-expressions. Many approaches typically rely on optical flow to capture motion between video frames. However, these methods exhibit limited variability in expression intensity across frames, which may not be effective for all individuals due to significant differences in their micro-expressions. To address this issue, we propose a novel framework called the Action Amplification Representation and Transformer Network (ARTNet) to adjust the motion amplitude, making it easier to recognize each individual’s micro-expressions. Firstly, we amplify the motion discrepancies between frames to enhance expression intensity. Subsequently, we calculate the optical flow of these amplified frames to depict micro-expressions more prominently. Finally, we use transformer layers to capture the relationships between different amplification features. Extensive experiments conducted on three diverse datasets confirm the efficacy of our proposed method.

1. Introduction

Facial expressions play a crucial role in discerning human emotions [1]. Therefore, capturing the state of micro-expressions promptly through automated means is highly valuable for addressing practical issues. Micro-expression recognition can be utilized to detect deceptive behavior [2], aiding in identifying falsehoods in various scenarios. This technology can assist in determining whether an interviewee or interrogated individual is concealing the truth. Compared to macro-expressions, micro-expressions have a shorter duration and lower amplitude of movement, making micro-expression recognition a challenging task. However, the study of micro-expressions continues to present many difficulties [3], such as their short duration and low range of motion. Moreover, it is challenging to determine a person’s emotional state by observing changes in a timely manner [4].
Because of the need to explore real human emotions, more and more researchers are studying how to monitor human micro-expressions. The small magnitude of micro-expressions’ movements leads to poor recognition accuracy. So, how to enhance their amplitude is a question worth investigating. Li Y et al. [5] employed a learning-based video motion magnification network to enhance facial movements in the apex frame. By employing this approach, the connections between various facial regions become more evident, thereby enhancing the effectiveness of subsequent micro-expression recognition tasks. Li et al. utilized Eulerian Motion Magnification [6] to enhance the apex frame extracted from original micro-expressions. This magnification technique effectively enhances the distinctions among various micro-expression categories, making it easier for the network to learn discriminative features. However, this amplification technique is generalized and does not effectively account for individual differences, resulting in insignificant extracted features, which may limit the performance of micro-expression recognition. In contrast, we amplify the motion amplitude of micro-expressions and utilize the optical flow method to capture the motion relationships between onset and apex frames, thereby ensuring high recognizability of micro-expressions.
To enhance the differences in micro-expressions, we propose the ARTNet to amplify micro-expressions. Our approach consists of three main steps. The first step is dedicated to amplifying the amplitude of motion of micro-expressions between the onset frame and the apex frame. Here, micro-expressions are initially scaled up by an appropriate factor, followed by the extraction of their optical flow information. In the next part, a deep learning approach is used to compute the optical flow information between the onset frame and the apex frame of the micro-expressions. As illustrated in Figure 1, motion amplification reinforces the small changes of the human face in micro-expressions, which can have an important impact on the improvement of MER. Finally, we focus on integrating both local fine-grained and global coarse-grained interactions at various image scales.
Figure 1. The onset and apex frames of micro-expressions are amplified at different levels, and then the optical flow information is extracted using deep learning, where α represents the magnification factor.
In summary, the key contributions of our work can be summarized as follows:
  • We present a novel approach to processing amplified micro-expression features, aiming to enhance the distinction between the micro-expression onset frame and apex frame. This addresses the issue of the limited range of micro-expression movements.
  • We introduce an innovative self-attention mechanism that prioritizes facial local areas over global areas like the eye and lip regions. Since micro-expression recognition heavily relies on local facial muscle movements, our model effectively captures essential features by focusing on specific facial regions using local attention while disregarding irrelevant areas.
  • Through experiments on three datasets, our proposed method demonstrates significant improvements over previous approaches, showcasing its effectiveness and superiority in micro-expression recognition tasks.

3. ARTNet

As illustrated in Figure 2, the proposed ARTNet consists of three main components: the Action Amplification Module (AA module), the Recurrent All-Pairs Field Transforms (RAFT module), and the Transformer and Block Aggregation Module (TBA module). The AA module selects the onset and apex frames of a micro-expression and applies motion amplification at an appropriate level. In the RAFT module, deep learning techniques are employed to extract optical flow information between the onset frame and the magnified frame, effectively capturing motion changes during the occurrence of the micro-expression. The TBA module incorporates a self-attention mechanism within the transformer layer of the low-level network to capture intricate details and features. Block aggregation processes combine smaller image blocks into larger ones to facilitate interactions and extract coarse-grained features. The resulting feature maps are then input into an MLP block for micro-expression classification. This modular and hierarchical design of the ARTNet model effectively extracts and integrates features at different scales and levels of granularity, ultimately enhancing micro-expression recognition performance.
Figure 2. The proposed ARTNet consists of three main modules: the AA module, the RAFT module, and the TBA module.

3.1. AA Module

Motion magnification is a valuable technique for enhancing subtle movements that may not be easily visible to the human eye but are nonetheless significant for analysis. According to the definition provided by Wadhwa et al. [24] regarding motion magnification, a solitary frame within a sequential video can be characterized as follows:
I ( x , t ) = f ( x + δ ( x , t ) ) .
where I ( x , t ) represents the image at position x and moment t of the luminance value, and δ ( x , t ) represents the motion deviation. The goal is to generate the amplified image I ˙ ( x , t ) . This can be expressed as
I ˙ ( x , t ) = f ( x + ( 1 + α ) δ ( x , t ) ) .
where α is the magnification factor.
Traditional motion amplification techniques primarily depend on optical and physical models, which present three significant limitations in enhancing subtle movements, such as in micro-expression analysis. First, these methods are highly sensitive to environmental noise; while amplifying target motion, they inadvertently enhance high-frequency noise, resulting in a significant reduction in the signal-to-noise ratio. Second, their algorithmic design typically assumes static backgrounds and simple scenes, making it difficult to distinguish genuine motion signals from background interference in dynamic or complex environments, ultimately reducing extraction accuracy. Lastly, traditional methods like Eulerian amplification are constrained by linear approximation theory, leading to theoretical upper limits on amplification coefficients that are inadequate for addressing nonlinear motion features. Collectively, these limitations restrict the practicality and reliability of traditional approaches in complex scenarios.
In the actual process, we use this pre-trained AA module [25] to zoom in on the motion between the onset and apex frames of a micro-expression by a factor of α 1 to α m . This can be formulated as
I a a α m = A ( I o n s e t , α m , I a p e x ) .
where I o n s e t is the onset frame of a micro-expression. α m stands for magnifying power. I a p e x is the apex frame of a micro-expression. I a a α m is using the AA model to zoom in on the motion between I o n s e t and I a p e x by a factor of α m to produce the enlarged picture. A ( ) is the AA module.
As illustrated in Figure 3, the AA module begins with the encoder extracting static features from both the onset and apex frames of the micro-expressions. Subsequently, the static information from the onset frame is subtracted from that of the apex frame to derive the motion information. This motion information undergoes operations such as convolution and the application of an activation function. Following this, an amplification factor is applied to enhance the motion information, after which feature extraction is performed on the amplified data, which is then added back to the onset frame. Finally, the decoder processes this combined information to generate the amplified frame. Notably, the motion changes between the amplified frame and the onset frame are more pronounced than those observed between the apex frame and the onset frame. The figure provides a clear illustration of the significant changes in the mouth region between the amplified frame and the onset frame.
Figure 3. Data from the micro-expression dataset after processing with the AA module.

3.2. RAFT Module

Researchers have highlighted the significance of optical flow features in motion estimation and have developed several enhanced techniques for extracting optical flow [11,26,27]. Optical flow represents the projection of an object’s movement in three-dimensional space onto a two-dimensional image plane. It is created by the relative velocity of the object and the camera, indicating the direction and speed of the object’s image pixels over a brief period. The optical flow approach is grounded on three key assumptions. The first assumption is constant brightness, which states that the pixel brightness of an object in an image remains unchanged between consecutive frames. The second assumption is short distance or short-time motion, where the time interval between adjacent frames is brief enough for the object to move minimally. The third assumption is spatial consistency, indicating that neighboring pixels exhibit similar motion. h ( x , y , t ) is the pixel value of the pixel point ( x , y ) at the moment of t. Then, according to the first two assumptions, it can be obtained as
h ( x , y , t ) = h ( x + d x , y + d y , t + d t ) .
Equation (4) passes through a first-order Taylor expansion, which yields
h ( x + d x , y + d y , t + d t ) = h ( x , y , t ) + h x d x + h y d y + h t d t .
From Equations (4) and (5), it can be inferred that
h t = h x d x d t h y d y d t .
The RAFT [28] algorithm, developed in the era of deep learning, aims to estimate optical flow by iteratively updating the flow field using GRU cyclic cells. This mimics the optimization process seen in traditional approaches. The RAFT network is composed of three main layers: the feature coding layer, the feature association layer, and the cyclic update operator. The feature coding layer extracts pixel-by-pixel features, while the feature association layer generates 4D correlation information for all pixels. The convergence layer produces lower-resolution correlation information. The cyclic update operator, based on GRU principles, uses 4D correlation information to iteratively update the optical flow field starting from an initial zero setting.
In the actual process, we utilize this pre-trained RAFT module to compute the optical flow map between the micro-expression’s onset frame and the magnified frame. This can be represented as
I ˙ raft α m = R I onset , I aa α m .
where I onset is the onset frame of a micro-expression. α m denotes magnifying power. I aa α m is an image generated through the application of the AA module. We use the RAFT module to calculate the optical flow between I onset and I apex to obtain I ˙ raft α m . R ( ) is the RAFT module.

3.3. TBA Module

The TBA module is composed of transformer layers and block aggregation layers, with each transformer independently processing the features of image blocks. Initially, horizontal and vertical optical flow motion maps are computed between the onset and apex frames of micro-expressions to effectively represent facial motion. This foundational step lays the groundwork for subsequent feature analysis, enabling the model to capture subtle dynamic changes in facial expressions. Next, the TBA module utilizes the multi-head attention mechanism of the transformer to model facial features. This approach accurately analyzes the dependencies among key facial regions, such as the lips, eyes, and nose, thereby improving the precision of feature extraction. Additionally, the block aggregation layer facilitates adaptive fusion of spatiotemporal features through a dynamic weight allocation strategy. In the preceding network, the self-attention mechanism of the transformer captures fine-grained features, while the block aggregation layer consolidates small image blocks into larger ones, allowing interaction between different blocks and the extraction of coarse-grained features. Lastly, the MLP layer within the model is employed for classifying the final feature map into distinct micro-expression categories.

3.3.1. Transformer Layer

The architecture consists of multiple transformer layers, where each layer utilizes visual tokens and key point tokens as input. Within each block, there is a Layer Normalization (LN) operation followed by a multi-head self-attention (MHA) module and a Feedforward Neural Network (FFN) module. MHA operation is structured in the following way:
M H A X l = concat soft   max Q · K T d k · V · M .
where Q, K, and V R S × d . During the transformation process of the transformer layer, S represents the sequence length and d is the dimension of the inputs. M R d × d refers to the combination of h attention heads. X l indicates the output tensor of l, which is the transformer layer. 1 d k is the scaling factor.
LN will be applied in each block as follows:
L N ( x ) = x μ δ o λ + β .
where μ is the mean of features. δ is standard deviation of the features. o is the element-wise dot. λ and β are learnable parameters.
The FFN layer can be expressed as
F F N ( x ) = max ( 0 , x W 1 + b 1 ) W 2 + b 2 .

3.3.2. Block Aggregation

Our ARTNet’s block aggregation function incorporates a hierarchical structure and feature fusion to capture multi-scale information, facilitating robust analysis of visual data across different scales. Additionally, our model leverages local attention for individual image blocks, leading to significant enhancements in performance. The accurate recognition of micro-expressions is heavily reliant on detecting localized facial muscle motion regions. By employing local attention and focusing on specific facial regions, our model effectively captures essential features for accurately inferring micro-expression states.
In the ARTNet, we enable the exchange of local and global features. Low-level block aggregation focuses on local facial areas to extract intricate details of facial dynamics, while high-level block aggregation facilitates global information exchange, capturing broader facial expressions. The block aggregation process in the ARTNet consists of several steps: it begins with a 3 × 3 convolutional layer, followed by Layer Normalization (LN) and a 3 × 3 max pooling operation. The facial optical flow map is initially divided into 16 blocks (4 × 4 feature maps). The first convolution combines features from these areas, reducing the block size to 2 × 2. A second convolution then exchanges information between the areas, further reducing the size to 1 × 1. This process extracts complete optical feature maps, which are input into an MLP layer for micro-expression classification. This hierarchical approach captures and combines features at different levels of detail, enhancing micro-expression recognition.

3.4. Loss Function

In this work, we utilize the cross-entropy loss function to train our model.
L = i ( w i log ( p t i ) ) .
p t = p c y ( 1 p c ) 1 y .
where w i is the weight assigned to each sample in the dataset, and y represents the ground truth label, and y i { 0 , 1 } .

4. Experiments

4.1. Datasets

We conducted experiments using three datasets: CASME II [29], SMIC [30], and SAMM [31]. The SAMM datasets do not pre-process for face cropping. In order to reduce the interference of the background, we use multi-task cascaded convolutional networks (MTCNNs) [32] to conduct preprocessing for face cropping. Especially, the SAMM and CASME II benchmark datasets include labeled onset, apex, and offset frames for each micro-expression sequence. However, the SMIC does not label the apex frame. We try to use intermediate position frame instead of apex frames. This is because we found that most of the apex frames are located in the middle part of the micro-expression video.

4.2. Evaluation Metrics

To minimize the influence of specific subjects, we utilize the leave-one-subject-out (LOSO) cross-validation approach. In each iteration, samples from one subject are reserved for testing, while samples from the remaining subjects are used for training. This process is repeated S times, where S = 27 in our experiments. We assess performance using the unweighted F1-score ( U F 1 ) and unweighted average recall ( U A R ). U F 1 calculates the average F1-score across all classes, providing an overall measure of classification accuracy that is not affected by class imbalances. U A R computes the average recall across all classes without considering class weights, with U F 1 defined as the average F1 across all n c classes.
In order to eliminate the impact of specific subjects during the process, we employ the leave-one-subject-out (LOSO) cross-validation approach. During each iteration, samples belonging to a single subject were kept aside for testing purposes, while all the remaining samples from other subjects were used for training the model. The experiment was repeated S times, where S represents the total number of subjects in the dataset. S is 27 in our experiments. To assess the performance, the unweighted F 1 -score ( U F 1 ) and unweighted average recall ( U A R ) metrics are employed. U F 1 calculates the average F1-score across all classes, providing an overall measure of classification accuracy that is not influenced by class imbalances. U A R computes the average recall across all classes without considering class weights. U F 1 is calculated as the average F 1 across all classes n c as
U F 1 = 1 n c j F 1 j .
where F 1 j = 2 s = 1 S T P j s 2 s = 1 S T P j s + s = 1 S F P j s + s = 1 S F N j s is the F 1 for the class. U A R is defined as the average recall as
U A R = 1 n c j Re c a l l j .
where Re c a l l j = s = 1 S T P j s s = 1 S F P j s + s = 1 S F N j s . T P j s , F P j s , and F N j s are true positives, false positives, and false negatives for j class of the s subject.

4.3. Implementation Details

The AA module is applied to focus on magnifying the motion between the onset and apex frames. Distinct values of α m are necessary for various datasets. For the CASME II dataset, α m is 3. For the SAMM dataset, α m is 2. For the SMIC dataset, α m is 7. Subsequently, the RAFT module calculates the optical flow between the onset and magnified frames of the micro-expression. The MTCNN is used to extract face landmark coordinates from the magnified images, which are essential for accurately localizing specific facial regions. Finally, the TBA module is employed to capture the relevant facial muscle movements related to micro-expressions, and these features are input to an MLP for classification.
Our experiments are conducted on a system with an Intel Core™ i7-9700 CPU, 16 GB RAM, and a GIGABYTE 3090Ti GPU using Ubuntu 20.04 and PyTorch 1.7. We optimize the model with the Adam optimizer, a learning rate of 0.00005, 800 epochs, and a batch size of 16.

4.4. Magnification Setting

Facial muscle movements become more noticeable at higher amplification levels, offering important visual cues for micro-expression recognition. However, excessive amplification can lead to facial distortion. To find an optimal balance between enhancing visibility and maintaining facial authenticity, we will conduct comparative experiments using the CASME II, SMIC, and SAMM datasets. In this study, we utilize UF1 and UAR metrics to systematically assess the effects of different amplification factors on micro-expression recognition performance.
As shown in Table 1, Table 2 and Table 3, with α m increasing, U F 1 and U A R show large increases in these datasets. In the CASME II dataset, U F 1 and U A R reach their maximum values when α m is 3, while, in the SMIC dataset, they peak at α m = 7, and, in the SAMM dataset, the maximum values occur at α m = 2. Compared to the optical flow of micro-expressions without magnification, both U F 1 and U A R are notably improved, which reflects the effectiveness of the magnification strategy.
Table 1. Assessment of various magnification levels on the CASME II dataset was conducted to evaluate their impact and effectiveness.
Table 2. Assessment of various magnification levels on the SMIC dataset was conducted to evaluate their impact and effectiveness.
Table 3. Assessment of various magnification levels on the SAMM dataset was conducted to evaluate their impact and effectiveness.

4.5. Comparison with the State of the Art

Conventional expression recognition techniques often use local binary patterns from three orthogonal planes (LBP-TOP) [8], which is effective for this purpose. LBP-TOP enhances micro-expression recognition by employing local binary patterns with six intersection points (LBP-SIP) for facial feature extraction. Bi-WOOF [11] utilizes the Bi-Weighted Optical Flow method to extract key facial characteristics from apex images, effectively capturing relevant facial motions and emphasizing critical details for micro-expression identification. To address cross-database challenges, Dual-Inception [33] introduces two inception networks that extract horizontal and vertical features from optical flow maps. STSTNet [34] uses a model with three shallow CNN layers to obtain high-level discriminative representations for micro-expression emotion classification. FeatRef [35] operates in two phases: first, it extracts horizontal and vertical muscle motion features using separate inception networks and then combines these features for classification through three attention-based networks. Finally, the classification branch integrates key features from the inception module to identify micro-expressions.
As shown in Table 4, both U F 1 and U A R of our ARTNet having high performance on all three datasets proves that the ARTNet can adapt to more complex situations. On the CASME II, both U F 1 and U A R of our framework ARTNet are higher than 0.7900. Especially, on the SAMM, the ARTNet improves by 3.35% and 1.30% on U F 1 and U A R over TFT [36]. On the SMIC, the ARTNet improves by 3.10% and 5.12% on U F 1 and U A R over TFT. This indicates that the ARTNet performs excellently in handling diverse micro-expression samples.
Table 4. The UF1 and UAR of different methods were evaluated using the leave-one-subject-out (LOSO) protocol on three datasets.

5. Conclusions

Due to their short duration and low amplitude of movement, micro-expressions are challenging to characterize as valid inputs for the network. To address this, we introduce a new adaptive network designed for action amplification in micro-expression recognition. This network comprises three key modules: the AA module, the RAFT module, and the TBA module. The AA module utilizes deep learning techniques to enhance the motion of micro-expression frames, specifically focusing on the onset and apex frames. Similarly, the RAFT module employs deep learning techniques to calculate the optical flow of micro-expressions at these critical frames. The TBA module further emphasizes essential components within the derived optical flow information, ultimately enhancing overall performance.
While this work provides an in-depth exploration of micro-expression recognition and proposes effective methods, several important considerations warrant attention for future research. First, the current reliance on the CASME II, SMIC, and SAMM datasets should be broadened to include samples from various age groups, as well as diverse social contexts and emotional states. Such an expansion would significantly enhance the model’s generalization capabilities. Furthermore, the high complexity of the ARTNet may hinder its practicality in real-time applications. Therefore, incorporating model compression techniques, such as pruning and quantization, is recommended to mitigate this complexity and facilitate more effective real-time implementation.

Author Contributions

Methodology, C.W.; software, L.S.; formal analysis, C.W.; resources, P.C.; data curation, Y.C.; writing—original draft preparation, C.W.; writing—review and editing, W.Z.; supervision, P.C.; project administration, P.C.; funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Institute of Energy, Hefei Comprehensive National Science Center (Anhui Energy Laboratory) under Grant No. 25KZS212.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, J.; Dong, Z.; Lu, S.; Wang, S.-J.; Yan, W.-J.; Ma, Y.; Liu, Y.; Huang, C.; Fu, X. CAS (ME)3: A third generation facial spontaneous micro-expression database with depth information and high ecological validity. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2782–2800. [Google Scholar] [CrossRef] [PubMed]
  2. Ekman, P.; Friesen, W.V. Nonverbal leakage and clues to deception. Psychiatry 1969, 32, 88–106. [Google Scholar] [CrossRef] [PubMed]
  3. Ben, X.; Ren, Y.; Zhang, J.; Wang, S.-J.; Kpalma, K.; Meng, W.; Liu, Y.-J. Video-based facial micro-expression analysis: A survey of datasets, features and algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5826–5846. [Google Scholar] [CrossRef] [PubMed]
  4. Yan, W.-J.; Wu, Q.; Liang, J.; Chen, Y.-H.; Fu, X. How fast are the leaked facial expressions: The duration of micro-expressions. J. Nonverbal Behav. 2013, 37, 217–230. [Google Scholar] [CrossRef]
  5. Li, Y.; Huang, X.; Zhao, G. Joint local and global information learning with single apex frame detection for micro-expression recognition. IEEE Trans. Image Process. 2020, 30, 249–263. [Google Scholar] [CrossRef]
  6. Wu, H.-Y.; Rubinstein, M.; Shih, E.; Guttag, J.; Durand, F.; Freeman, W. Eulerian video magnification for revealing subtle changes in the world. ACM Trans. Graph. (TOG) 2012, 31, 1–8. [Google Scholar] [CrossRef]
  7. Ben, X.; Jia, X.; Yan, R.; Zhang, X.; Meng, W. Learning effective binary descriptors for micro-expression recognition transferred by macro-information. Pattern Recognit. Lett. 2018, 107, 50–58. [Google Scholar] [CrossRef]
  8. Wang, Y.; See, J.; Phan, R.C.-W.; Oh, Y.-H. Lbp with six intersection points: Reducing redundant information in lbp-top for micro-expression recognition. In Computer Vision—ACCV 2014: 12th Asian Conference on Computer Vision, Singapore, Singapore, November 1–5, 2014, Revised Selected Papers, Part I 12; Springer: Berlin/Heidelberg, Germany, 2015; pp. 525–537. [Google Scholar]
  9. Wei, C.; Xie, L.; Ren, X.; Xia, Y.; Su, C.; Liu, J.; Tian, Q.; Yuille, A.L. Iterative reorganization with weak spatial constraints: Solving arbitrary jigsaw puzzles for unsupervised representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1910–1919. [Google Scholar]
  10. Zhao, G.; Pietikainen, M. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 915–928. [Google Scholar] [CrossRef]
  11. Liong, S.-T.; See, J.; Wong, K.; Phan, R.C.-W. Less is more: Micro-expression recognition from video using apex frame. Signal Process. Image Commun. 2018, 62, 82–92. [Google Scholar] [CrossRef]
  12. Fleet, D.; Weiss, Y. Optical flow estimation. In Handbook of Mathematical Models in Computer Vision; Springer: Berlin/Heidelberg, Germany, 2006; pp. 237–257. [Google Scholar]
  13. Liu, Y.; Du, H.; Zheng, L.; Gedeon, T. A neural micro-expression recognizer. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–4. [Google Scholar]
  14. Khor, H.-Q.; See, J.; Liong, S.-T.; Phan, R.C.W.; Lin, W. Dual-stream shallow networks for facial micro-expression recognition. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 36–40. [Google Scholar]
  15. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  16. Van Quang, N.; Chun, J.; Tokuyama, T. CapsuleNet for micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–7. [Google Scholar]
  17. Wang, S.-J.; Chen, H.-L.; Yan, W.-J.; Chen, Y.-H.; Fu, X. Face recognition and micro-expression recognition based on discriminant tensor subspace analysis plus extreme learning machine. Neural Process. Lett. 2014, 39, 25–43. [Google Scholar] [CrossRef]
  18. Wang, S.-J.; Yan, W.-J.; Li, X.; Zhao, G.; Zhou, C.-G.; Fu, X.; Yang, M.; Tao, J. Micro-expression recognition using color spaces. IEEE Trans. Image Process. 2015, 24, 6034–6047. [Google Scholar] [CrossRef]
  19. Xu, F.; Zhang, J.; Wang, J.Z. Microexpression identification and categorization using a facial dynamics map. IEEE Trans. Affect. Comput. 2017, 8, 254–267. [Google Scholar] [CrossRef]
  20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  21. Bertasius, G.; Wang, H.; Torresani, L. Is space-time attention all you need for video understanding? ICML 2021, 2, 4. [Google Scholar]
  22. Dong, X.; Long, C.; Xu, W.; Xiao, C. Dual graph convolutional networks with transformer and curriculum learning for image captioning. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 2615–2624. [Google Scholar]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  24. Wadhwa, N.; Rubinstein, M.; Durand, F.; Freeman, W.T. Phase-based video motion processing. ACM Trans. Graph. (ToG) 2013, 32, 1–10. [Google Scholar] [CrossRef]
  25. Oh, T.-H.; Jaroensri, R.; Kim, C.; Elgharib, M.; Durand, F.; Freeman, W.T.; Matusik, W. Learning-based video motion magnification. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 633–648. [Google Scholar]
  26. Patel, D.; Zhao, G.; Pietikäinen, M. Spatiotemporal integration of optical flow vectors for micro-expression detection. In Advanced Concepts for Intelligent Vision Systems: 16th International Conference, ACIVS 2015, Catania, Italy, October 26–29, 2015. Proceedings 16; Springer: Berlin/Heidelberg, Germany, 2015; pp. 369–380. [Google Scholar]
  27. Wang, S.-J.; Wu, S.; Qian, X.; Li, J.; Fu, X. A main directional maximal difference analysis for spotting facial movements from long-term videos. Neurocomputing 2021, 230, 382–389. [Google Scholar] [CrossRef]
  28. Teed, Z.; Deng, J. Raft: Recurrent all-pairs field transforms for optical flow. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 402–419. [Google Scholar]
  29. Yan, W.-J.; Li, X.; Wang, S.-J.; Zhao, G.; Liu, Y.-J.; Chen, Y.-H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef]
  30. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement; collection; baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (fg), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar]
  31. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. Samm: A spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 2018, 9, 116–129. [Google Scholar] [CrossRef]
  32. Jose, E.; Greeshma, M.; Haridas, M.T.P.; Supriya, M. Face recognition based surveillance system using facenet and mtcnn on jetson tx2. In Proceedings of the 2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS), Coimbatore, India, 15–16 March 2019; pp. 608–613. [Google Scholar]
  33. Zhou, L.; Mao, Q.; Xue, L. Dual-inception network for cross-database micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–5. [Google Scholar]
  34. Liong, S.-T.; Gan, Y.S.; See, J.; Khor, H.-Q.; Huang, Y.-C. Shallow triple stream three-dimensional cnn (ststnet) for micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–5. [Google Scholar]
  35. Zhou, L.; Mao, Q.; Huang, X.; Zhang, F.; Zhang, Z. Feature refinement: An expression-specific feature learning and fusion method for micro-expression recognition. Pattern Recognit. 2022, 122, 108275. [Google Scholar] [CrossRef]
  36. Wang, Z.; Yang, M.; Jiao, Q.; Xu, L.; Han, B.; Li, Y.; Tan, X. Two-level spatio-temporal feature fused two-stream network for micro-expression recognition. Sensors 2024, 24, 1574. [Google Scholar] [CrossRef] [PubMed]
  37. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  38. Sengupta, A.; Ye, Y.; Wang, R.; Liu, C.; Roy, K. Going deeper in spiking neural networks: VGG and residual architectures. Front. Neurosci. 2019, 13, 95. [Google Scholar] [CrossRef]
  39. Zhang, L.; Hong, X.; Arandjelović, O.; Zhao, G. Short and long range relation based spatio-temporal transformer for micro-expression recognition. IEEE Trans. Affect. Comput. 2022, 13, 1973–1985. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.