Next Article in Journal
Divergence Shepherd Feature Optimization-Based Stochastic-Tuned Deep Multilayer Perceptron for Emotional Footprint Identification
Previous Article in Journal
Correction: Sanjalawe et al. Reinforcement Learning-Guided Hybrid Metaheuristic for Energy-Aware Load Balancing in Cloud Environments. Algorithms 2025, 18, 715
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Analysis and Comparison of Machine Learning-Based Facial Expression Recognition Algorithms

College of Science, Beijing Forestry University, Beijing 100083, China
*
Authors to whom correspondence should be addressed.
Algorithms 2025, 18(12), 800; https://doi.org/10.3390/a18120800
Submission received: 25 October 2025 / Revised: 5 December 2025 / Accepted: 13 December 2025 / Published: 17 December 2025

Abstract

With the rapid development of artificial intelligence technology, facial expression recognition (FER) has gained increasingly widespread applications in digital human generation, humanoid robotics, mental health, and human–computer dialogue. Typical FER algorithms based on machine learning have been widely studied over the past few decades, which motivated our survey. In this study, we have surveyed the state of the art in FER across two categories: traditional machine learning-based (ML-based) and deep learning-based (DL-based) approaches. Each category is analyzed based on six subcategories. Then, twelve methods, including four ML-based models and eight DL-based models, are compared to evaluate FER performance across four datasets. The experimental results show that in validation sets, the average accuracy of HOG-SVM is 50.12%, which is the best performance for the four ML-based methods; in contrast, Poster has an average accuracy of 75.98%, which is the best result obtained among the eight DL-based methods. The most difficult expression to recognize is contempt, with recognition accuracies of 10.00% and 40.06% for ML-based and DL-based methods, respectively. The accuracy of the ML-based method for identifying neutral expression is the highest at 35.25%; the DL-based method has the highest accuracy in identifying surprise at 69.56%. From the theoretical analysis and comparative experimental results of existing methods, we can see that FER faces challenges, including inaccurate recognition in complex environments and unbalanced data categories, highlighting several future research directions, especially those involving the latest applications of digital humans and large language models.

1. Introduction

With the rapid development of artificial intelligence technology, digital humans (meta-humans) have been applied to news anchors, virtual e-commerce live streaming, AI-assisted teaching, and humanoid robots. Generating appropriate expressions for digital humans and robots will improve their visual impact. Therefore, the recognition, analysis, and generation of digital human facial expressions are topics worthy of study. The recognition and analysis of facial expressions provide the technical basis for facial expression generation.
Facial expression is one of the most intuitive external manifestations of human emotions and psychological states, playing a vital role in human communication throughout history. Facial expression recognition (FER) is an important research direction in computer vision and pattern recognition, aiming to equip computer systems with the ability to interpret the emotional information in human facial expressions. In the field of human–computer interaction, it can enable computer systems to perceive users’ emotions more naturally and intelligently, thereby optimizing the interactive experience. For example, intelligent customer service can adjust the response strategy according to users’ expressions and enhance quality of service [1]. In medical health, it can assist doctors in diagnosing patients’ mental states and help with mental health assessment and disease monitoring [2]. In security monitoring, it can provide timely warnings of potentially dangerous behaviors by identifying abnormal expressions, ensuring public safety. In recent years, the generation of digital humans and the widespread use of humanoid robots have also made facial expression recognition applicable in a large number of scenarios.
Since the 20th century, research on facial expression recognition has been widely conducted. Early on, due to limitations in computer performance and image processing technology, research on facial expression recognition focused on developing basic theory and exploring simple approaches. With the rapid development of computer technology, especially the iterative advancement of image processing, pattern recognition, machine learning (ML), and related technologies, FER research has gradually shifted from traditional methods based on prior knowledge to deep learning (DL) models. Traditional methods often use geometric features, such as the relative position and contour shape of facial organs, or texture features, such as local binary pattern (LBP) and its variants, to extract the features of expression, and then combine support vector machine (SVM) and adaptive boosting (AdaBoost) classifiers to realize expression classification [3]. With the development of deep learning, convolutional neural networks (CNNs), recurrent neural networks (RNNs), and their derived models have been widely used. CNNs’ powerful image feature extraction capabilities allow them to automatically learn multi-level features from expression images [4]; RNNs can effectively process expression sequence data and capture dynamic changes in expression over time [5]. In addition, some hybrid learning (HL) methods integrating multiple features or models have emerged in the field of deep learning technology, such as cascaded networks and multimodal models [6], aiming to fully leverage the advantages of different technologies and improve the performance of expression recognition technologies [7]. The general history of facial expression recognition is shown in Figure 1.
Figure 1 shows the development timeline of FER, which can be summarized into the following key stages:
i.
Theory and data foundation (1971–2000): In 1971, Ekman and Friesen [8] proposed the basic emotion theory applicable across cultures, before establishing the facial motion coding system (FACS) in 1978, defining action units (AUs), which laid the theoretical foundation for subsequent algorithm annotation. To support data-driven research, the Jaffe dataset was released in 1998, and the CK dataset was released in 2000 [9], establishing an early standard for expression recognition in controlled environments.
ii.
Machine learning and Feature Engineering (2009–2012): This period focused on the effectiveness of manual features. In 2009, Shan et al. [10] demonstrated that combining local binary patterns (LBPs) with SVMs achieved a breakthrough performance on a Cohn–Kanade dataset. In 2010, the release of the extended CK dataset [11] provided more complete annotations for action units and emotion, and these factors remain important benchmarks.
iii.
Deep learning and field challenges (2013–2019): The release of FER2013 in 2013 marked a shift in research focus to large-scale “in the wild” data. Subsequently, Mollahosseini et al. [12] successfully applied the concept and other deep networks to FER. In 2017, the release of AffectNet and RAF-DB further developed the field to address the complexity of real-world expressions (e.g., complex emotions). Self-supervised learning (such as SimCLR) also began to affect the FER pre-training paradigm in around 2019.
iv.
Transformer and multimodal integration (2021-present): Since 2021, vision transformers (VITs) have been systematically introduced into FER to capture global dependencies. The EmoSet released in 2023 represents a trend toward larger-scale, richer visual emotion data. The current research frontier is gradually shifting toward multimodal fusion (images, voices, and physiological signals) and exploring privacy-preserving technologies.
Although FER technology has made significant progress, it still faces many significant challenges in achieving wider applications. One problem is the changeable real environment. Dramatic changes in lighting conditions, facial gestures from different angles, and partial occlusion caused by wearing masks and glasses will seriously compromise the accuracy of expression recognition [13]. Another challenge is the richness and diversity of human expressions. In addition to basic expressions such as happiness, surprise and sadness, there are many subtle and fuzzy composite expressions, which are difficult to identify [14] accurately; the differences in the facial structure and muscle movement patterns of different individuals, as well as the diversity of expression and understanding in individuals from different cultural backgrounds, all add difficulty to the construction of a unified and accurate expression recognition model [15]. In addition, reducing the model’s parameter count and floating-point operations is a challenge. The lightweight model can be better deployed on some edge devices, reducing deployment costs [16]. Therefore, in the absence of a unified standard method, in this study, we survey the existing methods to help researchers further promote the development of expression recognition technology, as well as to select appropriate methods according to different needs. Unlike most existing reviews [17,18,19,20,21] that focus on text analysis, i.e., comparing data, conclusions, method names, ideas, and performance indicators, we add a comprehensive experimental comparison. The similarities and differences between our review and these existing reviews are listed in Table 1. As shown in the table, the highlights of this study include a detailed literature review, algorithmic comparisons across multiple datasets, and suggestions for future research directions using large language models (LLMs).
The main contributions of this study include the following three points:
  • We review some typical facial expression recognition methods by category, based on the theoretical ideas and modeling tools of the algorithms. The selection criteria for a reference to be included in this review are twofold. The first is that it selects at least one representative method from each category or subcategory. The second is that it selects a reference that proposed a method that was significant or showed significant technological progress.
  • We compare the performances of 12 FER algorithms based on 4 outdoor facial expression datasets. Comparisons of experiments using the same dataset on the same computer are lacking in many existing review articles.
  • We propose potential future research directions in the field of FER, especially by combining LLMs and digital human technology, based on summarizing the gap between the existing methods and the future demand of technology.
The structure of this paper is as follows: The second section introduces the standard preprocessing technology used for facial expression data. In the third section, the existing expression recognition methods are reviewed, including traditional feature extraction methods and deep learning methods. In the fourth section, several typical FER methods are summarized, and comparative experiments are conducted to reveal the performances of these methods. Finally, the fifth section summarizes the conclusion and points out future research directions.

2. Preprocessing

In FER, data preprocessing can improve model performance. Due to the poor quality of public facial expression datasets, problems such as face occlusion, non-uniform lighting (too dark or too bright), and blurred or low-resolution images often occur, significantly affecting the accuracy of feature extraction and classification. Therefore, it is usually necessary to preprocess the original image to reduce noise interference and improve data consistency. Here, we present several image preprocessing methods used in facial expression recognition.
Face detection and alignment are used to locate the face and align the positions of facial features (such as the eyes and mouth corners) [22]. Zhang et al. [23] proposed a multiple-task cascaded convolutional network (MTCNN) for face detection and alignment. This method achieves robust detection across multiple poses and complex lighting conditions by refining candidate boxes iteratively and combining face detection and keypoint localization. Wang et al. [24] introduced the image inpainting and the structural similarity index measure (SSIM) method to address the occlusion problem in expression images. They used completion technology based on structural constraints to fill the occluded area and ensured the overall consistency of the repaired image and the original face through structural similarity.
Normalization and scale adjustment are used to normalize the input image. After normalization, these input images are resized to a fixed size and normalized to a uniform gray range (e.g., [0,1] or a standard normal distribution), thereby avoiding training instability caused by resolution differences or inconsistent pixel value ranges. Cootes et al. [25] used the active appearance model (AAM) to realize the geometric normalization of the face. By jointly optimizing the shape and texture models, this method aligns different face images to a unified reference coordinate system, enabling comparisons of changes in expression in a structured space.
Illumination and contrast enhancement can be implemented using histogram equalization, gamma correction, or adaptive brightness adjustment, thereby alleviating feature distortion caused by changes in illumination intensity. Pizer et al. [26] used adaptive histogram equalization to improve image visual quality by enhancing local contrast. This method can maintain good texture details across different lighting conditions, thereby alleviating the interference of lighting differences on expression feature extraction.
Image enhancement and denoising can be achieved using smoothing and super-resolution reconstruction. Tomasi and Manduchi [27] proposed a bilateral filtering method combining spatial-domain and pixel-intensity-domain weights to remove noise while preserving image edges. When applied to facial images, this method can effectively reduce blur and noise and improve the identifiability of facial texture features.
In general, reasonable preprocessing can not only effectively alleviate adverse factors such as occlusion, illumination, and blur but also enhance the discriminability of feature expression to a certain extent, thereby improving the facial expression recognition performance of subsequent traditional machine learning or deep learning models.

3. Facial Expression Recognition Method

This section reviews the FER methods developed and applied in the past 20 years. These methods can be divided into two categories according to whether it is necessary to actively extract features: methods based on prior features (machine learning-based methods) and methods based on deep learning (deep learning-based methods). The former type relies on researchers to design features (such as facial texture and organ position) and extract features based on prior knowledge to recognize facial expressions, such as SVM algorithm, which offers fast speed and strong interpretability when constructing classifiers. However, its effect depends on the characteristics of the data and the design, and its generalizability is generally weak. The latter type is based on deep learning and can automatically learn features, such as a convolutional neural network, offering high recognition performance but weak interpretability and a high parameter learning cost. The main progress in the development of the FER methods is outlined in the following subsections. Although the models selected overlap with the above survey references, our survey covers both classic manual features and the latest architectures from 2024 to 2025, providing an effective technical supplement to the existing survey literature.

3.1. Methods Based on Prior Features

The prior features listed here refer to the statistics constructed by researchers based on specific theories, mainly including wavelet transform (WT), local binary patterns (LBPs), histogram of oriented gradients (HOG), the gray-level co-occurrence matrix (GLCM), local phase quantization (LPQ), and scale-invariant feature transform (SIFT). The methods are described in detail below. Figure 2 shows the general traditional machine learning method flow for classification tasks.

3.1.1. Wavelet Transform-Based Method

Wavelet transform has a good effect in terms of extracting the texture features of facial expressions, so it is also used in expression recognition. For example, using Gabor wavelets, a key complex function of the method is
G x , y ; λ , θ , ψ , σ , γ = exp x 2 + γ 2 y 2 2 σ 2 exp i 2 π x λ + ψ
where x , y are the coordinates of pixel points; λ is the wavelength of the sine wave, where the smaller the value, the finer the detection; θ is the direction of the filter; and ψ is the phase offset. When ψ = 0, it represents the even symmetric filter, and when ψ = π 2 , it represents the odd symmetric filter; σ is the standard deviation of Gaussian function, which controls the size of the filter window, while γ is the spatial aspect ratio.
We searched for articles from 2000 to 2024 on the “IEEE Xplore” platform using the keyword “facial expression recognition” and obtained 15,924 results. Then, we added the keyword “Gabor” to the advanced search function and obtained 1107 results. This shows that the Gabor filter plays a role in expression recognition due to its excellent texture analysis capabilities across multiple scales and orientations. In 2008, Bashyal et al. proposed a method of facial expression recognition using Gabor wavelet and learning vector quantization [28]. The method noted that the two-dimensional Gabor function is a plane wave with wave factor k , which is limited by a Gaussian envelope function with a relative width σ :
ψ k , x = k 2 σ 2 e x p k 2 x 2 2 σ 2 exp i k · x exp σ 2 2
where σ is set to π when the image has a resolution of 256 × 256. A set of discrete Gabor kernels is used, which are the spatial frequencies k = π 4 , π 8 , π 16 and six different directions from 0° to 180° (every 30°). A filter group consisting of 18 filters is obtained. The filter group is applied to 34 target points of the image to obtain the feature vector. Then, principal component analysis (PCA) is used to reduce the dimension. Finally, classification is performed using learning vector quantization.
Zhou et al. [29] proposed a quaternion wavelet transformer (QWTR) in 2025. This method is oriented toward facial expression recognition in a complex field environment. It aims to address the decline in recognition accuracy caused by factors such as illumination changes, head pose deviations, and occlusions. This method combines quaternion mathematical theory with transformer structure, and the overall framework is composed of three parts: quaternion histogram equalization (QHE) is used to enhance image brightness and contrast and maintain color structure; quaternion wavelet transform-based feature selection (QWT-FS) extracts high-quality emotional features by decomposing and filtering the most relevant wavelet coefficients; quaternion-valued transformer (QVT) integrates a quaternion CNN and a multi-head attention transformer module to jointly model facial emotional information from local and global levels. The experiment was carried out on four typical FER in-the-wild datasets (RAF-DB, AffectNet, SFEW, and ExpW). QWTR surpassed the current state-of-the-art model across many evaluation metrics, demonstrating excellent robustness and generalization. Particularly in challenging scenes such as illumination changes, skin color differences, occlusions, and negative postures, the model maintains stable, efficient recognition performance.

3.1.2. Local Binary Pattern-Based Method

Local binary pattern (LBP) is also used to extract texture features. The idea behind this method is to describe the local texture features of an image by comparing the gray values of each pixel with those of its adjacent pixels. Let the central pixel be x c , y c and its corresponding gray value be g c . The gray value of the P adjacent pixels of the pixel is g i i = 1,2 , , P :
s i = 1 , g i > g c 0 , o r t h e r w i s e
It is referred to as the 0–1 encoding of the ith-nearest-neighbor pixel. Then, it arranges the 0–1 codes of the P adjacent pixels in clockwise or counterclockwise order to obtain a binary code and convert it to decimal to obtain the LBP value of the central pixel. At present, many improved versions of the traditional LBP method have been developed, such as circular LBP and rotation-invariant LBP. In facial expression recognition, the face can be divided into multiple regions, LBP histograms are created on each region, and then histograms are concatenated to form feature vectors.
Benniu et al. [30] combined LBP (local binary pattern) and ORB (Oriented FAST and Rotated BRIEF) features for facial expression recognition on the CK+, JAFFE, and MMI databases, demonstrating that this feature extraction method was effective. ORB comprises four steps: fast detection of key points, Harris corner response filtering, calculation of key point directions, and a brief description of the image around each key point. First, let p be a candidate corner, t be the intensity threshold set in advance, and p i ( i = 1,2 , , n ) be n consecutive pixels around p . g p is the gray value of point p . Similarly, g p i is the gray value of point p i . If g p i g p > t for all p i values, then point p is a corner. Given the possibility of overlapping corners, a non-maximum suppression strategy is adopted. Next, the main direction of the feature point, that is, the corner point, is calculated. Then, the region around the key points is described using the brief descriptor. Next, select several pairs of pixels around the feature points, compare their gray values, and express the results with 0–1. Finally, a binary code is formed.

3.1.3. Directional Gradient Histogram-Based Method

The histogram of oriented gradients (HOG) is often used to describe the local characteristics of an image by counting the distribution of gradient directions in a local region.
Firstly, the image is preprocessed to convert the color image into a gray image. Then, gamma correction is performed, that is, a power transformation is performed on each pixel value. Let I ( x , y ) be the value of the pixel ( x , y ) . After gamma transformation, the new pixel value is I ( x , y ) = I ( x , y ) γ . Then, the Sobel operator is used to calculate the gradients G x and G y of the image in the horizontal and vertical directions. Then, the gradient amplitude G ( x , y ) is obtained by G x , y = G x 2 + G y 2 . According to θ = a r c t a n G y G x , calculate the gradient direction. The image can be divided into several patches, usually 8 × 8 or 16 × 16 pixels. In each patch, the gradient direction of each pixel is calculated and divided into nine intervals, either 0° to 180° or 0° to 360°. For each pixel, the gradient direction is used to determine the corresponding interval, and the gradient amplitude is added to it; finally, a nine-dimensional feature vector is formed. Pierluigi Carcagnì et al. [31] verified the practicability of the HOG feature extraction method and discussed how to adjust HOG parameters to make it more suitable for the facial expression recognition task.

3.1.4. Gray-Level Co-Occurrence Matrix-Based Method

For any pixel (x, y) in an image and another point ( x + a ,   y + b ) with a certain distance from it, let the gray value pairs of these points be ( g 1 , g 2 ) . Assuming that the gray values are divided into k intervals, there are k 2 intervals of such gray value pairs. Let ( x ,   y ) traverse the entire image, count the number of occurrences of each gray value pair, and arrange it into a square matrix. The number of occurrences is normalized to the probability of the occurrence of each gray value pair. Such a square matrix is called a gray-level co-occurrence matrix. Let ( a ,   b ) take a different combination, and then the gray-level co-occurrence relationships of pixel pairs at different distances and directions can be counted.
In 2014, Murty et al. [32] proposed a facial expression recognition method based on the fusion of the gray-level co-occurrence matrix (GLCM) and distinctive local binary pattern (DLBP) based on a 3 × 3 first-order compressed image (FCI), which is used to accurately recognize seven kinds of expressions neutral, happy, sad, surprised, angry, disgusted, and fearful. The method consists of three steps: Firstly, 5 × 5 subimages are compressed into 3 × 3 subimages to retain key information. Secondly, two DLBPs (the sum of the upper-triangle LBP, SUTLBP, and the sum of the lower-triangle LBP, SLTLBP) are extracted from the triangular pattern of the upper and lower parts of the 3 × 3 subimage. Finally, GLCM is constructed based on DLBP, and characteristic parameters such as contrast, uniformity, energy and correlation are calculated for recognition. This method overcomes the generalization problem caused by the unpredictable distribution of facial images in real-world environments, a limitation of traditional statistical methods (such as PCA and LDA), and addresses the limitations of traditional LBP due to lighting, thereby improving classification performance by combining structural and statistical features. On the database containing 213 female facial expression images, the k-nearest neighbor classifier (k = 1) achieved an accuracy of 96.67%, significantly better than those of the two-tier perceptron architecture developed by Zhengyou Zhang [33] and the expression analysis method developed by De la Torre F et al. [34], and verified its effectiveness.

3.1.5. Local Phase Quantization-Based Method

By analyzing the phase information of the local region, detailed features of the image can be captured. Compared with the direct use of pixel values or other amplitude-based feature descriptors, phase information is more robust to illumination changes and image blur. This is because illumination changes mainly affect the image amplitude, rather than its phase. Local Phase Quantization (LPQ) is based on quantizing the phase information of the local Fourier transform. The specific calculation process is as follows: For a fixed-size window, the pixels within it are subjected to the Short-Time Fourier Transform (STFT). Let the image be f ( x ) , the pixel be x , and its neighborhood be N x . Then, the local Fourier transform of pixel x is
F u , x = y N x f y . ω y x . e j 2 π u T y
where ω y x is the weighted window parameter, which is the commonly used Gaussian window, and u is the frequency vector. In most cases, four frequencies, recorded as u 1 , u 2 , u 3 , u 4 , are selected. Let
F x = F u 1 , x , F u 2 , x , F u 3 , x , F u 4 , x
For each frequency point u k , the symbols of the real and imaginary parts of the Fourier transform are calculated. Let
q j = 1   i f   g j 0 0   e l s e
where g j is the j-th term of the vector G ( x ) = [ R e { F x } , I m { F x } ] . Thus, an eight-bit binary number is obtained and converted into a decimal number. This decimal number is the LPQ value of pixel x . The image is then divided into several regions. The LPQ values for each region are calculated, and a histogram of the statistics is generated. Finally, the LPQ feature of the image is obtained by connecting the histograms of each region. At present, LPQ features have also been improved and varied.
Jiawaili et al. proposed a new facial expression recognition algorithm based on rotation-invariant local phase quantization (RI-LPQ) and sparse representation [35]. RI-LPQ was used to extract expression features, combined with the sparse representation-based classification (SRC) method. The test image was represented as a linear combination of training images, and the expression was determined using sparse representation residual analysis. Compared with 2DPCA+SVM, LDA+SVM, and other algorithms, the new method achieves better performance and recognizes facial expression images with different occlusions.
Nikan et al. proposed a face recognition method combining multi-scale LPQ and multi-resolution LBP [36]. First, the image is preprocessed to obtain a representation that is insensitive to illumination changes. Then, the image is divided into small sub-blocks, and LPQ and LBP features are extracted from them. By combining score-level and decision-level features, recognition performance under severe illumination changes is improved.

3.1.6. Method Based on SIFT

SIFT is a classic local feature extraction algorithm that was proposed by Lowe in 1999 [37] and further improved in 2004 [38]. Its core advantage lies in its robustness to image scaling, rotation, and illumination changes, as well as its good adaptability to angle changes and noise. In this method, the image is convolved with Gaussian kernels of different scales to generate a series of blurred images, and the differences between adjacent blurred images are calculated to obtain difference-of-Gaussian (DoG) images. In these DoG images, the local maxima or minima present within a 26-neighborhood area around each pixel at the current scale and adjacent scales are searched for candidate key points. Then, the gradient direction histogram in the neighborhood of the key points is calculated. These histograms are combined into a 128-dimensional feature vector and normalized. The SIFT feature is widely used in facial expression recognition. For example, Berretti et al. [39] proposed an expression recognition method based on a 3D facial model, achieving a recognition rate of 78.43% on the BU-3DFE dataset by automatically detecting and combining key points and SIFT descriptors using an SVM classifier. In addition, Shi et al. [40] proposed an improved SIFT algorithm to enhance the robustness of expression recognition under different lighting and pose conditions through shape decomposition and feature point constraint.

3.2. Deep Learning-Based Method

The deep learning-based method generally does not require researchers to design statistical features and directly uses the original image as input. Researchers mainly focus on the model architecture. The following are nine typical convolutional neural network models for facial expression recognition. Figure 3 shows the general process of expression recognition using a deep learning method.

3.2.1. Facial Expression Recognition Based on Traditional CNN

In recent years, various classic CNN architectures (such as VGGNet, ResNet, and Inception) have been introduced into the field of expression recognition, leading to significant performance improvement. In 2023, Luo et al. [41] proposed an improved CNN model by optimizing the convolution activation function and the classification layer loss function. Using the convolution activation function, the authors comprehensively consider the pixels of the facial expression image, the convolution operation times, and the pooling operation times and attain highly integrated feature extraction. Using the classification loss function, the weights and bias values of the convolution layer, pooling layer, and classification layer are defined, enhancing the flexibility of the model. In the experimental part, three gradient descent algorithms (e.g., Full Gradient Descent, FGD; Stochastic Gradient Descent, SGD; Mini-batch Gradient Descent, MGD) and six traditional models (LBP+SVM, CSLBP/RILPQ+SVM, DeeID, Deep Face, LeNet-5, and the improved CNN model) are proposed. The experimental results show that the improved algorithm achieves 87% classification and 88% recognition accuracy, outperforming other comparison methods. In 2024, Shehada et al. [42] proposed two methods to improve facial expression recognition performance. The first method combines negative emotions (anger, fear, sadness, disgust) into a single category to simplify the classification task. The second method introduces diverse datasets. In terms of training and testing, the authors used 80% of the FER-2013, IMDB, AffectNet, and RAF-DB dataset samples for training; 20% of the FER-2013 samples for validation; and all CK+ samples for testing, achieving 90.9% and 94.7% accuracy on the FER-2013 and CK+ datasets, respectively.
However, challenges remain in distinguishing similar negative expressions, such as anger and disgust. In 2024, Adiga et al. [43] proposed that, based on image preprocessing, a multi-layer CNN be used to automatically extract facial features, such as edges, textures, and shapes. During training, the probability of each emotion is computed using the softmax activation function, and the model is optimized with the Adam optimizer or the SGD optimizer using the cross-entropy loss function. The experimental results show that the model achieves high accuracy and generalization on the FER-2013 dataset, especially for happy, sad, surprised, and other emotions.

3.2.2. Residual Network-Based Method

Deep residual learning addresses the degradation problem during deep neural network training. This method uses the core strategy of residual mapping and shortcut connection. In expression recognition, ResNet is often used as a benchmark model for comparison with other methods. At the same time, ResNet is often used as a part of many neural networks to increase the depth of the network, enabling the model to achieve better results.
Istiqomah A A et al. [44] used the pre-trained ResNet-50 network as the feature extractor in the expression recognition task and introduced transfer learning to improve the performance of the model; the Kaggle public facial expression dataset was used, with 80% used for training and 20% used for testing, and edge detection and region detection were used in preprocessing, with an accuracy of 99.49% and an F1 score of 99.60%. Jiahang Li [45] systematically compared the performances of different depth variants of ResNet (ResNet-18, 50, 101, and 152) and the attention-enhancement variant SE-ResNet. They conducted experiments on the FER-2013 dataset and used image enhancement techniques (e.g., random cropping, horizontal flipping, and brightness and contrast adjustments) to improve the model’s generalization. The results showed that SE-ResNet-34 with medium depth achieved the best performance, with an accuracy of 70.80%, better than those of SE-ResNet-50 and SE-ResNet-101 with deep depth. Roy et al. [46] proposed ResEmoteNet, which uses a CNN to extract hierarchical features, employs squeeze-and-excitation (SE) blocks to enhance key facial feature channels selectively, and combines residual blocks to alleviate the problem of deep network gradient disappearance and improve the learning of complex features. On the open-source datasets FER-2013, RAF-DB, AffectNet-7, and ExpW, the accuracy of ResEmoteNet [46] was 79.79%, 94.76%, 72.39%, and 75.67%, respectively, surpassing the accuracies of APViT, S2D, and other methods.

3.2.3. Attention Mechanism-Based Method

In facial expression recognition, an attention mechanism can help the model to focus on highly correlated facial regions, such as eyes, mouth corners, and eyebrows, especially under complex conditions such as subtle expressions, facial occlusion, or lighting changes, thereby improving robustness. This mechanism can be implemented using two methods: the first method is embedding traditional convolutional neural network modules (such as SE and CBAM) to enhance feature expression; the second method uses a full-attention architecture (such as a transformer) to completely replace convolutional operations. Using the former method, channel attention mainly focuses on “which feature channels are more important”, while the latter method focuses on “which positions in the image are more critical”. This mechanism improves the model’s discrimination ability and effectively alleviates feature redundancy.
Considering the illumination change, occlusion, micro-expression, and other factors present during facial expression recognition, Shiwei et al. [47] proposed a fine-grained method of global and local feature fusion, which achieved good classification performance through multi-stage preprocessing, such as super-resolution processing, illumination and shadow processing, and texture enhancement, before using a double-branch convolutional neural network to fuse global and local features.
Because the fuzzy annotation in the dataset will affect the performance of the model, Shu Liu et al. [48] proposed solving FER tasks through the label distribution learning paradigm and developed a two-branch adaptive distribution fusion (ADA-DF) framework, which uses the attention module to give high weights to the determined samples and low weights to the uncertain samples. The model achieved high accuracy on RAF-DB, AffectNet, and SFEW.
Amine Bohi et al. [49] proposed using ConvNeXt as the basic framework to enhance the robustness of the model with respect to scale, rotation, and translation changes and integrated an SE module to model the dependency between channels, highlight key features, and add self-attention regularization to enhance the balance and discrimination of feature expressions by balancing the attention weight. Ablation experiments showed that each module of EmoNext [50] (STN, SE, and SA regular terms) achieves significant performance improvement.
In 2025, Cao et al. [51] proposed a fuzzy perception recognition framework, COA, to address the issue of “wrong labels” and “fuzzy labels” in the sample. The COA framework introduces the emotion extraction and expression description modules. The former extracts rich semantic features from multi-scale and local attention mechanisms, respectively, via the coupled flow structure; the latter generates tag pairs by accurately predicting the probability, accurately describes the multiple emotional attributes of fuzzy samples, and better guides the model to fit mixed features during training. In the study, experiments were conducted on five datasets: RAF-DB, FERPlus, AffectNet, SFEW, and CAER-S. The results show that COA has significant advantages and robustness in dealing with a high proportion of fuzzy samples.
In 2024, Ngwe et al. [52] addressed the challenge of FER under natural and complex conditions. They proposed the lightweight network Patt-Lite, which integrates a patch extraction block and an attention classifier based on the MobileNetV1 architecture to improve recognition performance. Experiments on CK+, RAF-DB, FER-2013, FERPlus, and other datasets showed that Patt-Lite achieves 100% accuracy on CK+ and 95.05%, 92.50%, and 95.55% accuracy on RAF-DB, FER-2013, and FERPlus, respectively; these outcomes are significantly better than those of existing lightweight methods and some transformer models. The parameter is only 1.10 M, demonstrating the method’s robustness under complex conditions.

3.2.4. Transformer-Based Method

The original transformer model is a deep learning architecture based on a self-attention mechanism proposed by Vaswani et al. [53] in 2017. Unlike the traditional RNNs or CNNs, the transformer completely discards the cyclic structure and convolutional operations and relies solely on the self-attention mechanism to model dependencies between elements at different positions in the input sequence. It can simultaneously focus on global and local information in the sequence, significantly improving the model’s ability to understand context semantics. For facial expression recognition, the transformer-based model can effectively capture the subtle changes in key areas of the face (such as eyes, mouth, and eyebrows) and their correlation, focus on the discriminant features related to expression through the multi-level attention mechanism, and alleviate the influence of interference factors such as light and posture change.
In recent years, researchers have further improved the accuracy and robustness of the model in expression recognition by improving the input representation of the transformer (such as introducing facial key point information) and optimizing the attention mechanism (such as combining local attention and global attention), making it an important research direction in this field. In 2024, Raman et al. [54] proposed a multimodal facial expression recognition (MMFER) framework based on a hierarchical cross-attention graph convolutional network (HCA-GCN), which improves recognition accuracy and robustness by fusing multimodal data such as visual images, audio signals, and contextual information. Firstly, a CNN, a Short-Time Fourier Transform (STFT) and a pre-trained transformer are used to extract modal features. Then, a cross-modal graph structure is constructed by layering graph convolutions. First, the internal features of the single mode are refined, and then the dynamic correlation between modes is captured. Finally, the dynamic weighted features of the layers are fused via cross-attention, and the resulting features are classified using softmax.
In 2025, Zhou et al. [55] proposed a high-performance facial expression recognition model, HPC, including a high-dynamic-range preprocessing module (HDR preprocessing module), a parallel attention vision transformer structure (parallel attention vision transformer), and a co-expression head. During the preprocessing stage, the HDR preprocessing module optimizes the input image through local contrast and detail enhancement technology to enhance the adaptability of the model to light and detail changes. During the feature processing stage, the parallel attention visual transformer structure uses the multi-head self-attention mechanism encoder to capture and process the multi-scale facial expression features effectively and realizes the detailed analysis of subtle expression differences. In the feature fusion stage, the cooperative expression head uses the cooperative expression mechanism to efficiently process and optimize the features of different expression states.
In 2024, Liu et al. [56] proposed a framework based on human expression-sensitive cues (HESPs) to capture subtle expression dynamics by improving the clip model, addressing the challenge of clip models being too complex to recognize unknown expression categories in open-set video facial expression recognition (OV-FER). The framework includes three modules: the text prompt module introduces the learnable emotion template and negative text prompt to optimize the expression ability of the text encoder for known/unknown emotions; the visual cue module uses the class activation map to locate the expression sensitive area, combining with the time global pool to aggregate the video sequence characteristics; and the open-set multi-task learning scheme enhances the ability of cross-modal feature alignment and unknown category generalization through known category loss, negative representation loss, and global prediction loss.
In 2025, Mao et al. [57] proposed a lightweight, improved version of Poster++ to address the high computational cost of the Poster [58] model in FER, achieving improved efficiency and performance by optimizing the dual-flow design, cross-fusion mechanism, and multi-scale feature extraction. The model removes the image from the original dual-flow structure and routes it only to the landmark branch, while retaining the landmark-to-image branch to focus on key expression features. At the same time, the window-based Multi-Head Cross Self-Attention (W-MCSA) mechanism replaces traditional cross attention, reducing computational complexity from quadratic to linear and enabling feature fusion by directly extracting multi-scale features from the backbone network and combining them with the two-layer transformer module.

3.2.5. Graph Structure-Based Method

In recent years, GNNs have directly modeled non-European spatial data, making them particularly suitable for processing graph-structured data composed of facial key points, providing a new idea for expression recognition. The graph-based expression recognition process comprises three steps:
(1)
Construct the face image structure: Hundreds of facial key points are extracted using face key point detection methods (such as Dlib, MediaPipe, or OpenFace), and each key point is used as a graph node. Then, the adjacency matrix is generated using k-nearest neighbors (KNNs), full connectivity, or prior knowledge of facial anatomical structure, so that the relationship between nodes reflects local facial geometry.
(2)
Graph neural network modeling: After obtaining the graph structure, the graph convolutional network (GCN) or its improved version is used to extract the features of the graph data. Node features (such as keypoint coordinates, texture features, or motion features) are updated via the graph convolutional layer to transmit and aggregate information between nodes. This enables the model to capture interactions among local facial regions, thereby enhancing its ability to model complex expressions.
(3)
Classification and identification: After several layers of graph convolution, node features are usually transformed into a graph-level representation through global pooling. Finally, the fully connected layer and softmax classifier are used to output specific expression categories.
Many studies have demonstrated the effectiveness of the graph-based expression recognition method. For example, the FERGCN model proposed by Liao et al. [59] uses the face’s key points as nodes, constructs an adjacency matrix to capture the face’s geometric relationships, and performs feature propagation via a graph convolutional network. This method can capture the dependence between local areas of the face, especially in highly sensitive regions such as the eyebrows, eyes, nose, and mouth corners. Liu et al. [60] proposed a video expression recognition method based on graph convolution. Firstly, CNNs are used to extract local features from different regions of the face (e.g., eyes, mouth, and eyebrows), which are modeled as graph nodes. The edges between nodes connect region features, and GCN learns the interactions between regions.
Furthermore, to capture the temporal dynamics of expression, this method combines an LSTM to model temporal features and enables joint learning of space and time. Antoniadis et al. [61] proposed an emotion GCN model to address the significant co-occurrence and semantic dependence among expressions. This method constructs a graph in the label space, treats different emotion categories as nodes, and establishes edges based on statistical co-occurrence or semantic relationships, enabling GCN to explicitly model dependencies between categories. In addition, the method combines two expression modeling methods—classification and regression—and further improves the model’s generalization ability through multi-task learning.

3.2.6. RNN-Based Approach

An RNN is a type of deep learning model designed for processing sequential data. Its core feature is that, by introducing cyclic connections, the model can retain historical input information and apply it to the current time calculation, thereby naturally capturing temporal dependence in the sequence. RNNs are often used to process video sequences, dynamic images, and other information that contain a time dimension. For facial expression recognition, especially dynamic expression recognition, RNNs can effectively model the evolution of expression over time, capture the transition characteristics from neutral expression to target expression, and determine the temporal correlation of the movement of various regions of the face (such as the speed of rising corners of the mouth, the amplitude of frowning, etc.). For example, by inputting the video frame sequence into an RNN, the model can learn the dynamic patterns of different expressions on the time axis and then improve the recognition ability of complex expressions (such as micro expressions) or fuzzy expressions.
Furthermore, traditional RNNs suffer from gradient disappearance or gradient explosion, making it challenging to handle long-term sequence dependence. By introducing input, forgetting, and output gates to control the inflow and outflow of information, LSTM can effectively alleviate the gradient problem and better capture the characteristics of long-term time series. Gated Recurrent Unit (GRU) simplifies LSTM gating and reduces computational complexity while maintaining performance. These improved models have been widely used in facial expression recognition, especially when combined with dynamic information such as facial keypoint sequences and optical flow, thereby significantly improving the model’s sensitivity to temporal changes and recognition stability. Güler et al. [62] proposed a multimodal emotion recognition framework integrating electroencephalogram (EEG) signals and facial expressions to address the limitations of single-modality approaches. They used deep learning models such as GRUs, LSTMs, and transformers to explore the synergistic effect of multi-source data on emotion classification. This method also needs to solve the problem of modal dimension mismatch. In 2025, Yuanlun Xie et al. [63] proposed an end-to-end low-illumination FER framework (LL-FER), which realizes the collaborative optimization of image enhancement and expression recognition by jointly training the low-illumination enhancement network (LLENet) and the FER network.

4. Experimental Comparison

Although existing reviews cover a wide range of methods, most rely on data from original studies for comparison, lacking a unified experimental environment. The core contribution of this study is the establishment of a unified benchmark framework that rigorously compares 12 methods across different technical paradigms within the same hardware environment, data partitioning, and hyperparameter settings.
The selection of the 12 assessed models is based on three criteria. The first criterion is the historical baseline. We have chosen traditional manual feature extraction methods such as LBP, GLCM, and HOG. Although they are no longer SOTA, as classic representatives of the ‘pre-deep learning era’, they provide important baselines for benchmark testing, are representative, and reflect the evolutionary representativeness of FER technology. Based on this criterion, four traditional machine learning methods, namely GLCM-SVM [32], HOG-SVM [64], LBP-SVM [30], and LPQ-SRC [35], are selected.
The second criterion is paradigm coverage. The chosen method should cover the mainstream architectures of the deep learning era to ensure the breadth of method evaluation. According to this criterion, eight DL-based models, namely ADA-DF [48], CNN-SIFT [65], EmoNext [50], FERGCN [59], GCN [66], ResEmoteNet [46], ResNet [67], and Poster (Pyramid Cross-Fusion Transformer) [58], are selected. Among them, ResNet represents the classic convolutional neural network architecture. Poster and EmoNext are the most advanced attention-based methods currently available. FERGCN is a structured modeling approach for non-Euclidean data. ADA-DF is a specialized method for solving the problems of “class imbalance” and “label ambiguity”.
The third criterion is reproducibility. To ensure the fairness of this benchmark, we have selected models that provide official open-source code, thereby ensuring the reliability of the experimental results. This enables us to reproduce experiments with new datasets, rather than relying solely on the metric values reported in the original references.
According to the three criteria, 12 models are selected, and their publication years, method characteristics, and environmental configurations are listed in Table 2.

4.1. Experimental Data

The comparative experiment was conducted using four datasets: AffectNet [68], FERPlus [69], FER-2013 [70], and RAF-DB [71]. The first two datasets both contain images of eight different facial expressions; the last two datasets both contain images with seven different facial expressions.

4.1.1. AffectNet

AffectNet (2017) [68] is one of the largest and most challenging facial expression recognition datasets currently in use, and it is used to support research on expression recognition, emotion analysis, and multidimensional emotion modeling. It was built by the research team of Ohio State University, who collected more than 1 million facial images from the Internet. These images were automatically crawled using a search engine with emotional keywords and manually labeled. AffectNet provides two annotation methods: The first method uses seven (or eight) categories of basic emotional labels, including neutral, happy, sad, angry, surprised, disgusted, fear and contempt. The second method annotates the continuous emotion dimension, that is, each image has the corresponding value of happiness and arousal, which can be used for regression tasks and emotion distribution modeling. About 450,000 images in the dataset have human-labeled labels, and more than 500,000 have automatically generated labels. AffectNet reflects the complexity of the “real world” in terms of emotional category imbalance, posture change, expression subtlety, etc. It is one of the most important benchmark datasets for evaluating deep learning expression recognition models.

4.1.2. FERPlus

FERPlus (face expression recognition plus 2016) [69] is an enhanced version of Microsoft Research relabeling based on the FER-2013 “wild” expression dataset produced in 2016: it retains the original 35,887 48 × 48 gray face images and the original training/public/private triad structure, but introduces the voting results of 10 crowdsourcing taggers for each image, as well as gives the count distributions for eight emotions (neutral, happy, surprised, sad, angry, disgusted, fear, contempt) and two auxiliary tags (“unknown” and “non face”); this ensures that either a single tag can be obtained with a majority vote or multi-tag or soft tag learning can be directly carried out using the probability distribution. Despite challenging low-resolution, real-world conditions, FERPlus has become a standard benchmark for studying noise-robust learning, expression distribution modeling, and depth network robustness.

4.1.3. FER-2013

FER-2013 (Facial Expression Recognition 2013) [70] is a widely used facial expression recognition dataset, which Kaggle initially released in the 2013 “challenges in representation learning” competition. The dataset contains 35,887 grayscale face images, each with 48 × 48 pixels. Its data were collected from Google image searches and cover seven basic emotions: anger, disgust, fear, happiness, sadness, surprise, and neutrality. The images are all face regions subjected to automatic detection and alignment. Although there is a certain degree of noise and annotation error, its large scale, standard division (28,709 training sets, 3589 verification sets, 3589 test sets), and wide use make it one of the most classic benchmark datasets in the field of expression recognition. FER-2013 is often used for training and evaluating deep learning models, such as CNNs, ResNets, and VGGs. It plays an important role in robustness research for expression recognition, emotion classification, multi-task learning, and related areas.

4.1.4. RAF-DB

RAF-DB (Real World Effective Faces Database 2017) [71] was proposed by Shan and Deng in the IEEE Transactions on Image Processing in 2017. It contains about 30,000 “in the wild” face images. Through crowdsourcing, each image is labeled by about 40 taggers on average, thus ensuring high reliability. The dataset includes 7 basic expressions (such as happy, sad, angry, etc.) and 12 complex expressions, and it provides face-keypoint information. It is widely used for facial expression recognition, complex emotion modeling, and robust learning and is an important benchmark dataset in the field.
Our experimental datasets were downloaded from Kaggle. Table 3 provides a summary of the original dataset, and Table 4 provides a summary of the number of various expressions in the datasets used in our experiment.

4.2. Implementation Details

In our experiments, we conducted analyses across two distinct computational setups tailored to model complexity. For traditional machine learning methods (GLCM-SVM, HOG-SVM, LBP-SVM, LPQ-SRC), we utilized a local workstation powered by an Intel(R) Core(TM) i9-14900HX CPU @ 2.20 GHz with 16 GB RAM. Due to the complexity of deep learning models and the large amounts of training data, they were trained on remote high-performance servers equipped with dedicated accelerators. Specifically, ADA-DF, EmoNext, ResEmoteNet, ResNet, and Poster were trained on a single NVIDIA RTX 4060 GPU (CUDA 12.4, PyTorch 2.5.1), while CNN-SIFT, FERGCN, and GCN were executed on a single NVIDIA Tesla V100 GPU (CUDA 12.8/9.0.176, PyTorch 2.8.0/1.1.0). The software environment spanned Python versions 3.7 to 3.11, ensuring compatibility with the original implementations.
The experimental settings of the four traditional methods are as follows:
  • GLCM-SVM: For the GLCM-SVM method, we implemented a hybrid feature extraction pipeline designed to capture texture statistics from quantized patterns. Input images were first transformed into Uniform LBP maps ( R = 1 ,   P = 8 ) and subsequently quantized into 15 discrete levels (DLBP). On these quantized maps, GLCMs were computed using three distances ( d     { 1 ,   2 ,   3 } ) and four orientations ( 0 °, 45°, 90°, 135°). Six statistical descriptors—Contrast, Dissimilarity, Homogeneity, ASM, Energy, and Correlation—were extracted and aggregated via their mean and standard deviation to form the final feature vector. The classification was performed using a linear SVM with a regularization parameter C = 1.0 and a maximum of 5000 iterations, utilizing a random 80/20 train–test split strategy.
  • HOG-SVM: The HOG-SVM approach utilized the histogram of oriented gradients descriptor implemented via OpenCV to capture local shape and edge information. Input images were resized to 128   ×   128 pixels. We configured the HOG descriptor with a cell size of 8   ×   8 pixels, a block size of 16   × 16 pixels, and a block stride of 8   ×   8 pixels. Gradient orientations were discretized into nine bins. Given the high dimensionality of the resulting feature vector, we employed an SVM with a Radial Basis Function (RBF) kernel. The regularization parameter C was set to 10, and the kernel coefficient γ was set to ‘scale’. Similarly to the GLCM method, the dataset was split randomly, with 80% for training and 20% for validation.
  • LBP-SVM: For the LBP-SVM method, we adopted a feature fusion strategy to combine global texture patterns with local keypoint descriptors. Texture features were extracted using uniform LBP ( R = 1 ,   P = 8 ) and represented as a normalized histogram. Simultaneously, ORB descriptors were computed with a maximum limit of 200 keypoints, which were then aggregated using mean pooling. These two feature sets were concatenated to form the final representation. Classification was conducted using a linear SVM ( C = 1.0 ,   m a x _ i t e r = 5000 ). Unlike the randomized splits used in other baseline methods, this experiment strictly followed the pre-defined training–validation splits provided by the respective datasets, with 80% for training and 20% for validation.
  • LPQ-SRC: The LPQ-SRC method employed local phase quantization features, which are known for their robustness to blur and illumination variations. The LPQ descriptor was configured with a local window size of 3   ×   3 , a frequency estimation parameter α = 1.0 , and a correlation coefficient ρ = 0.90 . The resulting feature codes were accumulated into a 256-dimensional histogram and normalized using a standard scaler. For classification, we utilized a sparse representation-based classifier (SRC). The overcomplete dictionary was constructed using all available training samples. The sparse coding problem was solved using the Orthogonal Matching Pursuit (OMP) algorithm with a sparsity constraint of K = 30 non-zero coefficients, and the final class label was determined based on the minimum reconstruction residual.
The training superparameters of the deep learning models are listed in Table 5. We optimized the solver configuration for different architectures: most models (ADA-DF, FERGCN, ResNet) use the Adam optimizer, while GCN and ResEmoteNet use SGD. Poster uses a more advanced SAM (sharpness-aware minimization) optimizer to improve its generalization ability. The learning rate falls within [ 4 × 10 5 , 5 × 10 3 ] according to the convergence characteristics of the model. All models were trained for 100 or 200 epochs, and the batch size was set to 16, 32, or 64 to accommodate the video memory limit.

4.3. Analysis of Experimental Results

The following two sections compare the 12 FER methods listed in Table 2 in terms of performance and classification effect. The comparative experiment is based on the four datasets of AffectNet, FERPlus, FER-2013, and RAF-DB introduced in Section 4.1.

4.3.1. Performance Comparison of Various Models

The experimental results of four traditional methods and eight deep learning methods across four natural datasets are listed in Table 6. In this table, we can see that the depth model (ADA-DF [48], EmoNext [50], GCN [66], ResNet [67], ResEmoteNet [46], and Poster [58]) can generally achieve high accuracy on the training set. Some methods such as AffectNet and RAF-DB can even exceed 95% while also maintaining strong performance on the validation set, especially on the “in the wild” datasets, and the validation accuracy of the depth model is significantly higher than those of the traditional methods, showing good generalization ability and robustness to complex environments (such as different lighting, posture, and occlusion). In contrast, the traditional prior feature methods (CNN-SIFT [65], GLCM-SVM [32], HOG-SVM [64], LBP-SVM [30] and LPQ-SRC [35]) perform pretty well on the training set. However, the accuracy on the validation set is generally low, especially on large-scale, real-world images, which are difficult to adapt to expression diversity and label noise, leading to significant performance degradation. In addition, it is worth noting that a deep learning model must not follow the assumption of the more complex, the better. Although ResEmoteNet [46] has a large parameter count, its performance across multiple datasets is not the best, and it exhibits some overfitting. Moreover, although the traditional HOG-SVM [64] method performs nearly perfectly on the training set, its performance on the verification set is clearly insufficient, indicating its limited generalization ability.
In contrast, the ADA-DF [48] model achieves better recognition performance while maintaining low complexity through its double-branch structure, thereby verifying the effectiveness of constructing the loss function based on the actual label distribution. Particularly on the RAF-DB dataset, the model’s recognition accuracy is 89.44%, demonstrating strong generalization. However, there is still room for improvement in its performance on other datasets. The GCN [66] model achieves good results across all expression recognition categories. Moreover, its performance is stable. The effectiveness of the Gabor feature in expression recognition is verified.
In expression recognition, in addition to accuracy, the number of parameters (trainable params) and computational complexity (MACS) of deep learning models are important indicators for evaluating their lightweight degree. Despite the high computational efficiency and low cost of traditional machine learning methods, the classification effect is poor; this study focuses on comparing the structural complexity of eight deep learning models: ADA-DF [48], CNN-SIFT [65], EmoNext [50], FERGCN [59], GCN [66], ResEmoteNet [46], Poster [58] and standard ResNet [67].
Table 7 shows significant differences in the number of parameters and calculations across different facial expression recognition models. High-performance deep networks such as ADA-DF [48], EmoNext [50], Poster [58], and ResEmoteNet [46] have tens of millions of parameters and G-level MACs. These models have strong expression capabilities and are suitable for training and reasoning on high-performance GPUs. The parameter count of the traditional convolutional network, ResNet [67], is relatively small. However, due to the computational characteristics of the convolutional operation, its MAC count reaches 193.3 M, which still imposes a significant computational burden. The graph convolutional models GCN [66] and FERGCN [59] exhibit lightweight characteristics, especially FERGCN [59], which has only 4.872 K parameters and 0.593 M MACs, making it very suitable for resource-constrained scenarios. CNN-SIFT [65], as a method combining a priori features with a shallow network, is at a medium level in terms of the number of parameters and calculations, taking into account specific performance and efficiency.
We take eight kinds of expressions as examples. The input image size for EmoNext and Poster is 3 × 224 × 224 pixels. In contrast, the GCN model uses an input image size of 1 × 90 × 90 pixels. Other models have an input image size of 3 × 120 × 120 pixels. Additionally, the parameters and floating-point operations of each model are compared.
Overall, there is an apparent trade-off between the capacity and computational cost of facial expression recognition models. A deep network with a high number of parameters and high MACs can provide stronger learning ability, but the cost of training and deployment is high; lightweight models such as FERGCN [59] occupy very low computing resources and are suitable for real-time or embedded applications, but their limited capacity may affect their performance in complex scenarios. Models representing a combination of the two, such as ResNet [67] and CNN-SIFT [65], strike a balance between performance and efficiency. Therefore, in practical applications, it is necessary to select appropriate models according to task complexity, computing resources, and real-time requirements.

4.3.2. Analysis of Identification of Each Category

We analyze the recognition accuracy of each model in each dataset. The recognition accuracy of a category is equal to the ratio of the number of correctly predicted pictures in the category to the total number of pictures in the category. The classification accuracies of each model on the validation sets of the four datasets are listed in Table 8, Table 9, Table 10 and Table 11.
Table 8, Table 9, Table 10 and Table 11 show that the deep learning method is clearly superior to the traditional prior-feature method in facial expression recognition as a whole. The depth model, represented by ResNet [67], ADA-DF [48], EmoNext [50], ResEmoteNet [46], Poster [58] and GCN [66], can achieve high recognition accuracy in most expression categories, especially in common expression categories such as happy, neutral, and surprise, which shows that the depth model has significant advantages in feature extraction and complex pattern modeling. In particular, the ADA-DF [48], Poster [58], and GCN [66] models achieve higher than 90% accuracy in some categories, reflecting their strong adaptability to high-quality annotations and grayscale images, as well as their ability to capture the fuzziness and diversity of expressions. From a global perspective, in “in the wild” datasets such as AffectNet and RAF-DB, the depth model can still maintain high performance. Recognition of some small sample categories, such as sadness, anger, and contempt, is slightly low. However, the overall accuracy is still significantly higher than those of traditional methods, reflecting the robustness of the depth method across complex scenes, varying lighting conditions, pose changes, and occlusions.
In contrast, traditional prior-feature methods (such as LBP-SVM [30], HOG-SVM [64], GLCM-SVM [32], CNN-SIFT [65], etc.) may achieve a certain level of accuracy on the training set. However, they are unstable on the verification set and have limited adaptability to rare expression categories and complex environments. The accuracy levels of some methods in a few categories are even below 20%, indicating that they struggle to capture the diversity and subtle changes in facial expressions fully.
Table 12 shows the average accuracies of both the ML-based and DL-based methods in recognizing different facial expressions. The average accuracy is calculated as a simple average of the data in Table 8, Table 9, Table 10 and Table 11. In Table 12, we can see that the most difficult expression to recognize is contempt, with recognition accuracies of 10.00% and 40.06% for ML-based and DL-based methods, respectively. The accuracy of the ML-based method for identifying neutral is the highest at 35.25%; the DL-based method has the highest accuracy for identifying surprise, with 69.56%.
Overall, these results show that deep feature extraction and structured modeling methods can more comprehensively and effectively represent facial expression features, which is the mainstream and the most effective technical approach in current facial expression recognition research, and they also provide a solid foundation for further research on compound expression recognition, multimodal emotion understanding, and robustness improvement.

4.4. Discussion

Our experiment reveals a significant phenomenon: traditional machine learning methods (such as HOG-SVM and LBP-SVM) experience severe performance decline on natural scene datasets such as AffectNet and RAF-DB. The fundamental reason is the limitation of the feature extraction mechanism. Traditional descriptors are highly dependent on strict pixel alignment and local texture consistency. However, in the field, uncontrollable illumination, pose variations, and occlusion distort the distribution of these rigid features. In contrast, deep learning models (such as ResNet and Poster) extract high-level semantic features with spatial invariance via hierarchical convolutions or an attention mechanism, thereby adapting to complex environmental interference.
In addition, through in-depth analysis of accuracy differences across various categories (see Table 8, Table 9, Table 10 and Table 11), we observed that all methods performed better on RAF-DB in the “happy” and “neutral” categories than in the “sad” and “disgust” categories. This performance deviation directly reflects the dataset’s long-tail distribution (as shown in Table 4). Due to the scarcity of minority samples, the model tends to overfit the majority during training to reduce the global loss, resulting in a fuzzy decision boundary for the minority. This shows that simply improving the model architecture cannot completely solve this problem, and future research must be combined with data resampling or cost-sensitive learning strategies.

5. Conclusions and Future Work

In this study, we surveyed facial expression recognition models from recent decades that employ both machine learning and deep learning methods. Machine learning-based methods were reviewed across six subcategories: wavelet transform, LBP, HOG, GLCM, LPQ, and SIFT. Deep learning-based methods were reviewed across six subcategories: CNNs, ResNets, attention mechanisms, transformers, graph structures, and RNNs. In this way, the technical details of various methods over the past few years can be seen at a glance, a feature that sets this survey apart from many existing surveys.
We examined four datasets and compared twelve models through a series of experiments. Our comparative experiments demonstrate the following: (1) The results indicate that deep learning-based methods outperform traditional machine learning methods on the validation sets. (2) Models with the highest number of parameters or floating-point operations (MACs) do not always perform the best. For example, although ResEmoteNet has the largest number of parameters (80.24 M), its accuracy on RAF-DB is lower than that of Poster with fewer parameters (71.8 M). This discovery emphasizes an important insight: efficient architecture design, such as effective feature fusion or attention mechanisms, is more critical than simply increasing network depth or parameter size. (3) Class imbalance remains the most significant bottleneck restricting the improvement of FER accuracy. All evaluated models, regardless of their architecture, exhibit significant performance degradation in categories with few samples. This indicates that the majority of classes largely dominate the current performance improvement.
Based on the literature review and the results of our comparative experiments, future research can be further developed in the following areas:
For dataset enhancement, in view of the widespread category imbalance in facial expression recognition tasks, we should strengthen the modeling and enhancement of a small number of emotion samples; at the same time, it is necessary to consider the complex factors in the real scene, such as blur, occlusion (e.g., faces with masks), illumination change, etc., to test the robustness of a model. In addition, facial expressions are related to ethnic and cultural habits. The observer’s cultural background will significantly adjust their perception and judgment of expression [72], so it is necessary to establish culturally relevant facial expression datasets to improve the specificity of expression recognition. Cross-cultural datasets can also be established to examine the generalization of an FER method.
Regarding model accuracy, although existing models have achieved good performance in expression recognition, some fuzzy expressions (as shown in Figure 4) are still not recognized accurately, indicating that they require further improvement.
Noting the outstanding performances of large language models in assisting scientific research over the past two years, exploring efficient feature-extraction and fusion mechanisms for FER with large language models is a promising direction for the coming years. In addition, because the emotional information conveyed by the face is limited, we can develop a multimodal emotion recognition model that combines it with other types of data containing emotional information to improve recognition accuracy.
In the training strategy, we can leverage multi-task learning, transfer learning, and label distribution-based modeling further to improve a model’s cross-dataset and cross-scene generalization performance and enhance its practicability and adaptability.
In terms of application prospects, with the rapid development of digital human technology and humanoid robots, facial expression recognition, as one of the key technologies for natural human–computer interaction, will offer broad application value across related fields. For example, in intelligent education, expression recognition can be used to monitor learners’ emotional states in real time. In medical and health care, it can provide new means for the auxiliary diagnosis of emotional disorders and pain intensity [73,74]. In intelligent customer service and human–computer interaction, it helps to improve the naturalness and affinity of the system. Finally, in safety monitoring and driving assistance, it can provide support for abnormal emotion or fatigue detection.

Author Contributions

Conceptualization, Y.L. and H.L.; validation, Y.L. and H.L.; formal analysis, Y.L. and Q.F.; investigation, Y.L. and H.L.; resources, Q.F.; data curation, Y.L.; writing—original draft preparation, Y.L. and H.L.; writing—review and editing, Y.L., H.L. and Z.Z.; visualization, Y.L.; supervision, Q.F. and H.L.; project administration, Q.F. and H.L.; funding acquisition, Q.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by State Key Laboratory of Aerodynamics, grant number SKLA-JSSX-2024-KFKT-05; and the Beijing Science and Technology Plan Project, grant number Z231100005923035.

Data Availability Statement

The data used in this study are openly available. AffectNet: https://www.kaggle.com/datasets/mstjebashazida/affectnet, accessed on 15 October 2025; FERPlus: https://www.kaggle.com/datasets/arnabkumarroy02/ferplus, accessed on 15 October 2025; FER-2013: https://www.kaggle.com/competitions/challenges-in-representation-learning-facial-expression-recognition-challenge/data, accessed on 15 October 2025; RAF-DB: https://aistudio.baidu.com/datasetdetail/339643, accessed on 15 October 2025.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AdaBoostAdaptive Boosting
CBAMConvolutional Block Attention Module
CLAHEContrast-Limited Adaptive Histogram Equalization
CNNConvolutional Neural Network
DLBPDistinctive Local Binary Pattern
FERFacial Expression Recognition
FGDFull Gradient Descent
GLCMGray Level Co-occurrence Matrix
GCNGraph Convolution Network
GNNGraph Neural Network
HCA-GCNHierarchical Cross-Attention Graph Convolution Network
HDRHigh Dynamic Range
HOGHistogram of Oriented Gradients
LBPLocal Binary Pattern
LDALinear Discriminant Analysis
LPQLocal Phase Quantization
MGDMini-batch Gradient Descent
MTCNNMulti-Task Cascaded Convolutional Networks
ORBOriented Fast And Rotated Brief
OV-FEROpen-Set Video Facial Expression Recognition
PCAPrincipal Component Analysis
QHEQuaternion Histogram Equalization
QVTQuaternion-valued Transformer
QWTRQuaternion Wavelet Transformer
QWT-FSQuaternion Wavelet Transform-based Feature Selection
RI-LPQRotation Invariant Local Phase Quantization
RNNRecurrent Neural Network
SESqueeze-and-Excitation
SGDStochastic Gradient Descent
SIFTScale-Invariant Feature Transform
SLTLBPSum of Lower Triangle LBP
SSIMStructural Similarity Index Measure
SUTLBPSum of Upper Triangle LBP
SVMSupport Vector Machine
W-MCSAWindow-based Multi-Head Cross Self-Attention
WTWavelet Transform

References

  1. Tombs, A.G.; Russell-Bennett, R.; Ashkanasy, N.M. Recognising Emotional Expressions of Complaining Customers: A Cross-Cultural Study. Eur. J. Mark. 2014, 48, 1354–1374. [Google Scholar] [CrossRef]
  2. Meyer-Lindenberg, H.; Moessnang, C.; Oakley, B.; Ahmad, J.; Mason, L.; Jones, E.J.H.; Hayward, H.L.; Cooke, J.; Crawley, D.; Holt, R.; et al. Facial Expression Recognition Is Linked to Clinical and Neurofunctional Differences in Autism. Mol. Autism 2022, 13, 43. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, S.; Zhao, X.; Lei, B. Facial Expression Recognition Based on Local Binary Patterns and Local Fisher Discriminant Analysis. WSEAS Trans. Signal Process. 2012, 8, 21–31. [Google Scholar]
  4. Li, S.; Deng, W. Deep Facial Expression Recognition: A Survey. IEEE Trans. Affect. Comput. 2022, 13, 1195–1215. [Google Scholar]
  5. Hasani, B.; Mahoor, M.H. Spatio-Temporal Facial Expression Recognition Using Convolutional Neural Networks and Conditional Random Fields. In Proceedings of the 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017. [Google Scholar]
  6. Khomidov, M.; Lee, J.-H. The Novel EfficientNet Architecture-Based System and Algorithm to Predict Complex Human Emotions. Algorithms 2024, 17, 285. [Google Scholar]
  7. Yang, H.; Ciftci, U.; Yin, L. Facial Expression Recognition by De-Expression Residue Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  8. Ekman, P.; Friesen, W.V. Constants Across Cultures in the Face and Emotion. J. Pers. Soc. Psychol. 1971, 17, 124–129. [Google Scholar] [CrossRef]
  9. Kanade, T.; Cohn, J.F.; Tian, Y. Comprehensive Database for Facial Expression Analysis. In Proceedings of the Fourth IEEE International Conference on Automatic Face and Gesture Recognition, Grenoble, France, 28–30 March 2000; pp. 46–53. [Google Scholar]
  10. Shan, C.; Gong, S.; McOwan, P.W. Facial Expression Recognition Based on Local Binary Patterns: A Comprehensive Study. Image Vis. Comput. 2009, 27, 803–816. [Google Scholar] [CrossRef]
  11. Lucey, P.; Cohn, J.F.; Kanade, T.; Saragih, J.; Ambadar, Z.; Matthews, I. The Extended Cohn-Kanade Dataset (CK+): A Complete Dataset for Action Unit and Emotion-Specified Expression. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), San Francisco, CA, USA, 13–18 June 2010; pp. 94–101. [Google Scholar]
  12. Mollahosseini, A.; Chan, D.; Mahoor, M.H. Going Deeper in Facial Expression Recognition Using Deep Neural Networks. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar]
  13. Li, Y.; Zeng, J.; Shan, S.; Chen, X. Occlusion Aware Facial Expression Recognition Using CNN With Attention Mechanism. IEEE Trans. Image Process. 2019, 28, 2439–2450. [Google Scholar]
  14. Du, S.; Martinez, A.M. Compound Facial Expressions of Emotion: From Basic Research to Clinical Applications. Dialogues Clin. Neurosci. 2015, 17, 443–455. [Google Scholar]
  15. Jack, R.E.; Garrod, O.G.B.; Yu, H.; Caldara, R.; Schyns, P.G. Facial Expressions of Emotion Are Not Culturally Universal. Proc. Natl. Acad. Sci. USA 2012, 109, 7241–7244. [Google Scholar]
  16. Yang, Q.; He, Y.; Chen, H.; Wu, Y.; Rao, Z. A Novel Lightweight Facial Expression Recognition Network Based on Deep Shallow Network Fusion and Attention Mechanism. Algorithms 2025, 18, 473. [Google Scholar] [CrossRef]
  17. Naga, P.; Marri, S.D.; Borreo, R. Facial Emotion Recognition Methods, Datasets and Technologies: A Literature Survey. Mater. Today Proc. 2023, 80, 2399–2403. [Google Scholar] [CrossRef]
  18. Kaur, M.; Kumar, M. Facial Emotion Recognition: A Comprehensive Review. Expert Syst. 2024, 41, e13670. [Google Scholar] [CrossRef]
  19. Kopalidis, T.; Solachidis, V.; Vretos, N.; Daras, P. Advances in Facial Expression Recognition: A Survey of Methods, Benchmarks, Models, and Datasets. Information 2024, 15, 135. [Google Scholar] [CrossRef]
  20. Ullah, S.; Ou, J.; Xie, Y.; Tian, W. Facial Expression Recognition (FER) Survey: A Vision, Architectural Elements, and Future Directions. PeerJ Comput. Sci. 2024, 10, e2024. [Google Scholar] [CrossRef]
  21. Rehman, A.; Mujahid, M.; Elyassih, A.; AlGhofaily, B.; Bahaj, S.A.O. Comprehensive Review and Analysis on Facial Emotion Recognition: Performance Insights into Deep and Traditional Learning with Current Updates and Challenges. Comput. Mater. Contin. 2025, 82, 41–72. [Google Scholar] [CrossRef]
  22. Haq, M.U.; Sethi, M.A.J.; Ben Aoun, N.; Alluhaidan, A.S.; Ahmad, S.; Farid, Z. CapsNet-FR: Capsule Networks for Improved Recognition of Facial Features. Comput. Mater. Contin. 2024, 79, 2169–2186. [Google Scholar] [CrossRef]
  23. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint Face Detection and Alignment Using Multi-Task Cascaded Convolutional Networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  24. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  25. Cootes, T.F.; Edwards, G.J.; Taylor, C.J. Active Appearance Models. In Computer Vision—ECCV 1998; Burkhardt, H., Neumann, B., Eds.; Springer: Berlin/Heidelberg, Germany, 1998; Volume 1407, pp. 484–498. [Google Scholar]
  26. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive Histogram Equalization and Its Variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  27. Tomasi, C.; Manduchi, R. Bilateral Filtering for Gray and Color Images. In Proceedings of the Sixth International Conference on Computer Vision (ICCV), Bombay, India, 4–7 January 1998; pp. 839–846. [Google Scholar]
  28. Bashyal, S.; Venayagamoorthy, G.K. Recognition of Facial Expressions Using Gabor Wavelets and Learning Vector Quantization. Eng. Appl. Artif. Intell. 2008, 21, 1056–1064. [Google Scholar] [CrossRef]
  29. Zhou, Y.; Pei, J.; Si, W.; Qin, J.; Heng, P.-A. Delving Into Quaternion Wavelet Transformer for Facial Expression Recognition in the Wild. IEEE Trans. Multimed. 2025, 27, 3895–3909. [Google Scholar] [CrossRef]
  30. Niu, B.; Gao, Z.; Guo, B. Facial Expression Recognition with LBP and ORB Features. Comput. Intell. Neurosci. 2021, 2021, 8828245. [Google Scholar] [CrossRef] [PubMed]
  31. Carcagnì, P.; Del Coco, M.; Leo, M.; Distante, C. Facial Expression Recognition and Histograms of Oriented Gradients: A Comprehensive Study. SpringerPlus 2015, 4, 645. [Google Scholar] [CrossRef]
  32. Satyanarayana Murty, G.; Kiran, J.S.; Kumar, V.V. Facial Expression Recognition Based on Features Derived From the Distinct LBP and GLCM. Int. J. Image Graph. Signal Process. 2014, 6, 68–77. [Google Scholar] [CrossRef]
  33. Zhang, Z. Feature-Based Facial Expression Recognition: Sensitivity Analysis and Experiments With a Multilayer Perceptron. Int. J. Pattern Recognit. Artif. Intell. 1999, 13, 893–911. [Google Scholar] [CrossRef]
  34. De la Torre, F.; Cohn, J.F. Facial Expression Analysis. In Visual Analysis of Humans: Looking at People; Moeslund, T.B., Hilton, A., Krüger, V., Sigal, L., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 377–409. [Google Scholar]
  35. Li, J.; Ying, Z. Facial Expression Recognition Based on Rotation Invariant Local Phase Quantization and Sparse Representation. In Proceedings of the Second International Conference on Instrumentation, Measurement, Computer, Communication and Control (IMCCC 2012), Harbin, China, 8–10 December 2012; pp. 798–801. [Google Scholar]
  36. Nikan, S.; Ahmadi, M. Local Gradient-Based Illumination Invariant Face Recognition Using Local Phase Quantisation and Multi-Resolution Local Binary Pattern Fusion. IET Image Process. 2015, 9, 12–21. [Google Scholar] [CrossRef]
  37. Lowe, D.G. Object Recognition from Local Scale-Invariant Features. In Proceedings of the Seventh IEEE International Conference on Computer Vision (ICCV), Kerkyra, Greece, 20–27 September 1999; pp. 1150–1157. [Google Scholar]
  38. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  39. Berretti, S.; Ben Amor, B.; Daoudi, M.; Del Bimbo, A. 3D Facial Expression Recognition Using SIFT Descriptors of Automatically Detected Keypoints. Vis. Comput. 2011, 27, 1021–1036. [Google Scholar] [CrossRef]
  40. Shi, Y.; Lv, Z.; Bi, N.; Zhang, C. An Improved SIFT Algorithm for Robust Emotion Recognition under Various Face Poses and Illuminations. Neural Comput. Appl. 2020, 32, 9267–9281. [Google Scholar] [CrossRef]
  41. Luo, Y.; Wu, J.; Zhang, Z.; Zhao, H.; Shu, Z. Design of Facial Expression Recognition Algorithm Based on CNN Model. In Proceedings of the 3rd IEEE International Conference on Power, Electronics and Computer Applications (ICPECA), Shenyang, China, 10–12 February 2023; pp. 883–886. [Google Scholar]
  42. Shehada, D.; Turky, A.; Rabie, T.; Hussain, A. Enhanced Lightweight Facial Emotion Recognition Systems for Visually Impaired People. In Proceedings of the 12th IEEE Conference on Systems, Process & Control (ICSPC), Malacca, Malaysia, 13–14 December 2024; pp. 268–273. [Google Scholar]
  43. Adiga, M.S.; Adarsh, M.J. Facial Expression Based Sentimental Analysis Using CNN. Int. J. Innov. Technol. Explor. Eng. 2024, 13, 45–49. [Google Scholar]
  44. Istiqomah, A.A.; Sari, C.A.; Susanto, A.; Rachmawanto, E.H. Facial Expression Recognition Using Convolutional Neural Networks with Transfer Learning Resnet-50. J. Appl. Inform. Comput. 2024, 8, 257–264. [Google Scholar] [CrossRef]
  45. Li, J. Comparison and Analysis of Deep Neural Networks in Facial Expression Recognition. Appl. Comput. Eng. 2023, 21, 93–98. [Google Scholar] [CrossRef]
  46. Roy, A.K.; Kathania, H.K.; Sharma, A.; Pandey, S.K.; Koundal, D. ResEmoteNet: Bridging Accuracy and Loss Reduction in Facial Emotion Recognition. IEEE Signal Process. Lett. 2024, 32, 491–495. [Google Scholar] [CrossRef]
  47. Li, S.; Wang, J.; Tian, L.; Wang, J.; Huang, Y. A Fine-Grained Human Facial Key Feature Extraction and Fusion Method for Emotion Recognition. Sci. Rep. 2025, 15, 6153. [Google Scholar] [CrossRef]
  48. Liu, S.; Xu, Y.; Wan, T.; Kui, X. A Dual-Branch Adaptive Distribution Fusion Framework for Real-World Facial Expression Recognition. In Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  49. Bohi, A.; Boudouri, Y.E.; Sfeir, I. A Novel Deep Learning Approach for Facial Emotion Recognition: Application to Detecting Emotional Responses in Elderly Individuals With Alzheimer’s Disease. Neural Comput. Appl. 2025, 37, 5235–5253. [Google Scholar] [CrossRef]
  50. Boudouri, Y.E.; Bohi, A. EmoNeXt: An Adapted ConvNeXt for Facial Emotion Recognition. In Proceedings of the 2023 IEEE 25th International Workshop on Multimedia Signal Processing (MMSP), Poitiers, France, 27–29 September 2023; pp. 1–6. [Google Scholar]
  51. Cao, X.; Luo, L.; Gu, Y.; Ren, F. Co-Dance with Ambiguity: An Ambiguity-Aware Facial Expression Recognition Framework for More Robustness. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 5431–5444. [Google Scholar] [CrossRef]
  52. Ngwe, J.L.; Lim, K.M.; Lee, C.P.; Ong, T.S.; Alqahtani, A. PAtt-Lite: Lightweight Patch and Attention MobileNet for Challenging Facial Expression Recognition. IEEE Access 2024, 12, 79327–79341. [Google Scholar] [CrossRef]
  53. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  54. Raman, D.R.; Kumar, P.M.; Sreenivas, T.V.; Manjunath, B.S. Multi-Modal Facial Expression Recognition Through a Hierarchical Cross-Attention Graph Convolutional Network. In Proceedings of the 2024 International Conference on Knowledge Engineering and Communication Systems (ICKECS), Chikkaballapur, India, 28–29 June 2024; pp. 1–6. [Google Scholar]
  55. Zhou, Y.; Kantathanawat, T.; Tuntiwongwanich, S.; Liu, C. High Dynamic Range Preprocessing, Parallel Attention Transformer and CoExpression Analysis for Facial Expression Recognition. Comput. Electr. Eng. 2025, 123, 110110. [Google Scholar] [CrossRef]
  56. Liu, Y.; Zhang, Z.; Li, Y.; Wang, Z. Open-Set Video-Based Facial Expression Recognition With Human Expression-Sensitive Prompting. In Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne, VIC, Australia, 21–25 October 2024; pp. 1–10. [Google Scholar]
  57. Mao, J.; Xu, R.; Yin, X.; Chang, Y.; Nie, B.; Huang, A. POSTER++: A Simpler and Stronger Facial Expression Recognition Network. Pattern Recognit. 2025, 157, 110951. [Google Scholar]
  58. Zheng, C.; Mendieta, M.; Chen, C. POSTER: A Pyramid Cross-Fusion Transformer Network for Facial Expression Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 15745–15754. [Google Scholar]
  59. Liao, L.; Zhu, Y.; Zheng, B.; Jiang, X.; Lin, J. FERGCN: Facial Expression Recognition Based on Graph Convolution Network. Mach. Vis. Appl. 2022, 33, 40. [Google Scholar] [CrossRef]
  60. Liu, D.; Zhang, H.; Zhou, P. Video-Based Facial Expression Recognition Using Graph Convolutional Networks. In Proceedings of the 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 6092–6099. [Google Scholar]
  61. Antoniadis, P.; Filntisis, P.P.; Maragos, P. Exploiting Emotional Dependencies With Graph Convolutional Networks for Facial Expression Recognition. In Proceedings of the 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021), Jodhpur, India, 15–19 December 2021; pp. 1–8. [Google Scholar]
  62. Güler, S.E.; Akbulut, F.P. Multimodal Emotion Recognition: Emotion Classification through the Integration of EEG and Facial Expressions. IEEE Access 2025, 13, 9954–9966. [Google Scholar] [CrossRef]
  63. Xie, Y.; Ou, J.; Wen, B.; Yu, Z.; Tian, W. A Joint Learning Method for Low-Light Facial Expression Recognition. Complex Intell. Syst. 2025, 11, 139. [Google Scholar] [CrossRef]
  64. Eng, S.K.; Ali, H.; Cheah, A.Y.; Chong, Y.F. Facial Expression Recognition in JAFFE and KDEF Datasets Using Histogram of Oriented Gradients and Support Vector Machine. IOP Conf. Ser. Mater. Sci. Eng. 2019, 705, 012031. [Google Scholar]
  65. Connie, T.; Al-Shabi, M.; Cheah, W.P.; Goh, M. Facial Expression Recognition Using a Hybrid CNN–SIFT Aggregator. In Multi-disciplinary Trends in Artificial Intelligence; Phon-Amnuaisuk, S., Ang, S.-P., Lee, S.-Y., Eds.; Springer International Publishing: Cham, Switzerland, 2017; Volume 10607, pp. 139–149. [Google Scholar]
  66. Jiang, P.; Wan, B.; Wang, Q.; Wu, J. Fast and Efficient Facial Expression Recognition Using a Gabor Convolutional Network. IEEE Signal Process. Lett. 2020, 27, 1954–1958. [Google Scholar] [CrossRef]
  67. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  68. Mollahosseini, A.; Hasani, B.; Mahoor, M.H. AffectNet: A Database for Facial Expression, Valence, and Arousal Computing in the Wild. IEEE Trans. Affect. Comput. 2019, 10, 18–31. [Google Scholar] [CrossRef]
  69. Barsoum, E.; Zhang, C.; Ferrer, C.C.; Zhang, Z. Training Deep Networks for Facial Expression Recognition With Crowd-Sourced Label Distribution. In Proceedings of the 18th ACM International Conference on Multimodal Interaction (ICMI), Tokyo, Japan, 12–16 November 2016; pp. 279–283. [Google Scholar]
  70. Goodfellow, I.J.; Erhan, D.; Carrier, P.L.; Courville, A.; Mirza, M.; Hamner, B.; Cukierski, W.; Tang, Y.; Thaler, D.; Lee, D.-H.; et al. Challenges in Representation Learning: A Report on Three Machine Learning Contests. In Proceedings of the International Conference on Neural Information Processing (ICONIP), Daegu, South Korea, 3–7 November 2013; pp. 117–124. [Google Scholar]
  71. Li, S.; Deng, W.; Du, J. Reliable Crowdsourcing and Deep Locality-Preserving Learning for Expression Recognition in the Wild. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2852–2861. [Google Scholar]
  72. Nagata, M.; Okajima, K. Effect of Observer’s Cultural Background and Masking Condition of Target Face on Facial Expression Recognition for Machine-Learning Dataset. PLoS ONE 2024, 19, e0313029. [Google Scholar]
  73. Ben Aoun, N. A Review of Automatic Pain Assessment from Facial Information Using Machine Learning. Technologies 2024, 12, 92. [Google Scholar] [CrossRef]
  74. Ben Aoun, N. Deep Learning-Based Pain Intensity Estimation from Facial Expressions. In Intelligent Systems Design and Applications; Abraham, A., Bajaj, A., Hanne, T., Hong, T.-P., Eds.; Lecture Notes in Networks and Systems; Springer: Cham, Switzerland, 2024; Volume 1047, pp. 495–504. [Google Scholar]
Figure 1. The general history of facial expression recognition [8,9,10,11,12].
Figure 1. The general history of facial expression recognition [8,9,10,11,12].
Algorithms 18 00800 g001
Figure 2. A flowchart of the operational process of most traditional machine learning FER algorithms. The input photo is derived from the AffectNet dataset.
Figure 2. A flowchart of the operational process of most traditional machine learning FER algorithms. The input photo is derived from the AffectNet dataset.
Algorithms 18 00800 g002
Figure 3. The general process of facial expression recognition using a deep learning method. The image is sourced from the AffectNet dataset. The second block is different from that in Figure 2.
Figure 3. The general process of facial expression recognition using a deep learning method. The image is sourced from the AffectNet dataset. The second block is different from that in Figure 2.
Algorithms 18 00800 g003
Figure 4. Four examples of recognition errors. Images are sourced from AffectNet dataset.
Figure 4. Four examples of recognition errors. Images are sourced from AffectNet dataset.
Algorithms 18 00800 g004
Table 1. A comparison of our survey with 5 survey references.
Table 1. A comparison of our survey with 5 survey references.
Ref.YearTechnique ReviewDatasetsFuture DirectionsComp Exp.
[17]2023FER steps: detection, feature extraction, augmentation11Mapping of the various elements involved in fulfilling the task of facial expression recognition in different studiesNo
[18]2024emotion theories; recognition pipeline; traditional methods; DL-based; applications7Cross-cultural; real-time emotion recognition; multimodal emotion; explainable AINo
[19]2024evolution timeline; pro-deep learning with pipeline; DL-based21Human identification mechanism; accelerated model; multiple modalities; thermal infrared imaging; model reduction; high performance on “in-the-wild” datasetNo
[20]2024FER basics; FER applications; IoT; Cloud computing0Compound emotions; cross cultural models; multi-pose FER; enhancing dataset; noise labelsNo
[21]2025ML-Based; DL-Based; HL-based7Provide more inclusive and diverse datasets; improve precision; mitigate bias; unsupervised and self-supervised learning methods.No
Ours--ML-Based; DL-Based. Both reviewed with 6 subcategories4Dataset enhancement; high-accuracy model; combining with a large language model; cross-dataset applicationsYes
Note: “Datasets” represents the number of datasets reviewed in each reference; “Comp exp.” is an abbreviation of comparative experiment.
Table 2. A comparison of the main information of 12 FER methods.
Table 2. A comparison of the main information of 12 FER methods.
IDMethodYearMethod CharacteristicsCategory
1GLCM-SVM [32]2014Improved the LBP texture feature and the GLCM statistical feature.ML-based
2HOG-SVM [64]2019HOG feature combined with the support vector machine classifier.ML-based
3LBP-SVM [30]2021Combined LBP and ORB features.ML-based
4LPQ-SRC [35]2012LPQ feature combined with the sparse representation classifier.ML-based
5ADA-DF [48]2023Double-branch feature extraction combined with the adaptive distribution fusion mechanism.DL-based
6CNN-SIFT [65]2017Combined deep–shallow feature fusion, CNN and SIFT hybrid features, local invariance, and deep semantics.DL-based
7EmoNext [50]2023Combined ConvNeXt-based optimization and a tailored training strategy.DL-based
8FERGCN [59]2022Combined graph convolution modeling, facial topology representation, and local dependency capture.DL-based
9GCN [66]2020Gabor–CNN fusion, with directional and frequency texture extraction and fast feature extraction.DL-based
10ResEmoteNet [46]2024Residual structure integration and improved loss design.DL-based
11ResNet [67]2016Residual connections and alleviated gradient vanishing.DL-based
12Poster [58]2023Double-branch structure, landmark–image cross-fusion, and pyramid multi-scale fusion.DL-based
Table 3. Image information of the four expression datasets.
Table 3. Image information of the four expression datasets.
IDDatasetNumber of CategoriesYearNumber of SamplesResolution
1AffectNet82017450,000Unfixed
2FERPlus8201635,887 112 × 112
3FER-20137201335,887 48 × 48
4RAF-DB7201715,339 100 × 100
Table 4. The number of images in each expression category in the four datasets.
Table 4. The number of images in each expression category in the four datasets.
ExpressionAffectNetFERPlusFER-2013RAF-DB
Anger5500932249533204
Contempt4250903000
Disgust43039021547867
Fear5500909851212460
Happy550099298989355
Neutral550012,99461981619
Sad550094496077877
Surprise5500945040025957
Table 5. Comparison of 4 DL-based FER methods.
Table 5. Comparison of 4 DL-based FER methods.
IDMethodOptimizerLearning RateBatch-SizeEpochs
5ADA-DF [48]Adam0.00164100
6CNN-SIFT [65]Adam0.00164100
7EmoNext [50]AdamW0.000132100
8FERGCN [59]Adam0.00132100
9GCN [66]SGD0.00564200
10ResEmoteNet [46]SGD0.00116100
11ResNet [67]Adam0.0002532100
12Poster [58]SAM0.0000432200
Note: The model ID is the same as that in Table 4.
Table 6. The training and verification accuracy of 12 models on four datasets.
Table 6. The training and verification accuracy of 12 models on four datasets.
MethodTrainVal
AffectNetFERPlusFER2013RAF-DBAve.AffectNetFERPlusFER2013RAF-DBAve.
GLCM-SVM [32]12.7016.2814.8237.1320.2312.1216.4913.9438.2520.20
HOG-SVM [64]99.4599.6399.78100.0099.7234.8851.1847.2367.1750.12
LBP-SVM [30]18.7118.3425.2238.5325.2020.0017.5224.6538.6625.21
LPQ-SRC [35]13.4013.9215.2818.6615.3212.3813.5214.9717.4114.57
ADA-DF [48]64.1896.1975.292.6582.0657.2571.5665.3799.5173.42
CNN-SIFT [65]53.9882.7080.6293.5877.7234.3857.0052.5865.5052.37
EmoNext [50]59.3490.2173.0575.2574.4656.7576.0375.6882.8772.83
FERGCN [59]28.2031.0636.4951.2736.7629.6331.2136.7752.5337.54
GCN [66]70.1196.0994.4094.3088.7359.2568.4269.3885.0370.52
ResEmoteNet [46]86.7998.5798.4298.8595.6650.8863.1258.9078.7562.91
ResNet [67]58.1486.7376.1492.4678.3733.6247.6939.2657.7544.58
Poster [58]97.2899.0199.2299.7798.8262.8876.1574.1590.7475.98
Note: The bold number indicates the maximum value of the column (optimal), while the underlined number represents the second largest value in the column (suboptimal).
Table 7. Parameter complexes of eight deep learning-based models.
Table 7. Parameter complexes of eight deep learning-based models.
ModelTrainable ParametersMACs
ADA-DF [48]22.05 M1.72 G
CNN-SIFT [65]7.48 M7.48 M
EmoNext [50]30.57 M4.57 G
FERGCN [59]4.872 K0.593 M
GCN [66]4.438 M10.74 M
Poster [58]71.8 M3.95 G
ResEmoteNet [46]80.24 M4.35 G
ResNet [67]57.91 K193.3 M
Note: The bold number indicates the best one among the column.
Table 8. The classification accuracies of each model on the validation set of AffectNet.
Table 8. The classification accuracies of each model on the validation set of AffectNet.
ModelNeutralHappySadSurpriseFearDisgustAngerContemptAve
GLCM-SVM [32]0.180.120.090.160.050.120.120.130.12
HOG-SVM [64]0.870.520.230.280.300.240.250.100.35
LBP-SVM [30]0.110.410.220.240.210.000.400.010.20
LPQ-SRC [35]0.050.080.000.040.830.020.020.020.13
ADA-DF [48]0.620.780.480.570.620.560.520.430.57
CNN-SIFT [65]0.360.370.270.190.150.010.230.050.20
EmoNext [50]0.480.790.580.510.620.490.510.560.57
FERGCN [59]0.170.740.170.440.250.020.400.060.28
GCN [66]0.500.740.690.630.630.470.570.510.59
Poster [58]0.560.630.550.670.680.700.640.650.64
ResEmoteNet [46]0.370.740.440.610.470.530.490.420.51
ResNet [67]0.280.510.300.340.310.260.360.330.34
Note: The bold number indicates the maximum value of the column (optimal), while the underlined number represents the second largest value in the column (suboptimal).
Table 9. The classification accuracies of each model on the validation set of FERPlus.
Table 9. The classification accuracies of each model on the validation set of FERPlus.
ModelNeutralHappySadSurpriseFearDisgustAngerContemptAve
GLCM-SVM [32]0.220.120.130.090.140.140.110.140.14
HOG-SVM [64]0.930.630.360.630.380.380.420.220.49
LBP-SVM [30]0.500.060.030.240.040.120.140.170.16
LPQ-SRC [35]0.020.040.220.320.030.030.350.010.13
ADA-DF [48]0.820.910.720.800.830.590.780.470.74
CNN-SIFT [65]0.860.590.500.650.520.490.520.350.56
EmoNext [50]0.840.910.720.860.750.630.780.570.76
FERGCN [59]0.180.680.030.490.310.210.170.360.30
GCN [66]0.860.880.740.860.580.470.770.240.68
Poster [58]0.760.910.600.780.890.850.620.810.78
ResEmoteNet [46]0.800.860.640.790.590.570.630.280.65
ResNet [67]0.640.620.370.580.510.470.400.320.49
Note: The bold number indicates the maximum value of the column (optimal), while the underlined number represents the second largest value in the column (suboptimal).
Table 10. The classification accuracies of each model on the validation set of FER-2013.
Table 10. The classification accuracies of each model on the validation set of FER-2013.
ModelNeutralHappySadSurpriseFearDisgustAngerAve
GLCM-SVM [32]0.180.270.140.140.070.190.100.16
HOG-SVM [64]0.880.640.190.530.230.330.210.43
LBP-SVM [30]0.040.890.080.010.000.000.010.15
LPQ-SRC [35]0.650.080.100.090.070.030.010.15
ADA-DF [48]0.690.870.540.700.390.170.640.57
CNN-SIFT [65]0.070.000.060.300.230.220.230.16
EmoNext [50]0.740.950.580.820.680.640.690.73
FERGCN [59]0.250.740.390.450.050.000.070.28
GCN [66]0.650.880.590.780.520.520.660.66
Poster [58]0.690.870.620.820.670.810.720.74
ResEmoteNet [46]0.580.810.490.740.450.480.480.58
ResNet [67]0.320.670.320.540.250.170.330.37
Note: The bold number indicates the maximum value of the column (optimal), while the underlined number represents the second largest value in the column (suboptimal).
Table 11. The classification accuracies of each model on the validation set of RAF-DB.
Table 11. The classification accuracies of each model on the validation set of RAF-DB.
ModelNeutralHappySadSurpriseFearDisgustAngerAve
GLCM-SVM [32]0.050.110.000.840.110.090.130.19
HOG-SVM [64]0.900.250.080.890.490.440.560.52
LBP-SVM [30]0.000.000.000.950.020.000.080.15
LPQ-SRC [35]0.060.090.220.030.360.010.180.14
ADA-DF [48]1.000.950.991.001.001.000.990.99
CNN-SIFT [65]0.550.290.160.850.560.400.620.49
EmoNext [50]0.840.430.630.930.810.640.790.72
FERGCN [59]0.290.000.000.870.080.110.660.29
GCN [66]0.830.620.570.940.780.760.870.77
Poster [58]0.900.860.800.960.910.880.850.88
ResEmoteNet [46]0.850.460.470.930.680.660.760.69
ResNet [67]0.500.160.030.860.380.240.570.39
Note: The bold number indicates the maximum value of the column (optimal), while the underlined number represents the second largest value in the column (suboptimal).
Table 12. The average accuracies of the 12 methods in recognizing different facial expressions.
Table 12. The average accuracies of the 12 methods in recognizing different facial expressions.
ModelsNeutralHappySadSurpriseFearDisgustAngerContempt
ML-based0.35250.26940.13060.34250.20810.13380.19310.1000
DL-based0.58910.66310.47000.69560.53590.46940.57250.4006
Note: The bold number indicates the maximum value of the column (optimal).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Zhou, Z.; Feng, Q.; Li, H. Analysis and Comparison of Machine Learning-Based Facial Expression Recognition Algorithms. Algorithms 2025, 18, 800. https://doi.org/10.3390/a18120800

AMA Style

Li Y, Zhou Z, Feng Q, Li H. Analysis and Comparison of Machine Learning-Based Facial Expression Recognition Algorithms. Algorithms. 2025; 18(12):800. https://doi.org/10.3390/a18120800

Chicago/Turabian Style

Li, Yuelong, Zhanyi Zhou, Quandong Feng, and Hongjun Li. 2025. "Analysis and Comparison of Machine Learning-Based Facial Expression Recognition Algorithms" Algorithms 18, no. 12: 800. https://doi.org/10.3390/a18120800

APA Style

Li, Y., Zhou, Z., Feng, Q., & Li, H. (2025). Analysis and Comparison of Machine Learning-Based Facial Expression Recognition Algorithms. Algorithms, 18(12), 800. https://doi.org/10.3390/a18120800

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop