Next Article in Journal
Impact Evaluation of DME Beacons on BeiDou B2a Signal Reception Performance
Next Article in Special Issue
Dual-Branch Multi-Dimensional Attention Mechanism for Joint Facial Expression Detection and Classification
Previous Article in Journal
A Parallel Plate Variable Capacitor-Based Wind Pressure Sensor: Closed-Form Solution and Numerical Design and Calibration
Previous Article in Special Issue
CAGNet: A Network Combining Multiscale Feature Aggregation and Attention Mechanisms for Intelligent Facial Expression Recognition in Human-Robot Interaction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Facial Landmark-Driven Keypoint Feature Extraction for Robust Facial Expression Recognition

1
Department of Electronic Engineering, Soongsil University, Seoul 06978, Republic of Korea
2
School of AI Convergence, Soongsil University, Seoul 06978, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3762; https://doi.org/10.3390/s25123762
Submission received: 25 April 2025 / Revised: 5 June 2025 / Accepted: 10 June 2025 / Published: 16 June 2025

Abstract

:

Highlights

What are the main findings?
  • We proposed Keypoint Features, a novel approach for feature representation based on facial landmarks, and NKF, a framework for robust FER, which was designed for clear component analysis.
  • We demonstrated state-of-the-art FER performance on benchmark datasets and performed a rigorous analysis of component contributions via ablation studies, and demonstrated strong generalization.
What are the implications of the main findings?
  • This work demonstrates the significant potential of keypoint-based FER, paving the way for its application in real-world scenarios and potential extension to other landmark-driven tasks.
  • This work provides a strong empirical foundation and performs an analysis of the components that can guide the future research in developing highly accurate and computationally efficient FER systems for unconstrained environments.

Abstract

Facial expression recognition (FER) is a core technology that enables computers to understand and react to human emotions. In particular, the use of face alignment algorithms as a preprocessing step in image-based FER is important for accurately normalizing face images in terms of scale, rotation, and translation to improve FER accuracy. Recently, FER studies have been actively leveraging feature maps computed by face alignment networks to enhance FER performance. However, previous studies were limited in their ability to effectively apply information from specific facial regions that are important for FER, as they either only used facial landmarks during the preprocessing step or relied solely on the feature maps from the face alignment networks. In this paper, we propose the use of Keypoint Features extracted from feature maps at the coordinates of facial landmarks. To effectively utilize Keypoint Features, we further propose a Keypoint Feature regularization method using landmark perturbation for robustness, and an attention mechanism that emphasizes all Keypoint Features using representative Keypoint Features derived from a nasal base landmark, which carries information for the whole face, to improve performance. We performed experiments on the AffectNet, RAF-DB, and FERPlus datasets using a simply designed network to validate the effectiveness of the proposed method. As a result, the proposed method achieved a performance of 68.17% on AffectNet-7, 64.87% on AffectNet-8, 93.16% on RAF-DB, and 91.44% on FERPlus. Furthermore, the network pretrained on AffectNet-8 had improved performances of 94.04% on RAF-DB and 91.66% on FERPlus. These results demonstrate that the proposed Keypoint Features can achieve comparable results to those of the existing methods, highlighting their potential for enhancing FER performance through the effective utilization of key facial region features.

Graphical Abstract

1. Introduction

Facial information is essential for enabling computers to understand human emotions in the field of human–computer interaction. In particular, facial expressions play a crucial role in understanding human affective states and intentions. Consequently, facial expression recognition (FER) methods have been actively studied, and their performance has been significantly enhanced due to rapid advancements in deep learning technologies. FER has increasingly become recognized as a core technology in providing user-customized experiences across various industries.
The sensors used for FER can be broadly divided into contact and non-contact types. Contact sensors, such as those capturing brain wave or electromyography data, can be difficult to reliably attach onto users’ skin, which limits their practicality for everyday use. However, non-contact sensors, including cameras and microphones, can capture image and audio data from a distance. These sensors have significantly fewer constraints compared with contact sensors. In particular, most of the information humans perceive from external environments is visual data, and facial expressions are essentially a form of visual non-verbal communication. Therefore, an approach focused on analyzing facial expression using image data could be a direct and efficient method for FER.
However, FER in the wild remains challenging due to factors such as large facial poses and occlusions. Although FER methods typically process images of individual faces, such as ID photos, general images frequently contain various background elements. This makes the preprocessing steps that focus explicitly on the face crucial for performance. Generally, a FER pipeline involves detecting facial regions using a face detection algorithm and subsequently normalizing the detected face using a face alignment algorithm that performs scaling, rotation, and translation based on predicted facial landmarks. These preprocessing steps allow the FER network to learn consistent feature representations, leading to stable inference in various environments.
Recently, Zheng et al. [1] proposed a method utilizing feature maps computed from a face alignment network. Their study showed that face alignment networks could function not only as preprocessing tools but could also directly contribute to enhancing FER performance through leveraging these feature maps. Although these approaches are effective, the comprehensive analyses and structural designs remain insufficient for fully utilizing face alignment information. Prior studies using face alignment networks either as preprocessing tools or feature extractors have shown limitations in achieving high FER accuracy, indicating that the face alignment information was not fully exploited.
When humans recognize another person’s emotions, they tend to focus on specific facial regions, such as the eyes, nose, and mouth. Similarly, several FER studies [1,2,3,4] have shown that their networks focused on these specific facial regions through visualization methods and emphasized that these regions provided significant information for FER.
Motivated by these observations and the limitations of previous approaches, we propose a novel approach that introduces the use of Keypoint Features and a dedicated network architecture to leverage them for FER. These features are designed to fully exploit the rich information provided by the face alignment network, thereby enhancing FER performance. Our core contribution is the introduction of Keypoint Features, which are meticulously derived from semantically meaningful facial regions within the combined feature maps of a backbone network and a face alignment network. Their locations are precisely determined using the predicted facial landmark coordinates. This approach enables a more effective and comprehensive utilization of fine-grained information from specific facial regions, thus surpassing the limitations of methods that rely solely on landmark coordinates for image alignment or holistic feature map interpretation. To fully utilize and optimize the proposed Keypoint Features, we designed a Network for utilizing Keypoint Features (NKF). This network comprises two essential components: a novel Keypoint Feature regularization method based on landmark perturbation and Representative Keypoint Feature Attention (RKFA). Landmark perturbation-based regularization, applied during training, enhances the model’s robustness to potential inaccuracies in predicted facial landmarks and aims to mitigate the negative effects of imperfect landmark detection in the wild. RKFA adaptively emphasizes the most salient facial cues by leveraging a representative Keypoint Feature derived from the nasal base landmark, which encapsulates information for the entire face, thereby improving FER performance.
Figure 1 illustrates the NKF’s flow. NKF primarily consists of a Global Feature Network and a Keypoint Feature Network. From an input face image, the Global Feature Network obtains feature maps from the backbone network. Simultaneously, it retrieves feature maps and predicts facial landmark coordinates from the face alignment network. These feature maps are then concatenated to produce global features. Subsequently, the Keypoint Feature Network extracts Keypoint Features from the global features based on the predicted facial landmark coordinates. The RKFA module then refines the Keypoint Features using a nasal base Keypoint Feature that contains information for the whole face. Finally, the network classifies the refined Keypoint Features for FER.
To validate the effectiveness of the proposed method, we evaluated it on widely adopted FER benchmark datasets, including AffectNet [5], RAF-DB [6], and FERPlus [7]. These datasets consist of facial images captured using standard camera sensors used in image-based FER studies.
The experimental results using standard protocols demonstrate that our method achieves a performance comparable to previous methods in these benchmarks. Specifically, we designed the NKF with a simple structure to clearly analyze the effectiveness of the proposed components and validated it through quantitative evaluations and ablation studies. In this study, we do not propose a novel sensor or a new dataset, but rather focus on developing a novel computational methodology for analyzing the image data produced by camera sensing technologies in the context of FER. Our primary contributions are as follows:
  • We propose a novel type of feature called Keypoint Features, which are extracted from feature maps at the coordinates of facial landmarks predicted by a face alignment network. This feature enables a more effective utilization of information from specific facial regions.
  • To enhance the quality of the extracted Keypoint Features, we introduce two methods: (i) Keypoint Feature regularization, which aims to reduce the negative effect of incorrectly predicted facial landmarks; and (ii) Representative Keypoint Feature Attention (RKFA), which uses a nasal base Keypoint Feature that includes whole-face information to improve FER performance.
  • We designed a deliberately simple network, excluding other advanced structures, to clearly analyze the effects of the proposed components. We empirically validated their contributions through experiments on multiple benchmark datasets and ablation studies.

2. Related Work

Facial expression recognition (FER) is a crucial area of affective computing. Broadly, approaches can be categorized into uni-modal and multi-modal methods. Uni-modal FER primarily relies on a single source of information, most commonly facial images, to recognize expressions. In contrast, multi-modal FER integrates diverse data streams for a more comprehensive understanding of emotions. For instance, recent advancements in multi-modal emotion recognition include systems that primarily rely on non-visual physiological signals, such as electroencephalography (EEG) and electrocardiography (ECG) [8]. However, acquiring such physiological signals often requires specialized close-proximity sensors, which can limit their applicability in daily scenarios. In contrast, other multi-modal approaches aim to improve the FER performance by leveraging combinations such as facial images and textual data to capture a richer emotional context [9,10]. While multi-modal methods are valuable for holistic emotion recognition, facial images inherently provide the most direct and rich source of information for facial expression recognition, as facial expressions are essentially a form of visual non-verbal communication. Thus, uni-modal, image-based FER methods remain critically important.
For facial expression recognition (FER), various methods have been proposed that leverage facial properties alongside standard deep network architectures, such as ResNet [11] and ViT [12], and attention modules like SE [13] and the self-attention mechanism of ViT. Because multiple facial expressions can appear on the same face, some studies [14,15] have aimed to emphasize expressive components while degrading facial identity. Other FER methods [16,17,18,19,20] have focused on loss functions designed to minimize intra-class distance and maximize inter-class separability in the expression feature space. Additional studies [21,22] addressed expression ambiguity by refining label distribution and feature representations. Because the proposed method focuses on utilizing face alignment for FER, we excluded the aforementioned methods that primarily address distribution alignment or representation learning to validate the contribution of face alignment more clearly.
Face alignment-based FER approaches can be broadly categorized based on their application in either the preprocessing or feature extraction stage. In our method, we incorporated face alignment into both stages. This section introduces previous face alignment-based FER methods and their limitations.

2.1. FER with Face Alignment as a Preprocessing Step

Most FER datasets focus on the expressions of individual faces. In general, a face image is cropped using a face detection method, and its scale, rotation, and translation are normalized using a face alignment method [23,24,25,26,27,28,29,30,31].
Most FER studies have employed face alignment in the preprocessing stage, which consequently limits the improvement in FER performance. Typically, the face alignment methods used for FER preprocessing have lower accuracy than more sophisticated modern methods. Commonly used methods include face detection-based approaches, such as MTCNN [32] and RetinaFace [33], or conventional algorithms, such as ERT [34], which utilize gradient boosting trees rather than deep neural networks.

2.2. FER with Face Alignment for Feature Extraction

Kollias et al. [35] fitted a 3D Morphable Model [36] using face alignment, subsequently generating synthetic-expression images based on the fitted model for use in FER. However, this method is limited because it only modifies the input images and does not directly enhance the feature maps of the FER network. Chen et al. [37] proposed a method to reduce expression ambiguity by constructing a landmark-based k-nearest neighbor graph that adjusts similar expressions to have similar label distributions. Liu et al. [38], following a strategy similar to that of Kollias et al., generated synthetic images while simultaneously training their network with landmark information via multi-task learning, enabling the network to suppress identity information. Ding et al. [39] enhanced the network’s feature extraction performance by generating attention maps derived from landmark locations. Furthermore, Zheng et al. [1], Mao et al. [2], and Chen et al. [40] employed MobileFaceNet [41], a face alignment network, to perform cross-attention between its generated feature maps and those from the backbone networks. However, these studies only utilized feature maps from the face alignment network, without leveraging landmark coordinates, thus limiting the potential benefits of face alignment.
In contrast, we propose a novel method that utilizes both feature maps and landmark coordinates from the face alignment network to enhance FER performance.

3. Proposed Method

The structure of the proposed Network for utilizing Keypoint Features (NKF) is shown in Figure 2. This network consists of two main components: a Global Feature Network and a Keypoint Feature Network. The Global Feature Network includes a backbone network that generates its own feature maps and a face alignment network that generates feature maps and detects facial landmarks. These two types of feature maps are then fused to produce Global Features. The Keypoint Feature Network extracts the Keypoint Features from the Global Features based on the coordinates of the predicted facial landmarks and uses them to perform FER.
Because this study focuses on strategies for utilizing Keypoint Features derived from face alignment, we briefly describe the relatively standard structure of the Global Feature Network and provide a more detailed explanation of the Keypoint Feature Network components in the following subsections.

3.1. Global Feature Network

The Global Feature Network generates feature maps, called global features, to perform FER. This network comprises a backbone network and a face alignment network to effectively extract facial information from input images. The backbone network computes feature maps dedicated to FER, whereas the face alignment network generates its own feature maps and predicts facial landmarks. These feature maps provide preliminary information for extracting Keypoint Features.
Because humans tend to recognize facial expressions based on facial shape rather than skin texture and color, it is crucial for a deep neural network used in FER to learn this characteristic to improve its performance. The feature maps from the face alignment network contain semantic information about facial landmarks, including the eyes, nose, and mouth, making them useful for understanding facial shape. Therefore, by fusing the feature maps from the face alignment network with those from the backbone network, rich information about the facial structure is incorporated into the combined feature maps.
Zheng et al. [1] and Mao et al. [2] improved FER performance through an attention mechanism, where feature maps from a face alignment network guided the backbone network’s attention toward facial landmarks. In contrast, our approach enhances FER performance by simply concatenating the feature maps from both networks. The fusion process is defined as Equation (1):
F G = F 3 × 3 , 2 D F B B N 1 F B B N 2 F 3 × 3 , 0 F F A N 1 F 3 × 3,1 U F F A N 2
In Equation (1), F 3 × 3 , i represents the i -th 3 × 3 convolution layer, and F j denotes the feature maps of the j -th block layer. Here, B B N and F A N refer to the backbone and face alignment networks, respectively. D and U represent downsampling and upsampling operations, respectively, and denotes feature map concatenation along the channel dimension. F G is the resulting global feature.
The dimensions of each feature map are as follows: F B B N 1 has dimensions of 128 × 28 × 28 and is downsampled to 128 × 14 × 14. F F A N 1 has dimensions of 128 × 14 × 14 and passed through F 3 × 3 , 0 without a change in size.  F F A N 2 has dimensions of 512 × 7 × 7, which is upsampled to 512 × 14 × 14 and then passed through F 3 × 3,1 to decrease the dimensions, resulting in 256 × 14 × 14. F B B N 2 has the dimensions of 256 × 14 × 14. Their concatenation results in a 768 × 14 × 14 feature map, which is then passed through F 3 × 3 , 2 to produce a global feature, F G , with the same dimension. Each convolution layer in Equation (1) does not include a batch-normalization layer or an activation layer. Both networks utilize residual block-based architectures, and the Global Feature Network exploits multi-level features to capture both low- and high-level semantic information, leading to better FER performance.

3.2. Keypoint Features

The Keypoint Feature Network extracts Keypoint Features from the previously generated global features, F G , and utilizes them for FER. Facial landmark coordinates correspond to specific facial regions, such as the eyes, nose, and mouth, which are critical for FER. Siqueira et al. [4] utilized Grad-CAM [42] to demonstrate that the feature regions around these landmarks significantly influence FER. Recent ViT-based studies [1,2,3] have demonstrated similar findings. This suggests that features extracted from landmark coordinates are highly informative in FER. Keypoint Features are defined as follows:
F K = F G x l m k ,     x l m k = x l m k ,   y l m k        
where x l m k denotes the facial landmark coordinates predicted by the face alignment network. The Keypoint Features F K are sampled from the global features F G at these coordinates. Each of the 68 landmarks provides a 768-dimensional feature vector, resulting in F K with a size of 68 × 768. Because landmark coordinates are floating-point values, they are typically quantized to integer grid indices for direct indexing. However, this can introduce quantization errors that degrade feature quality. To address this, we employed a bilinear interpolation:
F K = f b i l i n F G ,   x l m k
where f b i l i n represents the bilinear interpolation function.
Figure 3 shows the distributions of the Keypoint Features. Each landmark produced a distribution, totaling 68. The blue box in the figure highlights the Keypoint Features from the outer face contour, where ‘Happy’ and ‘Neutral’ expressions are relatively well-separated, while other expressions are harder to distinguish. In contrast, the red box illustrates the Keypoint Features around the upper lip, which exhibit clearer class separability. This suggests that mouth-region features are more discriminative for expression recognition than outer-contour features.
Although recognizing expressions based solely on facial landmarks is challenging, our Keypoint Features incorporate information about the entire facial region owing to the wide receptive fields obtained through the convolution layers of the Global Feature Network. Hence, even outer-region Keypoint Features contribute meaningfully, validating their overall effectiveness. A comparison of different Keypoint Feature types is presented in Section 4.3.2.
The final classification of the FER model employed two fully connected layers. After the first layer, batch normalization and a ReLU activation function were applied to generate 768 channel feature vectors, followed by a fully connected layer for classification.

3.3. Keypoint Feature Regularization

Since Keypoint Features are extracted based on facial landmark coordinates, they are inherently dependent on the accuracy of those coordinates. However, under in-the-wild conditions, landmark predictions can be imprecise owing to factors such as large pose variations or occlusions, which can degrade the FER performance. To mitigate this issue, we adopted a regularization method that introduced random perturbations to the landmark coordinates during training. Equation (4) illustrates how perturbed Keypoint Features are generated:
F K = f b i l i n F G ,   x p r e d + 1 t r a i n · U c ,   c w h e r e   1 t r a i n = 1 , if   training   phase   0 , e l s e
where U ( c , c ) denotes a uniform distribution within the range [ c , c ], and 1 t r a i n is an indicator function that equals 1 during training and 0 otherwise. During training, random offsets sampled from U c , c are added to the predicted landmark coordinates prior to feature extraction. By contrast, during inference, the original coordinates are used without perturbation. The hyperparameter c defines the maximum magnitude of the coordinate perturbation and can be tuned to optimize the robustness. This regularization method improves adaptability of the network to uncertainties in landmark detection and enhances its ability to generalize under various challenging conditions, such as diverse head poses and occlusions.
Figure 4 illustrates the process of Keypoint Feature regularization. During the training phase, random offsets are added to the landmark coordinates, thereby introducing intentional noise into the Keypoint Features. This method generates various inaccurate Keypoint Features to provide a regularization effect and improve the network’s generalization performance on noisy data. Section 4.3.2 presents a performance comparison across different combinations of the proposed components, confirming that Keypoint Feature regularization contributes to the FER performance gains.

3.4. Representative Keypoint Feature Attention Using the Nasal Base Landmark

In Section 3.2, we analyzed the distribution of Keypoint Features and observed that some, particularly those located along the outer facial contour, exhibited a relatively low classification performance. Nevertheless, these features still encode unique and valuable information, and discarding them entirely may lead to suboptimal results. To address this, we propose Representative Keypoint Feature Attention (RKFA), a mechanism that enhances the classification performance of all Keypoint Features while preserving their contributions.
The proposed RKFA approach leverages the Keypoint Features at the nasal base landmark, which is located near the geometric center of the face. Features extracted from this location effectively captured the global facial structure and consistently demonstrated a strong classification performance. Therefore, by leveraging this globally representative feature, we aim to enhance all other Keypoint Features.
RKFA is implemented as a residual attention module inspired by conventional attention methods [13,44,45], with a key difference in that it duplicates the nasal base Keypoint Feature and broadcasts it across all Keypoint Features to compute attention weights. The RKFA mechanism is formulated as follows:
F R = F K + 1 × Softmax F f c 1 F K N a s a l b a s e F K ,   1 R N × 1  
In Equation (5), F K N a s a l b a s e denotes the Keypoint Features at the nasal base landmark. The operator replicates this feature N times, where N is the number of landmarks. The duplicated feature is concatenated with the full set of Keypoint Features, F K , along the landmark dimension. This concatenated tensor is passed through a fully connected layer, F f c , to compute the individual attention scores. The scores are normalized via a SoftMax function to produce attention weights, which are then applied channel-wise to refine each Keypoint Feature. The resulting features, F R , are thus adaptively reweighted according to their importance, guided by the globally representative nasal base features.
Figure 5 shows the distributions of the Keypoint Features before and after applying RKFA. Notably, the nasal base Keypoint Feature demonstrates relatively clear class separability even prior to applying the attention mechanism, whereas Keypoint Features located along the facial contour show more ambiguous class boundaries. After applying RKFA, the distribution of these outer Keypoint Features becomes more structured, and class separation improves significantly.
This improvement highlights the effectiveness of the RKFA mechanism in enhancing the discriminative power of all Keypoint Features, including those with initially weaker representations. The performance gains resulting from RKFA are further validated experimentally in Section 4.3.2, which demonstrates its positive contribution to overall FER accuracy.

4. Experiments

This section evaluates the proposed NKF introduced in Section 3. We assessed the accuracy of NKF and compared it with the state-of-the-art methods using widely adopted FER benchmark datasets, including AffectNet [5], RAF-DB [6], and FERPlus [7]. Furthermore, we validated the effectiveness of each proposed component through ablation studies.

4.1. Dataset

AffectNet is currently one of the most widely used large-scale datasets for facial expression recognition, with a training set of 287,651 samples and a test set of 4000 samples. The dataset provides both categorical labels (e.g., basic facial expressions) and dimensional labels (valence–arousal), along with 68 facial landmark annotations. Performance evaluations on categorical labels are typically performed under two settings: AffectNet-7, which includes seven categories (neutral, happy, sad, surprise, fear, angry, and disgust), and AffectNet-8, which adds the contempt category to the previous seven.
RAF-DB is another popular FER dataset containing 12,271 training samples and 3068 test samples. It includes annotations for 7 basic expressions (surprise, fear, disgust, happy, sad, anger, and neutral) and 12 compound expressions derived from combinations of these basic emotions. Each image was annotated by 40 human raters. The dataset provides 5 manually labeled and 37 automatically extracted facial landmarks, along with demographic metadata, such as race, age, and gender. In this paper, we evaluated the FER performance using only seven basic expression categories.
FERPlus is a revised version of the FER2013 [46] dataset, in which the original labels were re-annotated by 10 individuals per sample to correct prior inaccuracies. It consists of 28,709 training samples and 3589 samples for validation and testing. The dataset includes annotations for eight expressions (neutral, happy, surprise, sad, anger, disgust, fear, and contempt), as well as ‘unknown’ and ‘non-face’ labels. In our experiments, we evaluated the classification performance on the eight valid expressions, excluding the ‘unknown’ and ‘non-face’ classes. The ground truth labels were determined by majority voting across the 10 annotators.

4.2. Implementation Details

We trained a dedicated face alignment network to support the NKF pipeline. This network, based on MobileFaceNet [41], a lightweight architecture, was trained on each dataset individually. To guide its training, we utilized the PFA network [47], a high-accuracy face alignment model, to distill landmark knowledge. This approach ensures precise landmark detection and effective face alignment. The network was trained for 200 epochs across all datasets using stochastic gradient descent (SGD) with a learning rate of 0.01, momentum of 0.9, batch size of 16, and mean squared error (MSE) loss.
For the FER task, our training strategy followed a framework similar to POSTER++ [2]. The backbone network used for expression classification was ResNet50 (IR50) [48], pretrained on the Ms-Celeb-1M [49] dataset. The weights of the face alignment network were frozen during the FER training. The input images were resized to 112 × 112 via bilinear interpolation. We trained the FER model using the Sharpness-Aware Minimization (SAM) optimizer [50,51] built on top of Adam [52], with a batch size of 64, weight decay of 1.0 × 10−4, and an exponential learning rate decay of 0.98 per epoch. The cross-entropy loss function was used.
Data augmentation was performed in multiple stages as follows. First, we applied ±15% random scaling, ±30° rotation, ±15 pixels random translation, and 50% random horizontal flipping. Subsequently, we randomly selected and applied one of the following: Gaussian noise, Gaussian blur, grayscale conversion, contrast adjustment, color jittering, power law transform, histogram equalization, JPEG compression artifacts, lighting adjustment, or identity mapping. Finally, Random Erasing was applied to enhance model robustness through regularization.
All NKF models were trained using only the training split of each dataset. Because the AffectNet training set exhibited a class imbalance, whereas the test set was relatively balanced, we employed a class-balanced sampling strategy during training to mitigate this discrepancy. Specifically, samples from underrepresented classes were assigned higher sampling probabilities, whereas dominant classes were downweighted. However, when fine-tuning the AffectNet pretrained model on RAF-DB or FERPlus, we maintained the original class imbalance.
We implemented our models using Python 3.10 and the PyTorch 2.5 deep learning framework. All experiments were conducted on a system equipped with an AMD Ryzen 7 7700X CPU (AMD, Santa Clara, CA, USA) and an NVIDIA RTX3090 Ti GPU (NVIDIA, Santa Clara, CA, USA), operating under Ubuntu 22.04 LTS.
The training details for each dataset are summarized in Table 1. The landmark perturbation value, denoted as c in Equation (4) of Section 3.3, was multiplied by the size of the feature map to control the perturbation scale. A smaller value leads to a less significant impact on the landmark coordinates, whereas a larger value increases the regularization strength, improving robustness and generalization.

4.3. Evaluation

4.3.1. Comparison with State-of-the-Art Methods

To validate the effectiveness of the proposed method, we compared the performance of our NKF framework with state-of-the-art FER methods on the AffectNet, RAF-DB, and FERPlus benchmark datasets. As shown in Table 2, the proposed NKF consistently achieves higher classification accuracy across all datasets and outperforms the existing state-of-the-art methods, thereby validating its superiority for FER tasks.
Table 2 compares the methods pretrained on AffectNet-8 with those that were not. The proposed NKF outperforms most existing state-of-the-art methods, except for its performance on the AffectNet-7 setting. Specifically, NKF achieved an accuracy that was 0.52% lower than that of Norface [38] on AffectNet-7 but surpassed DDAMFN [30] by 0.62% on AffectNet-8, Norface by 0.19% on RAF-DB, and S2D by 0.43% on FERPlus.
Norface utilizes additional generated images for training, which allows it to show an excellent performance on datasets such as AffectNet-7, where facial expressions are complex. Indeed, Norface demonstrated superior results on AffectNet-7, whereas NKF, trained without generated images, performed relatively worse on the same dataset. However, the proposed NKF, trained using only the images from the given dataset without generated images, achieved higher accuracy than Norface on the RAF-DB dataset, despite its comparatively worse performance on AffectNet-7.
Furthermore, when comparing models pretrained on AffectNet-8, NKF outperformed S2D by 1.57% on RAF-DB and 0.49% on FERPlus. These improvements represent a 2.44-fold increase on RAF-DB and a 1.38-fold increase on FERPlus in terms of the relative gains achieved through fine-tuning compared to S2D. These results highlight the impact of consistent landmark-based alignment across datasets, which effectively reduces domain discrepancies and maximizes the benefits of pretraining.
Figure 6 illustrates the classification performance of the proposed NKF on each dataset using confusion matrices, providing a detailed analysis of the prediction accuracy for individual expression classes. In all confusion matrices, except (d), each expression class shows a higher proportion of correctly classified samples than misclassified samples.
Although DDAMFN achieved a relatively high accuracy of 90.74% on the FERPlus dataset, it struggled significantly with the contempt class, frequently misclassifying it as neutral and anger. The proposed NKF also faces challenges in recognizing contempt; however, the NKF model pretrained on AffectNet-8 achieves the highest accuracy for this particular class, suggesting that the pretrained weights contribute positively to the recognition of nuanced expressions. Despite this improvement, the NKF’s accuracy for contempt remains lower than that for most other classes.
This lower accuracy is likely attributable to both the limited number of training samples for the contempt class and the subtle facial expression cues, which often resemble disgust or anger. These overlaps in visual features introduce ambiguity in neighboring classes, such as neutral, leading to frequent misclassifications. This trend, in which expressions with a less distinct visual separation from neutral are more prone to error, can also be observed for other ambiguous classes, such as disgust.
Table 3 compares the classification accuracy of NKF with that of state-of-the-art methods across individual expression classes for each dataset. In general, NKF demonstrates a higher overall accuracy than competing methods and consistently performs well, even in classes where other methods underperform. Although NKF occasionally shows slightly lower accuracy in certain high-performing classes compared to some specific approaches, it achieves the highest average accuracy overall.
These results highlight NKF’s strong generalization capabilities and the robustness of NKF in handling challenging expression classes, reinforcing its effectiveness as a reliable FER framework across varying datasets and emotional categories.
Additionally, we conducted a detailed attribute analysis of its performance on the RAF-DB dataset to evaluate its robustness and generalization capabilities. As presented in Table 4, the NKF generally shows a consistently good performance across gender, race, and age groups.
The NKF achieved over 90% accuracy for all groups, except for older adults aged 70+ years. This includes consistent accuracy across male, female, and ‘Unsure’ gender categories, as well as a good performance across all racial subgroups.
However, a decrease in performance was observed for older adults aged 70+ years, with accuracy falling into the 80% range. This decline is primarily attributed to age-related morphological changes, such as the development of wrinkles and skin folds, which can make the appearance of expressions ambiguous [58]. These physical alterations can make it more difficult for image-based FER systems to accurately interpret subtle cues of emotion in older faces.
This comprehensive attribute evaluation underscores the generalization capabilities of the NKF across varied populations and indicates the need for future improvement, particularly in enhancing performance for older-adult attributes.
Figure 7 compares the accuracies and FLOPs of various models on the RAF-DB and AffectNet-7 datasets. Although NKF incurs a relatively higher computational cost than lightweight architectures, it requires fewer FLOPs than POSTER++, despite using the same backbone [48] and face alignment networks [41]. Despite using fewer FLOPs, NKF achieves 94.04% accuracy with a 1.83% improvement on RAF-DB, 68.17% with a 0.68% improvement on AffectNet-7, and 64.87% with a 1.1% improvement on AffectNet-8.
For a more detailed comparative analysis, Table 5 presents the parameters, FLOPs, and accuracy across different models on the RAF-DB and AffectNet-7 datasets. In Table 5, ‘Global Net.’ refers to the combined computational cost of the backbone and face alignment networks, while ‘Local Net.’ denotes the proposed Keypoint Feature Network.
Although the NKF incurs a relatively higher computational cost than some lightweight architectures, Figure 7 demonstrates that the NKF demands fewer FLOPs than POSTER++, despite using the same backbone [48] and face alignment networks [41]. Regarding model complexity, the NKF generally exhibits a higher number of parameters than other models. This is primarily because we employed a straightforward fully connected layer to clearly demonstrate the effectiveness of the proposed Keypoint Features with a simple network structure. In particular, a substantial number of parameters are concentrated in the penultimate layer, which reduces the dimensionality from 68 × 768 to 768.
To further assess the practical efficiency of the NKF, we measured its inference speed. We report Frames Per Second (FPS) by distinguishing between the CPU and GPU times for full network inference. Our analysis suggests that both CPU-based inference on an AMD Ryzen 7 7700X CPU and GPU inference of the NKF on an NVIDIA RTX3090 Ti GPU are highly efficient, with both achieving real-time performance and indicating strong potential. On this hardware, our model achieved approximately 36.09 FPS for CPU inference and 244.62 FPS for GPU inference. Leveraging this efficiency, the NKF achieves 94.04% accuracy with a 1.83% improvement on RAF-DB, 68.17% with a 0.68% improvement on AffectNet-7, and 64.87% with a 1.1% improvement on AffectNet-8.

4.3.2. Ablation Study

To comprehensively analyze the effectiveness of the NKF, we conducted a series of accuracy-based performance comparisons using the AffectNet-8 dataset. With its substantial number of samples and class-balanced evaluation, AffectNet-8 is a suitable benchmark for assessing the generalization ability of various models.
Our ablation analysis begins with an evaluation of the individual contributions of each proposed component through comparative experiments involving different network configurations. Each configuration incrementally integrates specific elements of the NKF framework. In addition, we assessed the influence of landmark types and the positioning of keypoints used in the Representative Keypoint Feature Attention (RKFA) mechanism.
Table 6 shows the performance of the NKF models designed by incorporating the various proposed components. In all cases, the input consisted of the original facial images provided in the dataset. The baseline model, which does not utilize any of these components, is constructed by applying global average pooling to feature maps of the backbone network to form a feature vector, followed by classification using a fully connected layer.
According to Table 6, applying accurate face alignment using facial landmarks yields a 1.03% performance improvement, raising the accuracy from 62.94% to 63.97%, thereby emphasizing the importance of precise facial alignment. The NKF model, which integrates all proposed components, achieves a total accuracy increase of 1.45% (from 62.94% to 64.39%), validating the cumulative effectiveness of the method. Moreover, we observed performance improvements across all configurations that included any of the proposed modules, whether applied individually or in combination.
Finally, when starting with aligned images as the input and then applying all proposed components, the model achieved an additional 0.90% gain, improving from 63.97% to 64.87%. This further confirms that the combined application of accurate alignment, Keypoint Feature extraction, regularization, and RKFA yields synergistic improvements in FER accuracy.
Keypoint Features are particularly effective owing to the wide receptive field of the feature maps computed by the underlying networks. Consequently, even when a limited number of landmark coordinates are used, each extracted feature can still capture information from a broad spatial area. This characteristic enables a robust FER performance despite the relatively sparse spatial sampling.
Facial expression cues are predominantly concentrated in specific regions of the face, namely the eyes, nose, and mouth. Accordingly, the critical information required for facial expression recognition can be sufficiently captured using Keypoint Features extracted only from these localized facial regions.
Figure 8 and Table 7 provide an overview of the different landmark subsets employed for Keypoint Feature extraction in this study. The results quantitatively demonstrate the efficacy of the proposed method by comparing FER performance across various landmark configurations, validating that key facial regions alone can serve as an effective basis for expression classification.
Table 7 presents the FER performance of the NKF models designed using different landmark configurations. The results demonstrate that FER remains feasible even when utilizing a reduced set of Keypoint Features, as evidenced by the ‘Half’ landmark type. However, this reduced configuration led to an overall drop in accuracy compared to the ‘Full’ landmark setting.
For the ‘Inner’ type, which comprises only landmarks located in the central facial regions, the results confirm that landmarks around the eyes, nose, and mouth are more informative than those on the facial periphery. Nonetheless, a slight performance degradation was observed relative to the ‘Full’ setting, indicating that the Keypoint Features along the outer face contour also carry useful expression-related information. This suggests that leveraging Keypoint Features from the entire face contributes to a higher FER accuracy and supports the holistic design of the proposed NKF framework.
The Representative Keypoint Feature Attention (RKFA) mechanism enhances FER by attending to the most expressive Keypoint Features. In this study, the nasal base Keypoint Feature was selected as the representative feature owing to its strong performance and central facial location. As this feature lies near the geometric center of aligned facial images—typically between the nose and mouth—it may be substituted with other centrally located features on the feature map without a significant loss of information.
Figure 9 provides a visual comparison between the general facial landmarks, nasal base landmark, and center point of the feature map. As illustrated, the nasal base landmark is situated very close to the center of the facial image, supporting its role as a representative anchor for attention computation in the RKFA. Table 8 presents a performance comparison of the NKF variants designed using different attention mechanisms. Using Keypoint Features based on the center point of the feature map improved the performance compared to the version without RKFA, indicating that even spatially central features can provide meaningful guidance.
Notably, the RKFA configuration that utilized the nasal base Keypoint Feature achieved the best performance among all attention types. This result can be attributed to the location of the nasal base landmark, which is positioned between the nose and mouth, capturing regions more directly associated with expressive facial movements than the geometric center point.
These findings suggest that, for enhancing the FER performance, the semantic relevance of Keypoint Features to expression representation is more critical than relying solely on their spatial proximity to the center of the face. In other words, attending to regions rich in expression-related cues yields more effective attention guidance than using position alone.

5. Discussion

We proposed a novel FER framework, NKF, which leverages Keypoint Features extracted from specific facial landmarks and integrates additional components to enhance recognition performance. Prior studies have demonstrated that facial regions, such as the eyes, nose, and mouth, carry highly discriminative cues for FER. In line with this, our method extracts Keypoint Features from these regions and uses them for effective expression classification. As shown in Figure 3, the proposed Keypoint Features form well-separated clusters across expression classes, validating their discriminative power.
Moreover, the landmark perturbation-based regularization improved the model’s robustness by addressing errors from inaccurately detected landmarks and enhancing generalization performance. The Representative Keypoint Feature Attention (RKFA) mechanism further contributed by refining lower-performing Keypoint Features using the nasal base Keypoint Feature, thereby maintaining expression-relevant information while improving the overall classification accuracy.
Although the nasal base Keypoint Feature proved effective as a representative cue for RKFA in our experiments, we are mindful that its universal representativeness might be challenged under certain conditions, such as extreme facial poses or severe occlusions. For large poses, such as profile faces, information from the nasal base landmark may be less effective; moreover, occlusion of this region could render the information entirely invalid. However, network weights trained on augmented large poses or occluded face images can potentially mitigate these limitations to some extent. While our current data augmentation strategy includes techniques such as Random Erasing for occlusion and 2D image rotation to introduce variations for large poses, it does not specifically address full profile faces. In such challenging scenarios, the perceived ‘holistic facial information’ of the nasal base could otherwise be distorted or obscured, potentially limiting the effectiveness of RKFA. Future work could explore more dynamic strategies for selecting representative landmarks, perhaps by employing a learned mechanism to adaptively choose the most reliable and informative reference point based on specific characteristics of the input image, including large pose and occlusion level, thereby enhancing the robustness of RKFA across a wider range of unconstrained facial images.
The proposed NKF, which integrates Keypoint Features, landmark perturbation, and RKFA, outperforms several state-of-the-art methods across multiple datasets. Specifically, NKF attained 64.87% accuracy on AffectNet-8 (+0.62% over DDAMFN [30]), 93.16% on RAF-DB (+0.19% over Norface [38]), and 91.44% on FERPlus (+0.43% over S2D [40]), outperforming several state-of-the-art methods. In particular, using aligned images based on precise facial landmarks significantly enhanced the performance and contributed to reducing the domain gap across datasets. Consequently, fine-tuning the AffectNet-8 pretrained model yielded 94.04% accuracy on RAF-DB (+1.57% over S2D [40]) and 91.66% on FERPlus (+0.49% over S2D [40]). Enhanced performance of the NKF is primarily attributed to several core factors: (1) high-quality Keypoint Features extracted from semantically meaningful facial regions, (2) robust regularization via landmark perturbation, and (3) targeted feature enhancement by RKFA. These results are consistent with prior works, such as POSTER [1], POSTER++ [2], EAC [3], and EfficientFace [4], which also emphasize the importance of region-specific features in FER.
In the AffectNet-7 evaluation, the NKF underperformed slightly compared to Norface [35] by 0.52%. This suggests that while the NKF is effective with the available dataset, it may face limitations when dealing with more complex or diverse data distributions without additional data augmentation. It is important to note that Norface incorporates synthetic images generated from an external dataset, creating a non-comparable training environment for the model.
To overcome these limitations, future work could explore the integration of the NKF with Vision Transformer (ViT)-based architectures or employ refined landmark definitions for more complex datasets. However, given that ViT lacks strong inductive biases, its effective application necessitates large-scale pretraining on datasets such as JFT [59] or ImageNet-21k [60]. Furthermore, the fixed patch structure of ViT may present compatibility challenges for landmark-based feature extraction. Moving forward, we aim to enhance performance by utilizing large-scale face-specific datasets and embedding self-attention mechanisms from ViT into the NKF. Such endeavors could extend the capabilities of keypoint-based FER and serve as a crucial step toward optimizing both recognition accuracy and computational efficiency in real-world scenarios.
The NKF’s analysis of misclassified samples revealed specific challenging scenarios. As illustrated in Figure 10, we observed two primary contributing factors to recognition errors: inaccurately predicted facial landmarks and inherently ambiguous expressions.
The top row of Figure 10 presents examples of how factors, such as facial occlusion, can lead to mislocalized landmarks. Specifically, the first, second, and fourth columns illustrate faces with various forms of occlusion. Furthermore, in the third column, the eyebrow hair is nearly imperceptible, posing an additional challenge for accurate localization. These inaccurately localized Keypoint Features provide distorted facial information, which subsequently results in failures in downstream tasks.
Despite the facial landmarks being localized with relative accuracy in the first face shown in the bottom row, its FER result remains ambiguous. The first, second, and third examples in the bottom row feature ambiguous expressions, while the fourth example’s expression cannot be accurately recognized due to severe occlusion from a hand. Notably, even human perception struggles to guarantee accurate FER consistently in these challenging cases. Future work will consider an approach incorporating uncertainty-aware facial landmark prediction and FER.

6. Conclusions

In this paper, we introduce a novel facial expression recognition (FER) framework that leverages Keypoint Features derived from facial landmarks to enhance FER performance. Drawing inspiration from prior studies that emphasize the importance of specific facial regions, we extracted features from these landmark locations for expression classification. To further boost the effectiveness of the proposed Keypoint Features, we incorporated two techniques: landmark perturbation-based regularization, which improves robustness against detection noise, and Representative Keypoint Feature Attention (RKFA), which amplifies expressive regions using features from the nasal base landmark. As a result, NKF achieved impressive accuracies, attaining 64.87% on AffectNet-8, 93.16% on RAF-DB, and 91.44% on FERPlus. These results consistently outperform previous state-of-the-art methods. Furthermore, fine-tuning the NKF model pretrained on AffectNet-8 for RAF-DB and FERPlus led to additional performance improvements by reducing the domain gap between the datasets. Notably, the NKF demonstrated an excellent generalization ability, maintaining high accuracy even for difficult-to-recognize expressions, such as contempt. Despite these achievements, our approach has some limitations, including its reliance on the accuracy of face alignment and a relatively simple network architecture. Future work will address these issues by integrating global features via a Vision Transformer (ViT)-based structure and implementing a Keypoint weighting strategy that can estimate landmark uncertainty and increase resilience to occlusions and illumination variations. Through comprehensive empirical validation, we demonstrated the potential of Keypoint-based FER, laying a solid foundation for its extension to other landmark-driven tasks, such as human pose estimation. From an academic perspective, this work represents a meaningful step toward building FER systems that deliver a robust and high-accuracy performance in unconstrained real-world environments.

Author Contributions

Conceptualization, J.S.; methodology, J.S.; validation, Y.H.; formal analysis, J.S.; investigation, J.S.; resources, Y.H.; writing—original draft preparation, J.S.; writing—review and editing, Y.H.; visualization, J.S.; supervision, Y.H.; project administration, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data were obtained from third parties and are available from the authors with the permission of the respective data providers. Specifically, AffectNet is available upon request and approval from the original authors at http://mohammadmahoor.com/pages/databases/affectnet/, accessed on 1 January 2025, and RAF-DB can be accessed with permission from http://www.whdeng.cn/RAF/model1.html, accessed on 25 October 2021. FERPlus is publicly available and can be accessed at https://github.com/microsoft/FERPlus accessed on 25 April 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zheng, C.; Mendieta, M.; Chen, C. POSTER: A Pyramid Cross-Fusion Transformer Network for Facial Expression Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Paris, France, 2–6 October 2023; pp. 3146–3155. [Google Scholar]
  2. Mao, J.; Xu, R.; Yin, X.; Chang, Y.; Nie, B.; Huang, A.; Wang, Y. Poster++: A simpler and stronger Facial Expression Recognition network. Pattern Recognit. 2024, 148, 110951. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Wang, C.; Ling, X.; Deng, W. Learn from all: Erasing attention consistency for noisy label facial expression recognition. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 418–434. [Google Scholar]
  4. Zhao, Z.; Liu, Q.; Zhou, F. Robust lightweight Facial Expression Recognition network with label distribution training. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Online, 2–9 February 2021; Volume 35, pp. 3510–3519. [Google Scholar]
  5. Mollahosseini, A.; Hasani, B.; Mahoor, M.H. Affectnet: A database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 2017, 10, 18–31. [Google Scholar] [CrossRef]
  6. Li, S.; Deng, W.; Du, J. Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2852–2861. [Google Scholar]
  7. Barsoum, E.; Zhang, C.; Ferrer, C.C.; Zhang, Z. Training deep networks for Facial Expression Recognition with crowd-sourced label distribution. In Proceedings of the 18th ACM International Conference on Multimodal Interaction (ICMI), Tokyo, Japan, 12–16 November 2016; pp. 279–283. [Google Scholar]
  8. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  9. Wang, J.; Feng, Z.; Ning, X.; Lin, Y.; Chen, B.; Jia, Z. Two-stream Dynamic Heterogeneous Graph Recurrent Neural Network for Multi-label Multi-modal Emotion Recognition. IEEE Trans. Affect. Comput. 2025, 1–14. [Google Scholar] [CrossRef]
  10. Li, Y.; Wang, M.; Gong, M.; Lu, Y.; Liu, L. FER-former: Multimodal Transformer for Facial Expression Recognition. IEEE Trans. Multimed. 2024, 27, 2412–2422. [Google Scholar] [CrossRef]
  11. Li, H.; Niu, H.; Zhu, Z.; Zhao, F. Cliper: A unified vision-language framework for in-the-wild facial expression recognition. In Proceedings of the 2024 IEEE International Conference on Multimedia and Expo (ICME), Niagara Falls, ON, Canada, 15–19 July 2024; pp. 1–6. [Google Scholar]
  12. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16×16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 3–7 May 2021. [Google Scholar]
  13. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  14. Liu, X.; Vijaya Kumar, B.V.K.; You, J.; Jia, P. Adaptive deep metric learning for identity-aware facial expression recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPR Workshops), Honolulu, HI, USA, 21–26 July 2017; pp. 20–29. [Google Scholar]
  15. Meng, Z.; Liu, P.; Cai, J.; Han, S.; Tong, Y. Identity-aware convolutional neural network for facial expression recognition. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 558–565. [Google Scholar]
  16. Cai, J.; Meng, Z.; Khan, A.S.; Li, Z.; O’Reilly, J.; Tong, Y. Island loss for learning discriminative features in facial expression recognition. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 302–309. [Google Scholar]
  17. Li, Y.; Lu, Y.; Li, J.; Lu, G. Separate loss for basic and compound Facial Expression Recognition in the wild. In Proceedings of the Asian Conference on Machine Learning (ACML), Nagoya, Japan, 17–19 November 2019; pp. 897–911. [Google Scholar]
  18. Farzaneh, A.H.; Qi, X. Discriminant distribution-agnostic loss for Facial Expression Recognition in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPR Workshops), Online, 14–19 June 2020; pp. 406–407. [Google Scholar]
  19. Farzaneh, A.H.; Qi, X. Facial expression recognition in the wild via deep attentive center loss. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Online, 5–9 January 2021; pp. 2402–2411. [Google Scholar]
  20. Xuefeng, C.; Huang, L. A Lightweight Model Enhancing Facial Expression Recognition with Spatial Bias and Cosine-Harmony Loss. Computation 2024, 12, 201. [Google Scholar] [CrossRef]
  21. Zeng, J.; Shan, S.; Chen, X. Facial expression recognition with inconsistently annotated datasets. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 222–237. [Google Scholar]
  22. Wang, K.; Peng, X.; Yang, J.; Lu, S.; Qiao, Y. Suppressing uncertainties for large-scale facial expression recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Online, 14–19 June 2020; pp. 6897–6906. [Google Scholar]
  23. Otberdout, N.; Kacem, A.; Daoudi, M.; Ballihi, L.; Berretti, S. Deep Covariance Descriptors for Facial Expression Recognition. In Proceedings of the British Machine Vision Conference (BMVC), Newcastle, UK, 3–6 September 2018; p. 159. [Google Scholar]
  24. Fan, Y.; Lam, J.C.K.; Li, V.O.K. Video-based emotion recognition using deeply-supervised neural networks. In Proceedings of the 20th ACM International Conference on Multimodal Interaction (ICMI), Boulder, CO, USA, 16–20 October 2018; pp. 584–588. [Google Scholar]
  25. Yang, H.; Ciftci, U.; Yin, L. Facial expression recognition by de-expression residue learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 2168–2177. [Google Scholar]
  26. Acharya, D.; Huang, Z.; Pani Paudel, D.; Van Gool, L. Covariance pooling for facial expression recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPR Workshops), Salt Lake City, UT, USA, 18–22 June 2018; pp. 367–374. [Google Scholar]
  27. Chen, Y.; Wang, J.; Chen, S.; Shi, Z.; Cai, J. Facial motion prior networks for facial expression recognition. In Proceedings of the 2019 IEEE Visual Communications and Image Processing (VCIP), Taipei, Taiwan, 1–4 December 2019; pp. 1–4. [Google Scholar]
  28. Fu, Y.; Wu, X.; Li, X.; Pan, Z.; Luo, D. Semantic neighborhood-aware deep facial expression recognition. IEEE Trans. Image Process. 2020, 29, 6535–6548. [Google Scholar] [CrossRef] [PubMed]
  29. Wu, F.; Pang, C.; Zhang, B. FaceCaps for facial expression recognition. Comput. Animat. Virtual Worlds 2021, 32, e2021. [Google Scholar] [CrossRef]
  30. Zhang, S.; Zhang, Y.; Zhang, Y.; Wang, Y.; Song, Z. A dual-direction attention mixed feature network for facial expression recognition. Electronics 2023, 12, 3595. [Google Scholar] [CrossRef]
  31. Wang, K.; Peng, X.; Yang, J.; Meng, D.; Qiao, Y. Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 2020, 29, 4057–4069. [Google Scholar] [CrossRef]
  32. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  33. Deng, J.; Guo, J.; Ververas, E.; Kotsia, I.; Zafeiriou, S. RetinaFace: Single-Shot Multi-Level Face Localisation in the Wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Online, 14–16 June 2020. [Google Scholar]
  34. Kazemi, V.; Sullivan, J. One millisecond face alignment with an ensemble of regression trees. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 24–27 June 2014; pp. 1867–1874. [Google Scholar]
  35. Kollias, D.; Cheng, S.; Ververas, E.; Kotsia, I.; Zafeiriou, S. Deep neural network augmentation: Generating faces for affect analysis. Int. J. Comput. Vis. 2020, 128, 1455–1484. [Google Scholar] [CrossRef]
  36. Blanz, V.; Vetter, T. Face recognition based on fitting a 3D morphable model. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1063–1074. [Google Scholar] [CrossRef]
  37. Chen, S.; Wang, J.; Chen, Y.; Shi, Z.; Geng, X.; Rui, Y. Label distribution learning on auxiliary label space graphs for facial expression recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Online, 14–16 June 2020; pp. 13984–13993. [Google Scholar]
  38. Liu, H.; An, R.; Zhang, Z.; Ma, B.; Zhang, W.; Song, Y.; Hu, Y.; Chen, W.; Ding, Y. Norface: Improving facial expression analysis by identity normalization. In Proceedings of the European Conference on Computer Vision (ECCV), Milan, Italy, 29 September–4 October 2024; pp. 293–314. [Google Scholar]
  39. Ding, H.; Zhou, P.; Chellappa, R. Occlusion-adaptive deep network for robust facial expression recognition. In Proceedings of the 2020 IEEE International Joint Conference on Biometrics (IJCB), Houston, TX, USA, 28 September–1 October 2020; pp. 1–9. [Google Scholar]
  40. Chen, Y.; Li, J.; Shan, S.; Wang, M.; Hong, R. From static to dynamic: Adapting landmark-aware image models for Facial Expression Recognition in videos. IEEE Trans. Affect. Comput. 2024, 15, 4589–4599. [Google Scholar] [CrossRef]
  41. Chen, C. PyTorch Face Landmark: A Fast and Accurate Facial Landmark Detector. 2021. Available online: https://github.com/cunjian/pytorch_face_landmark (accessed on 24 October 2023).
  42. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  43. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  44. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  45. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  46. Goodfellow, I.J.; Erhan, D.; Carrier, P.L.; Courville, A.; Mirza, M.; Hamner, B.; Cukierski, W.; Tang, Y.; Thaler, D.; Lee, D.-H.; et al. Challenges in representation learning: A report on three machine learning contests. In Proceedings of the 20th International Conference of the Neural Information Processing (ICONIP 2013), Daegu, Republic of Korea, 3–7 November 2013; Proceedings, Part III 20. Springer: Daegu, Republic of Korea, 2013; pp. 117–124. [Google Scholar]
  47. So, J.; Han, Y. 3D face alignment through fusion of head pose information and features. Image Vis. Comput. 2024, 151, 105253. [Google Scholar] [CrossRef]
  48. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  49. Guo, Y.; Zhang, L.; Hu, Y.; He, X.; Gao, J. Ms-celeb-1m: A dataset and benchmark for large-scale face recognition. In Proceedings of the 14th European Conference of the Computer Vision (ECCV 2016), Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part III 14. Springer: Amsterdam, The Netherlands, 2016; pp. 87–102. [Google Scholar]
  50. Foret, P.; Kleiner, A.; Mobahi, H.; Neyshabur, B. Sharpness-aware Minimization for Efficiently Improving Generalization. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 3–7 May 2021. [Google Scholar]
  51. Kwon, J.; Kim, J.; Park, H.; Choi, I.K. ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks. In Proceedings of the 38th International Conference on Machine Learning (ICML), Virtual Event, 18–24 July 2021; pp. 5905–5914. [Google Scholar]
  52. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations (ICLR)—Poster Session, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  53. Li, H.; Wang, N.; Ding, X.; Yang, X.; Gao, X. Adaptively learning facial expression representation via cf labels and distillation. IEEE Trans. Image Process. 2021, 30, 2016–2028. [Google Scholar] [CrossRef] [PubMed]
  54. Wen, Z.; Lin, W.; Wang, T.; Xu, G. Distract your attention: Multi-head cross attention network for facial expression recognition. Biomimetics 2023, 8, 199. [Google Scholar] [CrossRef] [PubMed]
  55. Shi, J.; Zhu, S.; Wang, D.; Liang, Z. ARM: A lightweight module to amend facial expression representation. Signal Image Video Process. 2023, 17, 1315–1323. [Google Scholar] [CrossRef]
  56. Xue, F.; Wang, Q.; Guo, G. Transfer: Learning relation-aware facial expression representations with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Online, 10 March 2021; pp. 3601–3610. [Google Scholar]
  57. Bohi, A.; Boudouri, Y.E.; Sfeir, I. A novel deep learning approach for facial emotion recognition: Application to detecting emotional responses in elderly individuals with Alzheimer’s disease. Neural Comput. Appl. 2024, 36, 8483–8501. [Google Scholar] [CrossRef]
  58. Fölster, M.; Hess, U.; Werheid, K. Facial age affects emotional expression decoding. Front. Psychol. 2014, 5, 30. [Google Scholar] [CrossRef] [PubMed]
  59. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting Unreasonable Effectiveness of Data in Deep Learning Era. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 843–852. [Google Scholar]
  60. Ridnik, T.; Ben-Baruch, E.; Noy, A.; Zelnik, L. ImageNet-21K Pretraining for the Masses. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, Virtual Event, 22–29 September 2021; Volume 1. [Google Scholar]
Figure 1. Flow of the proposed FER method. Global features are computed from the backbone and face alignment networks. Keypoint features are then extracted from these global features based on facial landmark coordinates from the face alignment network, followed by refinement using Representative Keypoint Feature Attention, and finally, classification.
Figure 1. Flow of the proposed FER method. Global features are computed from the backbone and face alignment networks. Keypoint features are then extracted from these global features based on facial landmark coordinates from the face alignment network, followed by refinement using Representative Keypoint Feature Attention, and finally, classification.
Sensors 25 03762 g001
Figure 2. Overview of NKF structure for FER. The Global Feature Network detects facial landmarks and computes feature maps. The Keypoint Feature Network extracts the Keypoint Features from the Global Features at the facial landmark coordinates. Subsequently, they are improved by a Representative Keypoint Feature Attention module and finally classified for FER.
Figure 2. Overview of NKF structure for FER. The Global Feature Network detects facial landmarks and computes feature maps. The Keypoint Feature Network extracts the Keypoint Features from the Global Features at the facial landmark coordinates. Subsequently, they are improved by a Representative Keypoint Feature Attention module and finally classified for FER.
Sensors 25 03762 g002
Figure 3. t-SNE-based [43] 2D projection of Keypoint Features extracted from test dataset samples after training the proposed network on the RAF-DB training set. The points at the top left represent the locations of the 68 mean landmarks of the 3D Morphable Model [36]. The projection results at the top right display only red dots, excluding blue dots for ease of visual analysis, and are placed at the corresponding locations. The blue box at the bottom shows an enlarged view of the distribution of the Keypoint Features corresponding to the outer face contour, and the red box shows an enlarged view of the distribution of the Keypoint Features corresponding to the upper lip.
Figure 3. t-SNE-based [43] 2D projection of Keypoint Features extracted from test dataset samples after training the proposed network on the RAF-DB training set. The points at the top left represent the locations of the 68 mean landmarks of the 3D Morphable Model [36]. The projection results at the top right display only red dots, excluding blue dots for ease of visual analysis, and are placed at the corresponding locations. The blue box at the bottom shows an enlarged view of the distribution of the Keypoint Features corresponding to the outer face contour, and the red box shows an enlarged view of the distribution of the Keypoint Features corresponding to the upper lip.
Sensors 25 03762 g003
Figure 4. Generating inaccurately perturbed landmarks by adding random offsets during the training phase. The intentional incorporation of these inaccuracies in the landmark coordinates during training serves to extract Keypoint Features and enhance the network’s robustness and generalization performance. This process effectively makes the network less sensitive to potential errors in landmark detection.
Figure 4. Generating inaccurately perturbed landmarks by adding random offsets during the training phase. The intentional incorporation of these inaccuracies in the landmark coordinates during training serves to extract Keypoint Features and enhance the network’s robustness and generalization performance. This process effectively makes the network less sensitive to potential errors in landmark detection.
Sensors 25 03762 g004
Figure 5. t-SNE [40] 2D projection illustrating Keypoint Feature distributions for samples from the RAF-DB test dataset, pre- and post-RKFA. The upper distribution shows the state before RKFA, and the lower distribution shows the state after RFKA is applied, all of which are depicted on the left side. The red, blue, and green boxes show the enlarged distribution of nasal base Keypoint Features, face outline Keypoint Features, and refined Keypoint Features after applying RKFA, respectively.
Figure 5. t-SNE [40] 2D projection illustrating Keypoint Feature distributions for samples from the RAF-DB test dataset, pre- and post-RKFA. The upper distribution shows the state before RKFA, and the lower distribution shows the state after RFKA is applied, all of which are depicted on the left side. The red, blue, and green boxes show the enlarged distribution of nasal base Keypoint Features, face outline Keypoint Features, and refined Keypoint Features after applying RKFA, respectively.
Sensors 25 03762 g005
Figure 6. Confusion matrices of NKF evaluated on the FER datasets. (a) AffectNet-7 showing 68.17% accuracy of the NKF model, (b) AffectNet-8 showing 64.87% accuracy, (c) RAF-DB showing 93.12% accuracy, (d) FERPlus dataset showing 91.44% accuracy, (e) RAF-DB showing 94.04% accuracy pretrained on AffectNet-8, and (f) FERPlus showing 91.66% accuracy pretrained on AffectNet-8.
Figure 6. Confusion matrices of NKF evaluated on the FER datasets. (a) AffectNet-7 showing 68.17% accuracy of the NKF model, (b) AffectNet-8 showing 64.87% accuracy, (c) RAF-DB showing 93.12% accuracy, (d) FERPlus dataset showing 91.44% accuracy, (e) RAF-DB showing 94.04% accuracy pretrained on AffectNet-8, and (f) FERPlus showing 91.66% accuracy pretrained on AffectNet-8.
Sensors 25 03762 g006
Figure 7. FLOPs versus accuracy: (a) depicts the accuracy on RAF-DB, and (b) depicts the accuracy on AffectNet-7. The red, pink, blue, yellow, black, and green symbols denote NKF, POSTER++ [2], DDAMFN [30], LFNSB [20], TransFER [56], and DAN [54], respectively.
Figure 7. FLOPs versus accuracy: (a) depicts the accuracy on RAF-DB, and (b) depicts the accuracy on AffectNet-7. The red, pink, blue, yellow, black, and green symbols denote NKF, POSTER++ [2], DDAMFN [30], LFNSB [20], TransFER [56], and DAN [54], respectively.
Sensors 25 03762 g007
Figure 8. Three types of facial landmarks to extract Keypoint Features. Blue dots indicate unused landmarks; red dots indicate used landmarks. ‘Half’ denotes the selection of landmarks at odd-numbered indices. ‘Inner’ refers to the selection of landmarks within the face. ‘Full’ denotes the selection of all landmarks.
Figure 8. Three types of facial landmarks to extract Keypoint Features. Blue dots indicate unused landmarks; red dots indicate used landmarks. ‘Half’ denotes the selection of landmarks at odd-numbered indices. ‘Inner’ refers to the selection of landmarks within the face. ‘Full’ denotes the selection of all landmarks.
Sensors 25 03762 g008
Figure 9. Examples of Keypoint Features. A green dot indicates a facial landmark, a red dot indicates a nasal base landmark, and a blue dot indicates the center location of a feature map.
Figure 9. Examples of Keypoint Features. A green dot indicates a facial landmark, a red dot indicates a nasal base landmark, and a blue dot indicates the center location of a feature map.
Sensors 25 03762 g009
Figure 10. Examples of failed cases. The top row displays the input image with the overlaid predicted facial landmarks. The bottom row presents the same input image along with the predicted facial expression. The green dots denote the predictions of the face alignment network.
Figure 10. Examples of failed cases. The top row displays the input image with the overlaid predicted facial landmarks. The bottom row presents the same input image along with the predicted facial expression. The green dots denote the predictions of the face alignment network.
Sensors 25 03762 g010
Table 1. Settings of the difference between the datasets.
Table 1. Settings of the difference between the datasets.
ParametersAffectNetRAF-DBFERPlus
Initial Learning Rate1.0 × 10−65.0 × 10−65.0 × 10−6
Epoch50200200
Landmark Perturbation0.050.100.5
Random Erasing Scale(0.05, 0.05)(0.02, 0.1)(0.05, 0.05)
Table 2. Comparison with state-of-the-art methods. The best scores are in bold.
Table 2. Comparison with state-of-the-art methods. The best scores are in bold.
MethodAccuracy (%)
AffectNet-7AffectNet-8RAF-DBFERPlus
KTN [53]63.97-88.0790.49
DAN [54] 65.6962.0989.70-
EAC [3]65.32-90.3589.64
ARM [55]65.2061.3390.42-
TransFER [56]66.23-90.9190.83
LFNSB [20]66.5763.1291.07-
S2D [40]66.4263.7692.2191.01
EmoNeXt [57]67.4664.13--
POSTER++ [2]67.4963.7792.21-
DDAMFN [30]67.0364.2591.3590.74
Norface [38]68.69-92.97-
NKF(Ours)68.1764.8793.1691.44
S2D † [40]--92.57 (+0.36)91.17 (+0.16)
NKF † (Ours)--94.04 (+0.88)91.66 (+0.22)
† means pretrained on AffectNet-8.
Table 3. Comparison with state-of-the-art methods for each facial expression. The best scores for each expression and dataset are in bold.
Table 3. Comparison with state-of-the-art methods for each facial expression. The best scores for each expression and dataset are in bold.
DataMethodAccuracy (%)
NeuHapSadSurFeaDisAngConAvg
Aff7KTN [53]66.6086.8060.8055.2064.0060.0054.40-63.97
ARM [55]65.0087.0064.0061.0062.0053.0064.00-65.20
LFNSB [20] 68.7987.5069.4466.6760.7150.0057.89-65.86
POSTER++ [2]65.4089.4068.0066.0064.2055.4065.00-67.45
DDAMFN [30]66.2087.2066.4063.8066.0056.6063.00-67.03
NKF (Ours)64.4086.4065.8066.2068.0064.2062.20-68.17
Aff8ARM [55]63.0086.0064.0061.0062.0052.0063.0040.0061.33
LFNSB [20] 52.3882.3568.7580.0062.5044.4446.1557.1461.71
POSTER++ [2]60.6076.4066.8065.6063.0058.0060.2059.5263.76
DDAMFN [30]57.8081.8063.0065.6064.2060.0062.0059.6064.25
NKF (Ours)58.4078.4065.0065.0067.0063.2062.2059.7264.87
RAFKTN [53]88.5394.6087.2483.2868.9265.6281.48-81.38
ARM [55]97.9095.4083.9090.3070.3064.4077.20-82.77
LFNSB [20] 92.7996.7187.2487.8467.5774.3887.65-84.88
POSTER++ [2]92.0697.2292.8990.5868.9271.8888.27-85.97
DDAMFN [30]92.4996.6089.5591.0365.7572.3384.57-84.62
NKF(Ours)95.0097.8989.1290.8870.2779.3790.74-87.61
NKF † (Ours)94.1297.8994.5693.9272.9776.8890.74-88.73
FERPlusKTN [48]92.0795.8581.3892.9657.6953.3390.4130.7774.31
DDAMFN [30]92.7796.4079.4292.8655.2962.5093.2326.6774.89
NKF (Ours)93.2196.8682.5594.7056.6361.1190.4831.2575.85
NKF † (Ours)92.4897.2083.8594.9556.6361.1191.5850.0078.48
† means pretrained on AffectNet-8.
Table 4. Performance of NKF on RAF-DB across gender, race, and age groups.
Table 4. Performance of NKF on RAF-DB across gender, race, and age groups.
AttributeMethodAccuracy (%)
Gender-MaleFemaleUnsure--
NKF (Ours)91.8393.7795.98--
NKF † (Ours)92.8794.5196.98--
Race-CaucasianAfrican AmericanAsian--
NKF (Ours)92.6495.7394.20--
NKF † (Ours)93.3297.0195.86--
Age-0–34–1920–3940–6970+
NKF (Ours)96.9694.2492.3693.2386.52
NKF † (Ours)98.1895.6892.8494.2289.89
† means pretrained on AffectNet-8.
Table 5. Comparison of model complexity. Global Net. indicates the global networks and KF Net. indicates the Keypoint Feature Networks in the NKF. The best scores are highlighted in bold.
Table 5. Comparison of model complexity. Global Net. indicates the global networks and KF Net. indicates the Keypoint Feature Networks in the NKF. The best scores are highlighted in bold.
Method#Param(M)#FLOPs(G)Accuracy (%)
RAFAff7Aff8
TransFER [56]65.2015.3090.9166.23-
DAN [54]19.722.2389.7065.6962.09
POSTER++ [2]43.708.4092.2167.4963.77
DDAMFN [30]4.110.5591.3567.0364.25
LFNSB [20]2.680.3891.0766.5763.12
NKF †
(Ours)
Global Net.25.247.01 - - -
KF
Net.
Penultimate Layer40.110.04 - - -
Others1.190.08 - - -
All66.547.1394.0468.1764.87
† means pretrained on AffectNet-8.
Table 6. Evaluation of different components. The best scores are highlighted in bold.
Table 6. Evaluation of different components. The best scores are highlighted in bold.
ComponentChoice
InputOriginal-----
Accurately Aligned Image --
NetworkKeypoint Features--
KF Regularization----
RKFA----
Accuracy (%)62.9464.3963.9764.4464.5964.5764.87
Table 7. Evaluation of landmark types. The best scores are in bold.
Table 7. Evaluation of landmark types. The best scores are in bold.
Landmark TypeAccuracy (%)
Half, 34 landmarks64.51
Inner, 51 landmarks64.52
Full, 68 landmarks64.87
Table 8. Effectiveness of the RKFA. The best scores are in bold.
Table 8. Effectiveness of the RKFA. The best scores are in bold.
Attention TypeAccuracy (%)
Without RKFA64.59
Center point of the Feature Map64.69
Nasal base64.87
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

So, J.; Han, Y. Facial Landmark-Driven Keypoint Feature Extraction for Robust Facial Expression Recognition. Sensors 2025, 25, 3762. https://doi.org/10.3390/s25123762

AMA Style

So J, Han Y. Facial Landmark-Driven Keypoint Feature Extraction for Robust Facial Expression Recognition. Sensors. 2025; 25(12):3762. https://doi.org/10.3390/s25123762

Chicago/Turabian Style

So, Jaehyun, and Youngjoon Han. 2025. "Facial Landmark-Driven Keypoint Feature Extraction for Robust Facial Expression Recognition" Sensors 25, no. 12: 3762. https://doi.org/10.3390/s25123762

APA Style

So, J., & Han, Y. (2025). Facial Landmark-Driven Keypoint Feature Extraction for Robust Facial Expression Recognition. Sensors, 25(12), 3762. https://doi.org/10.3390/s25123762

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop