Next Article in Journal
Commercial Hoverboard Reverse Engineering and Repurposing for a Stabilized Platform: A Recyclable Solution for Modular Robotic Bases
Previous Article in Journal
Multi-Modal Joint Pulsed Eddy Current Sensor Signal Denoising Method Integrating Inductive Disturbance Mechanism
Previous Article in Special Issue
Building Segmentation in Urban and Rural Areas with MFA-Net: A Multidimensional Feature Adjustment Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced AlexNet with Gabor and Local Binary Pattern Features for Improved Facial Emotion Recognition

by
Furkat Safarov
1,
Alpamis Kutlimuratov
2,
Ugiloy Khojamuratova
3,
Akmalbek Abdusalomov
1 and
Young-Im Cho
1,*
1
Department of Computer Engineering, Gachon University, Sujeong-Gu, Seongnam-si 13120, Gyeonggi-Do, Republic of Korea
2
Department of Applied Informatics, Kimyo International University in Tashkent, Tashkent 100121, Uzbekistan
3
Department of Computer Science, CUNY Queens College, 65-30 Kissena Blvd Flushing, New York, NY 11374, USA
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3832; https://doi.org/10.3390/s25123832
Submission received: 10 March 2025 / Revised: 18 May 2025 / Accepted: 7 June 2025 / Published: 19 June 2025

Abstract

:
Facial emotion recognition (FER) is vital for improving human–machine interactions, serving as the foundation for AI systems that integrate cognitive and emotional intelligence. This helps bridge the gap between mechanical processes and human emotions, enhancing machine engagement with humans. Considering the constraints of low hardware specifications often encountered in real-world applications, this study leverages recent advances in deep learning to propose an enhanced model for FER. The model effectively utilizes texture information from faces through Gabor and Local Binary Pattern (LBP) feature extraction techniques. By integrating these features into a specially modified AlexNet architecture, our approach not only classifies facial emotions more accurately but also demonstrates significant improvements in performance and adaptability under various operational conditions. To validate the effectiveness of our proposed model, we conducted evaluations using the FER2013 and RAF-DB benchmark datasets, where it achieved impressive accuracies of 98.10% and 93.34% for the two datasets, with standard deviations of 1.63% and 3.62%, respectively. On the FER-2013 dataset, the model attained a precision of 98.2%, a recall of 97.9%, and an F1-score of 98.0%. Meanwhile, for the other dataset, it achieved a precision of 93.54%, a recall of 93.12%, and an F1-score of 93.34%. These results underscore the model’s robustness and its capability to deliver high-precision emotion recognition, making it an ideal solution for deployment in environments where hardware limitations are a critical concern.

1. Introduction

Emotions are fundamental human traits that play a critical role in social interaction. People express emotions through various channels, such as facial expressions [1,2], verbal communication [3,4], brain and heart signals [5,6], and bodily gestures [7]. Among these, facial expressions are particularly powerful, innate, and universally recognized as a means for humans to convey their emotions and intentions. Extensive research [8,9,10] has been carried out on automated facial expression analysis because of its relevance in fields such as interactive robotics, medical diagnostics, monitoring driver fatigue, and other human–machine interaction systems. These efforts have paved the way for the creation of advanced deep learning models for FER [11,12,13], which have significantly impacted both academic research and commercial industries. In academia, these models have improved the accuracy and speed of emotion recognition, while in industry, they have been applied in various domains, like emotion-sensitive systems, enhanced surveillance tools, and customer engagement analytics. The primary or basic facial expressions typically include emotions such as neutral, anger, disgust, fear, happiness, sadness, and surprise. These seven emotions are widely recognized as universal across various cultures and are commonly used in studies related to emotional recognition. Researchers often focus on these core emotions [14,15] because of the consistency in how they manifest across different populations and contexts, providing a solid foundation for the development of facial recognition models and systems.
The process of FER can be systematically divided into three critical stages: face detection, feature extraction, and classification. Initially, face detection serves as the foundational technology in facial recognition systems, pivotal in identifying and locating human faces within images [16,17]. Subsequently, feature extraction analyzes these detected faces to identify distinctive features that are crucial for recognizing different emotions [18,19]. Finally, the classification module [20,21,22,23] categorizes these emotions based on the extracted features, thereby completing the process of FER. Each of these modules plays a vital role in the accurate and efficient processing of facial emotions, ensuring that the system performs effectively across varied contexts and environments.
Moreover, the research in [24] introduces a novel parameter selection approach that employs swarm intelligence and a fitness function for the intelligent recognition of micro-emotions. Additionally, the study presents a method that utilizes geometric visual representations derived from facial landmark points. These representations form the basis of input features, which include normalized angles and distances calculated from the landmarks. These features are then analyzed using a Deep Neural Network model to enhance the accuracy of emotion recognition. Similarly, Palestra, Giuseppe, et al. [25] suggest an automated system for recognizing facial expressions, which uses a novel collection of 32 geometric features from one side of the face. This approach encompasses a broad range of geometric details to enhance the accuracy of identifying six different emotional expressions. In their study, Chen et al. [26] employ a softmax regression-based deep sparse auto-encoder network for FER in human–robot interactions, which enhances learning efficiency and dimensional complexity while accurately extracting features and classifying input signals.
While larger and more complex models for FER are being developed, industries demand solutions that are both accurate and lightweight for real-world deployment. Knowledge distillation (KD) has proven effective in addressing this need by transferring knowledge from high-performing teacher models to compact student models. Compared to other compression techniques, like pruning and quantization, KD is especially effective in preserving performance while reducing computational load. It enables student models to learn fine-grained representations and outputs, narrowing the gap with their teacher counterparts. This approach is particularly important for deploying FER models on resource-constrained devices, such as smartphones and Internet of Things platforms. KD enhances model generalization and robustness against variations in lighting, expression, and noise. It has been successfully applied in practical FER tasks, from surveillance to mobile emotion detection, offering a scalable and efficient solution. Ultimately, KD bridges the divide between high accuracy and low-latency requirements in FER systems.
Interactive KD [27] introduces dynamic teaching strategies by swapping in operations, where blocks in the student network are randomly replaced by those from the teacher network. This enhances the student’s ability to learn advanced feature transformations but increases computational demands and memory usage, limiting its suitability for resource-constrained settings. In a previous study [28], a KD method is proposed to transfer high-resolution features from a teacher to a student trained on low-resolution inputs. By leveraging multi-level teacher features, which are concatenated, weighted, and used to supervise a single-level student output, the model effectively extracts rich features from low-resolution data. A key limitation is the potential mismatch between the teacher’s high-resolution features and the student’s low-resolution input, which may hinder effective knowledge transfer and result in suboptimal student performance, especially when the resolution gap is significant.
This study aims to answer the following research questions (RQs):
-
RQ1: How can we enhance the efficiency and accuracy of FER using a lightweight deep learning model?
-
RQ2: To what extent does the integration of Gabor and LBP features and a modified AlexNet architecture improve the performance of FER in resource-constrained environments?
-
RQ3: Can the combination of depthwise-separable convolution and pointwise convolution significantly reduce the computational costs of FER models without sacrificing their performance in terms of emotion detection accuracy?
To overcome the shortcomings of the current KD and larger and more complex FER methods, a new model is created. This study introduces a lightweight and efficient FER model that leverages Gabor and LBP features, integrated with a modified AlexNet architecture. The proposed model is designed to enhance both accuracy and computational efficiency by refining the standard AlexNet-based network. These combinations enable the model to deliver competitive results while maintaining a compact structure ideal for real-world deployment.
This approach ensures high accuracy while reducing complexity, making it suitable for resource-constrained environments like mobile devices. The proposed method simplifies training and enhances scalability, providing a robust solution for real-world applications, such as emotion analysis and human–computer interactions. To extract meaningful texture information, Gabor and LBP features are extracted and fused. The resulting data is then fed into a modified AlexNet architecture. This integration leverages the complementary strengths of the two feature extraction techniques to enhance the model’s ability to interpret and analyze facial expressions.
A modified version of AlexNet serves as the core model, specifically adapted to improve its efficacy in classification tasks. The enhancement incorporates depthwise-separable convolutions within this modified AlexNet, a technique that significantly cuts down computational demands while preserving high accuracy. This adjustment renders the model particularly advantageous for environments with limited resources.
The following points outline the key contributions of this work:
-
A novel FER method was developed, which utilizes Gabor and LBP features to extract texture information, along with a modified AlexNet for classification;
-
The developed FER method was optimized for limited hardware resources;
-
A modified AlexNet model was successfully developed and validated for classification purposes in this research;
-
Experiments conducted on the FER2013 [29] and RAF-DB [30] datasets demonstrated superior performance, highlighting the effectiveness of the proposed approach.
These advancements significantly contribute to the progress of FER research. By presenting novel methodologies and actionable insights, the proposed model establishes a solid basis for future investigations and real-world implementations.

2. Proposed Method

Our proposed FER model workflow is illustrated in Figure 1, where the input consists of facial images. Initially, Gabor and LBP features are extracted from these images. Subsequently, the extracted features are fed into a modified AlexNet architecture, which is utilized to classify them. This process integrates advanced feature extraction techniques with a robust deep learning model to enhance the accuracy and efficiency of emotion recognition.

2.1. Feature Extraction

2.1.1. Gabor Feature

Gabor filters, which are linear filters, are extensively employed in image processing and analysis for applications such as edge detection, surface evaluation, feature extraction, and object recognition [31,32]. The Gabor filter, akin to the human visual system’s sensitivity to frequency and orientation, is effective in extracting nuanced facial features across various spatial positions and directions, making it ideal for facial emotion recognition tasks. A key benefit of Gabor filters is their ability to remain invariant to changes in scale, rotation, and translation [33]. Gabor wavelets are highly effective in facial emotion recognition because of their acute sensitivity to the dynamic changes that occur in crucial facial features, such as the eyes, mouth, and eyebrows. These specific regions are prone to significant alterations in grayscale intensity as facial muscles contract or relax during emotional expressions. The sensitivity of Gabor wavelets to these changes is due to their ability to detect fluctuations in both the real and imaginary components of the wavelet, tailored to capture different orientations and scales within the image. As facial expressions change, these fluctuations manifest in Gabor wavelets as varied responses. A Gabor filter, in turn, produces strong amplitude responses when it encounters variations in the wavelet components caused by motion in the facial features. This characteristic of Gabor filters makes them especially capable of distinguishing and isolating local facial features that are pivotal in differentiating between emotions. Gabor filters’ robustness in capturing the essence of emotional expressions through detailed textural and directional information significantly enhances the performance and reliability of emotion recognition systems. The kernel function for the two-dimensional Gabor wavelet can be expressed as follows (Equation (1)):
ψ u v = k u v 2 σ 2 × e k u v 2 z 2 2 σ 2 × ( e i k u v z e σ 2 2 )
In the Gabor wavelet kernel, u and v denote the direction and frequency, respectively, while z = (x, y) indicates the position of a specific pixel within the image. The symbol σ represents the filter’s bandwidth. The term | k u v | 2 σ 2 compensates for the energy spectrum’s attenuation, which is influenced by the frequency. For face emotion recognition, the Gabor feature is crucial as it is obtained by convolving the image that displays a facial expression with a Gabor wavelet kernel. This convolution process effectively captures the textural and edge-related characteristics of facial expressions, which are critical for identifying different emotions. In this convolution operation, the kernel function—specifically designed to detect specific frequencies and orientations—interacts with the image. If we denote k as the grayscale value at any given point (x, y) in the image, the resulting Gabor feature at that point is calculated using a specific Equation (2). This equation integrates the values of the image and the responses of the Gabor kernel, providing a robust feature set that enhances the accuracy and sensitivity of emotion recognition systems.
G u v ( x , y ) = I ( x , y ) ψ u v ( x , y )
where G u v ( x , y ) denotes the Gabor feature of the extracted image, ψ u v ( x , y ) is the kernel function for the two-dimensional Gabor wavelet, and the symbol ∗ indicates the convolution operation.

2.1.2. LBP Feature

LBP is exceptionally effective in capturing the texture information pertinent to FER [34,35,36]. This technique works by comparing each pixel with its neighbors, encoding these relationships into a binary pattern that effectively represents facial textures. This attribute is crucial in distinguishing between different facial expressions, as changes in emotion are often subtly reflected in changes in texture patterns on the face. The adaptability of LBP allows for customization in terms of sampling points and neighborhood radius, enhancing its ability to detect nuanced emotional expressions at various scales. When applied to face emotion recognition, LBP can be used in conjunction with preprocessing methods, like histogram equalization, to improve the visibility of key facial features under varying lighting conditions.
The LBP operator targets 3 × 3 pixel neighborhoods, treating the middle pixel as the central reference point, and evaluates the eight adjacent pixels according to a predefined threshold. To expand further, the mechanism of LBP involves examining each pixel within a small, localized grid. The central pixel in this grid serves as a benchmark, against which the gray levels of the adjacent eight pixels are compared. Each neighbor is given a binary value based on whether its intensity is higher or lower than the central pixel. These binary values are then arranged into an eight-digit binary number, reflecting the texture around the central pixel. This binary representation allows LBP to effectively capture and encode local textural patterns in image. The LBP operator is structured as follows (Equation (3)):
LBP ( x c , y c ) = n = 0 7 2 n s ( i n i c )
For the eight neighboring pixels around the central pixel c, where i n   and   i c represent the grayscale values at c and at each neighbor n, respectively, the function s(u) is defined as 1 if u ≥ 0 and 0 if u < 0.
An LBP operator that yields a binary code featuring at most one transition from 0 to 1 and another from 1 to 0 indicates the formation of what is termed a “uniform pattern”. These uniform patterns are particularly valuable in image analysis because they represent basic but essential geometric structures, such as the presence of edges and corners within the image. The simplicity and regularity of these patterns make them pivotal in distinguishing structured textures and shapes. The concept of uniform patterns is crucial for efficiently capturing the image’s structural essence without excessive complexity. In practical terms, the occurrence of uniform patterns allows for a significant reduction in the length of the feature vector. A shorter feature vector not only simplifies the computational demands but also improves the efficiency of the processing tasks. Moreover, by focusing on these uniform patterns, it is possible to implement a descriptor that remains consistent (invariant) across different image rotations.

2.2. FER with Modified AlexNet

In our study, we implemented significant modifications to the AlexNet architecture [37], a well-established convolutional neural network initially designed for general image classification tasks, to tailor it specifically for facial emotion recognition. This adaptation enhances AlexNet’s ability to interpret subtle facial expressions by refining its convolutional layers and tuning hyperparameters to focus on the nuanced features critical for distinguishing between different emotions. Our enhanced model leverages the depth and robustness of AlexNet while optimizing it to better capture the complex emotional states expressed in facial images, aiming to increase both the accuracy and sensitivity of emotion detection. AlexNet, recognized for its complex and deep structure, was reengineered into a version that is both more streamlined and computationally efficient, yet still retains its strong capability for feature extraction.
Depthwise-separable convolution layers were introduced to the classic AlexNet architecture, effectively reducing the number of parameters and computational costs. This modification maintains the network’s capacity to learn complex features without the usual overhead associated with standard convolution layers. The classic AlexNet employs standard convolutional layers that utilize multiple filters over all input channels, leading to significant computational expenses. These layers process extensive data simultaneously, which, while effective for feature extraction, requires substantial computational resources and can slow down processing speeds, particularly in systems with limited hardware capabilities.
Our model employs depthwise convolution [38], a method that applies an individual filter to each separate input channel. This approach significantly decreases the number of parameters and reduces the computational burden. By treating each channel independently, the model becomes more efficient, allowing for faster processing and reduced resource consumption. After the depthwise convolution, a 1 × 1 convolution, also known as pointwise convolution, is applied to integrate the outputs from the depthwise step. This process is designed to preserve the model’s capacity to capture complex features while additionally lowering computational expenses. By condensing the outputs through pointwise convolution, the model efficiently consolidates feature information, thereby optimizing the learning process and minimizing resource usage without compromising on performance.
So, our modified AlexNet architecture, as shown in Figure 2, features five depthwise-separable convolution layers, paired with five ReLU activation functions, includes three max pooling layers, and concludes with a flatten layer that connects to an output layer equipped with a softmax activation function.

3. Results

3.1. Datasets and Evaluation Metrics

This research was conducted using two well-known benchmark datasets, providing a reliable foundation for evaluating the proposed methods. By utilizing these datasets, this study ensures a robust and comprehensive assessment of performance, enabling comparisons with existing approaches and validating the effectiveness of the proposed techniques. This approach highlights the importance of leveraging standard benchmarks for consistency and reproducibility in research.
(1)
The FER2013 [29] dataset consists of grayscale images, each labeled with the corresponding emotion displayed by the person. It features 48 × 48 pixel images categorized into seven emotions (Figure 3): angry, disgust, fear, happy, sad, surprise, and neutral. The dataset is divided into three parts: 28,709 images for training, 3589 for public testing, and 3589 for private testing.
(2)
The RAF-DB [30] dataset comprises 15,339 facial images collected from real-world scenarios, showcasing a broad spectrum of emotions displayed by individuals in natural settings. The dataset includes a variety of emotions (Figure 4), such as anger, disgust, fear, happiness, sadness, surprise, and neutral expressions.
Figure 3. Emotion distribution in the FER2013 dataset.
Figure 3. Emotion distribution in the FER2013 dataset.
Sensors 25 03832 g003
Figure 4. Emotion distribution in the RAF-DB dataset.
Figure 4. Emotion distribution in the RAF-DB dataset.
Sensors 25 03832 g004
FER requires the use of specific metrics to accurately assess the performance of machine learning models. In this study, we employed several standard classification metrics to gain valuable insights into algorithm effectiveness and to facilitate comparative analysis. The metrics used comprise accuracy, precision, recall, and the F1-score [39], with their respective formulas provided in Equations (4)–(7).

3.1.1. Precision

Precision is a key metric in FER, used to evaluate a model’s ability to correctly identify true emotional expressions. It is determined by dividing the number of true positives (correctly predicted positive cases) by the total positive predictions, which include both true positives (TPs) and false positives (FPs).
Precision = TP TP + FP

3.1.2. Recall

Recall in FER measures the model’s capability to accurately detect relevant emotional expressions. It evaluates the fraction of actual positive instances that are correctly classified, highlighting the algorithm’s performance in identifying true emotional observations. False negatives (FNs) occur when the actual emotion is class A (e.g., happy), but the model predicts a different class (e.g., sad). So, FN stands for when the model fails to recognize that emotion.
Recall = TP TP + FN

3.1.3. F1-Score

The F-score, or F-measure, is a commonly used metric in FER that evaluates a model’s performance by integrating precision and recall. It is computed as the harmonic mean of precision and recall, as outlined in Equation (7).
F 1 = 2 × Precision × Recall Precsion + Recall

3.1.4. Accuracy

In FER, the accuracy metric is widely regarded as the standard for evaluating the performance of classification algorithms. It measures the ratio of correctly classified emotion categories to the total number of observations, making it a commonly preferred method for assessing accuracy in emotion classification tasks. True negative (TN) refers to cases where the model correctly predicts a different class when the actual class is not the target emotion.
Accuracy = TP + TN TP + TN + FP + FN

3.2. Experimental Settings and Results

We implemented the proposed model on a GeForce RTX 3090 Ti with 12 GB and an Intel Core i9-13900K, using the TensorFlow framework and Adam optimization. The training and testing regimen was extensive, covering 300 epochs with batches of 32. For smooth functionality and optimal processing speeds, the system operated on Windows 10, supported by 32 GB of RAM. The statistical analysis was conducted using the SciPy library in Python v.3.10. To guarantee a fair assessment of our model using the FER2013 and RAF-DB datasets, we implemented a comprehensive end-to-end training strategy. We meticulously divided our dataset, allocating 80% for training to provide sufficient data for the model’s learning and adaptation. The remaining 20% was strictly set aside for testing, allowing us to validate the model’s performance in novel situations. Cross-validation was not used in this study because of the high computational cost and the large size of the dataset, which made a single hold-out split both practical and sufficient for reliable evaluation.

Model Performances

Table 1 presents a comparison of model performance in terms of accuracy for different emotions on the two datasets: FER2013 and RAF-DB. It also includes the overall average accuracy for each dataset. On the FER2013 dataset, the highest accuracy is observed for the emotion “happy”, at 99.75%, closely followed by “neutral”, at 99.15%. Similarly, on the RAF-DB dataset, “happy” also has the highest accuracy, at 99.37%. The lowest accuracy on the FER2013 dataset is for “disgust”, at 94.50%. For the RAF-DB dataset, the lowest accuracy is significantly lower for “fear”, at 88.10%. The FER2013 dataset shows consistently higher accuracies across most emotions when compared to the RAF-DB dataset. The overall average accuracy for the FER2013 dataset is 98.10%, which is considerably higher than the 93.34% for the RAF-DB dataset. This indicates that the model performs better on the FER2013 dataset.
The emotions “fear” and “disgust” have notably lower accuracies on the RAF-DB dataset compared to other emotions, suggesting that these emotions may be more challenging to recognize accurately with the current model or may be less consistently labeled in the RAF-DB dataset. “Happy” is the best-recognized emotion in both datasets, indicating that features defining happiness are well-captured by the model.
Furthermore, Figure 5 and Figure 6 deliver an in-depth analysis of our model’s capabilities using the FER2013 and RAF-DB datasets. They illustrate crucial performance metrics, like precision, recall, and F1-score, underscoring the model’s ability to precisely discern various emotional expressions. Evaluating these metrics together provides a clear picture of the model’s strengths and areas for enhancement. This holistic analysis is essential, as it presents a full perspective on the model’s consistency and adaptability to diverse and intricate datasets. The insights derived from these figures demonstrate the model’s proficiency in navigating the complexities of FER, setting benchmarks for future research and improvements in this area.
The confusion matrices for facial expression recognition on the FER2013 and RAF-DB datasets are presented in Table 2 and Table 3, respectively. The tables highlight the effectiveness of our proposed method in accurately detecting various facial expressions. Our model achieved impressive accuracy, reaching over 94% on the FER2013 dataset and 88% on the RAF-DB dataset across all emotional categories. However, the confusion matrix reveals some challenges, with certain emotional categories exhibiting lower classification performance. Despite these challenges, the model’s overall accuracy remains strong, and further improvements can be made by addressing the specific misclassifications identified in the confusion matrix. This strong performance underscores the effectiveness of our modified neural network architecture, which incorporates optimizations like depthwise-separable convolutions. Coupled with thoughtful data augmentation and precise parameter tuning, these modifications play a pivotal role in achieving superior accuracy in facial expression recognition.
Figure 7 demonstrates random experiment results of the proposed FER model, showcasing an individual expressing four different emotions: neutral, surprise, happy, and angry. The model accurately identifies the neutral expression with 99.86% accuracy, recognizes surprise with 98.35% accuracy, perfectly identifies happiness with 100% accuracy, and detects anger with 98.58% accuracy. This visualization highlights the model’s robust ability to accurately discern and classify distinct facial expressions.

4. Discussion

In addition to discussing the proposed model’s performance, we present a comprehensive comparison of our model against existing models in the field of FER. This comparison is systematically organized in Table 4, which details the average accuracy performances on the FER2013 and RAF-DB datasets.
The comparative analysis of FER models, as shown in the table, highlights varying performance across the two benchmark datasets: FER2013 and RAF-DB. The earlier models, such as the RS-exception by Lia et al., set a baseline with accuracies of 69.02% on FER2013 and 82.98% on RAF-DB. In contrast, our model significantly outperforms these with accuracies of 98.10% on FER2013 and 93.34% on RAF-DB, demonstrating a stronger ability to capture a wider range of emotional expressions.
The Custom Lightweight CNN-based Model (CLCM) by Gursesli et al. [41] showed a modest improvement on RAF-DB, reaching 84.0% accuracy. However, it still lagged far behind our model, particularly on FER2013. This gap suggests that while lightweight architectures offer computational efficiency, they may fall short in capturing the full complexity of facial emotions as effectively as our approach. More advanced models, like EfficientNet_FER by Kalsum et al. [42] and the Deep CNN by Alsharekh M.F. [46], showed notable improvements—especially Alsharekh’s model, which achieved 89.20% accuracy on FER2013. These results come closer to our model’s performance, particularly in terms of generalization on FER2013. However, our model’s integration of customized feature extraction and architectural enhancements clearly provides a competitive edge.
Similarly, the attention-based models by Chen et al. [43] and Peng et al. [44] achieved strong results on RAF-DB, with 91.07% and 92.37% accuracy, respectively. These models excel at focusing on spatial details within facial regions, showcasing the power of attention mechanisms in complex recognition tasks. Nevertheless, even these high-performing models fall short of the benchmarks set by our proposed method.
Moreover, to assess the statistical significance of the performance differences between the FER2013 and RAF-DB datasets, we conducted a series of statistical analyses, as presented Table 5. The alpha level used for the statistical tests in our analysis was set to 0.05, which is the standard threshold for determining statistical significance. First, we calculated the standard deviations for the accuracy, precision, recall, and F1-score metrics for both datasets. The standard deviation of accuracy for the FER2013 dataset was 1.622%, while for the RAF-DB dataset, it was 3.632%. This indicates that the model’s performance on FER2013 was more stable and consistent compared to RAF-DB, which exhibited greater variability.
We performed the Mann–Whitney U test to assess whether the observed differences in accuracy between the FER2013 and RAF-DB datasets were statistically significant. The resulting U statistic was 41.000, with a p-value of 0.037. Since the p-value was less than 0.05, we concluded that FER2013 achieves higher accuracy compared to RAF-DB.
Similar findings were observed for precision, where the standard deviation for FER2013 was 1.558%, and for RAF-DB, it was 3.726%. The statistical analysis returned a U statistic of 41.000 and a p-value of 0.037, confirming that FER2013 outperforms RAF-DB in terms of precision.
For recall, the standard deviation for FER2013 was 1.687%, and for RAF-DB, it was 3.586%. The analysis yielded a U statistic of 42.000 and a p-value of 0.029, further demonstrating that FER2013 shows higher consistency and performance in recall.
Finally, for the F1-score, the standard deviation for FER2013 was 1.626%, and for RAF-DB, it was 3.642%. The results showed a U statistic of 41.000 and a p-value of 0.037, indicating that FER2013 also outperforms RAF-DB in terms of the F1-score.
While the Mann–Whitney U tests indicated that the model performed significantly better on the FER2013 dataset compared to the RAF-DB dataset, several limitations should be considered. The RAF-DB dataset’s greater diversity and complexity may contribute to higher variability and less stable performance. Additionally, class imbalance in RAF-DB could explain the increased standard deviation. The model’s better generalization on FER2013 might be influenced by dataset-specific factors, such as differences in facial expressions or lighting conditions, which may limit its performance on RAF-DB. The higher standard deviation in RAF-DB further suggests instability, possibly due to dataset distribution or inherent noise.
Furthermore, the evaluation metrics used may not fully capture the model’s robustness to environmental variations. A more comprehensive analysis of the model’s stability under different conditions could provide deeper insights. Lastly, differences in image resolution, category sizes, and preprocessing methods could also contribute to the observed performance disparity.
The results emphasize the effectiveness and robustness of our model across diverse datasets. By combining efficiency with a deeper understanding of emotional representation, our approach offers a promising direction for real-world FER.
To provide a deeper understanding of the contributions and practical implications of this study, we now revisit and discuss the main research questions posed at the beginning of this manuscript. This reflection allows us to evaluate how well the proposed methods address the intended objectives and to highlight the significance of the experimental findings in the context of FER.
-
Enhancing both efficiency and accuracy in FER represents a long-standing challenge in affective computing, particularly when deploying models in real-time, low-resource environments such as embedded systems, mobile devices, or surveillance systems. Our study proposes a lightweight yet expressive deep learning framework by modifying the classical AlexNet architecture. The key to achieving this efficiency–accuracy balance lies in network compression without performance degradation. We adopt depthwise-separable convolutions, which decompose standard convolutions into spatial and channel-wise operations. This reduces the number of parameters and floating-point operations significantly—by approximately 80%—compared to standard convolution layers. Despite this compression, we observed only marginal degradation in performance, and in some cases, even improved accuracy, suggesting better generalization due to reduced overfitting. The result is a model that maintains discriminative power for emotion classification while being computationally tractable for real-time inference.
-
The integration of handcrafted features—particularly Gabor filters and LBP—with deep learning architectures is a strategy rooted in hybrid feature representation, which seeks to combine domain knowledge with data-driven learning. This integration aims to augment the model’s robustness, especially when training data is limited or heterogeneous. Gabor filters model the response of the human visual cortex to spatial frequency and orientation, making them biologically inspired and well-suited for detecting local facial structure changes (e.g., eye wrinkles or smile contours). They are particularly effective at encoding directional and frequency-sensitive information—key cues in emotional expression. LBP, on the other hand, excels in capturing fine-grained texture variations through pixel intensity comparisons, which is critical for representing subtle expressions, such as disgust or fear. Our results showed that fusing these handcrafted features with deep feature maps significantly improved classification accuracy across both the FER2013 and RAF-DB datasets. This is particularly important in resource-constrained settings, where the depth of the network is limited and may struggle to learn complex low-level features from scratch. The hybrid approach effectively compensates for this limitation by injecting priors into the learning process, improving feature richness and reducing the model’s dependency on data quantity.
-
The integration of depthwise-separable convolution with pointwise convolution helps tackle one of the major challenges in deploying deep learning models by reducing computational cost. Traditional CNNs apply full 3D kernels across all input channels, which scales poorly as the number of channels increases. In contrast, depthwise convolution applies a single filter per input channel, followed by pointwise convolution (1×1) to combine the resulting features across channels. This drastically reduces both the parameter count and floating-point operations while maintaining expressive power. Our empirical findings show that this architectural refinement preserves or even enhances accuracy, particularly when combined with domain-specific enhancements like Gabor and LBP. The reduction in model complexity improves generalization by reducing overfitting, especially when the training data is noisy or unbalanced—common challenges in FER datasets. From a deployment perspective, this architectural decision makes our model viable for edge computing, where computational and memory budgets are minimal. Furthermore, the reduced model size facilitates on-device learning and privacy-preserving computation, which are emerging needs in privacy-sensitive applications.
The findings from our study underscore the value of combining lightweight network architectures with handcrafted features for robust and efficient FER. This hybrid strategy enhances feature richness, generalization ability, and computational feasibility, making the proposed model a strong candidate for real-world FER applications.

5. Conclusions

In conclusion, this study introduced a novel method for FER that integrates Gabor and LBP features for effective texture extraction, combined with a modified AlexNet architecture for classification. The method was optimized to ensure efficient operation on hardware with limited computational resources, making it suitable for real-world applications with such constraints. A modified AlexNet model was successfully developed, validated, and shown to deliver outstanding performance in classification tasks [47,48], demonstrating its potential for practical deployment. Extensive experiments conducted on the FER2013 and RAF-DB benchmark datasets demonstrated the effectiveness of the proposed approach. The model achieved impressive average accuracies of 98.10% and 93.34% for the two datasets, with corresponding standard deviations of 1.63% and 3.62%, respectively, in recognizing a wide range of facial emotions. On the FER-2013 dataset, the model excelled, achieving a precision of 98.2%, a recall of 97.9%, and an F1-score of 98.0%. In comparison, for the RAF-DB dataset, the model obtained a precision of 93.54%, a recall of 93.12%, and an F1-score of 93.34%. These findings highlight the model’s strong capability in recognizing facial emotions, with notably superior performance on the FER-2013 dataset compared to RAF-DB.
However, despite the promising results, this study has certain limitations. The current model primarily focuses on facial expressions, overlooking critical audio and textual cues that are essential for a more comprehensive emotion recognition system, particularly in scenarios where detecting emotions in videos may be challenging. Additionally, the model’s performance may degrade when applied to datasets with significant variations in lighting, occlusions, or pose, highlighting the need for further robustness improvements. Future work should explore the integration of multimodal data [9,49,50] and enhance the model’s adaptability to diverse real-world conditions.

Author Contributions

Conceptualization, F.S. and A.K.; data curation, A.K. and U.K.; formal analysis, F.S., A.K. and A.A.; funding acquisition, Y.-I.C.; investigation, U.K. and A.A.; methodology, F.S.; project administration, Y.-I.C.; resources, U.K. and A.A.; software, F.S. and A.K.; supervision, Y.-I.C.; validation, A.K.; writing—original draft, F.S. and A.K.; writing—review and editing, A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by Korean Agency for Technology and Standard under Ministry of Trade, Industry and Energy in 2024, project numbers are 20022362 (2410003714, Establishment of standardization basis for BCI and AI Interoperability).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study analyzed datasets that are publicly accessible at https://www.kaggle.com/datasets/msambare/fer2013 (accessed on 17 May 2025) and http://www.whdeng.cn/RAF/model1.html (accessed on 17 May 2025).

Acknowledgments

The authors, F.S., A.K., U.K. and A.A., acknowledge, with gratitude, the invaluable support, insightful feedback, and guidance provided by their supervisor, Young-Im Cho (Gachon University), throughout the preparation of this manuscript. The authors also extend their appreciation to the editor and anonymous reviewers for their constructive comments, which have significantly enhanced the academic quality and presentation of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, S.; Deng, W. Deep Facial Expression Recognition: A Survey. IEEE Trans. Affect. Comput. 2022, 13, 1195–1215. [Google Scholar] [CrossRef]
  2. Kołakowska, A.; Landowska, A.; Szwoch, M.; Szwoch, W.; Wróbel, M.R. Emotion Recognition and Its Applications. In Human-Computer Systems Interaction: Backgrounds and Applications 3; Hippe, Z., Kulikowski, J., Mroczek, T., Wtorek, J., Eds.; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2014; Volume 300. [Google Scholar] [CrossRef]
  3. Yi, M.-H.; Kwak, K.-C.; Shin, J.-H. KoHMT: A Multimodal Emotion Recognition Model Integrating KoELECTRA, HuBERT with Multimodal Transformer. Electronics 2024, 13, 4674. [Google Scholar] [CrossRef]
  4. Abdusalomov, A.; Kutlimuratov, A.; Nasimov, R.; Whangbo, T.K. Improved speech emotion recognition focusing on high-level data representations and swift feature extraction calculation. Comput. Mater. Contin. 2023, 77, 2915–2933. [Google Scholar] [CrossRef]
  5. Cavus, N.; Esmaili, P.; Sekeroglu, B.; Aşır, S.; Azar, N.A.N. Detecting emotions through EEG signals based on modified convolutional fuzzy neural network. Sci. Rep. 2024, 14, 10371. [Google Scholar] [CrossRef]
  6. Wang, L.; Hao, J.; Zhou, T.H. ECG Multi-Emotion Recognition Based on Heart Rate Variability Signal Features Mining. Sensors 2023, 23, 8636. [Google Scholar] [CrossRef]
  7. Shen, Z.; Cheng, J.; Hu, X.; Dong, Q. Emotion Recognition Based on Multi-View Body Gestures. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 3317–3321. [Google Scholar] [CrossRef]
  8. Ciraolo, D.; Fazio, M.; Calabrò, R.S.; Villari, M.; Celesti, A. Facial expression recognition based on emotional artificial intelligence for tele-rehabilitation. Biomed. Signal Process. Control. 2024, 92, 106096, ISSN 1746-8094. [Google Scholar] [CrossRef]
  9. Kalyta, O.; Barmak, O.; Radiuk, P.; Krak, I. Facial Emotion Recognition for Photo and Video Surveillance Based on Machine Learning and Visual Analytics. Appl. Sci. 2023, 13, 9890. [Google Scholar] [CrossRef]
  10. Chowdary, M.K.; Nguyen, T.N.; Hemanth, D.J. Deep learning-based facial emotion recognition for human–computer interaction applications. Neural Comput. Appl. 2023, 35, 23311–23328. [Google Scholar] [CrossRef]
  11. Akhand, M.A.H.; Roy, S.; Siddique, N.; Kamal, M.A.S.; Shimamura, T. Facial Emotion Recognition Using Transfer Learning in the Deep CNN. Electronics 2021, 10, 1036. [Google Scholar] [CrossRef]
  12. Mamieva, D.; Abdusalomov, A.B.; Kutlimuratov, A.; Muminov, B.; Whangbo, T.K. Multimodal Emotion Detection via Attention-Based Fusion of Extracted Facial and Speech Features. Sensors 2023, 23, 5475. [Google Scholar] [CrossRef]
  13. Savchenko, A.V.; Savchenko, L.V.; Makarov, I. Classifying emotions and engagement in online learning based on a single facial expression recognition neural network. IEEE Trans. Afect. Comput. 2022, 13, 2132–2143. [Google Scholar] [CrossRef]
  14. Chen, D.; Wen, G.; Li, H.; Chen, R.; Li, C. Multi-Relations Aware Network for In-the-Wild Facial Expression Recognition. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3848–3859. [Google Scholar] [CrossRef]
  15. Makhmudov, F.; Kultimuratov, A.; Cho, Y.-I. Enhancing Multimodal Emotion Recognition through Attention Mechanisms in BERT and CNN Architectures. Appl. Sci. 2024, 14, 4199. [Google Scholar] [CrossRef]
  16. Gao, G.; Chen, C.; Liu, K.; Mashhadi, A.; Xu, K. Automatic face detection based on bidirectional recurrent neural network optimized by improved Ebola optimization search algorithm. Sci. Rep. 2024, 14, 27798. [Google Scholar] [CrossRef]
  17. Shi, Y.; Zhang, H.; Guo, W.; Zhou, M.; Li, S.; Li, J.; Ding, Y. LighterFace Model for Community Face Detection and Recognition. Information 2024, 15, 215. [Google Scholar] [CrossRef]
  18. Charoqdouz, E.; Hassanpour, E. Feature Extraction from Several Angular Faces Using a Deep Learning Based Fusion Technique for Face Recognition. Int. J. Eng. 2023, 36, 1548–1555. [Google Scholar] [CrossRef]
  19. Yang, Y. Facial Features Extraction And Clustering With Machine Learning. In Proceedings of the EITCE’ 22: Proceedings of the 2022 6th International Conference on Electronic Information Technology and Computer Engineering; Xiamen, China, 21–23 October 2022, pp. 1806–1810. [CrossRef]
  20. Anand, M.; Babu, S. Multi-class Facial Emotion Expression Identification Using DL-Based Feature Extraction with Classification Models. Int. J. Comput. Intell. Syst. 2024, 17, 25. [Google Scholar] [CrossRef]
  21. Rakhimovich, A.M.; Kadirbergenovich, K.K.; Ishkobilovich, Z.M.; Kadirbergenovich, K.J. Logistic Regression with Multi-Connected Weights. J. Comput. Sci. 2024, 20, 1051–1058. [Google Scholar] [CrossRef]
  22. Nautiyal, A.; Bhardwaj, D.K.; Narula, R.; Singh, H. Real Time Emotion Recognition Using Image Classification. In Proceedings of the 2023 Fifteenth International Conference on Contemporary Computing (IC3-2023), Noida, India, 3–5 August 2023; pp. 8–12. [Google Scholar] [CrossRef]
  23. Lolaev, M.; Madrakhimov, S.; Makharov, K.; Saidov, D. Choosing allowability boundaries for describing objects in subject areas. IAES Int. J. Artif. Intell. 2024, 13, 329–336. [Google Scholar] [CrossRef]
  24. Vaijayanthi, S.; Arunnehru, J. Deep neural network-based emotion recognition using facial landmark features and particle swarm optimization. Automatika 2024, 65, 1088–1099. [Google Scholar] [CrossRef]
  25. Palestra, G.; Pettinicchio, A.; Del Coco, M.; Carcagnì, P.; Leo, M.; Distante, C. Improved Performance in Facial Expression Recognition Using 32 Geometric Features. In Image Analysis and Processing—ICIAP 2015; Murino, V., Puppo, E., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland; Volume 9280. [CrossRef]
  26. Chen, L.; Zhou, M.; Su, W.; Wu, M.; She, J.; Hirota, K. Softmax regression based deep sparse autoencoder network for facial emotion recognition in human-robot interaction. Inf. Sci. 2018, 428, 49–61. [Google Scholar] [CrossRef]
  27. Fu, S.; Li, Z.; Liu, Z.; Yang, X. Interactive knowledge distillation for image classification. Neurocomputing 2021, 449, 411–421. [Google Scholar] [CrossRef]
  28. Ma, T.; Tian, W.; Xie, Y. Multi-level knowledge distillation for low-resolution object detection and facial expression recognition. Knowl.-Based Syst. 2022, 240, 108136, ISSN 0950-7051. [Google Scholar] [CrossRef]
  29. Goodfellow, I.J.; Erhan, D.; Carrier, P.L.; Courville, A.; Mirza, M.; Hamner, B.; Cukierski, W.; Tang, Y.; Thaler, D.; Lee, D.H.; et al. Challenges in Representation Learning: A Report on Three Machine Learning Contests. In Proceedings of the Neural Information, Processing; Lee, M., Hirose, A., Hou, Z.G., Kil, R.M., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 117–124. [Google Scholar]
  30. Li, S.; Deng, W.; Du, J.P. Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  31. Munawar, H.S.; Aggarwal, R.; Qadir, Z.; Khan, S.I.; Kouzani, A.Z.; Mahmud, M.A.P. A Gabor Filter-Based Protocol for Automated Image-Based Building Detection. Buildings 2021, 11, 302. [Google Scholar] [CrossRef]
  32. Mehrotra, R.; Namuduri, K.R.; Ranganathan, N. Gabor filter-based edge detection. Pattern Recognit. 1992, 25, 1479–1494, ISSN 0031-3203. [Google Scholar] [CrossRef]
  33. Song, Z. Facial Expression Emotion Recognition Model Integrating Philosophy and Machine Learning Theory. Front. Psychol. 2021, 12, 759485. [Google Scholar] [CrossRef]
  34. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolutiongray-scale and rotation invariant texture classification withlocal binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  35. Niu, B.; Gao, Z.; Guo, B. Bingbing, Facial Expression Recognition with LBP and ORB Features. Comput. Intell. Neurosci. 2021, 2021, 8828245. [Google Scholar] [CrossRef]
  36. Garcia Freitas, P.; Da Eira, L.P.; Santos, S.S.; Farias, M.C.Q.D. On the Application LBPTexture Descriptors and Its Variants for No-Reference Image Quality Assessment. J. Imaging 2018, 4, 114. [Google Scholar] [CrossRef]
  37. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  38. Kuldashboy, A.; Umirzakova, S.; Allaberdiev, S.; Nasimov, R.; Abdusalomov, A.; Im Cho, Y. Efficient image classification through collaborative knowledge distillation: A novel AlexNet modification approach. Heliyon 2024, 10, e34376. [Google Scholar] [CrossRef] [PubMed]
  39. Powers, D.M. Evaluation: From precision, recall and Fmeasure to ROC, informedness, markedness and correlation. J. Mach. Learn. Res. 2011, 2, 37–63. [Google Scholar]
  40. Liao, L.; Wu, S.; Song, C.; Fu, J. RS-Xception: A Lightweight Network for Facial Expression Recognition. Electronics 2024, 13, 3217. [Google Scholar] [CrossRef]
  41. Gursesli, M.C.; Lombardi, S.; Duradoni, M.; Bocchi, L.; Guazzini, A.; Lanata, A. Facial Emotion Recognition (FER) Through Custom Lightweight CNN Model: Performance Evaluation in Public Datasets. IEEE Access 2024, 12, 45543–45559. [Google Scholar] [CrossRef]
  42. Kalsum, T.; Mehmood, Z. A Novel Lightweight Deep Convolutional Neural Network Model for Human Emotions Recognition in Diverse Environments. J. Sens. 2023, 2023, 6987708. [Google Scholar] [CrossRef]
  43. Chen, X.; Huang, L. A Lightweight Model Enhancing Facial Expression Recognition with Spatial Bias and Cosine-Harmony Loss. Computation 2024, 12, 201. [Google Scholar] [CrossRef]
  44. Peng, C.; Li, B.; Zou, K.; Zhang, B.; Dai, G.; Tsoi, A.C. An Innovative Neighbor Attention Mechanism Based on Coordinates for the Recognition of Facial Expressions. Sensors 2024, 24, 7404. [Google Scholar] [CrossRef]
  45. Soman, G.; Vivek, M.V.; Judy, M.V.; Papageorgiou, E.; Gerogiannis, V.C. Precision-Based Weighted Blending Distributed Ensemble Model for Emotion Classification. Algorithms 2022, 15, 55. [Google Scholar] [CrossRef]
  46. Alsharekh, M.F. Facial Emotion Recognition in Verbal Communication Based on Deep Learning. Sensors 2022, 22, 6105. [Google Scholar] [CrossRef]
  47. Zhao, Y.; Chen, D. Expression Recognition Using Improved AlexNet Network in Robot Intelligent Interactive System. J. Robot. 2022, 2022, 4969883. [Google Scholar] [CrossRef]
  48. Zhang, Y. DCAlexNet: Deep coupled AlexNet for micro facial expression recognition based on double face images. Comput. Biol. Med. 2025, 189, 109986, ISSN 0010-4825. [Google Scholar] [CrossRef] [PubMed]
  49. Salas-Cáceres, J.; Lorenzo-Navarro, J.; Freire-Obregón, D.; Castrillón-Santana, M. Multimodal emotion recognition based on a fusion of audiovisual information with temporal dynamics. Multimed. Tools Appl. 2024. [Google Scholar] [CrossRef]
  50. Makhmudov, F.; Kilichev, D.; Giyosov, U.; Akhmedov, F. Online Machine Learning for Intrusion Detection in Electric Vehicle Charging Systems. Mathematics 2025, 13, 712. [Google Scholar] [CrossRef]
Figure 1. Proposed method workflow.
Figure 1. Proposed method workflow.
Sensors 25 03832 g001
Figure 2. Modified AlexNet architecture.
Figure 2. Modified AlexNet architecture.
Sensors 25 03832 g002
Figure 5. Performance metrics by emotion on the FER2013 dataset.
Figure 5. Performance metrics by emotion on the FER2013 dataset.
Sensors 25 03832 g005
Figure 6. Performance metrics by emotion on the RAF-DB dataset.
Figure 6. Performance metrics by emotion on the RAF-DB dataset.
Sensors 25 03832 g006
Figure 7. Random experiment results of the proposed model.
Figure 7. Random experiment results of the proposed model.
Sensors 25 03832 g007
Table 1. Model accuracy across emotions on the FER2013 and RAF-DB datasets.
Table 1. Model accuracy across emotions on the FER2013 and RAF-DB datasets.
EmotionsAccuracy (%)Average Accuracy and Standard Deviation (%)
FER2013RAF-DBFER2013RAF-DB
Happy99.7599.3798.10 and 1.6393.34 and 3.62
Neutral99.1597.43
Sad99.1094.15
Fear98.4788.10
Angry98.2091.25
Surprise97.4092.20
Disgust94.5091.0
Table 2. Confusion matrix on the FER2013 dataset.
Table 2. Confusion matrix on the FER2013 dataset.
HappyNeutralSadFearAngrySurpriseDisgust
Happy1.000.000.000.000.000.000.00
Neutral0.010.990.000.000.000.000.00
Sad0.000.000.990.010.000.000.00
Fear0.000.000.000.980.020.000.00
Angry0.000.000.020.000.980.000.00
Surprise0.020.000.000.000.010.970.00
Disgust0.000.000.000.040.020.000.94
Table 3. Confusion matrix on the RAF-DB dataset.
Table 3. Confusion matrix on the RAF-DB dataset.
AngryDisgustFearHappyNeutralSadSurprise
Angry0.910.000.030.000.000.060.00
Disgust0.010.910.050.000.000.030.00
Fear0.030.000.880.000.000.090.00
Happy0.000.000.000.990.000.000.01
Neutral0.000.000.000.020.970.010.00
Sad0.010.000.050.000.000.940.00
Surprise0.000.000.000.060.000.020.92
Table 4. Comparative performance of the proposed model versus existing models on the FER-2013 and RAF-DB datasets.
Table 4. Comparative performance of the proposed model versus existing models on the FER-2013 and RAF-DB datasets.
ReferencesModelsAverage Accuracy (%)
FER2013RAF-DB
Lia et.al. [40]RS-exception 69.0282.98
Gursesli et.al. [41]Custom Lightweight CNN-based Model (CLCM)63.084.0
Kalsum et.al. [42]EfficientNet_FER69.8784.10
Chen et al. [43]Lightweight Facial Network with Spatial Bias-91.07
Peng et al. [44]Coordinate-based Neighborhood Attention Mechanism-92.37
Soman et al. [45]Weighted Ensemble Model76.20-
Alsharekh M.F. [46]Deep CNN89.20-
Our modelModified Alexnet with LBP and Gabor Features98.1093.34
Table 5. Statistical significance of performance differences between the FER2013 and RAF-DB datasets.
Table 5. Statistical significance of performance differences between the FER2013 and RAF-DB datasets.
MetricFER2013 Std Dev (%)RAF-DB Std Dev (%)U Statisticp-ValueSignificant (p < 0.05)
Accuracy1.6223.63241.00.037Yes
Precision1.5583.72641.00.037Yes
Recall1.6873.58642.00.029Yes
F1-Score1.6263.64241.00.037Yes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Safarov, F.; Kutlimuratov, A.; Khojamuratova, U.; Abdusalomov, A.; Cho, Y.-I. Enhanced AlexNet with Gabor and Local Binary Pattern Features for Improved Facial Emotion Recognition. Sensors 2025, 25, 3832. https://doi.org/10.3390/s25123832

AMA Style

Safarov F, Kutlimuratov A, Khojamuratova U, Abdusalomov A, Cho Y-I. Enhanced AlexNet with Gabor and Local Binary Pattern Features for Improved Facial Emotion Recognition. Sensors. 2025; 25(12):3832. https://doi.org/10.3390/s25123832

Chicago/Turabian Style

Safarov, Furkat, Alpamis Kutlimuratov, Ugiloy Khojamuratova, Akmalbek Abdusalomov, and Young-Im Cho. 2025. "Enhanced AlexNet with Gabor and Local Binary Pattern Features for Improved Facial Emotion Recognition" Sensors 25, no. 12: 3832. https://doi.org/10.3390/s25123832

APA Style

Safarov, F., Kutlimuratov, A., Khojamuratova, U., Abdusalomov, A., & Cho, Y.-I. (2025). Enhanced AlexNet with Gabor and Local Binary Pattern Features for Improved Facial Emotion Recognition. Sensors, 25(12), 3832. https://doi.org/10.3390/s25123832

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop