You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

2 December 2025

An Explainable Framework for Mental Health Monitoring Using Lightweight and Privacy-Preserving Federated Facial Emotion Recognition

,
,
and
1
Department of Electrical Engineering, University of Sharjah, Sharjah P.O. Box 27272, United Arab Emirates
2
Department of Engineering and Information Technology, University of Dubai, Dubai P.O. Box 14143, United Arab Emirates
3
School of Built Environment, Engineering and Computing, Leeds Beckett University, Leeds LS1 3HE, UK
4
College of Computing and Informatics, University of Sharjah, Sharjah P.O. Box 27272, United Arab Emirates
Sensors2025, 25(23), 7320;https://doi.org/10.3390/s25237320 
(registering DOI)
This article belongs to the Section Sensing and Imaging

Abstract

The continuous analysis of emotional cues through facial emotion recognition (FER) systems can support mental health evaluation and psychological well-being monitoring systems. Most FER systems face privacy and trust concerns due to their centralized data approaches and lack of transparency, making potential deployment difficult. To address these concerns, a federated, explainability-driven FER framework designed to provide trustworthy and privacy-preserving emotion recognition with potential applications in mental health monitoring is proposed in this paper. The proposed lightweight Convolutional Neural Network (CNN) enables real-time inference while preserving high accuracy. Comprehensive evaluations on RAF-DB, ExpW, and FER2013 datasets, show that the proposed model demonstrates improved cross-dataset generalization compared to related works, achieving average accuracies of 75.5% and 74.3% in centralized and federated settings, respectively. Quantitative perturbation-based metrics, including Insertion and Deletion Area Under Curve (IAUC and DAUC), Average Drop (AD), Increase in Confidence (IC), Average Drop in Accuracy (ADA), and Active Pixel Ratio, were employed to objectively evaluate the quality and reliability of the model Grad-CAM++ explanations. The results confirm that model explainability enhances transparency and is directly associated with improved model performance.

1. Introduction

Mental health plays a critical role in overall human well-being. However, millions of individuals worldwide remain undiagnosed or untreated due to limited access to care, social stigma, and shortages of trained professionals. The World Health Organization (WHO) [1] estimates that nearly one billion people suffer from mental illnesses like anxiety and depression, which continue to impose increasing social and economic burdens. In order to tackle these issues, the WHO Comprehensive Mental Health Action Plan 2013–2030 [2] emphasizes technology-enabled support, community-based intervention, and early detection. Early detection and ongoing monitoring of emotional states could support prompt intervention and individualized support. Advances in artificial intelligence (AI) have enabled emotion recognition from multiple modalities such as facial expressions, speech, and physiological signals. Among these, facial emotion recognition (FER) has become a popular, non-invasive, and informative technique for determining emotional states, which makes it suitable for assistive technology with potential applications in mental health monitoring. Psychological studies have shown that emotional balance strongly correlates with psychological resilience and overall well-being. Fredrickson and Losada’s work on the positivity ratio underscores that individuals who maintain a higher frequency of positive emotions tend to experience improved mental health outcomes [3,4]. These findings highlight the importance of reliable and continuous emotion assessment tools in supporting mental health care.
The deployment of facial emotion recognition (FER) in mental health applications remains constrained by two major challenges: privacy and interpretability. State-of-the-art FER systems rely on centralized data storage, which raises substantial privacy concerns when managing sensitive personal and clinical information. Additionally, deep learning models often function as “black boxes,” providing no transparency into their decision-making processes or the reliability of their predictions. In healthcare contexts, where accountability, explainability, and trust are essential, the lack of interpretability poses a major barrier to adoption [5].
Federated learning (FL) offers an effective solution to the privacy issue by enabling decentralized training across multiple devices without transferring raw data to a central server [6]. While FL enhances data confidentiality, it inherits the same interpretability challenges as conventional deep models [7]. Explainable artificial intelligence (XAI) methods can bridge this gap by revealing the underlying reasoning behind model predictions and highlighting salient features contributing to emotion recognition. This paper introduces a lightweight, federated learning framework designed for potential edge deployment and enhanced with quantitative explainability assessment for trustworthy and privacy-preserving facial emotion recognition. The approach employs multiple perturbation-based explainability metrics to objectively evaluate explanation quality and model reliability, such as Insertion and Deletion Area Under Curve (IAUC and DAUC), Average Drop (AD), Increase in Confidence (IC), Average Drop in Accuracy (ADA), and Active Pixel Ratio to analyze the relationship between highlighted facial regions and model performance. The lightweight design ensures computational efficiency and suitability for real-time emotion monitoring applications on edge devices, such as assistive or mental health support systems. Embedding interpretability into the optimization and evaluation pipeline promotes transparent and trustworthy decision-making, aligning the framework with ethical standards for mental health–related applications. The main contributions of this work can be summarized as follows:
  • Lightweight and Edge-Deployable FER Architecture: A compact CNN architecture optimized for efficient training and low-latency inference, achieving a strong balance between accuracy and computational cost. The model demonstrates superior cross-dataset generalization across Facial Emotion Recognition 2013 (FER2013), Real-world Affective Faces Database (RAF-DB), and Expression in the Wild (ExpW), reaching 75.5% and 74.3% average accuracy under centralized and federated configurations, respectively. With only 1.45 M parameters and 0.107 GFLOPs, it demonstrates stronger cross-dataset generalization than previously reported cross-dataset FER architectures across all three datasets, while remaining suitable for real-time emotion monitoring on resource-constrained edge devices.
  • Federated Learning with Integrated Explainability for Trustworthy Model Selection: A privacy-preserving federated learning framework that explicitly incorporates explainability into the evaluation and model optimization process. The framework combines federated training with multi-level interpretability analysis, enabling systematic comparison of model behavior across different configurations. By leveraging explainability to guide model selection, the approach enhances transparency, interpretability, and accountability, supporting its suitability for deployment in privacy-sensitive mental health monitoring systems.
  • Quantitative Explainability Assessment: A systematic evaluation of Gradient-weighted Class Activation Mapping++ (Grad-CAM++) explanations using perturbation metrics, including Insertion Area Under Curve (IAUC), Deletion Area Under Curve (DAUC), Average Drop (AD), Increase in Confidence (IC), Average Drop in Accuracy (ADA), and Active Pixel Ratio, to provide objective, reproducible, and comparable measures of explanation quality and model trustworthiness.
The remainder of this paper is organized as follows: Section 2 reviews related works. Section 3 details the proposed framework, including architecture design and metric formulation. Section 4 presents experimental results and explainability analysis. Section 5 presents cross-dataset evaluation and comparative analysis to assess generalization performance. Section 6 examines edge deployment and runtime performance. Section 7 discusses ethical and practical considerations for deploying FER systems in mental health contexts. Finally, Section 8 concludes the paper and outlines future research directions.

3. Proposed Methodology

3.1. FER for Mental Health

We formulate the task of emotion recognition for monitoring emotional states with potential support for mental health applications. Emotions such as anger, happiness, sadness, and neutrality are widely recognized as relevant indicators of mood patterns and potential psychological disorders. Accordingly, our study restricts the label space to these four clinically relevant emotions. This selection is supported by their established significance within the domain of mental health [48,49]. Happiness, or more specifically its absence, is a core feature of both depression and anxiety, reflecting impairments in responses to pleasurable experiences. Sadness is a hallmark of depressive disorders and is strongly associated with elevated suicide risk [50,51]. Additionally, various studies have consistently linked anger to depression, emphasizing its role in the onset and persistence of symptoms [52,53]. The inclusion of a neutral state provides an essential baseline for distinguishing affective responses. In particular, the contrast between happiness and neutral is crucial for identifying anhedonia, where emotional experiences are flattened toward neutrality. Our choice is further reinforced by prior FER for mental health studies that employed the same or closely aligned four emotion subset in related monitoring tasks [29,35]. This emphasis on capturing both positive and negative emotions aligns with the positivity ratio, which indicates that psychological well-being is promoted when the frequency of positive emotions is at least three times that of negative ones [3]. Validation on clinically annotated depression datasets would be ideal; however, such datasets are currently unavailable publicly due to privacy and ethical constraints. Our evaluation demonstrates the technical feasibility of the framework using established emotion categories with documented clinical relevance.

3.2. Federated Learning Collaborative Training

In this paper, we propose a privacy-preserving federated learning-based facial emotion recognition framework designed for assistive technology. Federated learning is incorporated to ensure data privacy and enhance the model’s generalization across diverse users and devices. The framework employs a lightweight deep learning model optimized for edge deployment, enabling real-time performance while maintaining privacy. FL is used to implement collaborative, iterative training across multiple devices. To emulate real-world scenarios where client data distributions are inherently heterogeneous with variations in ethnicity and demographic characteristics. Three popular FER datasets, FER2013, RAF-DB, and ExpW, were used for training and testing in a heterogeneous data environment. These datasets include naturally captured facial samples with diverse demographic and environmental characteristics. Figure 1 shows the high-level overview of the proposed system.
Figure 1. Proposed FL-based system for privacy-preserving FER.
The FL setup involves multiple clients and a central server. Each client trains a local copy of the model using its own data and sends only the model updates to the server. The server then aggregates these updates from all clients to produce an improved global model. This process repeats over several rounds until the model converges.

3.3. Datasets

To evaluate the proposed model, each of the three datasets utilized in this research contributes complementary heterogeneity in demographics, environment, and situational context:
  • Real-world Affective Faces Database (RAF-DB) [54]: contributes a total of 12,488 images split into 9983 training and 2505 test samples, capturing a wide demographic range with variations in age, ethnicity, gender, head poses, and illumination. The dataset follows its official split of approximately 80:20 (train:test), which we retained to ensure comparability with prior works.
  • Facial Emotion Recognition 2013 (FER2013) [55]: provides 16,482 images under challenging in-the-wild conditions such as occlusions and variable lighting, split into 13,269 training and 3213 test samples (80:20).
  • Expression in the Wild (ExpW) [56]: adds situational diversity with 79,650 images of both posed and spontaneous expressions, split into 55,739 training and 23,911 test samples (70:30).
Together, these datasets provide complementary demographic, environmental, and situational heterogeneity for training and evaluating robust FER models.

3.4. Explainability, Performance, and Trustworthiness-Driven CNN Architecture Optimization

To identify the best lightweight architecture for facial emotion recognition, a number of CNN models with different filter configurations were trained and evaluated within a federated learning framework. Several lightweight CNN architectures were examined with different filter block configurations. For the three-layer models, the configurations included (16, 32, 64), (32, 64, 128), (64, 128, 256), and (128, 128, 256), while the four-layer models used (16, 32, 64, 128), (32, 64, 128, 256), and (64, 128, 256, 512). The models under evaluation have parameter counts ranging from 164 K to 1.72 M, encompassing both ultra-lightweight and lightweight designs. This range allows for direct comparison with edge-optimized architectures like MobileNetV2 (3.5 M parameters, ≈14 MB) [57], SqueezeNet (1.24 M parameters), and ShuffleNet (1.0–2.3 M parameters) [58].
Using Grad-CAM++, we quantitatively evaluated the explanation quality of our lightweight architectures on the RAF-DB, ExpW, and FER2013 datasets. While most evaluations rely on standard performance metrics such as accuracy or on qualitative visual inspection of model interpretations, our approach utilized a quantitative assessment strategy for model explainability. The final configuration was selected based on both its strong performance and the reliability of its explanations.
The general architecture of the evaluated CNN models is illustrated in Figure 2. The network processes 48 × 48 grayscale facial images using a variable number of convolutional filter blocks, each consisting of a 3 × 3 convolutional layer followed by 2 × 2 max-pooling. The configurable filter depths (FB1–FB4) correspond to the different filter configurations evaluated across the three-layer and four-layer architectures. After feature extraction, the feature maps are flattened and passed through two fully connected layers with 256 and 128 units, followed by a dropout layer (rate = 0.2) for regularization. The final output layer employs a softmax activation function to classify the four target emotions: neutral, happy, sad, and angry.
Figure 2. Hierarchical CNN architecture for facial emotion recognition. The Feature Extraction Module is configurable via the filter sizes (FB1–FB4).
Training samples from the FER2013, RAF-DB, and ExpW datasets were merged and shuffled to emulate realistic heterogeneous data distributions. The combined dataset was then distributed among three clients in the federated training setup. Each client trained its local model for 15 epochs per round using a batch size of 64 and a learning rate of 0.001. At the end of each round, the clients transmitted their model weights to a central server, where aggregation was performed using the Federated Averaging (FedAvg) [59] algorithm. Model performance across all configurations was evaluated using accuracy, F1-score, recall, and precision. Training continued until the global model converged, when performance stabilized without any further notable improvement. This occurred after approximately 20 rounds.

Grad-CAM++ Explainability Evaluation

To assess the explainability and trustworthiness of the CNN architectures, Grad-CAM++ [60] was employed as the primary explanation method, followed by quantitative evaluation using perturbation-based metrics. Grad-CAM++ generates class-specific heatmaps that highlight the regions most influential to the model’s predictions by weighting feature maps with higher-order gradient information. The resulting activation maps were normalized to the range [ 0 ,   1 ] for quantitative analysis.
To compare interpretability across the candidate CNN architectures and to examine the contribution of each convolutional layer to the decision-making process, Grad-CAM++ overlays were generated for each convolutional block: Filter Block 1–Filter Block 3 (FB1–FB3) for the three-layer models and Filter Block 1–Filter Block 4 (FB1–FB4) for the four-layer model. The visualizations were produced using images correctly classified by all models from the FER2013, RAF-DB, and ExpW datasets.
For the final overlay, the Grad-CAM++ heatmap was combined with a channel-attention map to enhance localization consistency and suppress irrelevant activations. This integration, inspired by the Convolutional Block Attention Module (CBAM) [61], allows the network to simultaneously detect the important features (channel attention) and where they occur (spatial attention), thereby producing more informative and interpretable feature representations. All preprocessing steps and parameters were maintained consistently across all models.

3.5. Assessing Grad-CAM++ Explanations with Perturbation-Based Metrics

To systematically evaluate the faithfulness of the Grad-CAM++ explanations, we employed a set of perturbation-based quantitative metrics that measure how the model’s prediction changes under controlled modifications of the input image. These metrics provide an objective way to determine whether the highlighted regions in the explanation map truly influence the model’s decision-making. In particular, we employ three widely adopted families of perturbation-based metrics:
  • Insertion and Deletion AUC (IAUC/DAUC), which quantify how the model’s confidence evolves as the most salient regions are progressively revealed or removed.
  • Average Drop (AD) and Average Drop in Accuracy (ADA), which measure the reduction in confidence and accuracy when the important regions highlighted by the explanation are removed.
  • Increase in Confidence (IC), which counts cases where retaining only the salient regions leads to higher confidence than using the full image.
The full mathematical definitions and implementation details for all metrics are provided in Appendix A.

4. Results Analysis and Discussion

4.1. Optimization and Selection of CNN Architectures for Explainability Analysis

This subsection presents the optimization results of various CNN architectures, comparing them across the RAF-DB, ExpW, and FER2013 datasets. The main goal is to find the top-performing models that exhibit good generalization across these diverse datasets, as these will make solid candidates for further assessment of the quality of the explanation.
The performance of the three-layer and four-layer CNN architectures under various filter configurations is presented in Table 1, Table 2, Table 3 and Table 4, Table 5, Table 6, respectively. These tables report the accuracy, F1-score, recall, and precision for each configuration, with bold indicating the best-performing configuration for each metric.
Table 1. Performance comparison of CNN (3-layer) models on FER2013 dataset.
Table 2. Performance comparison of CNN (3-layer) models on RAF-DB dataset.
Table 3. Performance comparison of CNN (3-layer) models on ExpW dataset.
Table 4. Performance comparison of CNN (4-layer) models on FER2013 dataset.
Table 5. Performance comparison of CNN (4-layer) models on RAF-DB dataset.
Table 6. Performance comparison of CNN (4-layer) models on ExpW dataset.
Among the three-layer CNNs, two configurations showed improved performance. With an accuracy of 82.8% on RAF-DB, the (64, 128, 256) model outperformed the (128, 128, 256) model, which came in second with 80.7%. Both configurations attained similar accuracy values of 71.1% on FER2013, while (128, 128, 256) performed marginally better on ExpW. Furthermore, the (64, 128, 256) model demonstrated consistent improvements across all other metrics, including F1-score, recall, and precision. Considering the goal of identifying models that exhibit both high accuracy and strong generalization across diverse datasets, the (64, 128, 256) configuration was selected for further analysis of its explainability. Among the four-layer CNNs, the (64, 128, 256, 512) configuration achieved the best overall results, recording the highest values across all metrics on both RAF-DB and ExpW. It reached an accuracy of 80.6% on RAF-DB and 69.0% on ExpW, while maintaining comparable performance on FER2013 with 70.4%.
Incorporating the best model from each configuration allows for a more thorough examination of how interpretability is affected by increasing filter depth. Models that generalize well across datasets typically generate more targeted and semantically meaningful activation patterns, which further emphasizes the relationship between interpretability and accuracy. The confusion matrix for the selected model is shown in Figure 3. Overall, both architectures demonstrated comparable performance across datasets. However, the three-layer model showed enhanced discrimination for negative emotions such as sad and anger, while the four-layer model performed slightly better on neutral and happy expressions.
Figure 3. Normalized confusion matrices for Model 1 and Model 2 across FER2013, RAF-DB, and ExpW datasets.
To further examine interpretability and understand the contribution of each convolutional component, Grad-CAM++ overlays were generated for each Filter Block (FB): FB1–FB3 for the three-layer model and FB1–FB4 for the four-layer model, as illustrated in Table 7. The filter block-wise heatmaps show a consistent interpretability progression within the three-layer model. Low-level facial edges were mainly captured by FB1, whereas mid-level semantic areas like the corners of the mouth, eyes, and eyebrows were prioritized by FB2. Effective spatial focus and discriminative learning were demonstrated by FB3, which generated highly class relevant activation regions.
Table 7. Layer-wise Grad-CAM++ visualizations for the selected models, showing activation progression across feature blocks from shallow (FB1) to deep (FB4) layers.
In the four-layer model, a similar progression was observed across FB1–FB3; however, the additional FB4 produced diffuse activations that extended beyond salient facial regions. The overlay visualizations showed spatially dispersed responses and no clear gradient patterns in FB4’s feature maps, indicating weakened gradient flow and reduced semantic focus. These effects likely result from gradient degradation and overfitting caused by the extra depth and capacity, leading to redundant rather than informative features. This observation aligns with the findings of [32,62], which report that increasing model complexity often reduces the spatial accuracy and causal relevance of deeper layer explanations.
In order to assess interpretability and explainability of the best chosen models, Grad-CAM++ visualizations obtained from the third convolutional layer (FB3) were used to evaluate interpretability and explainability, as deeper layers generally exhibit reduced spatial precision and less meaningful activations [62]. The Grad-CAM++ overlays of the corresponding layers are shown in Table 8. The results demonstrate that both the (64, 128, 256) and (64, 128, 256, 512) models have the ability to highlight relevant facial features for the given emotions as demonstrated by the activation maps. Both models are able to give meaningful visual explanations for their predictions by consistently highlighting the key features for each emotion, such as furrowed eyebrows and tight lips for anger, drooping mouth and lowered eyelids for sadness, smiling mouth and eyes for happiness, and more diffuse activation maps for neutral expressions.
Table 8. Qualitative examples of Grad-CAM++ overlays extracted from the third convolutional block (FB3), demonstrating spatial focus patterns and interpretability characteristics for the evaluated models.
The (64, 128, 256) model generally produced more meaningful heatmaps compared to the (64, 128, 256, 512) model, suggesting better overall explainability, especially for anger and sad samples. However, for some happy and neutral samples, the (64, 128, 256, 512) model’s activation maps appear to be more concentrated on key facial regions. This aligns with the confusion matrix analysis that concluded that the (64, 128, 256) model performs better on anger and sad emotions, while the (64, 128, 256, 512) model performs slightly better on happy and neutral emotions.
The results indicate that the (64, 128, 256) model demonstrates superior overall explainability. However, a more extensive analysis on a larger dataset is required to thoroughly assess the generalizability and consistency of the model visualizations. To quantitatively evaluate the explainability of the models, a comprehensive assessment was conducted using perturbation-based metrics, including IAUC, DAUC, AD, ADA, and IC, computed on the Grad-CAM++ saliency maps. Higher IAUC, AD, ADA, and IC values, along with lower DAUC scores, reflect more faithful and relevant explanations. This quantitative approach complements the qualitative visual inspection and provides a systematic comparison of the models’ explainability across multiple datasets. It thereby supports more definitive conclusions regarding their consistency, generalizability, and trustworthiness for facial emotion recognition. The metrics were evaluated on 17,739 samples, combining the FER2013, RAF-DB, and ExpW test sets, where both models were assessed. Table 9 summarizes the performance comparison of the (64, 128, 256) and (64, 128, 256, 512) models across the different explainability metrics.
Table 9. Comparison of Grad-CAM++ metrics for all emotions across the models. Where (↑) and (↓) are indicators of higher-is-better and lower-is-better metrics, respectively.
For the anger and sad emotions, the (64, 128, 256) model consistently outperforms the (64, 128, 256, 512) model across nearly all explainability metrics. It achieves higher IAUC, AD, IC, and ADA values, along with lower active pixel ratios and post-masking accuracy, reflecting stronger causal relevance and more focused saliency distributions. Although the 4-layer model exhibits slightly lower DAUC values, the difference is marginal. The slightly higher DAUC observed in the 3-layer model indicates a slower confidence drop when key regions are removed, suggesting reliance on broader yet still meaningful feature activations. This behavior highlights the model’s ability to capture semantically relevant but spatially distributed cues, enhancing both robustness and interpretability. Overall, the (64, 128, 256) model demonstrates superior spatial coherence and explanatory consistency for both anger and sad emotions.
For the Happy emotion, the (64, 128, 256) model achieves higher IAUC and IC scores, along with lower DAUC and Active Pixel ratio compared to the (64, 128, 256, 512) model. These results indicate that the shallower model provides more concentrated and discriminative activation regions, focusing more effectively on the key facial cues that characterize happiness. Although the deeper model shows slightly higher AD and ADA values, suggesting broader activations, leading to a higher performance drop after masking, as they include widespread but less informative regions. This is supported by the higher active pixel ratio of 0.578, confirming that the deeper model activates larger portions of the face. In other words, the deeper model captures more extensive areas with reduced localization, making its explanations appear more diffuse. Overall, the 3-layer model produces clearer and more interpretable attention patterns, while the 4-layer model tends to capture stronger but less focused responses.
The (64, 128, 256, 512) model outperforms the (64, 128, 256) model in terms of IAUC and IC for the neutral emotion, suggesting more reliable activation when detecting neutral features. However, it shows higher DAUC and lower ADA, suggesting that the deeper model captures stronger responses; its activations are less spatially precise when features are removed. The relatively high active pixels of 0.510 vs. 0.451 active pixels of both models reflect the difficulty of identifying neutral expressions, which lack distinctive cues and frequently cause the model to activate broader and less discriminative areas in order to rule out emotional features rather than focus on specific facial indicators.
Overall, the comparative analysis indicates that the (64, 128, 256) model offers a better balance between performance and interpretability across emotions. Table 10 summarizes the comparative metrics and supports this observation, showing that the model consistently achieves higher IAUC, IC, and ADA scores and lower DAUC and Active Pixel ratios for most emotions. These findings confirm its ability to generate more compact and discriminative attention maps. Even for the neutral emotion, where the deeper model shows stronger global activations, the shallower model maintains competitive performance with clearer spatial focus. The slightly worse values in some deletion-based metrics, such as DAUC and AD for Happy and Sad, are likely due to the model’s lower activation ratio, as it focuses on a smaller number of key regions compared to deeper models that remove more pixels, which can cause a sharper drop in performance.
Table 10. Average performance comparison of (64, 128, 256) and (64, 128, 256, 512) architectures across emotion categories of explainability metrics. Where (↑) and (↓) are indicators of higher-is-better and lower-is-better metrics, respectively.
Overall, the comparative analysis indicates that the (64, 128, 256) model provides a better balance between performance and interpretability across emotions. Table 10 summarizes the comparative metrics and supports this observation, showing that the (64, 128, 256) model achieves a better balance between performance and interpretability across emotions. It consistently achieves higher IAUC, IC, and ADA scores and lower DAUC and active pixel ratios for most emotions, confirming that it generates more compact and discriminative attention maps. Even for the neutral emotion, where the deeper model shows stronger global activations, the shallow model maintains competitive performance with clearer spatial focus. The slightly worse values in some deletion-based metrics, such as DAUC and AD for happy and sad, are likely due to the model’s lower activation ratio as it focuses on a smaller number of key regions compared to the other models that remove more pixels that can cause a sharper drop in performance. To ensure the reliability of these findings, the obtained Grad-CAM++ explainability metrics for all emotion classes were compared against empirically observed ranges reported in prior works. Average AD values typically range from 0.195 to 0.598 and IC from 0.067 to 0.191 [60], while IAUC and DAUC span approximately from 0.16 to 0.49 and from 0.016 to 0.296, respectively [63]. These comparisons confirm the stability and interpretability validity of the evaluated models. Figure 4 illustrates the comparative distribution of Grad-CAM++ explainability metrics. The (64, 128, 256) model exhibits superior explainability, showing higher AD, IAUC, ADA, and IC values, reflected by larger radar areas, while achieving lower DAUC, masked accuracy, and active pixel ratio, indicated by smaller radar areas for relevant metrics.
Figure 4. Radar-plot comparison of 3-layer (64, 128, 256) and 4-layer (64, 128, 256, 512) CNN emotion recognition models. (Left): higher-is-better metrics (IAUC, AD, IC, ADA). (Right): lower-is-better metrics (DAUC, Active Pixels, Accuracy after masking).

4.2. Reproducibility and Statistical Evaluation

To ensure the reliability and reproducibility of our federated learning experiments, the reported results for the FedXAI were computed over ten random seeds { 1 ,   3 ,   7 ,   17 ,   21 ,   33 ,   42 ,   55 ,   77 ,   100 } and evaluated on the three datasets. This multi-seed evaluation captures stochastic variability originating from model initialization, batch sampling, and client update ordering. For each metric, we report the mean and standard deviation (Std) across runs. The proposed framework’s robustness and stability were measured using a 95% confidence interval (95% CI) over repeated trials. Table 11 shows consistently narrow confidence intervals and low standard deviations across the datasets confirming the stable and statistically reliable performance of the proposed FedXAI framework.
Table 11. Statistical evaluation across 10 seeds: mean, standard deviation, and 95% confidence intervals for Accuracy, F1, Recall, and Precision (all reported in %).

4.3. Discussion on Client Scaling and Data Characteristics

To ensure a fair comparison between model configurations, we adopt a federated setup where all clients receive an equal number of samples and the same emotion classes, using a shuffled combination of the three FER datasets to introduce a level of heterogeneity. This controlled design ensures that any variation in performance can be attributed directly to architectural and configuration differences rather than fluctuations in client or data distribution. The integration of three natural FER datasets introduces inherent variability in capture conditions and label characteristics, providing a form of non-IID behavior across clients. To further examine how the framework behaves under increasingly distributed conditions, we conducted additional experiments by varying the number of participating clients. This analysis allows us to assess how performance scales as data become more distributed across a larger number of clients. Table 12 shows that the model maintains consistent accuracy across 3, 6, 9, and 15 clients, with only small fluctuations across configurations, which is expected given the increase in data fragmentation and heterogeneity. Despite this, the model exhibits stable performance under moderate scaling.
Table 12. Accuracy Across Client Configurations Using the (64, 128, 256) CNN.
However, real-world federated environments typically involve a larger number of clients, higher degrees of data heterogeneity, and stronger non-IID conditions, which remain challenging for practical deployment. In future work, we will expand the framework to larger-scale settings and consider aggregation strategies tailored to more challenging non-IID and heterogeneous client scenarios.

5. Comparison with Cross-Dataset Evaluation Methods

To demonstrate the generalization capability of the proposed framework, we benchmark it against representative cross-dataset FER approaches on RAF-DB, ExpW, and FER2013. Unlike methods trained on a single dataset, these approaches are designed to operate across heterogeneous data sources. For fair evaluation, we also train our network in a centralized setup by combining and shuffling the three datasets and follow the same test protocol as in the federated configuration. Table 13 reports the performance of state-of-the-art cross-dataset FER models spanning both large-scale and lightweight architectures, as established in [64]. While large backbone networks such as ResNet-50 variants achieve competitive results, several methods, including Stepwise Adaptive Feature Norm (SAFN) [65], Conditional Adversarial Domain Adaptation (CADA) [66], and Enhanced Channel Attention Network (ECAN) [67], exhibited limited generalization, particularly on FER2013. Though attention-based models, such as Attention-Guided Residual Aggregation (AGRA) [68], Cross-Global Local Representation Learning (CGLRL) [69], and Label Dynamic Weight Matching (LDWM) [64], improve cross-domain learning, however, they show noticeable performance drops across the different test sets.
Table 13. Cross-dataset FER comparison on RAF-DB, ExpW, and FER2013.
In contrast, our lightweight convolutional architecture delivers the highest mean accuracy (75.5%) among all lightweight models and even surpasses all ResNet-50 based cross-domain baselines in average accuracy. Moreover, it achieves consistently strong results across datasets, including the challenging FER2013 benchmark, which many methods struggled with.
Importantly, the federated version of our model sustained strong performance under data heterogeneity, achieving 82.8% (RAF-DB), 68.8% (ExpW), and 71.3% (FER2013), with an overall average of 74.3%. The marginal 1.2% gap compared to centralized training highlights the robustness and efficiency of our federated optimization pipeline, confirming the feasibility of privacy-preserving facial emotion recognition on distributed edge devices.
Overall, these results establish our method as a practical and scalable solution for real-world FER, effectively balancing accuracy, computational efficiency, and privacy. The superior mean accuracy across datasets, combined with compact model size and stable cross-domain performance, positions the proposed framework as a strong candidate for potential deployment in assistive and resource-constrained environments.

6. Edge Deployment and Runtime Evaluation

The real-world potential deployment feasibility was validated using a Raspberry Pi 4 Model B equipped with a 1.5 GHz quad-core ARM Cortex-A72 processor and 8 GB RAM. After conversion to TensorFlow Lite for optimized edge inference, the model was evaluated on the FER2013 test set and demonstrated excellent runtime performance, as shown in Table 14. The system achieved an average inference latency of 13.92 ± 1.46 ms, corresponding to 71.84 FPS throughput, exceeding the 15–30 FPS requirement for real-time video analysis [70]. Resource usage remained low, consuming only 25.30 ± 8.51% of a single CPU core while maintaining a peak memory footprint of 156.10 MB. Importantly, the optimized model retained its classification accuracy at 71.1%. With a compact storage requirement of 5.5 MB and a runtime memory overhead of 1.00 MB, demonstrating that the model is highly suitable for resource-constrained edge environments.
Table 14. Runtime Performance Metrics on Raspberry Pi 4 Model B.
A comparative analysis further underscores the advantages of the proposed lightweight CNN. For instance, MobileNetV2 requires 513.6 ms per inference (approximately 2 FPS) on the same hardware [71] with 25.6% CPU utilization, whereas our model achieves significantly lower inference latency with a comparable CPU usage of 25.3%. The proposed system also outperforms Light-FER’s reported throughput of 4.85 FPS [72] and demonstrates improved efficiency over existing Raspberry Pi 4 benchmarks, with MobileNetV1 achieving 82.7 ms inference time and MobileNetV2 reaching 112.6 ms using an accelerator [73].
In comparison with widely used mobile architectures reported in recent FER studies, the efficiency of the proposed lightweight CNN is further supported through its faster inference and higher throughput on edge devices. For example, in [74], MobileNetV1 (4.3 M parameters, 600 M FLOPs) achieved 22.6 ms, whereas MobileNetV2 (3.5 M parameters, 312.86 M FLOPs) reported 25.9 ms on an Intel Xeon Silver 4214R CPU at 2.40 GHz with dual processors and 128 GB RAM. In another study evaluating lightweight CNN models [75], MobileNetV2 with 3.51 M parameters and 0.6016 GFLOPs achieved an inference time of 58.4 ± 3.5 ms, while ShuffleNetV2 with 3.99 M parameters and 0.9743 GFLOPs required 63.3 ± 6.2 ms. Their recognition performance reached 58% and 65% accuracy on FER2013 and 73% and 80% on RAF-DB, respectively. A third study on an improved MobileNetV2 [76] reported 67.9% accuracy on FER2013 with a mobile inference latency of 300 ms on Huawei P60 hardware.
In comparison, the proposed model is significantly lighter and faster, requiring only 1.45 M parameters and 107 MFLOPs, yet achieving 71.1% accuracy on FER2013 and 82.8% on RAF-DB. Moreover, it delivers real-time inference at 13.92 ms on a Raspberry Pi 4, a substantially more resource-constrained edge device than the high-end desktop systems or modern smartphones used in the prior studies. Structurally, the proposed network uses three convolutional layers with a principled (64, 128, 256) channel progression that utilizes hierarchical feature learning [77], in which each layer increases representational capacity while maintaining computational efficiency suitable for edge devices. This design avoids the architectural complexity of MobileNetV1, MobileNetV2, and ShuffleNetV2, which rely on deeper architectures with numerous depthwise-separable convolutions, inverted residual blocks, and channel-shuffle units that introduce additional overhead and reduce interpretability. The three-layer structure provides efficient feature extraction while preserving spatial locality, which is crucial for facial emotion recognition [78], making it particularly suitable for edge deployment where computational resources and power budgets are constrained [79].

7. Ethical and Practical Considerations for Deployment

Several practical and ethical issues must be considered for potential deployment of FER systems in mental health contexts. Publicly available datasets are not designed for clinical assessment and do not represent clinical contexts, as a result limiting their suitability to such applications [5]. Furthermore, expert evaluation is essential for trustworthy deployment of AI systems in clinical contexts to confirm the alignment of the model-highlighted regions with meaningful and relevant cues [5]. According to [80], both individual AI model decisions and overall system behavior should be open to review by auditors. Future work will incorporate qualitative, expert-driven assessment of explanations in collaboration with mental health professionals to evaluate whether the highlighted regions correspond to clinically meaningful cues and to assess the practical interpretability of the system in real clinical contexts.
Bias is another major ethical challenge, occurring when FER systems make discriminatory decisions that disadvantage certain groups due to dataset imbalance, cross-cultural differences in emotional expression, and variations associated with mental health conditions [5]. Bias can also arise at multiple stages of the AI pipeline, including annotation practices, modeling choices, and deployment processes [80].
As a result, bias mitigation should involve multi-level strategies at the data, algorithmic, and evaluation levels. At the data level, recommended practices include demographic auditing, balanced sampling, and ensuring adequate representation of diverse and clinically relevant groups [81,82], as emphasized in recent fairness and bias mitigation surveys [83,84]. At the algorithmic level, explainability methods help reveal biased decision patterns [82]. In addition, fairness-aware modeling approaches include sampling adjustments, transparency mechanisms, and mitigation techniques such as reweighting, adversarial debiasing, equalized-odds post-processing, and fair representation learning, which aim to reduce unfair performance differences across demographic groups [80]. At the evaluation level, ongoing fairness monitoring across demographic subgroups is essential for detecting performance disparities [5]. Independent auditing frameworks are also recommended to support accountability and oversight [80].
Although federated learning prevents raw data sharing, privacy risks such as inference attacks or leakage through model updates remain an ongoing concern. Prior work on emotion-recognition technologies suggests that these systems are also susceptible to unintended inferences or misuse of derived information [5,82]. As a result, additional safeguards, including secure aggregation, differential privacy, and transparent auditing, are important components of responsible deployment. Given the sensitivity of emotional and mental health related information, such protections are essential for safe deployment [85].
Informed consent is another important consideration for continuous emotion monitoring in mental health contexts. FER systems may involve ongoing or repeated data processing, which raises additional considerations for meaningful informed consent [82]. To support meaningful consent, systems should provide clear and accessible explanations of data use, model limitations, and the scope of analysis, along with user controls that allow individuals to pause or disable monitoring. These steps are aligned with recent recommendations for ethically deploying affective computing technologies [5]. Informed consent is another important consideration for continuous emotion monitoring in mental health contexts. FER systems involve ongoing and repeated data processing, which raises additional requirements for meaningful informed consent [82]. To support informed consent, systems should provide clear and accessible explanations of data use, model limitations, and the scope of analysis, along with user controls that allow individuals to pause or disable monitoring, in line with recent recommendations for ethically deploying affective computing technologies [5].
Accountability and transparency are also essential for responsible deployment. At the technical level, documentation such as model cards, which provide standardized information on a model’s purpose, data sources, evaluation metrics, and known limitations, can help clinicians and users understand system behavior [82]. Recent best-practice guidelines emphasize structured documentation, including dataset provenance and model-level reporting, to support transparency and enable external auditing [80]. At the clinical level, human oversight remains crucial, with clinicians reviewing AI-generated outputs and retaining the authority to override system recommendations. Explainability methods used in this work support accountability by enabling domain experts to inspect the basis of model decisions. Together, these practices help ensure that FER-based mental health tools are deployed safely and responsibly [5].

8. Conclusions

This work presented a federated and explainability-driven facial emotion recognition framework that enables trustworthy, privacy-preserving, and edge-deployable emotion analysis with potential applications in mental health monitoring. Unlike traditional centralized FER systems, the proposed approach preserves data confidentiality through federated learning while enhancing transparency through quantitative evaluation of explanation quality using Grad-CAM++ perturbation-based metrics, including IAUC, DAUC, AD, IC, ADA, and Active Pixel Ratio.
A lightweight CNN architecture was designed to achieve efficient training and inference on resource-constrained devices while maintaining high accuracy. Extensive experiments on RAF-DB, ExpW, and FER2013 datasets demonstrated that the proposed model achieved strong cross-dataset generalization among lightweight architectures, with 75.5% and 74.3% average accuracies under centralized and federated configurations, respectively, confirming the framework’s generalization and suitability for potential decentralized deployment.
By integrating interpretability into the model selection process, this study establishes a quantitative and transparent pathway for trustworthy FER systems. Future work will extend this framework toward personalized federated learning and temporal emotion modeling, enabling continuous and adaptive affective monitoring in real-world assistive and mental health contexts. Additionally, integrating multimodal data sources, such as audio cues, physiological signals, and contextual information, will enable more comprehensive and contextually aware emotion understanding. Future work will also include qualitative validation of the Grad-CAM++ explanations, in collaboration with mental health professionals, to establish clinical interpretability.

Author Contributions

Conceptualization, H.T., A.B. and A.H.; Conceptualization, D.S.; methodology, D.S., H.T., A.B. and A.H.; software, D.S.; validation, D.S., H.T., A.B. and A.H.; formal analysis, D.S.; investigation, D.S.; resources, H.T., A.B. and A.H.; writing—original draft preparation, D.S.; writing—review and editing, D.S., H.T., A.B. and A.H.; visualization, D.S.; supervision, H.T., A.B. and A.H.; project administration, D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Perturbation-Based Metrics

Appendix A.1. Insertion and Deletion Area Under Curve

In order to assess the faithfulness of saliency maps, we use the two perturbation-based metrics introduced by [86]: Insertion Area Under Curve (IAUC) and Deletion Area Under Curve (DAUC) metrics. This allows us to determine the reliability of the Grad-CAM++ explanations. The metrics measure the changes in the model’s confidence as image regions are gradually added or removed, starting with the most significant features.
Let the input grayscale image be I R H × W and its Grad-CAM++ heatmap be S R H × W , normalized to the range [ 0 ,   1 ] . The heatmap S is first flattened and sorted in descending order of importance to obtain an ordered index list σ .
At each step k, corresponding to revealing or masking a proportion p k = k N of the total N pixels, a new image I k is constructed as follows:
I k ( ins ) ( x , y ) = I ( x , y ) , if ( x , y ) σ 1 : k , 0 , otherwise ,
I k ( del ) ( x , y ) = 0 , if   ( x , y ) σ 1 : k , I ( x , y ) , otherwise ,
where I k ( ins ) denotes the progressively revealed image in insertion and I k ( del ) the progressively occluded image in deletion. At each step, the model’s confidence for the target class c is recorded as
f k ( ins ) = m c ( I k ( ins ) ) , f k ( del ) = m c ( I k ( del ) ) ,
where f k ( ins ) and f k ( del ) denote the model’s predicted confidence scores for class c at step k during the insertion and deletion processes, respectively. Here, m c ( · ) represents the model’s output probability for the target class c given the perturbed image input. IAUC and DAUC are then computed as the area under the corresponding confidence curves. In practice, these integrals are estimated discretely using the trapezoidal rule:
AUC k = 1 N 1 ( f k + 1 + f k ) 2 N .
A faithful explanation map should cause a sharp increase in confidence during insertion, yielding a higher IAUC and a sharp decrease in confidence during deletion, yielding a lower DAUC.

Appendix A.2. Average Drop

A good model captures the important features of an image. When we classify an image, the probability calculated for the correct class is expected to drop if we remove the important regions identified by the explanation map, due to the loss of discriminative features. We use this fact to check the performance of each model by comparing the confidence when the entire image is passed to the classifier versus when the important regions are removed. A better explanation map will identify the most relevant parts, such that removing them leads to a larger drop in confidence. We compute this metric as the average percentage drop in the model’s confidence when the important regions are removed [60,87]. The Average Drop (AD) is expressed as
Average Drop ( AD ) = 1 N i = 1 N y i c R i c y i c
where
  • y i c represents the model’s confidence for class c of the ith sample when the entire image is used as input.
  • R i c is the model’s confidence for class c of the ith sample when the important regions are removed.
  • N is the total number of images.
Higher AD values indicate that the explanation successfully identifies decision-critical regions. In addition, we evaluate the Average Drop in Accuracy (ADA), which measures how often the model’s prediction changes when important regions are removed. ADA is defined as
A D A = 1 N i = 1 N 1 ( P i c = T i c ) 1 ( R i c = T i c )
where:
  • 1 ( · ) is the indicator function that equals 1 if the condition is true and 0 otherwise.
  • P i c and R i c denote the predicted classes for the original and masked images, respectively.
  • T i c represents the true class label for the ith sample.
The ADA metric captures the proportion of correctly classified samples that become misclassified after removing the important regions. A higher ADA value implies that the explanation map effectively identifies regions critical for correct classification.

Appendix A.3. Increase in Confidence

Contrary to previous expectations, there may be scenarios where removing the context helps boost prediction confidence, especially when the context contains noise or unnecessary distractions. Keeping such cases in mind, we calculate the number of times that passing only the explanation map to the classifier increases its confidence compared to passing the entire image.
Formally, the Increase in Confidence (IC) is defined as [87]:
I C = 1 N i = 1 N 1 y i c < O i c
where:
  • y i c represents the model’s confidence for class c of the ith sample when the entire image is used as input.
  • O i c represents the model’s confidence for class c of the ith sample when only the important regions identified by the explanation map are kept, and the remaining regions are replaced with the image mean.
  • 1 ( · ) is an indicator function that returns 1 when the argument is true (i.e., when O i c > y i c ) and 0 otherwise.
Higher IC values indicate that the explanation successfully identifies decision-critical regions.

References

  1. World Health Organization. Mental Disorders. 2022. Available online: https://www.who.int/news-room/fact-sheets/detail/mental-disorders (accessed on 1 November 2025).
  2. World Health Organization. Comprehensive Mental Health Action Plan 2013–2030. 2021. Available online: https://www.who.int/publications/i/item/9789240031029 (accessed on 25 August 2025).
  3. Fredrickson, B.L.; Losada, M.F. Positive affect and the complex dynamics of human flourishing. Am. Psychol. 2005, 60, 678–686. [Google Scholar] [CrossRef]
  4. Ong, A.D.; Bergeman, C.; Chow, S.M. Positive emotions as a basic building block of resilience in adulthood. In Handbook of Adult Resilience; Guilford Press: New York, NY, USA, 2010; pp. 81–93. [Google Scholar]
  5. Barker, D.; Tippireddy, M.K.R.; Farhan, A.; Ahmed, B. Ethical Considerations in Emotion Recognition Research. Psychol. Int. 2025, 7, 43. [Google Scholar] [CrossRef]
  6. McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), Ft. Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  7. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. Found. Trends Mach. Learn. 2021, 14, 1–210. [Google Scholar] [CrossRef]
  8. Minaee, S.; Minaei, M.; Abdolrashidi, A. Deep-emotion: Facial expression recognition using attentional convolutional network. Sensors 2021, 21, 3046. [Google Scholar] [CrossRef] [PubMed]
  9. Hou, C.; Ai, J.; Lin, Y.; Guan, C.; Li, J.; Zhu, W. Evaluation of Online Teaching Quality Based on Facial Expression Recognition. Future Internet 2022, 14, 177. [Google Scholar] [CrossRef]
  10. Shi, C.; Tan, C.; Wang, L. A facial expression recognition method based on a multibranch cross-connection convolutional neural network. IEEE Access 2021, 9, 39255–39274. [Google Scholar] [CrossRef]
  11. Zhou, N.; Liang, R.; Shi, W. A lightweight convolutional neural network for real-time facial expression detection. IEEE Access 2020, 9, 5573–5584. [Google Scholar] [CrossRef]
  12. Kim, J.; Kang, J.K.; Kim, Y. A Resource Efficient Integer-Arithmetic-Only FPGA-Based CNN Accelerator for Real-Time Facial Emotion Recognition. IEEE Access 2021, 9, 104367–104381. [Google Scholar] [CrossRef]
  13. Zhao, G.; Wei, W.; Xie, X.; Fan, S.; Sun, K. An FPGA-Based BNN Real-Time Facial Emotion Recognition Algorithm. In Proceedings of the 2022 IEEE International Conference on Artificial Intelligence and Computer Applications (ICAICA), Dalian, China, 24–26 June 2022; pp. 20–24. [Google Scholar]
  14. Saurav, S.; Gidde, P.; Saini, R.; Singh, S. Dual integrated convolutional neural network for real-time facial expression recognition in the wild. Vis. Comput. 2022, 38, 1083–1096. [Google Scholar] [CrossRef]
  15. Saurav, S.; Saini, R.; Singh, S. EmNet: A deep integrated convolutional neural network for facial emotion recognition in the wild. Appl. Intell. 2021, 51, 5543–5570. [Google Scholar] [CrossRef]
  16. Zhao, G.; Yang, H.; Yu, M. Expression recognition method based on a lightweight convolutional neural network. IEEE Access 2020, 8, 38528–38537. [Google Scholar] [CrossRef]
  17. Huang, Z.Y.; Chiang, C.C.; Chen, J.H.; Chen, Y.C.; Chung, H.L.; Cai, Y.P.; Hsu, H.C. A study on computer vision for facial emotion recognition. Sci. Rep. 2023, 13, 8425. [Google Scholar] [CrossRef]
  18. Dias, W.; Andaló, F.; Padilha, R.; Bertocco, G.; Almeida, W.; Costa, P.; Rocha, A. Cross-dataset emotion recognition from facial expressions through convolutional neural networks. J. Vis. Commun. Image Represent. 2022, 82, 103395. [Google Scholar] [CrossRef]
  19. Liang, L.; Lang, C.; Li, Y.; Feng, S.; Zhao, J. Fine-grained facial expression recognition in the wild. IEEE Trans. Inf. Forensics Secur. 2020, 16, 482–494. [Google Scholar] [CrossRef]
  20. Reghunathan, R.K.; Ramankutty, V.K.; Kallingal, A.; Vinod, V. Facial Expression Recognition Using Pre-trained Architectures. Eng. Proc. 2024, 62, 22. [Google Scholar]
  21. Gupta, S.; Kumar, P.; Tekchandani, R.K. Facial emotion recognition based real-time learner engagement detection system in online learning context using deep learning models. Multimed. Tools Appl. 2023, 82, 11365–11394. [Google Scholar] [CrossRef]
  22. Hu, F.; He, K.; Wang, C.; Zheng, Q.; Zhou, B.; Li, G.; Sun, Y. STRFLNet: Spatio-Temporal Representation Fusion Learning Network for EEG-Based Emotion Recognition. IEEE Trans. Affect. Comput. 2025; in press. [Google Scholar] [CrossRef]
  23. Cai, M.; Chen, J.; Hua, C.; Wen, G.; Fu, R. EEG emotion recognition using EEG-SWTNS neural network through EEG spectral image. Inf. Sci. 2024, 680, 121198. [Google Scholar] [CrossRef]
  24. Ma, Y.; Shen, J.; Zhao, Z.; Liang, H.; Tan, Y.; Liu, Z.; Qian, K.; Yang, M.; Hu, B. What can facial movements reveal? Depression recognition and analysis based on optical flow using Bayesian networks. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 3459–3468. [Google Scholar] [CrossRef] [PubMed]
  25. Shangguan, Z.; Liu, Z.; Li, G.; Chen, Q.; Ding, Z.; Hu, B. Dual-stream multiple instance learning for depression detection with facial expression videos. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 31, 554–563. [Google Scholar] [CrossRef] [PubMed]
  26. Gue, J.X.; Chong, C.Y.; Lim, M.K. Facial Expression Recognition as markers of Depression. In Proceedings of the 2023 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Taipei, Taiwan, 31 October–3 November 2023; pp. 674–680. [Google Scholar]
  27. Ramis Guarinos, S.; Manresa Yee, C.; Buades Rubio, J.M.; Gaya-Morey, F.X. Explainable Facial Expression Recognition for People with Intellectual Disabilities. In Proceedings of the XXIII International Conference on Human Computer Interaction, Copenhagen, Denmark, 23–28 July 2023; pp. 1–7. [Google Scholar]
  28. Rathod, M.; Dalvi, C.; Kaur, K.; Patil, S.; Gite, S.; Kamat, P.; Kotecha, K.; Abraham, A.; Gabralla, L.A. Kids’ emotion recognition using various deep-learning models with explainable ai. Sensors 2022, 22, 8066. [Google Scholar] [CrossRef] [PubMed]
  29. Ruangdit, T.; Sungkhin, T.; Phenglong, W.; Phaisangittisagul, E. Integration of Facial and Speech Expressions for Multimodal Emotional Recognition. In Proceedings of the TENCON 2023–2023 IEEE Region 10 Conference (TENCON), Chiang Mai, Thailand, 31 October–3 November 2023; pp. 519–523. [Google Scholar]
  30. Hettiarachchi, H.; Ekanayake, M.; Kaveendhya, G.; Koralage, O.; Samarasekara, P.; Kasthurirathna, D. Gender influence on emotional recognition using facial expressions and voice. In Proceedings of the 2023 5th International Conference on Advancements in Computing (ICAC), Colombo, Sri Lanka, 7–8 December 2023; pp. 721–726. [Google Scholar]
  31. Cesarelli, M.; Martinelli, F.; Mercaldo, F.; Santone, A. Emotion recognition from facial expression using explainable deep learning. In Proceedings of the 2022 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, Intl Conf on Pervasive Intelligence and Computing, Intl Conf on Cloud and Big Data Computing, Intl Conf on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), Falerna, Italy, 12–15 September 2022; pp. 1–6. [Google Scholar]
  32. Kandeel, A.A.; Abbas, H.M.; Hassanein, H.S. Explainable model selection of a convolutional neural network for driver’s facial emotion identification. In Proceedings of the International Conference on Pattern Recognition, Milan, Italy, 10–15 January 2021; pp. 699–713. [Google Scholar]
  33. Deramgozin, M.; Jovanovic, S.; Rabah, H.; Ramzan, N. A hybrid explainable ai framework applied to global and local facial expression recognition. In Proceedings of the 2021 IEEE International Conference on Imaging Systems and Techniques (IST), Online, 24–26 August 2021; pp. 1–5. [Google Scholar]
  34. Zhu, H.; Yu, C.; Cangelosi, A. Explainable emotion recognition for trustworthy human–robot interaction. In Proceedings of the Workshop Context-Awareness Human-Robot Interaction Approaches Challenges ACM/IEEE HRI, Sapporo, Japan, 7–10 March 2022. [Google Scholar]
  35. Borriero, A.; Milazzo, M.; Diano, M.; Orsenigo, D.; Villa, M.C.; DiFazio, C.; Tamietto, M.; Perotti, A. Explainable Emotion Decoding for Human and Computer Vision. In Proceedings of the World Conference on Explainable Artificial Intelligence, Valletta, Malta, 17–19 July 2024; pp. 178–201. [Google Scholar]
  36. Punuri, S.B.; Kuanar, S.K.; Mishra, T.K.; Rao, V.V.R.M.; Reddy, S.S. Decoding Human Facial Emotions: A Ranking Approach using Explainable AI. IEEE Access 2024, 12, 186229–186245. [Google Scholar] [CrossRef]
  37. Manresa-Yee, C.; Ramis, S.; Gaya-Morey, F.X.; Buades, J.M. Impact of explanations for trustworthy and transparent artificial intelligence. In Proceedings of the XXIII International Conference on Human Computer Interaction, Copenhagen, Denmark, 23–28 July 2023; pp. 1–8. [Google Scholar]
  38. Rajpal, A.; Sehra, K.; Bagri, R.; Sikka, P. Xai-fr: Explainable ai-based face recognition using deep neural networks. Wirel. Pers. Commun. 2023, 129, 663–680. [Google Scholar] [CrossRef] [PubMed]
  39. Gaya-Morey, F.X.; Ramis-Guarinos, S.; Manresa-Yee, C.; Buades-Rubio, J.M. Unveiling the human-like similarities of automatic facial expression recognition: An empirical exploration through explainable ai. Multimed. Tools Appl. 2024, 83, 85725–85753. [Google Scholar] [CrossRef]
  40. Lorch, S.; Gebele, J.; Brune, P. Towards Trustworthy AI: Evaluating SHAP and LIME for Facial Emotion Recognition. In Proceedings of the 58th Hawaii International Conference on System Sciences, Big Island, HI, USA, 7–10 January 2025; pp. 7532–7546. [Google Scholar]
  41. Di Luzio, F.; Rosato, A.; Panella, M. An explainable fast deep neural network for emotion recognition. Biomed. Signal Process. Control 2025, 100, 107177. [Google Scholar] [CrossRef]
  42. Jaswanth, M.; Narayana, N.; Rahul, S.; Amudha, J.; Aiswariya Milan, K. Emotion and Advertising Effectiveness: A Novel Facial Expression Analysis Approach Using Federated Learning. In Proceedings of the 2023 IEEE 20th India Council International Conference (INDICON), Hyderabad, India, 14–17 December 2023; pp. 368–373. [Google Scholar]
  43. Zhang, C.; Li, M.; Wu, D. Federated multidomain learning with graph ensemble autoencoder GMM for emotion recognition. IEEE Trans. Intell. Transp. Syst. 2022, 24, 7631–7641. [Google Scholar] [CrossRef]
  44. Franco, D.; Oneto, L.; Navarin, N.; Anguita, D. Toward learning trustworthily from data combining privacy, fairness, and explainability: An application to face recognition. Entropy 2021, 23, 1047. [Google Scholar] [CrossRef]
  45. Ghosh, T.; Banna, M.H.A.; Nahian, M.J.A.; Kaiser, M.S.; Mahmud, M.; Li, S.; Pillay, N. A privacy-preserving federated-mobilenet for facial expression detection from images. In Proceedings of the International Conference on Applied Intelligence and Informatics, Reggio Calabria, Italy, 1–3 September 2022; pp. 277–292. [Google Scholar]
  46. Simić, N.; Suzić, S.; Milošević, N.; Stanojev, V.; Nosek, T.; Popović, B.; Bajović, D. Enhancing Emotion Recognition through Federated Learning: A Multimodal Approach with Convolutional Neural Networks. Appl. Sci. 2024, 14, 1325. [Google Scholar] [CrossRef]
  47. Qi, F.; Zhang, Z.; Yang, X.; Zhang, H.; Xu, C. Feeling Without Sharing: A Federated Video Emotion Recognition Framework Via Privacy-Agnostic Hybrid Aggregation. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 151–160. [Google Scholar] [CrossRef]
  48. Treadway, M.T.; Zald, D.H. Reconsidering anhedonia in depression: Lessons from translational neuroscience. Neurosci. Biobehav. Rev. 2011, 35, 537–555. [Google Scholar] [CrossRef]
  49. Pizzagalli, D.A. Depression, stress, and anhedonia: Toward a synthesis and integrated model. Annu. Rev. Clin. Psychol. 2014, 10, 393–423. [Google Scholar] [CrossRef]
  50. Rottenberg, J. Mood and emotion in major depression. Curr. Dir. Psychol. Sci. 2005, 14, 167–170. [Google Scholar] [CrossRef]
  51. Bylsma, L.M.; Morris, B.H.; Rottenberg, J. A meta-analysis of emotional reactivity in major depressive disorder. Clin. Psychol. Rev. 2008, 28, 676–691. [Google Scholar] [CrossRef]
  52. Martin, L.A.; Neighbors, H.W.; Griffith, D.M. The experience of symptoms of depression in men vs women: Analysis of the National Comorbidity Survey Replication. JAMA Psychiatry 2013, 70, 1100–1106. [Google Scholar] [CrossRef]
  53. Busch, F.N. Anger and depression. Adv. Psychiatr. Treat. 2009, 15, 271–278. [Google Scholar] [CrossRef]
  54. Li, S.; Zhang, M.; Liu, Z.; Chen, W.; Yang, Y.; Jiang, Z. RAF-DB: A Real-world Affective Faces Database. In Proceedings of the 2017 International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017; pp. 51–57. [Google Scholar]
  55. Goodfellow, I.; DeVries, W.; Fidler, S.; Grosse, R.; Mnih, V.; Taylor, G.; An, Y. The Facial Expression Recognition 2013 (FER2013) Dataset. arXiv 2013, arXiv:1307.6888. [Google Scholar]
  56. Zhang, Z.; Luo, P.; Loy, C.C.; Tang, X. From Facial Expression Recognition to Interpersonal Relation Prediction. arXiv 2016, arXiv:1609.06426v2. [Google Scholar] [CrossRef]
  57. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef]
  58. Zhang, X.; Zhou, X.; Qiao, Y. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
  59. Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated learning: Challenges, methods, and future directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
  60. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 839–847. [Google Scholar]
  61. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  62. Fong, R.; Patrick, M.; Vedaldi, A. Understanding deep networks via extremal perturbations and smooth masks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2950–2958. [Google Scholar]
  63. Gomez, T.; Fréour, T.; Mouchère, H. Metrics for saliency map evaluation of deep learning explanation methods. In Proceedings of the International Conference on Pattern Recognition and Artificial Intelligence, Paris, France, 1–3 June 2022; pp. 84–95. [Google Scholar]
  64. Gao, Y.; Cai, Y.; Bi, X.; Li, B.; Li, S.; Zheng, W. Cross-domain facial expression recognition through reliable global–local representation learning and dynamic label weighting. Electronics 2023, 12, 4553. [Google Scholar] [CrossRef]
  65. Xu, R.; Li, G.; Yang, J.; Lin, L. Larger norm more transferable: An adaptive feature norm approach for unsupervised domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1426–1435. [Google Scholar]
  66. Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional adversarial domain adaptation. Adv. Neural Inf. Process. Syst. 2018, 31, 1647–1657. [Google Scholar]
  67. Li, S.; Deng, W. A deeper look at facial expression dataset bias. IEEE Trans. Affect. Comput. 2020, 13, 881–893. [Google Scholar] [CrossRef]
  68. Chen, T.; Pu, T.; Wu, H.; Xie, Y.; Liu, L.; Lin, L. Cross-Domain Facial Expression Recognition: A Unified Evaluation Benchmark and Adversarial Graph Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 9887–9903. [Google Scholar] [CrossRef]
  69. Xie, Y.; Gao, Y.; Lin, J.; Chen, T. Learning consistent global-local representation for cross-domain facial expression recognition. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 2489–2495. [Google Scholar]
  70. Lee, J.; Hwang, K.i. YOLO with adaptive frame control for real-time object detection applications. Multimed. Tools Appl. 2022, 81, 36375–36396. [Google Scholar] [CrossRef]
  71. Dey, A.; Srivastava, S.; Singh, G.; Pettit, R.G. Real-Time Performance Benchmarking of TinyML Models in Embedded Systems (PICO: Performance of Inference, CPU, and Operations). In Proceedings of the 2025 28th International Symposium on Real-Time Distributed Computing (ISORC), Toulouse, France, 26–28 May 2025; pp. 279–284. [Google Scholar]
  72. Pascual, A.M.; Valverde, E.C.; Kim, J.i.; Jeong, J.W.; Jung, Y.; Kim, S.H.; Lim, W. Light-FER: A Lightweight Facial Emotion Recognition System on Edge Devices. Sensors 2022, 22, 9524. [Google Scholar] [CrossRef]
  73. Allan, A. Benchmarking TensorFlow Lite on the New Raspberry Pi 4, Model B. 2019. Available online: https://www.hackster.io/news/benchmarking-tensorflow-lite-on-the-new-raspberry-pi-4-model-b-3fd859d05b98 (accessed on 1 November 2025).
  74. Liao, L.; Wu, S.; Song, C.; Fu, J. RS-Xception: A lightweight network for facial expression recognition. Electronics 2024, 13, 3217. [Google Scholar] [CrossRef]
  75. Gursesli, M.C.; Lombardi, S.; Duradoni, M.; Bocchi, L.; Guazzini, A.; Lanata, A. Facial emotion recognition (FER) through custom lightweight CNN model: Performance evaluation in public datasets. IEEE Access 2024, 12, 45543–45559. [Google Scholar] [CrossRef]
  76. Zhu, Q.; Zhuang, H.; Zhao, M.; Xu, S.; Meng, R. A study on expression recognition based on improved mobilenetV2 network. Sci. Rep. 2024, 14, 8121. [Google Scholar] [CrossRef]
  77. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  78. Li, S.; Deng, W. Deep facial expression recognition: A survey. IEEE Trans. Affect. Comput. 2020, 13, 1195–1215. [Google Scholar] [CrossRef]
  79. Han, S.; Mao, H.; Dally, W.J. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv 2015, arXiv:1510.00149. [Google Scholar]
  80. Alvarez, J.M.; Colmenarejo, A.B.; Elobaid, A.; Fabbrizzi, S.; Fahimi, M.; Ferrara, A.; Ghodsi, S.; Mougan, C.; Papageorgiou, I.; Reyero, P.; et al. Policy advice and best practices on bias and fairness in AI. Ethics Inf. Technol. 2024, 26, 31. [Google Scholar] [CrossRef]
  81. Latif, S.; Ali, H.S.; Usama, M.; Rana, R.; Schuller, B.; Qadir, J. Ai-based emotion recognition: Promise, peril, and prescriptions for prosocial path. arXiv 2022, arXiv:2211.07290. [Google Scholar] [CrossRef]
  82. Martinez-Martin, N. What are important ethical implications of using facial recognition technology in health care? AMA J. Ethics 2019, 21, E180. [Google Scholar] [PubMed]
  83. Ferrara, E. Fairness and bias in artificial intelligence: A brief survey of sources, impacts, and mitigation strategies. Sci 2024, 6, 3. [Google Scholar] [CrossRef]
  84. González-Sendino, R.; Serrano, E.; Bajo, J. Mitigating bias in artificial intelligence: Fair data generation via causal models for transparent and explainable decision-making. Future Gener. Comput. Syst. 2024, 155, 384–401. [Google Scholar] [CrossRef]
  85. Benouis, M.; Andre, E.; Can, Y.S. Balancing Between Privacy and Utility for Affect Recognition Using Multitask Learning in Differential Privacy–Added Federated Learning Settings: Quantitative Study. JMIR Ment. Health 2024, 11, e60003. [Google Scholar] [CrossRef]
  86. Petsiuk, V.; Das, A.; Saenko, K. RISE: Randomized Input Sampling for Explanation of Black-box Models. arXiv 2018, arXiv:1806.07421. [Google Scholar] [CrossRef]
  87. Mitra, S.; Sukul, A.; Roy, S.K.; Singh, P.; Verma, V. Reliable or Deceptive? Investigating Gated Features for Smooth Visual Explanations in CNNs. arXiv 2024, arXiv:2404.19341. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.