Next Article in Journal
An Optimal Deep Hybrid Framework with Selective Kernel U-Net for Skin Lesion Detection and Classification
Previous Article in Journal
Hypoglossal Nerve Transection Induces Anxiety- and Depression-like Behaviors with HPA Axis Dysregulation in Rats
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AI-Based Angle Map Analysis of Facial Asymmetry in Peripheral Facial Palsy

by
Andreas Heinrich
1,*,
Gerd Fabian Volk
2,3,4,
Christian Dobel
2,3,4 and
Orlando Guntinas-Lichius
2,3,4
1
Department of Radiology, Jena University Hospital—Friedrich Schiller University, Am Klinikum 1, 07747 Jena, Germany
2
Department of Otorhinolaryngology, Jena University Hospital—Friedrich Schiller University, Am Klinikum 1, 07747 Jena, Germany
3
Facial-Nerve-Center Jena, Jena University Hospital—Friedrich Schiller University, Am Klinikum 1, 07747 Jena, Germany
4
Center for Rare Diseases, Jena University Hospital—Friedrich Schiller University, Am Klinikum 1, 07747 Jena, Germany
*
Author to whom correspondence should be addressed.
Bioengineering 2026, 13(4), 426; https://doi.org/10.3390/bioengineering13040426
Submission received: 13 March 2026 / Revised: 27 March 2026 / Accepted: 3 April 2026 / Published: 6 April 2026

Abstract

Peripheral facial palsy (PFP) causes pronounced facial asymmetry and functional impairment, highlighting the need for reliable, objective assessment. This study presents a novel, fully automated, reference-free method for quantifying facial symmetry using artificial intelligence (AI)-based facial landmark detection. A total of 405 datasets from 198 PFP patients were analyzed, each including nine standardized facial expressions covering both resting and dynamic movements. AI detected 478 landmarks per image, from which 225 paired landmarks were used to compute local asymmetry angles. Systematic evaluation identified 91 highly informative landmark pairs, primarily around the eyes, nose and mouth, which simplified the analysis and enhanced discriminatory power, while also enabling region-specific assessment of asymmetry. Statistical evaluation included Kruskal–Wallis H-tests across clinical scores and Spearman correlations, showing moderate to strong associations (0.32–0.73, p < 0.001). The fully automated pipeline produced reproducible results and demonstrated robustness to head rotation. Intuitive full-face angle maps allowed direct assessment of asymmetry without a reference image. This AI-driven approach provides a robust, objective, and visually interpretable framework for clinical monitoring, severity classification, and treatment evaluation in PFP, combining quantitative precision with practical applicability.

1. Introduction

Peripheral facial palsy (PFP) frequently results in pronounced facial asymmetry, functional limitations, and psychosocial impact [1,2,3,4,5,6,7]. Objective and reproducible assessment of facial symmetry is essential for monitoring recovery, guiding rehabilitation, and supporting clinical decision-making [8,9,10,11,12,13,14,15].
Artificial intelligence (AI)-based landmark approaches have become increasingly common for quantifying facial asymmetry, providing automated and reproducible measurements of facial features. For example, previous studies [16,17,18,19] automatically detected 68 facial landmarks to compute distance- or angle-based asymmetry indices, demonstrating good correlation with established clinical grading systems. Standardized head positioning and expression protocols have typically been required to ensure measurement reliability, which can limit applicability in clinical settings. A recent component-based model using dense facial landmarks correlates well with clinical scores after vestibular schwannoma surgery [20]. However, its visualizations may be unintuitive for clinicians or patients, potentially limiting practical interpretability. Previous work [21] introduced reference-based dynamic heatmaps as a robust approach for assessing facial asymmetry, using a neutral facial image as a reference and quantifying deviations in expressive facial states. This method provided reliable spatial mapping and objective scoring of asymmetrical regions, while being relatively insensitive to the precise localization of individual landmarks. By focusing on intra-individual changes relative to a personal neutral baseline, it captured dynamic facial movements and offered reproducible quantification suitable for clinical monitoring and research applications. However, despite these advances, a key limitation of existing approaches is that they may rely on a reference image, are not fully automated, or do not provide an intuitive representation of facial symmetry applicable across both resting and dynamic facial states. This limits their applicability in routine clinical settings and in scenarios where baseline images are unavailable.
Therefore, the aim of this study is to address this gap by developing and evaluating a fully automated, reference-free, landmark-based method for objective facial symmetry assessment in patients with PFP. By leveraging high-density facial landmarks and angle-based metrics, the proposed approach seeks to combine quantitative robustness with intuitive visualization, thereby extending current methods toward more flexible and clinically applicable use.

2. Materials and Methods

This retrospective study received approval from the Institutional Review Board of Jena University Hospital, Germany (protocol 2019-1539-BO). Written informed consent was not required due to the study design. All procedures adhered to the applicable ethical guidelines and the principles outlined in the Declaration of Helsinki.

2.1. Study Population and Image Acquisition

A total of 518 datasets from patients with unilateral PFP were retrospectively screened. They were collected between 26 March 2008 and 20 December 2011, and each consisted of nine standardized facial photographs capturing a range of expressions. Datasets were excluded if any of these nine images were missing. After applying these inclusion criteria, 405 datasets from 198 patients (94 female, 104 male) remained for analysis. Patient ages ranged from 4 to 90 years (mean 53 ± 19 years). Among these, 98 patients had multiple datasets acquired on different days during therapy.
Nine standardized facial photographs per patient captured a range of expressions, including:
(1)
Neutral;
(2)
Gentle eye closure;
(3)
Forced eye closure;
(4)
Frowning/forehead wrinkling;
(5)
Nose wrinkling;
(6)
Closed-mouth stretch;
(7)
Mouth stretch with teeth visible;
(8)
Lip pursing;
(9)
Downward movement of the mouth corners.
Most images were taken with Nikon DSLR cameras (mainly D90 and D100), with a few captured using Canon or Sony cameras. Flash and white balance settings varied across cases, and image resolutions ranged from approximately 700 × 1000 to 1800 × 2400 pixels. Randomization and blinding were not applicable due to the retrospective, observational nature of the study.

2.2. Image Preprocessing and Analysis

Facial images were analyzed in Python (v3.14.0, Python Software Foundation) [22] using an AI-based landmark detection approach (MediaPipe v0.10.33, Google Research) [23], which automatically located 478 facial landmarks per photograph with enhanced prediction accuracy. To standardize head alignment, 20 central landmarks (excluding eight mouth landmarks) were first used to define a midline through the center of the face (see Figure 1a, red points). A straight line was fitted to these points using ordinary least squares regression to estimate the facial midline. Each image was then rotated around its center to achieve approximate horizontal head alignment (see Figure 1b).
For the analysis of facial symmetry, 225 landmark pairs were defined on corresponding left and right face locations (450 points total) to calculate the angles of the connecting lines relative to a horizontal reference (see Figure 1c). For each pair L = x L , y L and R = x R , y R , the angle θ was calculated as:
θ =   180 π · a r c t a n 2 y R y L , x R x L
The set of all angles for a given image forms a high-dimensional asymmetry feature vector a = ( θ 1 , θ 2 , , θ N ) , representing local deviations for each landmark pair. Each angle encodes both direction and magnitude, providing a quantitative representation of facial asymmetry. Visual representations of these vectors were generated as angle maps (see Figure 1d).
From this asymmetry feature vector, a global asymmetry score was computed as the mean of the absolute angular deviations across all selected landmark pairs:
S c o r e A I =   1 N · i = 1 N θ i
Higher ScoreAI values indicate greater asymmetry. The reported standard deviations reflect variability of ScoreAI across samples within each severity group. Region-wise features were also computed for the eyes, nose and mouth by selecting subsets of the vector, enabling localized assessment of asymmetry patterns.
The fully automated pipeline, including landmark detection, image rotation, angle computation, and visualization, produced reproducible results without manual correction or additional preprocessing (e.g., cropping, resizing, or intensity normalization).
To assess robustness, each image was rotated from −25° to 25° in 1° increments, and the previously described pipeline was applied to each rotation. The mean absolute deviation of ScoreAI from the original image was then computed.

2.3. Landmark Filtering

In order to evaluate whether using fewer than 225 landmark pairs could improve the analysis (see Figure 1e–h), datasets with available Stennert index facial grading scores were examined. Of the 405 datasets, 266 included these scores and were used for landmark filtering and correlation analyses, while the remaining datasets were included only in the image-based symmetry analysis.
Facial nerve function was graded using the Stennert index, a commonly applied scoring system in Germany [24]. It evaluates facial symmetry both at rest (0 = no asymmetry, 4 = severe asymmetry) and during voluntary movement (0 = normal movement, 6 = complete or near-complete impairment).
For scoring at rest, four yes-or-no criteria are rated: Eyelid fissure difference 3 mm or more? Ectropion? Nasolabial fold smoothed (if formed on the healthy side)? Low corner of the mouth 3 mm or more? Every “yes” is counted with 1 point; every “no” with 0 points.
During movement, scoring is based on the following criteria: Frowning (wrinkle formation or raising of the eyebrow) not possible or only partly but less than 50% of the healthy side possible? Residual palpebral fissure in sleeping position? Residual palpebral fissure at tight eye closing? Teeth showing upper and lower canine teeth not visible? Teeth showing upper second incisor across entire width not visible? Corners of mouth (shortening of distance between philtrum and corner of mouth compared to healthy side) less than 50% compared to healthy side? Again, every “yes” is counted with 1 point; every “no” with 0 points.
One-way ANOVA was used to evaluate the discriminatory power of individual landmark pairs across Stennert index grades under resting and movement conditions. For each pair, absolute angular deviations were averaged across all images per patient to obtain patient-level values, and eta-squared (η2) was calculated to quantify the proportion of variance explained by grade as a measure of effect size. Systematic η2 thresholds (0.001–0.400 in 0.001 increments) were applied to identify the most informative pairs; only pairs exceeding the threshold in both conditions were retained. This threshold-based selection served as an exploratory feature-selection step.
To define the optimal number of landmark pairs, each subset of selected pairs was used to compute ScoreAI, which was then evaluated using the Kruskal–Wallis H-test across Stennert index grades to determine overall discriminatory power under both conditions. Finally, Spearman’s rank correlation coefficients were calculated between ScoreAI derived from the optimal subset of landmark pairs and the corresponding clinical Stennert scores for resting and movement conditions, to assess the monotonic association between objective symmetry measures and clinical severity. Statistical significance was assessed using the corresponding p-values, with significance set at p < 0.05. All analyses were performed in Python using standard statistical libraries.

2.4. Comparison with Other 68-Point Landmark Models

For comparison with other landmark-based approaches, the standard Dlib 68-point predictor (shape_predictor_68_face_landmarks.dat) and three Emotrics-based 68-point predictors (mee_shape_predictor_68_face_landmarks.dat, mee_shape_predictor_68_face_landmarks2.dat, and mee_shape_predictor_68_face_landmarks3.dat) were evaluated. From each set, 6 landmarks excluding four mouth landmarks were used to define the facial midline for head alignment. For asymmetry analysis, 29 landmark pairs were initially defined and evaluated. In addition, a reduced set excluding landmarks corresponding to the facial contour was tested, resulting in 21 landmark pairs. All landmarks were processed using the same analytical pipeline as the 478-point model, including midline-based head alignment, left-right pairing, and ScoreAI computation. This allowed a direct comparison between the 68-point configurations and the high-density 478-point model, which was also tested in reduced configurations of 140, 91, 50, and 21 landmark pairs.

3. Results

The method was successfully applied to all 405 datasets. Analyses requiring clinical Stennert index scores used a subset of 266 datasets. This subset allowed evaluation of the discriminatory power of landmark pair reductions in relation to clinical severity.

3.1. Discriminatory Power of Landmark Sets

To assess whether a reduced set of landmark pairs could differentiate between clinical severity levels, systematic reduction in landmark pairs was performed. The analysis showed that 91 pairs provided optimal separation between Stennert index grades. These pairs exhibited a minimum eta-squared (η2) value of 0.178, indicating a moderate to high discriminatory power across clinical severity levels. The Kruskal–Wallis H-test statistic initially increased with decreasing number of pairs, peaking at 91, and then declined as fewer pairs were included (Figure 2). Higher H-values indicate stronger separation of clinical severity levels, and p-values remained < 0.001 for all configurations, confirming statistically significant differentiation. These results suggest that the 91-pair configuration balances complexity and discriminatory performance effectively.

3.2. ScoreAI Distribution and Correlation with Clinical Severity

The relationship between ScoreAI and clinical severity was examined across nine standardized facial expressions. For the 91 selected pairs, ScoreAI increased with higher Stennert index grades, particularly for eye-, nose- and mouth-related movements. Expressions 5–7 under rest and movement showed the strongest correlation with clinical severity (Figure 3). This indicates that the method captures both static and dynamic asymmetry patterns. Summary statistics for all datasets (Table 1) and for the subset with Stennert scores (Table 2) further highlight that increasing clinical severity is associated with higher ScoreAI, reflecting reduced facial symmetry. Region-wise analyses for eye, nose and mouth landmarks are presented in Table A1, Table A2 and Table A3, respectively. These region-specific asymmetry scores allow evaluation of how each facial region contributes to overall facial asymmetry and correlates with clinical severity.
Spearman’s rank correlations confirmed moderate to strong associations between ScoreAI derived from the 91 most informative landmark pairs and Stennert index scores (see Table 3). In addition, region-specific analyses based on the eye (24 pairs), nose (23 pairs), and mouth (44 pairs) subsets showed that the mouth region was more strongly correlated with clinical severity, particularly during voluntary movement, whereas eye-region correlations were lower but remained significant. The nose region also showed moderate to strong correlations for expressions 5–7; expression 4 (frowning/forehead wrinkling) showed a weak correlation at rest (r = 0.15, p = 0.013) and was not significant during movement (r = 0.06, p = 0.359). Mouth-related expressions (6–7) demonstrated the highest correlations during movement. Overall, these results indicate that the angular measures reflect facial asymmetry and that full-face and region-specific analyses provide complementary information.

3.3. Visualization

To examine whether changes over time and across patients could be visualized, angle maps were generated for both longitudinal and cross-sectional data. Example data from a single patient demonstrated that ScoreAI and angle maps can track therapy-related improvements over time (Figure 4). Across the patient cohort, extreme and percentile cases illustrated the range of observed asymmetry patterns without revealing individual photographs (Figure 5). Dynamic expressions, such as mouth stretches, clearly highlighted affected sides, providing an intuitive visual representation of asymmetry across expression types.

3.4. Comparison with Other 68-Point Landmark-Based Models

The reduced MediaPipe configurations demonstrated the strongest and most consistent separation across Stennert grades in both rest and movement. The 21- and 50-pair models achieved the highest slopes, indicating the greatest sensitivity to differences in severity, while the 91-pair configuration provided a favorable balance between slope and stability, maintaining informative regional patterns in a clinically interpretable format (see Figure 6). The full 225-pair MediaPipe model offered a more detailed representation but showed lower slopes, reflecting weaker separation between severity levels.
In comparison, Dlib- and Emotrics-based configurations generally exhibited lower slopes and greater variability in ScoreAI across Stennert grades. Although some variants occasionally reached high maximum values, they showed less consistent increases across grades and wider ranges within individual severity levels, indicating less reliable differentiation (see Table 4 and Table A4). Across all methods, ScoreAI increased with higher Stennert grades, confirming that the angular measures capture the progression of facial asymmetry.

3.5. Robustness to Artificial Rotations

The 91-pair ScoreAI showed high stability across artificial head rotations from −25° to +25°. Absolute deviations of the ScoreAI from the original (unrotated) images were low, with average deviations of 0.22 ± 0.21° and median deviations of 0.17° across the evaluated range. Even at the extremes of the rotation range, the ScoreAI remained consistent, confirming robustness of the measure (see Table 5).

4. Discussion

The presented method provides a fully automated, reference-free approach for quantifying facial symmetry in patients with PFP, applicable to both resting and dynamic facial states. Angle-based measurements of paired AI-based landmarks provide a quantitative and reproducible measure, while visual angle maps allow intuitive identification of asymmetrical facial regions. Landmark pairs around the mouth, nose and eyes exerted the greatest influence on discriminating clinical Stennert index grades, which may reflect regions commonly affected in PFP, although this pattern may also be influenced by landmark distribution or expression-specific variability.
Automated assessment of facial palsy increasingly relies on landmark-based approaches. Lee et al. [16] used 68 CNN-detected landmarks to compute asymmetry scores from 10 to 12 distance ratios across six expressions in 128 patients, but were limited by a small, non-diverse training dataset (300 faces) and reliance on distance ratios. Kim et al. [17], Anping et al. [18], and Müller et al. [19] similarly applied 68 landmarks within Emotrics or Dlib frameworks to calculate linear and angular asymmetry parameters, yet all faced small cohorts (11–36 patients), limited expressions, and partial manual landmark adjustment. The present method leverages MediaPipe’s dense coverage of 478 facial points, trained on 30,000 in-the-wild images [25], enabling detailed spatial analysis and robust feature detection. This high-density mesh provides inherent redundancy as a key advantage for clinical robustness because localized errors in landmark detection or partial facial occlusions caused by hair or shadows are mitigated by averaging across the entire feature vector, which prevents single-point inaccuracies from disproportionately skewing the global asymmetry score. From 225 initial paired points, 91—mainly around the eyes, nose and mouth—were most discriminative for clinical grading. The Stennert index assesses facial regions in a way comparable to the Sunnybrook Facial Grading System, which has demonstrated high intra- and interrater reliability [14,26,27]. Although sensitivity and specificity are standard diagnostic metrics, they are not directly applicable here because the Stennert index is an ordinal rather than binary scale. Therefore, Spearman’s rank correlation was used to assess the association between angular measures and clinical severity. Effect size analysis confirmed that landmarks controlling eyelid closure, mouth opening, and lip corner movement carried the strongest discriminatory power, while regions such as the jaw contour or hairline contributed little, reflecting their limited functional relevance in facial palsy. The measured angles between these landmark pairs provide a quantitative assessment of facial asymmetry and serve as a proxy for surface-level manifestations of underlying muscle activation. While these geometric measures correlate with functional impairment, they reflect the cumulative effect of muscle tone and skin displacement rather than direct physiological activity. Guan et al. [20] reported similar findings with a component-based CNN for eyelid and mouth landmarks in 274 post-vestibular schwannoma patients, but their approach was restricted to discrete regions and did not allow an immediate, intuitive full-face visualization of asymmetry.
Compared to previous reference-based work [21], the current approach analyzes each image directly without requiring a neutral reference. In that earlier framework, expressive images were aligned to a neutral reference, and local spatial filtering was applied to the resulting difference maps to suppress minor landmark noise and emphasize dynamic deviations from the patient’s own baseline. This design proved highly robust when a neutral image was available. The new reference-free method, by contrast, computes angular symmetry measures directly from each frame, eliminating dependency on baseline alignment and potentially enabling application to single-image assessments, video sequences, retrospective photo collections, and telemedical use. These strategies address complementary aspects of the same problem: the reference-based pipeline enhances consistency and noise resilience, while the reference-free, high-density landmark approach improves spatial resolution and flexibility. The reference-based heatmap and the angle map each provide privacy-preserving, visually interpretable, and intuitive full-face visualizations of facial function.
The current framework is compatible with different AI-based landmark detection models, which may vary in performance depending on the application and over time as models are updated. Its dense landmark representation allows for both full-face and region-specific analyses by selecting corresponding subsets of the asymmetry feature vector, enabling localized assessment of asymmetry patterns without repeated image processing. In this study, full-face measurements integrating eye, nose, and mouth landmarks showed the highest overall discrimination across Stennert grades, whereas smaller configurations, such as the 21- and 50-pair MediaPipe sets, may perform particularly well when asymmetry is concentrated in specific regions, such as the mouth during voluntary movement. Because ScoreAI is calculated as the mean of the absolute angular deviations across the selected landmark pairs, reduced subsets can emphasize region-specific effects, while broader configurations capture complementary asymmetry information across the face. A further advantage is that all analyses—including global scores, region-specific scores, and landmark-pair visualizations—can be derived simultaneously from a single image processing step, producing a comprehensive dataset for flexible and detailed interpretation of facial asymmetry. The framework also allows for future refinement of ScoreAI, such as weighting landmark pairs differently or integrating additional facial features, which could further enhance its discriminative power and clinical applicability.
More broadly, objective facial-performance measures should be interpreted in the context of underlying peripheral nerve function, muscle activation patterns, and rehabilitation status. Future studies could integrate image-based asymmetry metrics with complementary physiological measures, such as electromyography or clinical grading, and explore how neurobiological, cognitive, and stress-related factors influence facial motor patterns [28,29,30,31,32]. Such multimodal approaches may provide a more comprehensive interpretation of facial asymmetry beyond purely geometric analysis.
Despite its strengths, several limitations should be acknowledged. The accuracy of symmetry quantification depends directly on precise landmark localization and correct head orientation. To standardize image orientation, a vertical midline was defined using central facial landmarks, which is crucial because even small deviations from a horizontal head position can substantially distort angular measurements. Consequently, the accuracy of this alignment remains a critical factor influencing the overall precision of the symmetry metrics. Although the MediaPipe model, not specifically trained on PFP patients, generally performed robustly [21,33], extreme asymmetries or atypical expressions may still induce local errors that propagate to the angular measures. While robustness to moderate head rotation was confirmed through systematic sensitivity analysis, future studies could further quantify the impact of extreme facial occlusion, atypical expressions, or model-specific landmark noise, which can propagate to angular measurements. However, the current high-density framework is designed to minimize such effects through feature averaging. In practice, occlusion is avoided in patient recordings or images are retaken, and landmark variability depends on the AI model used. Additional variability arises from the retrospective dataset, including differences in camera type, resolution, and illumination. While midline alignment mitigates some of these effects, residual variations may still influence measurements. Moreover, as the method inherently compares the affected with the unaffected side, it cannot be directly applied to cases of bilateral facial palsy. Future studies could improve landmark detection by retraining the model with a large number of standardized PFP images or by implementing a secondary correction network to identify and adjust inconsistent landmarks. Achieving this, however, requires a substantial dataset of well-standardized images of PFP patients.

5. Conclusions

In conclusion, the proposed approach provides a fully automated and reference-free method for quantifying facial asymmetry in patients with PFP. Based on a high-density landmark model and a reduced set of 91 informative landmark pairs, which integrate eye, nose, and mouth landmarks, the method demonstrated moderate to strong associations with clinical Stennert index scores and effective discrimination between severity levels, highlighting the added value of combining complementary facial regions for assessing asymmetry. The resulting angle maps enable intuitive visualization of asymmetry patterns across different facial expressions, supporting the interpretation of regional differences. While the approach shows potential as an objective and reproducible framework for facial symmetry assessment, further studies are required to validate its clinical applicability and to compare its performance with established methods.

Author Contributions

Conceptualization, A.H.; methodology, A.H.; software, A.H.; validation, A.H.; formal analysis, A.H.; investigation, A.H. and G.F.V.; resources, A.H. and O.G.-L.; data curation, A.H. and O.G.-L.; writing—original draft preparation, A.H.; writing—review and editing, A.H., G.F.V., C.D. and O.G.-L.; visualization, A.H.; supervision, A.H.; project administration, A.H.; funding acquisition, O.G.-L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of Jena University Hospital (protocol code 2019–1539-BO and date of approval 2019).

Informed Consent Statement

Patient consent was waived due to the retrospective nature of the study, as approved by the Institutional Review Board at Jena University Hospital. Written informed consent has been obtained from the patient to publish the identifiable photographs shown in Figure 1.

Data Availability Statement

The datasets analyzed during the current study are not publicly available due to privacy concerns. However, all relevant data supporting the findings of this study are included in the article. For further inquiries or data requests, please contact the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
CNNConvolutional Neural Network
PFPPeripheral facial palsy

Appendix A

Table A1. ScoreAI (±standard deviation) derived from 24 landmark pairs for the eye region, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions.
Table A1. ScoreAI (±standard deviation) derived from 24 landmark pairs for the eye region, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions.
nScoreAI (Eye Region) of Expressions [°]
123456789
Stennert index rest
0991.02 ± 0.641.13 ± 0.791.09 ± 0.811.07 ± 0.731.29 ± 0.961.11 ± 0.761.08 ± 0.861.13 ± 0.741.10 ± 0.71
1531.31 ± 0.821.43 ± 0.931.63 ± 1.032.06 ± 1.201.55 ± 0.981.42 ± 0.891.48 ± 1.001.52 ± 0.941.23 ± 0.75
2591.31 ± 0.771.59 ± 0.841.78 ± 1.002.07 ± 1.231.83 ± 1.221.56 ± 0.971.79 ± 1.161.69 ± 1.051.42 ± 0.88
3391.28 ± 0.691.71 ± 0.922.10 ± 1.272.23 ± 1.001.82 ± 1.112.12 ± 1.572.23 ± 1.591.60 ± 0.911.48 ± 0.91
4162.11 ± 1.012.82 ± 1.552.87 ± 1.492.34 ± 1.262.01 ± 1.302.73 ± 1.112.51 ± 0.822.19 ± 0.982.26 ± 1.28
Stennert index movement
0580.89 ± 0.600.90 ± 0.610.85 ± 0.500.87 ± 0.621.06 ± 0.750.92 ± 0.610.76 ± 0.540.99 ± 0.660.88 ± 0.63
1271.31 ± 0.741.30 ± 0.831.35 ± 0.921.46 ± 0.681.59 ± 1.021.38 ± 0.841.37 ± 1.001.52 ± 0.851.43 ± 0.81
2351.27 ± 0.881.21 ± 0.961.19 ± 0.811.66 ± 1.391.48 ± 0.911.13 ± 0.801.22 ± 0.791.37 ± 0.901.19 ± 0.72
3211.25 ± 0.681.84 ± 0.811.66 ± 0.601.71 ± 1.201.44 ± 0.911.47 ± 0.931.47 ± 0.901.47 ± 1.201.22 ± 0.75
4391.31 ± 0.751.41 ± 0.791.57 ± 0.952.37 ± 1.281.88 ± 1.211.51 ± 0.991.59 ± 1.121.82 ± 1.051.54 ± 0.94
5221.33 ± 0.591.60 ± 0.982.02 ± 1.072.05 ± 0.921.38 ± 0.871.36 ± 0.731.74 ± 0.991.30 ± 0.761.25 ± 0.59
6641.46 ± 0.902.11 ± 1.132.50 ± 1.342.19 ± 1.002.06 ± 1.292.40 ± 1.342.56 ± 1.331.76 ± 0.931.67 ± 1.06
The table also shows the number of datasets (n) used for each Stennert index category. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Table A2. ScoreAI (±standard deviation) derived from 23 landmark pairs for the nose region, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions.
Table A2. ScoreAI (±standard deviation) derived from 23 landmark pairs for the nose region, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions.
nScoreAI (Nose Region) of Expressions [°]
123456789
Stennert index rest
0990.63 ± 0.420.70 ± 0.430.68 ± 0.500.77 ± 0.490.93 ± 0.640.78 ± 0.740.84 ± 0.730.74 ± 0.520.71 ± 0.59
1530.68 ± 0.530.76 ± 0.631.01 ± 0.980.70 ± 0.461.72 ± 1.551.24 ± 1.031.49 ± 1.140.96 ± 0.890.91 ± 0.70
2590.84 ± 0.570.74 ± 0.510.91 ± 0.790.87 ± 0.652.22 ± 1.221.87 ± 1.342.23 ± 1.511.27 ± 0.731.14 ± 0.81
3391.03 ± 0.830.89 ± 0.541.32 ± 0.991.01 ± 0.712.63 ± 1.502.53 ± 1.112.45 ± 1.531.32 ± 0.891.36 ± 0.92
4161.51 ± 1.121.27 ± 0.871.93 ± 1.231.44 ± 0.853.07 ± 1.942.78 ± 1.832.98 ± 1.971.49 ± 1.281.84 ± 1.57
Stennert index movement
0580.63 ± 0.430.69 ± 0.450.67 ± 0.560.73 ± 0.470.77 ± 0.490.71 ± 0.440.65 ± 0.390.69 ± 0.480.62 ± 0.37
1270.62 ± 0.300.66 ± 0.290.65 ± 0.320.78 ± 0.501.01 ± 0.570.75 ± 0.460.79 ± 0.410.71 ± 0.330.71 ± 0.57
2350.65 ± 0.350.75 ± 0.490.81 ± 0.580.80 ± 0.470.98 ± 0.650.69 ± 0.440.95 ± 0.650.72 ± 0.440.63 ± 0.41
3210.78 ± 0.550.75 ± 0.550.98 ± 0.700.89 ± 0.701.79 ± 1.031.37 ± 0.901.50 ± 0.761.05 ± 0.740.97 ± 0.47
4390.98 ± 0.700.79 ± 0.640.96 ± 0.781.04 ± 0.672.49 ± 1.271.82 ± 1.342.03 ± 1.601.44 ± 0.831.22 ± 0.87
5220.80 ± 0.670.87 ± 0.480.92 ± 0.530.90 ± 0.612.15 ± 1.271.95 ± 1.112.59 ± 1.721.22 ± 0.881.15 ± 0.80
6641.00 ± 0.890.91 ± 0.701.46 ± 1.280.89 ± 0.732.77 ± 1.762.64 ± 1.492.77 ± 1.471.33 ± 1.051.52 ± 1.20
The table also shows the number of datasets (n) used for each Stennert index category. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Table A3. ScoreAI (±standard deviation) derived from 44 landmark pairs for the mouth region, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions.
Table A3. ScoreAI (±standard deviation) derived from 44 landmark pairs for the mouth region, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions.
nScoreAI (Mouth Region) of Expressions [°]
123456789
Stennert index rest
0991.00 ± 0.731.07 ± 0.711.31 ± 1.031.11 ± 0.811.58 ± 1.231.35 ± 1.531.27 ± 1.231.39 ± 0.881.30 ± 1.14
1531.11 ± 0.761.07 ± 0.901.95 ± 1.641.03 ± 0.842.87 ± 2.482.53 ± 2.042.30 ± 1.761.70 ± 1.361.74 ± 1.26
2591.60 ± 1.141.32 ± 1.001.99 ± 1.901.54 ± 1.213.83 ± 2.643.89 ± 2.654.07 ± 2.652.23 ± 1.392.01 ± 1.44
3391.96 ± 1.311.63 ± 1.072.36 ± 1.861.70 ± 1.134.60 ± 2.205.01 ± 2.224.36 ± 2.462.32 ± 1.652.30 ± 1.88
4163.09 ± 2.242.70 ± 2.043.83 ± 2.472.58 ± 1.706.03 ± 3.266.09 ± 3.944.88 ± 3.222.67 ± 1.863.27 ± 3.18
Stennert index movement
0580.96 ± 0.691.02 ± 0.681.19 ± 0.961.03 ± 0.771.20 ± 0.771.04 ± 0.760.90 ± 0.701.30 ± 0.761.07 ± 0.73
1270.94 ± 0.470.96 ± 0.571.35 ± 0.921.12 ± 0.811.78 ± 1.101.32 ± 0.721.23 ± 0.711.25 ± 0.781.41 ± 1.06
2351.13 ± 0.661.06 ± 0.851.47 ± 1.061.21 ± 0.951.85 ± 1.431.33 ± 1.141.55 ± 1.111.61 ± 1.091.31 ± 0.74
3211.46 ± 1.141.40 ± 0.941.94 ± 1.331.31 ± 0.942.97 ± 1.512.83 ± 1.582.69 ± 1.291.74 ± 0.991.62 ± 1.00
4391.68 ± 1.291.32 ± 1.101.98 ± 1.441.60 ± 1.193.75 ± 2.633.45 ± 2.573.42 ± 2.532.53 ± 1.592.11 ± 1.57
5221.36 ± 1.171.27 ± 1.091.78 ± 1.261.46 ± 1.093.75 ± 2.154.06 ± 2.294.28 ± 2.952.41 ± 1.712.35 ± 1.75
6642.05 ± 1.611.82 ± 1.402.96 ± 2.521.70 ± 1.405.26 ± 2.895.68 ± 2.874.88 ± 2.492.17 ± 1.592.61 ± 2.31
The table also shows the number of datasets (n) used for each Stennert index category. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Table A4. ScoreAI is reported for each landmark-based method, including the proposed MediaPipe configurations, as well as Dlib and Emotrics variants with 21 and 29 landmark (LM) pairs. Values are shown for rest (Stennert 0–4) and movement (Stennert 0–6) conditions.
Table A4. ScoreAI is reported for each landmark-based method, including the proposed MediaPipe configurations, as well as Dlib and Emotrics variants with 21 and 29 landmark (LM) pairs. Values are shown for rest (Stennert 0–4) and movement (Stennert 0–6) conditions.
ScoreAI [°]
MediaPipeDlibEmotrics-1Emotrics-2Emotrics-3
LM2259150212921292129212921
Stennert index rest
00.86 ± 0.581.10 ± 0.671.27 ± 0.591.39 ± 0.341.43 ± 0.941.55 ± 0.931.82 ± 1.582.16 ± 1.661.68 ± 1.341.93 ± 1.411.86 ± 1.602.21 ± 1.69
11.15 ± 0.831.54 ± 0.951.83 ± 0.842.03 ± 0.471.92 ± 1.312.17 ± 1.322.39 ± 2.112.88 ± 2.222.24 ± 1.842.62 ± 1.932.35 ± 2.152.85 ± 2.27
21.38 ± 1.071.99 ± 1.202.47 ± 1.032.82 ± 0.592.18 ± 1.492.47 ± 1.492.58 ± 2.223.10 ± 2.322.50 ± 2.102.94 ± 2.212.44 ± 2.152.94 ± 2.27
31.58 ± 1.272.30 ± 1.442.90 ± 1.223.34 ± 0.732.29 ± 1.582.65 ± 1.582.81 ± 2.553.44 ± 2.662.79 ± 2.393.35 ± 2.502.77 ± 2.613.40 ± 2.75
41.96 ± 1.613.04 ± 1.793.89 ± 1.524.53 ± 0.842.82 ± 2.043.28 ± 2.073.21 ± 2.743.92 ± 2.843.25 ± 2.913.92 ± 3.052.99 ± 2.713.67 ± 2.84
Stennert index movement
00.75 ± 0.490.93 ± 0.561.07 ± 0.491.16 ± 0.271.30 ± 0.851.38 ± 0.841.66 ± 1.401.93 ± 1.481.55 ± 1.231.75 ± 1.311.81 ± 1.532.13 ± 1.61
10.97 ± 0.621.17 ± 0.711.29 ± 0.631.40 ± 0.341.58 ± 1.041.73 ± 1.051.89 ± 1.632.27 ± 1.691.81 ± 1.462.10 ± 1.541.88 ± 1.612.25 ± 1.69
20.95 ± 0.671.21 ± 0.761.41 ± 0.681.57 ± 0.361.75 ± 1.191.99 ± 1.182.03 ± 1.802.44 ± 1.902.07 ± 1.692.44 ± 1.772.04 ± 1.832.46 ± 1.93
31.21 ± 0.851.64 ± 0.961.99 ± 0.832.23 ± 0.501.83 ± 1.202.03 ± 1.192.29 ± 1.992.75 ± 2.092.10 ± 1.652.46 ± 1.721.99 ± 1.732.39 ± 1.82
41.37 ± 1.041.97 ± 1.172.43 ± 1.002.72 ± 0.602.09 ± 1.422.40 ± 1.422.55 ± 2.253.10 ± 2.362.52 ± 2.142.99 ± 2.252.42 ± 2.222.94 ± 2.34
51.30 ± 1.111.98 ± 1.252.52 ± 1.082.86 ± 0.602.15 ± 1.552.41 ± 1.592.84 ± 2.483.43 ± 2.612.56 ± 2.182.99 ± 2.322.88 ± 2.713.53 ± 2.87
61.72 ± 1.392.54 ± 1.573.21 ± 1.343.73 ± 0.782.50 ± 1.732.88 ± 1.753.00 ± 2.653.65 ± 2.772.86 ± 2.463.43 ± 2.582.83 ± 2.593.46 ± 2.73

References

  1. Cuenca-Martínez, F.; Zapardiel-Sánchez, E.; Carrasco-González, E.; La Touche, R.; Suso-Martí, L. Assessing anxiety, depression and quality of life in patients with peripheral facial palsy: A systematic review. PeerJ 2020, 8, e10449. [Google Scholar] [CrossRef] [PubMed]
  2. Hotton, M.; Huggons, E.; Hamlet, C.; Shore, D.; Johnson, D.; Norris, J.H.; Kilcoyne, S.; Dalton, L. The psychosocial impact of facial palsy: A systematic review. Br. J. Health Psychol. 2020, 25, 695–727. [Google Scholar] [CrossRef] [PubMed]
  3. Volk, G.F.; Hesse, S.; Geißler, K.; Kuttenreich, A.M.; Thielker, J.; Dobel, C.; Guntinas-Lichius, O. Role of body dysmorphic disorder in patients with postparalytic facial synkinesis. Laryngoscope 2021, 131, E2518–E2524. [Google Scholar] [CrossRef]
  4. Osthues, M.; Kuttenreich, A.-M.; Volk, G.F.; Dobel, C.; Strauss, B.; Altmann, U.; Guntinas-Lichius, O. Continual rehabilitation motivation of patients with postparalytic facial nerve syndrome. Eur. Arch. Oto-Rhino-Laryngol. 2022, 279, 481–491. [Google Scholar] [CrossRef] [PubMed]
  5. Bruins, T.E.; van Veen, M.M.; Mooibroek-Leeuwerke, T.; Werker, P.M.N.; Broekstra, D.C.; Dijkstra, P.U. Association of Socioeconomic, Personality, and Mental Health Factors With Health-Related Quality of Life in Patients With Facial Palsy. JAMA Otolaryngol.—Head Neck Surg. 2020, 146, 331–337. [Google Scholar] [CrossRef]
  6. Bylund, N.; Hultcrantz, M.; Jonsson, L.; Marsk, E. Quality of Life in Bell’s Palsy: Correlation with Sunnybrook and House-Brackmann Over Time. Laryngoscope 2021, 131, E612–E618. [Google Scholar] [CrossRef]
  7. Nellis, J.C.; Ishii, M.; Byrne, P.J.; Boahene, K.D.O.; Dey, J.K.; Ishii, L.E. Association Among Facial Paralysis, Depression, and Quality of Life in Facial Plastic Surgery Patients. JAMA Facial Plast. Surg. 2017, 19, 190–196. [Google Scholar] [CrossRef]
  8. Volk, G.F.; Roediger, B.; Geißler, K.; Kuttenreich, A.-M.; Klingner, C.M.; Dobel, C.; Guntinas-Lichius, O. Effect of an intensified combined electromyography and visual feedback training on facial grading in patients with post-paralytic facial synkinesis. Front. Rehabil. Sci. 2021, 2, 746188. [Google Scholar] [CrossRef]
  9. Hahnemann, I.; Fron, J.; Ballmaier, J.; Guntinas-Lichius, O.; Volk, G.F. Electromyography as an objective outcome measure for the therapeutic effect of biofeedback training to reduce post-paralytic facial synkinesis. Healthcare 2025, 13, 550. [Google Scholar] [CrossRef]
  10. Ross, B.G.; Fradet, G.; Nedzelski, J.M. Development of a Sensitive Clinical Facial Grading System. Otolaryngol.—Head Neck Surg. 1996, 114, 380–386. [Google Scholar] [CrossRef] [PubMed]
  11. Renne, A.; Heaton, J.; Boahene, K.D. Associations of AI-Based Facial Metrics With Patient-Reported Outcomes in Idiopathic Facial Paralysis. Laryngoscope 2026. online ahead of print. [Google Scholar] [CrossRef]
  12. Hartley, H.; Blumenow, W.; Williams, R.; Fattah, A. 7—Assessment and Grading of Synkinesisand Facial Palsy. In Management of Post-Facial Paralysis Synkinesis; Azizzadeh, B., Nduka, C., Eds.; Elsevier: Amsterdam, The Netherlands, 2022; pp. 51–58. [Google Scholar]
  13. Miller, M.Q.; Hadlock, T.A.; Fortier, E.; Guarin, D.L. The Auto-eFACE: Machine Learning–Enhanced Program Yields Automated Facial Palsy Assessment Tool. Plast. Reconstr. Surg. 2021, 147, 467–474. [Google Scholar] [CrossRef]
  14. Fattah, A.Y.; Gurusinghe, A.D.R.; Gavilan, J.; Hadlock, T.A.; Marcus, J.R.; Marres, H.; Nduka, C.C.; Slattery, W.H.; Snyder-Warwick, A.K. Facial nerve grading instruments: Systematic review of the literature and suggestion for uniformity. Plast Reconstr Surg 2015, 135, 569–579. [Google Scholar] [CrossRef]
  15. Banks, C.A.; Bhama, P.K.; Park, J.; Hadlock, C.R.; Hadlock, T.A. Clinician-Graded Electronic Facial Paralysis Assessment: The eFACE. Plast. Reconstr. Surg. 2015, 136. [Google Scholar] [CrossRef]
  16. Lee, S.A.; Kim, J.; Lee, J.M.; Hong, Y.-J.; Kim, I.-J.; Lee, J.D. Automatic facial recognition system assisted-facial asymmetry scale using facial landmarks. Otol. Neurotol. 2020, 41, 1140–1148. [Google Scholar] [CrossRef] [PubMed]
  17. Kim, M.G.; Bae, C.R.; Oh, T.S.; Park, S.J.; Jeong, J.M.; Kim, D.Y. Reliability and validity of emotrics in the assessment of facial palsy. J. Pers. Med. 2023, 13, 1135. [Google Scholar] [CrossRef] [PubMed]
  18. Anping, S.; Guoliang, X.; Xuehai, D.; Jiaxin, S.; Gang, X.; Wu, Z. Assessment for facial nerve paralysis based on facial asymmetry. Australas. Phys. Eng. Sci. Med. 2017, 40, 851–860. [Google Scholar] [CrossRef]
  19. Müller, S.L.C.; Zeier, C.; Pfister, P.; Menzi, N.; Tafrishi, B.; Schaefer, D.J.; Plock, J.A.; Ismail, T.; Klein, H.J. Artificial Intelligence Analysis of Symmetry and Emotions in Facial Palsy Patients After Botulinum Toxin A Injections. Toxins 2025, 17, 597. [Google Scholar] [CrossRef]
  20. Guan, X.; Wang, H.; Zhang, D.; Hou, C.; Ma, S.; Zhang, J.; Sun, X.; Liu, W.; Lou, M.; Qu, Y.; et al. Continuous and componentized facial palsy measurement alignment and clinical interpretable model. npj Digit. Med. 2025, 8, 694. [Google Scholar] [CrossRef]
  21. Heinrich, A.; Volk, G.F.; Dobel, C.; Guntinas-Lichius, O. Deep learning-based automatic facial symmetry scoring in peripheral facial palsy. Sci. Rep. 2025, 15, 31531. [Google Scholar] [CrossRef] [PubMed]
  22. Python Software Foundation. Python Language Reference. Available online: https://www.python.org (accessed on 13 March 2026).
  23. Google Research. MediaPipe. Available online: https://ai.google.dev/edge/mediapipe/solutions/guide (accessed on 13 March 2026).
  24. Stennert, E.; Limberg, C.H.; Frentrup, K.P. Parese- und Defektheilungs-Index. Ein leicht anwendbares Schema zur objektiven Bewertung von Therapieerfolgen bei Fazialispraesen. Arch. Angiol. 1977, 25, 238–245. [Google Scholar]
  25. Grishchenko, I.; Ablavatski, A.; Kartynnik, Y.; Raveendran, K.; Grundmann, M. Attention mesh: High-fidelity face mesh prediction in real-time. arXiv 2020, arXiv:2006.10962. [Google Scholar] [CrossRef]
  26. Mengi, E.; Kara, C.O.; Ardıç, F.N.; Topuz, B.; Metin, U.; Alptürk, U.; Aydemir, G.; Şenol, H. Comparison of the Reliability of the House–Brackmann, Facial Nerve Grading System 2.0, and Sunnybrook Facial Grading System for the Evaluation of Patients with Peripheral Facial Paralysis. J. Int. Adv. Otol. 2024, 20, 14. [Google Scholar] [CrossRef] [PubMed]
  27. Kanerva, M.; Poussa, T.; Pitkäranta, A. Sunnybrook and House-Brackmann Facial Grading Systems: Intrarater repeatability and interrater agreement. Otolaryngol.—Head Neck Surg. 2006, 135, 865–871. [Google Scholar] [CrossRef] [PubMed]
  28. Martins, L.A.; Schiavo, A.; Paz, L.V.; Xavier, L.L.; Mestriner, R.G. Neural underpinnings of fine motor skills under stress and anxiety: A review. Physiol. Behav. 2024, 282, 114593. [Google Scholar] [CrossRef]
  29. Jamadar, S.D.; Behler, A.; Deery, H.; Breakspear, M. The metabolic costs of cognition. Trends Cogn. Sci. 2025, 29, 541–555. [Google Scholar] [CrossRef]
  30. Bress, K.S.; Cascio, C.J. Sensorimotor regulation of facial expression—An untouched frontier. Neurosci. Biobehav. Rev. 2024, 162, 105684. [Google Scholar] [CrossRef]
  31. Nazzi, C.; Avenanti, A.; Battaglia, S. The Involvement of Antioxidants in Cognitive Decline and Neurodegeneration: Mens Sana in Corpore Sano. Antioxidants 2024, 13, 701. [Google Scholar] [CrossRef]
  32. Garofalo, S.; Timmermann, C.; Battaglia, S.; Maier, M.E.; di Pellegrino, G. Mediofrontal Negativity Signals Unexpected Timing of Salient Outcomes. J. Cogn. Neurosci. 2017, 29, 718–727. [Google Scholar] [CrossRef]
  33. Veeravalli, S.S.; Bodapati, P. Facial key points and contour identification for facial paralysis images. J. Data Acquis. Process. 2024, 39, 561–574. [Google Scholar]
Figure 1. Workflow of the facial symmetry analysis method. (a) All 478 facial landmarks are initially detected (green and red points). A midline is estimated from the central red landmarks (blue line). (b) The image is rotated so that this midline becomes vertical, aligning the head horizontally. (c) Landmark pairs on the left and right sides of the face relative to the vertical midline are connected by lines, with color indicating the angle relative to the horizontal, transitioning from blue to red as the angle increases. (d) Visualization of the angle map without the patient photo, showing 225 landmark pairs. (eh) Angle maps with progressively reduced landmark sets of 140, 91, 50, and 21 pairs, highlighting that the most informative features are concentrated in the mouth, nose and eye regions. Written informed consent for publication of these identifiable photographs was obtained from the patient.
Figure 1. Workflow of the facial symmetry analysis method. (a) All 478 facial landmarks are initially detected (green and red points). A midline is estimated from the central red landmarks (blue line). (b) The image is rotated so that this midline becomes vertical, aligning the head horizontally. (c) Landmark pairs on the left and right sides of the face relative to the vertical midline are connected by lines, with color indicating the angle relative to the horizontal, transitioning from blue to red as the angle increases. (d) Visualization of the angle map without the patient photo, showing 225 landmark pairs. (eh) Angle maps with progressively reduced landmark sets of 140, 91, 50, and 21 pairs, highlighting that the most informative features are concentrated in the mouth, nose and eye regions. Written informed consent for publication of these identifiable photographs was obtained from the patient.
Bioengineering 13 00426 g001
Figure 2. Discriminatory power of landmark sets across Stennert index grades. Boxplots show the distribution of ScoreAI for each patient grouped by Stennert index levels at rest (top) and during voluntary movement (bottom) for five different landmark sets (225, 140, 91, 50, 21 pairs). The mean is marked with a blue diamond. The Kruskal–Wallis H-test statistic and associated p-value for each landmark set are indicated above the respective groups, reflecting the ability of each configuration to differentiate clinical severity. Larger H-values correspond to stronger separation between Stennert index grades, highlighting that the 91-pair configuration achieves optimal discriminatory performance.
Figure 2. Discriminatory power of landmark sets across Stennert index grades. Boxplots show the distribution of ScoreAI for each patient grouped by Stennert index levels at rest (top) and during voluntary movement (bottom) for five different landmark sets (225, 140, 91, 50, 21 pairs). The mean is marked with a blue diamond. The Kruskal–Wallis H-test statistic and associated p-value for each landmark set are indicated above the respective groups, reflecting the ability of each configuration to differentiate clinical severity. Larger H-values correspond to stronger separation between Stennert index grades, highlighting that the 91-pair configuration achieves optimal discriminatory performance.
Bioengineering 13 00426 g002
Figure 3. Distribution of ScoreAI across facial expressions and Stennert index levels. Top: Boxplots of ScoreAI for all nine expressions and the overall mean (“all”) for five landmark sets (225, 140, 91, 50, 21 pairs) across all patients, illustrating overall variability, with each set indicated by a different shade of gray. The mean is marked with a blue diamond. Middle: Boxplots for the 91-pair configuration, stratified by Stennert index at rest (levels 0–4), showing the distribution of ScoreAI across the nine expressions and overall mean for patients with available clinical scores. Bottom: Boxplots for the 91-pair configuration, stratified by Stennert index during voluntary movement (levels 0–6), displaying the same information. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Figure 3. Distribution of ScoreAI across facial expressions and Stennert index levels. Top: Boxplots of ScoreAI for all nine expressions and the overall mean (“all”) for five landmark sets (225, 140, 91, 50, 21 pairs) across all patients, illustrating overall variability, with each set indicated by a different shade of gray. The mean is marked with a blue diamond. Middle: Boxplots for the 91-pair configuration, stratified by Stennert index at rest (levels 0–4), showing the distribution of ScoreAI across the nine expressions and overall mean for patients with available clinical scores. Bottom: Boxplots for the 91-pair configuration, stratified by Stennert index during voluntary movement (levels 0–6), displaying the same information. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Bioengineering 13 00426 g003
Figure 4. Visualization of facial symmetry in a single patient across multiple time points and nine standardized facial expressions. Each row corresponds to one time point, ordered chronologically (ad), with the ScoreAI of all 91 landmark pairs displayed above each row. Each column represents one facial expression (1–9), and the associated angle map is color-coded from blue (0° deviation) to red (≥5° deviation) to indicate the degree of asymmetry between paired landmarks. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Figure 4. Visualization of facial symmetry in a single patient across multiple time points and nine standardized facial expressions. Each row corresponds to one time point, ordered chronologically (ad), with the ScoreAI of all 91 landmark pairs displayed above each row. Each column represents one facial expression (1–9), and the associated angle map is color-coded from blue (0° deviation) to red (≥5° deviation) to indicate the degree of asymmetry between paired landmarks. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Bioengineering 13 00426 g004
Figure 5. Visualization of facial symmetry across patients with peripheral facial palsy, showing extreme and percentile cases for each of the nine standardized facial expressions. Each row (ae) represents one target group based on ScoreAI values across 91 landmark pairs: maximum, 75th percentile, median, 25th percentile, and minimum. Columns correspond to the nine facial expressions (1–9). The angle maps are color-coded from blue (0° deviation) to red (≥5° deviation), indicating the degree of asymmetry between paired landmarks. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Figure 5. Visualization of facial symmetry across patients with peripheral facial palsy, showing extreme and percentile cases for each of the nine standardized facial expressions. Each row (ae) represents one target group based on ScoreAI values across 91 landmark pairs: maximum, 75th percentile, median, 25th percentile, and minimum. Columns correspond to the nine facial expressions (1–9). The angle maps are color-coded from blue (0° deviation) to red (≥5° deviation), indicating the degree of asymmetry between paired landmarks. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Bioengineering 13 00426 g005
Figure 6. Comparison of landmark detection AI models using the same original image as in Figure 4, row 7. Rows (ad) show different time points, and columns show the respective model configurations. The ScoreAI is displayed above each image, and the color bar shows the angular deviation scale from 0° to ≥5°.
Figure 6. Comparison of landmark detection AI models using the same original image as in Figure 4, row 7. Rows (ad) show different time points, and columns show the respective model configurations. The ScoreAI is displayed above each image, and the color bar shows the angular deviation scale from 0° to ≥5°.
Bioengineering 13 00426 g006
Table 1. Summary statistics of ScoreAI derived from 91 landmark pairs for each of the nine facial expressions and overall average (“all”) across all 405 datasets, as well as region-specific results for the eye, nose and mouth.
Table 1. Summary statistics of ScoreAI derived from 91 landmark pairs for each of the nine facial expressions and overall average (“all”) across all 405 datasets, as well as region-specific results for the eye, nose and mouth.
ScoreAI of Expressions [°]
123456789All
mean1.241.261.611.362.332.272.211.611.551.72
SD0.860.811.180.791.671.881.770.991.110.96
variance0.740.661.390.632.783.533.150.991.240.92
median0.991.071.301.141.861.601.591.341.281.48
max5.315.937.864.257.9510.1510.166.168.115.75
min0.220.230.260.250.290.180.200.210.260.44
region-specific eye analysis (24 landmark pairs)
mean1.231.471.581.821.561.491.551.461.351.50
SD0.771.01.081.21.071.091.110.970.880.70
region-specific nose analysis (23 landmark pairs)
mean0.830.811.010.851.771.541.691.061.071.18
SD0.670.590.910.591.461.411.520.850.940.78
region-specific mouth analysis (44 landmark pairs)
mean1.451.381.941.383.043.072.851.991.922.12
SD1.311.191.781.122.572.92.651.511.761.42
The table shows the mean, standard deviation (SD), variance, median, maximum (max), and minimum (min) symmetry scores. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Table 2. ScoreAI (±standard deviation) derived from 91 landmark pairs, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions. Higher ScoreAI indicates lower facial symmetry between the affected and unaffected sides. A trend of increasing ScoreAI with increasing Stennert index can be observed, particularly in nose wrinkling (expression 5) and mouth-related expressions (expressions 6 and 7) for movement, reflecting the association between higher facial palsy severity and reduced symmetry.
Table 2. ScoreAI (±standard deviation) derived from 91 landmark pairs, stratified by Stennert index for rest (0–4) and movement (0–6) across nine facial expressions. Higher ScoreAI indicates lower facial symmetry between the affected and unaffected sides. A trend of increasing ScoreAI with increasing Stennert index can be observed, particularly in nose wrinkling (expression 5) and mouth-related expressions (expressions 6 and 7) for movement, reflecting the association between higher facial palsy severity and reduced symmetry.
nScoreAI of Expressions [°]
123456789
Stennert index rest
0990.91 ± 0.511.00 ± 0.501.09 ± 0.661.01 ± 0.531.34 ± 0.861.14 ± 1.021.11 ± 0.881.16 ± 0.601.10 ± 0.74
1531.05 ± 0.541.09 ± 0.661.63 ± 1.081.22 ± 0.612.23 ± 1.601.91 ± 1.321.88 ± 1.211.46 ± 0.921.40 ± 0.76
2591.33 ± 0.721.25 ± 0.611.66 ± 1.181.51 ± 0.872.90 ± 1.662.77 ± 1.713.00 ± 1.741.84 ± 0.941.63 ± 0.90
3391.54 ± 0.871.46 ± 0.622.03 ± 1.201.67 ± 0.693.37 ± 1.393.62 ± 1.473.31 ± 1.681.88 ± 0.901.85 ± 1.07
4162.43 ± 1.492.37 ± 1.373.09 ± 1.592.23 ± 0.994.22 ± 2.024.37 ± 2.423.77 ± 2.052.25 ± 1.292.64 ± 1.99
Stennert index movement
0580.86 ± 0.480.91 ± 0.460.97 ± 0.570.91 ± 0.501.05 ± 0.470.92 ± 0.510.80 ± 0.431.06 ± 0.510.91 ± 0.46
1270.96 ± 0.320.98 ± 0.341.17 ± 0.591.13 ± 0.481.54 ± 0.691.19 ± 0.551.15 ± 0.491.19 ± 0.531.24 ± 0.69
2351.04 ± 0.421.02 ± 0.631.23 ± 0.701.23 ± 0.661.54 ± 0.951.12 ± 0.761.31 ± 0.741.32 ± 0.701.10 ± 0.45
3211.23 ± 0.741.35 ± 0.641.62 ± 0.811.31 ± 0.662.27 ± 1.072.11 ± 1.072.07 ± 0.871.49 ± 0.731.35 ± 0.59
4391.41 ± 0.861.21 ± 0.691.61 ± 0.931.66 ± 0.862.94 ± 1.662.53 ± 1.702.59 ± 1.742.07 ± 1.001.73 ± 0.95
5221.21 ± 0.801.26 ± 0.651.63 ± 0.811.47 ± 0.792.72 ± 1.312.82 ± 1.413.18 ± 1.901.82 ± 1.101.76 ± 0.99
6641.63 ± 1.091.66 ± 0.942.46 ± 1.601.62 ± 0.863.79 ± 1.824.05 ± 1.823.73 ± 1.581.85 ± 1.042.09 ± 1.44
The table also shows the number of datasets (n) used for each Stennert index category. Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Table 3. Spearman’s rank correlation coefficients and corresponding p-values for the association between the Stennert index (rest and movement) and ScoreAI derived from 91 landmark pairs across nine facial expressions in 266 datasets. Positive correlation coefficients indicate that higher clinical severity (higher Stennert index) is associated with higher ScoreAI, reflecting lower facial symmetry.
Table 3. Spearman’s rank correlation coefficients and corresponding p-values for the association between the Stennert index (rest and movement) and ScoreAI derived from 91 landmark pairs across nine facial expressions in 266 datasets. Positive correlation coefficients indicate that higher clinical severity (higher Stennert index) is associated with higher ScoreAI, reflecting lower facial symmetry.
Spearman’s Rank Correlation Coefficients of Expressions
123456789All
rest0.380.350.440.420.580.630.610.380.380.46
movement0.320.380.490.360.640.710.730.360.440.49
p-value<0.001
region-specific eye analysis (24 landmark pairs)
rest0.240.330.410.450.220.350.390.280.240.32
movement0.230.420.510.480.280.440.520.280.280.38
p-value<0.001
region-specific nose analysis (23 landmark pairs)
rest0.260.150.320.150.540.590.520.340.350.36
movement0.180.130.310.060.590.610.630.350.390.36
p-value rest<0.0010.013<0.0010.013<0.001
p-value movement0.0030.040<0.0010.359
region-specific mouth analysis (44 landmark pairs)
rest0.360.240.320.280.550.630.60.270.290.39
movement0.30.230.340.220.610.710.710.260.340.41
p-value<0.001
Facial expressions: (1) neutral, (2) gentle eye closure, (3) forced eye closure, (4) frowning/forehead wrinkling, (5) nose wrinkling, (6) closed-mouth stretch, (7) mouth stretch with teeth visible, (8) lip pursing, and (9) downward movement of the mouth corners.
Table 4. Comparison of landmark-based methods for facial asymmetry across Stennert index grades. ScoreAI, along with range and slope, is shown for rest (Stennert 0–4) and movement (Stennert 0–6) conditions, with methods ranked by slope to indicate the strength of separation between severity levels.
Table 4. Comparison of landmark-based methods for facial asymmetry across Stennert index grades. ScoreAI, along with range and slope, is shown for rest (Stennert 0–4) and movement (Stennert 0–6) conditions, with methods ranked by slope to indicate the strength of separation between severity levels.
At Rest (Stennert 0–4)During Facial Movement (Stennert 0–6)
Ranked Method
(LM Pairs)
ScoreAI [°]RangeSlopeRanked Method
(LM Pairs)
ScoreAI [°]RangeSlope
MediaPipe (21)2.82 ± 0.593.140.76MediaPipe (21)2.24 ± 0.492.580.42
MediaPipe (50)2.47 ± 1.042.630.63MediaPipe (50)1.99 ± 0.862.140.35
Emotrics-2 (21)2.95 ± 2.221.990.47Emotrics-1 (21)2.80 ± 2.131.720.29
MediaPipe (91)1.99 ± 1.211.950.47Emotrics-2 (21)2.59 ± 1.931.680.26
Emotrics-1 (21)3.10 ± 2.341.760.41MediaPipe (91)1.64 ± 1.001.610.26
Dlib (21)2.42 ± 1.481.730.39Emotrics-3 (21)2.74 ± 2.141.400.25
Emotrics-2 (29)2.49 ± 2.111.570.37Emotrics-1 (29)2.32 ± 2.031.340.23
MediaPipe (140)1.69 ± 1.171.510.36Dlib (21)2.12 ± 1.291.500.22
Emotrics-3 (21)3.01 ± 2.361.460.35Emotrics-2 (29)2.21 ± 1.831.320.21
Dlib (29)2.13 ± 1.471.400.32MediaPipe (140)1.41 ± 0.961.290.20
Emotrics-1 (29)2.56 ± 2.241.380.32Emotrics-3 (29)2.27 ± 2.031.070.19
Emotrics-3 (29)2.48 ± 2.241.130.27Dlib (29)1.89 ± 1.281.200.18
MediaPipe (225)1.39 ± 1.071.100.26MediaPipe (225)1.18 ± 0.880.970.14
Table 5. Absolute deviations of the ScoreAI (derived from the 91-pair asymmetry feature vector) from the original images across selected artificial head rotations for all 405 datasets.
Table 5. Absolute deviations of the ScoreAI (derived from the 91-pair asymmetry feature vector) from the original images across selected artificial head rotations for all 405 datasets.
Image Rotation [°]Deviation of ScoreAI from Original Image [°]Image Rotation [°]Deviation of ScoreAI from Original Image [°]
MeanMedianMeanMedian
10.15 ± 0.140.11−10.15 ± 0.140.11
50.18 ± 0.160.14−50.18 ± 0.170.13
100.21 ± 0.190.16−100.21 ± 0.180.16
150.24 ± 0.210.18−150.23 ± 0.210.18
200.27 ± 0.240.20−200.26 ± 0.230.19
250.29 ± 0.270.22−250.29 ± 0.260.22
Mean and median deviations are shown for representative rotations.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Heinrich, A.; Volk, G.F.; Dobel, C.; Guntinas-Lichius, O. AI-Based Angle Map Analysis of Facial Asymmetry in Peripheral Facial Palsy. Bioengineering 2026, 13, 426. https://doi.org/10.3390/bioengineering13040426

AMA Style

Heinrich A, Volk GF, Dobel C, Guntinas-Lichius O. AI-Based Angle Map Analysis of Facial Asymmetry in Peripheral Facial Palsy. Bioengineering. 2026; 13(4):426. https://doi.org/10.3390/bioengineering13040426

Chicago/Turabian Style

Heinrich, Andreas, Gerd Fabian Volk, Christian Dobel, and Orlando Guntinas-Lichius. 2026. "AI-Based Angle Map Analysis of Facial Asymmetry in Peripheral Facial Palsy" Bioengineering 13, no. 4: 426. https://doi.org/10.3390/bioengineering13040426

APA Style

Heinrich, A., Volk, G. F., Dobel, C., & Guntinas-Lichius, O. (2026). AI-Based Angle Map Analysis of Facial Asymmetry in Peripheral Facial Palsy. Bioengineering, 13(4), 426. https://doi.org/10.3390/bioengineering13040426

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop