Next Article in Journal
Real-Time Radar-Based Hand Motion Recognition on FPGA Using a Hybrid Deep Learning Model
Next Article in Special Issue
Predicting Performance in Working Memory During the Waking Period by Applying a Convolutional Neural Network to EEG Data in the N-Back Task: A Pilot Study
Previous Article in Journal
HIPER-CHAD: Hybrid Integrated Prediction-Error Reconstruction-Based Anomaly Detection for Multivariate Indoor Environmental Time-Series Data
Previous Article in Special Issue
Machine Learning-Based Risk Stratification for Sudden Cardiac Death Using Clinical and Device-Derived Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Neuromuscular Assessment: Machine-Learning-Based Facial Palsy Classification Using Surface Electromyography

Department of Neurosurgery and Neurotechnology, Eberhard Karls University, Hoppe-Seyler-Straße 3, 72076 Tübingen, Germany
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 173; https://doi.org/10.3390/s26010173
Submission received: 7 November 2025 / Revised: 19 December 2025 / Accepted: 22 December 2025 / Published: 26 December 2025
(This article belongs to the Special Issue Machine Learning in Biomedical Signal Processing)

Highlights

What are the main findings?
  • Time-domain EMG features during facial movements effectively reflect facial nerve dysfunction.
  • Ensemble ML models achieved up to ~84.8% accuracy in automated HB classification from EMG data in facial palsy.
What are the implications of the main findings?
  • EMG-based ML enables objective, non-invasive assessment of facial palsy severity.
  • This method can enhance diagnostic consistency and enable longitudinal monitoring in both clinical practice and research settings.

Abstract

Facial palsy (FP) impairs voluntary control of facial muscles, resulting in facial asymmetry and difficulties in emotional expression. Traditional assessment methods to define the severity of FP (e.g., House–Brackmann score, HB) rely on visual examinations and, therefore, are highly examiner-dependent. This study proposes an alternative approach using facial surface electromyography (EMG) for automated HB prediction. Time-domain EMG features were extracted during different facial movements (i.e., smile, close eyes, and raise forehead) and analyzed through nine different machine learning (ML) models in 58 subjects (51.98 ± 1.67 years, 20 male) with variable facial nerve function (HB 1: n = 16, HB 2–3: n = 32; HB 4–6: n = 10). Model performances were evaluated based on accuracy, precision, recall, and F1-score. Among the evaluated models, ensemble-based approaches—particularly a random forest model with 100 trees and a decision tree ensemble—proved to be the most effective with classification accuracies ranging from 81.7 to 84.8% and from 81.7 to 84.7%, depending on the evaluated facial movement. The results indicate that ensemble-based ML models can reliably distinguish between different FP grades using non-invasive EMG data. The approach offers a robust alternative to subjective clinical scoring, potentially improving diagnostic consistency and supporting longitudinal monitoring in clinical and research applications.

1. Introduction

Facial palsy (FP) is the most common cranial nerve dysfunction that leads to partial or complete loss of voluntary muscle movement, causing functional and esthetic impairments that severely impact quality of life (QoL) [1]. The House–Brackmann (HB) grading system is a widely used clinical tool for assessing the severity of FP [2,3]. Despite its clinical utility, the HB score has notable limitations: it relies heavily on visual inspection and clinician judgment, making it prone to subjectivity and inter-rater variability [4].
Recent years have witnessed a surge in image- and video-based facial analysis, which enables objective quantification of visible facial motion. These approaches have been applied to measure facial symmetry and expression dynamics, showing strong potential for standardized, reproducible grading of FP [5,6]. However, these methods typically assess visible outcomes of muscle activity rather than the underlying neuromuscular function. Electromyography (EMG), by contrast, directly measures electrical signals produced during muscle activation and provides complementary functional information about neuromuscular status, including subtle asymmetries that may not be apparent on visual inspection [7]. Therefore, it plays a pivotal role in diagnosing neuromuscular disorders, evaluating muscle fatigue, monitoring rehabilitation, and other applications [8,9,10,11,12,13]. Surface EMG (sEMG) provides a non-invasive means of acquiring these signals, making it suitable for routine clinical use and longitudinal assessments [14].
Advances in machine learning (ML) and biomedical signal processing have enabled EMG-based automated classification in various neuromuscular applications, including hand-movement recognition and detection of diseases such as amyotrophic lateral sclerosis [15,16,17]. Building on these advances, the integration of EMG signal analysis with ML-based classification approaches offers considerable potential for objective assessment of FP. While initial studies primarily focused on correlating EMG data with established grading scales [18,19,20], Holze et al. [21] proposed a semi-automated sEMG-ML approach as an objective alternative to HB grading and demonstrated encouraging classification accuracies. However, as only three machine-learning approaches were evaluated in a cohort of 28 patients, the generalizability of the findings remains limited.
Against this background, our proof-of-concept study investigates the effectiveness of nine machine learning models applied to surface EMG data for the classification of FP. This comprehensive evaluation aims to advance the development of objective, data-driven methods for FP grading, which may reduce subjectivity and enhance clinical decision-making.

2. Materials and Methods

2.1. Study Cohort

This prospective pilot study enrolled 51 patients with varying grades of facial function, as well as a control group of 7 healthy individuals, to investigate various ML techniques in assessing the FP severity on sEMG. The patients were recruited during outpatient (i.e., skull base and facial nerve consultation) and inpatient stays in our department. Patients with bilateral FP or incomplete sEMG data were excluded. Detailed cohort characteristics are presented in Table 1. The study was approved by the local ethics committee of Eberhard Karls University of Tübingen and conducted in accordance with the Declaration of Helsinki. Ethical considerations included obtaining informed consent from all participants.

2.2. Data Collection and Study Design

The FP severity was classified by the HB score, which categorizes facial function into ranges from 1 (normal) to 6 (total paralysis) based on different facial movements (e.g., eye closure) [3]. The grades of palsies were assigned by two experienced examiners, and differences were resolved in consensus discussion to prevent variability across different examiners, which is a problem of the HB score [4]. Due to an imbalanced distribution of FP grades, we required stratified sampling for model training and, therefore, divided patients into three groups: HB 1 (no FP), HB 2–3 (mild FP), HB 4–6 (severe FP). A differentiation of individual classification of each possible HB grade (1–6) was not evaluated due to the limited sample size. After positioning the sEMG electrodes (Reusable Ag/AgCl snap electrodes with 4 mm diameter, Biopac Systems, Inc., Goleta, CA, USA) in a bipolar setting according to the scheme in Figure 1, each participant underwent a 30-min sEMG session, during which they performed six predefined facial movements targeting the zygomatic, orbicularis oculi and frontal facial muscles: smile strongly, smile lightly, close eyes strongly, close eyes lightly, raise forehead strongly, and raise forehead lightly. Each movement task consisted of 20 repetitions with the same movement sequence of a 3-s MOVE interval and a 4-s REST interval (longer relaxation duration allows muscle recovery, prevents fatigue, and ensures baseline stability). Stimuli were presented via a computer screen in front of the patients and controlled by the Psychopy® software (version 2024.1.1) to ensure precise movement timing (Figure 1).
Prior to starting the experiment, subjects were instructed in detail about the meaning of the instructions. They had to perform each movement with either maximum or slight effort: strong or light smiling, full eye closure with or without squinting, and forehead raising with high or low intensity. Throughout the whole session, a trained investigator was present in the room and monitored the performance of each trial to ensure that all commands were understood and executed correctly. The EMG data were streamed with the commercial Neuro Omega V2.0 recording device (Alpha Omega Engineering, Nof HaGalil, Israel) at a sampling rate of 2000 Hz and transferred and stored on an external computer.

2.3. Signal Processing and Machine Learning Models

Raw EMG signals were analyzed offline using custom-written scripts in MATLAB (MathWorks Inc., Natick, MA, USA, R2022b). During preprocessing, a notch filter (49.8–50.2 Hz) was applied to eliminate powerline interference, followed by a band pass filter (10–250 Hz) to ensure a clean and physiologically relevant signal. Subsequently, 20 time series features were extracted for each individual interval and separately for each movement type. For an in-depth description of the sEMG features analyzed, please refer to the previous publication [22].
The preprocessed feature set served as input for supervised machine learning models. To obtain a broad and representative comparison of classical and modern machine learning techniques, the following nine models were examined: (1) SVM with linear and (2) radial basis function (RBF) kernels, (3 and 4) k-nearest neighbors (k-NN; k = 3 or 5), (5) a decision tree (DT) classifier, (6 and 7) random forest models (RF; 10 and 100 trees), (8) a DT ensemble (DTE), and (9) a neural network (NN) (Figure 1). All models were implemented and tested using MATLAB-integrated functions [23]. Model performance was primarily assessed using repeated stratified trial-level k-fold cross-validation. This approach was chosen instead of strictly subject-wise splitting because the number of subjects per HB grade was both limited and imbalanced. The dataset was split into five stratified folds (k = 5), corresponding to an 80/20 train–test ratio in each iteration. In each repetition, one fold served as the test set and the remaining four folds were used for training. This procedure was repeated 10 times, resulting in 50 independent train–test splits, implemented using MATLAB’s cvpartition (k-fold = 5, Stratify = true). This evaluation strategy follows established recommendations for reducing variance and bias in model selection, where k-fold cross-validation provides a good trade-off between bias and variance of the error estimate [23,24]. Hyperparameters were optimized using grid search [25].
For multiclass classification (HB 1, HB 2–3, and HB 4–6), we implemented SVMs in MATLAB using the fitcecoc function with templateSVM as the base learner. This error-correcting output codes (ECOC) framework decomposes the three-class problem into multiple binary SVMs (one-vs-one), whose outputs are combined to produce a single multiclass prediction. Within this setup, we evaluated both a linear SVM (model 1) and an SVM with an RBF kernel (model 2). These two methods differ in the way that a linear SVM uses a linear decision boundary (hyperplane) to separate the data, while an SVM with an RBF kernel uses a non-linear decision boundary by projecting the data into a higher-dimensional space. The RBF kernel is defined as follows:
k ( x i , x j ) = e x p ( | | x i x j | | 2 2 σ 2 )
where x i and x j are data points in the feature space, and σ is a parameter that controls the width of the Gaussian kernel. We evaluated both linear and RBF SVM classifiers and standardized all input features prior to training. A subject-wise grid search indicated that the best performance was obtained with C = 10 and automatic kernel scaling. Accordingly, the final linear and RBF SVM models both used C = 10.
The k-nearest neighbors (k-NN) classifier assigns each test sample to the majority class among its k closest training samples in feature space. We implemented k-NN using MATLAB’s fitcknn. A grid search guided the final setup, and we report results for k = 3 (best setting) and k = 5 using Euclidean distance.
The classification performance of a decision tree (DT) was evaluated by using fitctree, binary splits, and Gini impurity (model 5). This algorithm recursively divides the feature space into regions that maximize the homogeneity of the target variable within each region [26]. A grid search indicated that the best cross-validated performance was achieved with MaxNumSplits = 1, MinLeafSize = 1, and the GDI criterion, and these settings were used for all reported DT results.
Random forest classification was performed by using the TreeBagger function with 10 (model 6) and 100 trees (model 7). Each tree was trained on a random subset of the data using bootstrapping [27]. The final prediction was made by averaging the predictions of all individual trees for regression tasks or by majority voting for classification tasks. Bootstrap resampling and random feature subsetting at each split increased diversity between trees and therefore reduced variance of the ensemble, which was particularly beneficial given the moderate dataset size.
A DTE method with 100 trees (model 8) for FP classification was examined using fitcensemble, in which bagging was carried out using bootstrap aggregating. This means that each sample could contain some data points more than once and not contain others at all. A grid search showed that an ensemble of 100 pruned decision trees provided the best cross-validated performance, and this configuration was used for all reported DTE.
Finally, the MATLAB function fitcnet was used to implement FP classification by a feedforward, fully connected neural network (model 9). A grid search identified the best configuration as a three-layer network with 100, 25, and 10 neurons, ReLU activations, and an initial learning rate of 0.5. This architecture was used for all reported neural-network results. To reduce overfitting, we applied L2 weight regularization and allowed a high maximum number of training iterations to ensure convergence. Class labels were represented as one-hot vectors, and the network was trained using the categorical cross-entropy loss
L = i = 1 N c = 1 C y i , c log p i , c
where y i , c is the one-hot target for sample i and class c, and p i , c is the corresponding softmax probability.

2.4. Performance Ratios and Statistics

Statistical tests were performed using MATLAB (MathWorks Inc., Natick, MA, USA, R2022b). We used the following metrics to provide a comprehensive assessment of the aforementioned classification models: accuracy (ACC), precision, recall, and F1-score, which were reported averaged across all splits. Since our analysis involves a multi-class problem (HB 1 vs. HB 2–3 vs. HB 4–6), we used a one-vs-rest strategy (OvR; also referred to as one-vs-all in MATLAB documentation) to compute per-class metrics. In this formulation, each class is evaluated by treating that class as the positive category and all remaining classes as the negative category. This is the standard multiclass evaluation design used by MATLAB’s perfcurve and rocmetrics functions. To obtain overall performance, macro-averaged metrics were calculated by taking the mean of the individual OvR results across all three classes. This macro-averaging was applied to precision, recall, F1-score, and accuracy to ensure that each class contributed equally to the overall summary, regardless of class size. Furthermore, we conducted a binary classification subanalysis comparing HB 1 of healthy subjects and HB 1 of patients. The same repeated stratified 5-fold cross-validation with 10 repetitions was applied within this restricted dataset, meaning that training and test sets were re-partitioned exclusively within the HB 1 group. Stratification ensured that each fold preserved the proportion of healthy versus postoperative HB 1 cases, thereby preventing bias from potential imbalances. Finally, receiver operating characteristic (ROC) curves were generated to illustrate the model’s discriminatory ability by plotting the sensitivity (=recall) against the specificity. The area under the curve (AUC) of the ROC curve quantified the overall performance of the classifier, with a value closer to 1 indicating superior classification ability.

3. Results

3.1. Clinical Characteristics

This prospective study included 58 subjects, enrolling 42 patients with facial palsy HB grade 2–6, while 16 subjects had no FP (HB 1). The majority of FP were iatrogenic (39/42, 92.9%). The average duration of the FP was 25.24 ± 75.46 months. Surface EMG recordings were successfully conducted in all participants without any reported complications or adverse events.

3.2. ML-Based Facial Function Classification

Across the six facial movements, we compared nine ML models with respect to their ability to classify the degree of facial palsy (no FP, mild FP, severe FP) from sEMG. The random forest with 100 trees (RF100, model 7) and the bagged decision-tree ensemble (DTE, model 8) achieved the strongest and most consistent performance across all six movements. For RF100, accuracy ranged from 81.7 to 84.8%, recall from 81.0 to 85.0%, precision from 78.3 to 81.7%, and F1-score from 79.3 to 82.9%. The DTE achieved very similar values, with accuracy between 81.7 and 84.7%, recall between 80.9 and 84.8%, precision between 78.4 and 82.2%, and F1-score between 79.4 and 82.9%. Performance was comparable for smiling, blinking, and forehead movements, with slightly higher accuracies for the strong movements compared to their light counterparts. Detailed results for all nine models and six movements are reported in Table 2 and Figure 2.
The ROC analysis confirmed this advantage of tree-based ensembles. For RF100, AUC values across the six movements were approximately 0.9–0.95 for HB 1, 0.92–0.94 for HB 2–3, and 0.95–0.97 for HB 4–6. The DTE showed similarly high discrimination with AUCs of about 0.91–0.95 (HB 1), 0.92–0.95 (HB 2–3), and 0.93–0.98 (HB 4–6) (Figure 3). These findings indicate that tree-based ensemble classifiers are particularly well-suited for sEMG-based FP classification.
Conversely, the linear SVM (model 1) performed the worst among all ML models, with a very poor classification performance compared to the other models. This suggests that the linear SVM struggled to effectively capture the complex patterns within the EMG data.

3.3. Differences in HB 1 Facial Function in Healthy Subjects and Patients

In a dedicated subcohort analysis, we examined whether patients with clinically normal facial function after surgery (HB 1 patient) could be distinguished from healthy controls (HB 1 healthy) using the RF100 and DTE models. The results demonstrated that the two groups could be reliably distinguished based on sEMG signals, with classification accuracies ranging between 85.0% and 97.2% using the RF100 model, while the DTE reached 84.0–97.2%, with the best performance consistently observed for close eyes lightly (Table 3, Figure 4). Furthermore, ROC analyses showed excellent separability for both models, with AUC values of 0.93–0.99 for HB 1 healthy subjects and 0.93–0.99 for HB 1 patients across movements. Together, these results indicate that, despite a normal HB 1 grading on clinical examination, postoperative patients exhibit systematic differences in facial EMG activity compared with non-operated individuals.

4. Discussion

In clinical practice, the classification of FP traditionally relies on visual inspection and manual grading systems such as the House–Brackmann (HB) scale, which are prone to significant inter-rater variability [4]. To address this limitation, the use of photographic and video-based visual assessments has been proposed [6,28,29]. However, image-based classification methods quantify only the externally visible manifestations of FP and do not provide direct information about the underlying neuromuscular activation. By contrast, sEMG captures muscle activity with high temporal sensitivity, allowing the detection of subtle differences in activation amplitude, timing, and co-activation patterns that remain below the threshold of visual inspection [18,30]. Therefore, although sEMG does not yield mechanistic insight into the pathophysiology of FP, its sensitivity to functional changes offers complementary information that may help detect subclinical dysfunction or residual activation capacity.
The potential of sEMG in FP assessment has been explored in several promising studies, supporting its physiological validity [19,20,21,31,32]. Franz et al. [31] examined patients with FP following vestibular schwannoma surgery and identified significantly greater signal variability on the affected side, with a strong correlation between sEMG measurements and the Sunnybrook grading scale. Ryu et al. [20] demonstrated that sEMG not only correlated with clinical grading systems (HB, Sunnybrook, and Yanagihara) but also showed alignment with nerve conduction studies (NCS), thus supporting its physiological validity across multiple modalities. Additionally, a semi-automated sEMG- and ML-based system developed by Holze et al. [21] demonstrated promising classification results and represented an objective alternative to the subjective HB scale. Together, these findings support the utility of sEMG as a sensitive physiological biomarker of facial nerve function.
Our findings further build on this evidence. Machine-learning models trained on sEMG data reliably distinguished among clinically relevant severity levels (HB 1, 2–3, and 4–6). Ensemble-based methods showed robust performance, which was consistent with their established strength in handling noisy biomedical signals and complex non-linear relationships. In contrast, SVMs demonstrated comparatively lower performance in our dataset. Importantly, sEMG enabled discrimination between HB 1 in healthy individuals and HB 1 in postoperative vestibular schwannoma (VS) patients, despite the absence of visible facial asymmetry. This represents a functional advantage of sEMG-based assessment: it can uncover early or subclinical physiological differences that image-based methods are generally less sensitive to. While sEMG cannot directly characterize the underlying neural pathophysiology, it provides sensitive markers of neuromuscular integrity that may indicate early dysfunction or residual activation capacity. However, the sEMG-based classifiers employed in our study should be regarded as complementary functional information to imaging modalities and clinical grading systems, rather than replacements.
The strong performance of RF models in our study aligns with a growing body of work demonstrating the suitability of tree-based algorithms for EMG-based classification tasks: Holze et al. [21] demonstrated that logistic regression and k-NN outperformed SVM in both the binary classification task (‘normal’ vs. ‘impaired facial function’) and the multiclass classification of HB grades 1, 2, and 3. Beyond the context of FP, numerous biomedical applications have shown that combining ML with EMG signals yields high discriminative performance. For example, Yuan et al. [33] found that RFs outperformed artificial NNs and SVMs for sEMG-based Chinese sign language classification. Gokgoz et al. [34] demonstrated that among multiple tree-based classifiers, random forests best differentiated EMG patterns in myopathy and amyotrophic lateral sclerosis (ALS). These consistent results across neuromuscular contexts underscore the resilience, noise tolerance, and generalization capability of RF models—characteristics that align with their strong performance in our FP dataset.
Beyond diagnostic classification, the physiological sensitivity of sEMG offers several potential therapeutic applications in the FP classification. The capacity to detect nuanced discrepancies—like the difference between HB 1 healthy subjects and HB 1 patients—has meaningful clinical implications. sEMG-based monitoring may help detect early neuromuscular deterioration, track recovery, or refine prognostic estimations. Biofeedback or rehabilitation systems could leverage sEMG to identify residual muscle activity—even when clinically invisible—and, therefore, enable patients to selectively activate weak muscles and optimize targeted exercises. Continuous sEMG monitoring may allow clinicians to track recovery progress objectively, adjust therapy intensity, and identify early signs of neuromuscular deterioration. In this context, previous studies have already demonstrated the utility of invasive EMG in outcome prediction [35,36]. For instance, Rampp et al. [37] developed a predictive model integrating intraoperative EMG data with preoperative HB grade and tumor size to estimate postoperative facial nerve function. Their findings highlight the multifactorial nature of FP prognosis, with tumor size being a particularly influential factor in VS cases. Building upon this, ML may offer further support in therapeutic contexts by enabling the integration of physiological data into predictive or adaptive treatment frameworks. An example of this potential is provided by Zhu et al. [38], who combined sEMG signals with electroacupuncture in a deep-learning model, illustrating how advanced signal processing can be effectively paired with innovative treatment modalities.

Limitations

An important limitation of the study is the modest sample size, which led to class imbalance, particularly in the higher HB grades (HB 4–6). Class imbalance may bias ML models toward the majority class, leading to poor detection of infrequent but clinically relevant outcomes. Collapsing HB grades into three categories improved model stability but reduced granularity, and therefore, conclusions regarding fine-graded severity distinctions should be interpreted with caution. Furthermore, the healthy control group was younger than the patient cohort, and most FP cases were postoperative, meaning that age-related physiological differences and postoperative tissue changes could influence sEMG signals independently of nerve function. In addition, natural inter-individual asymmetry—such as variations in muscle mass, fiber orientation, or facial dominance—may affect measurements. Although recordings were standardized and supervised, sEMG remains susceptible to artifacts from movement variability, fatigue, environmental noise, or technical factors. However, the demographic and clinical characteristics of the cohort are representative of a neurosurgical FP population [39,40,41]. Furthermore, stratified sampling for model training, as well as extensive preprocessing, were performed, and all recording sessions were supervised by trained investigators. Nevertheless, normalization and preprocessing mitigate these effects but cannot eliminate them entirely. A further limitation of the present study is that the use of ML methods means that it is not possible to say exactly which of the 20 time series features were used by the models, i.e., which ones are the best. However, a previous study of ours investigated this and showed that 5/20 time series features seem to be the most appropriate [22]. Finally, long-term outcome data were not available, preventing the assessment of prognosis or recovery prediction. Future longitudinal studies with age-matched controls, non-surgical FP, and more balanced class distributions are needed to confirm generalizability and to explore the predictive potential of sEMG-based approaches.

5. Conclusions

This study provides an important contribution to the development of objective and automated systems for assessing facial palsy. The results highlight the potential of sEMG data, particularly in combination with ensemble-based machine learning models such as random forests, to detect subtle neuromuscular differences beyond what is visible in clinical inspection. Our findings suggest that sEMG-based approaches may support more precise grading, early detection, and individualized prognosis, offering a valuable complement to traditional image-based methods. It paves the way for real-time feedback systems that could aid healthcare professionals in monitoring patient progress, contributing to more tailored and effective treatment plans.

Author Contributions

K.M., I.M. and G.N. contributed to the conception and design of the study. K.M., I.M., A.P. and S.R. contributed to the acquisition of data. I.M. and K.M. contributed to the analysis, as well as the literature search and drafting of the manuscript. G.N., A.P., A.S., S.R., M.G. and M.T. contributed to the manuscript review. All authors have read and agreed to the published version of the manuscript.

Funding

The data was collected as part of a project on biofeedback for facial palsy funded by the Federal Ministry for Research, Technology, and Aeronautics of Germany (BMFTR, funding number: 16SV9236). The BMFTR did not influence data or reporting.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of the University of Tuebingen (protocol code: 326/2024BO1).

Informed Consent Statement

All participants gave written informed consent.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to ethical reasons.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest. Therefore, none of the authors has any conflict of interest to disclose.

Abbreviations

The following abbreviations are used in this manuscript:
ACCAccuracy
ALSAmyotrophic lateral sclerosis
ANNArtificial neural network
AUCArea under the curve
CNNConvolutional neural networks
DTDecision trees
DTEDecision tree ensemble
ECOCError-correcting output codes
EMGElectromyography
FPFacial palsy
HBHouse–Brackmann score
k-NNk-nearest neighbors
MLMachine learning
MNLRMultinomial logistic regression
NCSNerve conduction studies
NNNeural network
QoLQuality of life
OvROne-vs-rest
RBFRadial basis function
RFRandom forests
ROCReceiver operating characteristic
sEMGSurface electromyography
SVMSupport vector machines

References

  1. Walker, N.R.; Mistry, R.K.; Mazzoni, T. Facial Nerve Palsy. In Emergency Neurology, 2nd ed.; Springer: Cham, Switzerland, 2023; pp. 141–150. [Google Scholar] [CrossRef]
  2. Reitzen, S.D.; Babb, J.S.; Lalwani, A.K. Significance and Reliability of the House-Brackmann Grading System for Regional Facial Nerve Function. Otolaryngol.—Head Neck Surg. 2009, 140, 154–158. [Google Scholar] [CrossRef]
  3. House, J.W.; Brackmann, D.E. Facial Nerve Grading System. Otolaryngol.—Head Neck Surg. 1985, 93, 146–147. [Google Scholar] [CrossRef]
  4. Scheller, C.; Wienke, A.; Tatagiba, M.; Gharabaghi, A.; Ramina, K.F.; Scheller, K.; Prell, J.; Zenk, J.; Ganslandt, O.; Bischoff, B.; et al. Interobserver Variability of the House-Brackmann Facial Nerve Grading System for the Analysis of a Randomized Multi-Center Phase III Trial. Acta Neurochir. 2017, 159, 733–738. [Google Scholar] [CrossRef]
  5. Lugrís, U.; Pérez-Soto, M.; Michaud, F.; Cuadrado, J. Human Motion Capture, Reconstruction, and Musculoskeletal Analysis in Real Time. Multibody Syst. Dyn. 2024, 60, 3–25. [Google Scholar] [CrossRef]
  6. Eisenhardt, S.; Volk, G.F.; Rozen, S.; Parra-Dominguez, G.S.; Garcia-Capulin, C.H.; Sanchez-Yanez, R.E. Automatic Facial Palsy Diagnosis as a Classification Problem Using Regional Information Extracted from a Photograph. Diagnostics 2022, 12, 1528. [Google Scholar] [CrossRef] [PubMed]
  7. Jiang, M.; Rahmani, A.M.; Westerlund, T.; Liljeberg, P.; Tenhunen, H. Facial Expression Recognition with SEMG Method. In Proceedings of the 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing, Liverpool, UK, 26–28 October 2015; pp. 981–988. [Google Scholar] [CrossRef]
  8. Hogrel, J.Y. Clinical Applications of Surface Electromyography in Neuromuscular Disorders. Neurophysiol. Clin./Clin. Neurophysiol. 2005, 35, 59–71. [Google Scholar] [CrossRef]
  9. Goen, A. Classification of EMG Signals for Assessment of Neuromuscular Disorders. Int. J. Electron. Electr. Eng. 2014, 2, 242–248. [Google Scholar] [CrossRef]
  10. Sun, J.; Liu, G.; Sun, Y.; Lin, K.; Zhou, Z.; Cai, J. Application of Surface Electromyography in Exercise Fatigue: A Review. Front. Syst. Neurosci. 2022, 16, 893275. [Google Scholar] [CrossRef] [PubMed]
  11. Cifrek, M.; Medved, V.; Tonković, S.; Ostojić, S. Surface EMG Based Muscle Fatigue Evaluation in Biomechanics. Clin. Biomech. 2009, 24, 327–340. [Google Scholar] [CrossRef]
  12. Mokri, C.; Bamdad, M.; Abolghasemi, V. Muscle Force Estimation from Lower Limb EMG Signals Using Novel Optimised Machine Learning Techniques. Med. Biol. Eng. Comput. 2022, 60, 683–699. [Google Scholar] [CrossRef]
  13. Al-Ayyad, M.; Owida, H.A.; De Fazio, R.; Al-Naami, B.; Visconti, P. Electromyography Monitoring Systems in Rehabilitation: A Review of Clinical Applications, Wearable Devices and Signal Acquisition Methodologies. Electronics 2023, 12, 1520. [Google Scholar] [CrossRef]
  14. Wu, J.; Li, X.; Liu, W.; Wang, Z.J. SEMG Signal Processing Methods: A Review. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2019; Volume 1237, p. 032008. [Google Scholar] [CrossRef]
  15. Zhang, X.; Barkhaus, P.E.; Rymer, W.Z.; Zhou, P. Machine Learning for Supporting Diagnosis of Amyotrophic Lateral Sclerosis Using Surface Electromyogram. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 96–103. [Google Scholar] [CrossRef] [PubMed]
  16. Tannemaat, M.R.; Kefalas, M.; Geraedts, V.J.; Remijn-Nelissen, L.; Verschuuren, A.J.M.; Koch, M.; Kononova, A.V.; Wang, H.; Bäck, T.H.W. Distinguishing Normal, Neuropathic and Myopathic EMG with an Automated Machine Learning Approach. Clin. Neurophysiol. 2023, 146, 49–54. [Google Scholar] [CrossRef]
  17. Reffad, A.; Mebarkia, K. Motor Imagery Hand Movements Recognition Using SVM Classifier and Genetic Algorithm Optimization. In Proceedings of the 2022 19th IEEE International Multi-Conference on Systems, Signals and Devices, SSD, Setif, Algeria, 6–10 May 2022; Volume 2022, pp. 1125–1129. [Google Scholar] [CrossRef]
  18. Franz, L.; de Filippis, C.; Daloiso, A.; Biancoli, E.; Iannacone, F.P.; Cazzador, D.; Tealdo, G.; Marioni, G.; Nicolai, P.; Zanoletti, E. Facial Surface Electromyography: A Systematic Review on the State of the Art and Current Perspectives. Am. J. Otolaryngol.—Head Neck Med. Surg. 2024, 45, 104041. [Google Scholar] [CrossRef]
  19. Kim, G.H.; Park, J.H.; Kim, T.K.; Lee, E.J.; Jung, S.E.; Seo, J.C.; Kim, C.H.; Choi, Y.M.; Yoon, H.M. Correlation Between Accompanying Symptoms of Facial Nerve Palsy, Clinical Assessment Scales and Surface Electromyography. J. Acupunct. Res. 2022, 39, 297–303. [Google Scholar] [CrossRef]
  20. Ryu, H.M.; Lee, S.J.; Park, E.J.; Kim, S.G.; Kim, K.H.; Choi, Y.M.; Kim, J.U.; Song, B.Y.; Kim, C.H.; Yoon, H.M.; et al. Study on the Validity of Surface Electromyography as Assessment Tools for Facial Nerve Palsy. J. Pharmacopunct. 2018, 21, 258–267. [Google Scholar] [CrossRef] [PubMed]
  21. Holze, M.; Rensch, L.; Prell, J.; Scheller, C.; Simmermacher, S.; Scheer, M.; Strauss, C.; Rampp, S. Learning from EMG: Semi-Automated Grading of Facial Nerve Function. J. Clin. Monit. Comput. 2022, 36, 1509–1517. [Google Scholar] [CrossRef] [PubMed]
  22. Manzoor, I.; Popescu, A.; Stark, A.; Gorbachuk, M.; Spolaore, A.; Tatagiba, M.; Naros, G.; Machetanz, K. Surface Electromyographic Features for Severity Classification in Facial Palsy: Insights from a German Cohort and Implications for Future Biofeedback Use. Sensors 2025, 25, 2949. [Google Scholar] [CrossRef]
  23. Yerramilli, S.; Apley, D.W. Fractional Cross-Validation for Optimizing Hyperparameters of Supervised Learning Algorithms. Technometrics 2025, 67, 683–692. [Google Scholar] [CrossRef]
  24. Kohavi, R. A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection. In Proceedings of the 40th International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 20–25 August 1995. [Google Scholar]
  25. Pathak, A.K.; Chaubey, M.; Gupta, M. Randomized-Grid Search for Hyperparameter Tuning in Decision Tree Model to Improve Performance of Cardiovascular Disease Classification. arXiv 2024, arXiv:2411.18234. [Google Scholar] [CrossRef]
  26. Quinlan, J.R. Induction of Decision Trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  27. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  28. Sajid, M.; Shafique, T.; Baig, M.; Riaz, I.; Amin, S.; Manzoor, S. Automatic Grading of Palsy Using Asymmetrical Facial Features: A Study Complemented by New Solutions. Symmetry 2018, 10, 242. [Google Scholar] [CrossRef]
  29. Zhuang, Y.; McDonald, M.; Uribe, O.; Yin, X.; Parikh, D.; Southerland, A.M.; Rohde, G.K. Facial Weakness Analysis and Quantification of Static Images. IEEE J. Biomed. Health Inform. 2020, 24, 2260–2267. [Google Scholar] [CrossRef] [PubMed]
  30. Bertozzi, N.; Bianchi, B.; Salvagni, L.; Raposio, E. Activity Evaluation of Facial Muscles by Surface Electromyography. Plast. Reconstr. Surg. Glob. Open 2020, 8, E3081. [Google Scholar] [CrossRef] [PubMed]
  31. Franz, L.; Marioni, G.; Daloiso, A.; Biancoli, E.; Tealdo, G.; Cazzador, D.; Nicolai, P.; de Filippis, C.; Zanoletti, E. Facial Surface Electromyography: A Novel Approach to Facial Nerve Functional Evaluation after Vestibular Schwannoma Surgery. J. Clin. Med. 2024, 13, 590. [Google Scholar] [CrossRef]
  32. Machetanz, K.; Grimm, F.; Schäfer, R.; Trakolis, L.; Hurth, H.; Haas, P.; Gharabaghi, A.; Tatagiba, M.; Naros, G. Design and Evaluation of a Custom-Made Electromyographic Biofeedback System for Facial Rehabilitation. Front. Neurosci. 2022, 16, 666173. [Google Scholar] [CrossRef]
  33. Yuan, S.; Wang, Y.; Wang, X.; Deng, H.; Sun, S.; Wang, H.; Huang, P.; Li, G. Chinese Sign Language Alphabet Recognition Based on Random Forest Algorithm. In Proceedings of the 2020 IEEE International Workshop on Metrology for Industry 4.0 and IoT, MetroInd 4.0 and IoT 2020—Proceedings, Roma, Italy, 3–5 June 2020; pp. 340–343. [Google Scholar] [CrossRef]
  34. Gokgoz, E.; Subasi, A. Comparison of Decision Tree Algorithms for EMG Signal Classification Using DWT. Biomed. Signal Process. Control 2015, 18, 138–144. [Google Scholar] [CrossRef]
  35. Ozgur, A.; Semai, B.; Hidir, U.U.; Mehmet Fatih, O.; Tayfun, K.; Zeki, O. Which Electrophysiological Measure Is Appropriate in Predicting Prognosis of Facial Paralysis? Clin. Neurol. Neurosurg. 2010, 112, 844–848. [Google Scholar] [CrossRef]
  36. Petrides, G.A.; Hayler, R.; Lee, J.W.; Jankelowitz, S.; Low, T.H. Electromyography in the Prognostication of Recovery in Patients with Acute Peripheral Facial Nerve Palsy: A Systematic Review. Clin. Otolaryngol. 2023, 48, 563–575. [Google Scholar] [CrossRef]
  37. Rampp, S.; Holze, M.; Scheller, C.; Strauss, C.; Prell, J. Neural Networks for Estimation of Facial Palsy after Vestibular Schwannoma Surgery. J. Clin. Monit. Comput. 2023, 37, 575–583. [Google Scholar] [CrossRef] [PubMed]
  38. Zhu, P.; Wang, H.; Zhang, L.; Jiang, X. Deep Learning-Based Surface Nerve Electromyography Data of E-Health Electroacupuncture in Treatment of Peripheral Facial Paralysis. Comput. Math. Methods Med. 2022, 2022, 8436741. [Google Scholar] [CrossRef] [PubMed]
  39. Machetanz, K.; Roegele, M.; Liebsch, M.; Oberle, L.; Weinbrenner, E.; Gorbachuk, M.; Wang, S.S.; Tatagiba, M.; Naros, G. Predictive Value of Facial Motor–Evoked Potential and Electromyography for Facial Motor Function in Vestibular Schwannoma Surgery. Acta Neurochir. 2024, 166, 23. [Google Scholar] [CrossRef]
  40. Rinaldi, V.; Casale, M.; Bressi, F.; Potena, M.; Vesperini, E.; De Franco, A.; Silvestri, S.; Zini, C.; Salvinelli, F. Facial Nerve Outcome after Vestibular Schwannoma Surgery: Our Experience[*]. J. Neurol. Surg. Part B Skull Base 2012, 73, 21–27. [Google Scholar] [CrossRef]
  41. Ren, Y.; MacDonald, B.V.; Tawfik, K.O.; Schwartz, M.S.; Friedman, R.A. Clinical Predictors of Facial Nerve Outcomes After Surgical Resection of Vestibular Schwannoma. Otolaryngol.—Head Neck Surg. 2021, 164, 1085–1093. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Study design and data analysis pipeline demonstrating the electrode mounting (top left), time series features of surface electromyography (top right), machine learning models (middle), and classification of facial palsy severity (bottom). AAC: Average amplitude change, DASDV: Difference absolute standard value, HB: House–Brackmann score, IAV: Integral of absolute value, iEMG: Integrated EMG, KURT: Kurtosis, LOG: Log detector, MAV: Mean absolute value, MAX: Maximal amplitude, MMAV1: Modified mean absolute value 1, MMAV2: Modified mean absolute value 2, RMS: Root mean square, SKEW: Skewness, SSC: Slope sign change, SSI: Simple square integral, STD: Standard deviation, SVM: Support vector machine, VAR: Variance, VO: V-order, WAMP: Willison amplitude, WL: Waveform length, ZC: Zero crossing.
Figure 1. Study design and data analysis pipeline demonstrating the electrode mounting (top left), time series features of surface electromyography (top right), machine learning models (middle), and classification of facial palsy severity (bottom). AAC: Average amplitude change, DASDV: Difference absolute standard value, HB: House–Brackmann score, IAV: Integral of absolute value, iEMG: Integrated EMG, KURT: Kurtosis, LOG: Log detector, MAV: Mean absolute value, MAX: Maximal amplitude, MMAV1: Modified mean absolute value 1, MMAV2: Modified mean absolute value 2, RMS: Root mean square, SKEW: Skewness, SSC: Slope sign change, SSI: Simple square integral, STD: Standard deviation, SVM: Support vector machine, VAR: Variance, VO: V-order, WAMP: Willison amplitude, WL: Waveform length, ZC: Zero crossing.
Sensors 26 00173 g001
Figure 2. Graphical representation of the performance metrics across all movement types (averaged across all splits). Subplots demonstrate (A) the accuracy, (B) precision, (C) Recall and (D) the F1 score. DT: Decision tree, k-NN: k-nearest Neighbors, RF: Random forest, SVM: Support vector machine.
Figure 2. Graphical representation of the performance metrics across all movement types (averaged across all splits). Subplots demonstrate (A) the accuracy, (B) precision, (C) Recall and (D) the F1 score. DT: Decision tree, k-NN: k-nearest Neighbors, RF: Random forest, SVM: Support vector machine.
Sensors 26 00173 g002
Figure 3. Receiver operating characteristic (ROC) curves of the decision tree ensemble and random forest (100 trees) model differentiating HB 1, HB 2–3, and HB 4–6. Subplots (AC) demonstrate the ROC curves of DTE for different movements, while subplots (DF) demonstrate the ROC curves of RF(100). DTE: Decision tree ensemble, HB: House–Brackmann score, RF: Random forest.
Figure 3. Receiver operating characteristic (ROC) curves of the decision tree ensemble and random forest (100 trees) model differentiating HB 1, HB 2–3, and HB 4–6. Subplots (AC) demonstrate the ROC curves of DTE for different movements, while subplots (DF) demonstrate the ROC curves of RF(100). DTE: Decision tree ensemble, HB: House–Brackmann score, RF: Random forest.
Sensors 26 00173 g003
Figure 4. Receiver operating characteristic (ROC) curves of the decision tree ensemble (A–C) and random forest (100 trees, D–F) model differentiating HB 1 of healthy subjects and patients. DTE: Decision tree ensemble, HB: House–Brackmann score, RF: Random forest.
Figure 4. Receiver operating characteristic (ROC) curves of the decision tree ensemble (A–C) and random forest (100 trees, D–F) model differentiating HB 1 of healthy subjects and patients. DTE: Decision tree ensemble, HB: House–Brackmann score, RF: Random forest.
Sensors 26 00173 g004
Table 1. Cohort characteristics.
Table 1. Cohort characteristics.
Total
N = 58
Patients
N = 51
Healthy Subjects
N = 7
Age51.98 ± 1.67 years54.65 ± 12.48 years32.57 ± 15.67 years
Gender
Male20 (34.5%)17 (33.3%)3 (42.9%)
Female38 (65.5%)34 (66.7%)4 (57.1%)
HB Grade
116 (27.6%)9 (17.6%)7 (100%)
210 (17.2%)10 (19.6%)0 (0%)
322 (37.9%)22 (43.1%)0 (0%)
44 (6.9%)4 (7.8%)0 (0%)
54 (6.9%)4 (7.8%)0 (0%)
62 (3.4%)2 (3.9%)0 (0%)
Side of FP or Surgery
Right22 (37.9%)22 (43.1%)0 (0%)
Left29 (50.0%)29 (56.9%)0 (0%)
No FP/surgery7 (12.1%)0 (0%)7 (100%)
Etiology of FP
Idiopathic2 (3.4%)2 (2.0%)0 (0%)
Tumor1 (1.7%)1 (2.0%)0 (0%)
Iatrogenic48 (82.8%)48 (94.1%)0 (0%)
No FP/surgery7 (12.1%)0 (0%)7 (100%)
HB: House–Brackmann; FP: Facial palsy. Surgery always means posterior skull base surgery (i.e., VS surgery or meningioma).
Table 2. Performance metrics across all movement types and machine learning models differentiating HB 1, HB 2–3, and HB 4–6.
Table 2. Performance metrics across all movement types and machine learning models differentiating HB 1, HB 2–3, and HB 4–6.
MovementMetric (%)SVM (Lin.)SVM (RBF)k-NN (k = 3)k-NN (k = 5)RF (10)RF (100)DTDT EnsembleNN
Smile StrongAccuracy56.8 ± 1.783.4 ± 1.880.4 ± 2.579.6 ± 3.181.6 ± 1.983.7 ± 2.777.2 ± 3.283.7 ± 2.180.9 ± 3.1
Precision38.7 ± 2.480.1 ± 2.778.0 ± 3.177.4 ± 3.878.8 ± 2.680.7 ± 3.374.7 ± 3.780.6 ± 2.678.6 ± 4.5
Recall50.2 ± 1.483.7 ± 2.878.7 ± 3.277.9 ± 3.780.7 ± 2.283.1 ± 3.375.0 ± 3.583.0 ± 2.478.8 ± 3.5
F1-score35.1 ± 4.281.6 ± 2.478.2 ± 3.077.5 ± 3.779.5 ± 2.381.7 ± 3.274.7 ± 3.481.6 ± 2.378.4 ± 4.1
Smile LightAccuracy58.1 ± 1.580.1 ± 2.379. ± 2.777.7 ± 3.180.9 ± 2.282.9 ± 2.075.3 ± 3.182.9 ± 2.278.3 ± 2.9
Precision39.9 ± 2.175.5 ± 3.174.6 ± 3.072.63 ± 3.477.7 ± 2.779.7 ± 2.272.3 ± 3.7579.8 ± 3.076.2 ± 3.4
Recall61.8 ± 8.479.8 ± 2.977.5 ± 3.276.12 ± 3.980.0 ± 2.582.3 ± 2.672.9 ± 3.782.4 ± 2.576.4 ± 3.6
F1-score37.1 ± 3.777.1 ± 2.975.7 ± 3.073.9 ± 3.678.6 ± 2.480.8 ± 2.272.3 ± 3.680.7 ± 2.676.2 ± 3.4
Close Eyes StrongAccuracy69.2 ± 2.282.2 ± 2.380.3 ± 2.479.9 ± 2.582.7 ± 2.384.8 ± 2.177.5 ± 2.884.7 ± 1.880.2 ± 2.4
Precision57.0 ± 3.079.2 ± 3.277.8 ± 2.975.5 ± 3.179.5 ± 3.181.6 ± 2.975.7 ± 3.281.2 ± 2.379.0 ± 2.8
Recall72.7 ± 4.683.2 ± 2.580.2 ± 3.080.7 ± 3.182.3 ± 2.785 ± 2.876.1 ± 3.284.8 ± 2.278.8 ± 2.7
F1-score59.7 ± 3.580.1 ± 2.778.7 ± 2.877.5 ± 3.080.6 ± 2.782.9 ± 2.675.7 ± 3.082.9 ± 2.178.7 ± 2.6
Close Eyes LightAccuracy59.1 ± 1.576.1 ± 2.375.7 ± 3.075.1 ± 2.880.7 ± 2.182.8 ± 2.377.4 ± 2.982.9 ± 2.275.9 ± 3.0
Precision40.3 ± 1.970.1 ± 371.4 ± 3.969.5 ± 3.576.4 ± 2.978.5 ± 3.074.3 ± 3.478.5 ± 2.972.9 ± 3.9
Recall69.7 ± 1.175.6 ± 3.472.9 ± 3.873.2 ± 3.678.9 ± 2.981.5 ± 2.974.9 ± 3.481.8 ± 2.872.8 ± 3.6
F1-score37.3 ± 3.272.0 ± 3.071.9 ± 3.870.8 ± 3.577.0 ± 2.7379.7 ± 2.974.4 ± 3.079.7 ± 2.672.6 ± 3.8
Raise Forehead StrongAccuracy64.6 ± 2.776.8 ± 2.675.1 ± 2.674.8 ± 2.579.1 ± 2.981.7 ± 2.773.3 ± 3.081.7 ± 1.875.5 ± 2.6
Precision53.0 ± 3.972.3 ± 2.972.7 ± 3.171.0 ± 3.375.8 ± 3.378.3 ± 3.571.1 ± 3.478.4 ± 2.573.2 ± 2.7
Recall72.2 ± 5.577.6 ± 3.373.1 ± 2.973.0 ± 2.978.3 ± 3.381.0 ± 3.071.2 ± 3.580.9 ± 2.273.2 ± 3.0
F1-score54.8 ± 4.774.2 ± 3.072.8 ± 2.971.8 ± 3.076.8 ± 3.279.3 ± 3.271.0 ± 3.379.4 ± 2.373.0 ± 2.8
Raise Forehead LightAccuracy55.6 ± 1.776.1 ± 2.575.1 ± 2.874.0 ± 2.480.1 ± 2.483.3 ± 2.373.6 ± 3.083.6 ± 2.474.7 ± 2.5
Precision35.7 ± 3.872.2 ± 3.371.5 ± 3.569.2 ± 3.278.6 ± 3.281.7 ± 2.672.0 ± 3.782.2 ± 2.773.0 ± 3.1
Recall34.1 ± 1.877.6 ± 3.174.0 ± 3.373.5 ± 3.179.9 ± 2.783.3 ± 2.672.6 ± 3.483.4 ± 2.973.3 ± 2.7
F1-score28.1 ± 6.674.0 ± 3.172.4 ± 3.370.8 ± 3.179.0 ± 2.882.3 ± 2.572.1 ± 3.582.6 ± 2.773.0 ± 2.7
DT: Decision tree, k-NN: k-nearest Neighbors, RF: Random forest, SVM: Support vector machine.
Table 3. Performance metrics of the decision tree (DT) ensemble and the random forest (100 trees, RF) model, differentiating HB 1 of healthy subjects and patients.
Table 3. Performance metrics of the decision tree (DT) ensemble and the random forest (100 trees, RF) model, differentiating HB 1 of healthy subjects and patients.
MetricSmile StrongSmile LightClose Eyes StrongClose Eyes LightRaise Forehead StrongRaise Forehead Light
RF (100)Accuracy (%)86.9 ± 4.388.6 ± 3.590.0 ± 3.697.2 ± 2.285.0 ± 3.985.1 ± 3.6
Precision (%)86.6 ± 4.688.2 ± 3.789.6 ± 3.897.0 ± 2.284.5 ± 4.084.9 ± 3.8
Recall (%)87.2 ± 4.388.8 ± 3.690.4 ± 3.597.3 ± 2.185.4 ± 4.085.2 ± 3.6
F1-score86.6 ± 4.588.3 ± 3.789.8 ± 3.797.1 ± 2.284.7 ± 4.084.8 ± 3.8
AUC HB 1 healthy0.9390.9540.9620.9900.9280.927
AUC HB 1 patient0.9390.9540.9620.9900.9280.927
DT EnsembleAccuracy (%)86.1 ± 5.288.5 ± 3.789.9 ± 3.697.2 ± 2.184.0 ± 4.885.6 ± 4.5
Precision (%)85.7 ± 5.388.1 ± 3.989.4 ± 3.997.0 ± 2.383.5 ± 4.685.5 ± 4.6
Recall (%)86.5 ± 5.488.6 ± 3.890.4 ± 3.497.4 ± 1.984.7 ± 4.885.7 ± 4.4
F1-score85.8 ± 5.388.2 ± 3.889.6 ± 3.897.1 ± 2.183.6 ± 4.885.4 ± 4.6
AUC HB 1 healthy0.9300.9520.9630.9920.9260.927
AUC HB 1 patient0.9300.9520.9610.9920.9260.929
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Manzoor, I.; Popescu, A.; Ricchizzi, S.; Spolaore, A.; Gorbachuk, M.; Tatagiba, M.; Naros, G.; Machetanz, K. Automated Neuromuscular Assessment: Machine-Learning-Based Facial Palsy Classification Using Surface Electromyography. Sensors 2026, 26, 173. https://doi.org/10.3390/s26010173

AMA Style

Manzoor I, Popescu A, Ricchizzi S, Spolaore A, Gorbachuk M, Tatagiba M, Naros G, Machetanz K. Automated Neuromuscular Assessment: Machine-Learning-Based Facial Palsy Classification Using Surface Electromyography. Sensors. 2026; 26(1):173. https://doi.org/10.3390/s26010173

Chicago/Turabian Style

Manzoor, Ibrahim, Aryana Popescu, Sarah Ricchizzi, Aldo Spolaore, Mykola Gorbachuk, Marcos Tatagiba, Georgios Naros, and Kathrin Machetanz. 2026. "Automated Neuromuscular Assessment: Machine-Learning-Based Facial Palsy Classification Using Surface Electromyography" Sensors 26, no. 1: 173. https://doi.org/10.3390/s26010173

APA Style

Manzoor, I., Popescu, A., Ricchizzi, S., Spolaore, A., Gorbachuk, M., Tatagiba, M., Naros, G., & Machetanz, K. (2026). Automated Neuromuscular Assessment: Machine-Learning-Based Facial Palsy Classification Using Surface Electromyography. Sensors, 26(1), 173. https://doi.org/10.3390/s26010173

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop