Next Article in Journal
Resolution and Frequency Effects on UAVs Semi-Direct Visual-Inertial Odometry (SVO) for Warehouse Logistics
Previous Article in Journal
A GNSS/INS/LiDAR Integration Scheme for UAV-Based Navigation in GNSS-Challenging Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning and Wearable Sensors for the Early Detection of Balance Disorders in Parkinson’s Disease

by
Francesco Castelli Gattinara Di Zubiena
1,
Greta Menna
1,
Ilaria Mileti
2,
Alessandro Zampogna
3,
Francesco Asci
3,4,
Marco Paoloni
5,
Antonio Suppa
3,4,
Zaccaria Del Prete
1 and
Eduardo Palermo
1,*
1
Department of Mechanical and Aerospace Engineering, Sapienza University of Rome, 00184 Rome, Italy
2
Mechanical Measurements and Microelectronics (M3Lab) Laboratory, Engineering Department, University Niccolò Cusano, 00166 Rome, Italy
3
Department of Human Neurosciences, Sapienza University of Rome, 00185 Rome, Italy
4
IRCCS Neuromed Institute, 86077 Pozzilli, Italy
5
Department of Anatomical and Histological Sciences, Legal Medicine and Orthopedics, Sapienza University of Rome, 00185 Rome, Italy
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9903; https://doi.org/10.3390/s22249903
Submission received: 27 October 2022 / Revised: 6 December 2022 / Accepted: 14 December 2022 / Published: 16 December 2022
(This article belongs to the Section Wearables)

Abstract

:
Dynamic posturography combined with wearable sensors has high sensitivity in recognizing subclinical balance abnormalities in patients with Parkinson’s disease (PD). However, this approach is burdened by a high analytical load for motion analysis, potentially limiting a routine application in clinical practice. In this study, we used machine learning to distinguish PD patients from controls, as well as patients under and not under dopaminergic therapy (i.e., ON and OFF states), based on kinematic measures recorded during dynamic posturography through portable sensors. We compared 52 different classifiers derived from Decision Tree, K-Nearest Neighbor, Support Vector Machine and Artificial Neural Network with different kernel functions to automatically analyze reactive postural responses to yaw perturbations recorded through IMUs in 20 PD patients and 15 healthy subjects. To identify the most efficient machine learning algorithm, we applied three threshold-based selection criteria (i.e., accuracy, recall and precision) and one evaluation criterion (i.e., goodness index). Twenty-one out of 52 classifiers passed the three selection criteria based on a threshold of 80%. Among these, only nine classifiers were considered “optimum” in distinguishing PD patients from healthy subjects according to a goodness index ≤ 0.25. The Fine K-Nearest Neighbor was the best-performing algorithm in the automatic classification of PD patients and healthy subjects, irrespective of therapeutic condition. By contrast, none of the classifiers passed the three threshold-based selection criteria in the comparison of patients in ON and OFF states. Overall, machine learning is a suitable solution for the early identification of balance disorders in PD through the automatic analysis of kinematic data from dynamic posturography.

1. Introduction

Parkinson’s disease (PD) is the second most common neurodegenerative disorder worldwide [1]. It is primarily characterized by the death of the dopaminergic neurons in a specific region of the midbrain named substantia nigra pars compacta [2]. Indeed, the current therapeutic cornerstone in PD consists of replacement treatment with L-Dopa that significantly improves movements, although it is invariably associated with complications in the advanced stages of the disease (e.g., dyskinesia and fluctuations) [3]. Among parkinsonian motor symptoms, postural instability is one of the most debilitating [2,4]. Postural instability severely affects balance and causes frequent falls and injuries, leading to increased hospitalization, loss of independence and high mortality rates in PD [5]. The delayed recognition and controversial response to the dopaminergic therapy of postural instability is responsible for negative outcomes in patients with PD [5]. Accordingly, the early identification of postural instability is a primary concern in the clinical management of patients with PD in order to optimize therapeutic strategies, prevent injuries and support the individual’s autonomy.
The current challenge resides in the low sensitivity of routine clinical examinations to early recognize postural instability in PD [6]. Indeed, clinical assessment usually detects postural instability only when the patient becomes symptomatic by complaining of balance issues and recurrent falls [7]. Dynamic posturography is an instrumental technique to objectively assess postural control under challenging conditions resembling free-living situations [8]. In 2021, Kamieniarz et al. [9] performed a study to detect balance changes in early stage of PD. They found that simple spatiotemporal parameters of the center of pressure (COP) were not enough to detect differences between PD and a healthy control (HC). Instead, power spectrum density (PSD) of the COP and sample entropy could detect early postural issues even if sample entropy failed to differentiate different states of the disease (Hoehn & Yahr H&Y-II and H&Y-III). In 2021, Yu et al. [10] also performed a quantitative analysis of postural instability in PD to detect early balance issues. They found that limits of stability in H&Y-I were not different from those of a HC, while they detected a pattern of decreases in postural stability during the progress of the disease.
We have recently used dynamic posturography integrated with a network of wearable inertial sensors to examine postural control in a cohort of patients with PD without a clinically overt postural instability and without a history of falls [11]. Despite being clinically asymptomatic for postural instability, patients with PD showed abnormal reactive postural responses to axial rotations both under and not under L-dopa (i.e., ON and OFF state of therapy, respectively) [11]. Hence, our instrumental approach demonstrated a subclinical balance impairment in patients, pointing at the potential suitability of portable inertial sensors for the early recognition of postural instability in PD.
The use of wearable sensors is part of the broader goal of integrating and interconnecting sensors and actuators with control systems and smart materials [12]. These kinds of sensors are becoming prominent in healthcare for measuring various biomechanical parameters [13,14]. Wearable sensors are easy to set up, relatively inexpensive and can be used in real-time since the processing phase is much shorter than the computing time required by some standard systems (e.g., marker tracking algorithms) [2]. However, the large-scale, high-dimensional data captured by wearable sensors requires sophisticated signal processing to transform it into scientifically and clinically valuable information, thus precluding their routine application in daily practice [15]. Providing clinicians with quantitative and automatic measures of motor and postural performance through portable sensors would allow the early detection and management of postural instability in PD. By automatically managing large volumes of data, machine learning algorithms have shown remarkable success in making accurate predictions for complex problems, including healthcare issues [16,17,18].
Another important role in the recent development of healthcare studies is played by machine learning solutions, which are often fed with data from various wearable sensors for the prediction of different patient health conditions [19,20,21,22]. In this scenario, a large number of researchers have already studied PD through inertial data and a machine learning approach, mostly involving gait. In 2018, Aich et al. [23] proposed a method to detect the freezing of gait (FoG). They asked 51 PD patients (36 with FoG episodes and 25 without FoG episodes) to perform gait tasks. They collected the inertial data from a 3D accelerometer on knees and extracted some features (such as average step time, stride time, step length, stride length, walking speed). They then trained four classifiers: Support Vector Machine (SVM), k-Nearest Neighbours (kNN), Decision Tree (DT) and Naïve Bayes. An accuracy of 89% was reported in the classification of FoG patients from no FoG patients with the SVM (rbf kernel). In 2015, Parisi et al. [24] presented a procedure to assess the severity of PD, automatically assigning UPDRS scores through a machine learning algorithm. They used portable sensors situated on the chest and both thighs and asked 24 PD patients to perform leg agility tasks and gait tasks. The best performance was that of kNN, with a maximum accuracy of 62%. In 2018, Caramia et al. [25] implemented a method to automatically divide PD patients from healthy subjects, examining the inertial data of 27 healthy controls and 27 PD patients: in particular, 8 subjects belonging to stage I, 9 to stage II and 8 to stage III on the Hoehn and Yahr (H&Y) scale. They were asked to perform a gait task. Different machine learning algorithms were used: Naïve Bayes, Linear Discriminant Analysis and Support Vector Machine. Average classification accuracy ranged between 63 and 75% among classifiers. In 2020, Aich et al. [26] realized a procedure to detect the ON/OFF state. Twenty idiopathic PD patients were asked to perform a gait task. Their inertial data were recorded by a 3D accelerometer on the knees. The extracted features were fed to a Support Vector Machine, k-Nearest Neighbors, Random Forest and Naïve Bayes. They found the best accuracy (96.7%) for the Random Forest classifier. The studies show no evidence of a machine learning algorithm outperforming others. Typology of the extracted features determined the most accurate algorithm in each case.
Machine learning was also used for disease classification by considering other motor tasks. For example, in 2017 Jeon et al. [15] used machine learning for the assessment of the severity of PD in 85 PD patients. They collected inertial data from an accelerometer and a gyroscope situated on the wrist during a resting task and trained five types of classifiers: Decision Tree, Linear Discriminant Analysis, Random Forest (RF), k-Nearest Neighbors and Support Vector Machine. They found the best accuracy (85.5%) with the Decision Tree algorithm. In 2010, Cancela et al. [27] studied the automatic assessment of the severity of PD and the assignment of UPDRS score. They analyzed inertial data from IMUs situated on the limbs and the trunk of 20 idiopathic PD patients during daily life activity. They used four classifiers: Support Vector Machine, k-Nearest Neighbors, Decision Tree and Artificial Neural Network. The highest accuracy of 86% was achieved by SVM.
The use of machine learning recently also appeared in posturography. Exley et al. [28] used it to predict UPDRS motor symptoms using force platforms during quiet standing. Features were obtained from quiet standing data while UPDRS-III subscores were the target variables. They compared seven machine learning algorithms for the classification: ridge and lasso logistic regression, Decision Tree, k-Nearest Neighbors, Random Forest, Support Vector Machine (rbf kernel) and Extreme Gradient Boosting. For body bradykinesia and hypokinesia, postural stability, rigidity and tremor at rest subscores, different algorithms performed the best in each case, but all of them had an accuracy lower than 80%.
In 2021, Fadil et al. [29] tested 19 PD and 13 HC. Patients were asked to lie supine on a tilt table for 5 min. The table was then tilted to 70 degrees for 15 min inducing an orthostatic challenge. Then, patients stood upright on a force platform and COP data were recorded. They used these data to feed six machine learning algorithms: Random Forest, Decision Tree, Support Vector Machine, k-Nearest Neighbors, Gaussian naïve Bayes and Neural Network (NN). Best performing algorithm was RF with an accuracy of 81%. They also found that PD patients were better differentiated from HC using time domain features compared to other feature groups.
Most of the related literature studies are focused on the classification of PD gait disturbances in comparison to healthy subjects, neglecting postural instability, especially in the early stages of the disease. Moreover, despite the fact that studies for the automatic classification and scoring of PD severity using machine learning algorithms are copious, the classification of ON/OFF medication state still needs to be explored.
Intending to fill this gap in the research, this study examines machine learning approaches to objectively evaluate early postural instability in PD, comparing reactive postural responses in patients with respect to healthy subjects when yaw rotations are pro-vided, trying to detect early asymptomatic postural abnormalities and verify whether those changes improve following dopaminergic therapy (i.e., OFF and ON state of therapy). Towards this aim, we analyzed and compared the results of four types of machine learning algorithms with different kernel functions.

2. Materials and Methods

2.1. Subjects

Twenty parkinsonian patients (1 woman and 19 men with a mean age of 67.7 ± 8.6 years) and a healthy control of fifteen subjects of similar age (mean age 65.2 ± 3.4 years) were enrolled in this study. The Department of Human Neurosciences of Sapienza University of Rome, Italy, enrolled PD patients according to specific inclusion criteria: diagnosis of idiopathic PD based on current criteria; no clinically evident postural instability; last year with no history of falls; score 1–2 on the Hoehn and Yahr scale; walk and upright stance independence; balance not affected by dementia (Mini-Mental State Examination-MMSE > 24), dyskinesia and comorbidities induced by L-Dopa.
An expert neurologist evaluated PD motor symptoms through the following clinical scales: Hoehn & Yahr scale and MDS-UPDRS (Movement Disorders Society-Unified Parkinson’s disease rating scale) part III. For the clinical assessment of balance, the postural instability/gait difficulty score (PIGD) and the Berg balance scale (BBS) were evaluated.
PD patients were tested both on the ON and OFF medication states. The OFF state was evaluated at least 12 h after the usual dopaminergic treatment, while the ON state was assessed one hour after. For each patient, the L-Dopa Equivalent Daily Dose (LEDD) was calculated. Patients did not receive other neuropsychiatric medications possibly affecting balance at the time of the study. Written informed consent to the study, approved by the institutional review board following the Declaration of Helsinki, was given by all enrolled patients.
The number of subjects was chosen to guarantee an adequate dimensionality of the training, validation and testing datasets.
More information is reported in Table 1.

2.2. Experimental Setup

To perform the experiment, subjects stand on a robotic platform (RotoBit1D) in an upright position, with arms hanging vertically and feet externally rotated a preferred amount. The RotoBit1D is a flat, rigid and round-shaped robotic platform with a polyethylene rotating disk with a height of 0.15 m and a diameter of 0.5 m, which allows a comfortable upright bipedal stance without narrowing feet. The RotoBit1D is equipped with the following components: a SANYO DENKI servo motor (maximum torque = 1.96 Nm); a speed reducer; an incremental encoder; a toothed belt (PowerGrip HDT). This robotic platform has already been used in the study of various postural parameters [30,31] thanks to an ad-hoc LabVIEW software program (2014, National Instruments, Austin, TX, USA) that controls the robotic platform, providing sinusoidal rotation around the vertical (yaw) axis chosen for this study.
Kinematics of axial body segments was gathered using three IMUs (MTw, Xsens Technologies-NL), each equipped with a 3-axes accelerometer, a 3-axes gyroscope and a 3-axes magnetometer. Wearable sensors were placed on the head, trunk and pelvis of participants through elastic belts to avoid relative movement between sensor and body. In more detail, the IMU of the head was placed under the supraciliary arc over the frontal bone, the IMU of the trunk was placed on the sternum body under the suprasternal notch and the IMU of the pelvis was placed just below the anterior sacral promontory and centered on the median sacral crests. Each subject was instrumented by the same expert operator to guarantee consistent sensor positioning on the body locations. IMUs data were acquired by MT Manager software, which let to visualization of inertial and magnetic quantity, detect pitch, yaw and roll angles, and understanding of 3D sensor orientation in real-time during the experimental session.
Both systems, the RotoBit1D and the IMUs, were simultaneously triggered through the RotoBit1D LabVIEW software. More specifically, an external trigger, i.e., a square signal ranging from 0 to +3 V, was provided to the IMUs both at the start and end of each trial.
Experimental setup and trajectories of the sinusoidal perturbation are shown in Figure 1.

2.3. Experimental Protocol

Before each recording session, all subjects were asked to perform a functional calibration procedure (FC). The FC consists of standing and sitting tasks, each lasting 5 s and allow for overcoming incorrect misalignment between body and sensor orientation [32].
Subjects were asked to stand in their comfortable upright position on a robotic platform, facing a 2 m distant red cross placed in the wall in front of them. Feet were symmetrically placed over the center of the platform. Although all participants were instructed on the protocol before starting the experimental session, no instructions were provided on the postural strategies to be adopted to maintain balance.
After performing the FC procedure, each subject was elicited though sinusoidal perturbation around the vertical axis. Three sinusoidal perturbations were provided: (i) Low Perturbation (LP), consisting of sinusoidal perturbation with a frequency of 0.2 Hz, peak amplitude of ±55° and a peak angular acceleration of 0.25°/s2; (ii) Medium Perturbation (MP), consisting of sinusoidal perturbation with a frequency of 0.3 Hz, peak amplitude of ±55° and a peak angular acceleration of 0.40°/s2; (iii) High Perturbation (HP), consisting of a sinusoidal perturbation with a frequency of 0.5 Hz, peak amplitude of ±35° and a peak angular acceleration of 0.50°/s2.
Each perturbation was repeated three times and starting both on the right and left side of the subject. Thus, a total of 18 trials were performed for each participant. Types of tasks are summarized in Table 2.
To make the subjects blind to the specific perturbation and avoid habituation or anticipatory strategies, perturbations were delivered randomly. To avoid bias due to similar task sequence, the order of the trials across the subject was randomized. Additionally, a sigmoidal wave was added at the start and at the end of the sinusoidal trajectory in order to avoid sudden variation in starting/stopping velocity.

2.4. Data Analysis and Features Extraction

Similar to [11], two types of features were extracted from the IMUs data: the range of motion of each body segment (head, trunk and pelvis) in AP and ML directions and the reciprocal body segment rotations, represented by the Gain Ratio (G) and the Phase Shift (φ).
To estimate the displacement of the head, trunk and pelvis in both the medio-lateral (ML) and antero-posterior (AP) directions, the acceleration signal of the inertial sensors was rotated in the global coordinate frame. Through the quaternion-derived rotation matrix, gravitational acceleration was removed. Then, the acceleration signal was straightforwardly integrated. Next, a zero-lag first-order Butterworth high-pass filter filtered the velocity with a cut-off frequency of 0.2 Hz for the AP and ML components and the body displacement was obtained through a second integration and filtering process. In this way, the range of motion of each body segment in the ML (ROM-ML) and AP (ROM-AP) directions were computed and expressed in mm. The reciprocal body segment rotations were also examined. The angular rotation of the pelvis, trunk and head around the longitudinal axis was computed according to previously reported procedures [33]. The Fast Fourier Transform was used to calculate the Gain Ratio (G) and the Phase Shift (φ) indices. In detail, the G index is calculated as the ratio of the maximum amplitudes of fundamental waves from distal and proximal body signals at the same frequency. The φ index was calculated in the Fourier domain as the difference between the phase angles of the two signal’s Fourier transform at the frequencies with the maximum amplitude. Values of G < 1 are associated with a lower amplitude of the yaw angle of the distal segment compared to the proximal one, whereas G > 1 are associated with a larger amplitude. A delay is connected to values of φ > 0, while an anticipation in the yaw phase between distal and proximal segments is related to values of φ < 0. Values that are close to 0 represent a perfect phase match the two segments.
In summary, each subject was asked to perform six different types of tests (see Table 2) and four features were extracted (ROM-ML, ROM-AP, G-relative and φ-relative) for each body segment (head, trunk and pelvis). Thus, the total number of the extracted features was 72, as combinations of 3 body segments × 6 tests × 4 extracted features.

2.5. Machine Learning: Training and Validation

Four types of machine learning algorithm were considered with different kernel functions: (i) Fine (DTF), Medium (DTM) and Coarse (DTC) Decision Tree; (ii) Fine (KNNF), Medium (KNNM), Coarse (KNNCO), Cosine (KNNCS), Cubic (KNNCB) and Weighted (KNNW) K-Nearest Neighbour; (iii) Linear (SVML), Quadratic (SVMQ) and Cubic (SVMCB) Support Vector Machine; (iv) Artificial Neural Network (ANN). The neural network was a shallow ANN with 1 hidden layer with 10 neurons and 1 output layer with 2 neurons.
To train the classifiers, 50-fold cross validation was used in DT, kNN and SVM and Leave One Subject Out Cross-Validation (LOSOCV) was used in ANN. Machine learning algorithms were implemented using the MATLAB (v.2020b, MathWorks, Natick, MA, USA) program.
These algorithms were chosen as they are some of the most commonly used and easy to implement.
The classification task consisted of four separate classification experiments:
  • PD vs. HC: which distinguishes between PD patients and healthy subjects;
  • OFF vs. HC: which distinguishes between patients in OFF state and healthy subjects;
  • ON vs. HC: which distinguishes between patients in ON state and healthy subjects;
  • OFF vs. ON: which distinguishes between patients in OFF state and patients in ON state.
In total, 52 machine learning classifiers were tested, as combinations of 13 types of machine learning algorithms × 4 classification experiments.

2.6. Performance Evaluation

To individuate the best performing machine learning algorithms among the 52 tested, confusion matrices were first extracted. Then, three threshold-based selection criteria [34] and one evaluation criteria were applied, similar to [35]. For all criteria, an 80% threshold was set (Figure 2).
The first selection criterion was based on accuracy, which represents the percentage of correct classification. It was calculated as:
A C C = T P + T N P + N × 100
where TP, TN, P and N are the true positive, true negative, number of positive and negative, respectively.
The second selection criterion was based on the recall or True Positive Rate (TPR), which represents the probability that an actual positive will test positive. It was calculated as:
T P R = T P T P + F N × 100
T N R = T N T N + F P × 100
where TNR stands for True Negative Rate, FN stands for false negative and FP stands for false positive.
The third selection criterion was based on the precision or PPV, which represents the percentage of true positive with respect to all positive. It was calculated as:
P P V = T P T P + F P × 100
where FP is the number of false positive.
According to the reported selection criteria, both type I and II errors were taken into account when evaluating the robustness of the classifier. The threshold applied was chosen to be 80% since it is a typical value derived from the literature for accepting a classifier as good [35].
The evaluation criterion was based on the Goodness index (G), which represents the Euclidean distance between the evaluated point in the ROC space and the point [0 1], representing the perfect classifier. It was calculated as:
G = 1 T P R 2 + 1 T N R 2
where TPR and TNR are the true positive rate and true negative rate, respectively.
G has values included between 0 and √2. If G ≤ 0.25, a classifier can be considered as optimum, if 0.25 < G ≤ 0.70, it can be considered a good classifier, if G = 0.70 it can be considered a random classifier and if G > 0.70 it can be considered a bad classifier [35]. By analyzing G value results, the best-performing classifier could be evaluated.
Figure 3 summarizes the complete process flow of the experiments.

3. Results

The evaluation of the machine learning algorithms was carried out by considering accuracy, recall (or TPR) and precision (or PPV). The accuracy is used to evaluate the number of correct classifications, while the recall evaluates the number of data samples correctly identified in a particular class. Finally, the confidence to belong to a particular class for a particular prediction is defined by the precision. Furthermore, the goodness index was calculated and used to assess the best machine learning algorithm for each classifier and dataset combination.

3.1. First Selection Criterion

The accuracy of each classifier is reported in Table 3. To calculate TN, TP and accuracy, we defined the “positive” outcome for the different classifiers as follows: (i) for PD vs. HC “positive” is PD class; for OFF vs. HC, the OFF class; for ON vs. HC, the ON class; for OFF vs. ON, the ON class.
Twenty-nine out of the 52 machine learning classifiers passed the first selection criterion and reached an accuracy equal to or higher than 80%. More specifically, the PD vs. HC classifier achieved a maximum accuracy of 95.6%. The OFF vs. HC classifier came up to a maximum accuracy of 90.9%. The ON vs. HC classifier reached a maximum accuracy of 89.9%. The OFF vs. ON classifier exhibited a top accuracy of 84.2%. It can be observed that the fine k-Nearest Neighbors algorithm showed the best performance in terms of accuracy for all classifiers, except for the OFF vs. HC. In this case, the best accuracy was reached using the coarse Decision Tree.

3.2. Second Selection Criterion

The TPR of the classifiers that met the first selection criterion are reported in Table 3.
Twenty-eight out of the 29 classifiers passed the selection criterion based on the recall value. It can be noticed that the highest value of recall (100%) was reached by many algorithms among different classifiers.

3.3. Third Selection Criterion

Table 3 shows the PPV of the overall classification model that fulfilled the first and second selection criteria.
Twenty-one out of 28 classifiers passed the third precision-based selection criterion, reaching a PPV at least of 80%. In detail, the PD vs. HC classifier achieved a PPV of 97.0%. The OFF vs. HC classifier came up to a PPV of 94.7%. The ON vs. HC classifier reached a PPV of 87.9%. In the OFF vs. ON classifier, no tested algorithms passed the precision threshold.

3.4. Goodness Index

The goodness index of all machine learning algorithms that passed the three selection criteria is exposed in Table 3.
Twelve out of 21 could be catalogued as “good” classifiers because their G-index was 0.25 < G < 0.70. Nine out of 21 classifiers could be classified as “optimum” classifiers because their G-index was ≤ 0.25, such as DTm, kNNf, ANN for PD vs. HC; DTf, DTm, DTc, kNNf, SVMcu and ANN for OFF vs. HC; kNNf for ON vs. HC.

3.5. Best Performing Algorithms

The most important characteristics and parameters of the three best-performing algorithms, which present the lowest G-index for each classifier, are reported in Table 4. It can be noticed that the training time is short and the prediction speed is high.
The ROC curve and, in particular, the Area Under the Curve (AUC) are usually used for evaluating the accuracy of classifiers. The ROC curves and the AUC values of the best machine learning classifiers are reported in Figure 3. In more detail, the AUC of the classifier PD vs. HC is 0.95. The AUC of the classifier OFF vs. HC is 0.90. The AUC of the classifier ON vs. HC is 0.94 (Figure 4).

4. Discussion

In this study, we used and compared four types of machine learning algorithms, including Decision Tree (Fine, Medium, and Coarse), k-Nearest Neighbor (Fine, Medium, Coarse, Cosine, Cubic and Weighted), Support Vector Machine (Linear, Quadratic and Cubic) and Artificial Neural Network, to automatically classify patients with PD and healthy subjects based on reactive postural responses to yaw perturbations. Algorithms were fed with data obtained from portable inertial sensors placed on the body during the balance perturbation test. We have demonstrated that, when considering balance control, machine learning algorithms are able to accurately distinguish PD patients from healthy subjects, irrespectively of their state of therapy. By contrast, different therapeutic conditions are not efficiently identified in PD based on reactive postural responses, as demonstrated by the suboptimal performance of classifiers when comparing patients in ON and OFF state of therapy.
The methodological approach adopted in this study allowed us to exclude a number of confounding factors possibly leading to the misinterpretation of findings. To specifically focus on asymptomatic and subclinical balance disorders, only PD patients without a history of falls and without clinically overt postural instability have been recruited. All patients have been carefully characterized clinically by means of standardized scales, investigating both motor and cognitive functions. Functional calibration procedures have been performed before all recording sessions to prevent the misalignment between subjects’ bodies and sensors orientation. Finally, experimental sessions have been conducted randomly according to the therapeutic state as well as the direction and frequency of postural perturbations to avoid possible carry-over effects.
As a first finding, a relevant number of classifiers (i.e., 21 out of 52) successfully passed the three selection criteria (i.e., accuracy, recall and precision) provided for the selection of the best-performing algorithm by considering a threshold of 80%. Among these, most (i.e., 20 out of 21) were considered “good” classifiers based on their G-index. This result supports the hypothesis that subclinical changes of balance control occur in PD patients before postural instability becomes clinically evident, thus allowing the distinction from healthy subjects based on the analysis of reactive postural responses. Indeed, as already reported elsewhere [11], compared to healthy subjects, our cohort of patients showed increased displacement of axial body segments in the medio-lateral direction and reduced lumbo-sacral mobility during yaw perturbations, fully in line with a number of previous observations [36,37,38].
Only nine classifiers presented an “optimum” performance in distinguishing PD patients from healthy subjects according to their G-index. Among these, the kNNf was the most efficient classifier in the automatic detection of PD when comparing patients with healthy subjects, irrespective of their therapeutic state. Only the DTc showed a slightly better performance than the kNNf in the comparison between patients in OFF state and healthy subjects (i.e., G-index 0.12 vs. 0.18). Overall, the three best-performing algorithms (kNNf for PD vs. HC, DTc for OFF vs. HC and kNNf for ON vs. HC) showed high performance both in the training and prediction phases with a short training time and high prediction speed (Table 4). Indeed, the high accuracy values shown in Table 4 fully agree with the ROC-AUC values (between 0.90 and 0.97), in line with previous studies adopting a similar approach to assess other parkinsonian symptoms [15,39].
When looking at the performance of the best classifiers, an accuracy of 95.6% (kNNf), 90.9% (DTc) and 89.9% (kNNf) was reached in the distinction of healthy subjects from all patients, patients in OFF state and patients in ON state, respectively. These performances are similar or even better than those reported in previous studies adopting inertial data and machine learning to differentiate PD patients from healthy subjects based on specific motor symptoms (Table 5) [23,25,40,41,42].
It is likely that the mild differences in the classification performance between our and other studies mostly rely on the specific motor tasks adopted by various authors. Indeed, most previous studies primarily focused on gait abnormalities in PD patients, also including paroxysmal disorders such as FoG [23,25,40,41]. According to the high accuracy of our algorithms in classifying PD patients based on their balance control, our results expand previous findings by proposing reactive postural responses to yaw perturbations as alternative measures to recognize PD patients through machine learning algorithms. Moreover, our approach provides clinicians with a new sensitive tool to easily monitor postural abilities in PD patients over time and early recognize subclinical balance disorders.
It can be seen that more complex deep learning techniques can lead to better results that allow discrimination of PD even in real-life situations [42].
Concerning the automatic classification of PD patients in ON and OFF states, the kNNf reached an accuracy of 84.2% and a TPR of 95% in distinguishing the different therapeutic conditions. However, owing to a PPV of 78.1%, the third selection criterion has failed and, consequently, all machine learning algorithms for the detection of the OFF and ON states presented a suboptimal performance. This finding is apparently in contrast with previous studies successfully using machine learning to distinguish PD patients under and not under L-Dopa [26,43,44,45]. However, it should be considered that balance disorders are usually refractory to dopaminergic therapy [7,37,46,47]. By contrast, previous studies successfully classifying ON and OFF states in PD considered motor symptoms that are largely responsive to L-Dopa, such as gait disorders [26,43,44,45,48]. Therefore, the suboptimal performance of our machine learning algorithms in distinguishing patients in ON and OFF states would reflect unchanged reactive postural responses after the L-Dopa intake, suggesting that, in PD, neurodegeneration involves neurotransmitter systems other than the dopaminergic one since the early stages of the disease [11]. Accordingly, these findings further support the potential usefulness of our approach to early recognize balance disorders and implement alternative therapeutic strategies, including tailored rehabilitative training, to prevent harmful injuries in PD.
The use of Rotobit complicates the reproducibility of the procedure, but any other source of controlled axial perturbation to the patients could be applied in a clinical setting. However, building on the results of this study, it would be possible in the future to identify a methodology in which the patient is placed on the perturbating platform with the three low-cost and easy-to-position wearable sensors. Using the best-performing machine learning (kNN), this will allow us to conduct an evaluation to add to the others clinical assessments of diagnosis of PD. Furthermore, to strengthen the identification of postural problems, the number of sensors applied could be increased when needed. Therefore, the use of a simple system and basic machine learning algorithms could be an excellent support to clinicians for the early identification of postural disorders due to Parkinson’s disease.

5. Conclusions

This study has provided a new quantitative, reliable and useful tool based on dynamic posturography and machine learning to early recognize and objectively monitor balance control in patients with PD. The proposed data acquisition procedure is fast and simple, as is the equipment, which is composed of only three easy-to-place wearable inertial sensors. The inertial data can be stored, analyzed at a later time and used to train machine learning classifiers. This work points out how machine learning algorithms could successfully distinguish healthy subjects from PD patients without clinically overt postural instability, both in OFF and ON state, by analyzing reactive postural responses to yaw perturbations. Among the thirteen machine learning techniques used in this study, fine k-Nearest Neighbors outweighs all the other classifiers, detecting the two conditions (PD patients and healthy subjects) with a maximum accuracy of 95.6, recall of 99.0 and precision of 95.2%. Accordingly, this method may help to early recognize and address balance disorders in PD. In the future, a simulation model based on the data from this study could be developed to increase the possibility of validating the results. Furthermore, future developments foresee the possibility of applying this method to distinguish different stages of PD and monitor the disease progression, as well as to help in the differential diagnosis of atypical forms of parkinsonism, such as progressive supranuclear palsy and multiple system atrophy.

Author Contributions

Conceptualization, I.M., M.P., A.S. and E.P.; Data curation, G.M. and I.M.; Investigation, I.M., A.Z. and E.P.; Methodology, F.C.G.D.Z., G.M., I.M., A.Z., F.A., M.P., A.S., Z.D.P. and E.P.; Project administration, M.P., A.S. and E.P.; Resources, M.P., A.S. and E.P.; Software, F.C.G.D.Z. and G.M.; Supervision, A.S. and E.P.; Validation, I.M., A.S. and E.P.; Visualization, F.C.G.D.Z. and G.M.; Writing—original draft, F.C.G.D.Z. and G.M.; Writing—review and editing, F.C.G.D.Z., G.M., I.M., A.Z., F.A., M.P., A.S., Z.D.P. and E.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the fellowship BE_FOR_ERC 2019 granted to E. Palermo by Sapienza University of Rome, Italy.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Ethics Committee of Sapienza University of Rome (protocol code 0372/2022 and date of approval 4 May 2022).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The private data presented in this study are available on request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dorsey, E.R.; Elbaz, A.; Nichols, E.; Abbasi, N.; Abd-Allah, F.; Abdelalim, A.; Adsuar, J.C.; Ansha, M.G.; Brayne, C.; Choi, J.-Y.J.; et al. Global, regional, and national burden of Parkinson’s disease, 1990–2016: A systematic analysis for the Global Burden of Disease Study 2016. Lancet Neurol. 2018, 17, 939–953. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Davie, C.A. A review of Parkinson’s disease. Br. Med. Bull. 2008, 86, 109–127. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Armstrong, M.J.; Okun, M.S. Diagnosis and Treatment of Parkinson Disease: A Review. JAMA 2020, 323, 548–560. [Google Scholar] [CrossRef] [PubMed]
  4. Zampogna, A.; Cavallieri, F.; Bove, F.; Suppa, A.; Castrioto, A.; Meoni, S.; Pélissier, P.; Schmitt, E.; Bichon, A.; Lhommée, E.; et al. Axial impairment and falls in Parkinson’s disease: 15 years of subthalamic deep brain stimulation. NPJ Park. Dis. 2022, 8, 121. [Google Scholar] [CrossRef] [PubMed]
  5. Crouse, J.; Phillips, J.R.; Jahanshahi, M.; Moustafa, A.A. Postural instability and falls in Parkinson’s disease. Rev. Neurosci. 2016, 27, 549–555. [Google Scholar] [CrossRef]
  6. Nonnekes, J.; Goselink, R.; Weerdesteyn, V.; Bloem, B.R. The Retropulsion Test: A Good Evaluation of Postural Instability in Parkinson’s Disease? J. Park. Dis. 2015, 5, 43–47. [Google Scholar] [CrossRef]
  7. Palakurthi, B.; Burugupally, S.P. Postural instability in Parkinson’s disease: A review. Brain Sci. 2019, 9, 239. [Google Scholar] [CrossRef] [Green Version]
  8. Nonnekes, J.; de Kam, D.; Geurts, A.C.; Weerdesteyn, V.; Bloem, B.R. Unraveling the mechanisms underlying postural instability in Parkinson’s disease using dynamic posturography. Expert Rev. Neurother. 2013, 13, 1303–1308. [Google Scholar] [CrossRef]
  9. Kamieniarz, A.; Michalska, J.; Marszałek, W.; Stania, M.; Słomka, K.J.; Gorzkowska, A.; Juras, G.; Okun, M.S.; Christou, E.A. Detection of postural control in early Parkinson’s disease: Clinical testing vs. modulation of center of pressure. PLoS ONE 2021, 16, e0245353. [Google Scholar] [CrossRef]
  10. Yu, Y.; Liang, S.; Wang, Y.; Zhao, Y.; Zhao, J.; Li, H.; Wu, J.; Cheng, Y.; Wu, F.; Wu, J. Quantitative Analysis of Postural Instability in Patients with Parkinson’s Disease. Park. Dis. 2021, 2021, 5681870. [Google Scholar] [CrossRef]
  11. Zampogna, A.; Mileti, I.; Martelli, F.; Paoloni, M.; Del Prete, Z.; Palermo, E.; Suppa, A. Early balance impairment in Parkinson’s Disease: Evidence from Robot-assisted axial rotations. Clin. Neurophysiol. 2021, 132, 2422–2430. [Google Scholar] [CrossRef] [PubMed]
  12. Fortuna, L.; Buscarino, A. Smart Materials. Materials 2022, 15, 6307. [Google Scholar] [CrossRef] [PubMed]
  13. Mileti, I.; Taborri, J.; D’Alvia, L.; Parisi, S.; Ditto, M.C.; Peroni, C.L.; Scarati, M.; Priora, M.; Rossi, S.; Fusaro, E.; et al. Accuracy Evaluation and Clinical Application of an Optimized Solution for Measuring Spatio-Temporal Gait Parameters. In Proceedings of the 2020 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Bari, Italy, 1 June–1 July 2020; IEEE: Bari, Italy, 2020. [Google Scholar] [CrossRef]
  14. DrAlvia, L.; Scalona, E.; Palermo, E.; Del Prete, Z.; Pittella, E.; Pisa, S.; Piuzzi, E. Tetrapolar Low-Cost Systems for Thoracic Impedance Plethysmography. In Proceedings of the 2018 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 11–13 June 2018. [Google Scholar] [CrossRef]
  15. Jeon, H.; Lee, W.-W.; Park, H.; Lee, H.J.; Kim, S.K.; Kim, H.B.; Jeon, B.; Park, K.S. Automatic Classification of Tremor Severity in Parkinson’s Disease Using a Wearable Device. Sensors 2017, 17, 2067. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Borzì, L.; Mazzetta, I.; Zampogna, A.; Suppa, A.; Irrera, F.; Olmo, G. Predicting Axial Impairment in Parkinson’s Disease through a Single Inertial Sensor. Sensors 2022, 22, 412. [Google Scholar] [CrossRef] [PubMed]
  17. Borzì, L.; Mazzetta, I.; Zampogna, A.; Suppa, A.; Olmo, G.; Irrera, F. Prediction of Freezing of Gait in Parkinson’s Disease Using Wearables and Machine Learning. Sensors 2021, 21, 614. [Google Scholar] [CrossRef]
  18. Asci, F.; Scardapane, S.; Zampogna, A.; D’Onofrio, V.; Testa, L.; Patera, M.; Falletti, M.; Marsili, L.; Suppa, A. Handwriting Declines with Human Aging: A Machine Learning Study. Front. Aging Neurosci. 2022, 14, 889930. [Google Scholar] [CrossRef]
  19. Massaro, A.; Ricci, G.; Selicato, S.; Raminelli, S.; Galiano, A. Decisional Support System with Artificial Intelligence oriented on Health Prediction using a Wearable Device and Big Data. In Proceedings of the 2020 IEEE International Workshop on Metrology for Industry 4.0 & IoT, Roma, Italy, 3–5 June 2020. [Google Scholar]
  20. Majnarić, L.T.; Babič, F.; O’Sullivan, S.; Holzinger, A. AI and Big Data in Healthcare: Towards a More Comprehensive Research Framework for Multimorbidity. J. Clin. Med. 2021, 10, 766. [Google Scholar] [CrossRef]
  21. Massaro, A.; Maritati, V.; Savino, N.; Galiano, A. Neural Networks for Automated Smart Health Platforms oriented on Heart Predictive Diagnostic Big Data Systems. In Proceedings of the 2018 AEIT International Annual Conference, Bari, Italy, 3–5 October 2018. [Google Scholar]
  22. Dias, D.; Paulo Silva Cunha, J. Wearable Health Devices—Vital Sign Monitoring, Systems and Technologies. Sensors 2018, 18, 2414. [Google Scholar] [CrossRef]
  23. Aich, S.; Pradhan, P.M.; Park, J.; Sethi, N.; Vathsa, V.S.S.; Kim, H.-C. A Validation Study of Freezing of Gait (FoG) Detection and Machine-Learning-Based FoG Prediction Using Estimated Gait Characteristics with a Wearable Accelerometer. Sensors 2018, 18, 3287. [Google Scholar] [CrossRef] [Green Version]
  24. Parisi, F.; Ferrari, G.; Giuberti, M.; Contin, L.; Cimolin, V.; Azzaro, C.; Albani, G.; Mauro, A. Body-Sensor-Network-Based Kinematic Characterization and Comparative Outlook of UPDRS Scoring in Leg Agility, Sit-to-Stand, and Gait Tasks in Parkinson’s Disease. IEEE J. Biomed. Health Inform. 2015, 19, 1777–1793. [Google Scholar] [CrossRef]
  25. Caramia, C.; Torricelli, D.; Schmid, M.; Munoz-Gonzalez, A.; Gonzalez-Vargas, J.; Grandas, F.; Pons, J.L. IMU-Based Classification of Parkinson’s Disease From Gait: A Sensitivity Analysis on Sensor Location and Feature Selection. IEEE J. Biomed. Health Inform. 2018, 22, 1765–1774. [Google Scholar] [CrossRef] [PubMed]
  26. Aich, S.; Youn, J.; Chakraborty, S.; Pradhan, P.M.; Park, J.H.; Park, S.; Park, J. A Supervised Machine Learning Approach to Detect the on/off State in Parkinson’s Disease Using Wearable Based Gait Signals. Diagnostics 2020, 10, 421. [Google Scholar] [CrossRef] [PubMed]
  27. Cancela, J.; Pansera, M.; Arredondo, M.; Estrada, J.; Pastorino, M.; Pastor-Sanz, L.; Villalar, J. A comprehensive motor symptom monitoring and management system: The bradykinesia case. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2010, 2010, 1008–1011. [Google Scholar] [CrossRef]
  28. Exley, T.; Moudy, S.; Patterson, R.M.; Kim, J.; Albert, M.V. Predicting UPDRS Motor Symptoms in Individuals with Parkinson’s Disease from Force Plates Using Machine Learning. IEEE J. Biomed. Health Inform. 2022, 26, 3486–3494. [Google Scholar] [CrossRef] [PubMed]
  29. Fadil, R.; Huether, A.; Brunnemer, R.; Blaber, A.P.; Lou, J.-S.; Tavakolian, K. Early Detection of Parkinson’s Disease Using Center of Pressure Data and Machine Learning. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 31 October–4 November 2021; pp. 2433–2436. [Google Scholar] [CrossRef]
  30. Mileti, I.; Taborri, J.; Rossi, S.; Del Prete, Z.; Paoloni, M.; Suppa, A.; Palermo, E. Measuring age-related differences in kinematic postural strategies under yaw perturbation. In Proceedings of the 2018 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 11–13 June 2018. [Google Scholar] [CrossRef]
  31. Taborri, J.; Mileti, I.; Del Prete, Z.; Rossi, S.; Palermo, E. Yaw Postural Perturbation Through Robotic Platform: Aging Effects on Muscle Synergies. In Proceedings of the 2018 7th IEEE International Conference on Biomedical Robotics and Biomechatronics (Biorob), Enschede, The Netherlands, 26–29 August 2018; pp. 916–921. [Google Scholar] [CrossRef]
  32. Palermo, E.; Rossi, S.; Marini, F.; Patane, F.; Cappa, P. Experimental evaluation of accuracy and repeatability of a novel body-to-sensor calibration procedure for inertial sensor-based gait analysis. Measurement 2014, 52, 145–155. [Google Scholar] [CrossRef]
  33. Mileti, I.; Taborri, J.; Rossi, S.; Del Prete, Z.; Paoloni, M.; Suppa, A.; Palermo, E. Reactive Postural Responses to Continuous Yaw Perturbations in Healthy Humans: The Effect of Aging. Sensors 2020, 20, 63. [Google Scholar] [CrossRef] [Green Version]
  34. Juba, B.; Le, H.S. Precision-Recall versus Accuracy and the Role of Large Data Sets. [Online]. Available online: www.aaai.org (accessed on 5 December 2022).
  35. Taborri, J.; Palermo, E.; Rossi, S. Automatic Detection of Faults in Race Walking: A Comparative Analysis of Machine-Learning Algorithms Fed with Inertial Sensor Data. Sensors 2019, 19, 1461. [Google Scholar] [CrossRef] [Green Version]
  36. Beretta, V.; Vitório, R.; dos Santos, P.C.R.; Orcioli-Silva, D.; Gobbi, L.T.B. Postural control after unexpected external perturbation: Effects of Parkinson’s disease subtype. Hum. Mov. Sci. 2019, 64, 12–18. [Google Scholar] [CrossRef]
  37. Halmi, Z.; Dinya, E.; Málly, J. Destroyed non-dopaminergic pathways in the early stage of Parkinson’s disease assessed by posturography. Brain Res. Bull. 2019, 152, 45–51. [Google Scholar] [CrossRef]
  38. Nijhuis, L.B.O.; Allum, J.H.J.; Nanhoe-Mahabier, W.; Bloem, B.R. Influence of Perturbation Velocity on Balance Control in Parkinson’s Disease. PLoS ONE 2014, 9, e86650. [Google Scholar] [CrossRef] [Green Version]
  39. Tripoliti, E.E.; Tzallas, A.T.; Tsipouras, M.G.; Rigas, G.; Bougia, P.; Leontiou, M.; Konitsiotis, S.; Chondrogiorgi, M.; Tsouli, S.; Fotiadis, D.I. Automatic detection of freezing of gait events in patients with Parkinson’s disease. Comput. Methods Programs Biomed. 2013, 110, 12–26. [Google Scholar] [CrossRef] [PubMed]
  40. Klucken, J.; Barth, J.; Kugler, P.; Schlachetzki, J.; Henze, T.; Marxreiter, F.; Kohl, Z.; Steidl, R.; Hornegger, J.; Eskofier, B.; et al. Unbiased and Mobile Gait Analysis Detects Motor Impairment in Parkinson’s Disease. PLoS ONE 2013, 8, e56956. [Google Scholar] [CrossRef] [PubMed]
  41. Naghavi, N.; Miller, A.; Wade, E. Towards Real-Time Prediction of Freezing of Gait in Patients with Parkinson’s Disease: Addressing the Class Imbalance Problem. Sensors 2019, 19, 3898. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Atri, R.; Urban, K.; Marebwa, B.; Simuni, T.; Tanner, C.; Siderowf, A.; Frasier, M.; Haas, M.; Lancashire, L. Deep Learning for Daily Monitoring of Parkinson’s Disease Outside the Clinic Using Wearable Sensors. Sensors 2022, 22, 6831. [Google Scholar] [CrossRef]
  43. Keijsers, L.W.; Horstink, M.W.; Gielen, S.C. Automatic assessment of levodopa-induced dyskinesias in daily life by neural networks. Mov. Disord. 2003, 18, 70–80. [Google Scholar] [CrossRef]
  44. Keijsers, N.L.; Horstink, M.W.; Gielen, S.C. Ambulatory motor assessment in Parkinson’s disease. Mov. Disord. 2006, 21, 34–44. [Google Scholar] [CrossRef] [Green Version]
  45. Hssayeni, M.D.; Burack, M.A.; Jimenez-Shahed, J.; Ghoraani, B. Assessment of response to medication in individuals with Parkinson’s disease. Med. Eng. Phys. 2019, 67, 33–43. [Google Scholar] [CrossRef]
  46. Di Giulio, I.; George, R.J.S.; Kalliolia, E.; Peters, A.L.; Limousin, P.; Day, B.L. Maintaining balance against force perturbations: Impaired mechanisms unresponsive to levodopa in Parkinson’s disease. J. Neurophysiol. 2016, 116, 493–502. [Google Scholar] [CrossRef] [Green Version]
  47. De Kam, D.; Nonnekes, J.; Nijhuis, L.B.O.; Geurts, A.C.H.; Bloem, B.R.; Weerdesteyn, V. Dopaminergic medication does not improve stepping responses following backward and forward balance perturbations in patients with Parkinson’s disease. J. Neurol. 2014, 261, 2330–2337. [Google Scholar] [CrossRef]
  48. Suppa, A.; Kita, A.; Leodori, G.; Zampogna, A.; Nicolini, E.; Lorenzi, P.; Rao, R.; Irrera, F. l-DOPA and Freezing of Gait in Parkinson’s Disease: Objective Assessment through a Wearable Wireless System. Front. Neurol. 2017, 8, 406. [Google Scholar] [CrossRef]
Figure 1. (a) Experimental setup and sensor placement on subject; (b) Right and left sinusoidal trajectories provided through the Rotobit1D during the low perturbation condition. Dashed lines represent the sigmoidal trajectory of the beginning and the end of the acquisition. Continuous line represents the sinusoidal trajectory.
Figure 1. (a) Experimental setup and sensor placement on subject; (b) Right and left sinusoidal trajectories provided through the Rotobit1D during the low perturbation condition. Dashed lines represent the sigmoidal trajectory of the beginning and the end of the acquisition. Continuous line represents the sinusoidal trajectory.
Sensors 22 09903 g001
Figure 2. Flow chart for the identification of the best-performing classifiers.
Figure 2. Flow chart for the identification of the best-performing classifiers.
Sensors 22 09903 g002
Figure 3. Complete process flow of the experiments.
Figure 3. Complete process flow of the experiments.
Sensors 22 09903 g003
Figure 4. ROC curves and AUC values of the classifiers PD vs. HC (a), OFF vs. HC (b) and ON vs. HC (c).
Figure 4. ROC curves and AUC values of the classifiers PD vs. HC (a), OFF vs. HC (b) and ON vs. HC (c).
Sensors 22 09903 g004
Table 1. Clinical features of patients with Parkinson’s disease.
Table 1. Clinical features of patients with Parkinson’s disease.
PatientsPD OnsetH&YMDS-UPDRSIIIMMSEPIGDBERG
120171.52730156
2201024130647
320131.52630450
4201023030256
520131.51530255
6201024330256
720151.52430156
8201523729155
9201124525745
1020151.52529256
1120161.52228155
1220151.52829252
13201811429256
1420151.52130456
1520151.52125250
16201222430352
17201022530554
18201411230354
19201611030252
20201522830746
Mean-1.725.929.22.952.9
SD-0.59.81.51.93.6
Table 2. Resume of tasks provided by the Rotobit1D.
Table 2. Resume of tasks provided by the Rotobit1D.
PerturbationLowMediumHigh
Frequency (Hz)0.20.30.5
Peak amplitude (°)±55±55±35
Peak angular acceleration (°/s2)0.250.400.50
Table 3. Percentage values of each classifier for each criterion: accuracy (ACC), which represents the percentage of correct classification; recall (TPR), which represents the probability that an actual positive will test positive (positives are PD in PD vs. HC, OFF in OFF vs. HC and ON in ON vs. HC and OFF vs. ON); precision (PPV), which represents the percentage of true positive with respect to all positive; goodness index (G), which represents the Euclidean distance between the evaluated point in the ROC space and the point [0 1], representing the perfect classifier. Values exceeding the threshold and passing to the next selection are shown in green, those not exceeding the threshold are shown in white and values not passed are shown with “/”.
Table 3. Percentage values of each classifier for each criterion: accuracy (ACC), which represents the percentage of correct classification; recall (TPR), which represents the probability that an actual positive will test positive (positives are PD in PD vs. HC, OFF in OFF vs. HC and ON in ON vs. HC and OFF vs. ON); precision (PPV), which represents the percentage of true positive with respect to all positive; goodness index (G), which represents the Euclidean distance between the evaluated point in the ROC space and the point [0 1], representing the perfect classifier. Values exceeding the threshold and passing to the next selection are shown in green, those not exceeding the threshold are shown in white and values not passed are shown with “/”.
DTfDTmDTckNNfkNNmkNNckNNcokNNcukNNwSVMlSVMqSVMcuANN
PD vs. HCACC84.784.786.995.686.173.786.183.989.188.389.192.788.3
TPR89.189.193.199.0100/97.010010010099.010088.3
PPV90.090.089.595.284.2/86.082.187.186.387.79197.0
G0.300.290.310.140.52/0.450.610.420.430.390.280.16
OFF vs. HCACC88.988.990.989.979.860.677.875.883.879.883.888.982.8
TPR86.786.790.095.0////100/95.096.781.2
PPV94.594.594.789.1////78.9/81.486.693.3
G0.200.200.120.18//////0.330.230.23
ON vs. HCACC81.881.877.889.978.860.678.877.882.881.881.884.880.8
TPR95.095.0/96.7////10010096.798.379.7
PPV79.279.2/87.9////77.976.978.480.8/
G///0.20///////0.36/
OFF vs. ONACC66.766.757.584.250.038.350.854.275.063.377.377.553.3
TPR///95.0/////////
PPV///78.1/////////
G/////////////
Table 4. Best-performing machine learning classifier’s characteristics and parameters.
Table 4. Best-performing machine learning classifier’s characteristics and parameters.
PD vs. HCkNNfAccuracy95.6%
Goodness index0.14
Total misclassification cost6
Prediction speed660 obs/s
Training time1.7149 s
Number of neighbors1
Distance metricEuclidean
Distance weightEqual
OFF vs. HCDTcAccuracy90.9%
Goodness index0.12
Total misclassification cost9
Prediction speed620 obs/s
Training time1.3464 s
Max number of splits4
Split criterionGini’s diversity index
ON vs. HCkNNfAccuracy89.9%
Goodness index0.20
Total misclassification cost10
Prediction speed440 obs/s
Training time2.129 s
Number of neighbors1
Distance metricEuclidean
Distance weightEqual
Table 5. A comparison of this work with state-of-the-art models’ work for activity detection. PYP (Postural Yaw Perturbation), GT (Gait Task), TTHP (Toe Tapping with Heel Pin), CT (Circling Test), WLE (Walk-like events), LDA (Linear Discriminant Analysis), MLP (Multilayer Perceptron), 1D-CNN (One-dimensional Convolution Neural Network).
Table 5. A comparison of this work with state-of-the-art models’ work for activity detection. PYP (Postural Yaw Perturbation), GT (Gait Task), TTHP (Toe Tapping with Heel Pin), CT (Circling Test), WLE (Walk-like events), LDA (Linear Discriminant Analysis), MLP (Multilayer Perceptron), 1D-CNN (One-dimensional Convolution Neural Network).
AuthorObjectiveType of TaskML AlgorithmAccuracy
This workClassification of PD from HCPYPkNNf96%
Aich et al. [23]Detection of FoGGTSVM88%
Caramia et al. [25]Classification of PD from HCGTSVM80%
Klucken et al. [40]Classification of PD from HCGT, TTHP, CTLDA, AdaBoost, SVM81%
Naghavi et al. [41]Detection of FoGGTkNN, SVM, MLP97%
Atri et al. [42]Classification of PD from HCWLE1D-CNNs90%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Castelli Gattinara Di Zubiena, F.; Menna, G.; Mileti, I.; Zampogna, A.; Asci, F.; Paoloni, M.; Suppa, A.; Del Prete, Z.; Palermo, E. Machine Learning and Wearable Sensors for the Early Detection of Balance Disorders in Parkinson’s Disease. Sensors 2022, 22, 9903. https://doi.org/10.3390/s22249903

AMA Style

Castelli Gattinara Di Zubiena F, Menna G, Mileti I, Zampogna A, Asci F, Paoloni M, Suppa A, Del Prete Z, Palermo E. Machine Learning and Wearable Sensors for the Early Detection of Balance Disorders in Parkinson’s Disease. Sensors. 2022; 22(24):9903. https://doi.org/10.3390/s22249903

Chicago/Turabian Style

Castelli Gattinara Di Zubiena, Francesco, Greta Menna, Ilaria Mileti, Alessandro Zampogna, Francesco Asci, Marco Paoloni, Antonio Suppa, Zaccaria Del Prete, and Eduardo Palermo. 2022. "Machine Learning and Wearable Sensors for the Early Detection of Balance Disorders in Parkinson’s Disease" Sensors 22, no. 24: 9903. https://doi.org/10.3390/s22249903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop