You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

13 November 2023

Clinically Informed Automated Assessment of Finger Tapping Videos in Parkinson’s Disease

,
,
and
1
Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
2
Pacific Parkinson Research Centre, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
3
Department of Neurology, Faculty of Medicine, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Intelligent Sensors for Healthcare and Patient Monitoring

Abstract

The utilization of Artificial Intelligence (AI) for assessing motor performance in Parkinson’s Disease (PD) offers substantial potential, particularly if the results can be integrated into clinical decision-making processes. However, the precise quantification of PD symptoms remains a persistent challenge. The current standard Unified Parkinson’s Disease Rating Scale (UPDRS) and its variations serve as the primary clinical tools for evaluating motor symptoms in PD, but are time-intensive and prone to inter-rater variability. Recent work has applied data-driven machine learning techniques to analyze videos of PD patients performing motor tasks, such as finger tapping, a UPDRS task to assess bradykinesia. However, these methods often use abstract features that are not closely related to clinical experience. In this paper, we introduce a customized machine learning approach for the automated scoring of UPDRS bradykinesia using single-view RGB videos of finger tapping, based on the extraction of detailed features that rigorously conform to the established UPDRS guidelines. We applied the method to 75 videos from 50 PD patients collected in both a laboratory and a realistic clinic environment. The classification performance agreed well with expert assessors, and the features selected by the Decision Tree aligned with clinical knowledge. Our proposed framework was designed to remain relevant amid ongoing patient recruitment and technological progress. The proposed approach incorporates features that closely resonate with clinical reasoning and shows promise for clinical implementation in the foreseeable future.

1. Introduction

Parkinson’s Disease (PD) is the most-common neurodegenerative movement disorder characterized by various motor symptoms, including resting tremor, bradykinesia, muscular rigidity, and postural instability [1]. Among these symptoms, bradykinesia, the slowness of movement often associated with a reduced movement amplitude, is a fundamental feature and a key diagnostic criterion for Parkinsonism [2].
The Unified Parkinson’s Disease Rating Scale (UPDRS) is recognized as the definitive standard for assessing the severity of Parkinson’s Disease (PD). This scale has undergone a partial revision via the Movement Disorder Society (MDS), resulting in the MDS-UPDRS [3]. An essential element of the MDS-UPDRS is the motor examination (Part III), encompassing 18 motor tasks performed by the patient and evaluated by a rater, who scores each task based on the observed level of impairment.
A significant task in this scale, Item 4, is finger tapping, which is used to gauge bradykinesia. In this test, patients are instructed to tap their index finger to their thumb as quickly and widely as possible, completing ten taps. The rater observes and evaluates several aspects of this action, such as the speed and amplitude, any noticeable decrease in speed or amplitude, and the presence of hesitations or stops. Each side of the body is evaluated separately, and scores are assigned on a scale from 0 (normal) to 4 (severe), as outlined in Table 1.
Table 1. MDS-UDPRS criteria for the finger tapping task.
While the transition from the UPDRS to the MDS-UPDRS addressed several uncertainties and shortcomings of the original version, it did not completely overcome all challenges. Notably, the subjective nature of this manually administered clinical scale persists, resulting in potential scoring discrepancies by the same (intra-rater) or different (inter-rater) evaluators. Moreover, human assessments of this nature can be both time-intensive and costly.
Advancements in machine learning technologies have greatly improved the objective and automated assessment of Parkinson’s Disease (PD) [4,5,6,7]. Notably, wearable sensor technology has been utilized to measure bradykinesia during finger tapping [8,9], with several of these sensors now available for clinical use [10,11]. The use of machine learning in the assessment of Parkinsonism holds promise for clinical care, potentially providing a more-efficient, extensive, and remote monitoring system. Additionally, machine learning approaches could greatly benefit clinical trials by offering a quantitative, objective digital biomarker for Parkinsonism.
However, concerns regarding costs and potential issues with patient compliance regarding wearable sensors have led to growing interest in video-based analysis. This method offers a non-contact, unobtrusive alternative for the automated assessment of motor symptoms in PD [12,13]. That said, there are technical challenges associated with the remote interpretation of video recordings [5]. A significant limitation of existing video-based assessments of PD is their focus on only a subset of the clinically relevant features detailed in the UPDRS instructions, often resulting in limited clinical interpretation of the models. Given that UPDRS is grounded in decades of clinical expertise and encompasses a wide range of bradykinesia symptoms in Parkinsonism, both the comprehensiveness and explainability of any model are vital. Furthermore, many studies have relied on solely collecting data in a standardized video collection environment, raising concerns about the sustainability and adaptability of these models. Addressing these gaps, our study aimed to develop a machine-learning model for the automatic prediction of UPDRS bradykinesia scores from finger tapping videos. We focused on enhancing the robustness of our model by extracting features that are in strict accordance with the clinical guidelines provided in the UPDRS, employing a white-box model for better explainability. This approach also involves incorporating a variety of video collection environments to ensure broader applicability and resilience of the model in diverse settings.

3. Method

Recognizing that data collection is a persistent and long-term pursuit, with our data repository steadily growing, we present a framework, illustrated in Figure 1, to maintain the lifecycle of the proposed system. Our framework consists of three sequential and independent cycles: a data cycle, a model learning cycle, and an algorithm deployment cycle.
Figure 1. The main structure of the proposed framework. The pipeline consists of three distinct cycles: data cycle, model learning cycle, and algorithm deployment cycle.

3.1. Data Cycle

The data cycle consists of three parts; data collection, data preprocessing, and data annotation.
Data collection: Patient-generated videos can be sourced from diverse platforms, comprising self-recorded, home-based footage and professionally captured videos within laboratory settings by healthcare practitioners. This compilation may include a range of data forms: structured and unstructured data (encompassing text, images, and videos), as well as external data acquired via Application Programming Interfaces (APIs).
Data preprocessing: Following data collection, the preprocessing phase is undertaken to transform the data into a format suitable for further analysis and model training. This phase includes handling missing values, eliminating outliers, normalizing or scaling features, and ensuring data format uniformity. We employed a range of methods to process the initial video data, as illustrated in Figure 2.
Figure 2. Data preprocessing flowchart.
  • Upscaling: Upscaling is employed to increase the resolution or dimensions of images or videos. This enhancement improves the quality of low-resolution data, thereby strengthening the model’s ability to extract significant features [41,42].
  • Removing motion blur: Motion blur occurs in images or videos as a result of camera wobble or object motion [43]. When collecting multi-source data from various devices and environments, it is essential to remove motion blur to regain the sharpness and clarity of the data.
  • Denoising: Noise may be introduced during the image- or video-acquisition process, leading to a degradation in quality and potentially hindering the capacity to extract meaningful features. A denoising step is often employed to reduce noise from the data while retaining the details.
  • Frame interpolation: This technique generates additional frames in videos with low frame rates [44]. This process estimates the motion between existing frames and synthesizes new ones, thereby smoothing the motion and enhancing the temporal resolution.
These measures are designed to enhance data quality and facilitate the extraction of meaningful information for more-effective model training.
Data annotation: In this step, domain experts, typically clinicians, are involved in providing UPDRS scores or labels for the patient’s data, which serve as a reference for model training and evaluation [45]. It is imperative to establish precise and uniform guidelines for the scoring criteria in the annotation process. As shown in Table 1, the UPDRS documentation offers detailed instructions on how clinicians should allocate scores to patients’ data. Our study adopted a dual-annotator system. The primary annotator, an experienced Movement Disorder Specialist, annotated all the videos. In parallel, a secondary annotator, also skilled in movement disorders, was tasked with validation, acting as a quality control to guarantee the accuracy of the annotations. These annotators were blind to each other’s ratings throughout the process. To ascertain the annotations’ quality and reliability, we evaluated the inter-annotator agreement. It is essential to manage any discrepancies and disagreements among clinicians during the data-annotation process. This control is necessary to maintain the reliability and consistency of the annotated data.
In practice, the data cycle is iterative to allow the data repository to grow. As the data volume increases, this repetitive method facilitates continuous improvements and modifications to the models, leading to enhanced performance over time [45].

3.2. Model Learning Cycle

The model learning cycle consists of three key stages: hand pose estimation, feature extraction, and machine learning.
Initially, the 3D hand pose is extracted using the most-recent version of MediaPipe, updated in February 2023, to generate time series data of crucial joint positions [36]. MediaPipe was selected over other deep learning libraries for hand pose estimation due to its efficient performance in various recording contexts and its relatively lightweight architecture. The raw time series data of key joint positions are represented as K = { K j ( t ) } , where j means the j-th key point, j = 1 , 2 , . . , 20 , and t denotes the image frame index (for t = 1 , 2 , . . . , L ). We calculated the time series of the thumb–index finger distance, denoted as I with length L, which corresponds to the distance between points K 4 and K 8 in the finger tapping task. In the second stage, we extracted 15 specific features for each video, as detailed in Table 3. These features encompass two demographic elements and 13 kinematic features, which were extracted through signal processing of the time series data. These kinematic features are aligned with the guidelines of the UPDRS instructions.
Table 3. Fifteen features extracted from the finger tapping time series.
Lastly, based on these features, we trained a machine learning model.

3.2.1. Time Series Preprocessing

For the time series input I , p = [ p 1 , p 2 , . . . , p n , . . . , p N ] represents the peaks, p n represents the n-th peak, and A n represents the amplitude of the corresponding peak p n . Given a hand model, as shown in Figure 3, K j = ( k x , j , k y , j , k z , j ) represents the position of the j-th keypoint. The palm size at time t is estimated using K 0 and K 9 as:
P a l m t = ( k x , 0 k x , 9 ) 2 + ( k y , 0 k y , 9 ) 2 + ( k z , 0 k z , 9 ) 2 | t
Figure 3. Illustration of the hand keypoints.
Then, normalization is performed to make P a l m t more suitable for subjectwise comparison, as below:
I n o r m = I P a l m t
Following this, we applied Empirical Mode Decomposition (EMD) to decompose I n o r m into a series of Intrinsic Mode Functions (IMFs) and a residual component, as depicted in Figure 4. Each IMF captures a distinct oscillatory mode characterized by variable frequencies and amplitudes. In our approach, we selected the first three IMFs as the primary components of the signal, striking a balance between simplifying the data and preserving vital information. However, even after EMD-based decomposition, the signal may contain noise or unwanted artifacts, such as those arising from finger tremors. To address this, we employed a denoising and smoothing process to eliminate such disturbances.
Figure 4. Depiction of the time series data preprocessing pipeline. The input time series, denoted as I , corresponds to the distance between K 4 and K 8 during the finger tapping task. The X−axis indicates the frame index within an image sequence, reflecting the duration of the finger tapping action. The Y−axis represents the normalized amplitude of finger movement, adjusted according to the palm size of each subject.
After preprocessing and cleaning the time series data, the subsequent step involved the detection of peaks or specific events of interest within the signal. The peak-detection module was applied to find local maxima or minima, which served as key points in the data.

3.2.2. Feature Extraction

Feature vector: demographics. In our model, we incorporated two demographic factors: age and sex. This inclusion was based on the observation that older individuals may exhibit signs of bradykinesia related to aging, independent of Parkinson’s Disease (PD), and that both age and sex can influence the progression of PD [1,46,47]. While the Unified Parkinson’s Disease Rating Scale (UPDRS) instructs raters to evaluate based on observable symptoms, there is a possibility that human annotators may subconsciously consider these demographic characteristics. Therefore, we chose to include age and sex as features in our model. Furthermore, given our goal of assessing the model’s explainability, incorporating these demographic factors may offer a more-comprehensive understanding that might not be immediately evident through the UPDRS guidelines alone.
Feature vector: amplitude. Based on the time series data, we extracted 5 amplitude-related features. First, according to Equations (3) and (4), we extracted F a m p a v g and F a m p v a r as:
F a m p a v g = n = 1 N A n N
F a m p v a r = n = 1 N ( A n F a m p a v g ) 2 N
Then, based on the peaks P, we applied piecewise linear regression to model the distinct patterns of different regions of the data, as shown in Figure 5:
y = α 0 + F a m p α 1 x + F a m p α 2 ( x F a m p b p ) I ( x > F a m p b p ) + ϵ
where α 0 is the intercept, F a m p α 1 and F a m p α 2 represent the slope coefficients of two segments, respectively, and ϵ represents the error term or residual, accounting for the deviations of the actual data points from the fitted model. I ( · ) is an indicator function that takes a value of 1 if x is greater than a specific breakpoint F a m p b p and 0 otherwise.
Figure 5. Extraction of F a m p α 1 , F a m p α 2 , and F a m p b p .
Feature vector: velocity. We extracted 5 velocity-related features. Here, we have
V n = A n + A n 1 Δ t n
where Δ t n = t n t n 1 is the time interval between the n-th and ( n 1 ) -th peak. Then, we have
F v e l a v g = n = 1 N 1 V n N 1
F v e l v a r = n = 1 N 1 ( V n F v e l a v g ) 2 N 1
We also extracted another 3 features, F v e l α 1 , F v e l α 2 , and F v e l b p , from the piecewise linear regression fitting of the velocity time series, to estimate the velocity trend.
y = α 0 + F v e l α 1 x + F v e l α 2 ( x F v e l b p ) I ( x > F v e l b p ) + ϵ
Feature vector: halt and hesitation. In PD, “halts or hesitations” are characterized by brief instances of freezing or a delayed ability to initiate movements, often seen during repetitive actions. While the concept of the freezing of gait, essentially halts and hesitations during walking, is well-documented in PD [48], there is a scarcity of detailed definitions and quantitative assessments of “halts and hesitations” in upper limb movements during repetitive tasks.
Through extensive discussions to clarify “halts” and “hesitation” within our study, we defined “halts” as moments of complete stoppage, characterized by zero amplitude. In contrast, “hesitation” was identified as a noticeable reduction in velocity, registering values above zero, but significantly below a predefined empirical threshold. This distinction was quantitatively established in Equation (11), where the “hesitation” threshold was set at 20% of the average amplitude observed immediately before and after the occurrence. More specifically, we define the “halt and hesitation” in finger tapping tasks as follows: given a time series of the distance between the keypoints K 4 and K 8 , p = [ p 1 , p 2 , . . . , p n , . . . , p N ] represents the detected peaks. A curve fitting was applied to p , aiming to estimate the trend of the peaks. Assuming p n to be the peak when halt and hesitation happened (the red arrow in Figure 6), p ^ n represents the predicted peak based on the fitted model. The residual between the observed value and the predicted value is denoted by | p ^ n p n | . Then, the feature “halt and hesitation” could be defined as:
F h h 1 if   | p ^ n p n |   θ 0 if   | p ^ n p n |   < θ .
Figure 6. Illustration of halt and hesitation. The red row indicates the precise moment when the halt and hesitation occur.
As demonstrated in Figure 6, in this situation, the patient may experience a sudden interruption or drop in his/her ability to perform repetitive finger tapping movements. The threshold θ is defined as:
θ = α · A n + 1 + A n 1 2
where α is a parameter controlling the detection threshold. In our framework, we empirically set α to 0.2.

3.2.3. Machine Learning

We chose the Decision Tree approach as our primary classification method to improve the interpretability and quantification of assessments tailored to our specific task. Decision Trees offer clear interpretability and the ability to visually represent the model, facilitating an uncomplicated explanation of the decision-making process. Inherently proficient in managing categorical data, Decision Tree models are particularly suitable for modeling ordinal UPDRS scores. Additionally, the inherent transparency and interpretability of Decision Tree models correspond well with the explicit criteria set forth by the UPDRS. This alignment ensures that our model’s predictions are not only easily understandable, but also hold clinical relevance.
The Pearson correlation coefficients between the 15 features and the UPDRS scores are presented in Figure 7. Redundant or overlapping features can lead to inefficient splits and diminished predictive capability in the model. However, as we identified a high correlation solely between the frequency and velocity of tapping, we decided to maintain all 15 original features in our analysis. This approach was chosen to ensure that the unique contribution of each feature is preserved.
Figure 7. Correlations between the 15 extracted features and UPDRS scores.

3.3. Algorithm Deployment Cycle

Model experimentation: This initial phase involved the training and evaluation of various models using different algorithms, hyperparameters, and subsets of data. The purpose of model experimentation is to identify the most-promising model based on its performance across various evaluation metrics.
Model updating: After selecting a suitable model, further refinement and improvement are often necessary. This updating process includes retraining the chosen model with additional data or altered hyperparameters to boost its performance. The goal of this iterative procedure is to enhance the model’s accuracy, robustness, and ability to generalize.
Deployment: The deployment phase involves making the refined and updated model accessible in a production setting. This stage encompasses establishing the required infrastructure to host and operate the model, which may involve servers, cloud services, or containerization platforms. Additionally, it includes the development of the necessary APIs or interfaces to enable the user or system access to the model for predictions or analyses.
The deployment cycle in MLOps begins with model experimentation, where a variety of models are trained and assessed. Following the selection of the most-effective model, it enters the updating stage for further improvement. The enhanced model is then deployed in a production environment, ready for practical application.
It is crucial to recognize that this deployment cycle is often iterative and ongoing. As new data are introduced, models require updates and redeployment to meet shifting demands. Continuous monitoring, testing, and feedback are integral to the deployment cycle, ensuring the deployed model maintains its performance, reliability, and efficacy in real-world settings.

4. Experiments

4.1. Participants and Dataset

RGB, single-view videos were collected in two settings with different camera and supervision setups: in-lab and in-clinic. In the laboratory setting, patients were invited to perform the finger tapping task as outlined in the MDS-UPDRS guidelines. This task was carried out in a spacious laboratory room. A research assistant was present to guide the patients through the task, which was captured by eight cameras (Lorex 4K Bullet IP Camera/E841CA-E from Lorex Technology Inc., Toronto, ON, Canada ) mounted on the walls. The raw data from these cameras had a resolution of 3840 × 2160 px and a frame rate of 15/s. Of the eight camera angles, our primary focus was on the single camera view that provided the clearest, most-unobstructed perspective of the hand movements. While not a central aspect of this study, an inertial sensor was also attached to the patient’s hand during recording, visible as a black band on the palm in Figure 1. In the clinic setting, patients were recruited to perform various tasks from the MDS-UPDRS, including finger tapping, in a designated room within the waiting area of the Movement Disorders Clinic. These tasks were recorded using an AI camera, OAK-D (Luxonis, Littleton, CO, USA), installed on an iMac 24” desktop (Apple, Cupertino, CA, USA). In this setting, patients received little or no assistance from a research assistant in using the app for recording. The raw RGB data collected in the clinic had a resolution of 1920 × 1080 px and a frame rate of 60 fps.
A total of 65 patients diagnosed with PD were recruited to do the UDPRS assessment. The clinical diagnosis for each patient was confirmed based on the United Kingdom Parkinson’s Disease Society Brain Bank clinical diagnostic criteria. To ensure the quality of the data, specialists manually inspected the videos, applying several criteria to evaluate their suitability. Videos were excluded if they exhibited poor visibility of motion, such as the hands frequently moving out of the camera frame, a significant distance between the hands and the camera, inconsistent execution of tasks, or frequent interruptions (verbal or physical) either by the patients or bystanders. After the exclusion process based on these quality criteria, 75 videos from 50 patients were deemed suitable for assessing the effectiveness of the proposed method. This dataset comprised 20 subjects recorded in a laboratory setting and 30 subjects in a clinic environment, as detailed in Table 4. For the purposes of our study, 60 videos were designated for the training set and 15 videos for the testing set. The selection of testing subjects was randomized, ensuring a balanced representation of different categories. The Research Ethics Board of the University of British Columbia approved this study, and all participants provided written consent. All individuals were evaluated using the MDS-UPDRS, with the entire evaluation process captured on video. Due to regulations safeguarding patient privacy, these videos are not permitted to be made publicly available.
Table 4. Baseline characteristics of the participants (n = 50) and the videos (n = 75).

4.2. Evaluation Matrix

To evaluate the performance of our model, we employed well-known metrics such as accuracy, precision, recall, and the F1-score. These metrics are defined as follows:
A c c u r a c y = T P + T N T P + F N + F P + T N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where T P , T N , F P , and F N represent “true positive”, “true negative”, “false positive”, and “false negative”, respectively, and the F score (F1 when β = 1 and F2 when β = 2) is defined as:
F β = 1 + β 2 · Precision · Recall β 2 · Precision + Recall
In our evaluation matrix for accuracy, we considered two scenarios:
  • y p r e d = y g t , which is reported as “Accuracy(t1)”, represents that the prediction y p r e d is equal to the ground truth y g t .
  • y p r e d [ y g t 1 , y g t , y g t + 1 ] , reported as “Accuracy(t2)”, represents that the prediction is within a range of the ground truth, which could also be called “acceptable” accuracy. This situation takes the inter-annotator agreement into consideration, and we will further analyze this situation in the result part.

4.3. Method for Comparison

To assess the performance of our proposed method, we compared it against two recent studies: AEMPD-FCN [49] and FTTST [23]. It is important to note that there might be significant differences in data quality, volume, and time series length between our study and those in the referenced work, which could potentially influence the comparative results. Additionally, we explored the effectiveness of different classifiers by including results from three other methods. These methods all employ the same feature-extraction framework, but differ in their classification approaches, such as Support Vector Machine (SVM) and Random Forest (RF). Specifically, b a s e l i n e - S P is the baseline model that follows the UDPRS criteria precisely. To some extent, this criterion is overly strict for some data points situated at the boundaries. For example, a slight decrease in amplitude or velocity could be classified as decrementing features. F15- S V M - G A uses the 15 features as the input vector to the SVM classifier, where the Genetic Algorithm is used to search for the best parameters. F15- R F uses Random Forest for classification, and in F15- D T , the D T represents a Decision Tree.

5. Results and Discussion

5.1. Labeling UPDRS Finger Tapping Videos

In our study, we calculated the level of agreement between annotators, as shown in Table 5. “Complete”: This term refers to the percentage of instances where both annotators assigned an identical score to a specific piece of data. “Acceptable”: This category covers cases where the scores given by the two annotators differed by either 0 or 1. It represents a more-lenient agreement threshold. “None”: This indicates the percentage of cases where the discrepancy in scoring between the two annotators was more than 1. The data presented in Table 5 highlights the inherent difficulty in achieving complete consensus among human annotators, even when they are experts in their field, as has been previously observed [50]. Acknowledging this challenge, we report two types of accuracy metrics in our experiments, namely Accuracy(t1) and Accuracy(t2), as defined in Section 4.2. These metrics are intended to evaluate the efficacy of our proposed method, taking into account the varying levels of agreement among annotators.
Table 5. Inter-annotator agreement analysis in the proposed task.

5.2. Classification of UPDRS Finger Tapping Score

First, the main classification task was performed on the finger tapping features using the clinical ratings. The proposed method achieved a high accuracy rate, an Accuracy(t1) of 80% and an Accuracy(t2) of 93.3%. Table 6 lists the specific accuracy(t1 and t2), precision, recall, and F-score of the final prediction results on five UPDRS scores. The structure of the Decision Tree and the confusion matrix of the classification results are shown in Figure 8 and Figure 9, respectively.
Table 6. Performance results of UPDRS classification using different classifiers.
Figure 8. Structure of the proposed Decision Tree.
Figure 9. Confusion matrix of the predicted score labels and the true labels.
Clinical interpretation: Overall, the Decision Tree our model produced (as shown in (Figure 8)) aligns well with the MDS-UPDRS rating guidelines. These guidelines instruct raters to evaluate finger tapping based on: (i) amplitude, (ii) velocity, (iii) decrement in amplitude or velocity during repetitive tapping, and (iv) the presence of halts or hesitations [3]. The root node in our Decision Tree (Node 0) utilizes the “mean amplitude” feature, dividing the data into two paths and indicating amplitude as a pivotal factor in the finger tapping assessment. In the second layer, the nodes “vel_alpha2” and “vel_alpha1” reflect the velocity decrement, suggesting that greater decrements often correspond with lower UPDRS scores. This points to the relevance of the “decrement in amplitude or velocity” criterion in cases of mild bradykinesia, while in more-severe cases, the velocity notably reduces from the beginning of the task. Further nodes incorporate amplitude, velocity, and decrement aspects, aligning conceptually with the MDS-UPDRS guidelines, highlighting the model’s ability to offer more-quantifiable data and reduce intra-rater variability compared to human ratings.
Additional clinical insights emerged from the Decision Tree. Notably, “halts and hesitations” had minimal impact on the model. This could be due to the rarity and lesser correlation of upper limb freezing with PD severity compared to gait freezing. Our model’s interpretation suggests a non-linear relationship between these halts and overall bradykinesia severity [51]. It is important to recognize the limitations posed by the relatively small video dataset used for the AI training, potentially under-representing upper limb freezing episodes Furthermore, our study’s definition of halts and hesitations, based purely on significant amplitude reduction without a time duration criterion, may not fully align with clinical descriptions of a halt as a prolonged pause. Future studies, especially with larger datasets, might need a more-refined definition, possibly incorporating time-based thresholds for a better characterization of these phenomena. Lastly, while we considered demographic variables like age and gender, their absence in the Decision Tree underscores that UPDRS ratings focus on motor performance irrespective of demographic factors.

6. Limitations and Future Work

Our study marks a step forward in automated Parkinson’s Disease (PD) symptom assessment, yet it is important to recognize its limitations. Firstly, the limited number of patient videos used for training the model, though yielding promising accuracy, may not fully represent the broad spectrum of PD symptoms, including pronounced halts or hesitation phenomena. A more-extensive dataset will be critical for enhancing the model’s generalizability and robustness. Secondly, the study’s focus was solely on the finger tapping task, an important, but not complete aspect of bradykinesia evaluation. Other key tasks in assessing bradykinesia, such as hand movements and leg agility, were not included. Furthermore, PD involves a variety of motor symptoms beyond bradykinesia. Future research will aim to extend our video-based approach to cover a broader range of motor tasks from the UPDRS, enabling a more-comprehensive assessment of PD motor symptoms. Thirdly, the study’s setting was restricted to controlled clinical environments, which may not fully replicate the varied conditions in which the model might be used. Factors like changing lighting, patient positioning, and camera quality or angles, crucial for video quality, were not thoroughly examined. Additionally, our approach of manually excluding low-quality videos needs to be replaced by automated algorithms in future work. These algorithms should be capable of distinguishing between low- and high-quality videos based on predefined quality thresholds. Finally, while not directly related to our scientific methodology, the ethical aspects of patient privacy and data security are crucial. As we move towards real-world applications of this technology, a multidisciplinary approach involving legal, ethical, clinical, engineering, and patient stakeholders is essential. This collaborative effort is key to addressing the complex ethical landscape, ensuring the protection of patient privacy and data security.

7. Conclusions

The proposed innovative vision-based classification model offers a promising solution to automate UPDRS scoring accurately during the finger tapping test, a crucial aspect of motor evaluation in Parkinson’s Disease (PD). Our approach adheres to the UPDRS guidelines by extracting the relevant features and quantifying them, enabling a thorough analysis. The proposed method was designed to adapt to future changes in patient demographics, disease subtypes, and technological advancements. We are actively working to expand our data collection using a variety of devices and in different settings. This effort is crucial to enhance the model’s applicability and effectiveness across diverse contexts and technological platforms.
Our comprehensive experiments demonstrated the effectiveness and reliability of the proposed method, producing robust results with datasets obtained both in laboratory and clinical settings.

Author Contributions

Conceptualization and methodology, T.Y., M.J.M., and Z.J.W.; experiments, T.Y.; PD data collection, annotation, and curation, K.W.P. and M.J.M.; writing—original draft, T.Y. and K.W.P.; writing—review and editing, T.Y., K.W.P., Z.J.W., and M.J.M.; validation and visualization, T.Y. and K.W.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) and the Canadian Institutes of Health Research (CIHR). This research was partially supported by a Collaborative Health Research Project (CHRP) grant (MJM), and the John Nichol Chair in Parkinson’s Research.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the UBC Clinical Research Ethics Board of the Department of Medicine (H18-03548, 4/5/2019).

Data Availability Statement

The data are not allowed to be made public according to the relevant regulations protecting patient privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Balestrino, R.; Schapira, A. Parkinson disease. Eur. J. Neurol. 2020, 27, 27–42. [Google Scholar] [CrossRef] [PubMed]
  2. Fish, J. Unified Parkinson’s Disease Rating Scale. In Encyclopedia of Clinical Neuropsychology; Kreutzer, J.S., DeLuca, J., Caplan, B., Eds.; Springer: New York, NY, USA, 2011; pp. 2576–2577. [Google Scholar] [CrossRef]
  3. Goetz, C.G.; Tilley, B.C.; Shaftman, S.R.; Stebbins, G.T.; Fahn, S.; Martinez-Martin, P.; Poewe, W.; Sampaio, C.; Stern, M.B.; Dodel, R.; et al. Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS): Scale presentation and clinimetric testing results. Mov. Disord. Off. J. Mov. Disord. Soc. 2008, 23, 2129–2170. [Google Scholar] [CrossRef]
  4. Morinan, G.; Dushin, Y.; Sarapata, G.; Rupprechter, S.; Peng, Y.; Girges, C.; Salazar, M.; Milabo, C.; Sibley, K.; Foltynie, T.; et al. Computer vision quantification of whole-body Parkinsonian bradykinesia using a large multi-site population. NPJ Park. Dis. 2023, 9, 10. [Google Scholar] [CrossRef]
  5. Park, S.H.; Seo, N.Y.; Hwang, S.M.; Park, H.Y.; Jung, Y.J. Quantifying Finger-tapping-test Scores using a Three-dimensional Motion Analysis Program: A Preliminary Study. J. Magn. 2022, 27, 411–420. [Google Scholar] [CrossRef]
  6. Lu, M.; Poston, K.; Pfefferbaum, A.; Sullivan, E.V.; Fei-Fei, L.; Pohl, K.M.; Niebles, J.C.; Adeli, E. Vision-based estimation of MDS-UPDRS gait scores for assessing Parkinson’s disease motor severity. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020. Proceedings, Part III 23. pp. 637–647. [Google Scholar]
  7. Ayaz, Z.; Naz, S.; Khan, N.H.; Razzak, I.; Imran, M. Automated methods for diagnosis of Parkinson’s disease and predicting severity level. Neural Comput. Appl. 2023, 35, 14499–14534. [Google Scholar] [CrossRef]
  8. Mughal, H.; Javed, A.R.; Rizwan, M.; Almadhor, A.S.; Kryvinska, N. Parkinson’s disease management via wearable sensors: A systematic review. IEEE Access 2022, 10, 35219–35237. [Google Scholar] [CrossRef]
  9. Balakrishnan, A.; Medikonda, J.; Namboothiri, P.K.; Natarajan, M. Role of wearable sensors with machine learning approaches in gait analysis for Parkinson’s disease assessment: A review. Eng. Sci. 2022, 19, 5–19. [Google Scholar] [CrossRef]
  10. Joshi, R.; Bronstein, J.M.; Keener, A.; Alcazar, J.; Yang, D.D.; Joshi, M.; Hermanowicz, N. PKG movement recording system use shows promise in routine clinical care of patients with Parkinson’s disease. Front. Neurol. 2019, 10, 1027. [Google Scholar] [CrossRef] [PubMed]
  11. Hadley, A.J.; Riley, D.E.; Heldman, D.A. Real-world evidence for a smartwatch-based parkinson’s motor assessment app for patients undergoing therapy changes. Digit. Biomarkers 2021, 5, 206–215. [Google Scholar] [CrossRef] [PubMed]
  12. Schneider, R.B.; Myers, T.L.; Tarolli, C.G.; Amodeo, K.; Adams, J.L.; Jensen-Roberts, S.; Dorsey, E.R. Remote administration of the MDS-UPDRS in the time of COVID-19 and beyond. J. Park. Dis. 2020, 10, 1379–1382. [Google Scholar] [CrossRef]
  13. Park, K.W.; Wu, H.J.; Yu, T.; Mahal, R.; Mirian, M.S.; McKeown, M.J. Potential Pitfalls of Remote and Automated Video Assessments of Movements Disorders. Mov. Disord. 2023, 38, 504–506. [Google Scholar] [CrossRef] [PubMed]
  14. Tripathi, S.; Malhotra, A.; Qazi, M.; Chou, J.; Wang, F.; Barkan, S.; Hellmers, N.; Henchcliffe, C.; Sarva, H. Clinical Review of Smartphone Applications in Parkinson’s Disease. Neurologist 2022, 27, 183–193. [Google Scholar] [CrossRef]
  15. Jeon, H.; Lee, W.; Park, H.; Lee, H.J.; Kim, S.K.; Kim, H.B.; Jeon, B.; Park, K.S. Automatic classification of tremor severity in Parkinson’s disease using a wearable device. Sensors 2017, 17, 2067. [Google Scholar] [CrossRef]
  16. Mahadevan, N.; Demanuele, C.; Zhang, H.; Volfson, D.; Ho, B.; Erb, M.K.; Patel, S. Development of digital biomarkers for resting tremor and bradykinesia using a wrist-worn wearable device. NPJ Digit. Med. 2020, 3, 5. [Google Scholar] [CrossRef]
  17. Borzì, L.; Sigcha, L.; Rodríguez-Martín, D.; Olmo, G. Real-time detection of freezing of gait in Parkinson’s disease using multi-head convolutional neural networks and a single inertial sensor. Artif. Intell. Med. 2023, 135, 102459. [Google Scholar] [CrossRef] [PubMed]
  18. Liu, P.; Yu, N.; Yang, Y.; Yu, Y.; Sun, X.; Yu, H.; Han, J.; Wu, J. Quantitative assessment of gait characteristics in patients with Parkinson’s disease using 2D video. Park. Relat. Disord. 2022, 101, 49–56. [Google Scholar] [CrossRef] [PubMed]
  19. Talitckii, A.; Kovalenko, E.; Shcherbak, A.; Anikina, A.; Bril, E.; Zimniakova, O.; Semenov, M.; Dylov, D.V.; Somov, A. Comparative study of wearable sensors, video, and handwriting to detect Parkinson’s disease. IEEE Trans. Instrum. Meas. 2022, 71, 2509910. [Google Scholar] [CrossRef]
  20. Ngo, Q.C.; Motin, M.A.; Pah, N.D.; Drotár, P.; Kempster, P.; Kumar, D. Computerized analysis of speech and voice for Parkinson’s disease: A systematic review. Comput. Methods Programs Biomed. 2022, 226, 107133. [Google Scholar] [CrossRef]
  21. Aouraghe, I.; Khaissidi, G.; Mrabti, M. A literature review of online handwriting analysis to detect Parkinson’s disease at an early stage. Multimed. Tools Appl. 2023, 82, 11923–11948. [Google Scholar] [CrossRef]
  22. Yang, Y.; Yuan, Y.; Zhang, G.; Wang, H.; Chen, Y.C.; Liu, Y.; Tarolli, C.G.; Crepeau, D.; Bukartyk, J.; Junna, M.R.; et al. Artificial intelligence-enabled detection and assessment of Parkinson’s disease using nocturnal breathing signals. Nat. Med. 2022, 28, 2207–2215. [Google Scholar] [CrossRef]
  23. Guo, Z.; Zeng, W.; Yu, T.; Xu, Y.; Xiao, Y.; Cao, X.; Cao, Z. Vision-based finger tapping test in patients with Parkinson’s disease via spatial-temporal 3D hand pose estimation. IEEE J. Biomed. Health Inform. 2022, 26, 3848–3859. [Google Scholar] [CrossRef]
  24. Ospina, B.M.; Chaparro, J.A.V.; Paredes, J.D.A.; Pino, Y.J.C.; Navarro, A.; Orozco, J.L. Objective arm swing analysis in early-stage Parkinson’s disease using an RGB-D camera (Kinect®). J. Park. Dis. 2018, 8, 563–570. [Google Scholar] [CrossRef] [PubMed]
  25. Muñoz Ospina, B.; Valderrama Chaparro, J.A.; Arango Paredes, J.D.; Castaño Pino, Y.J.; Navarro, A.; Orozco, J.L. Age matters: Objective gait assessment in early parkinson’s disease using an RGB-D camera. Park. Dis. 2019, 2019, 5050182. [Google Scholar] [CrossRef] [PubMed]
  26. Fernández-González, P.; Carratalá-Tejada, M.; Monge-Pereira, E.; Collado-Vázquez, S.; Sánchez-Herrera Baeza, P.; Cuesta-Gómez, A.; Oña-Simbaña, E.D.; Jardón-Huete, A.; Molina-Rueda, F.; Balaguer-Bernaldo de Quirós, C.; et al. Leap motion controlled video game-based therapy for upper limb rehabilitation in patients with Parkinson’s disease: A feasibility study. J. Neuroeng. Rehabil. 2019, 16, 133. [Google Scholar] [CrossRef]
  27. Butt, A.H.; Rovini, E.; Dolciotti, C.; De Petris, G.; Bongioanni, P.; Carboncini, M.; Cavallo, F. Objective and automatic classification of Parkinson disease with Leap Motion controller. Biomed. Eng. Online 2018, 17, 1–21. [Google Scholar] [CrossRef]
  28. Khwaounjoo, P.; Singh, G.; Grenfell, S.; Özsoy, B.; MacAskill, M.R.; Anderson, T.J.; Çakmak, Y.O. Non-Contact Hand Movement Analysis for Optimal Configuration of Smart Sensors to Capture Parkinson’s Disease Hand Tremor. Sensors 2022, 22, 4613. [Google Scholar] [CrossRef]
  29. Moshkova, A.; Samorodov, A.; Voinova, N.; Volkov, A.; Ivanova, E.; Fedotova, E. Parkinson’s disease detection by using machine learning algorithms and hand movement signal from LeapMotion sensor. In Proceedings of the 2020 26th Conference of Open Innovations Association (FRUCT), Yaroslavl, Russia, 20–24 April 2020; pp. 321–327. [Google Scholar]
  30. Sarapata, G.; Dushin, Y.; Morinan, G.; Ong, J.; Budhdeo, S.; Kainz, B.; O’Keeffe, J. Video-based activity recognition for automated motor assessment of Parkinson’s disease. IEEE J. Biomed. Health Inform. 2023, 10, 5032–5041. [Google Scholar] [CrossRef]
  31. Islam, M.S.; Rahman, W.; Abdelkader, A.; Yang, P.T.; Lee, S.; Adams, J.L.; Schneider, R.B.; Dorsey, E.; Hoque, E. Using AI to Measure Parkinson’s Disease Severity at Home. arXiv 2023, arXiv:2303.17573. [Google Scholar]
  32. Park, K.W.; Lee, E.J.; Lee, J.S.; Jeong, J.; Choi, N.; Jo, S.; Jung, M.; Do, J.Y.; Kang, D.W.; Lee, J.G.; et al. Machine learning–based automatic rating for cardinal symptoms of Parkinson disease. Neurology 2021, 96, e1761–e1769. [Google Scholar] [CrossRef] [PubMed]
  33. Scott, B.; Seyres, M.; Philp, F.; Chadwick, E.K.; Blana, D. Healthcare applications of single camera markerless motion capture: A scoping review. PeerJ 2022, 10, e13517. [Google Scholar] [CrossRef] [PubMed]
  34. Moro, M.; Marchesi, G.; Hesse, F.; Odone, F.; Casadio, M. Markerless vs. marker-based gait analysis: A proof of concept study. Sensors 2022, 22, 2011. [Google Scholar] [CrossRef]
  35. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  36. Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.L.; Yong, M.G.; Lee, J.; et al. MediaPipe: A framework for building perception pipelines. arXiv 2004, arXiv:1906.08172. [Google Scholar]
  37. Fang, H.S.; Li, J.; Tang, H.; Xu, C.; Zhu, H.; Xiu, Y.; Li, Y.L.; Lu, C. Alphapose: Whole-body regional multi-person pose estimation and tracking in real-time. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 7157–7173. [Google Scholar] [CrossRef] [PubMed]
  38. Vignoud, G.; Desjardins, C.; Salardaine, Q.; Mongin, M.; Garcin, B.; Venance, L.; Degos, B. Video-Based Automated Assessment of Movement Parameters Consistent with MDS-UPDRS III in Parkinson’s Disease. J. Park. Dis. 2022, 12, 2211–2222. [Google Scholar] [CrossRef]
  39. Liu, W.; Lin, X.; Chen, X.; Wang, Q.; Wang, X.; Yang, B.; Cai, N.; Chen, R.; Chen, G.; Lin, Y. Vision-based estimation of MDS-UPDRS scores for quantifying Parkinson’s disease tremor severity. Med. Image Anal. 2023, 85, 102754. [Google Scholar] [CrossRef] [PubMed]
  40. Li, H.; Shao, X.; Zhang, C.; Qian, X. Automated assessment of Parkinsonian finger tapping tests through a vision-based fine-grained classification model. Neurocomputing 2021, 441, 260–271. [Google Scholar] [CrossRef]
  41. Liang, J.; Lugmayr, A.; Zhang, K.; Danelljan, M.; Van Gool, L.; Timofte, R. Hierarchical conditional flow: A unified framework for image super-resolution and image rescaling. arXiv 2021, arXiv:2108.05301. [Google Scholar]
  42. Pan, Z.; Li, B.; He, D.; Yao, M.; Wu, W.; Lin, T.; Li, X.; Ding, E. Towards bidirectional arbitrary image rescaling: Joint optimization and cycle idempotence. arXiv 2022, arXiv:2203.00911. [Google Scholar]
  43. Li, J.; Wang, L.; Xu, B. Research on Motion Blur Image of Infrared Target Deblurring Based on Wavelet Transform. J. Phys. Conf. Ser. 2020, 1570, 012080. [Google Scholar] [CrossRef]
  44. Shi, Z.; Xu, X.; Liu, X.; Chen, J.; Yang, M.H. Video frame interpolation transformer. arXiv 2022, arXiv:2111.13817. [Google Scholar]
  45. Mazurowski, M.A.; Dong, H.; Gu, H.; Yang, J.; Konz, N.; Zhang, Y. Segment anything model for medical image analysis: An experimental study. Med. Image Anal. 2023, 89, 102918. [Google Scholar] [CrossRef]
  46. Ho, P.W.L.; Leung, C.T.; Liu, H.; Pang, S.Y.Y.; Lam, C.S.C.; Xian, J.; Li, L.; Kung, M.H.W.; Ramsden, D.B.; Ho, S.L. Age-dependent accumulation of oligomeric SNCA/α-synuclein from impaired degradation in mutant LRRK2 knockin mouse model of Parkinson disease: Role for therapeutic activation of chaperone-mediated autophagy (CMA). Autophagy 2020, 16, 347–370. [Google Scholar] [CrossRef]
  47. Kang, K.W.; Choi, S.M.; Kim, B.C. Gender differences in motor and non-motor symptoms in early Parkinson disease. Medicine 2022, 101, e28643. [Google Scholar] [CrossRef] [PubMed]
  48. Nutt, J.G.; Bloem, B.R.; Giladi, N.; Hallett, M.; Horak, F.B.; Nieuwboer, A. Freezing of gait: Moving forward on a mysterious clinical phenomenon. Lancet Neurol. 2011, 10, 734–744. [Google Scholar] [CrossRef] [PubMed]
  49. Li, Z.; Lu, K.; Cai, M.; Liu, X.; Wang, Y.; Yang, J. An Automatic Evaluation Method for Parkinson’s Dyskinesia Using Finger Tapping Video for Small Samples. J. Med. Biol. Eng. 2022, 42, 351–363. [Google Scholar] [CrossRef]
  50. Williams, S.; Wong, D.; Alty, J.E.; Relton, S.D. Parkinsonian hand or clinician’s eye? Finger tap Bradykinesia interrater reliability for 21 movement disorder experts. J. Park. Dis. 2023, 13, 525–536. [Google Scholar] [CrossRef]
  51. Nieuwboer, A.; Vercruysse, S.; Feys, P.; Levin, O.; Spildooren, J.; Swinnen, S. Upper limb movement interruptions are correlated to freezing of gait in Parkinson’s disease. Eur. J. Neurosci. 2009, 29, 1422–1430. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.