Next Article in Journal
Systematic Review of Deep Learning Techniques in Skin Cancer Detection
Previous Article in Journal
Association Between Social Determinants of Health and Patient Portal Utilization in the United States
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Feature Selection and Classification for Parkinson’s Disease Using Deep Learning and Dynamic Bag of Features Optimization

1
Department of Computer Science and Engineering, Lovely Professional University, Jalandhar 144411, Punjab, India
2
Department of Artificial Intelligence and Machine Learning, Sreyas Institute of Engineering and Technology, Hyderabad 500068, Telangana, India
3
Department of Biomedical Sciences, Institute for Medical Science, Jeonbuk National University Medical School, Jeonju 54907, Republic of Korea
4
Department of Artificial Intelligence and Big Data, Woosong University, Daejeon 34606, Republic of Korea
*
Author to whom correspondence should be addressed.
BioMedInformatics 2024, 4(4), 2223-2250; https://doi.org/10.3390/biomedinformatics4040120
Submission received: 3 September 2024 / Revised: 14 October 2024 / Accepted: 1 November 2024 / Published: 12 November 2024

Abstract

Parkinson’s Disease (PD) is a neurological condition that worsens with time and is characterized bysymptoms such as cognitive impairment andbradykinesia, stiffness, and tremors. Parkinson’s is attributed to the interference of brain cells responsible for dopamine production, a substance regulating communication between brain cells. The brain cells involved in dopamine generation handle adaptation and control, and smooth movement. Convolutional Neural Networks are used to extract distinctive visual characteristics from numerous graphomotor sample representations generated by both PD and control participants. The proposed method presents an optimal feature selection technique based on Deep Learning (DL) and the Dynamic Bag of Features Optimization Technique (DBOFOT). Our method combines neural network-based feature extraction with a strong optimization technique to dynamically choose the most relevant characteristics from biological data. Advanced DL architectures are then used to classify the chosen features, guaranteeing excellent computational efficiency and accuracy. The framework’s adaptability to different datasets further highlights its versatility and potential for further medical applications. With a high accuracy of 0.93, the model accurately identifies 93% of the cases that are categorized as Parkinson’s. Additionally, it has a recall of 0.89, which means that 89% of real Parkinson’s patients are accurately identified. While the recall for Class 0 (Healthy) is 0.75, meaning that 75% of the real healthy cases are properly categorized, the precision decreases to 0.64 for this class, indicating a larger false positive rate.

Graphical Abstract

1. Introduction

Parkinson’s Disease (PD) is a neurological condition that worsens with time and mostly affects movement. Tremors, stiffness, bradykinesia (slowness of movement), and postural instability are some of its symptoms. Non-motor symptoms are also common and include sadness, sleep difficulties, and cognitive impairment [1,2]. The illness is brought on by the death of dopamine-producing neurons in the midbrain area known as the substantia nigra. One neurotransmitter that is essential for coordinating balanced and fluid muscular actions is dopamine. Although the precise origin of neuron degeneration is unknown, a mix of environmental and genetic variables is thought to be involved. After Alzheimer’s Disease, PD is the second most common neurodegenerative condition, affecting about 1% of the over-60 population [3]. PD is more common in males than in women, and its prevalence rises with age. PD cannot be definitively diagnosed; instead, the condition is managed with a combination of physical and neurological examinations, medical history, symptoms, and dopamine replacement medication response. While imaging tests such as Magnetic Resonance Imaging (MRI) and DaT scans are not commonly used to diagnose PD, they can be useful in ruling out other illnesses. Significant motor function impairments caused by PD make daily tasks including eating, dressing, and walking difficult. Patients may experience progressively worsening disabilities as the illness worsens. Patients’ and their families’ emotional well-being are significantly impacted by the prevalent conditions of depression, anxiety, and cognitive impairment.
Parkinson’s motor symptoms manifest when 60% to 80% of these cells are lost due to the inadequate production of dopamine. Recent research has identified a close connection between speech impairment and PD, leading to the development of classification algorithms for PD identification. Artificial Intelligence (AI) is leading the way in next-generation computing for data analytics, particularly in predictive edge analytics for high-risk diseases including PD. Deep Learning (DL) techniques play a crucial role in facilitating edge AI applications for real-time and enhanced data processing. In contrast to conventional methods, which mostly depend on the dynamic (kinematic and spatio-temporal) aspects of handwriting, we objectively assess the visual characteristics in this work in order to characterize graphomotor samples of PD patients. Even with Machine Learning (ML) advances, choosing the best attributes that greatly improve PD diagnostic accuracy continues to be a major difficulty. High dimensionality and superfluous or unnecessary characteristics are common problems with existing models, which can cause overfitting and decreased generalizability. Furthermore, a lot of research has concentrated on static feature sets rather than investigating dynamic strategies that may adjust to various patient data distributions. Research is still being done to find biomarkers for early diagnosis, improve our understanding of the mechanism of PD, and create medicines that alter the condition. DL and other AI advances have the potential to improve diagnosis, forecast the course of disease, and customize therapy regimens. Owing to the intricacy and variability of PD, feature selection (FS) and categorization provide a number of difficulties. The primary difficulties consist of:
  • High Dimensionality: Clinical, genetic, imaging, and biochemical indicators are only a few of the many aspects that biomedical data frequently contain. It is challenging to determine which features are most important for precise categorization because of this high dimensionality.
  • Heterogeneity of Data: Patients with PD might have very different symptoms and rates of progression, which results in heterogeneous data that makes feature selection more difficult.
  • Unbalanced Data: The distribution of PD datasets is frequently unbalanced, with fewer positive cases than healthy controls. This can cause bias in the classification models.
  • Redundancy and Noise: Biomedical data may have duplicate information and be noisy, which can impair the effectiveness of classification systems.
The proposed method makes use of the Dynamic Bag of Features Optimization Technique (DBOFOT) to overcome the shortcomings of previous studies in FS and classification for PD. By streamlining the classification process and dynamically choosing the most pertinent features, this technique seeks to improve overall performance. DBOFOT uses a dynamic FS procedure that changes depending on the particulars of the dataset. Long Short-Term Memory-DBOFOT (LSTM-DBOFOT) continuously modifies the feature set according to each feature’s relevance and contribution to classification accuracy, in contrast to static FS approaches that could ignore significant characteristics or keep unimportant ones [4]. The model can more efficiently manage the high complexity and variability of biomedical data dueto its adaptability. LSTM-DBOFOT improves classification accuracy by reducing noise and redundancy in the data by dynamically choosing the most relevant characteristics. By limiting the usage of only the most informative variables, the optimization strategy improves the model’s capacity to discriminate between PD patients and healthy controls. The shortcomings of earlier study [1,2] that had trouble with noisy and redundant data are addressed by this method. LSTM-DBOFOT uses optimization methods to make the FS process more efficient and less time-consuming. For managing large-scale biomedical datasets, which are typical in PD research, this efficiency is essential. By reducing the computing overhead of conventional FS techniques, the method enables more rapid and scalable analysis. This method improves the model’s interpretability by illuminating the chosen characteristics and how they contribute to the classification goal. For clinical applications, where knowing the importance of traits can help with improved diagnosis and treatment planning, this transparency is essential. Furthermore, by being dynamic, LSTM-DBOFOT addresses the variability problems in earlier studies and enhances the model’s generalizability across various datasets. Multimodal data sources including clinical, genetic, imaging, and gait data can be integrated with DBOFOT. A more complete picture of the patient’s state is provided by this integration, which results in more reliable and accurate categorization models. Studies that concentrate on a single data type have limitations; nevertheless, the model’s ability to handle many data kinds guarantees that it can utilize the entire range of information available.
Our model can effectively capture the temporal dependencies included in the data by integrating LSTM cells, which makes it a suitable fit for the intricate patterns linked to PD. By real-time feature input optimization, this hybrid methodology overcomes the drawbacks of conventional FS techniques and improves the performance of DL models. The major contribution includes:
  • The proposed method integrates an LSTM network with the BoF representation, offering a novel approach for PD prediction. It focuses on robust feature selection, and dynamic feature sets, improving classification accuracy.
  • The use ofDBOFOT has been used to handwriting-based PD identification. By allowing for adaptive feature combinations based on the significance of the data, it enhances sensitivity to minute handwriting differences that are common in individuals with Parkinson’s Disease.
  • The proposed work employs optimization techniques of transfer learning, cross-validation, or regularization to counter overfitting and ensure robust performance even with smaller datasets.
  • Handwriting patterns are subtle indications of motor deficits, and the Parkinson’s Disease handwriting (PaHaW) collection used in proposed analysis offers thorough information on handwriting patterns.
  • The deployment and continuous monitoring aspects of the proposed method are designed to maintain high accuracy and relevance over time.
The remaining contribution of this paper is structured as follows: Section 2 reviews various prediction methods. Section 3 details the implementation of the LSTM–BoF model and the dataset used, while Section 4 covers the evaluation procedure and analysis of the results on two different datasets. Finally, Section 5 concludes the work.

2. Literature Review

DL and ML models have been used more and more in the diagnosis of PD in recent years. Conventional ML methods, including Random Forest (RF) and Support Vector Machine (SVM), have demonstrated some effectiveness but are frequently limited in their scalability by the need for human feature extraction. By automatically learning hierarchical representations of data, DL models—in particular, CNN and LSTM networks—offer notable advances in feature extraction and classification tasks. The study of computer programs that use statistical models and algorithms to infer information through patterns and inference without explicit programming is known as ML [1]. As ML algorithms learning by training and they improve over time. These algorithms use methods for training models, then use the information they have learned to determine outputs on their own [2]. Moreover, ML systems are capable of adjusting to changing surroundings. According to the study by [3], the three most well-known signs of PD that may be seen in handwriting are “tremor”, “brachykinesia”, and “micrographia”. Maintaining the size and alignment of the generated graphomotor impressions becomes challenging for a patient with micrographia [4]. A supervised ML model that applies classification techniques to two-group classification issues is called an SVM. SVM is well-known for being quick and dependable, and it performs well when processing small datasets [3,4,5]. A prospective PD patient may take longer than normal to finish a graphomotor task if they have bradykinesia, or slowness of movement (caused by either motor or cognitive impairment) [6]. Uncontrollably moving back and forth, “tremors” are represented by asymmetrical character and drawing forms. The coexistence or independence of various Parkinsonian disorders is contingent upon the nature and course of the ailment.
The Decision Tree (DT) approach in ML systems partitions data into subsets with the training data divided into the smallest tree. In order to check the target classes in leaf nodes, this supervised classification approach uses split tests in internal nodes [7]. Because of their consistency and adaptability, decision trees are frequently employed in categorization [8,9,10,11,12]. A supervised learning technique that is both flexible and easy to use, RF frequently produces good results without requiring a lot of hyperparameter tweaking [13]. An RF is made up of randomly generated DTs that use majority vote to decide the expected output. Even if DTs alone might be less precise and dependable, the RF makes up for these drawbacks [14]. Unlike Logistic Regression (LR), which can only handle linear situations, SVM is superior at managing nonlinear problems. SVM determines maximum margin solutions, which efficiently handles outliers. When it comes to handling collinearity, DTs perform better than LR, particularly for categorical values. Compared to individual DTs, RF shave higher accuracy thanks to their ensemble of DTs. While DTs use hyperrectangles in input space to solve problems, SVM use kernel approaches to address nonlinear problems. SVM typically performs better than RF in classification problems [15].
Medical diagnosis has made extensive use of ML models [16,17,18,19]. This research examines different performances of ML for Alzheimer’s diagnosis. A genetic and irreversible brain disorder called Alzheimer’s syndrome gradually impairs critical abilities such as memory and reasoning [20]. To categorize PD using speech data, for example, in reference [21] the author used a CNN-based model, demonstrating a high accuracy rate by capturing subtle patterns that are frequently ignored by traditional ML approaches [21,22,23,24]. In a similar way, ref. [25] showed how well LSTM networks handle time-series biological data and performs better in the early stages of Parkinson’s Disease-identification. The symptoms usually appear in the mid-30s and mid-60s and range from altered sleep patterns to despair and anxiety. Alzheimer’s Disease causes memory loss and reduced cognitive ability, and its symptoms take ten to twenty years to manifest. The primary cause of Alzheimer’s Disease is dementia, which affects 40–50 million people worldwide and is predicted to increase to 131.5 million by 2050. Remarkably, 70% of dementia sufferers live in low-income nations [26].
Salmanpour et al. [27] used ML methods predict the cognitive effects of PI. A comprehensive review of the literature on FS and ML was carried out by Wan et al. [28]. ML is actively used in brain surgery to pinpoint the exact area that has to be treated. The post-diagnosis aspects of PD were also examined by the researchers.
Cavallo et al. [29] used motion data from patients’ upper limbs to create a predictive model for PD. Two groups participated in the study: those with PD and those who did not. The subjects had a device attached to their upper limbs, and they were given instructions on how to carry out different tasks. Frequency and spatiotemporal analyses were performed during data analysis. Different learning strategies were then used to complete the classification process. An alternate method [30] used different ML and feature extraction approaches to identify PD. The most useful criterion for diagnosing the illness, according to the researchers, was phonation. Optimum Path Forest, K-Nearest Neighbor (KNN), SVM, and Multilayer Perceptron (MLP) classifiers were used. Artificial Neural Networks (ANNs) were employed to reduce linguistic impairments, while SVM contributed to the classification process, facilitating the ML-based identification of PD.
Researchers presented an unsupervised method for PD identification in [31]. For clustering and prediction, they used an incremental Support Vector Regression (SVR) and a Self-Organizing Map (SOM), respectively. After dimension reduction using Partial Least Squares (PLS), the Unified PD Rating Scale (UPDRS) is predicted and SOM and SVR approaches are implemented in the suggested method. In a different work [32], authors used the KNN classifier technique for PD identification and investigated a Fuzzy-based C-means clustering algorithm to evaluate feature weights. To find the ideal K value, the KNN classifier made use of a weighted PD dataset with a range of K values.
In order to handle unbalanced data, Wang et al. [33,34] used sophisticated learning techniques to diagnose PD. These techniques included a weighted scheme and a nonlinear mapping of the kernel function. Features were chosen and variables were optimized using the Artificial Bee Colony (ABC) algorithm. Fisher Discriminant Ratio (FDR), Principal Component Analysis (PCA), and SVM were utilized for FS, dimension optimization, and classification, respectively, in a different method put forward by PI. Abdulhay et al. [35] used sensors beneath patients’ feet to collect physical signals and looked at tremor and gait symptoms for PD detection. With an accuracy rate of over 92%, gait features were extracted from the raw data in the PhysioNet database by analyzing the electrical pulses to identify anomalous peaks. Using voice inputs from a shared dataset, Yaman et al. [36] applied feature augmentation to find salient features for illness detection. A total of sixty-six characteristics were chosen to categorize PD. Instead of depending on MRI, motion, or voice data, a noteworthy study in [37,38,39,40,41,42] investigated the possibility of using handwriting inspection to detect PD.
High classification rates are demonstrated by ML-based techniques [43,44,45,46,47] for determining PD, according to a thorough evaluation of the literature. Although ML is useful foridentifying PD, its application is fraught with difficulties. While considering a lot of features improves identification rates, it also uses more resources in terms of computing cost and execution time. On the other hand, utilizing fewer features could make results less reliable.
According to recent research, computation time can be greatly decreased by using a simpler classifier, a lightweight feature extraction model, and a smaller number of features. Moreover, it has been noted that feature extraction from speech signals is comparatively easier than approaches based on motion [39] or MRI. To increase classification rates, a significant number of vocal features have been chosen by several researchers [40,41]. In contrast, speech analysis is a simpler process compared to the extraction of features from MRI images, which yields higher classification accuracy. A summarization of the existing works is given in Table 1.
PD is a progressive neurological disorder that impairs movement and significantly impacts quality of life, necessitating timely and accurate diagnosis for effective management. Traditional diagnostic methods, reliant on clinical examination, can be subjective and delay diagnosis. This work introduces a novel approach using advanced ML techniques to improve the early detection of PD by analyzing diverse patient data sources. The proposed system employs dynamic FS to enhance classification accuracy by reducing noise and redundancy, making it particularly effective for large-scale biomedical datasets. By integrating LSTM networks with the BoF technique, the model captures temporal patterns and complex correlations in data, improving interpretability and generalizability. The LSTM–BoF PD model is designed to enhance the discriminative power of features across various modalities, such as gait, imaging, and genetic data, offering a more accurate and comprehensive understanding of PD dynamics. This integrated approach aims to advance healthcare technologies by enabling a more reliable and efficient system for early PD diagnosis.
In the proposed system, we extend existing methods by combining a DL classification framework with an optimization method called DBoF. This hybrid technique overcomes the drawbacks of previous models that just use ML algorithms by ensuring appropriate feature selection and improved classification accuracy. Furthermore, by incorporating DL, the model can recognize more intricate patterns in the data, increasing its versatility and scalability across various datasets. Predictions become more accurate as a result of the DBoF approach, which improves the system’s capacity to eliminate noise and choose the most pertinent features. The model is highly suited forPDdiagnosis because of method’s special ability to handle large-scale and heterogeneous biological datasets. Moreover, it enables dynamic feature selection for ongoing improvement, guaranteeing the system’s sustained high performance.

3. Proposed Parkinson’s Disease Classification Model

The proposed technique combines the benefits of the BoF technique with LSTM networks. LSTM networks are perfect for evaluating time-series data, including voice recordings and sensor measurements, since they are well-suited for modeling temporal dependencies in sequential data. By expressing data segments as fixed-size feature vectors, the BoF technique, on the other hand, efficiently manages variability in input data and makes the LSTM model’s input preparation easier. The proposed architecture consists the components illustrated in Figure 1 and discussed subsequently.

3.1. Data Collection

To obtain a wide range of information about PD, extensive datasets such as voice recordings, clinical records, and sensor data (such as gait analysis and tremor data) are gathered. The Unified PD Rating Scale (UPDRS) scores and voice measures are part of the extensive PTD (Parkinson’s Telemonitoring Dataset) [13] from the UCI ML Repository [35], which is intended for tracking the advancement of PD. Nearly 6000 instances with 22 variables were collected in the dataset. Important UPDRS components such as UPDRS_1, UPDRS_2 and UPDRS_3 which stand for sub-scores pertaining to behavior, mood, daily living activities, and mentation, respectively, are included. It also contains additional significant features, such as total_UPDRS and motor_UPDRS (UPDRS part III), which offer an overall evaluation of the motor and total UPDRS scores. This dataset provides insights into how these characteristics may be utilized for remote telemonitoring and is useful for regression and prediction tasks that try to explore the association between voice biomarkers and the severity of PD.
Another dataset for the experimentation was considered from PD Handwriting Database (PaHaW) [31], and is dedicated to documenting and examining handwriting patterns in people who have PD. The collection contains dynamic handwriting samples that were taken using digitizing tablets that measure pen location, pressure, and speed in real-time from both PD patients and healthy control people. The purpose of these examples is to emphasize motor deficits such as bradykinesia and tremors, which are usually linked with PD. Typically, these activities include writing words or drawing spirals. Research aiming at comprehending the motor symptoms of PD and creating diagnostic instruments based on handwriting analysis will find great value in the PaHaW dataset.

3.2. Data Preprocessing

To guarantee that the model receives high-quality inputs, perform necessary preprocessing procedures such feature extraction, normalization, and data cleaning. Choosing the most pertinent characteristics, scaling the data, and addressing missing values are all included in this step. The proposed system performs Min Max Normalization defined in Equation (1) to ensure that the features are scaled to same level and it extracts feature by utilizing Equation (2).
ϰ 1 = x u σ
where x is the feature value of original data, μ is the mean of the feature x , and σ is the standard deviation of the data.
i = 1 I l o g x i cos π n i 0.5 k
where the Fourier transform’s magnitude at bin is represented by x , i represents the number of frequency bins, and n is the MFCC(Mel-Frequency Cepstral Coefficients) index.

3.3. Bag of Features Representation

Using a BoF technique, partition the data into fixed-size windows and encode each segment into a feature vector. In this step, a dictionary of features and transformations is created. Afterwards, variable-length data are converted into a uniform format that can be fed into an LSTM, and a dictionary of features is created as defined in Equation (3).
i = i N m i n u j C x 1 u j 2
where x is the data points, u is the cluster centroid, and C is the centroid of all data points. Every segment can then be encoded according to how frequently the dictionary features occur once it has been created: B o F x = [ f 1 , f 2 , , f n ] , where the frequency of the i -th feature in the segment is represented by f .

3.4. LSTM Representation

To represent temporal dependencies and identify patterns in the sequential data, an LSTM model is used. The model is composed of several LSTM layers, succeeded by thick layers that enhance the feature representation and yield precise forecasts. The LSTM is recursive, as indicated by its looping arrow. The term “cell state” refers to this condition. As a result, the data from the previous interval arestored in the cell state. The cell state is modified by a remember vector underneath the input modification gates and defined in Equation (4).
X t = f t X t + i t
After the output has passed through the LSTM layers, it is often routed into fully connected layers where the retrieved features are further processed. The final layer generates a probability distribution over all possible activity classes (e.g., sitting, running, and walking); the expected activity has the highest probability class. This layer is often a SoftMax layer. The network’s performance predictions are compared to the actual activity labels during training using a loss function, and backpropagation is used to adjust the parameters to lower the loss. Regular LSTMs [42,43,44,45] can capture temporal patterns and long-term associations, which enable them to accurately recognize and classify human behaviors based on periodic sensor data. The implementation of LSTM then is followed using Equation (5).
i t = σ w f i i h t 1 , x t + b f , i
f t = σ W f , f h t 1 , x t + b f , f
C t = tanh N f , c h t 1 , x t + b f , c
h t = h t ; h t
where, i t is the input gate, f t is the forget gate, C t is the cell state, σ is the Sigmoid activation function, tanh is the Hyperbolic Tangent activation function, W is the weights, and b is the bias term. Hidden state h t 1 is considered from current point x t . For every time step t, the forward and backward LSTM layers’ outputs are concatenated to create the final output using Equation (6).

3.5. Deployment and Monitoring

The key novelty lies in the incorporation of the DBOFOT, a dynamic feature optimization technique designed to adaptively adjust to the evolving characteristics of the input data. Traditional BoF approaches often treat features as static entities, disregarding the temporal nuances inherent in sequential data. DBOFOT, on the other hand, introduces a dynamic mechanism that tailors the feature extraction process based on the data’s temporal evolution. This adaptability is particularly pertinent in healthcare applications, where disease progression is often characterized by subtle temporal changes that traditional methods might overlook. As shown in Figure 2, the diagram shows a NN design that uses a Softmax layer, a BoF module, and LSTM cells for classification. After processing a series of inputs, the LSTM cells create hidden states and record temporal dependencies. The BoF module then aggregates these hidden states, condensing the features into a fixed-size vector. Class probabilities are produced by the Softmax layer using this aggregated vector as input. This output is further refined by a Dense layer, which results in the final categorization. The architecture is designed with skip connections and dilated convolutions. The network uses dilated convolutions to capture more context without sacrificing resolution and ( 1 × 1 ) convolutions to control the depth of the feature maps. BoF after feature extraction but before classification allows it to act as a bridge, converting the learned intricate features into a format that can be used and classified. The key to this is the use of dilated convolutions, which may be very helpful for collecting temporal or spatial aspects in data, since they enable the network to span a broader receptive field without increasing the number of parameters.
The combination of DBOFOT with a DL-based optimal FS method leverages the strengths of both paradigms. DL, particularly in the form of LSTM networks, excels at capturing intricate temporal dependencies within sequential data. The LSTM component of the proposed model facilitates the extraction of meaningful patterns and representations from the data, allowing for a more nuanced understanding of the underlying dynamics associated Parkinson’s Disease. Furthermore, the addition of stack sparse autoencoder NNs enhances the model’s capacity for optimal FS. Autoencoders are adept at learning efficient data representations and extracting essential features. By stacking these autoencoders, the model can hierarchically learn and represent complex features, enabling a more compact and discriminative feature space.This architecture offers a more dynamic and context-aware feature extraction process, which usually entails building a set of features and encoding incoming data depending on the existence of these characteristics. The architecture displayed here is probably more adaptive and able to extract more complicated patterns straight from the data, which might be regarded as a major benefit in applications requiring extensive feature representation on predetermined features.
For sequential handwriting, data taken from the PaHaW dataset make up the process. Time-series sequences are used to represent each sample, and they capture several attributes with time, including pen stroke location, pressure, acceleration, and velocity. Given that Parkinson’s Disease presents with subtle motor deficits, such as tremors and diminished fine motor abilities, these traits offer vital insights into motor control, which is crucial for recognizing the disease. Recurrent cells, such as GRU (gated recurrent units) or LSTM units, analyse the sequential input. Each input X i travels through these cells, which are displayed at the bottom of the picture, over subsequent time steps.
Due to their ability to recognize both the long-term patterns and the short-term relationships seen in the handwriting sequences, the LSTM/GRU cells are especially well-suited for this task. The forget gates, input gates, and memory cells are part of the internal structure of the cells, which makes sure the model stores and discards data selectively in order to detect minute inconsistencies in handwriting. The BoF layer receives the intermediate outputs from each recurrent unit at various time steps ( h 1 , h 2 , h n ) respectively. By combining and improving feature selection from these recurrent outputs, the BoF layer serves a critical function. Dynamic feature pooling reduces the dimensionality of the data by eliminating unnecessary information and preserving just the most important handwritten traits.
These steps integrate features from different modalities, perform transfer learning, and provide explainability by generating attention weights and visualizing attention heatmaps. Proposed algorithm incorporates mathematical expressions for dynamic BoF integration, adaptive learning rate, multi-modal feature fusion, transfer learning, and explainability.

4. Performance Evaluation

This section presents the performance evaluation of the proposed model by evaluating various metrics such as accuracy, precision, recall, and F1-score, which identify the model’s accuracy in identifying PD, and performance is evaluated using a dataset that included both patients and healthy persons.

4.1. Experiment Setup

To get ready for analysis, the data are preprocessed after they are gathered. In order to deal with missing values, data cleaning is conducted, which may entail mean imputation or the removal of incomplete information. Noise reduction techniques, such as low-pass filtering, are used for voice recordings. The model’s performance is then enhanced by normalization, which makes sure all input features have a comparable scale. Depending on the nature of data, multiple methods are used for feature extraction.
To find the most pertinent features, FS methods PCA and Correlation Analysis (CA) are applied by converting the original features into a collection of linearly uncorrelated components that represent the highest variance in the data, PCA is used to decrease the dimensionality of the data. This makes it easier to recognize and keep the most illuminating elements while removing the unnecessary ones. The associations between the features are then evaluated using CA, making sure that strongly correlated characteristics are not included at the same time, preventing multicollinearity and enhancing model performance. By combining these two methods, it is ensured that the chosen characteristics greatly increase the predictive capacity of the model. Windows of a specific size are used to segment the data. For instance, sensor data might be split into one-minute chunks and voice data into one-second parts. For every segment, feature vectors that capture the key elements of the data are computed. By finding similar patterns among the segments, k-means clustering creates a dictionary of features. The variable-length input data arethen converted into a fixed-length format by encoding each data segment according to the frequency of the dictionary features.
The model consists of numerous LSTM layers with 100 number of units. In order to retain computational efficiency while enabling the network to capture temporal relationships in the data, this design was selected to strike a compromise between model complexity and performance. After an input layer to receive sequences encoded using BoF. One or two thick layers with ReLU activation functions are placed after these layers to further enhance the feature representation. Using a Sigmoid activation function, the final output layer predicts whether PD is present or not. Performance is maximized by choosing the model’s hyperparameters, which include the learning rate (0.001), batch size (32), and number of epochs (50–100). The dataset is divided into test (15%), validation (15%), and training (70%). The training data is used to train the LSTM model. In order to prevent overfitting, the LSTM model is trained using the training set and its performance is tracked on the validation set. In order to keep the model from overfitting the data and training for an excessive amount of time, early stopping is used depending on the validation loss. Through backpropagation, the training process modifies the model’s weights and biases, progressively enhancing its predictive power. Input sequences encoded in BoF at the input layer. We assign 100 amount of unitsto two-to-three LSTM layers. A dense layer include one or two completely connected layers that use ReLU or another activation function. For binary classification (Parkinson’s presence/absence), we use a Sigmoid activation function.

4.2. Results Analysis

For PD to be effectively managed and treated, an accurate diagnosis is essential. This study compares the performance of a number of well-known ML techniques—ANN, Conventional LSTM models, Bayesian Classifiers (BC), and SVM—and the handwriting data analysis against the proposed LSTM–BoF model for PD detection in terms of the performance metrics. The proposed LSTM–BoF model was benchmarked against the following methods as part of our comparative analysis.
We use an adaptive selection technique that assesses each feature’s contribution to the overall classification job in realtime. This is accomplished through the use of a hybrid strategy that combines the automated feature extraction power of DL models with more conventional ML method known as Recursive Feature Elimination (RFE). The goal of the DBoF optimization is to keep the most informative characteristics while reducing redundancy.
Stochastic Gradient Descent (SGD), in conjunction with a L2 regularization method, is used in the optimization process. This guarantees that the model converges to an ideal solution and lessens overfitting. We also add early stopping depending on the validation loss in order to maximize the performance of the model. The hybrid technique combines the potent representation learning of DL models with the interpretability advantages of ML. The proposed approach reduces the computational complexity of the model and increases classification accuracy by striking a balance between automated DL-based optimization and manual FS [11,12].

4.2.1. Artificial Neural Network

A classic ML technique that imitates the composition and operation of the human brain is ANN. NN, which are modelled after the architecture of the brain, are made up of layers of interconnected nodes, or neurons. To produce an output, a neuron takes inputs, calculates a weighted sum, and applies an activation function. In our tests, we employed an ANN with three completely linked layers. There are 64 neurons in the input layer and 128 neurons in the hidden layer. One neuron handles binary categorization in the output layer. We used the Sigmoid function for the output layer to turn the output into a probability score and the ReLU for the hidden layers to introduce non-linearity as the activation functions. The network was trained at a learning rate of 0.001 using the Adam optimizer, and the error between the predicted and real labels was measured using the binary cross-entropy loss function.
The parameters used in the model are as follows:
  • Input Layer: Receives input data x i .
  • Hidden Layers: Perform computations z = i 1 n w i x i + b , followed by activation α = σ ( z ) .
  • Output Layer: Produces final predictions.
  • Weighted Sum: z = i 1 n w i x i + b
  • Activation Function: α = σ ( z ) .
  • Training: Adjusts weights to minimize loss L via gradient descent.
NNs are useful for many applications, such as picture classification and natural language processing, since they are excellent at deriving complicated patterns from input.

4.2.2. Conventional LSTM Model

Because LSTM models can capture long-term dependencies, they are frequently employed for sequential data analysis. Recurrent Neural Networks (RNNs) with LSTM networks are specifically engineered to tackle the vanishing gradient problem and effectively identify long-term dependencies within sequential data. In order to control information flow, LSTMs integrate a memory cell and a number of gating techniques. Two LSTM layers, each with 100 units, were used in the creation of the LSTM model. The model can successfully capture temporal dependencies in the data with this setting. A completely linked layer is placed after the LSTM layers for categorization. The LSTM cells employ normal tanh and sigmoid activations within their gates, whereas ReLU activation was applied in the thick layers. Binary cross-entropy was once more utilized as the loss function to direct learning throughout the model’s optimization utilizing the Adam optimizer at a learning rate of 0.001. In order to control information flow, LSTM integrate a memory cell and a number of gating techniques. The parameters used in the model are given below:
  • Memory Cell C t : Stores information over time.
  • Forget Gate f t : Determines what information to discard.
  • Input Gate it and Input Modulation C t : Control the flow of new information.
  • Output Gate o t : Filters information for the output.

4.2.3. Bayesian Classifier

Based on the Bayes theorem, these probabilistic models forecast that a sample will belong to a specific class. Through the application of Bayes theorem to determine the specific class thedata point belongs to, a BC aids withprediction. It refreshes the previous understanding of the class distribution with the observed data. BC can be used to forecast PD by combining past probability of the illness’s presence with the chance of observed symptoms or biomarkers (such protein expression levels or UPDRS scores). We used a Gaussian Naive Bayes model, which is predicated on the idea that the distribution of the features is normal. The model computes posterior probabilities for classification based on the probability of the characteristics given the class. Since this model is based on probabilistic principles for classification rather than neural network design, it is not affected by hyperparameters such as neuron counts or layers. This method offers a probabilistic framework for classification, making it possible to make reliable predictions even in the case of missing or ambiguous data. It works especially well when combining various kinds of biological and clinical data. The experiment results are shown in Table 2.
The Mean Squared Errors (MSE), shown in Table 2, reveals that the various UPDRS components have differing degrees of prediction error, with UPDRS 3 displaying the highest error and UPDRS 4 the lowest. The values of the Symmetric Mean Absolute Percentage Error (sMAPE), shown in the table, indicate a wide range of relative prediction accuracy, especially for UPDRS 4, which has a low MSE but a considerably high sMAPE, indicating big proportional mistakes. This shows that although some forecasts may have smaller absolute errors, they still require significant improvement in terms of relative accuracy, particularly when it comes to motor problems.

4.2.4. Support Vector Machine

SVM is an approach for supervised learning that divides classes in feature space using a hyperplane. The results obtained are shown in Table 3. A Radial Basis Function (RBF) kernel was used in the SVM model’s configuration to transfer the input characteristics into a higher-dimensional space where a hyperplane may divide the classes and manage the trade-off between maximizing the margin and reducing classification mistakes, the regularization value C was set at 1.0. The SVM model’s performance metrics show that different MDS-UPDRS score components have differing levels of prediction accuracy. According to the MSE, UPDRS 3 has the largest error, whilst UPDRS 4 has the lowest error. Despite having a low MSE, UPDRS 4 has the largest relative error according to the sMAPE, showing considerable proportional inaccuracy. The SVM model does rather well in terms of absolute mistakes, but there is a lot of variety in relative accuracy, especially when it comes to motor problems and everyday experiences, as evidenced by its average MSE of 58.26 and sMAPE of 98.84.

4.2.5. Proposed LSTMBoF Model

There are 232,741 entries and 5 columns in the train_proteins data. As shown in Table 4, the patient’s baseline appointment (0 months) was attended by ID 55. UPDRS Parts I, II, and III have scores of 10.0, 6.0, and 15.0, in that order. The medication state is not recorded (NaN), nor is Part IV recorded (NaN). The patient’s UPDRS Parts I, II, and III scores at three months are 10.0, 7.0, and 25.0, respectively. There is no information available for Part IV or medication state. After six months, the UPDRS Parts I, II, and III scores are 8.0, 10.0, and 34.0, respectively; Part IV and medication state data are not available. UPDRS Parts I, II, and III have scores of 8.0, 9.0, and 30.0 at nine months, respectively. Part IV has a score of 0.0, which indicates no motor impairment. The medication state is “On” and the UPDRS Parts I, II, and III scores at 12 months are 10.0, 10.0, and 41.0, respectively. Part IV has a score of 0.0.
Multiple peptides combine to form the lengthy molecules known as proteins. Each peptide’s abundance in proteins related to PD is noted by the clinic. In line with NPX for proteins, it displays the peptide concentration. The train_peptides Data Frame contains thesedata. This dataset offers comprehensive data on a patient’s PD progression across time, as determined by the MDS-UPDRS and shown in Table 4. It makes it possible to monitor changes in the patient’s non-motor and motor symptoms, as well as the impact of medication on their condition.
The train_peptides dataset consists of 981,834 elements divided into six columns. Table 4 shows the first five entries, and we can view all six dimensions of our dataset. The target label values of the patient will be plotted against each month (visit_month) using a randomly selected patient_id, as shown in Table 5. Using a randomly selected patient_id, we plot the patient’s first 40 UniProt protein entries and their NPX value against the patient’s visit month (visit_month).
The ratings for the first four UPDRS segments as shown in Figure 3, updrs_1, updrs_2, updrs_3, and updrs_4, are forecast by the clinic andrecorded during a patient visit. The dataset is now ready for model training in order to predict the four labels. Using the patient’s recorded protein and peptide data from that visit, we are able to forecast the target labels (updrs_1, updrs_2, updrs_3, updrs_4) for a particular visit.
Using visit ids (visit id) and protein ids (UniProt), we first group the rows in the train_proteins data. The mean of the NPX values of all the rows in a group will then be used to replace the NPX values of each row in that group. Similarly, we organize the rows in the train_peptides data according to their peptide IDs (Peptide) and visit ids (visit_id). Next, we substitute the mean of the Peptide Abundance values of all the rows in a group for each row’s Peptide Abundance values.
The Protein dataset is rearranged so that its UniProt1 values become columns and its unique visit_id values become indices. The NPX values for each visit (row) are recorded in the columns as shown in Figure 4, and they match to the various UniProt1 values that were noted for that visit. The Peptipe dataset has been rearranged so that the unique visit_id values serve as indexes and the dataset’s Peptide values serve as columns. The Peptide Abundance values for each visit (row), which correspond to the various Peptide values noted for that visit, are recorded in the columns as shown in Table 6.
As shown in Table 7, UPDRS 3 displays the lowest MSE (12.19), with decreased MSE values across all UPDRS components, indicating improved prediction accuracy for motor testing results. The sMAPE values reveal less proportional errors and better relative prediction accuracy, particularly for UPDRS 3 (33.1509). With a sMAPE of 52.77 and an MSE of 15.90 on average, the LSTM–BoF model demonstrates noticeably improved accuracy and overall performance.
Three pre-trained models and the proposed system were considered for the evaluation such as ANN, Conventional LSTM, BC, SVM, and the proposed LSTM model, as shown in Table 8. The performance of decision forest models trained on various datasets is displayed in the training results. With 205 cases used for testing and 863 examples for training, Model 1 produced a MSE of 25.7711. Model 2 had an MSE of 29.0788 after being trained on 856 samples, of which 212 were used for testing. But Model 3, which was trained on 845 samples and tested on 213 of them, had a noticeably higher MSE of 175.8958. On the other hand, with the lowest MSE of 9.4679, Model 4, which was trained on 471 samples and 98 for testing, showed the most promising performance. These findings highlight the importance of dataset size and composition in training models, with Model 4 showing the highest level of performance out of all the models tested.
The models labeled as updrs_1, updrs_2, updrs_3, and updrs_4 are assigned MSE values by the given labels, with corresponding MSE values of 25.7711, 29.0788, 175.8958, 9.4679, and 8.6676, respectively. It is estimated that the average MSE for these models is roughly 60.0534. These labels match the models that have been trained and assessed to predict UPDRS scores, a measure of how severe symptoms are of PD. Lower MSE values indicate greater performance. We manually separated 20% of the dataset for validation (called valid_ds) prior to training the dataset. The ANN can also be validated using the Out of Bag (OOB) score. An algorithm selects a subset of random samples from the training set to train ANN; the remaining samples are utilized to fine-tune the model. OOB data arethe subset of data that arenot selected. The OOB data areused to compute the OOB score, as shown in Figure 5.
The scores of each label are described in Table 9. The performance of decision forest models trained on various datasets is displayed in the training results. With 205 cases used for testing and 863 examples for training, Model 1 produced a mean squared error (MSE) of 25.7711. Model 2 had an MSE of 29.0788 after being trained on 856 samples, of which 212 were used for testing. But Model 3, which was trained on 845 samples and tested on 213 of them, had a noticeably higher MSE of 175.8958. On the other hand, with the lowest MSE of 9.4679, Model 4, which was trained on 471 samples and 98 for testing, showed the most promising performance. These findings highlight the importance of dataset size and composition in training models, with Model 4 showing the highest level of performance out of all the models tested. The performance of the algorithms is summarized in the Table 10. Accuracy, sensitivity, specificity, and AUC-ROC wereutilized to compare the performance of the proposed algorithm against existing models.
  • Accuracy: The ratio of accurately anticipated observations to total observations is known as accuracy. It shows how accurately both positive and negative classes are predicted by the model defined in Equation (7).
    A c c u r a c y = T P + T N T P + T N + F P + F N
    where: True Positives (TP): Positive cases that were correctly predicted, True Negatives (TN): Negative cases that were correctly predicted, False Positives (FP): Positive cases that were incorrectly expected, and False Negatives (FN): Negative cases that were incorrectly predicted.
  • Sensitivity: The ratio of accurately predicted positive observations to all actual positive observations is known as sensitivity, also known as recall. It gauges a model’s capacity to identify positive cases. It is defined in Equation (8).
    S e n s i t i v i t y = T P T P + F N
  • Specificity: The ratio of accurately predicted negative observations to all actual negative observations is known as specificity. It assesses the model’s capacity to recognize negative instances defined in Equation (9).
    S p e c i f i c i t y = T N T N + F P
  • AUC-ROC (Area Under the Receiver Operating Characteristic Curve): AUC-ROC is a metric used to assess how well classification models perform at different threshold values. The True Positive Rate (Sensitivity) is plotted against the False Positive Rate (1—Specificity) on the ROC curve. The area under this curve is denoted by AUC. The model performs better at differentiating between positive and negative classes the closer its AUC value is to 1.

4.3. Analysis of Parkinson Using Handwriting

We also used the PaHaW dataset [31] in our tests to examine the resilience and efficacy of our approach. The PD handwriting database includes several handwriting/drawing samples from 38 healthy controls (HC) and 33 PD patients.
Throughout the assessment procedure, we used stratified 10-fold cross-validation. Ten equal-sized subsets that were mutually exclusive were created from the dataset. The training data for each subset werethe union of the remaining subsets. Furthermore, we used data augmentation, producing thirty additional versions for every training data sample that was first used. Ten iterations of the entire process were performed until one test set was used for each fold. To calculate the overall accuracy, we then summed the accuracies of the various subgroups.
To prepare it for batch processing, the image is first enlarged to ( 210 × 210 ) pixels and then turned into an array with an extra dimension. Matplotlib is used to display the batches of augmented photos generated by an image data generator (Train_Generator), as shown in Figure 6. By subjecting machine learning models to a greater range of picture transformations, this data augmentation strategy contributes to the improvement of the training dataset and may increase the robustness and performance of the models.
Further, we can also check on detection using hand drawings. Although there is no known cure for PD, early identification and appropriate treatment can greatly reduce symptoms and enhance quality of life. For this reason, the condition is a valuable study subject, particularly when developing novel diagnostic methods. The study [44] discovered that asking a patient to draw a spiral and then tracking might be used to identify PD.
Table 11 gives the model’s learning progress over time by capturing the model’s accuracy and loss for both training and validation sets across each of the 65 periods. Drawing speed and pressure can be used for detection. Research [45] discovered that patients with PD drew more slowly and used less pressure when writing; this difference was particularly noticeable in those with more severe or advanced stages of the illness. Using this fact, tremors and muscular rigidity, two of the most prevalent Parkinson’s symptoms, have an immediate effect on how a hand-drawn image appears visually, as shown in Figure 7.
Figure 7 shows a sequence of spirals that were made by hand and are used to assess motor control skills. This technique is very helpful in differentiating between people who have problems with their motor control and people who do not. People are asked to draw spirals on paper or a digital device as part of the process. After that, these spirals are examined for particular patterns that point to the existence of problems with motor control. The spiral features that have been made by hand are each labeled with a forecast. These hypotheses probably match the output of a classification model that determines whether the drawing points to PD or shows normal motor control. A prediction, which seems to be a tuple of values, is labeled for each image. These numbers reflects the probability that the model allocated to the various classes (e.g., 1 for Parkinson’s, 0 for healthy).
A label such as PREDICTION: [0.50145745, 0.49854255], for instance, indicates that the model is marginally more certain that the picture belongs to the first class (healthy) than the second class (Parkinson’s). The spirals themselves serve as illustrations of motor control in pictures. Smooth Spirals a sign of good motor control and is usually connected to well people. An assessment of the model’s ability to discriminate between normal and abnormal motor control can be made by looking at the spirals as well as the predictions. When a spiral’s forecast is near [0.5, 0.5], it means that the model is unsure of its classification. More extreme prediction spirals (such [0.7, 0.3] or [0.3, 0.7]) indicate a higher level of categorization confidence, as shown in Figure 8. The forecasts’ distribution and the spirals that go along with it can reveal information about how well the model works. A robust model would be suggested by consistently accurate predictions with high confidence; on the other hand, frequent misclassifications or uncertainty might point to the need for improvement.
With the AUC-ROC of 0.86 as shown in Figure 9, the ANN model produced an accuracy of 86%. In terms of itsperformance, however, there wasalmost similar accuracy received for plain LSTM, but less sensitivity and specificity than the suggested LSTM–BoF. The 82% accuracy and 0.87 AUC-ROC of the conventional LSTM model indicated competitive performance, but in terms of sensitivity it was not as good as the suggested LSTM–BoF. Among the compared methods, the Bayesian classifier had the lowest AUC-ROC score of 0.81 and the highest accuracy of 79%. SVM performed exceptionally well, with the greatest AUC-ROC score of 0.91 and an accuracy of 86%. But in terms of sensitivity, it was not as good as the suggested LSTM–BoF.
One important indicator is how smooth the spiral lines are. Generally speaking, smooth, continuous lines indicate strong motor control. Uneven, jagged, or wavering lines frequently signify the existence of tremors, which are typical in a number of neurological disorders. Measurements are made of the tremor waves’ height (amplitude) and frequency (number of waves per unit length). Elevated frequency and amplitude are suggestive of more serious problems with motor control. Repeatedly deviating from the planned spiral trajectory indicates a hand movement incapacity, which is a typical sign in several circumstances. The precision, recall, F1-score, and support for each class are included in the classification report. These are crucial metrics for assessing a classifier’s effectiveness. An easy-to-understand indicator called the accuracy score determines how many of the model’s overall predictions were accurate. When combined, these three components offer a thorough assessment of the classifier’s performance using the test dataset.
The model’s performance in classifying PD is shown by the confusion matrix. The model demonstrated a high recall rate by correctly identifying 42 (true positives) out of 47 real Parkinson’s cases. However, it affected the precision by incorrectly predicting Parkinson’s in three healthy persons (false positives), as shown in Figure 10. Furthermore, five genuine Parkinson’s cases were overlooked by the model (false negatives), which is important for a precise diagnosis. Nine out of twelve examples (true negatives) were properly detected by the model for healthy persons.
The classification report indicates the performance of the model in distinguishing between healthy individuals (class 0) and those with Parkinson’s (class 1). The model has a precision of 0.64 and a recall of 0.75 for class 0, meaning it correctly identifies 64% of the predictions for healthy individuals and captures 75% of actual healthy cases. For class 1, the precision is 0.93 and recall is 0.89, indicating the model accurately predicts 93% of Parkinson’s cases and correctly identifies 89% of actual Parkinson’s instances. The overall accuracy of the model is 86.44%, as shown in Table 12, which suggests that it performs well when classifying the images with an overall balanced precision, recall, and F1-score, especially for identifying PD. Out of all the positive predictions the model makes, precision indicates the percentage of real positive predictions (erroneously classified as Parkinson’s). For example, our model’s accuracy for the Parkinson’s class was 0.93, which means that 93% of the people who were predicted to have the disease were really diagnosed.
Recall measures how effectively the model detects Parkinson’s Disease patients by dividing the number of true positives by the total number of actual positives. With a recall of 0.89 for the Parkinson’s class, 89% of the real Parkinson’s patients were correctly recognized by the model.
When working with unbalanced classes, the F1-Score offers a balanced statistic as it is the harmonic mean of precision and recall. The model shows a solid balance between accurately identifying people with the condition and reducing false positives, as seen by its F1-score of 0.91 for the Parkinson’s class.
As the ratio of accurate predictions (both for healthy and Parkinson’s patients) to the total number of occurrences, accuracy measures the overall correctness of the model for both groups. With an accuracy of 0.86, our model was able to classify 86% of the cases correctly.

4.4. Discussion

We compared both ML methods and classical handwriting prediction techniques. The proposed LSTM–BoF model outperformed the other machine learning models in the comparison test for PD detection, showing the greatest AUC-ROC score of 0.93, along with accuracy of 89%, sensitivity of 0.85, and specificity of 0.91 as shown in Figure 11. This suggests that the LSTM–BoF model is very good at properly classifying people as either healthy or Parkinson’s sufferers. The SVM model was a strong candidate, but with somewhat less balanced sensitivity (0.83) and specificity (0.88), demonstrating good performance with an AUC-ROC of 0.91 and an accuracy of 86%. With AUC-ROC values of 0.89 and 0.87, respectively, the ANN and Conventional LSTM models performed competitively.The handwriting analysis model, on the other hand, was employed to diagnose PD and had an overall accuracy of 86%. With a recall of 0.89 and a high precision of 0.93 for the class of Parkinson’s patients, the F1-Score for this class was a good 0.91. This shows how reliable the model is at identifying people who have PD. The healthy class, on the other hand, performed worse, with an F1-Score of 0.69, recall of 0.75, and accuracy of 0.64. With an F1-Score of 0.80 for the macro average between the two classes, balanced performance was shown, and an F1-Score of 0.87 for the weighted average throughout the dataset revealed the overall efficacy of the model.
According to experimental data, our approach outperforms current models in classification performance, which makes it a useful tool for PD early diagnosis. The framework’s adaptability to different datasets further highlights its versatility and potential for further medical applications.
This demonstrates the model’s effectiveness at categorizing Parkinson’s patients, even if it might still use more work to accurately identify healthy people. The proposed LSTM–BoF model has a better AUC-ROC (0.93) and overall accuracy (89%), demonstrating robust and balanced performance across several criteria. This model’s high specificity (0.91) indicates that it performs well not just in identifying people with PD but also in accurately identifying healthy persons.

5. Conclusions

The proposed LSTM–BoF algorithm performed better at PD identification than a number of well-known ML methods, such as ANN, conventional LSTM, Bayesian classifiers, and SVM. Through the use of BoF to dynamically integrate features, LSTM-specific learning rates, and multi-modal feature fusion, the LSTM–BoF algorithm outperformed conventional methods in terms of accuracy, sensitivity, specificity, and AUC-ROC. These results highlight the LSTM–BoF algorithm’s potential as a useful tool for precise and understandable PD diagnosis in clinical settings. By analyzing hand-drawn spirals, the LSTM–BoF model shows promising potential as a PD detector. The algorithm accurately classifies people as having PD or not, with an overall accuracy rate of 86.44%. For Parkinson’s cases, it demonstrates high recall (0.89) and precision (0.93), correctly recognizing the majority of cases with fewfalse positives. The model performs moderately for healthy persons, with a precision of 0.64 and a recall of 0.75. In general, the LSTM–BoF model performs well atdiagnosing PD. Nevertheless, it should be improved in terms of lowering false positives and improving the distinction between healthy cases.

Author Contributions

Conceptualization, A. and S.G.; methodology, A.S.M.S.H.; software, S.G.; validation, A., S.G. and A.S.M.S.H.; formal analysis, M.I.A.B.; investigation, M.I.A.B.; resources, A.; data curation, A.; writing—original draft preparation, S.G.; writing—review and editing, S.G.; visualization, M.I.A.B.; supervision, A.S.M.S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Woosong University Academic Research Fund, 2024, South Korea.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Simon, A.; Deo, M.; Selvam, V.; Babu, R. An overview of machine learning and its applications. Int. J. Electr. Sci. Eng. 2016, 1, 22–24. [Google Scholar]
  2. Alzubi, J.; Nayyar, A.; Kumar, A. Machine learning from theory to algorithms: An overview. J. Phys. Conf. Ser. 2018, 1142, 012012. [Google Scholar] [CrossRef]
  3. Çelik, O. A research on machine learning methods and its applications. J. Educ. Technol. Online Learn. 2018, 1, 25–40. [Google Scholar] [CrossRef]
  4. Attaran, M.; Deb, P. Machine learning: The new “big thing” for competitive advantage. Int. J. Knowl. Eng. Data Min. 2018, 5, 277–305. [Google Scholar] [CrossRef]
  5. Schmidt, J.; Marques, M.; Botti, S.; Marques, M. Recent advances and applications of machine learning in solid-state materials science. Npj Comput. Mater. 2019, 5, 83. [Google Scholar] [CrossRef]
  6. Masud, M.; EldinRashed, A.E.; Hossain, M.S. Convolutional neural network-based models for diagnosis of breast cancer. Neural Comput. Appl. 2020, 34, 11383–11394. [Google Scholar] [CrossRef]
  7. El-Hasnony, I.M.; Barakat, S.I.; Mostafa, R.R. Optimized ANFIS model using hybrid metaheuristic algorithms for Parkinson’s disease prediction in IoT environment. IEEE Access 2020, 8, 119252–119270. [Google Scholar] [CrossRef]
  8. Saleh, S.; Cherradi, B.; El Gannour, O.; Hamida, S.; Bouattane, O. Predicting patients with Parkinson’s disease using Machine Learning and ensemble voting technique. Multimed. Tools Appl. 2024, 83, 33207–33234. [Google Scholar] [CrossRef]
  9. Boateng, E.Y.; Abaye, D.A. A review of the logistic regression model with emphasis on medical research. J. Data Anal. Inf. Process. 2019, 7, 190–207. [Google Scholar] [CrossRef]
  10. Welton, T.; Septian, H.; Weiling, L.; Peik, Y.T.; Wenlu, H.; Robert, C.C.; Celeste, C.; Ee, W.L.; Kumar, M.P.; Louis, C.S.T.; et al. Classification of Parkinson’s disease by deep learning on midbrain MRI. Front. Aging Neurosci. 2024, 16, 1425095. [Google Scholar] [CrossRef]
  11. Mall, P.K.; Yadav, R.K.; Rai, A.K.; Narayan, V.; Srivastava, S. Early warning signs of Parkinson’s disease prediction using machine learning technique. J. Pharm. Negat. Results 2022, 13, 4784–4792. [Google Scholar]
  12. Patel, H.H.; Prajapati, P. Study and analysis of decision tree based classification algorithms. Int. J. Comput. Sci. Eng. 2018, 6, 74–78. [Google Scholar] [CrossRef]
  13. Govindu, A.; Palwe, S. Early detection of Parkinson’s disease using machine learning. Procedia Comput. Sci. 2023, 218, 249–261. [Google Scholar] [CrossRef]
  14. Ali, J.; Khan, R.; Ahmad, N.; Maqsood, I. Random forests and decision trees. Int. J. Comput. Sci. Issues (IJCSI) 2012, 9, 272. [Google Scholar]
  15. Chen, W.; Xie, X.; Wang, J.; Pradhan, B.; Hong, H.; Bui, D.T.; Duan, Z.; Ma, J. A comparative study of logistic model tree, random forest, and classification and regression tree models for spatial prediction of landslide susceptibility. Catena 2017, 151, 147–160. [Google Scholar] [CrossRef]
  16. Rohan, G.; Kumari, S.; Senapati, A.; Ambasta, R.K.; Kumar, P. New era of artificial intelligence and machine learning-based detection, diagnosis, and therapeutics in Parkinson’s disease. Ageing Res. Rev. 2023, 90, 102013. [Google Scholar]
  17. Masud, M.; Sikder, N.; Nahid, A.-A.; Bairagi, A.K.; AlZain, M.A. A machine learning approach to diagnosing lung and colon cancer using a deep learning-based classification framework. Sensors 2021, 21, 748. [Google Scholar] [CrossRef]
  18. Masud, M.; Alhumyani, H.; Alshamrani, S.S.; Cheikhrouhou, O.; Ibrahim, S.; Muhammad, G.; Hossain, M.S.; Shorfuzzaman, M. Leveraging deep learning techniques for malaria parasite detection using mobile application. Wirel. Commun. Mob. Comput. 2020, 2020, 8895429. [Google Scholar] [CrossRef]
  19. Masud, M.; Bairagi, A.K.; Nahid, A.A.; Sikder, N.; Rubaiee, S.; Ahmed, A.; Anand, D. A pneumonia diagnosis scheme based on hybrid features extracted from chest radiographs using an ensemble learning algorithm. J. Healthc. Eng. 2021, 2021, 8862089. [Google Scholar] [CrossRef]
  20. Fiest, K.M.; Roberts, J.I.; Maxwell, C.J.; Hogan, D.B.; Smith, E.E.; Frolkis, A.; Cohen, A.; Kirk, A.; Pearson, D.; Pringsheim, T.; et al. The prevalence and incidence of dementia due to alzheimer’s disease: A systematic review and meta-analysis. J. Can. Des Sci. Neurol. 2016, 43 (Suppl. S1), S51–S82. [Google Scholar] [CrossRef]
  21. Rubenstein, E.; Hartley, S.; Bishop, L. Epidemiology of dementia and alzheimer disease in individuals with down syndrome. JAMA Neurol. 2020, 77, 262. [Google Scholar] [CrossRef] [PubMed]
  22. Swathi, G.; Choudhary, S.; Rishitha, M.; Tejaswi, S.; Reddy, L.S.; Reddy, M.S. Drone—Assisted Image Forgery Detection Using Generative Adversarial Net—Based Module. In Advances in Aerial Sensing and Imaging; Wiley: Hoboken, NJ, USA, 2024; pp. 245–266. [Google Scholar]
  23. Reitz, C.; Brayne, C.; Mayeux, R. Epidemiology of alzheimer disease. Nat. Rev. Neurol. 2011, 7, 137–152. [Google Scholar] [CrossRef] [PubMed]
  24. Saeed, F.; Al-Sarem, M.; Al-Mohaimeed, M.; Emara, A.; Boulila, W.; Alasli, M.; Ghabban, F. Enhancing Parkinson’s disease prediction using machine learning and feature selection methods. Comput. Mater. Contin. 2022, 71, 5639–5658. [Google Scholar] [CrossRef]
  25. Qinghui, W.; Zeng, W.; Dai, X. Gait classification for early detection and severity rating of Parkinson’s disease based on hybrid signal processing and machine learning methods. Cogn. Neurodynamics 2024, 18, 109–132. [Google Scholar]
  26. Prince, M.; Anders, W.; Maëlenn, G.; Ali, G.C.; Wu, Y.T.; Prina, M. World Alzheimer Report 2015. The Global Impact of Dementia: An Analysis of Prevalence, Incidence, Cost and Trends. Available online: https://www.alzint.org/resource/world-alzheimer-report-2015/ (accessed on 20 December 2021).
  27. Salmanpour, M.R.; Shamsaei, M.; Saberi, A.; Setayeshi, S.; Klyuzhin, I.S.; VesnaSossi; Rahmim, A. Optimized machine learning methods for prediction of cognitive outcome in Parkinson’s disease. Comput. Biol. Med. 2019, 111, 1033–1047. [Google Scholar] [CrossRef]
  28. Wan, K.R.; Maszczyk, T.; Qi, A.A.; Dauwels, J.; King, N.K. A review on microelectrode recording selection of features for machine learning in deep brain stimulation surgery for Parkinson’s disease. Clin. Neurophys. 2019, 130, 145–154. [Google Scholar] [CrossRef]
  29. Cavallo, F.; Moschetti, A.; Esposito, D.; Maremmani, C.; Rovini, E. Upper limb motor pre-clinical assessment in Parkinson’s disease using machine learning. Parkinson. Relat. Disord. 2019, 63, 111–116. [Google Scholar] [CrossRef]
  30. Almeida, J.S.; Rebouças Filho, P.P.; Carneiro, T.; Wei, W.; Damaševičius, R.; Maskeliūnas, R.; de Albuquerque, V.H.C. Detecting Parkinson’s disease with sustained phonation and speech signals using machine learning techniques. Pattern Recog. Lett. 2019, 125, 55–62. [Google Scholar] [CrossRef]
  31. Zham, P.; Kumar, D.K.; Dabnichki, P.; Arjunan, S.P.; Raghav, S. Distinguishing different stages of Parkinson’s disease using composite index of speed and pen-pressure of sketching a spiral. Front. Neurol. 2017, 8, 435. [Google Scholar] [CrossRef]
  32. Akila, B.; Nayahi, J.J.V. Parkinson classification neural network with mass algorithm for processing speech signals. Neural Comput. Appl. 2024, 36, 10165–10181. [Google Scholar] [CrossRef]
  33. Polat, K. Classification of Parkinson’s disease using feature weighting method on the basis of fuzzy C-means clustering. Int. J. Syst. Sci. 2012, 43, 597–609. [Google Scholar] [CrossRef]
  34. Swathi, G.; Swathi, V.; Tiwari, A. Handwriting and Speech—Based Secured Multimodal Biometrics Identification Technique. In Multimodal Biometric and Machine Learning Technologies: Applications for Computer Vision; Wiley: Hoboken, NJ, USA, 2023; pp. 227–250. [Google Scholar]
  35. Singh, G.; Vadera, M.; Samavedham, L.; Lim, E.C.H. Machine learningbased framework for multi-class diagnosis of neurodegenerative diseases: A study on Pasrkinson’s disease. IFAC-PapersOnLine 2016, 49, 990–995. [Google Scholar] [CrossRef]
  36. Camacho, M.; Wilms, M.; Mouches, P.; Almgren, H.; Souza, R.; Camicioli, R.; Ismail, Z.; Monchi, O.; Forkert, N.D. Explainable classification of Parkinson’s disease using deep learning trained on a large multi-center database of T1-weighted MRI datasets. NeuroImage Clin. 2023, 38, 103405. [Google Scholar] [CrossRef] [PubMed]
  37. Yaman, O.; Ertam, F.; Tuncer, T. Automated Parkinson’s disease recognition based on statistical pooling method using acoustic features. Med. Hypoth. 2020, 135, 109483. [Google Scholar] [CrossRef] [PubMed]
  38. Kotsavasiloglou, C.; Kostikis, N.; Hristu-Varsakelis, D.; Arnaoutoglou, M. Machine learning-based classification of simple drawing movements in Parkinson’s disease. Biomed. Sig. Proc. Contr. 2017, 31, 174–180. [Google Scholar] [CrossRef]
  39. Ayaz, Z.; Naz, S.; Khan, N.H.; Razzak, I.; Imran, M. Automated methods for diagnosis of Parkinson’s disease and predicting severity level. Neural Comput. Appl. 2022, 35, 14499–14534. [Google Scholar] [CrossRef]
  40. Gao, C.; Sun, H.; Wang, T.; Tang, M.; Bohnen, N.I.; Müller, M.L.T.M.; Herman, T.; Giladi, N.; Kalinin, A.; Spino, C.; et al. Model-based and model-free machine learning techniques for diagnostic prediction and classification of clinical outcomes in Parkinson’s disease. Sci. Rep. 2018, 8, 7129. [Google Scholar] [CrossRef]
  41. Nahar, N.; Ara, F.; Neloy, M.A.I.; Biswas, A.; Hossain, M.S.; Andersson, K. Feature selection based machine learning to improve prediction of Parkinson disease. In Brain Informatics: 14th International Conference, BI 2021; Springer International Publishing: New York, NY, USA, 2021; pp. 496–508. [Google Scholar]
  42. D’Alessandro, T.; Stefano, C.D.; Fontanella, F.; Nardone, E.; Pace, C.D. From Handwriting Analysis to Alzheimer’s Disease Prediction: An Experimental Comparison of Classifier Combination Methods. In International Conference on Document Analysis and Recognition; Springer Nature: Cham, Switzerland, 2024; pp. 334–351. [Google Scholar]
  43. Gowroju, A.; Kumar, S. Robust deep learning technique: U-net architecture for pupil segmentation. In Proceedings of the 2020 11th IEEE Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON), Vancouver, BC, Canada, 4–7 November 2020; pp. 0609–0613. [Google Scholar]
  44. Gowroju, S.; Kumar, S. Robust pupil segmentation using UNET and morphological image processing. In Proceedings of the 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), Cairo, Egypt, 26–27 May 2021; pp. 105–109. [Google Scholar]
  45. Swathi, A.; Kumar, S. A smart application to detect pupil for small dataset with low illumination. Innov. Syst. Softw. Eng. 2021, 17, 29–43. [Google Scholar] [CrossRef]
  46. Gowroju, A.; Kumar, S. Review on secure traditional and machine learning algorithms for age prediction using IRIS image. Multimed. Tools Appl. 2022, 81, 35503–35531. [Google Scholar] [CrossRef]
  47. Chung, S.J. Diagnosis and treatment of hand tremor. J. Korean Med. Assoc. 2012, 55, 987–995. [Google Scholar] [CrossRef]
Figure 1. The proposed system architecture.
Figure 1. The proposed system architecture.
Biomedinformatics 04 00120 g001
Figure 2. Proposed fine-tuned LSTM-BoF model for PD classification.
Figure 2. Proposed fine-tuned LSTM-BoF model for PD classification.
Biomedinformatics 04 00120 g002
Figure 3. The clinical data with updrs_1, updrs_2, updrs_3, updrs_4 labels.
Figure 3. The clinical data with updrs_1, updrs_2, updrs_3, updrs_4 labels.
Biomedinformatics 04 00120 g003
Figure 4. Patients and their NPX values against their visit month. (a) NPX values at the baseline (month 0). (b) NPX values at the 24-month visit. (c) NPX values at 36 months. (d) NPX values at the 60-month visit.
Figure 4. Patients and their NPX values against their visit month. (a) NPX values at the baseline (month 0). (b) NPX values at the 24-month visit. (c) NPX values at 36 months. (d) NPX values at the 60-month visit.
Biomedinformatics 04 00120 g004aBiomedinformatics 04 00120 g004bBiomedinformatics 04 00120 g004c
Figure 5. Mean Squared Error during Out of Bag results.
Figure 5. Mean Squared Error during Out of Bag results.
Biomedinformatics 04 00120 g005
Figure 6. Preprocessed hand-written PD patient’s spiral drawing images. (a) There are three separate loops in this spiral, which begin with a small, tight centre and grow outward. The spiral’s breadth seems to be rather constant, although there are a few tiny abnormalities are the signs of mild motor control problems. (b) In comparison to the first image, the loops are a little bit more enlarged. With fewer line deviations and a smoother appearance, the spiral may indicate improved motor control or less tremor. (c) The spiral is uniformly wide throughout, but it seems less accurate with a more noticeable outward curvature, which might suggest more severe motor problems. (d) This spiral has been enlarged to reveal more information on a selected section of the spiral. This line exhibits increased jaggedness and abnormalities, indicating a more serious problem with motor control. It may be associated with the tremor or another Parkinson’s Disease symptom.
Figure 6. Preprocessed hand-written PD patient’s spiral drawing images. (a) There are three separate loops in this spiral, which begin with a small, tight centre and grow outward. The spiral’s breadth seems to be rather constant, although there are a few tiny abnormalities are the signs of mild motor control problems. (b) In comparison to the first image, the loops are a little bit more enlarged. With fewer line deviations and a smoother appearance, the spiral may indicate improved motor control or less tremor. (c) The spiral is uniformly wide throughout, but it seems less accurate with a more noticeable outward curvature, which might suggest more severe motor problems. (d) This spiral has been enlarged to reveal more information on a selected section of the spiral. This line exhibits increased jaggedness and abnormalities, indicating a more serious problem with motor control. It may be associated with the tremor or another Parkinson’s Disease symptom.
Biomedinformatics 04 00120 g006
Figure 7. Model predictions on spiral drawings for PD detection.
Figure 7. Model predictions on spiral drawings for PD detection.
Biomedinformatics 04 00120 g007
Figure 8. Predictions on spiral images: predicted probabilities of spiral drawings for PD detection.
Figure 8. Predictions on spiral images: predicted probabilities of spiral drawings for PD detection.
Biomedinformatics 04 00120 g008
Figure 9. ROC curve representing the accuracy of models.
Figure 9. ROC curve representing the accuracy of models.
Biomedinformatics 04 00120 g009
Figure 10. Confusion matrix of proposed system.
Figure 10. Confusion matrix of proposed system.
Biomedinformatics 04 00120 g010
Figure 11. Comparison of model performance using two different datasets: (a) classification accuracy of model’s performance using PTD dataset and (b) classification accuracy of handwriting analysis.
Figure 11. Comparison of model performance using two different datasets: (a) classification accuracy of model’s performance using PTD dataset and (b) classification accuracy of handwriting analysis.
Biomedinformatics 04 00120 g011aBiomedinformatics 04 00120 g011b
Table 1. Comparison of various methods in the literature.
Table 1. Comparison of various methods in the literature.
Ref.ApproachAdvantageLimitation
[1]T1-weighted MRI and DTI with SVMGood accuracy with MRI and DTI dataHigh computational resources required; expert-driven feature extraction can be subjective
[2]Deep learning framework using MRI dataHigh accuracy, automatic feature extractionLimited interpretability; requires large labeled datasets
[3]Genetic algorithms and SVM using clinical and genetic dataEffective FSComputationally intensive; difficulty generalizing results across populations
[4]Hybrid PCA and deep learning for gait dataImproved classification accuracyPotential loss of valuable information with PCA; challenge in integrating multimodal data
[5]Machine learning for predicting cognitive effects of Parkinson’sEffective prediction of cognitive impactNot specified
[6]Motion data from upper limbs with various learning strategiesComprehensive analysis with spatiotemporal featuresRequires specialized equipment for data collection
[7]Various ML and feature extraction approaches (phonation data)High accuracy with multiple classifiersNot specified
[8]Unsupervised method with incremental SVR and SOMEffective clustering and prediction
[9]k-NN classifier with fuzzy-based C-means clustering for feature weightsEffective feature weighting and classification
[10]Weighted scheme, nonlinear kernel function mapping, ABC algorithmEffective handling of unbalanced data
[11]FDR, PCA, and SVM for FS, dimension optimization, classificationSuccessful hybrid model
[12]Sensors under feet for gait and tremor detectionHigh accuracy (92%) with gait featuresRequires specialized sensors for data collection
[13]Feature augmentation from voice inputsEffective vocal FSNot specified
[14]Handwriting inspection for Parkinson’s detectionNovel approach
[15]Simpler classifier, lightweight feature extraction, fewer featuresReduced computation timeTrade-off between reliability and computational cost
Table 2. Results obtained from pre-trained Bayesian Classifier model.
Table 2. Results obtained from pre-trained Bayesian Classifier model.
TargetMSEsMAPE
UPDRS 131.3472.6704
UPDRS 245.4178.7865
UPDRS 3122.0378.1446
UPDRS 418.55125.4781
Table 3. Results obtained from pre-trained SVM model.
Table 3. Results obtained from pre-trained SVM model.
TargetMSEsMAPE
UPDRS 119.0363.44
UPDRS 230.5894.9
UPDRS 3173.380.68
UPDRS 410.14156.35
Average58.2698.84
Table 4. Sample PTD dataset.
Table 4. Sample PTD dataset.
Seqvisit_idpatient_idvisit_monthupdrs_1updrs_2updrs_3updrs_4upd23b_clinical_state_on_medication
055_055010.06.015.0NaNNaN
155_355310.07.025.0NaNNaN
255_65565.010.034.0NaNNaN
355_95598.09.030.00.0On
455_12551210.010.041.00.0On
Table 5. Sample training data.
Table 5. Sample training data.
Seqvisit_idvisit_monthpatient_idUniProtNPX
055_0055O0039111,254.3
155_0055O00533732,430.0
255_0055O0058439,585.8
355_0055O1449841,526.9
455_0055O1477331,238.0
Table 6. Peptides information considered for training.
Table 6. Peptides information considered for training.
Seqvisit_idvisit_monthpatient_idUniProtPeptidePeptide Abundance
055_0055O00391NEQEQPLGQWHLS11,254.3
155_0055O00533GNPEPTFSWTK102,060.0
255_0055O00533IEIPSSVQQVPTIIK174,185.0
355_0055O00533KPQSAVYSTGSNGILLC(UniMod_4)EAEGEPQPTIIK27,278.9
455_0055O00533SMEQNGPGLEYR30,838.7
Table 7. Results obtained from pre-trained proposed LSTM–BoF model.
Table 7. Results obtained from pre-trained proposed LSTM–BoF model.
TargetMSEsMAPE
UPDRS 123.4165.091
UPDRS 212.8160.0634
UPDRS 312.1933.1509
UPDRS 415.2NaN
Average15.9052.77
Table 8. Mean Square Error analysis of four models.
Table 8. Mean Square Error analysis of four models.
ModelAlgorithmLabelMSE
Model 1ANNupdrs_125.7711
Model 2Conventional LSTMupdrs_229.0788
Model 3Bayesian Classifierupdrs_3175.8958
Model 4SVMupdrs_49.4679
Model 5Proposed LSTM–BoFupdrs_58.6676
Table 9. Feature score description.
Table 9. Feature score description.
Feature NameIndexImportance Score
Q0648181036.0
P0418070321.0
P0760272917.0
FFLC(UniMod_4)QVAGDAK23012.0
P0506071211.0
O608886389.0
DSGEGDFLAEGGGVR1628.0
LEEQAQQIR4758.0
NVVYTC(UniMod_4)NEGYSLIGNPVAR6278.0
P050677138.0
P171747657.0
QQTHMLDVMQDHFSR8887.0
TQSSLVPALTDFVR10607.0
EVNVSPC(UniMod_4)PTQPC(UniMod_4)QLSK2176.0
QWAGLVEK8976.0
SSGLVSNAPGVQIR9886.0
WYFDVTEGK11636.0
AVLPTGDVIGDSAK845.0
AVC(UniMod_4)SQEAMTGPC(UniMod_4)R804.0
NLREGTC(UniMod_4)PEAPTDEC(UniMod_4)KPVK6094.0
Q6UXB88294.0
Table 10. Evaluation parameters of the proposed model using repository dataset.
Table 10. Evaluation parameters of the proposed model using repository dataset.
AlgorithmAccuracySensitivitySpecificityAUC-ROC
ANN0.840.810.870.89
Conventional LSTM0.820.790.850.87
Bayesian Classifier0.790.760.820.81
SVM0.860.830.880.91
Proposed LSTM–BoF0.890.850.910.93
Table 11. Training of proposed LSTM model for handwriting dataset.
Table 11. Training of proposed LSTM model for handwriting dataset.
EpochTraining LossTraining AccuracyValidation LossValidation Accuracy
12.555363.57%1.112228.57%
52.222252.72%0.700242.86%
100.681564.79%0.676828.57%
150.658064.79%0.705328.57%
200.772335.21%0.689157.14%
250.691252.31%0.691557.14%
300.692056.60%0.684071.43%
350.688947.69%0.696157.14%
400.685451.96%0.694542.86%
450.687353.09%0.677971.43%
500.695744.62%0.686571.43%
550.691559.32%0.682771.43%
600.687769.24%0.690271.43%
650.696786.74%0.688171.43%
Table 12. Evaluation parameters of proposed model of PaHaW dataset.
Table 12. Evaluation parameters of proposed model of PaHaW dataset.
ClassPrecisionRecallF1-ScoreSupport
0 (Healthy)0.640.750.6912
1 (Parkinson’s)0.930.890.9147
Accuracy 0.8659
Macro Avg0.790.820.8059
Weighted Avg0.870.860.8759
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aarti; Gowroju, S.; Begum, M.I.A.; Hosen, A.S.M.S. Optimal Feature Selection and Classification for Parkinson’s Disease Using Deep Learning and Dynamic Bag of Features Optimization. BioMedInformatics 2024, 4, 2223-2250. https://doi.org/10.3390/biomedinformatics4040120

AMA Style

Aarti, Gowroju S, Begum MIA, Hosen ASMS. Optimal Feature Selection and Classification for Parkinson’s Disease Using Deep Learning and Dynamic Bag of Features Optimization. BioMedInformatics. 2024; 4(4):2223-2250. https://doi.org/10.3390/biomedinformatics4040120

Chicago/Turabian Style

Aarti, Swathi Gowroju, Mst Ismat Ara Begum, and A. S. M. Sanwar Hosen. 2024. "Optimal Feature Selection and Classification for Parkinson’s Disease Using Deep Learning and Dynamic Bag of Features Optimization" BioMedInformatics 4, no. 4: 2223-2250. https://doi.org/10.3390/biomedinformatics4040120

APA Style

Aarti, Gowroju, S., Begum, M. I. A., & Hosen, A. S. M. S. (2024). Optimal Feature Selection and Classification for Parkinson’s Disease Using Deep Learning and Dynamic Bag of Features Optimization. BioMedInformatics, 4(4), 2223-2250. https://doi.org/10.3390/biomedinformatics4040120

Article Metrics

Back to TopTop