Next Article in Journal
Energy Homeostasis-Associated (Enho) mRNA Expression and Energy Homeostasis in the Acute Stress Versus Chronic Unpredictable Mild Stress Rat Models
Next Article in Special Issue
Epileptic Patient Activity Recognition System Using Extreme Learning Machine Method
Previous Article in Journal
MEG Node Degree for Focus Localization: Comparison with Invasive EEG
Previous Article in Special Issue
Automated Collateral Scoring on CT Angiography of Patients with Acute Ischemic Stroke Using Hybrid CNN and Transformer Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Early Prediction of Dementia Using Feature Extraction Battery (FEB) and Optimized Support Vector Machine (SVM) for Classification

1
Aging Research Center, Karolinska Institutet, 171 65 Stockholm, Sweden
2
Department of Health, Blekinge Institute of Technology, 371 79 Karlskrona, Sweden
3
Department of Electrical Engineering, University of Science and Technology Bannu, Bannu 28100, Pakistan
4
Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, Stoke-on-Trent ST4 2DE, UK
5
School of Health Sciences, University of Skövde, 541 28 Skövde, Sweden
*
Author to whom correspondence should be addressed.
Biomedicines 2023, 11(2), 439; https://doi.org/10.3390/biomedicines11020439
Submission received: 15 January 2023 / Revised: 30 January 2023 / Accepted: 31 January 2023 / Published: 2 February 2023
(This article belongs to the Special Issue Artificial Intelligence in Neurobiology and Neurologic Diseases)

Abstract

:
Dementia is a cognitive disorder that mainly targets older adults. At present, dementia has no cure or prevention available. Scientists found that dementia symptoms might emerge as early as ten years before the onset of real disease. As a result, machine learning (ML) scientists developed various techniques for the early prediction of dementia using dementia symptoms. However, these methods have fundamental limitations, such as low accuracy and bias in machine learning (ML) models. To resolve the issue of bias in the proposed ML model, we deployed the adaptive synthetic sampling (ADASYN) technique, and to improve accuracy, we have proposed novel feature extraction techniques, namely, feature extraction battery (FEB) and optimized support vector machine (SVM) using radical basis function (rbf) for the classification of the disease. The hyperparameters of SVM are calibrated by employing the grid search approach. It is evident from the experimental results that the newly pr oposed model (FEB-SVM) improves the dementia prediction accuracy of the conventional SVM by 6%. The proposed model (FEB-SVM) obtained 98.28% accuracy on training data and a testing accuracy of 93.92%. Along with accuracy, the proposed model obtained a precision of 91.80%, recall of 86.59, F1-score of 89.12%, and Matthew’s correlation coefficient (MCC) of 0.4987. Moreover, the newly proposed model (FEB-SVM) outperforms the 12 state-of-the-art ML models that the researchers have recently presented for dementia prediction.

1. Introduction

Dementia is a mental disorder marked by a progressive decline in cognitive functions that interferes with daily living skills such as visual perception, problem-solving, memory, and the ability to focus on a single topic. Dementia is more common in older adults, yet many consider it an inevitable outcome of aging. This impression of aging might be incorrect. A wide range of illnesses and injuries to the brain are the primary causes of dementia development [1]. The number of people with dementia is swiftly growing globally, and statistical forecasts indicate that 135 million individuals might have dementia by 2050 [2]. According to the World Health Organization, dementia is the sixth leading cause of death globally, and it is the leading cause of disability and dependency among the aged worldwide [3].
Current early-stage dementia diagnosis relies on pathology features or cognitive diagnostic procedures. Pathology features can be detected via neuroimaging. Magnetic resonance imaging (MRI) is employed to investigate the changes in neuron structure [4]. Electroencephalography (EEG) is utilized to assess event-related possibilities to determine the early stages of dementia [5]. However, such techniques are ineffective in identifying dementia. These prediction tests are relatively inexpensive and time-consuming. In addition, a recent study proposes employing computed tomography (CT) or electromagnetic resonance imaging (MRI) to rule out structural causes of the clinical phenotype [6]. It is estimated that primary-care physicians overlook between 29% and 76% of dementia patients or suspected dementia patients [7]. The use of cognitive tests to assess the early stages of dementia also has some disadvantages. It is difficult for paramedics to engage patients and urge them to participate in testing procedures because older individuals often fear visiting clinics. On the other hand, dementia tests performed at home are usually conducted by inexperienced relatives unfamiliar with the scales; as a result, many test findings are incorrect. ML algorithms provide a novel approach to this challenge. Because of developments in information technology, paramedics now have better access to patients’ lives and can detect decreased cognitive function at an earlier stage. ML algorithms may also provide expert medical knowledge. ML-based tools can provide excellent accuracy and a user-friendly approach to the early prediction of dementia. Scientists have built multiple automated diagnostic systems for various ailments using ML methods, such as heart failure [8,9,10], Parkinson’s [11], hepatitis [12], and clinical decision support systems [13].
Different automated diagnostic methods based on ML methodologies have been presented in the past for early diagnosis of dementia. F. A. Salem et al. developed an ML algorithm for dementia prediction using a regression model. They also studied ML techniques for unbalanced classes in the dataset. They utilized oversampling and undersampling to eliminate the bias in the ML model. The balanced random forest (RF) model was the most resilient probabilistic model using only 20 variables from the dataset. Their proposed model reported an F1-score of 0.82%, G-Mean of 0.88%, and AUC of 0.88% [14]. Dallora et al. [15] employed decision trees (DT) to evaluate predictive factors for the ten-year prediction of dementia. In their proposed technique, they employed a recursive feature elimination (RFE) feature selection method to determine the most important variables from a dataset for dementia classification. Their proposed approach, based on RFE and DT, had an AUC of 74.50%. Through feature engineering and genetic algorithms, F.G. Gutierrez et al. devised an ML approach for diagnosing AD and FTD. Their suggested method had an accuracy of 84% [16]. G. Mirzaei and H. Adeli investigated cutting-edge ML approaches for identifying and categorizing AD [17]. H. Hsiu et al. investigated ML techniques for the early detection of cognitive deterioration using a threefold cross-validation approach, and their suggested model achieved an accuracy of 70.32% [18]. A. Shahzad et al. proposed an ML model for pre-screening MCI using an inertial sensor-derived gait biomarker with a 71.67% accuracy rate [19].

Aim of Study

The previously proposed ML models suffer from lower accuracy and bias in ML models. Motivated by these factors, we proposed a novel feature extraction method to extract useful features from the dataset. Moreover, we optimized an SVM using a grid search algorithm. The proposed hybrid model uses two components: the feature extraction battery (FEB) and support vector machine (SVM), leading to the newly proposed model, namely, (FEB-SVM). To address the issue of bias in the ML model, we deployed the adaptive synthetic sampling (ADASYN) scheme to balance the classes in the dataset. To validate the effectiveness of the proposed model (FEB-SVM), we used different evaluation metrics, e.g., accuracy, precision, recall, PR curve, area under the curve (AUC), F1 score, and Matthew’s correlation coefficient (MCC). Moreover, we conducted three different experiments to evaluate the performance of the newly proposed model.
It is essential to understand that dementia is classified into various subtypes, the most frequent being AD and Vascular dementia, among others. Mixed comorbidities are uncommon, and AD commonly comports with Vascular or Lewy Bodies dementia. Furthermore, unusual forms of AD are sometimes misdiagnosed, according to [20]. The study mentioned here makes no difference between subtypes of dementia, and the word dementia refers to all kinds of dementia.

2. Materials and Methods

This research collected data from the Swedish National Study on Aging and Care (SNAC) for the experimental purpose of the proposed model (FEB-SVM). The SNAC is a longitudinal consortium that has been collecting multimodal data from the Swedish senior population to “create trustworthy, comparable, durable datasets” to be used for aging research and aged care [21]. The SNAC was developed as a multifunctional program to explore healthcare quality for the aging population. It comprises a database containing details regarding physical assessment, metacognition, social variables, lifestyle factors, medical records, and so on. Blekinge, Skåne, Nordanstig, and Kungsholmen are the sites from which the SNAC database is collected. They consider a couple of Swedish counties—municipal and borough. This research adopted the SNAC-Blekinge baseline examination, with data gathered from 2000 to 2003. In the literature, there is substantial evidence that environmental factors may impact dementia development [22,23]. This research is based on standard criteria and uses data from urban areas (Blekinge). The inclusion methodology used to eliminate individuals from this investigation is given as follows:
  • Individuals who presented a dementia diagnosis at the beginning of the study or before the ten-year mark.
  • Individuals with missing data in the outcome column.
  • Individuals with more than 10% incomplete data.
  • Individuals who expired (died) well before the ten-year study period.
From the 1402 participants in the SNAC-Blekinge baseline, after the application of the selection criteria, 726 participants (313 males and 413 females) were included, 91 (12.5%) of whom had dementia in the 10 years, and 635 (87.5%) who were free of dementia. The demographics of the sample population in the collected dataset are shown in Table 1. The factors chosen from the SNAC-Blekinge database were based on published research [24,25]. The collected dataset (SNAC-Blekinge) consists of 13 physical measurement parameters such as body mass index (BMI), pain in the last four weeks, heart rate sitting, heart rate lying, blood pressure on the right arm, hand strength in the right arm, hand strength in the left arm, feeling of safety when rising from a chair, assessment of rising from a chair, single-leg standing with right leg, single-leg standing with left leg, dental prosthesis, and several teeth.
It is important to remember that all of the features used in the SNAC were picked based on the evidence relevant to the aging process [21]. At the commencement of the study (2000–2003), 75 variables were identified from the seven categories: demographics, lifestyle, social, psychological, medical history, physical examination, blood tests, and the assessment of various health instruments connected to dementia examination. Medical practitioners provide the target variable employed by the proposed model to predict dementia 10 years after the SNAC baseline. The dementia diagnosis was made using the International Statistical Classification of Diseases and Related Health Problems-10th Revision (ICD-10) and the Diagnostic and Statistical Manual of Mental Disorders (DSM-IV). Table 2 provides a feature category and name from the selected dataset (SNAC-Blekinge).

Proposed Work

In this work, we designed an automated diagnostic technique for the early prediction of dementia using machine learning and data mining approaches. The suggested diagnostic system is divided into two modules: the first module extracts valuable features from datasets to avoid the problem of model overfitting and the second module works as a classifier to predict dementia. We developed a novel feature extraction method based on linear discriminate analysis (LDA), independent component analysis (ICA), principal component analysis (PCA), locally linear embedding (LLE), and t-distributed Stochastic Neighbor Embedding (t-SNE). The aforementioned feature extraction methods have been cascaded into a single component, which we named a “feature extraction battery” (FEB). Feature extraction begins with an initial set of measured data. It creates derived values (features) that are meant to be useful and non-redundant, allowing future learning and generalization phases and, in some situations, leading to improved human interpretations. Feature extraction helps reduce the dimensionality of the dataset, which eventually reduces the computational complexity of the machine-learning models. The extracted features from FEB are fed into the predictive module of the proposed diagnostic system for the prediction of dementia. We employed a support vector machine (SVM) as a predictive module, and the working of the proposed diagnostic system (Figure 1).
The first stage of the proposed diagnostic system is data preprocessing because data play a vital role in predictive ML models. The dataset is refined, standardized, and normalized. We deal with the missing values in the data preprocessing stage by employing K-nearest neighbors (KNN) imputation [26]. This technique finds the K-items comparable (near) to the missing data. The KNN replaces the mean or most common value of K in the missing data with the most comparable neighbors. The selected dataset for the experiments portrays highly imbalanced classes. Hence, KNN imputation is employed independently on missing data from the majority and minority classes. Through this technique, the chance of infecting the minority class with data from the dominant class was reduced. Following the resolution of missing values, we performed the StandardScaler function on the selected dataset. The StandardScaler function helps to standardize a dataset by eliminating the mean and scaling to unit variance. A sample’s average score λ is computed as follows:
Z = ( λ υ ) γ
where υ denotes the mean of training samples and γ is the standard deviation of the training samples. By calculating the relevant statistics on the training set samples, standardizing and scaling are performed independently on each feature. The mean and standard deviation are then saved for further use to transform on additional data.
When ML models are trained using all the feature space of a dataset, they tend to overfit, which means that ML models display improved performance on training data but poor performance on testing data [27,28]. This might be because the classifier learned superfluous or noisy features in the training data, or it could be due to a weak classifier with too many parameters. As a result, we should extract a subset of features from the dataset and a properly constructed classifier. In feature extraction methods, new features are constructed from the given dataset. Feature extraction decreases the resources necessary to explain a vast dataset. One of the primary issues with analyzing complex data is the number of variables involved. Analysis with many variables often necessitates a substantial memory and computing capacity. It may also lead a classification method to overfit training examples and generalize poorly to new samples. Feature extraction is a broad term encompassing ways of building variable combinations to avoid these issues while accurately summarizing the data. Many machine learning practitioners feel that well-optimized feature extraction is the key to good model design [29]. Therefore, we proposed a novel feature extraction method (FEB) to avoid the problem of model overfitting while simultaneously reducing data dimensionality. Reduced data dimensionality increases the performance of the proposed SVM-FEB in terms of time complexity. In FEB, we cascaded different feature extraction methods (LDA, PCA, ICA, LLE, TSNE) into a single module. The four feature extraction methods (PCA, ICA, LLE, TSN) construct a couple of new features, while LDA constructs only one new feature. The newly extracted FEB features are combined to generate an optimum dataset with low dimensionality. The proposed FEB constructs nine new features from the original dataset, which consists of 75 features.
After the feature extraction stage, we divided the dataset into two parts; one for training and the other for the testing purposes of the proposed SVM-FEB model. The classes in the dataset for the experiments are highly imbalanced, which means that the model would be biased toward the majority class. To address this issue, we used the ADASYN approach to tackle the imbalanced class issue [30]. The ADASYN approach uses a density distribution δ i as a criterion to automatically compute the number of synthetic samples necessary for minority data samples. δ i is a physical evaluation of the weight distribution of unique minority class instances depending on their level of learning difficulty. Following ADASYN, the final dataset will not only provide a balanced structure of classes in the data distribution (as defined by the coefficient) but also compels the learning algorithm to concentrate on complicated cases. As a consequence, the proposed system (FEB-SVM) is trained on balanced data, mitigating the risk of bias in the ML model. It is worth noting that the ADASYN approach is applied to training data following the data split. Suppose the ADASYN technique is used for the entire dataset (i.e., before data partitioning). In that case, the performance of the ML model will be skewed because samples from the testing dataset will also be included in the training dataset. Using ADASYN to balance the training dataset, we employed SVM for the classification task.
SVM is a powerful tool for classification and regression problems [31]. SVM attempts to construct a hyperplane with the greatest possible margin. In the case of a classification problem, the hyperplane h ( x ) = ( σ T x ) + γ , where γ denotes the bias and σ represents a weight vector that is built using training data and serves as a decision boundary for determining the class of a data point (a multidimensional feature vector). In the case of binary classification, SVM employs two support vectors and identifies the nearest vectors (data points) of two classes to create a margin. These vectors are referred to as support vectors. Margin is computed by taking the perpendicular distance between the lines going through the support vectors and multiplying by 2 | σ | 2 2 . The primary objective is to develop an optimized SVM predictive model to provide an ideal hyperplane with the highest margin. SVM employs a set of slack variables known as ν i , i = 1, 2, …, ℧, as well as a penalty parameter known as β , and attempts to maximize | | σ | | 2 2 while minimizing misclassification errors. This fact is mathematically expressed as follows:
min σ , γ , ν 1 2 σ 2 2 R e g u l a r i z e d + γ i = 1 ν i E r r o r
S . t y i ( σ x i + γ ) 1 ν i ν i 0 , i = 1 , 2 , ,
where ν is the slack variable that is used to calibrate the degree of misdiagnosis and penalized factor is the Euclidean norm, also known as the L2-norm.
The major difficulty is that a linear hyperplane cannot correctly partition the binary classes’ data points (i.e., with the lowest classification error). For this reason, the SVM employs a kernel technique in which the SVM model converts local data points into hyperdimensional points to convert non-separable data points into separable data points. Different kernels are used, including the radial basis function (RBF) kernel, linear kernel, sigmoid kernel, and polynomial kernel. These kernels are SVM model hyperparameters that must be adjusted for each task. To design the SVM model that works best on a dementia prediction challenge, we must carefully update or optimize its hyperparameters. Grid search is the main way to reach this purpose. As a result, we employed the grid search approach to tweak the SVM hyperparameters. Consequently, in this paper, we suggest an FEB decrease in the data dimensionality. The suggested SVM-FEB approach dynamically optimizes the SVM model’s hyperparameters using the grid search method.

3. Validation and Evaluation

The holdout validation technique has been extensively used in the literature as a benchmark for assessing the effectiveness of ML-based diagnostic systems [10,12,13]. In a holdout validation procedure, a dataset is partitioned into a couple of segments where one half is utilized for training while the remaining half is utilized for testing purposes of the proposed ML model. The dataset is split with a ratio of 70% for training the ML model and 30% used for testing. Thus, in our tests, we employed the data mentioned above partition criteria for the training and testing of the proposed SVM-FEB model. Following the data partition, we select evaluation measures to compare the performance of the proposed model with existing state-of-the-art ML models for dementia prediction. The assessment criteria used for evaluating the SVM-FEB model are accuracy, precision, recall, F1-score, Matthew’s correlation coefficient (MCC), and area under the curve (AUC) using PR plot. The evaluation metrics are presented mathematically as
A c c u r a c y = T + + T T + + T + F + + F
where T + stands for the No. of true positive, F + represents the No. of false positives, T denotes the number of true negatives, and F represents the No. of false negatives.
P r e c i s i o n = T + T + + F +
R e c a l l = T + T + + F
F 1 _ s c o r e = 2 T + 2 T + + F + + F
M C C = T + × T F + × F ( T + + F + ) ( T + + F ) ( T + F + ) ( T + F )
A binary classification problem is statistically examined. The F1-score is defined as the F-measure. The F1-score gives a score between 0 and 1, with 1 representing excellent predictions and 0 representing the worst. MCC is used to determine whether or not a test is correct. The value range for MCC is between 1 and −1, with 1 being the best prognosis and −1 representing the worst prediction.

4. Experimental Results

Three kinds of experiments were conducted on the dementia dataset to examine the efficacy of the newly developed model (FEB-SVM). The first experiment used the grid search algorithm to construct and optimize traditional ML models using all the dataset’s features (75). The proposed SVM-FEB approach is built in the second experiment. At the same time, additional state-of-the-art ML models are constructed in the third experiment while utilizing the same dementia data and a novel, suggested feature extraction module (FEB). All computation tasks were carried out on an Intel (R) Core (TM) i5-8250U CPU running at 1.60 GHz using Windows 10 64bit. The Python software package is employed to carry out all of the experiments.

4.1. Experiment No.1: Performance of ML Models Using All Features

In this experiment, we set up multiple ML models, i.e., naive Bayes (NB), logistic regression (LR), decision tree (DT), random forest (RF), k-nearest neighbors (KNN), and support vector machine (SVM) with various kernels (rbf, linear, polynomial, sigmoid), which are implemented in Python. The efficiency of the constructed ML models was assessed using all the dataset’s features (75). It is noteworthy that the classes in the dataset are balanced through the ADASYN technique. Table 3 shows the accuracy, precision, recall, and MCC of dementia prediction. The SVM with a polynomial kernel obtained the best dementia diagnosis accuracy of 88.59 percent. However, the SVM training accuracy is lower than the test accuracy, indicating that the ML model overfits. Therefore, we constructed a feature extraction battery (FEB) to prevent the problem of model overfitting.

4.2. Experiment No.2: Performance of Proposed Model SVM-FEB

In the proposed model (FEB-SVM), features are extracted from the dataset using a novel feature extraction battery (FEB). From FEB, we obtained the nine features that were extracted from the whole dataset. The classes were found to be highly unbalanced; so, we deployed the adaptive synthetic (ADASYN) oversampling method. After balancing the classes in the dataset, we optimized the hyperparameters of the SVM with the rbf kernel. We employed a grid research algorithm for the optimization of SVM hyperparameters. The optimal values of the hyperparameters of SVM are set by exploiting an exhaustive grid search algorithm. Table 4 shows that the proposed model SVM-FEB achieved the best dementia prediction accuracy, 93.92%, where C = 10 and G = 0.1 (C = Cost, G = Gamma) are the values of hyperparameters that are searched through using the grid research technique. The comparison of Table 3 and Table 4 shows that the proposed model SVM-FEB improves the performance of traditional SVM by 8%. Furthermore, traditional SVM with rbf used all the features of the dataset (see Table 3), while in the proposed model SVM-FEB, only 9 features are used for the prediction of dementia (see Table 4).
To validate the efficiency of the proposed model (FEB-SVM), we also used a precision–recall (PR) curve. In total, two types of experiments were conducted. In the first experiment, a simple SVM with an rbf kernel is tested against the PR-curve; in contrast, in the second experiment, the proposed SVM-FEB is tested. The important parameter in the PR plot is the area under the curve (AUC), in which a model having more area under the curve is considered better. Figure 2b shows the PR curve plot of the traditional SVM model with an AUC of 88% while Figure 2a presents the PR plot of the proposed SVM-FEB model with an AUC of 93%. Hence, the proposed model has better dementia prediction accuracy in comparison to the simple SVM model.
Furthermore, we analyzed the training and testing accuracy of the proposed model (SVM-FEB) with other ML models. From Figure 3, it can be observed that the proposed (SVM-FEB) model achieved the highest testing accuracy compared with the rest of the ML models.

4.3. Experiment No.3: Performance of ML Models Based on FEB

In this experiment, we compared the newly developed model (SVM-FEB) performance with other ML models using a novel feature extraction model (FEB). We selected the following ML models, i.e., Naive Bayes (NB), K-Nearest Neighbors (KNN), Random Forest (RF), Decision Tree (DT), Logistic Regression (LR), and Support Vector Machine (SVM) with different kernels (linear, sigmoid, ref, linear). The selected ML models’ hyperparameters were optimized using a grid search algorithm. For a fair comparison, we used balanced classes in the dataset, which were obtained through the ADASYN technique. Table 5 presents the results of each ML model along with the values of tuned hyperparameters, i.e., D—depth, E—edge, Ne—number of estimators, G—gamma, and K—number of neighbors. The performance of each model is assessed across different evaluation metrics such as accuracy on training data (ACC._train), accuracy on test data (ACC._test), precision, recall, f1_score, and Matthew’s correlation coefficient (MCC). The proposed model (SVM-FEB) has achieved the highest accuracy of 93.92% in comparison with other ML models using the same feature extraction module (FEB), as shown in Table 5. Furthermore, we can also compare the results of Table 5 with Table 3 where the same ML models are used using all dataset features. From the comparative analysis of both tables, it can be seen that the novel proposed feature extraction method (FEB) has significantly improved the performance of ML models. Thus, the proposed feature extraction model reduced the complexity of the ML models because ML models used only 9 features compared with all the available feature space of 75 features. The proposed feature extraction module (FEB) and optimized SVM with an rbf kernel obtained improved accuracy results compared with conventional SVM.

4.4. Comparison of Dementia Prediction Methods

We evaluated the performance of numerous ML-based automated diagnostic systems that were proposed by researchers in the literature for dementia prediction. Table 6 summarizes the performance of previously proposed ML-based approaches for dementia prediction along with our proposed model. Compared with the recently proposed models such as F. A. Salem et al. [14], F. G. Gutierrez et al. [16], G. Mirzaei and H. Adeli [18], and A. Shahzad et al. [19] and A. Javeed et al. (2022) [32], our newly developed model (FEB-SVM) performed considerably better.

5. Conclusions

In this study, we addressed the problems of low accuracy and bias in ML models for dementia prediction, which researchers in the recent past have raised. Unfortunately, dementia is a rare, occurring disease, so classes in the dementia datasets are significantly imbalanced, causing bias in ML models. Therefore, we deployed the adaptive synthetic sampling (ADASYN) technique to resolve this issue. For improved accuracy, we proposed a novel feature extraction (FEB) method that extracts the valuable features from the dataset so that the proposed model would not learn the noisy features from the dataset and avoid model overfitting problems. The feature extraction (FEB) module extracts 9 features from the dataset, comprising 75 features. The FEB helps to improve the accuracy and reduce the computational complexity of the proposed model. For classification purposes, SVM is deployed with different kernel functions, and hyperparameters of the SVM are fine-tuned using a grid research algorithm. From the experimental results, it is evident that the proposed model (FEB-SVM) improved the performance of conventional SVM by 6% for dementia prediction. Moreover, the proposed model outperformed the 12 recently proposed models based on ML for dementia prediction, which the researchers presented. However, the proposed method (SVM-FEB) has significant shortcomings that the researchers must overcome. The newly proposed FEB’s constructed features cannot help identify the features causing dementia in older persons because the newly generated FEB is based on feature extraction techniques. Therefore, new methods should be proposed in the future for identifying the features that cause dementia problems based on feature selection algorithms [39]. Furthermore, a machine learning model based on meta-heuristics and deep learning should be built to improve dementia prediction accuracy.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/biomedicines11020439/s1, Table S1: List of the input variables with description, possible values and number of missing values. References [40,41,42,43,44,45,46,47,48,49] are cited in the Supplementary Materials.

Author Contributions

Conceptualization and validated by P.A.; data curation and formal analysis by A.L.D.; methodology and write up by A.J.; supervised by J.S.B. A.I. has validated the results, L.A. has provided the technical support and co-supervision, and the manuscript writing and proofreading are done by H.T.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was carried out in accordance with the Declaration of Helsinki and was approved by the Research Ethics Committee at Blekinge Institute of Technology (BTH).

Informed Consent Statement

Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

Data can be available in Supplementary Materials.

Acknowledgments

The first author’s learning process was supported by the National E-Infrastructure for Aging Research (NEAR), Sweden. NEAR is working on improving the health condition of older adults in Sweden.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lo, R.Y. The borderland between normal aging and dementia. Tzu-Chi Med. J. 2017, 29, 65. [Google Scholar] [CrossRef] [PubMed]
  2. Vrijsen, J.; Matulessij, T.; Joxhorst, T.; de Rooij, S.E.; Smidt, N. Knowledge, health beliefs and attitudes towards dementia and dementia risk reduction among the Dutch general population: A cross-sectional study. BMC Public Health 2021, 21, 857. [Google Scholar] [CrossRef] [PubMed]
  3. Who. Dementia. Available online: https://www.who.int/news-room/fact-sheets/detail/dementia (accessed on 21 December 2022).
  4. Duchesne, S.; Caroli, A.; Geroldi, C.; Barillot, C.; Frisoni, G.B.; Collins, D.L. MRI-based automated computer classification of probable AD versus normal controls. IEEE Trans. Med. Imaging 2008, 27, 509–520. [Google Scholar] [CrossRef] [PubMed]
  5. Lai, C.L.; Lin, R.T.; Liou, L.M.; Liu, C.K. The role of event-related potentials in cognitive decline in Alzheimer’s disease. Clin. Neurophysiol. 2010, 121, 194–199. [Google Scholar] [CrossRef]
  6. Patel, T.; Polikar, R.; Davatzikos, C.; Clark, C.M. EEG and MRI data fusion for early diagnosis of Alzheimer’s disease. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–25 August 2008; pp. 1757–1760. [Google Scholar]
  7. Patnode, C.D.; Perdue, L.A.; Rossom, R.C.; Rushkin, M.C.; Redmond, N.; Thomas, R.G.; Lin, J.S. Screening for cognitive impairment in older adults: Updated evidence report and systematic review for the US Preventive Services Task Force. JAMA 2020, 323, 764–785. [Google Scholar] [CrossRef]
  8. Javeed, A.; Rizvi, S.S.; Zhou, S.; Riaz, R.; Khan, S.U.; Kwon, S.J. Heart risk failure prediction using a novel feature selection method for feature refinement and neural network for classification. Mob. Inf. Syst. 2020, 2020, 884315. [Google Scholar] [CrossRef]
  9. Javeed, A.; Zhou, S.; Yongjian, L.; Qasim, I.; Noor, A.; Nour, R. An intelligent learning system based on random search algorithm and optimized random forest model for improved heart disease detection. IEEE Access 2019, 7, 180235–180243. [Google Scholar] [CrossRef]
  10. Javeed, A.; Khan, S.U.; Ali, L.; Ali, S.; Imrana, Y.; Rahman, A. Machine learning-based automated diagnostic systems developed for heart failure prediction using different types of data modalities: A systematic review and future directions. Comput. Math. Methods Med. 2022, 2022, 9288452. [Google Scholar] [CrossRef]
  11. Ali, L.; Zhu, C.; Golilarz, N.A.; Javeed, A.; Zhou, M.; Liu, Y. Reliable Parkinson’s disease detection by analyzing handwritten drawings: Construction of an unbiased cascaded learning system based on feature selection and adaptive boosting model. IEEE Access 2019, 7, 116480–116489. [Google Scholar] [CrossRef]
  12. Akbar, W.; Wu, W.-p.; Saleem, S.; Farhan, M.; Saleem, M.A.; Javeed, A.; Ali, L. Development of Hepatitis Disease Detection System by Exploiting Sparsity in Linear Support Vector Machine to Improve Strength of AdaBoost Ensemble Model. Mob. Inf. Syst. 2020, 2020, 8870240. [Google Scholar] [CrossRef]
  13. Javeed, A.; Ali, L.; Mohammed Seid, A.; Ali, A.; Khan, D.; Imrana, Y. A Clinical Decision Support System (CDSS) for Unbiased Prediction of Caesarean Section Based on Features Extraction and Optimized Classification. Comput. Intell. Neurosci. 2022, 2022, 1901735. [Google Scholar] [CrossRef] [PubMed]
  14. Salem, F.A.; Chaaya, M.; Ghannam, H.; Al Feel, R.E.; El Asmar, K. Regression based machine learning model for dementia diagnosis in a community setting. Alzheimers Dement. 2021, 17, e053839. [Google Scholar] [CrossRef]
  15. Dallora, A.L.; Minku, L.; Mendes, E.; Rennemark, M.; Anderberg, P.; Sanmartin Berglund, J. Multifactorial 10-year prior diagnosis prediction model of dementia. Int. J. Environ. Res. Public Health 2020, 17, 6674. [Google Scholar] [CrossRef] [PubMed]
  16. Garcia-Gutierrez, F.; Delgado-Alvarez, A.; Delgado-Alonso, C.; Díaz-Álvarez, J.; Pytel, V.; Valles-Salgado, M.; Gil, M.J.; Hernández-Lorenzo, L.; Matías-Guiu, J.; Ayala, J.L.; et al. Diagnosis of Alzheimer’s disease and behavioural variant frontotemporal dementia with machine learning-aided neuropsychological assessment using feature engineering and genetic algorithms. Int. J. Geriatr. Psychiatry 2022, 5667. [Google Scholar] [CrossRef]
  17. Mirzaei, G.; Adeli, H. Machine learning techniques for diagnosis of alzheimer disease, mild cognitive disorder, and other types of dementia. Biomed. Signal Process. Control 2022, 72, 103293. [Google Scholar] [CrossRef]
  18. Hsiu, H.; Lin, S.K.; Weng, W.L.; Hung, C.M.; Chang, C.K.; Lee, C.C.; Chen, C.T. Discrimination of the Cognitive Function of Community Subjects Using the Arterial Pulse Spectrum and Machine-Learning Analysis. Sensors 2022, 22, 806. [Google Scholar] [CrossRef]
  19. Shahzad, A.; Dadlani, A.; Lee, H.; Kim, K. Automated Prescreening of Mild Cognitive Impairment Using Shank-Mounted Inertial Sensors Based Gait Biomarkers. IEEE Access 2022, 10, 15835. [Google Scholar] [CrossRef]
  20. World Health Organization. Dementia: A Public Health Priority; World Health Organization: Geneva, Switzerland, 2012.
  21. Lagergren, M.; Fratiglioni, L.; Hallberg, I.R.; Berglund, J.; Elmstahl, S.; Hagberg, B.; Holst, G.; Rennemark, M.; Sjolund, B.M.; Thorslund, M.; et al. A longitudinal study integrating population, care and social services data. The Swedish National study on Aging and Care (SNAC). Aging Clin. Exp. Res. 2004, 16, 158–168. [Google Scholar] [CrossRef]
  22. Nunes, B.; Silva, R.D.; Cruz, V.T.; Roriz, J.M.; Pais, J.; Silva, M.C. Prevalence and pattern of cognitive impairment in rural and urban populations from Northern Portugal. BMC Neurol. 2010, 10, 42. [Google Scholar] [CrossRef]
  23. Killin, L.O.; Starr, J.M.; Shiue, I.J.; Russ, T.C. Environmental risk factors for dementia: A systematic review. BMC Geriatr. 2016, 16, 175. [Google Scholar] [CrossRef] [Green Version]
  24. Yu, J.T.; Xu, W.; Tan, C.C.; Andrieu, S.; Suckling, J.; Evangelou, E.; Pan, A.; Zhang, C.; Jia, J.; Feng, L.; et al. Evidence-based prevention of Alzheimer’s disease: Systematic review and meta-analysis of 243 observational prospective studies and 153 randomised controlled trials. J. Neurol. Neurosurg. Psychiatry 2020, 91, 1201–1209. [Google Scholar] [CrossRef] [PubMed]
  25. Arvanitakis, Z.; Shah, R.C.; Bennett, D.A. Diagnosis and management of dementia. JAMA 2019, 322, 1589–1599. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, S. Nearest neighbor selection for iteratively kNN imputation. J. Syst. Softw. 2012, 85, 2541–2552. [Google Scholar] [CrossRef]
  27. Loughrey, J.; Cunningham, P. Overfitting in wrapper-based feature subset selection: The harder you try the worse it gets. In Proceedings of the International Conference on Innovative Techniques and Applications of Artificial Intelligence, 13 December 2004; Springer: Cambridge, UK, 2004; pp. 33–43. [Google Scholar]
  28. Pourtaheri, Z.K.; Zahiri, S.H. Ensemble classifiers with improved overfitting. In Proceedings of the 2016 1st Conference on Swarm Intelligence and Evolutionary Computation (CSIEC), Bam, Iran, 9–11 March 2016; pp. 93–97. [Google Scholar]
  29. Alpaydin, E. Introduction to Machine Learning; MIT Press: Cambridge, MA, USA, 2020. [Google Scholar]
  30. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 1322–1328. [Google Scholar]
  31. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  32. Javeed, A.; Dallora, A.L.; Berglund, J.S.; Anderberg, P. An Intelligent Learning System for Unbiased Prediction of Dementia Based on Autoencoder and Adaboost Ensemble Learning. Life 2022, 12, 1097. [Google Scholar] [CrossRef] [PubMed]
  33. Cho, P.C.; Chen, W.H. A double layer dementia diagnosis system using machine learning techniques. In Proceedings of the International Conference on Engineering Applications of Neural Networks, London, UK, 20–23 September 2012; pp. 402–412. [Google Scholar]
  34. Gurevich, P.; Stuke, H.; Kastrup, A.; Stuke, H.; Hildebrandt, H. Neuropsychological testing and machine learning distinguish Alzheimer’s disease from other causes for cognitive impairment. Front. Aging Neurosci. 2017, 9, 114. [Google Scholar] [CrossRef] [PubMed]
  35. Stamate, D.; Alghamdi, W.; Ogg, J.; Hoile, R.; Murtagh, F. A machine learning framework for predicting dementia and mild cognitive impairment. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 671–678. [Google Scholar]
  36. Visser, P.J.; Lovestone, S.; Legido-Quigley, C. A metabolite-Based Machine Learning Approach to Diagnose Alzheimer-Type Dementia in Blood: Results from the European Medical Information Framework for Alzheimer Disease Biomarker Discovery Cohort; Wiley Online Library: Hoboken, NJ, USA, 2019. [Google Scholar]
  37. Karaglani, M.; Gourlia, K.; Tsamardinos, I.; Chatzaki, E. Accurate blood-based diagnostic biosignatures for Alzheimer’s disease via automated machine learning. J. Clin. Med. 2020, 9, 3016. [Google Scholar] [CrossRef]
  38. Ryzhikova, E.; Ralbovsky, N.M.; Sikirzhytski, V.; Kazakov, O.; Halamkova, L.; Quinn, J.; Zimmerman, E.A.; Lednev, I.K. Raman spectroscopy and machine learning for biomedical applications: Alzheimer’s disease diagnosis based on the analysis of cerebrospinal fluid. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2021, 248, 119188. [Google Scholar] [CrossRef]
  39. Javeed, A.; Dallora, A.L.; Berglund, J.S.; Arif, A.A.; Liaqata, L.; Anderberg, P. Machine Learning for Dementia Prediction: A Systematic Review and Future Research Directions. J Med Syst. 2023, 47, 1573. [Google Scholar] [CrossRef]
  40. Antonovsky, A. The structure and properties of the sense of coherence scale. Soc. Sci. Med. 1993, 36, 725–733. [Google Scholar] [CrossRef]
  41. Wechsler, D. WAIS-III Administration and Scoring Manual; The Psychological Corporation: San Antonio, TX, USA, 1997. [Google Scholar]
  42. Livingston, G.; Blizard, B.; Mann, A. Does sleep disturbance predict depression in elderly people? A study in inner London. Br. J. Gen. Pract. 1993, 43, 445–448. [Google Scholar] [PubMed]
  43. Brooks, R.; Group, E. EuroQol: The current state of play. Health Policy 1996, 37, 53–72. [Google Scholar] [CrossRef] [PubMed]
  44. Katz, S. Assessing Self-maintenance: Activities of Daily Living, Mobility, and Instrumental Activities of Daily Living. J. Am. Geriatr. Soc. 1983, 31, 721–727. [Google Scholar] [CrossRef] [PubMed]
  45. Lawton, M.P.; Brody, E.M. Assessment of Older People: Self-Maintaining and Instrumental Activities of Daily Living. Gerontologist 1969, 9, 179–186. [Google Scholar] [CrossRef] [PubMed]
  46. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. “Mini-Mental State”. A Practical Method for Grading the Cognitive State of Patients for the Clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef] [PubMed]
  47. Agrell, B.; Dehlin, O. The clock-drawing test. Age Ageing 1998, 27, 399–403. [Google Scholar] [CrossRef]
  48. Jenkinson, C.; Layte, R. Development and testing of the UK SF-12 (short form health survey). J. Health Serv. Res. Policy 1997, 2, 14–18. [Google Scholar] [CrossRef] [PubMed]
  49. Montgomery, S.A.; Åsberg, M. A New Depression Scale Designed to be Sensitive to Change. Br. J. Psychiatry 1979, 134, 382–389. [Google Scholar] [CrossRef]
Figure 1. Schematic overview of the proposed model SVM-FEB.
Figure 1. Schematic overview of the proposed model SVM-FEB.
Biomedicines 11 00439 g001
Figure 2. Performance comparison based on AUC. (a) PR curve of proposed model; (b) PR curve of the conventional SVM.
Figure 2. Performance comparison based on AUC. (a) PR curve of proposed model; (b) PR curve of the conventional SVM.
Biomedicines 11 00439 g002
Figure 3. Performance comparison of proposed model with other ML models.
Figure 3. Performance comparison of proposed model with other ML models.
Biomedicines 11 00439 g003
Table 1. Demographic overview of the data samples.
Table 1. Demographic overview of the data samples.
Age_GroupMaleFemaleHealthy_SubjectDementia_Cases
60828216402
66759517006
72507412410
7841509117
8135468119
8426426822
8704192314
90+00050501
Total31341372691
Table 2. Overview of features in the dataset.
Table 2. Overview of features in the dataset.
Feature_ClassFeature_NamesTotal
DemographicAge, Gender02
LifestyleLight Exercise, Alcohol Consumption, Alcohol Quantity, Work Status, Physical Workload, Present Smoker, Past Smoker, Number of Cigarettes a Day, Social Activities, Physically Demanding Activities, Leisure Activities11
SocialEducation, Religious Belief, Religious Activities, Voluntary Association, Social Network, Support Network, Loneliness07
Physical ExaminationBody Mass Index (BMI), Pain in the last 4 weeks, Heart Rate Sitting, Heart Rate Lying, Blood Pressure on the Right Arm, Hand Strength in Right Arm in a 10 s Interval, Hand Strength in Left Arm in a 10 s Interval, Feeling of Safety from Rising from a Chair, Assessment of Rising from a Chair, Single-Leg Standing with Right Leg, Single-Leg Standing with Left Leg, Dental Prosthesis, Number of Teeth13
PsychologicalMemory Loss, Memory Decline, Memory Decline 2, Abstract Thinking, Personality Change, Sense of Identity06
Health InstrumentsSense of Coherence, Digit Span Test, Backwards Digit Span Test, Livingston Index, EQ5D Test, Activities of Daily Living, Instrumental Activities of Daily Living, Mini-Mental State Examination, Clock Drawing Test, Mental Composite Score of the SF-12 Health Survey, Physical Composite Score of the SF-12 Health Survey, Comprehensive Psychopathological Rating Scale12
Medical HistoryNumber of Medications, Family History of Importance, Myocardial Infarction, Arrhythmia, Heart Failure, Stroke, TIA/RIND, Diabetes Type 1, Diabetes Type 2, Thyroid Disease, Cancer, Epilepsy, Atrial Fibrillation, Cardiovascular Ischemia, Parkinson’s Disease, Depression, Other Psychiatric Diseases, Snoring, Sleep Apnea, Hip Fracture, Head Trauma, Developmental Disabilities, High Blood Pressure22
Biochemical TestHemoglobin Analysis, C-Reactive Protein Analysis02
Table 3. Performance of ML models on balance data using all features.
Table 3. Performance of ML models on balance data using all features.
ModelAcc._trainAcc._testPrecisionRecallF1_scoreMCC95% CI 1
NB66.7840.3656.0063.0040.000.16950.77, 0.87
LR90.6577.2555.0057.0077.000.11240.82, 0.91
DT10077.9849.0049.0078.000.11570.78, 0.88
RF10088.0755.0051.0088.000.40580.85, 0.93
KNN89.1786.6951.0050.0087.000.38510.82, 0.92
Adaboost95.1580.2766.0051.0080.000.18340.83, 0.91
SVM_rbf99.3085.7888.0056.0086.000.21310.84, 0.93
SVM_sigmoid78.5474.2856.0063.0074.000.19860.83, 0.92
SVM_linear93.5480.7558.0061.0081.000.20420.82, 0.91
SVM_poly73.2488.5945.0050.0089.000.23310.83, 0.92
1 CI = Confidence Intervals.
Table 4. Performance of ML models on balance data while utilizing FEB.
Table 4. Performance of ML models on balance data while utilizing FEB.
ModelHyper.Acc._trainAcc._testPrecisionRecallF1_scoreMCC
SVM_rbfC:100, G:0.199.7790.8892.3987.6289.500.4178
SVM_rbfC:10, G:0.197.4891.1692.8186.5990.000.4216
SVM_rbfC:100, G:0.199.8891.7992.2285.5691.000.4487
SVM_rbfC:10, G:198.2891.8391.8086.5989.120.4387
SVM_rbfC:10, G:198.5092.4689.6785.0590.000.4725
SVM_rbfC:300, G:0.0195.4692.0293.8887.1192.000.4747
SVM_rbfC:10, G:110092.7089.3795.3692.000.4852
SVM_rbfC:10, G:0.0198.7492.9593.8286.0892.500.4810
SVM_rbfC:100, G:0.198.4193.3192.3487.1193.000.4853
SVM_rbfC:10, G:0.198.2893.9291.8086.5989.120.4987
SVM_linearC: 0.1, G: 0184.5482.9996.1577.3185.710.3630
SVM_sigmoidC: 10, G: 0.00183.2382.9796.5271.6482.000.3359
SVM_polyC:01, G:0188.2384.5795.7590.2884.000.3732
Table 5. Performance of ML models on balance data using FEB.
Table 5. Performance of ML models on balance data using FEB.
ModelHyper.Acc._trainAcc._testPrecisionRecallF1_scoreMCC
NBVar:0.00679.4075.2262.0077.0075.000.3642
LRC:100, S: newton, p:l284.4678.3664.0079.0078.000.3954
DTD: 03, E:0487.8086.0096.0274.7485.000.3374
RFD:10, Ne:100. E:0197.9590.1594.7984.5389.000.4286
KNNLf:01, K:01, P:0210090.5492.9881.9587.000.4324
AdaboostLr:01, Ne: 1086.7785.6379.3886.7686.000.3567
SVM_LinearC: 0.1, G: 0184.5484.9996.1577.3185.710.3630
SVM_SigmoidC: 10, G: 0.00183.2382.9796.5271.6482.000.3359
SVM_PolyC:01, G:0188.2384.5795.7590.2884.000.3732
SVM_rbfC:10, G:0.198.2893.9291.8086.5989.120.4987
Table 6. Comparison of classification accuracies with previously proposed methods for dementia prediction.
Table 6. Comparison of classification accuracies with previously proposed methods for dementia prediction.
Study (Year)MethodAccuracy (%)Balancing
P. C. Cho and W. H. Chen (2012) [33]PNNs83.00No
P. Gurevich et al. (2017) [34]SVM89.00Yes
D. Stamate et al. (2018) [35]Gradient Boosting88.00Yes
Visser et al. (2019) [36]XGBoost+ RF88.00No
Dallora et al. (2020) [15]DT74.50Yes
M. Karaglani et al. (2020) [37]RF84.60No
E. Ryzhikova et al. (2021) [38]ANN + SVM84.00No
F. A. Salem et al. (2021) [14]RF88.00Yes
F. G. Gutierrez et al. (2022) [16]GA84.00No
G. Mirzaei and H. Adeli (2022) [18]MLP70.32No
A. Shahzad et al. (2022) [19]SVM71.67No
A. Javeed et al. (2022) [32]Autoencoder + Adaboost90.23Yes
Proposed Model (2023)FEB + SVM93.92Yes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Javeed, A.; Dallora, A.L.; Berglund, J.S.; Idrisoglu, A.; Ali, L.; Rauf, H.T.; Anderberg, P. Early Prediction of Dementia Using Feature Extraction Battery (FEB) and Optimized Support Vector Machine (SVM) for Classification. Biomedicines 2023, 11, 439. https://doi.org/10.3390/biomedicines11020439

AMA Style

Javeed A, Dallora AL, Berglund JS, Idrisoglu A, Ali L, Rauf HT, Anderberg P. Early Prediction of Dementia Using Feature Extraction Battery (FEB) and Optimized Support Vector Machine (SVM) for Classification. Biomedicines. 2023; 11(2):439. https://doi.org/10.3390/biomedicines11020439

Chicago/Turabian Style

Javeed, Ashir, Ana Luiza Dallora, Johan Sanmartin Berglund, Alper Idrisoglu, Liaqat Ali, Hafiz Tayyab Rauf, and Peter Anderberg. 2023. "Early Prediction of Dementia Using Feature Extraction Battery (FEB) and Optimized Support Vector Machine (SVM) for Classification" Biomedicines 11, no. 2: 439. https://doi.org/10.3390/biomedicines11020439

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop