Next Article in Journal
Has COVID-19 Modified the Weight of Known Systemic Inflammation Indexes and the New Ones (MCVL and IIC) in the Assessment as Predictive Factors of Complications and Mortality in Acute Pancreatitis?
Next Article in Special Issue
Cardio-Renal Metabolic Syndrome: Interdisciplinary Diagnostic Methods
Previous Article in Journal
Neonatal Sepsis Caused by Streptococcus gallolyticus Complicated with Pulmonary Hypertension: A Case-Report and a Systematic Literature Review
Previous Article in Special Issue
Diagnosing Arterial Stiffness in Pregnancy and Its Implications in the Cardio-Renal-Metabolic Chain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Approach for Metabolic Syndrome Diagnosis Using Explainable Data-Augmentation-Based Classification

by
Mohammed G. Sghaireen
1,*,
Yazan Al-Smadi
2,
Ahmad Al-Qerem
2,
Kumar Chandan Srivastava
3,*,
Kiran Kumar Ganji
4,
Mohammad Khursheed Alam
4,
Shadi Nashwan
5 and
Yousef Khader
6
1
Department of Prosthetic Dentistry, College of Dentistry, Jouf University, Sakaka 72345, Saudi Arabia
2
Department of Computer Science, Faculty of Information Technology, Zarqa University, Zarqa 13110, Jordan
3
Department of Oral Maxillofacial Surgery & Diagnostic Sciences, College of Dentistry, Jouf University, Sakaka 72345, Saudi Arabia
4
Department of Preventive Dentistry, College of Dentistry, Jouf University, Sakaka 72345, Saudi Arabia
5
Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakaka 72345, Saudi Arabia
6
Department of Public Health, Jordan University of Science & Technology, Irbid 22110, Jordan
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(12), 3117; https://doi.org/10.3390/diagnostics12123117
Submission received: 21 October 2022 / Revised: 2 December 2022 / Accepted: 3 December 2022 / Published: 10 December 2022

Abstract

:
Metabolic syndrome (MetS) is a cluster of risk factors including hypertension, hyperglycemia, dyslipidemia, and abdominal obesity. Metabolism-related risk factors include diabetes and heart disease. MetS is also linked to numerous cancers and chronic kidney disease. All of these variables raise medical costs. Developing a prediction model that can quickly identify persons at high risk of MetS and offer them a treatment plan is crucial. Early prediction of metabolic syndrome will highly impact the quality of life of patients as it gives them a chance for making a change to the bad habit and preventing a serious illness in the future. In this paper, we aimed to assess the performance of various algorithms of machine learning in order to decrease the cost of predictive diagnoses of metabolic syndrome. We employed ten machine learning algorithms along with different metaheuristics for feature selection. Moreover, we examined the effects of data augmentation in the prediction accuracy. The statistics show that the augmentation of data after applying feature selection on the data highly improves the performance of the classifiers.

1. Introduction

Major and dramatic changes in the development of nations, economies, cultures, and the environment, as well as an enhancement of the standard of living, have resulted from breakthroughs in science and technology throughout the past century. However, these technological advances have had unintended consequences, including shifts and disruptions in people’s daily routines, the natural world, cultural practices, and societal and economic structures. Because of this, the population is more susceptible to various internal and external risk factors, any one of which may in turn generate pathological states that may ultimately lead to diseases. Infection with pathogenic bacteria, free radicals, carcinogens, toxic chemicals, pollutants, and genetic abnormalities are only some of the many risk factors that might contribute to the onset of these conditions [1]. Changes in way of life, nutrition, and exercise, as well as inactivity, have all been linked to the rise of metabolic illnesses [2]. The aforementioned risk factors may have contributed to the development of diseases such as metabolic syndrome, cardiovascular disorders, diabetes mellitus, cerebrovascular diseases, food-borne diseases, infectious diseases, and cancer [3]. Therefore, focusing on health indicators is an attractive opportunity to explore the health state of individual and population subjects in relation to biochemical changes in the body.
Central obesity, high blood pressure, high blood sugar, and abnormal lipid profiles are the four main components of metabolic syndrome (MetS) [4]. It is important to note that rapid economic growth, an aging population, changes in lifestyle, and obesity are all contributing to the rising prevalence of MetS. The global prevalence of MetS is estimated to be between 20 and 25% [5]. MetS has been linked to an elevated danger of developing diabetes, heart disease, cancer, and mortality [6]. MetS poses a growing clinical and public health burden all over the world [7]. That is why it is crucial to implement effective measures to prevent and manage the spread of MetS. Data mining of medical checkup data can help identify patients at high risk of MetS at an early stage, advancing the timing of disease prevention and control from the later stages of disease development to the earlier stages of disease development. Preventing and controlling MetS requires the development of risk prediction models using data from physical examinations. Models for predicting the likelihood of a disease occurring are called disease risk prediction models [8]. These models are developed to identify those at high risk for a certain disease so that preventative or early intervention measures can be taken. Therefore, it is of considerable practical importance to build a MetS risk prediction model so that at-risk individuals can be identified and treated as soon as possible. Factors such as sex, age, and family history [9], as well as modifiable factors such as diet, physical activity level, and blood pressure [10] contribute to the onset and progression of MetS [11]. Modifiable risk factors are those that can, in theory, be altered. The best way to prevent and manage metabolic syndrome is to identify and address the underlying causes of the condition. However, identifying risk factors for MetS is complicated by interactions between risk factors [12]. In order to successfully deal with complex interactions between variables, machine learning, which is algorithm-based data analysis technology, is equipped with potent data analysis skills. These considerations led to the development of a machine-learning-based risk prediction model for MetS in this study. Figure 1 illustrates the metabolic syndrome diagnosis using machine learning [13].
In this paper, to establish a simple and practical risk prediction model for MetS, we built different machine learning models, based on easily available indicators such as demographic characteristics, anthropometric indicators, living habits, and family history of the subjects and then used a SHAP tool to explain and visualize the model. The interpretable MetS risk prediction model can help uncover risk factors, identify high-risk individuals, and provide methodological references for the prevention and control of MetS.
The key contributions are as follows: (1) Comparing the performance of numerous ancient and new statistical, bagging, and boosting machine learning models; (2) using the synthetic minority oversampling technique (SMOTE) as a data-augmentation strategy to solve the problem of unbalanced classes and avoid bias in machine learning models; (3) using multiple metaheuristics algorithms for feature selection to highlight the metabolic indicators with the best discriminative potential for machine learning models; (4) following feature selection, the Shapley additive explanations (SHAP) method is used to analyze machine learning model results and highlight the most significant metabolic indicators.
The rest of the paper is organized as follows: Section 2 focuses on relevant studies in the field. Section 3 illustrates the proposed methodology, data gathered, classifiers used, and metaheuristics methods, whereas Section 4 shows the statistical measures used for model assessment as well as the hyperparameter tuning process. Moreover, Section 5 discusses the experimental findings and SHAP’s explanations of machine learning models. Lastly, Section 6 concludes the study and suggests next directions.

2. Research Background and Related Works

In the field of metabolic syndrome data prediction and classification, a lot of new study methods, including experimental, investigative, empirical, and comparative research techniques, have been developed lately. Today, because of the availability of a large number of data sets that detail a variety of medical examinations, medical information, as well as the symptoms and indicators of particular diseases, has become an aid in the process of predicting diseases in their early stages. Metabolic syndrome is a prime example of this. The authors of [14] proposed a novel framework for the classification of metabolic data. This new framework makes use of random forests, C 4.5 classifiers, and JRip classifiers. They used a dataset that included information on the lifestyles and blood tests of 2942 patients living in Mexico City. In addition, Chi-squared was used as a feature selection approach in order to exclude irrelevant aspects of the data. According to the results, triglycerides, HDL cholesterol, and waist circumference are some of the greatest indicators of whether or not someone has metabolic syndrome. Regardless of the circumstances, the best results were obtained from random forests. The study in [15] utilizes clinical and genetic data from a Korean community that does not have an excessively overweight population. A framework was created with the use of the WEKA tool to simplify the employment of five different machine learning classifiers. These classifiers are known as naive Bayes, random forests, decision trees, neural networks, and support vector machines. They used data from a total of 10,349 different people. A few examples of the clinical variables that are discussed in the dataset are triglyceride levels, high-density lipoprotein cholesterol levels, and alcohol intake. In comparison to the other approaches, naive Bayes was successful in obtaining the largest area under the curve (0.69).
Using deep learning: In [16], a study of metabolic syndrome scores was proposed, and as part of that research, a dataset was offered that described the health of 3577 students in Birjand. The levels of glucose and triglycerides in the blood taken first thing in the morning are two blood tests that may be used as indicators. The distribution of the data, on the other hand, mandated the use of the synthetic minority oversampling technique for the purpose of achieving data parity. A linear discriminant analysis model was used in the process of feature extraction. The CART classifier outperforms neural networks and support vector machines in terms of four statistical metrics; however, these other two methods are more common. The prediction models, on the other hand, came to the conclusion that the most discriminating factors are waist circumference and high-density lipoprotein. In [17], a dataset consisting of 67,730 Chinese patients who had had a medical examination was used to evaluate the performance of random forest, XGBoost, and stacking classifiers. There were 32 predictors of physical medical tests, blood tests, and ages included in the information that was acquired. However, by employing cross-validation with 10 folds and the area under the curve metric, according to the statistics, XGBoost had the most area under the curve (93%) of any algorithm. In addition, the shapely additive explanation (SHAP) was used in the data to determine the relevance of the attributes. The fasting triglyceride level, abdominal obesity, and body mass index were shown to be the most significant indicators by the SHAP analysis for metabolic data prediction.
The study in [18] evaluates a dataset that describes medical information and tests for 17,182 patients using six of the most effective machine learning classifiers, including logistic regression, extreme gradient boosting, K-nearest neighbors, light gradient boosting, decision trees, and linear analysis. However, three statistical criteria were utilized to evaluate the models. Light gradient boosting exceeded the others, with an area under the curve of 86%. Furthermore, SHAP analysis revealed that waist circumference, triglycerides, and HDL-cholesterol had the strongest discriminative power for predicting metabolic data. In [19], decision trees and support vector machines were tested on an Isfahan cohort study dataset. The dataset described 16 metabolic syndrome indicators. However, for data balance, the synthetic minority oversampling technique was used. The WEKA tool and three statistical metrics were used in the experiment. According to the results, the support vector machine outperforms decision trees by 75% in terms of accuracy.
Using one-dimensional neural networks and eight additional machine learning classifiers: A study was conducted in [20] to evaluate a dataset that describes the lifestyles, medical information, and blood tests of 1991 middle-aged Korean patients. The dataset described several metabolic syndrome indicators such as age, smoking status, sleep time, and waist circumference. However, SMOTE was utilized for data balance. In addition, an area under the curve measure was utilized to evaluate the models. When compared to others, the findings demonstrate that XGBoost has the greatest performance, with an AUC of 85%. Furthermore, waist-to-hip ratio and BMI were shown to be the most important indicators for metabolic prediction. The study in [21] collected a dataset of 39,134 Chinese metabolic syndrome patients. The data set contained information on 19 different diagnostic tests, including alkaline phosphatase, prior diabetes, uric acid, and eosinophil percentage. The developed framework, on the other hand, assessed the employment of logistic regression, random forest, and extreme gradient boosting classifiers. Recursive feature elimination was also employed to select features. With an accuracy of 99.7%, XGBoost surpassed the others. In addition, the LIME library was utilized to display the significance of features. Fasting triglycerides, central adiposity, and systolic blood pressure were shown to be the most significant indications.
Using gradient boosted trees and logistic regression: In [22], an experimental study was conducted to investigate the usage of a Japanese metabolic syndrome dataset. The health insurance union data covered certain patients’ demographics and medical examinations, such as high-density lipoprotein cholesterol, anemia, and smoking. However, the proposed machine learning classifiers were fine-tuned. The area under the curve measure was also employed to evaluate the models. The findings, on the other hand, show that gradient-boosted trees performed best, with an AUC of 89.4%. Furthermore, other metabolic indicators, such as diastolic and systolic blood pressure, were shown to have no impact. The study in [23] was carried out in order to investigate the use of metabolic data from 5646 patients in Bangkok. On the other hand, in order to evaluate how well the random forest classifier worked, four different statistical methods were used. Additionally, ten-fold cross-validation and principal component analysis were carried out on the data. The research indicates that random forest had the best accuracy, coming in at 98.11%. Furthermore, it was shown that the triglyceride level was the single most important feature.
Table 1 summarizes the related research in terms of data gathered, machine learning classifiers employed, and key metabolic syndrome indicators. However, in comparison to prior contributions, in this paper, we compare the performance of ten statistical, boosting, and bagging machine learning classifiers on a metabolic syndrome dataset that includes 29 separate diagnostic procedures and medical tests. We also utilize five metaheuristic algorithms for feature selection, as well as SMOTE for data balance. Furthermore, we apply the Shapley additive explanations tool to explain the outputs of machine learning models at different data observation samples.

3. Proposed Metabolic Classification Framework

In the following sections, we illustrate the proposed framework. In this paper, we propose a new framework for metabolic data classification as shown in Figure 2. The designed framework evaluates the use of ten distinct machine learning classifiers such as logistic regression (LR), support vector machine (SVM), K-nearest neighbors (KNNs), decision trees (DTs), random forest (RFs), adaptive boosting (AdaBoost), gradient boosting (GB), stochastic gradient boosting (SGB), categorical boosting (CatBoost), and extreme gradient boosting (XGBoost) over a metabolic dataset obtained from the Kaggle repository. The dataset contains 29 distinct features that describe several patients’ statuses with a total of 12,012 records. Data preprocessing, on the other hand, was performed to clean the missing values (e.g., null values). Additionally, we noted that the metabolic dataset had a large target classes distribution that indicated imbalanced classes. Therefore, we followed the use of the synthetic minority oversampling technique (SMOTE) as a data resampling technique to balance the data classes. In this study, we evaluate the use of five distinct nature-inspired algorithms for feature selection such as particle swarm optimization, genetic algorithm, firefly algorithm, ant colony optimization, and bat algorithm. Finally, we employ the use of shapely additive explanation (SHAP) to explain machine learning models’ outcomes. SHAP [24] is an open-source python library that is a game-theory-based tool that visualizes how much a single or a group of observations adds to the predictive models by calculating the SHAP value; hence, based on the average predicted values from each classifier, it can be used to highlight and rank the most important features. The SHAP value can be computed as follows:
V = E f x = E f x f x  
where E represents the mean of predicted values of f x .
For evaluating machine learning classifiers, 20% of the data was maintained as a testing set. Our goal is to build accurate classification models for MetS analysis. The proposed framework has the potential to identify high-risk patients and give methodological references for MetS prevention and treatment.

3.1. Metabolic Data Collection and Analysis

Researchers have employed a variety of machine learning metabolic datasets over many years. In this paper, we follow the use of a metabolic dataset that is available at the Kaggle repository. The dataset includes details on 28 different diagnostic procedures, patient demographics, and indicators of metabolic syndrome. The features of the dataset are summarized in Table 2. Nonetheless, we recorded that out of a total of 348,348 data samples (data total population) in the dataset, 23.301 had at least one missing value. However, missing values were cleaned via data preprocessing. In this study, we follow the use of dropping all data records with missing values. In addition, the dataset was free of outliers. In this work, we construct ten machine learning models with different meta-heuristics algorithms for data evaluation and classification.

3.2. Synthetic Minority Oversampling Technique for Data Balancing

Due to the presence of unbalanced data, often referred to as imbalanced target classes, the classifiers that are used by machine learning systems may be biased in favor of one category over another. On the other hand, there is a significant disparity in the distribution of target classes among metabolic datasets. Because of this, we made use of synthetic minority oversampling technique (SMOTE) to guarantee that every piece of data was dispersed fairly. SMOTE [25] develops synthetic data samples to increase minority class data samples by first locating the K nearest neighbors, then calculating the distance between those neighbors, and then increasing that distance by a random integer that falls between 0 and 1.

3.3. Machine Learning Classifiers

3.3.1. Logistic Regression

The objective of a logistic regression model is the same as that of a linear regression model, which is to make predictions about the dependent variable by examining the relationship between the independent variables. However, rather than making continuous predictions, it may be used to answer questions with yes or no. Given that h θ x = 1 / ( 1 + e θ T · x ) , the sigmoid function is used to fit an s-shaped line with a predicted value between 0 and 1. This allows for the possibility that y = 1, which is correct given the value of x. In addition, the sigmoid function is used to fit an s-shaped line with a predicted value between 0 and 1 [26]. When it is predicted that Y will be 0.5, a value of 0.5 for x indicates that there is a probability of 50% that Y will be 1. A low x number suggests a low likelihood that y = 1, whereas a high x value indicates a high probability that y = 1, and an intermediate x value indicates a 50% chance that y will be 1.

3.3.2. K-Nearest Neighbors

The K-nearest neighbors (KNNs) classifier [27] is a comparison method that is based on the distance between the two sets of data. Because it keeps data samples all through the learning phase and then learns throughout the assessment phase, the KNNs classifier is an algorithm that is known for its slowness. When KNNs is seated, it goes through a variety of different stages. We first determine the value of K, which is the number of samples that are the closest to the new data sample, and then apply a distance function to determine how much the new data sample deviates from the training samples. The majority of class values are utilized to determine the prediction value. Based on the value of K, the distance values are sorted, and the samples that are physically nearest to each other are chosen.

3.3.3. Decision Trees

DTs [28] are a classifier that group the characteristics that are present in the training data based on the judgments that are made by trees. Predictions in DTs are made by drawing conclusions in response to a collection of questions that have a significant influence on the intended result of the prediction. These inferences are based on the answers that individuals provide to the questions. This classifier is referred to as a non-parametric approach since it does not need the use of a mapping function to link predictors and outcomes. The term “impurity” is used in DTs to refer to the procedure of determining the decision values that are optimal for each depth level. On the other hand, the function known as the Gini index may be used to determine how impure the tree is.

3.3.4. Random Forests

For classification, RFs [29] use a bagging ensemble. They learn the basics by creating many decision trees from the training data. Similarly, RFs are not a parametric model. To create classification output, however, just a small random sample of rows is selected from the whole data set, and a set of decision trees is constructed for each subset. Decision trees may vary in size. A meta-classifier is a choice on the classification that is made using the basic learners’ judgments as input, and it is indicated by a majority vote in RFs. They can handle simple and complex predictor functions, as well as numeric and categorical data.

3.3.5. Support Vector Machine

The number of features that are input into a support vector machine help to identify the appropriate decision boundary in high-dimensional space. The classifier is the one who uses this boundary after it has been determined. The purpose of the support vector machine, given that there is a wide variety of hyperplanes from which to choose, is to select the hyperplane that maximizes margins over all of the data samples [30]. In addition, the classification may be carried out either in a linear or non-linear way. When the class distribution is known, support vector machines often perform better and are less likely to succumb to the problem of overfitting. On the other hand, if one makes use of the biggest margins available, it is feasible that new data may be correctly fitted and categorized.

3.3.6. Adaptive Boosting

An ensemble of machines is used in the classification method known as AdaBoost [31], which is a boosting classifier. Unlike random forests and decision trees, this kind of forest is ordered. When using AdaBoost, a number of decision trees are joined in a manner that is decentralized. Each one of them is referred to as a “stump,” and it is made up of a single node in addition to two leaves. A forest of stumps is a collection of tree trunks that have been chopped down to their bases and left standing together. A stump, on the other hand, is not a very good learner when it comes to learning how to classify things. The purpose of AdaBoost is to combine numerous weak learners for classification to work together on a classification problem. Because the output of one tree may influence the output of the next, the AdaBoost algorithm places a significant amount of weight on the order in which stumps are generated. When it comes time to rank the stumps, each data sample is assigned a weight (w) that is proportional to the total number of samples as a whole. The formula s w = w o l d × e a m o u n t   o f   s a y   α is used on a regular basis in order to make adjustments to the sample weight. The Gini index is another method that may be used to evaluate the relevance of a stump; typically, a lower Gini score indicates more significance.

3.3.7. Gradient Boosting

GB [32] is a method for classification and regression that is based on boosting ensembles and makes use of pseudo residuals (PR). Gradient boosting differs from AdaBoost in that it does not begin with a stump, but rather with a single leaf that reflects the median values of the predicted classes. AdaBoost begins with a stump. In contrast to that, GB creates a tree of a certain size, very similar to how AdaBoost performs it, with the key difference being that each tree may be far more substantial than a stump. In GB, decision trees serve as the base learners. The probability ratio is the loss function that may be determined by contrasting the actual values with the projected values. This strategy is a powerful one, since it allows prediction errors in GB to be reduced to a minimum via the consistent updating of PR values across trees.

3.3.8. Stochastic Gradient Boosting

Friedman first described the method that uses an ensemble of learners and is known as stochastic gradient boosting [33]. The homogenous hybrid approach, which involves boosting and bagging, was the impetus behind it. The process of weighing the base learners may be used to perform both regression and classification. On the other hand, the SGB base learners with the highest adoption are decision trees. On the other hand, in order to train each tree, a randomized selection of records from the data sample is employed. SGB is a potent method for lowering the chance of overfitting via the random elimination of a sample of input data. This is accomplished when compared to bagging procedures, which are not as effective. SGB is used throughout a wide range of areas, and this practice dates back many decades.

3.3.9. Extreme Gradient Boosting

XGBoost [34] is one of the ensemble methods that make use of boosting. Other ensemble methods include adaptive and gradient boosting. This new technology has replaced the older technique known as gradient boosting. In the same vein as its predecessor, it was intended to handle enormous and complicated data sets. In contrast to adaptive and gradient boosting, XGBoost makes use of innovative regression trees in its algorithm, changing into what is now referred to as reborn trees. However, in contrast to gradient boosting, XGBoost trees begin with a single leaf rather than several leaves. In addition, the regularization and gradient boosting steps of XGBoost are not extras that may be skipped. XGBoost has been one of the most successful boosters on the market for a significant amount of time. The discriminative accuracy of its predictions is quite high, and it is not difficult to put into practice. In addition to this, it has the capability of managing large datasets with imbalanced target classes.

3.3.10. Categorical Boosting

CatBoost [35] is an improvement on gradient boosting. In 2018, it was delivered by Anna Veronika Dorogush and the rest of the Yandex business team. It was conceived with the intention of being preferable to bagging and stacking for a number of different kinds of data. It is based on a method known as boosting, and it was designed to handle information of both category and numerical kinds. CatBoost is a free software package that enables highly rapid calculations on central processing units (CPUs) as well as graphics processing units (GPUs). It is very effective for usage with relatively small datasets despite its ease of implementation. In addition to this, it is a technique that is based on decision trees, which helps avoid overfitting. CatBoost has been in a head-to-head battle with a variety of gradient boosting algorithms during the course of its existence.

3.4. Metaheuristics Features Selection

3.4.1. Genetic Algorithm (GA)

Utilizing a genetic algorithm, which is a population-based meta-heuristic, is one method that may be used to address optimization problems [36]. The organization of chromosomes had a significant impact on the development of the algorithm, which may be thought of as a kind of search. The following is a list of the steps involved in GA: Establish the population by means of its fitness function, subsequent selection, and subsequent reproduction. Gene parameters are first allocated to individuals whose genomes have been randomly created before being utilized to produce chromosomes. The idea that each individual has something to offer to the process of finding a solution to a problem is one of the core principles behind GA. Instead, the fitness function is used to assign a numerical value to each individual, which indicates whether or not they are fit for reproduction. This value might be positive or negative. After the choosing process is complete, the crossover function could be used to create a new individual as the next step.

3.4.2. Ant Colony Optimization (ACO)

To this day, ACO remains one of the most popular meta-heuristic search-based algorithms, used in a broad variety of fields. This new method is a refinement of the ant system published by Marco Dorigo in 1992. ACO [37] was conceptualized by seeing how ant colonies go about their foraging. The goal of ant colony optimization is to find the quickest route from the anthill to the food source. Each path represents a potential answer. However, after the paths have been established, ants use organic chemical molecules called pheromones to guide them along the most direct route to their nest. ACO has been used to and improved upon a wide variety of models.

3.4.3. Particle Swarm Optimization (PSO)

The PSO optimization approach obtains its inspiration from the behaviors of birds that are found in their natural environments [38]. Within the context of this population-based research approach, each individual bird is referred to as a particle, while the whole flock is referred to as a swarm (i.e., population). In PSO, each “particle” is effectively the ideal response to the problem that is being addressed. The positions of the particles, however, will be rearranged once the first randomization has been performed. PSO has been used widely in medical data classification. In [39], PSO was used for metabolic syndrome risk quantification.

3.4.4. Firefly Algorithm (FA)

The use of FA, a kind of meta-heuristic method, may be used for a wide range of optimization problems. It is an algorithm based on natural phenomena, namely the flashing behavior of fireflies. The firefly algorithm is founded on the idea that fireflies are attracted to one another as a swarm, with the less dazzling ones following the brighter ones [40]. This is because the brighter fireflies attract the attention of the other fireflies. Each firefly represents a potential answer, and the closer that firefly goes to the optimal solution, the brighter it grows, which encourages other fireflies to follow in its footsteps. FA has been used often in medical data collection for a great many years. The authors of [41] used FA in order to enhance the classification capabilities of the AdaBoost algorithm for liver diseases.

3.4.5. Bat Algorithm (BA)

The BA optimizer is a meta-heuristic approach that may be used to solve optimization problems. It is an algorithm that draws its inspiration from nature and is based on the way microbats use echolocation [42]. A type of bat known as the micro bat makes use of echolocation, which may be thought of as a kind of sonar, to help it locate prey. The idea behind echolocation sonar is that little bats emit a loud sound in the shape of a wave, and the wave is reflected back to the microbats by the prey it is directed towards. The process of selecting features to classify medical data has traditionally made extensive use of BA. In the study [43], the bat algorithm was altered to accommodate the classification of breast cancer data.

4. Evaluation Metrics

Some statistical evaluation metrics, such as the area under the curve, precision, recall, root mean squared error as prediction error function, and receiver operating characteristic curve, can be used to determine the degree of relationship between the performance of machine learning models and the amount of data. In this work, we evaluate ten different machine learning classifiers by employing different metrics of precision, recall, area under the curve (AUC), and testing accuracy that are based on a confusion matrix, as shown in Equations (1)–(4). Despite the fact that these metrics show the performance of classifiers in terms of successfully classified data and data that was mistakenly categorized, in order to enhance the accuracy of the classifiers, we made use of the GridSearchCV method to determine the optimal values for the boosting-based classifiers’ hyperparameters [44]. As can be seen in Table 3, we choose how many estimators to use and how many different values for the learning rate to include when defining the grid search space. In addition, in order to conduct the experiment, we formulated two distinct loss algorithms and functions along with five-fold cross-validation. The following are the four key basic elements of the confusion matrix:
  • True positive (TP): Presents the number of infected patients that have been classified correctly as infected patients.
  • True negative (TN): Presents the number of non-infected patients that have been classified correctly as non-infected patients.
  • False positive (FP): Presents the number of misclassified non-infected patients that are infected.
  • False negative (FN): Presents the number of misclassified infected patients that are non-infected patients.
A c c u r a c y =   T P + T N T P + F P + T N + F N        
P r e c i s i o n = T P T P + F P  
R e c a l l = T P T P + F N  
A U C = T r u e   p o s i t i v e   r a t e   S e n s i t i v i t y = T P T P + F N
  F a l s e   p o s i t i v e   r a t e = 1 s p e c i f i c i t y = 1 F P T N + F P

5. Metabolic Data Classification Experimental Results and Model Explanations

The next section discusses the experimental findings that were obtained by using the framework that was devised. In this investigation, five different metaheuristics algorithms for feature selection were carried out. Table 4 demonstrates the fine-tuning that was performed on the proposed metaheuristic algorithms. There are a total of six hyperparameters that required adjustment: chaotic type (µ), number of iterations (β), heuristic rate (η), number of generations (α), crossover function probability (ϓ), and mutation probability (Ψ). The selected features have been shown to have the greatest discriminative power for prediction models that make use of metaheuristics methods. Figure 3 shows the total number of selected features utilizing metaheuristics. For clarity, we found that subject age, gene a, gene b, gene c, gene d, breathing rate, high blood pressure, high triglyceride, comprehensive metabolic panel test results, maternal pregnancy record, premature delivery, per MCL quantity of white blood cells, maternal abortion count, and reduced HDL were among the most informative features and indicators for metabolic syndrome based on metaheuristics performance.
The testing accuracy results of several classifiers that were evaluated on the metabolic syndrome dataset using metaheuristic approaches are shown in Table 5. The data, taken as a whole, show an increase in accuracy. Using the genetic and bat algorithms, however, the KNNs classifier outperforms others with a 94.4% accuracy rate. To be clear, this is due to the KNNs algorithm performing better with fewer features, and so the number of features was lowered following the feature selection process utilizing metaheuristics techniques. Another explanation is that KNNs is a distance-based and sluggish algorithm by nature, therefore it outperforms others in the evaluation stage owing to data distribution and training, which is not surprising. On the other hand, the performance of boosting ensemble-based classifiers was superior to that of statistics and bagging classifiers. It was found that the GB, SGB, and CatBoost classifiers had the highest improvement in results, with an average accuracy range increase of between 33% and 35%. This was established by comparing the new findings to the pure results. However, this is not surprising given the fact that boosting-based classifiers decrease bias by increasing variance. In contrast to boosting classifiers, the logistic regression classifier exhibited a 15% gain in accuracy. This is because logistic regression is a linear model, but metabolic data distribution has nonlinear decision boundaries. Additionally, in order to determine which metaheuristic algorithms provided the best results, the average testing accuracy was computed. Notably, the genetic algorithm and the bat optimizer were superior to other methods, since they had an average testing accuracy of 78.1%. In addition, it was discovered that the classifiers KNNs, DTs, GB, SGB, and CatBoost represented significant results when comparing precision and recall, as is shown in Table 6 and Table 7. However, by calculating the average value of precision and recall, it was found that particle swarm optimization, firefly algorithm, and ant colony optimizer had the highest outcome improvement, with an average value of 80% for precision and 79% for recall. Table 8 shows the findings of employing AUC to evaluate the effectiveness of the classifier. However, it is noteworthy that RFs, GB, SGB, and CatBoost classifiers exhibited the best performance, with an AUC of 96.9% as the highest result.
For the purpose of this study, we made use of the SHAP library to explain the results of machine learning models. On the other hand, the bee swarm plot was used to demonstrate the most discriminative metabolic syndrome indicators in comparison to the prediction models. Figure 4 shows the most informative features determined by DTs, KNNs, GB, CatBoost, and SGB with the use of the optimizers PSO, FA, and ANT. These classifiers demonstrated the most significant increase in terms of testing accuracy, precision, recall, and area under the curve (AUC). Nonetheless, we found that the comprehensive metabolic panel test, the quantity of MCL white blood cells, the subject’s age, high blood pressure, and breathing rate were the most important metabolic indicators via SHAP.
Furthermore, when comparing the importance ranking of features from one classifier to another, it is worth noting that these rankings may differ depending on the feature selection methods utilized and the process of tweaking their hyperparameters. Nonetheless, in this study, we investigated the application of SMOTE as a data-augmentation strategy that minimizes the sensitivity of classifiers to new data samples. As a result, our goal was to build accurate classification models that are as stable as possible. However, another restriction is that we highlighted other significant variables that might give sufficient and effective metabolic indicators, such as body mass index and waist circumference, that did not exist in the dataset employed.
Figure 5 shows a SHAP analysis of KNNs performance that was accomplished using genetic and bat optimizers. KNNs achieved a score of 94.4% accuracy throughout testing, which was higher than any other classifier. Nevertheless, in order to explain the results of the model, we make use of two different SHAP plots: the global force plot, which shows model outcomes over a variety of data observations, and the local force plot, which illustrates model outcomes across a single data observation. In spite of this, we chose two observations from the data at random, 200 and 700, to investigate which metabolic indicators are the most significant. Subject age of 9, maternal abortion count, per MCL quantity of white blood cells of 3.569, and previous maternal pregnancy record indicators have a positive impact and contribution to the KNNs classifier, according to the SHAP analysis of KNNs, which was based on the SHAP base value of 0.939. On the other hand, the KNNs classifier was used in the bee swarm plot, which presents an illustration of the most important features overall. The variables that were found to have the highest feature values were age, the quantity of white blood cells per MCL, a comprehensive metabolic panel, the number of maternal abortions, and the breathing rate.
To go further for medical considerations, first we filtered metabolic indicators using a straightforward, methodical technique that consisted of two main stages: choosing features using metaheuristics algorithms and then interpreting them using the SHAP tool. Figure 6 shows the intersection of the most significant common features selected using five metaheuristic algorithms and a set of the most important features rated as having the highest discriminating power using SHAP using a Venn diagram. However, based on common features selected by the suggested approaches, such as age, hypertension, white blood cell count, comprehensive medical tests, and raised respiratory rate, it can be inferred that machine learning models obtain the highest accuracy in prediction and classification. However, medical considerations must be taken into account to explain why these qualities were chosen. Therefore, we aim in this study to take medical considerations into account.
For medical considerations, we discuss the ranking of top-importance features as determined by SHAP values below. It should also be noted that the top five most important metabolic indicators are the patient’s age, comprehensive metabolic panel blood tests, per MCL quantity of white blood cells, breathing rate, and high blood pressure. However, when it comes to patient age, we found that the patient’s risk of developing metabolic syndrome increases with age. Seven years of research published in [45] found that the likelihood of having metabolic syndrome has increased since the 1990s. People above the age of 50 were more likely to be obese than those younger. It is also worth mentioning that as people become older, some disorders emerge that have a high likelihood of producing metabolic syndromes, such as insulin resistance, the development and advent of heart disease, and vascular diseases. These disorders, which worsen with age, have been linked to the development of metabolic syndrome [46]. Therefore, it may be argued that age plays a significant influence in predicting whether or not an individual is impacted by metabolic disease. However, when it comes to comprehensive metabolic panel blood testing, we found that it covers glucose, a type of sugar that the body needs for energy, and that a high glucose rate may indicate a risk of developing metabolic disease. It also includes measures for carbon dioxide, potassium, chloride, triglyceride levels, and cholesterol. Comprehensive metabolic blood tests have identified these as the most common metabolic syndrome criteria [47]. Furthermore, triglyceride and cholesterol levels were defined as features in the utilized dataset. As a result, based on data distribution, they provide a strong indicator of the significance of the whole metabolic blood test characteristic. As a result, it was evaluated as the first and third most important indication according to the majority of classifiers.
Furthermore, we found that due to their influence on the body’s immunity, white blood cells are one of the first lines of defense in the body. As a result, the higher the number of white blood cells, the stronger the body’s resistance to diseases, and vice versa. Nonetheless, the studies in [48,49] emphasize a relationship between white blood cells and metabolic syndrome. Additionally, they noticed a correlation between white blood cells and other tests, such as high blood pressure, which was identified as another crucial metabolic indication using machine learning algorithms in this study. Some other measurements related to the number of white blood cells are insulin, triglycerides, and body mass index. Therefore, it can be concluded that people with metabolic syndrome have more white blood cells, in addition to some observations on the high levels of some proteins such as C-reactive [50]. The metabolic syndrome, on the other hand, is heavily influenced by breathing rate. Increasing the rate of breathing raises the body’s metabolism, which enhances the body’s ability to burn extra fat and vice versa.
For high blood pressure, which is also known as hypertension, it was found by prior studies that there is a strong correlation between blood pressure and metabolic syndrome [51], particularly in the case of severe hypertension. It is stressed, however, that increasing body mass and weight causes a rise in blood pressure. It is also correlated to other tests such as heart pulse rate, where an overweight person’s heart pulse rate rises to allow blood to circulate to the body, which raises blood pressure. However, it should be noted that some studies found that the relationship between high blood pressure and metabolism is still not fully understood, and therefore they found that the body’s insulin resistance and visceral obesity scale are classified as risks leading to high blood pressure in metabolic syndrome [52]. As a result, machine learning algorithms recognized high blood pressure as one of the most significant metabolic indicators, owing to its strong association with other data such as pulse rate.
To summarize and highlight the proposed methodology’s state of the art in comparison to prior methods, the goal of this work is to highlight the highest indices of discriminative metabolism by interpreting machine learning models using the SHAP approach. Other research revealed significant metabolite indicators that must be considered when developing effective prediction models using feature selection approaches. However, this is not enough to explain all predicted values generated using machine learning models. Therefore, by obtaining the average predictive values for the collection of predicted indicators and evaluating the performance of classifiers at certain samples, we may achieve a reasonable degree of interpretation to understand the reasons behind machine learning models’ preference for some features over others. Furthermore, earlier research has concentrated on the application of a variety of specialized types of machine learning models. However, in this study, we investigated and compared the application of ten machine learning classifiers from all classes, including statistical, bagging, and boosting, to build and find the most accurate models for data classification.
Moreover, when compared to prior studies, we applied data augmentation approaches such as SMOTE to lower the degree of sensitivity of machine learning models to future data. Additionally, it has the advantage of not copying data records when compared to other data resampling methods; instead, it generates synthetic data samples, in addition to the examination of the use of natural-derived algorithms such as metaheuristics, which is regarded as one of the contemporary ways of selecting attributes when compared to other methods such as filter and wrappers techniques. However, we noted that the majority of prior research concentrated on the usage of certain data quality in a given region and culture. As a result, the goal of this study was to employ a dataset that focuses on metabolism at the public level without regard to culture or geography.

6. Conclusions

Metabolic syndrome is not a condition that stands on its own as a disease. Instead, it is a collection of risk factors, including high blood pressure, high blood sugar, abnormal cholesterol levels, and fat accumulation in the abdominal region. In this paper, we propose a new framework for classifying metabolic syndrome data. The framework designed has the ability to identify high-risk individuals and provide references for MetS prevention and therapy. Ten machine learning models of various statistical, bagging, and boosting types were evaluated in this study using different patients’ medical data, blood tests, four gene types, and metabolic panel indicators. Five metaheuristics algorithms were used for feature selection: particle swarm optimization, genetic algorithm, bat algorithm, firefly algorithm, and ant colony optimization. The main contribution of this paper, however, is to explain the outputs of machine learning models and highlights the most important metabolic indicators using the Shapley additive explanation library. In addition, we used SMOTE as a data resampling technique for data balance. The findings show that KNNs outperforms others, with a testing accuracy of 94.4% and AUC of 84.4%. Furthermore, we found that patient age, comprehensive metabolic panel blood test, per MCL quantity of white blood cells, breathing rate, and high blood pressure are among the top five most informative metabolic indicators when compared to machine learning models. Additionally, we concluded that GA and BA surpass others, with an average testing accuracy of 78.1%.

Author Contributions

Conceptualization, M.G.S., Y.A.-S., A.A.-Q. and K.C.S.; methodology, M.G.S., Y.A.-S. and A.A.-Q.; resources, K.K.G., M.K.A., S.N. and Y.K.; writing—original draft preparation, M.G.S., Y.A.-S., A.A.-Q. and K.C.S.; writing—review and editing, M.G.S., Y.A.-S., A.A.-Q., K.C.S., K.K.G., M.K.A., S.N. and Y.K.; project administration, M.G.S.; funding acquisition, M.G.S. and M.K.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Deanship of Scientific Research at Jouf University under Grant number DSR2022-RG-0157.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vasile, C.; Baican, M. Progresses in Food Packaging, Food Quality, and Safety—Controlled-Release Antioxidant and/or Antimicrobial Packaging. Molecules 2021, 26, 1263. [Google Scholar] [CrossRef] [PubMed]
  2. Martinez-Ferran, M.; de la Guía-Galipienso, F.; Sanchis-Gomar, F.; Pareja-Galeano, H.J.N. Metabolic impacts of confinement during the COVID-19 pandemic due to modified diet and physical activity habits. Nutrients 2020, 12, 1549. [Google Scholar] [CrossRef] [PubMed]
  3. Devnath, L.; Fan, Z.; Luo, S.; Summons, P.; Wang, D. Detection and Visualisation of Pneumoconiosis Using an Ensemble of Multi-Dimensional Deep Features Learned from Chest X-rays. Int. J. Environ. Res. Public Health 2022, 19, 11193. [Google Scholar] [CrossRef] [PubMed]
  4. Tang, Y.; Zhao, T.; Huang, N.; Lin, W.; Luo, Z.; Ling, C. Identification of Traditional Chinese Medicine Constitutions and Physiological Indexes Risk Factors in Metabolic Syndrome: A Data Mining Approach. Evid.-Based Complement. Altern. Med. 2019, 2019, 1686205. [Google Scholar] [CrossRef] [PubMed]
  5. Tanner, R.M.; Brown, T.M.; Muntner, P. Epidemiology of obesity, the metabolic syndrome, and chronic kidney disease. Curr. Hypertens. Rep. 2012, 14, 152–159. [Google Scholar] [CrossRef]
  6. Chen, Y.-G.; Yang, C.-W.; Chung, C.-H.; Ho, C.-L.; Chen, W.-L.; Chien, W.-C. The association between metabolic risk factors, nonalcoholic fatty liver disease, and the incidence of liver cancer: A nationwide population-based cohort study. Hepatol. Int. 2022, 16, 807–816. [Google Scholar] [CrossRef]
  7. Cefalu, W.T.; Rodgers, G.P. COVID-19 and metabolic diseases: A heightened awareness of health inequities and a renewed focus for research priorities. Cell Metab. 2021, 33, 473–478. [Google Scholar] [CrossRef]
  8. Ibrahim, M.S.; Pang, D.; Randhawa, G.; Pappas, Y. Risk models and scores for metabolic syndrome: Systematic review protocol. BMJ Open 2019, 9, e027326. [Google Scholar] [CrossRef]
  9. Kong, S.; Cho, Y.S. Identification of female-specific genetic variants for metabolic syndrome and its component traits to improve the prediction of metabolic syndrome in females. BMC Med. Genet. 2019, 20, 99. [Google Scholar] [CrossRef]
  10. Abd El–Wahab, E.W.; Shatat, H.Z.; Charl, F. Adapting a prediction rule for metabolic syndrome risk assessment suitable for developing countries. J. Prim. Care Community Health 2019, 10, 2150132719882760. [Google Scholar] [CrossRef]
  11. Yao, F.; Zhao, L.; Yang, Y.; Piao, W.; Fang, H.; Ju, L.; Guo, Q.; Yu, D. Association between Early Life Famine Exposure and Metabolic Syndrome in Adulthood. Nutrients 2022, 14, 2881. [Google Scholar] [CrossRef]
  12. Hoyas, I.; Leon-Sanz, M. Nutritional challenges in metabolic syndrome. J. Clin. Med. 2019, 8, 1301. [Google Scholar] [CrossRef] [Green Version]
  13. Lawson, C.E.; Martí, J.M.; Radivojevic, T.; Jonnalagadda, S.V.R.; Gentz, R.; Hillson, N.J.; Peisert, S.; Kim, J.; Simmons, B.A.; Petzold, C.J.; et al. Machine learning for metabolic engineering: A review. Metab. Eng. 2020, 63, 34–60. [Google Scholar] [CrossRef]
  14. Gutiérrez-Esparza, G.O.; Vázquez, O.I.; Vallejo, M.; Hernández-Torruco, J. Prediction of Metabolic Syndrome in a Mexican Population Applying Machine Learning Algorithms. Symmetry 2020, 12, 581. [Google Scholar] [CrossRef] [Green Version]
  15. Choe, E.K.; Rhee, H.; Lee, S.; Shin, E.; Oh, S.-W.; Lee, J.-E.; Choi, S.H. Metabolic Syndrome Prediction Using Machine Learning Models with Genetic and Clinical Information from a Nonobese Healthy Population. Genom. Inform. 2018, 16, e31. [Google Scholar] [CrossRef] [Green Version]
  16. Saffarian, M.; Babaiyan, V.; Namakin, K.; Taheri, F.; Kazemi, T. Developing a Novel Continuous Metabolic Syndrome Score: A Data Mining Based Model. J. AI Data Min. 2021, 9, 193–202. [Google Scholar]
  17. Yang, H.; Yu, B.; Ouyang, P.; Li, X.; Lai, X.; Zhang, G.; Zhang, H. Machine learning-aided risk prediction for metabolic syndrome based on 3 years study. Sci. Rep. 2022, 12, 2248. [Google Scholar] [CrossRef]
  18. Tavares, L.D.; Manoel, A.; Donato, T.H.R.; Cesena, F.; Minanni, C.A.; Kashiwagi, N.M.; da Silva, L.P.; Amaro, E., Jr.; Szlejf, C. Prediction of metabolic syndrome: A machine learning approach to help primary prevention. Diabetes Res. Clin. Pract. 2022, 191, 110047. [Google Scholar] [CrossRef]
  19. Karimi-Alavijeh, F.; Jalili, S.; Sadeghi, M. Predicting metabolic syndrome using decision tree and support vector machine methods. ARYA Atheroscler. 2016, 12, 146–152. [Google Scholar]
  20. Kim, J.; Mun, S.; Lee, S.; Jeong, K.; Baek, Y. Prediction of metabolic and pre-metabolic syndromes using machine learning models with anthropometric, lifestyle, and biochemical factors from a middle-aged population in Korea. BMC Public Health 2022, 22, 664. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Zhang, X.; Razbek, J.; Li, D.; Xia, W.; Bao, L.; Mao, H.; Daken, M.; Cao, M. Opening the black box: Interpretable machine learning for predictor finding of metabolic syndrome. BMC Endocr. Disord. 2022, 22, 214. [Google Scholar] [CrossRef] [PubMed]
  22. Shimoda, A.; Ichikawa, D.; Oyama, H. Prediction models to identify individuals at risk of metabolic syndrome who are unlikely to participate in a health intervention program. Int. J. Med Inform. 2018, 111, 90–99. [Google Scholar] [CrossRef] [PubMed]
  23. Worachartcheewan, A.; Shoombuatong, W.; Pidetcha, P.; Nopnithipat, W.; Prachayasittikul, V.; Nantasenamat, C. Predicting Metabolic Syndrome Using the Random Forest Method. Sci. World J. 2015, 2015, 581501. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  25. Zhu, T.; Lin, Y.; Liu, Y. Synthetic minority oversampling technique for multiclass imbalance problems. Pattern Recognit. 2017, 72, 327–340. [Google Scholar] [CrossRef]
  26. Liew, B.X.W.; Kovacs, F.M.; Rügamer, D.; Royuela, A. Machine learning versus logistic regression for prognostic modelling in individuals with non-specific neck pain. Eur. Spine J. 2022, 31, 2082–2091. [Google Scholar] [CrossRef]
  27. Bansal, M.; Goyal, A.; Choudhary, A. A comparative analysis of K-Nearest Neighbor, Genetic, Support Vector Machine, Decision Tree, and Long Short Term Memory algorithms in machine learning. Decis. Anal. J. 2022, 3, 100071. [Google Scholar] [CrossRef]
  28. Rivera-Lopez, R.; Canul-Reich, J.; Mezura-Montes, E.; Cruz-Chávez, M.A. Induction of decision trees as classification models through metaheuristics. Swarm Evol. Comput. 2022, 69, 101006. [Google Scholar] [CrossRef]
  29. Hosseinpour-Zarnaq, M.; Omid, M.; Biabani-Aghdam, E. Fault diagnosis of tractor auxiliary gearbox using vibration analysis and random forest classifier. Inf. Process. Agric. 2022, 9, 60–67. [Google Scholar] [CrossRef]
  30. Wang, M.; Liang, Y.; Hu, Z.; Chen, S.; Shi, B.; Heidari, A.A.; Zhang, Q.; Chen, H.; Chen, X. Lupus nephritis diagnosis using enhanced moth flame algorithm with support vector machines. Comput. Biol. Med. 2022, 145, 105435. [Google Scholar] [CrossRef]
  31. Kordi, M.; Dehghan, M.J.; Shayesteh, A.A.; Azizi, A. The impact of artificial intelligence algorithms on management of patients with irritable bowel syndrome: A systematic review. Inform. Med. Unlocked 2022, 29, 100891. [Google Scholar] [CrossRef]
  32. Liu, Y.; Méric, G.; Havulinna, A.S.; Teo, S.M.; Åberg, F.; Ruuskanen, M.; Sanders, J.; Zhu, Q.; Tripathi, A.; Verspoor, K.; et al. Early prediction of incident liver disease using conventional risk factors and gut-microbiome-augmented gradient boosting. Cell Metab. 2022, 34, 719–730.e4. [Google Scholar] [CrossRef]
  33. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  34. Abbasniya, M.R.; Sheikholeslamzadeh, S.A.; Nasiri, H.; Emami, S. Classification of Breast Tumors Based on Histopathology Images Using Deep Features and Ensemble of Gradient Boosting Methods. Comput. Electr. Eng. 2022, 103, 108382. [Google Scholar] [CrossRef]
  35. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar] [CrossRef]
  36. Whitley, D. An overview of evolutionary algorithms: Practical issues and common pitfalls. Inf. Softw. Technol. 2001, 43, 817–831. [Google Scholar] [CrossRef]
  37. Neto, R.T.; Filho, M.G. Literature review regarding Ant Colony Optimization applied to scheduling problems: Guidelines for implementation and directions for future research. Eng. Appl. Artif. Intell. 2013, 26, 150–161. [Google Scholar] [CrossRef]
  38. Marini, F.; Walczak, B.J. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
  39. Kakudi, H.A.; Loo, C.K.; Pasupa, K. Risk quantification of metabolic syndrome with quantum particle swarm optimisation. In Proceedings of the 26th International Conference on World Wide Web Companion, Perth, Australia, 3–7 May 2017; pp. 1141–1147. [Google Scholar]
  40. Fister, I.; Fister, I., Jr.; Yang, X.-S.; Brest, J. A comprehensive review of firefly algorithms. Swarm Evol. Comput. 2013, 13, 34–46. [Google Scholar] [CrossRef] [Green Version]
  41. Ardam, S.; Soleimanian Gharehchopogh, F. Diagnosing liver disease using firefly algorithm based on Adaboost. J. Health Adm. 2019, 22, 61–77. [Google Scholar]
  42. Jayabarathi, T.; Raghunathan, T.; Gandomi, A.H. The Bat Algorithm, Variants and Some Practical Engineering Applications: A Review. Nat.-Inspired Algorithms Appl. Optim. 2017, 744, 313–330. [Google Scholar]
  43. Jeyasingh, S.; Veluchamy, M. Modified bat algorithm for feature selection with the wisconsin diagnosis breast cancer (WDBC) dataset. Asian Pac. J. Cancer Prev. 2017, 18, 1257. [Google Scholar] [PubMed]
  44. Ranjan, G.; Verma, A.K.; Radhika, S. K-nearest neighbors and grid search cv based real time fault monitoring system for industries. In Proceedings of the 2019 IEEE 5th International Conference for Convergence in Technology (I2CT), Bombay, India, 29–31 March 2019; pp. 1–5. [Google Scholar]
  45. Kraja, A.T.; Borecki, I.B.; North, K.; Tang, W.; Myers, R.H.; Hopkins, P.N.; Arnett, D.; Corbett, J.; Adelman, A.; Province, M.A. Longitudinal and age trends of metabolic syndrome and its risk factors: The Family Heart Study. Nutr. Metab. 2006, 3, 41. [Google Scholar] [CrossRef] [PubMed]
  46. Guarner-Lans, V.; Rubio-Ruiz, M.E.; Pérez-Torres, I.; de MacCarthy, G.B. Relation of aging and sex hormones to metabolic syndrome and cardiovascular disease. Exp. Gerontol. 2011, 46, 517–523. [Google Scholar] [CrossRef] [PubMed]
  47. Ladla, A.; Tongkrajai, P.; Srisaenpang, S.; Siviroj, P.; Yutthakasemsunt, S.; Tiamkao, S.; Chotmongkol, V.; Sawanyawisuth, K. Which diagnostic criteria of metabolic syndrome are predictors of cardiovascular diseases in elderly populations? J. Clin. Transl. Endocrinol. 2020, 23, 100248. [Google Scholar] [CrossRef]
  48. Hedayati, M.T.; Montazeri, M.; Rashidi, N.; Abdolmaleki, E.Y.; Shafiee, M.-A.; Maleki, A.; Farmani, M.; Montazeri, M. White blood cell count and clustered components of metabolic syndrome: A study in western Iran. Casp. J. Intern. Med. 2021, 12, 59–64. [Google Scholar]
  49. Friedman, G.D.; Tekawa, I.; Grimm, R.H.; Manolio, T.; Shannon, S.G.; Sidney, S. The Leucocyte Count: Correlates and Relationship to Coronary Risk Factors: The CARDIA Study. Int. J. Epidemiol. 1990, 19, 889–893. [Google Scholar] [CrossRef]
  50. Park, J.T.; Chang, T.I.; Kim, D.K.; Choi, H.Y.; Lee, J.E.; Kim, H.W.; Chang, J.H.; Park, S.Y.; Kim, E.; Yoo, T.-H.; et al. Association of white blood cell count with metabolic syndrome in patients undergoing peritoneal dialysis. Metabolism 2009, 58, 1379–1385. [Google Scholar] [CrossRef]
  51. Iwasaki, Y.; Arisawa, K.; Katsuura-Kamano, S.; Uemura, H.; Tsukamoto, M.; Kadomatsu, Y.; Okada, R.; Hishida, A.; Tanaka, K.; Hara, M.; et al. Associations of Nutrient Patterns with the Prevalence of Metabolic Syndrome: Results from the Baseline Data of the Japan Multi-Institutional Collaborative Cohort Study. Nutrients 2019, 11, 990. [Google Scholar] [CrossRef] [Green Version]
  52. Morse, S.A.; Zhang, R.; Thakur, V.; Reisin, E. Hypertension and the metabolic syndrome. Am. J. Med. Sci. 2005, 330, 303–310. [Google Scholar] [CrossRef]
Figure 1. Metabolic syndrome diagnosis using machine learning.
Figure 1. Metabolic syndrome diagnosis using machine learning.
Diagnostics 12 03117 g001
Figure 2. An overview of the proposed framework for metabolic data classification.
Figure 2. An overview of the proposed framework for metabolic data classification.
Diagnostics 12 03117 g002
Figure 3. The total number of selected features utilizing metaheuristics algorithms.
Figure 3. The total number of selected features utilizing metaheuristics algorithms.
Diagnostics 12 03117 g003
Figure 4. The most informative metabolic indicators using PSO, FA, and ANT optimizers. Blue dots represent a low contribution to the predictive model, whereas red dots represent high impact. (A) Shows quantity of white blood cells as the most important feature according to DTs. (B) Shows the subject age as the most important feature according to KNNs. (C) Shows quantity of white blood cells as the most important feature by GB. (D,E) Show the comprehensive metabolic panel test as the most informative feature according to CatBoost and SGB.
Figure 4. The most informative metabolic indicators using PSO, FA, and ANT optimizers. Blue dots represent a low contribution to the predictive model, whereas red dots represent high impact. (A) Shows quantity of white blood cells as the most important feature according to DTs. (B) Shows the subject age as the most important feature according to KNNs. (C) Shows quantity of white blood cells as the most important feature by GB. (D,E) Show the comprehensive metabolic panel test as the most informative feature according to CatBoost and SGB.
Diagnostics 12 03117 g004
Figure 5. SHAP analysis over the KNNs classifier utilizing GA and BA optimizers. (A) Shows the most informative metabolic indicators found via KNNs. Blue dots show negative impact, whereas red dots show positive impact. (B) Shows SHAP global and local force plots at data observations of 200 and 700.
Figure 5. SHAP analysis over the KNNs classifier utilizing GA and BA optimizers. (A) Shows the most informative metabolic indicators found via KNNs. Blue dots show negative impact, whereas red dots show positive impact. (B) Shows SHAP global and local force plots at data observations of 200 and 700.
Diagnostics 12 03117 g005
Figure 6. An analysis of the intersection of the most important metabolic indicators found via metaheuristics algorithms and SHAP. (A) Shows a set of common selected features by ACO, GA, PSO, FA, and BA algorithms as the most important indicators. (B) Shows as set of classified features as the highest discriminative power for prediction by SHAP values. (C) Shows an intersection for the final most informative features between metaheuristic algorithms and SHAP.
Figure 6. An analysis of the intersection of the most important metabolic indicators found via metaheuristics algorithms and SHAP. (A) Shows a set of common selected features by ACO, GA, PSO, FA, and BA algorithms as the most important indicators. (B) Shows as set of classified features as the highest discriminative power for prediction by SHAP values. (C) Shows an intersection for the final most informative features between metaheuristic algorithms and SHAP.
Diagnostics 12 03117 g006
Table 1. Summary of the related works.
Table 1. Summary of the related works.
PaperDatasetClassifiersMetaheuristicsMetabolic Best IndicatorsResults
[14]2942 patients living in Mexico CityRFs, C 4.5, JRipNO
  • Triglycerides
  • HDL cholesterol
  • Waist circumference
RFs outperforms others.
[15]10,349 of non-obese Korean patientsNB, RFs, DTs, ANN, SVMNONONB best performing with AUC (69%).
[16]3577 students in BirjandCART, ANN, SVMNO
  • Waist circumference
  • High-density lipoprotein
CART classifiers outperform others.
[17]67,730 Chinese patientsRFs, XGBoost, stackingNO
  • Fasting triglyceride level
  • Body mass index
  • Abdominal obesity
XGBoost best results with AUC (93%).
[18]Medical tests of 17,182 patientsLG, XGBoost, KNNs, LGB, DTs, linear analysisNO
  • Waist circumference
  • Triglycerides
  • HDL-cholesterol
LGB best results with AUC (86%).
[19]Isfahan cohort study datasetSVM, DTsNONOSVM best results with an accuracy of 75%.
[20]Medical tests of 1991 middle-aged Korean patients1D NN, DTs, NB, KNNs, XGBoost, RFs, LG, SVM, ANNNO
  • Waist-to-hip ratio
  • BMI
XGBoost best results with an AUC (85%).
[21]39,134 Chinese metabolic syndrome patientsLG, RFs, XGBoostNO
  • Fasting triglycerides
  • Central adiposity
  • Systolic blood pressure
XGBoost with an accuracy of 99.7%.
[22]Japanese metabolic syndrome datasetGBT, RFs, LGNONOGBT performs the best with an AUC of 89.4%
[23]Medical tests of 5646 patients in BangkokRFsNO
  • Triglyceride level
RFs with an accuracy of 98.11%.
Table 2. Metabolic dataset description and analysis.
Table 2. Metabolic dataset description and analysis.
Metabolic Dataset Description
FeaturesDescription
Subject IDDescribes the patient’s ID.
Subject ageDescribes the patient’s age.
Gene ADescribes the patient’s Gene A in DNA.
Gene BDescribes the patient’s Gene B in DNA.
Gene CDescribes the patient’s Gene C in DNA.
Gene DDescribes the patient’s Gene D in DNA.
Per MCL quantity of blood cellsDescribes the patient’s blood cells per microliter. The typical range of adults is between 4.35 to 5.65 million blood cells.
Breathing rateDescribes the patient’s breathing rate. It is a measurement to check if the patient has breathing difficulty.
Pulse rateDescribes the patient’s heart pulse rate.
Diagnostic testingDescribes if the patient has any medical test records.
Carrier testingDescribes if the patient has had any carrier test (a type of genetic test that is used to determine if the patient is a carrier of specific diseases) before.
Enzyme testDescribes if the patient has had any enzyme test (a blood test that measures if the patient has a specific disease) before.
Insulin testDescribes if the patient has any insulin test records.
Thyroid testDescribes if the patient has records of any thyroid tests (a type of blood test that is used to measure thyroid performance).
GenderDescribes the patient’s gender (male/female).
Gastrin defectDescribes if the patient has a gastrin hormone defect or not.
Neural anomalyDescribes if the patient has any neural anomaly tests.
Presence of severe allergiesDescribes if the patient has any allergies.
Premature deliveryDescribes if the patient has any premature delivery record (indicates an early baby birth).
Assistance needed in fertilityDescribes if the patient has needed any assistance in fertility.
Previous maternal pregnancy recordDescribes if the patient has any previous maternal pregnancy record.
Maternal abortion countDescribes the patient’s total number of abortions.
Per MCL quantity of white blood cellsDescribes the patient’s white blood cells per microliter.
CMP resultsStands for the comprehensive metabolic panel, which is a blood test that provides information about body metabolism.
High triglyceride levelDescribes if the patient has high triglyceride.
Reduced HDLDescribes if the patient has a low cholesterol level, which indicates a potential for heart disease.
High BPDescribes if the patient has high blood pressure.
Metabolic syndrome typeTarget classes.
Total number of features28
Total number of records12,012
Total number of data in population348,348
Table 3. GridSearchCV search space values for classifier hyperparameter tuning.
Table 3. GridSearchCV search space values for classifier hyperparameter tuning.
ParametersGrid-Search Space
Number of estimators50, 70, 90, 100, 120, 150, 180, 200
Learning rate0.001, 0.01, 0.1, 1, 10
Loss functionDeviance, exponential
AlgorithmsSAMME, SAMME.R
Table 4. Metaheuristics algorithms for feature selection hyperparameter tuning.
Table 4. Metaheuristics algorithms for feature selection hyperparameter tuning.
Meta-Heuristics AlgorithmsHyperparametersPopulation Size
ACOµ = logistic map, β = 20, η = 0.720
GAα = 20, ϓ = 0.6, ψ = 0.03320
PSOβ = 20, ψ = 0.0120
FAµ = logistic map, ψ = 0.01, β = 2020
BAµ = logistic map, β = 20, ψ = 0.0120
Table 5. Classifier testing accuracy results over metabolic dataset utilizing metaheuristics.
Table 5. Classifier testing accuracy results over metabolic dataset utilizing metaheuristics.
ClassifiersPure(PSO/FA/ANT)(GA/BAT)
Logistic regression0.517270.6690.6746
KNNs0.46230.9300.9449
SVM0.51100.7240.7236
DTs0.44690.8210.8169
RFs0.51100.6600.6578
AdaBoost0.510.66 0.6578
GB0.52850.8790.8887
SGB0.52760.8450.8588
CatBoost0.53260.79540.8349
XGBoost0.51260.6610.6495
Accuracy (AVG)0.50740.7550.7814
Table 6. Classifier performance using precision metric.
Table 6. Classifier performance using precision metric.
Classifiers(PSO/FA/ANT)(GA/BAT)
Logistic regression0.660.67
KNNs0.930.94
SVM0.720.72
DTs0.830.82
RFs0.680.71
AdaBoost0.820.71
GB0.880.89
SGB0.840.86
CatBoost0.790.84
XGBoost0.670.66
Precision (AVG)0.80.79
Table 7. Classifier performance using recall metric.
Table 7. Classifier performance using recall metric.
Classifiers(PSO/FA/ANT)(GA/BAT)
Logistic regression0.670.67
KNNs0.930.94
SVM0.720.72
DTs0.820.82
RFs0.660.67
AdaBoost0.820.67
GB0.880.89
SGB0.840.86
CatBoost0.800.83
XGBoost0.670.66
Recall (AVG)0.790.78
Table 8. Classifier AUC results.
Table 8. Classifier AUC results.
Classifiers(PSO/FA/ANT)(GA/BAT)
Logistic regression0.8370.838
KNNs0.8680.844
SVM0.8390.852
DTs0.8020.797
RFs0.9470.953
AdaBoost0.7610.758
GB0.9690.959
SGB0.9690.963
CatBoost0.9410.967
XGBoost0.8960.857
AUC (AVG)0.8880.883
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sghaireen, M.G.; Al-Smadi, Y.; Al-Qerem, A.; Srivastava, K.C.; Ganji, K.K.; Alam, M.K.; Nashwan, S.; Khader, Y. Machine Learning Approach for Metabolic Syndrome Diagnosis Using Explainable Data-Augmentation-Based Classification. Diagnostics 2022, 12, 3117. https://doi.org/10.3390/diagnostics12123117

AMA Style

Sghaireen MG, Al-Smadi Y, Al-Qerem A, Srivastava KC, Ganji KK, Alam MK, Nashwan S, Khader Y. Machine Learning Approach for Metabolic Syndrome Diagnosis Using Explainable Data-Augmentation-Based Classification. Diagnostics. 2022; 12(12):3117. https://doi.org/10.3390/diagnostics12123117

Chicago/Turabian Style

Sghaireen, Mohammed G., Yazan Al-Smadi, Ahmad Al-Qerem, Kumar Chandan Srivastava, Kiran Kumar Ganji, Mohammad Khursheed Alam, Shadi Nashwan, and Yousef Khader. 2022. "Machine Learning Approach for Metabolic Syndrome Diagnosis Using Explainable Data-Augmentation-Based Classification" Diagnostics 12, no. 12: 3117. https://doi.org/10.3390/diagnostics12123117

APA Style

Sghaireen, M. G., Al-Smadi, Y., Al-Qerem, A., Srivastava, K. C., Ganji, K. K., Alam, M. K., Nashwan, S., & Khader, Y. (2022). Machine Learning Approach for Metabolic Syndrome Diagnosis Using Explainable Data-Augmentation-Based Classification. Diagnostics, 12(12), 3117. https://doi.org/10.3390/diagnostics12123117

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop