Next Article in Journal
Analysis of the Effects of Day-Time vs. Night-Time Surgery on Renal Transplant Patient Outcomes
Next Article in Special Issue
Outpatient Readmission in Rheumatology: A Machine Learning Predictive Model of Patient’s Return to the Clinic
Previous Article in Journal
Motor Adaptation Impairment in Chronic Cannabis Users Assessed by a Visuomotor Rotation Task
Previous Article in Special Issue
Exploration of Artificial Intelligence Use with ARIES in Multiple Myeloma Research
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Physician-Friendly Machine Learning: A Case Study with Cardiovascular Disease Risk Prediction

Department of Electrical and Computer Engineering, University of Houston, Houston, TX 77004, USA
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2019, 8(7), 1050; https://doi.org/10.3390/jcm8071050
Submission received: 18 June 2019 / Revised: 14 July 2019 / Accepted: 15 July 2019 / Published: 18 July 2019
(This article belongs to the Special Issue The Future of Artificial Intelligence in Clinical Medicine)

Abstract

:
Machine learning is often perceived as a sophisticated technology accessible only by highly trained experts. This prevents many physicians and biologists from using this tool in their research. The goal of this paper is to eliminate this out-dated perception. We argue that the recent development of auto machine learning techniques enables biomedical researchers to quickly build competitive machine learning classifiers without requiring in-depth knowledge about the underlying algorithms. We study the case of predicting the risk of cardiovascular diseases. To support our claim, we compare auto machine learning techniques against a graduate student using several important metrics, including the total amounts of time required for building machine learning models and the final classification accuracies on unseen test datasets. In particular, the graduate student manually builds multiple machine learning classifiers and tunes their parameters for one month using scikit-learn library, which is a popular machine learning library to obtain ones that perform best on two given, publicly available datasets. We run an auto machine learning library called auto-sklearn on the same datasets. Our experiments find that automatic machine learning takes 1 h to produce classifiers that perform better than the ones built by the graduate student in one month. More importantly, building this classifier only requires a few lines of standard code. Our findings are expected to change the way physicians see machine learning and encourage wide adoption of Artificial Intelligence (AI) techniques in clinical domains.

1. Introduction

Machine learning and artificial intelligence (AI) have witnessed tremendous progress in the past five years. AI algorithms have gained significant interest from clinical researchers. As an example, a recent survey indicates that nearly half of the interviewed healthcare organizations are using or planning to use artificial intelligence in imaging [1]. Unfortunately, developing machine learning algorithms traditionally requires a significant amount of time and understanding of how the underlying algorithms work. For example, tuning and training of deep neural networks take weeks to months. Most state-of-the-art deep networks have been manually designed by human experts who have advanced degrees and long-term training in computer science and artificial intelligence [2,3,4,5]. Such requirements pose a great challenge for clinical researchers who want to use AI tools to validate important biomedical questions.
Motivated by this limitation, researchers recently investigated more automated AI techniques [6,7,8,9,10,11,12]. These methods are collectively known as Auto Machine Learning (AutoML). The idea is to automate the process of building an AI model that gives competitive performances on any given dataset. This includes automation of data pre-processing, feature extraction, hyper-parameter tuning, and algorithm selection. Here, a feature means a compact vector containing information about the input data essential for making the final prediction. The emergence of AutoML is potentially transformative to the biomedical and clinical domains. By removing the high technical barrier, AutoML could enable physicians to use AI techniques more broadly in their work and research.
AutoML can be viewed as the end-to-end process of searching for the best AI model configuration on an arbitrarily given dataset. Each configuration is the result of making multiple choices regarding which algorithm, optimization method, or hyper-parameter to use. Due to the vast number of configurations in the search space, finding the best model is computationally expensive. As computer processing power improves thanks to advanced technologies such as graphics processing units (GPU) and tensor processing units (TPU), and more efficient search algorithms, AutoML methods have been able to scale up dramatically. Recent papers showed that classifiers built by automated techniques have reached or even surpassed those designed by human experts. Notable examples include AmoebaNet [13] which outperforms most state-of-the-art architectures on a large-scale natural image dataset [14], and Auto-Sklearn [12] which has shown competitive performances on multiple heterogeneous datasets. Major technological companies, including Google and Microsoft, aware of the vast potential of AutoML across industries, have recently started to build their own AutoML platforms. Despite the great potential, AutoML has not been well-studied in biomedical applications. Our paper will make the following contributions:
  • For the first time, we investigate the use of AutoML for building classifiers of cardiovascular diseases.
  • We compare AutoML performances against that of a graduate student with significant experience in machine learning and computer programming.
  • We provide extensive experimental results on two cardiovascular datasets.
This study will inform physicians and biomedical researchers on an important and emerging machine learning tool. Our findings will shed light on what benefit AutoML can bring, how easy it is to use the tool, and how well it performs compared to a human expert. Although we evaluate our method on cardiovascular data, the findings are expected to hold for other kinds of biomedical data.
Given that cardiovascular diseases are the leading cause of deaths in the world today [15] and the availability of tremendous amount of cardiovascular data, there have been numerous studies in the past to get machine learning models to deduce patterns in the data to allow for early detection of heart diseases. Multiple standalone machine learning models and hybrid models have been proposed [16].
Vembandasamy et al. [17] propose the use of Naive Bayes classifier for prediction of heart disease on a dataset from a leading diabetic research institute in Chennai, India containing 500 records and 10 attributes. The Naive Bayes classifier attained accuracy of 86.4%. Shouman et al. [18] propose the Decision Tree classifier on the benchmark Heart UCI (University of California, Irvine, CA, USA) dataset by applying several tuning techniques to Decision Trees like different combinations of discretization, tree types, voting, etc. to identify a reliable, robust and accurate method of classification. The final reported accuracy is 84.1%. Srinivas et al. [19] propose more complicated data mining algorithms. The technique involves the extraction of significant patterns from the dataset, choosing patterns with values greater than a prescribed threshold and using five different mining goals. The reported accuracy is 83.7%. Tomar et al. [20] use Least Squares Twin Support Vector Machines [21] for diagnosis of heart diseases using the grid-search approach for hyperparameter selection and F-scores as the evaluation metric on the heart UCI dataset. Reported accuracy is 85.59%. Several ensemble classifiers, which are a weighted combination of simple classifiers have also been seen to work well with heart disease prediction. Pouriyeh et al. [22] use the Decision Tree classifier, Naïve Bayes classifier, Multilayer Perceptron, K-Nearest Neighbor classifier, Single Conjunctive Rule Learner and Radial Basis Function with Support Vector Machines both individually and in combination on the Heart UCI dataset. In addition, bagging, boosting and stacking techniques have been applied on each of the above-mentioned classifiers. The best performing classifier was reported to be a combination of the Support Vector Machine and the Multilayer Perceptron and the reported accuracy is 84.81%. Bashir et al. [23] propose the use of an ensemble classifier that uses an enhanced bagging approach with the multi-objective weighted voting scheme. Five different base classifiers including Naïve Bayes, linear regression, quadratic discriminant analysis, instance-based learner and support vector machines are used. Five different heart disease datasets are used. The experimental evaluation shows that the proposed framework achieves diagnosis accuracy of 84.16%.
There are several challenges associated with manually training and evaluating machine learning models. Most important is the difficulty in correctly identifying the nature (continuous or categorical) of all features to preprocess them accordingly before passing them into machine learning models. The required expertise and time associated with this task are also significantly high. This study proposes to use auto machine learning as a solution to the above-mentioned problems and attempts to quantify the performance and time benefits that auto machine learning has to offer over a manually built solution.
Section 2 of this paper discusses the processes and techniques involved in the experimental stage. Section 2.1 describes Auto-Sklearn and the techniques adopted by it to find and fine-tune the machine learning model best suited for the dataset. Section 2.2 provides a description of the datasets, the nature of preprocessing and the train-test procedure applied on each of them by the graduate student. Furthermore, Section 2.3 goes on to elucidate challenges that the graduate student faced during manual training and how they were addressed. Section 3 discusses the results drawn from manual training and Auto-Sklearn and provides a comparison between the corresponding evaluation metrics obtained post training. Finally, Section 4, the concluding section, describes how the ease of use and superior performance of the AutoML tool as described in this paper could greatly impact the clinical domain.

2. Materials and Methods

Our experiments focus on evaluating the performances of AutoML techniques on two cardiovascular datasets. In particular, we will use Auto-Sklearn [12], which is one of the state-of-the-art generic AutoML frameworks. Auto-Sklearn has consistently won several AutoML competitions over the years. Moreover, its user interface is friendly and a non-technical person without much prior knowledge in machine learning can quickly achieve mastery. In what follows, we will provide a brief description of Auto-Sklearn and the experimental setup.

2.1. Algorithmic Description of Auto-Sklearn

Auto-Sklearn was proposed in [12]. The name was motivated by Scikit-Learn [24], a popular generic machine learning toolbox. Auto-Sklearn automates the process of building an AI model by utilizing a large number of machine learning classifiers (14 in total) and pre-processing steps (14 feature processing methods, and four data preprocessing methods) in the Scikit-Learn toolbox. This includes logistic regressions, support vector machines, random forests [25], boosting, and neural networks. Figure 1 shows the graphical illustration of the pipeline. Given the training data, Auto-Sklearn first selects an appropriate set of data preprocessing steps such as rescaling or imputation of missing values. It then passes the processed data to the feature processing block, which further normalizes the data or reduces their dimensions using standard techniques such as principal component analysis [26] and independent component analysis [27]. Finally, data are passed to the estimator block, which selects and trains machine learning algorithms to predict desirable outputs from input data samples.
Auto-Sklearn defines AutoML as the process of automatically producing test-set predictions (without any human intervention) given a fixed computational budget. Here, computational budget means computer run time or computer memory usage. Auto-Sklearn combines traditional machine learning techniques with a Bayesian optimization framework to search for the best combination of AI models and parameters. It also introduces several notable improvements compared to previous approaches [28]. First, it uses prior experience on other datasets to create a good model initialization for a new dataset. The central intuition is that domain experts derive knowledge from previous tasks. Motivated by this observation, Auto-Sklearn employs a similar strategy. It collects a set of 38 meta-features, or vector descriptions of dataset properties that would help to determine appropriate algorithms that would likely perform well on a particular dataset. Examples of meta-features include statistics about the number of data samples, data dimensions, classes, and skewness. Based on these features, Auto-Sklearn makes a rough suggestion for what algorithms, pre-processing, and other hyper-parameters will work well on a particular dataset. Bayesian optimization further refines and improves the model. Second, instead of outputting one model, Auto-Sklearn uses a weighted combination of multiple best-performing models. This is similar to the ensemble method in random forests [25] that combines multiple random trees to reduce the prediction variance. Empirical studies found that this modification significantly improves the robustness of the final model [12].
A non-technical person will find Auto-Sklearn intuitive and easy to learn. Figure 2 shows the code for training a classifier for an arbitrary dataset. It essentially contains only four lines of code. The first line loads the Auto-Sklearn library, assuming that this library is already pre-installed in the computer. The second line of code creates an instance of the classifier. One can think of this as a placeholder for the final classifier. The third line of code calls the function .fit to train (also known as fitting) the final classifier given the training data X_train and the corresponding labels y_train. The last line calls the function .predict to make the predictions on the test data X_test.

2.2. Process of Building AutoML Models

The process of building an AutoML model contains two main steps. The first step is to prepare data into an appropriate format and load them into the computer’s memory. In our case, we store our training and test data in .csv tables whose rows and columns represent different patients and features, respectively. We then load those tables into the computer memory using the standard csv library in Python. Readers can refer to [29] for more information about the process of reading .csv files. Let X_train and y_train respectively denote the training features and labels that we have loaded into the computer memory. The second step is to run AutoML code described in Figure 2. This script will import Auto-Sklearn and build a classifier to predict the outcome for each patient. The classifier can be used to make prediction on new patients through the command classifier.predict(x_test), where x_test is the test data from a new patient as shown in Figure 2. In this paper, we use a computer with a 16-core processor (i9-7960X, 16 Cores, 2.80 GHz) and a Titan-V graphic processing unit (NVIDIA GeForce Titan V, 12 GB HBM2 Memory, NVIDIA, Santa Clara, CA, USA). However, we believe that any modern computers will not have problems with running the AutoML code since Auto-Sklearn is written in Python which can compile across different hardware platforms and operating systems.

2.3. Datasets and Manual Preprocessing by a Graduate Student

Training, testing and evaluation are performed on two different cardiovascular datasets, the Heart UCI (University of California, Irvine, CA, USA) dataset and the Cardiovascular Disease Dataset. In what follows, the two datasets, including the nature and meaning of each of their features, are described. Given the vast difference in the nature of these two datasets, subsequent sections address the training procedures and the challenges independently for the two datasets.

2.3.1. Dataset Description

The Heart UCI dataset contains data of patient records with the target field referring to the presence or absence of heart disease. The database has 76 attributes, but only 13 attributes are used for our experiments to make our results comparable to previous machine learning papers. Table 1 shows the selected attributes and their properties. This dataset has in total 303 records, which is relatively small given that a typical machine learning dataset contains several thousand to hundreds of thousands of data points. There have been multiple works [18,30,31,32,33,34,35,36] investigating the performances of different machine learning algorithms on this dataset. The popularity of this dataset makes it easy to know how competitive the results of the graduate student are as well as how the performance of the AutoML method compares to human–experts’ systems. The target variable in this dataset is ‘Target’ in Table 1. Of the 303 records, 138 records are that of patients with Target 0 and 165 records with Target 1.
The cardiovascular disease dataset consists of 70,000 records of patients’ data with the target (Cardio) describing the presence or absence of heart disease using 11 features as described in Table 2 [37]. The input features are of three types: objective (containing factual information), examination (containing the results of a medical examination) and subjective (containing information given by the patient). The target variable in this dataset is ‘Cardio’ in Table 2. Of the 70,000 records, 35,021 records are that of patients with Cardio 0 and 34,979 records are that of patients with Cardio 1.

2.3.2. Manual Data Preprocessing

Data preprocessing is important to ensure the quality of the final machine learning model. The graduate student performs a number of preprocessing steps. The steps include:
  • Looking for missing data and performing missing data imputation (both datasets did not have any missing data).
  • Identifying continuous and categorical features.
  • Identifying ordinal categorical features (where the categories have a natural ordered relationship between each other) and integer encoding them.
  • Identifying nominal categorical features (where there is no natural ordered relationship between categories), and one-hot encoding them.
  • Feature scaling, so as to bring all feature values to a similar dynamic range, which allows for faster convergence of learning algorithms optimized with the gradient descent method.
  • Feature selection, which is the process of choosing relevant features from the given set and eliminating features that do not contribute much to the prediction of the target variable. This reduces training time and improves performance. Statistical tests allow ranking features according to their relation with the target. The F-statistical test (to capture linear relationships between features and target) and mutual information test (to capture linear and nonlinear relationships between features and target) are used independently on the datasets to evaluate the best set of features. In addition, recursive feature elimination, which is a greedy optimization algorithm that intends to find the best subset of features by repeatedly creating models and ranking the model’s performance with each subset, is also used. The above-mentioned statistical techniques are applied to capture complicated and nonlinear relationships of each feature with the target. Once the best set of features are extracted from training data independently using the above techniques, models are fitted on the new feature subsets and the corresponding cross validation performances are evaluated independently and compared so as to make a fool-proof decision on which feature-selection technique works well with an algorithm.

2.3.3. Train and Test Procedure

Our experiments split each dataset into three distinct sets: (1) training set, (2) validation set, and (3) test set. The graduate student uses the validation set for determining the best hyper-parameters and machine learning algorithms. Once the best model and its hyper-parameters are found for a particular dataset, the student merges the training and the validation sets together and performs training on the merged set to get the final model. Note that Auto-Sklearn does not need the validation set since the hyper-parameters are selected automatically by the framework. Therefore, we merge the training set and validation set together and use that as the training data for Auto-Sklearn. We compare the graduate student’s results to that of Auto-Sklearn using the test set, which has never been used in any way during the training step.
The Heart-UCI dataset, with only 303 samples (a sample is a data point or record in the dataset that is, one trial subject’s information), demands a careful split ratio to ensure that there are enough test samples to provide a fair representation of the dataset when the model is put to test, while at the same time having enough data samples to train and validate the model performance. One-hundred samples are set aside for testing while retaining 203 for training and validation. Cross validation is performed using the k-fold cross validation technique, wherein the data are split into k groups, trained and fitted on k-1 groups and validated on one group. This procedure repeats until all k groups are validated once. The mean performance of the model on the k folds serves as the estimated cross validation performance. k is chosen to be 5 in order to ensure that each fold is large enough to be representative of the whole dataset.
The Cardiovascular disease dataset has 70,000 records, of which 14,000 records serve as the test dataset and the remaining 56,000 records are used for training and cross validation. With most machine learning models used to train this dataset, k-fold cross validation is used with k = 10, but, in the case of certain models (specifically Neural networks, Support vector machines and Bagged K-Nearest Neighbors), k-fold cross validation consumes high execution times (due to large dataset size), and the hold out cross validation scheme (where a portion of the training data are sampled before training and serves as the cross validation set) is applied in those cases. Both datasets are randomly sampled to ensure that each target subgroup receives proper representation within sets.
The graduate student uses ten main machine learning models in Scikit-Learn to build classifiers for the two datasets. These models include the logistic regression model, the support vector machines with different kernels, the decision tree and ensemble tree models, and the boosting and bagging classifiers with appropriate models as base classifiers. Finally, an ensemble classifier that combines the best performing base classifiers is trained. The datasets are manually trained and tested over a span of 30 days until satisfactory performance that works well in terms of bias and variance is obtained. The best model found fit for the Heart UCI dataset after tuning and cross validating over 17 days is found to be the hyperparameter tuned Linear Support Vector classifier (with features selected using the Recursive Feature Selection technique), while the best model for the Cardiovascular Disease Dataset after 15 days of training was found to be the bagged and hyper-parameter tuned decision tree model.

2.4. Challenges during Training Faced by a Graduate Student

Heart UCI Dataset: Given the small size of the dataset, over-fitting of models to training data poses a major challenge. High variance in Logistic Regression and Support vector models is addressed by carefully tuning the regularization parameter and setting it to the value at which cross-validation performance begins to fall and training performance begins to increase rapidly.
The decision tree model also over-fits the data as expected. While careful hyper-parameter tuning of tree depth significantly improves performance, other model averaging techniques are also used to enhance generalization. Multiple ensemble tree models, which construct multiple (hyper-parameter selected) trees with sample and feature subsets, are constructed and suitably hyper-parameter tuned.
Given the many parameters (number of hidden layers, number of nodes in each of these hidden layers and the maximum number of iterations, among many others) that need to be tuned in a Multi-Layer Perceptron (or Artificial Neural Network), the challenge is tackled by manually comparing the training and validation performances of the model. The model is kept simple and small, given the high chance of over-fitting. Once the model performance is satisfactory with default maximum iterations, the model shape and size are fixed, and the maximum number of iterations is tuned to ensure the convergence of Gradient Descent.
Cardiovascular Disease Dataset: The biggest challenge associated with training this dataset is that of high training time, limiting the number of models that can be trained. Of all the trained models, the Support Vector Machine models are seen to consume the most time given their computational complexity. As a solution, Principal Component Analysis (PCA) is applied to the input features to obtain dimensionality reduction of the input and lower the computational complexity and speed up the training process.
Given the high computation time, building and evaluating models over large hyper-parameter ranges are time expensive. For example, hyper-parameter tuning of the Adaboost classifier with decision trees as the base classifier takes close to a day. Some common practices are resorted to while selecting hyper-parameters of time expensive models. For example, the value of k is set to be odd and equal to the square root of the number of samples for K-Nearest Neighbors, and the number of nodes in the hidden layer of Multi-Layer Perceptron is set to the average of input and output nodes.

2.5. Comparison of AutoML and Human

Qualification of the graduate student: We design several experiments to compare the performance between AutoML and the graduate student. The student is highly qualified for developing a machine learning model. First, the student has taken a machine learning course and received an A letter grade for the overall performance. The course covers in-depth machine learning theories as well as multiple programming assignments. For this reason, we expect the graduate student to serve as a strong baseline for comparison with AutoML performances. Furthermore, the machine learning models developed by this student for the two datasets produce competitive results compared to prior work as described in more detail in the next section.
Metrics for comparing AutoML and the graduate student: The student was given roughly two weeks for working on each dataset. The first dataset takes 18 days, which is three days longer than the second dataset due to additional time for setting up the project and becoming familiar with the Scikit-Learn. We study how the best results obtained by the student compare to that produced by AutoML. To gain an insight into the student’s progress, we record the classification accuracies, the areas under ROC and PR curves, and major challenges faced by the student over different days. Analyzing where the student spent most of the time and technical challenges will enable a deep understanding of how AutoML will benefit the development of machine learning models.

3. Results and Discussion

3.1. AutoML Benefits Complex Datasets More

In this section, we compare the classification accuracies obtained by the student with that of AutoML. Since the student selects the final model based on the best validation accuracy, this experiment will show how quickly the student can find a good model for a particular dataset. Figure 3 and Figure 4 show the classification over 15–18 day periods for UCI Heart and Cardiovascular Disease datasets, respectively. For both datasets, AutoML achieves competitive validation accuracies compared to that of the student. On the UCI-Heart dataset, the student was able to find a good model from the first day. Since this dataset only has a small number of samples (303 in total), simple classifiers such as linear support vector machine and logistic regression tend to work well. Moreover, the small dataset size makes it faster to run an algorithm and thus reduce the overall development time. On the Cardiovascular Disease dataset, it took the student significantly longer time (seven days) to find a good model. This could be because the second dataset is more complex, demonstrated by the lower validation accuracy of linear classifiers compared to the previous dataset. Moreover, the number of data points is also significantly larger (70,000 of Cardiovascular Disease vs. 303 of UCI-Heart). Our experiment suggests that the time-saving factor is larger for more complex datasets when using AutoML instead of manual model search.

3.2. Comparison of AutoML’s and Graduate Student’s Test-Set Performances

Once the final models are selected for the two datasets based on the best validation accuracies, the student performed inference on the test sets to obtain the final performance measures. Table 3 and Table 4 compare the final models’ classification accuracy, area under ROC curve, and area under precision recall curve to that of AutoML models. Note that AutoML models were evaluated on exactly the same test sets to make the results comparable. AutoML achieves slightly better mean accuracy for the UCI-Heart dataset, and similar accuracy for the Cardiovascular Disease dataset compared to the student. In addition, AutoML achieves significantly better areas under curves on both datasets. This suggests that AutoML classifiers generalize much better than their manual counterparts. Most importantly, AutoML only takes 30 min to build a competitive classifier for each dataset, compared to long periods of time (432 h for UCI and 360 h for Cardiovascular datasets) taken by the graduate student to develop similar classifiers.
Other state-of-the-art studies on the Heart UCI dataset have shown results comparable with those of the graduate student and AutoML. Table 5 presents performance accuracies from training different machine learning models on the Heart UCI dataset. The accuracies obtained by the student and AutoML are on par with those reported in the recent literature. This further supports the claim that the AutoML method is able to quickly find competitive classifiers with minimal human effort.
The Heart UCI dataset contains 76 features, but only 13 most-important features are included since most studies and published papers utilize them to build machine learning models on. This makes it possible to compare our results to these published papers, in order to serve as the baseline to check if the results obtained by the Graduate student and Auto-sklearn are competent enough. However, the potential downside to reduced feature space is loss of information. The other features (not included within these 13 attributes) include information on the subject’s response to exercise Electrocardiogram and cigarette smoking habits among others [38].

4. Conclusions

This study intends to propose the use of AutoML for adoption in the clinical domain by breaking the perception that machine learning is accessible to trained experts only. For the first time, we evaluate the performance of an AutoML library (Auto-Sklearn) on two cardiovascular disease datasets and compare the results to that obtained by a graduate student after a month of effort in training multiple classifiers on the datasets. These two cardiovascular datasets contain clinical data from trial subjects and whether or not they have cardiovascular disease, so that, given a new subject’s data, the model (learned patterns from given data) can predict the presence or absence of cardiovascular disease with a reasonably good accuracy. The results indicate that the graduate student and AutoML report similar accuracies on the two datasets, on par with other state-of-the-art studies. The area under curves for AutoML is significantly higher indicating that the model built by AutoML generalizes better than that of the graduate student. In addition, the time taken by AutoML to produce these results is just around 30 min per dataset, which is significantly less compared to about 400 h taken by the graduate student. The number of lines of code for AutoML is also significantly lesser compared to the several hundred code lines used by the graduate student, hence justifying the ease of use. Thus, our experimental results strongly suggest that AutoML is a promising approach that enables non-technical users to quickly build competitive machine learning models that work as well as those designed by humans with experience in machine learning. This finding is expected to change the way biomedical researchers and physicians view machine learning. The development of AutoML technology is likely to make machine learning tools more accessible and speed up the research discovery process in the clinical community. Although this study focuses on cardiovascular disease datasets, we conjecture that the key findings related to the efficiency and efficacy of AutoML will hold for other biomedical datasets. In the future, we will investigate the effects of AutoML on other clinically relevant tasks such as tumor detection and segmentation from medical images. Another important advantage of AutoML techniques is that they can incorporate additional constraints when searching for AI models. For example, physicians might want to maximize the classification accuracy while ensuring that the classifier’s sensitivity is higher than a certain threshold. Such constraints are hard to optimize in the traditional AI framework. Our future work will evaluate this complex scenario. We expect that the advantage of AutoML will be more prominent when the complexity of the task increases.

Supplementary Materials

The following are available online at https://www.mdpi.com/2077-0383/8/7/1050/s1, Table S1: Validation performances of different algorithms on Heart UCI dataset over days, Table S2: Validation performances of different algorithms on Cardiovascular disease dataset over days, Table S3: Descriptions and parameter settings of algorithms used by the graduate student on the datasets.

Author Contributions

Conceptualization, H.V.N. and P.Y.; methodology, M.P. and H.V.N.; software, M.P. and H.V.N.; validation, M.P. and H.V.N.; formal analysis, M.P. and H.V.N.; investigation, M.P. and H.V.N.; data curation, M.P.; writing—original draft preparation, H.V.N., M.P., P.Y., and G.C.; writing—review and editing, H.V.N., M.P., P.Y., and G.C.; visualization, M.P. and H.V.N.; supervision, H.V.N.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AutoMLAuto Machine Learning
UCIUniversity of California, Irvine
GPUGraphic Processing Unit
TPUTensor Processing Unit
ROCReciever Operating Characteristic
PRPrecision–Recall
PCAPrincipal Component Analysis

References

  1. Bresnick, J. AI for Imaging Analytics Intrigues Healthcare Orgs, However, Starts Slow. 2018. Available online: https://healthitanalytics.com/news/ai-for-imaging-analytics-intrigues-healthcare-orgs-yet-starts-slow (accessed on 10 June 2019).
  2. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  3. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  4. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  5. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  6. Baker, B.; Gupta, O.; Naik, N.; Raskar, R. Designing neural network architectures using reinforcement learning. arXiv Preprint 2016, arXiv:1611.02167. [Google Scholar]
  7. Liu, C.; Zoph, B.; Neumann, M.; Shlens, J.; Hua, W.; Li, L.J.; Fei-Fei, L.; Yuille, A.; Huang, J.; Murphy, K. Progressive neural architecture search. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 19–34. [Google Scholar]
  8. Miikkulainen, R.; Liang, J.; Meyerson, E.; Rawal, A.; Fink, D.; Francon, O.; Raju, B.; Shahrzad, H.; Navruzyan, A.; Duffy, N.; et al. Evolving deep neural networks. In Artificial Intelligence in the Age of Neural Networks and Brain Computing; Elsevier: Amsterdam, The Netherlands, 2019; pp. 293–312. [Google Scholar]
  9. Pham, H.; Guan, M.Y.; Zoph, B.; Le, Q.V.; Dean, J. Faster discovery of neural architectures by searching for paths in a large model. In Proceedings of the ICLR 2018, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  10. Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.L.; Tan, J.; Le, Q.V.; Kurakin, A. Large-scale evolution of image classifiers. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, Sydney, NSW, Australia, 6–11 August 2017; pp. 2902–2911. [Google Scholar]
  11. Thornton, C.; Hutter, F.; Hoos, H.H.; Leyton-Brown, K. Auto-WEKA: Combined selection and hyperparameter optimization of classification algorithms. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 847–855. [Google Scholar]
  12. Feurer, M.; Klein, A.; Eggensperger, K.; Springenberg, J.; Blum, M.; Hutter, F. Efficient and robust automated machine learning. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 2962–2970. [Google Scholar]
  13. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. arXiv Preprint 2018, arXiv:1802.01548. [Google Scholar]
  14. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  15. Mozaffarian, D.; Benjamin, E.J.; Go, A.S.; Arnett, D.K.; Blaha, M.J.; Cushman, M.; Das, S.R.; de Ferranti, S.; Després, J.P.; Fullerton, H.J.; et al. Heart disease and stroke statistics-2016 update a report from the American Heart Association. Circulation 2016, 133, e38–e48. [Google Scholar] [PubMed]
  16. Chitra, R.; Seenivasagam, V. Review of heart disease prediction system using data mining and hybrid intelligent techniques. ICTACT J. Soft Comput. 2013, 3, 605–609. [Google Scholar]
  17. Vembandasamy, K.; Sasipriya, R.; Deepa, E. Heart diseases detection using Naive Bayes algorithm. IJISET-Int. J. Innov. Sci. Eng. Technol. 2015, 2, 441–444. [Google Scholar]
  18. Shouman, M.; Turner, T.; Stocker, R. Using decision tree for diagnosing heart disease patients. In Proceedings of the Ninth Australasian Data Mining Conference-Volume 121, Ballarat, Australia, 1–2 December 2011; pp. 23–30. [Google Scholar]
  19. Srinivas, K.; Rani, B.K.; Govrdhan, A. Applications of data mining techniques in healthcare and prediction of heart attacks. Int. J. Comput. Sci. Eng. (IJCSE) 2010, 2, 250–255. [Google Scholar]
  20. Tomar, D.; Agarwal, S. Feature selection based least square twin support vector machine for diagnosis of heart disease. Int. J. Bio-Sci. Bio-Technol. 2014, 6, 69–82. [Google Scholar] [CrossRef]
  21. Fung, G.; Mangasarian, O.L. Incremental support vector machine classification. In Proceedings of the 2002 SIAM International Conference on Data Mining, Arlington, VA, USA, 11–13 April 2002; pp. 247–260. [Google Scholar]
  22. Pouriyeh, S.; Vahid, S.; Sannino, G.; De Pietro, G.; Arabnia, H.; Gutierrez, J. A comprehensive investigation and comparison of Machine Learning Techniques in the domain of heart disease. In Proceedings of the 2017 IEEE Symposium on Computers and Communications (ISCC), Heraklion, Greece, 3–6 July 2017; pp. 204–207. [Google Scholar]
  23. Bashir, S.; Qamar, U.; Khan, F.H. BagMOOV: A novel ensemble for heart disease prediction bootstrap aggregation with multi-objective optimized voting. Aust. Phys. Eng. Sci. Med. 2015, 38, 305–323. [Google Scholar] [CrossRef] [PubMed]
  24. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  25. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  26. Jolliffe, I. Principal Component Analysis; Springer: New York, NY, USA, 2011. [Google Scholar]
  27. Hyvärinen, A.; Karhunen, J.; Oja, E. Independent Component Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2004; Volume 46. [Google Scholar]
  28. Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H. The WEKA data mining software: An update. ACM SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar] [CrossRef]
  29. CSV File Reading and Writing Tutorial. Available online: https://docs.python.org/3/library/csv.html (accessed on 10 July 2019).
  30. Zhou, Z.H.; Jiang, Y. NeC4. 5: Neural ensemble based C4. 5. IEEE Trans. Knowl. Data Eng. 2004, 16, 770–773. [Google Scholar] [CrossRef]
  31. Bouckaert, R.R.; Frank, E. Evaluating the replicability of significance tests for comparing learning algorithms. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, 26–28 May 2004; pp. 3–12. [Google Scholar]
  32. To, G.B.; Brown, G.; To, G.; Brown, G. Diversity in Neural Network Ensembles; University of Birmingham: Birmingham, UK, 2004. [Google Scholar]
  33. Chai, X.; Deng, L.; Yang, Q.; Ling, C.X. Test-cost sensitive naive bayes classification. In Proceedings of the Fourth IEEE International Conference on Data Mining (ICDM’04), Brighton, UK, 1–4 November 2004; pp. 51–58. [Google Scholar]
  34. Mason, L.; Bartlett, P.L.; Baxter, J. Improved generalization through explicit optimization of margins. Mach. Learn. 2000, 38, 243–255. [Google Scholar] [CrossRef]
  35. Duch, W.; Adamczak, R.; Grabczewski, K. A new methodology of extraction, optimization and application of crisp and fuzzy logical rules. IEEE Trans. Neural Netw. 2001, 12, 277–306. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Wang, S.J.; Mathew, A.; Chen, Y.; Xi, L.F.; Ma, L.; Lee, J. Empirical analysis of support vector machine ensemble classifiers. Expert Syst. Appl. 2009, 36, 6466–6476. [Google Scholar] [CrossRef] [Green Version]
  37. Ulianova, S. Cardiovascular Disease Dataset. Available online: https://www.kaggle.com/sulianova/cardiovascular-disease-dataset (accessed on 10 July 2019).
  38. Janosi, A.; Steinbrunn, W.; Pfisterer, M.; Detrano, R. Heart Disease Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/Heart+Disease (accessed on 10 July 2019).
Figure 1. The Auto-Sklearn pipeline [12] contains three main building blocks: (a) Data preprocessor, (b) Feature preprocessor, and (c) Estimator or machine learning algorithms.
Figure 1. The Auto-Sklearn pipeline [12] contains three main building blocks: (a) Data preprocessor, (b) Feature preprocessor, and (c) Estimator or machine learning algorithms.
Jcm 08 01050 g001
Figure 2. Python code for using Auto-Sklearn to train a classifier for any dataset.
Figure 2. Python code for using Auto-Sklearn to train a classifier for any dataset.
Jcm 08 01050 g002
Figure 3. Validation accuracy over 18 days by the graduate student on the Heart UCI dataset.
Figure 3. Validation accuracy over 18 days by the graduate student on the Heart UCI dataset.
Jcm 08 01050 g003
Figure 4. Validation accuracy over 15 days by the graduate student on the Cardiovascular Disease dataset.
Figure 4. Validation accuracy over 15 days by the graduate student on the Cardiovascular Disease dataset.
Jcm 08 01050 g004
Table 1. Thirteen attributes of the Heart UCI (University of California, Irvine, CA, USA) dataset.
Table 1. Thirteen attributes of the Heart UCI (University of California, Irvine, CA, USA) dataset.
AttributeTypeDescription
AgeContinuousAge in years
CpDiscreteChest pain type (4 values)
TrestbpsContinuousResting blood pressure (in mm Hg on admission to the hospital)
CholContinuousSerum cholestoral in mg/dL
FbsDiscreteFasting blood sugar > 120 mg/dL 1 = true; 0 = false
RestecgDiscreteResting electrocardiographic results (values 0,1,2)
ThalachContinuousMaximum heart rate achieved
ExangDiscreteExercise induced angina (1 = yes; 0 = no)
OldpeakContinuousST depression induced by exercise relative to rest
SlopeDiscreteThe slope of the peak Exercise ST segment (values 0,1,2)
CaDiscreteNumber of major vessels (0–4) colored by flourosopy
ThalDiscreteNature of defect, values (0–3)
TargetDiscretePresence or absence of heart disease, values (1,0)
Table 2. Twelve attributes of Cardiovascular Diseases dataset.
Table 2. Twelve attributes of Cardiovascular Diseases dataset.
AttributeTypeDescription
AgeContinuousAge of the patient in days
GenderDiscrete1: women, 2: men
Height (cm)ContinuousHeight of the patient in cm
Weight (kg)ContinuousWeight of the patient in kg
Ap_hiContinuousSystolic blood pressure
Ap_loContinuousDiastolic blood pressure
CholesterolDiscrete1: normal, 2: above normal, 3: well above normal
GlucDiscrete1: normal, 2: above normal, 3: well above normal
SmokeDiscretewhether patient smokes or not
AlcoDiscreteAlcohol intake-Binary feature
ActiveDiscretePhysical activity-Binary feature
CardioDiscretePresence or absence of cardiovascular disease
Table 3. Comparison of AutoML and the graduate student’s classification performances and total time on UCI test set.
Table 3. Comparison of AutoML and the graduate student’s classification performances and total time on UCI test set.
AccuracyAUC-ROCAUC-PRTotal Ttime (h)
Graduate student0.840.820.80432
AutoML0.850.930.940.5
Table 4. Comparison of AutoML and graduate student’s classification performances and total time on the Cardiovascular test set.
Table 4. Comparison of AutoML and graduate student’s classification performances and total time on the Cardiovascular test set.
AccuracyAUC-ROCAUC-PRTotal Time (h)
Graduate student0.740.730.68360
AutoML0.740.80.790.5
Table 5. Accuracies reported by previous studies on the Heart UCI Dataset compared to accuracies of the graduate student and AutoML.
Table 5. Accuracies reported by previous studies on the Heart UCI Dataset compared to accuracies of the graduate student and AutoML.
AuthorReported Accuracy
Shouman et al. [18]0.841
Duch et al. [35]0.856
Wang et al. [36]0.8337
Srinivas et al. [19]0.837
Tomar and Agarwal [20]0.8559
Graduate student (this paper)0.84
AutoML (this paper)0.85

Share and Cite

MDPI and ACS Style

Padmanabhan, M.; Yuan, P.; Chada, G.; Nguyen, H.V. Physician-Friendly Machine Learning: A Case Study with Cardiovascular Disease Risk Prediction. J. Clin. Med. 2019, 8, 1050. https://doi.org/10.3390/jcm8071050

AMA Style

Padmanabhan M, Yuan P, Chada G, Nguyen HV. Physician-Friendly Machine Learning: A Case Study with Cardiovascular Disease Risk Prediction. Journal of Clinical Medicine. 2019; 8(7):1050. https://doi.org/10.3390/jcm8071050

Chicago/Turabian Style

Padmanabhan, Meghana, Pengyu Yuan, Govind Chada, and Hien Van Nguyen. 2019. "Physician-Friendly Machine Learning: A Case Study with Cardiovascular Disease Risk Prediction" Journal of Clinical Medicine 8, no. 7: 1050. https://doi.org/10.3390/jcm8071050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop