Next Article in Journal
Pediatric and Adolescent Seizure Detection: A Machine Learning Approach Exploring the Influence of Age and Sex in Electroencephalogram Analysis
Next Article in Special Issue
Comparing ANOVA and PowerShap Feature Selection Methods via Shapley Additive Explanations of Models of Mental Workload Built with the Theta and Alpha EEG Band Ratios
Previous Article in Journal
Machine Learning Models and Technologies for Evidence-Based Telehealth and Smart Care: A Review
Previous Article in Special Issue
Whole Slide Image Understanding in Pathology: What Is the Salient Scale of Analysis?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effect of Data Missingness on Machine Learning Predictions of Uncontrolled Diabetes Using All of Us Data

1
Information and Computer Sciences Department, University of Hawaiʻi at Mānoa, 2500 Campus Road, Honolulu, HI 96822, USA
2
Hawaiʻi Digital Health Laboratory, 1680 East-West Road, Honolulu, HI 96822, USA
*
Author to whom correspondence should be addressed.
BioMedInformatics 2024, 4(1), 780-795; https://doi.org/10.3390/biomedinformatics4010043
Submission received: 21 January 2024 / Revised: 7 February 2024 / Accepted: 26 February 2024 / Published: 6 March 2024
(This article belongs to the Special Issue Feature Papers in Medical Statistics and Data Science Section)

Abstract

:
Electronic Health Records (EHR) provide a vast amount of patient data that are relevant to predicting clinical outcomes. The inherent presence of missing values poses challenges to building performant machine learning models. This paper aims to investigate the effect of various imputation methods on the National Institutes of Health’s All of Us dataset, a dataset containing a high degree of data missingness. We apply several imputation techniques such as mean substitution, constant filling, and multiple imputation on the same dataset for the task of diabetes prediction. We find that imputing values causes heteroskedastic performance for machine learning models with increased data missingness. That is, the more missing values a patient has for their tests, the higher variance there is on a diabetes model AUROC, F1, precision, recall, and accuracy scores. This highlights a critical challenge in using EHR data for predictive modeling. This work highlights the need for future research to develop methodologies to mitigate the effects of missing data and heteroskedasticity in EHR-based predictive models.

Graphical Abstract

1. Introduction

Diabetes is a health condition characterized by chronic hyperglycemia and resulting from issues with insulin secretion and action [1]. The onset of diabetes increases the risk for a number of health complications such as cardiovascular disease, kidney disease, retinopathy, and neuropathy [2,3]. The longer one has diabetes, the more complications are likely to occur [4]. Diabetes affects 464 million people in the world as of 2021, and it is predicted to increase to 638 million by 2045 [5]. Diabetes disproportionately affects minority populations [4,6].
Diabetes has also been studied using machine learning [7,8,9]. Oikonomou et al. [10] provide a comprehensive overview of how machine learning has been applied to precision diabetes care, particularly in cardiovascular risk prediction among diabetic patients. Their work underscores the significant potential of machine learning in transforming diabetes care by leveraging large datasets to identify risk factors and predict outcomes with high accuracy.
In recent years, the application of machine learning to electronic health records (EHR) has emerged as a promising tool for enhancing our understanding of diabetes and improving prediction models for its management. The integration of machine learning with EHR data offers a new frontier in diabetic research. Prior studies have shown that machine learning models can effectively predict the progression to pre-diabetes and type 2 diabetes using EHR data, emphasizing the role of established risk factors and identifying novel factors for further research [11]. Cahn et al. highlighted the use of machine learning models to improve the prediction of incident diabetes utilizing patient data from EHR, underscoring the potential for targeted interventions [12]. Additionally, leveraging large health records datasets has enabled significant progress in diabetes forecasting using machine learning, as demonstrated by research conducted using the health records of patients in Ontario, Canada [13]. This approach not only offers predictive insights, but also helps identify critical features contributing to diabetes onset.
Building upon this line of research, we study the prediction of diabetes using EHR data from the National Institutes of Health (NIH)’s All of Us (AoU) dataset. The program is a result of the Precision Medicine Initiative Cohort Program [14]. The cohort consists of over 1 million volunteers who contributed their biospecimen samples (such as blood and urine), physical measurements, and extensive surveys on health and lifestyle [15]. The overarching goal of All of Us is to advance precision medicine—a personalized approach to disease prevention and treatment that considers individual differences in lifestyle, environment, and biology. This approach is intended to overcome the limitations of a one-size-fits-all model in health care by factoring individual variation. The All of Us Research Program stands out for its commitment to diversity, striving to include participants from various racial and ethnic backgrounds, age groups, geographic regions, and health statuses to ensure the dataset reflects the broad diversity of the U.S. population [16]. By harnessing the power of big data and emphasizing inclusivity and participant engagement, the All of Us Research Program aspires to revolutionize our understanding of health and pave the way for more effective, personalized healthcare solutions.
We focus, in particular, on measuring the effect of data missingness on the prediction of health outcomes such as diabetes using All of Us data. We apply several data imputation techniques and measure their effect on various model performance metrics. This characterization of data missingness on large EHR datasets can inform future efforts that apply imputation strategies to such data.

2. Materials and Methods

2.1. Dataset

We used the National Institutes of Health (NIH) All of Us dataset. We selected 47 features from Abegaz et al.’s work [17]. We list them in Table 1 alongside the proportion of missing values per feature. For each measurement type, we created two features: one for the average reading and another for the number of times the feature is read.
The total size of the dataset is 162,453, with 56,655 positive and 105,798 negative data points. The stratified train/test split is 80/20, yielding 129,962 training patients of whom 45,324 are positive and 84,638 are negative, and 32,491 test patients consisting of 11,331 positive and 21,160 negative patients.

2.2. Modeling

To increase uniformity for ease of comparison while maintaining a robust search for well-performing models, we employed Autosklearn2.0 [18,19]. This meta-model has a search space consisting of every model within Scikit-Learn and subsequently searches over hyperparameter space per model. The training is conducted on four CPUs, 26 GB of RAM, 3 h of training time, 6572 MB of memory per job, log loss as the objective function, and no limit to the number of models on disk.
We compare the following six imputation methods alongside an oversampling preprocessing step.
No Imputation: This method involves not performing any imputation on the dataset, leaving the missing values as they are. In this approach, the model chosen must inherently be capable of handling missing data. Techniques such as decision trees or certain ensemble methods can often process datasets with missing values directly. This method is based on the assumption that the model can interpret and manage the missingness in the data without any explicit intervention.
Automatic Imputation (via Autosklearn): This approach employs Autosklearn, an automated machine learning tool, to determine the best imputation method for the dataset. Autosklearn explores various imputation strategies as part of its preprocessing pipeline and selects the one that optimizes model performance. This method leverages the power of automated machine learning to identify the most effective imputation technique, which could range from simple strategies like mean or median substitution to more complex ones, based on the characteristics of the data.
Constant Fill: In this approach, missing values are filled with a constant value. This constant could be a number outside the normal range of values (such as −1) to differentiate imputed values from real ones. The advantage of this method is its simplicity and the clear demarcation it provides, which can be helpful in certain analytical contexts.
Mean Substitution: Mean substitution involves replacing missing values in a dataset with the mean value of the respective column. This method assumes that the missing values are randomly distributed and that the mean is a representative statistic for the missing data. It is a straightforward approach but may not always be suitable, particularly in cases where the data distribution is skewed or the mean is not a good representation of the central tendency.
Median Substitution: Similar to mean substitution, median substitution replaces missing values with the median of the respective column. This method is particularly useful in datasets where the distribution is skewed or there are outliers, as the median is less affected by extreme values than the mean. It is a robust approach that can provide a better central tendency estimate in certain types of data distributions.
Multiple Imputation with Bayesian Ridge: This is a more sophisticated approach where multiple imputation is performed using Bayesian Ridge regression. In this method, missing values are estimated based on observed data, with the Bayesian Ridge regression model used to predict the missing values. Specifically, one begins by denoting one column of the training input f and the other columns X f . A Bayesian Ridge regression model is then fitted on ( X f , f ) . This is conducted for every feature and can be repeated so that in the next round, the previous rounds’ predictions can be used to make better predictions of the missing value. In this paper, we use 15 imputation rounds. The number of imputation rounds, 15, is chosen arbitrarily. The higher the number, the more accurate the imputation should be. For a dataset as large as All of Us, we chose to keep it lower. This technique considers the uncertainty in the imputation process by creating several imputed datasets and combining the results, leading to more accurate and reliable imputation compared to single imputation methods.
Each of these imputation methods has its strengths and weaknesses and is suitable for different types of datasets and missing data patterns. The choice of imputation method can significantly impact the performance of the subsequent analysis or machine learning models.
Random oversampling is a technique used to address class imbalance in a dataset, particularly in situations where the dataset has a disproportionate number of instances in different classes. This imbalance can lead to biased or inaccurate model performance, as the model may tend to favor the majority class.
In random oversampling, the idea is to balance the dataset by increasing the size of the underrepresented class (minority class). This is accomplished by randomly duplicating instances from the minority class until the number of instances in both the minority and majority classes is approximately equal. This method creates additional samples from the minority class not by generating new samples but by resampling from the existing samples.
In total, there are 12 different models to test with the same underlying classifier.

2.3. Model Evaluation

Model performance is a catch-all term to describe the plethora of different metrics used to compare a model’s predictions to the actual outcome. We can summarize the comparison of a classification model’s predictions as compared to the number of actual classes in a confusion matrix.
We use the following abbreviations in the definitions of our performance metrics:
T P = True Positive F N = False Negative F P = False Positive T N = True Negative P = Positive = T P + F N N = Negative = F P + T N
We have the corresponding normalized quantities associated with the above counts:
T P R = True Positive Rate = T P / P F N R = False Negative Rate = F N / P F P R = False Positive Rate = F P / N T N R = True Negative Rate = T N / N
We may now define four of the five metrics:
Balanced Accuracy = 1 2 T P T P + F N + T N T N + F P Precision = T P T P + F P Recall = T P T P + F N F 1 = 2 T P 2 T P + F P + F N
The final of the five metrics consists of the probability output of a model. Given an input, a model has a probability associated with the class and a threshold such that inputs with a probability larger than the threshold are predicted to be a member of the class. There are certain points on this curve that we know the values for.
If the threshold is set to 0, then the model predicts all inputs as positive. Thus, the true positive rate is 1 and the false positive rate is 1. If the threshold is set to 1, then the model predicts all inputs as negative. Thus, the true positive rate is 0 and the false positive rate is 0. This defines a curve in the space with coordinates ( F P R , T P R ) parameterized by the probability threshold with endpoints ( 0 , 0 ) and ( 1 , 1 ) . This curve is called the Receiver Operating Characteristic (ROC) curve, and its integral is called the Area Under the ROC (AUROC).

2.4. Model Fairness Evaluation

Given the standard metrics above, we can consider some fairness metrics that are measured as discrepancies of some performance metric between members of a privileged group and the remaining groups. In this dataset, there are two primary sensitive attributes that fall into this regime: gender and race. In order to define these differences, we must introduce new notation. The exact notation will differ based on the source [20,21,22,23]. The quantities below will be numerically equivalent to those in the previous literature while remaining consistent with the notation used in this paper. Let μ S denote the metric μ on the subset S within the data. For example, FPR P will denote the False Positive Rate on the privileged group, whereas FPR U will denote the False Positive Rate on the unprivileged group. We let y i represent the test result for patient i and y ^ i represent the model’s prediction for patient i. The final fairness metric shown below is described in detail by Speicher et al. [24].
Average Odds Difference = 1 2 ( FPR U FPR P ) + ( TPR P TPR U ) Average Odds Error = 1 2 | FPR U FPR P |   +   | TPR U TPR P | Class Imbalance = ( P U + N U ) ( P P + N P ) P + N Equal Opportunity Difference = TPR U TPR P Statistical Parity Difference = ( TPR U + FPR U ) ( TPR P + FPR P ) Between Group Generalized Entropy Error = 1 2 n i = 1 n y ^ i y i + 1 1 n i = 1 n y ^ i y i + 1 2 1

2.5. Measuring the Effect of Data Missingness

We are interested in measuring the effect on the model’s performance as the number of missing features varies. One expects that a higher number of missing features would lead to lower overall performance. Since the number of missing features is a large range, we can study the trend by fitting an ordinary least squares line between the performance versus the number of missing features. Our procedure is as follows:
  • Given a model fitted on the training data:
  • Select a subset of the testing data with a specified number of missing features.
  • Evaluate the model’s performance on that subset.
  • Plot the performance versus the number of missing features.
  • Evaluate the F-test for the slope of the line and the Breuch–Pagan test for the heteroskedasticity of the residuals around the line.

3. Results

3.1. Data Missingness

We constructed a simple (but interpretable) linear regression model that predicts the number of missing features given race and gender. The coefficients are shown in Table 2. We observe that race and gender are predictors of missingness.

3.2. Model Performance

Figure 1 and Figure 2 outline each imputation method’s overall performance on the dataset when stratified by different sensitive attributes. For each imputation method, we measured the AUROC, balanced accuracy, F1, precision, and recall on the total population, each gender category, each racial category, and across the different missing feature bucketed groups. We then reran the analysis with an extra step of oversampling to balance the dataset for the number of people with diabetes.
Figure 3 and Figure 4 compare the fairness metrics, average odds difference, average odds error, between-group generalized entropy error, class imbalance, equal opportunity difference, mean difference, and statistical parity difference. These are fairness metrics, which means that for a sensitive attribute, we denote one group to be privileged and one to be unprivileged. We evaluated the imputation methods on the model discrepancy across groups. Since there is no obvious privileged group for the missing feature sub-populations, we only compared gender (with male being the privileged group) and race (with white being the privileged group).

3.3. Effect of Data Missingness

We next seek to understand the effect of data missingness. In the previous section, the 0.2-quantile missing feature sub-populations had their AUROC, balanced accuracy, F1, precision, and recall tabulated. We may visualize how the models perform more easily by plotting the models performance as a grouped bar chart, both without (Figure 5) and with (Figure 6) oversampling.
We plotted the line of best fit for each machine learning metric as a function of the number of missing features and across imputation strategies, both without oversampling (Figure 7, Table 3) and with oversampling (Figure 8, Table 4). We observed a statistically significant negative slope in all of the performance metrics and models except for the following imputation methods using balanced accuracy: impute mean, impute naive, impute median, impute ridge. Furthermore, any model apart from “No Imputation” and “Auto Impute” demonstrated statistically significant heteroskedasticity.

4. Discussion

We observe that imputation methods homogenize the amount of information per patient. That is, without imputation, the models have a sharp performance loss, whereas imputation makes the slope less steep at the cost of increasing heteroskedasticity. We also note that every statistical test agrees between the oversampled and non-oversampled models. This trend underscores the sensitivity of predictive models to the method of handling missing data in electronic health records (EHR). The negative slope indicates that as the degree of imputation increases—implying more data are being estimated rather than observed—the accuracy, precision, and recall of the models tend to decrease. This phenomenon can be attributed to the fact that imputation, despite being a necessary process to address missing data, introduces a level of uncertainty or noise. This noise can distort the underlying patterns within the data, leading to less reliable predictions from the models.
We are not the first paper to study diabetes prediction using the All of Us dataset. A paper by Abegaz et al. studied the application of machine learning algorithms to predict diabetes in the All of Us dataset [17]. Their work presents the AUROC, recall, precision, and F1 scores stratified by gender of the random forest, XGBoost, logistic regression, and weighted ensemble models. Our work builds upon those foundations in three ways. First, we note that all of the models in Abegaz et al.’s work can be found in Scikit-Learn. Hence, we performed a deep search over all Scikit-learn models to find the best performing ones. Second, we presented our results for further substrata of the dataset. One of the most important features of AoU is the diversity of people within the dataset. We highlighted the five performance metrics on the total testing dataset on each gender, on each race, and on groups bucketed by the number of missing features. We also presented the models’ performance on a number of fairness measurements when the sub-populations have a clear privileged group. Third, our largest deviation from the previous work was to show how the performance of a model changes as one changes the number of missing features.
The model performance in Figure 1 and Figure 2 has been trained for only three hours (as opposed to the multiday- or multiweek-long training that some deep neural network solutions provide) and yields modest results. Our best performing model is the “Auto Impute” model. We may compare the performance of that model to Abegaz et al.’s work. “Auto Impute” has a higher AUROC, comparable precision, and worse recall and F1. We note, however, that these are not clinically ready. Further improvements need to be made in order to prefer this to a HbA1c test for diabetes testing. Since the multiple imputer only used 15 iterations, the algorithm likely did not stabilize and caused the performance to drop. We emphasize that the primary objective of our research was not to maximize the performance of machine learning models applied to AoU data, but instead to study the effects of data missingness and imputation strategies on model performance.
Our analysis also highlights the presence of statistically significant heteroskedastic variance in model performances across imputation methods. Heteroskedasticity, in this context, refers to the irregular variability in the performance of predictive models, dependent on the amount and pattern of missing data being imputed. This irregular variance poses a significant challenge in predictive modeling, as it implies that the error terms (or the differences between predicted and actual values) are not uniformly distributed. Models thus exhibit different levels of accuracy and reliability depending on the specific characteristics of the missing data in each patient record.
The presence of heteroskedastic variance can be particularly problematic in clinical settings. It implies that for some patients, especially those with more extensive or particular patterns of missing data, the predictions made by the models could be less reliable. This inconsistency could lead to disparities in clinical decision-making, potentially affecting the quality of care provided to certain patient groups. Since the “Auto Imputation” model has the largest Y-intercept and one of the most negative slopes, it might be most beneficial to use the “Auto Impute” method for patients with few missing values in a clinical setting. For patients with a lot of missing values, one may use another imputation method with a less steep slope or perform a cost–benefit analysis of ordering more tests to make the model more performant.
These findings highlight the critical need for developing more robust imputation techniques that can minimize the introduction of noise and ensure uniform model performance across varying degrees of missing data. It also underscores the importance of considering the nature and pattern of missing data when applying machine learning models in healthcare settings. Future research should focus on exploring advanced imputation methods, possibly incorporating domain knowledge or utilizing more sophisticated algorithms, to mitigate the effects of data missingness on predictive model performance. In conclusion, while imputation is a necessary step in dealing with incomplete datasets for some models, our study indicates that current methods have significant limitations.
Addressing these limitations is crucial for the development of reliable and consistent machine learning models for clinical predictions, ultimately enhancing the quality of patient care and health outcomes. Our analysis on data missingness revealed that individuals who are male and persons of color would be disproportionately affected by a loss in performance with respect to data missingness. This is due to the number of missing features being more highly correlated with males and non-white people.
Future work can be conducted to ensure the robustness of the findings. A number of unanswered questions remain, such as: (1) does heteroskedasticity depend on certain features included in the model over another? (2) Do these findings pertain to more modern and complex deep learning models? (3) What other forms of data augmentation can be performed to reduce heteroskedasticity? Another comparison of interest is exploring whether the testing dataset holds more missing values than the training dataset and how the performance differs compared to the case of having roughly similar missing values between training and testing. If the testing dataset does not require many labels, then hospitals could save time and money by not measuring every missing value.

Author Contributions

Conceptualization, Z.J. and P.W.; methodology, Z.J.; software, Z.J.; validation, Z.J. and P.W.; formal analysis, Z.J.; investigation, P.W.; resources, Z.J.; data curation, Z.J.; writing—original draft preparation, Z.J.; writing—review and editing, Z.J. and P.W.; visualization, Z.J.; supervision, P.W.; project administration, P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from the National Institutes of Health’s All of Us and are available at https://www.researchallofus.org/ (accessed on 12 January 2024) with the permission of National Institutes of Health’s All of Us. Obtaining access to the data involves institutional agreement, verification of identity, and mandatory training.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. World Health Organization. ICD-11: International Classification of Diseases 11th Revision: The Global Standard for Diagnostic Health Information; World Health Organization: Geneva, Switzerland, 2019. [Google Scholar]
  2. Cole, J.B.; Florez, J.C. Genetics of diabetes mellitus and diabetes complications. Nat. Rev. Nephrol. 2020, 16, 377–390. [Google Scholar] [CrossRef] [PubMed]
  3. Association, A.D. Diagnosis and classification of diabetes mellitus. Diabetes Care 2010, 33, S62–S69. [Google Scholar] [CrossRef] [PubMed]
  4. Group, T.S. Long-term complications in youth-onset type 2 diabetes. N. Engl. J. Med. 2021, 385, 416–426. [Google Scholar] [CrossRef] [PubMed]
  5. Rooney, M.R.; Fang, M.; Ogurtsova, K.; Ozkan, B.; Echouffo-Tcheugui, J.B.; Boyko, E.J.; Magliano, D.J.; Selvin, E. Global prevalence of prediabetes. Diabetes Care 2023, 46, 1388–1394. [Google Scholar] [CrossRef] [PubMed]
  6. Haw, J.S.; Shah, M.; Turbow, S.; Egeolu, M.; Umpierrez, G. Diabetes complications in racial and ethnic minority populations in the USA. Curr. Diabetes Rep. 2021, 21, 1–8. [Google Scholar] [CrossRef] [PubMed]
  7. Khanam, J.J.; Foo, S.Y. A comparison of machine learning algorithms for diabetes prediction. ICT Express 2021, 7, 432–439. [Google Scholar] [CrossRef]
  8. Hasan, M.K.; Alam, M.A.; Das, D.; Hossain, E.; Hasan, M. Diabetes prediction using ensembling of different machine learning classifiers. IEEE Access 2020, 8, 76516–76531. [Google Scholar] [CrossRef]
  9. Krishnamoorthi, R.; Joshi, S.; Almarzouki, H.Z.; Shukla, P.K.; Rizwan, A.; Kalpana, C.; Tiwari, B. A novel diabetes healthcare disease prediction framework using machine learning techniques. J. Healthc. Eng. 2022, 2022, 1684017. [Google Scholar] [CrossRef] [PubMed]
  10. Oikonomou, E.K.; Khera, R. Machine learning in precision diabetes care and cardiovascular risk prediction. Cardiovasc. Diabetol. 2023, 22, 259. [Google Scholar] [CrossRef] [PubMed]
  11. Anderson, J.P.; Parikh, J.R.; Shenfeld, D.K.; Ivanov, V.; Marks, C.; Church, B.; Laramie, J.; Mardekian, J.; Piper, B.; Willke, R.; et al. Reverse Engineering and Evaluation of Prediction Models for Progression to Type 2 Diabetes. J. Diabetes Sci. Technol. 2016, 10, 6–18. [Google Scholar] [CrossRef] [PubMed]
  12. Cahn, A.; Shoshan, A.; Sagiv, T.; Yesharim, R.; Goshen, R.; Shalev, V.; Raz, I. Prediction of progression from pre-diabetes to diabetes: Development and validation of a machine learning model. Diabetes/Metabolism Res. Rev. 2020, 36, e3252. [Google Scholar] [CrossRef] [PubMed]
  13. Ravaut, M.; Sadeghi, H.; Leung, K.K.; Volkovs, M.; Rosella, L. Diabetes Mellitus Forecasting Using Population Health Data in Ontario, Canada. arXiv 2019, arXiv:abs/1904.04137. [Google Scholar]
  14. Hudson, K.; Lifton, R.; Patrick-Lake, B.; Burchard, E.G.; Coles, T.; Collins, R.; Conrad, A. The Precision Medicine Initiative Cohort Program—Building a Research Foundation for 21st Century Medicine; Precision Medicine Initiative (PMI) Working Group Report to the Advisory Committee to the Director, National Institutes of Health: Bethesda, MD, USA, 2015. [Google Scholar]
  15. Sankar, P.L.; Parker, L.S. The Precision Medicine Initiative’s All of Us Research Program: An agenda for research on its ethical, legal, and social issues. Genet. Med. 2017, 19, 743–750. [Google Scholar] [CrossRef] [PubMed]
  16. Mapes, B.M.; Foster, C.S.; Kusnoor, S.V.; Epelbaum, M.I.; AuYoung, M.; Jenkins, G.; Lopez-Class, M.; Richardson-Heron, D.; Elmi, A.; Surkan, K.; et al. Diversity and inclusion for the All of Us research program: A scoping review. PLoS ONE 2020, 15, e0234962. [Google Scholar] [CrossRef] [PubMed]
  17. Abegaz, T.M.; Ahmed, M.; Sherbeny, F.; Diaby, V.; Chi, H.; Ali, A.A. Application of Machine Learning Algorithms to Predict Uncontrolled Diabetes Using the All of Us Research Program Data. Healthcare 2023, 11, 1138. [Google Scholar] [CrossRef] [PubMed]
  18. Feurer, M.; Eggensperger, K.; Falkner, S.; Lindauer, M.; Hutter, F. Auto-Sklearn 2.0: Hands-free AutoML via Meta-Learning. arXiv 2020, arXiv:2007.04074 [cs.LG]. [Google Scholar]
  19. Feurer, M.; Klein, A.; Eggensperger, K.; Springenberg, J.; Blum, M.; Hutter, F. Efficient and Robust Automated Machine Learning. Adv. Neural Inf. Process. Syst. 2015, 28, 2962–2970. [Google Scholar]
  20. Bellamy, R.K.E.; Dey, K.; Hind, M.; Hoffman, S.C.; Houde, S.; Kannan, K.; Lohia, P.; Martino, J.; Mehta, S.; Mojsilovic, A.; et al. AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias. arXiv 2018, arXiv:1810.01943. [Google Scholar]
  21. Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A. A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
  22. Caton, S.; Haas, C. Fairness in machine learning: A survey. ACM Comput. Surv. 2020. [Google Scholar] [CrossRef]
  23. Barocas, S.; Hardt, M.; Narayanan, A. Fairness and Machine Learning: Limitations and Opportunities; MIT Press: Cambridge, MA, USA, 2023. [Google Scholar]
  24. Speicher, T.; Heidari, H.; Grgic-Hlaca, N.; Gummadi, K.P.; Singla, A.; Weller, A.; Zafar, M.B. A unified approach to quantifying algorithmic unfairness: Measuring individual &group unfairness via inequality indices. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 2239–2248. [Google Scholar]
Figure 1. Performance of the models, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label). The color denotes the magnitude of the metric, warmer colors indicating higher performance. The text color is adjusted to be readable given the background color.
Figure 1. Performance of the models, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label). The color denotes the magnitude of the metric, warmer colors indicating higher performance. The text color is adjusted to be readable given the background color.
Biomedinformatics 04 00043 g001
Figure 2. Performance of the models when oversampling, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label). The color denotes the magnitude of the metric, warmer colors indicating higher performance.
Figure 2. Performance of the models when oversampling, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label). The color denotes the magnitude of the metric, warmer colors indicating higher performance.
Biomedinformatics 04 00043 g002
Figure 3. Performance of the models, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label).
Figure 3. Performance of the models, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label).
Biomedinformatics 04 00043 g003
Figure 4. Performance of the models when oversampling, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label). The color denotes the magnitude of the metric, with warmer colors indicating better performance. The text color is adjusted to be readable given the background color.
Figure 4. Performance of the models when oversampling, with the columns denoting the specific metric, across the evaluated sub-population (left label) and the imputation method (right label). The color denotes the magnitude of the metric, with warmer colors indicating better performance. The text color is adjusted to be readable given the background color.
Biomedinformatics 04 00043 g004
Figure 5. Machine learning performance exhibited by different imputation methods grouped by 0.2 quantiles.
Figure 5. Machine learning performance exhibited by different imputation methods grouped by 0.2 quantiles.
Biomedinformatics 04 00043 g005
Figure 6. Machine learning performance exhibited by different imputation methods using an oversampling preprocessing step grouped by 0.2 quantiles.
Figure 6. Machine learning performance exhibited by different imputation methods using an oversampling preprocessing step grouped by 0.2 quantiles.
Biomedinformatics 04 00043 g006
Figure 7. Best fit lines of machine learning metrics as a function of the number of missing features. The shading is the residual of the best fit line. The best fit line is colored green if we reject the null hypothesis that the line has a slope of zero. The shading is colored green if we reject the null hypothesis that the residuals have constant variance.
Figure 7. Best fit lines of machine learning metrics as a function of the number of missing features. The shading is the residual of the best fit line. The best fit line is colored green if we reject the null hypothesis that the line has a slope of zero. The shading is colored green if we reject the null hypothesis that the residuals have constant variance.
Biomedinformatics 04 00043 g007
Figure 8. Best fit lines of machine learning metrics as a function of the number of missing features. The shading is the residual of the best fit line. All models contain an oversampling step. The best fit line is colored green if we reject the null hypothesis that the line has a slope of zero. The shading is colored green if we reject the null hypothesis that the residuals have constant variance.
Figure 8. Best fit lines of machine learning metrics as a function of the number of missing features. The shading is the residual of the best fit line. All models contain an oversampling step. The best fit line is colored green if we reject the null hypothesis that the line has a slope of zero. The shading is colored green if we reject the null hypothesis that the residuals have constant variance.
Biomedinformatics 04 00043 g008
Table 1. Model input features and missingness proportion for the total dataset for training and testing subsets.
Table 1. Model input features and missingness proportion for the total dataset for training and testing subsets.
TotalTrainingTesting
Age0.0000000.0000000.000000
Median income0.0000000.0000000.000000
Deprivation index0.0000000.0000000.000000
Chloride0.0914480.0912570.092210
Bicarbonate0.6530570.6533680.651811
Alanine aminotransferase0.1442020.1437500.146010
Albumin0.1384770.1378250.141085
Alkaline phosphatase0.1406810.1402180.142532
Anion gap0.2228890.2226800.223723
Aspartate aminotransferase0.1451620.1449270.146102
Basophils0.1555160.1549910.157613
Bilirubin0.1596030.1592770.160906
Height0.0068700.0069400.006586
Weight0.0087900.0088260.008649
Calcium0.0942300.0938510.095750
Carbon dioxide0.1779280.1774900.179681
HDL0.3310190.3304970.333108
LDL0.3515790.3509870.353944
Creatinine0.0835380.0833010.084485
Eosinophil0.1510780.1506060.152965
Erythrocytes0.1046150.1042770.105968
Heart rate0.0089750.0089410.009110
Leukocyte0.0895640.0892030.091010
Lymphocytes0.1440790.1437650.145333
MCH0.1394000.1390790.140685
MCHC0.1400340.1396950.141393
MCV0.1691260.1688420.170263
Monocytes0.1468790.1464200.148718
Neutrophils0.1433710.1428960.145271
Platelets0.1157070.1152260.117633
Potassium0.1030880.1029840.103506
Respiratory rate0.3100530.3089130.314610
Sodium0.0929130.0926810.093841
Triglyceride0.3394210.3390680.340833
Urea nitrogen0.1128390.1127330.113262
Vomiting0.0000000.0000000.000000
Myocardial infarction0.0000000.0000000.000000
Arthritis0.0000000.0000000.000000
Polyuria0.0000000.0000000.000000
Aspirin0.0989700.0989520.099043
Beta blockers0.0989700.0989520.099043
Steroids0.0989700.0989520.099043
Acetaminophen0.0989700.0989520.099043
Statin0.0989700.0989520.099043
Opioids0.0989700.0989520.099043
Nicotine0.0989700.0989520.099043
Paraesthesia0.0989700.0989520.099043
Table 2. Linear regression coefficients.
Table 2. Linear regression coefficients.
Sensitive AttributeCoefficient
Female−0.54
Male0.39
Gender Other0.14
Black1.31
White−1.34
Middle Eastern0.48
Asian−0.09
Race Other−0.35
Table 3. Tabular representation of Figure 7. We display the Y-intercept and slope of the lines of best fit for the estimator performance on a given metric. The F-Test p-value gives the probability of the null hypothesis that the line of best fit has a slope of zero. The Breusch–Pagan p-value, which gives the probability that the error of the line has constant variance, is also given.
Table 3. Tabular representation of Figure 7. We display the Y-intercept and slope of the lines of best fit for the estimator performance on a given metric. The F-Test p-value gives the probability of the null hypothesis that the line of best fit has a slope of zero. The Breusch–Pagan p-value, which gives the probability that the error of the line has constant variance, is also given.
EstimatorMetricY-InterceptSlopeF-TestBreusch–Pagan
p-Valuep-Value
No ImputationBalanced Accuracy0.938288−0.0143310.0000000.372675
Precision1.038484−0.0159070.0000000.683840
Recall0.688205−0.0107700.0000000.380535
F10.805173−0.0124960.0000000.934875
AUROC0.938288−0.0143310.0000000.372675
Auto ImputeBalanced Accuracy0.962040−0.0147390.0000000.352425
Precision1.023924−0.0157650.0000000.681243
Recall0.742006−0.0116670.0000000.657899
F10.844082−0.0131650.0000000.760720
AUROC0.962040−0.0147390.0000000.352425
Impute MeanBalanced Accuracy0.744528−0.0004830.5171610.000141
Precision0.887014−0.0065340.0000780.000000
Recall0.680919−0.0057090.0000680.001215
F10.759957−0.0062770.0000070.000021
AUROC0.825048−0.0044090.0000890.000099
Impute NaiveBalanced Accuracy0.736591−0.0005480.4318440.000462
Precision0.867793−0.0061540.0000950.000001
Recall0.649528−0.0051190.0001860.005948
F10.733024−0.0057880.0000160.000159
AUROC0.812685−0.0042570.0001060.000334
Impute MedianBalanced Accuracy0.743865−0.0006200.3840450.000024
Precision0.890143−0.0067350.0000250.000000
Recall0.665581−0.0056310.0000310.005150
F10.752403−0.0063180.0000020.000065
AUROC0.824385−0.0045460.0000300.000131
Impute RidgeBalanced Accuracy0.6952960.0002470.7248790.000051
Precision0.875240−0.0065660.0000360.000002
Recall0.565600−0.0040390.0027790.007240
F10.671835−0.0050790.0001640.000335
AUROC0.775816−0.0036790.0006050.000196
Table 4. Tabular representation of Figure 8 displaying the Y-intercept and slope of the lines of best fit for the estimator performance with oversampling on a given metric. The F-Test p-value, which gives the probability of the null hypothesis that the line of best fit has a slope of zero, is given. The Breusch–Pagan p-value, which gives the probability that the error of the line has constant variance, is also given.
Table 4. Tabular representation of Figure 8 displaying the Y-intercept and slope of the lines of best fit for the estimator performance with oversampling on a given metric. The F-Test p-value, which gives the probability of the null hypothesis that the line of best fit has a slope of zero, is given. The Breusch–Pagan p-value, which gives the probability that the error of the line has constant variance, is also given.
EstimatorMetricY-InterceptSlopeF-TestBreusch–Pagan
p-Valuep-Value
No ImputationBalanced Accuracy0.929096−0.0143280.0000000.485545
Precision1.016151−0.0157120.0000000.913311
Recall0.681390−0.0109740.0000000.125074
F10.793971−0.0125730.0000000.761271
AUROC0.929096−0.0143280.0000000.485545
Auto ImputeBalanced Accuracy0.927235−0.0141830.0000000.510401
Precision0.985620−0.0152220.0000000.935550
Recall0.686184−0.0107580.0000000.247551
F10.793458−0.0123720.0000000.783637
AUROC0.927235−0.0141830.0000000.510401
Impute MeanBalanced Accuracy0.726784−0.0008870.2570000.003557
Precision0.850450−0.0052280.0007940.000000
Recall0.583068−0.0038410.0037610.017359
F10.676989−0.0045790.0002970.000400
AUROC0.775188−0.0034110.0014670.000415
Impute NaiveBalanced Accuracy0.7062280.0002010.7486660.000180
Precision0.856446−0.0057070.0003320.000000
Recall0.588463−0.0036210.0073000.003331
F10.683514−0.0045330.0006380.000028
AUROC0.780858−0.0034070.0016580.000215
Impute MedianBalanced Accuracy0.7026670.0000420.9434920.000159
Precision0.846575−0.0057290.0002160.000000
Recall0.597699−0.0041810.0011870.017877
F10.691866−0.0052320.0000110.000528
AUROC0.778030−0.0036160.0005130.000478
Impute RidgeBalanced Accuracy0.716799−0.0010770.1726840.003920
Precision0.854813−0.0064360.0000260.000025
Recall0.562891−0.0040000.0027770.017178
F10.666654−0.0051180.0000370.000330
AUROC0.763755−0.0035370.0008800.000269
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jabbar, Z.; Washington, P. The Effect of Data Missingness on Machine Learning Predictions of Uncontrolled Diabetes Using All of Us Data. BioMedInformatics 2024, 4, 780-795. https://doi.org/10.3390/biomedinformatics4010043

AMA Style

Jabbar Z, Washington P. The Effect of Data Missingness on Machine Learning Predictions of Uncontrolled Diabetes Using All of Us Data. BioMedInformatics. 2024; 4(1):780-795. https://doi.org/10.3390/biomedinformatics4010043

Chicago/Turabian Style

Jabbar, Zain, and Peter Washington. 2024. "The Effect of Data Missingness on Machine Learning Predictions of Uncontrolled Diabetes Using All of Us Data" BioMedInformatics 4, no. 1: 780-795. https://doi.org/10.3390/biomedinformatics4010043

Article Metrics

Back to TopTop