Next Article in Journal
Silver Diamine Fluoride (SDF) Efficacy in Arresting Cavitated Caries Lesions in Primary Molars: A Systematic Review and Metanalysis
Previous Article in Journal
Addressing Psychosocial Factors in Cognitive Impairment Screening from a Holistic Perspective: The DeCo-Booklet Methodology Design and Pilot Study
Previous Article in Special Issue
A Novel Algorithm to Estimate the Significance Level of a Feature Interaction Using the Extreme Gradient Boosting Machine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of Integrating Machine Learning in Comparative Effectiveness Research of Oral Anticoagulants in Patients with Atrial Fibrillation

1
College of Pharmacy, Kyung Hee University, Seoul 02447, Korea
2
Health Outcomes Division, College of Pharmacy, The University of Texas at Austin, Austin, TX 78712, USA
3
Department of Regulatory Science, Graduate School, Kyung Hee University, Seoul 02447, Korea
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(19), 12916; https://doi.org/10.3390/ijerph191912916
Submission received: 31 August 2022 / Revised: 2 October 2022 / Accepted: 6 October 2022 / Published: 9 October 2022
(This article belongs to the Special Issue Machine Learning Analytics for Cardiovascular Diseases)

Abstract

:
We aimed to compare the ability to balance baseline covariates and explore the impact of residual confounding between conventional and machine learning approaches to derive propensity scores (PS). The Health Insurance Review and Assessment Service database (January 2012–September 2019) was used. Patients with atrial fibrillation (AF) who initiated oral anticoagulants during July 2015–September 2018 were included. The outcome of interest was stroke/systemic embolism. To estimate PS, we used a logistic regression model (i.e., a conventional approach) and a generalized boosted model (GBM) which is a machine learning approach. Both PS matching and inverse probability of treatment weighting were performed. To evaluate balance achievement, standardized differences, p-values, and boxplots were used. To explore residual confounding, E-values and negative control outcomes were used. In total, 129,434 patients were identified. Although all baseline covariates were well balanced, the distribution of continuous variables seemed more similar when GBM was applied. E-values ranged between 1.75 and 2.70 and were generally higher in GBM. In the negative control outcome analysis, slightly more nonsignificant hazard ratios were observed in GBM. We showed GBM provided a better ability to balance covariates and had a lower impact of residual confounding, compared with the conventional approach in the empirical example of comparative effectiveness analysis.

1. Introduction

The era of big data and machine learning seems to be entering a stage in which it is possible to produce additional high-quality evidence and overcome the common limitations of randomized controlled trials (RCTs), which include short follow-up periods and restricted patients who do not represent real-world patients [1,2]. However, despite its importance, healthcare big data are criticized for their incomplete information on potential unmeasured confounders, which may result in biased estimates of the association between drug exposure and outcomes in comparative effectiveness research [1].
Propensity score (PS) methods are most commonly used to minimize confounding in exposure–outcome associations [3]. However, compared with RCTs, the true PS model remains unknown. Thus, there is a possibility of residual unmeasured confounding or misspecification of the PS model [4].
Recently, machine-learning-based PS methods were introduced to reduce additional confounding [4]. Machine learning methods, such as boosting, have iterative and automated variable selection processes that show good performance in terms of bias reduction or mean squared error when compared with simple logistic regression models [3].
However, studies on machine learning and residual unmeasured confounding are limited. An empirical example of a comparative effectiveness study may provide insights into applying machine learning, and could become an important basis for understanding ways to minimize residual confounding. This could potentially help guide physicians in better healthcare decision making.
This study had two objectives: to compare the ability to balance baseline covariates and to explore and compare the impact of residual unmeasured confounding between the conventional PS and machine-learning-based PS. We used an empirical example, a comparative effectiveness study into patients with atrial fibrillation (AF), a commonly studied topic globally [5,6,7,8,9,10,11], to provide practical insight into the comparative effectiveness research field.

2. Materials and Methods

2.1. Study Scheme and Data Source

This study utilized an empirical example of a retrospective cohort study to compare the effectiveness of oral anticoagulants (OACs) in patients with AF in Korea. The study scheme is illustrated in Figure 1.
This study used the Health Insurance Review and Assessment Service (HIRA) database (1 January 2012–30 September 2019), which covers the entire Korean population via the country’s universal health insurance system [12]. This database includes the diagnosis, procedure, and medication codes as variables for each medical or pharmacy claim [12]. The diagnoses were coded according to the Korean Standard Classification of Diseases, based on the International Classification of Diseases 10th Revision (ICD-10) codes [12].

2.2. Study Population and Clinical Outcomes

The study population of interest comprised OAC-naïve patients with AF. Patients who met all the following inclusion criteria were included in this study: initiation of OACs (apixaban, dabigatran, edoxaban, rivaroxaban, or warfarin) between 1 July 2015 and 30 September 2018; ≥1 AF diagnosis code (ICD-10: I48) prior to or on the index date; and age of ≥18 years on the index date.
The following patients were excluded: those in whom OACs (i.e., apixaban, dabigatran, edoxaban, rivaroxaban, or warfarin) were initiated prior to the index date; >1 OAC received at the index date; diagnosis or procedure codes for rheumatic mitral valvular heart disease; diagnosis codes for mitral valve stenosis, venous thromboembolism, end-stage chronic kidney disease, kidney transplant, dialysis, pericarditis, thyrotoxicosis, or hypertrophic cardiomyopathy within one year prior to the index date; both standard and reduced doses for the same index OAC were prescribed at the index date.
In this population, the clinical outcome was stroke or systemic embolism (S/SE). Stroke included ischemic and hemorrhagic stroke. A detailed S/SE definition that was defined based on clinical expert opinion during the follow-up period, is described in Supplementary Table S1. S/SE was defined using a main diagnosis code based on a previous study [13].

2.3. Study Outcomes

The study outcomes were: (1) the balance of baseline covariates between groups and (2) the impact of residual confounding. We evaluated the balance of baseline covariates between groups using absolute standardized differences, side-by-side boxplots, and a statistical significance test (p-value). We evaluated the impact of residual confounding using E-values and negative control outcomes. The E-values for the S/SE were calculated. Pneumonia and urinary tract infections were considered as negative control outcomes. Detailed methods regarding these study outcomes are described in the following sections.

2.4. Statistical Analyses—PS Estimation and Balance

We considered the following comparisons: (1) standard dose (10 mg per day) of apixaban vs. warfarin; (2) reduced dose (5 mg per day) of apixaban vs. warfarin; (3) standard dose (300 mg per day) of dabigatran vs. warfarin; (4) reduced dose (220 mg per day) of dabigatran vs. warfarin; (5) standard dose (60 mg per day) of edoxaban vs. warfarin; (6) reduced dose (30 mg per day) of edoxaban vs. warfarin; (7) standard dose (20 mg per day) of rivaroxaban vs. warfarin; and (8) reduced dose (15 mg per day) of rivaroxaban vs. warfarin. Considering the comparability of the results with other studies, the standard- and reduced-dose groups were defined using the recommended dose per day for each non-vitamin K antagonist oral anticoagulants (NOACs) [14]. Reduced doses are recommended if patients have clinical factors (e.g., older age, lower body weight, renal impairment, or concomitant use of a P-glycoprotein inhibitor) that could affect bleeding risk [14].
We used logistic regression to estimate the conventional PS, and the generalized boosted model (GBM) to estimate the machine-learning-based PS for each comparison. GBM iteratively generates many simple regression tree models, combining them together to estimate the PS. Similar to logistic regression, GBM models the log odds of the probability of being in the treatment group, g(x) = log(p(x)/(1 − p(x))); however, the g(x) is iteratively updated to find the model with the maximized log-likelihood of g(x) [15,16]. More details on the PS estimation process were published previously [15,16]. For comparison, we used GBM among several machine learning approaches (e.g., LASSO or super learner) because it is the most frequently used in comparative effectiveness studies of patients with AF [6,7,17]. In addition, PS adjustment with GBM reportedly provides better results than other machine-learning approaches [18]. Thus, we believed that using a method that is known to have better performance and is used more frequently could provide more informative results. In addition, it is not necessary to predict the treatment accurately in PS models [19]. It is important to include all confounders in the model. Including variables that accurately predict the treatment but have no bearing on the outcome will cause increased variance in the estimated treatment effect without reducing bias or confounding [19,20]. Therefore, the predictive performance of the PS models cannot be translated into the extent of bias in the treatment effect or the level of balance between the two treatment groups. Thus, we did not evaluate the predictive performance (e.g., confusion matrix, accuracy, precision, or receiver operating characteristics curve).
To estimate PS, the variables in Supplementary Table S2 were defined using data from one year of the pre-index period. The same variables (covariates) were used for both the logistic regression and the GBM. All variables used for PS estimation were based on RCTs, observational studies, and clinical expert opinion [21,22,23,24,25]. For the logistic regression, a linear relationship was assumed, and no interaction terms were included. For the GBM, the number of trees was 20,000, the interaction depth was 2, the shrinkage parameter was set to 0.01, and the stopping method that selected the optimal number of iterations minimized the standardized differences between the groups [26]. We used these settings because they were introduced as starting points in previously published studies [15,26]. In these studies, it was recommended that the number of tree settings should be large enough to achieve balance, and if balance is not achieved, researchers should consider a more complex model. In this study, the balance was achieved by using these settings and the best balance was found before 20,000 iterations, ranging between 9895 and 19,754 for the eight comparisons. Model building was automated through an iterative process of adding terms to achieve the best balance between the groups [27].
After estimating the PS, we performed PS matching (PSM) and inverse probability of treatment weighting (IPTW). For PSM, a 0.2 standard deviation of the logit of the PS was used as a caliper. The IPTW weight for the average treatment effect for the treated (ATT) was used in this study. Although IPTW can estimate both the average treatment effect (ATE) and ATT, ATT was more appropriate for this study, as some patients receiving warfarin did not meet the standard or reduced dose criteria for patients receiving NOACs [17]. After matching or weighting, balance was evaluated using absolute standardized differences for all covariates [28] and side-by-side boxplots for continuous variables. An absolute standardized difference of <0.1 was considered an acceptable balance. Additionally, a statistical significance test (p-value) was applied. If variables did not reach an acceptable level of balance (i.e., standardized difference < 0.1 or p-value < 0.05), they were included in the outcome model to be adjusted.

2.5. Statistical Analyses—E-Value and Negative Control Outcome

E-values and negative control outcomes were used to detect residual unmeasured confounding [1]. E-values for the hazard ratio (HR) scale for a rare outcome (where outcome < 15%) were calculated for each PS method. In a rare outcome setting, the risk ratio approximates HR [29]. E-values and confidence intervals (CIs) were calculated using a web-based tool [30,31,32].
To compare the E-values of the conventional and machine learning methods (hereinafter referred to as E-values from the main analysis), we assessed whether the difference between E-values from the main analysis was larger than the “meaningful difference Δ” for each comparison. Meaningful difference ∆ was defined as the difference between the E-value from the main analysis and the E-value defined as an anchor (reference point). E-values as anchors were derived from the analyses by excluding age and sex from the PS model. The exclusion of age and sex was needed to create a scenario in which a strong unmeasured confounder existed. Age and sex are classical confounders and factors used to predict stroke risk in patients with AF [14,33]. Age- or sex-related variables in the CHA2DS2-VASc and HAS-BLED scores were also excluded from the PS model. To define one anchor for each comparison, a method that maximized the balance of covariates was chosen. The meaningful difference ∆ was expressed as an absolute value, as this analysis aimed to assess the magnitude of the impact of a scenario wherein an unmeasured confounder is as powerful as a known measured confounder, such as age or sex.
If the difference in E-values from the main analysis was larger than the meaningful difference ∆, it was considered meaningfully different. If the E-values from the main analysis were not meaningfully different, the percentage of maximum possible coverage (i.e., maximum possible difference in E-values from the main analysis divided by the meaningful difference ∆) was calculated to explore the extent to which the method with the highest E-value for each comparison covered a meaningful difference ∆. The summarized process of this approach is presented in Table 1.
In this study, the negative control outcomes were pneumonia and urinary tract infections (the definitions are provided in Supplementary Table S3). These outcomes were previously considered in a comparative effectiveness study of NOACs in patients with AF [34]. HRs with 95% CIs were calculated using a Cox proportional hazard model. The proportional hazard assumption was assessed using log–log plots. To ensure comparability between different methods, if the proportional hazard assumption was not met, HRs were not reported.
Statistical significance was set at p < 0.05. PS estimation using GBM was performed using R Studio version 1.1.463 (R Foundation for Statistical Computing, Vienna, Austria). To perform GBM, we used the twang (twang: Toolkit for Weighting and Analysis of Nonequivalent Groups) package in R [26]. The R codes are listed in Supplementary Table S4. All other analyses were performed using the SAS Enterprise Guide 6.1 M1 (SAS Institute Inc., Cary, NC, USA).

3. Results

3.1. Baseline Characteristics and Balance

In total, 129,434 patients with AF who were initiated on OACs were identified (Figure 2). The mean and median follow-up periods for each cohort are presented in Supplementary Table S5. Among all treatment groups, the follow-up period was the longest for the standard and reduced doses of dabigatran.
The baseline characteristics of each index treatment group before and after the application of the PS method are summarized in Supplementary Tables S6–S13. After applying PSM and IPTW, all the baseline characteristics were well balanced. However, although PSM and IPTW were applied, some variables were statistically significant (p < 0.05; Supplementary Tables S14–S21). Among these methods, IPTW (GBM) was the only method without statistically significant variables. As all PS methods achieved acceptable balance levels (standardized differences < 0.10), figures with their x-axis (i.e., standardized difference) from 0 to 0.1, are also presented for a closer comparison (Supplementary Figures S1–S8). The conventional approach, especially IPTW (logistic), showed larger standardized differences more often than the other methods. IPTW (GBM) usually showed the smallest standardized difference among all the comparisons. Among the continuous variables used for PS, only age and CHA2DS2-VASc score are displayed in side-by-side boxplots, as they were the most relevant for effectiveness outcomes (Supplementary Figures S9–S16). PSM (GBM) and IPTW (GBM) produced a more similar distribution between the treatment groups than the conventional approaches.

3.2. E-Value

The E-values were calculated as a sensitivity analysis to assess the results’ robustness to potential residual unmeasured confounding [30]. The E-values for the HRs and for one limit of the CI closest to null were estimated in this study and are presented in Table 2. Additionally, the E-values as an anchor and meaningful difference ∆ are presented in Table 2. Because the IPTW (GBM) showed a maximized balance among all PS methods, the E-value as an anchor was derived from it.
Conditional on the measured confounders considered in this study, the range of E-values for adjusted HRs was 1.75–2.70; this indicates that unmeasured confounders associated with exposure and outcome with HRs of at least 1.75–2.70 can nullify the observed HRs; however, this was not the case for weaker unmeasured confounders. The E-values were generally higher in the IPTW (GBM) method than in the other methods (Table 2).
E-values as anchors were generally lower than those from the main analysis, except for Comparison 3 (i.e., the standard dose of dabigatran vs. warfarin). This seems to be related to the fact that Comparison 3 was the only one wherein the NOAC group had a lower age and a lower proportion of women than the warfarin group. Comparisons 1, 3, and 5 showed that the differences in the E-values from the main analysis were larger than the meaningful difference ∆. This indicates that the E-values from the main analysis are meaningfully different in Comparisons 1, 3, and 5. The percentages of maximum possible coverage of GBM were 29%, 64%, 30%, and 16% of the meaningful difference ∆ in Comparisons 2, 4, 7, and 8, respectively (Table 2).

3.3. Negative Control Outcome

For negative control outcome analysis, effect estimates were comparable, and the a priori null hypothesis of neutral associations was generally not rejected in most comparisons (Figure 3). For all comparisons, non-significant HRs were observed slightly more frequently in GBM than in the conventional approach.

4. Discussion

Several studies on the comparative effectiveness analysis of OACs in AF patients, including recent studies using the GBM approach, were published worldwide. However, little is known about the advantages or strengths of the GBM approach over simple logistic regression, which is more widely used in this research area. To the best of our knowledge, this study is the first to show that GBM, which is machine learning approach, demonstrated a better ability to balance covariates and had a lower impact of residual confounding than the conventional approach, in the empirical example of comparative effectiveness analysis in patients with AF.
We compared the GBM (machine-learning-based PS) to logistic regression (conventional PS) for balance and residual confounding. Compared with other methods, the IPTW (GBM) showed better balance achievement in terms of standardized difference, statistical significance, and E-value. In the negative control outcome analysis, the smallest number of statistically significant associations was observed in the PSM (GBM). Considering the results on balance ability and residual confounding in this study, GBM would be useful in comparative effectiveness studies in patients with AF. In this study, we attempted to find the GBM-based PS that achieved the best balance through an iterative process. The number of unbalanced covariates was meaningfully reduced in the GBM approaches than in the conventional approaches, even in comparisons of reduced dose NOACs versus warfarin in which it was more difficult to achieve balance.
Our results may not be generalizable to all diseases or exposure–outcome scenarios, as other issues may be found, such as those related to achieving comparability between groups or research questions [35,36]. However, in many settings, the machine learning approach is expected to perform well, especially in balance, owing to its automated computational system [3,18,27]. Nevertheless, findings regarding the balance ability of the conventional versus machine learning approach were inconsistent [35,37,38], which may be due to the different study settings and tuning parameters for machine learning [35]. Alam et al. found that machine-learning-based PS was comparable or inferior to logistic regression in terms of balance, bias, and mean squared error; moreover, considering computational cost, machine learning has no clear advantage [35]. Harder et al. examined some combinations of PS estimation and application methods and found that the performance of machine-learning-based PS could be better when combined with the weighting method [37]. Lee et al. calculated PS weights using logistic regression and machine learning, and found that logistic regression tended to provide a higher number of extreme weights in all scenarios with various degrees of additivity and linearity [38]. It is noteworthy that both conventional and machine learning approaches for deriving PS have advantages and disadvantages. The conventional approach is simple to implement but has key challenges related to selecting an appropriate set of interaction and polynomial terms, as recommended by Rosenbaum et al. and Dehejia et al. [3,18,39,40,41]. Thus, the conventional approach sometimes requires a very time-consuming process to achieve a better balance, which is performed by humans [3]. In contrast, the machine learning approach is known to have the advantage of being able to identify an appropriate set of interaction and polynomial terms using an automated iterative process to achieve the best balance [3]. However, the machine learning approach also has limitations, such as requiring computational storage and time for the iterative process [42].
The concept of E-value methodology was introduced and used since 2016 [30,43]. A systematic review reported that a substantial overlap was found in E-values between studies concluding that residual confounding is likely to affect the results and those concluding that residual confounding is unlikely to affect the results [30]. In this review, the median E-values for the point estimate were 1.88, 1.82, and 2.02 for 43 studies concluding that confounding was likely to affect, 348 studies concluding that confounding was unlikely to affect, and 125 studies that provided no comment, respectively [30]. As there are no cut-offs that can be used to interpret the magnitude of residual unmeasured confounding, the result of this new E-value approach should be interpreted cautiously [44].
VanderWeele et al. suggested the following considerations when researchers use and interpret E-values: (1) reporting an E-value for estimate and CI, (2) providing discussion regarding specific known unmeasured confounders that could not be accounted for in the analysis, and (3), where possible, comparing E-values to previous studies including confounders that were considered as known unmeasured confounders in the researcher’s study and/or comparing E-values with previous studies that adjusted different combinations of measured confounders from the researcher’s study [44]. We thought that the known unmeasured confounders in this study may be cognitive impairment or frailty; however, it was reported that patients with AF who are comorbid with or without cognitive impairment or frailty experience similar effectiveness (i.e., stroke/transient ischemic attack/non-central nervous system embolism) when using OACs [45]. Thus, although the possibility of some residual unmeasured confounding cannot be completely excluded, the E-values, as a sensitivity analysis in this study, indicate that the study results are likely to be robust to potential residual unmeasured confounding.
This study has several limitations. First, HIRA claims data were collected for reimbursement and not for research purposes. Thus, coding errors may be found in some claims related to defining covariates or outcomes. To minimize this limitation, codes based on previous studies or clinical expert opinions were used [11,13]. This limitation is expected to have a similar impact on the two groups.
Second, only logistic regression and GBM were used for conventional and machine-learning-based PS, respectively. Because these two models are increasingly used to derive PS in the field of AF research [6,7,8,17], it can provide insight into a practical and user-friendly comparison based on real experiences. Further research is needed to provide suggestions for which analytical methods are better in each study context. In addition, research comparing machine learning methods with econometric methods is also needed (e.g., the Heckman selection model or instrumental variable method), as these econometric methods were suggested to account for unmeasured confounding [1,46,47]. Given that machine learning/artificial intelligence and big data analysis proved useful for limiting confounding, the concept of this study could be further extended to other research areas, for example, more efficient targeting of candidate genes by minimizing and evaluating confounding in genetic research on cardiovascular diseases [48,49,50].
Third, other PS methods (e.g., stratification or covariate adjustment using PS) were not considered in the comparison. This is because PSM and IPTW were reported to be able to remove more systematic differences in covariates between groups compared with other PS methods (i.e., covariate adjustment using PS or stratification) [4]. Moreover, both PSM and IPTW were recommended for use by researchers for estimating survival outcomes [51]. However, considering the characteristics of PSM and IPTW, the exclusion of unmatched subjects in PSM results can cause loss of generalizability and precision [36]. In addition, PSM or IPTW may not be appropriate in other studies [36]. Thus, researchers should choose an appropriate PS method based on the characteristics of their study [36].
Fourth, to compare E-values between methods, meaningful difference ∆ was defined by the researchers of this study. Only age- and sex-related variables were considered to derive the E-value as an anchor in IPTW (GBM). Therefore, our results should be interpreted based on the definitions used in this study. Despite these limitations, this analysis is meaningful as it provides insights into the possibility of another interpretation approach for E-values.
Fifth, we could not perform a runtime complexity analysis for our GBM, O(k*m), where k is the depth of the tree and m is the number of decision trees in the tree [52,53]. This could provide useful information for evaluating the efficiency of algorithms and the trade-off between performance and runtime [54]. Although we did not perform this analysis, our runtime to estimate PS using GBM was approximately 45 min on the system with an Intel Xeon 4-core 3.00 GHz CPU and 8 GB RAM, whereas it was 8–10 min when using conventional logistic regression. Future machine learning studies should consider complexity analysis to help readers find a good balance between performance and runtime. Although a longer running time was needed to perform PS using GBM, ML has a useful feature in that it can identify the PS model that provides the best balance [3]; this benefit may outweigh the longer running time needed for ML because of advancements in computer processing power.

5. Conclusions

This study compared the ability to balance baseline covariates and explored the impact of residual confounding between conventional and machine-learning-based PS in the context of a comparative effectiveness study into patients with AF. We showed that GBM provided a better ability to balance covariates and had a lower impact on residual confounding, compared with the conventional approach, in the empirical example of comparative effectiveness analysis in patients with AF. Considering that the PS method is widely used, this study highlights the need for a more suitable and efficient adjustment method. This study serves as a basis for developing more advanced practical applications of machine learning and artificial intelligence in future observational studies.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijerph191912916/s1. Figure S1: Absolute standardized differences in comparison 1; Figure S2: Absolute standardized differences in comparison 2; Figure S3: Absolute standardized differences in comparison 3; Figure S4: Absolute standardized differences in comparison 4; Figure S5: Absolute standardized differences in comparison 5; Figure S6: Absolute standardized differences in comparison 6; Figure S7: Absolute standardized differences in comparison 7; Figure S8: Absolute standardized differences in comparison 8; Figure S9: Side-by-side boxplot in comparison 1; Figure S10: Side-by-side boxplot in comparison 2; Figure S11: Side-by-side boxplot in comparison 3; Figure S12: Side-by-side boxplot in comparison 4; Figure S13: Side-by-side boxplot in comparison 5; Figure S14: Side-by-side boxplot in comparison 6; Figure S15: Side-by-side boxplot in comparison 7; Figure S16: Side-by-side boxplot in comparison 8; Table S1: Definitions for stroke or systemic embolism; Table S2: Propensity score variables during pre-index period; Table S3: Definitions for negative control outcomes; Table S4: R codes for generalized boosted model to estimate propensity score; Table S5: Length of follow-up period in each comparison; Table S6: Baseline characteristics for comparison 1 (standard dose of apixaban vs warfarin); Table S7: Baseline characteristics for comparison 2 (reduced dose of apixaban vs warfarin); Table S8: Baseline characteristics for comparison 3 (standard dose of dabigatran vs warfarin); Table S9: Baseline characteristics for comparison 4 (reduced dose of dabigatran vs warfarin); Table S10: Baseline characteristics for comparison 5 (standard dose of edoxaban vs warfarin); Table S11: Baseline characteristics for comparison 6 (reduced dose of edoxaban vs warfarin); Table S12: Baseline characteristics for comparison 7 (standard dose of rivaroxaban vs warfarin); Table S13: Baseline characteristics for comparison 8 (reduced dose of rivaroxaban vs warfarin); Table S14: Absolute standardized difference of baseline characteristics in comparison 1; Table S15: Absolute standardized difference of baseline characteristics in comparison 2; Table S16: Absolute standardized difference of baseline characteristics in comparison 3; Table S17: Absolute standardized difference of baseline characteristics in comparison 4; Table S18: Absolute standardized difference of baseline characteristics in comparison 5; Table S19: Absolute standardized difference of baseline characteristics in comparison 6; Table S20: Absolute standardized difference of baseline characteristics in comparison 7; Table S21: Absolute standardized difference of baseline characteristics in comparison 8.

Author Contributions

Conceptualization, S.H. and H.S.S.; methodology, S.H. and H.S.S.; formal analysis, S.H.; investigation, S.H. and H.S.S.; data curation, S.H. and H.S.S.; writing—original draft preparation, S.H. and H.S.S.; writing—review and editing, S.H. and H.S.S.; supervision, H.S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2020R1F1A1069526). This work was supported by a grant (21153MFDS601) from the Ministry of Food and Drug Safety in 2021. This work was supported by an NRF (National Research Foundation of Korea) grant funded by the Korean Government (NRF-2019-Global Ph.D. Fellowship Program). The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Institutional Review Board Statement

The Pusan National University Institutional Review Board determined that this study was exempted from ethical review as it utilized publicly available secondary data (PNU IRB/2020_29_HR).

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are available at https://opendata.hira.or.kr/ (accessed on 27 May 2020) with permission from the Health Insurance Review and Assessment Service. Data were provided by the Health Insurance Review and Assessment Service after reviewing the researcher’s request for academic purposes.

Acknowledgments

This article was adapted from part of a doctoral thesis submitted to Pusan National University by the first author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Uddin, M.J.; Groenwold, R.H.; Ali, M.S.; de Boer, A.; Roes, K.C.; Chowdhury, M.A.; Klungel, O.H.J. Methods to control for unmeasured confounding in pharmacoepidemiology: An overview. Int. J. Clin. Pharm. 2016, 38, 714–723. [Google Scholar] [CrossRef]
  2. Monti, S.; Grosso, V.; Todoerti, M.; Caporali, R. Randomized controlled trials and real-world data: Differences and similarities to untangle literature data. Rheumatology 2018, 57, vii54–vii58. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. McCaffrey, D.F.; Griffin, B.A.; Almirall, D.; Slaughter, M.E.; Ramchand, R.; Burgette, L.F. A tutorial on propensity score estimation for multiple treatments using generalized boosted models. Stat. Med. 2013, 32, 3388–3414. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Austin, P.C. An introduction to propensity score methods for reducing the effects of confounding in observational studies. Multivar. Behav. Res. 2011, 46, 399–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Cha, M.-J.; Choi, E.-K.; Han, K.-D.; Lee, S.-R.; Lim, W.-H.; Oh, S.; Lip, G.Y. Effectiveness and safety of non-vitamin K antagonist oral anticoagulants in Asian patients with atrial fibrillation. Stroke 2017, 48, 3040–3048. [Google Scholar] [CrossRef]
  6. Chan, Y.H.; See, L.C.; Tu, H.T.; Yeh, Y.H.; Chang, S.H.; Wu, L.S.; Lee, H.F.; Wang, C.L.; Kuo, C.F.; Kuo, C.T. Efficacy and safety of apixaban, dabigatran, rivaroxaban, and warfarin in Asians with nonvalvular atrial fibrillation. J. Am. Heart Assoc. 2018, 7, e008150. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Larsen, T.B.; Skjøth, F.; Nielsen, P.B.; Kjældgaard, J.N.; Lip, G.Y. Comparative effectiveness and safety of non-vitamin K antagonist oral anticoagulants and warfarin in patients with atrial fibrillation: Propensity weighted nationwide cohort study. BMJ 2016, 353, i3189. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Yao, X.; Abraham, N.S.; Sangaralingham, L.R.; Bellolio, M.F.; McBane, R.D.; Shah, N.D.; Noseworthy, P.A. Effectiveness and safety of dabigatran, rivaroxaban, and apixaban versus warfarin in nonvalvular atrial fibrillation. J. Am. Heart Assoc. 2016, 5, e003725. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Chan, Y.-H.; Lee, H.-F.; Chao, T.-F.; Wu, C.-T.; Chang, S.-H.; Yeh, Y.-H.; See, L.-C.; Kuo, C.-T.; Chu, P.-H.; Wang, C.-L.; et al. Real-world comparisons of direct oral anticoagulants for stroke prevention in Asian patients with non-valvular atrial fibrillation: A systematic review and meta-analysis. Cardiovasc. Drugs Ther. 2019, 33, 701–710. [Google Scholar] [CrossRef] [PubMed]
  10. Wang, K.-L.; Lip, G.Y.; Lin, S.-J.; Chiang, C.-E. Non–vitamin K antagonist oral anticoagulants for stroke prevention in Asian patients with nonvalvular atrial fibrillation: Meta-analysis. Stroke 2015, 46, 2555–2561. [Google Scholar] [CrossRef]
  11. Bang, O.Y.; On, Y.K.; Lee, M.-Y.; Jang, S.-W.; Han, S.; Han, S.; Won, M.-M.; Park, Y.-J.; Lee, J.-M.; Choi, H.-Y.; et al. The risk of stroke/systemic embolism and major bleeding in Asian patients with non-valvular atrial fibrillation treated with non-vitamin K oral anticoagulants compared to warfarin: Results from a real-world data analysis. PLoS ONE 2020, 15, e0242922. [Google Scholar] [CrossRef]
  12. Kim, J.-A.; Yoon, S.; Kim, L.-Y.; Kim, D.-S. Towards actualizing the value potential of Korea Health Insurance Review and Assessment (HIRA) data as a resource for health research: Strengths, limitations, applications, and strategies for optimal use of HIRA data. J. Korean Med. Sci. 2017, 32, 718–728. [Google Scholar] [CrossRef] [PubMed]
  13. Ministry of Health and Welfare. Production of Statistics of Quality of Health Care in 2015–16 [Korean]. 2017. Available online: http://www.google.co.kr/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&cad=rja&uact=8&ved=2ahUKEwjtrIK7gvPsAhW9L6YKHXR4AUAQFjAAegQIAhAC&url=http%3A%2F%2Fwww.mohw.go.kr%2Freact%2Fmodules%2Fdownload.jsp%3FBOARD_ID%3D60010%26CONT_SEQ%3D356240%26FILE_SEQ%3D296869&usg=AOvVaw1v2GXUM7OHTEc6mOB_fx1V (accessed on 18 October 2020).
  14. Joung, B.; Lee, J.M.; Lee, K.H.; Kim, T.-H.; Choi, E.-K.; Lim, W.-H.; Kang, K.-W.; Shim, J.; Lim, H.E.; Park, J. 2018 Korean guideline of atrial fibrillation management. Korean Circ. J. 2018, 48, 1033–1080. [Google Scholar] [CrossRef] [Green Version]
  15. McCaffrey, D.F.; Ridgeway, G.; Morral, A.R. Propensity score estimation with boosted regression for evaluating causal effects in observational studies. Psychol. Methods 2004, 9, 403–425. [Google Scholar] [CrossRef] [Green Version]
  16. Coffman, D.L.; Zhou, J.; Cai, X. Comparison of methods for handling covariate missingness in propensity score estimation with a binary exposure. BMC Med. Res. Methodol. 2020, 20, 168. [Google Scholar] [CrossRef] [PubMed]
  17. Allan, V.; Ramagopalan, S.V.; Mardekian, J.; Jenkins, A.; Li, X.; Pan, X.; Luo, X. Propensity score matching and inverse probability of treatment weighting to address confounding by indication in comparative effectiveness research of oral anticoagulants. J. Comp. Eff. Res. 2020, 9, 603–614. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Ferri-García, R.; Rueda, M.d.M. Propensity score adjustment using machine learning classification algorithms to control selection bias in online surveys. PLoS ONE 2020, 15, e0231500. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Alves, M.F. Causal Inference for the Brave and True. 11—Propensity Score. Available online: https://matheusfacure.github.io/python-causality-handbook/11-Propensity-Score.html# (accessed on 26 November 2021).
  20. Brookhart, M.A.; Schneeweiss, S.; Rothman, K.J.; Glynn, R.J.; Avorn, J.; Stürmer, T. Variable selection for propensity score models. Am. J. Epidemiol. 2006, 163, 1149–1156. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Connolly, S.J.; Ezekowitz, M.D.; Yusuf, S.; Eikelboom, J.; Oldgren, J.; Parekh, A.; Pogue, J.; Reilly, P.A.; Themeles, E.; Varrone, J.; et al. Dabigatran versus warfarin in patients with atrial fibrillation. N. Engl. J. Med. 2009, 361, 1139–1151. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Giugliano, R.P.; Ruff, C.T.; Braunwald, E.; Murphy, S.A.; Wiviott, S.D.; Halperin, J.L.; Waldo, A.L.; Ezekowitz, M.D.; Weitz, J.I.; Špinar, J. Edoxaban versus warfarin in patients with atrial fibrillation. N. Engl. J. Med. 2013, 369, 2093–2104. [Google Scholar] [CrossRef]
  23. Granger, C.B.; Alexander, J.H.; McMurray, J.J.; Lopes, R.D.; Hylek, E.M.; Hanna, M.; Al-Khalidi, H.R.; Ansell, J.; Atar, D.; Avezum, A. Apixaban versus warfarin in patients with atrial fibrillation. N. Engl. J. Med. 2011, 365, 981–992. [Google Scholar] [CrossRef] [Green Version]
  24. Noseworthy, P.A.; Yao, X.; Abraham, N.S.; Sangaralingham, L.R.; McBane, R.D.; Shah, N.D. Direct comparison of dabigatran, rivaroxaban, and apixaban for effectiveness and safety in nonvalvular atrial fibrillation. Chest 2016, 150, 1302–1312. [Google Scholar] [CrossRef]
  25. Patel, M.R.; Mahaffey, K.W.; Garg, J.; Pan, G.; Singer, D.E.; Hacke, W.; Breithardt, G.; Halperin, J.L.; Hankey, G.J.; Piccini, J.P. Rivaroxaban versus warfarin in nonvalvular atrial fibrillation. N. Engl. J. Med. 2011, 365, 883–891. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Ridgeway, G.; McCaffrey, D.; Morral, A.; Burgette, L.; Griffin, B.A. Toolkit for Weighting and Analysis of Nonequivalent Groups: A Tutorial for the Twang Package; RAND Corporation: Santa Monica, CA, USA, 2017. [Google Scholar]
  27. Parast, L.; McCaffrey, D.F.; Burgette, L.F.; de la Guardia, F.H.; Golinelli, D.; Miles, J.N.; Griffin, B.A.J. Optimizing variance-bias trade-off in the TWANG package for estimation of propensity scores. Health Serv. Outcomes Res. Methodol. 2017, 17, 175–197. [Google Scholar] [CrossRef]
  28. Austin, P.C. Using the standardized difference to compare the prevalence of a binary variable between two groups in observational research. Commun. Stat. Simul. Comput. 2009, 38, 1228–1234. [Google Scholar] [CrossRef]
  29. Cusson, A.; Infante-Rivard, C. Bias factor, maximum bias and the E-value: Insight and extended applications. Int. J. Epidemiol. 2020, 49, 1509–1516. [Google Scholar] [CrossRef]
  30. Blum, M.R.; Tan, Y.J.; Ioannidis, J.P.A. Use of E-values for addressing confounding in observational studies—An empirical assessment of the literature. Int. J. Epidemiol. 2020, 49, 1482–1494. [Google Scholar] [CrossRef]
  31. Mathur, M.B.; Ding, P.; Riddell, C.A.; VanderWeele, T.J. Website and R package for computing E-values. Epidemiology 2018, 29, e45. [Google Scholar] [CrossRef] [PubMed]
  32. VanderWeele, T.J.; Ding, P. Sensitivity analysis in observational research: Introducing the E-value. Ann. Intern. Med. 2017, 167, 268–274. [Google Scholar] [CrossRef] [Green Version]
  33. Torp-Pedersen, C.; Goette, A.; Nielsen, P.B.; Potpara, T.; Fauchier, L.; John Camm, A.; Arbelo, E.; Boriani, G.; Skjoeth, F.; Rumsfeld, J.; et al. ‘Real-world’ observational studies in arrhythmia research: Data sources, methodology, and interpretation. A position document from European Heart Rhythm Association (EHRA), endorsed by Heart Rhythm Society (HRS), Asia-Pacific HRS (APHRS), and Latin America HRS (LAHRS). EP Eur. 2020, 22, 831–832. [Google Scholar]
  34. Lip, G.Y.H.; Skjøth, F.; Nielsen, P.B.; Kjældgaard, J.N.; Larsen, T.B. Effectiveness and Safety of Standard-Dose Nonvitamin K Antagonist Oral Anticoagulants and Warfarin Among Patients with Atrial Fibrillation with a Single Stroke Risk Factor: A Nationwide Cohort Study. JAMA Cardiol. 2017, 2, 872–881. [Google Scholar] [CrossRef]
  35. Alam, S.; Moodie, E.E.M.; Stephens, D.A. Should a propensity score model be super? The utility of ensemble procedures for causal adjustment. Stat. Med. 2019, 38, 1690–1702. [Google Scholar] [CrossRef]
  36. Elze, M.C.; Gregson, J.; Baber, U.; Williamson, E.; Sartori, S.; Mehran, R.; Nichols, M.; Stone, G.W.; Pocock, S.J. Comparison of propensity score methods and covariate adjustment: Evaluation in 4 cardiovascular studies. J. Am. Coll. Cardiol. 2017, 69, 345–357. [Google Scholar] [CrossRef]
  37. Harder, V.S.; Stuart, E.A.; Anthony, J.C. Propensity score techniques and the assessment of measured covariate balance to test causal associations in psychological research. Psychol. Methods 2010, 15, 234–249. [Google Scholar] [CrossRef] [Green Version]
  38. Lee, B.K.; Lessler, J.; Stuart, E.A. Improving propensity score weighting using machine learning. Stat. Med. 2010, 29, 337–346. [Google Scholar] [CrossRef] [Green Version]
  39. Dehejia, R.H.; Wahba, S. Causal effects in nonexperimental studies: Reevaluating the evaluation of training programs. J. Am. Stat. Assoc. 1999, 94, 1053–1062. [Google Scholar] [CrossRef]
  40. Dehejia, R.H.; Wahba, S. Propensity score-matching methods for nonexperimental causal studies. Rev. Econ. Stat. 2002, 84, 151–161. [Google Scholar] [CrossRef] [Green Version]
  41. Rosenbaum, P.R.; Rubin, D.B. Reducing bias in observational studies using subclassification on the propensity score. J. Am. Stat. Assoc. 1984, 79, 516–524. [Google Scholar] [CrossRef]
  42. Ridgeway, G. Generalized Boosted Models: A Guide to the Gbm Package. Updated 2020; pp. 1–15. Available online: https://cran.r-project.org/web/packages/gbm/vignettes/gbm.pdf (accessed on 11 November 2021).
  43. Ding, P.; VanderWeele, T.J. Sensitivity Analysis Without Assumptions. Epidemiology 2016, 27, 368–377. [Google Scholar] [CrossRef] [Green Version]
  44. VanderWeele, T.J.; Mathur, M.B. Commentary: Developing best-practice guidelines for the reporting of E-values. Int. J. Epidemiol. 2020, 49, 1495–1497. [Google Scholar] [CrossRef]
  45. Madhavan, M.; Holmes, D.N.; Piccini, J.P.; Ansell, J.E.; Fonarow, G.C.; Hylek, E.M.; Kowey, P.R.; Mahaffey, K.W.; Thomas, L.; Peterson, E.D.; et al. Association of frailty and cognitive impairment with benefits of oral anticoagulation in patients with atrial fibrillation. Am. Heart J. 2019, 211, 77–89. [Google Scholar] [CrossRef]
  46. DeMaris, A. Combating unmeasured confounding in cross-sectional studies: Evaluating instrumental-variable and Heckman selection models. Psychol. Methods 2014, 19, 380. [Google Scholar] [CrossRef] [Green Version]
  47. Schneeweiss, S. Sensitivity analysis and external adjustment for unmeasured confounders in epidemiologic database studies of therapeutics. Pharmacoepidemiol. Drug Saf. 2006, 15, 291–303. [Google Scholar] [CrossRef]
  48. Hai, Q.; Ritchey, B.; Robinet, P.; Alzayed, A.M.; Brubaker, G.; Zhang, J.; Smith, J.D. Quantitative trait locus mapping of macrophage cholesterol metabolism and CRISPR/Cas9 editing implicate an ACAT1 truncation as a causal modifier variant. Arterioscler. Thromb. Vasc. Biol. 2018, 38, 83–91. [Google Scholar] [CrossRef]
  49. Ritchey, B.; Hai, Q.; Han, J.; Barnard, J.; Smith, J.D. Genetic variant in 3′ untranslated region of the mouse pycard gene regulates inflammasome activity. eLife 2021, 10, e68203. [Google Scholar] [CrossRef]
  50. Hsu, J.; Gore-Panter, S.; Tchou, G.; Castel, L.; Lovano, B.; Moravec, C.S.; Pettersson, G.B.; Roselli, E.E.; Gillinov, A.M.; McCurry, K.R. Genetic control of left atrial gene expression yields insights into the genetic susceptibility for atrial fibrillation. Circ. Genom. Precis. Med. 2018, 11, e002107. [Google Scholar] [CrossRef] [Green Version]
  51. Austin, P.C. The use of propensity score methods with survival or time-to-event outcomes: Reporting measures of effect similar to those used in randomized experiments. Stat. Med. 2014, 33, 1242–1258. [Google Scholar] [CrossRef] [Green Version]
  52. Aljofey, A.; Rasool, A.; Jiang, Q.; Qu, Q. A Feature-Based Robust Method for Abnormal Contracts Detection in Ethereum Blockchain. Electronics 2022, 11, 2937. [Google Scholar] [CrossRef]
  53. Chadrasekhar, G. Time and Space Complexity of Machine Learning Models. Available online: https://pub.towardsai.net/time-and-space-complexity-of-machine-learning-models-df9b704e3e9c (accessed on 16 September 2022).
  54. Martindale, N.; Ismail, M.; Talbert, D.A. Ensemble-based online machine learning algorithms for network intrusion detection systems using streaming data. Information 2020, 11, 315. [Google Scholar] [CrossRef]
Figure 1. Study scheme. (A) Patient selection and follow-up. (B) Overall process.
Figure 1. Study scheme. (A) Patient selection and follow-up. (B) Overall process.
Ijerph 19 12916 g001aIjerph 19 12916 g001b
Figure 2. Patient-selection flow.
Figure 2. Patient-selection flow.
Ijerph 19 12916 g002
Figure 3. Hazard ratios for negative control outcomes.
Figure 3. Hazard ratios for negative control outcomes.
Ijerph 19 12916 g003aIjerph 19 12916 g003b
Table 1. Summarized process for comparing E-values in this study.
Table 1. Summarized process for comparing E-values in this study.
(1)
Calculate E-values from main analysis (E-values from conventional and machine learning approaches).
(2)
Calculate E-value as an anchor (E-value derived from age/sex excluded PS model) for each comparison.
(3)
Calculate meaningful difference Δ: E-value from main analysis (1) − E-value as an anchor (2) = (1) − (2).
(4)
Calculate difference in E-values from main analysis (e.g., E-value from IPTW (GBM) − E-value from IPTW (logistic)).
(5)
If difference in E-values from main analysis (4) ≥ meaningful difference Δ (3), E-values from main analysis are meaningfully different.
(6)
If difference in E-values from main analysis (4) < meaningful difference Δ (3), E-values from main analysis are not meaningfully different.
(7)
Calculate maximum possible coverage for the comparisons for (6). m a x i m u m   p o s s i b l e   c o v e r a g e = ( m a x i m u m   E v a l u e   f r o m   m a i n   a n a l y s i s ) ( m i n i m u m   E v a l u e   f r o m   m a i n   a n a l y s i s ) m e a n i n g f u l   d i f f e r e n c e   Δ
Table 2. E-values of the estimated hazard ratios for stroke or systemic embolism.
Table 2. E-values of the estimated hazard ratios for stroke or systemic embolism.
ComparisonsE-value for the Hazard Ratios (1–3)
(E-Value for the Limit of Confidence Interval
Closest to the Null)
E-Value
as an Anchor (4)
Meaningful Difference
Δ (5)
Maximum Possible
Coverage
(%) (6)
PSM
(Logistic)
PSM
(GBM)
IPTW
(Logistic)
IPTW
(GBM)
IPTW
(GBM)
Comparison 12.24 (1.92)2.23 (1.90)2.32 (2.04)2.35 (2.07)2.320.03-
Comparison 21.96 (1.66)2.01 (1.71)2.00 (1.78)2.08 (1.85)1.670.4129
Comparison 32.13 (1.76)2.19 (1.81)2.19 (1.82)2.30 (1.93)2.360.06-
Comparison 41.75 (1.47)1.89 (1.60)1.89 (1.64)1.93 (1.68)1.650.2864
Comparison 52.45 (2.05)2.27 (1.88)2.52 (2.13)2.43 (2.05)2.420.01-
Comparison 62.70 (2.32)2.67 (2.28)2.53 (2.21)2.56 (2.24)2.100.4637
Comparison 72.01 (1.73)2.04 (1.75)2.02 (1.78)2.04 (1.80)1.940.1030
Comparison 82.05 (1.76)2.08 (1.78)2.04 (1.80)2.10 (1.85)1.720.3816
Comparison 1, standard dose of apixaban vs warfarin; Comparison 2, reduced dose of apixaban vs warfarin; Comparison 3, standard dose of dabigatran vs warfarin; Comparison 4, reduced dose of dabigatran vs warfarin; Comparison 5, standard dose of edoxaban vs warfarin; Comparison 6, reduced dose of edoxaban vs warfarin; Comparison 7, standard dose of rivaroxaban vs warfarin; Comparison 8, reduced dose of rivaroxaban vs warfarin. (1) Conventional approach: PSM (logistic) and IPTW (logistic). (2) Machine learning approach: PSM (GBM) and IPTW (GBM). (3) Larger E-value indicates that the stronger unmeasured confounder associations can nullify the observed hazard ratios. (4) E-values derived from analysis that excludes age and sex variables from the PS model. (5) Meaningful difference Δ was defined by the researchers of this study as the difference between E-value from main analysis and E-value defined as an anchor. Interpretation under the definition used in this study is needed. (6) Maximum possible coverage (%) = [(maximum E-value from main analysis) − (minimum E-value from main analysis)]/(meaningful difference Δ) × 100. It was calculated for the comparisons that E-values from main analysis were not meaningfully different.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Han, S.; Suh, H.S. Impact of Integrating Machine Learning in Comparative Effectiveness Research of Oral Anticoagulants in Patients with Atrial Fibrillation. Int. J. Environ. Res. Public Health 2022, 19, 12916. https://doi.org/10.3390/ijerph191912916

AMA Style

Han S, Suh HS. Impact of Integrating Machine Learning in Comparative Effectiveness Research of Oral Anticoagulants in Patients with Atrial Fibrillation. International Journal of Environmental Research and Public Health. 2022; 19(19):12916. https://doi.org/10.3390/ijerph191912916

Chicago/Turabian Style

Han, Sola, and Hae Sun Suh. 2022. "Impact of Integrating Machine Learning in Comparative Effectiveness Research of Oral Anticoagulants in Patients with Atrial Fibrillation" International Journal of Environmental Research and Public Health 19, no. 19: 12916. https://doi.org/10.3390/ijerph191912916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop