Next Article in Journal
NRX-101 (D-Cycloserine + Lurasidone) Is Active against Drug-Resistant Urinary Pathogens In Vitro
Previous Article in Journal
What Is the Rate of Antimicrobial Resistance of a Prosthetic Joint Infection in a Major Orthopaedic Centre?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Brave New World of Artificial Intelligence: Its Use in Antimicrobial Stewardship—A Systematic Review

by
Rafaela Pinto-de-Sá
1,
Bernardo Sousa-Pinto
2,3 and
Sofia Costa-de-Oliveira
1,3,*
1
Division of Microbiology, Department of Pathology, Faculty of Medicine, University of Porto, Alameda Prof. Hernâni Monteiro, 4200-319 Porto, Portugal
2
Department of Community Medicine, Information and Health Decision Sciences, Faculty of Medicine, University of Porto, 4200-319 Porto, Portugal
3
Center for Health Technology and Services Research—CINTESIS@RISE, Faculty of Medicine, University of Porto, 4200-319 Porto, Portugal
*
Author to whom correspondence should be addressed.
Antibiotics 2024, 13(4), 307; https://doi.org/10.3390/antibiotics13040307
Submission received: 20 February 2024 / Revised: 24 March 2024 / Accepted: 25 March 2024 / Published: 28 March 2024

Abstract

:
Antimicrobial resistance (AMR) is a growing public health problem in the One Health dimension. Artificial intelligence (AI) is emerging in healthcare, since it is helpful to deal with large amounts of data and as a prediction tool. This systematic review explores the use of AI in antimicrobial stewardship programs (ASPs) and summarizes the predictive performance of machine learning (ML) algorithms, compared with clinical decisions, in inpatients and outpatients who need antimicrobial prescriptions. This review includes eighteen observational studies from PubMed, Scopus, and Web of Science. The exclusion criteria comprised studies conducted only in vitro, not addressing infectious diseases, or not referencing the use of AI models as predictors. Data such as study type, year of publication, number of patients, study objective, ML algorithms used, features, and predictors were extracted from the included publications. All studies concluded that ML algorithms were useful to assist antimicrobial stewardship teams in multiple tasks such as identifying inappropriate prescribing practices, choosing the appropriate antibiotic therapy, or predicting AMR. The most extracted performance metric was AUC, which ranged from 0.64 to 0.992. Despite the risks and ethical concerns that AI raises, it can play a positive and promising role in ASP.

1. Introduction

One Health is, according to the One Health High-Level Expert Panel, “an integrated, unifying approach that aims to sustainably balance and optimise the health of people, animals and ecosystems” [1]. This inextricable link between these actors applies to various fields of health and, inherently, to the growth of antimicrobial resistance (AMR).
AMR is a growing public health problem due to its effect in reducing the effectiveness of antimicrobial therapy and increasing the severity, incidence, and cost of infection [2]. AMR’s emergence, evolution, and spread stem from (i) the widespread and inadequate antimicrobial use in animals and clinical practice, (ii) contaminated environments, (iii) and insufficient infection control measures [3]. This increases the threat of the emergence of super-resistant bacteria [4]. The rapid development and dissemination of the mechanisms of resistance through antibiotic resistance genes (ARGs) to antibiotics used in the clinical setting, adding to the slow and infrequent access to new antimicrobials in recent years, makes AMR one of the most severe threats to global public health in the 21st century.
AMR levels are detected by antimicrobial susceptibility testing (AST). However, this method involves culture of the microorganisms, which can take 2–5 days. This delay in the prescription of the most effective antimicrobials leads to the prolongation of empiric therapy, contributing to the rise of AMR, so measures must be taken to combat this, including improved communication and education about the topic, adequate hygiene for infection control, surveillance practices, antimicrobial stewardship, swifter methods for AMR identification, and the use vaccines and bacteriophages [2,3].
Antimicrobial stewardship programs (ASPs) are a set of interventions aimed at optimizing the use of antimicrobials and, therefore, reducing costs, improving therapeutic outcomes, and reducing AMR [5]. ASPs were introduced in 1974 by McGowan and Finland [6], are applied to human healthcare, animal health, and the environment, and involve the optimal selection, dosage, and duration of therapy as well as the control of its use, which can be achieved with programs that recommend the appropriate adjustments. Typically, an ASP may involve pharmacists and infectious diseases physicians, and the tools available for these teams include limiting formularies, restricting certain classes of antimicrobials, cycling of antibiotics, decision support, and staff education about the optimal antimicrobial considering the patient [5]. These interventions are primarily used in hospital settings such as in intensive care units (ICUs), pediatrics, and neutropenic patients [7,8,9]. Still, efforts should be made for their application in outpatient settings to achieve a significant impact on the reduction of AMR [10]. The measurement of the impacts of ASPs can be categorized into antibiotic use, process and quality measures, costs, and clinical outcome measures, with the latter being the most relevant focus in practice [11]. There are challenges in implementing ASPs, including a lack of motivation for change and awareness, a lack of oversight and control of antimicrobial use in many countries, and over-the-counter therapy [12].
Artificial intelligence (AI) began developing in the 1950s, and its first use in healthcare was in the form of expert systems, which were based on rules provided by medical experts, but were never applied in practice [13]. Machine learning (ML) was developed to overcome the limitation of expert systems that need a large number of rules captured, since ML can find new rules from the data provided, based on their quality and volume [13], benefitting mainly from the enormous amount of health data gathered after the implementation of electronic health records. As some real patient situations are more complex and heterogeneous than a single guideline or the experience of an expert, ML can be a tool used to help decision-making in these situations, since it can analyze a great number of electronic records in a way similar to experts’ logical deduction. ML algorithms can be supervised or unsupervised, and some examples include support vector machines, artificial neural networks, random forests, decision trees, and logistic regression [10,14]. Previous studies have shown that this technology has been used in numerous healthcare fields, including infectious diseases [13]. It has been proved to be useful in prediction [15] and early detection [16] of sepsis, diagnosis of infection [17], prediction of treatment success [18], prediction of antimicrobial resistance [19], and treatment selection [20], meaning that it may be an effective tool to put into practice in antimicrobial stewardship teams, bettering their programs.
This systematic review aims to explore the use of AI in ASPs and summarizes the predictive performance of ML algorithms used in antimicrobial stewardship, compared with clinical and antimicrobial stewardship teams’ decisions, in inpatients and outpatients who need antimicrobial prescription. Studies were selected and screened from January 2010 until December 2022 in the electronic bibliographic databases of PubMed, Scopus, and Web of Science by using a combination of terms such as artificial intelligence, antimicrobial resistance, and stewardship. The protocol of this review was registered in the PROSPERO database (CRD42023470594).

2. Results

2.1. Characteristics of the Included Studies

A total of 4658 citations were identified from the three databases and, after removing the duplicates, 2839 were eligible for screening. A total of 1086 articles were assessed for eligibility and eighteen [20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37] were included in this systematic review (Figure 1). Most studies were excluded because they did not study the application of machine learning models nor their predictive performance or because they were not applied to hospital inpatients and outpatients with infections, such as studies in vitro or regarding drug development.
Characteristics of the eighteen included studies are available in Table 1. All the studies were published since 2016 and in English. One of the studies is an abstract presentation at a congress in video format [37]. One of the studies was from a low-/middle-income country [36], with the rest being from high-income countries. The number of features included in the machine learning algorithms ranged from 6 to 788. The patients included were from different settings; one (5.5%) study was designed for outpatients [35], and two were only applied to ICU patients [26,29]. The number of patients ranged from 48 (on a validation set) to 382,943. Two [20,34] of the studies had a prospective design, with the remaining being retrospective observational studies.
The most common ML algorithms used were logistic regression (12.1%), random forest (12.1%), support vector machine (7.6%), and k-nearest neighbors (6.1%) (Figure 2). The measurements used for predictive performance were not consistent between different studies, but the area under the curve (AUC) (15.9%), sensitivity (9.1%), specificity (8.0%), and precision (6.8%) were the most regularly used (Figure 3).
The features included in the algorithms were divided into the following groups: demographics, adult patients, pediatric patients, clinical, laboratory/microbiological, comorbidities, type of infection, and ICU. The most used features were demographical followed by laboratory/microbiological. Information about the features used in each study is available in Table 2.
The most common validation method was k-fold cross-validation (fivefold and tenfold) to avoid overfitting. Not all included studies provided information about handling missing data or methods to avoid overfitting, and two studies did not reference the model validation method [20,37]. Corbin C.K. et al. [22] replicated the process on an external validation cohort in Boston.

2.2. Risk of Bias/Quality Assessment

All the studies were rated as being of “fair quality” by the NIH Quality Assessment Tool for Observational Cohort and Cross-Sectional Studies; fourteen studies were rated as 57.1% and four [22,26,32,35] were rated as 64.3%. The participation rate, variation in amount or level of exposure, and loss to follow-up criteria were not applied to any of the studies. Only one study [35] provided a sample size justification or power description. No study reported information about blinding the assessors, and only three studies [22,26,32] met the criterion on the statistical adjustments of potential confounding variables. The answer to each of the fourteen criteria, as well as the quality rating, are available in Table 3.
The risk of bias (ROB) and the applicability for model prediction of the eighteen included studies were also assessed by PROBAST (Table 4). Only two studies were ranked as being of “low concern” in the analysis domain [22,24]; six studies were defined as being of “unclear concern” [20,21,23,25,26,29], and ten were ranked as being of “high concern” [27,28,30,31,32,33,34,35,36,37]. In these studies, no information was provided regarding how missing data had been handled. Overall, only one study was rated as having a low ROB [24]. Regarding applicability, one study ranked as being of “high concern” and as having a high ROB due to the lack of participant information and lack of definition of the inclusion and exclusion criteria [27].

2.3. Predictive Performance of Artificial Intelligence Algorithms

The most evaluated performance metric was AUC, which ranged from 0.64 to 0.992 (the highest value was obtained by the multilayer perceptron). This algorithm also achieved the highest sensitivity (0.967) and specificity (0.992) for auditing appropriate surgical antimicrobial prophylaxis. The highest precision was achieved by the gradient boosted tree, with an average precision of 0.99 for the selection of vancomycin + meropenem. The other main results are available in Table 1. All the studies concluded that ML algorithms were useful to assist antimicrobial stewardship teams in multiple tasks such as identifying inappropriate prescribing practices [20], choosing the appropriate antibiotic therapy [22,23,34,36], auditing surgical antimicrobial prophylaxis [24], predicting personal risk of treatment-induced emergence of resistance [25], estimating patient outcomes under the contrasting scenarios of stopping or continuing antibiotic treatment [26], predicting AMR [27], and identifying patients at low risk of bacterial infections [29].
Regarding the choice of the most appropriate antibiotic therapy, the model with the best performance was random forest, with an area under the curve of 0.80 (95% CI 0.66–0.94) for the prediction of susceptibility to ceftriaxone, 0.74 (0.59–0.89) for ampicillin and gentamicin, and 0.85 (0.70–1.00) for susceptibility to neither [36].
For the identification of inappropriate prescribing practices of piperacillin-tazobactam, the algorithm applied was the supervised learning module of APSS (antimicrobial prescription surveillance system). It obtained an overall positive predictive value of 74% (95% CI, 68–79), with sensitivity (recall) of 96% (92–98) and accuracy of 79% (74–83) [20].
Logistic regression achieved a 67% reduction in second-line antibiotics relative to clinicians and an 18% reduction in inappropriate antibiotic therapy [35].

3. Discussion

3.1. Main Findings

A systematic review of the utility of AI in antimicrobial stewardship for inpatients and outpatients who needed antimicrobial decisions was conducted, and eighteen studies were included. Logistic regression and random forest were the most used algorithms. AUC was the most common predictive performance measure, and the highest value was obtained by the multilayer perceptron [24]. The most studied application of AI in ASPs was the use of AI for choosing the appropriate antibiotic therapy. In one study, the algorithm used was a semi-supervised decision support system [21]; the remaining algorithms applied supervised ML algorithms, which are generally used to make predictions. All the studies concluded that AI algorithms can help choose the best antimicrobial therapy, benefiting, for example, the control of AMR. These results are aligned with what has been found about AI use in infectious diseases, since other systematic reviews summarize its applicability in antimicrobial susceptibility testing [14], predicting antimicrobial resistance [38], prediction of treatment success, diagnosis of infection, and prediction of sepsis [13].

3.2. AI and Antimicrobial Stewardship

Although AI can be helpful in addressing the large amount of data gathered nowadays and performing repetitive tasks, there are some risks and ethical concerns that must be considered, for example, the possibility of the algorithm making associations between features and outcomes that are not relevant or are without physiological/clinical rationale, the blind obedience/overdependence on AI, liability, or accountability in case of mistakes [39]. Clinical decisions are complex and include factors about the patient, the disease, the economy, or the environment, so the algorithm should not uniquely make the final decision. “Black box” is an aspect of AI that raises concerns, since these algorithms cannot explain the underlying mechanism to generate outputs, and we may not know the source of data input. This has a significant impact on transparency and trust [40,41]. In response to the rise of AI health technologies, the WHO published six regulatory areas of AI for health, including the transparency of development processes, external data validation, cybersecurity, and data protection [42]. The WHO emphasizes the need for collaboration between regulators, patients, healthcare professionals, industry, and governments to ensure the compliance of AI models with regulation. The application of AI on antimicrobial stewardship programs is still very limited, as seen in the few studies included in this systematic review. The methodological heterogeneity and the reduced number of diseases in which AI has been applied on ASPs restrict the widespread use of ML models in antimicrobial stewardship. Tools based on AI for this purpose are still in a development phase before they can be safely implemented in healthcare.
Addressing the perception among some clinicians that the use of AI in antimicrobial stewardship is more of a mirage than a reality necessitates a clear discussion on its evident benefits. Implementing AI requires a calculated investment in technology and skilled data analysts, with the scale dependent on each hospital’s needs. A thorough cost/benefit analysis is vital, showcasing the expenses and expected advancements in healthcare efficiency and patient care quality. Embracing AI, despite initial doubts, is crucial for the evolution of antimicrobial stewardship, moving the perception of antimicrobial stewardship from skepticism to accepted implementation.

3.3. Limitations of the Studies Included

The research on AI applications in ASP is mostly from high-income countries, which can introduce bias on the algorithms and inequalities in healthcare because it does not represent the entire population [43]. This may happen because low- and middle-income countries may face more challenges to implement systems allowing for the collection of large amounts of structured health data, access to health is scarcer, and the financial support for implementing AI algorithms needs to be improved. Efforts should be made to include data from these populations in training and validation datasets.
There needs to be a publicly recognized tool for quality and risk assessment of ML prediction models. PROBAST and the National Institute of Health (NIH) Quality Assessment Tool for Observational Cohort and Cross-Sectional Studies were used for a more complete assessment. For most studies, there was a lack of information about sample size justification or power description and a poor description of the statistical adjustment of confounders. This is a concern, since AI algorithms can provide biased results if the information input is subject to uncontrolled biases. Bolton et al. [26] consider that the models may have “learned” the association between less severe patients receiving fewer antibiotics and, therefore, having a shorter ICU length of stay, causing some confounding. PROBAST assessment of ROB raises concerns, since only one study was ranked as “low concern”. This is mainly due to the analysis domain, as not all the included studies provided information about the handling of missing data or the methods to avoid overfitting, and two studies [20,37] did not report information on validation methods. One of the studies performed external validation of the model [22], which raises concerns about the generalizability of the algorithms used in the other studies. G. Eickelberg and colleagues [29] state that their future research will focus on external validation and clinical utility assessment of the models. The lack of participant information and the definition of the inclusion and exclusion criteria also raise concerns about the applicability and biases of the study’s conclusions [27]. It is relevant to note that providing participants’ information can minimize or highlight biases that can influence the application of the algorithms in specific populations in which they were not studied. Kanjilal et al. [35] admit this limitation in their study. The features selected for the ML algorithms were adequate, since they gather information that influences therapy decisions and patients’ outcomes, mainly raising low concerns. Studies on AI use in health should provide all the features included so there is more transparency and understanding of the processes involved. This will allow for analysis of whether the features have a medical reasoning behind the clinical outcome.

3.4. Limitations of the Review

There are some limitations to this review. The literature search was limited to PubMed, Web of Science, and Scopus articles, with no other bibliographic databases having been searched. Although this is a recent research topic, this information can be quickly complemented with more recent data.
Publication bias is a possible limitation of this review, since it is likely that the studies with more favorable results have higher chances of being accepted for publication. Due to the diversity of the included studies (including differences in outcomes, assessed features, and the algorithms used), we could not perform meta-analysis.
It must be kept in mind that the AI algorithms are not implemented to substitute the healthcare professionals who make up antimicrobial stewardship teams but rather to assist in decision-making, mainly when a considerable amount of health data are gathered every day.
In the future, it would be interesting to research the integration of AI in ASPs, its adoption by healthcare professionals, usability and applicability, and their knowledge about the potential of using AI as a tool [44].

4. Materials and Methods

The systematic review was carried out in accordance with the Cochrane Handbook for Systematic Reviews of Interventions [45]; in addition, we followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) [46] checklist for the review (Supplementary Materials, Table S1).

4.1. Data Source and Search Strategy

The electronic bibliographic databases of PubMed, Scopus, and Web of Science were searched using a combination of MeSH terms and/or keywords regarding broad domains such as artificial intelligence, antimicrobial resistance, and stewardship. For this search strategy, the following query was used: (“artificial intelligence” OR “machine learning” OR “deep learning”) AND (“antibiotic resistance” OR “antibiotic resistant” OR “antifungal resistance” OR “antifungal resistant” OR “antimicrobial resistance” OR “antimicrobial resistant” OR “antibiotic susceptibility” OR “antifungal susceptibility” OR “antimicrobial susceptibility” OR “drug resistance” OR “drug resistant”). Additionally, and to avoid any bibliography loss, the terms (“artificial intelligence” OR “machine learning” OR “deep learning”) AND (stewardship) were included. Studies were selected and screened from January 2010 until December 2022, when the search results were last consulted. The search included all publication types except reviews or systematic reviews, and no language restrictions were applied.

4.2. Eligibility Criteria

Studies were included in this review if they assessed the performance of artificial intelligence models in ASP applied to hospital inpatients and outpatients with infections that needed antimicrobial treatment. We excluded (1) studies conducted only in vitro; (2) studies addressing non-infectious diseases such as cancer, epilepsy, or other neurologic diseases; (3) studies addressing the application of AI in food or animal production, drug development, disease diagnostic or survival or studies focusing on HIV, parasitic diseases, or tuberculosis; and (4) studies not focusing on bacterial infections.
This review intended to study the performance of AI algorithms for antimicrobial stewardship. The question being addressed can be expressed as follows:
P: Inpatients and outpatients who need an antimicrobial prescription;
I: Machine learning models used in antimicrobial stewardship;
C: Clinical or antimicrobial stewardship teams’ decision;
O: Predictive performance of ML algorithms (area under the curve (AUC), sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), etc.).

4.3. Data Extraction and Synthesis

The extracted studies were uploaded to EndNoteTM20 and Rayyan software [47] for duplicate removal, quality assessment, and further selection. Studies were selected first by title and abstract screening and then by full text reading. Both processes were independently performed by two reviewers (RPS and SCO) in a blinded, standardized manner. Eighteen studies were included in the systematic review (Figure 1).
A form was developed to extract the data from the included studies uniformly and consistently. We retrieved data on the study type, year of publication, country, study time frame, target population (demographic data), number of patients, hospital setting, type of infection, study objective, ML algorithms used, training data sets, number of features, data source (clinical and/or laboratory data), predictors, performance validation and metrics (AUC, sensitivity, specificity, etc.), and clinical outcome. Two authors (RPS and SCO) extracted data from primary studies independently.

4.4. Risk of Bias (ROB) Assessment

To evaluate the risk of bias of the studies included in this review, the National Institute of Health (NIH) Quality Assessment Tool for Observational Cohort and Cross-Sectional Studies and PROBAST (a tool to assess the risk of bias and applicability of prediction model studies) were used [48,49,50]. A three-point scale was used to grade the potential source of bias as good, fair, or poor. Regarding PROBAST, the risk of bias and applicability were assessed focusing on four domains (participants, predictors, outcomes, and analysis), which were evaluated for each included study. The risk of bias was defined as “high risk/concern” if the item’s answer was “No” or “Probably no” and “Unclear risk” if relevant information was absent. No studies were excluded based on quality. ROB assessment was performed independently by all authors.

4.5. Data Analysis

The predictive performance of the AI algorithms was extracted as some of these metrics: area under the curve, specificity, sensitivity, precision, accuracy (Table 1).
A meta-analysis was not conducted, due to the heterogeneity between the populations, algorithms, features, and aim of the studies included.

5. Conclusions

This systematic review focuses on various tasks where AI can be a supplemental tool for antimicrobial stewardship teams, benefiting the patient and the healthcare providers. It can assist in the identification of inappropriate prescriptions, the choice of appropriate antibiotic therapy, or the estimation of patient outcomes. This is essential in the One Health dimension, because preventing AMR and multiresistant microorganisms in humans interdependently benefits the health of animals, plants, and ecosystems. The supervised machine learning module of antimicrobial prescription surveillance systems and random forest could be useful tools for guiding the most appropriate antibiotic therapy. AI can assist antimicrobial stewardship teams, aiming at better control of AMR; thus, AI can be a valuable tool against this growing global health issue.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/antibiotics13040307/s1, Table S1: PRISMA 2020 Checklist.

Author Contributions

Conceptualization, R.P.-d.-S. and S.C.-d.-O.; methodology, R.P.-d.-S. and S.C.-d.-O.; validation, R.P.-d.-S., B.S.-P. and S.C.-d.-O.; formal analysis, R.P.-d.-S. and S.C.-d.-O.; data curation, R.P.-d.-S. and S.C.-d.-O.; writing—original draft preparation, R.P.-d.-S.; writing—review and editing, S.C.-d.-O. and B.S.-P.; supervision, S.C.-d.-O.; funding acquisition, S.C.-d.-O. All authors have read and agreed to the published version of the manuscript.

Funding

SCO acknowledges national funds through the FCT and IP within the scope of the project RISE-LA/P/0053/2020. This article was supported by National Funds through FCT—Fundação para a Ciência e a Tecnologia, I.P., within CINTESIS, R&D Unit (reference UIDP/4255/2020).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tripartite and UNEP Support OHHLEP’s Definition of ‘One Health’. Available online: https://www.who.int/news/item/01-12-2021-tripartite-and-unep-support-ohhlep-s-definition-of-one-health (accessed on 20 September 2023).
  2. O’Neill, J. Tackling Drug-Resistant Infections Globally: Final Report and Recommendations. Rev. Antimicrob. Resist. Arch. Pharm. Pract. 2016, 7, 110. [Google Scholar]
  3. Aslam, B.; Khurshid, M.; Arshad, M.I.; Muzammil, S.; Rasool, M.; Yasmeen, N.; Shah, T.; Chaudhry, T.H.; Rasool, M.H.; Shahid, A.; et al. Antibiotic Resistance: One Health One World Outlook. Front. Cell. Infect. Microbiol. 2021, 11, 771510. [Google Scholar] [CrossRef]
  4. González-Zorn, B.; Escudero, J.A. Ecology of Antimicrobial Resistance: Humans, Animals, Food and Environment. Int. Microbiol. 2012, 15, 101–109. [Google Scholar] [CrossRef]
  5. Rice, L.B. Antimicrobial Stewardship and Antimicrobial Resistance. Med. Clin. N. Am. 2018, 102, 805–818. [Google Scholar] [CrossRef]
  6. McGowan, J.E.; Finland, M. Usage of Antibiotics in a General Hospital: Effect of Requiring Justification. J. Infect. Dis. 1974, 130, 165–168. [Google Scholar] [CrossRef]
  7. Lanckohr, C.; Bracht, H. Antimicrobial Stewardship. Curr. Opin. Crit. Care 2022, 28, 551–556. [Google Scholar] [CrossRef]
  8. Donà, D.; Barbieri, E.; Daverio, M.; Lundin, R.; Giaquinto, C.; Zaoutis, T.; Sharland, M. Implementation and Impact of Pediatric Antimicrobial Stewardship Programs: A Systematic Scoping Review. Antimicrob. Resist. Infect. Control 2020, 9, 3. [Google Scholar] [CrossRef]
  9. Contejean, A.; Abbara, S.; Chentouh, R.; Alviset, S.; Grignano, E.; Gastli, N.; Casetta, A.; Willems, L.; Canouï, E.; Charlier, C.; et al. Antimicrobial Stewardship in High-Risk Febrile Neutropenia Patients. Antimicrob. Resist. Infect. Control 2022, 11, 52. [Google Scholar] [CrossRef]
  10. Rabaan, A.A.; Alhumaid, S.; Mutair, A.A.; Garout, M.; Abulhamayel, Y.; Halwani, M.A.; Alestad, J.H.; Bshabshe, A.A.; Sulaiman, T.; AlFonaisan, M.K.; et al. Application of Artificial Intelligence in Combating High Antimicrobial Resistance Rates. Antibiotics 2022, 11, 784. [Google Scholar] [CrossRef] [PubMed]
  11. Brotherton, A.L. Metrics of Antimicrobial Stewardship Programs. Med. Clin. N. Am. 2018, 102, 965–976. [Google Scholar] [CrossRef]
  12. McEwen, S.A.; Collignon, P.J. Antimicrobial Resistance: A One Health Perspective. Microbiol. Spectr. 2018, 6, 6.2.10. [Google Scholar] [CrossRef]
  13. Peiffer-Smadja, N.; Rawson, T.M.; Ahmad, R.; Buchard, A.; Georgiou, P.; Lescure, F.-X.; Birgand, G.; Holmes, A.H. Machine Learning for Clinical Decision Support in Infectious Diseases: A Narrative Review of Current Applications. Clin. Microbiol. Infect. 2020, 26, 584–595. [Google Scholar] [CrossRef] [PubMed]
  14. Weis, C.V.; Jutzeler, C.R.; Borgwardt, K. Machine Learning for Microbial Identification and Antimicrobial Susceptibility Testing on MALDI-TOF Mass Spectra: A Systematic Review. Clin. Microbiol. Infect. 2020, 26, 1310–1317. [Google Scholar] [CrossRef]
  15. McCoy, A.; Das, R. Reducing Patient Mortality, Length of Stay and Readmissions through Machine Learning-Based Sepsis Prediction in the Emergency Department, Intensive Care Unit and Hospital Floor Units. BMJ Open Qual. 2017, 6, e000158. [Google Scholar] [CrossRef]
  16. Calvert, J.S.; Price, D.A.; Chettipally, U.K.; Barton, C.W.; Feldman, M.D.; Hoffman, J.L.; Jay, M.; Das, R. A Computational Approach to Early Sepsis Detection. Comput. Biol. Med. 2016, 74, 69–73. [Google Scholar] [CrossRef]
  17. Saybani, M.R.; Shamshirband, S.; Golzari Hormozi, S.; Wah, T.Y.; Aghabozorgi, S.; Pourhoseingholi, M.A.; Olariu, T. Diagnosing Tuberculosis With a Novel Support Vector Machine-Based Artificial Immune Recognition System. Iran. Red. Crescent Med. J. 2015, 17. [Google Scholar] [CrossRef]
  18. Maiellaro, P.; Cozzolongo, R.; Marino, P. Artificial Neural Networks for the Prediction of Response to Interferon Plus Ribavirin Treatment in Patients with Chronic Hepatitis C. Curr. Pharm. Des. 2004, 10, 2101–2109. [Google Scholar] [CrossRef]
  19. Li, S.; Tang, B.; He, H. An Imbalanced Learning Based MDR-TB Early Warning System. J. Med. Syst. 2016, 40, 164. [Google Scholar] [CrossRef] [PubMed]
  20. Beaudoin, M.; Kabanza, F.; Nault, V.; Valiquette, L. Evaluation of a Machine Learning Capability for a Clinical Decision Support System to Enhance Antimicrobial Stewardship Programs. Artif. Intell. Med. 2016, 68, 29–36. [Google Scholar] [CrossRef] [PubMed]
  21. De Vries, S.; Ten Doesschate, T.; Totté, J.E.E.; Heutz, J.W.; Loeffen, Y.G.T.; Oosterheert, J.J.; Thierens, D.; Boel, E. A Semi-Supervised Decision Support System to Facilitate Antibiotic Stewardship for Urinary Tract Infections. Comput. Biol. Med. 2022, 146, 105621. [Google Scholar] [CrossRef]
  22. Corbin, C.K.; Sung, L.; Chattopadhyay, A.; Noshad, M.; Chang, A.; Deresinksi, S.; Baiocchi, M.; Chen, J.H. Personalized Antibiograms for Machine Learning Driven Antibiotic Selection. Commun. Med. 2022, 2, 38. [Google Scholar] [CrossRef]
  23. Feretzakis, G.; Loupelis, E.; Sakagianni, A.; Kalles, D.; Martsoukou, M.; Lada, M.; Skarmoutsou, N.; Christopoulos, C.; Valakis, K.; Velentza, A.; et al. Using Machine Learning Techniques to Aid Empirical Antibiotic Therapy Decisions in the Intensive Care Unit of a General Hospital in Greece. Antibiotics 2020, 9, 50. [Google Scholar] [CrossRef]
  24. Shi, Z.-Y.; Hon, J.-S.; Cheng, C.-Y.; Chiang, H.-T.; Huang, H.-M. Applying Machine Learning Techniques to the Audit of Antimicrobial Prophylaxis. Appl. Sci. 2022, 12, 2586. [Google Scholar] [CrossRef]
  25. Stracy, M.; Snitser, O.; Yelin, I.; Amer, Y.; Parizade, M.; Katz, R.; Rimler, G.; Wolf, T.; Herzel, E.; Koren, G.; et al. Minimizing Treatment-Induced Emergence of Antibiotic Resistance in Bacterial Infections. Science 2022, 375, 889–894. [Google Scholar] [CrossRef] [PubMed]
  26. Bolton, W.J.; Rawson, T.M.; Hernandez, B.; Wilson, R.; Antcliffe, D.; Georgiou, P.; Holmes, A.H. Machine Learning and Synthetic Outcome Estimation for Individualised Antimicrobial Cessation. Front. Digit. Health 2022, 4, 997219. [Google Scholar] [CrossRef]
  27. Feretzakis, G.; Sakagianni, A.; Loupelis, E.; Kalles, D.; Skarmoutsou, N.; Martsoukou, M.; Christopoulos, C.; Lada, M.; Petropoulou, S.; Velentza, A.; et al. Machine Learning for Antibiotic Resistance Prediction: A Prototype Using Off-the-Shelf Techniques and Entry-Level Data to Guide Empiric Antimicrobial Therapy. Healthc. Inform. Res. 2021, 27, 214–221. [Google Scholar] [CrossRef]
  28. Bystritsky, R.J.; Beltran, A.; Young, A.T.; Wong, A.; Hu, X.; Doernberg, S.B. Machine Learning for the Prediction of Antimicrobial Stewardship Intervention in Hospitalized Patients Receiving Broad-Spectrum Agents. Infect. Control Hosp. Epidemiol. 2020, 41, 1022–1027. [Google Scholar] [CrossRef]
  29. Eickelberg, G.; Sanchez-Pinto, L.N.; Luo, Y. Predictive Modeling of Bacterial Infections and Antibiotic Therapy Needs in Critically Ill Adults. J. Biomed. Inform. 2020, 109, 103540. [Google Scholar] [CrossRef]
  30. Chowdhury, A.S.; Lofgren, E.T.; Moehring, R.W.; Broschat, S.L. Identifying Predictors of Antimicrobial Exposure in Hospitalized Patients Using a Machine Learning Approach. J. Appl. Microbiol. 2020, 128, 688–696. [Google Scholar] [CrossRef]
  31. Moehring, R.W.; Phelan, M.; Lofgren, E.; Nelson, A.; Dodds Ashley, E.; Anderson, D.J.; Goldstein, B.A. Development of a Machine Learning Model Using Electronic Health Record Data to Identify Antibiotic Use Among Hospitalized Patients. JAMA Netw. Open 2021, 4, e213460. [Google Scholar] [CrossRef]
  32. Goodman, K.E.; Heil, E.L.; Claeys, K.C.; Banoub, M.; Bork, J.T. Real-World Antimicrobial Stewardship Experience in a Large Academic Medical Center: Using Statistical and Machine Learning Approaches to Identify Intervention “Hotspots” in an Antibiotic Audit and Feedback Program. Open Forum Infect. Dis. 2022, 9, ofac289. [Google Scholar] [CrossRef]
  33. Mancini, A.; Vito, L.; Marcelli, E.; Piangerelli, M.; De Leone, R.; Pucciarelli, S.; Merelli, E. Machine Learning Models Predicting Multidrug Resistant Urinary Tract Infections Using “DsaaS”. BMC Bioinform. 2020, 21, 347. [Google Scholar] [CrossRef] [PubMed]
  34. Wong, J.G.; Aung, A.-H.; Lian, W.; Lye, D.C.; Ooi, C.-K.; Chow, A. Risk Prediction Models to Guide Antibiotic Prescribing: A Study on Adult Patients with Uncomplicated Upper Respiratory Tract Infections in an Emergency Department. Antimicrob. Resist. Infect. Control 2020, 9, 171. [Google Scholar] [CrossRef] [PubMed]
  35. Kanjilal, S.; Oberst, M.; Boominathan, S.; Zhou, H.; Hooper, D.C.; Sontag, D. A Decision Algorithm to Promote Outpatient Antimicrobial Stewardship for Uncomplicated Urinary Tract Infection. Sci. Transl. Med. 2020, 12, eaay5067. [Google Scholar] [CrossRef]
  36. Oonsivilai, M.; Mo, Y.; Luangasanatip, N.; Lubell, Y.; Miliya, T.; Tan, P.; Loeuk, L.; Turner, P.; Cooper, B.S. Using Machine Learning to Guide Targeted and Locally-Tailored Empiric Antibiotic Prescribing in a Children’s Hospital in Cambodia. Wellcome Open Res. 2018, 3, 131. [Google Scholar] [CrossRef] [PubMed]
  37. Artificial Intelligence to Guide Antibiotic Choice in Recurrent Uti: Is It the Right Way for Improving Antimicrobial Stewardship? UROLUTS. Available online: https://uroluts.uroweb.org/webcast/artificial-intelligence-to-guide-antibiotic-choice-in-recurrent-uti-is-it-the-right-way-for-improving-antimicrobial-stewardship/ (accessed on 22 December 2023).
  38. Tang, R.; Luo, R.; Tang, S.; Song, H.; Chen, X. Machine Learning in Predicting Antimicrobial Resistance: A Systematic Review and Meta-Analysis. Int. J. Antimicrob. Agents 2022, 60, 106684. [Google Scholar] [CrossRef]
  39. Greenhalgh, T.; Wherton, J.; Papoutsi, C.; Lynch, J.; Hughes, G.; A’Court, C.; Hinder, S.; Fahy, N.; Procter, R.; Shaw, S. Beyond Adoption: A New Framework for Theorizing and Evaluating Nonadoption, Abandonment, and Challenges to the Scale-Up, Spread, and Sustainability of Health and Care Technologies. J. Med. Internet Res. 2017, 19, e367. [Google Scholar] [CrossRef]
  40. Liyanage, H.; Liaw, S.-T.; Jonnagaddala, J.; Schreiber, R.; Kuziemsky, C.; Terry, A.L.; De Lusignan, S. Artificial Intelligence in Primary Health Care: Perceptions, Issues, and Challenges: Primary Health Care Informatics Working Group Contribution to the Yearbook of Medical Informatics 2019. Yearb. Med. Inform. 2019, 28, 041–046. [Google Scholar] [CrossRef] [PubMed]
  41. Beil, M.; Proft, I.; Van Heerden, D.; Sviri, S.; Van Heerden, P.V. Ethical Considerations about Artificial Intelligence for Prognostication in Intensive Care. Intensive Care Med. Exp. 2019, 7, 70. [Google Scholar] [CrossRef]
  42. WHO Regulatory Considerations on Artificial Intelligence for Health; World Health Organization: Geneva, Switzerland, 2023; ISBN 978-92-4-007887-1.
  43. Gianfrancesco, M.A.; Tamang, S.; Yazdany, J.; Schmajuk, G. Potential Biases in Machine Learning Algorithms Using Electronic Health Record Data. JAMA Intern. Med. 2018, 178, 1544. [Google Scholar] [CrossRef]
  44. Abdullah, Y.I.; Schuman, J.S.; Shabsigh, R.; Caplan, A.; Al-Aswad, L.A. Ethics of Artificial Intelligence in Medicine and Ophthalmology. Asia-Pac. J. Ophthalmol. 2021, 10, 289–298. [Google Scholar] [CrossRef] [PubMed]
  45. Cochrane Handbook for Systematic Reviews of Interventions, 2nd ed.; Higgins, J. (Ed.) John Wiley & Sons: Hoboken, NJ, USA, 2019. [Google Scholar]
  46. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. Syst. Rev. 2021, 10, 89. [Google Scholar] [CrossRef] [PubMed]
  47. Ouzzani, M.; Hammady, H.; Fedorowicz, Z.; Elmagarmid, A. Rayyan—A Web and Mobile App for Systematic Reviews. Syst. Rev. 2016, 5, 210. [Google Scholar] [CrossRef] [PubMed]
  48. Study Quality Assessment Tools|NHLBI, NIH. Available online: https://www.nhlbi.nih.gov/health-topics/study-quality-assessment-tools (accessed on 30 November 2023).
  49. Wolff, R.F.; Moons, K.G.M.; Riley, R.D.; Whiting, P.F.; Westwood, M.; Collins, G.S.; Reitsma, J.B.; Kleijnen, J.; Mallett, S.; for the PROBAST Group. PROBAST: A Tool to Assess the Risk of Bias and Applicability of Prediction Model Studies. Ann. Intern. Med. 2019, 170, 51. [Google Scholar] [CrossRef]
  50. Moons, K.G.M.; Wolff, R.F.; Riley, R.D.; Whiting, P.F.; Westwood, M.; Collins, G.S.; Reitsma, J.B.; Kleijnen, J.; Mallett, S. PROBAST: A Tool to Assess Risk of Bias and Applicability of Prediction Model Studies: Explanation and Elaboration. Ann. Intern. Med. 2019, 170, W1. [Google Scholar] [CrossRef]
Figure 1. PRISMA flowchart representing the systematic search of the relevant studies.
Figure 1. PRISMA flowchart representing the systematic search of the relevant studies.
Antibiotics 13 00307 g001
Figure 2. Frequency of the most used AI algorithms.
Figure 2. Frequency of the most used AI algorithms.
Antibiotics 13 00307 g002
Figure 3. Frequency of the most used performance metrics (AUC—area under the curve; PPV—positive predictive value; NPV—negative predictive value; TP—true positive; FP—false positive; FN—false negative; TN—true negative).
Figure 3. Frequency of the most used performance metrics (AUC—area under the curve; PPV—positive predictive value; NPV—negative predictive value; TP—true positive; FP—false positive; FN—false negative; TN—true negative).
Antibiotics 13 00307 g003
Table 1. Characteristics of the included studies.
Table 1. Characteristics of the included studies.
StudyYear of PublicationCountryNo. CentersStudy Time FrameTarget PopulationNo. PatientsInfection SiteNo. FeaturesObjectiveAlgorithmPerformance MeasurementMain Results
[20]2016Canada1February to November 2012Patients monitored by APSS who received at least one prescription of piperacillin–tazobactam at the Centre Hospitalier Universitaire de Sherbrooke421 hospitalizationsNot specifiedNot specifiedTo evaluate the ability of the algorithm to discover rules for identifying inappropriate prescriptions of piperacillin-tazobactamSupervised learning module of APSS, temporal induction of classification models algorithmPPV, sensitivity, accuracy, precisionThe combined system achieved an overall PPV (precision) of identifying confirmed inappropriate prescriptions of 74% (95% CI, 68–79), with sensitivity (recall) of 96% (95% CI, 92–98), and accuracy of 79% (95% CI, 74–83).
[21]2022Netherlands1January 2017–December 2018Inpatients of the UMC Utrecht906 cultures from 810 patientsUTI36To report on the design and evaluation of a CDSS to predict UTI before the urine culture results are availableCDSS using the RESSEL method; supervised models implemented in the Scikit-learn package: LR, SVM, RF, XGB and k-NNAccuracy, sensitivity, specificity, PPV, NPV, AUC, Nneg, NposThe predictive performance of the best-performing semi-supervised model (RF enhanced with RESSEL) had an accuracy of 76.77 (±0.97), sensitivity of 81.28 (±1.16), specificity of 70.75 (±1.85), and AUC of 80.02 (±1.00).
[22]2022USA5Stanford hospitals: January 2009–December 2019; Boston hospitals: 2007–2016Patients who presented to Stanford emergency departments, Massachusetts General Hospital, and Brigham and Women’s Hospital in BostonStanford: N = 8342 infections from 6920 adult patients. Boston: N = 15,806 uncomplicated urinary tract infections from 13,862 unique female patients. Our dataset is split by time into training, validation, and test sets containing Ntrain = 5804 patient infections from 2009 to 2017, Nval = 1218 patient infections from 2018, and Ntest = 1320 patient infections from 2019.Stanford: unspecified infection; Boston: UTIBoston: The total number of features used in this portion of the analysis was 788. Stanford: In total, the sparse feature matrix contained 43,220 columns.To investigate the utility of ML-based clinical decision support for antibiotic prescribing stewardship.LR, RF, gradient boosted tree, lasso, ridgeAUROC, prevalence, average precision, antibiogram coverage rateStanford dataset: personalized antibiograms reallocate clinician antibiotic selections with a coverage rate of 85.9%, similar to clinician performance (84.3% p = 0.11). The best model class for selection of vancomycin+meropenem was gradient boosted tree, with average precision of 0.99 [0.99, 0.99] and AUROC of 0.73 [0.65, 0.81].
Boston dataset: personalized antibiograms coverage rate of 90.4%, a significant improvement over clinicians (88.1% p < 0.0001). The best model class for the selection of levofloxacin was LASSO, with an average precision of 0.96 [0.95, 0.96] and AUROC of 0.64 [0.60, 0.67].
[23]2020Greece1January 2017–December 2018ICU patients in a public tertiary hospital345Invasive, respiratory, urinary, mucocutaneous, and wound infections23,067 (binary, numerical, and categorical in total)To compare the performance of eight ML algorithms to assess antibiotic susceptibility predictionsML toolkit: WEKA—Data Mining Software in Java Workbench; LIBLINEAR LR and linear SVM; SVMs; SMO; instance-based learning (k-NN); J48; RF; RIPPER; MLPTP rate, FP rate, precision, recall, F-measure, mmc, AUROC, precision-recall plotThe best performances were obtained with the RIPPER algorithm (F-measure of 0.678) and the MLP classifier (AUROC of 0.726).
[24]2022Taiwan25May 2013 to May 2014Patients with healthcare-associated infections receiving at least one antimicrobial drug7377Healthcare-associated infection (bloodstream, urinary, pneumonia and surgical site infection).26To develop accurate and efficient ML models for auditing appropriate surgical antimicrobial prophylaxisSupervised ML classifiers (Auto-WEKA (Bayesian optimisation method), MLP (artificial neural network), decision tree, SimpleLogistic (LogitBoost e CART algorithm), bagging, SMOTE and AdaBoost)TP rate, TN rate, FP, FN, AUC, precision, specificity, sensitivity, weighted average for the multiclass model, execution timeThe ML technique with the best performance metrics was the MLP, with a sensitivity of 0.967, specificity of 0.992, precision of 0.967, and AUC of 0.992.
[25]2022Israel1June 2007 to January 2019Patients with UTI and wound infections from Maccabi Healthcare Services (MHS) with at least one record of a positive wound infection culture140,349 UTI and 7365 wound infections.UTI and woundNot specifiedTo understand and predict the personal risk of treatment-induced gain of resistanceMLPersonal predicted riskChoosing the antibiotic treatment with the minimal ML-predicted risk of emergence of resistance reduces the overall risk of emergence of resistance by 70% for UTIs and 74% for wound infections compared to the risk for physician-prescribed treatments.
[26]2022USA12008 to 2019Patients who received intravenous antibiotic treatment for a duration between 1 and 21 days during an ICU stay, at Beth Israel Deaconess Medical Centre, Boston18,988 (22,845 unique stays)Respiratory (pneumonia) and UTI43To estimate patients’ ICU LOS and mortality outcomes for any given day under the alternative scenarios of if they were to stop vs. continue antibiotic treatmentAI-based CDSS: recurrent neural network autoencoder and a synthetic control-based approach. It uses a bidirectional LSTM autoencoder; PyTorch was used to create a bidirectional LSTM RNNPatients’ ICU LOS (days, mean delta, root mean squared error), mortality outcomes, to stop vs. continue ATB treatment (mean days reduction); day(s), mean delta (days, p-value), MAPE, MAE, RMSE, AUROCThe model reliably estimates patient outcomes under the contrasting scenarios of stopping or continuing ATB treatment: impact days where the potential effect of the unobserved scenario was assessed showed that stopping ATB therapy earlier had a statistically significant shorter LOS (mean reduction 2.71 days, p-value < 0.01). No impact on mortality was observed.
[27]2021Greece1January to December 2018Patients admitted to the internal medicine wards of a public hospital499 patients (11,496 instances)Not specified6 (attributes of sex, age, sample type, Gram stain, 44 antimicrobial substances, and the antibiotic susceptibility results)To assess the effectiveness of AutoML-trained models to predict AMRAutoML techniques using Microsoft Azure AutoML; SMOTE; algorithms: StackEnsemble, VotingEnsemble, MaxAbsScaler, LightGBM, SparseNormalizer, XGBoostClassifierAUROC, AUCW, APSW, F1W, and ACCThe stack ensemble technique achieved the best results in the original and balanced dataset, with an AUCW metric of 0.822 and 0.850, respectively.
[28]2020EUA1December 2015 to August 2017Patients hospitalized who received at least one antimicrobial from a list of those routinely tracked by the ASP at University of California, San Francisco Medical Centre9651Bloodstream, UTI, etc.More than 200To predict whether antibiotic therapy required stewardship intervention on any given day compared to the criterion standard of note left by the antimicrobial stewardship team in the patient’s chartLR and boosted tree modelsAUROC, Brier score, sensitivity, specificity, PPV, and NPVLogistic regression and boosted tree models had AUROCs of 0.73 (95% CI, 0.69–0.77) and 0.75 (95% CI, 0.72–0.79) (p = 0.07), respectively.
[29]2020Israel12001 to 2012ICU adult patients are patients suspected of having a community-acquired bacterial infection10,290 patients (12,232 ICU encounters)Non-specified bacterial infectionNot specifiedTo identify ICU patients with low risk of bacterial infection as candidates for earlier EAT discontinuationML algorithms, including ridge regression, RF, SVC, XG Boost, K- NN, and MLPAUROC, NPV, F1, precision, recall, high sensitivity threshold, TN, FP, FN, TPUsing structured longitudinal data collected up to 24, 48, and 72 h after starting EAT, the best models identified patients at low risk of bacterial infections with AUROCs up to 0.8 and negative predictive values > 93%. The T = 24 h RF model was the best performing model within this timepoint: AUC of 0.774, F1 of 0.424, NPV of 0.944, precision of 0.277, recall of 0.905, high sensitivity threshold of 0.258.
[30]2019USA27October 2015 to September 2017Patients from the Duke Antimicrobial Stewardship Outreach Network (DASON) (Duke University School of Medicine)382,943Not specifiedMore than 100 features, including demographic data, length of stay, comorbidity, etc.To identify patient- and facility-level predictors of antimicrobial usage in hospitalized patients using an ML approach, which can be used to inform a risk adjustment model to facilitate assessment of antimicrobial utilizationSVR and CB modelsRoot-mean-square error valuesBoth the SVR and CB models show better predictive accuracy than the null LM and null NB-GLM models (null statistical models) for all SAAR (external comparator) groups. CB performed better than SVR, according to the RMSE values (5.51 vs. 7.17 for all antibiotics, respectively).
[31]2021USA3October 2015 to September 2017Adult and pediatric inpatient from Duke University Health System170,294Not specified204To evaluate whether variables derived from the electronic health records accurately identify inpatient antimicrobial useA 2-stage RF ML modelingAUROC and absolute errorModels accurately identified antimicrobial exposure in the testing dataset: the majority of AUCs were above 0.8, with a mean AUC of 0.85.
[32]2022USA1July 2017 to December 2019Patient with antimicrobial orders from University of Maryland Medical Centre17,503Sepsis/bacteremia, bone/joint, central nervous system, cardiac/vascular, gastrointestinal genitourinary, respiratory, nonsurgical prophylaxis, skin and soft tissue infection, mycobacterial infection, neutropenia, surgical prophylaxis33To understand which patient and treatment characteristics are associated with either a higher or lower likelihood of intervention in a PAF program and to develop prediction models to identify antimicrobial orders that may be safely excluded from the reviewLR, RFSensitivity, specificity, C-statistic, the out-of-bag error rateThe RF model had a C-statistic of 0.76 (95% CI, 0.75–0.77), with a sensitivity and specificity of 78% and 58%, respectively. This model would reduce review caseloads by 49%.
[33]2018Italy1March 2012 to 2019Patients with nosocomial (UTI) from Principe di Piemonte Hospital in Senigallia1486UTI6 (5 predictors + MDR resistance)To design, develop, and evaluate, with a real antibiotic stewardship dataset, a predictive model useful for predicting MDR UTI onset after patient hospitalizationCatboost, support vector machine, and NNAccuracy, AUROC, AUC-PRC, F1 score, sensitivity, specificity, MCC. FP, FN, TP, and TNThe ML method catboost had the best predictive results (MCC of 0.909; sensitivity of 0.904; F1 score of 0.809; AUC-PRC of 0.853, AUROC of 0.739; ACC of 0.717).
[34]2020Singapore1June 2016 to November 2018Patients with uncomplicated URTI at the emergency department at Tan Tock Seng Hospital715Upper respiratory tract infections50 (univariate analysis), 8 included in the algorithmTo develop prediction models based on local clinical and laboratory data to guide antibiotic prescribing for adult patients with uncomplicated upper respiratory tract infectionsLR models, LASSO, and CARTAUC, sensitivity, specificity, PPV, NPVThe AUC on the validation set for the models was similar: LASSO: 0.70 [95% CI: 0.62–0.77], LR: 0.72 [95% CI: 0.65–0.79], decision tree: 0.67 [95% CI: 0.59–0.74].
[35]2020USA22007 to 2016Patients presenting with uncomplicated UTI at Massachusetts General Hospital and the Brigham and Women’s Hospital in Boston10,053 (training dataset); 3629 (test set)UTI8To predict antibiotic susceptibility using electronic health record data and build a decision algorithm for recommending the narrowest possible antibiotic to which a specimen is susceptibleLR, decision tree, and RF modelsAUROC, FN ratesDecision trees and RF were excluded based on their poor validation set performance and relative lack of interpretability. The LR model provided antibiotic stewardship for a common infectious syndrome by maximizing reductions in broad-spectrum antibiotic use while maintaining optimal treatment outcomes. The algorithm achieved a 67% reduction in the use of second-line antibiotics relative to clinicians and reduced inappropriate antibiotic therapy by 18%, close to the rate of clinicians.
[36]2019Cambodia1February 2013 to January 2016Children with at least one positive blood culture from Angkor Hospital for Children195 (training set); 48 (model validation)Bloodstream35To predict Gram stains and whether bacterial pathogens could be treated with standard empiric antibiotic regimensRF, LR, decision trees constructed via recursive partitioning, boosted decision trees using adaptive boosting, linear SVM, polynomial SVM, radial SVM, and k-NNAUROCThe RF method had the best predictive performance overall: AUC of 0.80 (95% CI 0.66–0.94) for predicting susceptibility to ceftriaxone, 0.74 (0.59–0.89) for susceptibility to ampicillin and gentamicin, 0.85 (0.70–1.00) for susceptibility to neither, and 0.71 (0.57–0.86) for Gram stain result.
[37]2022Italy2January 2012 to December 2020Women affected by recurrent UTI who had undergone antimicrobial treatment for uncomplicated lower UTI1043Recurrent UTINot specifiedTo define an NN for predicting the clinical and microbiological efficacy of antimicrobial treatment of a large cohort of women affected by recurrent UTIs for use in everyday clinical practiceNNSensitivity, specificity, HRThe use of artificial NN in women with recurrent cystitis showed a sensitivity of 87.8% and specificity of 97.3% in predicting the clinical and microbiological efficacy of the prescribed antimicrobial treatment.
Notes: APSS—antimicrobial prescription surveillance system; ICU—intensive care unit; CI—confidence interval; CDSS—clinical decision support system; UTI—urinary tract infection; RESSEL—reliable semi-supervised ensemble learning; RF—random forest; PPV—positive predictor value; NPV—negative predictive value; AUC—area under the curve; AUROC—area under the ROC curve; Nneg—the number of UTI-negative labeled cultures; Npos—the number of UTI-positive labelled cultures; N—number; Ntrain—number in training set; Nval—number in validation set; Ntest—number in test set; TP—true positive; TN—true negative; FP—false positive; FN—false negative; Mmc—a correlation coefficient; LOS—length of stay; MAPE—mean absolute; MAE—mean absolute error; RMSE—root-mean-squared error; AUCW—area under the curve-weighted; APSW—average precision score-weighted; F1W—F1 score-weighted; ACC—accuracy; ML—machine learning; LR—logistic regression; SVM—support vector machine; XGB—eXtreme Gradient Boosting; NN—nearest neighbors; SMO—sequential minimal optimization; MLP—multilayer perceptron; LSTM—long short-term memory; RNN—recurrent neural network; ATB—antibiotic; AutoML—automated machine learning; SMOTE—synthetic minority oversampling technique; EAT—empiric antibiotic therapy; SVC—support vector classifier; SVR—support vector regression; CB—cubist regression; LM—linear model; null NB-GLM model—negative binomial generalized linear model; SAAR—standardized antimicrobial administration ratio; PAF—prospective audit with feedback; MDR—multidrug resistant; AUC-PRC—area under precision recall curve; MCC—Matthews correlation coefficient; HR—hazard ratio. Scikit-learn package—available at https://scikit-learn.org/stable/ (accessed on 24 March 2024). WEKA—Data Mining Software—WEKA 3.6. Auto-WEKA—2.0. Pytorch—https://pytorch.org/ (accessed on 24 March 2024). Microsoft Azure AutoML—https://learn.microsoft.com/en-us/azure/?product=popular (accessed on 24 March 2024). SMOTE—https://learn.microsoft.com/en-us/azure/machine-learning/component-reference/smote?view=azureml-api-2 (accessed on 24 March 2024).
Table 2. Characteristics of the features of the included studies.
Table 2. Characteristics of the features of the included studies.
StudyDemographicsAdultPaediatricClinicalLaboratory/MicrobiologicalComorbiditiesType of InfectionICU
[20]YesYesNoYesYesNoNoYes
[21]YesYesYesYesYesYesYesNo
[22]YesYesNoYesYesYesYesNo
[23]YesYesNoNoYesNoYesYes
[24]YesYesYesYesNoNoYesYes
[25]YesYesNoYesYesYesYesNot specified
[26]YesYesNot specifiedYesYesNoNoYes (only ICU patients)
[27]YesYesNoNoYesNoYesNo
[28]YesYesNoYesYesNoYesYes
[29]YesYesNoYesYesYesYesYes (only ICU patients)
[30]YesYesNoYesYesYesYesNot specified
[31]YesYesYesYesYesYesNot specifiedYes
[32]YesYesNoYesYesNoYesNot specified
[33]YesYesNot specifiedNoYesNoYesNot specified
[34]YesYesNoYesYesYesYesNo
[35]YesYesNoYesYesYesYesYes
[36]YesNoYesYesYesNoYesYes
[37]YesYesNoYesYesNot specifiedYesNot specified
Note: ICU—intensive care unit.
Table 3. Risk of bias assessment of the included studies by NIH Quality Assessment Tool for Observational Cohort and Cross-Sectional Studies.
Table 3. Risk of bias assessment of the included studies by NIH Quality Assessment Tool for Observational Cohort and Cross-Sectional Studies.
Criteria\
Study
1234567891011121314Quality Rating
[20]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[21]YesYesNAYesNRYesYesNAYesYesYesNRNANoFair (57.1%)
[22]YesYesNAYesNRYesYesNAYesYesYesNRNAYesFair (64.3%)
[23]YesYesNAYesNRYesYesNAYesYesYesNRNANoFair (57.1%)
[24]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[25]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[26]YesYesNAYesNRYesYesNAYesYesYesNRNAYesFair (64.3%)
[27]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[28]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[29]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[30]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[31]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[32]YesYesNAYesNRYesYesNAYesYesYesNRNAYesFair (64.3%)
[33]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[34]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[35]YesYesNAYesYesYesYesNAYesYesYesNRNANoFair (64.3%)
[36]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
[37]YesYesNAYesNRYesYesNAYesYesYesNRNANRFair (57.1%)
Notes: NA—not applicable; NR—not reported.
Table 4. Risk of bias and applicability assessment by PROBAST.
Table 4. Risk of bias and applicability assessment by PROBAST.
StudyRisk of BiasApplicabilityOverall
1.
Participants
2.
Predictors
3.
Outcome
4.
Analysis
1.
Participants
2.
Predictors
3.
Outcome
Risk of BiasApplicability
[20]-++???+-?
[21]+?+?+++?+
[22]+?+++++?+
[23]+++?+++?+
[24]+++++++++
[25]??+???+??
[26]+++?+++?+
[27]-++--++--
[28]+++-+?+-?
[29]+?+?+?+??
[30]+++-+?+-?
[31]+++-++?-?
[32]+++-+++-+
[33]+++-+?+-?
[34]+++-+++-+
[35]++?-++?-?
[36]++?-++?-?
[37]+++-++?-?
Notes: + low concern, - high concern, ? unclear.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pinto-de-Sá, R.; Sousa-Pinto, B.; Costa-de-Oliveira, S. Brave New World of Artificial Intelligence: Its Use in Antimicrobial Stewardship—A Systematic Review. Antibiotics 2024, 13, 307. https://doi.org/10.3390/antibiotics13040307

AMA Style

Pinto-de-Sá R, Sousa-Pinto B, Costa-de-Oliveira S. Brave New World of Artificial Intelligence: Its Use in Antimicrobial Stewardship—A Systematic Review. Antibiotics. 2024; 13(4):307. https://doi.org/10.3390/antibiotics13040307

Chicago/Turabian Style

Pinto-de-Sá, Rafaela, Bernardo Sousa-Pinto, and Sofia Costa-de-Oliveira. 2024. "Brave New World of Artificial Intelligence: Its Use in Antimicrobial Stewardship—A Systematic Review" Antibiotics 13, no. 4: 307. https://doi.org/10.3390/antibiotics13040307

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop