Next Article in Journal
Post-Bariatric Surgery Abdominoplasty Ameliorates Psychological Well-Being in Formerly Obese Patients: A Cross-Sectional Recall Study
Previous Article in Journal
Regenerative Surgery of the Nonunion of Metacarpals and Phalanges: Amniotic Membrane and Bone Micro-Grafts as a Novel Treatment Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Machine Learning and Artificial Intelligence in Intensive Care Medicine: Critical Recalibrations from Rule-Based Systems to Frontier Models

1
Sciense, New York, NY 10027, USA
2
Pediatric Intensive Care Unit, Cliniques Universitaires Saint-Luc, 1200 Bruxelles, Belgium
3
Department of Intensive Care, Hôpital Universitaire de Bruxelles (HUB), Université Libre de Bruxelles (ULB), Anderlecht, 1070 Brussels, Belgium
4
Department of Neurosurgery, CHR Citadelle, 4000 Liège, Belgium
5
Department of Clinical Neurosciences, University of Cambridge, Cambridge CB2 1TN, UK
6
Bioelectronics Laboratory, Department of Electrical Engineering, University of Cambridge, Cambridge CB2 1TN, UK
7
Institut Neuroscience des Systèmes (INS), Aix-Marseille Université, 13005 Marseille, France
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2025, 14(12), 4026; https://doi.org/10.3390/jcm14124026
Submission received: 28 April 2025 / Revised: 31 May 2025 / Accepted: 4 June 2025 / Published: 6 June 2025
(This article belongs to the Section Intensive Care)

Abstract

Artificial intelligence (AI) and machine learning (ML) are rapidly transforming clinical decision support systems (CDSSs) in intensive care units (ICUs), where vast amounts of real-time data present both an opportunity and a challenge for timely clinical decision-making. Here, we trace the evolution of machine intelligence in critical care. This technology has been applied across key ICU domains such as early warning systems, sepsis management, mechanical ventilation, and diagnostic support. We highlight a transition from rule-based systems to more sophisticated machine learning approaches, including emerging frontier models. While these tools demonstrate strong potential to improve predictive performance and workflow efficiency, their implementation remains constrained by concerns around transparency, workflow integration, bias, and regulatory challenges. Ensuring the safe, effective, and ethical use of AI in intensive care will depend on validated, human-centered systems supported by transdisciplinary collaboration, technological literacy, prospective evaluation, and continuous monitoring.

1. Introduction

The intensive care unit (ICU) represents both an ideal testing ground and a significant challenge for artificial intelligence (AI) applications in medicine. ICUs generate vast quantities of high-dimensional data, from continuous physiological monitoring, frequent laboratory testing, imaging studies, and detailed clinical documentation, which creates both an opportunity and a necessity for advanced computational approaches to data analysis and decision support [1]. The complexity of critically ill patients, characterized by multisystem disease, rapid clinical changes, and high-stakes decision-making under significant time constraints, necessitates the expertise of highly specialized clinicians who must continuously update their knowledge and skills to effectively integrate evolving technologies and maintain optimal patient care standards [2].
Clinical decision support systems (CDSSs) aim to augment human capabilities by providing actionable insights at the point of care [3]. While traditional CDSSs have often relied on rule-based logic derived from established guidelines or expert consensus, the advent of machine learning (ML) techniques has enabled a paradigm shift toward data-driven approaches capable of detecting complex patterns and subtle signals that might elude human perception [4,5]. This evolution in CDSS technology aligns with the growing recognition that conventional scoring systems in critical care have limitations in capturing the dynamic nature of critical illness and personalizing risk assessment [6,7].
Although AI tools have shown promise in many areas of medicine, their use in ICU settings remains limited and inconsistent. Key barriers include a lack of model transparency, unclear regulatory pathways, and limited clinician trust [8]. These challenges underscore the urgent need for stronger clinical evidence, rigorous validation studies, and ethical design principles to guide the safe and effective integration of AI into critical care practice [9].
This review examines the integration of machine intelligence in intensive care medicine, from early rule-based systems to current applications leveraging advanced machine learning and natural language processing for AI-driven CDSS. We evaluate the evidence supporting various AI/ML modalities across specific domains of critical care practice, including early warning systems, sepsis prediction and management, mechanical ventilation optimization, and clinical documentation. Implementation challenges, spanning technical integration, clinical workflow adaptation, ethical considerations, and regulatory frameworks, are critically analyzed. Finally, we discuss future directions that may influence the trajectory of AI in critical care, stressing the need for transparent, clinically validated systems that complement rather than supplant clinical judgment.

2. Historical Evolution of AI in Critical Care

Machine intelligence in biomedicine emerged during the 1960s with seminal efforts to develop expert systems emulating human expert decision-making. In the 1970s, such systems were applied in critical care with initiatives such as MYCIN, developed to guide antimicrobial therapy for severe infections (Figure 1 and Table 1) [10]. These systems encoded medical knowledge through explicitly programmed IF–THEN rules, allowing the computer to mimic clinical reasoning within narrow domains. Despite showing promise in research settings, these early systems faced practical limitations in handling the complexity and uncertainty inherent in critical care, and few achieved widespread clinical implementation [11].
The transition from rule-based to data-driven approaches began in the 1990s, with the development of ICU-specific prognostic models using statistical methods, such as APACHE (Acute Physiology and Chronic Health Evaluation) and SAPS (Simplified Acute Physiology Score) [34]. While groundbreaking, these models relied on relatively simple logistic regression techniques and required periodic manual recalibration to maintain accuracy across changing patient populations [24].
The modern era of AI in critical care emerged in the 2010s with the convergence of several key factors: (I) widespread adoption of electronic health records (EHRs) creating large, structured clinical datasets; (II) advances in computational power enabling complex model training; (III) development of more sophisticated machine learning algorithms capable of handling high-dimensional data with non-linear relationships; and (IV) the creation of publicly available critical care databases such as the Medical Information Mart for Intensive Care (MIMIC), which fostered open research and benchmarking [35,36].

3. AI/ML Paradigms and Technologies

Contemporary AI in ICU employs distinct machine learning paradigms, each aligned with specific clinical changes (Figure 2 and Table 2).

3.1. Supervised Learning

Supervised learning models use labeled patient data that include patient features (such as vital signs or lab test results) along with known clinical outcomes to train predictive models [42]. It is the most common AI method in critical care. This approach includes many ICU decision-support tools that learn from large datasets of ICU patient records. These tools can estimate a patient’s mortality risk, predict organ failure, or generate early warnings of deterioration [43]. Modern supervised methods, such as ensemble classifiers and deep neural networks, often outperform traditional scoring systems in capturing complex nonlinear relationships in the data. For example, a supervised sepsis early-warning system that can alert clinicians hours before a patient meets clinical sepsis criteria, effectively providing a critical lead-time for intervention [44].
That said, supervised models are only as good as their training data. They require extensive, high-quality labeled datasets and can inherit biases present in those data. If certain patient groups or outcomes are underrepresented or recorded with bias, the model’s predictions may be skewed or less generalizable [45,46]. Thus, careful dataset curation, bias correction, and continuous performance monitoring are essential when deploying supervised learning in intensive care. These measures help ensure that the model’s recommendations remain accurate and fair across diverse patient populations and clinical scenarios.

3.2. Unsupervised Learning

Unsupervised learning methodologies focus on finding hidden structure in unlabelled datasets to identify latent patterns, offering valuable insights by uncovering novel patient phenotypes. The goal is to let the data themselves reveal natural clusters of patients or important variables, rather than relying on prior assumptions. This approach is especially valuable for complex conditions such as sepsis and acute respiratory distress syndrome (ARDS) [29,30].
Key unsupervised techniques in critical care include the following:
  • Clustering algorithms (e.g., k-means, hierarchical clustering): These group patients with similar clinical profiles or characteristics.
  • Dimensionality reduction techniques (e.g., principal component analysis (PCA) or deep autoencoders), which compress high-dimensional data (such as hundreds of lab values and vital signs) into a smaller set of informative features, reducing complexity while preserving important information.
Recent research has applied unsupervised learning methods to uncover meaningful patient subgroups within ICU populations, including the following:
  • Sepsis Phenotypes: Using clustering analyses on sepsis patient data, researchers have uncovered distinct subgroups of sepsis patients that were not previously recognized. These data-driven clusters differ significantly in their clinical characteristics and outcomes [29].
  • ARDS Subphenotypes: unsupervised analyses have identified at least two consistent subphenotypes, often described informally as “hyperinflammatory” and “hypoinflammatory”. Each of these subgroups has its own pattern of biomarker levels, its own severity of illness, and a different response to treatments [38].
In summary, unsupervised learning in intensive care is a powerful tool for knowledge discovery. By uncovering latent patterns in complex ICU data, it can generate new hypotheses, such as novel disease subtypes, enabling more personalized approaches to patient management and deepening our understanding of complex disease beyond what traditional analyses can achieve [29,30,38].

3.3. Reinforcement Learning

Reinforcement learning (RL) is an approach to AI where an agent learns by interacting with its environment. The agent makes a sequence of decisions and improves its strategy based on feedback (rewards or penalties). Unlike supervised learning that relies on fixed labeled examples, RL continuously refines its decisions through trial and error. This approach is particularly well-suited to the ICU setting, where clinicians must continually adjust treatments as patient conditions change [47,48].
RL algorithms have been applied to tasks like drug titration and ventilator management. The goal in these applications is to discover an optimal treatment strategy that maximizes patient benefit throughout an ICU stay [49,50].
Classical RL algorithms such as Q-learning estimate the value of each action in a given clinical state, for example, evaluating the long-term benefit of giving an extra dose of a vasopressor to a hypotensive patient. More advanced deep RL methods (such as Deep Deterministic Policy Gradient (DDPG) or Proximal Policy Optimization (PPO)) use neural networks to handle complex, continuous action spaces (for example, continuously adjusting an infusion rate), effectively approximating the optimal policy [49,50]. In recent studies, researchers applied RL to mechanical ventilation by training an agent to adjust ventilator settings. The agent learned to prioritize lung-protection strategies more consistently than standard protocols. In sedation management, researchers developed an RL algorithm to dynamically recommend doses of the sedative dexmedetomidine, aiming to minimize delirium risk. The “AI Clinician” exemplifies this approach, utilizing RL to derive optimal fluid and vasopressor strategies from retrospective ICU data, demonstrating promise in managing pathological conditions in which fluid therapy is relevant, such as sepsis [27].
Practical translation to bedside implementation remains constrained by computational complexity, safety concerns, and integration challenges [39]. Current research increasingly focuses on Human-in-the-Loop RL, whereby clinicians interact with the learning system during training or inference phases. This approach enables models to incorporate expert feedback, enhance interpretability, and ensure safety in high-stakes environments, such as critical care [51,52].

3.4. Data Infrastructure and Computational Frameworks

The implementation of these AI paradigms in clinical practice relies on robust data infrastructure and computational frameworks. Comprehensive data repositories, including the MIMIC databases, such as the MIMIC-II and MIMIC-III databases, have been instrumental in advancing ICU AI research by providing extensive, detailed patient records that support robust model training and validation [23]. These repositories contain rich, multimodal data that capture the complexity of critical care, including physiological time series, laboratory results, medication administrations, and clinical documentation. Building upon these data foundations, various machine learning frameworks have been used to address the unique challenges of critical care data. Traditional algorithms, such as random forests and gradient boosting machines, effectively handle structured clinical data, while deep learning architectures—particularly convolutional neural networks (CNNs) and recurrent neural networks (RNNs)—excel at analyzing high-dimensional and temporal ICU data. These advanced architectures have demonstrated promising performance in modelling the complex temporal dynamics characteristic of critical illness progression [23,26,53]. As the field matures, privacy-preserving methods, such as federated learning, have emerged to address the ethical and regulatory challenges associated with healthcare data utilization. These approaches enable collaborative model training across multiple institutions while maintaining patient confidentiality, facilitating the development of generalized AI models across diverse populations without requiring centralized data aggregation [12]. This technological advancement represents a crucial step toward responsible AI deployment in healthcare settings.
Despite the strengths of open ICU datasets and federated learning, significant data infrastructure challenges continue to limit the real-time deployment of AI in clinical settings. Public ICU datasets (such as MIMIC) are constrained by their single-center scope and outdated information. For instance, MIMIC-IV data ends in 2019, before the COVID-19 pandemic [45,54]. In practice, many hospitals lack robust pipelines for streaming high-frequency ICU data (e.g., continuous vital signs and waveforms) into analytics engines. This issue is further compounded by limited on-site computational resources, such as the absence of GPU servers and unreliable network connectivity, which restrict the deployment of complex models at the bedside [55].

4. AI Applications in ICUs: Implementation and Challenges

AI technologies are increasingly being deployed across various ICU domains, from early warning systems and sepsis prediction to ventilator management and diagnostic support. These applications demonstrate the growing impact of machine intelligence in frontline care, each with distinct implementation approaches and challenges. Table 3 summarizes key performance benchmarks, documented benefits, and operational considerations across these domains.

4.1. Early Warning Systems

Early detection of clinical deterioration represents one of the most mature applications of AI in critical care. Traditional early warning score systems, such as the Modified Early Warning Score (MEWS) and the National Early Warning Score (NEWS), rely on simple aggregation of abnormal vital signs. However, these traditional scores reflect the patient’s current risk, without accounting for potential risk trajectories. By contrast, AI-based approaches can incorporate complex interactions between variables, temporal trends, and subtle patterns preceding acute events, on a beat-to-beat basis [57]. Deep learning early warning systems show a substantially higher sensitivity than conventional methods at equivalent specificity, with some studies reporting up to a 250% relative increase [56].
Numerous ML-based early warning systems have demonstrated superior performance compared to traditional scoring methods, with few exceptions [68]. For example, the Electronic Cardiac Arrest Risk Triage (eCART) system was developed using gradient-boosted machines on data from more than 250,000 hospitalizations, significantly outperformed MEWS in predicting cardiac arrest and ICU transfer (area under the curve, AUC of 0.85 vs. 0.70) [58]. When implemented as part of a rapid response team activation system, eCART was associated with a 16% reduction in cardiac arrests [37].
More sophisticated approaches using deep learning on continuous monitoring data have pushed prediction horizons further. Hyland et al. developed a recurrent neural network model analyzing minute-by-minute vital signs that predicted clinical interventions with an AUC of 0.90 up to 6 h before the event [31]. These systems increasingly incorporate strategies to balance sensitivity with specificity and to provide explanations for their predictions to support clinical adoption [69].
Beyond ICU-specific models, recent work in general medical risk prediction has introduced methodological innovations that could be adapted to critical care. For example, HeartEnsembleNet integrates multiple base classifiers with the SHapley Additive exPlanations (SHAP)-based interpretability, achieving state-of-the-art performance for cardiovascular-risk assessment while retaining model explainability [70]. Similarly, a multi-layer perceptron trained with Adadelta, RMSProp, and AdaMax optimizers reached approximately 95% accuracy in early stroke detection tasks [71]. While these systems were developed outside ICU settings, they illustrate transferable strategies that may further enhance early-warning performance when tailored to the dynamic and high-noise environment of the ICU.

4.2. Sepsis Care

Sepsis represents an important target for AI applications, given its prevalence, mortality impact, and the established benefit of early intervention. Several machine learning models have been developed for early sepsis detection, with reported promising lead times of 4–12 h before clinical recognition [72].
The Targeted Real-time Early Warning System (TREWS) exemplifies the potential of AI for sepsis care. This random forest-based model uses EHR data to identify patients at risk of sepsis who have undergone rigorous prospective evaluation [73]. In a multi-site study involving over 590,000 patient encounters, Adams et al. demonstrated that when clinicians acted on TREWS alerts within 3 h, in-hospital mortality was reduced by 3.3% absolute percentage points compared to unaddressed alerts [32]. This study provides rare prospective evidence that an AI-driven alert, when integrated into clinical workflow with appropriate response protocols, can meaningfully impact patients’ outcomes.
However, implementation experience has been mixed. The widely deployed Epic Sepsis Model (ESM) has faced criticism after independent validation revealed poor performance (sensitivity 33%, positive predictive value 12%) [59]. This illustrates the importance of transparent, peer-reviewed validation before the widespread implementation of AI tools in critical care.
Beyond detection, AI approaches to sepsis management have also emerged. The “AI Clinician” by Komorowski et al. used reinforcement learning to suggest individualized fluid and vasopressor dosing strategies [27]. Retrospective analysis suggested that when actual clinical decisions matched the AI’s recommendations, mortality was lower. While interesting, these findings remain to be validated in prospective interventional studies.
The TREWS sepsis alert system at Johns Hopkins continuously monitors vital signs, laboratory results, and clinical notes to detect sepsis. Compared with conventional methods, TREWS detected sepsis 6 h earlier and reduced time to first antibiotic by 1.85 h (median). In a two-year study of over 500,000 patient encounters, this was associated with a 20% relative reduction in sepsis mortality. The system provided alert explanations and adapted to unit workflows, leading to adoption by thousands of providers in routine care [32,74,75].

4.3. Ventilation Management

Mechanical ventilation represents another promising domain for AI applications, given the complexity of ventilator management and the potential impact of optimized settings on patient outcomes. AI-guided systems, including fully automated closed-loop devices and decision support tools, optimize mechanical ventilation by adjusting key respiratory parameters and reducing intervention frequency.
Reinforcement learning algorithms have been developed to recommend personalized ventilator settings based on patient characteristics and physiological responses. Peine et al. demonstrated that a reinforcement learning-based approach could recommend positive end-expiratory pressure (PEEP) levels that achieved better oxygenation, while minimizing harmful high plateau pressures compared to standard care in a retrospective analysis [39]. Similar models have been developed for weaning protocols, where ML algorithms incorporate multiple parameters to predict successful extubation with greater accuracy than traditional criteria [76].
Deep learning models analyzing ventilator waveforms have demonstrated high accuracy (>90%) in automatically detecting various forms of asynchrony, potentially enabling real-time monitoring and adjustment [61]. Gholami et al. used random forest classifiers to identify patient–ventilator cycling asynchrony with accuracy comparable to expert clinicians but offering continuous monitoring capability [60].
AI systems have also been implemented to monitor adherence to lung-protective ventilation strategies and suggest adjustments when patients deviate from recommended parameters. For instance, fully automated ventilator systems may reduce driving pressure and mechanical power while improving lung compliance [77]. These systems analyze not just current settings but also trends in respiratory mechanics to customize recommendations based on individual patient characteristics [78].
While most of these applications remain in research settings or limited deployments, full closed-loop systems face both technical and regulatory hurdles, making human-in-the-loop approaches more likely in the near term [79].
At Chi Mei Medical Center in Tainan City, Taiwan, an AI decision support system assisted ICU mechanical ventilation weaning using a two-stage machine learning model to identify optimal extubation timing. Patients managed with AI guidance spent 21 fewer hours on mechanical ventilation compared to standard care. The bedside dashboard tool achieved high staff acceptance, with respiratory therapists and intensivists reporting confidence in the AI’s weaning recommendations. This approach reduced ventilator-associated complications and eased ICU resource utilization. Initial clinician reluctance was addressed through transparent model design and frontline staff involvement in implementation [80].

4.4. Diagnostic Support

Machine learning algorithms yield promising diagnostic performance for complex cases and for interpreting diagnostic results. Deep learning models have reached expert-level performance in interpreting standard ICU imaging studies. For chest radiographs, CNNs can detect pneumothorax, misplaced tubes and lines, pulmonary oedema, and consolidation, with high accuracy [40]. These tools can prioritize critical findings for radiologist review or provide immediate feedback when specialist interpretation is unavailable.
Beyond standard vital sign monitoring, AI systems can extract additional diagnostic information from continuous waveforms. Advanced analysis of electrocardiograms can detect subtle signs of ischemia or electrolyte abnormalities, while arterial waveform analysis can provide insights into fluid responsiveness and cardiac function [81].
Large language models (LLMs) and knowledge-based systems can generate differential diagnoses based on clinical data, potentially identifying rare conditions or unusual presentations. Chen et al. [82] demonstrated that an LLM fine-tuned on clinical cases could generate appropriate differential diagnoses for complex ICU scenarios with performance comparable to junior critical care physicians. Similarly, in a study by McDuff et al., a large language model optimized for diagnostic reasoning outperformed unassisted clinicians in standalone performance (top 10 accuracy of 59.1% vs. 33.6%) [62].
ML models can identify patterns in laboratory data that suggest specific diagnoses or predict complications. For example, unsupervised learning approaches have been used to identify distinct metabolic phenotypes in critically ill patients that correlate with outcomes and responses to specific interventions [83].
The integration of these capabilities into coherent diagnostic support systems remains an active area of development, with emphasis on explanatory features that allow clinicians to understand the rationale behind AI suggestions.

4.5. Documentation and Workflow

Administrative and documentation burdens represent significant challenges in modern ICU practice. AI applications in this domain aim to reduce cognitive load and time waste, allowing clinicians to focus more on direct patient care.
Natural language processing (NLP) technologies have revolutionized the analysis of unstructured clinical documentation. Transformer-based models such as Bidirectional Encoder Representations from Transformers (BERT) efficiently extract actionable clinical information from extensive physician and nursing notes, converting narrative text into structured, analyzable data [33,83]. This application significantly enhances clinical decision support by incorporating previously inaccessible information embedded in clinical narratives.
NLP systems can generate structured documentation from clinician–patient interactions, team discussions, or even ambient conversations during rounds [84]. Furthermore, automatic speech recognition technologies could reduce documentation burdens by accurately transcribing clinical dictations. Transformer-based models, such as ConstDecoder, enhance transcription quality in acoustically challenging ICU environments, demonstrating practical utility and substantial reduction in documentation errors [64].
AI-powered search and summarizing tools can rapidly retrieve relevant information from voluminous patient records [85]. These can be useful both for research purposes and for in-depth clinical evaluations of patients who have had prolonged hospital stays.
Machine learning models can analyze patterns of ICU workflow to identify inefficiencies and suggest process improvements. AI-based predictive models for ICU length of stay and discharge readiness may improve resource utilization and patient flow [66]. Overall, specific efficiency improvements include a 15% increase in data analysis speed, a 10% faster alert generation, and a 5% improvement in decision-making speed [67].
These applications encounter fewer regulatory barriers compared to those designed for direct clinical decision support, and they may serve as a gateway for AI adoption in ICUs.

4.6. Clinical Decision Support

AI-CDSS aims to guide complex interventions, including antibiotic prescription and fluid management. AI-supported antimicrobial stewardship tools exemplify successful clinical implementation with demonstrable positive impact on patient outcomes [13].
Prognostic scoring and risk stratification systems incorporate AI models into ICU workflows, surpassing traditional scoring systems in both accuracy and adaptability [2,3,24].
These AI-driven tools continuously refine prognostic assessments based on real-time patient data, enabling dynamic risk stratification that adapts to evolving clinical scenarios [6,7,18,24].
Clinical evaluations at Our Lady of the Lake Regional Medical Center implemented IntelliSep, an FDA-cleared AI diagnostic tool for sepsis management. The system analyzes immune cell activation from a single blood sample, generating risk stratification scores within minutes of ICU or emergency department admission. Implementation was associated with a 20% reduction in sepsis mortality and a nearly 2-day decrease in average ICU length of stay for septic patients. Over six months, the hospital performed 1800 fewer blood cultures and reduced nursing labor by nine full days. The tool was integrated into existing clinical dashboards, with staff training on score interpretation [86].
Beyond sepsis management, decision support systems extend to other critical domains, including hemodynamic management, glucose control, and transfusion decisions. These systems aim to implement evidence-based practices while accounting for individual patient variations and contextual factors that might warrant deviation from standard protocols.

4.7. Operational Efficiency

At the operational level, resource optimization and workflow management applications employ AI to predict patient acuity and ICU demands, optimizing resource allocation, including bed utilization, staffing, and patient triage, thereby enhancing operational efficiency across critical care units [61].
These systems demonstrate how AI can simultaneously improve clinical outcomes and operational efficiency, addressing the dual challenges of clinical complexity and resource constraints in modern healthcare systems. By predicting patient flow, length of stay, and required resources, these tools help administrators make data-driven decisions about staffing levels, bed allocations, and patient transfers.
Computer vision (CV) technologies enable continuous patient monitoring via non-invasive sensors [87,88]. Advanced algorithms, such as YOLOv8, quantify patient mobility and room activity, enabling improved prediction of clinical events, including delirium and physiological instability, thereby reducing clinician workload [67].

4.8. Frontier Models

The latest evolution in AI technology relevant to critical care involves foundation models, large-scale neural networks pre-trained on massive datasets that can be fine-tuned for specific downstream tasks. LLMs and clinical adaptations, such as the Medical Pathways Language Model (Med-PaLM), represent this new paradigm [41]. These models have demonstrated remarkable capabilities in medical knowledge representation, clinical reasoning, and natural language understanding that could transform multiple aspects of critical care decision support [89].
Foundation models offer several potential advantages for critical care applications, including (I) transfer learning from general clinical knowledge to specific ICU tasks with relatively small fine-tuning datasets; (II) multimodal integration of text, tabular data, and potentially imaging; (III) more natural interaction through conversational interfaces; and (IV) sophisticated contextual understanding of complex clinical scenarios [90]. However, they also present unique challenges, including concerns about hallucination (generating plausible but factually incorrect information), transparency limitations, and difficulties in real-time deployment within existing EHR infrastructures [91].

4.9. Implementation Challenges

Implementing AI in intensive care units presents complex challenges that span technical, clinical, ethical, and regulatory domains (Table 4). The technical integration of AI systems is complicated by the fragmented nature of ICU infrastructure, where multiple monitoring and information systems often operate in isolation with limited interoperability [92]. This creates substantial barriers to assembling the high-quality, synchronized data streams necessary for robust AI model development and deployment.
As machine learning requires large volumes of high-quality training data, it is essential to ensure that datasets are representative of the target patient populations. Healthcare environments can contain various types of bias and noise that can cause models trained in one hospital setting to perform poorly elsewhere. A prominent concern is alert fatigue, as ICU clinicians already contend with numerous alerts from monitoring systems. False alarms, often caused by missing values due to irregular laboratory sampling, device disconnections, along with artifacts during continuous monitoring, can further mislead algorithms and contribute to work overload [85,96]. The temporal alignment of diverse data elements presents additional complexities.
Effective AI decision support depends on both timely recommendations and user-friendly interface design. Because clinical decisions often occur under intense time pressure, AI systems must provide recommendations at the right moment and present information in a way that is both comprehensive and cognitively accessible. Achieving this requires a nuanced understanding of clinical processes, decision points, and user interface design tailored to urgent care settings [97,98].
Many ML models are considered “black boxes”, generating predictions without offering clear explanations. This lack of transparency poses a significant challenge in critical care settings, where clinicians must take full responsibility for decisions that directly affect patient outcomes [99]. Researchers have explored methods to explain ML models, including SHAP and LIME, which generate interpretations of model behavior [100]; however, these explanations are often unreliable or even misleading. Others advocate for inherently interpretable models, though these may sacrifice some predictive accuracy and require considerable effort to build [101].
Contrastive explanations represent another promising approach, explaining why an algorithm produces one output rather than another [102]. Knowledge-augmented AI systems that combine data-driven learning with explicit medical knowledge graphs or causal models can provide explanations that align with clinical reasoning patterns [103].
Critical care AI applications are subject to heightened ethical and regulatory scrutiny due to their potential impact on vulnerable patients. A primary consideration concerns fairness and bias, as AI systems trained on historical data may perpetuate or amplify existing disparities in critical care [104].
While AI models have shown impressive performance, most applications rely on retrospective data collected for research purposes. Validating their real-world utility requires prospective clinical trials to assess model reliability in real-world heterogeneous and noisy clinical environments. Bridging the gap between AI research and clinical practice will also require multi-center validation, improved model explainability, and workflow-centered implementation strategies to ensure meaningful bedside benefits [105,106].
Regulatory challenges persist as machine learning models evolve rapidly through the incorporation of new data [107]. The FDA’s regulatory framework for AI/ML-based Software considers both intended use and associated risk level. Liability questions remain when AI recommendations contribute to adverse outcomes, with clinicians currently retaining ultimate decision responsibility [108].
Establishing secure systems for collecting, storing, and sharing EHRs is complex. Privacy-preserving methods, such as third-party cloud services, can help safeguard patient information. Federated learning offers a promising solution by enabling decentralized AI training across institutions or devices connected through the Internet of Things (IoT) without transferring raw patient data, addressing interoperability and privacy concerns [109]. Similarly, Steganography techniques encrypt medical records within multiple (DICOM) images using optimized Least Significant Bit (LSB) substitution, producing hidden data that are resilient against tampering and unauthorized access [110].

5. Future Directions

5.1. Prospective, Real-World Validation

The evolution of artificial intelligence in critical care is progressing along several complementary trajectories that collectively promise to transform intensive care practice. A pivotal advancement is the transition from retrospective validation toward rigorous prospective evaluation, including randomized controlled trials of AI interventions.
The TREWS sepsis study represents an early exemplar of this approach, and similar trials are now underway for various other applications [32]. These studies are crucial for establishing the real-world clinical impact of AI tools and defining their appropriate role within the broader context of critical care delivery.
Future trial designs must address the unique challenges of AI evaluation, including the following:
  • Determining appropriate control conditions (e.g., standard care versus non-AI decision support).
  • Handling continuous model updates during trial periods.
  • Measuring both direct outcomes and secondary effects on workflow and team dynamics.
  • Evaluating cost-effectiveness and resource utilization.
This level of rigorous validation is critical for distinguishing clinically meaningful AI implementations from those that merely demonstrate statistical significance without meaningful real-world clinical utility.

5.2. Integrated Multimodal Platforms

Concurrent with this evolution in clinical validation, the field is moving toward more sophisticated multimodal and integrated AI systems. Rather than isolated point solutions addressing single clinical problems, future AI platforms will likely integrate multiple data modalities and functions to provide comprehensive support across the critical care workflow. A fully developed ICU AI platform might simultaneously provide continuous risk monitoring for multiple complications (such as sepsis, respiratory failure, and acute kidney injury), treatment optimization recommendations based on individual patient trajectories, automated documentation and information retrieval, and support for team communication and handoffs. This integration could significantly reduce alert fragmentation and provide more contextually relevant decision support by understanding the full clinical picture rather than focusing on isolated physiological parameters. Architectural approaches that utilize foundation models as central coordination layers, with specialized modules for specific tasks, show particular promise in achieving this level of integration while maintaining performance across diverse clinical tasks [111].

5.3. Human–AI Co-Intelligence

As these integrated systems mature, defining optimal human–AI collaboration models becomes increasingly important. The ideal partnership between human clinicians and AI systems remains an active area of research, with different models of collaboration potentially appropriate for different clinical tasks. In certain contexts, AI may primarily serve as a screening tool that prioritizes cases for human review, enabling clinicians to focus their attention where it is most needed. In other scenarios, AI can serve as decision support, providing recommendations for human approval while augmenting clinical judgment without replacing it. For narrowly defined, low-risk tasks, AI may eventually function as an autonomous agent, handling routine aspects of care while freeing human attention for more complex cases. Perhaps most promisingly, AI could evolve as a cognitive extension that handles information processing and pattern recognition, while humans focus on complex reasoning, ethical judgment, and interpersonal care—strengths that remain uniquely human. Research on team cognition and adaptive automation will inform these evolving collaborative relationships, helping to define interaction paradigms that maximize the complementary strengths of human expertise and computational power [104].

5.4. AI Literacy

The increasing integration of AI into critical care necessitates a corresponding evolution in education and training for critical care professionals. Future clinicians will require competencies that extend beyond traditional medical knowledge to include a basic understanding of AI/ML principles and limitations, critical appraisal of AI-generated recommendations, recognition of situations where AI may be unreliable, and effective communication about AI inputs to patient care decisions. Medical education institutions are beginning to incorporate these topics into their training curricula, but significant work remains to define core competencies and develop effective evaluation methods [112]. This educational transformation represents not merely an addition to existing training requirements but a fundamental reconceptualization of critical care practice in an era of human–AI collaboration. Addressing these educational needs proactively will be essential to realize the potential benefits of AI while mitigating risks associated with inappropriate reliance on or rejection of computational assistance.
The convergence of these developments—rigorous clinical validation, integrated multimodal systems, refined human–AI collaboration models, and evolved educational approaches—suggests a future where AI becomes seamlessly integrated into critical care practice. This integration promises to enhance clinical decision-making, reduce cognitive burden on healthcare providers, improve resource utilization, and ultimately optimize patient outcomes. However, realizing this vision requires continued transdisciplinary collaboration among clinicians, data scientists, human factors engineers, educators, ethicists, and patients themselves. By addressing the technical, clinical, ethical, and educational dimensions of AI implementation in parallel, the critical care community can navigate toward a future where technology augments rather than replaces the human elements of care that remain essential to the discipline.

6. Conclusions

Artificial intelligence is poised to transform intensive care medicine through enhanced prediction, personalization, and workflow optimization. Current applications span early warning systems, sepsis management, mechanical ventilation, diagnostic support, and clinical documentation, with varying levels of evidence and implementation maturity. While significant challenges remain in technical integration, clinical workflow adaptation, explainability, and regulation, promising results from prospective studies suggest that thoughtfully designed and implemented AI tools can meaningfully improve critical care outcomes.
The future of AI in intensive care will likely involve more integrated, multimodal systems that collaborate with rather than replace human clinicians. Success will require ongoing interdisciplinary collaboration between critical care professionals, data scientists, human factor engineers, and ethicists. As the field evolves, maintaining focus on patient-centered outcomes while addressing issues of transparency, equity, and appropriate validation will be essential to realizing the full potential of AI-augmented critical care.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm14124026/s1, Table S1: Key Technological Milestones in Biomedical AI/ML Applications [16,21,113,114,115,116,117,118,119,120,121,122,123,124,125,126]; Table S2: Key Technological Milestones in AI/ML Applications for Intensive Care Medicine [23,24,27,45,127,128,129,130,131].

Author Contributions

Conceptualization, P.H. and S.B.; methodology, P.H., S.B., A.N. and M.S.; validation, P.H., S.E.H. and F.S.T.; formal analysis, P.H., M.S. and M.A.B.; investigation, A.N. and M.S.; resources, P.H., S.B., A.N. and M.A.B.; data curation, P.H., S.B. and F.S.T.; writing—original draft preparation, P.H. and S.B.; writing—review and editing, P.H. and S.E.H.; supervision, S.B., S.E.H. and M.S.; project administration, S.B., S.E.H. and A.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study is a literature review that analyzed publicly available data and published research and, therefore, did not require ethical approval.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets analyzed during this study are publicly available. All data supporting the conclusions of this article are included within the article and its supplementary materials. The publicly available datasets analyzed are accessible through the sources cited in the references.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kramer, A.A.; Zimmerman, J.E.; Knaus, W.A. Severity of Illness and Predictive Models in Society of Critical Care Medicine’s First 50 Years: A Tale of Concord and Conflict. Crit. Care Med. 2021, 49, 728–740. [Google Scholar] [CrossRef] [PubMed]
  2. Vincent, J.-L.; Creteur, J. Paradigm Shifts in Critical Care Medicine: The Progress We Have Made. Crit. Care 2015, 19, S10. [Google Scholar] [CrossRef] [PubMed]
  3. Shortliffe, E.H.; Sepúlveda, M.J. Clinical Decision Support in the Era of Artificial Intelligence. JAMA 2018, 320, 2199. [Google Scholar] [CrossRef]
  4. Yu, K.-H.; Beam, A.L.; Kohane, I.S. Artificial Intelligence in Healthcare. Nat. Biomed. Eng. 2018, 2, 719–731. [Google Scholar] [CrossRef]
  5. Gutierrez, G. Artificial Intelligence in the Intensive Care Unit. Crit. Care 2020, 24, 101. [Google Scholar] [CrossRef]
  6. Keegan, M.T.; Gajic, O.; Afessa, B. Severity of Illness Scoring Systems in the Intensive Care Unit. Crit. Care Med. 2011, 39, 163–169. [Google Scholar] [CrossRef] [PubMed]
  7. Sekulic, A.D.; Trpkovic, S.V.; Pavlovic, A.P.; Marinkovic, O.M.; Ilic, A.N. Scoring Systems in Assessing Survival of Critically Ill ICU Patients. Med. Sci. Monit. 2015, 21, 2621–2629. [Google Scholar] [CrossRef]
  8. Lekadir, K.; Feragen, A.; Fofanah, A.J.; Frangi, A.F.; Buyx, A.; Emelie, A.; Lara, A.; Porras, A.R.; Chan, A.-W.; Navarro, A.; et al. FUTURE-AI: International Consensus Guideline for Trustworthy and Deployable Artificial Intelligence in Healthcare. BMJ 2023, 388, e081554. [Google Scholar] [CrossRef]
  9. Michard, F.; Mulder, M.P.; Gonzalez, F.; Sanfilippo, F. AI for the Hemodynamic Assessment of Critically Ill and Surgical Patients: Focus on Clinical Applications. Ann. Intensive Care 2025, 15, 26. [Google Scholar] [CrossRef]
  10. Shortliffe, E.H.; Davis, R.; Axline, S.G.; Buchanan, B.G.; Green, C.C.; Cohen, S.N. Computer-Based Consultations in Clinical Therapeutics: Explanation and Rule Acquisition Capabilities of the MYCIN System. Comput. Biomed. Res. 1975, 8, 303–320. [Google Scholar] [CrossRef]
  11. Miller, R.A. Medical Diagnostic Decision Support Systems--Past, Present, And Future: A Threaded Bibliography and Brief Commentary. J. Am. Med. Inform. Assoc. 1994, 1, 8–27. [Google Scholar] [CrossRef] [PubMed]
  12. Alabdulkarim, A.; Al-Rodhaan, M.; Tian, Y. Privacy-Preserving Healthcare System for Clinical Decision-Support and Emergency Call Systems. Commun. Netw. 2017, 9, 249–274. [Google Scholar] [CrossRef]
  13. Durand, C.; Alfandari, S.; Béraud, G.; Tsopra, R.; Lescure, F.-X.; Peiffer-Smadja, N. Clinical Decision Support Systems for Antibiotic Prescribing: An Inventory of Current French Language Tools. Antibiotics 2022, 11, 384. [Google Scholar] [CrossRef]
  14. Adams, I.D.; Chan, M.; Clifford, P.C.; Cooke, W.M.; Dallos, V.; de Dombal, F.T.; Edwards, M.H.; Hancock, D.M.; Hewett, D.J.; McIntyre, N. Computer Aided Diagnosis of Acute Abdominal Pain: A Multicentre Study. Br. Med. J. Clin. Res. Ed. 1986, 293, 800–804. [Google Scholar] [CrossRef] [PubMed]
  15. Crichton, N.J.; Fryer, J.G.; Spicer, C.C. Some Points on the Use of ‘Independent Bayes’ to Diagnose Acute Abdominal Pain. Stat. Med. 1987, 6, 945–959. [Google Scholar] [CrossRef]
  16. Miller, R.A.; Pople, H.E.; Myers, J.D. Internist-I, an Experimental Computer-Based Diagnostic Consultant for General Internal Medicine. N. Engl. J. Med. 1982, 307, 468–476. [Google Scholar] [CrossRef]
  17. Shwe, M.A.; Middleton, B.; Heckerman, D.E.; Henrion, M.; Horvitz, E.J.; Lehmann, H.P.; Cooper, G.F. Probabilistic Diagnosis Using a Reformulation of the INTERNIST-1/QMR Knowledge Base. I. The Probabilistic Model and Inference Algorithms. Methods Inf. Med. 1991, 30, 241–255. [Google Scholar] [CrossRef]
  18. Mumtaz, H.; Ejaz, M.K.; Tayyab, M.; Vohra, L.I.; Sapkota, S.; Hasan, M.; Saqib, M. APACHE Scoring as an Indicator of Mortality Rate in ICU Patients: A Cohort Study. Ann. Med. Surg. 2023, 85, 416–421. [Google Scholar] [CrossRef]
  19. Samwald, M.; Fehre, K.; de Bruin, J.; Adlassnig, K.-P. The Arden Syntax Standard for Clinical Decision Support: Experiences and Directions. J. Biomed. Inform. 2012, 45, 711–718. [Google Scholar] [CrossRef]
  20. Miller, R.A. A History of the INTERNIST-1 and Quick Medical Reference (QMR) Computer-Assisted Diagnosis Projects, with Lessons Learned. Yearb. Med. Inform. 2010, 19, 121–136. [Google Scholar] [CrossRef]
  21. Barnett, G.O.; Cimino, J.J.; Hupp, J.A.; Hoffer, E.P. DXplain. An Evolving Diagnostic Decision-Support System. JAMA 1987, 258, 67–74. [Google Scholar] [CrossRef] [PubMed]
  22. Subbe, C.P.; Kruger, M.; Rutherford, P.; Gemmel, L. Validation of a Modified Early Warning Score in Medical Admissions. QJM Int. J. Med. 2001, 94, 521–526. [Google Scholar] [CrossRef]
  23. Lee, J.; Scott, D.J.; Villarroel, M.; Clifford, G.D.; Saeed, M.; Mark, R.G. Open-Access MIMIC-II Database for Intensive Care Research. Conf. Proc. IEEE Eng. Med. Biol. Soc. 2011, 2011, 8315–8318. [Google Scholar] [CrossRef]
  24. Pirracchio, R.; Petersen, M.L.; Carone, M.; Rigon, M.R.; Chevret, S.; van der Laan, M.J. Mortality Prediction in Intensive Care Units with the Super ICU Learner Algorithm (SICULA): A Population-Based Study. Lancet Respir. Med. 2015, 3, 42–52. [Google Scholar] [CrossRef]
  25. Mayampurath, A.; Hagopian, R.; Venable, L.; Carey, K.; Edelson, D.; Churpek, M. Comparison of Machine Learning Methods for Predicting Outcomes after In-Hospital Cardiac Arrest. Crit. Care Med. 2022, 50, e162–e172. [Google Scholar] [CrossRef]
  26. Miotto, R.; Li, L.; Kidd, B.A.; Dudley, J.T. Deep Patient: An Unsupervised Representation to Predict the Future of Patients from the Electronic Health Records. Sci. Rep. 2016, 6, 26094. [Google Scholar] [CrossRef] [PubMed]
  27. Komorowski, M.; Celi, L.A.; Badawi, O.; Gordon, A.C.; Faisal, A.A. The Artificial Intelligence Clinician Learns Optimal Treatment Strategies for Sepsis in Intensive Care. Nat. Med. 2018, 24, 1716–1720. [Google Scholar] [CrossRef] [PubMed]
  28. Komorowski, M.; Celi, L.A.; Badawi, O.; Gordon, A.C.; Faisal, A.A. Understanding the Artificial Intelligence Clinician and Optimal Treatment Strategies for Sepsis in Intensive Care. arXiv 2019, arXiv:1903.02345. [Google Scholar]
  29. Seymour, C.W.; Kennedy, J.N.; Wang, S.; Chang, C.-C.H.; Elliott, C.F.; Xu, Z.; Berry, S.; Clermont, G.; Cooper, G.; Gomez, H.; et al. Derivation, Validation, and Potential Treatment Implications of Novel Clinical Phenotypes for Sepsis. JAMA 2019, 321, 2003. [Google Scholar] [CrossRef]
  30. Ibrahim, Z.M.; Wu, H.; Hamoud, A.; Stappen, L.; Dobson, R.J.B.; Agarossi, A. On Classifying Sepsis Heterogeneity in the ICU: Insight Using Machine Learning. J. Am. Med. Inform. Assoc. 2020, 27, 437–443. [Google Scholar] [CrossRef]
  31. Hyland, S.L.; Faltys, M.; Hüser, M.; Lyu, X.; Gumbsch, T.; Esteban, C.; Bock, C.; Horn, M.; Moor, M.; Rieck, B.; et al. Early Prediction of Circulatory Failure in the Intensive Care Unit Using Machine Learning. Nat. Med. 2020, 26, 364–373. [Google Scholar] [CrossRef]
  32. Adams, R.; Henry, K.E.; Sridharan, A.; Soleimani, H.; Zhan, A.; Rawat, N.; Johnson, L.; Hager, D.N.; Cosgrove, S.E.; Markowski, A.; et al. Prospective, Multi-Site Study of Patient Outcomes after Implementation of the TREWS Machine Learning-Based Early Warning System for Sepsis. Nat. Med. 2022, 28, 1455–1460. [Google Scholar] [CrossRef]
  33. Singhal, K.; Tu, T.; Gottweis, J.; Sayres, R.; Wulczyn, E.; Hou, L.; Clark, K.; Pfohl, S.; Cole-Lewis, H.; Neal, D.; et al. Towards Expert-Level Medical Question Answering with Large Language Models. arXiv 2023, arXiv:2305.09617. [Google Scholar] [CrossRef] [PubMed]
  34. Knaus, W.A.; Draper, E.A.; Wagner, D.P.; Zimmerman, J.E. APACHE II: A Severity of Disease Classification System. Crit. Care Med. 1985, 13, 818–829. [Google Scholar] [CrossRef] [PubMed]
  35. Johnson, A.E.W.; Stone, D.J.; Celi, L.A.; Pollard, T.J. The MIMIC Code Repository: Enabling Reproducibility in Critical Care Research. J. Am. Med. Inform. Assoc. 2018, 25, 32–39. [Google Scholar] [CrossRef] [PubMed]
  36. Pollard, T.J.; Johnson, A.E.W.; Raffa, J.D.; Celi, L.A.; Mark, R.G.; Badawi, O. The eICU Collaborative Research Database, a Freely Available Multi-Center Database for Critical Care Research. Sci. Data 2018, 5, 180178. [Google Scholar] [CrossRef]
  37. Kang, M.A.; Churpek, M.M.; Zadravecz, F.J.; Adhikari, R.; Twu, N.M.; Edelson, D.P. Real-Time Risk Prediction on the Wards: A Feasibility Study. Crit. Care Med. 2016, 44, 1468–1473. [Google Scholar] [CrossRef]
  38. Calfee, C.S.; Delucchi, K.; Parsons, P.E.; Thompson, B.T.; Ware, L.B.; Matthay, M.A. Subphenotypes in Acute Respiratory Distress Syndrome: Latent Class Analysis of Data from Two Randomised Controlled Trials. Lancet Respir. Med. 2014, 2, 611–620. [Google Scholar] [CrossRef]
  39. Peine, A.; Hallawa, A.; Bickenbach, J.; Dartmann, G.; Fazlic, L.B.; Schmeink, A.; Ascheid, G.; Thiemermann, C.; Schuppert, A.; Kindle, R.; et al. Development and Validation of a Reinforcement Learning Algorithm to Dynamically Optimize Mechanical Ventilation in Critical Care. NPJ Digit. Med. 2021, 4, 32. [Google Scholar] [CrossRef]
  40. Rajpurkar, P.; Irvin, J.; Ball, R.L.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.P.; et al. Deep Learning for Chest Radiograph Diagnosis: A Retrospective Comparison of the CheXNeXt Algorithm to Practicing Radiologists. PLoS Med. 2018, 15, e1002686. [Google Scholar] [CrossRef]
  41. Singhal, K.; Azizi, S.; Tu, T.; Mahdavi, S.S.; Wei, J.; Chung, H.W.; Scales, N.; Tanwani, A.; Cole-Lewis, H.; Pfohl, S.; et al. Large Language Models Encode Clinical Knowledge. Nature 2023, 620, 172–180. [Google Scholar] [CrossRef] [PubMed]
  42. Johnson, A.E.W.; Bulgarelli, L.; Shen, L.; Gayles, A.; Shammout, A.; Horng, S.; Pollard, T.J.; Hao, S.; Moody, B.; Gow, B.; et al. MIMIC-IV, a Freely Accessible Electronic Health Record Dataset. Sci. Data 2023, 10, 1. [Google Scholar] [CrossRef]
  43. Shillan, D.; Sterne, J.A.C.; Champneys, A.; Gibbison, B. Use of Machine Learning to Analyse Routinely Collected Intensive Care Unit Data: A Systematic Review. Crit. Care 2019, 23, 284. [Google Scholar] [CrossRef] [PubMed]
  44. Delahanty, R.J.; Alvarez, J.; Flynn, L.M.; Sherwin, R.L.; Jones, S.S. Development and Evaluation of a Machine Learning Model for the Early Identification of Patients at Risk for Sepsis. Ann. Emerg. Med. 2019, 73, 334–344. [Google Scholar] [CrossRef]
  45. Johnson, A.E.W.; Pollard, T.J.; Shen, L.; Lehman, L.H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Anthony Celi, L.; Mark, R.G. MIMIC-III, a Freely Accessible Critical Care Database. Sci. Data 2016, 3, 160035. [Google Scholar] [CrossRef]
  46. Chen, I.Y.; Szolovits, P.; Ghassemi, M. Can AI Help Reduce Disparities in General Medical and Mental Health Care? AMA J. Ethics 2019, 21, E167–E179. [Google Scholar] [CrossRef]
  47. Jayaraman, P.; Desman, J.; Sabounchi, M.; Nadkarni, G.N.; Sakhuja, A. A Primer on Reinforcement Learning in Medicine for Clinicians. NPJ Digit. Med. 2024, 7, 337. [Google Scholar] [CrossRef] [PubMed]
  48. Liu, S.; See, K.C.; Ngiam, K.Y.; Celi, L.A.; Sun, X.; Feng, M. Reinforcement Learning for Clinical Decision Support in Critical Care: Comprehensive Review. J. Med. Internet Res. 2020, 22, e18477. [Google Scholar] [CrossRef]
  49. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous Control with Deep Reinforcement Learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  50. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal Policy Optimization Algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  51. Washington, P. A Perspective on Crowdsourcing and Human-in-the-Loop Workflows in Precision Health. J. Med. Internet Res. 2024, 26, e51138. [Google Scholar] [CrossRef] [PubMed]
  52. Holzinger, A. Interactive Machine Learning for Health Informatics: When Do We Need the Human-in-the-Loop? Brain Inf. 2016, 3, 119–131. [Google Scholar] [CrossRef] [PubMed]
  53. Tonekaboni, S.; Joshi, S.; McCradden, M.D.; Goldenberg, A. What Clinicians Want: Contextualizing Explainable Machine Learning for Clinical End Use. arXiv 2019, arXiv:1905.05134. [Google Scholar]
  54. Sadeghi, S.; Hempel, L.; Rodemund, N.; Kirsten, T. Salzburg Intensive Care Database (SICdb): A Detailed Exploration and Comparative Analysis with MIMIC-IV. Sci. Rep. 2024, 14, 11438. [Google Scholar] [CrossRef] [PubMed]
  55. Akhtar, U.; Khattak, A.M.; Lee, S. Challenges in Managing Real-Time Data in Health Information System (HIS). In Inclusive Smart Cities and Digital Health; Chang, C.K., Chiari, L., Cao, Y., Jin, H., Mokhtari, M., Aloulou, H., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2016; Volume 9677, pp. 305–313. ISBN 978-3-319-39600-2. [Google Scholar]
  56. Cho, K.-J.; Kwon, O.; Kwon, J.; Lee, Y.; Park, H.; Jeon, K.-H.; Kim, K.-H.; Park, J.; Oh, B.-H. Detecting Patient Deterioration Using Artificial Intelligence in a Rapid Response System. Crit. Care Med. 2020, 48, e285–e289. [Google Scholar] [CrossRef]
  57. Churpek, M.M.; Yuen, T.C.; Winslow, C.; Meltzer, D.O.; Kattan, M.W.; Edelson, D.P. Multicenter Comparison of Machine Learning Methods and Conventional Regression for Predicting Clinical Deterioration on the Wards. Crit. Care Med. 2016, 44, 368–374. [Google Scholar] [CrossRef]
  58. Churpek, M.M.; Yuen, T.C.; Winslow, C.; Robicsek, A.A.; Meltzer, D.O.; Gibbons, R.D.; Edelson, D.P. Multicenter Development and Validation of a Risk Stratification Tool for Ward Patients. Am. J. Respir. Crit. Care Med. 2014, 190, 649–655. [Google Scholar] [CrossRef] [PubMed]
  59. Wong, A.; Otles, E.; Donnelly, J.P.; Krumm, A.; McCullough, J.; DeTroyer-Cooley, O.; Pestrue, J.; Phillips, M.; Konye, J.; Penoza, C.; et al. External Validation of a Widely Implemented Proprietary Sepsis Prediction Model in Hospitalized Patients. JAMA Intern. Med. 2021, 181, 1065. [Google Scholar] [CrossRef]
  60. Gholami, B.; Phan, T.S.; Haddad, W.M.; Cason, A.; Mullis, J.; Price, L.; Bailey, J.M. Replicating Human Expertise of Mechanical Ventilation Waveform Analysis in Detecting Patient-Ventilator Cycling Asynchrony Using Machine Learning. Comput. Biol. Med. 2018, 97, 137–144. [Google Scholar] [CrossRef]
  61. Sottile, P.D.; Albers, D.; Higgins, C.; Mckeehan, J.; Moss, M.M. The Association Between Ventilator Dyssynchrony, Delivered Tidal Volume, and Sedation Using a Novel Automated Ventilator Dyssynchrony Detection Algorithm. Crit. Care Med. 2018, 46, e151–e157. [Google Scholar] [CrossRef]
  62. McDuff, D.; Schaekermann, M.; Tu, T.; Palepu, A.; Wang, A.; Garrison, J.; Singhal, K.; Sharma, Y.; Azizi, S.; Kulkarni, K.; et al. Towards Accurate Differential Diagnosis with Large Language Models. Nature 2025, 254. [Google Scholar] [CrossRef] [PubMed]
  63. Alsentzer, E.; Murphy, J.; Boag, W.; Weng, W.-H.; Jindi, D.; Naumann, T.; McDermott, M. Publicly Available Clinical. In Proceedings of the 2nd Clinical Natural Language Processing Workshop, Minneapolis, MN, USA, 7 June 2019; Association for Computational Linguistics: New York, NY, USA, 2019; pp. 72–78. [Google Scholar]
  64. Huang, Z.; Sun, A.; Bai, H. Automatic Speech Recognition Error Correction on ICU Clinical Narration Dataset; Stanford University: Stanford, CA, USA, 2023. [Google Scholar]
  65. Archana, D.; Singh, S.K. Artificial Intelligence in Critical Care: Enhancing Decision-Making and Patient Outcomes. Healthc. Bull. 2023, 5, 128–134. [Google Scholar]
  66. Bertsimas, D.; Pauphilet, J.; Stevens, J.; Tandon, M. Predicting Inpatient Flow at a Major Hospital Using Interpretable Analytics. Manuf. Serv. Oper. Manag. 2022, 24, 2809–2824. [Google Scholar] [CrossRef]
  67. Siegel, S.; Zhang, J.; Bandyopadhyay, S.; Nerella, S.; Silva, B.; Baslanti, T.; Bihorac, A.; Rashidi, P. Leveraging Computer Vision in the Intensive Care Unit (ICU) for Examining Visitation and Mobility. arXiv 2024, arXiv:2403.06322. [Google Scholar]
  68. Muralitharan, S.; Nelson, W.; Di, S.; McGillion, M.; Devereaux, P.J.; Barr, N.G.; Petch, J. Machine Learning-Based Early Warning Systems for Clinical Deterioration: Systematic Scoping Review. J. Med. Internet Res. 2021, 23, e25187. [Google Scholar] [CrossRef] [PubMed]
  69. Lundberg, S.M.; Nair, B.; Vavilala, M.S.; Horibe, M.; Eisses, M.J.; Adams, T.; Liston, D.E.; Low, D.K.-W.; Newman, S.-F.; Kim, J.; et al. Explainable Machine-Learning Predictions for the Prevention of Hypoxaemia during Surgery. Nat. Biomed. Eng. 2018, 2, 749–760. [Google Scholar] [CrossRef] [PubMed]
  70. Zaidi, S.A.J.; Ghafoor, A.; Kim, J.; Abbas, Z.; Lee, S.W. HeartEnsembleNet: An Innovative Hybrid Ensemble Learning Approach for Cardiovascular Risk Prediction. Healthcare 2025, 13, 507. [Google Scholar] [CrossRef]
  71. Uppal, M.; Gupta, D.; Juneja, S.; Gadekallu, T.R.; El Bayoumy, I.; Hussain, J.; Lee, S.W. Enhancing Accuracy in Brain Stroke Detection: Multi-Layer Perceptron with Adadelta, RMSProp and AdaMax Optimizers. Front. Bioeng. Biotechnol. 2023, 11, 1257591. [Google Scholar] [CrossRef]
  72. Fleuren, L.M.; Klausch, T.L.T.; Zwager, C.L.; Schoonmade, L.J.; Guo, T.; Roggeveen, L.F.; Swart, E.L.; Girbes, A.R.J.; Thoral, P.; Ercole, A.; et al. Machine Learning for the Prediction of Sepsis: A Systematic Review and Meta-Analysis of Diagnostic Test Accuracy. Intensive Care Med. 2020, 46, 383–400. [Google Scholar] [CrossRef]
  73. Boussina, A.; Shashikumar, S.P.; Malhotra, A.; Owens, R.L.; El-Kareh, R.; Longhurst, C.A.; Quintero, K.; Donahue, A.; Chan, T.C.; Nemati, S.; et al. Impact of a Deep Learning Sepsis Prediction Model on Quality of Care and Survival. NPJ Digit. Med. 2024, 7, 14. [Google Scholar] [CrossRef]
  74. Henry, K.E.; Kornfield, R.; Sridharan, A.; Linton, R.C.; Groh, C.; Wang, T.; Wu, A.; Mutlu, B.; Saria, S. Human–Machine Teaming Is Key to AI Adoption: Clinicians’ Experiences with a Deployed Machine Learning System. NPJ Digit. Med. 2022, 5, 97. [Google Scholar] [CrossRef]
  75. Henry, K.E.; Adams, R.; Parent, C.; Soleimani, H.; Sridharan, A.; Johnson, L.; Hager, D.N.; Cosgrove, S.E.; Markowski, A.; Klein, E.Y.; et al. Factors Driving Provider Adoption of the TREWS Machine Learning-Based Early Warning System and Its Effects on Sepsis Treatment Timing. Nat. Med. 2022, 28, 1447–1454. [Google Scholar] [CrossRef] [PubMed]
  76. Parreco, J.; Hidalgo, A.; Parks, J.J.; Kozol, R.; Rattan, R. Using Artificial Intelligence to Predict Prolonged Mechanical Ventilation and Tracheostomy Placement. J. Surg. Res. 2018, 228, 179–187. [Google Scholar] [CrossRef] [PubMed]
  77. Buiteman-Kruizinga, L.A.; Mkadmi, H.E.; Serpa Neto, A.; Kruizinga, M.D.; Botta, M.; Schultz, M.J.; Paulus, F.; Van Der Heiden, P.L.J. Effect of INTELLiVENT-ASV versus Conventional Ventilation on Ventilation Intensity in Patients with COVID-19 ARDS—An Observational Study. J. Clin. Med. 2021, 10, 5409. [Google Scholar] [CrossRef]
  78. Ghassemi, M.; Naumann, T.; Schulam, P.; Beam, A.L.; Chen, I.Y.; Ranganath, R. A Review of Challenges and Opportunities in Machine Learning for Health. AMIA Jt. Summits Transl. Sci. Proc. 2020, 2020, 191–200. [Google Scholar]
  79. Pinsky, M.R.; Bedoya, A.; Bihorac, A.; Celi, L.; Churpek, M.; Economou-Zavlanos, N.J.; Elbers, P.; Saria, S.; Liu, V.; Lyons, P.G.; et al. Use of Artificial Intelligence in Critical Care: Opportunities and Obstacles. Crit. Care 2024, 28, 113. [Google Scholar] [CrossRef] [PubMed]
  80. Liu, C.-F.; Hung, C.-M.; Ko, S.-C.; Cheng, K.-C.; Chao, C.-M.; Sung, M.-I.; Hsing, S.-C.; Wang, J.-J.; Chen, C.-J.; Lai, C.-C.; et al. An Artificial Intelligence System to Predict the Optimal Timing for Mechanical Ventilation Weaning for Intensive Care Unit Patients: A Two-Stage Prediction Approach. Front. Med. 2022, 9, 935366. [Google Scholar] [CrossRef]
  81. Fan, E.; Del Sorbo, L.; Goligher, E.C.; Hodgson, C.L.; Munshi, L.; Walkey, A.J.; Adhikari, N.K.J.; Amato, M.B.P.; Branson, R.; Brower, R.G.; et al. An Official American Thoracic Society/European Society of Intensive Care Medicine/Society of Critical Care Medicine Clinical Practice Guideline: Mechanical Ventilation in Adult Patients with Acute Respiratory Distress Syndrome. Am. J. Respir. Crit. Care Med. 2017, 195, 1253–1263. [Google Scholar] [CrossRef] [PubMed]
  82. Chen, L.; Teotia, R.; Verdone, A.; Cardall, A.; Tyagi, L.; Shen, Y.; Chopra, S. Fine-Tuning In-House Large Language Models to Infer Differential Diagnosis from Radiology Reports. arXiv 2024, arXiv:2410.09234. [Google Scholar]
  83. Nilius, H.; Tsouka, S.; Nagler, M.; Masoodi, M. Machine Learning Applications in Precision Medicine: Overcoming Challenges and Unlocking Potential. TrAC Trends Anal. Chem. 2024, 179, 117872. [Google Scholar] [CrossRef]
  84. Pyne, Y.; Wong, Y.M.; Fang, H.; Simpson, E. Analysis of ‘One in a Million’ Primary Care Consultation Conversations Using Natural Language Processing. BMJ Health Care Inform. 2023, 30, e100659. [Google Scholar] [CrossRef]
  85. Namli, T.; Anıl Sınacı, A.; Gönül, S.; Herguido, C.R.; Garcia-Canadilla, P.; Muñoz, A.M.; Esteve, A.V.; Ertürkmen, G.B.L. A Scalable and Transparent Data Pipeline for AI-Enabled Health Data Ecosystems. Front. Med. 2024, 11, 1393123. [Google Scholar] [CrossRef] [PubMed]
  86. Hollenbeak, C.S.; Henning, D.J.; Geeting, G.K.; Ledeboer, N.A.; Faruqi, I.A.; Pierce, C.G.; Thomas, C.B.; O’Neal, H.R. Costs and Consequences of a Novel Emergency Department Sepsis Diagnostic Test: The IntelliSep Index. Crit. Care Explor. 2023, 5, e0942. [Google Scholar] [CrossRef]
  87. Lindroth, H.; Nalaie, K.; Raghu, R.; Ayala, I.N.; Busch, C.; Bhattacharyya, A.; Moreno Franco, P.; Diedrich, D.A.; Pickering, B.W.; Herasevich, V. Applied Artificial Intelligence in Healthcare: A Review of Computer Vision Technology Application in Hospital Settings. J. Imaging 2024, 10, 81. [Google Scholar] [CrossRef] [PubMed]
  88. Prince, S.J.D. Computer Vision: Models, Learning, and Inference; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  89. Casey, A.; Davidson, E.; Poon, M.; Dong, H.; Duma, D.; Grivas, A.; Grover, C.; Suárez-Paniagua, V.; Tobin, R.; Whiteley, W.; et al. A Systematic Review of Natural Language Processing Applied to Radiology Reports. BMC Med. Inform. Decis. Mak. 2021, 21, 179. [Google Scholar] [CrossRef]
  90. Soenksen, L.R.; Ma, Y.; Zeng, C.; Boussioux, L.; Villalobos Carballo, K.; Na, L.; Wiberg, H.M.; Li, M.L.; Fuentes, I.; Bertsimas, D. Integrated Multimodal Artificial Intelligence Framework for Healthcare Applications. NPJ Digit. Med. 2022, 5, 149. [Google Scholar] [CrossRef] [PubMed]
  91. Brundage, M.; Avin, S.; Wang, J.; Belfield, H.; Krueger, G.; Hadfield, G.; Khlaaf, H.; Yang, J.; Toner, H.; Fong, R.; et al. Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims. arXiv 2020, arXiv:2004.07213. [Google Scholar] [CrossRef]
  92. Michard, F.; Teboul, J.L. Predictive Analytics: Beyond the Buzz. Ann. Intensive Care 2019, 9, 46. [Google Scholar] [CrossRef] [PubMed]
  93. Alwi, R.; Prowse, P.; Gaamangwe, T. Proactive Role of Clinical Engineering in the Adoption of ISO/IEC 80001-1 within Healthcare Delivery Organization. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 5623–5626. [Google Scholar]
  94. Duda, S.N.; Kennedy, N.; Conway, D.; Cheng, A.C.; Nguyen, V.; Zayas-Cabán, T.; Harris, P.A. HL7 FHIR-Based Tools and Initiatives to Support Clinical Research: A Scoping Review. J. Am. Med. Inform. Assoc. 2022, 29, 1642–1653. [Google Scholar] [CrossRef] [PubMed]
  95. Wilkinson, M.D.; Dumontier, M.; Aalbersberg, I.J.; Appleton, G.; Axton, M.; Baak, A.; Blomberg, N.; Boiten, J.-W.; Da Silva Santos, L.B.; Bourne, P.E.; et al. The FAIR Guiding Principles for Scientific Data Management and Stewardship. Sci. Data 2016, 3, 160018. [Google Scholar] [CrossRef]
  96. Van Der Sijs, H.; Aarts, J.; Vulto, A.; Berg, M. Overriding of Drug Safety Alerts in Computerized Physician Order Entry. J. Am. Med. Inform. Assoc. 2006, 13, 138–147. [Google Scholar] [CrossRef]
  97. Sutton, R.T.; Pincock, D.; Baumgart, D.C.; Sadowski, D.C.; Fedorak, R.N.; Kroeker, K.I. An Overview of Clinical Decision Support Systems: Benefits, Risks, and Strategies for Success. NPJ Digit. Med. 2020, 3, 17. [Google Scholar] [CrossRef] [PubMed]
  98. Sittig, D.F.; Wright, A.; Osheroff, J.A.; Middleton, B.; Teich, J.M.; Ash, J.S.; Campbell, E.; Bates, D.W. Grand Challenges in Clinical Decision Support. J. Biomed. Inform. 2008, 41, 387–392. [Google Scholar] [CrossRef] [PubMed]
  99. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 1135–1144. [Google Scholar]
  100. Lundberg, S.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar]
  101. Rudin, C. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
  102. Miller, T. Explanation in Artificial Intelligence: Insights from the Social Sciences. Artif. Intell. 2019, 267, 1–38. [Google Scholar] [CrossRef]
  103. Gao, Y.; Li, R.; Croxford, E.; Caskey, J.; Patterson, B.W.; Churpek, M.; Miller, T.; Dligach, D.; Afshar, M. Leveraging Medical Knowledge Graphs Into Large Language Models for Diagnosis Prediction: Design and Application Study. JMIR AI 2025, 4, e58670. [Google Scholar] [CrossRef]
  104. Topol, E.J. High-Performance Medicine: The Convergence of Human and Artificial Intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef]
  105. Ahmed, M.I.; Spooner, B.; Isherwood, J.; Lane, M.; Orrock, E.; Dennison, A. A Systematic Review of the Barriers to the Implementation of Artificial Intelligence in Healthcare. Cureus 2023, 15, e46454. [Google Scholar] [CrossRef]
  106. Windecker, D.; Baj, G.; Shiri, I.; Kazaj, P.M.; Kaesmacher, J.; Gräni, C.; Siontis, G.C.M. Generalizability of FDA-Approved AI-Enabled Medical Devices for Clinical Use. JAMA Netw. Open 2025, 8, e258052. [Google Scholar] [CrossRef]
  107. Center for Devices and Radiological Health. Artificial Intelligence and Machine Learning in Software as a Medical Device. FDA 2025. [Google Scholar]
  108. Price, L.; Nicholson, W., II. Medical AI and Contextual Bias. Harv. J. Law. Technol. 2019, 33, 65–116. [Google Scholar]
  109. Kaissis, G.A.; Makowski, M.R.; Rückert, D.; Braren, R.F. Secure, Privacy-Preserving and Federated Machine Learning in Medical Imaging. Nat. Mach. Intell. 2020, 2, 305–311. [Google Scholar] [CrossRef]
  110. Riaz, H.; Naqvi, R.A.; Ellahi, M.; Usman, M.A.; Usman, M.R.; Jeong, D.; Lee, S.W. Robust Steganography Technique for Enhancing the Protection of Medical Records in Healthcare Informatics. IEEE J. Biomed. Health Inform. 2025, 1–13. [Google Scholar] [CrossRef] [PubMed]
  111. Lu, M.Y.; Williamson, D.F.K.; Chen, T.Y.; Chen, R.J.; Barbieri, M.; Mahmood, F. Data-Efficient and Weakly Supervised Computational Pathology on Whole-Slide Images. Nat. Biomed. Eng. 2021, 5, 555–570. [Google Scholar] [CrossRef]
  112. Paranjape, K.; Schinkel, M.; Nannan Panday, R.; Car, J.; Nanayakkara, P. Introducing Artificial Intelligence Training in Medical Education. JMIR Med. Educ. 2019, 5, e16048. [Google Scholar] [CrossRef] [PubMed]
  113. Turing, A.M. I.—Computing machinery and intelligence. Mind 1950, LIX, 433–460. [Google Scholar] [CrossRef]
  114. McCarthy, J.; Minsky, M.L.; Rochester, N.; Shannon, C.E. A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence: 31 August 1955. AI Mag. 2006, 27, 12–14. [Google Scholar] [CrossRef]
  115. Weizenbaum, J. ELIZA—A Computer Program for the Study of Natural Language Communication between Man and Machine. Commun. ACM 1966, 9, 36–45. [Google Scholar] [CrossRef]
  116. Lindsay, R.K.; Buchanan, B.G.; Feigenbaum, E.A.; Lederberg, J. DENDRAL: A Case Study of the First Expert System for Scientific Hypothesis Formation. Artif. Intell. 1993, 61, 209–261. [Google Scholar] [CrossRef]
  117. Computer-Based Medical Consultations: Mycin; Shortliffe, E.H., Ed.; Elsevier: Amsterdam, The Netherlands, 1976; ISBN 978-0-444-00179-5. [Google Scholar]
  118. Fukushima, K. Neocognitron: A Self-Organizing Neural Network Model for a Mechanism of Pattern Recognition Unaffected by Shift in Position. Biol. Cybern. 1980, 36, 193–202. [Google Scholar] [CrossRef]
  119. Freer, T.W.; Ulissey, M.J. Screening Mammography with Computer-Aided Detection: Prospective Study of 12,860 Patients in a Community Breast Center. Radiology 2001, 220, 781–786. [Google Scholar] [CrossRef] [PubMed]
  120. Brown, M.P.S.; Grundy, W.N.; Lin, D.; Cristianini, N.; Sugnet, C.W.; Furey, T.S.; Ares, M.; Haussler, D. Knowledge-Based Analysis of Microarray Gene Expression Data by Using Support Vector Machines. Proc. Natl. Acad. Sci. USA 2000, 97, 262–267. [Google Scholar] [CrossRef] [PubMed]
  121. Ferrucci, D.; Brown, E.; Chu-Carroll, J.; Fan, J.; Gondek, D.; Kalyanpur, A.A.; Lally, A.; Murdock, J.W.; Nyberg, E.; Prager, J.; et al. Building Watson: An Overview of the DeepQA Project. AI Mag. 2010, 31, 59–79. [Google Scholar] [CrossRef]
  122. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  123. Abràmoff, M.D.; Lavin, P.T.; Birch, M.; Shah, N.; Folk, J.C. Pivotal Trial of an Autonomous AI-Based Diagnostic System for Detection of Diabetic Retinopathy in Primary Care Offices. NPJ Digit. Med. 2018, 1, 39. [Google Scholar] [CrossRef]
  124. Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly Accurate Protein Structure Prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef] [PubMed]
  125. Tu, T.; Schaekermann, M.; Palepu, A.; Saab, K.; Freyberg, J.; Tanno, R.; Wang, A.; Li, B.; Amin, M.; Cheng, Y.; et al. Towards Conversational Diagnostic Artificial Intelligence. Nature 2025, 1–9. [Google Scholar] [CrossRef]
  126. Warner Homer, R. Computer-Assisted Medical Decision-Making; Academic Press: Cambridge, MA, USA, 1979. [Google Scholar]
  127. Knaus, W.A.; Zimmerman, J.E.; Wagner, D.P.; Draper, E.A.; Lawrence, D.E. APACHE—Acute Physiology and Chronic Health Evaluation: A Physiologically Based Classification System. Crit. Care Med. 1981, 9, 591–597. [Google Scholar] [CrossRef]
  128. Rutledge, G.W.; Thomsen, G.E.; Farr, B.R.; Tovar, M.A.; Polaschek, J.X.; Beinlich, I.A.; Sheiner, L.B.; Fagan, L.M. The Design and Implementation of a Ventilator-Management Advisor. Artif. Intell. Med. 1993, 5, 67–82. [Google Scholar] [CrossRef]
  129. Moody, G.B.; Mark, R.G. A Database to Support Development and Evaluation of Intelligent Intensive Care Monitoring. In Proceedings of the Computers in Cardiology, Indianapolis, IN, USA, 8–11 September 1996; IEEE: New York, NY, USA, 1996; pp. 657–660. [Google Scholar]
  130. Henry, K.E.; Hager, D.N.; Pronovost, P.J.; Saria, S. A Targeted Real-Time Early Warning Score (TREWScore) for Septic Shock. Sci. Transl. Med. 2015, 7, 299ra122. [Google Scholar] [CrossRef]
  131. Johnson, A.; Bulgarelli, L.; Pollard, T.; Horng, S.; Celi, L.A.; Mark, R. MIMIC-IV, version 3.1; MIT Laboratory for Computational Physiology: Cambridge, MA, USA, 2025. [Google Scholar]
Figure 1. Key technological milestones in AI and ML applications for biomedicine and intensive care medicine. The upper timeline (blue) depicts landmark developments in general biomedical applications, spanning from early theoretical foundations to current specialized systems as shown in Supplementary Table S1. The lower timeline (purple) illustrates parallel innovations specific to intensive care medicine, highlighting progression from basic clinical decision support to sophisticated predictive analytics, as shown in Supplementary Table S2. A concise description of each timepoint, with corresponding scientific references, is provided in the Supplementary Material. Abbreviations: APACHE, Acute Physiology and Chronic Health Evaluation; CAD, computer-aided diagnosis; ELIZA, early natural language processing system; HELP, Health Evaluation through Logical Processing; MIMIC, Medical Information Mart for Intensive Care; SICULA, Super ICU Learner Algorithm; SVM, support vector machine; TREWScore, Targeted Real-time Early Warning Score.
Figure 1. Key technological milestones in AI and ML applications for biomedicine and intensive care medicine. The upper timeline (blue) depicts landmark developments in general biomedical applications, spanning from early theoretical foundations to current specialized systems as shown in Supplementary Table S1. The lower timeline (purple) illustrates parallel innovations specific to intensive care medicine, highlighting progression from basic clinical decision support to sophisticated predictive analytics, as shown in Supplementary Table S2. A concise description of each timepoint, with corresponding scientific references, is provided in the Supplementary Material. Abbreviations: APACHE, Acute Physiology and Chronic Health Evaluation; CAD, computer-aided diagnosis; ELIZA, early natural language processing system; HELP, Health Evaluation through Logical Processing; MIMIC, Medical Information Mart for Intensive Care; SICULA, Super ICU Learner Algorithm; SVM, support vector machine; TREWScore, Targeted Real-time Early Warning Score.
Jcm 14 04026 g001
Figure 2. Mind-map illustrating key aspects of AI/ML integration in ICUs.
Figure 2. Mind-map illustrating key aspects of AI/ML integration in ICUs.
Jcm 14 04026 g002
Table 1. Timeline of machine intelligence evolution in the biomedical field, particularly in intensive care medicine.
Table 1. Timeline of machine intelligence evolution in the biomedical field, particularly in intensive care medicine.
YearTechnologyExampleKey Innovation
1959Logical FrameworkLedley and Lusted’s “Reasoning Foundations of Medical Diagnosis”First formal approach to medical decision-making using symbolic logic and probability [12]
1972Rule-based Expert SystemMYCINFirst major clinical decision support system for infectious disease diagnosis and antibiotic selection [13]
1976Bayesian Networkde Dombal’s Acute Abdominal Pain Diagnosis SystemApplied Bayesian probability for differential diagnosis of abdominal pain [14,15]
1981Causal-Associational NetworkINTERNIST-I/QMRComprehensive internal medicine diagnostic system with disease-finding relationships [16,17]
1985Severity Scoring SystemAPACHE IIFirst widely adopted ICU mortality prediction model using physiologic variables [18]
1993Decision Support FrameworkArden SyntaxStandardized representation for sharing medical knowledge and decision rules [19]
1994Integrated Clinical SystemHELP SystemHospital-wide clinical decision support integrated with electronic records [20]
1997Probabilistic Expert SystemDXplainDiagnostic decision support system using probabilistic reasoning [21]
2001Real-time Alerting SystemMedical Emergency Team TriggersEarly implementation of rule-based deterioration detection [22]
2006Clinical Data RepositoryMIMIC-II DatabaseFirst major open-access critical care database enabling ML research [23]
2010Supervised Machine LearningSuperLearner ICU Mortality PredictionEnsemble ML methods outperform traditional scoring systems [24]
2014Early Warning SystemeCART ML-based deterioration prediction with demonstrated clinical impact [25]
2016Deep LearningDeepPatientDeep learning for disease prediction using EHR data [26]
2018Reinforcement LearningAI ClinicianRL for sepsis treatment optimization using retrospective ICU data [27,28]
2019Unsupervised LearningSepsis PhenotypingML-identified sepsis subtypes with differential treatment responses [29,30]
2020Temporal Deep LearningCirculatory Failure Early WarningLong Short-Term Memory networks for predicting hemodynamic instability from continuous data [31]
2022Validated AI ImplementationTREWS Sepsis SystemFirst prospectively validated AI system showing mortality reduction [32]
2023Foundation ModelsMed-PaLM/Clinical LLMsLarge language models demonstrating medical reasoning and knowledge [33]
2024Multimodal AI SystemsIntegrated Clinical AI PlatformsSystems combining multiple AI modalities for comprehensive decision support [33]
Note. APACHE II = Acute Physiology and Chronic Health Evaluation II; EHR = electronic health records; DXplain = Diagnostic Explanation System; eCART = Electronic Cardiac Arrest Risk Triage; HELP System = Health Evaluation through Logical Processing; INTERNIST-I/QMR = Quick Medical Reference for INTERNIST-I; LLM = Large language model; Med-PaLM = Medical Pathways Language Model; MIMIC = Medical Information Mart for Intensive Care; MYCIN = A Rule-Based Expert System for Infectious Diseases; TREWS = Targeted Real-time Early Warning System.
Table 2. Summary of AI/ML applications, modalities, and methods in intensive care.
Table 2. Summary of AI/ML applications, modalities, and methods in intensive care.
ICU ApplicationsData ModalitiesAI/ML ParadigmsMethods/ArchitecturesExample Tools/Studies
Early warning and deterioration predictionTime-series, tabularSupervised learningRandom Forests, RNNs, TransformerseCART [37], Hyland et al. [31], SICULA [24]
Sepsis prediction and triageTabular, TextSupervised learningXGBoost, ClinicalBERTTREWS [32], Sepsis NLP Alerts [33]
ICU discharge, length of stay (LOS) predictionTabular, time-seriesSupervised learningGradient Boosting, Deep MLPsReadmission models, ICU LOS tools
Phenotyping (e.g., ARDS, sepsis)Tabular, time-seriesUnsupervised learningK-means, LCA, AutoencodersSeymour et al. [29], Calfee et al. [38]
Mechanical ventilation managementTime-seriesReinforcement learningDeep Q-Networks, Policy GradientAI Clinician, Peine et al. [39]
Diagnostic support (imaging)ImagingSupervised learning (Deep Learning)CNNs, Vision TransformersCheXNeXt, MIMIC-CXR [40]
Clinical documentation/NLPTextSupervised learning, pretrainingTransformers, BERT, GPTAutoNote, Med-PaLM
Multimodal clinical reasoningText + tabular + imagingTransfer learning/fine-tuningFoundation Models (GPT-4, Med-PaLM)Med-PaLM, LLM-driven ICU copilots [41]
Note. ARDS = Acute Respiratory Distress Syndrome; BERT = Bidirectional Encoder Representations from Transformers; ClinicalBERT = Clinical adaptation of BERT for healthcare-specific tasks; CNNs = Convolutional Neural Networks; Deep Q-Networks = Deep Q-Learning Networks; eCART = Electronic Cardiac Arrest Risk Triage; Foundation Models = Large, generalizable AI models pre-trained on diverse datasets; GPT = Generative Pre-trained Transformer; GPT-4 = Fourth-generation version of GPT; LCA = Latent Class Analysis; LLM = Large language model; LOS = Length of stay; MIMIC-CXR = Medical Information Mart for Intensive Care Chest X-ray dataset; Med-PaLM = Medical Pathways Language Model; NLP = Natural Language Processing; Policy Gradient = Reinforcement Learning Policy Optimization Algorithm; RNNs = Recurrent Neural Networks; SICULA = Super ICU Learner Algorithm; TREWS = Targeted Real-Time Early Warning System; XGBoost = Extreme Gradient Boosting.
Table 3. Comparative performance, operational impact, and socio-technical considerations of human versus AI systems across key ICU domains.
Table 3. Comparative performance, operational impact, and socio-technical considerations of human versus AI systems across key ICU domains.
ICU DomainHuman Benchmark AI Benchmark (Best Published)Documented Benefit (Effect Size)Running Cost Key AI AdvantagesLimitations and Interaction ChallengesReferences
Early-warning/deteriorationMEWS AUROC ≈ 0.70 (ward vital checks every 4 h)eCART GBM AUROC ≈ 0.85; RNN model AUROC 0.90 at 6 h horizonΔ + 0.15–0.20 AUROC; sensitivity ↑ ≈ 250% at matched specificity; median alert 2–6 h earlierLow → MediumContinuous streaming; multivariate pattern detection; earlier rescue team activationHigh-quality signal feed; alert fatigue; model drift; opacityCho et al. [56]; Churpek et al. [57,58]; Hyland et al. [31]
Sepsis detection and managementClinical recognition delay 2–6 h; guideline adherence variableTREWS RF AUROC 0.87; alert acted ≤ 3 h dropped mortality by 3.3 pp4–12 h earlier detection; 18.7% relative mortality ↓Medium → HighStandardised, 24/7 screening; complex pattern handling; RL prototypes personalise fluids/pressorsInter-site performance spread (Epic model: Se 33%, PPV 12%); data bias; causal opacityAdams et al. [32]; Wong et al. [59]; Komorowski et al. [27]
Mechanical ventilationManual titration; >20% of ARDS cases plateau > 30 cmH2O; dyssynchrony often missedVentAI RL policy increased protective-vent use by 203%; waveform AI detects cycling asynchrony with > 90% accuracyModel-guided settings associated with 2–3 pp survival gain (retrospective); ↓ driving pressureMedium → HighPatient-specific closed-loop optimisation; real-time waveform analytics; workload reliefLimited prospective safety trials; autonomy liability; cross-device generalisabilityPeine et al. [39]; Gholami et al. [60]; Sottile et al. [61]
Diagnostic support (CXR ± LLM)9 radiologists: AUROC 0.832 across 14 findings; variable fatigueCheXNeXt * AUROC 0.846; read 420 images in 1.5 min; LLM differential top 10 accuracy of 59.1% vs. 33.6% (juniors)Similar or better accuracy in <1% of the time; broader differential listsHigh → MediumSpeed; consistency; 24/7 availability; heat-map explanationsLimited clinical context; rare/atypical cases; automation biasRajpurkar et al. [40]; McDuff et al. [62]
Documentation and workflow30–50% of shift on notes; 50 min after-hours EHR timeNLP/ASR tools cut note time by 15–20%; after-hours EHR ↓ 30%; alert generation ↑ 10% speed8% of shifts returned to bedside care; error rate ↓Low → MediumAdministrative burden relief; real-time structured data; searchable notesPrivacy; nuance capture; language diversity; dependency on templatesAlsentzer et al. [63]; Huang et al. [64]; Archana et al. [65]
Resource optimisationExperience-based staffing, reactive bed assignmentLOS/discharge ML models improve census forecasting; CV room activity analytics predict delirium↑ forecast accuracy → better bed and staff utilisation (exact % context-specific)MediumPro-active, data-driven planning; multi-variable optimizationSudden exogenous shocks; fairness in allocation; staff acceptanceBertsimas et al. [66]; Siegel et al. [67]
‡ Human benchmark denotes the best-reported clinician or rule-based performance in the comparator arms of the cited studies; AI benchmark is the highest-quality, peer-reviewed model with external or prospective validation for the same task; documented benefit is the absolute effect size taken directly from the human; * CheXNeXt, a convolutional neural network to concurrently detect the presence of 14 different pathologies in chest radiograph—AI comparison (e.g., ΔAUROC, lead-time gained, mortality, or minutes saved); † running cost categories are qualitative—Low = stand-alone software/SaaS, Medium = software plus moderate integration and compute, High = bedside hardware or continuous GPU inference with governance overhead. See the Abbreviations section for term definitions. Abbreviations: ARDS = Acute Respiratory Distress Syndrome; ASR = Automatic Speech Recognition; AUROC = Area Under the Receiver Operating Characteristic Curve; CheXNeXt = A convolutional neural network to detect 14 pathologies; CV = Computer Vision; CXR = Chest X-Ray; eCART = Electronic Cardiac Arrest Warning System; GBM = Gradient-Boosting Machine; LLMs = Large Language Models; LOS = Length of Stay; MEWS = Modified Early Warning Score; PPV = Positive Predictive Value; pp = percentage points; RF = Random Forest; RL = Reinforcement Learning; RNN = Recurrent Neural Network; Se = Sensitivity; TREWS = Targeted Real-Time Early Warning System.
Table 4. Implementation challenges for AI-CDSS in intensive care.
Table 4. Implementation challenges for AI-CDSS in intensive care.
DomainChallengeCritical Care RelevanceMitigation StrategiesRelevant Frameworks/Principles
Data and InfrastructureData fragmentation and heterogeneityICU systems (monitors, EHRs, devices) are poorly integrated, leading to incomplete or unsynchronized inputsInteroperability layers; data harmonization; standard ontologies (e.g., FHIR)ISO 80001, HL7, FAIR principles [93,94,95]
Data quality and artifact noiseICU data (e.g., vitals, labs) are prone to gaps, outliers, and artifacts that affect model performanceRobust preprocessing; outlier detection; real-time signal cleaningGood Clinical Data Management Practice (GCDMP)
Model DesignGeneralizability across ICUsModels trained in one setting (e.g., tertiary ICU) often perform poorly in others due to population differencesExternal validation; domain adaptation; multicenter training datasetsTRIPOD-AI, MINIMAR
Updating and model driftStatic models degrade over time; adaptive models pose safety and validation risksVersion control; locked models with scheduled retraining; monitoring performanceFDA Predetermined Change Control Plan (PCCP)
Clinical Workflow IntegrationPoor timing or placement of AI supportAlerts or predictions may come too late or disrupt natural decision pointsWorkflow-mapped deployment; UI design aligned with ICU care pathwaysHuman-centered design principles
Alert fatigueICU teams already face alarm overload; AI tools may add to the burden without benefitAlert prioritization; threshold tuning; suppressive logic during emergenciesAAMI standards; ISO/IEC usability standards
Low clinical adoptionAI-CDSS often bypassed if they are misaligned with team roles or too opaqueCo-design with ICU teams; agile iterations; training and onboardingImplementation Science best practices
Trust and TransparencyBlack box models and explainabilityClinicians must understand the rationale to use or defend AI-assisted decisionsSHAP/LIME; counterfactuals; interpretable surrogate modelsEU AI Act (Transparency and Explainability clauses)
Misalignment with clinical reasoningAI outputs may not fit into clinicians’ diagnostic frameworksKnowledge-augmented models; explanation interfaces using clinical logicACM Code of Ethics, Human–AI Teaming Guidelines
Governance and RegulationBias and inequityAI may reinforce disparities based on race, sex, and comorbiditiesSubgroup analysis; fairness audits; demographic-aware trainingFDA GMLP; GDPR Art. 22; IEEE P7003
Accountability and liabilityUnclear who is responsible when AI recommendations contribute to harmMaintain clinician-in-the-loop control; document AI influence in decision logsMalpractice law (jurisdiction-specific); legal gray zones
Privacy and data useICU data are highly sensitive; secondary use for training raises legal/ethical concernsDe-identification; federated learning; institutional review board (IRB) approvalHIPAA, GDPR, Institutional Ethics Committees
Note. AAMI = Association for the Advancement of Medical Instrumentation; ACM = Association for Computing Machinery; AI-CDSS = Artificial Intelligence Clinical Decision Support Systems; EHR = Electronic Health Record; EU AI Act = European Union Artificial Intelligence Act; FAIR = Findable, Accessible, Interoperable, Reusable (data principles); FDA = U.S Food and Drug Administration; FDA GMLP = U.S. FDA Good Machine Learning Practice guidelines; FHIR = Fast Healthcare Interoperability Resources; GCDMP = Good Clinical Data Management Practice; GDPR = General Data Protection Regulation; HIPAA = Health Insurance Portability and Accountability Act; HL7 = Health Level Seven International; IEEE P7003 = IEEE Standard for Algorithmic Bias Considerations; IRB = Institutional Review Board; ISO = International Organization for Standardization; IEC = International Electrotechnical Commission; LIME = Local Interpretable Model-agnostic Explanations; MINIMAR = Minimum Information for Medical AI Reporting; PCCP = Predetermined Change Control Plan; SHAP = SHapley Additive exPlanations; TRIPOD-AI = Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis—Artificial Intelligence extension; UI = User Interface.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hadweh, P.; Niset, A.; Salvagno, M.; Al Barajraji, M.; El Hadwe, S.; Taccone, F.S.; Barrit, S. Machine Learning and Artificial Intelligence in Intensive Care Medicine: Critical Recalibrations from Rule-Based Systems to Frontier Models. J. Clin. Med. 2025, 14, 4026. https://doi.org/10.3390/jcm14124026

AMA Style

Hadweh P, Niset A, Salvagno M, Al Barajraji M, El Hadwe S, Taccone FS, Barrit S. Machine Learning and Artificial Intelligence in Intensive Care Medicine: Critical Recalibrations from Rule-Based Systems to Frontier Models. Journal of Clinical Medicine. 2025; 14(12):4026. https://doi.org/10.3390/jcm14124026

Chicago/Turabian Style

Hadweh, Pierre, Alexandre Niset, Michele Salvagno, Mejdeddine Al Barajraji, Salim El Hadwe, Fabio Silvio Taccone, and Sami Barrit. 2025. "Machine Learning and Artificial Intelligence in Intensive Care Medicine: Critical Recalibrations from Rule-Based Systems to Frontier Models" Journal of Clinical Medicine 14, no. 12: 4026. https://doi.org/10.3390/jcm14124026

APA Style

Hadweh, P., Niset, A., Salvagno, M., Al Barajraji, M., El Hadwe, S., Taccone, F. S., & Barrit, S. (2025). Machine Learning and Artificial Intelligence in Intensive Care Medicine: Critical Recalibrations from Rule-Based Systems to Frontier Models. Journal of Clinical Medicine, 14(12), 4026. https://doi.org/10.3390/jcm14124026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop