Next Article in Journal
Clinical Significance of Rotational Thromboelastometry (ROTEM) for Detection of Early Coagulopathy in Trauma Patients: A Retrospective Study
Previous Article in Journal
Utility of Serum Biomarkers of Myocardial Fibrosis in High-Gradient Severe Aortic Stenosis: An Explorative Cardiovascular Magnetic Resonance Imaging-Based Study
Previous Article in Special Issue
Comparison of Vendor-Pretrained and Custom-Trained Deep Learning Segmentation Models for Head-and-Neck, Breast, and Prostate Cancers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Enhancing Radiologist Productivity with Artificial Intelligence in Magnetic Resonance Imaging (MRI): A Narrative Review

1
Department of Diagnostic Imaging, National University Hospital, 5 Lower Kent Ridge Rd, Singapore 119074, Singapore
2
AIO Innovation Office, National University Health System, 3 Research Link, #02-04 Innovation 4.0, Singapore 117602, Singapore
3
Department of Diagnostic Radiology, Yong Loo Lin School of Medicine, National University of Singapore, 10 Medical Drive, Singapore 117597, Singapore
4
National University Spine Institute, Department of Orthopaedic Surgery, National University Health System, 1E Lower Kent Ridge Road, Singapore 119228, Singapore
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(9), 1146; https://doi.org/10.3390/diagnostics15091146
Submission received: 4 February 2025 / Revised: 6 April 2025 / Accepted: 24 April 2025 / Published: 30 April 2025
(This article belongs to the Special Issue Deep Learning in Medical Image Segmentation and Diagnosis)

Abstract

:
Artificial intelligence (AI) shows promise in streamlining MRI workflows by reducing radiologists’ workload and improving diagnostic accuracy. Despite MRI’s extensive clinical use, systematic evaluation of AI-driven productivity gains in MRI remains limited. This review addresses that gap by synthesizing evidence on how AI can shorten scanning and reading times, optimize worklist triage, and automate segmentation. On 15 November 2024, we searched PubMed, EMBASE, MEDLINE, Web of Science, Google Scholar, and Cochrane Library for English-language studies published between 2000 and 15 November 2024, focusing on AI applications in MRI. Additional searches of grey literature were conducted. After screening for relevance and full-text review, 67 studies met inclusion criteria. Extracted data included study design, AI techniques, and productivity-related outcomes such as time savings and diagnostic accuracy. The included studies were categorized into five themes: reducing scan times, automating segmentation, optimizing workflow, decreasing reading times, and general time-saving or workload reduction. Convolutional neural networks (CNNs), especially architectures like ResNet and U-Net, were commonly used for tasks ranging from segmentation to automated reporting. A few studies also explored machine learning-based automation software and, more recently, large language models. Although most demonstrated gains in efficiency and accuracy, limited external validation and dataset heterogeneity could reduce broader adoption. AI applications in MRI offer potential to enhance radiologist productivity, mainly through accelerated scans, automated segmentation, and streamlined workflows. Further research, including prospective validation and standardized metrics, is needed to enable safe, efficient, and equitable deployment of AI tools in clinical MRI practice.

1. Introduction

In the last two decades, artificial intelligence (AI), particularly machine learning (ML) and natural language processing (NLP), have significantly transformed the field of radiology [1]. AI applications have demonstrated promise in reducing the diagnostic workload of radiologists while potentially enhancing productivity and diagnostic accuracy across various modalities. Magnetic resonance imaging (MRI) plays a crucial role in modern diagnostics due to its superior soft-tissue contrast, multi-planar capabilities, and lack of ionizing radiation, making it essential for diagnosing a wide range of conditions, from neurological disorders to musculoskeletal injuries [2]. As of 2022, there are an estimated 50,000 MRI systems installed worldwide, with over 95 million MRI scans performed yearly [3]. Despite the widespread use and critical importance of MRI in clinical diagnostics, there has been limited systematic assessment of how AI applications affect clinical productivity within MRI workflows across various specialties. Despite novel advancements in deep learning (DL) for computer imaging with increased emphasis on quality optimization whilst maintaining accuracy [4,5,6,7,8], it remains to be seen if these can be translated into medical imaging (specifically MRI) whilst optimizing time gains or losses.
The complexity of MRI interpretation significantly contributes to radiologists’ increased workload and the potential for diagnostic delays. Radiologists are tasked with analyzing large volumes of intricate images—often in the thousands—requiring the analysis of one image every 3–4 s during a typical 8 h workday to meet workload demands [9]. Furthermore, interobserver variability can result in inconsistent diagnoses, underscoring the need for tools that standardize and improve diagnostic accuracy [10]. AI holds promise in addressing these challenges by automating routine tasks, optimizing scan protocols, and aiding in image interpretation, ultimately enhancing workflow efficiency and reducing diagnostic errors [2,11].
Previous reviews have primarily focused on specific subspecialties or examined AI applications across multiple imaging modalities, often offering limited insight into the implications of AI specifically within MRI [12,13,14,15]. The heterogeneity in study designs, endpoints, and performance metrics makes it difficult to conduct a formal meta-analysis with robust reasoning [16]. Given the wide variations in study designs, AI models, clinical settings, and outcome measures, a narrative review is more suitable. This approach allows for the synthesis of findings and insights despite the lack of uniformity, providing a flexible framework to explore the impact of AI on MRI.
Measuring productivity in radiology is multifaceted. While time savings and increased throughput are often-cited benefits of AI, other aspects such as diagnostic accuracy, workflow efficiency, and reduction in radiologist fatigue are equally important [17,18]. However, there is no clear consensus on which metrics should be used to assess these outcomes. Studies often focus on specific applications of AI—such as lesion detection, scan protocolling and scan prioritization—which limits the ability to make an overall assessment of AI’s contribution to productivity [19,20,21]. Furthermore, AI models developed in specific institutions or datasets may not generalize well to broader clinical practice, raising concerns about bias and reproducibility [22,23]. Ethical considerations, including data privacy and the need for transparency in AI algorithms, also play a significant role in adoption of AI in MRI [24]. Radiologist acceptance of AI tools depends on their reliability and ability to integrate seamlessly into existing workflows, highlighting the importance of collaborative efforts between AI developers and clinicians [25].
This review aims to provide a comprehensive overview of AI applications in MRI across different clinical specialties, with a particular focus on their potential to enhance radiologist productivity during scanning and/or interpretation. We will examine studies that assess AI tools in terms of time savings, accuracy improvements, and workflow integration, while also identifying gaps in the literature where further research is needed. By addressing the current landscape and identifying areas for improvement, we hope to inform future research and guide the practical integration of AI tools in MRI.

2. Materials and Methods

2.1. Literature Search Strategy

In order to enhance the transparency and completeness of our review, we employed a systematic search with our narrative review methodology. Initial search strategy was performed according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines [26].
On 15 November 2024, a literature search of the major databases was conducted (PubMed, EMBASE, MEDLINE, Web of Science, Google Scholar, and Cochrane library) using the following strategy: (productivity OR productiveness OR prolificacy) AND (radiology OR radiologist) AND (artificial intelligence OR AI) AND (magnetic resonance imaging OR MRI). Limits were applied to only include English language studies published between the year 2000 and 15 November 2024.
Additional grey literature searches have also been conducted via hand search (OATD, OpenGrey, OCLC, NIH Clinical Trials, TRIP medical database).

2.2. Study Screening and Selection Criteria

A two-stage screening process was used. Studies were first screened by title and abstract (A.N). A full text review was then performed for any potentially eligible studies. Any controversies at either stage were reviewed by multiple authors (W.O., A.L, J.T.P.D.H) during assessment for study eligibility.
The inclusion criteria were as follows: studies focusing on the topic of using AI in MRI to enhance productivity, English studies, and studies performed on human subjects. The exclusion criteria were as follows: non-original research (for example, review articles and editorial correspondences), unpublished work, non-peer reviewed work, conference abstracts, and case reports. Studies focused primarily on other imaging modalities (for example radiographs, CT, or nuclear medical imaging) were also excluded.

2.3. Data Extraction and Reporting

The selected studies were extracted and compiled onto a spreadsheet using Microsoft Excel Version 16.81 (Microsoft Corporation, Washington, DC, USA). The following data were extracted:
  • Study details: title, authorship, year of publication and journal name.
  • Application and primary outcome measure.
  • Study details: sample size, region studied, MRI sequences used.
  • Artificial intelligence technique used.
  • Key results and conclusion.

3. Results

3.1. Search Results

Our initial literature search identified 630 studies following screening to the specified criteria and removal of any duplicates. Subsequently, 134 studies which did not meet the date range, 215 of incorrect article types (as specified in the exclusion criteria), and 4 non-English publications were initially excluded during screening by title and abstract alone. This led to 277 studies being selected for full screening. Four additional studies were also identified from citation searching and assessed for eligibility during the full screening. The end result was the inclusion of 66 studies in the present review (see Figure 1 for a detailed flowchart). The articles were summarized based on data extracted in Table 1 and Table 2.
We classified the included studies based on the following themes: Reducing scan times to improve efficiency (13/66, 19.7%), automating segmentation (26/66, 39.4%), optimizing worklist triage and workflow processes (7/66, 10.6%), decreasing reading times (16/66, 24.2%), and time-saving and workload reduction (6/66, 9.1%). In terms of overall productivity gains, around half of the studies (35/66, 53.1%) demonstrated this clearly across one or more of the themes identified. The remaining studies were deemed to have unclear productivity gains based on the limitations highlighted in Table 2 and subsequent discussion, with a few productivity losses (3/66, 4.5%) due to the use of AI also seen.

3.2. Overview of How Can AI Help in MRI Interpretation

Artificial intelligence (AI) refers to the use of computers and technology to perform tasks that mimic human intelligence and critical thinking [91]. First introduced at the 1956 Dartmouth Conference, AI gained significant traction in medicine during the third AI boom in the 2010s, driven by advancements in machine learning (ML) and deep learning (DL) [92]. ML involves training models to make predictions using existing datasets, while DL employs artificial neural networks modeled on the brain’s architecture to solve complex problems [93,94]. The exponential growth in digital imaging and reporting data have further accelerated advancements in medical imaging, particularly through the application of convolutional neural networks (CNNs), a key subset of DL. CNNs have revolutionized medical imaging by enabling the development of computer-aided diagnostic (CAD) systems, which assist radiologists in detecting lesions and tumors across various imaging modalities, including MRI [95]. These systems aim to enhance diagnostic accuracy and efficiency, reducing workload and interobserver variability. Many CAD systems rely on CNN models to deliver expert-level performance across different areas of imaging research [96]. The relationship between AI, ML, DL, and CNN is depicted in Figure 2, highlighting the interconnected nature of these technologies and their role in advancing medical imaging.
While AI has demonstrated success in improving accuracy, enhancing efficiency is equally critical to facilitate the seamless integration of these models into clinical practice [97]. The clinical application of AI in MRI can be broadly categorized into five key steps: image acquisition, image processing, task-specific applications, reporting, and integrated diagnostics. This workflow is further illustrated in Figure 3.

3.3. Reducing Scan Times to Improve Efficiency

Implementing more efficient protocols and eliminating unnecessary sequences can significantly reduce scan times, streamlining the initial stages of image acquisition and processing. This, in turn, eases the pressure on subsequent steps, such as reducing MRI reading times, ultimately helping to alleviate radiologist fatigue.
One key example is the improvements in AI-driven lung morphometry using diffusion-weighted MRI (DW-MRI). A deep cascade of residual dense networks (DC-RDN) has enabled the rapid production of high-quality images [32]. This approach reduced the required breath-holding time from 17.8 s to 4.7 s, with each slice (64 × 64 × 5) reconstructed within 7.2 ms.
In one study [88] exploring an MRI acceleration framework to reconstruct dynamic 2-D cardiac MRI sequences from undersampled data, a deep cascade of convolutional neural networks (CNNs) was successfully used to accelerate data acquisition. This approach resulted in improvements in both image quality and reconstruction speed. Each complete dynamic sequence was reconstructed in under 10 s, with each 2-D image frame taking only 23 ms. Faster reconstruction enables more timely image availability for diagnosis and intervention, directly contributing to increased radiologist throughput.
For assessing hepatobiliary phase (HBP) adequacy in gadoxetate disodium (EOB)-enhanced MRI, a convolutional neural network (CNN) was shown to evaluate HBP adequacy in real time, potentially reducing examination times by up to 48% for certain patient groups [40]. In gastric cancer workups, a deep learning-accelerated single-shot breath-hold (DSLB) T2-weighted imaging technique demonstrated superior quality, faster acquisition times, and improved staging accuracy compared to current BLADE T2-weighted imaging methods [46]. DSLB reduced the acquisition time of T2-weighted imaging from a mean of 4 min 42 s per patient to just 18 s per patient while delivering significantly better image quality, with 9.43-fold, 8.00-fold, and 18.31-fold improvements in image quality scores across three readers compared to the standard BLADE.
In spinal MRI, one study [56] utilized a deep learning-based reconstruction algorithm combined with an accelerated protocol for lumbar MRI, reducing the average acquisition time by 32.3% compared to standard protocols while maintaining image quality. A separate study [55] on lumbar central canal stenosis (LCCS) demonstrated improved measurement accuracy using sagittal lumbar MR slices alone, compared to models that combined sagittal and axial slices. This approach showed robustness across moderate-to-severe LCCS cases, suggesting that automated LCCS assessment using sagittal T2 MRI alone could reduce the need for additional axial imaging, thereby shortening scan times.
However, limitations remain. Prior studies on DL techniques for MRI acceleration [98,99] did not specify improvements across all full spine MRI examination sequences, and their clinical utility was constrained by the lack of multicenter validation. Despite these caveats, DL-based reconstruction in musculoskeletal (MSK) MRI offers a promising direction for reducing examination times. For instance, one study [100] demonstrated that a 10 min accelerated 3D SPACE MRI protocol for the knee delivered comparable image quality and diagnostic performance to a 20 min 2D TSE MRI protocol.
A novel study [89] on MRI of the knee introduced a variational network framework that integrates data consistency with deep learning-based regularization, enabling the reconstruction of high-quality images from undersampled MRI data. This method outperformed non-AI reconstruction algorithms by a factor of 4 and achieved reconstruction times under 193 ms on a single graphics card, requiring no parameter tuning once trained. These characteristics make it easily integrable into clinical workflows.
DL models [19,66,67,68] have also been developed to accelerate reconstruction times for various brain MRI sequences while maintaining accuracy and improving image quality. These models enhance the signal-to-noise ratio (SNR) and reduce artifacts, such as those caused by head motion. In one study [69], an accelerated DL reconstruction method improved image quality and reduced sequence scan times by 60% for brain MRI.
Not all studies demonstrated reduced scan times despite improvements in image quality. For instance, a DL-based reconstruction model for intracranial magnetic resonance angiography (MRA) significantly improved SNR and contrast-to-noise ratio (CNR) for basilar artery analysis [76], but increased scan times, potentially affecting clinical productivity. Similarly, for shoulder MRI, while DL-based reconstructions significantly improved SNR and CNR, they showed inconsistencies in reducing scan times when applied [60,61].

3.4. Automating Segmentation

A significant majority of the included studies focused on automating segmentation in MRI across various specialties. Manual segmentation is often time-consuming and labor-intensive for radiologists [50]. AI models provide a promising solution by easing this workload and enabling radiologists to deliver more concise and time-efficient imaging-based diagnoses.
Multiple studies [27,28,29] focusing on breast tumor have demonstrated the successful development of AI programs for automating tumor segmentation in dynamic contrast-enhanced MRI (DCE-MRI). One such program achieved a 20-fold reduction in manual annotation time for expert radiologists, while another showed the potential to reduce unnecessary biopsies of benign breast lesions by up to 36.2%. Automated tumor segmentation not only enhances productivity but also minimizes interobserver variability, leading to more consistent and accurate diagnoses.
For lung upper airway segmentation in static and dynamic 2D and 3D MRI, a study utilizing the VGG19 architecture [33] demonstrated a deep learning segmentation system that was statistically indistinguishable from manual segmentations, accounting for human variability. This represents a significant improvement over earlier models using classic augmentation, enabling more efficient structure identification. However, time savings were not detailed, leaving questions about overall efficiency in broader clinical applications.
In head, neck, and lung oncology, AI-based approaches to radiation planning have shown significant time savings with automated segmentation models combing various imaging modalities, including MRI [90,101]. These models eliminated labor-intensive, observer-dependent manual delineation of organs-at-risk (OARs). One study reported average model runtimes of 56 s and 3.8 min, compared to 12.4 min for clinical target volume (CTV) delineation and 30.7 min for OAR delineation by junior radiation oncologists. These time reductions allow radiologists and radiation oncologists to focus on complex decision-making tasks and patient care.
Studies on hepatocellular carcinoma detection and liver segmentation [34,35,36,102] have shown that radiomic and deep learning approaches can automate segmentation and delineation of pathological subtypes across MRI sequences, achieving 10- to 20-fold improvements in measurement time compared to manual methods.
AI pipelines for MRI have also demonstrated potential in automating segmentation and reducing contouring time for renal lesions [47,48]. In one study, combining MRI and AI for renal lesion differentiation during follow-up resulted in cost savings over 10 years, with total estimated costs of $8054 for the MRI strategy and $7939 for the MRI + AI strategy. This approach also yielded marginal gains in quality-adjusted life years (QALYs), demonstrating potential for reallocating radiologist resources to other tasks. Additionally, a novel automated deep learning-based abdominal multiorgan segmentation (ALAMO) technique [49] successfully segmented multiple OARs for MR-only radiation therapy and MR-guided adaptive radiotherapy. This technique offers potential time savings and reduces fatigue by replacing manual delineation with comparable accuracy.
For shoulder pathologies such as rotator cuff tears, deep learning frameworks utilizing parameters like occupation ratio and fatty infiltration [58] have achieved fast and accurate segmentation of the supraspinatus muscle and fossa from shoulder MRI. These frameworks enhance diagnostic efficiency and objectivity, with some performance metrics surpassing those of experienced radiologists [59].
In brain MRI segmentation, skull stripping is a critical preprocessing step that removes non-brain tissue (e.g., skull, skin, and eyeballs) to reduce computational burden and misclassification risks, enabling more targeted analysis of brain tissue [103]. One deep learning model using anatomical context-encoding network (ACEnet) [64] could segment an MRI head scan of 256 × 256 × 256 within approximately 9 s on a NVIDIA TITAN XP GPU, facilitating potential real-time applications. The model achieved Dice scores greater than 0.976. Another application [65] focused on cerebrospinal fluid (CSF) volume segmentation, demonstrating strong reliability in measuring age-related brain atrophy. This application automatically segmented brain and intracranial CSF and measured their volumes and volume ratios within 1 min, using a popular 3D workstation in Japan. For brain pathologies, a CNN model [70] accurately segmented meningiomas and provided volumetric assessments during serial follow-ups, achieving a median performance of 88.2% (within the inter-expert variability range of 82.6–91.6%). The study also reported a 99% reduction in segmentation time, averaging 2 s per segmentation (p < 0.001).
Automated segmentation has also enabled assessments of generalized parameters, such as tissue volume and distribution. One deep learning pipeline [62] automated measurements across various landmarks in the lower limbs. While this study reported time gains rather than savings, the potential for improved clinical value in diagnosing or monitoring lipoedema or lymphedema represents a potential productivity enhancement. For rarer hand tumors, a CNN-based segmentation system [63] improved tumor analysis on hand MRIs, though its accuracy (71.6%) was limited by a small sample size and narrow tumor selection.
Advances in image segmentation have extended to niche subspecialties in MRI. In neonatology, models integrating clinical variables (e.g., gestational age and sex) with automated segmentation techniques [77,78,79] have been used to screen for pathologies such as hypoxic–ischemic encephalopathy (HIE) and predict neurodevelopmental outcomes up to two years later. In nasopharyngeal carcinoma (NPC) treatment, a deep learning-based segmentation model [80] accurately delineated gross tumor volume (GTV), surpassing previous methods and improving clinicians’ efficiency in creating radiation plans for targeted therapies.
AI models have also been applied to sleep apnea evaluation. A study [22] using deep learning-based segmentation of the pharynx, tongue, and soft palate on mid-sagittal MRI efficiently identified sleep apnea-related structures. Results were consistent with intra-observer variability, and the model’s processing time of only 2 s on a modern GPU demonstrates its potential for real-time clinical applications, such as computer-aided reporting.

3.5. Optimizing Worklist Triage and Workflow Processes

A multicenter study [30] developed an automated sorting tool for cardiac MRI examinations by linking examination times with appropriate post-processing tools. This tool directed images that did not require quantitative post-processing to PACS more quickly for interpretation. By streamlining the workflow, radiologists received cardiac MRI images sooner, enhancing clinical efficiency. However, the tool demonstrated poor accuracy when handling images with high variability in acquisition parameters, such as perfusion imaging across different centers, highlighting a key limitation.
AI has also made significant progress in prostate MRI, particularly in the early detection and classification of prostate cancer [45]. One proposed model [42] aimed to optimize worklist triage by automating the grading of extra-prostatic extension using T2-weighted MRI, ADC maps, and high b-value DWI sequences. This approach shows promise for improving workflow efficiency in prostate cancer evaluations.
In the workup for hepatocellular carcinoma (HCC), an AI model demonstrated the ability to differentiate between pathologically proven HCC and non-HCC lesions, including those with atypical imaging features on MRI [36]. This CNN-based model achieved an overall accuracy of 87.3% and a computational time of less than 3 ms, suggesting its potential as a clinical support tool for prioritizing worklists and streamlining MRI workflows.
Workflow efficiencies can also be achieved through improved and automated protocoling of MRI studies. A recent study [104] using a secure institutional large language model demonstrated significantly better clinical detail in spine MRI request forms and an accurate protocol suggestion rate of 78.4%. Another study [86] on musculoskeletal MRI evaluated a CNN-based chatbot-friendly short-text classifier, which achieved 94.2% accuracy. The study suggested extending this text-based framework to other repetitive radiologic tasks beyond image interpretation to enhance overall efficiency.
For breast malignancy evaluation using MRI, artificial intelligence has shown great potential in improving diagnostic accuracy and workflow efficiency. A notable multicenter study [87] utilized machine learning algorithms to identify axillary-tumor radiomic signatures in patients with invasive breast cancer. This approach successfully detected axillary lymph node metastases at early stages, which is crucial for surgical decision-making. Integrating such AI tools into clinical workflows could help radiologists prioritize scans more effectively and reduce the need for additional surgical interventions in select patients.
Beyond assisting with report generation, several studies have targeted workflow improvements in integrated diagnostics, including decision support and outcome prediction. While these advancements currently have limited impact on routine reporting practices, the growing demand for real-time data and follow-up reporting could significantly burden radiologists if workflow efficiencies are not implemented [105].
Various AI models [37,38] have also demonstrated success in predicting mortality outcomes and early recurrence pre- and post-surgery. These models identified deep learning signatures linked to risk factors such as increased microvascular invasion and tumor number, outperforming conventional clinical nomograms in accuracy. However, most studies did not discuss how these innovations could be seamlessly integrated into clinical practice.
One AI model [39] focusing on hepatocellular carcinoma (HCC) described an automated risk assessment framework with a runtime of just over one minute per patient (0.7 s for automated liver segmentation, 65.4 s for extraction of radiomic features, and 0.42 s for model prediction). If implemented, such frameworks could enhance clinical reporting while maintaining workflow efficiency.

3.6. Decreasing Reading Times

Reading times can vary significantly depending on procedure type, individual radiologist, and specialty [106]. However, reducing reading time is a straightforward way to measure increased efficiency achieved by radiologists in a typical clinical workflow, assuming workload and report accuracy remain broadly consistent.
The use of AI has advanced to improve reporting times for post-processed images across various specialties. One CNN-based model [31], used for aortic root and valve measurements, achieved an average speed improvement of 100 times compared to expert cardiologists in the UK during image interpretation.
In spinal MR imaging, AI has brought multifaceted improvements, particularly in tasks where measurements by reporting radiologists are repetitive and time-consuming. One study [50] introduced a deep learning model for detecting and classifying lumbar spine central canal and lateral recess stenosis, achieving accuracy comparable to a panel of expert subspecialized radiologists. A subsequent study [51] focusing on productivity metrics for this model demonstrated significant reductions in reporting times. Similar performance metrics have been observed in other DL models [52,53,54] across various centers using different training datasets, such as SpineNet and Deep Spine. However, a key limitation is that time savings, if any, were not always clearly specified.
In musculoskeletal (MSK) imaging, AI applications have also reduced reading times, particularly for detecting ligament and meniscal injuries in the knee. A multicenter study [57] evaluating an AI assistant program reported marked improvements in accuracy (over 96%) and time savings in diagnosing anterior cruciate ligament (ACL) ruptures on MRI, particularly among trainee radiologists. Diagnostic time for trainees improved to 10.6 s (from 13.8 s) for routine tasks and 13.0 s (from 37.5 s) for difficult tasks, compared to 8.5 s (from 10.5 s) and 13.6 s (from 23.3 s), respectively, for expert radiologists.
In neuroradiological reporting of multiple sclerosis (MS) lesions, prior deep learning methods were limited due to error susceptibility. A multicenter study [85] introduced a novel automation framework called “Jazz”, which identified two to three times more new lesions than standard clinical reports. Reading time using “Jazz” averaged 2 min and 33 s, though the study did not compare this to the time required for a standard radiologist-generated report without AI assistance. Another multicenter study [72] tested a deep learning model to automatically differentiate active from inactive MS lesions using non-contrast FLAIR MRI images. The model achieved an average accuracy of up to 85%, but a significant limitation was its reliance on manually selected regions of interest (ROIs) already diagnosed and categorized by radiologists, thereby restricting productivity gains in decreased reading time. In other neurological presentations, various AI models [73,74,75] have demonstrated improvements in MRI related to Parkinson’s disease, with fully automated networks achieving accuracies of up to 93.5% and processing times of under one minute to distinguish between patients with Parkinson’s disease and healthy controls.
For prostate MRI scans, a DL-based computer-aided diagnosis (CAD) system [81] improved diagnostic accuracy for suspicious lesions by 2.9–4.4% and reduced average reading time by 21%, from 103 to 81 s. However, a study [82] examining an AI clinical decision support tool for brain MRI found significant improvements in performance among non-specialists in diagnosis, producing top differential diagnoses, and identifying rare diseases, while specialist performance showed no significant differences with or without the AI tool. This underscores the need to benchmark AI models against expert performance to justify their integration into radiologists’ workflows.
In prostate multiparametric MRI (mpMRI), a retrospective multicenter study [43] reported a significant reduction in radiologist reading time (by 351 s) when grading was AI-assisted. Another study [41] on prostate segmentation showed that, for prostate cancer diagnosis, AI-based CAD methods improved consistency in internal (κ = 1.000; 0.830) and external (κ = 0.958; 0.713) tests compared to radiologists (κ = 0.747; 0.600). The AI-first read (8.54 s/7.66 s) was faster than both the readers (92.72 s/89.54 s) and the concurrent-read method (29.15 s/28.92 s). Nonetheless, there is criticism that the training datasets currently used to validate deep learning CAD models for prostate cancer workup are inadequate. Insufficient training data have been shown to result in worse performance compared to expert panels of radiologists [44].

3.7. Time-Saving and Workload Reduction

Several studies have explored radiologist time-saving and workload reduction strategies that extend beyond previously discussed themes.
A deep learning model [20] utilizing time-resolved angiography with stochastic trajectories (TWIST) sequences was developed to identify normal ultrafast breast MRI examinations and exclude them from the radiologist’s workload when integrated into the workflow. The trained model achieved an AUC of 0.81, reducing workload by 15.7% and scan times by 16.6% through elimination of unnecessary additional sequences. This demonstrates AI’s potential to streamline imaging protocols and enhance workflow efficiency.
To address challenges in obtaining high-quality training data for AI models, one study [83] evaluated manual prostate MRI segmentation performed by radiology residents. The study found that residents achieved segmentation accuracy comparable to experts, suggesting that residents could be employed to create training datasets. This approach could alleviate work pressures on expert radiologists while maintaining the high-quality data necessary for AI development.
Exploring broader applications, a CNN [71] addressed limitations in learning from small or suboptimal MRI datasets using a progressive growing generative adversarial network augmentation model. By generating realistic MRIs of brain tumors, the model enabled accurate classification of gliomas, meningiomas, and pituitary tumors, achieving an accuracy of 98.54%. This augmentation model offers a novel method to reduce the burden on radiologists by minimizing the need to manually generate large training datasets for deep learning frameworks.
In spine MRI, a study [84] investigated the effectiveness of AI-generated reports. The study reported improvements in patient-centered summaries and workflow efficiency by assisting radiologists in producing concise reports. However, the technology also presented challenges, with 1.12% artificial hallucinations and 7.4% potentially harmful translations identified. These error rates highlight that, without optimization, the current implementation of such technology could increase radiologists’ workload, as they would need to verify AI-generated reports before sharing them with patients.

4. Discussion

This review highlights the wide-ranging AI solutions applied to MRI across multiple specialties. In particular, many studies focused on accelerating scan times, automating segmentation, optimizing workflows (e.g., MRI form vetting), and reducing radiologist reading times, with each contributing to improved productivity in different ways. In the following sections, we examine these findings in greater detail, beginning with a discussion of the dominant AI architectures employed. We then address the strengths and limitations of the current evidence, delve into the ethical and regulatory framework which is essential for real-world adoption, and conclude with future directions and recommendations that emphasize broader validation, feedback loops, and next-generation AI modalities.

4.1. Dominant AI Architectures and Approaches

In our review, the studies employed a range of AI-based techniques, with convolutional neural network (CNN) architectures being the most frequently utilized. CNNs can be customized [95] for a wide array of tasks, including image processing, lesion detection, and characterization. Notable architectures used in the models include ResNet, DenseNet, and EfficientNet, which address specific limitations of earlier CNN models. ResNet, for instance, mitigates the vanishing gradient issue, while EfficientNet reduces computational and memory loads, leading to enhanced performance in both accuracy and efficiency [20,47,48,54,72,107]. Additionally, U-Net, a widely adopted CNN-based architecture, enables rapid and precise image segmentation and is commonly integrated into studies focused on automating segmentation tasks in MRI.
Beyond CNNs, other AI-based techniques, including some emerging technologies, were also implemented. For example, certain models utilized machine learning-driven intelligent automation software [85]. Classical machine learning (ML) methods continue to excel in simpler feature extraction tasks and often demonstrate more consistent performance across certain imaging modalities [108]. However, DL-based neural networks, particularly CNNs, have been shown to outperform classical ML approaches in most MRI-related applications [109]. A key advancement driving the success of DL is the integration of self-supervised learning (SSL).
Unlike traditional supervised learning, which relies on manually labeled datasets for “ground truth” supervisory signals, SSL enables models to generate implicit labels from unstructured data. By leveraging pretext tasks and downstream tasks, SSL-based ML models can effectively analyze large volumes of unlabeled medical imaging data while reducing the need for extensive human fine-tuning [110]. This approach has facilitated previously impractical developments, such as automated anatomical tracking in medical imaging [111]. Beyond deep learning-based applications, SSL has also been incorporated into novel ML frameworks for MRI, including dynamic fetal MRI, functional MRI for autism and dementia diagnosis, and cardiac MRI, all of which have demonstrated promising results [112,113,114]. Although SSL may not directly enhance daily clinical workflow efficiency or reduce workload, it has the potential to accelerate the development of novel ML imaging techniques aimed at improving these aspects.
Another significant development since 2017 is the adoption of Transformer-based architectures for imaging tasks. Originally designed for sequence-to-sequence prediction, the Transformer architecture employs a self-attention mechanism that dynamically adjusts to input content, offering notable advantages over traditional CNNs in image processing tasks [115]. This innovation has also contributed to the emergence of large language models (LLMs), such as OpenAI’s Generative Pre-Trained Transformer (GPT), which leverage Transformer architecture and extensive text-based corpora to perform a wide range of natural language processing (NLP) tasks. Other foundational LLMs, including Bidirectional Encoder Representations from Transformers (BERT), have served as precursors to domain-specific models like Med-PaLM, which focuses on healthcare applications [116]. Unlike earlier ML architectures that relied on either SSL or unsupervised learning, LLMs primarily utilize SSL to derive labels from input data.
While LLMs have gained traction in various fields, including medical chatbot development [117], their application in MRI remains limited. Among the reviewed studies, only one implemented an LLM-based approach for MRI analysis [84], highlighting the need for further exploration in this domain. Future research could investigate the potential of LLMs to automate text-intensive radiology tasks, such as report generation and clinical protocol, thereby enhancing efficiency in radiological workflows.

4.2. Strengths and Limitations of Current Evidence

Some studies included in this review were sourced outside the initial search strategy to allow a more comprehensive discussion of certain topics. Because the journal archives and grey literature sources were limited to those commonly available in public or academic settings, the search was not exhaustive. Consequently, AI applications under vendor development or within non-commercial settings may have been overlooked. In addition, by adopting a broad scope that examines AI developments in MRI across multiple specialties, this review may have sacrificed depth in favor of breadth compared to more narrowly focused analyses.
The heterogeneity observed in the included research limited the feasibility of conducting a formal meta-analysis, thereby reducing the ability to draw robust, statistically aggregated conclusions. Variations in study design, patient population data, and MRI protocols (across either training and testing datasets) in turn reduce the reproducibility of the models discussed. Through issues such as data leakage, the same results are prevented from being obtained under conditions defined by other clinical settings [118]. With that said, the majority of studies (52/66, 78.8%) described attempts at internal cross-validation using different splits of training and test datasets. This would have helped prevent overfitting, thereby assisting in more reliable model generalization [119]. Apart from discrepancies in data use or preprocessing, computational costs and hardware variations are major sources of irreproducibility. Our included studies either did not discuss this information, or failed to simulate results using lower performance hardware or software. This posed a challenge in making direct comparisons between them, or deriving relevance for clinical practice.
A lack of external validation characterized the majority of included studies (48/66, 72.7%). This is not an uncommon limitation, with a 2019 meta-analysis [120] reporting that only 6% of 516 studies on novel AI applications in medical imaging included external validation as a key component. This is despite the fact that such validation is a definitive step to ensuring that models can generalize to diverse clinical settings, patient populations, and imaging protocols. In addition, aside from one study [48], the long-term impact and cost-effectiveness of AI applications were not thoroughly compared across different specialties.
Finally, the retrospective nature of many studies raises questions about their applicability to real-world clinical settings [121]. The potential sources of bias, including differences in patient demographics and potential conflicts of interest, were also not discussed in detail, which raises questions about the fairness and equity of certain AI solutions. Although this review sought to highlight productivity metrics enhanced by AI, most studies reported outcomes related primarily to time savings and diagnostic accuracy. Radiologist-specific productivity indicators, including reductions in repetitive tasks or improved workflow satisfaction, were often relegated to secondary outcomes, complicating cross-study comparisons.

4.3. Recommended Enhancements

4.3.1. Regulatory Frameworks and Compliance

Regulatory and compliance factors are crucial for a successful and consistent translation of AI tools into routine MRI practice. Concerns related to patient privacy, data security, and algorithmic transparency must be balanced with the potential productivity gains [122]. Regulatory guidelines established by organizations such as the United States Food and Drug Administration (FDA) and CE marking often mandate rigorous validation of AI algorithms, including prospective trials and post-market surveillance [123,124]. However, significant heterogeneity persists among vendors regarding the deployment, pricing, and maintenance of these AI products for specific clinical applications [125]. This variability poses challenges for stakeholders in selecting the most appropriate AI solution.
To address these complexities, several radiological organizations, including the American College of Radiology (ACR), Canadian Association of Radiologists (CAR), European Society of Radiology (ESR), Royal Australian and New Zealand College of Radiologists (RANZCR), and Radiological Society of North America (RSNA), have released joint guidelines to support the development, implementation, and use of AI tools [126]. These guidelines prioritize patient safety, ethical considerations, seamless integration with existing clinical workflows, and the stability of AI models through ongoing monitoring. Nevertheless, as these guidelines do not directly govern AI products or vendor selection, the responsibility falls on individual radiology departments or institutions to establish tailored frameworks. These frameworks must enable the selection and integration of suitable AI solutions while adhering to region-specific data standards [127].

4.3.2. Generalizability and Multi-Center Validation

To ensure fairness and equity in AI-driven MRI applications, future research should prioritize minimizing biases related to demographics and disease prevalence. For smaller institutions that may lack the resources to compile large datasets, multicenter collaboration and external validation are crucial. Although many AI models achieve high accuracy within their home institutions, differences in imaging protocols, sequences, and MRI hardware often remain underexplored, leading to model underspecification and limited generalizability [128,129]. By incorporating multicenter validation, models can be trained on larger, more diverse datasets from multiple regions, enabling the development of algorithms that adapt to varying clinical populations and imaging protocols. In cases where legislative constraints hinder multicenter data sharing, synthetic data augmentation offers an alternative strategy to enhance dataset diversity. Publicly available, anonymized datasets, such as PROSTATEx, have already been utilized in some of the reviewed studies [70,81] and were recently combined with regional datasets to facilitate a novel multicenter, multi-scanner validation study for multiparametric MRI in prostate cancer [130].
Another approach to improving generalizability is the adoption of a federated learning (FL) framework, which has gained increasing recognition in medical AI research [131]. FL enables institutions to collaboratively develop machine learning models while maintaining data privacy, as each institution retains local control over its dataset and model updates are orchestrated through a centralized server. This framework allows institutions to selectively share or receive data within a trusted execution environment. However, FL presents challenges, including concerns about data heterogeneity and the protection of sensitive medical information [132]. Additionally, stress testing is an effective strategy to address model underspecification and improve robustness. By simulating variations in imaging parameters, introducing artificial noise, and modifying datasets to mimic real-world conditions, stress testing can identify model vulnerabilities. While studies on stress testing in MRI-based AI remain limited, findings from other imaging domains suggest its potential benefits [133].
From an implementation perspective, several infrastructure-related challenges must be addressed to enable real-world AI deployment in MRI. Many hospitals lack in-house machine learning expertise, necessitating collaborations with external research groups or commercial organizations. These partnerships often rely on third-party cloud-based computing systems, raising concerns regarding data security and privacy [134]. Furthermore, seamless integration of AI tools into existing radiology information systems (RIS) and picture archiving and communication systems (PACS) is essential for widespread adoption. Given the complexity and variability of radiology IT infrastructures, the development of open-source, vendor-agnostic PACS-AI platforms holds promise in addressing these integration challenges [135,136].
Beyond technical considerations, adequate human IT support and radiologist training are critical to ensuring the successful implementation of AI in clinical practice. Without appropriate infrastructure and user engagement, even the most advanced AI models may struggle to demonstrate tangible benefits in real-world settings. A recent systematic review [137] highlighted a significant gap in AI education among radiologists and radiographers, despite widespread enthusiasm for its adoption. This challenge is further exacerbated by the “black box” nature of certain AI models, which obscure decision-making processes and hinder practitioners’ ability to validate outputs [138]. Moreover, standardized training programs for imaging professionals in AI applications remain largely absent [139]. Increasing support for “white box” AI systems, which provide greater interpretability, alongside the development of dedicated educational platforms, could enhance transparency and trust. By equipping healthcare professionals with the knowledge to assess AI-generated recommendations and communicate their implications effectively, these efforts could ultimately foster greater public confidence in AI-driven medical technologies.

4.3.3. Ethical Considerations

The expansion of digital healthcare has heightened concerns regarding ethical data usage and patient confidentiality. As hospitals and imaging centers rapidly upgrade IT systems to integrate AI tools, cybersecurity measures sometimes fail to keep pace. This has led to well-documented breaches of medical record databases and instances of fraudulent claims, exacerbating patient mistrust in AI-driven healthcare solutions [122]. In response, some regions have incorporated ethical AI guidelines into their legal frameworks [140]. For instance, the European Parliament’s AI Act, adopted in 2024 and set for full enforcement by 2026, complements existing legislation such as the General Data Protection Regulation (GDPR) and the Data Act. The AI Act establishes requirements for safety, reliability, and individual rights while fostering innovation and competitiveness. As the continued expansion of AI in healthcare relies on access to large-scale patient datasets, implementing robust regulatory frameworks is essential to maintaining public trust.
Hospitals and imaging centers must establish clear strategies for integrating AI-driven workflow changes to optimize efficiency while maintaining high standards of patient care. One key approach involves redesigning tasks to allow radiologists to focus on complex image interpretation and direct patient interactions while delegating routine or repetitive processes to automated AI tools [141]. However, for successful implementation, radiologists should play a central role in defining AI system performance parameters and must have the ability to pause or override the system if clinical targets are not met [142]. Ensuring transparency and control over AI-assisted decision-making will be essential for fostering trust and adoption among healthcare professionals.
Additionally, defining clear roles, responsibilities, and liabilities for clinicians involved in AI-integrated workflows is crucial for smooth and effective implementation [143]. AI adoption is expected to contribute to a rising clinical workload, making it imperative that AI-driven workflows align seamlessly with existing administrative and logistical processes [144]. In many cases, these supporting processes may also require AI-driven optimizations to prevent additional administrative burdens on radiologists and radiographers. A well-coordinated, symbiotic relationship between AI automation and hospital workflow management will be essential to maximizing the benefits of AI in radiology while ensuring operational efficiency and clinician satisfaction.

4.3.4. Current Recommendations

The prominence of segmentation among the included studies is unsurprising, given that it represents one of the earliest applications of AI in medical imaging [145]. However, a key limitation in comparing these studies is the heterogeneity of computational metrics and the lack of standardization for clinical implementation. Additionally, over a third of the segmentation studies (10/26, 38.5%) did not report reasoning time alongside accuracy improvements, limiting the ability to assess their potential clinical applicability. Despite these limitations, common trends in convolutional neural network (CNN) usage were observed across the included segmentation studies. UNet, an encoder–decoder CNN with skip connections and a fully convolutional architecture, enables the preservation of spatial information and achieves high precision in pixel-level segmentation. As a result, it has become the mainstream choice for medical image segmentation [146].
The majority of the included segmentation studies employed UNet, either in customized forms or in combination with other CNNs for specific segmentation tasks. One study focusing on liver segmentation [102] utilized UNet++, an enhanced variant of UNet that incorporates denser skip connections between different stages to reduce the semantic gap between the encoder and decoder. Although this approach improved efficiency, the dice similarity coefficient (DSC) for liver tumor segmentation remained below 0.7, indicating room for further accuracy improvements. Two studies that demonstrated overall productivity gains [47,62] applied U-Net architectures with an EfficientNet encoder. EfficientNet encoders require up to eight times fewer parameters compared to similar networks, making them well-suited for building efficient models with limited computational resources [147]. Despite these resource optimizations, both studies maintained high segmentation accuracy, reporting strong DSC scores between 0.97 and 0.99. This suggests that the combination of U-Net architectures with EfficientNet encoders is an effective approach for achieving both accuracy and computational efficiency in medical image segmentation.
Among the included studies focusing on reduced reading times, the choice of AI model for lesion classification was largely determined by the anatomical region and disease pathology under investigation. ResNet is one of the most widely recommended deep learning models for classification due to its superior feature extraction capabilities [148]. It was employed in several studies [54,72] examining brain and spine pathologies, with one study identifying ResNet50 as the most effective pre-trained model for discriminating lesions in multiple sclerosis (MS), achieving an AUC of 0.81. However, the same study demonstrated even greater accuracy with a custom-designed network, achieving an AUC of 0.90. This aligns with a broader trend observed in other studies comparing CNN performances, where task-specific architectures often outperformed standard pre-trained models. Despite variations in model selection, all studies reported measurable improvements in classification accuracy with AI assistance. However, studies that failed to clarify productivity gains did not provide a comparison of AI processing times against human reading times, limiting their clinical applicability.
Rather than focusing solely on model selection, our analysis suggests that future research should prioritize simulated AI applications in time-constrained settings. These studies should incorporate a diverse range of readers with varying levels of experience, not just expert radiologists, to better reflect real-world clinical workflows. An example of this approach is seen in one of the included studies [51], which focused on automated classification of lumbar spine central canal and lateral recess stenosis. Conducted under standardized reading conditions, the study found that AI-assisted reporting required 47 to 71 s, compared to 124 to 274 s for unassisted radiologists. Additionally, DL-assisted general and in-training radiologists demonstrated improved interobserver agreement for four-class neural foraminal stenosis, with κ values of 0.71 and 0.70 (with DL) vs. 0.39 and 0.39 (without DL), respectively (both p < 0.001). Beyond the measurable gains in efficiency and accuracy, the study highlights the advantages of an “AI and human collaboration model”, which should be a focal point for future research.
Among all productivity themes analyzed, the use of AI for reducing MRI scanner time demonstrated the highest success in achieving measurable productivity gains, with 84.6% (11/13) of the included studies reporting positive outcomes. Even in studies where productivity losses were observed [55,76], improvements in image quality—such as enhanced signal-to-noise ratio (SNR) and contrast-to-noise ratio (CNR)—were still achieved. However, these studies did not clarify how such enhancements would translate to improved classification or diagnostic accuracy. Despite the demonstrated benefits, only a limited number of deep learning-based reconstruction (DLR) techniques are currently available for clinical MRI scanners [149]. Moreover, most of these techniques remain vendor-specific, including Advanced Intelligent Clear-IQ Engine (AiCE; Canon Medical Systems), AIR Recon DL (GE Healthcare), and Deep Resolve (Siemens). Another significant barrier to broader adoption is the lack of publicly available training datasets, which restricts independent advancements and customization for different clinical settings [150].
To overcome these challenges, our study recommends further vendor-agnostic research aimed at improving image quality, particularly for older MRI scanners from various manufacturers. One example is a multicenter study [69] that investigated deep learning reconstruction for brain MRI using SubtleMR, an FDA-cleared, vendor-agnostic CNN trained on thousands of MRI datasets spanning multiple vendors, scanner models, field strengths, and clinical sites. By leveraging such diverse datasets, studies can provide stronger evidence that deep learning-based processing enhances MRI efficiency and diagnostic value across different clinical environments.
Because of the limited number of studies demonstrating novel improvements in workflow processes or workload reduction, our paper advocates for increased integration of Transformer-based architectures and large language models (LLMs) in future machine learning (ML) applications for MRI. These advancements could facilitate automated draft report generation, standardization of clinical protocols, and triaging of complex cases by leveraging information from request forms and electronic medical records [151]. Given their low customization costs and compatibility with existing IT infrastructure [152], LLMs have the potential to surpass other AI tools in overcoming current technological barriers. Recent innovations, such as voice assistants, could further enhance the utility of LLMs by enabling conversational interactions. Integrating LLMs into existing speech recognition dictation systems used in radiology departments may offer real-time productivity benefits, fostering seamless collaboration between AI and human radiologists. This approach could help mitigate potential time losses or increased workload associated with text-based reporting [153].
Despite the early success of LLMs in other imaging modalities, such as radiograph interpretation [154], recent studies evaluating their applicability in MRI have highlighted several challenges. One major limitation is that human–LLM interactions have primarily been assessed in controlled environments rather than real-world clinical settings, where frequent interruptions and high workloads influence radiologists’ behavior [155]. Additionally, the phenomenon of “hallucinations” in generative AI remains a significant barrier to its adoption in MRI-reporting workflows [84]. Future LLM must address these shortcomings to ensure reliability, safety, and clinical trustworthiness. As with other emerging AI applications, the successful integration of LLMs in MRI will require close collaboration among AI developers, clinical radiologists, healthcare administrators, and regulatory bodies. Establishing best practices for ethical and efficient implementation will also be critical for ensuring the widespread adoption of LLMs across MRI and other medical specialties [156].

4.3.5. Future Directions

Looking ahead, more standardized and multicenter prospective studies are needed to validate AI models under real-world conditions. Such trials could compare AI-driven MRI workflows with conventional practices, providing clearer insights into time savings, cost-effectiveness, and overall patient outcomes. In parallel, guidelines such as the recent PRISMA extensions for AI research [157] should continue to evolve in step with rapidly expanding AI research and applications. Establishing standardized productivity metrics, such as reduced turnaround times, minimized radiologist fatigue, and improved throughput would allow more meaningful comparisons across different studies.
Beyond technical measures, future research should also focus on user experiences, radiologist satisfaction, and patient-centered outcomes. While improvements in conventional metrics remain valuable, AI imaging has significant potential to reduce the fatigue and cognitive overload commonly experienced by radiologists. Burnout is a pressing concern, with one systematic review [158] finding rates as high as 88% in some regions. Although AI could theoretically mitigate burnout by streamlining tasks and maintaining diagnostic performance during prolonged reading sessions, a cross-sectional study [159] indicated that frequent AI use may paradoxically raise burnout risks if ease of adoption, radiologist satisfaction, and workflow integration are not properly addressed. Thus, tackling these interlinked factors is vital for the long-term viability of AI tools in MRI.
Lastly, initiatives that include iterative feedback loops, where radiologists refine AI outputs, could enhance model accuracy, yield institution-specific solutions, and build trust in AI recommendations. However, integrating AI into radiological workflows is not easy and requires significant resources and time, highlighting the need for strategies to continuously retrain and adapt models to evolving clinical settings [121]. Diversifying data sources, for instance by involving radiology residents in dataset creation [160], could lessen the load on senior radiologists and bolster model robustness. In addition, incorporating AI education into the radiology residency curriculum will equip future radiologists with the necessary technical competencies [161]. Further evidence is also required to support vendor-neutral platforms that can seamlessly integrate AI systems from multiple developers, thereby improving MRI access and facilitating enhanced combined reporting [162,163].

5. Conclusions

This review highlights significant advances in productivity metrics achieved through AI applications in MRI across various specialties, particularly in automating segmentation and reducing both scanning and reading times. Although many studies showed improvements in diagnostic accuracy, quantitative descriptions of efficiency gains were often limited. To enable meaningful clinical translation, future AI models must prioritize maintaining or improving productivity metrics without compromising diagnostic accuracy. In order to tackle underexplored themes in optimizing workflow and reducing workload, our study recommends increased focus and incorporation of LLM advancements in MRI workflow in order to accelerate radiologist engagement with substantial volumes of textual information encountered in day-to-day clinical practice. Good practices seen in other imaging modalities, where topics such as large-scale workload reduction or refined worklist triage have seen more maturity, should be explored and adopted in MRI [164,165].
Collaborative efforts among researchers, clinicians, and industry partners remain essential for developing standardized benchmarks and guidelines to measure productivity. Encouragingly, emerging initiatives are also addressing reproducibility concerns by integrating feedback loops for continuous model retraining and optimization. Exploring alternative training data sources could further reduce radiologist workload while retaining the credibility of reported productivity metrics. Ultimately, striking the right balance between efficacy, safety, and ease of implementation will be key to ensuring that AI-driven tools can meaningfully enhance radiologist productivity in MRI.

Author Contributions

Conceptualization, methodology, validation, data curation, writing—review and editing, A.N., W.O., A.L., A.M., Y.H.T., Y.J.L., S.J.O. and J.T.P.D.H.; software, resources, A.N., N.W.L., J.J.H.T., N.K. and J.T.P.D.H.; writing—original draft preparation, A.N. and J.J.H.T.; project administration, A.N., A.L., Y.H.T., S.J.O. and J.T.P.D.H.; supervision, funding acquisition, J.T.P.D.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by: Direct Funding from MOH/NMRC: This research is supported by the Singapore Ministry of Health National Medical Research Council under the NMRC Clinician Innovator Award (CIA); Grant Title: Deep learning pipeline for augmented reporting of MRI whole spine (CIAINV23jan-0001, MOH-001405) (J.T.P.D.H.).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Alowais, S.A.; Alghamdi, S.S.; Alsuhebany, N.; Alqahtani, T.; Alshaya, A.I.; Almohareb, S.N.; Aldairem, A.; Alrashed, M.; Bin Saleh, K.; Badreldin, H.A.; et al. Revolutionizing Healthcare: The Role of Artificial Intelligence in Clinical Practice. BMC Med. Educ. 2023, 23, 689. [Google Scholar] [CrossRef] [PubMed]
  2. Najjar, R. Redefining Radiology: A Review of Artificial Intelligence Integration in Medical Imaging. Diagnostics 2023, 13, 2760. [Google Scholar] [CrossRef]
  3. GE Healthcare—Committing to Sustainability in MRI. Available online: https://www.gehealthcare.com/insights/article/committing-to-sustainability-in-mri (accessed on 15 November 2024).
  4. Zhou, M.; Zhao, X.; Luo, F.; Luo, J.; Pu, H.; Tao, X. Robust RGB-T Tracking via Adaptive Modality Weight Correlation Filters and Cross-modality Learning. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 95. [Google Scholar] [CrossRef]
  5. Zhou, M.; Lan, X.; Wei, X.; Liao, X.; Mao, Q.; Li, Y. An End-to-End Blind Image Quality Assessment Method Using a Recurrent Network and Self-Attention. IEEE Trans. Broadcast. 2023, 69, 369–377. [Google Scholar] [CrossRef]
  6. Zhou, M.; Chen, L.; Wei, X.; Liao, X.; Mao, Q.; Wang, H. Perception-Oriented U-Shaped Transformer Network for 360-Degree No-Reference Image Quality Assessment. IEEE Trans. Broadcast. 2023, 69, 396–405. [Google Scholar] [CrossRef]
  7. Zhou, M.; Shen, W.; Wei, X.; Luo, J.; Jia, F.; Zhuang, X.; Jia, W. Blind Image Quality Assessment: Exploring Content Fidelity Perceptibility via Quality Adversarial Learning. Int. J. Comput. Vis. 2025. [Google Scholar] [CrossRef]
  8. Zhou, M.; Wu, X.; Wei, X.; Tao, X.; Fang, B.; Kwong, S. Low-Light Enhancement Method Based on a Retinex Model for Structure Preservation. IEEE Trans. Multimed. 2024, 26, 650–662. [Google Scholar] [CrossRef]
  9. McDonald, R.J.; Schwartz, K.M.; Eckel, L.J.; Diehn, F.E.; Hunt, C.H.; Bartholmai, B.J.; Erickson, B.J.; Kallmes, D.F. The Effects of Changes in Utilization and Technological Advancements of Cross-Sectional Imaging on Radiologist Workload. Acad. Radiol. 2015, 22, 1191–1198. [Google Scholar] [CrossRef] [PubMed]
  10. Quinn, L.; Tryposkiadis, K.; Deeks, J.; De Vet, H.C.W.; Mallett, S.; Mokkink, L.B.; Takwoingi, Y.; Taylor-Phillips, S.; Sitch, A. Interobserver Variability Studies in Diagnostic Imaging: A Methodological Systematic Review. Br. J. Radiol. 2023, 96, 20220972. [Google Scholar] [CrossRef]
  11. Khalifa, M.; Albadawy, M. AI in Diagnostic Imaging: Revolutionising Accuracy and Efficiency. Comput. Methods Programs Biomed. Update 2024, 5, 100146. [Google Scholar] [CrossRef]
  12. Potočnik, J.; Foley, S.; Thomas, E. Current and Potential Applications of Artificial Intelligence in Medical Imaging Practice: A Narrative Review. J. Med. Imaging Radiat. Sci. 2023, 54, 376–385. [Google Scholar] [CrossRef] [PubMed]
  13. Younis, H.A.; Eisa, T.A.; Nasser, M.; Sahib, T.M.; Noor, A.A.; Alyasiri, O.M.; Salisu, S.; Hayder, I.M.; Younis, H.A. A Systematic Review and Meta-Analysis of Artificial Intelligence Tools in Medicine and Healthcare: Applications, Considerations, Limitations, Motivation and Challenges. Diagnostics 2024, 14, 109. [Google Scholar] [CrossRef]
  14. Gitto, S.; Serpi, F.; Albano, D.; Risoleo, G.; Fusco, S.; Messina, C.; Sconfienza, L.M. AI Applications in Musculoskeletal Imaging: A Narrative Review. Eur. Radiol. Exp. 2024, 8, 22. [Google Scholar] [CrossRef]
  15. Vogrin, M.; Trojner, T.; Kelc, R. Artificial Intelligence in Musculoskeletal Oncological Radiology. Radiol. Oncol. 2021, 55, 1–6. [Google Scholar] [CrossRef]
  16. Wenderott, K.; Krups, J.; Zaruchas, F.; Weigl, M. Effects of Artificial Intelligence Implementation on Efficiency in Medical Imaging—A Systematic Literature Review and Meta-Analysis. NPJ Digit. Med. 2024, 7, 265. [Google Scholar] [CrossRef]
  17. Stec, N.; Arje, D.; Moody, A.R.; Krupinski, E.A.; Tyrrell, P.N. A Systematic Review of Fatigue in Radiology: Is It a Problem? Am. J. Roentgenol. 2018, 210, 799–806. [Google Scholar] [CrossRef]
  18. Ranschaert, E.; Topff, L.; Pianykh, O. Optimization of Radiology Workflow with Artificial Intelligence. Radiol. Clin. N. Am. 2021, 59, 955–966. [Google Scholar] [CrossRef] [PubMed]
  19. Clifford, B.; Conklin, J.; Huang, S.Y.; Feiweier, T.; Hosseini, Z.; Goncalves Filho, A.L.M.; Tabari, A.; Demir, S.; Lo, W.-C.; Longo, M.G.F.; et al. An Artificial Intelligence-Accelerated 2-Minute Multi-Shot Echo Planar Imaging Protocol for Comprehensive High-Quality Clinical Brain Imaging. Magn. Reson. Med. 2022, 87, 2453–2463. [Google Scholar] [CrossRef] [PubMed]
  20. Jing, X.; Wielema, M.; Cornelissen, L.J.; van Gent, M.; Iwema, W.M.; Zheng, S.; Sijens, P.E.; Oudkerk, M.; Dorrius, M.D.; van Ooijen, P.M.A. Using Deep Learning to Safely Exclude Lesions with Only Ultrafast Breast MRI to Shorten Acquisition and Reading Time. Eur. Radiol. 2022, 32, 8706–8715. [Google Scholar] [CrossRef]
  21. Chen, G.; Fan, X.; Wang, T.; Zhang, E.; Shao, J.; Chen, S.; Zhang, D.; Zhang, J.; Guo, T.; Yuan, Z.; et al. A Machine Learning Model Based on MRI for the Preoperative Prediction of Bladder Cancer Invasion Depth. Eur. Radiol. 2023, 33, 8821–8832. [Google Scholar] [CrossRef]
  22. Ivanovska, T.; Daboul, A.; Kalentev, O.; Hosten, N.; Biffar, R.; Völzke, H.; Wörgötter, F. A Deep Cascaded Segmentation of Obstructive Sleep Apnea-Relevant Organs from Sagittal Spine MRI. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 579–588. [Google Scholar] [CrossRef] [PubMed]
  23. Li, Y.; Wang, Y.; Qi, H.; Hu, Z.; Chen, Z.; Yang, R.; Qiao, H.; Sun, J.; Wang, T.; Zhao, X.; et al. Deep Learning–Enhanced T1 Mapping with Spatial-Temporal and Physical Constraint. Magn. Reson. Med. 2021, 86, 1647–1661. [Google Scholar] [CrossRef]
  24. Herington, J.; McCradden, M.D.; Creel, K.; Boellaard, R.; Jones, E.C.; Jha, A.K.; Rahmim, A.; Scott, P.J.H.; Sunderland, J.J.; Wahl, R.L.; et al. Ethical Considerations for Artificial Intelligence in Medical Imaging: Data Collection, Development, and Evaluation. J. Nucl. Med. 2023, 64, 1848–1854. [Google Scholar] [CrossRef]
  25. Martín-Noguerol, T.; Paulano-Godino, F.; López-Ortega, R.; Górriz, J.M.; Riascos, R.F.; Luna, A. Artificial Intelligence in Radiology: Relevance of Collaborative Work between Radiologists and Engineers for Building a Multidisciplinary Team. Clin. Radiol. 2021, 76, 317–324. [Google Scholar] [CrossRef] [PubMed]
  26. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, J.; Cui, Z.; Shi, Z.; Jiang, Y.; Zhang, Z.; Dai, X.; Yang, Z.; Gu, Y.; Zhou, L.; Han, C.; et al. A Robust and Efficient AI Assistant for Breast Tumor Segmentation from DCE-MRI via a Spatial-Temporal Framework. Patterns 2023, 4, 100826. [Google Scholar] [CrossRef]
  28. Pötsch, N.; Dietzel, M.; Kapetas, P.; Clauser, P.; Pinker, K.; Ellmann, S.; Uder, M.; Helbich, T.; Baltzer, P.A.T. An A.I. Classifier Derived from 4D Radiomics of Dynamic Contrast-Enhanced Breast MRI Data: Potential to Avoid Unnecessary Breast Biopsies. Eur. Radiol. 2021, 31, 5866–5876. [Google Scholar] [CrossRef]
  29. Zhao, Z.; Du, S.; Xu, Z.; Yin, Z.; Huang, X.; Huang, X.; Wong, C.; Liang, Y.; Shen, J.; Wu, J.; et al. SwinHR: Hemodynamic-Powered Hierarchical Vision Transformer for Breast Tumor Segmentation. Comput. Biol. Med. 2024, 169, 107939. [Google Scholar] [CrossRef]
  30. Lim, R.P.; Kachel, S.; Villa, A.D.M.; Kearney, L.; Bettencourt, N.; Young, A.A.; Chiribiri, A.; Scannell, C.M. CardiSort: A Convolutional Neural Network for Cross Vendor Automated Sorting of Cardiac MR Images. Eur. Radiol. 2022, 32, 5907–5920. [Google Scholar] [CrossRef]
  31. Parker, J.; Coey, J.; Alambrouk, T.; Lakey, S.M.; Green, T.; Brown, A.; Maxwell, I.; Ripley, D.P. Evaluating a Novel AI Tool for Automated Measurement of the Aortic Root and Valve in Cardiac Magnetic Resonance Imaging. Cureus 2024, 16, e59647. [Google Scholar] [CrossRef]
  32. Duan, C.; Deng, H.; Xiao, S.; Xie, J.; Li, H.; Zhao, X.; Han, D.; Sun, X.; Lou, X.; Ye, C.; et al. Accelerate Gas Diffusion-Weighted MRI for Lung Morphometry with Deep Learning. Eur. Radiol. 2022, 32, 702–713. [Google Scholar] [CrossRef]
  33. Xie, L.; Udupa, J.K.; Tong, Y.; Torigian, D.A.; Huang, Z.; Kogan, R.M.; Wootton, D.; Choy, K.R.; Sin, S.; Wagshul, M.E.; et al. Automatic Upper Airway Segmentation in Static and Dynamic MRI via Anatomy-Guided Convolutional Neural Networks. Med. Phys. 2022, 49, 324–342. [Google Scholar] [CrossRef]
  34. Wang, J.; Peng, Y.; Jing, S.; Han, L.; Li, T.; Luo, J. A deep-learning approach for segmentation of liver tumors in magnetic resonance imaging using UNet++. BMC Cancer 2023, 23, 1060. [Google Scholar] [CrossRef] [PubMed]
  35. Gross, M.; Huber, S.; Arora, S.; Ze’evi, T.; Haider, S.P.; Kucukkaya, A.S.; Iseke, S.; Kuhn, T.N.; Gebauer, B.; Michallek, F.; et al. Automated MRI Liver Segmentation for Anatomical Segmentation, Liver Volumetry, and the Extraction of Radiomics. Eur. Radiol. 2024, 34, 5056–5065. [Google Scholar] [CrossRef] [PubMed]
  36. Oestmann, P.M.; Wang, C.J.; Savic, L.J.; Hamm, C.A.; Stark, S.; Schobert, I.; Gebauer, B.; Schlachter, T.; Lin, M.; Weinreb, J.C.; et al. Deep Learning–Assisted Differentiation of Pathologically Proven Atypical and Typical Hepatocellular Carcinoma (HCC) versus Non-HCC on Contrast-Enhanced MRI of the Liver. Eur. Radiol. 2021, 31, 4981–4990. [Google Scholar] [CrossRef] [PubMed]
  37. Yan, M.; Zhang, X.; Zhang, B.; Geng, Z.; Xie, C.; Yang, W.; Zhang, S.; Qi, Z.; Lin, T.; Ke, Q.; et al. Deep Learning Nomogram Based on Gd-EOB-DTPA MRI for Predicting Early Recurrence in Hepatocellular Carcinoma after Hepatectomy. Eur. Radiol. 2023, 33, 4949–4961. [Google Scholar] [CrossRef]
  38. Wei, H.; Zheng, T.; Zhang, X.; Wu, Y.; Chen, Y.; Zheng, C.; Jiang, D.; Wu, B.; Guo, H.; Jiang, H.; et al. MRI Radiomics Based on Deep Learning Automated Segmentation to Predict Early Recurrence of Hepatocellular Carcinoma. Insights Imaging 2024, 15, 120. [Google Scholar] [CrossRef]
  39. Gross, M.; Haider, S.P.; Ze’evi, T.; Huber, S.; Arora, S.; Kucukkaya, A.S.; Iseke, S.; Gebauer, B.; Fleckenstein, F.; Dewey, M.; et al. Automated Graded Prognostic Assessment for Patients with Hepatocellular Carcinoma Using Machine Learning. Eur. Radiol. 2024, 34, 6940–6952. [Google Scholar] [CrossRef]
  40. Cunha, G.M.; Hasenstab, K.A.; Higaki, A.; Wang, K.; Delgado, T.; Brunsing, R.L.; Schlein, A.; Schwartzman, A.; Hsiao, A.; Sirlin, C.B.; et al. Convolutional Neural Network-Automated Hepatobiliary Phase Adequacy Evaluation May Optimize Examination Time. Eur. J. Radiol. 2020, 124, 108837. [Google Scholar] [CrossRef]
  41. Liu, G.; Pan, S.; Zhao, R.; Zhou, H.; Chen, J.; Zhou, X.; Xu, J.; Zhou, Y.; Xue, W.; Wu, G. The Added Value of AI-Based Computer-Aided Diagnosis in Classification of Cancer at Prostate MRI. Eur. Radiol. 2023, 33, 5118–5130. [Google Scholar] [CrossRef]
  42. Simon, B.D.; Merriman, K.M.; Harmon, S.A.; Tetreault, J.; Yilmaz, E.C.; Blake, Z.; Merino, M.J.; An, J.Y.; Marko, J.; Law, Y.M.; et al. Automated Detection and Grading of Extraprostatic Extension of Prostate Cancer at MRI via Cascaded Deep Learning and Random Forest Classification. Acad. Radiol. 2024, 31, 4096–4106. [Google Scholar] [CrossRef] [PubMed]
  43. Wang, K.; Xing, Z.; Kong, Z.; Yu, Y.; Chen, Y.; Zhao, X.; Song, B.; Wang, X.; Wu, P.; Wang, X.; et al. Artificial Intelligence as Diagnostic Aiding Tool in Cases of Prostate Imaging Reporting and Data System Category 3: The Results of Retrospective Multi-Center Cohort Study. Abdom. Radiol. 2023, 48, 3757–3765. [Google Scholar] [CrossRef] [PubMed]
  44. Hosseinzadeh, M.; Saha, A.; Brand, P.; Slootweg, I.; de Rooij, M.; Huisman, H. Deep Learning–Assisted Prostate Cancer Detection on Bi-Parametric MRI: Minimum Training Data Size Requirements and Effect of Prior Knowledge. Eur. Radiol. 2022, 32, 2224–2234. [Google Scholar] [CrossRef] [PubMed]
  45. Oerther, B.; Engel, H.; Nedelcu, A.; Schlett, C.L.; Grimm, R.; von Busch, H.; Sigle, A.; Gratzke, C.; Bamberg, F.; Benndorf, M. Prediction of Upgrade to Clinically Significant Prostate Cancer in Patients under Active Surveillance: Performance of a Fully Automated AI-Algorithm for Lesion Detection and Classification. Prostate 2023, 83, 871–878. [Google Scholar] [CrossRef]
  46. Li, Q.; Xu, W.-Y.; Sun, N.-N.; Feng, Q.-X.; Hou, Y.-J.; Sang, Z.-T.; Zhu, Z.-N.; Hsu, Y.-C.; Nickel, D.; Xu, H.; et al. Deep Learning-Accelerated T2WI: Image Quality, Efficiency, and Staging Performance against BLADE T2WI for Gastric Cancer. Abdom. Radiol. 2024, 49, 2574–2584. [Google Scholar] [CrossRef]
  47. Goel, A.; Shih, G.; Riyahi, S.; Jeph, S.; Dev, H.; Hu, R.; Romano, D.; Teichman, K.; Blumenfeld, J.D.; Barash, I.; et al. Deployed Deep Learning Kidney Segmentation for Polycystic Kidney Disease MRI. Radiol. Artif. Intell. 2022, 4, e210205. [Google Scholar] [CrossRef]
  48. Marka, A.W.; Luitjens, J.; Gassert, F.T.; Steinhelfer, L.; Burian, E.; Rübenthaler, J.; Schwarze, V.; Froelich, M.F.; Makowski, M.R.; Gassert, F.G. Artificial Intelligence Support in MR Imaging of Incidental Renal Masses: An Early Health Technology Assessment. Eur. Radiol. 2024, 34, 5856–5865. [Google Scholar] [CrossRef]
  49. Chen, Y.; Ruan, D.; Xiao, J.; Wang, L.; Sun, B.; Saouaf, R.; Yang, W.; Li, D.; Fan, Z. Fully Automated Multiorgan Segmentation in Abdominal Magnetic Resonance Imaging with Deep Neural Networks. Med. Phys. 2020, 47, 4971–4982. [Google Scholar] [CrossRef]
  50. Hallinan, J.T.P.D.; Zhu, L.; Yang, K.; Makmur, A.; Algazwi, D.A.R.; Thian, Y.L.; Lau, S.; Choo, Y.S.; Eide, S.E.; Yap, Q.V.; et al. Deep Learning Model for Automated Detection and Classification of Central Canal, Lateral Recess, and Neural Foraminal Stenosis at Lumbar Spine MRI. Radiology 2021, 300, 130–138. [Google Scholar] [CrossRef]
  51. Lim, D.S.W.; Makmur, A.; Zhu, L.; Zhang, W.; Cheng, A.J.L.; Sia, D.S.Y.; Eide, S.E.; Ong, H.Y.; Jagmohan, P.; Tan, W.C.; et al. Improved Productivity Using Deep Learning–Assisted Reporting for Lumbar Spine MRI. Radiology 2022, 305, 160–166. [Google Scholar] [CrossRef]
  52. McSweeney, T.P.; Tiulpin, A.; Saarakkala, S.; Niinimäki, J.; Windsor, R.; Jamaludin, A.; Kadir, T.; Karppinen, J.; Määttä, J. External Validation of SpineNet, an Open-Source Deep Learning Model for Grading Lumbar Disk Degeneration MRI Features, Using the Northern Finland Birth Cohort 1966. Spine 2023, 48, 484–491. [Google Scholar] [CrossRef] [PubMed]
  53. Grob, A.; Loibl, M.; Jamaludin, A.; Winklhofer, S.; Fairbank, J.C.T.; Fekete, T.; Porchet, F.; Mannion, A.F. External Validation of the Deep Learning System “SpineNet” for Grading Radiological Features of Degeneration on MRIs of the Lumbar Spine. Eur. Spine J. 2022, 31, 2137–2148. [Google Scholar] [CrossRef]
  54. Bharadwaj, U.U.; Christine, M.; Li, S.; Chou, D.; Pedoia, V.; Link, T.M.; Chin, C.T.; Majumdar, S. Deep Learning for Automated, Interpretable Classification of Lumbar Spinal Stenosis and Facet Arthropathy from Axial MRI. Eur. Radiol. 2023, 33, 3435–3443. [Google Scholar] [CrossRef]
  55. van der Graaf, J.W.; Brundel, L.; van Hooff, M.L.; de Kleuver, M.; Lessmann, N.; Maresch, B.J.; Vestering, M.M.; Spermon, J.; van Ginneken, B.; Rutten, M.J.C.M. AI-Based Lumbar Central Canal Stenosis Classification on Sagittal MR Images Is Comparable to Experienced Radiologists Using Axial Images. Eur. Radiol. 2024, 35, 2298–2306. [Google Scholar] [CrossRef] [PubMed]
  56. Yoo, H.; Yoo, R.-E.; Choi, S.H.; Hwang, I.; Lee, J.Y.; Seo, J.Y.; Koh, S.Y.; Choi, K.S.; Kang, K.M.; Yun, T.J. Deep Learning–Based Reconstruction for Acceleration of Lumbar Spine MRI: A Prospective Comparison with Standard MRI. Eur. Radiol. 2023, 33, 8656–8668. [Google Scholar] [CrossRef]
  57. Wang, D.; Liu, S.; Ding, J.; Sun, A.; Jiang, D.; Jiang, J.; Zhao, J.; Chen, D.; Ji, G.; Li, N.; et al. A Deep Learning Model Enhances Clinicians’ Diagnostic Accuracy to More Than 96% for Anterior Cruciate Ligament Ruptures on Magnetic Resonance Imaging. Arthroscopy 2024, 40, 1197–1205. [Google Scholar] [CrossRef] [PubMed]
  58. Ro, K.; Kim, J.Y.; Park, H.; Cho, B.H.; Kim, I.Y.; Shim, S.B.; Choi, I.Y.; Yoo, J.C. Deep-Learning Framework and Computer Assisted Fatty Infiltration Analysis for the Supraspinatus Muscle in MRI. Sci. Rep. 2021, 11, 15065. [Google Scholar] [CrossRef]
  59. Ni, M.; Zhao, Y.; Zhang, L.; Chen, W.; Wang, Q.; Tian, C.; Yuan, H. MRI-Based Automated Multitask Deep Learning System to Evaluate Supraspinatus Tendon Injuries. Eur. Radiol. 2024, 34, 3538–3551. [Google Scholar] [CrossRef]
  60. Feuerriegel, G.C.; Weiss, K.; Kronthaler, S.; Leonhardt, Y.; Neumann, J.; Wurm, M.; Lenhart, N.S.; Makowski, M.R.; Schwaiger, B.J.; Woertler, K.; et al. Evaluation of a Deep Learning-Based Reconstruction Method for Denoising and Image Enhancement of Shoulder MRI in Patients with Shoulder Pain. Eur. Radiol. 2023, 33, 4875–4884. [Google Scholar] [CrossRef]
  61. Liu, J.; Li, W.; Li, Z.; Yang, J.; Wang, K.; Cao, X.; Qin, N.; Xue, K.; Dai, Y.; Wu, P.; et al. Magnetic Resonance Shoulder Imaging Using Deep Learning–Based Algorithm. Eur. Radiol. 2023, 33, 4864–4874. [Google Scholar] [CrossRef]
  62. Nowak, S.; Henkel, A.; Theis, M.; Luetkens, J.; Geiger, S.; Sprinkart, A.M.; Pieper, C.C.; Attenberger, U.I. Deep Learning for Standardized, MRI-Based Quantification of Subcutaneous and Subfascial Tissue Volume for Patients with Lipedema and Lymphedema. Eur. Radiol. 2023, 33, 884–892. [Google Scholar] [CrossRef] [PubMed]
  63. Hu, X.; Liu, Z.; Zhou, H.; Fang, J.; Lu, H. Deep HT: A Deep Neural Network for Diagnose on MR Images of Tumors of the Hand. PLoS ONE 2020, 15, e0237606. [Google Scholar] [CrossRef]
  64. Li, Y.; Li, H.; Fan, Y. ACEnet: Anatomical Context-Encoding Network for Neuroanatomy Segmentation. Med. Image Anal. 2021, 70, 101991. [Google Scholar] [CrossRef]
  65. Yamada, S.; Otani, T.; Ii, S.; Kawano, H.; Nozaki, K.; Wada, S.; Oshima, M.; Watanabe, Y. Aging-Related Volume Changes in the Brain and Cerebrospinal Fluid Using Artificial Intelligence-Automated Segmentation. Eur. Radiol. 2023, 33, 7099–7112. [Google Scholar] [CrossRef] [PubMed]
  66. Dar, S.U.H.; Öztürk, Ş.; Özbey, M.; Oguz, K.K.; Çukur, T. Parallel-Stream Fusion of Scan-Specific and Scan-General Priors for Learning Deep MRI Reconstruction in Low-Data Regimes. Comput. Biol. Med. 2023, 167, 107610. [Google Scholar] [CrossRef] [PubMed]
  67. Jun, Y.; Cho, J.; Wang, X.; Gee, M.; Grant, P.E.; Bilgic, B.; Gagoski, B. SSL-QALAS: Self-Supervised Learning for Rapid Multiparameter Estimation in Quantitative MRI Using 3D-QALAS. Magn. Reson. Med. 2023, 90, 2019–2032. [Google Scholar] [CrossRef]
  68. Kemenczky, P.; Vakli, P.; Somogyi, E.; Homolya, I.; Hermann, P.; Gál, V.; Vidnyánszky, Z. Effect of Head Motion-Induced Artefacts on the Reliability of Deep Learning-Based Whole-Brain Segmentation. Sci. Rep. 2022, 12, 1618. [Google Scholar] [CrossRef]
  69. Bash, S.; Wang, L.; Airriess, C.; Zaharchuk, G.; Gong, E.; Shankaranarayanan, A.; Tanenbaum, L.N. Deep Learning Enables 60% Accelerated Volumetric Brain MRI While Preserving Quantitative Performance: A Prospective, Multicenter, Multireader Trial. Am. J. Neuroradiol. 2021, 42, 2130–2137. [Google Scholar] [CrossRef]
  70. Boaro, A.; Kaczmarzyk, J.R.; Kavouridis, V.K.; Harary, M.; Mammi, M.; Dawood, H.; Shea, A.; Cho, E.Y.; Juvekar, P.; Noh, T.; et al. Deep Neural Networks Allow Expert-Level Brain Meningioma Segmentation and Present Potential for Improvement of Clinical Practice. Sci. Rep. 2022, 12, 15462. [Google Scholar] [CrossRef]
  71. Gab Allah, A.M.; Sarhan, A.M.; Elshennawy, N.M. Classification of Brain MRI Tumor Images Based on Deep Learning PGGAN Augmentation. Diagnostics 2021, 11, 2343. [Google Scholar] [CrossRef]
  72. Amini, A.; Shayganfar, A.; Amini, Z.; Ostovar, L.; HajiAhmadi, S.; Chitsaz, N.; Rabbani, M.; Kafieh, R. Deep Learning for Discrimination of Active and Inactive Lesions in Multiple Sclerosis Using Non-Contrast FLAIR MRI: A Multicenter Study. Mult. Scler. Relat. Disord. 2024, 87, 105642. [Google Scholar] [CrossRef] [PubMed]
  73. Cui, X.; Chen, N.; Zhao, C.; Li, J.; Zheng, X.; Liu, C.; Yang, J.; Li, X.; Yu, C.; Liu, J.; et al. An Adaptive Weighted Attention-Enhanced Deep Convolutional Neural Network for Classification of MRI Images of Parkinson’s Disease. J. Neurosci. Methods 2023, 394, 109884. [Google Scholar] [CrossRef] [PubMed]
  74. Sun, F.; Lyu, J.; Jian, S.; Qin, Y.; Tang, X. Accurate Measurement of Magnetic Resonance Parkinsonism Index by a Fully Automatic and Deep Learning Quantification Pipeline. Eur. Radiol. 2023, 33, 8844–8853. [Google Scholar] [CrossRef] [PubMed]
  75. Li, C.; Hui, D.; Wu, F.; Xia, Y.; Shi, F.; Yang, M.; Zhang, J.; Peng, C.; Feng, J.; Li, C. Automatic Diagnosis of Parkinson’s Disease Using Artificial Intelligence Base on Routine T1-Weighted MRI. Front. Med. 2024, 10, 1303501. [Google Scholar] [CrossRef]
  76. Yasaka, K.; Akai, H.; Sugawara, H.; Tajima, T.; Akahane, M.; Yoshioka, N.; Kabasawa, H.; Miyo, R.; Ohtomo, K.; Abe, O.; et al. Impact of Deep Learning Reconstruction on Intracranial 1.5 T Magnetic Resonance Angiography. Jpn. J. Radiol. 2022, 40, 476–483. [Google Scholar] [CrossRef]
  77. Tian, T.; Gan, T.; Chen, J.; Lu, J.; Zhang, G.; Zhou, Y.; Li, J.; Shao, H.; Liu, Y.; Zhu, H.; et al. Graphic Intelligent Diagnosis of Hypoxic-Ischemic Encephalopathy Using MRI-Based Deep Learning Model. Neonatology 2023, 120, 441–449. [Google Scholar] [CrossRef]
  78. Lew, C.O.; Calabrese, E.; Chen, J.V.; Tang, F.; Chaudhari, G.; Lee, A.; Faro, J.; Juul, S.; Mathur, A.; McKinstry, R.C.; et al. Artificial Intelligence Outcome Prediction in Neonates with Encephalopathy (AI-OPiNE). Radiol. Artif. Intell. 2024, 6, e240076. [Google Scholar] [CrossRef]
  79. She, J.; Huang, H.; Ye, Z.; Huang, W.; Sun, Y.; Liu, C.; Yang, W.; Wang, J.; Ye, P.; Zhang, L.; et al. Automatic Biometry of Fetal Brain MRIs Using Deep and Machine Learning Techniques. Sci. Rep. 2023, 13, 17860. [Google Scholar] [CrossRef]
  80. Zhang, Y.; Xu, H.-R.; Wen, J.-H.; Hu, Y.-J.; Diao, Y.-L.; Chen, J.-L.; Xia, Y.-F. A Novel LVPA-UNet Network for Target Volume Automatic Delineation: An MRI Case Study of Nasopharyngeal Carcinoma. Heliyon 2024, 10, e30763. [Google Scholar] [CrossRef]
  81. Winkel, D.J.; Tong, A.; Lou, B.; Kamen, A.; Comaniciu, D.; Disselhorst, J.A.; Rodríguez-Ruiz, A.; Huisman, H.; Szolar, D.; Shabunin, I.; et al. A Novel Deep Learning Based Computer-Aided Diagnosis System Improves the Accuracy and Efficiency of Radiologists in Reading Biparametric Magnetic Resonance Images of the Prostate: Results of a Multireader, Multicase Study. Investig. Radiol. 2021, 56, 605–613. [Google Scholar] [CrossRef]
  82. Rudie, J.D.; Duda, J.; Duong, M.T.; Chen, P.-H.; Xie, L.; Kurtz, R.; Ware, J.B.; Choi, J.; Mattay, R.R.; Botzolakis, E.J.; et al. Brain MRI Deep Learning and Bayesian Inference System Augments Radiology Resident Performance. J. Digit. Imaging 2021, 34, 1049–1058. [Google Scholar] [CrossRef]
  83. Langkilde, F.; Masaba, P.; Edenbrandt, L.; Gren, M.; Halil, A.; Hellström, M.; Larsson, M.; Naeem, A.A.; Wallström, J.; Maier, S.E.; et al. Manual Prostate MRI Segmentation by Readers with Different Experience: A Study of the Learning Progress. Eur. Radiol. 2024, 34, 4801–4809. [Google Scholar] [CrossRef]
  84. Park, J.; Oh, K.; Han, K.; Lee, Y.H. Patient-Centered Radiology Reports with Generative Artificial Intelligence: Adding Value to Radiology Reporting. Sci. Rep. 2024, 14, 13218. [Google Scholar] [CrossRef]
  85. Federau, C.; Hainc, N.; Edjlali, M.; Zhu, G.; Mastilovic, M.; Nierobisch, N.; Uhlemann, J.-P.; Paganucci, S.; Granziera, C.; Heinzlef, O.; et al. Evaluation of the Quality and the Productivity of Neuroradiological Reading of Multiple Sclerosis Follow-up MRI Scans Using an Intelligent Automation Software. Neuroradiology 2024, 66, 361–369. [Google Scholar] [CrossRef]
  86. Lee, Y.H. Efficiency Improvement in a Busy Radiology Practice: Determination of Musculoskeletal Magnetic Resonance Imaging Protocol Using Deep-Learning Convolutional Neural Networks. J. Digit. Imaging 2018, 31, 604–610. [Google Scholar] [CrossRef] [PubMed]
  87. Yu, Y.; He, Z.; Ouyang, J.; Tan, Y.; Chen, Y.; Gu, Y.; Mao, L.; Ren, W.; Wang, J.; Lin, L.; et al. Magnetic Resonance Imaging Radiomics Predicts Preoperative Axillary Lymph Node Metastasis to Support Surgical Decisions and Is Associated with Tumor Microenvironment in Invasive Breast Cancer: A Machine Learning, Multicenter Study. EBioMedicine 2021, 69, 103460. [Google Scholar] [CrossRef] [PubMed]
  88. Schlemper, J.; Caballero, J.; Hajnal, J.V.; Price, A.N.; Rueckert, D. A Deep Cascade of Convolutional Neural Networks for Dynamic MR Image Reconstruction. IEEE Trans. Med. Imaging 2018, 37, 491–503. [Google Scholar] [CrossRef] [PubMed]
  89. Hammernik, K.; Klatzer, T.; Kobler, E.; Recht, M.P.; Sodickson, D.K.; Pock, T.; Knoll, F. Learning a Variational Network for Reconstruction of Accelerated MRI Data. Magn. Reson. Med. 2017, 79, 3055. [Google Scholar] [CrossRef]
  90. Dai, X.; Lei, Y.; Wang, T.; Zhou, J.; Roper, J.; McDonald, M.; Beitler, J.J.; Curran, W.J.; Liu, T.; Yang, X. Automated Delineation of Head and Neck Organs at Risk Using Synthetic MRI-Aided Mask Scoring Regional Convolutional Neural Network. Med. Phys. 2021, 48, 5862–5873. [Google Scholar] [CrossRef]
  91. Amisha; Malik, P.; Pathania, M.; Rathaur, V.K. Overview of Artificial Intelligence in Medicine. J. Fam. Med. Prim. Care 2019, 8, 2328–2331. [Google Scholar] [CrossRef]
  92. Kobayashi, H. Potential for Artificial Intelligence in Medicine and Its Application to Male Infertility. Reprod. Med. Biol. 2024, 23, e12590. [Google Scholar] [CrossRef] [PubMed]
  93. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine Learning for Medical Imaging. RadioGraphics 2017, 37, 505–515. [Google Scholar] [CrossRef]
  94. Zaharchuk, G.; Gong, E.; Wintermark, M.; Rubin, D.; Langlotz, C.P. Deep Learning in Neuroradiology. Am. J. Neuroradiol. 2018, 39, 1776–1784. [Google Scholar] [CrossRef]
  95. Hosny, A.; Parmar, C.; Quackenbush, J.; Schwartz, L.H.; Aerts, H.J.W.L. Artificial Intelligence in Radiology. Nat. Rev. Cancer 2018, 18, 500–510. [Google Scholar] [CrossRef] [PubMed]
  96. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional Neural Networks: An Overview and Application in Radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef]
  97. Cabitza, F.; Campagner, A.; Sconfienza, L.M. Studying Human-AI Collaboration Protocols: The Case of the Kasparov’s Law in Radiological Double Reading. Health Inf. Sci. Syst. 2021, 9, 8. [Google Scholar] [CrossRef]
  98. Almansour, H.; Herrmann, J.; Gassenmaier, S.; Afat, S.; Jacoby, J.; Koerzdoerfer, G.; Nickel, D.; Mostapha, M.; Nadar, M.; Othman, A.E. Deep Learning Reconstruction for Accelerated Spine MRI: Prospective Analysis of Interchangeability. Radiology 2022, 306, e212922. [Google Scholar] [CrossRef] [PubMed]
  99. Hallinan, J.T.P.D. Deep Learning for Spine MRI: Reducing Time Not Quality. Radiology 2022, 306, e222410. [Google Scholar] [CrossRef]
  100. Fritz, J.; Raithel, E.; Thawait, G.K.; Gilson, W.; Papp, D.F. Six-Fold Acceleration of High-Spatial Resolution 3D SPACE MRI of the Knee Through Incoherent k-Space Undersampling and Iterative Reconstruction—First Experience. Investig. Radiol. 2016, 51, 400–409. [Google Scholar] [CrossRef]
  101. Han, Z.; Wang, Y.; Wang, W.; Zhang, T.; Wang, J.; Ma, X.; Men, K.; Shi, A.; Gao, Y.; Bi, N. Artificial Intelligence-Assisted Delineation for Postoperative Radiotherapy in Patients with Lung Cancer: A Prospective, Multi-Center, Cohort Study. Front. Oncol. 2024, 14, 1388297. [Google Scholar] [CrossRef]
  102. Wang, F.; Chen, Q.; Chen, Y.; Zhu, Y.; Zhang, Y.; Cao, D.; Zhou, W.; Liang, X.; Yang, Y.; Lin, L.; et al. A Novel Multimodal Deep Learning Model for Preoperative Prediction of Microvascular Invasion and Outcome in Hepatocellular Carcinoma. Eur. J. Surg. Oncol. 2023, 49, 156–164. [Google Scholar] [CrossRef]
  103. Fatima, A.; Shahid, A.R.; Raza, B.; Madni, T.M.; Janjua, U.I. State-of-the-Art Traditional to the Machine- and Deep-Learning-Based Skull Stripping Techniques, Models, and Algorithms. J. Digit. Imaging 2020, 33, 1443–1464. [Google Scholar] [CrossRef] [PubMed]
  104. Hallinan, J.T.P.D.; Leow, N.W.; Ong, W.; Lee, A.; Low, Y.X.; Chan, M.D.Z.; Devi, G.K.; Loh, D.D.L.; He, S.S.; Nor, F.E.M.; et al. MRI Spine Request Form Enhancement and Auto Protocoling Using a Secure Institutional Large Language Model. Spine J. 2024, 25, 505–514. [Google Scholar] [CrossRef] [PubMed]
  105. Beauchamp, N.J.; Bryan, R.N.; Bui, M.M.; Krestin, G.P.; McGinty, G.B.; Meltzer, C.C.; Neumaier, M. Integrative Diagnostics: The Time Is Now—A Report from the International Society for Strategic Studies in Radiology. Insights Imaging 2023, 14, 54. [Google Scholar] [CrossRef] [PubMed]
  106. Forsberg, D.; Rosipko, B.; Sunshine, J.L. Radiologists’ Variation of Time to Read Across Different Procedure Types. J. Digit. Imaging 2017, 30, 86–94. [Google Scholar] [CrossRef]
  107. Nowak, S.; Mesropyan, N.; Faron, A.; Block, W.; Reuter, M.; Attenberger, U.I.; Luetkens, J.A.; Sprinkart, A.M. Detection of Liver Cirrhosis in Standard T2-Weighted MRI Using Deep Transfer Learning. Eur. Radiol. 2021, 31, 8807–8815. [Google Scholar] [CrossRef]
  108. Nai, Y.H.; Loi, H.Y.; O’Doherty, S.; Tan, T.H.; Reilhac, A. Comparison of the performances of machine learning and deep learning in improving the quality of low dose lung cancer PET images. Jpn. J. Radiol. 2022, 40, 1290–1299. [Google Scholar] [CrossRef]
  109. Chauhan, D.; Anyanwu, E.; Goes, J.; Besser, S.A.; Anand, S.; Madduri, R.; Getty, N.; Kelle, S.; Kawaji, K.; Mor-Avi, V.; et al. Comparison of machine learning and deep learning for view identification from cardiac magnetic resonance images. Clin. Imaging. 2022, 82, 121–126. [Google Scholar] [CrossRef]
  110. Zhang, C.; Zheng, H.; Gu, Y. Dive into the details of self-supervised learning for medical image analysis. Med. Image Anal. 2023, 89, 102879. [Google Scholar] [CrossRef]
  111. Frueh, M.; Kuestner, T.; Nachbar, M.; Thorwarth, D.; Schilling, A.; Gatidis, S. Self-supervised learning for automated anatomical tracking in medical image data with minimal human labeling effort. Comput. Methods Programs Biomed. 2022, 225, 107085. [Google Scholar] [CrossRef]
  112. Xu, J.; Turk, E.A.; Grant, P.E.; Golland, P.; Adalsteinsson, E. STRESS: Super-Resolution for Dynamic Fetal MRI using Self-Supervised Learning. Med. Image Comput. Comput. Assist. Interv. 2021, 12907, 197–206. [Google Scholar] [CrossRef] [PubMed]
  113. Peng, L.; Wang, N.; Xu, J.; Zhu, X.; Li, X. GATE: Graph CCA for Temporal Self-Supervised Learning for Label-Efficient fMRI Analysis. IEEE Trans. Med. Imaging 2023, 42, 391–402. [Google Scholar] [CrossRef]
  114. Hasan, S.M.K.; Linte, C. STAMP: A Self-training Student-Teacher Augmentation-Driven Meta Pseudo-Labeling Framework for 3D Cardiac MRI Image Segmentation. Med. Image Underst. Anal. 2022, 13413, 371–386. [Google Scholar] [CrossRef]
  115. Azad, R.; Kazerouni, A.; Heidari, M.; Aghdam, E.K.; Molaei, A.; Jia, Y.; Jose, A.; Roy, R.; Merhof, D. Advances in medical image analysis with vision Transformers: A comprehensive review. Med. Image Anal. 2024, 91, 103000. [Google Scholar] [CrossRef] [PubMed]
  116. Singhal, K.; Azizi, S.; Tu, T.; Mahdavi, S.S.; Wei, J.; Chung, H.W.; Tanwani, A.; Cole-Lewis, H.; Pfohl, S.; Payne, P.; et al. Large language models encode clinical knowledge. Nature 2023, 620, 172–180. [Google Scholar] [CrossRef]
  117. Kim, J.K.; Chua, M.; Rickard, M.; Lorenzo, A. ChatGPT and Large Language Model (LLM) Chatbots: The Current State of Acceptability and a Proposal for Guidelines on Utilization in Academic Medicine. J. Pediatr. Urol. 2023, 19, 598–604. [Google Scholar] [CrossRef]
  118. Han, H. Challenges of reproducible AI in biomedical data science. BMC Med. Genom. 2025, 18 (Suppl. 1), 8. [Google Scholar] [CrossRef] [PubMed]
  119. Bradshaw, T.J.; Huemann, Z.; Hu, J.; Rahmim, A. A Guide to Cross-Validation for Artificial Intelligence in Medical Imaging. Radiol. Artif. Intell. 2023, 5, e220232. [Google Scholar] [CrossRef]
  120. Kim, D.W.; Jang, H.Y.; Kim, K.W.; Shin, Y.; Park, S.H. Design Characteristics of Studies Reporting the Performance of Artificial Intelligence Algorithms for Diagnostic Analysis of Medical Images: Results from Recently Published Papers. Korean J. Radiol. 2019, 20, 405–410. [Google Scholar] [CrossRef]
  121. Kelly, C.J.; Karthikesalingam, A.; Suleyman, M.; Corrado, G.; King, D. Key Challenges for Delivering Clinical Impact with Artificial Intelligence. BMC Med. 2019, 17, 195. [Google Scholar] [CrossRef]
  122. Filkins, B.L.; Kim, J.Y.; Roberts, B.; Armstrong, W.; Miller, M.A.; Hultner, M.L.; Castillo, A.P.; Ducom, J.C.; Topol, E.J.; Steinhubl, S.R. Privacy and Security in the Era of Digital Health: What Should Translational Researchers Know and Do about It? Am. J. Transl. Res. 2016, 8, 1560. [Google Scholar]
  123. Hill, D.L.G. AI in Imaging: The Regulatory Landscape. Br. J. Radiol. 2024, 97, 483. [Google Scholar] [CrossRef]
  124. Quaranta, M.; Amantea, I.A.; Grosso, M. Obligation for AI Systems in Healthcare: Prepare for Trouble and Make It Double? Rev. Socionetwork Strateg. 2023, 17, 275–295. [Google Scholar] [CrossRef]
  125. van Leeuwen, K.G.; Schalekamp, S.; Rutten, M.J.C.M.; van Ginneken, B.; de Rooij, M. Artificial Intelligence in Radiology: 100 Commercially Available Products and Their Scientific Evidence. Eur. Radiol. 2021, 31, 3797. [Google Scholar] [CrossRef] [PubMed]
  126. Brady, A.P.; Allen, B.; Chong, J.; Kotter, E.; Kottler, N.; Mongan, J.; Oakden-Rayner, L.; Pinto dos Santos, D.; Tang, A.; Wald, C.; et al. Developing, Purchasing, Implementing and Monitoring AI Tools in Radiology: Practical Considerations. A Multi-Society Statement from the ACR, CAR, ESR, RANZCR & RSNA. J. Med. Imaging Radiat. Oncol. 2024, 68, 7–26. [Google Scholar] [CrossRef] [PubMed]
  127. Alis, D.; Tanyel, T.; Meltem, E.; Seker, M.E.; Seker, D.; Karakaş, H.M.; Karaarslan, E.; Öksüz, İ. Choosing the Right Artificial Intelligence Solutions for Your Radiology Department: Key Factors to Consider. Diagn. Interv. Radiol. 2024, 30, 357. [Google Scholar] [CrossRef]
  128. Huisman, M.; Hannink, G. The AI Generalization Gap: One Size Does Not Fit All. Radiol. Artif. Intell. 2023, 5, e230246. [Google Scholar] [CrossRef]
  129. Eche, T.; Schwartz, L.H.; Mokrane, F.Z.; Dercle, L. Toward Generalizability in the Deployment of Artificial Intelligence in Radiology: Role of Computation Stress Testing to Overcome Underspecification. Radiol. Artif. Intell. 2021, 3, e210097. [Google Scholar] [CrossRef]
  130. Giganti, F.; da Silva, N.M.; Yeung, M.; Davies, L.; Frary, A.; Rodriguez, M.F.; Sushentsev, N.; Ashley, N.; Andreou, A.; Bradley, A.; et al. AI-powered prostate cancer detection: A multi-centre, multi-scanner validation study. Eur. Radiol. 2025. [Google Scholar] [CrossRef]
  131. Rehman, M.H.U.; Hugo Lopez Pinaya, W.; Nachev, P.; Teo, J.T.; Ourselin, S.; Cardoso, M.J. Federated learning for medical imaging radiology. Br. J. Radiol. 2023, 96, 20220890. [Google Scholar] [CrossRef]
  132. Aouedi, O.; Sacco, A.; Piamrat, K.; Marchetto, G. Handling Privacy-Sensitive Medical Data With Federated Learning: Challenges and Future Directions. IEEE J. Biomed. Health Inform. 2023, 27, 790–803. [Google Scholar] [CrossRef] [PubMed]
  133. Young, A.T.; Fernandez, K.; Pfau, J.; Reddy, R.; Cao, N.A.; von Franque, M.Y.; Johal, A.; Wu, B.V.; Wu, R.R.; Chen, J.Y.; et al. Stress Testing Reveals Gaps in Clinic Readiness of Image-Based Diagnostic Artificial Intelligence Models. NPJ Digit. Med. 2021, 4, 10. [Google Scholar] [CrossRef]
  134. Lovejoy, C.A.; Arora, A.; Buch, V.; Dayan, I. Key considerations for the use of artificial intelligence in healthcare and clinical research. Future Healthc. J. 2022, 9, 75–78. [Google Scholar] [CrossRef] [PubMed]
  135. Theriault-Lauzier, P.; Cobin, D.; Tastet, O.; Langlais, E.L.; Taji, B.; Kang, G.; Chong, A.Y.; So, D.; Tang, A.; Gichoya, J.W.; et al. A Responsible Framework for Applying Artificial Intelligence on Medical Images and Signals at the Point of Care: The PACS-AI Platform. Can. J. Cardiol. 2024, 40, 1828–1840. [Google Scholar] [CrossRef]
  136. Sohn, J.H.; Chillakuru, Y.R.; Lee, S.; Lee, A.Y.; Kelil, T.; Hess, C.P.; Seo, Y.; Vu, T.; Joe, B.N. An Open-Source, Vender Agnostic Hardware and Software Pipeline for Integration of Artificial Intelligence in Radiology Workflow. J. Digit. Imaging 2020, 33, 1041–1046. [Google Scholar] [CrossRef] [PubMed]
  137. Arkoh, S.; Akudjedu, T.N.; Amedu, C.; Antwi, W.K.; Elshami, W.; Ohene-Botwe, B. Current Radiology Workforce Perspective on the Integration of Artificial Intelligence in Clinical Practice: A Systematic Review. J. Med. Imaging Radiat. Sci. 2025, 56, 101769. [Google Scholar] [CrossRef]
  138. Naik, N.; Hameed, B.M.Z.; Shetty, D.K.; Swain, D.; Shah, M.; Paul, R.; Aggarwal, K.; Brahim, S.; Patil, V.; Smriti, K.; et al. Legal and Ethical Consideration in Artificial Intelligence in Healthcare: Who Takes Responsibility? Front. Surg. 2022, 9, 862322. [Google Scholar] [CrossRef]
  139. Doherty, G.; McLaughlin, L.; Hughes, C.; McConnell, J.; Bond, R.; McFadden, S. A Scoping Review of Educational Programmes on Artificial Intelligence (AI) Available to Medical Imaging Staff. Radiography 2024, 30, 474–482. [Google Scholar] [CrossRef]
  140. Bouderhem, R. Shaping the future of AI in healthcare through ethics and governance. Humanit. Soc. Sci. Commun. 2024, 11, 416. [Google Scholar] [CrossRef]
  141. Zayas-Cabán, T.; Haque, S.N.; Kemper, N. Identifying Opportunities for Workflow Automation in Health Care: Lessons Learned from Other Industries. Appl. Clin. Inform. 2021, 12, 686. [Google Scholar] [CrossRef]
  142. Nair, M.; Svedberg, P.; Larsson, I.; Nygren, J.M. A Comprehensive Overview of Barriers and Strategies for AI Implementation in Healthcare: Mixed-Method Design. PLoS ONE 2024, 19, e0305949. [Google Scholar] [CrossRef] [PubMed]
  143. Petitgand, C.; Motulsky, A.; Denis, J.L.; Régis, C. Investigating the Barriers to Physician Adoption of an Artificial Intelligence- Based Decision Support System in Emergency Care: An Interpretative Qualitative Study. Stud. Health Technol. Inform. 2020, 270, 1001–1005. [Google Scholar] [CrossRef] [PubMed]
  144. Maleki Varnosfaderani, S.; Forouzanfar, M. The Role of AI in Hospitals and Clinics: Transforming Healthcare in the 21st Century. Bioengineering 2024, 11, 337. [Google Scholar] [CrossRef]
  145. Avanzo, M.; Stancanello, J.; Pirrone, G.; Drigo, A.; Retico, A. The Evolution of Artificial Intelligence in Medical Imaging: From Computer Science to Machine and Deep Learning. Cancers 2024, 16, 3702. [Google Scholar] [CrossRef]
  146. Al Qurri, A.; Almekkawy, M. Improved UNet with Attention for Medical Image Segmentation. Sensors 2023, 23, 8589. [Google Scholar] [CrossRef] [PubMed]
  147. Siddique, N.; Paheding, S.; Reyes Angulo, A.A.; Alom, M.Z.; Devabhaktuni, V.K. Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation. J. Med. Imaging 2022, 9, 064004. [Google Scholar] [CrossRef]
  148. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Transfer learning for medical image classification: A literature review. BMC Med. Imaging 2022, 22, 69. [Google Scholar] [CrossRef]
  149. Kiryu, S.; Akai, H.; Yasaka, K.; Tajima, T.; Kunimatsu, A.; Yoshioka, N.; Akahane, M.; Abe, O.; Ohtomo, K. Clinical Impact of Deep Learning Reconstruction in MRI. Radiographics 2023, 43, e220133. [Google Scholar] [CrossRef]
  150. Lin, D.J.; Johnson, P.M.; Knoll, F.; Lui, Y.W. Artificial Intelligence for MR Image Reconstruction: An Overview for Clinicians. J Magn Reson Imaging. 2021, 53, 1015–1028. [Google Scholar] [CrossRef]
  151. Keshavarz, P.; Bagherieh, S.; Nabipoorashrafi, S.A.; Chalian, H.; Rahsepar, A.A.; Kim, G.H.J.; Hassani, C.; Raman, S.S.; Bedayat, A. ChatGPT in Radiology: A Systematic Review of Performance, Pitfalls, and Future Perspectives. Diagn. Interv. Imaging 2024, 105, 251–265. [Google Scholar] [CrossRef]
  152. Wang, W.T.; Tan, N.; Hanson, J.A.; Crubaugh, C.A.; Hara, A.K. Initial Experience with a COVID-19 Screening Chatbot Before Radiology Appointments. J. Digit. Imaging 2022, 35, 1303. [Google Scholar] [CrossRef] [PubMed]
  153. Al-Aiad, A.; Momani, A.K.; Alnsour, Y.; Alsharo, M. The impact of speech recognition systems on the productivity and the workflow in radiology departments: A systematic review. In Proceedings of the AMCIS 2020 TREOs, Salt Lake City, UT, USA, 10–14 August 2020. [Google Scholar]
  154. Srivastav, S.; Ranjit, M.; Pérez-García, F.; Bouzid, K.; Bannur, S.; Castro, D.C.; Schwaighofer, A.; Sharma, H.; Ilse, M.; Salvatelli, V.; et al. MAIRA at RRG24: A specialised large multimodal model for radiology report generation. In Proceedings of the 23rd Workshop on Biomedical Natural Language Processing, Bangkok, Thailand, 16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 597–602. [Google Scholar]
  155. Kim, S.H.; Wihl, J.; Schramm, S.; Berberich, C.; Rosenkranz, E.; Schmitzer, L.; Serguen, K.; Klenk, C.; Lenhart, N.; Zimmer, C.; et al. Human-AI collaboration in large language model-assisted brain MRI differential diagnosis: A usability study. Eur. Radiol. 2025. [Google Scholar] [CrossRef] [PubMed]
  156. Kwong, J.C.C.; Wang, S.C.Y.; Nickel, G.C.; Cacciamani, G.E.; Kvedar, J.C. The Long but Necessary Road to Responsible Use of Large Language Models in Healthcare Research. NPJ Digit. Med. 2024, 7, 177. [Google Scholar] [CrossRef]
  157. Cacciamani, G.E.; Chu, T.N.; Sanford, D.I.; Abreu, A.; Duddalwar, V.; Oberai, A.; Kuo, C.C.J.; Liu, X.; Denniston, A.K.; Vasey, B.; et al. PRISMA AI Reporting Guidelines for Systematic Reviews and Meta-Analyses on AI in Healthcare. Nat. Med. 2023, 29, 14–15. [Google Scholar] [CrossRef] [PubMed]
  158. Fawzy, N.A.; Tahir, M.J.; Saeed, A.; Ghosheh, M.J.; Alsheikh, T.; Ahmed, A.; Lee, K.Y.; Yousaf, Z. Incidence and Factors Associated with Burnout in Radiologists: A Systematic Review. Eur. J. Radiol. Open 2023, 11, 100530. [Google Scholar] [CrossRef]
  159. Liu, H.; Ding, N.; Li, X.; Chen, Y.; Sun, H.; Huang, Y.; Liu, C.; Ye, P.; Jin, Z.; Bao, H.; et al. Artificial Intelligence and Radiologist Burnout. JAMA Netw. Open 2024, 7, e2448714. [Google Scholar] [CrossRef]
  160. Wiggins, W.F.; Caton, M.T.; Magudia, K.; Glomski, S.A.; George, E.; Rosenthal, M.H.; Gaviola, G.C.; Andriole, K.P. Preparing Radiologists to Lead in the Era of Artificial Intelligence: Designing and Implementing a Focused Data Science Pathway for Senior Radiology Residents. Radiol. Artif. Intell. 2020, 2, e200057. [Google Scholar] [CrossRef]
  161. van Kooten, M.J.; Tan, C.O.; Hofmeijer, E.I.S.; van Ooijen, P.M.A.; Noordzij, W.; Lamers, M.J.; Kwee, T.C.; Vliegenthart, R.; Yakar, D. A Framework to Integrate Artificial Intelligence Training into Radiology Residency Programs: Preparing the Future Radiologist. Insights Imaging 2024, 15, 15. [Google Scholar] [CrossRef]
  162. Kim, B.; Romeijn, S.; van Buchem, M.; Mehrizi, M.H.R.; Grootjans, W. A Holistic Approach to Implementing Artificial Intelligence in Radiology. Insights Imaging 2024, 15, 22. [Google Scholar] [CrossRef]
  163. Filice, R.W.; Mongan, J.; Kohli, M.D. Evaluating Artificial Intelligence Systems to Guide Purchasing Decisions. J. Am. Coll. Radiol. 2020, 17, 1405–1409. [Google Scholar] [CrossRef]
  164. Topff, L.; Ranschaert, E.R.; Bartels-Rutten, A.; Negoita, A.; Menezes, R.; Beets-Tan, R.G.H.; Visser, J.J. Artificial Intelligence Tool for Detection and Worklist Prioritization Reduces Time to Diagnosis of Incidental Pulmonary Embolism at CT. Radiol. Cardiothorac. Imaging 2023, 5, e220163. [Google Scholar] [CrossRef] [PubMed]
  165. Baltruschat, I.; Steinmeister, L.; Nickisch, H.; Saalbach, A.; Grass, M.; Adam, G.; Knopp, T.; Ittrich, H. Smart Chest X-Ray Worklist Prioritization Using Artificial Intelligence: A Clinical Workflow Simulation. Eur. Radiol. 2021, 31, 3837–3845. [Google Scholar] [CrossRef] [PubMed]
Figure 1. PRISMA flowchart showing the two-step study screening process. Adapted from PRISMA Group, 2020 [26].
Figure 1. PRISMA flowchart showing the two-step study screening process. Adapted from PRISMA Group, 2020 [26].
Diagnostics 15 01146 g001
Figure 2. The relationship between AI, ML, DL, and CNN. ML lies within the field of AI and allows computer systems to learn without explicit programming or knowledge. As a subsection of machine learning, deep learning uses computational models similar to the neuronal architecture within the brain, simulating multilayer neural networks in order to resolve complex tasks. A CNN automatically learns and adapts to spatial hierarchies of features through backpropagation using multiple building blocks, allowing analysis of 2D and 3D medical images.
Figure 2. The relationship between AI, ML, DL, and CNN. ML lies within the field of AI and allows computer systems to learn without explicit programming or knowledge. As a subsection of machine learning, deep learning uses computational models similar to the neuronal architecture within the brain, simulating multilayer neural networks in order to resolve complex tasks. A CNN automatically learns and adapts to spatial hierarchies of features through backpropagation using multiple building blocks, allowing analysis of 2D and 3D medical images.
Diagnostics 15 01146 g002
Figure 3. Outline of AI application in radiology workflow in typical clinical setting. AI has potential in reducing scan times during image acquisition and processing, support specific image-based task processes such as segmentation/diagnosis/staging, reduce reading time during reporting, and improve integrated diagnostic processes beyond the reporting phase, including deciding appropriate treatment plans and evaluating prognosis in patients.
Figure 3. Outline of AI application in radiology workflow in typical clinical setting. AI has potential in reducing scan times during image acquisition and processing, support specific image-based task processes such as segmentation/diagnosis/staging, reduce reading time during reporting, and improve integrated diagnostic processes beyond the reporting phase, including deciding appropriate treatment plans and evaluating prognosis in patients.
Diagnostics 15 01146 g003
Table 1. Summary of selected studies.
Table 1. Summary of selected studies.
No.Ref No.Application and Primary Outcome MeasureSample Size *Region
Studied
MRI
Sequences
Used
AI
Technique Used
Key Results
1[27]Develop AI assistant to automatically segment breast tumors by capturing dynamic changes in multi-phase DCE-MRI with spatial-temporal framework2190BreastDCE-MRIUNetDSC 92.6% internal + external testing set
2[28]Evaluate temporally and spatially resolved (4D) radiomics approach to distinguish benign from malignant enhancing breast lesions—avoiding unnecessary biopsies329BreastDCE-MRINeural Network (PCA AI classifier)AUC 80.6% (training dataset) + 83.5% (testing dataset)
3[29]Propose a SwinHR scheme for breast tumor segmentation on DCE-MRI, which integrates prior hemodynamic knowledge of DCE-MRI and Swin Transformer with RLK blocks to extract temporal and spatial information2246BreastDCE-MRICNNDice 0.75–0.81 across different datasets
4[30]Develop image-based automatic deep learning method to classify CMR images by sequence type and imaging plane for improved clinical post-processing efficiency334Chest (Cardiac)Cine-CMR, T1w/T2w/T2*w, Perfusion, LGECNNAccuracy and F1-scores = 81.8% and 0.82 for SVT, 94.3% and 0.94 MVT on the hold-out test set
5[31]Evaluating AI tool in interpreting MR imaging to produce aortic root and valve measurements by comparing accuracy and efficiency with cardiologists35Chest (Cardiac)Cine-CMRCNNICC = 0.98 between AI assessment of aortic root and valve with cardiologists
6[32]Accelerate multiple b-value gas DW-MRI for lung morphometry using deep learning101Chest (Lungs)DW-MRIDC
-RDN
Improved reconstruction acceleration factor of 4
7[33]Develop and test a comprehensive deep learning-based segmentation system to improve automatic upper airway segmentation in static and dynamic MRI160Chest (Lungs)Static/dynamic 2D + 3D MRUNetDSC 0.84 to 0.89 across different
sequences
8[34]Implement a deep-learning approach for segmentation of liver tumors in magnetic resonance imaging using UNet++105Abdomen (Liver)T1w arterial phase, T2wUNet
++
DSC liver and tumor segmentations > 0.9 and >0.6 respectively
9[35]Develop and evaluate a DCNN for automated liver segmentation, volumetry, and radiomic feature extraction on contrast-enhanced portal venous MRI470Abdomen (Liver)3D T1w portal venousDCNNDSC internal + external + public testing set 0.93–0.97
10[36]Train deep learning model to differentiate pathologically proven HCC and non-HCC lesions including lesions with atypical imaging features on MRI.118Abdomen (Liver)Multi-phasic T1wCNNSensitivities/specificities for HCC + non-HCC lesions = 92.7%/82.0% and 82.0%/92.7%; AOC = 0.912
11[37]Explore the feasibility of deep learning (DL) features derived from gadoxetate disodium (Gd-EOB-DTPA) MRI, qualitative features, and clinical variables for predicting early recurrence.285Abdomen (Liver)Contrast enhanced MRIVGGNet-19DL vs. clinical nomogram AUC 0.91–0.95 vs. 0.72–0.75 for training and validation set
12[38]Investigate the utility of DL automated segmentation-based MRI radiomic features and clinical-radiological characteristics in predicting early recurrence after curative resection of single HCC434Abdomen (Liver)T2wUNetAUC 0.743 vs. 0.55–0.64 in widely adopted BCLC and CNLC systems
13[39]Develop and validate a ML mortality risk quantification method for HCC patients using clinical data and liver radiomics on baseline MRI555Abdomen (Liver)T1w breath-hold sequences before + after contrastCNNC-indices of 0.8503 and 0.8234
14[40]Develop and evaluate a CNN based algorithm to evaluate hepatobiliary phase (HBP) adequacy of gadoxetate disodium (EOB)-enhanced MRI and reduce examination time484Abdomen (Liver)EOB-enhanced MRI-HBPCNNKappa 0.67–0.80 and AUC = 0.95–0.97 for internal and external test set
15[41]Develop AI model for prostate segmentation and PCa detection, and explore CAD compared to conventional PI-RADS assessment100Abdomen (Prostate)Not specifiedCNNProstate mean DSCs fine + coarse segmentation = 0.91 + 0.88. PCa diagnosis AI CAD > consistency in internal (kappa 1.00) + external (kappa 0.96) tests vs. radiologists (kappa 0.75)
16[42]Utilize a DL AI workflow for automated EPE grading from prostate T2W MRI, ADC map, and High B DWI.634Abdomen (Prostate)T2W MRI, ADC map, and High B DWICNNSensitivity of 0.67, specificity of 0.73,
accuracy of 0.72
17[43]Study the effect of artificial intelligence (AI) on the diagnostic performance of radiologists in interpreting prostate mpMRI images of the PI-RADS category 3.87Abdomen (Prostate)mpMRICNNAI with radiologists => diagnostic specificity + accuracy (0.695 vs. 0.000 and 0.736 vs. 0.322, p  <  0.001)
18[44]Assess PI-RADS-trained DL algorithm performance and investigate the effect of data size and prior knowledge on the detection of csPCa in biopsy-naïve men with a suspicion of PCa.2734Abdomen (Prostate)mpMRIUNetDL sensitivity for detecting PI-RADS ≥ 4 lesions = 87%; AUC of 0.88.
19[45]To evaluate whether an AI-driven algorithm can detect clinically significant PCa (csPCa) in patients under AS56Abdomen (Prostate)mpMRICNNPatient sensitivity/specificity = 92.5%/31% detection of ISUP ≥ 1 and 96.4%/25% detection of ISUP ≥ 2
20[46]Investigate image quality, efficiency, and diagnostic performance of a DLSB against BLADE for T2WI for GC112Abdomen
(Stomach)
DLSB
- T2w
CNNW test 0.679–0.869 subjective IQS; ICC 0.597–0.670 measure SNR and CNR; Kappa 0.683–0.703 agreement cT stage
21[47]Develop, validate, and deploy deep learning for automated TKV measurement on T2WI MRI studies of ADPKD129Abdomen (Kidney)T2wUNet,
EfficientNet
DSC > 0.97 + Bland–Altman mean percentage difference < 3.6%.
22[48]Analyzes potential cost-effectiveness of integrating an AI–assisted system into differentiation of incidental renal lesions as benign or malignant on MRI follow-upNAAbdomen (Kidney)T2w and T1 post contrastResNet8.76 and 8.77 QALYs for MRI vs. MRI  +  AI strategy
23[49]Develop a DL technique for fully automated segmentation of multiple OARs on clinical abdominal MRI with high accuracy, reliability, and efficiency102AbdomenT1-VIBEU-NetDSC 0.87–0.96
24[50]Develop a DL model for automated detection and classification of lumbar central canal, lateral recess, and neural foraminal stenosis446SpineAxial T2w, Sagittal T1wCNNInternal testing classification = kappa 0.89–0.95. External testing classification = kappa 0.95–0.96
25[51]To assess the speed and interobserver agreement of radiologists for reporting lumbar spinal stenosis with and without DL assistance25SpineAxial T2w, sagittal T1wCNNKappa 0.71 + 0.70 (DL) vs. 0.39 + 0.39 (no DL) for DL-assisted general + in-training radiologists (p < 0.001)
26[52]To externally validate SpineNet predictions for disk degeneration (DD) using Pfirrmann classification and Modic changes (MCs)1331SpineT2wSpineNet (DCNN)DD and MC accuracy = 78% and 86%. PG = Lin concordance correlation coefficient 0.86 and kappa 0.68
27[53]To evaluate SpineNet (SN) ratings compared with those of an expert radiologist in analysis of degenerative features in MRI scans882SpineNot specifiedSpineNet (DCNN)Kappa 0.63–0.77 for PG, kappa 0.07–60 for SL, kappa 0.17–0.57 for CCS
28[54]Evaluate a deep learning model for automated and interpretable classification of central canal stenosis, neural foraminal stenosis, and facet arthropathy from lumbar spine MRI200SpineT2WI axialV-Net, ResNet-50Dural sac + intervertebral disk = DSC 0.93 and 0.94. Localization foramen and facet = 0.72 and 0.83. CCS = kappa 0.54. Neural foraminal stenosis and facet arthropathy = AUC 0.92 and 0.93
29[55]Develop and validate an AI-based model that automatically classifies lumbar central canal stenosis (LCCS) using sagittal T2w MRIs186SpineSagittal T2w MRICNNMulticlass model = kappa 0.86 vs. 0.73–0.85 for readers. Binary model = AUC 0.98; sensitivity 93% + specificity 91% vs. 98–99% and 54–74% for readers
30[56]To compare the image quality and diagnostic performance between standard TSE MRI and accelerated MRI with deep learning (DL)-based image reconstruction for degenerative lumbar spine diseases50SpineT1WI, T2WICNNDL_coarse + DL_fine = significantly higher SNRs T1WI and T2WI, higher CNRs on T1WI, similar CNRs on T2WI
31[57]To develop a DL model to accurately detect ACL ruptures on MRI and evaluate effect on accuracy and efficiency of clinicians3800 MRIsMSKNot specifiedCNNAUC 0.987, sensitivity + specificity 95.1%
32[58]To analyze occupation ratio using a DL framework and study fatty infiltration of supraspinatus muscle using automated region-based Otsu thresholding technique for rotator cuff tears240MSKT1WI fast spin echoCNNMean DSC/accuracy/sensitivity/specificity/relative area difference segmented lesion = 0.97/99.84/96.89/99.92/0.07
33[59]To establish an automated, multitask, MRI-based deep learning system for the detailed evaluation of supraspinatus tendon (SST) injuries3087MSKNot specifiedVGG16AUC of VGG16 = 0.92, 0.97–0.99 for RC-MTL classifiers. ICCs of the radiologists = 0.97–0.99
34[60]Evaluate performance of automated reconstruction algorithm to reconstruct denoised images from undersampled MR images in patients with shoulder pain38MSKTSE fat saturation, sagittal T2w TSE, coronal T1w TSECNNKappa 0.95; Improved SNR and CNR (p < 0.04)
35[61]Investigate feasibility of DL-MRI in shoulder imaging400MSKPD, T2wCNNDL-MRI (Kendall W: 0.588~0.902); non-DL-MRI (Kendall W: 0751~0.865)
36[62]DL assessment of lower extremities in patients with lipedema or lymphedema. Develop pipeline for landmark detection methods and segmentation45MSK3D DIXON MR lymphangiographyEfficientNet-B1
UNet
Z-deviation (landmark) = 4.5 mm and segmentation DSC = 0.989
37[63]To propose a deep neural network for diagnosis of MR Images of tumors of the hand in order to better define preoperative diagnosis and standardize surgical treatment221MSKT1w axial/sagittal/coronalDeepLabv3+ DCNN, TensorFlowAverage confidence level of 71.6% in segmentation of hand tumors
38[64]Develop an Anatomical Context-Encoding Network (ACEnet) to incorporate 3D spatial and anatomical contexts in 2D CNNs for efficient and accurate segmentation of brain structures from MRI30BrainT1wUNetSkull stripping DSC > 0.976
39[65]To verify reliability of volumes automatically segmented using AI-based application and evaluate changes in brain and CSF volume with healthy aging133Brain3D T1wUNetMean ICC 0.986; 95% CI, 0.981–0.990
40[66]To evaluate inference times of novel PSFNet that synergistically fuses SS (scan specific) and SG (scan general) priors for performant MRI reconstruction in low-data regimesNABrainNot specifiedPSFNet3.1 dB higher PSNR, 2.8% higher SSIM, and 0.3 × lower RMSE vs. baselines
41[67]Rapid estimation of multiparametric T1, T2, PD, 3D-QALAS measurements using SSLNABrainT1w, T2w, PD, and inversion efficiency mapsSSL-based QALASRMSE dictionary matching and SSL-QALAS methods = 8.62, 8.75, 9.98, and 2.42% for T1, T2, PD, and IE sequences
42[68]Compare effect of head motion-induced artifacts on the consistency of MR segmentation performed by DL and non-DL methods110BrainT1wReSeg, FastSurfer, KwykDSC 0.91–0.96 for DL vs. 0.88–0.91 non-DL
43[69]Evaluate and compare both image quality and quantitative image-analysis consistency of 60% accelerated volumetric MR imaging sequences with SubtleMR with standard of care (SOC)40Brain3D T1wSubtleMR CNNBoth FAST-DL and SOC statistically superior to FAST for all analyzed features (p < 0.001)
44[70]To demonstrate a 3D CNN that performs expert-level, automated meningioma segmentation and volume estimation on MRI10,099BrainT1wUNetTumor segmentation DSC = 85.2% vs. experts 80.0–90.3%
45[71]Examine efficacy of new approach to classification of brain tumor MRIs through VGG19 features extractor and PGGAN augmentation model233BrainT1w CEVGG19 + PGGANVGG19 + model multiplied dataset ~ 6-fold + 96.59% accuracy. PGGAN model > accuracy VGG19 + model by >1.96%
46[72]Investigate the potential of DL to differentiate active and inactive lesions in MS using non-contrast FLAIR-type MRI data130BrainT1w FLAIRMultiple CNNsAccuracy 85%, sensitivity 95%, specificity 75%, AUC = 0.90. ResNet50 most effective
47[73]To propose a DL framework that combines multi-scale attention guidance and multi-branch feature processing modules to diagnose PD by learning sMRI T2w504BrainT2wDCNNAccuracy 92%, sensitivity 94%, specificity 90% F1 score 95% for identifying PD and HC
48[74]To propose a fully automatic pipeline for measuring the magnetic resonance parkinsonism index (MRPI) using deep learning methods400BrainT1wnnUNet, HRNetInter-rater APE on external datasets = 11.31%
49[75]Automatically diagnose PD from HCs with segmentation using VB-net and radiomics brain features246BrainT1w3D VB-netAUCs 0.988 and 0.976 in training and testing set
50[76]Evaluate whether DLR improves the image quality of intracranial MRA at 1.5 T40BrainAxial and coronal MIP imagesAI Clear IQ Engine—DLRSNR and CNR for basilar artery > in DLR images (p < 0.001)
51[77]Develop and validate intelligent hypoxic–ischemic encephalopathy (HIE) identification model186 (HIE) + 219 (HC)BrainNADLCRNAUC 0.868/0.813/0.798 for training/internal/independent cohorts
52[78]DL model to predict 2-year neurodevelopmental outcomes in neonates with HIE using MRI414BrainT1w, T2w, and diffusion tensor imagingOPiNE (CNN)AUC 0.74 and 63% accuracy in-distribution test set + AUC 0.77 and 78% accuracy out-of-distribution test set
53[79]Automatic segmentation-based method for 2D biometric measurements fetal brain on MRI268BrainT2w SSFSE, (BSSFP), T1w and DWInnUNetCorrelation coefficients CBPD, TCD, LAD and RAD = 0.977, 0.990, 0.817, 0.719
54[80]LVPA-UNet model based on 2D-3D architecture to improve GTV segmentation for radiotherapy in nasopharyngeal carcinoma1010ENTT1wLVPA-UNet>DSC by 2.14%, precision by 2.96%, recall by 1.01%, reduction HD95 0.543 mm
55[81]Evaluate DL-CAD system on radiologists’ interpretation accuracy and efficiency in reading bi-parametric prostate magnetic resonance imaging scans100Abdomen (Prostate)Not specifiedDL-CADDL-CAD vs. radiologist = AUC 0.88 vs. 0.84
56[82]Determine if AI systems for brain MRI could be used as a clinical decision support tool to augment radiologist performance390BrainT1w, FLAIR, GRE, T1-post diffusionCNNs + Bayesian networkRadiology resident performance > with ARIES for = TDx (55% vs. 30%; p < 0.001) and T3DDx (79% vs. 52%; p = 0.002)
57[83]To evaluate the learning progress of less experienced readers in prostate MRI segmentation—in preparation for datasets in AI development100 MRIsAbdomen (Prostate)T2w, DWIRECOMIAUsing DSC > 0.8 as threshold, residents = 99.2%; novices = 47.5%/68.3%/84.0% with subsequent rounds of teaching on accurate segmentations
58[84]Efficacy of AI-generated radiology reports and contribution to the advancement of radiology workflow685 MRI reportsNANAGPT-3.5-turboHigh scores comprehension (2.71 ± 0.73) + patient-friendliness (4.69 ± 0.48) (p < 0.001). 1.12% hallucinations and 7.40% potentially harmful translations
59[85]Evaluate intelligent automation software (Jazz) in assessing new, slowly expanding, and contrast-enhancing MS lesions117 MRIsNeurologyT1wJazzKappa 0.5
60[86]Evaluate protocol determination with CNN classifier based on short-text classification and comparing protocols determined by MSK radiologists6275 MRIsMSKNAConvNetCNN and radiologist = Kappa 0.88. Routine protocols and tumor protocols = AUC 0.977
61[22]Efficient approach for segmentation relevant for diagnosis and treatment of OSAS181ENTT1w, T2w TSEUNetDSC (test set) = 0.89, 0.87, 0.79 (tongue, pharynx, soft palate)
62[87]Develop an efficient preoperative MRI radiomics evaluation approach of ALN status1088BreastT2w, DWI-ADC, T1 + contrastML random forest algorithmALN-tumor radiomic signature AUC = 0.87–0.88 in training, external, prospective-retrospective validation cohort
63[20]Investigate automatically identifying normal scans in ultrafast breast MRI with AI to increase efficiency and reduce workload438BreastTWIST sequencesResNet-34AUC 0.81
64[88]Framework for reconstructing dynamic sequences of 2D CMR from undersampled data using CNNs5Chest (Cardiac)2D CMRCNNPreserves anatomical structures up to 11-fold under sampling
65[89]Fast and high-quality reconstruction of clinical accelerated multi-coil MR data by learning a VN using DL50MSK2D TSECNNVN reconstructions > standard algorithms and acceleration factor 4
66[90]Develop DL based automated OAR delineation method achieving reliable expert performance70Head and NeckT1wUNetAverage DSC 0.86 (0.73–0.97)—6% better than competing methods
Not available (NA), Artificial intelligence (AI), magnetic resonance imaging (MRI), T1-weighted (T1w), T2-weighted (T2w), T2*-weighted (T2*w), area under the curve (AUC), convolutional neural network (CNN), deep learning (DL), deep convolutional neural network (DCNN), deep cascade of residual dense network (DC-RDN), principal component analysis (PCA), diffusion-weighted imaging (DWI), dice similarity coefficient (DSC), dynamic contrast-enhanced (DCE), cardiac magnetic resonance (CMR), late gadolinium enhancement (LGE), root mean squared error (RMSE), peak signal to noise ratio (PSNR), Structural similarity (SSIM), healthy control (HC), hepatocellular carcinoma (HCC), prostate imaging-reporting and data system (PI-RADs), intraclass and concordance correlation coefficients (ICC), prostate cancer (PCa), self-supervised learning (SSL), computer-aided diagnosis (CAD), proton density (PD), multiparametric MRI (mpMRI), International Society of Urological Pathology (ISUP), turbo spin echo (TSE), ear nose throat (ENT), spondylolisthesis (SL), Pfirrmann grading (PG), central canal stenosis (CCS), anterior cruciate ligament (ACL), musculoskeletal (MSK), adaptive radiology interpretation and education system (ARIES). * Numbers are patients, unless stated otherwise.
Table 2. Productivity measures of selected studies.
Table 2. Productivity measures of selected studies.
No.Relevant
Study Finding(s)
Study Center SizeReduced Scan TimeAutomated
Segmentation
Optimized Worklist Triage and WorkflowTime Savings and Workload ReductionDecreased Reading TimeOverall
Productivity Gain
1Time for manual annotation reduced by factor of 20, while maintaining physician accuracy.MNoYesNoNoNoGain
2Fewer unnecessary biopsies of benign breast lesions by ~36.2%, p < 0.001 at the cost of up to 4.5% (n = 4) false negative low-risk cancersSNoYesNoNoNoUnclear—no dedicated reproducibility analysis of the automated lesion segmentation and feature extraction
3Experimental evidence showed superiority to other available automated segmentation methodsMNoYesNoNoNoUnclear—ablation studies used but no reasoning time or comparison to manual methods
4Fully automated multivendor model that classifies cardiac MR images directly from scanner for post processing, validated on external data with high accuracyMNoNoYesNoNoUnclear—no reasoning time and only trained to recognize standard cardiac anatomy
5AI delivered measurements in 2.6 s vs. mean of 334.5 s by cardiologists, without compromising accuracy.SNoNoNoYesNoGain
6Breath-holding reduced 17.8 s to 4.7 s; slice reconstruction within 7.2 msSYesNoNoNoNoGain
7Segmentation time per subject/per slice for static 3D and dynamic 2D MRI less than 1.2 and 0.01 s respectivelySNoYesNoNoNoGain
8Automated segmentation of liver and liver tumor saved >160 and >20 s, respectively, on average per caseSNoYesNoNoNoUnclear—Despite time gains, DSC liver tumor segmentation < 0.7, raising concerns with accuracy
9Total training time 6.35 h. Diverse training dataset of different hepatic etiologies. Mean run time of 0.52 s. MNoYesNoNoNoGain
10CNN accuracy of 87.3% and computational time of <3 msSNoNoYesNoNoGain
11Proposed DL nomogram superior to clinical nomogram in predicting early recurrence for HCC patients post hepatectomyMNoNoYesNoNoUnclear—no reasoning time provided
12DL-assisted automated segmentation using radiomic signatures allows better prediction of early HCC recurrenceSNoYesNoNoNoUnclear—no reasoning time provided
131.11 min risk prediction framework on averageSNoNoYesNoNoGain
1448% examinations achieved <20 min HPB delay normally appliedMYesNoNoNoNoGain
15AI-first read (8.54 s) was faster than readers (92.72 s) and the concurrent-read method (29.15 s)MNoYesNoNoYesGain
16Improved accuracy compared against expert genitourinary radiologistSNoNoYesNoNoUnclear—no reasoning time; only compared with the performance of one expert radiologist
17AI reduced reading time (mean: 351 s, p < 0.001)MNoNoNoNoYesGain
18DL sensitivity for the detection of Gleason > 6 lesions was 85% compared to 91% for a consensus panel of expert radiologistsMNoNoNoNoYesLoss—no improvement in reasoning time shown, poor accuracy vs. expert unless high training case count met
19Accurate automated AI-algorithm for prostate lesion detection and classificationSNoNoYesNoNoUnclear—no reasoning time provided; adopting to clinical practice unclear
20Vs. BLADE, DLSB reduced acquisition time of T2WI from 4 min 42 s/patient to 18 s/patient, with better overall image and comparable accuracySYesNoNoNoNoGain
2151% reduction in radiologist time for model-assisted segmentationSNoYesNoNoNoGain
22Additional use of AI-based algorithm may be cost-effective alternative in differentiation of incidental renal lesions on MRINANoYesNoNoNoUnclear—integration into clinical practice not clarified
23Inference completed within 1 min for a three-dimensional volume of 320 × 288 × 180SNoYesNoNoNoGain
24DL model detecting and classifying lumbar spinal stenosis at MRI comparable with radiologists for central canal and lateral recess stenosisSNoNoNoNoYesUnclear—improvements on interpretation time not discussed
25DL-assisted radiologists = reduced interpretation time per spine MRI study. Mean of 124–274 s reduced to 47–71 s (p < 0.001)SNoNoNoNoYesGain
2620.83% of disks rated differently by SpineNet compared to human raters, but only 0.85% of disks had a grade difference > 1SNoNoNoNoYesUnclear—improvements on interpretation time not discussed
27SpineNet recorded > pathological features in PG vs. radiologistSNoNoNoNoYesUnclear—improvements on interpretation time not discussed
28Accurate automated grading lumbar spinal stenosis and facet arthropathy using axial sequences onlySNoNoNoNoYesUnclear—improvements on interpretation time not discussed
29Accurate AI-based lumbar central canal stenosis classification on sagittal MR imagesSNoNoNoNoYesUnclear—improvements on interpretation time not discussed, no external test set
30Accelerated protocol reduced MRI acquisition time by 32.3% compared to standard protocolSYesNoNoNoNoGain
31AI assistance significantly improved accuracy (>96%) with reduction in diagnostic time observed (up to 9.7 s in expert or 24.5 s in trainee groups)MNoNoNoNoYesGain
32Analysis using freeware computer program that analyzed muscle atrophy and fatty infiltration MR images in <than 1 sSNoYesNoNoNoGain
33Automated multitask DL system in diagnosing SST injuries with comparable accuracy to expertsSNoNoNoNoYesUnclear—improvements on interpretation time not discussed
34Reconstruction performed on the scanner in ~80 s for CS DL reconstructions compared to approximately ~40 s for standard CS reconstructionsSYesNoNoNoNoLoss
35Scan time of DL-MRI (6 min 1 s) ~50% decreased compared with that of non-DL-MRI (11 min 25 s)SYesNoNoNoNoGain
36Prediction time/patient for first + second landmark detection = 2.7 s + 0.1 s. Mean prediction time/patient for tissue segmentation 8 sSNoYesNoNoNoGain
37Automated segmentation of hand tumors with 71.6% accuracySNoYesNoNoNoUnclear—poor accuracy, improvements on segmentation time not discussed
38Slight computation cost of 16.6% increase from baseline; MRI head scan segmentation within ~9 sSNoYesNoNoNoGain
39Brain and intracranial CSF volume ratios measured within 1 minSNoYesNoNoNoGain
40PSFNet required order of magnitude lower samples and faster inference vs. SG methodsSYesNoNoNoNoGain
41Reconstruction multiparametric maps using pre-trained SSL-QALAS model < 10 s, no external dictionary required. Scan-specific tuning with target subject < 15 minSYesNoNoNoNoGain
42ReSeg DCNN inference time 11 s with >consistent segmentationsSNoYesNoNoNoGain
4360% reduction sequence scan time using FAST-DL—from 4 to 6 min 56 s to 2 min 40–44 s, with >qualitySYesNoNoNoNoGain
44Reduced segmentation time by 99% (2 s/segmentation p < 0.001) + more accurateSNoYesNoNoNoGain
45Accurate DL model helping radiologists classify brain tumor types on MRI with less training data neededSNoNoNoYesNoUnclear—training time not compared, limited brain tumor types, inconsistent image sizes due to computational limits
46Multiple CNNs allow accurate discerning active vs. inactive lesions without reliance on contrast agents on FLAIR imagesMNoNoNoNoYesUnclear—improvements on interpretation time not discussed
47DL model can accurately distinguish PD and HC using T2w MRISNoNoNoNoYesUnclear—improvements on interpretation time not discussed
48Automated DL pipeline can accurately predict MRPI comparable with manual measurementsSNoYesNoNoNoUnclear—limited training data, segmentation time not discussed
49Automated segmentation of brain into 109 regions with distinguishing PD from HC in 1 minSNoNoNoNoYesGain
50DLR = higher quality 1.5 T intracranial MRA images with improved visualization of arteriesSYesNoNoNoNoLoss—multifold increase in imaging time, benefits high image quality on disease detectability not detailed
51DLCRN model allowing accurate identification of HIE in neonates using MRISNoYesNoNoNoUnclear—improvement on segmentation time and clinical application not detailed
52DL analysis neonatal brain MRI yielded high performance for predicting 2-year neurodevelopmental outcomesMNoYesNoNoNoUnclear—segmentation times and potential real time clinical application not discussed
53Accurate deep segmentation network using 2D biometric measurements of fetal brain on MRISNoYesNoNoNoUnclear—only normal fetal brains in training data, segmentation time vs. manual methods not detailed
54Accurate GTV segmentation using T1w MRI for radiotherapy in nasopharyngeal carcinomaMNoYesNoNoNoUnclear—segmentation time not discussed
55Median reading time reduced by 21% from 103 to 81 s (p < 0.001)MNoNoNoNoYesGain
56Improved diagnostic accuracy of radiology resident using ARIES support toolSNoNoNoYesNoUnclear—time gains or losses with support tool not discussed
57Segmentation time of experienced resident 7 min 22 s—7 min 50 s per caseSNoNoNoYesNoGain
58Good efficacy AI-generated radiology in terms of report summary accuracy and patient-friendlinessSNoNoNoYesNoUnclear—artificial hallucinations and harmful translations might equate to more work to verify reports
59Significantly more MS lesions detected vs. baseline. Reading time 2 min 33 s/case.MNoNoNoNoYesGain
60Training time approximately 10 minSNoNoYesNoNoGain
61Consistent with expert readings. Processing one dataset = ~2 s using modern GPUSNoYesNoNoNoGain
62Accurate signature that could be preoperatively used for identifying patients with ALN metastasis in early-stage invasive breast cancerMNoNoNoYesNoUnclear—potential time gains or losses with clinical application not discussed
63Workload and scanning time reductions = 15.7% and 16.6%SYesNoNoYesNoGain
64Reconstruction < 10 s with each complete dynamic sequence and <23 ms for 2D caseSYesNoNoNoNoGain
65Reconstruction time of 193 ms on single graphics cardSYesNoNoNoNoGain
66Accurate DL framework for automated delineation of OARs in HN radiotherapy treatment planningSNoYesNoNoNoUnclear—segmentation time not discussed
Single center study (S), multicenter study (M), hepatocellular carcinoma (HCC).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nair, A.; Ong, W.; Lee, A.; Leow, N.W.; Makmur, A.; Ting, Y.H.; Lee, Y.J.; Ong, S.J.; Tan, J.J.H.; Kumar, N.; et al. Enhancing Radiologist Productivity with Artificial Intelligence in Magnetic Resonance Imaging (MRI): A Narrative Review. Diagnostics 2025, 15, 1146. https://doi.org/10.3390/diagnostics15091146

AMA Style

Nair A, Ong W, Lee A, Leow NW, Makmur A, Ting YH, Lee YJ, Ong SJ, Tan JJH, Kumar N, et al. Enhancing Radiologist Productivity with Artificial Intelligence in Magnetic Resonance Imaging (MRI): A Narrative Review. Diagnostics. 2025; 15(9):1146. https://doi.org/10.3390/diagnostics15091146

Chicago/Turabian Style

Nair, Arun, Wilson Ong, Aric Lee, Naomi Wenxin Leow, Andrew Makmur, Yong Han Ting, You Jun Lee, Shao Jin Ong, Jonathan Jiong Hao Tan, Naresh Kumar, and et al. 2025. "Enhancing Radiologist Productivity with Artificial Intelligence in Magnetic Resonance Imaging (MRI): A Narrative Review" Diagnostics 15, no. 9: 1146. https://doi.org/10.3390/diagnostics15091146

APA Style

Nair, A., Ong, W., Lee, A., Leow, N. W., Makmur, A., Ting, Y. H., Lee, Y. J., Ong, S. J., Tan, J. J. H., Kumar, N., & Hallinan, J. T. P. D. (2025). Enhancing Radiologist Productivity with Artificial Intelligence in Magnetic Resonance Imaging (MRI): A Narrative Review. Diagnostics, 15(9), 1146. https://doi.org/10.3390/diagnostics15091146

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop