Next Article in Journal
A Unified Deep Learning Ensemble Framework for Voice-Based Parkinson’s Disease Detection and Motor Severity Prediction
Previous Article in Journal
Design and Bench Testing of a Novel, Pediatric, Non-Invasive, Bubble Bilevel Positive Pressure Ventilation Device
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive RAG-Assisted MRI Platform (ARAMP) for Brain Metastasis Detection and Reporting: A Retrospective Evaluation Using Post-Contrast T1-Weighted Imaging

1
Artificial Intelligence Center, China Medical University Hospital, Taichung 40447, Taiwan
2
Department of Medical Imaging, China Medical University Hospital, Taichung 40447, Taiwan
3
Department of Nuclear Medicine and PET Center, China Medical University Hospital, Taichung 40447, Taiwan
4
Department of Bioinformatics and Medical Engineering, Asia University, Taichung 41354, Taiwan
5
Graduate Institute of Biomedical Sciences, School of Medicine, College of Medicine, China Medical University, Taichung 40447, Taiwan
6
Division of Cardiovascular Medicine, Department of Medicine, China Medical University Hospital, School of Medicine, China Medical University, Taichung 40447, Taiwan
7
School of Medicine, China Medical University, Taichung 40447, Taiwan
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(7), 698; https://doi.org/10.3390/bioengineering12070698
Submission received: 15 May 2025 / Revised: 18 June 2025 / Accepted: 19 June 2025 / Published: 26 June 2025
(This article belongs to the Section Biosignal Processing)

Abstract

This study aimed to develop and evaluate an AI-driven platform, the Adaptive RAG Assistant MRI Platform (ARAMP), for assisting in the diagnosis and reporting of brain metastases using post-contrast axial T1-weighted (AX_T1+C) MRI. In this retrospective study, 2447 cancer patients who underwent MRI between 2010 and 2022 were screened. A subset of 100 randomized patients with confirmed brain metastases and 100 matched non-cancer controls were selected for evaluation. ARAMP integrates quantitative radiomic feature extraction with an adaptive Retrieval-Augmented Generation (RAG) framework based on a large language model (LLM, GPT-4o), incorporating five authoritative medical references. Three board-certified neuroradiologists and an independent LLM (Gemini 2.0 Pro) assessed ARAMP performance. Metrics of the assessment included Pre-/Post-Trained Inference Difference, Inter-Inference Agreement, and Sensitivity. Post-training, ARAMP achieved a mean Inference Similarity score of 67.45%. Inter-Inference Agreement among radiologists averaged 30.20% (p = 0.01). Sensitivity for brain metastasis detection improved from 0.84 (pre-training) to 0.98 (post-training). ARAMP also showed improved reliability in identifying brain metastases as the primary diagnosis post-RAG integration. This adaptive RAG-based framework may improve diagnostic efficiency and standardization in radiological workflows.

1. Introduction

Brain metastasis management represents a complex clinical challenge. Brain metastases affect approximately 20–40% of cancer patients and contribute substantially to morbidity and mortality [1]. These metastases, which develop when cancer cells spread from primary tumors to the brain tissue, often indicate advanced disease stages and are associated with limited therapeutic options. The presence of brain metastases necessitates early and accurate diagnosis to guide clinical decision-making, treatment strategy selection, and ultimately improve patient outcomes [2,3]. Magnetic resonance imaging (MRI) remains the gold standard for detecting brain metastases, owing to its superior sensitivity in soft tissue differentiation and high spatial resolution for detecting small lesions [4]. However, despite continuous technological advances in imaging acquisition and processing, diagnostic workflows continue to face substantial challenges in both efficiency and reliability.
Current clinical imaging diagnostic practice relies heavily on radiologists’ interpretation of complex and variable imaging presentations, a task that demands considerable expertise and time. Radiologists must carefully evaluate multiple imaging series, assess various lesion characteristics on MRI, and give a diagnosis or multiple differential diagnoses. Imaging interpretation is expertise-dependent and inherently subjective, with inter-interpreter variability. The complex and lengthy imaging diagnosis process may potentially delay treatment initiation and compromise patient care and clinical outcomes [5]. Furthermore, the increasing volume of medical imaging data, coupled with the growing complexity of diseases, creates substantial workflow bottlenecks in clinical practice. All of the above exacerbate a common challenge in the global medical community: the shortage of radiologists. These limitations underscore the pressing need for innovative imaging diagnostic tools that can enhance both accuracy and efficiency in clinical workflows while maintaining consistent diagnostic standards.
Artificial intelligence (AI) and radiomics have emerged as promising solutions to address these challenges. AI technologies, particularly deep learning algorithms, enable automated tumor detection, segmentation, and classification, facilitating the process of reading complex imaging data with remarkable speed and consistency. Radiomics complements these capabilities by extracting quantitative imaging features—including texture, shape, and intensity characteristics—to distinguish malignant from benign lesions and predict their clinical outcome with greater precision [6,7]. These approaches can potentially identify subtle imaging biomarkers that might escape human detection. Nevertheless, clinical implementation of these technologies faces several critical obstacles, including the need for standardized data acquisition protocols, limited model interpretability, concerns about generalizability across different patient populations, imaging protocols, and imaging scanners, and stringent validation requirements to ensure clinical safety and efficacy [8,9,10,11].
Retrieval-Augmented Generation (RAG) frameworks represent a recent advancement in radiology that potentially overcomes several limitations of conventional AI systems [12]. RAG incorporates external medical knowledge bases into AI models, enhancing diagnostic precision by leveraging established clinical expertise and enabling real-time retrieval of relevant clinical data. This sophisticated approach effectively addresses various challenges, including bias mitigation in model predictions, reduction of interpretation variability, and improvement of AI system transparency, while supporting personalized diagnostics based on individual patient characteristics and clinical contexts [13]. However, RAG systems face a significant limitation known as “hallucination,” where the generated outputs may contain inaccurate or unverifiable information derived from imperfect knowledge integration or retrieval processes. This limitation is particularly crucial in medical applications, where erroneous outputs can directly impact patient safety and clinical outcomes. To address these concerns, researchers have developed domain-specific RAG models designed to enhance accuracy and reliability in medical contexts through specialized knowledge bases and rigorous validation protocols [14].
In recent years, applying large language models (LLMs) such as GPT-4o and Google Gemini in the medical field has garnered widespread attention, particularly in evaluating their performance in professional examinations. Research indicates that these models demonstrate strong capabilities when processing text-based medical knowledge questions. For example, a study by [15] found that GPT-4o achieved a high score of 82.1% on text-only multiple-response questions (focusing on recall and understanding) from the European Diploma in Radiology (EDiR) examination, significantly outperforming actual candidates (radiologists, average 49.4%), especially excelling in identifying correct answers (high true positive rate).
However, their performance presents a different picture when evaluating LLMs’ ability to handle complex medical questions requiring visual input [15]. Ref [16] evaluated the capabilities of GPT-4o and Gemini regarding image-based questions from a neurosurgery board examination question bank. The results showed that GPT-4o’s accuracy (51.45%) surpassed Gemini’s (39.58%), particularly on second-order questions requiring diagnostic reasoning, and performed better in subspecialties like pathology/histology and radiology [16]. Notably, the study also found that the presence or absence of images had impacts on the models’ accuracy, suggesting that the models rely more on the textual context of the questions rather than genuinely analyzing the image content to arrive at answers.
While LLMs have reached or even surpassed human expert levels in handling text-based medical knowledge recall, their capabilities remain significantly limited in tasks requiring the integration of image interpretation and complex clinical reasoning. This indicates that image analysis and multimodal integration capabilities are key challenges for future development of multi-modal AI in the medical community.
In this study, we present a novel brain MRI Assistant platform that leverages adaptive RAG technology, namely Adaptive RAG-Assisted MRI Platform (ARAMP), for the diagnosis and reporting of brain metastases. This platform integrates advanced feature extraction for analyzing post-contrast T1-weighted axial view (AX_T1+C) MRI images with adaptive learning mechanisms, enabling seamless incorporation of multimodal data. Through real-time retrieval of relevant knowledge and contextualized generation of diagnostic insights, the system addresses key challenges in neuroradiology, such as differentiating metastases from other disease conditions and reducing diagnostic variability. The primary objective of this study was to develop and evaluate the performance of ARAMP in detecting brain metastases on AX_T1+C brain MRI images and generating imaging reports. To ensure transparency and reproducibility, all items in the CLAIM 2024 Checklist were addressed and documented in Supplementary Materials Table S6. This includes details on the study design, data preprocessing, radiologist consensus, evaluation metrics, and clinical implications.

2. Materials and Methods

2.1. Study Population

This retrospective study was approved by the Institutional Review Board of China Medical University Hospital (CMUH) (IRB No. CMUH111-REC1-206). The requirement for informed consent was waived due to the retrospective design and the use of de-identified patient data.
This retrospective study utilized data from patients who underwent brain MRI examinations at our institution between 2010 and 2022. Patient records were electronically queried and cross-referenced with the Major Illness Database (MajorIllnessDim) and the Cancer Registry Database (CaRecord) to identify individuals with a confirmed cancer diagnosis. Inclusion criteria were (1) age greater than or equal to 20 years and (2) a confirmed cancer diagnosis within 60 days prior to the MRI examination or within one year after the examination. Patients not meeting these criteria were excluded from the study. The selection was stratified by cancer type to reflect the distribution observed in the overall cohort.
Detailed inclusion and exclusion criteria, as well as sampling procedures, are documented in Supplementary Materials Table S6 (Items 8, 10, 21, and 35).
From this cohort, we randomly selected 100 cases for in-depth analysis, ensuring that the sample reflected the original cohort in terms of cancer type, age, and gender distribution.
The sampling was stratified by primary cancer diagnosis, with matched age and sex distribution to enhance representativeness (see Table 1 and Flowchart in Supplementary Materials Figure S1). To evaluate the performance of our large language model (LLM) model in identifying potential indicators of brain metastases by post-contrast trans-axial T1-wieghted (AX_T1+C) MRI with intravenous administration of a gadolinium-based contrast agent, a non-cancer control cohort was also established. This cohort consisted of 100 individuals randomly selected from a pool of patients who had undergone MRI examination containing AX_T1+C imaging but did not have a documented cancer diagnosis within the specified timeframe. This sampling design allowed balanced radiomic feature comparison and enabled controlled input for GPT-4o inference. All patient DICOM datasets were de-identified and processed through a standardized radiomics pipeline, generating both numerical features and image composites (10-slice PNG format), as illustrated in Supplementary Materials Table S2 and the workflow diagram in Supplementary Materials File S1, Figure S2.
These control subjects were matched to the cancer cohort in terms of age and gender distribution to minimize potential confounding factors and ensure baseline comparability for subsequent analyses. The matching process involved replicating the proportional distribution of age and gender observed in the cancer patient cohort within the control cohort.

2.2. Subgroup Analysis

To further investigate the relationship between imaging features and the presence of brain metastases, we conducted subgroup analyses using the same stratified subsample of 100 cancer patients previously described in Section 2.1. This group was randomly selected from the initial cohort of 2447 patients, with age and gender distributions matched to the broader cancer population to ensure representativeness. Among the 100 patients, 52 were male, and 48 were female. The age distribution was balanced across gender and spanned a wide range, with the largest groups falling between the ages of 40 and 70 years old.
This stratified sampling approach was designed to preserve representativeness and minimize demographic bias, as required by CLAIM Items 8 and 10. Detailed metadata, including age, gender, and diagnosis codes, were cross-referenced using the institutional cancer registry. The clinical cancer diagnoses within this subsample were diverse, with lung cancer comprising the majority (n = 65). Other cancer types included breast cancer (n = 11), colorectal cancer (n = 3), esophageal carcinoma (n = 3), and various other cancers, such as melanoma, non-Hodgkin’s disease, and prostate cancer. This composition reflects national epidemiological trends for brain metastases in East Asia, ensuring external relevance of the test set (see Supplementary Materials Table S6, CLAIM Item 22). The complete demographic and clinical distribution used for radiomic analysis and GPT-4o inference input is listed in Table 1. The full distribution of age, gender, and clinical diagnoses is detailed in Table 1.
For each patient in this subsample, the corresponding MRI images and reports were carefully reviewed by three board-certified radiologists, two with over 8 years and more than 40 years of experience in neuroradiology, respectively, to determine if there was any documented evidence of tumor metastasis (Mets). This manual review served as the gold standard for comparison with the LLM model’s predictions. To ensure consistency and reduce interpretive variability, the review protocol was standardized across cases and supported by structured report templates. All radiologist ratings were recorded in Supplementary Materials Table S5.

Inter-Rater Reliability Analysis

To evaluate the effectiveness of model training and the stability of its outputs, two intraclass correlation coefficient (ICC) analyses were conducted using a two-way mixed effects model (Type C, consistency), based on physician ratings using a 5-point Likert scale. Scores were converted to a 0–100% scale for normalization and comparative analysis. Full rating matrices for pre- and post-training outputs, as well as Gemini-generated responses, are detailed in Supplementary Materials Table S5. Corresponding CLAIM items are addressed in Supplementary Materials Table S6 (Items 36, 37, and 43).

2.3. Data Preparation for Radiomic and Multimodal Analysis

2.3.1. Data Preprocessing

All directories containing DICOM files within the institutional database were systematically identified using the PyDicom version 3.0 function. This function recursively scanned the designated storage locations to locate all folders containing DICOM images. Subsequently, PyDicom functions were employed to filter and retrieve DICOM files specifically related to the AX_T1+C MRI. These functions utilized predefined criteria to identify and extract only the relevant DICOM files from the larger pool of imaging data, establishing a comprehensive list of file paths for subsequent processing. After selection, all DICOM files underwent de-identification and were processed using PyRadiomics. Images were resampled to isotropic 1 mm3 voxels using B-spline interpolation, and intensity normalization was applied (scale = 100, voxel shift = 1000). Feature extraction included original, LoG (σ = 3, 5 mm), and wavelet (Haar) transformations, with discretization bin width set to 25. These parameters are detailed in Supplementary Materials Table S2.

2.3.2. Feature Extraction

Individual DICOM image files were loaded and processed using PyDicom, a specialized library for handling DICOM image data. This process involved extracting the image pixel data, which represents the raw intensity values for each pixel in the image. Feature extraction was then performed using the PyRadiomics library’s feature extractor module. This module provides a standardized and validated framework for extracting a wide range of quantitative imaging features from medical images.
In this study, parameters within the feature extractor module were configured to enable extraction of four specific categories of imaging features:
(1)
First-order statistical features: These features capture basic intensity distribution characteristics within the image, such as mean, standard deviation, skewness, and kurtosis.
(2)
Shape features: These features quantify the geometric properties of the region of interest (ROI), including volume, surface area, sphericity, and maximum diameter.
(3)
LoG features: These features are based on the Laplacian of Gaussian (LoG) filter, which enhances edges and boundaries within the image. LoG features capture the texture and heterogeneity of the ROI.
(4)
Wavelet features: These features decompose the image into different frequency subbands, capturing texture information at multiple scales.
To ensure cross-patient comparability, all images were spatially resampled to an iso-tropic voxel spacing of [1, 1, 1] mm using B-spline interpolation (interpolator: sitkBSpline). Intensity normalization was applied with a voxel array shift of 1000 and scaling factor of 100. Laplacian of Gaussian (LoG) features were extracted using σ values of 3 and 5 mm, while wavelet decomposition was performed using Haar basis functions at level 1. The bin width for gray level discretization was set to 25. These parameter settings were applied uniformly to all images and are summarized in Supplementary Materials Table S2.
Features were calculated for each patient’s image data in conjunction with a hypothetical label mask, which represents the region of interest (ROI) within the image. These extracted features were then organized and stored in a pandas DataFrame, a tabular data structure suitable for further analysis and machine learning applications.

2.3.3. Image Enhancement and Normalization

To optimize the visual representation of the images and mitigate the impact of potential noise or artifacts, image values were enhanced and normalized by utilizing the 1st and 99th percentiles of the image minimum and maximum values. This approach effectively compressed the numerical range of the image data to a standardized interval of [0, 1], ensuring that the images were represented within a consistent and comparable scale.
This percentile-based contrast stretching was applied prior to PNG image generation to ensure consistent grayscale representation for subsequent GPT-4o multimodal inference.
All enhanced images were saved in de-identified PNG format, with 10 contiguous slices combined into each composite image. This transformation pipeline is illustrated in Supplementary Materials File S1, Figure S2.
The complete set of enhancement parameters, image normalization strategy, and saving conventions used in this study were aligned with preprocessing standards defined in CLAIM 2024 (Supplementary Materials Table S6, Items 13 and 24).

2.3.4. DICOM-to-PNG Conversion

Following the enhancement and normalization procedures described in Section 2.3.3, the processed DICOM images were subsequently converted into the more widely compatible PNG format to facilitate visualization and downstream model input. This conversion process involved sequentially arranging the DICOM images, with a maximum of 10 layers combined into a single PNG image. The arrangement within the PNG followed a two rows by five columns matrix structure, effectively organizing the image layers into a visually coherent grid. OpenCV, a comprehensive library for image processing and computer vision tasks, was used to save these converted images in PNG format.
This layout design ensures consistency across inputs and enables rapid visual inspection of brain lesion morphology by both human reviewers and GPT-4o’s vision encoder.
The PNG images were de-identified and stored alongside structured metadata (e.g., radiomic features, patient ID, and axial location) in a standardized folder structure to support reproducibility and linkage with inference outputs.
The complete image conversion and arrangement process is depicted in Supplementary Materials File S1, Figure S2 and complies with CLAIM 2024 standards for image format compatibility and interpretability (Supplementary Materials Table S6, Item 24).

2.4. Feature Extraction and De-Identification

2.4.1. Feature Classification

Features extracted from the AX_T1+C DICOM images of both the cancer and non-cancer groups were classified as “Meta” (abnormal), indicating potential tumor metastasis, or “No_abnormalities” (normal), suggesting the absence of detectable abnormalities. These classified features were then stored in separate Excel files (Meta.xlsx and No_abnormalities.xlsx) to maintain organization and facilitate subsequent analyses. The labeling process was based on radiologist consensus review and corresponding clinical reports. The “Meta” group consisted of 100 cases with confirmed brain metastases, while the “No_abnormalities” group included 100 age- and gender-matched controls with no known cancer history.
A total of 1051 radiomic features were computed for each case, including original, Laplacian of Gaussian (LoG), and wavelet-derived features. Feature naming and trans-formation categories are documented in Supplementary Materials Table S1.
These labeled feature sets formed the core input for downstream GPT-4o-based inference and were version-controlled to ensure reproducibility. CLAIM checklist Items 16 and 21 (see Supplementary Materials Table S6) are addressed by this structured classification and data segregation process.

2.4.2. Data Merging and De-Identification

Following the initial classification and storage, the two Excel files (Meta.xlsx and No_abnormalities.xlsx) were merged into a single, comprehensive dataset named No_abnormalities & Brain_Meta (De-identification).xlsx. This merged dataset contained both abnormal (Meta) and normal (No_abnormalities) samples, providing a unified data source for model training and evaluation. Prior to merging, all files were verified for format consistency, and feature columns were harmonized across the two classes to ensure compatibility. To ensure patient privacy and comply with data protection regulations, patient IDs and other sensitive information were removed from this dataset. Only the classification labels (Meta or No_abnormalities) were retained to preserve the essential information for model training and analysis. The final dataset included 200 subjects with 1051 radiomic features per case and was structured in tabular format for compatibility with both GPT-4o-prompting pipelines and external model evaluation.
The full data processing workflow is visualized in Supplementary Materials File S1 Figure S2, and this protocol adheres to the transparency and safety criteria described in CLAIM 2024 (see Supplementary Materials Table S6, Items 18, 21, and 43).

2.5. Creating the Brain ARAMP

2.5.1. Reference Integration

To enhance MRI interpretation, a retrieval-augmented generation (RAG) system was implemented.
This framework enabled GPT-4o to dynamically access five authoritative medical references [17,18,19,20,21], providing domain-specific knowledge to support structured lesion analysis and clinical reasoning. The RAG implementation employed a version-controlled, keyword-indexed knowledge base constructed from core neuroradiology textbooks (e.g., Osborn’s Brain, ESNR Textbook). During inference, lesion descriptors such as “ring-enhancing lesion with peritumoral edema” triggered semantic-based retrieval of relevant literature. The update frequency of this knowledge base aligned with published guideline revisions. The literature base was version-controlled and aligned with updated clinical guidelines.

2.5.2. Dataset Validation

A structured dataset (No_abnormalities & Brain_Meta.xlsx) containing the extracted imaging features corresponding classifications was used for validation. The dataset included a “TYPE” column indicating feature categories (e.g., original_shape_VoxelVolume), allowing for easy identification and retrieval of specific feature categories. The “No_abnormalities” and “Brain_Meta” feature values enabled direct comparison and analysis of the feature distributions between the two groups. Radiomic features were extracted using PyRadiomics v3.0, with key settings including voxel resampling to [1, 1, 1] mm (B-spline interpolation), intensity normalization (scale: 100, shift: 1000), discretization bin width = 25, LoG filtering (σ = 3 and 5 mm), and Haar wavelet decomposition (level 1). Feature classes included first-order statistics, shape descriptors, and texture-based features. Full extraction parameters are provided in Supplementary Materials Table S2.

2.5.3. Image Data and Workflow

Composite images, each containing up to 10 sequentially arranged MR images, were used as the input for the GPT model. These composite images provided a comprehensive representation of the patient’s brain anatomy and pathology, supporting structured diagnostic reasoning by the model. A structured workflow, consisting of four distinct steps, was implemented to guide the brain MRI process by the GPT model. This process was applied to a focused evaluation cohort consisting of 100 cancer patients with confirmed brain metastases, as determined by radiologist consensus and original MRI reports. The inclusion of only metastasis-positive cases was intentional to assess reasoning improvement rather than classification accuracy; thus, specificity and overall accuracy were not computed.
  • Step 1: Image Definition and Assessment: this initial step involved defining the imaging type (MRI), specifying the organ being scanned (brain), and setting appropriate MRI scan parameters (e.g., T1-weighted, T2-weighted; axial, sagittal, coronal views; pre-contrast and post-contrast) based on the clinical context and diagnostic objectives.
  • Step 2: Lesion Analysis and Differential Diagnosis: in this step, the model analyzed any detected lesions, characterizing their size, intensity, and enhancement patterns. It then developed a differential diagnosis based on these lesion characteristics, dynamically retrieving and referencing relevant diagnostic criteria from the integrated medical literature [17,18,19,20,21]. Lesion descriptors served as semantic triggers within the RAG system to retrieve guideline-based diagnostic criteria, enabling precise matching with literature-defined imaging patterns.
  • Step 3: Treatment Strategy for Brain Metastases: if brain metastases were suspected or confirmed, the model proceeded to formulate a potential treatment plan based on the imaging findings and diagnosis. This involved matching the imaging features with established treatment criteria and dynamically integrating relevant treatment recommendations from the medical literature [17,18,19,20,21]. Literature retrieval was tailored through prompt-linked indexing aligned with oncology treatment standards, allowing context-specific adaptation.
  • Step 4: Conclusion and Follow-Up: in the final step, the model summarized the diagnostic findings, proposed a recommended treatment plan (if applicable), and defined appropriate follow-up requirements, tailoring imaging intervals and plans using dynamic references to the medical literature [17,18,19,20,21]. Each output included traceable reasoning and literature-cited justifications to enhance interpretability and transparency.
This inference protocol reflects ARAMP’s intended role as a decision-support tool applied after radiological abnormalities are identified, focusing on interpretive accuracy rather than primary detection.

2.6. Inquiry Procedure and Evaluation

To evaluate the performance of the developed Brain ARAMP, a structured inquiry procedure was conducted. The evaluation focused on comparing ARAMP’s accuracy and consistency to interpretations by expert radiologists, alongside performance benchmarks from a foundation LLM (Gemini) and relevant baseline models. This evaluation was conducted exclusively on a cohort of 100 patients with imaging-confirmed brain metastases, as verified through radiologist consensus review. Control cases were not included in this phase, as the analysis emphasized interpretive refinement rather than classification. This procedure involved two phases of inquiries:
  • GPT-4o Inquiry: An initial inquiry was made in GPT-4o, a large language model, to generate a baseline analysis report for each set of “Meta” images. This baseline assessment served as a reference point for comparison with the subsequent RAG-enhanced analysis.
  • Brain ARAMP Inquiries: Following the GPT-4o inquiry, a series of four inquiries was performed in the Brain ARAMP. These inquiries were conducted at different stages of the model’s training process:
    (1)
    After Training: The first inquiry assessed the model’s performance immediately after the initial training phase.
    (2)
    Post-Training 1, 2, and 3: The subsequent inquiries evaluated the model’s performance after three rounds of iterative post-training refinement, allowing for assessment of the model’s learning and improvement over time.
During each inquiry, a standardized prompt was applied to consistent analysis across all cases. The prompt explicitly instructed the model to execute the full diagnostic workflow comprehensively and step by step, spanning Step 1 (Image Definition) to Step 4 (Follow-Up Planning). To ensure transparency and facilitate evaluation, the prompt also required the model to embed all relevant programming code, intermediate results, and metadata validations in its output. Additionally, the model was instructed to provide clear reasoning for every decision, referencing retrieved dataset values, authoritative literature, or clinical guidelines [17,18,19,20,21].
It is important to clarify that ARAMP was evaluated as a reasoning-based support tool—not a detection model. Its role is to assist in post-detection interpretation of abnormal MR images, particularly in patients with known or suspected metastatic disease. This evaluation strategy, focused solely on metastasis-confirmed cases, allowed for precise assessment of interpretive enhancement while avoiding misclassification bias from unlabeled negatives. Expert evaluation of the model’s diagnostic performance was conducted using several metrics:
  • Pre-Trained Inference Difference: Assesses the discrepancy between initial model outputs (prior to structured prompting) and expert-confirmed diagnoses, illustrating baseline diagnostic variance.
  • Inter-Inference Agreement: Measures consistency among different post-training model outputs across three prompt iterations, indicating the stability and reproducibility of ARAMP’s diagnostic reasoning.
  • Post-Trained Inference Difference: Evaluates deviation between refined model outputs (after full RAG-enhanced prompting) and expert diagnoses, serving as a proxy for overall interpretive accuracy.
All metrics were scored on a 0–100% scale, where 0% indicates total discordance or incorrect reasoning, and 100% denotes complete agreement with expert assessment.

2.7. Statistical Analysis

Descriptive statistics (e.g., mean ± standard deviation, median [range], frequencies, and percentages) were used to summarize the demographic and clinical characteristics of the patient subgroups, as presented in Table 1. All statistical analyses were performed using IBM SPSS Statistics for Windows, version 26.0 (IBM Corp., Armonk, NY, USA). Given the focus on interpretive improvement rather than statistical significance testing, no inferential statistical tests (e.g., Wilcoxon, t-test) were applied. The pre- and post-training comparisons were based on aggregated descriptive measures and scoring trends across expert raters, as shown in Supplementary Materials Tables S2 and S3. Inter-rater agreement was assessed descriptively using intraclass correlation coefficients (ICC), as summarized in Table 2.

3. Result

3.1. Patient Characteristics

A total of 2447 patients (cancer group) with confirmed cancer diagnoses who underwent post-contrast axial T1-weighted brain MRI (AX_T1+C) examinations between 2010 and 2022 were identified from the institutional PACS system. The median age was 64 years (range, 21–95 years); 1323 of the patients (54.1%) were male.
From this cohort, a random subset of 100 patients was selected for detailed analysis. These patients had documented brain metastases (Mets) in their clinical MRI reports, further confirmed by expert consensus review (100% agreement).
In addition, 100 non-cancer individuals were selected as a control group, matched to the cancer group by age and gender distribution.
Although not included in model performance comparison, this control group underwent radiomic feature extraction to support construction of the RAG-based reference matrix for differential prompting.
Together, the 200 cases (100 cancer, 100 control) comprised the full dataset used for radiomic analysis and feature-informed GPT inference.
The demographic characteristics of the selected cancer subsample are presented in Table 1.
Radiologist-based evaluations demonstrated improved clarity and diagnostic consistency in ARAMP-generated reports, with detailed results presented in Supplementary Materials Table S5.

3.2. DICOM Feature Extraction and Image Conversion

A total of 1051 quantitative imaging features were extracted from the AX_T1+C DICOM images using the PyRadiomics library. These features included first-order statistical features (e.g., mean, skewness, kurtosis), shape-based features (e.g., volume, sphericity), as well as Laplacian of Gaussian (LoG) and wavelet-transformed features.
All extraction parameters, including voxel resampling, normalization, bin width, and enabled feature classes, are detailed in Supplementary Materials Table S2.
The extracted features were not used to train a conventional deep learning model, but rather served as structured input for GPT-4o in a retrieval-augmented diagnostic reasoning framework.
Following intensity normalization, axial DICOM images were converted into PNG format. Composite images were then constructed, with each composite comprising up to 10 sequentially arranged MRI slices, representing axial anatomical context per case. These composite PNGs were paired with their corresponding radiomic features and served as multimodal input for GPT-based zero-shot inference.

3.3. GPT Model Performance

This study evaluated the performance of the Brain ARAMP in identifying and reporting brain metastases among 100 cancer patients. Two key measures, derived from independent ratings by three board-certified radiologists and supplemented by Gemini benchmark outputs, are summarized in Table 2 and Table 3:
  • Post-Trained Inference Difference: The model’s post-trained inference difference was 67.45%, reflecting a substantial improvement in interpretive quality following implementation of the ARAMP prompting framework. In this context, “training” refers to a structured, multi-round prompting process—incorporating radiomic features, composite images, and literature-assisted reasoning via Retrieval-Augmented Generation (RAG)—rather than any form of parameter fine-tuning or supervised model updates.
  • Inter-Inference Agreement: The average inter-inference agreement between the three expert radiologists was 30.20%, reflecting modest exact-match consistency in their scoring of the model’s diagnostic outputs. To formally assess inter-rater reliability, the intraclass correlation coefficient (ICC) was computed, yielding a statistically significant result (ICC = 0.192, p = 0.01).
This agreement suggests that the model’s post-training reports were meaningfully distinct and consistently perceived as superior, supporting the reliability of the ARAMP-enhanced GPT workflow.
Table 3. Evaluator-level post-trained inference similarity and exact-match agreement.
Table 3. Evaluator-level post-trained inference similarity and exact-match agreement.
EvaluatorPost-Trained Inference Difference
(Mean ± SD)
Inter-Inference Agreement
(Mean ± SD)
Dr. A56.20 ± 25.5422.80 ± 8.54
Dr. B73.20 ± 17.5740.40 ± 13.33
Dr. C75.60 ± 18.7735.40 ± 11.67
Gemini68.00 ± 22.7422.00 ± 6.03
Average of Drs. A–C68.33 ± 20.6332.87 ± 11.18
Note: All values are presented as percentages (mean ± standard deviation, SD).
In addition, comparisons with the expert-reviewed gold standard revealed that the model achieved a sensitivity of 0.84 before training and 0.98 after training, reflecting a marked improvement in correctly identifying positive cases (see Table 4). As previously defined, “training” here refers to the structured prompting strategy incorporating radiomic features, composite MR images, and retrieval-augmented literature support, without any parameter fine-tuning.
Specificity and overall accuracy were not reported, due to the absence of negative cases in the dataset; in this context, accuracy is mathematically equivalent to sensitivity, and specificity could not be meaningfully calculated.
Furthermore, integration of the RAG framework significantly improved the model’s reliability in prioritizing brain metastases as the top-ranked differential diagnosis, as detailed in Table 5.

3.3.1. Mention Type Definitions

  • Listed as First: Brain metastasis was explicitly ranked as the first differential diagnosis.
  • Listed but Not First: Brain metastasis was included in the differential diagnoses, but not was not listed first.
  • Not Mentioned: Brain metastasis was not included in the differential diagnosis at all.

3.3.2. Post-Trained Inference Difference

This analysis aimed to assess whether the implementation of structured prompting strategies—as previously defined in the ARAMP workflow—led to measurable improvements in reporting quality. For each MRI examination, the GPT model generated one baseline report (pre-training) and three post-training reports, yielding four distinct outputs per MRI examination. These reports were independently evaluated by three board-certified neuroradiologists, each scoring them on predefined clinical criteria. ICC was calculated to assess the extent to which the ratings reflected meaningful differences among the reports. The results indicated low but statistically significant agreement between raters: ICC (single measures) = 0.286, 95% CI [0.161, 0.415], p = 0.000; ICC (average measures) = 0.546, 95% CI [0.366, 0.681], p = 0.000. While overall rater agreement was limited, the statistically significant ICC values suggest that the post-training reports were, to some extent, distinguishable from pre-training reports in terms of quality (Table 6).

3.3.3. Inter-Inference Agreement

To examine the stability of model outputs given identical input, three post-training reports generated from the same image were evaluated by the two radiologists. ICC values indicated statistically significant inter-rater agreement (Table 6): ICC (single measures) = 0.192, 95% CI [0.071, 0.324], p = 0.01; ICC (average measures) = 0.417, 95% CI [0.186, 0.590], p = 0.01. These results reflect statistically significant consistency between the raters in evaluating the similarity of the model outputs, indicating a fair-to-moderate level of stability in the model’s output based on physician ratings.
To qualitatively evaluate diagnostic improvement, structured annotations were obtained from three board-certified neuroradiologists—referred to as Reviewer A, Reviewer B, and Reviewer C––across 100 MRI cases (MRI1–MRI100). Each reviewer independently assessed the AI-generated diagnostic reports, focusing on correctness, granularity, and lesion-level consistency.
Reviewer A provided only three annotations. All were brief statements indicating a general preference for the pre-trained model (e.g., “pre-trained better”) and did not include specific lesion-level descriptions or model critique. His feedback served as high-level impressions/indications, rather than detailed evaluations.
Reviewer B offered the most comprehensive and detailed feedback. Across the dataset, 82 positive and 43 negative annotations were provided. Among them, 34 cases received both positive and negative remarks, 48 were annotated only with positive comments, and 9 were marked only with negative critiques. The remaining nine cases did not receive any annotations from Reviewer B. Positive feedback highlighted improvements in lesion localization, morphological detail, and accurate classification of lesion multiplicity. Examples included enhanced recognition of supratentorial/infratentorial spread, vasogenic edema, and ring-enhancing features. Negative comments focused on hallucinated findings, lateralization errors (e.g., “left diagnosed as right”), and missed lesions in key anatomical regions such as the cerebellum or skull.
Reviewer C contributed 21 annotations. These comments provided comparative judgments of pre- versus post-trained outputs and included lesion-level observations, such as “No lesion. Post-train is better,” “tiny lesion only identified on post-train,” or “Any certain lesion?”. In many cases, Reviewer C indicated whether a lesion was properly identified, missed, or ambiguously described. Several annotations offered insight into subtle or uncertain imaging findings, especially in borderline or very small lesions. His feedback provided contextual interpretation of model performance in diagnostically challenging scenarios.
This multi-reviewer annotation strategy not only ensured clinically grounded validation of AI-generated outputs, but also enabled a nuanced understanding of model strengths and limitations from diverse expert perspectives. The combined qualitative and quantitative evaluations supported a robust diagnostic performance assessment and informed iterative model refinement.

3.4. Illustrative Case

An example composite PNG was generated from 10 slices of AX_T1+C MR images from a patient with imaging diagnosis of brain metastases. A composite PNG with 10 slices of MR images is uploaded in every iteration of training due to capacity limitation. Depending on the slice thickness of the MR protocol, a patient’s images may need 6 to 18 iterations to complete the training (Figure 1).

4. Discussion

This study investigated the potential of a novel imaging framework, ARAMP (Adaptive RAG Assistant MRI Platform), to assist in detecting brain metastases in cancer patients undergoing brain MRI examinations. Our findings demonstrate the feasibility of extracting quantitative MRI data. By combining GPT-4o, structured radiomic features, and a retrieval-augmented generation (RAG) mechanism, ARAMP demonstrated promising diagnostic performance with both quantitative and qualitative validation.
The moderate inter-inference agreement observed suggests that the model’s diagnostic output is reasonably consistent and reproducible. Although a 30.20% exact-match agreement may initially appear low—suggesting that roughly 70% of rater responses were not perfectly aligned—it is important to interpret this figure in context. As clarified in the revised captions for Table 2 and Table 3, these values represent strict exact match rates (mean ± SD) across three independent raters, making this a deliberately conservative agreement measure. In tasks involving narrative diagnostic reasoning from AI systems, even subtle wording differences can lead to a mismatch, despite underlying clinical consensus. Similarly, the inter-round agreement scores reported in Table 3 (mean 32.87% ± 11.18%) reflect the average pairwise exact-match rate among repeated evaluations of the same post-trained outputs. These scores further underscore the challenge of achieving identical phrasing across multiple raters or evaluation rounds, especially when judging free-text outputs derived from LLMs. Importantly, this metric captures rater consistency in evaluating model outputs rather than model variability itself.
Prior studies of free-text or natural language outputs in clinical AI applications have reported similarly modest inter-rater agreement (kappa = 0.77–0.91 for human evaluators, 0.69–0.87 for human–machine agreement) due to inherent interpretive variability in clinical concept extraction and natural language processing tasks [22]. To supplement exact-match metrics, we also computed the intraclass correlation coefficient (ICC = 0.192, p = 0.01; see Table 6), which showed a statistically significant, though modest, level of consistency beyond chance. Together, these results suggest that, while exact-match agreement was limited, the AI-generated inferences were sufficiently reproducible to support their clinical utility.
Moreover, the substantial improvement in diagnostic accuracy following the training interventions (as shown by the increase from pre-trained to post-trained performance of 31.2 to 69.8%) underscores the effectiveness of ARAMP in enhancing the model’s learning and performance.
The integration of authoritative medical references through the RAG system allows the model to access and utilize a wealth of domain-specific knowledge, offering interpretive capabilities beyond those of rule-based or standalone models lacking external knowledge grounding. This dynamic integration of literature enables the model to adapt to new information and refine its diagnostic capabilities over time. Compare these findings in detail with specific results from studies like Pristoupil et al. [15] and Sau et al. [16] highlights how the ARAMP approach potentially addresses the limitations of LLMs in image-based tasks, as noted by Sau et al. Also compare with other AI/radiomics studies for brain metastasis detection [9,10], which discuss differences in methodology and performance.
We made some observations regarding model hallucination, and the adaptive RAG mechanism, combined with the structured workflow and reference grounding, appeared to mitigate this risk in our study.
While our results are encouraging, this study has limitations. First, the retrospective design introduces potential biases related to patient selection and data quality. Second, the sample size in the subgroup analysis (n = 100) may limit the generalizability of our findings regarding detailed performance metrics. The study was conducted at a single institution; the gold standard relied on internal MRI and radiology reports, which might have inherent limitations. Third, the model was tested exclusively on AX_T1+C sequences. This design choice reflects clinical consensus that contrast-enhanced T1-weighted imaging offers optimal sensitivity for brain metastasis detection, as supported by prior studies such as Grøvik et al. (2020), Pennig et al. (2022), and Zhang et al. (2022) [23,24,25]. While additional sequences like FLAIR or DWI may provide supplementary specificity in complex cases, AX_T1+C remains the most validated modality for lesion conspicuity and radiomic analysis. Recent studies, such as Machura et al. (2024) [26], have demonstrated that incorporating longitudinal multi-sequence MRI may improve detection consistency in deep learning ensembles. Although our current study did not evaluate FLAIR or DWI, we acknowledge their potential value, and the integration of multimodal imaging data will be a focus of future ARAMP iterations.
Because the evaluation cohort consisted solely of metastasis-positive cases confirmed by expert review, specificity and overall accuracy metrics were not calculated. Including unlabeled or unverified control scans could introduce verification bias or misclassification, thereby compromising analytic validity. Sensitivity, which measures true positive identification within known abnormal cases, was selected as the most appropriate metric to assess interpretive refinement. This approach aligns with ARAMP’s post-detection role, focusing on diagnostic reasoning rather than binary classification. Due to the absence of negative (failing) cases in the evaluation dataset, overall accuracy was not a meaningful metric and was not reported. Sensitivity was used instead to assess the model’s ability to correctly identify passing cases. Prospective studies are also needed, in the future, to evaluate the real-world clinical impact.
Despite these limitations, this study highlights the potential of LLM-based systems with adaptive RAG to assist radiologists in the interpretation of brain MRI images and the detection of brain metastases. Further research is warranted to explore the clinical utility of this technology in improving diagnostic accuracy and, subsequently, patient outcomes.
This study presents development and evaluation of ARAMP, a system integrating a large language model (GPT-4o) with a retrieval-augmented generation (RAG) framework and multiple domain-specific medical references. The system supports information retrieval and AI-assisted content generation. We assessed ARAMP’s efficacy in generating radiology reports for MRI studies, comparing its outputs before and after training by GPT-4o, using both human expert review and independent LLM-based inference.
Radiological workloads continue to increase amid a global shortage of trained radiologists. Recent advancements in LLMs and RAG architectures offer potential solutions to alleviate interpretive burden. This study investigates the clinical relevance and performance of ARAMP, an LLM-RAG integrated platform designed to assist radiologists in MRI interpretation. ARAMP leverages GPT-4o in conjunction with curated medical references [17,18,19,20,21] to perform AI-assisted MRI interpretation. Three board-certified radiologists with varying years of clinical experience independently evaluated the generated reports. An independent LLM (Gemini), which was not involved in the training process, was also used to assess inference performance. Key metrics included model interpretative score (pre- vs. post-training), inter-rater consistency, and parity between human and machine inference. Post-training, the performance of ARAMP improved significantly, with interpretative scores increasing from 59.90 to 67.45. The average inter-inference consistency among the three radiologists was 30.20 out of 100 (p = 0.01, Table 2), indicating a moderately high agreement level. Differences between pre- and post-training inference results were comparable between radiologists (average: 68.00/100.00) and Gemini (68.33/100.00), suggesting reliable model generalization and parity in independent assessment. These results provide empirical evidence that LLM-based models like ARAMP can achieve near-human performance in evaluating AI-generated content (Table 3). Moreover, Gemini’s independent assessment validates the generalizability of the post-trained ARAMP model across different inference platforms.
ARAMP demonstrates promising performance in AI-assisted MRI interpretation, supported by expert evaluation and independent model validation. Its integration into radiological practice could help address the increasing demand for imaging interpretation and improve clinical efficiency. The implementation of ARAMP in clinical settings may offer significant benefits by streamlining MRI interpretation workflows. As a previewing mechanism, it has the potential to reduce the diagnostic workload of radiologists—particularly important in an era marked by radiologist shortages worldwide.

5. Conclusions

Our proposed LLM model, Brain ARAMP, demonstrates potential to assist in the detection of brain metastases in cancer patients undergoing MRI examinations. The model demonstrated promising diagnostic accuracy and statistically significant inter-inference agreement (indicating moderate-to-high consistency) among expert radiologists. Future research should focus on validating these findings in larger, more diverse patient populations with brain tumors. The research scope can also prospectively extend to evaluating the model’s impact on clinical decision-making and patient outcomes.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/bioengineering12070698/s1, Supplementary File S1. PyRadiomics YAML Configuration File. Supplementary Table S1. Complete Radiomic Feature Summary. Supplementary Table S2. PyRadiomics Feature Extraction Settings. Supplementary Table S3. Comparison of GPT-4o Diagnostic Outputs Before and After ARAMP Integration. Supplementary Table S4. ARAMP vs Traditional AI. Supplementary Table S5. Radiologist Scoring. Supplementary Table S6. CLAIM2024 Completed.

Author Contributions

These authors’ individual contributions were as follows. K.-C.W. and W.-Y.G. were responsible for the study design. K.-C.W. and P.-C.Y. collected the data. All authors performed statistical analyses, interpreted data, and drafted the article. W.-Y.G., F.Y.C., K.-L.C., and W.-C.S. provided some intellectual content. C.-H.K. contributed to the critical review of the manuscript. S.-S.C., as the corresponding author, was responsible for the overall supervision of the project and the final revision of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China Medical University Hospital (DMR-114-001, DMR-114-006). The funders had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Institutional Review Board Statement

This retrospective study was approved by the Institutional Review Board of China Medical University Hospital (CMUH) on 15 December 2022 (IRB No. CMUH111-REC1-206). The requirement for informed consent was waived by the Institutional Review Board of CMUH due to the retrospective nature of the study and the use of de-identified patient data.

Informed Consent Statement

Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The datasets generated and/or analyzed during the current study are not publicly available, due to patient privacy regulations, but are available from the corresponding author on reasonable request and with institutional permission.

Acknowledgments

The authors would like to thank the Department of Medical Imaging at China Medical University Hospital for providing access to the imaging data. We also thank Wan-Yuo Guo for insightful comments during manuscript preparation.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, I.; Yamamoto, M.; Knisely, J.P. CH11. Multiple brain metastases. In Adult CNS Radiation Oncology: Principles and Practice; Chang, E.L., Brown, P.D., Lo, S.S., Sahgal, A., Suh, J.H., Eds.; Springer: Berlin/Heidelberg, Germany, 2024; pp. 475–495. [Google Scholar]
  2. Robin, A.M.; Kalkanis, S.N. CH10. 87C10 Brain Metastases. In Surgical Neuro-Oncology; Lonser, R., Elder, B., Eds.; Oxford Academic: Oxford, UK, 2018; pp. 87–96. [Google Scholar]
  3. Chilukuri, S.; Jalali, R. Brain metastasis: Momentum towards understanding the molecular milieu. Neurol. India 2019, 67, 755–756. [Google Scholar] [PubMed]
  4. Mirawati, D.K.; Soewondo, W. MRI sequencing of brain metastasis from lung malignancy: Literature review. GSC Adv. Res. Rev. 2023, 15, 105–109. [Google Scholar] [CrossRef]
  5. Kosiorowska, P.; Pasieka, K.; Perenc, H.; Majka, K. Overview of medical analysis capabilities in radiology of current Artificial Intelligence models. Qual. Sport 2024, 20, 53933. [Google Scholar] [CrossRef]
  6. Aerts, H.J.; Velazquez, E.R.; Leijenaar, R.T.; Parmar, C.; Grossmann, P.; Carvalho, S.; Bussink, J.; Monshouwer, R.; Haibe-Kains, B.; Rietveld, D.; et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 2014, 5, 4006. [Google Scholar] [CrossRef]
  7. Yang, H.C.; Wu, C.C.; Lee, C.C.; Huang, H.E.; Lee, W.K.; Chung, W.Y.; Wu, H.-M.; Guo, W.-Y.; Wu, Y.-T.; Lu, C.-F. Prediction of pseudoprogression and long-term outcome of vestibular schwannoma after Gamma Knife radiosurgery based on preradiosurgical MR radiomics. Radiother. Oncol. 2021, 155, 123–130. [Google Scholar] [CrossRef]
  8. Hosny, A.; Parmar, C.; Quackenbush, J.; Schwartz, L.H.; Aerts, H.J.W.L. Artificial intelligence in radiology. Nat. Rev. Cancer 2018, 18, 500–510. [Google Scholar] [CrossRef] [PubMed]
  9. Tong, E.; McCullagh, K.L.; Iv, M. Advanced Imaging of Brain Metastases: From Augmenting Visualization and Improving Diagnosis to Evaluating Treatment Response. Front. Neurol. 2020, 11, 270. [Google Scholar] [CrossRef] [PubMed]
  10. Ocaña-Tienda, B.; Pérez-Beteta, J.; Villanueva-García, J.D.; Romero-Rosales, J.A.; Molina-García, D.; Suter, Y.; Asenjo, B.; Albillo, D.; de Mendivil, A.O.; Pérez-Romasanta, L.A.; et al. A comprehensive dataset of annotated brain metastasis MR images with clinical and radiomic data. Sci. Data 2023, 10, 208. [Google Scholar] [CrossRef] [PubMed]
  11. Najjar, R. Redefining Radiology: A Review of Artificial Intelligence Integration in Medical Imaging. Diagnostics 2023, 13, 2760. [Google Scholar] [CrossRef] [PubMed]
  12. Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.-T.; Rocktäschel, T.; Riedel, S.; Kiela, D. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. NeurIPS 2020, 33, 9459–9474. [Google Scholar]
  13. Yang, R.; Ning, Y.; Keppo, E.; Liu, M.; Hong, C.; Bitterman, D.S.; Ong, J.C.L.; Ting, D.S.W.; Liu, N. Retrieval-Augmented Generation for Generative Artificial Intelligence in Medicine. arXiv 2024, arXiv:2406.12449. [Google Scholar]
  14. Zakka, C.; Shad, R.; Chaurasia, A.; Dalal, A.R.; Kim, J.L.; Moor, M.; Fong, R.; Phillips, C.; Alexander, K.; Ashley, E.; et al. Almanac—Retrieval-Augmented Language Models for Clinical Medicine. NEJM AI 2024, 1, AIoa2300068. [Google Scholar] [CrossRef] [PubMed]
  15. Pristoupil, J.; Oleaga, L.; Junquero, V.; Merino, C.; Sureyya, O.S.; Kyncl, M.; Burgetova, A.; Lambert, L. Generative pre-trained transformer 4o (GPT-4o) in solving text-based multiple response questions for European Diploma in Radiology (EDiR): A comparative study with radiologists. Insights Imaging 2025, 16, 66. [Google Scholar] [CrossRef] [PubMed]
  16. Sau, S.; George, D.D.; Singh, R.; Kohli, G.S.; Li, A.; Jalal, M.I.; Singh, A.; Furst, T.J.; Rahmani, R.; Vates, G.E.; et al. Accuracy and quality of ChatGPT-4o and Google Gemini performance on image-based neurosurgery board questions. Neurosurg. Rev. 2025, 48, 320. [Google Scholar] [CrossRef] [PubMed]
  17. Bostwick, P. Medical Terminology: Learning Through Practice, 2nd ed.; McGraw Hill LLC: Columbus, OH, USA, 2014. [Google Scholar]
  18. Nader, R.; Sabbagh, A.J.; Elbabaa, S.K. Neurosurgery Case Review: Questions and Answers, 2nd ed.; Georg Thieme Verlag: Stuttgart, Germany, 2019. [Google Scholar]
  19. Citow, J.S.; Spinner, R.J.; Puffer, R.C. Neurosurgery Oral Board Review, 3rd ed.; Georg Thieme Verlag: Stuttgart, Germany, 2019. [Google Scholar]
  20. Osborn, A.G.; Hedlund, G.L.; Salzman, K.L. Osborn’s Brain, 2nd ed.; Elsevier: Amsterdam, The Netherlands, 2017. [Google Scholar]
  21. Barkhof, F.; Jager, R.; Thurnher, M.; Cañellas, A.R. Clinical Neuroradiology: The ESNR Textbook, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  22. Oommen, C.; Howlett-Prieto, Q.; Carrithers, M.D.; Hier, D.B. Inter-rater agreement for the annotation of neurologic signs and symptoms in electronic health records. Front. Digit. Health 2023, 5, 1075771. [Google Scholar] [CrossRef] [PubMed]
  23. Grovik, E.; Yi, D.; Iv, M.; Tong, E.; Rubin, D.; Zaharchuk, G. Deep learning enables automatic detection and segmentation of brain metastases on multisequence MRI. Radiology 2020, 297, 670–678. [Google Scholar] [CrossRef] [PubMed]
  24. Pennig, L.; Shahzad, R.; Caldeira, L.; Lennartz, S.; Thiele, F.; Goertz, L.; Zopfs, D.; Maintz, D.; Perkuhn, M.; Borggrefe, J. Automated detection of brain metastases on T1-weighted contrast-enhanced MRI using an ensemble of convolutional neural networks: Impact of volume-aware loss and sampling strategy. J. Magn. Reson. Imaging 2022, 56, 535–545. [Google Scholar] [CrossRef]
  25. Zhang, M.; Young, G.S.; Chen, H.; Li, J.; Qin, L.; McFaline-Figueroa, J.R.; Reardon, D.A.; Cao, X.; Wu, X.; Xu, X. Deep-learning detection of cancer metastases to the brain on MRI. Neuro-Oncology 2022, 24, 1594–1602. [Google Scholar] [CrossRef] [PubMed]
  26. Machura, B.; Kucharski, D.; Bozek, O.; Eksner, B.; Kokoszka, B.; Pekala, T.; Radom, M.; Strzelczak, M.; Zarudzki, L.; Gutiérrez-Becker, B.; et al. Deep learning ensembles for detecting brain metastases in longitudinal multi-modal MRI studies. Comput. Med. Imaging Graph. 2024, 116, 102401. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustrative case of brain metastases evaluated by the ARAMP system. (A) Composite PNG image showing 10 sequential post-contrast T1-weighted axial (AX_T1+C) MR slices. (B) Three representative AX_T1+C slices demonstrating multiple metastatic lesions (examples indicated by red circles in panel (C). Note the lesion in the right cerebellum showing significant surrounding vasogenic edema (red circle). (C) Magnified view of the slices in (B), with metastatic lesions clearly marked by red circles. These lesions typically appear as enhancing nodules on post-contrast T1-weighted images.
Figure 1. Illustrative case of brain metastases evaluated by the ARAMP system. (A) Composite PNG image showing 10 sequential post-contrast T1-weighted axial (AX_T1+C) MR slices. (B) Three representative AX_T1+C slices demonstrating multiple metastatic lesions (examples indicated by red circles in panel (C). Note the lesion in the right cerebellum showing significant surrounding vasogenic edema (red circle). (C) Magnified view of the slices in (B), with metastatic lesions clearly marked by red circles. These lesions typically appear as enhancing nodules on post-contrast T1-weighted images.
Bioengineering 12 00698 g001
Table 1. Demographic and clinical characteristics of the 100-patient subsample used for metastasis evaluation.
Table 1. Demographic and clinical characteristics of the 100-patient subsample used for metastasis evaluation.
CharacteristicsValue
Gender (male/female)52/48
Age (male/female)
   20–301/1
   30–402/2
   40–505/6
   50–6012/12
   60–7017/14
   70–8012/9
   80+3/4
Clinical diagnosis
   Biliary or pancreatic cancer1
   Breast cancer11
   Cervical cancer1
   Colorectal cancer3
   Esophageal carcinoma3
   Gastric cancer1
   Head and neck cancer (excluding NPC_brain_spine_thyroid cancer)2
   Liver cancer1
   Lung cancer65
   Melanoma1
   Non-Hodgkin’s disease3
   Other cancers3
   Prostate cancer2
   Secondary malignant neoplasm1
   Upper lower respiratory tract cancer (excluding lung cancer)1
   Urologic cancer (excluding renal_bladder_prostate cancer)1
Note. Age distribution is shown as male/female counts.
Table 2. Summary of pre- and post-trained inference performance and inter-rater agreement.
Table 2. Summary of pre- and post-trained inference performance and inter-rater agreement.
MeasurePre-TrainedPost-TrainedInter-Inference Agreement
Mean ± SD59.90 ± 17.9967.45 ± 21.6030.20 ± 12.98
Sum of Converted Scores11,98026,98012,080
Range (Min–Max)20–10020–10020–100
Note: Scores were converted from a 5-point scale (3 raters × 100 cases = 300 ratings per group) into percentage categories (20–100%). Values represent the average across all evaluators and cases.
Table 4. Comparison of instance level model performance before and after ARAMP training.
Table 4. Comparison of instance level model performance before and after ARAMP training.
MetricBeforePost
True Positive (TP)8498
False Negative (FN)162
True Negative (TN)00
False Positive (FP)00
Sensitivity0.840.98
SpecificityNANA
Precision84%98%
Note: Sensitivity improved after training. Specificity could not be calculated, due to the absence of negative cases in the dataset.
Table 5. Distribution of brain metastasis mentions across model outputs.
Table 5. Distribution of brain metastasis mentions across model outputs.
Mention TypeBeforePost 1Post 2Post 3
Listed as First78959595
Listed but Not First6333
Not Mentioned16222
Note: Before refers to the initial model output without retrieval augmentation. Post 1–Post 3 are reports generated after applying RAG (retrieval-augmented generation) prompting.
Table 6. Intraclass correlation coefficient (ICC) for physician ratings.
Table 6. Intraclass correlation coefficient (ICC) for physician ratings.
Analysis ContextICC TypeIntraclass Correlation b95% CI
(Lower–Upper)
F (df1, df2)p-Value
Post-trained inference differencesSingle measures0.286 a0.161~0.4152.200 (99, 198)0.00
Average measures0.546 c0.366~0.6812.200 (99, 198)0.00
Inter-inference agreement of post-trained modelSingle measures0.192 a0.071~0.3241.714 (99, 198)0.01
Average measures0.417 c0.186~0.5901.714 (99, 198)0.01
Note: Two-way mixed effects model where people effects are random and measures effects are fixed. a The estimator is the same, whether the interaction effect is present or not. b Type C intraclass correlation coefficients using a consistency definition. Between-measure variance is excluded from the denominator variance. c This estimate is computed assuming the interaction effect is absent, because it is not estimable otherwise.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, K.-C.; Chew, F.Y.; Cheng, K.-L.; Shen, W.-C.; Yeh, P.-C.; Kao, C.-H.; Guo, W.-Y.; Chang, S.-S. Adaptive RAG-Assisted MRI Platform (ARAMP) for Brain Metastasis Detection and Reporting: A Retrospective Evaluation Using Post-Contrast T1-Weighted Imaging. Bioengineering 2025, 12, 698. https://doi.org/10.3390/bioengineering12070698

AMA Style

Wu K-C, Chew FY, Cheng K-L, Shen W-C, Yeh P-C, Kao C-H, Guo W-Y, Chang S-S. Adaptive RAG-Assisted MRI Platform (ARAMP) for Brain Metastasis Detection and Reporting: A Retrospective Evaluation Using Post-Contrast T1-Weighted Imaging. Bioengineering. 2025; 12(7):698. https://doi.org/10.3390/bioengineering12070698

Chicago/Turabian Style

Wu, Kuo-Chen, Fatt Yang Chew, Kang-Lun Cheng, Wu-Chung Shen, Pei-Chun Yeh, Chia-Hung Kao, Wan-Yuo Guo, and Shih-Sheng Chang. 2025. "Adaptive RAG-Assisted MRI Platform (ARAMP) for Brain Metastasis Detection and Reporting: A Retrospective Evaluation Using Post-Contrast T1-Weighted Imaging" Bioengineering 12, no. 7: 698. https://doi.org/10.3390/bioengineering12070698

APA Style

Wu, K.-C., Chew, F. Y., Cheng, K.-L., Shen, W.-C., Yeh, P.-C., Kao, C.-H., Guo, W.-Y., & Chang, S.-S. (2025). Adaptive RAG-Assisted MRI Platform (ARAMP) for Brain Metastasis Detection and Reporting: A Retrospective Evaluation Using Post-Contrast T1-Weighted Imaging. Bioengineering, 12(7), 698. https://doi.org/10.3390/bioengineering12070698

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop