Next Article in Journal
Comparison of Cottle-Area-2 and Cottle-Area-3 in Computed Tomography Scans of Patients with Nasal Obstruction and Controls
Previous Article in Journal
Platelet Satellitism in a Patient with Underlying Infection, Immune Thrombocytopenic Purpura (ITP) and Multiple Sclerosis
Previous Article in Special Issue
Lobish: Symbolic Language for Interpreting Electroencephalogram Signals in Language Detection Using Channel-Based Transformation and Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Do LLMs Have ‘the Eye’ for MRI? Evaluating GPT-4o, Grok, and Gemini on Brain MRI Performance: First Evaluation of Grok in Medical Imaging and a Comparative Analysis

1
Department of Neurosurgery, Sincan Training and Research Hospital, Ankara 06949, Turkey
2
Department of Neurosurgery, Kulu State Hospital, Konya 42780, Turkey
3
Ankara Medipol University Faculty of Medicine, Ankara 06050, Turkey
4
Department of Neurosurgery, Adiyaman Training and Research Hospital, Adiyaman 02100, Turkey
5
Department of Neurosurgery, Gazi University Faculty of Medicine, Ankara 06560, Turkey
6
Department of Radiology, Stanford University School of Medicine, Stanford, CA 94305, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(11), 1320; https://doi.org/10.3390/diagnostics15111320
Submission received: 2 May 2025 / Revised: 17 May 2025 / Accepted: 21 May 2025 / Published: 24 May 2025
(This article belongs to the Special Issue Artificial Intelligence in Neuroimaging 2024)

Abstract

Background/Objectives: Large language models (LLMs) are revolutionizing the world and the field of medicine while constantly improving themselves. With recent advancements in image interpretation, evaluating the reasoning capabilities of these models and benchmarking their performance on brain MRI tasks has become crucial, as they may be utilized—albeit off-label—for patient care by both neurosurgeons and non-neurosurgeons. Methods: ChatGPT-4o, Grok, and Gemini were presented with 35,711 slices of brain MRI, including various pathologies and normal MRIs. Models were asked to identify the MRI sequence and determine the presence of pathology. Their individual performances were measured and compared with one another. Results: GPT refused to answer 28.02% of the slices despite three attempts, whereas Grok and Gemini provided responses on the first attempt for every slice. Gemini achieved 74.54% pathology prediction and 46.38% sequence prediction accuracy. GPT-4o achieved 74.33% pathology prediction and 85.98% sequence prediction accuracy for questions that it had answered (53.50% and 61.67% in total, respectively). Grok achieved 65.64% pathology prediction and 66.23% sequence prediction accuracy. Conclusions: The image interpretation capabilities of the investigated LLMs are limited for now and require further refinement before competing with specifically trained and fine-tuned dedicated applications. Amongst them, Gemini outperforms the others in pathology prediction while Grok outperforms others in sequence prediction. These limitations should be kept in mind if use during patient care is planned.

1. Introduction

Artificial intelligence (AI) has made a large impact in every aspect of life in the past few years, including medicine and neurosurgical image interpretation. Although Cellular Neural Network systems were used in the early 2000s [1], this method has largely been abandoned in favor of Convolutional Neural Network (CNN) systems, which have been used for brain segmentation since 2015 [2]. In the field of neurosurgery, these radiological assessment capabilities of the AI are considered especially beneficial and showed promise for early warning systems in emergency settings [3]. Recently, AI applications in every field have gained great popularity when AI became a part of every household with large language models (LLMs).
Although LLMs usually use a very different and more recent method for image interpretation called Vision Transformer (ViT) [4], this new method is comparable to older and tested CNNs even now. With the integration of ViT into LLMs, image interpretation has rapidly become a central feature of major models in recent months.
ChatGPT-4o [5], xAI’s Grok [6], and Google’s Gemini (previously Bard) [7] all recently integrated image processing and interpretation capabilities. While their documentation states they are not validated to use for medical purposes, this does not preclude testing these capabilities in a research setting.
Especially in emergency departments and primary care settings, experienced neuroradiologists or neurosurgeons are not always readily available to interpret neuroradiological images. Moreover, even in urgent situations, image reporting can sometimes take hours and weeks in elective settings. This has driven the adoption of AI support in neurosurgical patient care across various settings, with numerous efforts dedicated to developing AI-based early warning systems. Widely accessible large language models (LLMs), with their intuitive and user-friendly interfaces, hold significant potential in this domain. Therefore, this study was conducted to assess the ability of these models to interpret brain magnetic resonance imaging (MRI) using a large-scale database, previously constructed from real-world data, for AI research.

2. Materials and Methods

2.1. Dataset Used

An expanded private version of previously published Gazi Brains 2020 Dataset [8] that contained segmented MRI data of real patients was used. An AI algorithm was developed using this dataset and was recently patented during the review process of this study [9]. This version contained fluid-attenuated inversion recovery (FLAIR), T2, T1 and T1c (contrast enhanced—when available) MRI scans of a total of 500 patients, including those who did not have any pathology (Normal, n = 93) and who had meningioma (n = 149), high-grade glioma (HGG, n = 60), cavernoma (n = 59), schwannomas (n = 38), post-operative patients who had various pathologies (n = 33), low-grade glioma (LGG, n = 30), metastatic disease (n = 19), and, classified as ‘Others’, a couple samples from various other pathologies (including: epidermoid tumors, hemorrhage, dermoid tumors, germinomas, pituitary tumors, arteriovenous malformations, hemangioblastomas, stroke patients, osteomas, plasmocytomas, and lymphomas, n = 19). A total of 1793 independent sequences and 35,711 slices were investigated. Slices that contained contrast-enhancing or non-enhancing parts of a tumor or lesion, peritumor edema, acute ischemia, encephalomalasic areas, cyst associated with a tumor, blood, or post-operative cavities were considered pathological and classified as ground truth positive. No ethical approval was necessary for this research as it was conducted using a pre-existing dataset according to local and regional guidelines.

2.2. Questioning LLMs

An OpenApi Software Development Kit (SDK) v.1.55.3 [10] was used to direct prompts to each LLM’s API. Image conversion to JPEG was performed using the OpenCV library with a quality setting of 95 in their original voxel resolution and converted to base64 format using the standard Python library (Python v.3.10.0) [11]. base64 formatted images were presented to LLMs with a standard prompt instructing the LLM as to what to do. The part of the code used that includes the prompt is presented in Appendix A. Other inference parameters were not explicitly set and were thus left to each model’s default behavior. All questioning was conducted between 20 November 2024 and 10 December 2024. None of the investigated services received any major update in this period to the best of the authors’ knowledge. The builds used were as follows: GPT-4o (gpt-4o-2024-08-06, released 6 August 2024), Gemini (gemini-1.5-pro-002, released 24 September 2024), and Grok (grok-vision-beta-0.1.0, released 23 November 2024). If an answer in the correct format was not provided, the algorithm was programmed to make 2 additional attempts before recording a “Not answered” status. Each slice was presented in an individual session with no memory function enabled to recall previous slices (i.e., every request was sent as a separate session), and the LLMs’ prediction regarding the presence of a pathology in the presented image and the prediction of the MRI sequence was recorded. Each LLM was tested in its native state without providing any additional training or fine-tuning.

2.3. Statistical Analysis

For the evaluation of the data obtained in this study and the statistical analysis, the “Statistical Package for Social Sciences” (SPSS) package program version 27.0 (Chicago, IL, USA) was used. Descriptive statistics were presented as numbers and percentages. When evaluating sequence predictions, even though both options were presented in the prompt, ‘T1’ and ‘T1c’ were considered correct reciprocally, since distinguishing them simply by assessing a single slice is not always possible. Results were aggregated at the ‘sequence-level’ to reflect clinical use. If any slice was predicted as pathological, the whole sequence was considered to be classified as pathological by that LLM, since there are pathologies actually visible in a couple of slices only. The differences between groups for categorical variables were evaluated using the Pearson Chi-square; results were reported and interpreted with adjusted standardized residuals (cut off: absolute 2.5). Fisher’s exact test was used for 2 × 2 tables. The Baptista–Pike method was used for odds ratio (OR) confidence intervals (95%CI); if the Baptista–Pike method could not be used, the Gart adjusted logit interval was used instead, and those results are marked with a ‘G’. Cochran’s Q test was used for comparing all three LLMs’ correct answer statuses together, and McNemar’s test with Bonferroni correction was used for individual (pairwise) comparisons between LLMs. Asymptotic significances are reported with χ2 statistics when the sample size adequacy assumption of this test was met, and exact p values are reported otherwise. Sensitivity, specificity, Positive Predictive Value (PPV), Negative Predictive Value (NPV), F Scores (F0.5: (precision weighted) − F1: (DICE) − F2: (recall weighted)) and accuracy (percentage correct) were calculated for each LLM [12]. Diagnostic values were evaluated using Receiver Operating Characteristic (ROC) analysis, and the area under the curves (AUCs) were calculated with confidence intervals. The results were evaluated with a confidence interval of 95.0% (95%CI) and a significance level of p ≤ 0.05; for post hoc analyses that require multiple comparisons, Bonferroni corrected p values are reported and evaluated.

3. Results

3.1. Dataset Descriptives and Answer Rates

Details of the analyzed dataset are presented in Table 1. Grok and Gemini provided a correctly formatted answer on the first attempt for every slice (number of requests sent = number of slices). However, GPT sometimes refused to provide a straightforward answer or sometimes completely refused to answer. GPT occasionally provided a valid response to a slice that it had previously refused to answer. Ultimately GPT refused to answer 10,006 slices (28.02%). A total of 71,542 requests were sent to GPT for the purposes of this study (2.003 requests/slice). Excluding 30,018 requests used for slices that were never answered by GPT, 41,524 requests were used to make GPT evaluate 22,581 slices (1.893 requests/slice), meaning GPT refused to answer 18,943 times for a slice that it eventually answered upon repeated requests. Reasons to refuse varied from ethical claims to confessing inability to decide (differing liberally between attempts, even for repeat attempts of the same slice; therefore, a systematic analysis of this refusal was not possible due to the probabilistic nature of this phenomenon). GPT was more likely to answer when the ground truth was positive [Fisher’s exact p < 0.001, OR = 1.19 (1.12–1.27)G]. GPT was considered to have answered incorrectly for the questions that it completely refused to answer when directly comparing with other LLMs. But individual metrics were calculated and presented both excluding those slices and including those slices as incorrect when possible.

3.2. Performance of GPT-4o

The performance of GPT was evaluated both by excluding the questions that it refused to answer and by considering them incorrect. A total of 33 (0.44%) T1c answers were considered correct for T1 sequence, and 3111 (41.42%) T1 answers were considered correct for T1c. Correct prediction of the sequence was associated with correct prediction of the pathology [Fisher’s exact p = 0.025], but the association was strangely inverse [OR: 0.91 (0.84–0.98)]. A confusion matrix that was based on the questions that it answered is presented in Table 2. Associations between correct pathology and sequence prediction were investigated separately for each sequence. Only the FLAIR sequence (lowest sequence prediction accuracy) showed an unexplained inverse association [Fisher’s exact p < 0.001, OR = 0.49 (0.43–0.55)]; T2 and T1 showed the expected positive correlation [Fisher’s exact p < 0.001 (both), OR = 1.48 (1.24–1.77), and OR = 1.51 (1.24–1.84), respectively], and the T1c sequence showed independence [Fisher’s exact p = 0.229]. Accuracy and sequence prediction accuracy were calculated as 53.50% (19,107/35,711) and 61.67% (22,023/35,711), respectively, when the slices for which it refused to answer were considered incorrect. GPT significantly performed differently for sequences (Adjusted residual) from best to worst as follows: T1 (7.7), FLAIR (6.6), T1c (−7.3), and T2 (−7.6) [χ2(3) = 159.746, p < 0.001].

3.3. Performance of Grok

The confusion matrix is presented in Table 3. Grok has clearly struggled to identify the FLAIR sequence (accuracy lower than random chance—25%), but its pathology prediction accuracy for this sequence was comparable to other sequences. For T1 sequence images, 135 (1.44%) T1c answers and 6022 (80.18%) T1 answers for T1c images were considered correct. In total, correct prediction of pathology seemed independent from sequence prediction [Fisher’s exact p = 0.064]; however, this was the result of the inverse association observed in FLAIR [Fisher’s exact p < 0.001, OR = 0.39 (0.35–0.43)G]; for T2 [Fisher’s exact p < 0.001, OR = 1.66 (1.50–1.83)], T1 [Fisher’s exact p < 0.001, OR = 1.73 (1.52–1.96)], and T1c [Fisher’s exact p < 0.001, OR = 1.39 (1.21–1.58)], correct prediction of the results was associated with correct identification of the sequence. Grok significantly performed differently for sequences (Adjusted residual) from best to worst as follows: T1 (5.7), T2 (1.5), FLAIR (−0.2), and T1c (−7.5) [χ2(3) = 70.356, p < 0.001].

3.4. Performance of Gemini

The confusion matrix is presented in Table 4. Gemini performed even worse than a die roll (25%) for predicting the FLAIR and T2 sequences but performed way better than random for T1 and T1c (their cut-off would be 50% because they are considered reciprocally correct). For T1 sequence images, 2420 (25.90%) T1c answers and 2820 (37.55%) T1 answers for T1c images were considered correct. Although in total, correct prediction of the pathology seemed independent from correct prediction of the sequence [Fisher’s exact p = 0.394], the FLAIR sequence showed an inverse association [Fisher’s exact p < 0.001, OR = 0.29 (0.26–0.32)], T2 showed independence [Fisher’s exact p = 0.972], and T1 and T1c showed positive correlation [Fisher’s exact p < 0.001 (both), OR = 3.51 (3.15–3.92), and OR = 2.63 (2.36–2.92), respectively]. Grok significantly performed differently for sequences (Adjusted residual) from best to worst as follows: T2 (7.3), FLAIR (5.6), T1 (−0.3), and T1c (−13.6) [χ2(3) = 208.026, p < 0.001].

3.5. Comparing LLMs and ROC Analyses

LLMs’ overall accuracy showed significant difference as the Cochran’s Q test shows [χ2(2) = 253,752.657, p < 0.001]. Post hoc analysis using McNemar’s test (with Bonferroni correction) showed that Grok is superior to GPT [χ2(1) = 1410.252, p < 0.001], and Gemini is superior to Grok [χ2(1) = 1085.637, p < 0.001] and GPT [χ2(1) = 4735.196, p < 0.001]. These outcomes were the same in every sequence individually (χ2 statistics omitted for readability).
Sequence prediction accuracy was significantly different between LLMs. Post hoc analysis showed that GPT is superior to Gemini even though slices for which GPT did not provide an answer were considered wrong; Grok is superior to Gemini and GPT. When sequences were analyzed individually, some deviations from this order was observed; therefore, the individual results and χ2 statistics are presented in Table 5.
GPT not answering to some slices created a challenge for ROC analysis. The first analysis was conducted excluding those slices for all LLMs (although, undoubtedly, these results would be biased in favor of GPT). AUCs (95%CI) were, respectively, 60.5% (59.8–61.2%), 57.9% (57.2–58.6%), and 57.8% (57.1–58.4%) for Gemini, Grok, and GPT, respectively. The AUC difference from Gemini was significant for Grok [z = 6.956, p < 0.001] and GPT [z = 7.948, p < 0.001]. The difference between Grok and GPT was not significant [z = 0.334, p = 0.738].
The second solution involved including those slices in some way. To do that, three different situations (answering positive for all of them [GPT (+)]; answering negative for all of them [GPT (−)], assigning them an intermediate unique value [GPT 0.5]) for GPT was evaluated. ROC Curves and Precision–Recall curves are presented in Figure 1. Individual AUCs and comparisons are presented in Table 6. Gemini significantly outperformed all, and GPT (+) was significantly outperformed by all. Assuming the intermediate value (0.5) resulted in the highest AUC for GPT.

3.6. Accuracy in Different Pathologies

A total of 6997 slices that contain at least one pathology were present in the dataset. GPT answered 5232 of them, and detailed accuracy distributions for different pathologies are presented in Table 7. LLMs’ accuracy showed significant difference [χ2(2) = 1009.917, p < 0.001]. Pairwise analysis showed Gemini is superior to GPT [χ2(1) = 329.376, p < 0.001]; Grok is superior to Gemini [χ2(1) = 223.779, p < 0.001] and GPT [χ2(1) = 885.939, p < 0.001]. These outcomes were the same in every pathology individually, except schwannoma (GPT vs. Gemini ns. after Bonferroni correction), cavernoma (GPT vs. Gemini ns.), and others (Grok vs. Gemini after Bonferroni correction) (χ2 statistics omitted for readability).

3.7. Slice Difficulty

None of the LLMs were able to correctly answer for 4975 (13.4%) slices, 11,482 (32.2%) were classified correctly by two LLMs at the same time, and 13.474 (37.7%) were correctly classified by all LLMs. Detailed distributions are presented in Figure 2. Representative MRI slices illustrating model predictions have been included as Appendix B to improve transparency.

3.8. Sequence-Level Analysis

Individual accuracies of all LLMs for various pathologies are presented in Table 8. There were no sequences that GPT completely refused to answer for any slices of, and its responses are based on the slices that it had answered.
Grok only classified one sequence of one patient as all-slices normal among 299 completely normal brain MRIs, showing extreme overestimation for this matter. This constant overestimation has falsely presented high accuracy for sequences that have actual pathology. Statistical tests were run separately for ground truth positives and negatives.
Ground truth positive sequence (n = 1480) performance was significantly different among LLMs [χ2(2) = 431.807, p < 0.001]. As post hoc analysis showed, Gemini is superior to GPT [χ2(1) = 100.363, p < 0.001], and Grok is superior to GPT [χ2(1) = 362.024, p < 0.001] and Gemini [χ2(1) = 152.910, p < 0.001].
Ground truth negative sequence (n = 313) performance was also significantly different among all LLMs [χ2(2) = 431.807, p < 0.001]. Additionally, Gemini is superior to Grok [χ2(1) = 100.363, p < 0.001], and GPT is superior to Gemini [χ2(1) = 7062, p < 0.024] and Grok [χ2(1) = 92.092, p < 0.001], as post hoc analyses indicate. Obviously, these results are biased in favor of GPT since it did not respond to some slices, reducing the possibility of misclassifying as positive.
GPT overestimated (i.e., classified more slices than the number actually pathological) in 418 (23.3%) sequences and underestimated (i.e., classified less) in 1097 (61.2%). Gemini overestimated in 651 (36.3%) and underestimated in 813 (45.3%), whereas Grok overestimated in 1331 (74.2%) and underestimated in 309 (17.2%).

4. Discussion

The results of this study indicate that Grok is the most accurate sequence predictor, while Gemini is the most reliable for pathology detection amongst the investigated LLMs for brain MRI interpretation. Although Grok clearly struggled to identify the FLAIR sequence, overall it performed well in identifying, especially for T2 and T1 sequences. Gemini was unreliable in identifying sequences, performing worse than random prediction for FLAIR and T2, but it was the most accurate and had the best AUC in the ROC analysis. Identifying FLAIR was an important strength of GPT when compared to others. Although, distinguishing FLAIR from T2 sequences may depend heavily on the anatomical region being imaged, particularly concerning the visibility of cerebrospinal fluid, which poses a potential limitation for both human readers and automated models. GPT’s unanswered slices were mostly ground truth negative slices, possibly due to security systems designed to avoid missing potentially important pathologies. Nevertheless, GPT is still outperformed by Gemini significantly (and Grok, although this is statistically insignificant) when those responses were considered intermediate or negative. Both of the other LLMs would be outperforming GPT, should those responses be positive. Grok has the highest accuracy in slices that contain actual pathology, showing higher sensitivity at the expense of reduced specificity.
A key challenge encountered in this study was the slices for which GPT did not provide an answer. This behavior was unique to GPT and required careful consideration. Considering all unanswered slices as incorrect would unfairly penalize GPT. The exact mechanism behind GPT’s refusal likely involves the rule-based reward system described in OpenAI’s official documentation [13]. Other investigated LLMs mention various safety systems but do not detail them (they also complied with every request sent for the purposes of the present study). However, even when unanswered slices were excluded from analysis (introducing severe bias favoring GPT), Gemini still demonstrated better performance in the ROC analysis. It should be noted that although differences between models were statistically significant, all AUC values remained relatively low (below 60%), indicating limited overall discriminative power. The sharp drop in Precision–Recall curves at higher recall levels likely reflects class imbalance, given that normal slices outnumber abnormal ones by approximately four to one. Another limitation was the binary classification system. Since not all pathologies are clearly visible on every MRI sequence, the only reliable question for single-slice analysis was whether the slice was normal or abnormal. We addressed this limitation by grouping slices according to the specific pathology present and assessing them separately. No clinical information about patients was provided to any of the LLMs, as this study aimed to evaluate MRI interpretation capabilities rather than comprehensive clinical decision-making skills. This may have reduced the performance for all models. The importance of presenting clinical context may be assessed in dedicated studies.
Grok’s real time integration with its social media platform X [6] makes it unique, adding extra interest to its results and comparisons to other LLMs. However, reports regarding the performance of Grok in the field of medicine are very limited, with only 3 relevant results among 14 total in PubMed. Two of those reports investigate patient information related capabilities, and one is about medical exam questions. None of these studies are actually investigating medical images or any image interpretation capabilities, placing them outside the scope for the purposes of the present study. This study is a first in this aspect for Grok.
There are no reports investigating Gemini alone. All existing reports compare Gemini with GPT and other LLMs. Among those, only Günay et al. provided actual clinical images (of ECGs) and asked the LLM to interpret the image [14]. Although they concluded that the LLMs’ performance is not adequate to rely on them, their results also show that GPT-4o is superior to Gemini Advanced. Another real image interpretation study was the Hindy et al. study that investigated gram-stained bacteria identification capabilities, concluding that GPT-4o outperforms Gemini 1.5 Pro; although, both LLMs are unreliable [15]. In the field of neuroradiology, Gupta et al. compared GPT-4 to its predecessor GPT-3.5 and Gemini for text based multiple choice questions, and although GPT-4 outperformed the other two, its best accuracy remained at 64.89% [16]. The other few studies neither contained actual medical images nor were related to neuroradiology, placing them outside the scope.
The most investigated LLM is perhaps GPT, with dozens of articles for each version of it. However, since the image interpretation is a recent feature, evidence regarding this capability is still very limited. One of the most relevant studies was by Zhang et al. [17]. They pre-trained GPT-4 with images of cerebral hemorrhages and then asked it to segment the bleeding areas. Even though they had pre-trained the model and presented the model only with the slices they deemed most representative of the condition, they concluded that GPT showed promising results, but it is not adequate just yet for use in a clinical setting. Pre-training appears to significantly improve the performance of GPT and other LLMs [18].
Two studies with some level of similarity in design investigated multimodal radiological diagnostic capabilities of GPT-4v (earlier version with image interpretation) both concluding that GPT excels at identifying the modality. Brin et al.’s study only included 43 head CT images (out of 103 total CT slices), and although head CT results were not presented separately, the overall diagnostic accuracy for CT was reported as 36.4% [19]. On the other hand, Strotzer et al. reported individual results, and GPT achieved 60% accuracy for MS, 56% for brain tumors, 54% for brain hemorrhage, and 46% for ischemic stroke in their multimodal study [20].
On the other hand, the current gold standard method for pathology prediction in MRI is specifically trained and fine-tuned dedicated models based on CNN or ViT. Recently, ViT-based models have outperformed CNN-based models with accuracies reported above 98–99% [21,22]. This recent report of ViT-based models indicates that the LLMs that use ViT as we discussed are easily able to adapt to these improvements to achieve similar results in the near future. Subsequently, LLMs may prove very useful and easy to use as triage support or decision support systems, once their reliability is established.

5. Conclusions

This is the first study to investigate the neuroradiology performance of ChatGPT-4o, Gemini and Grok each individually and in direct comparison. The image interpretation capabilities of these LLMs are limited for now and require improvement before they can compete with specifically trained and fine-tuned CNN- or ViT-based applications. Gemini outperformed others in pathology prediction (with the highest AUC), while Grok did better in sequence prediction tasks. Notably, Grok showed significantly higher accuracy than the other models when an actual pathology is present, though at the cost of reduced specificity. Future studies could leverage multi-class classification to provide a more clinically meaningful evaluation beyond the binary classification approach used here.

Author Contributions

Conceptualization: A.S., M.C.S. and E.C.; Data curation: A.S. and B.S.; Formal Analysis: A.S.; Investigation: A.S., M.C.S., B.S., G.E., O.Y.T., K.N. and Z.D.; Methodology: A.S., M.C.S. and B.S.; Project administration: A.S., M.C.S. and E.C.; Resources: G.E., O.Y.T., K.N. and Z.D.; Software: B.S.; Supervision: E.C.; Validation: A.S., M.C.S., B.S. and E.C.; Visualization: A.S.; Writing—original draft: A.S. and M.C.S.; Writing—review and editing: A.S., M.C.S., B.S., G.E., O.Y.T., K.N., Z.D. and E.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. All expenses were covered by the authors themselves. However, various free trial offers that the investigated LLM services provide to any user without any commitment during the investigation period (November–December 2024) were used to reduce costs. Any information regarding the details or even existence of this study was not disclosed to any parties related to these services.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. No ethical approval was necessary for this research as it was conducted using a pre-existing dataset according to local and regional guidelines.

Informed Consent Statement

Not applicable since this research as it was conducted using a pre-existing dataset.

Data Availability Statement

A part of the investigated dataset was previously distributed as the Gazi Brains 2020 dataset [8]. The rest of the dataset is currently confidential. Additional data are available upon reasonable request.

Acknowledgments

The authors would like to thank Public Health Specialist Merve BUKE SAHIN, M.D. for reviewing, commenting on, and confirming the statistical methods used and results reported in this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

LLMLarge Language Model
MRIMagnetic Resonance Imaging/Image
AIArtificial Intelligence
CNNConvolutional Neural Network
ViTVision Transformer
FLAIRFluid-Attenuated Inversion Recovery
HGGHigh Grade Glioma
LGGLow Grade Glioma
SDKSoftware Development Kit
OROdds Ratio
CIConfidence Interval
PPVPositive Predictive Value
NPVNegative Predictive Value
ROCReceiver Operating Characteristic
AUCArea Under the Curve
ECGElectrocardiogram
CTComputerized Tomography

Appendix A

Part of the python code used that shows the prompt used. Please note that lines starting with # are activated one at a time to direct requests to their respective LLMs after making necessary adjustments to the base_url variable.
  • response = completions_with_backoff(
  • # model = “gpt-4o”,
  • # model = “grok-vision-beta”,
  • # model = “gemini-1.5-pro”,
  • messages= [
  • {“role”: “system”, “content”: “You are a consultant neurosurgeon.”},
  • {
  • “role”: “user”,
  • “content”: [
  • {“type”: “text”, “text”: “You are given an axial MRI slice. Please evaluate whether there is any pathology. Your answer should be in the form of ‘true’ or ‘false,’ and you should also predict the MRI sequence of this slice. Format your answer as ‘‘True’ or ‘False’/MRI sequence’ without any additional descriptions. Possible MRI sequences for this task are ‘T1’, ‘T2’, ‘FLAIR’, ‘T1c’. You must provide an answer with the requested format.”},
  • {
  • “type”: “image_url”,
  • “image_url”: {
  • “url”: f”data:image/jpeg;base64,{base64_image}”,
  • },
  • },
  • ],
  • }
  • ],
  • max_tokens = 300,
  • )

Appendix B

Figure A1. Sample images from the dataset that demonstrate some disagreement cases. TP: True Positive; FN: False Negative; FP: False Positive. (A): Only GPT TP; (B): Only Grok TP; (C): Only Gemini TP; (D): Only GPT FN; (E): Only Grok FN; (F): Only Gemini FN. (G): All FP; (H): All FN; (I): All TP.
Figure A1. Sample images from the dataset that demonstrate some disagreement cases. TP: True Positive; FN: False Negative; FP: False Positive. (A): Only GPT TP; (B): Only Grok TP; (C): Only Gemini TP; (D): Only GPT FN; (E): Only Grok FN; (F): Only Gemini FN. (G): All FP; (H): All FN; (I): All TP.
Diagnostics 15 01320 g0a1

References

  1. Döhler, F.; Mormann, F.; Weber, B.; Elger, C.E.; Lehnertz, K. A Cellular Neural Network Based Method for Classification of Magnetic Resonance Images: Towards an Automated Detection of Hippocampal Sclerosis. J. Neurosci. Methods 2008, 170, 324–331. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, W.; Li, R.; Deng, H.; Wang, L.; Lin, W.; Ji, S.; Shen, D. Deep Convolutional Neural Networks for Multi-Modality Isointense Infant Brain Image Segmentation. NeuroImage 2015, 108, 214–224. [Google Scholar] [CrossRef] [PubMed]
  3. Aydoseli, A.; Unal, T.C.; Kardes, O.; Doguc, O.; Dolas, I.; Adiyaman, A.E.; Ortahisar, E.; Silahtaroglu, G.; Aras, Y.; Sabanci, P.A.; et al. An Early Warning System Using Machine Learning for the Detection of Intracranial Hematomas in the Emergency Trauma Setting. Turk. Neurosurg. 2022, 32, 459–465. [Google Scholar] [CrossRef] [PubMed]
  4. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929v2. [Google Scholar]
  5. Hello GPT-4o. Available online: https://openai.com/index/hello-gpt-4o/ (accessed on 5 December 2024).
  6. xAI. Available online: https://x.ai/ (accessed on 5 December 2024).
  7. Introducing Gemini: Google’s Most Capable AI Model Yet. Available online: https://blog.google/technology/ai/google-gemini-ai/ (accessed on 5 December 2024).
  8. Sagiroglu, S.; Celtikci, E.; Demirezen, M.U.; Terzi, R.; Asan, A.; Celtikci, P.; Sozer, A.; Sahin, M.C.; Atici, B.M.; Ucar, M.; et al. Gazi Brains 2020 Dataset 2020. Available online: https://www.synapse.org/Synapse:syn25926092 (accessed on 2 May 2025).
  9. Demirezen, M.U.; Terzi, R.; Aşan, A.; Çeltikçi, E.; Sağiroğlu, Ş.; Börcek, A.Ö.; Atici, B.M. Beyin Görüntülerindeki Aykırılığın veya Anomalilerin Zeki Bir Sistem ile Otomatik Tespiti ve Bildirim Sistemi. Patient TR2021009196A1, 21 December 2022. [Google Scholar]
  10. OPENAPI Initative. Available online: https://learn.openapis.org/ (accessed on 5 December 2024).
  11. Base64—Base16, Base32, Base64, Base85 Data Encodings. Available online: https://docs.python.org/3/library/base64.html (accessed on 1 December 2024).
  12. Taha, A.A.; Hanbury, A. Metrics for Evaluating 3D Medical Image Segmentation: Analysis, Selection, and Tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef] [PubMed]
  13. Improving Model Safety Behavior with Rule-Based Rewards. Available online: https://openai.com/index/improving-model-safety-behavior-with-rule-based-rewards/ (accessed on 13 December 2024).
  14. Günay, S.; Öztürk, A.; Yiğit, Y. The Accuracy of Gemini, GPT-4, and GPT-4o in ECG Analysis: A Comparison with Cardiologists and Emergency Medicine Specialists. Am. J. Emerg. Med. 2024, 84, 68–73. [Google Scholar] [CrossRef] [PubMed]
  15. Hindy, J.-R.; Souaid, T.; Kovacs, C.S. Capabilities of GPT-4o and Gemini 1.5 Pro in Gram Stain and Bacterial Shape Identification. Future Microbiol. 2024, 19, 1283–1292. [Google Scholar] [CrossRef] [PubMed]
  16. Gupta, R.; Hamid, A.M.; Jhaveri, M.; Patel, N.; Suthar, P.P. Comparative Evaluation of AI Models Such as ChatGPT 3.5, ChatGPT 4.0, and Google Gemini in Neuroradiology Diagnostics. Cureus 2024, 16, e67766. [Google Scholar] [CrossRef] [PubMed]
  17. Zhang, D.; Ma, Z.; Gong, R.; Lian, L.; Li, Y.; He, Z.; Han, Y.; Hui, J.; Huang, J.; Jiang, J.; et al. Using Natural Language Processing (GPT-4) for Computed Tomography Image Analysis of Cerebral Hemorrhages in Radiology: Retrospective Analysis. J. Med. Internet Res. 2024, 26, e58741. [Google Scholar] [CrossRef] [PubMed]
  18. Yu, S.; Sun, W.; Mi, D.; Jin, S.; Wu, X.; Xin, B.; Zhang, H.; Wang, Y.; Sun, X.; He, X. Artificial Intelligence Diagnosing of Oral Lichen Planus: A Comparative Study. Bioengineering 2024, 11, 1159. [Google Scholar] [CrossRef] [PubMed]
  19. Brin, D.; Sorin, V.; Barash, Y.; Konen, E.; Glicksberg, B.S.; Nadkarni, G.N.; Klang, E. Assessing GPT-4 Multimodal Performance in Radiological Image Analysis. Eur. Radiol. 2025, 35, 1959–1965. [Google Scholar] [CrossRef] [PubMed]
  20. Strotzer, Q.D.; Nieberle, F.; Kupke, L.S.; Napodano, G.; Muertz, A.K.; Meiler, S.; Einspieler, I.; Rennert, J.; Strotzer, M.; Wiesinger, I.; et al. Toward Foundation Models in Radiology? Quantitative Assessment of GPT-4V’s Multimodal and Multianatomic Region Capabilities. Radiology 2024, 313, e240955. [Google Scholar] [CrossRef] [PubMed]
  21. Asiri, A.A.; Shaf, A.; Ali, T.; Shakeel, U.; Irfan, M.; Mehdar, K.M.; Halawani, H.T.; Alghamdi, A.H.; Alshamrani, A.F.A.; Alqhtani, S.M. Exploring the Power of Deep Learning: Fine-Tuned Vision Transformer for Accurate and Efficient Brain Tumor Detection in MRI Scans. Diagnostics 2023, 13, 2094. [Google Scholar] [CrossRef] [PubMed]
  22. Ali Al-Hamza, K. ViT-BT: Improving MRI Brain Tumor Classification Using Vision Transformer with Transfer Learning. Int. J. Soft Comput. Eng. 2024, 14, 16–26. [Google Scholar] [CrossRef]
Figure 1. ROC Curve and Precision–Recall curve for all model variations. GPT 0.5 = missing data were replaced with an intermediate value (0.5); GPT (+) = missing answers were assumed as positive; GPT (−) = missing answers were assumed as negative.
Figure 1. ROC Curve and Precision–Recall curve for all model variations. GPT 0.5 = missing data were replaced with an intermediate value (0.5); GPT (+) = missing answers were assumed as positive; GPT (−) = missing answers were assumed as negative.
Diagnostics 15 01320 g001
Figure 2. Correct classification rates of MRI slices by ChatGPT-4o, Grok, and Gemini, including model combinations and cases where all models failed.
Figure 2. Correct classification rates of MRI slices by ChatGPT-4o, Grok, and Gemini, including model combinations and cases where all models failed.
Diagnostics 15 01320 g002
Table 1. Descriptive information regarding the investigated dataset.
Table 1. Descriptive information regarding the investigated dataset.
Origin PatientSequence
FLAIR (n = 9511, 26.63%)T2 (n = 9347, 26.17%)T1 (n = 9343, 26.16%)T1c (n = 7510, 21.03%)Total (n = 35,711)
PositiveNegativePositiveNegativePositiveNegativePositiveNegativePositiveNegative
Normal01854 (100%)01819 (100%)01819 (100%)0429 (100%)05921 (100%)
HGG499 (40.54%)732 (59.46%)499 (40.94%)720 (59.06%)499 (41.%)718 (59.%)499 (41.07%)716 (58.93%)1996 (40.88%)2886 (59.12%)
LGG116 (20.21%)458 (79.79%)116 (20.42%)452 (79.58%)116 (20.42%)452 (79.58%)115 (20.32%)451 (79.68%)463 (20.34%)1813 (79.66%)
Meningioma577 (20.88%)2187 (79.12%)577 (21.27%)2136 (78.73%)577 (21.27%)2136 (78.73%)572 (21.72%)2061 (78.28%)2303 (21.28%)8520 (78.72%)
Schwannoma102 (14.55%)599 (85.45%)102 (15.04%)576 (84.96%)102 (14.96%)580 (85.04%)102 (14.96%)580 (85.04%)408 (14.87%)2335 (85.13%)
Cavernoma128 (12.46%)899 (87.54%)128 (12.62%)886 (87.38%)128 (12.67%)882 (87.33%)97 (12.29%)692 (87.71%)481 (12.53%)3359 (87.47%)
Metastatic Disease130 (34.39%)248 (65.61%)130 (34.57%)246 (65.43%)130 (34.76%)244 (65.24%)130 (34.67%)245 (65.33%)520 (34.6%)983 (65.4%)
Other84 (24.78%)255 (75.22%)84 (25.07%)251 (74.93%)84 (25.07%)251 (74.93%)71 (24.15%)223 (75.85%)323 (24.79%)980 (75.21%)
Post-op130 (20.22%)513 (79.78%)129 (20.64%)496 (79.36%)129 (20.64%)496 (79.36%)115 (21.82%)412 (78.18%)503 (20.79%)1917 (79.21%)
TOTAL1766 (18.57%)7745 (81.43%)1765 (18.88%)7582 (81.12%)1765 (18.89%)7578 (81.11%)1701 (22.65%)5809 (77.35%)6997 (19.59%)28,714 (80.41%)
Table 2. Performance metrics and confusion table of GPT-4o. Care should be taken interpreting this table, since this excludes slices that the GPT refused to answer. TP: True Positive, TN: True Negative, FP: False Positive, FN: False Negative, PPV: Positive Predictive Value, NPV: Negative Predictive Value, LR: Likelihood Ratio (Positive or Negative), DOR: Diagnostic Odds Ratio.
Table 2. Performance metrics and confusion table of GPT-4o. Care should be taken interpreting this table, since this excludes slices that the GPT refused to answer. TP: True Positive, TN: True Negative, FP: False Positive, FN: False Negative, PPV: Positive Predictive Value, NPV: Negative Predictive Value, LR: Likelihood Ratio (Positive or Negative), DOR: Diagnostic Odds Ratio.
TPTNSensitivitySpecificityLR+DORF1 (DICE)AccuracySequence Accuracy
FPFNPPVNPVLR−F0.5F2
FLAIR643 (8.32%)4720 (61.11%)0.4280.7591.7762.3550.35369.43%5652/7724 (73.17%)
1501 (19.43%)860 (11.13%)0.3000.8460.7540.3190.394
T2354 (5.80%)4334 (70.98%)0.2910.8872.5753.2230.33376.78%5469/6106 (89.57%)
554 (9.07%)864 (14.15%)0.3900.8340.7990.3650.307
T1129 (1.92%)5189 (77.14%)0.1000.9552.2222.3590.15579.05%6727/7285 (92.34%) 1
242 (3.60%)1167 (17.35%)0.3480.8160.9420.2330.117
T1c435 (8.45%)3303 (64.16%)0.3580.8402.2382.9290.38272.61%4733/5148 (91.94%) 1
629 (12.22%)781 (15.17%)0.4090.8090.7640.3980.367
TOTAL1561 (6.07%)17,546 (68.26%)0.2980.8572.0842.5450.32174.33%22,581/26,263 (85.98%)
2926 (11.38%)3672 (14.29%)0.3480.8270.8190.3370.307
1 T1 and T1c answers were considered interchangeably.
Table 3. Performance metrics and confusion table of Grok. TP: True Positive, TN: True Negative, FP: False Positive, FN: False Negative, PPV: Positive Predictive Value, NPV: Negative Predictive Value, LR: Likelihood Ratio (Positive or Negative), DOR: Diagnostic Odds Ratio.
Table 3. Performance metrics and confusion table of Grok. TP: True Positive, TN: True Negative, FP: False Positive, FN: False Negative, PPV: Positive Predictive Value, NPV: Negative Predictive Value, LR: Likelihood Ratio (Positive or Negative), DOR: Diagnostic Odds Ratio.
TPTNSensitivitySpecificityLR+DORF1 (DICE)AccuracySequence Accuracy
FPFNPPVNPVLR−F0.5F2
FLAIR847 (8.91%)5388 (56.65%)0.4800.6961.5792.1140.34165.56%1610/9511 (16.93%)
2357 (24.78%)919 (9.66%)0.2640.8540.7470.2900.413
T2762 (8.15%)5432 (58.11%)0.4320.7161.5211.9180.32666.27%7314/9347 (78.25%)
2150 (23.00%)1003 (10.73%)0.2620.8440.7930.2840.382
T1505 (5.41%)5852 (62.64%)0.2860.7721.2541.3560.25268.04%8231/9343 (88.10%) 1
1726 (18.47%)1260 (13.49%)0.2260.8230.9250.2360.272
T1c759 (10.11%)3895 (51.86%)0.4460.6711.3561.6420.34761.97%6495/7510 (86.48%) 1
1914 (25.49%)942 (12.54%)0.2840.8050.8260.3060.400
TOTAL2873 (8.05%)20,567 (57.59%)0.4110.7161.4471.7580.31965.64%23,650/35,711 (66.23%)
8147 (22.81%)4124 (11.55%)0.2610.8330.8230.2820.369
1 T1 and T1c answers were considered interchangeably correct.
Table 4. Performance metrics and confusion table of Gemini. TP: True Positive, TN: True Negative, FP: False Positive, FN: False Negative, PPV: Positive Predictive Value, NPV: Negative Predictive Value, LR: Likelihood Ratio (Positive or Negative), DOR: Diagnostic Odds Ratio.
Table 4. Performance metrics and confusion table of Gemini. TP: True Positive, TN: True Negative, FP: False Positive, FN: False Negative, PPV: Positive Predictive Value, NPV: Negative Predictive Value, LR: Likelihood Ratio (Positive or Negative), DOR: Diagnostic Odds Ratio.
TPTNSensitivitySpecificityLR+DORF1 (DICE)AccuracySequence Accuracy
FPFNPPVNPVLR−F0.5F2
FLAIR606 (6.37%)6688 (70.32%)0.3430.8642.5223.3180.35376.69%2135/9511 (22.45%)
1057 (11.11%)1160 (12.20%)0.3640.8520.7600.3600.347
T2572 (6.12%)6658 (71.23%)0.3240.8782.6563.4490.35177.35%1340/9347 (14.34%)
924 (9.89%)1193 (12.76%)0.3820.8480.7700.3690.334
T1396 (4.24%)6557 (70.18%)0.2240.8651.6591.8490.24874.42%7618/9343 (81.54%) 1
1021 (10.93%)1369 (14.65%)0.2790.8270.8970.2660.233
T1c669 (8.91%)4473 (59.56%)0.3930.7701.7092.1690.36168.47%5468/7510 (72.81%) 1
1336 (17.79%)1032 (13.74%)0.3340.8130.7880.3440.380
TOTAL2243 (6.28%)24,376 (68.26%)0.3210.8492.1262.6580.33174.54%16,561/35,711 (46.38%)
4338 (12.15%)4754 (13.31%)0.3410.8370.8000.3370.325
1 T1 and T1c answers were considered interchangeably correct.
Table 5. Sequence prediction accuracy comparison results.
Table 5. Sequence prediction accuracy comparison results.
Cochran’s QGPT vs. GrokGPT vs. GeminiGrok vs. GeminiLLM Order
FLAIRχ2(2) = 4863.681
p < 0.001
χ2(1) = 3365.557,
p < 0.001
χ2(1) = 2566.381,
p < 0.001
χ2(1) = 122.088,
p < 0.001
GPT > Gemini > Grok
T2χ2(2) = 7041.433
p < 0.001
χ2(1) = 1031.342,
p < 0.001
χ2(1) = 2924.384,
p < 0.001
χ2(1) = 5229.658,
p < 0.001
Grok > GPT > Gemini
T1χ2(2) = 1588.105
p < 0.001
χ2(1) = 1358.836,
p < 0.001
χ2(1) = 628.697,
p < 0.001
χ2(1) = 186.248,
p < 0.001
Grok > Gemini > GPT
T1cχ2(2) = 1156.113
p < 0.001
χ2(1) = 1146.864,
p < 0.001
χ2(1) = 171.306,
p < 0.001
χ2(1) = 461.498,
p < 0.001
Grok > Gemini > GPT
TOTALχ2(2) = 196,901.084
p < 0.001
χ2(1) = 189.132,
p < 0.001
χ2(1) = 1741.563,
p < 0.001
χ2(1) = 3759.616,
p < 0.001
Grok > GPT > Gemini
Table 6. ROC analysis results and comparisons. GPT 0.5 = missing data replaced with intermediate values (0.5); GPT (+) = missing data considered positive; GPT (−) = missing data considered negative.
Table 6. ROC analysis results and comparisons. GPT 0.5 = missing data replaced with intermediate values (0.5); GPT (+) = missing data considered positive; GPT (−) = missing data considered negative.
AUC
(95%CI)
vs. Geminivs. Grokvs. GPT 0.5vs. GPT (−)
Gemini58.5%
(57.9–59.1%)
Grok56.3%
(55.7–57.0%)
z = 6.462,
p < 0.001 *
GPT 0.556.2%
(55.5–56.9%)
z = 5.950,
p < 0.001 *
z = 0.262,
p = 0.793
GPT (−)56.1%
(55.5–56.6%)
z = 8.427,
p < 0.001 *
z = 0.867,
p = 0.386
z = 0.733,
p = 0.463
GPT (+)54.3%
(53.7–55.0%)
z = 10.841,
p < 0.001 *
z = 5.004,
p < 0.001 *
z = 20.593,
p < 0.001 *
z = 5.952,
p < 0.001 *
* marks statistical significance
Table 7. Accuracy on pathological slices.
Table 7. Accuracy on pathological slices.
PathologyNumber of Positive SlicesGPT 1 AccuracyGrok AccuracyGemini AccuracyGPT Accuracy 2 (n)
HGG1996 39.8%57.7%51.4%48.9% (1625)
LGG463 13.8%30.0%21.0%18.8% (341)
Meningioma230317.5%36.1%26.3%24.6% (1644)
Schwannoma408 10.8%25.7%15.0%14.4% (305)
Cavernoma481 12.9%29.1%16.2%17.4% (356)
Metastatic Disease520 20.0%42.1%31.5%26.8% (388)
Other323 8.7%34.7%27.9%13.5% (207)
Post-op503 11.9%34.8%24.1%16.4% (366)
TOTAL6997 22.3%41.1%32.1%29.8% (5232)
1 Decline to answer is considered as wrong answer for GPT. 2 Only answered slices considered (n = number of answered slices).
Table 8. Accuracy on sequence-based analysis.
Table 8. Accuracy on sequence-based analysis.
SubgroupNumber of SequencesGPT AccuracyGrok AccuracyGemini Accuracy
Normal29942.5%0.3%31.1%
HGG23692.8%100.0%97.9%
LGG11269.6%92.9%86.6%
Meningioma54870.6%98.9%85.6%
Schwannoma14067.9%98.6%86.4%
Cavernoma19370.5%99.0%77.7%
Metastatic Disease7279.2%100.0%91.7%
Other6674.2%98.5%87.9%
Post-op12766.1%94.5%85.8%
TOTAL179368.7%81.9%77.7%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sozer, A.; Sahin, M.C.; Sozer, B.; Erol, G.; Tufek, O.Y.; Nernekli, K.; Demirtas, Z.; Celtikci, E. Do LLMs Have ‘the Eye’ for MRI? Evaluating GPT-4o, Grok, and Gemini on Brain MRI Performance: First Evaluation of Grok in Medical Imaging and a Comparative Analysis. Diagnostics 2025, 15, 1320. https://doi.org/10.3390/diagnostics15111320

AMA Style

Sozer A, Sahin MC, Sozer B, Erol G, Tufek OY, Nernekli K, Demirtas Z, Celtikci E. Do LLMs Have ‘the Eye’ for MRI? Evaluating GPT-4o, Grok, and Gemini on Brain MRI Performance: First Evaluation of Grok in Medical Imaging and a Comparative Analysis. Diagnostics. 2025; 15(11):1320. https://doi.org/10.3390/diagnostics15111320

Chicago/Turabian Style

Sozer, Alperen, Mustafa Caglar Sahin, Batuhan Sozer, Gokberk Erol, Ozan Yavuz Tufek, Kerem Nernekli, Zuhal Demirtas, and Emrah Celtikci. 2025. "Do LLMs Have ‘the Eye’ for MRI? Evaluating GPT-4o, Grok, and Gemini on Brain MRI Performance: First Evaluation of Grok in Medical Imaging and a Comparative Analysis" Diagnostics 15, no. 11: 1320. https://doi.org/10.3390/diagnostics15111320

APA Style

Sozer, A., Sahin, M. C., Sozer, B., Erol, G., Tufek, O. Y., Nernekli, K., Demirtas, Z., & Celtikci, E. (2025). Do LLMs Have ‘the Eye’ for MRI? Evaluating GPT-4o, Grok, and Gemini on Brain MRI Performance: First Evaluation of Grok in Medical Imaging and a Comparative Analysis. Diagnostics, 15(11), 1320. https://doi.org/10.3390/diagnostics15111320

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop