Next Article in Journal
From Microscopes to Monitors: Unique Opportunities and Challenges in Digital Pathology Implementation in Remote Canadian Regions
Next Article in Special Issue
EGFR Amplification in Diffuse Glioma and Its Correlation to Language Tract Integrity
Previous Article in Journal
Data Analytics and Machine Learning Models on COVID-19 Medical Reports Enhanced with XAI for Usability
Previous Article in Special Issue
Preoperative Adult-Type Diffuse Glioma Subtype Prediction with Dynamic Contrast-Enhanced MR Imaging and Diffusion Weighted Imaging in Tumor Cores and Peritumoral Tissue—A Standardized Multicenter Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multidisciplinary, Clinical Assessment of Accelerated Deep-Learning MRI Protocols at 1.5 T and 3 T After Intracranial Tumor Surgery and Their Influence on Residual Tumor Perception

by
Christer Ruff
1,*,
Till-Karsten Hauser
1,
Constantin Roder
2,3,
Daniel Feucht
2,3,
Paula Bombach
3,4,5,
Leonie Zerweck
1,
Deborah Staber
1,
Frank Paulsen
6,
Ulrike Ernemann
1 and
Georg Gohla
1
1
Department of Diagnostic and Interventional Neuroradiology, Eberhard Karls-University Tübingen, D-72076 Tübingen, Germany
2
Department of Neurosurgery, University of Tuebingen, D-72076 Tuebingen, Germany
3
Center for Neuro-Oncology, Comprehensive Cancer Center Tuebingen-Stuttgart, University Hospital of Tuebingen, Eberhard Karls University of Tuebingen, D-72070 Tuebingen, Germany
4
Department of Neurology and Interdisciplinary Neuro-Oncology, University Hospital Tuebingen, D-72076 Tuebingen, Germany
5
Hertie Institute for Clinical Brain Research, Center of Neuro-Oncology, Eberhard Karls University Tuebingen, D-72076 Tuebingen, Germany
6
Department of Radiation Oncology, University Hospital Tuebingen, D-72076 Tuebingen, Germany
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(15), 1982; https://doi.org/10.3390/diagnostics15151982
Submission received: 11 May 2025 / Revised: 31 July 2025 / Accepted: 4 August 2025 / Published: 7 August 2025
(This article belongs to the Special Issue Advanced Brain Tumor Imaging)

Abstract

Background/Objectives: Postoperative MRI is crucial for detecting residual tumor, identifying complications, and planning subsequent therapy. This study evaluates accelerated deep learning reconstruction (DLR) versus standard clinical protocols for early postoperative MRI following tumor resection. Methods: This study uses a multidisciplinary approach involving a neuroradiologist, neurosurgeon, neuro-oncologist, and radiotherapist to evaluate qualitative aspects using a 5-point Likert scale, the preferred reconstruction variant and potential residual tumor of DLR and conventional reconstruction (CR) of FLAIR, T1-weighted non-contrast and contrast-enhanced (T1), and coronal T2-weighted (T2) sequences for 1.5 and 3 T MRI. Quantitative analysis included the image quality metrics Structural Similarity Index (SSIM), Multi-Scale SSIM (MS-SSIM), Feature Similarity Index (FSIM), Noise Quality Metric (NQM), signal-to-noise ratio (SNR), and Peak SNR (PSNR) with CR as a reference. Results: All raters strongly preferred DLR over CR. This was most pronounced for FLAIR images at 1.5 and 3 T (91% at 1.5 T and 97% at 3 T) and least pronounced for T1 at 1.5 T (79% for non-contrast-enhanced and 84% for contrast-enhanced sequences) and for T2 at 3 T (69%). DLR demonstrated superior qualitative image quality for all sequences and field strengths (p < 0.001), except for T2 at 3 T, which was observed across all raters (p = 0.670). Diagnostic confidence was similar at 3 T with better but non-significant differences for T2 (p = 0.134) and at 1.5 T with better but non-significant differences for non-contrast-enhanced T1 (p = 0.083) and only marginally significant results for FLAIR (p = 0.033). Both the SSIM and MS-SSIM indicated near-perfect similarity between CR and DLR. FSIM performs worse in terms of consistency between CR and DLR. The image quality metrics NQM, SNR, and PSNR showed better results for DLR. Visual assessment of residual tumor was similar at 3 T but differed at 1.5 T, with more residual tumor detected with DLR, especially by the neurosurgeon (n = 4). Conclusions: An accelerated DLR protocol demonstrates clinical feasibility, enabling high-quality reconstructions in challenging postoperative MRIs. DLR sequences received strong multidisciplinary preference, underscoring their potential to improve neuro-oncologic decision making and suitability for clinical implementation.

1. Introduction

Postoperative imaging following intracranial tumor resection is standard for evaluating residual disease, detecting complications, and planning further treatment [1,2,3,4,5]. A comprehensive MRI protocol should include both non-enhanced and contrast-enhanced T1-weighted and T2-weighted sequences, as well as fluid-attenuated inversion recovery (FLAIR) imaging [6]. Misclassification postoperatively can lead to an inaccurate assessment of tumor progression or treatment response in follow-up examinations. Consequently, precise visualization and recognition of residual tumor are imperative to distinguish postoperative changes from residual tumor areas in subsequent scans. Postoperative MRI is challenging. Air–tissue interfaces, metallic implants, and tissue distortion create susceptibility artifacts, especially on conventional EPI-based diffusion-weighted imaging (DWI) sequences. Moreover, the mental state of tumor patients who have undergone surgery is diminished, thereby emphasizing the necessity for quick diagnostic examinations. Prolonged scan times can result in physical discomfort for the patient and, therefore, introduce image motion artifacts or even premature termination of the examination [7].
The trade-off between high-resolution imaging and scan time limits conventional MRI techniques. A critical consideration is achieving the optimal balance between the three interdependent imaging parameters: image resolution, signal-to-noise ratio (SNR), and acquisition time. While higher resolution enables the visualization of finer details, it is often accompanied by a reduction in signal-to-noise ratio (SNR) and/or an increase in acquisition time [8]. Concurrently, a minimum level of SNR is imperative to discern the signal of interest from system noise. Accurate delineation of tumor boundaries, a crucial element for effective treatment planning and monitoring, can be compromised by image quality issues.
Artificial intelligence (AI) applications in medical imaging have garnered significant attention, as they hold considerable promise in overcoming the obstacles mentioned above [9]. This has resulted in substantial advancements in image quality and the integration of neural network technologies, including neuro-oncological imaging [7,10]. In particular, deep learning (DL) has emerged as a crucial technique to overcome the limitations of traditional acceleration methods in imaging. State-of-the-art DL models are designed to enhance clinical diagnosis by improving accuracy, speed, and overall capabilities [11]. Models such as deep neural networks, variational autoencoders, and generative adversarial networks have been demonstrated to accelerate MRI acquisition, reduce patient discomfort, and enhance workflow [11]. Convolutional neural networks with attention mechanisms and U-Net architectures can improve MRI spatial resolution and thereby enhance anatomical visualization. AI has shown considerable promise in various applications in brain tumor management, encompassing diagnostic imaging, prognostication, and treatment planning. This potential extends beyond the conventional uses of AI, such as MRI acceleration and enhancement [12,13]. Integrating AI into radiology and pathology workflows can advance neuro-oncology. In glioblastoma and IDH-mutant gliomas, deep learning algorithms have reduced acquisition times by approximately 30% and improved subjective image quality [7,10,14,15].
A recent study evaluated an experimental approach to employing DL reconstruction (DLR) for intraoperative MRI in patients under general anesthesia [16]. However, currently, there is a lack of research on the use of DLR in the postoperative setting. It is essential to recognize the role of expert groups in developing consensus-based guidelines for scheduled postoperative surveillance MRI. Organizations such as the National Institute of Health and Care Excellence (NICE, postoperative scan < 72 h) and the European Association of Neuro-Oncology (EANO, postoperative scan < 48 h) provide timing recommendations for postoperative MRI to confirm the extent of resection or residual enhancing tumor [17]. This time window is recommended to be observed postoperatively to avert the occurrence of nonneoplastic contrast enhancement due to surgical manipulation [18,19].
A multidisciplinary review of postoperative imaging using a DL protocol has not yet been reported. DLR could shorten scan times, improve image quality, and alter residual-disease assessment by changing image appearance and increasing SNR in vulnerable postoperative patients. Therefore, this study compared an accelerated DL protocol with our conventional 1.5 Tesla (T) and 3 T reconstructions. We evaluated whether DLR facilitates the detection of residual tumors after intracranial tumor resection in a multidisciplinary setting.

2. Materials and Methods

2.1. Study Design

The institutional review board approved this monocentric retrospective study under code 549/2024BO2. All study procedures followed the Declaration of Helsinki and its subsequent amendments. This study comprised 33 patients who underwent cerebral MRI at 1.5 T or 3 T within the first 48 h following tumor resection for therapy assessment between July 2023 and June 2024. Exclusion criteria included MR examinations without DL datasets, spine imaging, and non-tumorous lesions (see Figure 1).

2.2. Imaging Protocol and Deep Learning Reconstruction Algorithm

All examinations were performed with a 1.5 T (MAGNETOM SOLA, Siemens, Siemens Healthineers, Erlangen, Germany) or a 3 T (MAGNETOM Vida Fit, Siemens Healthineers, Erlangen, Germany) clinical MRI scanner. Patients were scanned in the supine position with a 20-channel head coil on each scanner. The standard clinical protocol for postoperative imaging consists of FLAIR in the axial plane, T1-weighted imaging with and without contrast, T2-weighted imaging in the coronal plane, T2*-weighted imaging, diffusion-weighted imaging with two different b-values in the axial plane (0 s/mm2 and 1000 s/mm2), and corresponding apparent diffusion coefficient mapping. According to the standard clinical protocol, spin-echo sequences at 1.5 T were used for conventionally reconstructed T1-weighted sequences, while turbo spin-echo sequences (TSE) at 3 T were used for brain imaging. For the accelerated 1.5 T DL protocol, T1-weighted TSE sequences were used.
The undersampled DLR sequences were acquired and reconstructed separately for FLAIR, T1-weighted contrast-enhanced, and T2-weighted TSE imaging immediately after the corresponding conventional sequence type, with the same planes and slice thickness. The DLR algorithm Deep Resolve Boost by Siemens Healthineers (Erlangen, Germany) used in this study was cleared by the U.S. Food and Drug Administration (FDA) and has since been adopted in multiple investigations [1]. The software (XA50) offers a variational network approach, which alternates physics-based data consistency steps with learned convolutional neural network (CNN) regularization in a fixed, unrolled cascade. The CNN component, referred to as the “Deep, Iterative, Hierarchical Network,” extends a Down Up architecture through nested hierarchical blocks, allowing multi-resolution processing while conserving memory. A separate calibration acquisition supplies coil sensitivity maps and a bias field estimate to ensure homogeneous image intensity during reconstruction. Training relied on more than 25,000 fully sampled 2D slices obtained from volunteer examinations on a variety of clinical 1.5 T and 3 T MAGNETOM scanners (Siemens Healthineers, Erlangen, Germany). Once trained, the model was integrated into the scanner’s on-board reconstruction pipeline, enabling real-time assessment of its clinical utility. The Deep Resolve Boost algorithm only supports DLR for TSE readouts but not for SE. Hence, a SE sequence including DLR with improved acquisition time could not be acquired at 1.5 T.
For conventional reconstructed axial FLAIR (FLAIRCR), axial T1-weighted non-contrast-enhanced (T1CR) and contrast-enhanced (T1CECR), and coronal T2-weighted images (T2CR), the acceleration factor was set to two phase encoding (PE) steps. In contrast, the acceleration factor was set to four PE steps for the DLR sequences: axial FLAIR (FLAIRDLR), axial T1-weighted non-contrast-enhanced (T1DLR) and contrast-enhanced (T1CEDLR), and coronal T2-weighted imaging (T2DLR). DLR was performed offline using the accessible standard clinical hardware infrastructure of the MRI workstation, with no discernible time impact on the clinical workflow. Table 1 and Table 2 provide a comprehensive overview of the acquisition parameters for these sequences and their DLR versions for 1.5 T and 3 T.

2.3. Qualitative Image Quality Analysis

The evaluation team consisted of four experienced raters: a neuroradiologist (Rater 1) with ten years of experience, a neurosurgeon (Rater 2) with sixteen years of experience, a neuro-oncologist (Rater 3) with eleven years of experience, and a radiation oncologist (Rater 4) with thirty years of experience. Before the qualitative evaluation of the patients included in this study, a training phase took place with five patients outside the study cohort to familiarize the raters with postoperative MRIs and the rating scores. The raters then independently assessed 33 image datasets, including CR and DLR images featuring axial FLAIR, coronal T2-weighted TSE, and axial T1-weighted imaging with and without contrast agent. The evaluations were conducted in a randomized order. Raters were blinded to the reconstruction type, clinical and radiological reports, and each other’s assessments, with all patient- or sequence-identifying markers eradicated to ensure anonymity. The study readings were performed in a certified reading room under controlled conditions on a dedicated workstation (GE Centricity PACS RA 1000, version 7.0.2; General Electric (GE) Healthcare, Chicago, IL, USA) and on certified diagnostic radiology monitors (RadiForce RX350, Eizo Corporation, Hakusan, Ishikawa, Japan). The four raters evaluated the datasets based on multiple parameters using a detailed Likert scale, ranging from 1 to 5, with 5 being the highest rating to ensure sufficient image quality gradation. The evaluation criteria for the 33 image datasets included image quality, tumor conspicuity (i.e., the delineation of the tumor), and diagnostic confidence. The following score was used to evaluate the overall image quality qualitatively: 1 = non-diagnostic images, 2 = poor image quality (significant blurring, noise, or artifacts that hinder interpretation, but some structures are still visible), 3 = fair image quality (noticeable blurring, noise, or artifacts that moderately affect interpretation), 4 = good image quality (minimal blurring, noise, or artifacts, allowing for effective interpretation), and 5 = excellent image quality (exceptionally clear with no blurring, noise, or artifacts, providing optimal conditions for interpretation). The visualization of the resection defect was rated with the following score: 1 = very poor conspicuity (the resection defect is almost indistinguishable from the surrounding tissue, making identification very difficult), 2 = poor conspicuity (the resection defect is faintly visible but blends significantly with the surrounding tissue, leading to uncertainty in identification), 3 = fair conspicuity (the resection defect is moderately visible but could be confused with surrounding structures, requiring careful analysis), 4 = good conspicuity (the resection defect is clearly visible and distinguishable from the surrounding tissue with minimal effort), and 5 = excellent conspicuity (the resection defect stands out prominently and is easily distinguishable from the surrounding tissue, ensuring confident identification). The degree of diagnostic confidence was evaluated with the following scheme: 1 = non-diagnostic images, 2 = low diagnostic confidence (confidence in the diagnosis is low, with significant doubt regarding findings and interpretation), 3 = moderate diagnostic confidence (there is moderate confidence in the diagnosis, with some uncertainty that requires further verification), 4 = high diagnostic confidence (there is high confidence in the diagnosis, with clear findings and minimal doubt), and 5 = excellent confidence (there is very high confidence in the diagnosis, with clear, definitive findings and no doubt). Each score incrementally reflected increasing levels of approval or quality, allowing for a nuanced assessment of the images. These assessments of both conventional and accelerated datasets were conducted in a random order across separate sessions, with at least a two-week washout period between each session to mitigate bias.

2.4. Rater Preference

In the preference analysis, each of the four experienced raters independently reviewed the image datasets two weeks after their initial analysis. The datasets encompassed FLAIR, T2-weighted TSE, non-enhanced, and contrast-enhanced T1-weighted sequences, which were evaluated on a per-patient basis. The raters were assigned to indicate their preferred sequence reconstruction type (CR vs. DLR) for individual sequence evaluations and an overall assessment of the entire dataset. The raters were unaware of the patients’ clinical information and the specific imaging techniques for each sequence.

2.5. Quantitative Image-Quality Indices

Image quality metrics (IQMs) were used to assess structural fidelity and noise sensitivity, utilizing publicly available formulas in MATLAB R2015b (The MathWorks, Inc., Natick, MA, USA). These include the Structural Similarity Index (SSIM) and the Multi-Scale SSIM (MS-SSIM), which were used to quantify the preservation of anatomical detail in DLR versus CR images as a reference. The Noise Quality Metric (NQM) provided a global measure of noise robustness, while the Feature Similarity Index (FSIM) captured the fidelity of high-frequency components, which are critical for delineating subtle postoperative changes. SNR and peak signal-to-noise ratio (PSNR) were also calculated [20]. The MATLAB source code for all IQM methods, except PSNR and SSIM, was obtained from the original authors [21,22,23,24]. We calculated PSNR and SSIM using the built-in functions of MATLAB.

2.6. Evaluation of Residual Tumor

The neuroradiologist, neurosurgeon, neuro-oncologist, and radiation oncologist independently reviewed each patient’s imaging datasets to assess residual tumor detection. These datasets consisted of FLAIR, T2-weighted TSE, and T1-weighted sequences (both non-enhanced and contrast-enhanced) with DLR or CR to determine the presence or absence of residual tumor compared to the preoperative images. Evaluations were performed on a per-patient basis.

2.7. Statistical Analysis

Statistical analysis was conducted using SPSS Statistics Version 30 (IBM, Armonk, NY, USA). Continuous variables were presented as mean ± standard deviation (SD), and ordinal variables as median and interquartile range (IQR). Image quality comparisons between groups were performed using the Wilcoxon signed-rank test. p-values from Wilcoxon signed-rank tests were reported unadjusted, in line with prior DLR image quality studies. Interrater reliability was evaluated using the Kendall coefficient of concordance (W). Kendall’s W is calculated by contrasting the number of concordant and discordant pairs, yielding a value between −1 and 1, where 0 indicates no correlation, 1 denotes perfect agreement, and −1 represents perfect disagreement. The effect sizes for Kendall’s W were categorized as follows: none (0 to <0.10), fair (0.10 to <0.30), moderate (0.30 to <0.50), large (0.50 to <0.70), and substantially large to almost perfect agreement (0.70–1.00). The significance level was set at alpha = 0.05.

3. Results

This retrospective study included 33 patients who underwent postoperative MRI after resection of an intracranial tumor, and all tumors were histopathologically confirmed. The mean age of the patients was 53.9 ± 20.3 years, including 14 males and 19 females. In total, 27% (9/25) of the patients had structural epilepsy due to the intracranial tumor. Comprehensive clinical data were available for all participants during the imaging study. Patient characteristics are shown in Table 3.
DLR sequences significantly reduced acquisition times compared to CR at 1.5 T and 3 T. At 1.5 T (Table 1), FLAIR, T2-weighted, T1-weighted, non-contrast, and contrast-enhanced T1-weighted sequences showed significant time savings ranging from 30% to 53%, with an overall mean reduction of 42%. At 3 T (Table 2), DLR sequences reduced scan times by approximately 27% to 31%, with an average decrease of 29%.

3.1. Image Quality-Based Analysis

The analysis of qualitative image quality, visualization of the resection defect, and diagnostic confidence performed by the multidisciplinary team is presented in detail in Table 4, Table 5, Table 6 and Table 7 for FLAIR, T2-weighted, as well as T1-weighted and contrast-enhanced T1-weighted sequences. Examples of DLR and CR images from FLAIR, T2, non-contrast, and contrast-enhanced T1-weighted sequences at 1.5 T and 3 T are shown in Figure 2, Figure 3, Figure 4 and Figure 5.
For FLAIR imaging, overall image quality and visualization of the resection cavity improved significantly with DLR across all raters and both field strengths (p < 0.001), as shown in Table 4. In addition, diagnostic confidence was enhanced considerably (pooled p = 0.033 for 1.5 T, p = 0.002 for 3 T). However, rater-based analysis revealed discrepancies. Some raters reported only a tendency for significant changes in the visualization of the resection cavity, while others showed no significant changes in diagnostic accuracy, especially the neuroradiologist. Analysis of pooled data from 1.5 T and 3 T scans revealed significant ratings in all domains by all raters, except for the neuroradiologist and radiation oncologist, for diagnostic confidence (p = 0.157 and p = 0.480, respectively).
For T2-weighted imaging, overall image quality improved significantly with DLR for all raters at 1.5 T (p < 0.001 to p = 0.003) but not at 3 T (0.083 for the neurosurgeon to 0.565 for the neuroradiologist), as shown in Table 5. Visualization of the resection cavity was rated as superior for DLR by the neurosurgeon (p = 0.034) and the neuro-oncologist (p = 0.005). Diagnostic confidence did not change significantly for 1.5 T for all raters except the neuro-oncologist (p = 0.025). Visualization of the resection cavity and diagnostic confidence also did not change considerably for all raters at 3 T (p = 0.134 to 1).
For T1-weighted sequences, overall image quality was significantly improved with DLR at 1.5 T, except for the neurosurgeon, who rated the non-contrast-enhanced images similarly to CR (Table 6). Mixed results are shown for visualization of the resection cavity, and no significant changes are shown for diagnostic confidence. At 3 T, DLR has a more obvious tendency compared to CR.
For contrast-enhanced T1-weighted sequences, overall image quality, visualization of the resection cavity, and diagnostic confidence were rated as superior for DLR compared to CR at 1.5 T, except for the neurosurgeon, who rated image quality and diagnostic confidence as equal. At 3 T, overall image quality and diagnostic confidence were rated superior for DLR by all raters.

3.2. Pooled Analysis of Qualitative Image Quality Assessment

Image quality was consistently rated higher for DLR imaging compared to CR imaging across all sequences by all raters in a pooled data analysis for 1.5 T (all p < 0.001), as shown in Figure 6A. Similar results are displayed for 3 T, except for T2-weighted images (p = 0.670). Regarding diagnostic confidence for 1.5 T, there was no significant difference between CR and DLR imaging for T1 (p = 0.083) but slightly significant differences for FLAIR (p = 0.033) and more significant changes for T2 (p = 0.007) and T1CE (p < 0.001) in favor of DLR, Figure 6A. Regarding diagnostic confidence for 3 T, there was a significant difference between CR and DLR imaging in favor of DLR for FLAIR (p = 0.002), T1, and T1CE (each p < 0.001). There were no significant differences for T2 (p = 0.134), as shown in Figure 6B. A more detailed insight into the results, including the mean (standard deviation) and median (interquartile ranges), can be found in Table 4, Table 5, Table 6 and Table 7 for each sequence.

3.3. Interrater Intraprotocol Agreement of Image Quality-Based Analysis

Interrater intraprotocol agreement among the four raters increased with DLR across most sequences and was more pronounced at 3 T than at 1.5 T, suggesting improved clarity and consistency in interpretation. Regarding overall image quality, at 1.5 T, Kendall’s W remained in the fair to moderate range. For 1.5 T, Kendall’s W for overall image qualityCR was 0.225 (0.131–0.331) vs. 0.231 (0.136–0.349) for the DLR protocol, visualization of the resection cavityCR was 0.271 (0.150–0.362) vs. 0.280 (0.167–0.398), and diagnostic confidenceCR was 0.463 (0.368–0.560) vs. 0.361 (0.324- 0.469). At 3 T, agreement was higher with DLR, particularly for the visualization of the resection cavity (0.484–0.730 for DLR vs. 0.173–0.546 for CR. Kendall’s W for overall image qualityCR at 3 T increased from 0.276 (0.126–0.360) to 0.413 (0.135–0.605) for the DLR protocol, and diagnostic confidenceCR was 0.463 (0.193–0.518) for CR vs. 0.469 (0.394–0.576) for the DLR protocol. Table 8 shows the intraprotocol interrater agreement results for Kendall’s W for all protocols, including pooled data.

3.4. Rater Preference Between DLR and Conventionally Reconstructed Images

All four raters of the multidisciplinary team, comprising a neuroradiologist, a neurosurgeon, a neuro-oncologist, and a radiation oncologist (respectively, named rater 1 to 4), preferred DLR over CR images as a whole image set at 1.5 T and 3 T. The pooled data are shown in Figure 7. Individual ratings are shown in Supplementary Figure S1. Specifically, all raters (pooled) preferred FLAIRDLR over FLAIRCR (91% for 1.5 T, 95% for 3 T), T2DLR over T2CR (91% for 1.5 T, 69% for 3 T), T1DLR over T1CR (79% for 1.5 T, 93% for 3 T), and T1CEDLR over T1CECR (79% for 1.5 T, 95% for 3 T). Overall preference for the DLR protocol was 91% for 1.5 T and 97% for 3 T.

3.5. Quantitative Image Quality Assessment

Detailed IQM results are documented in Table 9 for 1.5 T, 3 T, and pooled data. IQM results demonstrate high agreement for the entire image when CR is used as the reference and DLR as the comparison for SSIM and MS-SSIM at 1.5 T and 3 T. Regarding image quality metrics, both the Structural Similarity Index and Multi-Scale SSIM (on a 0–1 scale) were close to 1 for the entire image, indicating near-perfect similarity for both 1.5 and 3 T, despite the different sequence types for T1-weighted sequences at 1.5 T (SE for CR vs. TSE for DLR). The SSIM achieves a value of 0.9988 ± 0.0006 for FLAIRpooled, 0.9947 ± 0.0032 for T2pooled, 0.9882 ± 0.0047 for T1pooled, and 0.9838 ± 0.0060 for T1CEpooled, while the MS-SSIM achieves values of 0.9999 ± 0.0001, 0.9991 ± 0.0006, 0.9986 ± 0.0007, and 0.9980 ± 0.0009, respectively.
FSIM performs worse in terms of consistency between CR and DLR. It was found to be 0.8337 ± 0.0298 for FLAIRpooled, 0.8284 ± 0.0441 for T2pooled, 0.7584 ± 0.0555 for T1pooled, and 0.7427 ± 0.0509 for T1CEpooled. At 1.5 T, FSIM was 0.6860 ± 0.0763 for T1 and 0.6599 ± 0.0685 for T1CE.
Comparison of DLR versus CR sequences using NQM, SNR, and PSNR consistently showed higher values for DLR images. These findings align with the improved subjective image quality ratings, resection cavity delineation, and diagnostic confidence. However, as with FSIM, discrepancies in the T1-weighted sequences at 1.5 T to 3 T have also been observed. These discrepancies have resulted in reduced values for NQM, SNR, and PSNR in the comparison between CR and DLR. NQM was found to be 14.6983 ± 2.0582 for FLAIRpooled, 14.2008 ± 2.8076 for T2pooled, 9.5293 ± 2.5095 for T1pooled, and 9.1932 ± 2.1332 for T1CEpooled. SNR and PSNR achieved values of 14.2934 ± 2.1457 and 67.5261± 2.9485 for the whole image for FLAIRpooled, 13.4269 ± 2.1329 and 59.1513 ± 2.8639 for T2pooled, 10.0246 ± 1.7514 and 58.9975 ± 3.8395 for T1pooled, and 9.9720 ± 1.6580 and 56.6507 ± 3.4399 for T1CEpooled.

3.6. Evaluation of Residual Tumor

Table 10 demonstrates the results of the assessment of the detection rate between CR and DLR images of postoperative residual tumor. The analysis of residual tumor tissue revealed that CR does not detect more lesions than the DLR protocol.
The intrarater visual evaluation of residual tumor was comparable at 3 T between DLR and CR. In one case, the neurosurgeon (rater 2) evaluated residual postoperative tumor tissue in DLR images, whereas this lesion was not perceptible and was not rated in the CR series. For 1.5 T, residual tumor tissue was detected in DLR images and not in the corresponding CR sequences in no case for the neuroradiologist (rater 1), in four cases by the neurosurgeon (rater 2), in one case by the neuro-oncologist (rater 3), and in two cases by the radiation oncologist (rater 4). In all other cases, no residual tumor lesion was detected exclusively by either protocol.

4. Discussion

Although DLR is entering routine practice, real-world evidence remains limited. Because evaluations of therapeutic efficacy and related therapeutic decisions in neuro-oncology are multidisciplinary, the successful adoption of DLR depends on acceptance across specialties, not radiology alone. A review of the extant literature reveals that most studies on the qualitative and diagnostic performance of novel DL MRI have focused exclusively on radiologists. Integrating interdisciplinary research and collaboration between radiologists and other specialties can yield artificial techniques more closely aligned with clinical requirements.
Consequently, the present study quantitatively evaluated the differences between DLR and CR. In addition, four experienced raters from different specialties (i.e., neuroradiologist, neurosurgeon, neurooncologist, and radiotherapist) were asked to score qualitative assessments, such as image quality in postoperative MRIs, to evaluate specialty-dependent preferences and their effect on perceived residual tumor.
DL-based reconstruction methods have shown promise in decreasing the total scan time compared to traditional acceleration methods on standardized imaging setups by up to 60% [25,26]. DLR reduced scan time by 30–53% at 1.5 T and 27–31% at 3 T without compromising image quality or diagnostic confidence—findings that mirror earlier neuroradiological and non-neuroradiological assessments [7,10,27,28,29,30,31,32,33,34,35,36,37]. Across all specialties, DLR outperformed CR on both subjective and objective image quality metrics, underscoring its clinical utility in postoperative imaging.
This study offers several potential advantages, notably the evaluation conducted by a diverse team of medical professionals. This multidisciplinary approach ensures that the findings remain relevant and applicable across the diverse specialties involved in neuro-oncological imaging and clinical decision making. Surgical, medical, and radiation oncologists rely heavily on multiple imaging modalities for accurate diagnosis and effective cancer treatment planning and implementation [38]. Collaboration among cancer subspecialists is essential, as no single specialist can work effectively in isolation. Integrating services is crucial for optimizing treatment outcomes, minimizing complications, and enhancing long-term patient outcomes. Our data suggest that DLR can shorten scans, improve image quality, and maintain—or even increase—diagnostic confidence in neuro-oncological imaging protocols, consistent with prior work [7,10]. Combining quantitative metrics with qualitative evaluations provides a robust and comprehensive understanding of the imaging technique.
Lee et al. demonstrated that radiologists’ preference for DLR varies with experience, attributable in part to sharper edge rendition and unfamiliar image appearance, among other factors [37]. All raters in our cohort, across specialties and sequences, consistently preferred DLR over CR in most cases and judged its image quality to be good to excellent. Since neuroradiologists and other specialties rely on imaging daily for therapy decisions, the implications of this concordant preference are clinically meaningful. All raters agreed that DLR produced superior image quality with better delineated resection cavities, highlighted small enhancing foci, and improved contrast on T1-weighted images. There is a variation in preference for T1-weighted sequences at 1.5 T and T2-weighted sequences at 3 T. T2DLR was preferred more often by all raters at 1.5 T than at 3 T, even though most still preferred DLR. Conversely, T1CR, both non- and contrast-enhanced, was preferred over T1DLR in some cases at 1.5 T, whereas this was not the case at 3 T. The differences between the sequences are noteworthy and warrant further investigation. This discrepancy could be due to several factors. For T1-weighted images at 1.5 T, this may be explained by the different sequence types (SE vs. TSE) [39].
Furthermore, prior research has highlighted that DLR images exhibit lower noise levels than images obtained through traditional full-scan techniques. This trait can make DLR images unnatural to seasoned radiologists [40,41]. The DLR algorithm tends to produce images where the internal structure (e.g., of the tumor) appears more uniform and has a softer contrast than conventional images. Additionally, individual preferences and familiarity with the types of images commonly seen in regular practice also play a role in these differing evaluations. The qualitative image quality rating is supported by the image quality metrics SNR and PSNR, which are better for DLR than CR. PSNR expresses the log-scaled ratio of the maximum signal intensity to the reconstruction error. Higher PSNR, therefore, indicates reduced noise and sharper delineation. This is in line with other studies that have also shown improved SNR for DLR images [42,43,44,45].
Regarding image quality metrics, both SSIM and MS-SSIM were close to 1.0 (on a 0–1 scale) at both field strengths, indicating high similarity—even though 1.5 T T1-weighted scans used different sequence types (SE for CR vs. TSE for DLR). SSIM compares a test image (DLR) to a reference image (CR) by evaluating local luminance, contrast, and structural patterns. By evaluating multiple resolution levels and using an advanced pooling strategy, MS-SSIM further refines SSIM’s local measurements of image fidelity. The FSIM incorporates low-level features such as phase congruency and gradient magnitude and ranged from 0.824 to 0.848 at 3 T and from 0.660 to 0.842 at 1.5 T (on a 0–1 scale), with lower values for the 1.5 T T1-weighted images, likely reflecting sequence differences (SE for CR vs. TSE for DLR). This comparatively lower value underscores FSIM’s heightened sensitivity to subtle image differences. By typically enhancing high-frequency structure and edge sharpness, DLR boosts phase congruency. Thus, FSIM can vary markedly even when SSIM and MS-SSIM show only minor differences. The NQM index similarly accounts for luminance and contrast variations but also integrates the influence of spatial frequencies, viewing distance, and contrast masking. Similar effects for T1-weighted sequences at 1.5 T are also seen for NQM, with values of 3.1430 for contrast-enhanced T1-weighted sequences and 4.2984 for native T1-weighted sequences compared to 14.3721 for FLAIR and 14.1331 for T2-weighted TSE sequences. Although the NQM at 1.5 T is lower than FLAIR and the T2-weighted TSE sequence, a positive value still indicates improved image quality in DLR. Mason et al. reported that both FSIM and NQM showed consistently high correlations with radiologists’ subjective assessments of image quality, highlighting their effectiveness in capturing perceptual differences [20]. We could not verify this result in our study, which had considerably fewer datasets and only two comparison images.
The immediate postoperative MRI scan, obtained within 48 h of surgery, has been recommended by the European Association of Neuro-Oncology (EANO) and used, for example, as the baseline MRI in the Response Assessment in Neuro-Oncology (RANO) for newly diagnosed gliomas, including high-grade gliomas (RANO-HGG) and immunotherapy response assessment (iRANO) [17,46,47]. In contrast, the modified RANO (mRANO) criteria recommend using the first post-radiotherapy MRI as the baseline for newly diagnosed gliomas to minimize the effects of pseudoprogression and avoid challenges related to postoperative changes, corticosteroid variability, scan timing, and imaging techniques [48,49,50,51,52]. In support of this approach, Youssef et al. showed a greater correlation between progression-free survival and overall survival when the post-radiation scan was used as a baseline compared with the postoperative scan [53]. However, the difference was not statistically significant. If the patient is clinically stable, RANO 2.0 recommends a post-radiotherapy MRI around 4 weeks (21–35 days) after treatment as the baseline scan for newly diagnosed gliomas [54]. RANO 2.0 no longer treats the immediate postoperative scan as the response baseline; however, it remains valuable for detecting complications and gauging the extent of resection [55].
Accordingly, our findings demonstrate that an accelerated DLR protocol in the immediate postoperative setting is clinically relevant. Enhanced image quality may facilitate differentiation between actual residual tumor, postoperative change, and contrast leakage from the resection cavity. Evaluating how DLR influences individual assessments of residual tumor is essential for judging the comparability and interchangeability of reconstruction methods in routine practice. Importantly, no universally accepted imaging gold standard exists to define actual tumor extent or to determine whether a given reconstruction over- or underestimates it. Subjective comparisons between CR and DLR are crucial to ensure that DLR does not introduce spurious residual tumor findings. This study found that residual tumor was not detected in cases where DLR did not occur within the individual rater in the absence of CR alone. In the absence of these conditions, the detection of the residual tumor showed consistency among all raters, with DLR showing uniformity with CR images in the remaining cases at 3 T. This finding suggests that the residual tumor detected on DLR images was already visible on CR images, indicating that no additional findings were present. However, at both field strengths, there is variability in residual tumor detection between the disciplines. The reported non-detection of residual tumor is not the only issue; it may also reflect overinterpretation during evaluation, particularly by the neurosurgeon in four cases at 3 T (regardless of whether CR or DLR was used). At 1.5 T, visual assessment showed greater variability, with more residual tumor identified using DLR, especially by the neurosurgeon (n = 4). The neuro-oncologist identified one additional finding with DL images compared to the CR images, and the radiation oncologist identified two. This reflects the subjective nature of interpreting postoperative MRIs. In all other cases, DLR was comparable to CR in terms of the conspicuity of residual tumor. In summary, while DLR partially improved residual tumor perception for non-neuroradiologists, expert radiological interpretation—and hence downstream clinical decision making—was equivalent between DLR and CR.
The findings in this study should be interpreted with multiple limitations in mind. Firstly, this study included a relatively limited cohort of 33 patients from a single site, which was determined by power analysis. Despite this limitation, image quality and diagnostic confidence were significantly better for DLR among all patients. Additional patients from multiple sites would be needed to further validate the study’s generalizability. This is particularly evident and important in assessing residual tumor in individual disciplines at 1.5 T, presumably due to the better signal from the DLR, which might lead to a different assessment by those less experienced in imaging. The accelerated T1-weighted TSE sequence offers suboptimal gray-white contrast, but this is acceptable for the intended task of depicting enhancing lesions [39]. Secondly, an objective reference standard to confirm the presence of residual tumor is missing. Confirming residual tumor intraoperatively—particularly in low-grade gliomas—can be challenging, which is why some centers occasionally employ intraoperative MRI in selected brain-tumor cases. Furthermore, early post-therapy scans are deprioritized in mRANO and RANO 2.0 due to pseudoprogression and timing confounders. These sources would likely have added bias rather than resolved ground-truth uncertainty in this immediate postoperative cohort. Consequently, we concentrated on pre- versus immediate postoperative imaging to capture how different specialists perceive residual tumor—one of the primary aims of this study. Thirdly, focusing exclusively on postoperative MRIs may introduce selection bias, but it allowed us to isolate DLR performance in this clinically relevant cohort [11,56,57]. Furthermore, it is essential to highlight that CR images were set as the baseline and reference for the quantitative analysis. However, no universally accepted imaging standard definitively determines the actual size of tumors or residual tumor and which reconstruction technique potentially over- or underestimates residual tumor. Fourthly, qualitative assessments were conducted exclusively by a multidisciplinary team of experienced raters. This study did not include evaluations by inexperienced raters or comparisons between experienced and inexperienced raters, which could have provided additional insights. Qualitative evaluations are inherently susceptible to individual biases, which can potentially affect the reliability of the findings. We intentionally limited the present analysis to subspecialty-trained neuroradiologists and other senior clinicians to better isolate the technical performance of DLR from rater-experience effects in the context of postoperative MRIs. Additional specialists from these medical disciplines are therefore necessary to mitigate individual tendencies. They would further strengthen the findings of this study, potentially affect the conclusion, and help to better differentiate between the raters. Fifthly, despite blinding, experts could likely infer the reconstruction type, potentially introducing recognition bias. Sixthly, CR and DLR sequences were acquired sequentially, one immediately after the other, which is currently necessary when setting CR as the ground truth. However, this approach introduces a risk of non-identical image alignment, particularly in cases where patient motion occurs between acquisitions. We did not randomize the acquisition order of CR and DLR sequences to minimize potential motion- and alignment-related bias. In the current study, we did not observe any relevant motion artifacts. Lastly, this study was limited to one manufacturer’s MRI scanners, which restricted the generalizability of the results to other settings. “True” reproducibility—i.e., across platforms from different manufacturers—cannot be realized within a single-centered setting.

5. Conclusions

An accelerated deep learning reconstruction (DLR) protocol for postoperative brain MRI proved clinically feasible, cutting acquisition time by 42% at 1.5 T and 29.6% at 3 T. Across all sequences (FLAIR, T2, T1, T1CE), DLR achieved higher qualitative and quantitative image quality scores than conventional reconstruction, and at 1.5 T it occasionally altered individual raters’ perception of residual tumor. Based on the results of this study, we believe DL-based reconstruction in postoperative MRI could lead to easier-to-interpret images in some cases. Validation in larger, dedicated postoperative cohorts is still required. Furthermore, our multidisciplinary approach indicates that not only neuroradiologists but also neurosurgeons, neuro-oncologists, and radiotherapists involved in treating and assessing treatment response preferred DLR over CR images. The strong preference for DLR sequences among all raters highlights the potential for these advanced imaging techniques to enhance diagnostic and therapeutic decision making in neuro-oncological practice and promote the adoption of this approach for implementation in clinical routine.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/diagnostics15151982/s1, Figure S1: Individual visual preferences in the review of paired magnetic resonance image sets following tumor resection at 1.5 T (A) and 3 T (B) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). CR = conventional reconstruction; DLR = deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; T2 = T2-weighted images; T1 = T1-weighted images. CE = contrast-enhanced; and T = Tesla.

Author Contributions

Conceptualization, C.R. (Christer Ruff), T.-K.H., F.P. and G.G.; validation, C.R. (Constantin Roder), D.F., F.P., L.Z., D.S., P.B. and G.G.; formal analysis, C.R. (Christer Ruff), T.-K.H., F.P., L.Z. and G.G.; investigation, C.R. (Christer Ruff), T.-K.H., C.R. (Constantin Roder), F.P., P.B. and G.G.; resources, U.E.; data curation, T.-K.H., P.B. and G.G.; writing—original draft, C.R. (Christer Ruff) and G.G.; writing—review and editing, C.R. (Christer Ruff), T.-K.H., C.R. (Constantin Roder), D.F., F.P., L.Z., D.S., P.B., U.E. and G.G.; visualization, C.R. (Christer Ruff) and G.G.; supervision, U.E. and G.G.; project administration, U.E.; methodology, T.-K.H., C.R. (Constantin Roder), D.F., F.P. and L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Ethics Committee of Tübingen (protocol code 549/2024BO2; 3 December 2024).

Informed Consent Statement

Patient consent was waived due to the retrospective nature of this study. All MRI images related to this study were anonymized.

Data Availability Statement

In order to safeguard the confidentiality of the participants, the data pertaining to this study are currently withheld from public access. The data can be shared upon request.

Acknowledgments

We acknowledge support from the Open Access Publication Fund of the University of Tuebingen.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNRContrast-to-noise Ratio
CRConventional Reconstruction
DLDeep Learning
DLRDeep Learning-based Reconstruction
FDAU.S. Food and Drug Administration
FLAIRFluid-attenuated Inversion Recovery
FSIMFeature Similarity Index
MS-SSIMMulti-Scale SSIM
NQMNoise Quality Metric
PATParallel Acquisition Technique
PSNRPeak SNR
SDStandard Deviation
SESpin Echo
SNRSignal-to-noise Ratio
SSIMStructural Similarity Index
TSETurbo Spin Echo

References

  1. Rykkje, A.M.; Carlsen, J.F.; Larsen, V.A.; Skjøth-Rasmussen, J.; Christensen, I.J.; Nielsen, M.B.; Poulsen, H.S.; Urup, T.H.; Hansen, A.E. Prognostic relevance of radiological findings on early postoperative MRI for 187 consecutive glioblastoma patients receiving standard therapy. Sci. Rep. 2024, 14, 10985. [Google Scholar] [CrossRef]
  2. INTERVAL-GB Collaborative; Neurology and Neurosurgery Interest Group (NANSIG); British Neurosurgical Trainee Research Collaborative (BNTRC). Imaging timing after surgery for glioblastoma: An evaluation of practice in Great Britain and Ireland (INTERVAL-GB)—A multi-centre, cohort study. J. Neuro-Oncol. 2024, 169, 517–529. [Google Scholar] [CrossRef]
  3. Garcia-Ruiz, A.; Naval-Baudin, P.; Ligero, M.; Pons-Escoda, A.; Bruna, J.; Plans, G.; Calvo, N.; Cos, M.; Majós, C.; Perez-Lopez, R. Precise enhancement quantification in post-operative MRI as an indicator of residual tumor impact is associated with survival in patients with glioblastoma. Sci. Rep. 2021, 11, 695. [Google Scholar] [CrossRef] [PubMed]
  4. Kessler, A.T.; Bhatt, A.A. Brain tumour post-treatment imaging and treatment-related complications. Insights Imaging 2018, 9, 1057–1075. [Google Scholar] [CrossRef]
  5. Kiesel, B.; Prihoda, R.; Borkovec, M.; Mercea, P.A.; Steindl, A.; Berghoff, A.S.; Furtner, J.; Leitner, J.; Roetzer, T.; Preusser, M.; et al. Postoperative Magnetic Resonance Imaging After Surgery of Brain Metastases: Analysis of Extent of Resection and Potential Risk Factors for Incomplete Resection. World Neurosurg. 2020, 143, e365–e373. [Google Scholar] [CrossRef]
  6. Srinivasan, S.; Dasgupta, A.; Chatterjee, A.; Baheti, A.; Engineer, R.; Gupta, T.; Murthy, V. The Promise of Magnetic Resonance Imaging in Radiation Oncology Practice in the Management of Brain, Prostate, and GI Malignancies. JCO Glob. Oncol. 2022, 8, e2100366. [Google Scholar] [CrossRef]
  7. Gohla, G.; Hauser, T.-K.; Bombach, P.; Feucht, D.; Estler, A.; Bornemann, A.; Zerweck, L.; Weinbrenner, E.; Ernemann, U.; Ruff, C. Speeding Up and Improving Image Quality in Glioblastoma MRI Protocol by Deep Learning Image Reconstruction. Cancers 2024, 16, 1287. [Google Scholar] [CrossRef]
  8. Plenge, E.; Poot, D.H.J.; Bernsen, M.; Kotek, G.; Houston, G.; Wielopolski, P.; van der Weerd, L.; Niessen, W.J.; Meijering, E. Super-resolution methods in MRI: Can they improve the trade-off between resolution, signal-to-noise ratio, and acquisition time? Magn. Reson. Med. 2012, 68, 1983–1993. [Google Scholar] [CrossRef]
  9. Armato, S.G.; Drukker, K.; Hadjiiski, L. AI in medical imaging grand challenges: Translation from competition to research benefit and patient care. Br. J. Radiol. 2023, 96, 20221152. [Google Scholar] [CrossRef]
  10. Ruff, C.; Bombach, P.; Roder, C.; Weinbrenner, E.; Artzner, C.; Zerweck, L.; Paulsen, F.; Hauser, T.-K.; Ernemann, U.; Gohla, G. Multidisciplinary quantitative and qualitative assessment of IDH-mutant gliomas with full diagnostic deep learning image reconstruction. Eur. J. Radiol. Open 2024, 13, 100617. [Google Scholar] [CrossRef]
  11. Hossain, M.B.; Shinde, R.K.; Oh, S.; Kwon, K.-C.; Kim, N. A Systematic Review and Identification of the Challenges of Deep Learning Techniques for Undersampled Magnetic Resonance Image Reconstruction. Sensors 2024, 24, 753. [Google Scholar] [CrossRef]
  12. Monsour, R.; Borkowski, A. Neuroimaging in the Era of Artificial Intelligence: Current Applications. Fed. Pr. 2022, 39 (Suppl. 1), S14–S20. [Google Scholar] [CrossRef] [PubMed]
  13. Philip, A.K.; Samuel, B.A.; Bhatia, S.; Khalifa, S.A.M.; El-Seedi, H.R. Artificial Intelligence and Precision Medicine: A New Frontier for the Treatment of Brain Tumors. Life 2022, 13, 24. [Google Scholar] [CrossRef] [PubMed]
  14. Wu, J.; Li, C.; Gensheimer, M.; Padda, S.; Kato, F.; Shirato, H.; Wei, Y.; Schönlieb, C.-B.; Price, S.J.; Jaffray, D.; et al. Radiological tumor classification across imaging modality and histology. Nat. Mach. Intell. 2021, 3, 787–798. [Google Scholar] [CrossRef] [PubMed]
  15. Zheng, Y.; Carrillo-Perez, F.; Pizurica, M.; Heiland, D.H.; Gevaert, O. Spatial cellular architecture predicts prognosis in glioblastoma. Nat. Commun. 2023, 14, 4122. [Google Scholar] [CrossRef]
  16. Ottesen, J.A.; Storas, T.; Vatnehol, S.A.S.; Løvland, G.; Vik-Mo, E.O.; Schellhorn, T.; Skogen, K.; Larsson, C.; Bjørnerud, A.; Groote-Eindbaas, I.R.; et al. Deep learning-based Intraoperative MRI reconstruction. Eur. Radiol. Exp. 2025, 9, 29. [Google Scholar] [CrossRef]
  17. Weller, M.; van den Bent, M.; Preusser, M.; Le Rhun, E.; Tonn, J.C.; Minniti, G.; Bendszus, M.; Balana, C.; Chinot, O.; Dirven, L.; et al. EANO guidelines on the diagnosis and treatment of diffuse gliomas of adulthood. Nat. Rev. Clin. Oncol. 2021, 18, 170–186. [Google Scholar] [CrossRef]
  18. Albert, F.K.; Forsting, M.; Sartor, K.; Adams, H.-P.; Kunze, S. Early postoperative magnetic resonance imaging after resection of malignant glioma: Objective evaluation of residual tumor and its influence on regrowth and prognosis. Neurosurgery 1994, 34, 45–61. [Google Scholar] [CrossRef]
  19. Henegar, M.M.; Moran, C.J.; Silbergeld, D.L. Early postoperative magnetic resonance imaging following nonneoplastic cortical resection. J. Neurosurg. 1996, 84, 174–179. [Google Scholar] [CrossRef]
  20. Mason, A.; Rioux, J.; Clarke, S.E.; Costa, A.; Schmidt, M.; Keough, V.; Huynh, T.; Beyea, S. Comparison of Objective Image Quality Metrics to Expert Radiologists’ Scoring of Diagnostic Quality of MR Images. IEEE Trans. Med. Imaging 2020, 39, 1064–1072. [Google Scholar] [CrossRef]
  21. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale Structural Similarity for Image Quality Assessment. In Proceedings of the 37th IEEE Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 9–12 November 2003. [Google Scholar]
  23. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A feature similarity index for image quality assessment. IEEE Trans. Image Process 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed]
  24. Damera-Venkata, N.; Kite, T.D.; Geisler, W.S.; Evans, B.L.; Bovik, A.C. Image quality assessment based on a degradation model. IEEE Trans. Image Process 2000, 9, 636–650. [Google Scholar] [CrossRef] [PubMed]
  25. Bash, S.; Wang, L.; Airriess, C.; Zaharchuk, G.; Gong, E.; Shankaranarayanan, A.; Tanenbaum, L. Deep Learning Enables 60% Accelerated Volumetric Brain MRI While Preserving Quantitative Performance: A Prospective, Multicenter, Multireader Trial. Am. J. Neuroradiol. 2021, 42, 2130–2137. [Google Scholar] [CrossRef]
  26. Rudie, J.D.; Gleason, T.; Barkovich, M.J.; Wilson, D.M.; Shankaranarayanan, A.; Zhang, T.; Wang, L.; Gong, E.; Zaharchuk, G.; Villanueva-Meyer, J.E. Clinical Assessment of Deep Learning-based Super—Resolution for 3D Volumetric Brain MRI. Radiol. Artif. Intell. 2022, 4, e210059. [Google Scholar] [CrossRef]
  27. Almansour, H.; Herrmann, J.; Gassenmaier, S.; Afat, S.; Jacoby, J.; Koerzdoerfer, G.; Nickel, D.; Mostapha, M.; Nadar, M.; Othman, A.E. Deep Learning Reconstruction for Accelerated Spine MRI: Prospective Analysis of Interchangeability. Radiology 2023, 306, e212922. [Google Scholar] [CrossRef]
  28. Gassenmaier, S.; Afat, S.; Nickel, M.D.; Mostapha, M.; Herrmann, J.; Almansour, H.; Nikolaou, K.; Othman, A.E. Accelerated T2-Weighted TSE Imaging of the Prostate Using Deep Learning Image Reconstruction: A Prospective Comparison with Standard T2-Weighted TSE Imaging. Cancers 2021, 13, 3593. [Google Scholar] [CrossRef]
  29. Wessling, D.; Herrmann, J.; Afat, S.; Nickel, D.; Othman, A.E.; Almansour, H.; Gassenmaier, S. Reduction in Acquisition Time and Improvement in Image Quality in T2-Weighted MR Imaging of Musculoskeletal Tumors of the Extremities Using a Novel Deep Learning-Based Reconstruction Technique in a Turbo Spin Echo (TSE) Sequence. Tomography 2022, 8, 1759–1769. [Google Scholar] [CrossRef]
  30. Estler, A.; Hauser, T.-K.; Mengel, A.; Brunnée, M.; Zerweck, L.; Richter, V.; Zuena, M.; Schuhholz, M.; Ernemann, U.; Gohla, G. Deep Learning Accelerated Image Reconstruction of Fluid-Attenuated Inversion Recovery Sequence in Brain Imaging: Reduction of Acquisition Time and Improvement of Image Quality. Acad. Radiol. 2024, 31, 180–186. [Google Scholar] [CrossRef]
  31. Kim, M.; Kim, H.S.; Kim, H.J.; Park, J.E.; Park, S.Y.; Kim, Y.-H.; Kim, S.J.; Lee, J.; Lebel, M.R. Thin-Slice Pituitary MRI with Deep Learning-based Reconstruction: Diagnostic Performance in a Postoperative Setting. Radiology 2021, 298, 114–122. [Google Scholar] [CrossRef]
  32. Lee, D.H.; Park, J.E.; Nam, Y.K.; Lee, J.; Kim, S.; Kim, Y.-H.; Kim, H.S. Deep learning-based thin-section MRI reconstruction improves tumour detection and delineation in pre- and post-treatment pituitary adenoma. Sci. Rep. 2021, 11, 21302. [Google Scholar] [CrossRef]
  33. Rastogi, A.; Brugnara, G.; Foltyn-Dumitru, M.; Mahmutoglu, M.A.; Preetha, C.J.; Kobler, E.; Pflüger, I.; Schell, M.; Deike-Hofmann, K.; Kessler, T.; et al. Deep-learning-based reconstruction of undersampled MRI to reduce scan times: A multicentre, retrospective, cohort study. Lancet Oncol. 2024, 25, 400–410. [Google Scholar] [CrossRef] [PubMed]
  34. Estler, A.; Hauser, T.-K.; Brunnée, M.; Zerweck, L.; Richter, V.; Knoppik, J.; Örgel, A.; Bürkle, E.; Adib, S.D.; Hengel, H.; et al. Deep learning-accelerated image reconstruction in back pain-MRI imaging: Reduction of acquisition time and improvement of image quality. Radiol. Med. 2024, 129, 478–487. [Google Scholar] [CrossRef] [PubMed]
  35. Gassenmaier, S.; Afat, S.; Nickel, D.; Mostapha, M.; Herrmann, J.; Othman, A.E. Deep learning—accelerated T2-weighted imaging of the prostate: Reduction of acquisition time and improvement of image quality. Eur J Radiol 2021, 137, 109600. [Google Scholar] [CrossRef] [PubMed]
  36. Gassenmaier, S.; Afat, S.; Nickel, D.; Kannengiesser, S.; Herrmann, J.; Hoffmann, R.; Othman, A.E. Application of a Novel Iterative Denoising and Image Enhancement Technique in T1-Weighted Precontrast and Postcontrast Gradient Echo Imaging of the Abdomen: Improvement of Image Quality and Diagnostic Confidence. Investig. Radiol. 2021, 56, 328–334. [Google Scholar] [CrossRef]
  37. Lee, E.J.; Hwang, J.; Park, S.; Bae, S.H.; Lim, J.; Chang, Y.-W.; Hong, S.S.; Oh, E.; Da Nam, B.; Jeong, J.; et al. Utility of accelerated T2-weighted turbo spin-echo imaging with deep learning reconstruction in female pelvic MRI: A multi-reader study. Eur. Radiol. 2023, 33, 7697–7706. [Google Scholar] [CrossRef]
  38. Gaudino, S.; Giordano, C.; Magnani, F.; Cottonaro, S.; Infante, A.; Sabatino, G.; La Rocca, G.; Della Pepa, G.M.; D’alessandris, Q.G.; Pallini, R.; et al. Neuro-Oncology Multidisciplinary Tumor Board: The Point of View of the Neuroradiologist. J. Pers. Med. 2022, 12, 135. [Google Scholar] [CrossRef]
  39. Sugahara, T.; Korogi, Y.; Ge, Y.; Shigematsu, Y.; Liang, L.; Yoshizumi, K.; Kitajima, M.; Takahashi, M. Contrast enhancement of intracranial lesions: Conventional T1-weighted spin-echo versus fast spin-echo MR imaging techniques. Am. J. Neuroradiol. 1999, 20, 1554–1559. [Google Scholar]
  40. Recht, M.P.; Zbontar, J.; Sodickson, D.K.; Knoll, F.; Yakubova, N.; Sriram, A.; Murrell, T.; Defazio, A.; Rabbat, M.; Rybak, L.; et al. Using Deep Learning to Accelerate Knee MRI at 3 T: Results of an Interchangeability Study. Am. J. Roentgenol. 2020, 215, 1421–1429. [Google Scholar] [CrossRef]
  41. Antun, V.; Renna, F.; Poon, C.; Adcock, B.; Hansen, A.C. On instabilities of deep learning in image reconstruction and the potential costs of AI. Proc. Natl. Acad. Sci. USA 2020, 117, 30088–30095. [Google Scholar] [CrossRef]
  42. Suh, P.S.; Park, J.E.; Roh, Y.H.; Kim, S.; Jung, M.; Koo, Y.S.; Lee, S.-A.; Choi, Y.; Kim, H.S. Improving Diagnostic Performance of MRI for Temporal Lobe Epilepsy With Deep Learning-Based Image Reconstruction in Patients With Suspected Focal Epilepsy. Korean J. Radiol. 2024, 25, 374–383. [Google Scholar] [CrossRef]
  43. Lee, K.L.; Kessler, D.A.; Dezonie, S.; Chishaya, W.; Shepherd, C.; Carmo, B.; Graves, M.J.; Barrett, T. Assessment of deep learning-based reconstruction on T2-weighted and diffusion-weighted prostate MRI image quality. Eur. J. Radiol. 2023, 166, 111017. [Google Scholar] [CrossRef] [PubMed]
  44. Allen, T.J.; Bancroft, L.C.H.; Unal, O.; Estkowski, L.D.; Cashen, T.A.; Korosec, F.; Strigel, R.M.; Kelcz, F.; Fowler, A.M.; Gegios, A.; et al. Evaluation of a Deep Learning Reconstruction for High-Quality T2-Weighted Breast Magnetic Resonance Imaging. Tomography 2023, 9, 1949–1964. [Google Scholar] [CrossRef] [PubMed]
  45. Kaniewska, M.; Deininger-Czermak, E.; Getzmann, J.M.; Wang, X.; Lohezic, M.; Guggenberger, R. Application of deep learning-based image reconstruction in MR imaging of the shoulder joint to improve image quality and reduce scan time. Eur. Radiol. 2023, 33, 1513–1525. [Google Scholar] [CrossRef]
  46. Wen, P.Y.; Macdonald, D.R.; Reardon, D.A.; Cloughesy, T.F.; Sorensen, A.G.; Galanis, E.; DeGroot, J.; Wick, W.; Gilbert, M.R.; Lassman, A.B.; et al. Updated response assessment criteria for high-grade gliomas: Response assessment in neuro-oncology working group. J. Clin. Oncol. 2010, 28, 1963–1972. [Google Scholar] [CrossRef] [PubMed]
  47. Okada, H.; Weller, M.; Huang, R.; Finocchiaro, G.; Gilbert, M.R.; Wick, W.; Ellingson, B.M.; Hashimoto, N.; Pollack, I.F.; Brandes, A.A.; et al. Immunotherapy response assessment in neuro-oncology: A report of the RANO working group. Lancet Oncol. 2015, 16, e534–e542. [Google Scholar] [CrossRef]
  48. Ellingson, B.M.; Wen, P.Y.; Cloughesy, T.F. Modified Criteria for Radiographic Response Assessment in Glioblastoma Clinical Trials. Neurotherapeutics 2017, 14, 307–320. [Google Scholar] [CrossRef]
  49. Brandes, A.A.; Franceschi, E.; Tosoni, A.; Blatt, V.; Pession, A.; Tallini, G.; Bertorelle, R.; Bartolini, S.; Calbucci, F.; Andreoli, A.; et al. MGMT promoter methylation status can predict the incidence and outcome of pseudoprogression after concomitant radiochemotherapy in newly diagnosed glioblastoma patients. J. Clin. Oncol. 2008, 26, 2192–2197. [Google Scholar] [CrossRef]
  50. Hagiwara, A.; Schlossman, J.; Shabani, S.; Raymond, C.; Tatekawa, H.; Abrey, L.E.; Garcia, J.; Chinot, O.; Saran, F.; Nishikawa, R.; et al. Incidence, molecular characteristics, and imaging features of “clinically-defined pseudoprogression” in newly diagnosed glioblastoma treated with chemoradiation. J. Neurooncol 2022, 159, 509–518. [Google Scholar] [CrossRef]
  51. Wick, W.; Chinot, O.L.; Bendszus, M.; Mason, W.; Henriksson, R.; Saran, F.; Nishikawa, R.; Revil, C.; Kerloeguen, Y.; Cloughesy, T. Evaluation of pseudoprogression rates and tumor progression patterns in a phase III trial of bevacizumab plus radiotherapy/temozolomide for newly diagnosed glioblastoma. Neuro-Oncology 2016, 18, 1434–1441. [Google Scholar] [CrossRef]
  52. Taal, W.; Brandsma, D.; De Bruin, H.G.; Bromberg, J.E.C.; Swaak-Kragten, A.T.; Smitt, P.A.E.S.; Van Es, C.A.; Van Den Bent, M. Incidence of early pseudo-progression in a cohort of malignant glioma patients treated with chemoirradiation with temozolomide. Cancer 2008, 113, 405–410. [Google Scholar] [CrossRef]
  53. Youssef, G.; Rahman, R.; Bay, C.; Wang, W.; Lim-Fat, M.J.; Arnaout, O.; Bi, W.L.; Cagney, D.N.; Chang, Y.-S.; Cloughesy, T.F.; et al. Evaluation of Standard Response Assessment in Neuro-Oncology, Modified Response Assessment in Neuro-Oncology, and Immunotherapy Response Assessment in Neuro-Oncology in Newly Diagnosed and Recurrent Glioblastoma. J. Clin. Oncol. 2023, 41, 3160–3171. [Google Scholar] [CrossRef]
  54. Wen, P.Y.; Bent, M.v.D.; A Vogelbaum, M.; Chang, S.M. RANO 2.0: The revised Response Assessment in Neuro-Oncology (RANO) criteria for high- and low-grade glial tumors in adults designed for the future. Neuro-oncology 2024, 26, 2–4. [Google Scholar] [CrossRef] [PubMed]
  55. Karschnia, P.; Young, J.S.; Dono, A.; Häni, L.; Sciortino, T.; Bruno, F.; Juenger, S.T.; Teske, N.; A Morshed, R.; Haddad, A.F.; et al. Prognostic validation of a new classification system for extent of resection in glioblastoma: A report of the RANO resect group. Neuro-Oncology 2023, 25, 940–954. [Google Scholar] [CrossRef] [PubMed]
  56. Noordman, C.R.; Yakar, D.; Bosma, J.; Simonis, F.F.J.; Huisman, H. Complexities of deep learning-based undersampled MR image reconstruction. Eur. Radiol. Exp. 2023, 7, 58. [Google Scholar] [CrossRef] [PubMed]
  57. Zeng, G.; Guo, Y.; Zhan, J.; Wang, Z.; Lai, Z.; Du, X.; Qu, X.; Guo, D. A review on deep learning MRI reconstruction without fully sampled k-space. BMC Med. Imaging 2021, 21, 195. [Google Scholar] [CrossRef]
Figure 1. Study flowchart and patient enrollment.
Figure 1. Study flowchart and patient enrollment.
Diagnostics 15 01982 g001
Figure 2. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with fluid-attenuated inversion recovery (FLAIR) images demonstrates non-inferior image quality and visualization of the resection defect, in addition to reduced image noise, compared to conventional reconstruction (CR). The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; and CE = contrast-enhanced.
Figure 2. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with fluid-attenuated inversion recovery (FLAIR) images demonstrates non-inferior image quality and visualization of the resection defect, in addition to reduced image noise, compared to conventional reconstruction (CR). The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; and CE = contrast-enhanced.
Diagnostics 15 01982 g002
Figure 3. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with T2-weighted images demonstrates non-inferior image quality and visualization of the resection defect, in addition to reduced image noise, compared to conventional reconstruction (CR). Note the DLR algorithm’s tendency to produce images of the tumor’s internal structure that appear more uniform and softer than CR images. The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique.
Figure 3. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with T2-weighted images demonstrates non-inferior image quality and visualization of the resection defect, in addition to reduced image noise, compared to conventional reconstruction (CR). Note the DLR algorithm’s tendency to produce images of the tumor’s internal structure that appear more uniform and softer than CR images. The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique.
Diagnostics 15 01982 g003
Figure 4. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with T1-weighted images demonstrates non-inferior image quality and visualization of the resection defect, in addition to reduced image noise, compared to conventional reconstruction (CR). According to the standard clinical protocol, spin-echo sequences are used at 1.5 T for the conventionally reconstructed sequences. In contrast, turbo spin-echo sequences are used for the DLR boost for technical reasons. The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique; SSIM = Structural Similarity Index; and MS-SSIM = Multi-Scale SSIM.
Figure 4. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with T1-weighted images demonstrates non-inferior image quality and visualization of the resection defect, in addition to reduced image noise, compared to conventional reconstruction (CR). According to the standard clinical protocol, spin-echo sequences are used at 1.5 T for the conventionally reconstructed sequences. In contrast, turbo spin-echo sequences are used for the DLR boost for technical reasons. The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique; SSIM = Structural Similarity Index; and MS-SSIM = Multi-Scale SSIM.
Diagnostics 15 01982 g004
Figure 5. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with contrast-enhanced T1-weighted images (T1CE) demonstrates non-inferior image quality and visualization of the resection defect, as well as reduced image noise, compared to conventional reconstruction (CR). The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique; CE = contrast-enhanced; SSIM = Structural Similarity Index; and MS-SSIM = Multi-Scale SSIM.
Figure 5. Postoperative imaging of a 55-year-old female patient with a glioblastoma (1.5 T) and an 80-year-old female patient with two histologically confirmed uterine carcinoma metastases in the left frontal and parietal lobes (3 T). The utilization of deep learning reconstruction (DLR) in conjunction with contrast-enhanced T1-weighted images (T1CE) demonstrates non-inferior image quality and visualization of the resection defect, as well as reduced image noise, compared to conventional reconstruction (CR). The Structural Similarity Index (SSIM) and Multi-Scale SSIM (MS-SSIM) are utilized to quantitatively assess the preservation of anatomical details in DLR images compared to CR images as reference images. CR: conventional reconstruction; DLR: deep learning reconstructed technique; CE = contrast-enhanced; SSIM = Structural Similarity Index; and MS-SSIM = Multi-Scale SSIM.
Diagnostics 15 01982 g005
Figure 6. Multidisciplinary rating of overall image quality and diagnostic confidence of deep learning reconstructed images (DLR) and conventional recorded (CR) images by four experienced raters in a pooled data analysis for 1.5 T ((A), n = 17) and 3 T ((B), n = 16). Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). Likert scale ranging from 1 to 5, with 5 being the best rating. Significant differences are indicated. The dotted lines represent the median. CR = conventional reconstruction; DLR = deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; T2 = T2-weighted images; T1 = T1-weighted images. CE = contrast-enhanced; and T = Tesla.
Figure 6. Multidisciplinary rating of overall image quality and diagnostic confidence of deep learning reconstructed images (DLR) and conventional recorded (CR) images by four experienced raters in a pooled data analysis for 1.5 T ((A), n = 17) and 3 T ((B), n = 16). Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). Likert scale ranging from 1 to 5, with 5 being the best rating. Significant differences are indicated. The dotted lines represent the median. CR = conventional reconstruction; DLR = deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; T2 = T2-weighted images; T1 = T1-weighted images. CE = contrast-enhanced; and T = Tesla.
Diagnostics 15 01982 g006
Figure 7. Pooled visual preferences in the review of paired magnetic resonance image sets following tumor resection at 1.5 T ((A), n = 17) and 3 T ((B), n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). CR = conventional reconstruction; DLR = deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; T2 = T2-weighted images; T1 = T1-weighted images. CE = contrast-enhanced; and T = Tesla.
Figure 7. Pooled visual preferences in the review of paired magnetic resonance image sets following tumor resection at 1.5 T ((A), n = 17) and 3 T ((B), n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). CR = conventional reconstruction; DLR = deep learning reconstructed technique; FLAIR = fluid-attenuated inversion recovery images; T2 = T2-weighted images; T1 = T1-weighted images. CE = contrast-enhanced; and T = Tesla.
Diagnostics 15 01982 g007
Table 1. MRI acquisition parameters at 1.5 Tesla.
Table 1. MRI acquisition parameters at 1.5 Tesla.
ParametersFLAIRCRFLAIRDLRT2(TSE)CRT2(TSE)DLRT1(SE)CRT1(TSE)DLRT1(SE)CECRT1(TSE)CEDLR
Field of view (mm)230230170170230230230230
Voxel size (mm)0.4 × 0.4 × 4.00.4 × 0.4 × 4.00.3 × 0.3 × 3.00.3 × 0.3 × 3.00.4 × 0.4 × 4.00.4 × 0.4 × 4.00.4 × 0.4 × 4.00.4 × 0.4 × 4.0
Slice thickness (mm)44444444
Number of slices3636343436363636
Base resolution256256256256256256256256
Parallel imaging factorn.a.4n.a.4n.a.3n.a.3
Acceleration modeNoneGRAPPANoneGRAPPANoneGRAPPANoneGRAPPA
TR (ms)8800880035503550450600450600
TE (ms)118118104104129.6129.6
Averages11221212
Concatenations22222323
Acquisition time (min)2:581:232:451:232:281:432:281:43
Time savings using DLR
sequences in min (%)
1:35 (53%) 1:22 (50%) 0:45 (30%) 0:45 (30%)
Time saving on average
in min (%)
1:07 (42%)
Total time saving in min 4:27
FLAIR = fluid-attenuated inversion recovery; CR = conventional reconstruction; DLR = deep learning reconstructed technique; CE = contrast-enhanced; GRAPPA = GeneRalized Autocalibrating Partial Parallel Acquisition (parallel imaging technique); TR = repetition time; TE = echo time; SE = spin echo; TSE = turbo spin echo; mm = millimeter; ms = millisecond; n.a. = not applicable; and min = minute.
Table 2. MRI acquisition parameters at 3 Tesla.
Table 2. MRI acquisition parameters at 3 Tesla.
ParametersFLAIRCRFLAIRDLRT2(TSE)CRT2(TSE)DLRT1(TSE)CRT1(TSE)DLRT1(TSE)CECRT1(TSE)CEDLR
Field of view (mm)230230230230230230230230
Voxel size (mm)0.4 × 0.4 × 4.00.4 × 0.4 × 4.00.2 × 0.2 × 3.00.2 × 0.2 × 3.00.4 × 0.4 × 4.00.4 × 0.4 × 4.00.4 × 0.4 × 4.00.4 × 0.4 × 4.0
Slice thickness (mm)44444444
Number of slices3636343436363636
Base resolution320320384384304304304304
Parallel imaging factor24242424
Acceleration modeGRAPPAGRAPPAGRAPPAGRAPPAGRAPPAGRAPPAGRAPPAGRAPPA
TR (ms)88008800764076402000200020002000
TE (ms)818187879.79.79.79.7
Averages11221111
Concatenations22112222
Acquisition time (min)2:401:572:111:361:541:191:541:19
Time savings using DLR
sequences in min (%)
0:43 (27%) 0:35 (27%) 0:35 (31%) 0:35 (31%)
Time saving on average
in min (%)
0:37 (29%)
Total time saving in min 2:28
FLAIR = fluid-attenuated inversion recovery; CR = conventional reconstruction; DLR = deep learning reconstructed technique; CE = contrast-enhanced; GRAPPA = GeneRalized Autocalibrating Partial Parallel Acquisition (parallel imaging technique); TR = repetition time; TE = echo time; TSE = turbo spin echo; mm = millimeter; ms = millisecond; and min = minute.
Table 3. Patient’s characteristics.
Table 3. Patient’s characteristics.
CharacteristicsValues
Number of examinationsn = 33
1.5 Tn = 17
3 Tn = 16
Age, mean ± standard deviation (years)53.9 ± 20.3
1.5 T56.4 ± 21.7
3 T51.1 ± 19.6
Sex (male vs. female)n = 14 (42%) vs. n = 19 (58%)
1.5 Tn = 8 (47%) vs. n = 9 (53%)
3 Tn = 6 (37%) vs. n = 10 (63%)
Structural epilepsyn = 9/33 (27%)
1.5 Tn = 3/17 (18%)
3 Tn = 6/16 (38%)
Tumor type (lesions), 1.5 T
Metastasis6
Subependymoma1
DNET1
Pilocytic astrocytoma1
Astrocytoma grade 21
Astrocytoma grade 41
Oligodendroglioma grade 32
Glioblastoma4
Tumor type (lesions), 3 T
Metastasis5
Ependymoma1
Hemangiopericytoma1
Ganglioglioma grade 11
Midline glioma grade 41
Astrocytoma grade 21
Glioblastoma6
Previous surgeriesn = 10/33 (30%)
1.5 Tn = 4/17 (24%)
3 Tn = 6/16 (38%)
Contrast-enhancing tumor partsn = 27/33 (82%)
1.5 Tn = 12/17 (71%)
3 Tn = 15/16 (94%)
Karnofsky performance scale index
(KPS in %), median [range]
80 [20–100]
1.5 T80 [40–90]
3 T80 [20–100]
ECOG performance status scale,
median [range]
1 [0–4]
1.5 T1 [0–4]
3 T1 [0–4]
ECOG = Eastern Cooperative Oncology Group; T = Tesla.
Table 4. Multidisciplinary qualitative rating of the overall image quality, the visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed FLAIR images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values are calculated using the Wilcoxon signed-rank test.
Table 4. Multidisciplinary qualitative rating of the overall image quality, the visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed FLAIR images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values are calculated using the Wilcoxon signed-rank test.
1.5 T (n = 17)3 T (n = 16)Pooled (n = 33)
FLAIRCRFLAIRDLR FLAIRCRFLAIRDLR FLAIRCRFLAIRDLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
FLAIRCR vs. FLAIRDLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
FLAIRCR vs. FLAIRDLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
FLAIRCR vs. FLAIRDLR
Overall image qualityRater 14 (4–5)4.24 ± 0.4375 (4.5–5)4.76 ± 0.4370.0034 (4–5)4.31 ± 0.6025 (5–5)4.88 ± 0.3420.0074 (4–5)4.27 ± 0.5175 (5–5)4.82 ± 0.392<0.001
Rater 24 (4–4.5)4.24 ± 0.4375 (5–5)4.94 ± 0.243<0.0014 (4–5)4.31 ± 0.4795 (5–5)4.94 ± 0.2500.0024 (4–5)4.27 ± 0.4525 (5–5)4.94 ± 0.242<0.001
Rater 34 (4–4)4.00 ± 0.5005 (4–5)4.53 ± 0.6240.0134 (4–4)4.00 ± 0.3655 (4–5)4.56 ± 0.6290.0074 (4–4)4.00 ± 0.4335 (4–5)4.55 ± 0.617<0.001
Rater 44 (4–4)4.18 ± 0.3935 (4–5)4.76 ± 0.4370.0024 (4–5)4.25 ± 0.5775 (4.25–5)4.75 ± 0.4470.0114 (4–4.5)4.21 ± 0.4855 (4.5–5)4.76 ± 0.435<0.001
All Raters4 (4–4)4.16 ± 0.4445 (5–5)4.75 ± 0.469<0.0014 (4–5)4.22 ± 0.5195 (5–5)4.75 ± 0.453<0.0014 (4–4)4.19 ± 0.4815 (5–5)4.77 ± 0.460<0.001
Visualization of the resection
cavity
Rater 14 (4–5)4.24 ± 0.5625 (4–5)4.53 ± 0.5140.0594 (4–5)4.25 ± 0.6834 (4–5)4.38 ± 0.5000.3174 (4–5)4.24 ± 0.6145 (4–5)4.45 ± 0.5060.035
Rater 24 (4–5)4.29 ± 0.4705 (4–5)4.65 ± 0.4930.0344.5 (4–5)4.44 ± 0.6295 (5–5)4.88 ± 0.3420.0884 (4–5)4.36 ± 0.5495 (4–4)4.76 ± 0.435<0.001
Rater 34 (3.5–4)3.82 ± 0.5294 (4–4.5)4.12 ± 0.6000.0594 (3–4)3.62 ± 0.5004 (4–4)3.94 ± 0.4430.0594 (3–4)3.73 ± 0.5174 (4–4)4.03 ± 0.5290.008
Rater 44 (4–4.5)4.18 ± 0.5294 (4–5)4.47 ± 0.5140.0254 (3.25–5)4.13 ± 0.8064 (4–5)4.31 ± 0.6020.1804 (4–5)4.15 ± 0.6674 (4–4)4.39 ± 0.5560.011
All Raters4 (4–4)4.13 ± 0.5444 (4–5)4.44 ± 0.557<0.0014 (4–5)4.11 ± 0.7154 (4–5)4.38 ± 0.577<0.0014 (4–5)4.12 ± 0.6304 (4–5)4.41 ± 0.556<0.001
Diagnostic
confidence
Rater 14 (4–5)4.47 ± 0.5145 (4–5)4.53 ± 0.5140.5644 (4–5)4.31 ± 0.4794.5 (4–5)4.50 ± 0.5160.1804 (4–5)4.39 ± 0.4965 (4–5)4.52 ± 0.5080.157
Rater 24 (4–5)4.29 ± 0.4704 (4–5)4.47 ± 0.5140.0835 (4–5)4.50 ± 0.6325 (5–5)4.88 ± 0.3420.0144 (4–5)4.39 ± 0.5565 (4–5)4.67 ± 0.4790.003
Rater 34 (3–4)3.59 ± 0.6184 (4–4)3.86 ± 0.4850.0254 (3–4)3.62 ± 0.5004 (3–4)3.75 ± 0.5770.3174 (3–4)3.61 ± 0.5664 (3.5–4)3.82 ± 0.5280.020
Rater 44 (4–5)4.47 ± 0.5144 (4–5)4.41 ± 0.5070.5644 (4–4.75)4.25 ± 0.4474 (4–5)4.44 ± 0.5120.1804 (4–5)4.36 ± 0.4894 (4–5)4.42 ± 0.5020.480
All Raters4 (4–5)4.21 ± 0.6364 (4–5)4.32 ± 0.5580.0334 (4–5)4.17 ± 0.6064 (4–5)4.39 ± 0.6630.0024 (4–5)4.19 ± 0.6194 (4–5)4.36 ± 0.594<0.001
CR = conventional reconstruction; DLR = deep learning reconstruction; Mdn = median; IQR = interquartile range; M = mean; SD = standard deviation; FLAIR = fluid-attenuated inversion recovery; and T = Tesla.
Table 5. Multidisciplinary qualitative rating of the overall image quality, the visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed T2-weighted images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values are calculated using the Wilcoxon signed-rank test.
Table 5. Multidisciplinary qualitative rating of the overall image quality, the visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed T2-weighted images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values are calculated using the Wilcoxon signed-rank test.
1.5 T (n = 17)3 T (n = 16)Pooled (n = 33)
T2CRT2DLR T2CRT2DLR T2CRT2DLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T2CR vs. T2DLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T2CR vs. T2DLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T2CR vs. T2DLR
Overall image qualityRater 14 (4–4)4.12 ± 0.3325 (4–5)4.65 ± 0.4930.0034 (4–5)4.38 ± 0.5004 (4–5)4.31 ± 0.4790.5654 (4–4.5)4.24 ± 0.4354 (4–5)4.48 ± 0.5080.021
Rater 24 (4–4)4.18 ± 0.3935 (5–5)4.94 ± 0.243<0.0015 (4.25–5)4.75 ± 0.4475 (5–5)4.94 ± 0.2500.0834 (4–5)4.45 ± 0.5065 (5–5)4.94 ± 0.242<0.001
Rater 34 (4–4)3.94 ± 0.2435 (4–5)4.53 ± 0.5140.0024 (4–4)4.06 ± 0.5744 (4–4)3.94 ± 0.5740.5274 (4–4)4.00 ± 0.4334 (4–5)4.24 ± 0.6140.074
Rater 44 (4–4)4.00 ± 0.3545 (4–5)4.59 ± 0.5070.0024 (4–5)4.31 ± 0.4794 (4–5)4.19 ± 0.6550.4144 (4–4)4.15 ± 0.4424 (4–5)4.39 ± 0.6090.046
All Raters4 (4–4)4.06 ± 0.3405 (4–5)4.68 ± 0.471<0.0014 (4–5)4.37 ± 0.5494 (4–5)4.34 ± 0.6230.6704 (4–4)4.21 ± 0.4795 (4–5)4.52 ± 0.573<0.001
Visualization of the resection
cavity
Rater 14 (4–5)4.35 ± 0.4935 (4–5)4.59 ± 0.5070.1024 (4–4.75)4.13 ± 0.6194 (4–4.75)4.19 ± 0.5440.5644 (4–5)4.24 ± 0.5614 (4–5)4.39 ± 0.5560.096
Rater 24 (4–5)4.35 ± 0.6065 (4.5–5)4.71 ± 0.5880.0345 (5–5)4.81 ± 0.4035 (5–5)4.94 ± 0.2500.1575 (4–5)4.58 ± 0.5615 (5–5)4.82 ± 0.4650.011
Rater 34 (3.5–4)3.76 ± 0.4374 (4–5)4.24 ± 0.5620.0054 (3–4)3.75 ± 0.6834 (4–4)3.87 ± 0.5000.4144 (3–4)3.76 ± 0.5614 (4–4)4.06 ± 0.5560.008
Rater 44 (4–5)4.24 ± 0.5625 (4–5)4.53 ± 0.5140.0594 (4–5)4.25 ± 0.6834 (4–4.75)4.19 ± 0.5440.5644 (4–5)4.24 ± 0.6144 (4–5)4.36 ± 0.5490.206
All Raters4 (4–5)4.18 ± 0.5715 (4–5)4.51 ± 0.560<0.0014 (4–5)4.23 ± 0.7074 (4–5)4.30 ± 0.6090.2854 (4–5)4.20 ± 0.6394 (4–5)4.41 ± 0.592<0.001
Diagnostic
confidence
Rater 15 (4–5)4.53 ± 0.6245 (4–5)4.65 ± 0.4930.4144 (4–5)4.37 ± 0.6194.5 (4–5)4.38 ± 0.71915 (4–5)4.45 ± 0.6175 (4–5)4.52 ± 0.6190.527
Rater 24 (4–5)4.35 ± 0.6065 (4–5)4.53 ± 0.6240.0835 (5–5)4.81 ± 0.4035 (5–5)4.94 ± 0.2500.1575 (4–5)4.58 ± 0.5615 (4.5–5)4.73 ± 0.5170.025
Rater 34 (3–4)3.71 ± 0.5884 (4–4)4.00 ± 0.5000.0254 (3–4)3.69 ± 0.6024 (3.25–4)3.88 ± 0.6190.1804 (3–4)3.70 ± 0.5854 (4–4)3.94 ± 0.5560.011
Rater 45 (4–5)4.53 ± 0.6245 (4–5)4.65 ± 0.4930.4144 (4–5)4.25 ± 0.5774 (4–5)4.31 ± 0.7040.6554 (4–5)4.39 ± 0.6095 (4–5)4.48 ± 0.6190.366
All Raters4 (4–5)4.28 ± 0.6884.5 (4–5)4.46 ± 0.5840.0074 (4–5)4.28 ± 0.6784.5 (4–5)4.38 ± 0.7010.1344 (4–5)4.28 ± 0.6804.5 (4–5)4.42 ± 0.6430.003
CR = conventional reconstruction; DLR = deep learning reconstruction; Mdn = median; IQR = interquartile range; M = mean; SD = standard deviation; T2 = T2-weighted images; and T = Tesla.
Table 6. Multidisciplinary qualitative rating of the overall image quality, the visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed T1-weighted images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values are calculated using the Wilcoxon signed-rank test.
Table 6. Multidisciplinary qualitative rating of the overall image quality, the visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed T1-weighted images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values are calculated using the Wilcoxon signed-rank test.
1.5 T (n = 17)3 T (n = 16)Pooled (n = 33)
T1CRT1DLR T1CRT1DLR T1CRT1DLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T1CR vs. T1DLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T1CR vs. T1DLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T1CR vs. T1DLR
Overall image
quality
Rater 14 (4–4)4.06 ± 0.2435 (4–5)4.71 ± 0.4700.0024 (4–4)4.13 ± 0.3425 (5–5)4.81 ± 0.403<0.0014 (4–4)4.09 ± 0.2925 (4.5–5)4.76 ± 0.435<0.001
Rater 24 (4–5)4.41 ± 0.5065 (4–5)4.65 ± 0.4930.3174 (4–4.75)4.25 ± 0.4475 (5–5)5.00<0.0014 (4–5)4.33 ± 0.4795 (5–5)4.82 ± 0.3920.002
Rater 34 (4–4)3.88 ± 0.4855 (4–5)4.41 ± 0.7120.0204 (3–4)3.56 ± 0.5124 (4–4)4.06 ± 0.5740.0214 (3–4)3.73 ± 0.5174 (4–5)4.24 ± 0.6630.001
Rater 44 (4–4)4.06 ± 0.2435 (5–5)4.82 ± 0.363<0.0014 (4–4)4.06 ± 0.4435 (4.25–5)4.75 ± 0.4470.0024 (4–4)4.06 ± 0.3485 (5–5)4.79 ± 0.425<0.001
All Raters4 (4–4)4.10 ± 0.4285 (4–5)4.65 ± 0.540<0.0014 (4–4)4.00 ± 0.5045 (4–5)4.66 ± 0.541<0.0014 (4–4)4.05 ± 0.4675 (4–5)4.65 ± 0.538<0.001
Visualization of
the resection
cavity
Rater 14 (4–4)4.00 ± 0.6125 (4–5)4.65 ± 0.4930.0024 (3–4.75)3.81 ± 0.8344 (3.25–5)4.13 ± 0.8060.0254 (3–4)3.91 ± 0.7235 (4–5)4.39 ± 0.704<0.001
Rater 24 (4–5)4.41 ± 0.5074 (4–5)4.47 ± 0.5140.6654 (4–4.75)4.19 ± 0.5445 (5–5)4.88 ± 0.342<0.0014 (4–5)4.30 ± 0.5295 (4–5)4.67 ± 0.4790.003
Rater 34 (4–4)3.88 ± 0.4854 (4–4.5)4.12 ± 0.6000.1023.5 (3–4)3.50 ± 0.5164 (3–4)3.81 ± 0.6550.0594 (3–4)3.70 ± 0.5294 (4–4)3.97 ± 0.6370.013
Rater 44 (4–4)4.00 ± 0.6125 (4–5)4.59 ± 0.507<0.0013.5 (3–4.75)3.75 ± 0.8564 (3–5)4.00 ± 0.8940.0464 (3–4)3.88 ± 0.7404 (4–5)4.30 ± 0.770<0.001
All Raters4 (4–4)4.07 ± 0.5814 (4–5)4.46 ± 0.558<0.0014 (3–4)3.81 ± 0.7324 (4–5)4.20 ± 0.800<0.0014 (3.25–4)3.95 ± 0.6694 (4–5)4.33 ± 0.695<0.001
Diagnostic
confidence
Rater 14 (4–5)4.41 ± 0.5074 (4–5)4.47 ± 0.5140.3174 (4–4.75)4.19 ± 0.5444 (4–5)4.38 ± 0.5000.0834 (4–5)4.30 ± 0.5294 (4–5)4.42 ± 0.5020.046
Rater 24 (4–5)4.41 ± 0.5074 (4–5)4.35 ± 0.4930.5644 (4–4.75)4.19 ± 0.5445 (5–5)4.88 ± 0.432<0.0014 (4–5)4.30 ± 0.5295 (4–5)4.61 ± 0.4960.008
Rater 34 (3–4)3.76 ± 0.5624 (4–4)4.06 ± 0.5560.0594 (3–4)3.69 ± 0.4794 (3–4)3.69 ± 0.60214 (3–4)3.73 ± 0.5174 (3.5–4)3.88 ± 0.6000.096
Rater 44 (4–5)4.41 ± 0.5074 (4–5)4.47 ± 0.5140.3174 (4–4)3.94 ± 0.5745 (4–5)4.56 ± 0.6290.0024 (4–5)4.18 ± 0.5845 (4–5)4.52 ± 0.566<0.001
All Raters4 (4–5)4.25 ± 0.5834 (4–5)4.34 ± 0.5360.0834 (4–4)4.00 ± 0.5634 (4–5)4.37 ± 0.678<0.0014 (4–4)4.13 ± 0.5854 (4–5)4.36 ± 0.607<0.001
CR = conventional reconstruction; DLR = deep learning reconstruction; Mdn = median; IQR = interquartile range; M = mean; SD = standard deviation; T1 = T1-weighted images; and T = Tesla.
Table 7. Multidisciplinary qualitative rating of the overall image quality, visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed contrast-enhanced T1-weighted images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values were calculated using the Wilcoxon signed-rank test.
Table 7. Multidisciplinary qualitative rating of the overall image quality, visualization of the resection cavity, and diagnostic confidence of conventional and deep learning reconstructed contrast-enhanced T1-weighted images at 1.5 Tesla (T, n = 17) and 3 T (n = 16) by four experienced raters. Raters consist of a neuroradiologist (Rater 1), a neurosurgeon (Rater 2), a neuro-oncologist (Rater 3), and a radiation oncologist (Rater 4). The p-values were calculated using the Wilcoxon signed-rank test.
1.5 T (n = 17)3 T (n = 16)Pooled (n= 33)
T1CECRT1CEDLR T1CECRT1CEDLR T1CECRT1CEDLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T1CECR vs. T1CEDLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T1CECR vs. T1CEDLR
Mdn (IQR)M ± SDMdn (IQR)M ± SDp-Value
T1CECR vs. T1CEDLR
Overall image qualityRater 14 (4–4)4.06 ± 0.2435 (5–5)4.82 ± 0.393<0.0014 (4–4)4.19 ± 0.4035 (4.25–5)4.75 ± 0.4470.0034 (4–4)4.12 ± 0.3315 (5–5)4.79 ± 0.415<0.001
Rater 24 (4–4)4.18 ± 0.3934 (4–5)4.47 ± 0.5120.0964 (4–4.75)4.25 ± 0.4475 (5–5)5.00<0.0014 (4–4)4.21 ± 0.4155 (4–5)4.73 ± 0.453<0.001
Rater 34 (4–4)4.06 ± 0.2435 (5–5)4.82 ± 0.393<0.0014 (3.25–4)3.75 ± 0.4474 (4–5)4.31 ± 0.6020.0074 (3–4)3.70 ± 0.4674 (4–5)4.30 ± 0.529<0.001
Rater 44 (3–4)3.65 ± 0.4934 (4–5)4.29 ± 0.470<0.0014 (4–4)4.13 ± 0.3425 (4–5)4.69 ± 0.4790.0034 (4–4)4.09 ± 0.2925 (4.5–5)4.76 ± 0.435<0.001
All Raters4 (4–4)3.99 ± 0.4055 (4–5)4.60 ± 0.493<0.0014 (4–4)4.08 ± 0.4475 (4–5)4.69 ± 0.500<0.0014 (4–4)4.03 ± 0.4295 (4–5)4.64 ± 0.496<0.001
Visualization of the resection
cavity
Rater 14 (4–4)4.06 ± 0.5565 (5–5)4.82 ± 0.393<0.0014 (4–5)4.19 ± 0.7504 (4–5)4.31 ± 0.7040.4144 (4–5)4.12 ± 0.6505 (4–5)4.58 ± 0.6140.001
Rater 24 (4–5)4.29 ± 0.4705 (4–5)4.59 ± 0.5070.0254 (4–4.75)4.25 ± 0.4475 (5–5)4.94 ± 0.250<0.0014 (4–5)4.21 ± 0.4525 (4.5–5)4.76 ± 0.435<0.001
Rater 34 (3–4)3.65 ± 0.6064 (4–4)3.12 ± 0.4850.0054 (3–4)3.69 ± 0.4794 (4–4)4.00 ± 0.6320.0594 (3–4)3.67 ± 0.5404 (4–4)4.06 ± 0.556<0.001
Rater 44 (4–4)4.06 ± 0.5565 (4.5–5)4.76 ± 0.437<0.0014 (3–5)4.06 ± 0.8544 (4–5)4.19 ± 0.7500.1574 (4–5)4.06 ± 0.7044 (5–5)4.48 ± 0.667<0.001
All Raters4 (4–4)4.01 ± 0.5865 (4–5)4.57 ± 0.527<0.0014 (4–4.75)4.05 ± 0.6774 (4–5)4.36 ± 0.698<0.0014 (4–4)4.03 ± 0.6295 (4–5)4.47 ± 0.623<0.001
Diagnostic
confidence
Rater 14 (4–5)4.41 ± 0.5075 (4.5–5)4.76 ± 0.4370.0144 (4–4.75)4.13 ± 0.6195 (4–5)4.56 ± 0.6290.0084 (4–5)4.27 ± 0.5745 (4–5)4.67 ± 0.540<0.001
Rater 24 (4–5)4.35 ± 0.4934 (4–5)4.47 ± 0.5140.1574 (4–4.75)4.25 ± 0.4475 (5–5)4.94 ± 0.250<0.0014 (4–5)4.30 ± 0.4675 (4–5)4.70 ± 0.467<0.001
Rater 34 (3–4)3.59 ± 0.6184 (4–4)4.00 ± 0.5000.0084 (3.2–-4)3.75 ± 0.4474 (4–4.75)4.06 ± 0.6800.0254 (3–4)3.67 ± 0.5404 (4–4)4.03 ± 0.585<0.001
Rater 44 (4–5)4.41 ± 0.5075 (4–5)4.71 ± 0.4700.0254 (4–4)4.06 ± 0.5745 (4–5)4.50 ± 0.6320.0084 (4–5)4.24 ± 0.5615 (4–5)4.61 ± 0.556<0.001
All Raters4 (4–5)4.19 ± 0.6295 (4–5)4.49 ± 0.560<0.0014 (4–4)4.05 ± 0.5475 (4–5)4.52 ± 0.642<0.0014 (4–4)4.12 ± 0.5935 (4–5)4.50 ± 0.599<0.001
CR = conventional reconstruction; DLR = deep learning reconstruction; Mdn = median; IQR = interquartile range; M = mean; SD = standard deviation; T1 = T1-weighted images; CE = contrast-enhanced; and T = Tesla.
Table 8. Mean (range) interrater intraprotocol agreement using Kendall’s W.
Table 8. Mean (range) interrater intraprotocol agreement using Kendall’s W.
Kendall’s W
1.5 T (n = 17)3 T (n = 16)Pooled (n = 33)
Overall image qualityCR0.225 (0.131–0.331)0.276 (0.126–0.360)0.238 (0.128–0.304)
FLAIRCR0.1330.1260.128
T2CR0.1310.3400.222
T1CR0.3030.3600.298
T1CECR0.3310.2790.304
Overall image qualityDLR0.231 (0.136–0.349)0.413 (0.135–0.605)0.264 (0.188–0.349)
FLAIRDLR0.2770.1350.188
T2DRL0.1620.6050.349
T1DLR0.1360.5870.293
T1CEDLR0.3490.3250.226
Visualization of the
resection cavityCR
0.271 (0.150–0.362)0.303 (0.173–0.546)0.262 (0.228–0.359)
FLAIRCR0.1500.3030.222
T2CR0.2720.5460.359
T1CR0.2980.1900.228
T1CECR0.3620.1730.237
Visualization of the
resection cavityDLR
0.280 (0.167–0.393)0.561 (0.484–0.730)0.329 (0.274–0.366)
FLAIRDLR0.2480.5020.359
T2DRL0.1670.7300.366
T1DLR0.3100.4840.274
T1CEDLR0.3930.5270.315
Diagnostic confidenceCR0.463 (0.368–0.560)0.463 (0.193–0.518)0.364 (0.313–0.405)
FLAIRCR0.4160.4530.405
T2CR0.3680.5180.360
T1CR0.5090.1930.313
T1CECR0.5600.1970.377
Diagnostic confidenceDLR0.361 (0.324–0.469)0.469 (0.394–0.576)0.354 (0.287–0.403)
FLAIRDLR0.3660.4950.403
T2DRL0.3240.3940.287
T1DLR0.2840.5760.365
T1CEDLR0.4690.4080.362
T = Tesla; FLAIR = fluid-attenuated inversion recovery; CR = conventional reconstruction; DLR = deep learning reconstruction; T2 = T2-weighted images; T1 = T1-weighted images; and CE = contrast-enhanced.
Table 9. Results of image quality metric (IQM) analysis between conventionally reconstructed and deep learning reconstructed MRI sequences.
Table 9. Results of image quality metric (IQM) analysis between conventionally reconstructed and deep learning reconstructed MRI sequences.
SequenceIQM1.5 T (n = 17)
Value (M ± SD)
3 T (n = 16)
Value (M ± SD)
Pooled (n = 33)
Value (M ± SD)
FLAIRSSIM0.9987 ± 0.00060.9989 ± 0.00060.9988 ± 0.0006
MS-SSIM0.9999 ± 0.00010.9999 ± 0.00010.9999 ± 0.0001
FSIM0.8420 ± 0.03160.8242 ± 0.02780.8337± 0.0298
NQM14.3721 ± 2.164515.0680 ± 1.937814.6983 ± 2.0582
SNR13.6127 ± 2.021415.0649 ± 2.286614.2934 ± 2.1457
PSNR67.2675 ± 3.248767.8192 ± 2.608367.5261± 2.9485
T2SSIM0.9965 ± 0.00290.9924 ± 0.00350.9947 ± 0.0032
MS-SSIM0.9995 ± 0.00040.9985 ± 0.00080.9991 ± 0.0006
FSIM0.8159 ± 0.04280.8440 ± 0.04580.8284 ± 0.0441
NQM14,1331 ± 2.843514.2798 ± 2.765714.2008 ± 2.8076
SNR13.8342 ± 2.207712.9179 ± 2.039513.4269 ± 2.1329
PSNR61.1301 ± 3.112756.6777 ± 2.552859.1513 ± 2.8639
T1SSIM0.9801 ± 0.00780.9982 ± 0.00100.9882 ± 0.0047
MS-SSIM0.9978 ± 0.00110.9997 ± 0.00020.9986 ± 0.0007
FSIM0.6860 ± 0.07630.8475 ± 0.02980.7584 ± 0.0555
NQM4.2984 ± 2.739615.9673 ± 2.22639.5293 ± 2.5095
SNR6.0580 ± 1.668414.9067 ± 1.853610.0246 ± 1.7514
PSNR55.3725 ± 3.721063.4592 ± 3.985458.9975 ± 3.8395
T1CESSIM0.9726 ± 0.00900.9957 ± 0.00280.9838 ± 0.0060
MS-SSIM0.9967 ± 0.00130.9994 ± 0.00040.9980 ± 0.0009
FSIM0.6599 ± 0.06850.8311 ± 0.03220.7427 ± 0.0509
NQM3.1430 ± 2.156215.6468 ± 2.10879.1932 ± 2.1332
SNR5.2700 ± 1.438614.9875 ± 1.89219.9720 ± 1.6580
PSNR52.6580 ± 3.093060.9097 ± 3.809956.6507 ± 3.4399
Data are presented as mean ± standard deviation. IQM = image quality metric; T = Tesla; M = mean; SD = Standard Deviation; SSIM = Structural Similarity Index; MS-SSIM = Multi-Scale SSIM; FSIM = Feature Similarity Index; NQM = Noise Quality Metric; SNR = signal-to-noise ratio; and PSNR = peak signal-to-noise ratio.
Table 10. Detection rate in absolute number of residual tumor following resection of intracranial tumors on 1.5 T (n = 17) and 3 T (n = 16) between DLR and CR protocols.
Table 10. Detection rate in absolute number of residual tumor following resection of intracranial tumors on 1.5 T (n = 17) and 3 T (n = 16) between DLR and CR protocols.
DLR > CRDLR = CRDLR < CR
Lesion1.5T (n = 17)Rater 1011/170
Rater 24/176/170
Rater 31/176/170
Rater 42/179/170
Lesion3T (n = 16)Rater 107/160
Rater 21/1610/160
Rater 303/160
Rater 406/160
Lesionpooled (n = 33)Rater 1018/330
Rater 25/3316/330
Rater 31/339/330
Rater 42/3315/330
DLR > CR, DLR protocol detects more lesions compared to CR protocol; DLR = CR, residual tumor tissue was detected in both DLR and CR protocols; DLR < CR, CR detects more lesions compared to DLR protocol. T = Tesla; CR = conventional reconstruction; DLR = deep learning reconstruction.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ruff, C.; Hauser, T.-K.; Roder, C.; Feucht, D.; Bombach, P.; Zerweck, L.; Staber, D.; Paulsen, F.; Ernemann, U.; Gohla, G. Multidisciplinary, Clinical Assessment of Accelerated Deep-Learning MRI Protocols at 1.5 T and 3 T After Intracranial Tumor Surgery and Their Influence on Residual Tumor Perception. Diagnostics 2025, 15, 1982. https://doi.org/10.3390/diagnostics15151982

AMA Style

Ruff C, Hauser T-K, Roder C, Feucht D, Bombach P, Zerweck L, Staber D, Paulsen F, Ernemann U, Gohla G. Multidisciplinary, Clinical Assessment of Accelerated Deep-Learning MRI Protocols at 1.5 T and 3 T After Intracranial Tumor Surgery and Their Influence on Residual Tumor Perception. Diagnostics. 2025; 15(15):1982. https://doi.org/10.3390/diagnostics15151982

Chicago/Turabian Style

Ruff, Christer, Till-Karsten Hauser, Constantin Roder, Daniel Feucht, Paula Bombach, Leonie Zerweck, Deborah Staber, Frank Paulsen, Ulrike Ernemann, and Georg Gohla. 2025. "Multidisciplinary, Clinical Assessment of Accelerated Deep-Learning MRI Protocols at 1.5 T and 3 T After Intracranial Tumor Surgery and Their Influence on Residual Tumor Perception" Diagnostics 15, no. 15: 1982. https://doi.org/10.3390/diagnostics15151982

APA Style

Ruff, C., Hauser, T.-K., Roder, C., Feucht, D., Bombach, P., Zerweck, L., Staber, D., Paulsen, F., Ernemann, U., & Gohla, G. (2025). Multidisciplinary, Clinical Assessment of Accelerated Deep-Learning MRI Protocols at 1.5 T and 3 T After Intracranial Tumor Surgery and Their Influence on Residual Tumor Perception. Diagnostics, 15(15), 1982. https://doi.org/10.3390/diagnostics15151982

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop