Next Article in Journal
PU-DZMS: Point Cloud Upsampling via Dense Zoom Encoder and Multi-Scale Complementary Regression
Previous Article in Journal
Research on the Accessibility of Different Colour Schemes for Web Resources for People with Colour Blindness
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review on Deep Learning Methods for Glioma Segmentation, Limitations, and Future Perspectives

by
Cecilia Diana-Albelda
*,
Álvaro García-Martín
and
Jesus Bescos
Video Processing and Understanding Lab, Escuela Politécnica Superior, Universidad Autónoma de Madrid, 28049 Madrid, Spain
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(8), 269; https://doi.org/10.3390/jimaging11080269
Submission received: 4 July 2025 / Revised: 28 July 2025 / Accepted: 7 August 2025 / Published: 11 August 2025
(This article belongs to the Section Medical Imaging)

Abstract

Accurate and automated segmentation of gliomas from Magnetic Resonance Imaging (MRI) is crucial for effective diagnosis, treatment planning, and patient monitoring. However, the aggressive nature and morphological complexity of these tumors pose significant challenges that call for advanced segmentation techniques. This review provides a comprehensive analysis of Deep Learning (DL) methods for glioma segmentation, with a specific focus on bridging the gap between research performance and practical clinical deployment. We evaluate over 80 state-of-the-art models published up to 2025, categorizing them into CNN-based, Pure Transformer, and Hybrid CNN-Transformer architectures. The primary objective of this paper is to critically assess these models not only on their segmentation accuracy but also on their computational efficiency and suitability for real-world medical environments by incorporating hardware resource considerations. We present a comparison of model performance on the BraTS datasets benchmark and introduce a suitability analysis for top-performing models based on their robustness, efficiency, and completeness of tumor region delineation. By identifying current trends, limitations, and key trade-offs, this review offers future research directions aimed at optimizing the balance between technical performance and clinical usability to improve diagnostic outcomes for glioma patients.

1. Introduction

Gliomas are among the most aggressive and lethal primary brain tumors, characterized by low survival rates and irregular growth patterns [1]. Accurate segmentation of these tumors is essential for diagnosis, treatment planning, and monitoring disease progression [2,3]. However, manual segmentation performed by radiologists is time-consuming, labor-intensive, and prone to inter-observer variability [4]. Given the importance of early and precise detection, the development of automated segmentation methods for this task has become a crucial research area [5,6].
Deep Learning (DL) techniques have emerged as a powerful tool for medical image analysis, including the semantic segmentation of brain tumors such as gliomas [7,8,9]. Current approaches in this field can be broadly categorized into three main types: (1) CNN-based methods, which excel in extracting spatial features; (2) Pure Transformer methods, which capture global context but require substantial computational resources; and (3) Hybrid CNN-Transformer, which aim to combine the strengths of both architectures.
Despite the significant progress achieved by these models, further research is needed to refine them, ensuring a balance between performance, computational efficiency, and clinical applicability, ultimately assisting radiologists in making faster and more reliable diagnoses.
Numerous surveys have explored brain tumor image analysis [10,11], distinguishing between classification [12,13,14] and segmentation tasks [15,16,17]. However, few have specifically focused on glioma segmentation [5,8,18] and these predominantly discuss Machine Learning and Deep Learning algorithms in a general context [6,19]. More importantly, the vast majority of the existing surveys share a fundamental limitation: they focus almost exclusively on segmentation accuracy while overlooking crucial details for real-world deployment, such as the number of trainable parameters and hardware requirements [20]. This omission creates a significant gap between state-of-the-art academic performance and the solutions that are genuinely feasible in a clinical environment.
To bridge this gap, the primary objective of this study is to provide a comprehensive and critical overview of the state-of-the-art, evaluating Deep Learning methods not only on their performance but also on their practical viability. To achieve this, our contributions are as follows:
  • A comparison of more than 80 Deep Learning methods for glioma segmentation up to 2025, providing insights into their effectiveness and efficiency.
  • An analysis of the number of tunable parameters of each method and their adaptability in clinical applications.
  • An identification of current trends and limitations within the field, along with a proposal for future research directions and suggestions for improvement.
This review represents a step forward, as it fills a significant gap in the current literature by incorporating hardware considerations into the evaluation of Deep Learning methods for glioma segmentation. By doing so, it guides the development of feasible technologies that can realistically improve glioma diagnosis and treatment outcomes.

Literature Search Strategy

To conduct this review, a structured and methodical approach has been adopted to identify, select, and analyze the most relevant scientific literature on glioma segmentation using deep learning techniques. The search focused on studies mainly published between 2020 and early 2025, with the aim of comparing the performance and efficiency of CNN-based, Transformer-based, and hybrid CNN-Transformer architectures.
We searched several major scientific databases, including IEEE Xplore, ScienceDirect, SpringerLink, and Wiley Online Library. These were complemented by queries in Google Scholar to capture recent conference proceedings, preprints, and emerging methods. The search strategy employed combinations of the following keywords:
  • “glioma segmentation”;
  • “brain tumor” AND “deep learning”;
  • “glioma” AND “CNN”, “U-Net”, “Transformer”, “Vision Transformer”, “SAM”, “multimodal segmentation”, “survey”.
The selection process comprised two screening phases, resulting in the inclusion of over 80 peer-reviewed papers that form the foundation of our comparative analysis. First, we conducted a preliminary screening of titles and abstracts to exclude studies unrelated to glioma segmentation or those lacking a methodological contribution in deep learning. Second, a full-text evaluation was performed to retain studies that (a) focus specifically on glioma or brain tumor segmentation, (b) report experimental results on public benchmarks such as BraTS, and (c) provide quantitative performance metrics and model complexity indicators such as parameter count.
To ensure transparency and methodological rigor, our review approach was guided by the principles of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) framework [21], a widely recognized guideline designed to enhance the clarity, reproducibility, and comprehensiveness of literature reviews.

2. Glioma Segmentation

Glioma segmentation refers to the process of delineating tumor regions from brain Magnetic Resonance Imaging (MRI) scans, a crucial step in the diagnosis and treatment of gliomas [22].
As illustrated in Figure 1, glioma segmentation has four main clinical applications: (1) Computer-aided Diagnosis, where AI-assisted tools help radiologists detect tumors; (2) Surgery Planning and Robotic Assistance, which rely on precise tumor boundaries to guide neurosurgical procedures; (3) Radiotherapy Planning, ensuring accurate targeting of radiation to tumor areas while sparing healthy tissue; and (4) Monitoring Disease Progression, enabling clinicians to track tumor evolution or response to treatment. Given the aggressive nature of many gliomas, robust segmentation methods are essential for improving patient outcomes [23].

2.1. Glioma

Gliomas are the most common primary tumors of the central nervous system (CNS), arising from glial or stem cells [24]. They are categorized into two types: infiltrating (diffuse) gliomas [25], which spread extensively into adjacent brain tissue, and circumscribed gliomas [26,27], which have clearer boundaries and are typically easier to treat.
Traditionally, the World Health Organization (WHO) classified gliomas from Grade I to IV based on their histological characteristics and degree of malignancy [28], with glioblastoma (GB) representing the most aggressive form (Grade IV), known for its rapid growth, necrosis, and resistance to standard therapies [29,30]. However, due to the limitations regarding inter-observer variability of purely histological classification, the 2021 WHO CNS tumor classification introduced a more robust, integrated diagnostic framework that combines histology with molecular and genetic markers [31].
The diverse characteristics of gliomas present significant challenges in treatment and image-based segmentation. Variability in tumor appearance, growth patterns, and different imaging techniques across institutions can hinder accurate diagnosis and treatment planning [32,33]. Deep Learning-driven automated segmentation tools are being developed to provide consistent and precise tumor delineation, aiding radiologists in decision-making and contributing to more efficient treatment.

2.2. MRI for Glioma Segmentation

MRI is the gold standard for glioma segmentation, providing detailed anatomical and functional insights into brain tumors [34]. Glioma imaging typically involves a multi-parametric MRI (mp-MRI) protocol that includes T1-weighted (T1), T2-weighted (T2), contrast-enhanced T1 (T1c), and Fluid-Attenuated Inversion Recovery T2 (T2-FLAIR) images, each highlighting distinct tissue properties [35]. In addition to these conventional sequences, some studies have also explored the potential contribution of diffusion-weighted imaging (DWI) [36], which provides information about water molecule motion and has shown promise when combined with anatomical and post-contrast images for segmentation tasks.
MRI images are acquired by exposing the patient to a strong magnetic field, followed by the application of radio-frequency pulses that interact with hydrogen protons (H+) in brain tissues [37]. The response of these protons generates the desired signal, with T1 and T2 images reflecting longitudinal and transverse relaxation times, respectively.
Accurate segmentation of gliomas requires integrating information from all conventional MRI modalities: T2 and T2-FLAIR help delineate peritumoral edema and infiltrative regions; T1c highlights active tumor areas, indicating regions of vascularization; and T1 assists in differentiating necrotic regions, a crucial factor in GB segmentation. Given these complementary sources of information, Deep Learning models for glioma segmentation must leverage multimodality integration, as relying on a single modality leads to suboptimal results [38].

3. Deep Learning Methods for Glioma Segmentation

Glioma segmentation methods based on Deep Learning can be categorized into three main groups: CNN-based, Pure Transformer, and Hybrid CNN-Transformer approaches.
CNN-based methods extract spatial features using convolutional layers, outperforming at local feature learning but struggling with long-range spatial dependencies. Pure Transformer methods, on the other hand, replace convolutions with self-attention, capturing global image context but demanding significantly more computational resources. Hybrid CNN-Transformer methods combine both approaches, aiming to balance precision and efficiency.

3.1. CNN-Based

Convolutional Neural Networks (CNNs), such as the U-Net architecture, are essential in medical image analysis, especially in glioma segmentation. As shown in Figure 2, these networks typically consist of convolutional layers for feature extraction, pooling to reduce spatial dimensions, and upsampling layers to restore the output images to the input size. Skip connections, which merge features from different layers, are a critical component for a detailed segmentation.
Among the CNN-based methods, 3D U-Net [39] and V-Net [40] focus on efficiently handling volumetric data, enhancing data management, and utilizing advanced loss functions to address class imbalances. Cascading CNNs, like those described in [41,42], and enhanced architectures like dResU-Net [43] and Dense Unet+ [44], incorporate techniques like residual blocks for robust feature extraction. However, they still exhibit a high dependence on the training dataset. In contrast, nnU-Net [45] dynamically adapts its architecture to each dataset, improving performance without requiring manual tuning. Additionally, methods like RFNet [46] and MAF-Net [47] offer adaptable solutions to the absence of specific modalities, while others employ innovative fusion techniques to better handle the multimodal nature of brain MRI data [48,49,50].
Recent innovations in glioma segmentation also include MM-BiFPN [51], which uses a fusion network with bidirectional pyramids, and GAM-Net [52], which integrates a dual convolution encoder and a gradient-driven decoder. Similarly, studies like [53,54] focus on leveraging variational auto-encoders and the Fuzzy Feature Contrast Maximization method to improve tumor delineation.
Despite these advances, the models still struggle with long-range dependencies, prompting the rise of Transformers in current segmentation techniques, as they effectively manage spatial and contextual relationships in medical images [47,55,56,57].

3.2. Pure Transformer

Figure 3 showcases the typical architecture of Pure Transformers for glioma segmentation. This approach processes images by converting them into sequences of patches and employs the Encoder Transformer Block, which includes multi-headed self-attention and position-wise feed-forward networks, to capture global dependencies across the entire brain scan. These Transformers utilize normalization layers and residual connections to enhance training stability, enabling a comprehensive understanding of the entire image, leveraging the strengths of attention mechanisms to produce highly accurate segmentation [58].
In the realm of Pure Transformer-based glioma segmentation, models like BTSwin U-Net [59] and Swin-Unet [60] enhance contextual information and spatial dependencies by utilizing Swin Transformers within U-shaped architectures. These models, while improving segmentation precision, require extensive training and substantial computational resources. Similarly, EMSViT [61] optimizes Transformer technology for medical imaging by integrating diverse convolution sizes for efficiency, while other alternatives [62,63,64] improve 3D data processing and resolution to preserve volumetric information.
Addressing more specialized challenges, MMCFormer [65] tackles the problem of missing modalities by employing co-training with 3D Transformer blocks to improve prediction reliability, even with incomplete datasets.
These models illustrate a trajectory towards optimizing Transformer technology for medical imaging, effectively capturing spatial and contextual information. However, they also incur a high computational burden, posing challenges for practical clinical application. To address this, the development of Hybrid CNN-Transformer architectures has emerged, aiming to balance the deep contextual insights of Transformers with the computational efficiency of CNNs.

3.3. Hybrid CNN-Transformer

Hybrid CNN-Transformer architectures are designed to integrate the spatial precision of CNNs with the long-range context capture of Transformers (see Figure 4). This integration aims to enhance segmentation accuracy by leveraging both local and global features, addressing the complex demands of tumor segmentation [66,67].
The fusion of CNN and Transformer technologies was pioneered by TransBTS [68]. These models enhance contextual and spatial relationships, despite facing high computational demands [69,70]. In response, more efficient architectures have recently been developed [71,72] to better handle multimodal data and missing modalities.
UNETR [73] represents a shift towards fully integrating Transformer capabilities for volumetric data and outperforming many state-of-the-art methods in most medical segmentation tasks. However, it usually struggles with fine details. In contrast, CFNet [74] focuses on multimodal brain tumor segmentation using hierarchical feature fusion from coarse to fine, enhancing segmentation through cross-modality attention, multi-scale context perception, and multimodal feature fusion. Moreover, Net [75] and UNetFormer [76] also try to mitigate the problem by simplifying training and enhancing accuracy but typically can lead to overfitting on the training set. To mitigate this, approaches like SegTransVAE [77] and TransBTSV2 [78] have focused on feature capturing and reducing overfitting.
With the ongoing need for generalization, federated learning and ensemble methods have been explored [79,80], although they encounter practical challenges related to domain adaptation in clinical settings. Alternatively, BiTr-Unet [81], Swin UNETR [60,82] optimize CNN and Transformer integration for medical applications, trying to minimize the existing domain gaps of the training and testing sets. These advancements are further extended by [83,84], refining architectures and employing topology-aware losses to enhance segmentation coherence.
Emerging models explore novel encoding and fusion strategies to balance local and global information [85,86,87,88,89,90], improving segmentation of irregularly shaped tumors through more advanced attention modules [91,92] or recalibration of features [93,94,95].
TransXAI [96] introduces a hybrid approach combining Vision Transformers (ViT) [97,98,99,100] and CNNs with explainable AI techniques to create interpretable heat maps, aiding medical professionals in understanding model decisions without compromising accuracy. In a more integrated application, SGS-Net [101] combines glioma segmentation with survival risk prediction, training both a segmentation decoder and a Cox model for survival analysis concurrently.
Further advancements are seen in models like 3D U-TFA [102], which incorporates Transformers and Feature Enhancement Attention within a 3D U-Net architecture to detect GB with vague boundaries more accurately. On the other hand, Arouse-Net [103] employs dilated convolutions and a focused attention mechanism on tumor edges, achieving faster processing times and improved metrics compared with most of the state-of-the-art models.
Finally, SAM [104,105,106,107,108] revolutionizes the standards towards prompt-driven interactive foundational models. Moreover, many methods have focused their research on efficient fine-tuning and optimization of SAM variants [109,110,111]. However, SAM is built to process 2D images, hence posing a challenge for volumetric medical data. To address this, Ref. [7] adapts the SAM2 [112] model for video semantic segmentation to be applied in the clinical domain.
These advancements illustrate a trend toward more generalizable, efficient, and interpretable segmentation solutions, addressing the complexities of brain tumor imaging with sophisticated architectural innovations.

3.4. Clinical Deployment

Despite the significant advancements in DL models for glioma segmentation, their deployment in real-world clinical environments remains limited due to a set of persistent challenges [113]. Although most models report strong performance on public datasets, these curated collections usually do not reflect the variability, artifacts, and limitations of routine hospital imaging [114]. As a result, the transition from research prototypes to clinically viable systems demands more than high segmentation accuracy; it requires robustness, efficiency, and practical integration capabilities.
One of the primary barriers to clinical deployment is the heterogeneity of MRI data across institutions [115]: differences in scanner hardware, acquisition protocols, image quality, and sequence availability can significantly degrade model generalization. Several studies highlighted this issue by comparing DL performance on retrospective clinical data against standardized datasets [116], revealing a substantial drop in segmentation accuracy when models trained on curated data were exposed to real-world inputs, reinforcing the importance of training on heterogeneous, multi-institutional datasets to ensure domain robustness.
Postoperative segmentation poses an additional challenge due to anatomical distortions, blood artifacts, resection cavities, and inconsistent sequences [117]. Algorithms initially designed for preoperative use tend to fail in these altered scenarios. Addressing this, some studies, such as RH-GlioSeg-nnU-Net [118], develop models specifically trained on postoperative data. On the other hand, Ref. [119] incorporates uncertainty estimation and non-tumorous tissue segmentation to improve reliability and clinical utility. Moreover, it provides an interactive tool enabling clinicians to visualize uncertainty maps and correct predictions in a short time, highlighting real-world usability.
Label quality has also proven critical, as ground truth manual segmentation is highly susceptible to inter-observer variability [115]. Due to this, uncertainty-aware loss functions and inter-observer validation have been proposed to mitigate label noise and improve reproducibility in training.
From a practical standpoint, computational constraints must be considered for deployment in routine workflows. Several models address this by employing lightweight architectures or hybrid pipelines that operate efficiently on standard hardware [111], avoiding the need for high-end GPUs. For example, Ref. [120] proposes a cascaded CNN with auto-encoder regularization that achieves full-case processing in under 10 min, including data routing, preprocessing, and segmentation. Other models eliminate the need for MRI-CT registration or minimize input requirements by demonstrating high segmentation accuracy using only T1c images [121].
In conclusion, the successful clinical deployment of DL-based glioma segmentation methods depends not only on segmentation accuracy but also on their robustness to variable data, resilience in postoperative settings, efficient resource use, and seamless integration into clinical systems. Importantly, low computational cost and reduced model complexity (i.e., limiting the number of trainable parameters) enable smoother integration into hospital environments with constrained resources, while also facilitating continuous refinement of the models over time as new clinical data becomes available.

4. Performance Analysis

In this section, we analyze the performance of more than 80 Deep Learning methods for glioma segmentation, focusing on the main publicly available databases for this task from the BraTS challenge collections. The analysis begins with a broad quantitative comparison of the different methods in terms of Dice Score [122], Hausdorff Distance [123], and the number of trainable parameters. Building on this review, we then identify the highest-performing models (those achieving a mean Dice of 90.00% or higher) for a focused qualitative evaluation. This subsequent analysis assesses their strengths, limitations, and overall suitability against several criteria for real-world clinical applicability. The ultimate goal is to synthesize the current state-of-the-art and inform a discussion on the most promising candidates for deployment in clinical environments based on specific operational requirements.

4.1. Glioma Databases

In the field of medical imaging, databases such as The Cancer Imaging Archive (TCIA) [124] and the Human Protein Atlas [125] provide invaluable data across various types of cancers, including brain tumors. For glioma MRI imaging specifically, BraTS (Brain Tumor Segmentation Challenge) [4], The Cancer Genome Atlas (TCGA) [126], and the Federated Tumor Segmentation (FeTS) [127] are particularly notable. BraTS stands out as the most extensive and frequently used in the development of Deep Learning algorithms for brain tumor image analysis, offering a rich and complete dataset of annotated brain MRI glioma scans and differentiating three tumor regions: the Whole Tumor (WT), which encompasses the full extension of the glioma; the Enhancing Tumor (ET), which captures active cells of the tumor that are infiltrating surrounding tissue; and the Tumor Core (TC), which combines ET with necrotic cells, typically placed in the center of the glioma (see Figure 5).
Over the years, as shown in Table 1, the scale of the BraTS datasets has significantly increased in terms of the number of training, testing, and validation cases from 2012 to 2024, which allows for a broader evaluation of algorithm performance [23,128,129,130].
This paper will utilize the BraTS benchmark to compare the different state-of-the-art methods, as it represents the most standardized dataset for glioma MRI imaging.

4.2. Segmentation Metrics

In glioma segmentation research, the literature employs metrics such as sensitivity, specificity [131], precision, F1 score, and the Jaccard index [132] to evaluate algorithm performance. However, the Dice Score [122] and the Hausdorff Distance with a confidence score of 95% (HD95) [123] are the most widely accepted metrics to assess segmentation effectiveness, particularly in the BraTS challenge [133,134]. The Dice Score (described in Equation (1)) measures the overlap between the predicted segmentation and the actual tumor, reflecting accuracy in tumor volume delineation, while HD95 (described in Equation (2)) evaluates the maximum distance at the worst-case boundary points between the predicted and actual segments.
Dice ( A , B ) = 2 | A B | | A | + | B |
where
  • A is the set of voxels in the ground truth segmentation.
  • B is the set of voxels in the predicted segmentation.
  • | A | and | B | are the number of voxels in each set.
  • | A B | is the number of voxels that both segmentations share (overlap).
HD 95 ( A , B ) = max percentile 95 min b B d ( a , b ) , percentile 95 min a A d ( b , a )
where
  • A and B are the sets of boundary points from the ground truth and predicted segmentations, respectively.
  • d ( a , b ) is the Euclidean distance between points a A and b B .
  • min b B d ( a , b ) is the minimum distance from a point a to any point in B.
  • percentile 95 denotes the 95th percentile of all such minimum distances.
In this paper, we use the Dice Score and the HD95 in order to compare different Deep Learning algorithms for glioma segmentation. Additionally, we study the number of trainable parameters for each method, considering the balance between model complexity and performance to determine their suitability for clinical applications. This approach helps identify models that are not only effective but also practical for deployment in medical settings.

4.3. Quantitative Review of Existing Methods

The segmentation of gliomas using Deep Learning methods has progressed significantly, driven by advancements in several architectures evaluated in the BraTS challenges.
CNN approaches, as presented in Table 2, typically achieve Dice Scores that mostly exceed 80%, with notable performers [44,135] reaching up to 95% in certain sub-regions, indicating high accuracy in tumor volume segmentation. However, the majority of these CNN-based approaches also exhibit high variation in HD95 distances, suggesting differences in their ability to precisely delineate tumor boundaries. Moreover, the underlying models vary greatly in their number of trainable parameters, ranging from as few as 6.92 M to over 111 M, suggesting the difficulty in trade-off between model performance and computational load.
Pure Transformer methods, as seen in Table 3, tend to maintain competitive Dice Scores but yet achieve lower performance compared with CNNs. Moreover, they often require a significantly higher number of tunable parameters, as seen in models like the HRSTNet-V4 [63] (266.33 M). This suggests that, while Pure Transformer approaches are capable of capturing complex spatial relationships, they do so at the expense of increased computational demand.
As depicted in Table 4, Hybrid CNN-Transformer algorithms showcase high Dice Scores and manage to keep HD95 distances low with comparable results to those achieved by CNNs while containing parameter counts to more manageable levels in many cases. For instance, GBT-SAM [111] achieves a Dice Score of 93.54% for the WT region with just 6.40 M trainable parameters, illustrating the effectiveness of integrating both architectures.
In conclusion, CNN-based approaches remain a strong baseline for their parameter efficiency and robust performance. Those based on Pure Transformers, while powerful, face challenges in deployment due to their high parameter counts. Alternatively, hybrid architectures offer a promising efficiency versus effectiveness compromise, suggesting that future developments could benefit from further exploration of this integrated approach.

4.4. Candidates for Real Deployment

While a vast number of deep learning models have been proposed for glioma segmentation, achieving high accuracy is only the first step toward clinical integration, while practical deployment requires models to be not only precise but also robust, efficient, and comprehensive. To this end, we provide a curated selection of models that demonstrate outstanding segmentation capability, defined as achieving a mean Dice Score of 90.00% or higher across all evaluated tumor regions. The goal is to analyze the strengths and limitations of these top-performing models to determine their viability as clinical support tools.
Table 5 presents a detailed comparison of these candidates. Beyond their architectural description, we evaluate them against four crucial conditions that directly impact their practical utility for assisting radiologists:
  • (C1) Robustness to Data Variability: A method is considered robust if it has been validated on at least two different datasets or incorporates specific domain generalization techniques to ensure stable performance across varied clinical data.
  • (C2) Postoperative/Low-Quality Data Handling: This condition marks methods designed or proven to handle challenging data, such as images with post-surgical changes or those from low-field or older imaging scanners, which can be encountered in clinical workflows.
  • (C3) Computational Efficiency: A model is deemed efficient for this comparison if it has fewer than 50 million trainable parameters. Where available, we also include specific metrics like GFLOPs and inference latency to provide a clearer picture of the required computational resources.
  • (C4) All Tumor Regions Delineated: This indicates whether the model provides segmentation for all three key sub-regions (WT, TC, and ET), as comprehensive delineation is essential for accurate diagnosis and treatment planning.
This structured evaluation aims to bridge the gap between theoretical performance and clinical applicability, highlighting models that best balance accuracy with the practical demands of a hospital environment.
Table 5. Comparison of glioma segmentation methods for clinical deployment readiness. The conditions (C1, C2, C3, C4) refer to (C1) Robustness to Data Variability, (C2) Postoperative/Low-Quality Data Handling, (C3) Computational Efficiency, and (C4) all tumor regions are delineated. The symbol “✓” indicates that the method addresses the condition.
Table 5. Comparison of glioma segmentation methods for clinical deployment readiness. The conditions (C1, C2, C3, C4) refer to (C1) Robustness to Data Variability, (C2) Postoperative/Low-Quality Data Handling, (C3) Computational Efficiency, and (C4) all tumor regions are delineated. The symbol “✓” indicates that the method addresses the condition.
MethodDescription + StrengthsLimitationsC1C2C3C4
Dense Unet+ [44]CNN-based. Effective 4-modality MRI fusion; Weighted connections reduce semantic gaps; ROI-focused training for efficiency; Lower computational complexity for inference ( 1.16 s in convolutional layers).Longer training time than baseline UNet; ResBlock coefficients require further optimization.
Kuntiyellannagari et al. [135]CNN-based. Advanced noise reduction via hybrid filter; Ensemble of three models for improved accuracy; Novel optimization algorithm (MAVOA) refines results.Limited generalizability and interpretability; High computational demand from its hybrid filter.
Enhanced Unet [142]CNN-based. Good generalization across multiple datasets; Optimized and simple architecture without complex additions; High computational efficiency ( 1.5 s per image).Exclusive reliance on the FLAIR modality; Needs validation on a wider range of glioma grades; Only provides results for WT tumor region.
CBAM-TransUNet [146]Hybrid CNN-Transformer. Combines U-Net, Swin Transformer, and CBAM; Attention module at the bottleneck to focus on key features; Includes a thorough robustness and ablation analysis.Aggressive data cropping; Use of a fixed, non-optimized weight in the loss function; Limited reproducibility.
Futrega et al. [83]Hybrid CNN-Transformer. Extensive ablation study; Includes a highly optimized post-processing strategy.Highly specialized for BraTS21; Only provides results for WT tumor region.
Unet Former [76]Hybrid CNN-Transformer. Self-supervised pretraining for 3D medical images; Fully reproducible; Flexible architecture offering an accuracy and efficiency ( 149.50 GFLOPs) trade-off.pretraining requires massive datasets and high-end hardware; pretraining effectiveness only tested on CT scans.
GBT-SAM [111]Hybrid CNN-Transformer. Generalization across four tumor domains; Multimodal integration of full mp-MRI; Modeling 3D inter-slice correlation; Parameter-efficient method (6.4 M trainable parameters).Dependent on bounding box prompts; Architecture specialized for four-modality MRI; Only provides results for WT tumor region.
CFNet [74]Hybrid CNN-Transformer. Modules for coarse-to-fine multimodal feature fusion; Rigorous ablation studies; Fully reproducible; High computational efficiency (35.47 GFLOPs).Fails to address the impact of MRI artifacts; Operates on 2D slices, losing direct 3D volumetric context.
MAT [148]Hybrid CNN-Transformer. Three-dimensional architecture using axial attention and self-distillation training that improves performance on small datasets; Smoother segmentation boundaries; Fully reproducible; Lightweight approach (11.7 M parameters).Non-isotropic image resizing during preprocessing can distort anatomical geometry; Limited ablation studies.
Arouse-Net [103]Hybrid CNN-Transformer. Attention mechanism specifically enhances tumor edges; Effective use of dilated convolutions to expand the receptive field of the model; High computational efficiency (1 s inference time).Insufficient experimental validation; Limited generalizability; Lacks reproducibility.

4.5. Discussion

The analysis of Deep Learning methods for glioma segmentation across the three most used BraTS datasets is presented in Figure 6. This visualization chronologically maps different algorithms by their Dice Scores, with bubble size representing the number of trainable parameters and color indicating the architectural family. This figure, in conjunction with the detailed results in Table 2, Table 3 and Table 4, suggests significant trends in the evolution of CNN, Pure Transformer, and Hybrid models.
As observed, Hybrid architectures (blue bubbles) have become dominant in recent years, frequently achieving the highest Dice Scores. Notably, many of these top-performing models manage their computational complexity effectively, often using fewer parameters than older or less efficient models. This suggests a successful synergy, where the convolutional layers’ feature extraction capabilities are effectively enhanced by the transformers’ ability to model long-range dependencies, leading to improved segmentation accuracy. Conversely, CNN models (pink bubbles) demonstrate consistent and reliable performance, often with remarkable parameter efficiency. This makes them strong, practical candidates for clinical environments where computational resources may be limited. Pure Transformers (green bubbles), while powerful, show that higher parameter counts do not inherently guarantee superior Dice Scores, indicating that architectural optimization remains a key challenge for this model family.
While these performance metrics are crucial, the transition to a real-world clinical setting demands an analysis beyond Dice Scores. For this purpose, we evaluated the top-performing models (mean Dice 90 %) against four criteria for clinical readiness, as detailed in Table 5. A critical finding from this analysis is that no single method perfectly satisfies all conditions, hence revealing the gap that still exists between research benchmarks and clinical applicability. Based on this analysis, the choice of the optimal model is highly dependent on the specific clinical use case.
For workflows where the primary objective is a rapid, interactive segmentation of the whole tumor (WT), whether to guide surgery or for a quick initial assessment, GBT-SAM [111] emerges as a particularly compelling candidate. Its high parameter efficiency (C3), proven robustness across different data domains (C1), and ability to handle low-quality data (C2) make it highly adaptable. Its prompt-based interactive nature could also empower radiologists by keeping them directly involved in the segmentation process. However, this method does not provide delineation of different regions inside the tumor area.
Consequently, when the clinical need is a fully automated, comprehensive delineation of all tumor sub-regions (WT, TC, and ET) for detailed radiotherapy planning or treatment response assessment, Dense Unet+ [44] presents a more robust and suitable solution. It uniquely satisfies the crucial conditions of robustness to data variability (C1), computational efficiency (C3), and complete tumor segmentation (C4).
Ultimately, the selection of a tool must align with the specific needs, workflow, and available hardware of the implementing medical center. Every method analyzed in Table 5 has its own set of strengths and limitations. Therefore, significant research is still required.

5. Conclusions

Key advancements and current limitations in glioma segmentation are summarized in this section. Moreover, it includes a discussion on future research directions in this area.

5.1. Recent Developments and Current Limitations

In recent years, automatic glioma segmentation has experienced rapid progress through the development of pure Transformer architectures and hybrid CNN-Transformer models.
CNN-based models, such as 3D U-Net and V-Net, have notably advanced the handling of volumetric MRI data and delivered solid performance across multiple tumor subregions [39,40]. More adaptive designs like nnU-Net [45] have further enhanced versatility by introducing automatic architecture and preprocessing configurations tailored to specific datasets. However, performance variability observed across datasets such as BraTS17 through BraTS20 suggests that CNNs remain sensitive to dataset-specific distributions, raising concerns about their generalizability to diverse clinical scenarios. Notably, Dense Unet+ [44] achieved some of the best results overall, with a Dice score of 95.80 for the whole tumor region, 95.50 for the tumor core, and 93.70 for the enhancing tumor, outperforming many more recent architectures.
Pure Transformer-based models have introduced a new paradigm in medical image segmentation by effectively capturing long-range dependencies. Methods like VT-UNet [62] demonstrate this potential by achieving high Dice scores and low HD95 values in tumor core segmentation. Still, such improvements often come at the cost of large model sizes and increased computational demand.
Hybrid CNN-Transformer architectures like SwinBTS [2] and 3D U-TFA [102] have demonstrated state-of-the-art segmentation scores, with Dice(WT) exceeding 95 and competitive HD95 values. GBT-SAM [111], for example, achieved a Dice score of 93.54 (WT), alongside a surprisingly low parameter count of 6.4 million, reflecting the potential efficiency of hybrid designs, though such low complexity is not representative of most models in this category.
However, despite these impressive results on curated research datasets, our analysis reveals a critical gap between the high accuracy reported in research benchmarks and true medical applicability. The deployment of these models in real-world environments remains limited due to a persistent set of challenges [113] that include the poor generalization of models to heterogeneous clinical data, the inability to perform postoperative segmentation, and the demanding computational resources required by many state-of-the-art architectures [115]. Therefore, significant future research is needed to address these practical barriers, focusing on enhancing model robustness and efficiency to ensure the successful translation of these powerful tools into routine clinical practice.

5.2. Future Directions

Recent developments in glioma segmentation research highlight several key directions that are likely to shape future work. These directions reflect not only the evolution of model architectures and performance but also a growing emphasis on building systems that are practical, robust, and ready for real-edge deployment. Drawing on recent empirical patterns, three main research challenges emerge:
  • Enhancing generalization across domains and datasets.
  • Minimizing computational cost without compromising accuracy.
  • Refining hybrid architectures for clinical integration.
First, generalization remains a considerable challenge. Deep learning models often suffer from “domain shift”, where a model trained on data from one hospital performs poorly when tested on data from another due to variations in scanners, acquisition protocols, and patient populations. This gap between performance on curated research data and real-world clinical data is a major barrier to adoption. Future research must prioritize methods that ensure models are robust to these variations. Two key strategies are emerging to address this:
  • Domain Generalization techniques aim to train a single, robust model on data from multiple sources that can generalize well to completely unseen hospitals without needing retraining. This involves methods like advanced data augmentation, feature alignment, and disentanglement to learn domain-invariant features.
  • Federated Learning offers a paradigm-shifting solution by training a shared global model across multiple institutions without ever centralizing or sharing sensitive patient data. Each institution trains the model locally, and only the model updates (weights or gradients) are sent to a central server for aggregation. This approach not only preserves data privacy but also naturally exposes the model to a diverse range of data, which has been shown to significantly boost the performance and robustness of brain tumor segmentation models. Validating models trained with these methods across multiple independent institutions (cross-institutional validation) is becoming the gold standard for proving their real-world clinical readiness.
Second, improving computational efficiency remains essential for real-world deployment. Although historically the best-performing models have relied on large parameter counts, recent trends suggest that competitive segmentation accuracy can be achieved with more compact designs. This opens the door for research into model compression, lightweight attention mechanisms, and neural architecture search aimed at producing architectures that are both high-performing and suitable for resource-constrained hardware.
Finally, as this review has shown, hybrid CNN-Transformer architectures have become increasingly prominent and now dominate the upper range of segmentation performance. Their flexibility positions them as strong candidates for clinical translation. Future directions will likely focus on modular, adaptive variants of these hybrids, possibly enhancing model transparency through interpretability and uncertainty estimation frameworks to build greater trust with clinicians.

Author Contributions

Conceptualization, C.D.-A., Á.G.-M. and J.B.; methodology, C.D.-A.; software, C.D.-A.; validation, C.D.-A.; formal analysis, C.D.-A.; investigation, C.D.-A.; resources, C.D.-A.; data curation, C.D.-A.; writing—original draft preparation, C.D.-A.; writing—review and editing, C.D.-A., Á.G.-M. and J.B.; visualization, C.D.-A.; supervision, Á.G.-M. and J.B.; project administration, Á.G.-M. and J.B.; funding acquisition, Á.G.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been partially funded by the SEGA-CV (TED2021-131643A-I00) and the HVD (PID2021-125051OB-I00) projects of the Ministerio de Ciencia e Innovación of the Spanish Government.

Informed Consent Statement

Since our data originates from an international challenge, all datasets have been anonymized and preprocessed to ensure privacy and eliminate concerns related to image registration [4]. This preprocessing phase maintains the data integrity while adhering to ethical and legal standards; hence, patient consent was held by the challenge organizers.

Conflicts of Interest

The atuthors declare no conflicts of interest.

References

  1. Weller, M.; Wick, W.; Aldape, K.; Brada, M.; Berger, M.; Pfister, S.M.; Nishikawa, R.; Rosenthal, M.; Wen, P.Y.; Stupp, R.; et al. Glioma. Nat. Rev. Dis. Prim. 2015, 1, 1–18. [Google Scholar] [CrossRef]
  2. Harshini, E.; Chinnam, S.K. SWIN BTS Using Deep Learning. In Proceedings of the 2024 Second International Conference on Data Science and Information System (ICDSIS), Hassan, India, 17–18 May 2024; pp. 1–8. [Google Scholar]
  3. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.R.; Xu, D. Swin unetr: Swin transformers for semantic segmentation of brain tumors in mri images. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 272–284. [Google Scholar]
  4. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
  5. El Hachimy, I.; Kabelma, D.; Echcharef, C.; Hassani, M.; Benamar, N.; Hajji, N. A comprehensive survey on the use of deep learning techniques in glioblastoma. Artif. Intell. Med. 2024, 154, 102902. [Google Scholar] [CrossRef]
  6. Ranjbarzadeh, R.; Caputo, A.; Tirkolaee, E.B.; Ghoushchi, S.J.; Bendechache, M. Brain tumor segmentation of MRI images: A comprehensive review on the application of artificial intelligence tools. Comput. Biol. Med. 2023, 152, 106405. [Google Scholar] [CrossRef]
  7. Zhu, J.; Qi, Y.; Wu, J. Medical sam 2: Segment medical images as video via segment anything model 2. arXiv 2024, arXiv:2408.00874. [Google Scholar] [CrossRef]
  8. Ghadimi, D.J.; Vahdani, A.M.; Karimi, H.; Ebrahimi, P.; Fathi, M.; Moodi, F.; Habibzadeh, A.; Khodadadi Shoushtari, F.; Valizadeh, G.; Mobarak Salari, H.; et al. Deep Learning-Based Techniques in Glioma Brain Tumor Segmentation Using Multi-Parametric MRI: A Review on Clinical Applications and Future Outlooks. J. Magn. Reson. Imaging 2025, 61, 1094–1109. [Google Scholar] [CrossRef] [PubMed]
  9. Biratu, E.S.; Schwenker, F.; Debelee, T.G.; Kebede, S.R.; Negera, W.G.; Molla, H.T. Enhanced region growing for brain tumor MR image segmentation. J. Imaging 2021, 7, 22. [Google Scholar] [CrossRef]
  10. Krishnapriya, S.; Karuna, Y. A survey of deep learning for MRI brain tumor segmentation methods: Trends, challenges, and future directions. Health Technol. 2023, 13, 181–201. [Google Scholar] [CrossRef]
  11. Kaur, R.; Doegar, A. Brain tumor segmentation using deep learning: Taxonomy, survey and challenges. In Brain Tumor MRI Image Segmentation Using Deep Learning Techniques; Elsevier: Amsterdam, The Netherlands, 2022; pp. 225–238. [Google Scholar]
  12. Chauhan, P.; Lunagaria, M.; Verma, D.K.; Vaghela, K.; Diwan, A.; Patole, S.; Mahadeva, R. Analyzing Brain Tumour Classification Techniques: A Comprehensive Survey. IEEE Access 2024, 12, 136389–136407. [Google Scholar] [CrossRef]
  13. Saeedi, S.; Rezayi, S.; Keshavarz, H.; Niakan Kalhori, S.R. MRI-based brain tumor detection using convolutional deep learning methods and chosen machine learning techniques. BMC Med. Inform. Decis. Mak. 2023, 23, 16. [Google Scholar] [CrossRef]
  14. Biratu, E.S.; Schwenker, F.; Ayano, Y.M.; Debelee, T.G. A survey of brain tumor segmentation and classification algorithms. J. Imaging 2021, 7, 179. [Google Scholar] [CrossRef] [PubMed]
  15. Umarani, C.M.; Gollagi, S.; Allagi, S.; Sambrekar, K.; Ankali, S.B. Advancements in deep learning techniques for brain tumor segmentation: A survey. Inform. Med. Unlocked 2024, 50, 101576. [Google Scholar] [CrossRef]
  16. Wang, P.; Yang, Q.; He, Z.; Yuan, Y. Vision transformers in multi-modal brain tumor MRI segmentation: A review. Meta-Radiol. 2023, 1, 100004. [Google Scholar] [CrossRef]
  17. Magadza, T.; Viriri, S. Deep learning for brain tumor segmentation: A survey of state-of-the-art. J. Imaging 2021, 7, 19. [Google Scholar] [CrossRef]
  18. Rasool, N.; Bhat, J.I. Glioma brain tumor segmentation using deep learning: A review. In Proceedings of the 2023 10th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 15–17 March 2023; pp. 484–489. [Google Scholar]
  19. Raghavendra, U.; Gudigar, A.; Paul, A.; Goutham, T.; Inamdar, M.A.; Hegde, A.; Devi, A.; Ooi, C.P.; Deo, R.C.; Barua, P.D.; et al. Brain tumor detection and screening using artificial intelligence techniques: Current trends and future perspectives. Comput. Biol. Med. 2023, 163, 107063. [Google Scholar] [CrossRef]
  20. Abidin, Z.U.; Naqvi, R.A.; Haider, A.; Kim, H.S.; Jeong, D.; Lee, S.W. Recent deep learning-based brain tumor segmentation models using multi-modality magnetic resonance imaging: A prospective survey. Front. Bioeng. Biotechnol. 2024, 12, 1392807. [Google Scholar] [CrossRef]
  21. Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G.; The PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. Int. J. Surg. 2010, 8, 336–341. [Google Scholar] [CrossRef]
  22. Porz, N.; Bauer, S.; Pica, A.; Schucht, P.; Beck, J.; Verma, R.K.; Slotboom, J.; Reyes, M.; Wiest, R. Multi-modal glioblastoma segmentation: Man versus machine. PLoS ONE 2014, 9, e96873. [Google Scholar] [CrossRef]
  23. Bakas, S.; Reyes, M.; Jakab, A.; Bauer, S.; Rempfler, M.; Crimi, A.; Shinohara, R.T.; Berger, C.; Ha, S.M.; Rozycki, M.; et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. arXiv 2018, arXiv:1811.02629. [Google Scholar] [CrossRef]
  24. Weller, M.; Wen, P.Y.; Chang, S.M.; Dirven, L.; Lim, M.; Monje, M.; Reifenberger, G. Glioma. Nat. Rev. Dis. Prim. 2024, 10, 33. [Google Scholar] [CrossRef] [PubMed]
  25. Claes, A.; Idema, A.J.; Wesseling, P. Diffuse glioma growth: A guerilla war. Acta Neuropathol. 2007, 114, 443–458. [Google Scholar] [CrossRef]
  26. Esteban-Rodríguez, I.; López-Muñoz, S.; Blasco-Santana, L.; Mejías-Bielsa, J.; Gordillo, C.H.; Jiménez-Heffernan, J.A. Cytological features of diffuse and circumscribed gliomas. Cytopathology 2024, 35, 534–544. [Google Scholar] [CrossRef]
  27. Chen, J.; Dahiya, S.M. Update on circumscribed gliomas and glioneuronal tumors. Surg. Pathol. Clin. 2020, 13, 249–266. [Google Scholar] [CrossRef]
  28. Chen, R.; Smith-Cohn, M.; Cohen, A.L.; Colman, H. Glioma subclassifications and their clinical significance. Neurotherapeutics 2017, 14, 284–297. [Google Scholar] [CrossRef]
  29. Wirsching, H.G.; Weller, M. Glioblastoma. Malignant Brain Tumors: State-of-the-Art Treatment; Springer: Berlin/Heidelberg, Germany, 2017; pp. 265–288. [Google Scholar]
  30. Davis, M.E. Glioblastoma: Overview of disease and treatment. Clin. J. Oncol. Nurs. 2016, 20, S2. [Google Scholar] [CrossRef] [PubMed]
  31. Bai, J.; Varghese, J.; Jain, R. Adult glioma WHO classification update, genomics, and imaging: What the radiologists need to know. Top. Magn. Reson. Imaging 2020, 29, 71–82. [Google Scholar] [CrossRef]
  32. Adewole, M.; Rudie, J.D.; Gbdamosi, A.; Toyobo, O.; Raymond, C.; Zhang, D.; Omidiji, O.; Akinola, R.; Suwaid, M.A.; Emegoakor, A.; et al. The brain tumor segmentation (brats) challenge 2023: Glioma segmentation in sub-saharan africa patient population (brats-africa). arXiv 2023, arXiv:2305.19369v1. [Google Scholar]
  33. Kazerooni, A.F.; Khalili, N.; Liu, X.; Haldar, D.; Jiang, Z.; Zapaishchykova, A.; Pavaine, J.; Shah, L.M.; Jones, B.V.; Sheth, N.; et al. BraTS-PEDs: Results of the multi-consortium international pediatric brain tumor segmentation challenge 2023. arXiv 2024, arXiv:2407.08855. [Google Scholar] [CrossRef]
  34. Despotović, I.; Goossens, B.; Philips, W. MRI segmentation of the human brain: Challenges, methods, and applications. Comput. Math. Methods Med. 2015, 2015, 450341. [Google Scholar] [CrossRef]
  35. Treister, D.; Kingston, S.; Hoque, K.E.; Law, M.; Shiroishi, M.S. Multimodal magnetic resonance imaging evaluation of primary brain tumors. Semin. Oncol. 2014, 41, 478–495. [Google Scholar] [CrossRef] [PubMed]
  36. Khorasani, A.; Kafieh, R.; Saboori, M.; Tavakoli, M.B. Glioma segmentation with DWI weighted images, conventional anatomical images, and post-contrast enhancement magnetic resonance imaging images by U-Net. Phys. Eng. Sci. Med. 2022, 45, 925–934. [Google Scholar] [CrossRef]
  37. Katti, G.; Ara, S.A.; Shireen, A. Magnetic resonance imaging (MRI)—A review. Int. J. Dent. Clin. 2011, 3, 65–70. [Google Scholar]
  38. Diana-Albelda, C.; Alcover-Couso, R.; García-Martín, Á.; Bescos, J. How SAM Perceives Different mp-MRI Brain Tumor Domains? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 4959–4970. [Google Scholar]
  39. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  40. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  41. Vu, M.H.; Nyholm, T.; Löfstedt, T. TuNet: End-to-end hierarchical brain tumor segmentation using cascaded networks. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 5th International Workshop, BrainLes 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, 17 October 2019; Springer: Berlin/Heidelberg, Germany, 2020; pp. 174–186. [Google Scholar]
  42. Wang, G.; Li, W.; Ourselin, S.; Vercauteren, T. Automatic brain tumor segmentation using cascaded anisotropic convolutional neural networks. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: Third International Workshop, BrainLes 2017, Held in Conjunction with MICCAI 2017, Quebec City, QC, Canada, 14 September 2017; Springer: Berlin/Heidelberg, Germany, 2018; pp. 178–190. [Google Scholar]
  43. Raza, R.; Bajwa, U.I.; Mehmood, Y.; Anwar, M.W.; Jamal, M.H. dResU-Net: 3D deep residual U-Net based brain tumor segmentation from multimodal MRI. Biomed. Signal Process. Control 2023, 79, 103861. [Google Scholar] [CrossRef]
  44. Çetiner, H.; Metlek, S. DenseUNet+: A novel hybrid segmentation approach based on multi-modality images for brain tumor segmentation. J. King Saud-Univ.-Comput. Inf. Sci. 2023, 35, 101663. [Google Scholar] [CrossRef]
  45. Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S.; et al. nnu-net: Self-adapting framework for u-net-based medical image segmentation. arXiv 2018, arXiv:1809.10486. [Google Scholar]
  46. Ding, Y.; Yu, X.; Yang, Y. RFNet: Region-aware fusion network for incomplete multi-modal brain tumor segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 19–25 June 2021; pp. 3975–3984. [Google Scholar]
  47. Huang, Z.; Lin, L.; Cheng, P.; Peng, L.; Tang, X. Multi-modal brain tumor segmentation via missing modality synthesis and modality-level attention fusion. arXiv 2022, arXiv:2203.04586. [Google Scholar]
  48. Wang, Y.; Zhang, Y.; Hou, F.; Liu, Y.; Tian, J.; Zhong, C.; Zhang, Y.; He, Z. Modality-pairing learning for brain tumor segmentation. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 6th International Workshop, BrainLes 2020, Held in Conjunction with MICCAI 2020, Lima, Peru, 4 October 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 230–240. [Google Scholar]
  49. Liu, Y.; Mu, F.; Shi, Y.; Cheng, J.; Li, C.; Chen, X. Brain tumor segmentation in multimodal MRI via pixel-level and feature-level image fusion. Front. Neurosci. 2022, 16, 1000587. [Google Scholar] [CrossRef]
  50. Tong, J.; Wang, C. A dual tri-path CNN system for brain tumor segmentation. Biomed. Signal Process. Control 2023, 81, 104411. [Google Scholar] [CrossRef]
  51. Syazwany, N.S.; Nam, J.H.; Lee, S.C. MM-BiFPN: Multi-modality fusion network with Bi-FPN for MRI brain tumor segmentation. IEEE Access 2021, 9, 160708–160720. [Google Scholar] [CrossRef]
  52. Wang, Y.; Chen, J.; Bai, X. Gradient-assisted deep model for brain tumor segmentation by multi-modality MRI volumes. Biomed. Signal Process. Control 2023, 85, 105066. [Google Scholar] [CrossRef]
  53. Myronenko, A. 3D MRI brain tumor segmentation using autoencoder regularization. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 4th International Workshop, BrainLes 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 16 September 2018; Springer: Berlin/Heidelberg, Germany, 2019; pp. 311–320. [Google Scholar]
  54. Sahoo, A.K.; Parida, P.; Muralibabu, K.; Dash, S. An improved DNN with FFCM method for multimodal brain tumor segmentation. Intell. Syst. Appl. 2023, 18, 200245. [Google Scholar] [CrossRef]
  55. Oktay, O. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  56. Akbar, A.S.; Fatichah, C.; Suciati, N. Unet3D with multiple atrous convolutions attention block for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 182–193. [Google Scholar]
  57. Zhao, L.; Ma, J.; Shao, Y.; Jia, C.; Zhao, J.; Yuan, H. MM-UNet: A multimodality brain tumor segmentation network in MRI images. Front. Oncol. 2022, 12, 950706. [Google Scholar] [CrossRef]
  58. Xing, Z.; Yu, L.; Wan, L.; Han, T.; Zhu, L. NestedFormer: Nested modality-aware transformer for brain tumor segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 140–150. [Google Scholar]
  59. Liang, J.; Yang, C.; Zhong, J.; Ye, X. BTSwin-Unet: 3D U-shaped symmetrical Swin transformer-based network for brain tumor segmentation with self-supervised pre-training. Neural Process. Lett. 2023, 55, 3695–3713. [Google Scholar] [CrossRef]
  60. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 205–218. [Google Scholar]
  61. Sagar, A. EMSViT: Efficient multi scale vision transformer for biomedical image segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 39–51. [Google Scholar]
  62. Peiris, H.; Hayat, M.; Chen, Z.; Egan, G.; Harandi, M. A robust volumetric transformer for accurate 3D tumor segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 162–172. [Google Scholar]
  63. Wei, C.; Ren, S.; Guo, K.; Hu, H.; Liang, J. High-resolution Swin transformer for automatic medical image segmentation. Sensors 2023, 23, 3420. [Google Scholar] [CrossRef] [PubMed]
  64. Peiris, H.; Hayat, M.; Chen, Z.; Egan, G.; Harandi, M. Hybrid window attention based transformer architecture for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Singapore, 18 September 2022; pp. 173–182. [Google Scholar]
  65. Karimijafarbigloo, S.; Azad, R.; Kazerouni, A.; Ebadollahi, S.; Merhof, D. Mmcformer: Missing modality compensation transformer for brain tumor segmentation. In Proceedings of the Medical Imaging with Deep Learning, Paris, France, 3–5 July 2024; pp. 1144–1162. [Google Scholar]
  66. Liang, J.; Yang, C.; Zeng, M.; Wang, X. TransConver: Transformer and convolution parallel network for developing automatic brain tumor segmentation in MRI images. Quant. Imaging Med. Surg. 2022, 12, 2397. [Google Scholar] [CrossRef]
  67. Li, X.; Ma, S.; Tang, J.; Guo, F. TranSiam: Fusing multimodal visual features using transformer for medical image segmentation. arXiv 2022, arXiv:2204.12185. [Google Scholar] [CrossRef]
  68. Wenxuan, W.; Chen, C.; Meng, D.; Hong, Y.; Sen, Z.; Jiangyun, L. Transbts: Multimodal brain tumor segmentation using transformer. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; pp. 109–119. [Google Scholar]
  69. Jun, E.; Jeong, S.; Heo, D.W.; Suk, H.I. Medical transformer: Universal brain encoder for 3D MRI analysis. arXiv 2021, arXiv:2104.13633. [Google Scholar] [CrossRef] [PubMed]
  70. Li, S.; Sui, X.; Luo, X.; Xu, X.; Liu, Y.; Goh, R. Medical image segmentation using squeeze-and-expansion transformers. arXiv 2021, arXiv:2105.09511. [Google Scholar]
  71. Fang, F.; Yao, Y.; Zhou, T.; Xie, G.; Lu, J. Self-supervised multi-modal hybrid fusion network for brain tumor segmentation. IEEE J. Biomed. Health Inform. 2021, 26, 5310–5320. [Google Scholar] [CrossRef]
  72. Zhou, T.; Canu, S.; Vera, P.; Ruan, S. Feature-enhanced generation and multi-modality fusion based deep neural network for brain tumor segmentation with missing MR modalities. Neurocomputing 2021, 466, 102–112. [Google Scholar] [CrossRef]
  73. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.R.; Xu, D. Unetr: Transformers for 3d medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 574–584. [Google Scholar]
  74. Cheng, Y.; Zheng, Y.; Wang, J. CFNet: Automatic multi-modal brain tumor segmentation through hierarchical coarse-to-fine fusion and feature communication. Biomed. Signal Process. Control 2025, 99, 106876. [Google Scholar] [CrossRef]
  75. Liu, J.; Zheng, J.; Jiao, G. Transition Net: 2D backbone to segment 3D brain tumor. Biomed. Signal Process. Control 2022, 75, 103622. [Google Scholar] [CrossRef]
  76. Hatamizadeh, A.; Xu, Z.; Yang, D.; Li, W.; Roth, H.; Xu, D. Unetformer: A unified vision transformer model and pre-training framework for 3d medical image segmentation. arXiv 2022, arXiv:2204.00631. [Google Scholar]
  77. Pham, Q.D.; Nguyen-Truong, H.; Phuong, N.N.; Nguyen, K.N.; Nguyen, C.D.; Bui, T.; Truong, S.Q. Segtransvae: Hybrid cnn-transformer with regularization for medical image segmentation. In Proceedings of the 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–5. [Google Scholar]
  78. Li, J.; Wang, W.; Chen, C.; Zhang, T.; Zha, S.; Wang, J.; Yu, H. TransBTSV2: Towards better and more efficient volumetric segmentation of medical images. arXiv 2022, arXiv:2201.12785. [Google Scholar] [CrossRef]
  79. Nalawade, S.; Ganesh, C.; Wagner, B.; Reddy, D.; Das, Y.; Yu, F.F.; Fei, B.; Madhuranthakam, A.J.; Maldjian, J.A. Federated learning for brain tumor segmentation using MRI and transformers. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 444–454. [Google Scholar]
  80. Shi, Y.; Micklisch, C.; Mushtaq, E.; Avestimehr, S.; Yan, Y.; Zhang, X. An ensemble approach to automatic brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 138–148. [Google Scholar]
  81. Jia, Q.; Shu, H. Bitr-unet: A cnn-transformer combined network for mri brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 3–14. [Google Scholar]
  82. Dobko, M.; Kolinko, D.I.; Viniavskyi, O.; Yelisieiev, Y. Combining CNNs with transformer for multimodal 3D MRI brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 232–241. [Google Scholar]
  83. Futrega, M.; Milesi, A.; Marcinkiewicz, M.; Ribalta, P. Optimized U-Net for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 15–29. [Google Scholar]
  84. Yang, H.; Shen, Z.; Li, Z.; Liu, J.; Xiao, J. Combining global information with topological prior for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 204–215. [Google Scholar]
  85. Gai, D.; Zhang, J.; Xiao, Y.; Min, W.; Zhong, Y.; Zhong, Y. RMTF-Net: Residual mix transformer fusion net for 2D brain tumor segmentation. Brain Sci. 2022, 12, 1145. [Google Scholar] [CrossRef]
  86. Zhang, Y.; He, N.; Yang, J.; Li, Y.; Wei, D.; Huang, Y.; Zhang, Y.; He, Z.; Zheng, Y. mmformer: Multimodal medical transformer for incomplete multimodal learning of brain tumor segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 107–117. [Google Scholar]
  87. Wang, P.; Liu, S.; Peng, J. AST-Net: Lightweight hybrid transformer for multimodal brain tumor segmentation. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montréal, QC, Canada, 21–25 August 2022; pp. 4623–4629. [Google Scholar]
  88. Huang, L.; Zhu, E.; Chen, L.; Wang, Z.; Chai, S.; Zhang, B. A transformer-based generative adversarial network for brain tumor segmentation. Front. Neurosci. 2022, 16, 1054948. [Google Scholar] [CrossRef]
  89. Zhou, T. Modality-level cross-connection and attentional feature fusion based deep neural network for multi-modal brain tumor segmentation. Biomed. Signal Process. Control. 2023, 81, 104524. [Google Scholar] [CrossRef]
  90. Vatanpour, M.; Haddadnia, J. TransDoubleU-Net: Dual Scale Swin Transformer With Dual Level Decoder for 3D Multimodal Brain Tumor Segmentation. IEEE Access 2023, 11, 125511–125518. [Google Scholar] [CrossRef]
  91. Pang, X.; Zhao, Z.; Wang, Y.; Li, F.; Chang, F. LGMSU-Net: Local Features, Global Features, and Multi-Scale Features Fused the U-Shaped Network for Brain Tumor Segmentation. Electronics 2022, 11, 1911. [Google Scholar] [CrossRef]
  92. Liang, J.; Yang, C.; Zeng, L. 3D PSwinBTS: An efficient transformer-based Unet using 3D parallel shifted windows for brain tumor segmentation. Digit. Signal Process. 2022, 131, 103784. [Google Scholar] [CrossRef]
  93. Tian, W.; Li, D.; Lv, M.; Huang, P. Axial attention convolutional neural network for brain tumor segmentation with multi-modality MRI scans. Brain Sci. 2022, 13, 12. [Google Scholar] [CrossRef] [PubMed]
  94. Gao, H.; Miao, Q.; Ma, D.; Liu, R. Deep mutual learning for brain tumor segmentation with the fusion network. Neurocomputing 2023, 521, 213–220. [Google Scholar] [CrossRef]
  95. Lu, Y.; Chang, Y.; Zheng, Z.; Sun, Y.; Zhao, M.; Yu, B.; Tian, C.; Zhang, Y. GMetaNet: Multi-scale ghost convolutional neural network with auxiliary MetaFormer decoding path for brain tumor segmentation. Biomed. Signal Process. Control. 2023, 83, 104694. [Google Scholar] [CrossRef]
  96. Zeineldin, R.A.; Karar, M.E.; Elshaer, Z.; Coburger, J.; Wirtz, C.R.; Burgert, O.; Mathis-Ullrich, F. Explainable hybrid vision transformers and convolutional network for multimodal glioma segmentation in brain MRI. Sci. Rep. 2024, 14, 3713. [Google Scholar] [CrossRef]
  97. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  98. Alcover-Couso, R.; Escudero-Viñolo, M.; SanMiguel, J.C.; Bescós, J. Gradient-based Class Weighting for Unsupervised Domain Adaptation in Dense Prediction Visual Tasks. arXiv 2024, arXiv:2407.01327. [Google Scholar] [CrossRef]
  99. Alcover-Couso, R.; SanMiguel, J.C.; Escudero-Viñolo, M.; Martínez, J.M. Layer-wise Model Merging for Unsupervised Domain Adaptation in Segmentation Tasks. arXiv 2024, arXiv:2409.15813. [Google Scholar] [CrossRef]
  100. Alcover-Couso, R.; Escudero-Vinolo, M.; SanMiguel, J.C.; Martinez, J.M. Soft labelling for semantic segmentation: Bringing coherence to label down-sampling. arXiv 2023, arXiv:2302.13961. [Google Scholar]
  101. Cheng, J.; Kuang, H.; Yang, S.; Yue, H.; Liu, J.; Wang, J. Segmentation-Guided Deep Learning for Glioma Survival Risk Prediction with Multimodal MRI. B Data Min. Anal. 2025, 8, 364–382. [Google Scholar] [CrossRef]
  102. Wu, S.; Chen, Z.; Sun, P. 3D U-TFA: A deep convolutional neural network for automatic segmentation of glioblastoma. Biomed. Signal Process. Control. 2025, 99, 106829. [Google Scholar] [CrossRef]
  103. Li, H.; Qi, X.; Hu, Y.; Zhang, J. Arouse-Net: Enhancing Glioblastoma Segmentation in Multi-Parametric MRI with a Custom 3D Convolutional Neural Network and Attention Mechanism. Mathematics 2025, 13, 160. [Google Scholar] [CrossRef]
  104. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 18–22 June 2023; pp. 4015–4026. [Google Scholar]
  105. Zhang, K.; Liu, D. Customized segment anything model for medical image segmentation. arXiv 2023, arXiv:2304.13785. [Google Scholar] [CrossRef]
  106. Cheng, J.; Ye, J.; Deng, Z.; Chen, J.; Li, T.; Wang, H.; Su, Y.; Huang, Z.; Chen, J.; Jiang, L.; et al. Sam-med2d. arXiv 2023. [Google Scholar] [CrossRef]
  107. Alcover-Couso, R.; Escudero-Viñolo, M.; SanMiguel, J.C.; Bescos, J. VLMs meet UDA: Boosting Transferability of Open Vocabulary Segmentation with Unsupervised Domain Adaptation. arXiv 2024, arXiv:2412.09240. [Google Scholar] [CrossRef]
  108. Montalvo, J.; García-Martín, Á.; Carballeira, P.; SanMiguel, J.C. Unsupervised Class Generation to Expand Semantic Segmentation Datasets. arXiv 2025, arXiv:2501.02264. [Google Scholar] [CrossRef]
  109. Wu, J.; Ji, W.; Liu, Y.; Fu, H.; Xu, M.; Xu, Y.; Jin, Y. Medical sam adapter: Adapting segment anything model for medical image segmentation. arXiv 2023, arXiv:2304.12620. [Google Scholar] [CrossRef]
  110. Ma, J.; He, Y.; Li, F.; Han, L.; You, C.; Wang, B. Segment anything in medical images. Nat. Commun. 2024, 15, 654. [Google Scholar] [CrossRef] [PubMed]
  111. Diana-Albelda, C.; Alcover-Couso, R.; García-Martín, Á.; Bescos, J.; Escudero-Viñolo, M. GBT-SAM: A Parameter-Efficient Depth-Aware Model for Generalizable Brain tumour Segmentation on mp-MRI. arXiv 2025, arXiv:2503.04325. [Google Scholar]
  112. Ravi, N.; Gabeur, V.; Hu, Y.T.; Hu, R.; Ryali, C.; Ma, T.; Khedr, H.; Rädle, R.; Rolland, C.; Gustafson, L.; et al. Sam 2: Segment anything in images and videos. arXiv 2024, arXiv:2408.00714. [Google Scholar]
  113. Domadia, S.G.; Thakkar, F.N.; Ardeshana, M.A. Recent advancement in learning methodology for segmenting brain tumor from magnetic resonance imaging—A review. Multimed. Tools Appl. 2023, 82, 34809–34845. [Google Scholar] [CrossRef]
  114. Kurmukov, A.; Dalechina, A.; Saparov, T.; Belyaev, M.; Zolotova, S.; Golanov, A.; Nikolaeva, A. Challenges in building of deep learning models for glioblastoma segmentation: Evidence from clinical data. In Public Health and Informatics; IOS Press: Amsterdam, The Netherlands, 2021; pp. 298–302. [Google Scholar]
  115. Bonada, M.; Rossi, L.F.; Carone, G.; Panico, F.; Cofano, F.; Fiaschi, P.; Garbossa, D.; Di Meco, F.; Bianconi, A. Deep learning for MRI segmentation and molecular subtyping in glioblastoma: Critical aspects from an emerging field. Biomedicines 2024, 12, 1878. [Google Scholar] [CrossRef]
  116. Perkuhn, M.; Stavrinou, P.; Thiele, F.; Shakirin, G.; Mohan, M.; Garmpis, D.; Kabbasch, C.; Borggrefe, J. Clinical evaluation of a multiparametric deep learning model for glioblastoma segmentation using heterogeneous magnetic resonance imaging data from clinical routine. Investig. Radiol. 2018, 53, 647–654. [Google Scholar] [CrossRef]
  117. Bianconi, A.; Rossi, L.F.; Bonada, M.; Zeppa, P.; Nico, E.; De Marco, R.; Lacroce, P.; Cofano, F.; Bruno, F.; Morana, G.; et al. Deep learning-based algorithm for postoperative glioblastoma MRI segmentation: A promising new tool for tumor burden assessment. Brain Inform. 2023, 10, 26. [Google Scholar] [CrossRef]
  118. Cepeda, S.; Romero, R.; Luque, L.; García-Pérez, D.; Blasco, G.; Luppino, L.T.; Kuttner, S.; Esteban-Sinovas, O.; Arrese, I.; Solheim, O.; et al. Deep learning-based postoperative glioblastoma segmentation and extent of resection evaluation: Development, external validation, and model comparison. Neuro-Oncol. Adv. 2024, 6, vdae199. [Google Scholar] [CrossRef]
  119. Holtzman Gazit, M.; Faran, R.; Stepovoy, K.; Peles, O.; Shamir, R.R. Post-operative glioblastoma multiforme segmentation with uncertainty estimation. Front. Hum. Neurosci. 2022, 16, 932441. [Google Scholar] [CrossRef]
  120. Lotan, E.; Zhang, B.; Dogra, S.; Wang, W.; Carbone, D.; Fatterpekar, G.; Oermann, E.; Lui, Y. Development and practical implementation of a deep learning–based pipeline for automated pre-and postoperative glioma segmentation. Am. J. Neuroradiol. 2022, 43, 24–32. [Google Scholar] [CrossRef] [PubMed]
  121. Hochreuter, K.M.; Ren, J.; Nijkamp, J.; Korreman, S.S.; Lukacova, S.; Kallehauge, J.F.; Trip, A.K. The effect of editing clinical contours on deep-learning segmentation accuracy of the gross tumor volume in glioblastoma. Phys. Imaging Radiat. Oncol. 2024, 31, 100620. [Google Scholar] [CrossRef] [PubMed]
  122. Bertels, J.; Eelbode, T.; Berman, M.; Vandermeulen, D.; Maes, F.; Bisschops, R.; Blaschko, M.B. Optimizing the dice score and jaccard index for medical image segmentation: Theory and practice. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 92–100. [Google Scholar]
  123. Zhao, C.; Shi, W.; Deng, Y. A new Hausdorff distance for image matching. Pattern Recognit. Lett. 2005, 26, 581–586. [Google Scholar] [CrossRef]
  124. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [PubMed]
  125. Pontén, F.; Jirström, K.; Uhlen, M. The Human Protein Atlas—a tool for pathology. J. Pathol. J. Pathol. Soc. Great Br. Irel. 2008, 216, 387–393. [Google Scholar] [CrossRef]
  126. Weinstein, J.N.; Collisson, E.A.; Mills, G.B.; Shaw, K.R.; Ozenberger, B.A.; Ellrott, K.; Shmulevich, I.; Sander, C.; Stuart, J.M. The cancer genome atlas pan-cancer analysis project. Nat. Genet. 2013, 45, 1113–1120. [Google Scholar] [CrossRef]
  127. Pati, S.; Baid, U.; Zenk, M.; Edwards, B.; Sheller, M.; Reina, G.A.; Foley, P.; Gruzdev, A.; Martin, J.; Albarqouni, S.; et al. The federated tumor segmentation (fets) challenge. arXiv 2021, arXiv:2105.05874. [Google Scholar] [CrossRef]
  128. Satushe, V.; Vyas, V.; Metkar, S.; Singh, D.P. Advanced CNN Architecture for Brain Tumor Segmentation and Classification using BraTS-GOAT 2024 Dataset. Curr. Med. Imaging 2025, e15734056344235. [Google Scholar] [CrossRef] [PubMed]
  129. Bonato, B.; Nanni, L.; Bertoldo, A. Advancing precision: A comprehensive review of MRI segmentation datasets from brats challenges (2012–2025). Sensors 2025, 25, 1838. [Google Scholar] [CrossRef] [PubMed]
  130. Ghaffari, M.; Sowmya, A.; Oliver, R. Automated brain tumor segmentation using multimodal brain scans: A survey based on models submitted to the BraTS 2012–2018 challenges. IEEE Rev. Biomed. Eng. 2019, 13, 156–168. [Google Scholar] [CrossRef]
  131. Binney, N.; Hyde, C.; Bossuyt, P.M. On the origin of sensitivity and specificity. Ann. Intern. Med. 2021, 174, 401–407. [Google Scholar] [CrossRef]
  132. Müller, D.; Soto-Rey, I.; Kramer, F. Towards a guideline for evaluation metrics in medical image segmentation. BMC Res. Notes 2022, 15, 210. [Google Scholar] [CrossRef]
  133. LaBella, D.; Schumacher, K.; Mix, M.; Leu, K.; McBurney-Lin, S.; Nedelec, P.; Villanueva-Meyer, J.; Shapey, J.; Vercauteren, T.; Chia, K.; et al. Brain tumor segmentation (brats) challenge 2024: Meningioma radiotherapy planning automated segmentation. arXiv 2024, arXiv:2405.18383. [Google Scholar] [CrossRef]
  134. Moawad, A.W.; Janas, A.; Baid, U.; Ramakrishnan, D.; Saluja, R.; Ashraf, N.; Jekel, L.; Amiruddin, R.; Adewole, M.; Albrecht, J.; et al. The Brain Tumor Segmentation-Metastases (BraTS-METS) Challenge 2023: Brain Metastasis Segmentation on Pre-treatment MRI. arXiv 2024, arXiv:2306.00838v3. [Google Scholar]
  135. Kuntiyellannagari, B.; Dwarakanath, B. Glioma segmentation using hybrid filter and modified African vulture optimization. Bull. Electr. Eng. Inform. 2025, 14, 1447–1455. [Google Scholar] [CrossRef]
  136. Akil, M.; Saouli, R.; Kachouri, R. Fully automatic brain tumor segmentation with deep learning-based selective attention using overlapping patches and multi-class weighted cross-entropy. Med. Image Anal. 2020, 63, 101692. [Google Scholar]
  137. Chang, Y.; Zheng, Z.; Sun, Y.; Zhao, M.; Lu, Y.; Zhang, Y. DPAFNet: A residual dual-path attention-fusion convolutional neural network for multimodal brain tumor segmentation. Biomed. Signal Process. Control. 2023, 79, 104037. [Google Scholar] [CrossRef]
  138. Li, X.; Jiang, Y.; Li, M.; Zhang, J.; Yin, S.; Luo, H. MSFR-Net: Multi-modality and single-modality feature recalibration network for brain tumor segmentation. Med. Phys. 2023, 50, 2249–2262. [Google Scholar] [CrossRef] [PubMed]
  139. Rastogi, D.; Johri, P.; Donelli, M.; Kadry, S.; Khan, A.A.; Espa, G.; Feraco, P.; Kim, J. Deep learning-integrated MRI brain tumor analysis: Feature extraction, segmentation, and Survival Prediction using Replicator and volumetric networks. Sci. Rep. 2025, 15, 1437. [Google Scholar] [CrossRef]
  140. de los Reyes, A.M.; Lord, V.H.; Buemi, M.E.; Gandía, D.; Déniz, L.G.; Alemán, M.N.; Suárez, C. Combined use of radiomics and artificial neural networks for the three-dimensional automatic segmentation of glioblastoma multiforme. Expert Syst. 2024, 41, e13598. [Google Scholar] [CrossRef]
  141. Beser-Robles, M.; Castellá-Malonda, J.; Martínez-Gironés, P.M.; Galiana-Bordera, A.; Ferrer-Lozano, J.; Ribas-Despuig, G.; Teruel-Coll, R.; Cerdá-Alberich, L.; Martí-Bonmatí, L. Deep learning automatic semantic segmentation of glioblastoma multiforme regions on multimodal magnetic resonance images. Int. J. Comput. Assist. Radiol. Surg. 2024, 19, 1743–1751. [Google Scholar] [CrossRef] [PubMed]
  142. Amri, Y.; Slama, A.B.; Mbarki, Z.; Selmi, R.; Trabelsi, H. Automatic Glioma Segmentation Based on Efficient U-Net Model using MRI Images. Intell.-Based Med. 2025, 11, 100216. [Google Scholar] [CrossRef]
  143. Pinaya, W.H.; Tudosiu, P.D.; Gray, R.; Rees, G.; Nachev, P.; Ourselin, S.; Cardoso, M.J. Unsupervised brain imaging 3D anomaly detection and segmentation with transformers. Med. Image Anal. 2022, 79, 102475. [Google Scholar] [CrossRef]
  144. Hu, Z.; Li, L.; Sui, A.; Wu, G.; Wang, Y.; Yu, J. An efficient R-transformer network with dual encoders for brain glioma segmentation in MR images. Biomed. Signal Process. Control. 2023, 79, 104034. [Google Scholar] [CrossRef]
  145. Wen, L.; Sun, H.; Liang, G.; Yu, Y. A deep ensemble learning framework for glioma segmentation and grading prediction. Sci. Rep. 2025, 15, 4448. [Google Scholar] [CrossRef]
  146. Chen, X.; Yang, L. Brain tumor segmentation based on CBAM-TransUNet. In Proceedings of the 1st ACM Workshop on Mobile and Wireless Sensing for Smart Healthcare, Sydney, Australia, 21 October 2022; pp. 33–38. [Google Scholar]
  147. Hou, Q.; Peng, Y.; Wang, Z.; Wang, J.; Jiang, J. MFD-Net: Modality Fusion Diffractive Network for Segmentation of Multimodal Brain Tumor Image. IEEE J. Biomed. Health Inform. 2023, 27, 5958–5969. [Google Scholar] [CrossRef] [PubMed]
  148. Liu, C.; Kiryu, H. 3D medical axial transformer: A lightweight transformer model for 3D brain tumor segmentation. In Proceedings of the Medical Imaging with Deep Learning, Paris, France, 3–5 July 2024; pp. 799–813. [Google Scholar]
  149. Deng, G.; Zou, K.; Ren, K.; Wang, M.; Yuan, X.; Ying, S.; Fu, H. SAM-U: Multi-box prompts triggered uncertainty estimation for reliable SAM in medical image. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 368–377. [Google Scholar]
Figure 1. Clinical Applications of Glioma Segmentation. This diagram shows the medical applications of glioma segmentation, highlighting its role in: Computer-aided Diagnosis, Monitoring Disease Progression, Surgery (Planning and Robotic Assistance), and Radiotherapy Planning.
Figure 1. Clinical Applications of Glioma Segmentation. This diagram shows the medical applications of glioma segmentation, highlighting its role in: Computer-aided Diagnosis, Monitoring Disease Progression, Surgery (Planning and Robotic Assistance), and Radiotherapy Planning.
Jimaging 11 00269 g001
Figure 2. CNN Architecture. Utilizes convolution, pooling, and upsampling layers for feature extraction and segmentation.
Figure 2. CNN Architecture. Utilizes convolution, pooling, and upsampling layers for feature extraction and segmentation.
Jimaging 11 00269 g002
Figure 3. Pure Transformer Architecture. Incorporates multi-head attention blocks to capture long-range dependencies from volumetric image patches.
Figure 3. Pure Transformer Architecture. Incorporates multi-head attention blocks to capture long-range dependencies from volumetric image patches.
Jimaging 11 00269 g003
Figure 4. Hybrid CNN-Transformer Architecture. Merges CNN-based feature extraction with Transformer-based global contextual understanding.
Figure 4. Hybrid CNN-Transformer Architecture. Merges CNN-based feature extraction with Transformer-based global contextual understanding.
Jimaging 11 00269 g004
Figure 5. Glioma Structure. This figure displays segmented gliomas on MRI scans, where color coding is used to highlight different tumor regions: blue is the edema (ED), green denotes enhancing tumor (ET), and pink is necrotic cells (NCR). The joint of the ET and NCR areas represents the tumor core region (TC), while the union of the three segmented zones creates the whole tumor (WT).
Figure 5. Glioma Structure. This figure displays segmented gliomas on MRI scans, where color coding is used to highlight different tumor regions: blue is the edema (ED), green denotes enhancing tumor (ET), and pink is necrotic cells (NCR). The joint of the ET and NCR areas represents the tumor core region (TC), while the union of the three segmented zones creates the whole tumor (WT).
Jimaging 11 00269 g005
Figure 6. Dice Scores Across BraTS Datasets by Year. This figure shows Dice Scores for glioma segmentation within BraTS19, BraTS20, and BraTS23 datasets. Bubble size indicates the number of model trainable parameters, with color coding distinguishing between CNN (pink), Pure Transformer (green), and Hybrid CNN-Transformer (blue) architectures. This visualization highlights performance trends and model diversity over the years.
Figure 6. Dice Scores Across BraTS Datasets by Year. This figure shows Dice Scores for glioma segmentation within BraTS19, BraTS20, and BraTS23 datasets. Bubble size indicates the number of model trainable parameters, with color coding distinguishing between CNN (pink), Pure Transformer (green), and Hybrid CNN-Transformer (blue) architectures. This visualization highlights performance trends and model diversity over the years.
Jimaging 11 00269 g006
Table 1. Evolution of the BraTS Challenge Datasets. This table summarizes the annual distribution of training, validation, and testing cases within the BraTS benchmark from 2012 to 2024, illustrating the expansion and evolution of the dataset over time.
Table 1. Evolution of the BraTS Challenge Datasets. This table summarizes the annual distribution of training, validation, and testing cases within the BraTS benchmark from 2012 to 2024, illustrating the expansion and evolution of the dataset over time.
YearTotal CasesTrainValidationTest
20125035N/A15
20136035N/A25
2014238200N/A38
2015253200N/A53
2016391200N/A191
201747728546146
201854228566191
2019626335125166
2020660369125166
202120401251219570
202220001251219530
202320401251219570
202422001540220440
Table 2. CNN-based Approaches. This table compares CNN-based algorithms for glioma segmentation, grouped chronologically by dataset. It reports the Dice Scores for Whole Tumor (WT), Enhancing Tumor (ET), and Tumor Core (TC); the 95th percentile Hausdorff Distance (HD95); and the number of trainable parameters (in millions). The best-performing method per metric and dataset group (when more than one method is available) is highlighted in bold. Note: An asterisk (*) indicates that the number of parameters was not explicitly reported by the authors but was estimated based on the model architecture. ↑↓: Lower or upper values are better for each metric.
Table 2. CNN-based Approaches. This table compares CNN-based algorithms for glioma segmentation, grouped chronologically by dataset. It reports the Dice Scores for Whole Tumor (WT), Enhancing Tumor (ET), and Tumor Core (TC); the 95th percentile Hausdorff Distance (HD95); and the number of trainable parameters (in millions). The best-performing method per metric and dataset group (when more than one method is available) is highlighted in bold. Note: An asterisk (*) indicates that the number of parameters was not explicitly reported by the authors but was estimated based on the model architecture. ↑↓: Lower or upper values are better for each metric.
IdAlgorithmDice(WT)↑Dice(ET)↑Dice(TC)↑HD95↓#Params(M)↓Dataset
3Wang et al. [42]90.5078.5983.7816.5020.00 *Brats17
4Att. Unet [55]87.2474.5176.857.6134.90Brats18
6Myronenko [53]91.0082.3386.685.107.70 *Brats18
8DenseMultiOCM [136]86.0073.2073.338.7535.00 *Brats18
39DPAFNet [137]90.5079.5083.905.0528.00 *Brats18
41MSFR-Net [138]90.9080.7085.804.8237.85Brats18
13D U-Net [39]85.3672.2171.0512.9419.00Brats19
2V-Net [40]85.1872.4373.468.9237.70Brats19
7TuNet [41]90.3478.4281.124.7730.50 *Brats19
50Tong et al. [50]88.5075.1077.60 61.84Brats19
86AR-CA [120]83.0072.0084.00 15.00 *Brats19
5nnUnet [45]90.7081.4084.805.9541.20Brats20
9RFNet [46]86.9861.4778.23 35.00 *Brats20
12Modality-pairing [48]92.4086.3089.804.9242.00 *Brats20
17MM-BiFPN [51]83.5877.9581.47 21.41Brats20
20MAF-Net [47]88.0041.8067.90 75.00 *Brats20
26Dres-Unet [43]86.6080.0483.57 30.47Brats20
42MM-Unet [57]85.0076.2076.508.47111.40Brats20
43Liu et al. [49]89.5077.4581.786.4064.00 *Brats20
61Sahoo et al. [54]90.36 85.75 17.00 *Brats20
62GAM-Net [52]89.9175.8084.025.3088.00 *Brats20
78Kuntiyellannagari et al. [135]97.0091.0096.00 147.00 *Brats20
84Rastogi et al. [139]87.5686.4686.66 40.00 *Brats20
87GBManalizer [140]89.3380.6983.726.840.004Brats20
88RH-GlioSeg-nnU-Net [118]88.0078.0072.00 41.20 *Brats20
35MAAB [56]89.0778.0280.7319.5924.00 *Brats21
64Dense Unet+ [44]95.8093.7095.50 6.92Brats23
75Beser-Robles et al. [141]71.0081.0079.00 19.00Brats23
79Enhanced Unet [142]91.87 4.1246.00 *Brats23
Table 3. Pure Transformer Approaches. This table compares Transformer-based algorithms for glioma segmentation, grouped chronologically by dataset. It reports Dice Scores for Whole Tumor (WT), Enhancing Tumor (ET), and Tumor Core (TC); the 95th percentile Hausdorff Distance (HD95); and the number of trainable parameters (in millions). The best-performing method per metric and dataset group (only when more than one method is available) is highlighted in bold. Note: An asterisk (*) indicates that the number of parameters was not explicitly reported by the authors but was estimated based on the model architecture. ↑↓: Lower or upper values are better for each metric.
Table 3. Pure Transformer Approaches. This table compares Transformer-based algorithms for glioma segmentation, grouped chronologically by dataset. It reports Dice Scores for Whole Tumor (WT), Enhancing Tumor (ET), and Tumor Core (TC); the 95th percentile Hausdorff Distance (HD95); and the number of trainable parameters (in millions). The best-performing method per metric and dataset group (only when more than one method is available) is highlighted in bold. Note: An asterisk (*) indicates that the number of parameters was not explicitly reported by the authors but was estimated based on the model architecture. ↑↓: Lower or upper values are better for each metric.
IdAlgorithmDice(WT)↑Dice(ET)↑Dice(TC)↑HD95↓#Params(M)↓Dataset
28Pinaya et al. [143]61.70 48.91 *Brats18
77MMCFormer [65]85.0064.7079.20 8.57Brats18
27BTSwin-Unet [59]90.2878.3881.735.1835.60Brats19
29EMSViT [61]90.2879.2482.235.4947.50 *Brats19
56Swin-Unet [60]86.8975.6276.637.8927.10Brats19
45NestedFormer [58]92.2080.0086.405.0510.48Brats20
44VT-Unet [62]92.2486.3189.533.5187.00Brats21
58HRSTNet-4 [63]91.9082.9287.628.94266.33Brats23
72CR-Swin2-VT [64]91.3881.7185.409.9790.00 *Brats23
Table 4. Hybrid CNN-Transformer Approaches. This table compares Hybrid CNN-Transformer algorithms for glioma segmentation, ordered chronologically by dataset. It reports Dice Scores for Whole Tumor (WT), Enhancing Tumor (ET), and Tumor Core (TC); the 95th percentile Hausdorff Distance (HD95); and the number of trainable parameters in millions. The best-performing method per metric and dataset group (only when more than one method is available) is highlighted in bold. Note: An asterisk (*) indicates that the number of parameters was not explicitly reported by the authors but was estimated based on the model architecture. ↑↓: Lower or upper values are better for each metric.
Table 4. Hybrid CNN-Transformer Approaches. This table compares Hybrid CNN-Transformer algorithms for glioma segmentation, ordered chronologically by dataset. It reports Dice Scores for Whole Tumor (WT), Enhancing Tumor (ET), and Tumor Core (TC); the 95th percentile Hausdorff Distance (HD95); and the number of trainable parameters in millions. The best-performing method per metric and dataset group (only when more than one method is available) is highlighted in bold. Note: An asterisk (*) indicates that the number of parameters was not explicitly reported by the authors but was estimated based on the model architecture. ↑↓: Lower or upper values are better for each metric.
IdAlgorithmDice(WT)↑Dice(ET)↑Dice(TC)↑HD95↓#Params(M)↓Dataset
40ERTN [144]83.2073.5977.935.1395.00 *Brats17
16Zhou et al. [72]82.9059.1074.907.1037.00 *Brats18
46mmFormer [86]89.6085.8077.607.85106.00Brats18
47LGMSU-Net [91]87.35 69.02Brats18
53Tongxue Zhou [89]86.5087.0079.403.6036.00Brats18
10TransBTS [68]88.4277.5880.967.1730.60Brats19
14SegTran [70]89.5074.0081.70 93.10Brats19
15Fang et al. [71]92.6783.5489.471.9515.37Brats19
19Transition-Net [75]91.2574.8584.4614.1544.00 *Brats19
22TransConver [66]90.1978.4082.574.749.00Brats19
25TransBTSv2 [68]90.4280.2484.874.8715.30Brats19
55Gao et al. [94]90.1080.1084.004.735.90Brats19
57GMetaNet [95]90.2078.4082.504.846.10Brats19
73TransXAI [96]88.2074.5078.206.1987.00 *Brats19
13Medical Transformer [69]87.3358.8269.69 2.41Brats20
18UneTR [73]89.9078.8084.205.2592.58Brats20
23TranSiam [67]89.34 5.657.98Brats20
30Nalawade et al. [79]87.4072.1077.3027.0964.00 *Brats20
38RMTF-Net [85]81.80 59.00 *Brats20
51AST-Net [87]90.4077.8084.2014.2310.50Brats20
52Huang et al. [88]90.3070.8081.5015.99144.86Brats20
67TransDoubleU-Net [90]92.8779.1686.5110.7793.00 *Brats20
68GSG U-net [145]91.2885.8885.825.3960.00 *Brats20
76SwinBTS [2]95.0685.3683.3010.0364.00 *Brats20
80SGS-Net [101]89.7976.1276.29 85.00 *Brats20
82CFNet [74]91.6090.2990.461.5149.18Brats20
21Unet Former [76]93.2288.8092.108.4958.96Brats21
24SegTransVAE [77]90.5285.4892.604.1044.70Brats21
31Shi et al. [80]89.1581.9473.8113.09143.00 *Brats21
32BiTr-Unet [81]90.9781.8784.3413.0143.50Brats21
33Dobko et al. [82]86.9884.9692.5610.0594.00 *Brats21
34Futrega et al. [83]91.63 43.00 *Brats21
36Swin UneTR [3]92.6085.8088.505.2161.98Brats21
37COTRNet [84]89.3477.6080.2117.2446.51Brats21
48CBAM-TransUNet [146]93.0887.7691.494.0196.00Brats21
493D PSwinBTS [92]92.6482.6286.7210.7820.40Brats21
54AABTS-Net [93]91.1077.7083.804.4275.00 *Brats21
60Med-SA [109]88.70 9.5013.00Brats22
59SAM [104]74.60 27.51636.00Brats23
63SAM-Med2D [106]82.90 16.20636.00Brats23
65MFD-Net [147]92.7085.4088.707.76105.00 *Brats23
66SAMed [105]77.30 19.0718.81Brats23
69MAT [148]93.2185.0591.914.7711.70Brats23
70MedSAM [110]83.60 14.90636.00Brats23
71SAM-U [149]81.00 17.260.00Brats23
74Diana-Albelda et al. [38]61.90 32.0012.00Brats23
813D U-TFA [102]95.0685.3683.3010.0311.31 *Brats23
83GBT-SAM [111]93.54 6.40Brats23
85Arouse-Net [103]93.5089.3089.50 45.00 *Brats23
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Diana-Albelda, C.; García-Martín, Á.; Bescos, J. A Review on Deep Learning Methods for Glioma Segmentation, Limitations, and Future Perspectives. J. Imaging 2025, 11, 269. https://doi.org/10.3390/jimaging11080269

AMA Style

Diana-Albelda C, García-Martín Á, Bescos J. A Review on Deep Learning Methods for Glioma Segmentation, Limitations, and Future Perspectives. Journal of Imaging. 2025; 11(8):269. https://doi.org/10.3390/jimaging11080269

Chicago/Turabian Style

Diana-Albelda, Cecilia, Álvaro García-Martín, and Jesus Bescos. 2025. "A Review on Deep Learning Methods for Glioma Segmentation, Limitations, and Future Perspectives" Journal of Imaging 11, no. 8: 269. https://doi.org/10.3390/jimaging11080269

APA Style

Diana-Albelda, C., García-Martín, Á., & Bescos, J. (2025). A Review on Deep Learning Methods for Glioma Segmentation, Limitations, and Future Perspectives. Journal of Imaging, 11(8), 269. https://doi.org/10.3390/jimaging11080269

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop