Next Article in Journal
An Interaction–Engagement–Intention Model: How Artificial Intelligence and Augmented Reality Transform the User–Platform Interaction Paradigm
Previous Article in Journal
Risk Analysis Method of Aviation Critical System Based on Bayesian Networks and Empirical Information Fusion
Previous Article in Special Issue
Optimized AI Methods for Rapid Crack Detection in Microscopy Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Glioblastoma Multiforme Diagnosis: Semantic Segmentation and Survival Modeling Using MRI and Genotypic Data

1
Department of Management Information Systems, National Chung Hsing University, Taichung 402202, Taiwan
2
Department of Neurosurgery, Neurological Institute, Taichung Veterans General Hospital, Taichung 407219, Taiwan
*
Authors to whom correspondence should be addressed.
Electronics 2025, 14(12), 2498; https://doi.org/10.3390/electronics14122498
Submission received: 3 May 2025 / Revised: 15 June 2025 / Accepted: 18 June 2025 / Published: 19 June 2025
(This article belongs to the Special Issue Image Segmentation, 2nd Edition)

Abstract

:
Glioblastoma multiforme (GBM) is the most aggressive and common primary brain tumor. Magnetic resonance imaging (MRI) provides detailed visualization of tumor morphology, edema, and necrosis. However, manually segmenting GBM from MRI scans is time-consuming, subjective, and prone to inter-observer variability. Therefore, automated and reliable segmentation methods are crucial for improving diagnostic accuracy. This study employs an image semantic segmentation model to segment brain tumors in MRI scans of GBM patients. The MRI recall images include T1-weighted imaging (T1WI) and fluid-attenuated inversion recovery (FLAIR) sequences. To enhance the performance of the semantic segmentation model, image preprocessing techniques were applied before analyzing and comparing commonly used segmentation models. Additionally, a survival model was constructed using discrete genotype attributes of GBM patients. The results indicate that the DeepLabV3+ model achieved the highest accuracy for semantic segmentation, with an accuracy of 77.9% on T1WI image sequences, while the U-Net model achieved 80.1% accuracy on FLAIR image sequences. Furthermore, in constructing the survival model using a discrete attribute dataset, the dataset was divided into three subsets based on different missing value handling strategies. This study found that replacing missing values with 1 resulted in the highest accuracy, with the Bernoulli Bayesian model and the multinomial Bayesian model achieving an accuracy of 94.74%. This study integrates image preprocessing techniques and semantic segmentation models to improve the accuracy and efficiency of brain tumor segmentation while also developing a highly accurate survival model. The findings aim to assist physicians in saving time and facilitating preliminary diagnosis and analysis.

1. Introduction

Glioblastoma multiforme (GBM), also known simply as glioblastoma, is the most aggressive and most common primary brain tumor in adults. GBM originates from astrocytes, the star-shaped glial cells in the brain and spinal cord, and is classified by the World Health Organization as a Grade IV glioma—a designation reserved for the most malignant tumors [1]. GBM accounts for 48.6% of all malignant tumors in the central nervous system. This high percentage highlights GBM as the most prevalent central nervous system cancer [2]. The median overall survival for GBM patients is low at just 15 months [3]. Malignant gliomas (with GBM being a major subset) are responsible for 2.5% of all cancer-related deaths, indicating GBM’s broader impact on public health [4].
Magnetic resonance imaging (MRI) is a non-invasive medical imaging technique that uses a strong magnetic field and radio waves to generate signals from hydrogen atoms within the body. These signals are then processed by a computer to produce high-resolution images of organs and tissues. In the diagnostic process, MRI plays a crucial role in establishing a differential diagnosis during a patient’s initial consultation. It can be used to accurately identify the location and size of brain tumors. MRI scans can reveal characteristics of GBM, and commonly used MRI sequences include T1-weighted imaging (T1WI), T2-weighted imaging, fluid-attenuated inversion recovery (FLAIR), and T1-weighted contrast-enhanced [5,6]. The T1WI is a sequence that primarily reflects the T1 relaxation times of tissues. This sequence provides basic anatomical information and helps distinguish between different brain tissues. The T2-weighted imaging, on the other hand, is based on T2 relaxation times. It is sensitive to water content, making edema and certain parts of the tumor appear bright. FLAIR is a special type of T2-weighted imaging with an added inversion recovery pulse that suppresses free fluid signals, especially from cerebrospinal fluid. FLAIR is particularly valuable for segmenting edematous regions, which often contain infiltrative tumor cells beyond the enhancing core. T1-weighted contrast-enhanced adds a gadolinium-based contrast agent that shortens the T1 relaxation time of tissues where it accumulates. It shows strong enhancement in the tumor core region.
Image segmentation, the process of dividing an image into meaningful regions, plays a crucial role in the quantitative analysis of brain tumors. Specifically, segmentation of brain tumor MRI involves identifying and delineating tumor boundaries, which is essential for measuring tumor volume, monitoring growth over time, and planning surgical interventions. However, manual segmentation by radiologists is time-consuming, subjective, and prone to inter-observer variability, especially given the complexity and heterogeneity of brain tumors [7]. These limitations highlight the need for automated, robust, and accurate segmentation methods to assist clinicians in making informed decisions.
Early efforts in brain tumor segmentation relied heavily on traditional image processing techniques, such as thresholding, region growing, and clustering [8]. While these methods offer simplicity and interpretability, they are often sensitive to noise, artifacts, and heterogeneous appearances in MRI scans, resulting in inconsistent performance across datasets. To overcome the limitations of traditional methods, researchers turned to machine learning approaches, which offer more adaptive and data-driven solutions. These methods typically involve two stages: feature extraction and classification. Features such as texture, intensity, and spatial information are extracted from MRI images and fed into classifiers like support vector machine (SVM) or random forest (RF). For instance, Bahadure et al. [9] proposed an SVM-based method using texture and intensity features for tumor segmentation, achieving significant improvements over traditional methods. However, these techniques require elaborate feature engineering, which is both time-consuming and dependent on expert knowledge. Moreover, handcrafted features may fail to capture the complex patterns present in brain tumor images, limiting their generalization capability.
With the rise of deep learning, particularly following the success of AlexNet [10] in 2012 in image classification tasks, convolutional neural networks (CNNs) began to be applied in medical image analysis. Unlike earlier approaches, CNNs can automatically learn hierarchical features directly from raw image data, eliminating the need for manual feature design. In 2015, the introduction of U-Net [11] marked a milestone in brain tumor segmentation. U-Net, specifically designed for medical image segmentation, uses an encoder–decoder architecture with skip connections, enabling it to effectively capture fine-grained features. U-Net has become the cornerstone of brain tumor segmentation and has inspired numerous adaptations.
The evolution of medical image segmentation has progressed from manual feature design to traditional machine learning-based automated classification and, ultimately, to deep learning methods featuring end-to-end automatic feature learning and segmentation. Although the emergence of deep learning has significantly improved the accuracy and automation of segmentation tasks, its performance still varies across different medical imaging datasets. Traditional methods continue to hold value in specific applications or when integrated as part of hybrid approaches. The application of image processing techniques can enhance the performance of deep learning models.
The application of image processing techniques can enhance the performance of deep learning models. Ullah et al. [12] evaluated various preprocessing methods, including bias field correction (BFC), intensity normalization, histogram equalization, and Gibbs ringing artifact removal (GRAR), to assess their impact on tumor segmentation in brain MRI images. The results showed that these preprocessing techniques significantly improved the segmentation accuracy of the 3D U-Net model. Hussein et al. [13] proposed using contrast limited adaptive histogram equalization (CLAHE) to enhance the contrast of X-ray images, followed by classification using a CNN. Their approach effectively increased the accuracy of lung disease classification. Similarly, Marciniak et al. [14] assessed the impact of CLAHE on the classification of retinal abnormalities in optical coherence tomography images. Their findings indicated that CLAHE preprocessing enhanced deep learning classification accuracy, with improvements of up to 4.75%.
Therefore, the objective of this study is to enhance the segmentation performance of deep learning models on GBM MRI scans by integrating image processing techniques with deep learning architectures. The methodology involves using MRI sequences as the dataset. Preprocessing is performed using several image enhancement techniques, including adaptive histogram equalization (AHE), intensity normalization, BFC, and GRAR. Subsequently, deep learning segmentation models (U-Net, U-Net++, DeepLabV3+, and LinkNet) are applied to segment brain tumors. The accuracy of these models is evaluated to assist clinicians in making more effective preliminary diagnoses.
Furthermore, the study incorporates discrete patient attribute data and applies machine learning techniques, including decision tree (DT), RF, extreme gradient boosting (XGBoost), SVM, multi-layer perceptron (MLP), K-nearest neighbors (KNN), and naive Bayes (NB), to analyze gene expression and anti-cancer drug-related attributes, aiming to predict patient survival outcomes and support clinical decision-making.
The primary contributions of this study are delineated as follows:
  • This study explores the combination of image processing techniques and deep learning models for segmenting GBM MRI images and finds that such integration can significantly improve segmentation performance.
  • For T1WI, the highest Dice coefficient of 0.779 is achieved when applying Intensity Normalization as preprocessing followed by the DeepLabV3+ model. For FLAIR imaging, the highest Dice coefficient of 0.801 is obtained when using GRAR preprocessing combined with the U-Net model.
  • The DeepLabV3+ model shows Dice coefficient improvements of 0.156 and 0.149 on T1WI and FLAIR images, respectively, when using intensity normalization as preprocessing.
  • Using gene data from GBM patients, this study predicts survival outcomes. When missing values are replaced with 1, the NB model achieves an accuracy of 0.9474.

2. Related Works

2.1. Medical Image Features

Medical image features refer to quantifiable attributes extracted from imaging modalities such as MRI, computed tomography, X-ray, or ultrasound. These features (such as color, texture, shape, and grayscale intensity) are essential for assisting clinicians in tasks like tumor identification and tissue structure evaluation. Prior studies have demonstrated that such features play a crucial role in medical image analysis, typically categorized into intensity, shape, and texture features [15]. Intensity features describe the distribution of pixel values in the image and are often used to differentiate tissues or detect lesions. Metrics like mean, maximum, minimum, standard deviation, skewness, and kurtosis are frequently used to summarize pixel intensity [16]. Shape features focus on the geometric properties and boundary characteristics of lesions or organs, providing information such as volume, surface area, aspect ratio, and boundary irregularity [17]. Texture features capture the spatial distribution and regularity of grayscale values, aiding in distinguishing between different tissue types. Common methods for extracting texture features include the gray-level co-occurrence matrix, gray-level run-length matrix, gray-level size zone matrix, and wavelet transform analysis [17,18].
These features can be extracted manually (referred to as hand-crafted features) or learned automatically using deep learning models. With the advancement of deep learning, especially CNNs, it has become possible to automatically learn more abstract and discriminative features from large-scale datasets. These deep learning-based features, which emerge through layers of convolution and pooling, may be less interpretable but tend to outperform traditional features in complex segmentation tasks.

2.2. Medical Image Segmentation

Historically, due to computational limitations, early medical image segmentation methods relied on basic techniques such as thresholding, region growing, edge detection, and clustering [8]. Thresholding segments images based on pixel intensity, while region growing starts from a seed point and groups neighboring pixels with similar properties. Edge detection methods like the Sobel and Canny algorithms identify sharp intensity changes that indicate boundaries, and clustering approaches such as K-means group pixels based on similarity in intensity or position.
With the emergence of machine learning, segmentation shifted toward model-based classification using features extracted from image regions or pixels. Classifiers such as SVM and RF were commonly applied to predict pixel labels. These traditional machine learning approaches had advantages in interpretability and required less training data, but they were limited by their reliance on manually designed features and their difficulty in handling complex or highly variable image content.
Since 2012, deep learning has revolutionized medical image segmentation, largely replacing traditional methods due to its superior performance and automation capabilities. Deep learning models can learn hierarchical, high-level features directly from raw image data without the need for manual design. For instance, in the application of deep learning methods to retinal vessel segmentation, the presence of numerous capillaries and complex tree-like structures in retinal vessels makes it difficult for traditional segmentation techniques to perform effectively [19]. Therefore, deep learning approaches are necessary to improve segmentation accuracy in such cases. Luu et al. [20] developed a deep learning model based on an improved U-Net architecture for brain tumor segmentation in MRI images, achieving top performance in the Brain Tumor Segmentation 2021 challenge. This demonstrates how modern deep learning approaches have become the dominant technique in medical image segmentation, delivering substantial improvements over classical methods. Ali et al. [21] investigated the application of deep learning in diagnosing neurological disorders using brain MRI images. The results demonstrated that deep learning models outperformed traditional machine learning methods in terms of classification accuracy, especially when dealing with high-dimensional and complex feature data.

2.3. Deep Learning for Brain Tumor Image Segmentation

The application of deep learning in medical image segmentation has become a research hotspot in recent years, especially for processing complex medical images such as MRI, computed tomography, and X-rays. Fully convolutional networks (FCNs) [22], introduced in 2015, replaced the fully connected layers in traditional CNNs with convolutional layers, enabling end-to-end pixel-level prediction. This means that the output of an FCN is a segmentation map of the same size as the input image, where each pixel is assigned a class label. Later, U-Net was developed as an improvement upon the FCN and became particularly suitable for high-precision segmentation in medical imaging.
Chen et al. [23] and Noori et al. [24] used U-Net for brain tumor image segmentation. In addition, Islam and colleagues [25] not only used U-Net for segmenting brain tumor images but also employed it to predict patient survival. These studies demonstrate that U-Net is highly suitable for brain tumor image segmentation. The name “U-Net” originates from the U-shaped architecture formed by its encoder and decoder paths. This design has proven particularly effective in medical image segmentation and other applications requiring precise pixel-level classification.
In the related literature, Cui et al. [26] applied U-Net++ [27] for brain tumor segmentation in CT images. Wu et al. [28] used U-Net++ for the automatic segmentation of pelvic organs after hysterectomy, while Micallef, Seychell, and others [29] utilized U-Net++ for automated brain tumor image segmentation. These studies show that U-Net++ is an effective model for segmenting brain tumor MRI images.
Roy Choudhury et al. [30] employed DeepLabV3+ [31] for brain tumor image segmentation. Wang [32] used both DeepLabV3+ and U-Net for leaf image segmentation and classification. Meanwhile, Liu, Zhang, and colleagues [33] used DeepLabV3+ in a study on bladder cancer image segmentation. These examples suggest that DeepLabV3+ is a viable model for brain tumor image segmentation.
Regarding research on LinkNet [34], Sobhaninia et al. [35] and Naz et al. [36] adopted LinkNet as the model architecture for analyzing brain tumor images. Hemanth and colleagues [37] used LinkNet for brain tumor object detection. These findings indicate that the LinkNet model is also applicable for brain tumor image segmentation.

2.4. Digital Image Processing

One of the primary tasks in digital image processing is image enhancement and restoration, as digital images are often affected by noise, blurring, and other distortions. Image processing techniques are therefore particularly crucial. Given the widespread application of digital imaging in medical contexts, image processing plays an essential and irreplaceable role in improving image quality [38].
AHE [39] is a computer image processing technique used to enhance local contrast in images. It works by computing histograms of specified regions and redistributing the image brightness values based on those histograms, thereby modifying the image’s contrast. This method is particularly suited for improving local contrast and enhancing image edges to reveal more details. While AHE can also reduce noise, it may degrade the visual quality of MRI images.
Intensity normalization [40] is a method that maps image intensity values into a standardized, predefined range. During the MRI acquisition process, different scanners or scanning parameters may be used at different times to scan different subjects or even the same subject. This can lead to significant variations in intensity, which severely affect MRI image consistency. To mitigate such variability, intensity normalization is applied. This method leverages geometric moments of images to find a set of parameters that eliminate the effects of various transformations, converting the original raw image into a unique standard form. This often involves affine transformations such as translation, rotation, and scaling, to normalize images to the same dimensional scale or equivalent intensity range.
BFC [41] is an image processing technique primarily used to correct intensity inhomogeneity in images. In medical imaging (especially MRI) non-uniform brightness may arise due to hardware limitations of the scanner, physiological factors of the patient, or other unknown causes. Such inconsistencies result in the same tissue appearing with different grayscale values across the image. Bias field correction serves as a preprocessing step in MRI to correct these distortions. In this study, the method used is N4ITK, which was proposed by Tustison, to perform bias field correction.
GRAR [42] is a method for eliminating artifacts in MRI images. Gibbs artifacts are typically the result of using the Fourier transform to reconstruct MRI signals into images. These artifacts appear as false oscillations or ripples near regions with abrupt intensity changes, such as edges or interfaces. To eliminate Gibbs ringing from MRI images, the local sub-voxel shift technique is employed.

3. Materials and Methods

The hospital provides data of GBM patients, including MRI and genetic data. The MRI data is filtered and annotated to form the raw dataset, which then undergoes four image processing techniques: AHE, intensity normalization, BFC, and GRAR. Each technique generates a corresponding dataset. The datasets are then subjected to image segmentation using four semantic segmentation models: U-Net, U-Net++, DeepLabV3+, and LinkNet.
Additionally, the genetic data is filtered to create a usable dataset, with missing values replaced by 0, 1, and 2, creating three datasets. These three datasets are then used with various machine learning classification models to predict patient survival. The machine learning models include DT [43], RF [44], XGBoost [45], SVM [46], MLP [47], KNN [48], and NB [49]. NB includes Gaussian naive Bayes (GNB), multinomial naive Bayes (MNB) and Bernoulli naive Bayes (BNB). The research workflow is shown in Figure 1.

3.1. Datasets

The dataset used in this study is sourced from the Taichung Veterans General Hospital in Taiwan (IRB No: CF17263B-3), which provided MRI scans of patients with glioblastoma. The dataset includes MRI scans from 69 patients, with each MRI consisting of T1WI and FLAIR image sequences. Each sequence contains 1695 images, totaling 3390 images.
The inclusion criteria for patients were being at least 20 years of age and having a diagnosis of GBM, regardless of gender. The exclusion criteria included patients under the age of 20, those with concurrent tumors or other specific medical conditions, and members of vulnerable populations (including children, minors, pregnant women, individuals with mental disorders, critically ill patients, and those with behavioral disorders).
The hospital also provided genetic data, including discrete attribute data on the patients’ gender, survival status, use of the anti-cancer drug Avastin, and the presence of genes such as VEGF, IDH, hTERT, MGMT, p53, and p21. A total of 222 patients are included in this dataset.
The following are some genes or factors associated with GBM.
  • VEGF (vascular endothelial growth factor) [50] primarily binds to receptors on vascular endothelial cells (VEGF receptors, VEGFR), thereby promoting tumor growth.
  • IDH (isocitrate dehydrogenase) [51] refers to a group of enzymes divided into IDH1 and IDH2. Mutations in IDH are associated with the development of glioblastoma.
  • hTERT (human telomerase reverse transcriptase) [52] is commonly considered an oncogene when mutated.
  • MGMT (O6-methylguanine-DNA methyltransferase) [53] is a DNA repair gene. MGMT is often involved in promoter methylation, which can interfere with the DNA repair process.
  • p53 [54] refers to a family of homologous proteins known as tumor suppressors.
  • p21 [55], also known as CDKN1A (cyclin-dependent kinase inhibitor 1A), is a tumor suppressor protein involved in regulating the cell cycle.
The presence and status of these GBM-related genes or factors are typically revealed in a patient’s genetic test results.

3.2. Screening and Labeling of MRI Images

The MRI data underwent a preprocessing pipeline, including image selection and tumor annotation, to prepare for model training. MRI scans consist of a series of consecutive slices captured from the top of the head downward. However, not all slices clearly depict brain tumors or associated edema, which are critical for effective model training. To streamline subsequent research, this study manually selected slices that clearly show brain tumors or edema to construct a curated dataset.
The selected MRI slices were annotated using LabelMe, an open-source tool that supports drawing polygons, rectangles, circles, lines, and points. LabelMe offers versatile functionality and can be used for data annotation tasks such as classification, object detection, semantic segmentation, and instance segmentation. In this study, LabelMe was used to manually delineate the contours of brain tumors and edema, producing precise annotations to support downstream research and model development.

3.3. Screening and Handling Missing Values of Genetic Data

In this study, the patients’ tabular data was also filtered prior to building the survival prediction models to ensure data usability. The dataset obtained from the hospital contains information on 222 patients. However, many fields contained missing values. After filtering, the missing values were replaced with 0, 1, and 2, respectively, in order to construct usable datasets for model training.

3.4. Image Processing

Image processing techniques play a crucial and indispensable role in enhancing the quality of medical imaging. In this study, MRI is used as the dataset, including T1WI and FLAIR image sequences. Four image processing techniques (AHE, intensity normalization, BFC, and GRAR) are applied to both T1WI and FLAIR sequences. These preprocessing steps are essential for the subsequent development of image segmentation models.

3.4.1. AHE

AHE divides the image into multiple regions and applies histogram equalization independently to each tile. Afterward, bilinear interpolation is performed to smooth the boundaries between adjacent regions.
The new intensity value I is presented in Formula (1):
I i = C D F i C D F m i n C D F m a x C D F m i n × L 1 ,
where i is the pixel value of the image, C D F i is the cumulative distribution function (CDF) of pixel value i , C D F m i n is the CDF of the minimum pixel value, C D F m a x is the CDF of the maximum pixel value, and L is the total number of pixels.

3.4.2. Intensity Normalization

The calculation method of intensity normalization is as follows (2).
I = X I m i n I m a x I m i n ,
where I is the new intensity value, X is the original intensity value, I m a x is the maximum intensity value, and I m i n is the minimum intensity value.

3.4.3. BFC

The N4ITK algorithm operates on the assumption that the observed MRI image v is modeled as shown in Formula (3):
v = u f + n ,
where u is the true underlying image without bias, f is the multiplicative bias field, n is additive noise.
Assuming the noise is negligible after log transformation, the model simplifies to shown in Formula (4):
v ^ = u ^ + f ^ ,
where v ^ = log v , u ^ = log u , and f ^ = log f .
The corrected image is then obtained by subtracting the estimated bias field from the observed image, as shown in Formula (5):
u ^ n = v ^ n f ^ n ,
where n denotes the n -th iteration.
This process continues until convergence criteria are met, typically when changes between iterations fall below a predefined threshold.

3.4.4. GRAR

The core idea of GRAR is to reduce the Gibbs effect by averaging the image after subtle spatial shifts. This method of local sub-pixel shifting and averaging can effectively eliminate Gibbs artifacts while preserving edge information.
The new intensity value I is presented in Formula (6):
I = F T 1 F T I x · G x + F T I y · G y ,
where F T · denotes the Fourier transform, I x is the corrected image in the x dimension, I y is the corrected image in the y dimension, G x is the weighting function in the x dimension, and   G y is the weighting function in the x dimension.
G x and G y are as shown in Formula (7):
G x = 1 + cos k y 1 + cos k y + 1 + cos k x ,     G y = 1 + cos k x 1 + cos k x + 1 + cos k y ,
where k y is the k radian of the y -axis, and k x is the k radian of the x -axis. When G x + G y = 1 , the artifacts can be eliminated.

3.5. Semantic Segmentation

This study employs four semantic segmentation models, including U-Net, U-Net++, DeepLabV3+, and LinkNet.

3.5.1. U-Net

The architecture of U-Net (Figure 2) is structured as a symmetric, it comprises two main components: a contracting path and an expansive path. The contracting path functions similarly to a traditional convolutional network. It involves repeated application of two 3 × 3 convolutions, each followed by a rectified linear unit (ReLU) activation, and then a 2 × 2 max pooling operation with stride 2 for downsampling. As the network goes deeper, the number of feature channels doubles, enabling the model to capture increasingly abstract and complex features. On the other hand, the expansive path mirrors the structure of the contracting path but in reverse. It begins with a 2 × 2 up-convolution (also known as a transposed convolution) that reduces the number of feature channels by half. Each upsampling step is followed by a concatenation with the corresponding feature map from the contracting path, facilitated by skip connections. These skip connections are critical, as they reintroduce spatial information lost during downsampling. After concatenation, the model applies two more 3 × 3 convolutions followed by ReLU activations, continuing the pattern from the encoder. The final layer is a 1 × 1 convolution, which maps the feature representation to the desired number of classes, typically resulting in a pixel-wise segmentation output.

3.5.2. U-Net++

The fundamental idea in U-Net++ (Figure 3) is to gradually evolve feature maps before they are passed to the decoder. Instead of simply copying encoder features to the decoder through a single skip connection, U-Net++ uses intermediate convolution layers that sit between the encoder and decoder blocks. These layers progressively refine the encoder’s output at each spatial resolution, resulting in a series of convolution blocks forming short paths and deeper convolution chains that improve feature alignment. Each of these intermediate nodes receives features not only from its immediate predecessor but also from earlier layers, creating a dense network of connections across the architecture.

3.5.3. DeepLabV3+

The encoder in DeepLabV3+ employs Xception, where atrous convolutions are applied to extract dense features from the input image. To further aggregate multi-scale context, it incorporates the atrous spatial pyramid pooling (ASPP) module. ASPP consists of parallel atrous convolutions with different dilation rates, enabling the model to capture features at various scales simultaneously. This spatial pyramid of filters helps the network become more sensitive to objects of different sizes in a single forward pass, as shown in Figure 4.

3.5.4. LinkNet

What sets LinkNet apart is how it handles the decoding phase. Instead of relying on complex mechanisms to recover spatial resolution, LinkNet uses simple yet effective decoder blocks that are directly linked to their corresponding encoder blocks. This one-to-one linkage is where the model gets its name. The decoder receives two sources of information: the upsampled output from the previous decoder layer and a bypassed, compressed version of the encoder’s output from the same resolution level. These bypass connections are not raw feature maps but are first passed through 1 × 1 convolutions, reducing their channel dimensionality and helping preserve essential spatial features without overwhelming the decoder. Each decoder block then combines the upsampled features with these lightweight encoder features, followed by a few convolution operations to refine the result. This fusion process is repeated at each scale until the spatial resolution is restored to match the original input size. As shown in Figure 5.

4. Results and Discussion

4.1. The Results of Screening and Labeling of MRI Images

The dataset obtained from the hospital included 1695 T1WI sequence images and 1695 FLAIR sequence images from 69 patients. Through manual curation and selection, we compiled a usable dataset by retaining only the images that clearly captured brain tumors or edema. As a result, the final dataset consists of 433 T1WI sequence images and 433 FLAIR sequence images from the same 69 patients.

4.2. The Results of Screening and Handling Missing Values of Genetic Data

The genetic data obtained from the hospital originally contained data from 222 patients. If a record had missing values in more than two fields, it was excluded from the dataset. As a result, data from only 61 patients were retained after filtering. The missing values were replaced with 0, 1, and 2, where 0 indicates absence, 1 indicates presence, and 2 indicates unknown. As a result, three gene datasets were created: Gene_0, Gene_1, and Gene_2.

4.3. Image Processing Results

This study utilized MRI image sequences of brain tumor patients, including T1WI and FLAIR sequences. To enhance the performance of subsequent model training, in addition to using the original images, four types of image preprocessing were applied to the MRI sequences: AHE, intensity normalization, BFC, and GRAR.
For the T1WI sequence, the original image dataset is referred to as the T1_OG dataset. After applying the four preprocessing methods to the original dataset, the following datasets were generated: T1_AHE, T1_Norm, T1_BFC, and T1_GRAR. For the FLAIR sequence, the original image dataset is referred to as the FLAIR_OG dataset. Similarly, after applying the four preprocessing methods, the following datasets were generated: FLAIR_AHE, FLAIR_Norm, FLAIR_BFC, and FLAIR_GRAR.

4.4. Semantic Segmentation Results

In this study, four semantic segmentation models (U-Net, U-Net++, DeepLabV3+, and LinkNet) were used. The hyperparameters were set as follows: batch size = 8, epochs = 50, optimizer = “Adam”, learning rate = 1 × 10−4, and the loss function was defined as a combination of 50% binary cross-entropy loss and 50% Dice loss. The dataset was split into training, validation, and testing sets in a 5:2:3 ratio. Evaluation metrics included Dice coefficient, intersection over union (IoU), recall, and precision. Each image in the testing set was evaluated, and the results were averaged to obtain the final performance scores.

4.4.1. Comparison Results of Model Performance for T1WI Sequence

The four semantic segmentation models were tested on five different T1WI datasets. The results are presented in Table 1.
As shown in Table 1, the test results on the T1_OG dataset indicate that the DeepLabV3+ model achieved the highest Dice coefficient, reaching 0.623. On the T1_AHE dataset, the U-Net model performed best, with a Dice coefficient of 0.631. For the T1_Norm dataset, the DeepLabV3+ model again achieved the highest Dice coefficient, reaching 0.779. On the T1_BFC dataset, the DeepLabV3+ model outperformed the others, with a Dice coefficient of 0.684. Among all the T1WI datasets, the highest Dice coefficient was observed in the T1_Norm dataset using the DeepLabV3+ model.
Specifically, in the T1_Norm dataset, DeepLabV3+ achieved the highest Dice coefficient of 0.779, while the second-best model was U-Net with a Dice coefficient of 0.709. To evaluate whether the accuracy of DeepLabV3+ was statistically superior to that of U-Net, a paired t-test was conducted. With a significance level (α) set at 0.05, the paired t-test yielded a t-value of 3.4759 and a p-value of 6.9414 × 10−4. Since p < α, the difference between the two models is statistically significant. Therefore, the DeepLabV3+ model’s accuracy is statistically better than that of the U-Net model.
We also examined the effect of preprocessing on the T1-weighted images and found that the greatest improvement in Dice coefficient occurred when intensity normalization was applied in combination with the DeepLabV3+ model, resulting in a performance gain of 0.156.
In addition, during the training of the DeepLabV3+ model on the T1_Norm dataset, we observed the training and validation loss curves, as shown in Figure 6. Starting from epoch 15, the training loss continued to decrease, while the validation loss plateaued. The increasing gap between the two curves indicates that the model was becoming increasingly specialized in fitting the training data, with limited improvement on the validation data. This suggests that the model showed signs of slight overfitting.

4.4.2. Comparison Results of Model Performance for FLAIR Sequence

The four semantic segmentation models were tested on five different FLAIR datasets. The results are presented in Table 2.
The four semantic segmentation models were also tested on five different FLAIR image datasets. The results are shown in Table 2. According to the results, the U-Net model consistently achieved the highest Dice coefficient across all datasets. Specifically, it reached 0.717 on the FLAIR_OG dataset, 0.739 on the FLAIR_AHE dataset, 0.781 on the FLAIR_Norm dataset, 0.725 on the FLAIR_BFC dataset, and 0.801 on the FLAIR_GRAR dataset—the highest among all.
We also examined the impact of preprocessing on the FLAIR images and found that the greatest improvement in Dice coefficient was observed when intensity normalization was combined with the DeepLabV3+ model, resulting in a performance gain of 0.149.
In the FLAIR_GRAR dataset, U-Net achieved the highest Dice coefficient of 0.801, followed by DeepLabV3+ with a Dice coefficient of 0.724. A paired t-test was conducted to verify whether the accuracy of the U-Net model was statistically superior to that of DeepLabV3+. With the significance level α set at 0.05, the t-test produced a t-value of 6.0327 and a p-value of 1.5686 × 10−8. Since p < α, the difference between the two models is statistically significant, indicating that the U-Net model is statistically more accurate than the DeepLabV3+ model.
Additionally, during training with the FLAIR_GRAR dataset, the U-Net model’s training and validation loss curves were observed, as shown in Figure 7. The training loss steadily and consistently decreased, eventually falling below 0.1, indicating that the model learned well on the training data. The validation loss dropped sharply in the early epochs, reaching close to the training loss around epochs 5 to 10. However, from epoch 10 onward, the validation loss began to fluctuate and showed limited further improvement, eventually stabilizing at around 0.17. This behavior suggests that the model exhibited slight signs of overfitting.

4.4.3. Discussion on MRI Semantic Segmentation

From the above experimental results, we found that in the T1WI image sequence, the model with the highest accuracy was DeepLabV3+ combined with intensity normalization, achieving a Dice coefficient of 0.779. The average performance of the DeepLabV3+ model on T1WI images also outperformed other methods, indicating that DeepLabV3+ is a more effective model when using T1WI sequences for training. The semantic segmentation result of one T1WI image sequence using the DeepLabV3+ model is shown in Figure 8.
In the FLAIR image sequence, the highest accuracy was achieved by the U-Net model using the GRAR technique, with a Dice coefficient of 0.801. Similarly, the average performance of the U-Net model with FLAIR images was better than other methods, suggesting that U-Net is the more suitable model for training when FLAIR sequences are used. The semantic segmentation result of one FLAIR image sequence using the U-Net model is shown in Figure 9.
T1WI sequences provide high-resolution anatomical details, offering clear delineation of brain structures. DeepLabV3+ leverages atrous convolution and a robust encoder–decoder architecture, enabling it to capture multi-scale contextual information effectively. This makes it particularly adept at segmenting structures with well-defined boundaries, as commonly seen in T1WI images. Studies have shown that DeepLabV3+ achieves high accuracy and Dice coefficient when applied to T1WI sequences, highlighting its suitability for tasks requiring precise structural segmentation.
FLAIR sequences are designed to suppress cerebrospinal fluid signals, enhancing the visibility of lesions such as white matter hyperintensities. U-Net’s architecture, characterized by its symmetric encoder–decoder structure and skip connections, allows for the preservation of spatial information and fine details. This makes it highly effective in detecting and segmenting lesions that appear prominently in FLAIR images. Research indicates that U-Net models achieve superior segmentation performance on FLAIR sequences.
Moreover, when comparing the various image preprocessing techniques, we observed that using GRAR and intensity normalization as preprocessing methods led to performance improvements across all models. In previous literature, Ullah et al. [12] also demonstrated through experiments that GRAR is an important preprocessing technique that can enhance brain MRI segmentation performance.
When dealing with the complex task of brain tumor medical image segmentation, each model exhibits certain limitations. U-Net relies on transposed convolutions and skip connections for upsampling, but it may fail to capture sufficient detail in cases where tumor boundaries are indistinct or blurry. U-Net++ enhances representational capacity through nested skip connections across different resolutions; however, this complexity increases the risk of overfitting, especially when the medical imaging dataset is small. DeepLabV3+ incorporates atrous convolutions and an ASPP module to improve multi-scale feature extraction, but its performance remains unstable when segmenting very small or poorly defined tumor regions. LinkNet, lacking a multi-scale context module like ASPP, performs poorly in scenarios involving blurry tumor boundaries or multiple small tumor regions.

4.5. Survival Classification Results

In this study, nine survival classification models (DT, RF, XGBoost, SVM, MLP, KNN, GNB, MNB, and BNB) were trained and tested using genetic data. The hyperparameters were set as follows: DT used criterion = “gini”; RF used estimators = 100; XGBoost used estimators = 100; learning rate = 0.3; SVM used a linear kernel (kernel = “linear”); MLP had hidden layer sizes of (16, 5) with solver = “Adam” and activation = “logistic”; KNN used neighbors = 3; MNB used alpha = 1.0; and BNB used binarize = 3 and alpha = 1.0. The dataset was split into 70% for training and 30% for testing.
The results of survival models built using various classification algorithms, after replacing missing values with zero, are shown in Table 3. Among them, the Bernoulli Bayesian model achieved the highest accuracy and F1-score, reaching 0.8947 and 0.9444, respectively. Based on these two predictive metrics, it can be concluded that this model demonstrates the highest accuracy and relatively superior performance in predicting patient survival.
Table 4 presents the results of survival models built using various classification algorithms based on patient data where missing values were replaced with 1. It can be observed that both the Bernoulli Bayesian model and the multinomial Bayesian model achieved the highest Accuracy and F1-Score, reaching 0.9474 and 0.9730, respectively. Based on these two predictive indicators, it can be concluded that these models are the most accurate and relatively superior in predicting patient survival.
Table 5 presents the results of survival models built using various classification algorithms based on patient data where missing values were replaced with 2. It can be observed that the KNN model achieved the highest accuracy and F1-Score, reaching 0.9231 and 0.9600, respectively. Based on these two predictive indicators, it can be concluded that this model demonstrates the highest accuracy and relatively superior performance in predicting patient survival.
From the results above, it can be observed that replacing missing values with different values in genetic data has a significant impact on the analysis outcomes. When the missing values are replaced with 0 or 1, BNB achieves the best performance. However, when the missing values are replaced with 2, representing “unknown”, KNN yields the best results.

5. Conclusions

This study demonstrates that the choice of preprocessing methods significantly influences the performance of deep learning models in brain tumor segmentation. Specifically, the DeepLabV3+ model achieved the best segmentation accuracy on T1WI when combined with intensity normalization, while the U-Net model performed best on FLAIR images with GRAR. These findings highlight the importance of selecting appropriate combinations of models and preprocessing techniques to optimize performance in medical image segmentation tasks.
Preprocessing techniques such as GRAR and intensity normalization consistently improved segmentation outcomes across multiple models, including U-Net, U-Net++, DeepLabV3+, and LinkNet. However, the limited sample size in this study posed a challenge, increasing the risk of overfitting and reducing the model’s generalization ability. Future work can address this limitation through data augmentation or by leveraging Generative Adversarial Networks to expand the training data.
Moreover, traditional evaluation metrics like Dice coefficient may not fully capture the accuracy of tumor boundary and shape delineation. Incorporating additional metrics such as Hausdorff distance and average surface distance could offer a more comprehensive assessment of segmentation quality.
In the genetic data analysis, handling missing values remains a critical issue. Instead of simple binary imputation (e.g., 0 or 1), introducing a third category to represent “unknown” may improve classification accuracy. Additionally, integrating biological pathway analysis tools could provide deeper insights and enhance the biological relevance of survival predictions.
One key limitation of this study is that MRI images and genetic data were not obtained from the same patients, preventing the development of a multimodal model. Future research will focus on collecting both imaging and genetic data from individual patients to enable the construction of integrated models that can more accurately predict survival outcomes through multimodal learning.

Author Contributions

Conceptualization, M.-H.T., W.-Y.C., Y.-H.T. and B.-H.H.; methodology, Y.-H.T. and B.-H.H.; software, B.-H.H.; validation, Y.-H.T., W.-Y.C., B.-H.H. and C.-C.S.; formal analysis, Y.-H.T. and B.-H.H.; investigation, W.-Y.C. and B.-H.H.; resources, W.-Y.C. and C.-C.S.; data curation, W.-Y.C. and B.-H.H.; writing—original draft preparation, Y.-H.T. and B.-H.H.; writing—review and editing, Y.-H.T.; visualization, Y.-H.T. and B.-H.H.; supervision, M.-H.T.; project administration, M.-H.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Chung Hsing University and Taichung Veterans General Hospital grant number TCVGH-NCHU1137630 and TCVGH-NCHU1147627 and the APC was funded by National Chung Hsing University and Taichung Veterans General Hospital.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Institutional Review Board I & II of Taichung Veterans General Hospital (CF17263B-3) on 28 October 2020.

Informed Consent Statement

Informed consent for participation was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to individual privacy concerns.

Acknowledgments

We are grateful to Taichung Veterans General Hospital for providing the data of glioblastoma patients.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jiang, H.; Wang, X.; Chen, X.; Zhang, S.; Ren, Q.; Li, M.; Ren, X.; Lin, S.; Cui, Y. Unraveling the heterogeneity of WHO grade 4 gliomas: Insights from clinical, imaging, and molecular characterization. Discov. Oncol. 2025, 16, 111. [Google Scholar] [CrossRef] [PubMed]
  2. Grochans, S.; Cybulska, A.M.; Simińska, D.; Korbecki, J.; Kojder, K.; Chlubek, D.; Baranowska-Bosiacka, I. Epidemiology of glioblastoma multiforme–literature review. Cancers 2022, 14, 2412. [Google Scholar] [CrossRef]
  3. Koshy, M.; Villano, J.L.; Dolecek, T.A.; Howard, A.; Mahmood, U.; Chmura, S.J.; Weichselbaum, R.R.; McCarthy, B.J. Improved survival time trends for glioblastoma using the SEER 17 population-based registries. J. Neuro-Oncol. 2012, 107, 207–212. [Google Scholar] [CrossRef]
  4. Sabouri, M.; Dogonchi, A.F.; Shafiei, M.; Tehrani, D.S. Survival rate of patient with glioblastoma: A population-based study. Egypt. J. Neurosurg. 2024, 39, 42. [Google Scholar] [CrossRef]
  5. Wang, Y.; Xu, J.; Guan, Y.; Ahmad, F.; Mahmood, T.; Rehman, A. MSegNet: A multi-view coupled cross-modal attention model for enhanced MRI brain tumor segmentation. Int. J. Comput. Intell. Syst. 2025, 18, 63. [Google Scholar] [CrossRef]
  6. Shukla, G.; Alexander, G.S.; Bakas, S.; Nikam, R.; Talekar, K.; Palmer, J.D.; Shi, W. Advanced magnetic resonance imaging in glioblastoma: A review. Chin. Clin. Oncol. 2017, 6, 40. [Google Scholar] [CrossRef] [PubMed]
  7. Visser, M.; Müller, D.M.J.; Van Duijn, R.J.M.; Smits, M.; Verburg, N.; Hendriks, E.J.; de Munck, J.C. Inter-rater agreement in glioma segmentations on longitudinal MRI. NeuroImage Clin. 2019, 22, 101727. [Google Scholar] [CrossRef]
  8. Padmapriya, T.; Sriramakrishnan, P.; Kalaiselvi, T.; Somasundaram, K. Advancements of MRI-based brain tumor segmentation from traditional to recent trends: A review. Curr. Med. Imaging Rev. 2022, 18, 1261–1275. [Google Scholar]
  9. Bahadure, N.B.; Ray, A.K.; Thethi, H.P. Image analysis for MRI based brain tumor detection and feature extraction using biologically inspired BWT and SVM. Int. J. Biomed. Imaging 2017, 2017, 9749108. [Google Scholar] [CrossRef]
  10. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Comm. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  11. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  12. Ullah, F.; Ansari, S.U.; Hanif, M.; Ayari, M.A.; Chowdhury, M.E.H.; Khandakar, A.A.; Khan, M.S. Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net. Sensors 2021, 21, 7528. [Google Scholar] [CrossRef] [PubMed]
  13. Hussein, F.; Mughaid, A.; AlZu’bi, S.; El-Salhi, S.M.; Abuhaija, B.; Abualigah, L.; Gandomi, A.H. Hybrid CLAHE-CNN Deep Neural Networks for Classifying Lung Diseases from X-ray Acquisitions. Electronics 2022, 11, 3075. [Google Scholar] [CrossRef]
  14. Marciniak, T.; Stankiewicz, A. Impact of Histogram Equalization on the Classification of Retina Lesions from OCT B-Scans. Electronics 2024, 13, 4996. [Google Scholar] [CrossRef]
  15. Zhang, W.; Guo, Y.; Jin, Q. Radiomics and Its Feature Selection: A Review. Symmetry 2023, 15, 1834. [Google Scholar] [CrossRef]
  16. Vijithananda, S.M.; Jayatilake, M.L.; Hewavithana, B.; Gonçalves, T.; Rato, L.M.; Weerakoon, B.S.; Kalupahana, T.D.; Silva, A.D.; Dissanayake, K.D. Feature extraction from MRI ADC images for brain tumor classification using machine learning techniques. BioMed. Eng. Online 2022, 21, 52. [Google Scholar] [CrossRef]
  17. Nissar, A.; Mir, A.H. Proficiency evaluation of shape and WPT radiomics based on machine learning for CT lung cancer prognosis. Egypt. J. Radiol. Nucl. Med. 2024, 55, 50. [Google Scholar] [CrossRef]
  18. Zhang, H.; Huang, D.; Wang, Y.; Zhong, H.; Pang, H. CT radiomics based on different machine learning models for classifying gross tumor volume and normal liver tissue in hepatocellular carcinoma. Cancer Imaging 2024, 24, 20. [Google Scholar] [CrossRef]
  19. Wang, J.; Li, X.; Ma, Z. Multi-Scale Three-Path Network (MSTP-Net): A new architecture for retinal vessel segmentation. Measurement 2025, 250, 117100. [Google Scholar] [CrossRef]
  20. Luu, H.M.; Park, S.-H. Extending nn-UNet for brain tumor segmentation. In International MICCAI Brainlesion Workshop; Springer: Berlin/Heidelberg, Germany, 2022; pp. 173–186. [Google Scholar]
  21. Ali, S.S.A.; Memon, K.; Yahya, N.; Khan, S. Deep learning frameworks for MRI-based diagnosis of neurological disorders: A systematic review and meta-analysis. Artif. Intell. Rev. 2025, 58, 171. [Google Scholar] [CrossRef]
  22. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  23. Chen, W.; Liu, B.; Peng, S.; Sun, J.; Qiao, X. S3D-UNET: Separable 3D U-Net for brain tumor segmentation. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Quebec City, QC, Canada, 2019; Volume 11384, pp. 358–368. [Google Scholar]
  24. Noori, M.; Bahri, A.; Mohammadi, K. Attention-guided version of 2D UNet for automatic brain tumor segmentation. In Proceedings of the 2019 9th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 24–25 October 2019; pp. 269–275. [Google Scholar]
  25. Islam, M.; Vibashan, V.S.; Jose, V.J.M.; Wijethilake, N.; Utkarsh, U.; Ren, H. Brain Tumor Segmentation and Survival Prediction Using 3D Attention UNet. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Springer: New York, NY, USA, 2020; pp. 262–272. [Google Scholar]
  26. Cui, H.; Liu, X.; Huang, N. Pulmonary vessel segmentation based on orthogonal fused u-net++ of chest CT images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 293–300. [Google Scholar]
  27. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A nested UNet architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  28. Wu, S.; Wang, Z.; Liu, C.; Zhu, C.; Wu, S.; Xiao, K. Automatical segmentation of pelvic organs after hysterectomy by using dilated convolution u-net++. In Proceedings of the 2019 IEEE 19th International Conference on Software Quality, Reliability and Security Companion (QRS-C), Sofia, Bulgaria, 22–26 July 2019; pp. 362–367. [Google Scholar]
  29. Micallef, N.; Seychell, D.; Bajada, C.J. Exploring the U-Net++ Model for Automatic Brain Tumor Segmentation. IEEE Access 2021, 9, 125523–125539. [Google Scholar] [CrossRef]
  30. Roy Choudhury, A.; Vanguri, R.; Jambawalikar, S.R.; Kumar, P. Segmentation of brain tumors using DeepLabv3+. In Proceedings of the International MICCAI Brainlesion Workshop, Granada, Spain, 16 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 154–167. [Google Scholar]
  31. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  32. Wang, C.; Du, P.; Wu, H.; Li, J.; Zhao, C.; Zhu, H. A cucumber leaf disease severity classification method based on the fusion of DeepLabV3+ and U-Net. Comput. Electron. Agric. 2021, 189, 106373. [Google Scholar] [CrossRef]
  33. Liu, H.; Zhang, Q.; Liu, Y. Image Segmentation of Bladder Cancer Based on DeepLabv3+. In Proceedings of the 2021 Chinese Intelligent Systems Conference, Fuzhou, China, 16–17 October 2021; Springer: Singapore, 2022; Volume 805, pp. 614–621. [Google Scholar]
  34. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  35. Sobhaninia, Z.; Rezaei, S.; Noroozi, A.; Ahmadi, M.; Zarrabi, H.; Karimi, N.; Samavi, S. Brain tumor segmentation using deep learning by type specific sorting of images. arXiv 2018, arXiv:1809.07786. [Google Scholar]
  36. Naz, A.R.; Naseem, U.; Razzak, I.; Hameed, I.A. Deep AutoEncoder-Decoder Framework for Semantic Segmentation of Brain Deep AutoEncoder-Decoder Framework for Semantic Segmentation of Brain Tumor. Aust. J. Intell. Inf. Process. Syst. 2019, 15, 54–60. [Google Scholar]
  37. Hemanth, G.; Janardhan, M.; Sujihelen, L. Design and implementing brain tumor detection using machine learning approach. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019; pp. 1289–1294. [Google Scholar]
  38. Qu, C.; Tangge, Y.; Yang, G. Digital signal processing techniques for image enhancement and restoration. In Proceedings of the Third International Conference on Advanced Algorithms and Signal Image Processing (AASIP 2023), Kuala, Lumpur, 30 June–2 July 2023; SPIE: Bellingham, WA, USA, 2023; Volume 12799, pp. 663–668. [Google Scholar]
  39. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  40. Sun, X.; Shi, L.; Luo, Y.; Yang, W.; Li, H.; Liang, P.; Li, K.; Mok, V.C.T.; Chu, W.C.W.; Wang, D. Histogram-based normalization technique on human brain magnetic resonance images from different acquisitions. Biomed. Eng. Online 2015, 14, 73. [Google Scholar] [CrossRef]
  41. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef]
  42. Kellner, E.; Dhital, B.; Kiselev, V.G.; Reisert, M. Gibbs-ringing artifact removal based on local subvoxel-shifts. Magn. Reson. Med. 2016, 76, 1574–1581. [Google Scholar] [CrossRef] [PubMed]
  43. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  44. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  45. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar]
  46. Vapnik, V.; Chervonenkis, A. A note on class of perceptron. Autom. Remote Control 1964, 25, 103–109. [Google Scholar]
  47. Rumelhart, D.; Hinton, G.; Williams, R. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  48. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  49. Rish, I. An empirical study of the naive Bayes classifier. In Proceedings of the IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence, Seattle, WA, USA, 4 August 2001; Volume 3, pp. 41–46. [Google Scholar]
  50. Ribatti, D. Tumor refractoriness to anti-VEGF therapy. Oncotarget 2016, 7, 46668. [Google Scholar] [CrossRef]
  51. Cairns, R.A.; Mak, T.W. Oncogenic isocitrate dehydrogenase mutations: Mechanisms, models, and clinical opportunities. Cancer Discov. 2013, 3, 730–741. [Google Scholar] [CrossRef] [PubMed]
  52. Ding, X.; Nie, Z.; She, Z.; Bai, X.; Yang, Q.; Wang, F.; Geng, X.; Tong, Q. The regulation of ROS-and BECN1-mediated autophagy by human telomerase reverse transcriptase in glioblastoma. Oxidative Med. Cell. Longev. 2021, 2021, 6636510. [Google Scholar] [CrossRef]
  53. Hegi, M.E.; Diserens, A.C.; Gorlia, T.; Hamou, M.F.; De Tribolet, N.; Weller, M.; Kros, J.M.; Hainfellner, J.A.; Mason, W.; Mariani, L.; et al. MGMT gene silencing and benefit from temozolomide in glioblastoma. N. Engl. J. Med. 2005, 352, 997–1003. [Google Scholar] [CrossRef]
  54. Nigro, J.M.; Baker, S.J.; Preisinger, A.C.; Jessup, J.M.; Hosteller, R.; Cleary, K.; Signer, S.H.; Davidson, N.; Baylin, S.; Devilee, P.; et al. Mutations in the p53 gene occur in diverse human tumour types. Nature 1989, 342, 705–708. [Google Scholar] [CrossRef]
  55. El-Deiry, W.S.; Tokino, T.; Velculescu, V.E.; Levy, D.B.; Parsons, R.; Trent, J.M.; Lin, D.; Mercer, W.E.; Kinzler, K.W.; Vogelstein, B. WAF1, a potential mediator of p53 tumor suppression. Cell 1993, 75, 817–825. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions, and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions, or products referred to in the content.
Figure 1. The architecture of the research.
Figure 1. The architecture of the research.
Electronics 14 02498 g001
Figure 2. The architecture of the U-Net model [11].
Figure 2. The architecture of the U-Net model [11].
Electronics 14 02498 g002
Figure 3. The architecture of the U-Net++ model [27].
Figure 3. The architecture of the U-Net++ model [27].
Electronics 14 02498 g003
Figure 4. The architecture of the DeepLabV3+ model [31].
Figure 4. The architecture of the DeepLabV3+ model [31].
Electronics 14 02498 g004
Figure 5. The architecture of the LinkNet model [34].
Figure 5. The architecture of the LinkNet model [34].
Electronics 14 02498 g005
Figure 6. The training of the DeepLabV3+ model on the T1_Norm dataset.
Figure 6. The training of the DeepLabV3+ model on the T1_Norm dataset.
Electronics 14 02498 g006
Figure 7. The training of the U-Net model on the FLAIR_GRAR dataset.
Figure 7. The training of the U-Net model on the FLAIR_GRAR dataset.
Electronics 14 02498 g007
Figure 8. Semantic segmentation of a T1WI image sequence using DeepLabV3+. GT represents the ground truth and Pred represents the prediction.
Figure 8. Semantic segmentation of a T1WI image sequence using DeepLabV3+. GT represents the ground truth and Pred represents the prediction.
Electronics 14 02498 g008
Figure 9. Semantic segmentation of a FLAIR image sequence using U-Net. GT represents the ground truth and Pred represents the prediction.
Figure 9. Semantic segmentation of a FLAIR image sequence using U-Net. GT represents the ground truth and Pred represents the prediction.
Electronics 14 02498 g009
Table 1. The test results on T1WI images.
Table 1. The test results on T1WI images.
DatasetModelDiceIoURecallPrecision
T1_OGU-Net0.6090.4380.5930.625
U-Net++0.5060.3390.5320.483
DeepLabV3+0.6230.4520.5990.648
LinkNet0.4270.2710.4070.449
T1_AHEU-Net0.6310.4610.6180.644
U-Net++0.5250.3560.5010.552
DeepLabV3+0.6040.4330.6220.587
LinkNet0.4240.2690.4020.448
T1_NormU-Net0.7090.5490.7010.717
U-Net++0.5830.4110.5910.575
DeepLabV3+0.7790.6380.7490.812
LinkNet0.4270.2710.4100.445
T1_BFCU-Net0.6550.4870.6380.673
U-Net++0.5190.3500.4930.548
DeepLabV3+0.6840.520.6930.675
LinkNet0.4260.2710.4050.449
T1_GRARU-Net0.6810.5160.7020.661
U-Net++0.5380.3680.5420.534
DeepLabV3+0.7110.5520.7480.678
LinkNet0.4280.2720.4430.414
Table 2. The test results on FLAIR images.
Table 2. The test results on FLAIR images.
DatasetModelDiceIoURecallPrecision
FLAIR_OGU-Net0.7170.5590.7310.704
U-Net++0.5710.4000.5860.557
DeepLabV3+0.6310.4610.6630.602
LinkNet0.4630.3010.4460.481
FLAIR_AHEU-Net0.7390.5860.7600.719
U-Net++0.5720.4010.5510.595
DeepLabV3+0.6880.5240.6790.697
LinkNet0.4510.2910.4940.415
FLAIR_NormU-Net0.7810.6410.7640.799
U-Net++0.6260.4560.6580.597
DeepLabV3+0.7800.6390.7590.802
LinkNet0.5070.3400.5230.492
FLAIR_BFCU-Net0.7250.5690.7390.712
U-Net++0.6020.4310.6350.572
DeepLabV3+0.6820.5170.7270.642
LinkNet0.4720.3090.5030.445
FLAIR_GRARU-Net0.8010.6680.8260.777
U-Net++0.6400.4710.6690.613
DeepLabV3+0.7240.5670.7090.74
LinkNet0.4860.3210.4950.477
Table 3. The results of survival models for replacing missing values with zero.
Table 3. The results of survival models for replacing missing values with zero.
ModelAccuracyPrecisionRecallF1
DT0.68420.92310.70590.8000
RF0.73680.87500.82350.8495
XGBoost0.78950.93570.83330.8824
SVM0.84210.84210.86350.9143
MLP0.78950.84210.86350.9143
KNN0.76920.91330.90910.8696
GNB0.73680.77680.93330.8485
MNB0.84210.88890.94120.9143
BNB0.89470.89470.90210.9444
Table 4. The results of survival models for replacing missing values with one.
Table 4. The results of survival models for replacing missing values with one.
ModelAccuracyPrecisionRecallF1
DT0.73680.91250.86670.8367
RF0.78950.88330.93750.8824
XGBoost0.78680.92350.87350.8485
SVM0.84210.84210.86350.9143
MLP0.78950.85950.83330.8824
KNN0.84620.90620.93750.9167
GNB0.73680.93330.77780.8485
MNB0.94740.94740.93750.9730
BNB0.94740.94740.93750.9730
Table 5. The results of survival models for replacing missing values with two.
Table 5. The results of survival models for replacing missing values with two.
ModelAccuracyPrecisionRecallF1
DT0.63160.78570.73330.7586
RF0.73680.87500.82350.8485
XGBoost0.78680.86670.81250.8387
SVM0.78950.78950.85250.8824
MLP0.84210.84210.81330.9143
KNN0.92310.92730.91740.9600
GNB0.72680.86350.87500.8585
MNB0.84210.85210.83750.9143
BNB0.78850.82740.81750.8824
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsai, Y.-H.; Cheng, W.-Y.; Huang, B.-H.; Shen, C.-C.; Tsai, M.-H. Optimizing Glioblastoma Multiforme Diagnosis: Semantic Segmentation and Survival Modeling Using MRI and Genotypic Data. Electronics 2025, 14, 2498. https://doi.org/10.3390/electronics14122498

AMA Style

Tsai Y-H, Cheng W-Y, Huang B-H, Shen C-C, Tsai M-H. Optimizing Glioblastoma Multiforme Diagnosis: Semantic Segmentation and Survival Modeling Using MRI and Genotypic Data. Electronics. 2025; 14(12):2498. https://doi.org/10.3390/electronics14122498

Chicago/Turabian Style

Tsai, Yu-Hung, Wen-Yu Cheng, Bo-Hua Huang, Chiung-Chyi Shen, and Meng-Hsiun Tsai. 2025. "Optimizing Glioblastoma Multiforme Diagnosis: Semantic Segmentation and Survival Modeling Using MRI and Genotypic Data" Electronics 14, no. 12: 2498. https://doi.org/10.3390/electronics14122498

APA Style

Tsai, Y.-H., Cheng, W.-Y., Huang, B.-H., Shen, C.-C., & Tsai, M.-H. (2025). Optimizing Glioblastoma Multiforme Diagnosis: Semantic Segmentation and Survival Modeling Using MRI and Genotypic Data. Electronics, 14(12), 2498. https://doi.org/10.3390/electronics14122498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop