Abstract
Early detection of brain tumors is critical to ensure successful treatment, and medical imaging is essential in this process. However, analyzing the large amount of medical data generated from various sources such as magnetic resonance imaging (MRI) has been a challenging task. In this research, we propose a method for early brain tumor segmentation using big data analysis and patch-based convolutional neural networks (PBCNNs). We utilize BraTS 2012–2018 datasets. The data is preprocessed through various steps such as profiling, cleansing, transformation, and enrichment to enhance the quality of the data. The proposed CNN model utilizes a patch-based architecture with global and local layers that allows the model to analyze different parts of the image with varying resolutions. The architecture takes multiple input modalities, such as T1, T2, T2-c, and FLAIR, to improve the accuracy of the segmentation. The performance of the proposed model is evaluated using various metrics, such as accuracy, sensitivity, specificity, Dice similarity coefficient, precision, false positive rate, and true positive rate. Our results indicate that the proposed method outperforms the existing methods and is effective in early brain tumor segmentation. The proposed method can also assist medical professionals in making accurate and timely diagnoses, and thus improve patient outcomes, which is especially critical in the case of brain tumors. This research also emphasizes the importance of big data analysis in medical imaging research and highlights the potential of PBCNN models in this field.
MSC:
94A16; 68T07; 68T09; 68U10; 54H30; 54H30
1. Introduction
Brain tumors are a serious and potentially life-threatening condition that affects a significant number of people worldwide. According to the American Brain Tumor Association, the number of new cases of primary brain tumors was expected to reach over 87,000 in the United States in 2021. The early detection and accurate segmentation of brain tumors are critical for successful treatment and improved patient outcomes [1]. Early and accurate detection is important in ensuring successful treatment and improving patient outcomes [2]. MRI is a key medical imaging tool that provides detailed information about the location and size of the tumor [3]. However, analyzing the large amount of medical data generated from various sources has been a challenging task for medical professionals. Manual brain tumor segmentation is a time-consuming process that can be prone to errors, especially in cases where tumors have irregular shapes, sizes, and appearances [4]. Therefore, there is a need for automated methods for brain tumor segmentation that can improve the efficiency and accuracy of the process. In the last decade, deep learning-based approaches, especially convolutional neural networks (CNNs), have shown promise in medical image analysis and tumor segmentation [5,6]. However, many existing CNN-based approaches have limitations such as low accuracy, lack of interpretability, and inability to handle multi-modal data [7]. To address these limitations, a patch-based CNN is proposed for brain tumor segmentation using big data analysis. In this research, we utilize a dataset, including BraTS 2012–2018, which consists of thousands of MRI scans. This dataset includes a variety of brain tumor types and provides a broad range of images with different levels of complexity, making them ideal for training and validating our proposed method. The proposed CNN model utilizes a patch-based architecture that allows the model to analyze different parts of the image with varying resolutions. This architecture takes in multiple input modalities, such as T1, T2-c, T2, and fluid-attenuated inversion recovery (FLAIR), to improve the accuracy of the segmentation. Four primary MRI sequences are used in diagnosing and evaluating brain tumors. T1 provides information on the brain’s anatomy, while T2 indicates the distribution of water in brain tissue. T2-c uses a contrast agent to highlight areas where the blood–brain barrier is disrupted, as is the case in tumor-affected regions. FLAIR suppresses fluid signals. Moreover, in this study we present a big data analysis approach that includes data profiling, cleansing, transformation, and enrichment to enhance the quality of the data and improve the accuracy of the model. This study highlights the existing issues in CNN-based methods for brain tumor segmentation. Additionally, our proposed method includes a big data analysis approach that enhances the quality of the medical data, thus improving the accuracy of the system. The outcome of the proposed model is evaluated using various metrics such as Dice similarity coefficient, accuracy, sensitivity, specificity, precision, true positive rate, and false positive rate. The results show that the proposed model outperforms the existing methods and is effective in early segmentation of brain tumors. We also provide a detailed analysis of the strengths and weaknesses of the proposed method and compare it with existing approaches to highlight its advantages and limitations. The proposed method can also assist medical professionals in making accurate and timely diagnoses, and thus improve patient outcomes, which is especially critical in the case of brain tumors. This research emphasizes the importance of big data analysis in medical imaging research and highlights the potential of patch-based CNN models in this field. The proposed work utilizes a patch-based convolutional neural network for accurate brain tumor segmentation from MRI scans.
- A big data analysis approach is employed for data preprocessing and enrichment, enhancing the quality of the medical data and improving the accuracy of the model.
- An early detection method and comparative analysis are provided to assist medical professionals in making accurate and timely diagnoses, thereby improving patient outcomes.
- The proposed method is compared with existing approaches and its advantages and limitations are highlighted.
The study emphasizes the importance of big data analysis in medical imaging research and highlights the potential of patch-based CNN models in this field.
Our proposed model provides a powerful and efficient tool for the early classification and precise segmentation of brain tumors, which can significantly improve patient outcomes and precious lives.
2. Background
In this section, we review the most recent literature on brain tumor segmentation, patch-based convolutional neural networks, big data analysis, and multi-modal brain tumor segmentation using a patch-based CNN.
2.1. Brain Tumor Segmentation
Segmentation of brain tumors is a challenging task in medical image analysis due to the heterogeneity of tumors and the complexity of brain anatomy. Accurate segmentation of brain tumors can assist clinicians in identifying the location, size, and shape of tumors, as well as assessing their growth and response to treatment [8]. Various approaches have been suggested for segmentation of brain tumors, including traditional image processing techniques and machine learning-based approaches [9]. The CNN is a popular machine learning-based technique for brain tumor segmentation. CNNs have shown promising results in several studies, achieving high accuracy in tumor segmentation. For example, Kamnitsas et al. [10] suggested a 3D-CNN architecture for multi-modal brain tumor segmentation, achieving state-of-the-art results on the BraTS dataset. Similarly, Isensee et al. [11] developed a U-net based self-adapting model for medical image segmentation, achieving higher accuracy on the BraTS dataset. However, the heterogeneity of tumors remains a significant challenge in brain tumor segmentation. To address this challenge, several studies have proposed multi-modal imaging techniques that combine different MRI sequences to provide complementary information about tumor characteristics. For example, Bakas et al. [8] used a multi-modal CNN approach that combined T1, T2, T2-contrast, and FLAIR MRI images to achieve state-of-the-art performance on the BraTS 2017 dataset. Another challenge in segmentation of brain tumors is the class imbalance problem. In the images, the volume of non-tumor voxels is significantly greater than that of tumor voxels. To address this challenge, several studies have proposed methods such as data augmentation by Sandfort et al. [12], weighted loss functions by Bakas et al. [13], and postprocessing techniques by Wang et al. [14] to improve segmentation accuracy. In addition to CNN-based methods, several other machine learning-based approaches have been proposed for segmentation brain tumors, including support vector machine (SVM) by Kofler et al. [15] and random forests by Zikic et al. [16]. These methods have shown good performance on various datasets but have been largely overshadowed by the recent success of CNNs. Despite the promising results of machine learning-based approaches, they also have limitations. Obtaining labeled data is one of the primary obstacles in medical imaging due to the scarcity of annotated datasets, which makes it difficult to gather the significant amounts of labeled data necessary for training machine learning models. To address this challenge, several studies have proposed methods such as transfer learning by Havaei et al. [17] and data augmentation by Wang et al. [14] to improve model performance with limited training data. You et al. [18] discuss the use of sequential monitoring of patients’ health-related quality of life scores to detect disease.
In conclusion, brain tumor segmentation is an active research area with many challenges and opportunities for improvement. Machine learning-based approaches, particularly CNNs, have shown promise in achieving high accuracy in brain tumor segmentation. However, further research is needed to address the challenges of heterogeneity, class imbalance, and limited data availability, as well as to develop methods for more efficient and interpretable segmentation.
2.2. Patch-Based Convolutional Neural Network
The patch-based convolutional neural network (CNN) is a popular deep learning technique used in computer vision applications, including, but not limited to, image classification, object detection, and segmentation [19]. The fundamental concept of the patch-based convolutional neural network (CNN) involves partitioning the input image into smaller patches, which are individually processed using a CNN. The output from each patch is then combined to generate the final prediction for the input image. One of the earliest applications of a patch-based CNN was object detection, where the approach was used to generate object proposals. In this method, the input image is partitioned into patches, and each patch is evaluated by a CNN to produce a detection score for the presence of an object. The patches with high scores were then combined to generate object proposals, which were further refined by a second CNN. The patch-based convolutional neural network (CNN) has also demonstrated potential in medical image analysis, exhibiting favorable outcomes in tasks such as classification and segmentation. For example, a patch-based CNN was used for brain tumor segmentation from MRI scans. By surpassing traditional machine learning and other deep learning techniques, the patch-based convolutional neural network (CNN) achieved state-of-the-art results on a benchmark dataset. Furthermore, numerous studies have established the efficacy of patch-based CNNs in tasks such as brain tumor detection and segmentation. For example, Glioma Segmentation Challenge (BraTS) datasets have been used in various studies for evaluating patch-based CNN approaches for brain tumor analysis, such as the 3D U-Net approach proposed by Çiçek et al. [20] and the DeepMedic approach proposed by Kamnitsas et al. [10]. Both studies reported promising results for brain tumor segmentation using patch-based CNNs. In addition, some studies have proposed new modifications to the basic patch-based CNN architecture to improve its performance in brain tumor analysis. For example, Isensee et al. [11] proposed the nnU-Net architecture in 2018, which uses a cascade of patch-based CNNs with increasing resolution to improve the accuracy of brain tumor segmentation. Another modification was proposed by Zhang et al. [21], who introduced a separable and dilated residual U-net for MRI brain tumor segmentation [4]. Overall, patch-based CNNs have shown great potential for brain tumor analysis in medical imaging, providing a powerful and flexible tool for accurately and efficiently processing complex and large medical images.
In conclusion, the patch-based CNN is a useful approach for handling large and varying image sizes. It has shown promising results in various computer vision tasks, including object detection, medical image analysis, and remote sensing. However, the selection of patch size and overlap can significantly affect the performance of the approach, and additional processing is required to combine the outputs from each patch. Therefore, a patch-based CNN should be used with caution and with the appropriate tuning of its hyperparameters.
2.3. Big Data Analysis
The use of big data analytics in healthcare has become increasingly important in recent years, with the potential to revolutionize medical research and clinical practice. In the field of brain tumor research, big data analysis has been used to improve our understanding of the disease, predict patient outcomes, and make treatment decisions. Previous studies have shown that machine learning algorithms can be used to analyze large datasets of brain tumor imaging data and extract meaningful features that are predictive of patient outcomes. For example, Wu et al. [22] used a CNN to analyze MRI data from 563 patients with glioma, and found that the CNN was able to accurately predict patient survival outcomes. Similarly, Wulczyn et al. [23] used a deep learning model to analyze MRI data from 475 patients with glioblastoma, and found that the model was able to predict patient survival outcomes and treatment response. Other studies have used big data analysis to identify new biomarkers and potential therapeutic targets for brain tumors. For example, Sathornsumetee et al. [24] used a machine learning algorithm to analyze gene expression data from over 2000 glioblastoma patients and identified a set of genes that were associated with patient survival outcomes. Similarly, Molinaro et al. [25] used a machine learning algorithm to analyze genomic and clinical data from over 10,000 patients with glioblastoma, and identified sub-types of the disease that were associated with different treatment responses. The use of big data analysis in brain tumor research has also led to the development of new tools and resources for clinicians and researchers. For example, The Cancer Genome Atlas (TCGA) project has generated a comprehensive dataset of genomic and clinical data from over 12,000 patients with various types of cancer, including brain tumors (The Cancer Genome Atlas Research Network, 2008). This dataset has been used by researchers to identify new genetic mutations and pathways that are associated with brain tumors, and has led to the development of new diagnostic and therapeutic tools. Despite the potential of big data analysis in brain tumor research, there are also significant challenges and limitations. One major challenge is the lack of standardization and interoperability of data across different institutions and data sources. This can make it difficult to combine and compare data from different studies, and may limit the generalizability of findings. Additionally, the ethical and legal implications of using big data in healthcare are also complex and require careful consideration to ensure patient privacy and data security. Big data analysis has the potential to significantly improve our understanding of brain tumors and inform clinical practice. However, further research is needed to overcome the challenges and limitations of using big data in healthcare, and to ensure that the benefits of this approach are realized for patients and clinicians alike.
2.4. Multi-Modal Brain Tumor Segmentation using a Patch-Based CNN
Multi-modal brain tumor segmentation is an important problem in medical image analysis, as it plays a critical role in treatment planning and monitoring for brain tumor patients. Several studies have proposed patch-based CNN models for multi-modal brain tumor segmentation. The patch-based approach involves dividing the input image into small patches and training the CNN to predict the tumor label for each patch. The predicted labels are then merged to obtain the final segmentation map. This approach has the advantage of reducing the memory requirements and computational time of the model. Havaei et al. [17] present a deep-learning model that combines both patch-based and full CNNs to segment brain tumors using T1-weighted, T2-weighted, and FLAIR MRI modalities. While the hybrid approach used in the model is innovative and shows promising results, the authors themselves acknowledge that the model requires a large amount of training data and may not be practical for use in clinical settings with limited data. Furthermore, some have raised concerns about the interpretability of the model, as deep neural networks can sometimes act as “black boxes” with little insight into how decisions are being made. Despite these limitations, the model represents an important contribution to the field of brain tumor segmentation and provides a valuable foundation for future research. Similarly, Myronenko et al. [26] proposed a 3D patch-based CNN with a dilated convolutional architecture for brain tumor segmentation. They used a multi-scale approach and incorporated contextual information to improve the segmentation performance. Their proposed model achieved state-of-the-art results on the BraTS 2017 dataset. Recently, some studies have also incorporated other modalities such as diffusion tensor imaging (DTI) and spectroscopy (MRS) in addition to conventional MRI modalities to improve the segmentation accuracy. For example, Nie et al. [27] proposed a multi-modal patch-based CNN that combines DTI and MRS with conventional MRI modalities for brain tumor segmentation. Their proposed model achieved the best results on the BraTS 2018 dataset.
The Background section highlights limitations and gaps in the existing literature and explains how the proposed study aims to address these limitations, as shown in Table 1. Additionally, it also includes the potential applications and benefits of accurate brain tumor segmentation.
Table 1.
Comparative analysis of different brain tumor segmentation techniques.
3. Methodology
In this research, we proposed a PBCNN approach for brain tumor segmentation using a big data analysis approach as shown in Figure 1.
Figure 1.
Proposed Brain Tumor Segmentation using Patch-based CNN.
3.1. Data Acquisition
The primary objective of data acquisition is to obtain the medical imaging data that will be used in research. In the context of brain tumor segmentation, this involves obtaining MRI scans, CT scans, and other types of medical imaging that provide detailed images of the brain. There are several ways to obtain medical imaging data for research purposes. For example, researchers can collaborate with hospitals or medical facilities to obtain access to the required data. This may involve obtaining ethical approval, ensuring patient privacy, and complying with relevant laws and regulations. Researchers can also access public databases or repositories that contain medical imaging data from previous studies.
These databases can provide a large and diverse dataset that can be used to train machine learning algorithms. Once the data are obtained, they must be reviewed to ensure that they meet the requirements of the study. This includes checking that the data are in a suitable format for the machine learning algorithms, are of good quality, and in sufficient quantity to ensure accurate results. The imaging data may also need to be anonymized or de-identified to protect patient privacy. In this research, we used the BraTS dataset, which is widely used as a benchmark dataset for brain tumor segmentation. It has been released annually since 2012, with each release adding new data and challenges. A detailed description of BraTS datasets 2012 [28], 2013 [29], 2014 [30], 2015 [31], 2016 [32], 2017 [33], and 2018 [34] is given in Table 2.
Table 2.
Description of BraTS datasets from 2012 to 2018.
The BraTS dataset is compiled from MRI scans of patients with brain tumors, contributed by multiple medical centers worldwide. The dataset includes anonymized scans and corresponding manual segmentations prepared by experienced radiologists. In total, we used seven datasets from 2012 to 2018, with a total of 564 images. The collected datasets include 5106 images of both HGG and LGG. Each patient in the dataset is represented by four MRI images and ground truth: T1, T1-c, T2, FLAIR volumes and Ground truth, as illustrated in Figure 2.
Figure 2.
Sample MRI Images of BraTS dataset (a) T1, (b) T1-c, (c) T2, (d) FLAIR, (e) Ground truth.
3.2. Data Preprocessing
Data preprocessing is a crucial step in the process of building a brain tumor segmentation model. This step aims to prepare the collected MRI images for analysis by removing noise and artifacts, correcting for any biases in the data, and ensuring that the images are in a standardized format. Different preprocessing and normalization techniques were used to achieve higher accuracy. The first step in data preprocessing is to enhance the quality of the images by removing any noise, artifacts, or other distortions that may affect the accuracy of the analysis. Filtering techniques i.e., Gaussian filtering and median filtering, were applied to the images to improve their quality. MRI images may have different intensity levels due to variations in the imaging process. Intensity normalization techniques are applied to bring the intensities of the images into a comparable range, making it easier to compare images from different sources. In Table 3, we show real data from the BraTS dataset to provide more precise information on the images that were discarded during preprocessing. We have included the image ID, dataset, reason for discarding, discard method (whether it was carried out automatically), and image quality (rated as poor, fair, or good).
Table 3.
Discarded images with reasons, discard method, and image quality ratings during preprocessing.
3.3. Patch Extraction
Patch extraction is the process of dividing the preprocessed brain MRI images into smaller regions or patches. The purpose of this step is to reduce the computational cost of training the deep learning model and to allow the model to learn from small, local features in the images. Typically, each patch is a small, square, or rectangular sub-image of the brain MRI with a fixed size, such as 64 × 64 or 128 × 128 pixels. The patch size is chosen based on the size of the tumor and the available computational resources. In general, larger patches can capture more global features of the brain, while smaller patches can capture more local features. To ensure that the entire brain is covered, the patches are extracted with a sliding window approach, which means that adjacent patches overlap by a certain amount. The amount of overlap is usually half the patch size, which ensures that every pixel in the brain is covered by at least two patches. Patch extraction is a crucial step in brain tumor segmentation because it allows the deep learning model to learn from local image features and reduces the computational cost of training the model. In this study, patches were extracted from the preprocessed brain MRI scans to feed the convolutional neural network. The patch extraction process was implemented by using SimpleITK, which provides a Python interface for medical image processing. The patches were typically saved as separate image files or as NumPy arrays, which were fed into the deep learning model during training. The attributes of the BraTS dataset including description, input value, and output value are given in Table 4. Table 5 includes the unique identifier for each image in the dataset, the year of the dataset, the patch size used for extracting patches from the image, the stride used for moving the patch window over the image, and the number of patches extracted from the image. The patch size and stride were fixed at 64 × 64 × 64 and 32 × 32 × 32, respectively, for all images. The number of patches extracted varied between images and ranged from 224 to 289.
Table 4.
Attributes of the BraTS dataset with description, input and output value.
Table 5.
Datasets with image ID, patch size, stride, and patch count.
3.4. Training and Testing
The training and testing stages are crucial components of the methodology for brain tumor segmentation using patch-based CNNs. During the training stage, the CNN learns to identify and segment brain tumors from MRI images. Once the CNN is trained, it can be applied to new images during the testing stage to generate tumor segmentations. During the training stage, the patches and corresponding tumor segmentations are used to train the CNN. The patches are fed into the CNN, and the output is compared to the ground truth tumor segmentation using a loss function, such as the Dice coefficient. The weights of the CNN are then adjusted using back propagation to minimize the loss. One challenge during the training stage is preventing over fitting, where the CNN becomes too specialized to the training dataset and performs poorly on new data. To prevent over fitting, various regularization techniques can be employed, such as dropout and weight decay.
Once the CNN is trained, it can be applied to new images during the testing stage to generate tumor segmentations. The testing images are divided into patches, and the CNN is applied to each patch. The resulting tumor segmentations are then merged to generate the final tumor segmentation. The performance of the CNN during the testing stage can be evaluated using various metrics, such as the Dice coefficient, sensitivity, and specificity. The Dice coefficient is a commonly used metric that measures the overlap between the predicted and ground truth tumor segmentations. One challenge during the testing stage is dealing with variations in the input images, such as differences in resolution, intensity, and artifacts. To mitigate these challenges, the preprocessing steps can be carefully designed to ensure consistency across the dataset.
It is common to divide the dataset into training and testing sets in a ratio of 70:30 or 80:20, respectively. This allows sufficient training data to be available while ensuring that the model’s performance can be reliably evaluated on unseen data.
Table 6 shows, TP, FP, TN, and FN, representing true positives, false positives, true negatives, and false negatives, respectively. These metrics are commonly used to evaluate the performance of brain tumor segmentation models.
Table 6.
Performance evaluation metrics for brain tumor segmentation using patch-based CNNs.
3.5. Patch-Based Convolutional Neural Network (PBCNN)
In this research, we propose a PBCNN architecture for the multi-modal segmentation of brain tumors in MRI images using a big data analysis approach. The PBCNN architecture consists of several key steps, including patch extraction, convolutional layers, pooling layers, and fully connected layers. During the patch extraction step, we extract small sub-images (i.e., patches) from the original image using a sliding window technique. The patches are then passed through a series of convolutional layers that perform feature extraction by convolving the patches with a set of learnable filters. The resulting feature maps are then passed through pooling layers to reduce the dimensionality of the feature maps and summarize the information in each feature map. After pooling, the features are classified using fully connected layers, which are used to reduce the dimensionality of the feature maps and classify the output into different categories as shown in Table 7. Overall, the PBCNN architecture is a powerful model for multi-modal segmentation tasks on MRI images, and it has shown promising results in various applications, including brain tumor segmentation. Our big data analysis approach enables the use of large datasets, which can improve the performance of the model and the accuracy of the segmentation results. Convolutional layers in PBCNNs perform feature extraction on small patches of images using a set of learned filters as shown in Table 8 and Table 9. The filters slide over the input patches and produce a dot product between the filter and the patch at each location, resulting in a feature map.
Table 7.
Model Architecture of proposed Patch-Based Convolutional Neural Network.
Table 8.
Architecture of Convolutional Layers with Input and Output Data.
Table 9.
Design Architecture of Convolutional and Pooling Layers with Configuration.
These feature maps are then passed through activation functions to introduce non-linearity, and pooling layers are used to reduce the dimensionality of the feature maps. Convolutional layers are critical in PBCNNs for learning spatially invariant features and performing accurate image analysis tasks as shown in Table 10.
Table 10.
Features and descriptions of BraTS datasets (2012–2018).
4. Results and Discussion
We proposed a PBCNN for brain tumor segmentation using a big data analysis approach. The proposed method achieved a high level of accuracy in segmenting brain tumors in MRI scans, demonstrating the potential of deep learning techniques for medical image analysis. We evaluated the proposed method on a large dataset of 3304 MRI scans, consisting of 1652 scans with brain tumors and 1652 scans without brain tumors. The dataset was randomly divided into a training and a validation set. We used a patch-based approach to train our CNN, where each input image patch was classified as either tumor or non-tumor. The CNN was trained using the Adam optimizer with a learning rate of 0.001 for 50 epochs.
4.1. Patches Extraction
In this research, patch-based convolutional neural networks were utilized for brain tumor segmentation. The first step in this process was to extract patches from the brain MRI images. The size of the patches used in this study was 240 × 240, which was chosen based on the size of the tumors in the dataset. A total of 15,000 patches were extracted from the original MRI images, with each patch containing either tumor or normal tissue.
Table 11 shows the distribution of the patches used in the training, validation, and testing sets. As can be seen, the majority of the patches were used for training, with a smaller portion used for validation and testing. This distribution was chosen to ensure that the model was trained on a large enough dataset to learn the features of the tumors and could be validated and tested on independent data. Overall, patch extraction was a crucial step in the process of brain tumor segmentation using a patch-based convolutional neural network. The size and distribution of the patches used in this study were carefully selected to ensure that the model was trained on a large dataset to learn the features of the tumors and could be validated and tested on independent data.
Table 11.
Distribution of Patches in the Training, Validation, and Testing Sets.
4.2. Using Whole Images
We also evaluated the performance of our model using a whole MRI. The results are summarized in Table 12. Our model achieved high sensitivity and specificity for all tumor types, indicating that it can accurately detect and exclude regions that do not contain tumors. The precision and recall values were also high, indicating that our model achieved a good balance between true positives and false positives.
Table 12.
Performance of the CNN for brain tumor segmentation without patch-based CNN.
4.3. Patch-Based CNN
The results demonstrate that the proposed patch-based CNN can accurately segment brain tumors in MRI scans. One advantage of the PBCNN is that it allows the model to capture local features in the MRI scans. This is particularly important for brain tumor segmentation, as tumors can have complex and heterogeneous shapes and appearances. The result of extracted patches with various strides is shown in Figure 3. By using patches, the model can capture the characteristics of different regions of the brain in a more granular way, which can lead to more accurate segmentation results as shown in Table 13.
Figure 3.
Sample of Extracted Patches.
Table 13.
Performance of the proposed patch-based CNN for brain tumor segmentation.
Another advantage of patch-based CNNs is that they can be trained more efficiently than full-image CNNs. This is because patches require less memory and computational resources than full images. By training on patches, our model was able to take advantage of the large amount of data in the BraTS dataset without running into issues with computational limitations. The performance of the proposed PBCNN model was evaluated using various metrics, including DSC and sensitivity.
The evaluation was performed on the BraTS dataset, which includes MRI scans of brain tumor patients from 2012 to 2018. The model was trained on the training set and evaluated on the validation set to select the best hyperparameters. The final evaluation was performed on the test set to assess the generalization performance of the model. The DSC measures the overlap between the predicted and ground truth segmentation masks, with a value of 1 indicating a perfect overlap. Sensitivity measures the ability of the model to detect the true positive cases, with a higher value indicating better performance. The proposed PBCNN model achieved an average DSC score of 0.87 and a sensitivity of 0.83, indicating high accuracy in brain tumor segmentation. These results demonstrate the effectiveness of the PBCNN model in segmenting brain tumors and the potential of the big data analysis approach in medical image analysis. The performance of the proposed PBCNN model was evaluated using various evaluation metrics on BraTS datasets from 2012 to 2018. The evaluation metrics are summarized in Table 14.
Table 14.
Performance Metrics for BraTS Dataset.
The proposed PBCNN model achieved high accuracy and achieved state-of-the-art performance on all the BraTS datasets from 2012 to 2018. The results demonstrate that the PBCNN model is a highly effective method for brain tumor segmentation in MRI images.
4.4. Comparison of the PBCNN with State-of-the-Art Methods
The proposed PBCNN is compared with U-Net [35], DeepLab V3+ [36], and ResUNet++ [37] and shows favorable results. As shown in Table 15, the proposed method achieved the highest Dice coefficient of 0.91 and the highest accuracy of 0.96 among all the compared methods. The proposed method also outperformed the other methods in terms of sensitivity and specificity as shown in Figure 4. These results suggest that the proposed patch-based CNN architecture can accurately segment brain tumors and can be an effective tool for assisting medical professionals in clinical decision making.
Table 15.
Performance comparison of the proposed PBCNN with State-of-the-Art Methods.
Figure 4.
Distribution of MRI Patches with various Configurations.
Table 16 shows the Dice coefficient for brain tumor segmentation using a CNN without a patch-based approach and the proposed patch-based CNN. The results show a significant improvement in the Dice coefficient for all tumor types when using the proposed patch-based CNN approach.
Table 16.
Comparison of brain tumor segmentation results between CNN and proposed Patch-based CNN.
The results of the study demonstrate the potential of deep learning methods for automated brain tumor segmentation. The accurate segmentation of brain tumors is critical for medical professionals to make informed decisions regarding patient treatment. The proposed method can assist medical professionals in identifying the tumor’s location, size, and shape, which is essential for effective treatment planning. However, the proposed method has some limitations. First, the training of the model is computationally expensive, and requires a powerful GPU for training. Second, the proposed method requires a large amount of annotated data for training, which can be time-consuming and expensive to acquire.
5. Conclusions
In this study we present a novel patch-based convolutional neural network (PBCNN) approach for segmenting brain tumors in magnetic resonance imaging (MRI) scans. Leveraging the large amount of data available in the BraTS dataset from 2012 to 2018, this approach extracts patches from both tumor and normal tissue regions to train the CNN. The results of the study demonstrate that the PBCNN approach is highly effective for brain tumor segmentation. Compared to other state-of-the-art methods such as U-Net, DeepLab V3+, and ResUnet++, the PBCNN approach achieves higher segmentation accuracy and performance on a separate testing set. The use of patches enables the network to learn features at a more local level, which is important for capturing the heterogeneity and variability of brain tumors. Furthermore, the large amount of data available in the BraTS dataset allows for better generalization and performance compared to smaller datasets. These findings highlight the potential of using PBCNNs for medical image analysis tasks and the importance of leveraging big datasets for training deep learning models. The PBCNN approach presented in this study can have a significant impact on the diagnosis and treatment of brain tumors, ultimately improving patient outcomes. Future research can explore the application of this approach to other medical imaging modalities and disease types, further expanding the scope of deep learning in healthcare.
Author Contributions
The contributions of the authors are as follows: conceptualization, A.S.; methodology F.U. and M.A.; software, F.U. and F.A.; validation, F.A. and A.S.; draft preparation, A.S., F.U. and F.A.; review and editing, M.A. and A.S.; visualization, F.U.; supervision, A.S.; funding acquisition, F.A. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
Datasets analyzed during the current study are available on the BraTS website. The detailed description, online repository, and references [28,29,30,31,32,33,34] are provided in Table 2.
Acknowledgments
We thank our families and colleagues who provided us with moral support.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Akhavan, D.; Alizadeh, D.; Wang, D.; Weist, M.R.; Shepphird, J.K.; Brown, C.E. CAR T cells for brain tumors: Lessons learned and road ahead. Immunol. Rev. 2019, 290, 60–84. [Google Scholar] [CrossRef] [PubMed]
- Baker, A.W.; Ilieş, I.; Benneyan, J.C.; Lokhnygina, Y.; Foy, K.R.; Lewis, S.S.; Wood, B.A.; Baker, E.; Crane, L.; Crawford, K.L. 93. Early Recognition and Response to Increases in Surgical Site Infections (SSI) using Optimized Statistical Process Control (SPC) Charts–the Early 2RIS Trial: A Multicenter Stepped Wedge Cluster Randomized Controlled Trial (RCT). Open Forum Infect. Dis. 2021, 8, S59–S60. [Google Scholar] [CrossRef]
- Sharma, P.; Diwakar, M.; Choudhary, S. Application of edge detection for brain tumor detection. Int. J. Comput. Appl. 2012, 58, 21–25. [Google Scholar] [CrossRef]
- Paul, T.U.; Bandhyopadhyay, S.K. Segmentation of brain tumor from brain MRI images reintroducing K–means with advanced dual localization method. Int. J. Eng. Res. Appl. 2012, 2, 226–231. [Google Scholar]
- Greenspan, H.; Van Ginneken, B.; Summers, R.M. Guest editorial deep learning in medical imaging: Overview and future promise of an exciting new technique. IEEE Trans. Med. Imaging 2016, 35, 1153–1159. [Google Scholar] [CrossRef]
- Yeganeh, A.; Shadman, A.; Shongwe, S.C.; Abbasi, S.A. Employing evolutionary artificial neural network in risk-adjusted monitoring of surgical performance. Neural Comput. Appl. 2023, 1–17. [Google Scholar] [CrossRef]
- Wang, S.; Yin, Y.; Wang, D.; Wang, Y.; Jin, Y. Interpretability-based multimodal convolutional neural networks for skin lesion diagnosis. IEEE Trans. Cybern. 2021, 52, 12623–12637. [Google Scholar] [CrossRef]
- Bakas, S.; Reyes, M.; Jakab, A.; Bauer, S.; Rempfler, M.; Crimi, A.; Shinohara, R.T.; Berger, C.; Ha, S.M.; Rozycki, M. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. arXiv 2018, arXiv:1811.02629. [Google Scholar]
- Mittal, M.; Goyal, L.M.; Kaur, S.; Kaur, I.; Verma, A.; Hemanth, D.J. Deep learning based enhanced tumor segmentation approach for MR brain images. Appl. Soft Comput. 2019, 78, 346–354. [Google Scholar] [CrossRef]
- Kamnitsas, K.; Ledig, C.; Newcombe, V.F.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef]
- Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S. nnU-net: Self-adapting framework for u-net-based medical image segmentation. arXiv 2018, arXiv:1809.10486. [Google Scholar]
- Sandfort, V.; Yan, K.; Pickhardt, P.J.; Summers, R.M. Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks. Sci. Rep. 2019, 9, 16884. [Google Scholar] [CrossRef]
- Bakas, S.; Makris, D.; Hunter, G.J.; Fang, C.; Sidhu, P.S.; Chatzimichail, K. Automatic identification of the optimal reference frame for segmentation and quantification of focal liver lesions in contrast-enhanced ultrasound. Ultrasound Med. Biol. 2017, 43, 2438–2451. [Google Scholar] [CrossRef] [PubMed]
- Wang, W.; Chen, C.; Ding, M.; Yu, H.; Zha, S.; Li, J. Transbts: Multimodal brain tumor segmentation using transformer. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; pp. 109–119. [Google Scholar]
- Kofler, F.; Berger, C.; Waldmannstetter, D.; Lipkova, J.; Ezhov, I.; Tetteh, G.; Kirschke, J.; Zimmer, C.; Wiestler, B.; Menze, B.H. Brats toolkit: Translating brats brain tumor segmentation algorithms into clinical and scientific practice. Front. Neurosci. 2020, 14, 125. [Google Scholar] [CrossRef] [PubMed]
- Zikic, D.; Glocker, B.; Konukoglu, E.; Criminisi, A.; Demiralp, C.; Shotton, J.; Thomas, O.M.; Das, T.; Jena, R.; Price, S.J. Decision forests for tissue-specific segmentation of high-grade gliomas in multi-channel MR. In Proceedings of the MICCAI (3), Nice, France, 1–5 October 2012; pp. 369–376. [Google Scholar]
- Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef]
- You, L.; Qiu, A.; Huang, B.; Qiu, P. Early detection of high disease activity in juvenile idiopathic arthritis by sequential monitoring of patients’ health-related quality of life scores. Biom. J. 2020, 62, 1343–1356. [Google Scholar] [CrossRef]
- Hou, L.; Samaras, D.; Kurc, T.M.; Gao, Y.; Davis, J.E.; Saltz, J.H. Patch-based convolutional neural network for whole slide tissue image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2424–2433. [Google Scholar]
- Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; pp. 424–432. [Google Scholar]
- Zhang, J.; Lv, X.; Sun, Q.; Zhang, Q.; Wei, X.; Liu, B. SDResU-net: Separable and dilated residual U-net for MRI brain tumor segmentation. Curr. Med. Imaging 2020, 16, 720–728. [Google Scholar] [CrossRef]
- Wu, W.; Li, D.; Du, J.; Gao, X.; Gu, W.; Zhao, F.; Feng, X.; Yan, H. An intelligent diagnosis method of brain MRI tumor segmentation using deep convolutional neural network and SVM algorithm. Comput. Math. Methods Med. 2020, 2020, 6789306. [Google Scholar] [CrossRef] [PubMed]
- Wulczyn, E.; Steiner, D.F.; Xu, Z.; Sadhwani, A.; Wang, H.; Flament-Auvigne, I.; Mermel, C.H.; Chen, P.-H.C.; Liu, Y.; Stumpe, M.C. Deep learning-based survival prediction for multiple cancer types using histopathology images. PLoS ONE 2020, 15, e0233678. [Google Scholar] [CrossRef] [PubMed]
- Sathornsumetee, S.; Reardon, D.A.; Desjardins, A.; Quinn, J.A.; Vredenburgh, J.J.; Rich, J.N. Molecularly targeted therapy for malignant glioma. Cancer 2007, 110, 13–24. [Google Scholar] [CrossRef]
- Molinaro, A.M.; Taylor, J.W.; Wiencke, J.K.; Wrensch, M.R. Genetic and molecular epidemiology of adult diffuse glioma. Nat. Rev. Neurol. 2019, 15, 405–417. [Google Scholar] [CrossRef] [PubMed]
- Moreno Lopez, M.; Ventura, J. Dilated convolutions for brain tumor segmentation in MRI scans. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: Third International Workshop, BrainLes 2017, Held in Conjunction with MICCAI 2017, Quebec City, QC, Canada, 14 September 2017; pp. 253–262. [Google Scholar]
- Nie, D.; Lu, J.; Zhang, H.; Adeli, E.; Wang, J.; Yu, Z.; Liu, L.; Wang, Q.; Wu, J.; Shen, D. Multi-channel 3D deep feature learning for survival time prediction of brain tumor patients using multi-modal neuroimages. Sci. Rep. 2019, 9, 1103. [Google Scholar] [CrossRef] [PubMed]
- BRATS—SICAS Medical Image Repository. 2012. Available online: https://www.smir.ch/BRATS/Start2012 (accessed on 2 February 2023).
- BRATS—SICAS Medical Image Repository. 2013. Available online: https://www.smir.ch/BRATS/Start2013 (accessed on 12 February 2023).
- BRATS 2014: Brain Tumor Image Segmentation Challenge. 2014. Available online: https://www.smir.ch/BRATS/Start2014 (accessed on 13 February 2023).
- BRATS 2015: Brain Tumor Image Segmentation Challenge. 2015. Available online: https://www.smir.ch/BRATS/Start2015 (accessed on 13 February 2023).
- BRATS 2016: Brain Tumor Image Segmentation Challenge. 2016. Available online: https://www.smir.ch/BRATS/Start2016 (accessed on 16 February 2023).
- Multimodal Brain Tumor Segmentation Challenge 2017. Available online: https://www.med.upenn.edu/sbia/brats2017/data.html (accessed on 10 February 2023).
- Multimodal Brain Tumor Segmentation Challenge 2018. Available online: https://www.med.upenn.edu/sbia/brats2018/data.html (accessed on 9 February 2023).
- Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.-W.; Heng, P.-A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [PubMed]
- Ahuja, S.; Panigrahi, B.; Gandhi, T.K. Fully automatic brain tumor segmentation using DeepLabv3+ with variable loss functions. In Proceedings of the 2021 8th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 26–27 August 2021; pp. 522–526. [Google Scholar]
- Maji, D.; Sigedar, P.; Singh, M. Attention Res-UNet with Guided Decoder for semantic segmentation of brain tumors. Biomed. Signal Process. Control 2022, 71, 103077. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).



