Next Article in Journal
Optimization Design of Surface Wave Electromagnetic Acoustic Transducers Based on Simulation Analysis and Orthogonal Test Method
Next Article in Special Issue
RDCTrans U-Net: A Hybrid Variable Architecture for Liver CT Image Segmentation
Previous Article in Journal
Feasibility and Accuracy of Thoracolumbar Pedicle Screw Placement Using an Augmented Reality Head Mounted Device
Previous Article in Special Issue
Advanced Kidney Volume Measurement Method Using Ultrasonography with Artificial Intelligence-Based Hybrid Learning in Children
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Framework for Segmenting Brain Tumors Using MRI and Synthetically Generated CT Images

by
Kh Tohidul Islam
*,
Sudanthi Wijewickrema
and
Stephen O’Leary
Department of Surgery (Otolaryngology), Faculty of Medicine, Dentistry and Health Sciences, University of Melbourne, Melbourne, VIC 3010, Australia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(2), 523; https://doi.org/10.3390/s22020523
Submission received: 23 November 2021 / Revised: 26 December 2021 / Accepted: 28 December 2021 / Published: 11 January 2022
(This article belongs to the Collection Artificial Intelligence (AI) in Biomedical Imaging)

Abstract

:
Multi-modal three-dimensional (3-D) image segmentation is used in many medical applications, such as disease diagnosis, treatment planning, and image-guided surgery. Although multi-modal images provide information that no single image modality alone can provide, integrating such information to be used in segmentation is a challenging task. Numerous methods have been introduced to solve the problem of multi-modal medical image segmentation in recent years. In this paper, we propose a solution for the task of brain tumor segmentation. To this end, we first introduce a method of enhancing an existing magnetic resonance imaging (MRI) dataset by generating synthetic computed tomography (CT) images. Then, we discuss a process of systematic optimization of a convolutional neural network (CNN) architecture that uses this enhanced dataset, in order to customize it for our task. Using publicly available datasets, we show that the proposed method outperforms similar existing methods.

1. Introduction

Although vast amounts of medical images are produced each day, most are still interpreted through visual analysis on a slice-by-slice basis [1]. This requires experience, is time consuming, expensive, prone to human error, and most importantly, is inadequate for the processing of large-scale specimens [2]. Automated image processing/analysis techniques have gained popularity as alternatives that overcome these concerns.
Segmentation is a standard method of image analysis where an image is separated into different regions, typically based on the characteristics of its pixels (for two-dimensional (2-D) images) or voxels (for three-dimensional (3-D) volumes) [3]. Applications of segmentation in medical imaging include the identification of pixels/voxels representing organs [4] and pathologies such as tumors [5]. Here, we focus on the latter application: the segmentation of tumors. Using information from multiple imaging modalities can enhance the segmentation process, as different modalities provide different information on target regions [6]. As such, multi-modal medical image segmentation has the potential to deliver more reliable and accurate segmentation results. However, integrating such information efficiently still remains a challenge [7].
Deep learning techniques have gained considerable interest in the field of multi-modal medical image segmentation due to their self-learning and generalization ability over large-scale data [8]. Furthermore, such methods have shown improved performance when compared to single-mode segmentation methods in many instances [9]. As multi-modal images originate from different imaging sources, some pre-processing (for example, co-registration) may be required prior to their use in a deep learning model [10]. Multi-modal fusion strategies can then be used to integrate the images to facilitate segmentation [11].
Fusion methods for deep-learning-based multi-modal image segmentation can broadly be categorized as: input-, layer-, and decision-level fusion [8]. In input-level fusion, multi-modal images are combined channel-by-channel and serve as multi-channel inputs, representing integrated image features [12]. For layer-level fusion, multi-modal images are used to train individual networks to perform segmentation. Those networks are interconnected at different layers to learn shared features from input images. In decision-level fusion, each image is given as input to its respective network, and the learned features are fused into a decision layer to perform the segmentation task. Figure 1 illustrates the differences between the three types of fusion methods.
Zhou et al. [13] used four different modalities of magnetic resonance imaging (MRI) images (T1-weighted, T2-weighted, contrast-enhanced (CE) T1-weighted, and fluid attenuation inversion recovery (FLAIR)) to perform multi-modal brain tumor segmentation using the BraTS [14] dataset. They used three convolutional neural networks (CNNs) in cascade to perform a coarse-to-fine segmentation in one pass. The initial network produced a coarse mask for the possible whole tumor region. The second network was used to predict the precise region of the whole tumor. The final network was used as post-processing to enhance the tumor region. They used three modified versions of the FusionNet [15] network architecture in the Caffe [16] platform. The three networks were identical except for their final segmentation layers. The authors used input-level fusion with an input patch size of 32-by-32-by-16-by-4, where the final dimension represented the four MRI modalities.
Similarly, Wang et al. [17] also proposed a multi-modal segmentation method using the BraTS dataset [14] to segment brain tumors. They identified the entire tumor, tumor core, and enhancing tumor core (the anatomical structure of the tumor) separately. To obtain a combined feature set, they combined the four modalities of MRI images (T1-weighted, T2-weighted, CE-T1-weighted, and FLAIR) as multi-channel inputs. Then, they divided their segmentation task into a sequence of three successive binary segmentation tasks according to the hierarchical framework of brain tumors. To avoid the complexity of a 3-D deep learning model, they trained three segmentation models for the three different orthogonal views (sagittal, axial, and coronal). They used the average of the predictions of the three models as the final prediction.
Guo et al. [9] proposed a method of lung tumor segmentation from MRI, computed tomography (CT), and positron emission tomography (PET). They performed multi-modal fusion strategies at the input, layer, and decision level. They concluded that performing image fusion at the input- and layer-levels was more beneficial than fusion at the decision-level. In order to develop and validate their method, they used a publicly available soft-tissue sarcoma (STS) dataset [18], which was initially generated from a well-known cancer imaging archive [19]. For input-level fusion, they considered 28-by-28-by-k image patches, where k represented the number of modalities in the input layer. In layer-level fusion, every modality’s network layers were learned from corresponding features independently. Then, they were used to train a single network to perform the segmentation task. For decision-level fusion, the output of each network’s prediction was used to decide the final segmentation through voting.
Choudhury et al. [20] investigated DeepLabv3+ [21] along with the Xception [22] CNN (as the base network) to perform brain tumor segmentation for multi-modal MRI images in the BraTS dataset. Similar to Wang et al. [17], they used various combinations of T1-weighted, T2-weighted, CE-T1-weighted, and FLAIR images to perform segmentation of the whole tumor, tumor core, and enhancing tumor. Instead of using the 3-D volume as an input to the network, they also used 2-D orthogonal slices extracted from it. They prepared the dataset as a three-channel input (using different combinations of orthogonal slices) for 18 DeepLabv3+ models. Then, they used the outputs of these models as inputs to three more DeepLabv3+ models to segment each of the three regions under consideration.
Sun et al. [6] investigated different kernel sizes in a 3-D fully convolutional neural network (FCNN) architecture for brain tumor segmentation. The input to their FCN was four-channel images formed from four different MRI modalities (T1-weighted, T2-weighted, CE-T1-weighted, and FLAIR). To reduce the sensitivity to network initialization and to speed up the training of the FCN, they introduced a 3-D batch normalization layer after every convolutional layer [23]. Their findings indicated that smaller kernel sizes may provide better segmentation performance.
In this paper, we explore different deep learning architectures and training/testing strategies for the multi-modal segmentation of brain tumors using a publicly available dataset. We enhance this dataset by developing a method of generating CT images from MRI. Using experimental results, we show that using generated CT images improves the performance of deep learning models.

2. Materials and Methods

Here, for the task of segmenting brain tumors using 3-D multi-modal images, we used data from a publicly available dataset that contained four types of MRI images. We investigated whether the segmentation accuracy could be improved by introducing an additional modality to this dataset, namely CT images. For this purpose, we developed a method of generating CT images from MRI. Figure 2 shows an overview of the proposed methodology.

2.1. Experimental Setup

The MATLAB numeric computing environment (academic), including the image processing and deep learning toolboxes, was used to implement all methods. An Intel Xeno Silver 4108 CPU (1.80 GigaHertZ) processor with 5 gigabytes of NVIDIA QuADro P2000 GPU graphics memory and 16 gigabytes of physical memory, running the education version of 64 bit Microsoft Windows, was used as the main experimental platform. Some experiments were conducted on the University of Melbourne’s high-performance computing (HPC) system [24].

2.2. Synthetic CT Generation

Generating one (synthetic) image modality from another is a convenient way to avoid intricacies involved in obtaining multiple images of the same patient. Pre-processing steps that enable the analysis of multiple images, such as co-registration, are also not required in this case. As such, there has been some interest in the field of synthetic data generation [25,26,27,28]. Here, we developed 2-D and 3-D CNNs for this purpose and used the retrospective image registration evaluation (RIRE) dataset (this dataset can be downloaded from: http://www.insight-journal.org/rire/download_data.php, accessed on 1 October 2021) containing the 3-D MRI and CT pairs of 16 patients to train them [29]. We used a modified U-Net [30,31] architecture, as it has been shown to perform well in similar tasks [27,32,33,34,35]. Prior to training, we co-registered each image pair using an existing multi-modal image registration method [36]. To train the 3-D CNN, we used the volumes themselves, while we used axial slices of the volumes to train the 2-D CNN.
As the existing dataset was relatively small, we augmented it by generating synthetic data based on the original data. To this end, we performed one of five different transformations (nil, rotate 90°, flip left–right, flip top–bottom, rotate 90°, and flip left–right) on each image, using the same transformation for each corresponding MRI and CT pair. To further augment the 2-D dataset of axial slices, we used random transformations (rotations in the range of [−20° 20°], translations in the range of [ 10 10 ] , and reflections in the left–right and top–bottom directions). We used voxel patches of 32-by-32-by-13 (instead of the full volumes) in the 3-D CNN to avoid memory issues. We divided the dataset into three subgroups: training (70%), validation (15%), and testing (15%). Then, we optimized our training strategy with respect to this dataset.
A U-Net architecture typically consists of encoder and decoder sub-networks connected by a bridge. These sub-networks comprise multiple stages (depth) with each stage having multiple layers. There are two sets of convolutional (with additional batch normalization for 3-D) and rectified linear unit (ReLU) layers [37], followed by a 2-by-2 (or 2-by-2-by-2 for 3-D) max-pooling layer in each encoder sub-network stage. There is a transposed convolution layer for up-sampling, followed by two sets of convolutional and ReLU layers in each decoder sub-network stage. Another two sets of convolution (with additional batch normalization for 3-D) and ReLU layers are included in the bridge connection section. As the original U-Net architecture was developed for image segmentation, we replaced the softmax and segmentation output layers with a regression output layer to enable image-to-image regression. Examples of 2-D and 3-D U-Net architectures with depths of three are illustrated in Figure 3.
To optimize our network, we first tested depths from two to five and selected a depth of three, as it showed the best performance. Then, we used different training optimizers (stochastic gradient descent with momentum (SGDM) [38], root mean square propagation (RMSProp), and Adam [39]) with multiple training settings. We used root-mean-squared-error (RMSE) as the loss function along with different initial learning rates (0.1, 0.01, 0.001, and 0.0001). We chose 0.001 as the initial learning rate as it provided more balanced training and validation results after a few complete training cycles. To avoid memory issues, we set a minimum batch with one observation at each iteration. This minimum batch was used to evaluate the loss function gradient and update the training weights. The maximum number of epochs (complete passes over the training set) was set to 10 and was increased by a factor of 2 until it reached 160 and then increased to 200. We selected the weights from the network trained on the combination of loss function and maximum number of epochs that had the lowest validation loss.

2.3. Tumor Segmentation

Several state-of-the-art deep learning network architectures, mostly CNNs, dominate the field of medical image segmentation [40,41,42]. We considered two such CNNs: 2-D U-Net [30] and 3-D U-Net [31]. These were chosen as the best for our task after multiple training and testing strategies with different CNN architectures (AlexNet [43], VGG-16 and 19 [44], ResNet-18 and 50 [45], Inception-ResNet-v2 [46], DenseNet-201 [47], and DeepLabv3+ [21]).
We used the publicly available BraTS dataset (this dataset can be downloaded from: http://medicaldecathlon.com/, accessed on 1 October 2021) [48] in this experiment. It contains 750 MRI images (T1-weighted, T2-weighted, CE-T1-weighted, and FLAIR) of gliomas, the most commonly occurring brain malignancy. Figure 4 shows some images from this dataset along with their corresponding labels and orthogonal slices. From this dataset, we only used the 484 training volumes, as the remaining (test) data were not labeled. As in the case of CT generation (Section 2.2), we used the full volumes for the 3-D CNN and axial slices for the 2-D CNN. We augmented this dataset using the same strategies discussed in Section 2.2 (except for co-registration, as the dataset was already co-registered). Percentages of 70%, 15%, and 15% of the data were used for training, validation, and testing, respectively.
For the 3-D CNN, we used volume patches of 32-by-32-by-32 as input and output. For the 2-D CNN, we pre-processed the data to remove slices with a large number of background pixels (those with less than 120 labeled pixels). We used the remaining slices (25,051 out of a total of 60,729) in the training and testing of the 2-D U-Net.
First, we used the four different MRI scan modalities (T1-weighted, T2-weighted, CE-T1-weighted, and FLAIR) together to perform the segmentation task. Then, we paired each of the MRI modalities with the synthetic CT to obtain four combinations of input. Finally, we combined all five modalities to form a new set of inputs. We modified the input layer to allow for the respective number of channels (four, two, or five for the above input sets, respectively). The 2-D and 3-D U-Net architectures used for segmentation were the same as those shown in Figure 3 except that their regression output layers were replaced with softmax and segmentation output layers. The training strategies for segmentation were identical to those discussed in Section 2.2 except for the loss function. We used cross entropy as well as Tversky and generalized Dice loss functions [49,50]. The two latter loss functions were selected because they have proven effective for imbalanced datasets such as ours.
Additionally, we explored different fusion strategies: input-, layer-, and decision-level. For input-level fusion, we used one network with input images fused (see Figure 1 for an example). For layer- and decision-level fusion, we aligned two identical networks side by side and used a concatenation layer to connect them. For layer-level fusion, we connected the networks after their first ReLU layers to the second last ones iteratively. We chose to connect them after the second ReLU layers as it gave the most robust results. For decision-level fusion, we connected the networks right after the last ReLU layers. The following layers (of the combined network) after the connections were left unchanged.

3. Results

3.1. Synthetic CT Generation

Metrics such as the structural similarity (SSIM) index, peak signal-to-noise ratio (PSNR), and mean-squared error (MSE) [51] provide quantitative measures for the similarity between two images. We used these to compare the synthetic CT images generated by our method to ground truth CTs. Table 1 shows the performance (mean and standard deviation of a five-fold cross validation) of the 2-D and 3-D models with respect to different training optimizers. Note that the higher the value of SSIM and PSNR, the better the performance, while the opposite is true for MSE. Figure 5 shows some example outputs for the different training optimizers.

3.2. Tumor Segmentation

3.2.1. Evaluation Metrics

The confusion matrix compares segmentation results with ground truth data and presents true positives ( T P ), true negatives ( T N ), false positives ( F P ), and false negatives ( F N ) in a C-by-C matrix (for C segmentation classes). We used several variants of some segmentation performance metrics based on the confusion matrix (Equations (1) to (3)) in our study: global accuracy, mean accuracy, mean intersection over union (IoU), weighted IoU, and mean BF (boundary F1) (additional information on these metrics can be obtained from: https://au.mathworks.com/help/vision/ref/evaluatesemanticsegmentation.html, accessed on 1 October 2021).
A c c u r a c y = T P + T N T P + F N + F P + T N
I o U s c o r e = T P T P + F P + F N
B F ( b o u n d a r y F 1 ) = 2 × T P T P + F P × T P T P + F N T P T P + F P + T P T P + F N
Global accuracy, the accuracy over all classes, and mean accuracy, the average over all classes, provide computationally inexpensive and quick estimations. IoU (or Jaccard similarity coefficient) is one of the most commonly used metrics, with the mean IoU representing the average IoU over all classes. Weighted average IoU (weighted by the number of pixels per class) is useful when class sizes are imbalanced. The BF score has been seen to correlate better with human perception, as the contour of the BF matching score shows how well the predicted boundary of each class aligns with the actual boundary [52]. A mean BF score is the average BF score of all classes in all images of a dataset.

3.2.2. Comparison of Different Loss Functions

We observed that the Adam optimizer performed best in the tumor segmentation task. As such, we tested its performance on different loss functions with all four MRI modalities of the original dataset provided as different input channels. Table 2 shows the results (mean and standard deviation of a five-fold cross validation). Figure 6 shows some 2-D segmentation results along with their segmentation boundaries and segmented regions. Similar to previous studies where the segmented region was small with respect to the background [53], the Dice loss function performed best. The 3-D methods showed more precise results than their 2-D counterparts, likely because volumetric data retain 3-D information that is lost when using 2-D slices separately as input.

3.2.3. Comparison of Different Fusion Strategies

Using the network that showed the best results from the previous section (3-D U-Net with Adam optimizer and Dice loss function), we compared the effect of different fusion strategies on tumor segmentation performance. To this end, we considered all five image modalities (original MRI images as well as the synthetic CT we generated) together. As seen in Table 3, input-level fusion showed the best results when compared to layer- and decision-level fusion. This has been observed in other studies and could be due to the fact that earlier fusion may enable the network to learn shared representations, enabling it to learn better correlations across modalities [54].

3.2.4. Comparison with Synthetic CT

Here, we evaluated the segmentation performance for the different MRI and synthetic CT combinations for input-level fusion, as it performed best in the previous experiment. Table 4 shows the results of this experiment, and Figure 7 visualizes some 3-D segmentation results. We observed that the combination of all five modalities performed best. This implies that our CT generation method was successful in providing additional information to supplement the segmentation task.

3.2.5. Comparison of Different Network Architectures

We modified several state-of-the-art deep neural network architectures (AlexNet [43], VGG-16 and 19 [44], ResNet-18 and 50 [45], Inception-ResNet-v2 [46], DenseNet-201 [47], and DeepLabv3+ [21]) to accept 3-D images and perform the tumor segmentation task. For DeepLabv3+ [21], we used ResNet-18 [45], ResNet-50 [45], and Inception-ResNet-v2 [46] as base networks. As DeepLabv3+ [21] with ResNet-50 [45] showed the best results, we used that in our experiment. As seen from Table 5, the proposed method outperformed all of the other considered networks.

3.2.6. Comparison with Similar Methods

We compared the proposed method with similar existing methods (discussed earlier). We re-implemented these methods and in some cases, modified their architectures to suit our dataset. For example, for the network presented by Guo et al. [9], we used input-level fusion as they observed it generally worked better. For the network introduced by Sun et al. [6], we used a kernel size of 5-by-5-by-5 as they found it showed higher segmentation performance. In all other cases, we used their baseline methods. As seen from Table 6, the proposed method performed better than the others on every metric. This could be due to the fact that we introduced new information in the form of a new image modality to the dataset and because we optimized our method systematically.

4. Conclusions

We discussed how a segmentation task, namely brain tumor segmentation, can be improved by (1) adding a synthetic CT modality to an existing MRI dataset and (2) optimizing network configurations. To generate synthetic CT images, we used a modified U-Net architecture. We experimentally demonstrated that the proposed method outperformed other similar state-of-the-art methods. In future work, we will explore how this work can be extended to other segmentation tasks/datasets. We will also investigate how other complex CNN architectures can be utilized for this purpose. Furthermore, we will consider multi-class segmentation, for example, to identify different regions of a tumor. A natural extension to this would be to classify segmented areas (for example, to identify the type of tumor). We also aim to explore how our synthetic image generator could be utilized for other modalities (for example, to obtain T1-weighted MRI images from T2-weighted ones).

Author Contributions

Conceptualization, K.T.I.; data curation, K.T.I. and S.W.; formal analysis, K.T.I. and S.W.; funding acquisition, S.W. and S.O.; investigation, K.T.I. and S.W.; methodology, K.T.I. and S.W.; project administration, S.W. and S.O.; resources, S.W. and S.O.; software, K.T.I. and S.W.; supervision, S.W. and S.O.; validation, K.T.I., S.W. and S.O.; visualization, K.T.I.; writing—original draft, K.T.I.; writing—review and editing, K.T.I., S.W. and S.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the University of Melbourne and the Australian Government under the Melbourne Research Scholarship (MRS) and Research Training Program (RTP) Scholarship.

Institutional Review Board Statement

Ethical review and approval were waived for this study, as we used freely available de-identified databases.

Informed Consent Statement

Patient consent was waived for this study, as we used freely available de-identified databases.

Data Availability Statement

The datasets used in this study are freely available. Associated code can be accessed from https://github.com/khtohidulislam/ImageSegmentationCodes, accessed on 4 January 2022.

Acknowledgments

This research was supported by The University of Melbourne’s Research Computing Services and the Petascale Campus Initiative.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
2-Dtwo-dimensional
3-Dthree-dimensional
BFboundary F1
CEcontrast-enhanced
CNNconvolutional neural network
CTcomputed tomography
FCNNfully convolutional neural network
FLAIRfluid attenuation inversion recovery
FNfalse negatives
FPfalse positives
HPChigh-performance computing
IoUintersection over union
MRImagnetic resonance imaging
MSEmean-squared error
PETpositron emission tomography
PSNRpeak signal-to-noise ratio
ReLUrectified linear unit
RIREretrospective image registration evaluation
RMSEroot-mean-squared-error
RMSProproot mean square propagation
SGDMstochastic gradient descent with momentum
SSIMstructural similarity
STSsoft-tissue sarcoma
TNtrue negatives
TPtrue positives

References

  1. de Azevedo-Marques, P.M.; Ferreira, J.R. Medical Image Analyst: A Radiology Career Focused on Comprehensive Quantitative Imaging Analytics to Improve Healthcare. Acad. Radiol. 2021, 29, S1076633221000891. [Google Scholar] [CrossRef] [PubMed]
  2. Alalwan, N.; Abozeid, A.; ElHabshy, A.A.; Alzahrani, A. Efficient 3D Deep Learning Model for Medical Image Semantic Segmentation. Alex. Eng. J. 2021, 60, 1231–1239. [Google Scholar] [CrossRef]
  3. Shen, F.; Zeng, G. Semantic image segmentation via guidance of image classification. Neurocomputing 2019, 330, 259–266. [Google Scholar] [CrossRef]
  4. Bhandari, B.; Alsadoon, A.; Prasad, P.W.C.; Abdullah, S.; Haddad, S. Deep learning neural network for texture feature extraction in oral cancer: Enhanced loss function. Multimed. Tools Appl. 2020, 79, 27867–27890. [Google Scholar] [CrossRef]
  5. Ranjbarzadeh, R.; Bagherian Kasgari, A.; Jafarzadeh Ghoushchi, S.; Anari, S.; Naseri, M.; Bendechache, M. Brain tumor segmentation based on deep learning and an attention mechanism using MRI multi-modalities brain images. Sci. Rep. 2021, 11, 10930. [Google Scholar] [CrossRef]
  6. Sun, J.; Peng, Y.; Guo, Y.; Li, D. Segmentation of the multimodal brain tumor image used the multi-pathway architecture method based on 3D FCN. Neurocomputing 2021, 423, 34–45. [Google Scholar] [CrossRef]
  7. Tan, W.; Tiwari, P.; Pandey, H.M.; Moreira, C.; Jaiswal, A.K. Multimodal medical image fusion algorithm in the era of big data. Neural Comput. Appl. 2020. [Google Scholar] [CrossRef]
  8. Zhou, T.; Ruan, S.; Canu, S. A review: Deep learning for medical image segmentation using multi-modality fusion. Array 2019, 3–4, 100004. [Google Scholar] [CrossRef]
  9. Guo, Z.; Li, X.; Huang, H.; Guo, N.; Li, Q. Deep Learning-Based Image Segmentation on Multimodal Medical Imaging. IEEE Trans. Radiat. Plasma Med Sci. 2019, 3, 162–169. [Google Scholar] [CrossRef]
  10. Na, X.; Ma, L.; Milanova, M.; Yang, M.Q. Brain Tumor Segmentation Using Deep Neural Networks and Survival Prediction. In Advances in Artificial Intelligence and Applied Cognitive Computing; Arabnia, H.R., Ferens, K., de la Fuente, D., Kozerenko, E.B., Olivas Varela, J.A., Tinetti, F.G., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 89–99. [Google Scholar] [CrossRef]
  11. Kaur, M.; Singh, D. Multi-modality medical image fusion technique using multi-objective differential evolution based deep neural networks. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 2483–2493. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, X.; Wu, Y.; Song, G.; Li, Z.; Zhang, Y.; Fan, Y. A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med. Image Anal. 2018, 43, 98–111. [Google Scholar] [CrossRef] [PubMed]
  13. Zhou, C.; Ding, C.; Lu, Z.; Wang, X.; Tao, D. One-Pass Multi-task Convolutional Neural Networks for Efficient Brain Tumor Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Springer International Publishing: Cham, Switzerland, 2018; Volume 11072, pp. 637–645. [Google Scholar] [CrossRef]
  14. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS). IEEE Trans. Med. Imaging 2015, 34, 1993–2024. [Google Scholar] [CrossRef] [PubMed]
  15. Quan, T.M.; Hildebrand, D.G.C.; Jeong, W.K. FusionNet: A Deep Fully Residual Convolutional Neural Network for Image Segmentation in Connectomics. Front. Comput. Sci. 2021, 3. [Google Scholar] [CrossRef]
  16. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture for Fast Feature Embedding. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–4 November 2014; pp. 675–678. [Google Scholar] [CrossRef]
  17. Wang, G.; Li, W.; Ourselin, S.; Vercauteren, T. Automatic Brain Tumor Segmentation Based on Cascaded Convolutional Neural Networks With Uncertainty Estimation. Front. Comput. Neurosci. 2019, 13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Vallières, M.; Freeman, C.R.; Skamene, S.R.; El Naqa, I. A radiomics model from joint FDG-PET and MRI texture features for the prediction of lung metastases in soft-tissue sarcomas of the extremities. Phys. Med. Biol. 2015, 60, 5471–5496. [Google Scholar] [CrossRef]
  19. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [Green Version]
  20. Roy Choudhury, A.; Vanguri, R.; Jambawalikar, S.R.; Kumar, P. Segmentation of Brain Tumors Using DeepLabv3+. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Bakas, S., Kuijf, H., Keyvan, F., Reyes, M., van Walsum, T., Eds.; Springer International Publishing: Cham, Switzerland, 2019; Volume 11384, pp. 154–167. [Google Scholar] [CrossRef]
  21. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; Volume 11211, pp. 833–851. [Google Scholar] [CrossRef] [Green Version]
  22. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar] [CrossRef] [Green Version]
  23. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  24. Meade, B.; Lafayette, L.; Sauter, G.; Tosello, D. Spartan HPC-Cloud Hybrid: Delivering Performance and Flexibility; University of Melbourne: Melbourne, Australia, 2017. [Google Scholar] [CrossRef]
  25. Han, X. MR-based synthetic CT generation using a deep convolutional neural network method. Med. Phys. 2017, 44, 1408–1419. [Google Scholar] [CrossRef] [Green Version]
  26. You, C.; Cong, W.; Wang, G.; Yang, Q.; Shan, H.; Gjesteby, L.; Li, G.; Ju, S.; Zhang, Z.; Zhao, Z.; et al. Structurally-Sensitive Multi-Scale Deep Neural Network for Low-Dose CT Denoising. IEEE Access 2018, 6, 41839–41855. [Google Scholar] [CrossRef]
  27. Wang, Y.; Liu, C.; Zhang, X.; Deng, W. Synthetic CT Generation Based on T2 Weighted MRI of Nasopharyngeal Carcinoma (NPC) Using a Deep Convolutional Neural Network (DCNN). Front. Oncol. 2019, 9, 1333. [Google Scholar] [CrossRef] [PubMed]
  28. Palmér, E.; Karlsson, A.; Nordström, F.; Petruson, K.; Siversson, C.; Ljungberg, M.; Sohlin, M. Synthetic computed tomography data allows for accurate absorbed dose calculations in a magnetic resonance imaging only workflow for head and neck radiotherapy. Phys. Imaging Radiat. Oncol. 2021, 17, 36–42. [Google Scholar] [CrossRef]
  29. West, J.B.; Fitzpatrick, J.M.; Wang, M.Y.; Dawant, B.M.; Maurer, C.R., Jr.; Kessler, R.M.; Maciunas, R.J.; Barillot, C.; Lemoine, D.; Collignon, A.M.F.; et al. Comparison and evaluation of retrospective intermodality image registration techniques. In Medical Imaging 1996: Image Processing; Loew, M.H., Hanson, K.M., Eds.; SPIE: Newport Beach, CA, USA, 1996; pp. 332–347. [Google Scholar] [CrossRef]
  30. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Lecture Notes in Computer Science; Springer International Publishing: Munich, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  31. Özgün, Ç.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016; Springer International Publishing: Athens, Greece, 2016; pp. 424–432. [Google Scholar] [CrossRef] [Green Version]
  32. Chen, S.; Qin, A.; Zhou, D.; Yan, D. Technical Note: U-net-generated synthetic CT images for magnetic resonance imaging-only prostate intensity-modulated radiation therapy treatment planning. Med. Phys. 2018, 45, 5659–5665. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Gupta, D.; Kim, M.; Vineberg, K.A.; Balter, J.M. Generation of Synthetic CT Images From MRI for Treatment Planning and Patient Positioning Using a 3-Channel U-Net Trained on Sagittal Images. Front. Oncol. 2019, 9. [Google Scholar] [CrossRef]
  34. Liu, L.; Johansson, A.; Cao, Y.; Dow, J.; Lawrence, T.S.; Balter, J.M. Abdominal synthetic CT generation from MR Dixon images using a U-net trained with ‘semi-synthetic’ CT data. Phys. Med. Biol. 2020, 65, 125001. [Google Scholar] [CrossRef] [PubMed]
  35. Farjam, R.; Nagar, H.; Zhou, X.K.; Ouellette, D.; Formenti, S.C.; DeWyngaert, J.K. Deep learning-based synthetic CT generation for MR-only radiotherapy of prostate cancer patients with 0.35T MRI linear accelerator. J. Appl. Clin. Med. Phys. 2021, 22, 93–104. [Google Scholar] [CrossRef]
  36. Islam, K.T.; Wijewickrema, S.; O’Leary, S. A deep learning based framework for the registration of three dimensional multi-modal medical images of the head. Sci. Rep. 2021, 11, 1860. [Google Scholar] [CrossRef] [PubMed]
  37. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  38. Robbins, H.; Monro, S. A Stochastic Approximation Method. Ann. Math. Stat. 1951, 22, 400–407. [Google Scholar] [CrossRef]
  39. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar]
  40. Jurdi, R.E.; Petitjean, C.; Honeine, P.; Cheplygina, V.; Abdallah, F. High-level prior-based loss functions for medical image segmentation: A survey. Comput. Vis. Image Underst. 2021, 210, 103248. [Google Scholar] [CrossRef]
  41. Sarvamangala, D.R.; Kulkarni, R.V. Convolutional neural networks in medical image understanding: A survey. Evol. Intell. 2021. [Google Scholar] [CrossRef]
  42. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 1. [Google Scholar] [CrossRef] [PubMed]
  43. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  44. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  46. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-ResNet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  47. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  48. Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. Brain Tumor Segmentation and Radiomics Survival Prediction: Contribution to the BRATS 2017 Challenge. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Bakas, S., Kuijf, H., Menze, B., Reyes, M., Eds.; Springer International Publishing: Cham, Switzerland, 2018; Volume 10670, pp. 287–297. [Google Scholar] [CrossRef] [Green Version]
  49. Salehi, S.S.M.; Erdogmus, D.; Gholipour, A. Tversky Loss Function for Image Segmentation Using 3D Fully Convolutional Deep Networks. In Machine Learning in Medical Imaging; Springer International Publishing: Cham, Switzerland, 2017; pp. 379–387. [Google Scholar] [CrossRef] [Green Version]
  50. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer International Publishing: Cham, Switzerland, 2017; pp. 240–248. [Google Scholar] [CrossRef] [Green Version]
  51. Zou, K.H.; Warfield, S.K.; Bharatha, A.; Tempany, C.M.; Kaus, M.R.; Haker, S.J.; Wells, W.M.; Jolesz, F.A.; Kikinis, R. Statistical validation of image segmentation quality based on a spatial overlap index1. Acad. Radiol. 2004, 11, 178–189. [Google Scholar] [CrossRef] [Green Version]
  52. Csurka, G.; Larlus, D.; Perronnin, F. What is a good evaluation measure for semantic segmentation? Procedings of the British Machine Vision Conference 2013; British Machine Vision Association: Bristol, UK, 2013; Volume 27, pp. 32.1–32.11. [Google Scholar] [CrossRef] [Green Version]
  53. Zhang, Y.; Liu, S.; Li, C.; Wang, J. Rethinking the Dice Loss for Deep Learning Lesion Segmentation in Medical Images. J. Shanghai Jiaotong Univ. (Sci.) 2021, 26, 93–102. [Google Scholar] [CrossRef]
  54. Huang, S.C.; Pareek, A.; Seyyedi, S.; Banerjee, I.; Lungren, M.P. Fusion of medical imaging and electronic health records using deep learning: A systematic review and implementation guidelines. NPJ Digit. Med. 2020, 3, 136. [Google Scholar] [CrossRef]
Figure 1. Multi-modal deep learning networks for the segmentation of brain tumors from computed tomography (CT) and magnetic resonance imaging (MRI) images. Here, the three sections (top to bottom) represent the input-, layer-, and decision-level fusion networks, respectively.
Figure 1. Multi-modal deep learning networks for the segmentation of brain tumors from computed tomography (CT) and magnetic resonance imaging (MRI) images. Here, the three sections (top to bottom) represent the input-, layer-, and decision-level fusion networks, respectively.
Sensors 22 00523 g001
Figure 2. An overview of the proposed method. The first section (top) shows the process of synthetic CT generation from MRI. The second section (bottom) illustrates how the four MRI modalities along with the generated synthetic CT are used to perform the segmentation task.
Figure 2. An overview of the proposed method. The first section (top) shows the process of synthetic CT generation from MRI. The second section (bottom) illustrates how the four MRI modalities along with the generated synthetic CT are used to perform the segmentation task.
Sensors 22 00523 g002
Figure 3. Two-dimensional (2-D) and three-dimensional (3-D) U-Net architectures with encoder depths of three.
Figure 3. Two-dimensional (2-D) and three-dimensional (3-D) U-Net architectures with encoder depths of three.
Sensors 22 00523 g003
Figure 4. Example 3-D images and labels. From left to right: T1-weighted, T2-weighted, contrast-enhanced (CE)-T1-weighted, FLAIR, and labels. Corresponding orthogonal slices are shown in the second row.
Figure 4. Example 3-D images and labels. From left to right: T1-weighted, T2-weighted, contrast-enhanced (CE)-T1-weighted, FLAIR, and labels. Corresponding orthogonal slices are shown in the second row.
Sensors 22 00523 g004
Figure 5. Example results for synthetic CT generation. The first, second, and third rows represent SGDM, RMSProp, and Adam optimizers, respectively. Columns from left to right show the input MRI, corresponding ground truth CT, predicted CT, and the complement of the absolute difference image for each optimizer.
Figure 5. Example results for synthetic CT generation. The first, second, and third rows represent SGDM, RMSProp, and Adam optimizers, respectively. Columns from left to right show the input MRI, corresponding ground truth CT, predicted CT, and the complement of the absolute difference image for each optimizer.
Sensors 22 00523 g005
Figure 6. Example 2-D segmentation results. From top to bottom: input images, segmentation boundaries, and segmented regions.
Figure 6. Example 2-D segmentation results. From top to bottom: input images, segmentation boundaries, and segmented regions.
Sensors 22 00523 g006
Figure 7. Example 3-D segmentation results. From top to bottom: ground truth labeled volume, predicted labeled volume, center slice of the ground truth labels, and center slice of the predicted labels. For the first and second columns, the brain volume was made transparent make the tumor region visible. The third and fourth columns represent the center slice of the ground truth and predicted labels along the depth direction.
Figure 7. Example 3-D segmentation results. From top to bottom: ground truth labeled volume, predicted labeled volume, center slice of the ground truth labels, and center slice of the predicted labels. For the first and second columns, the brain volume was made transparent make the tumor region visible. The third and fourth columns represent the center slice of the ground truth and predicted labels along the depth direction.
Sensors 22 00523 g007
Table 1. Synthetic computed tomography (CT) generation performance (mean and standard deviation of a five-fold cross validation). Best results for each metric are shown in bold.
Table 1. Synthetic computed tomography (CT) generation performance (mean and standard deviation of a five-fold cross validation). Best results for each metric are shown in bold.
DimensionOptimizerSSIMPSNRMSE
2-DSGDM0.9163 ± 0.02129.0613 ± 0.66880.7135 ± 1.856
RMSProp0.9459 ± 0.01831.0253 ± 0.62151.3513 ± 1.027
Adam0.9617 ± 0.00932.5197 ± 0.30936.4010 ± 0.346
3-DSGDM0.9588 ± 0.02332.5921 ± 0.76635.7988 ± 0.841
RMSProp0.9746 ± 0.01835.0689 ± 0.66320.2392 ± 0.383
Adam0.9827 ± 0.00936.9242 ± 0.35113.2027 ± 0.125
Table 2. Segmentation performance (mean and standard deviation of a five-fold cross validation) with different loss functions. Best results for each metric are shown in bold.
Table 2. Segmentation performance (mean and standard deviation of a five-fold cross validation) with different loss functions. Best results for each metric are shown in bold.
DimensionLoss FunctionGlobal AccuracyMean AccuracyMean IoUWeighted IoUMean BF
2-DCross Entropy0.9525 ± 0.0280.9464 ± 0.0270.7312 ± 0.0240.9578 ± 0.0280.7512 ± 0.023
Tversky Loss0.9530 ± 0.0280.9465 ± 0.0270.7415 ± 0.0230.9580 ± 0.0280.7655 ± 0.023
Dice Loss0.9542 ± 0.0270.9473 ± 0.0260.7450 ± 0.0240.9574 ± 0.0270.7683 ± 0.024
3-DCross Entropy0.9549 ± 0.0250.9378 ± 0.0220.7409 ± 0.0230.9606 ± 0.0260.7987 ± 0.023
Tversky Loss0.9545 ± 0.0240.9485 ± 0.0210.7875 ± 0.0220.9613 ± 0.0260.8035 ± 0.021
Dice Loss0.9580 ± 0.0240.9520 ± 0.0200.8170 ± 0.0210.9625 ± 0.0250.8085 ± 0.020
Table 3. Segmentation performance (mean and standard deviation of a five-fold cross validation) with different fusion strategies. Best results for each metric are shown in bold.
Table 3. Segmentation performance (mean and standard deviation of a five-fold cross validation) with different fusion strategies. Best results for each metric are shown in bold.
Fusion LevelGlobal AccuracyMean AccuracyMean IoUWeighted IoUMean BF
Decision0.8842 ± 0.0290.8690 ± 0.0280.6828 ± 0.0220.8725 ± 0.0280.7074 ± 0.023
Layer0.9053 ± 0.0200.8832 ± 0.0200.6957 ± 0.0160.8968 ± 0.0210.7290 ± 0.017
Input0.9849 ± 0.0090.9579 ± 0.0080.8410 ± 0.0090.9706 ± 0.0100.8986 ± 0.009
Table 4. Segmentation performance (mean and standard deviation of a five-fold cross validation) for different combinations of magnetic resonance imaging (MRI) and synthetic CT. Best results for each metric are shown in bold.
Table 4. Segmentation performance (mean and standard deviation of a five-fold cross validation) for different combinations of magnetic resonance imaging (MRI) and synthetic CT. Best results for each metric are shown in bold.
CT WithGlobal AccuracyMean AccuracyMean IoUWeighted IoUMean BF
T1-weighted0.9405 ± 0.0120.9255 ± 0.0120.7138 ± 0.0090.9140 ± 0.0110.7450 ± 0.009
T2-weighted0.9415 ± 0.0120.9218 ± 0.0120.7174 ± 0.0090.9145 ± 0.0110.7455 ± 0.009
CE-T1-weighted0.9418 ± 0.0120.9263 ± 0.0120.7181 ± 0.0090.9157 ± 0.0110.7468 ± 0.009
FLAIR0.9422 ± 0.0120.9281 ± 0.0120.7153 ± 0.0090.9184 ± 0.0110.7480 ± 0.009
All0.9849 ± 0.0090.9579 ± 0.0080.8410 ± 0.0090.9706 ± 0.0100.8986 ± 0.009
Table 5. Segmentation performance (mean and standard deviation of a five-fold cross validation) with different network architectures. Best results for each metric are shown in bold.
Table 5. Segmentation performance (mean and standard deviation of a five-fold cross validation) with different network architectures. Best results for each metric are shown in bold.
Network ArchitectureGlobal AccuracyMean AccuracyMean IoUWeighted IoUMean BF
AlexNet [43]0.8847 ± 0.0270.8408 ± 0.0250.7505 ± 0.0230.8595 ± 0.0260.7682 ± 0.023
VGG-16 [44]0.9238 ± 0.0260.8857 ± 0.0260.7750 ± 0.0220.9074 ± 0.0260.8185 ± 0.024
VGG-19 [44]0.9355 ± 0.0280.8920 ± 0.0270.7852 ± 0.0240.9217 ± 0.0280.8250 ± 0.025
ResNet-18 [45]0.9641 ± 0.0200.9287 ± 0.0190.8074 ± 0.0170.9430 ± 0.0200.8583 ± 0.018
ResNet-50 [45]0.9715 ± 0.0190.9404 ± 0.0190.8170 ± 0.0160.9543 ± 0.0190.8732 ± 0.017
Inception-ResNet-v2 [46]0.9735 ± 0.0190.9460 ± 0.0180.8175 ± 0.0160.9542 ± 0.0190.8842 ± 0.018
DeepLabv3+ [21]0.9825 ± 0.0090.9505 ± 0.0090.8358 ± 0.0080.9680 ± 0.0090.8963 ± 0.009
DenseNet-201 [47]0.9782 ± 0.0150.9490 ± 0.0140.8325 ± 0.0120.9590 ± 0.0140.8783 ± 0.013
Proposed0.9849 ± 0.0090.9579 ± 0.0080.8410 ± 0.0090.9706 ± 0.0100.8986 ± 0.009
Table 6. Segmentation performance (mean and standard deviation of a five-fold cross validation) with similar methods. Best results for each metric are shown in bold.
Table 6. Segmentation performance (mean and standard deviation of a five-fold cross validation) with similar methods. Best results for each metric are shown in bold.
Network ArchitectureGlobal AccuracyMean AccuracyMean IoUWeighted IoUMean BF
Zhou et al. [13]0.9753 ± 0.0080.9557 ± 0.0080.8305 ± 0.0070.9684 ± 0.0080.8855 ± 0.008
Wang et al. [17]0.9825 ± 0.0080.9560 ± 0.0070.8384 ± 0.0060.9688 ± 0.0070.8873 ± 0.007
Guo et al. [9]0.9814 ± 0.0080.9531 ± 0.0070.8305 ± 0.0060.9654 ± 0.0080.8824 ± 0.007
Choudhury et al. [20]0.9810 ± 0.0080.9515 ± 0.0070.8375 ± 0.0060.9650 ± 0.0070.8955 ± 0.007
Sun et al. [6]0.9780 ± 0.0100.9435 ± 0.0090.8190 ± 0.0080.9650 ± 0.0090.8385 ± 0.008
Proposed0.9849 ± 0.0090.9579 ± 0.0080.8410 ± 0.0090.9706 ± 0.0100.8986 ± 0.009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Islam, K.T.; Wijewickrema, S.; O’Leary, S. A Deep Learning Framework for Segmenting Brain Tumors Using MRI and Synthetically Generated CT Images. Sensors 2022, 22, 523. https://doi.org/10.3390/s22020523

AMA Style

Islam KT, Wijewickrema S, O’Leary S. A Deep Learning Framework for Segmenting Brain Tumors Using MRI and Synthetically Generated CT Images. Sensors. 2022; 22(2):523. https://doi.org/10.3390/s22020523

Chicago/Turabian Style

Islam, Kh Tohidul, Sudanthi Wijewickrema, and Stephen O’Leary. 2022. "A Deep Learning Framework for Segmenting Brain Tumors Using MRI and Synthetically Generated CT Images" Sensors 22, no. 2: 523. https://doi.org/10.3390/s22020523

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop