Next Article in Journal
Multiparametric Contrast-Enhanced Ultrasound in Early Prediction of Response to Neoadjuvant Chemotherapy and Recurrence-Free Survival in Breast Cancer
Next Article in Special Issue
An Automatic Needle Puncture Path-Planning Method for Thermal Ablation of Lung Tumors
Previous Article in Journal
A Nomogram Based on Consolidation Tumor Ratio Combined with Solid or Micropapillary Patterns for Postoperative Recurrence in Pathological Stage IA Lung Adenocarcinoma
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparative Analysis of Image Processing Techniques for Enhanced MRI Image Quality: 3D Reconstruction and Segmentation Using 3D U-Net Architecture

by
Chee Chin Lim
1,2,*,
Apple Ho Wei Ling
1,
Yen Fook Chong
2,
Mohd Yusoff Mashor
1,2,
Khalilalrahman Alshantti
3 and
Mohd Ezane Aziz
3
1
Faculty of Electronic Engineering & Technology, Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
2
Sport Engineering Research Centre (SERC), Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
3
Department of Radiology, Universiti Sains Malaysia, Kubang Kerian 16150, Kelantan, Malaysia
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(14), 2377; https://doi.org/10.3390/diagnostics13142377
Submission received: 31 May 2023 / Revised: 29 June 2023 / Accepted: 5 July 2023 / Published: 14 July 2023
(This article belongs to the Special Issue Artificial Intelligence and Robotics in Interventional Radiology)

Abstract

:
Osteosarcoma is a common type of bone tumor, particularly prevalent in children and adolescents between the ages of 5 and 25 who are experiencing growth spurts during puberty. Manual delineation of tumor regions in MRI images can be laborious and time-consuming, and results may be subjective and difficult to replicate. Therefore, a convolutional neural network (CNN) was developed to automatically segment osteosarcoma cancerous cells in three types of MRI images. The study consisted of five main stages. First, 3692 DICOM format MRI images were acquired from 46 patients, including T1-weighted, T2-weighted, and T1-weighted with injection of Gadolinium (T1W + Gd) images. Contrast stretching and median filter were applied to enhance image intensity and remove noise, and the pre-processed images were reconstructed into NIfTI format files for deep learning. The MRI images were then transformed to fit the CNN’s requirements. A 3D U-Net architecture was proposed with optimized parameters to build an automatic segmentation model capable of segmenting osteosarcoma from the MRI images. The 3D U-Net segmentation model achieved excellent results, with mean dice similarity coefficients (DSC) of 83.75%, 85.45%, and 87.62% for T1W, T2W, and T1W + Gd images, respectively. However, the study found that the proposed method had some limitations, including poorly defined borders, missing lesion portions, and other confounding factors. In summary, an automatic segmentation method based on a CNN has been developed to address the challenge of manually segmenting osteosarcoma cancerous cells in MRI images. While the proposed method showed promise, the study revealed limitations that need to be addressed to improve its efficacy.

1. Introduction

Osteosarcoma is one of the most prevalent bone tumors, affecting mostly children and adolescents, typically those between 5 and 25 years of age who are experiencing puberty growth spurts [1]. Osteosarcoma usually develops from osteoblasts, which are the cells that make bones, and most commonly affects the distal femur, proximal tibia, and proximal humerus. The American Cancer Society relies on information from the SEER (Surveillance, Epidemiology, and End Results Program) database maintained by the National Cancer Institute (NCI) to provide 5-year relative survival rates statistics for people diagnosed with osteosarcoma between 2010 and 2016 with localized (74%), regional (66%), and distant (27%). If all SEER stages were combined, the 5-year relative survival rate was 60% [2]. Normal, active children and adolescents frequently experience pain and swelling in their limbs, which is more likely to be caused by normal bumps and bruises. Despite the fact that osteosarcoma weakens developing bones, fractures are uncommon, with the exception of the rare telangiectatic osteosarcomas, which weaken bones more than other types of osteosarcoma and are more likely to produce fractures where the osteosarcoma tumor is present [3].
Magnetic resonance imaging (MRI) provides more accurate anatomical information for medical examinations than other medical imaging techniques such as X-rays, ultrasound, and CT images [4]. MRI is an advanced medical imaging technique that provides a wealth of information about the anatomy of human soft tissues. The accuracy of osteosarcoma tumor segmentation in MRI images is important not only for treatment planning for neoadjuvant chemotherapy but also for evaluating subsequent treatment effects. However, the manual method of radiologists outlining malignant tissue from each slice is time-consuming, subjective, and often produces non-repeatable findings. Additionally, the difficulty in identifying osteosarcoma is high due to aspects such as size, location, shape, and texture features. Therefore, this study aims to develop an automatic segmentation algorithm for osteosarcoma tumor cells using convolutional neural networks (CNN) to segment the 3D model for osteosarcoma tumor cells. After the osteosarcoma cancerous cells are segmented automatically, the performance of the auto-segmentation algorithm will be compared to manual segmentation of the region of interest (ROI) of the 3D osteosarcoma tumor cells.
This paper is organized as follows. Section 2 focuses on the literature review of image processing for medical images such as MRI and CT images to review current trends in image processing techniques. Section 3 focuses on the methods used for segmenting the osteosarcoma cancerous cells, which include image pre-processing, image quality assessment (IQA), and segmentation. Section 4 presents the experimental results and discussions of the image pre-processing, IQA, image segmentation, and validation of the proposed model. A conclusion is drawn in Section 5.

2. Related Works

This paper presents a deep learning method for segmenting osteosarcoma bone cancerous cells from MRI images. Prior to analysis, the MRI images undergo image processing. Raw images that contain unwanted noise are processed to extract useful information. This image processing involves pre-processing and image segmentation.

2.1. Image Pre-Processing

In this paper, a method is introduced to segment osteosarcoma bone cancerous cells using deep learning. Prior to processing the MRI images, image pre-processing is performed to eliminate artifacts and improve signal quality without losing information. Artifacts can significantly affect the MRI image of the osteosarcoma, making it more difficult to analyze and interpret.
Rajeshwari and Sharmila (2013) [5] described a two-phase pre-processing method. In the first phase, a median filter was used to remove film artifacts. In the second phase, an algorithm was introduced to remove unwanted parts using morphological operations. This reduced false positive results in subsequent processing stages and improved results over tracking algorithms that preserve regions of interest and remove film artifacts. This pre-processing method also helps to prevent over-segmentation, which could retain the tumor during further processing.
Suhas and Venugopal (2017) [6] evaluated the performance of MRI image denoising techniques. They applied various filters, including median, Gaussian, max, min, and arithmetic mean filters, to MRI brain and spinal cord images. A new strategy for modifying the existing median filter by adding features is proposed. Experimental results showed that the proposed method, along with the other three image filtering algorithms, improved the root mean square error (RMSE), signal-to-noise ratio (SNR), peak signal-to-noise ratio (PSNR), and other statistical characteristics used to assess output image quality. The proposed method successfully preserved the structural characteristics of the medical image while reducing the majority of the noise.
Intensity normalization is an important preprocessing step for MRI images. In the paper by Mohan and Subashini (2018) [7], six intensity normalization techniques were proposed, including contrast stretching, histogram stretching, histogram equalization, histogram normalization, intensity scaling, and Gaussian kernel normalization. However, according to Loizou et al. (2009) [8], histogram equalization has limited success on medical images due to the removal of minor details. In contrast, Mohan and Subashini (2018) [7] state that histogram equalization is successful, but only when applied to certain types of medical images. To address the limitations of histogram equalization, adaptive and spatially variable processing techniques have been developed. The Wiener filter is an example of such a technique, which can handle spatial and local changes in the image with flexibility.

2.2. Segmentation Technique Using CNN in Deep Learning

Deep learning (DL) is a sub-category of machine learning that utilizes multilayer networks to analyze complex patterns in raw image input data. In recent years, DL has proven to be a beneficial tool in various imaging applications.
Singh et al. (2020) [9] reviewed the current applications of interpretable deep learning in different medical imaging tasks. This review discusses various methods, challenges, and areas that require further research in clinical deployment from the practical perspective of deep learning researchers designing systems for clinical end-users.
Debelee et al. (2020) [10] surveyed several deep learning-based techniques for identifying and segmenting tumors in breast cancer, cervical cancer, brain tumors, colon cancer, and lung cancer. Deep learning approaches were found to be the most advanced in tumor identification, segmentation, feature extraction, and classification as a result of the evaluation process.
Çiçek et al. (2016) [11] suggested a volume-to-volume segmentation network called the 3D U-Net, which is a 2D U-Net extension. The 3D U-Net uses dual paths: an analysis path to extract features and a synthesis path for up-sampling to generate full-resolution segmentation. Additionally, the 3D U-Net developed a shortcut link between layers with the same resolution in the early and late stages of the analysis and synthesis path.
Chen et al. (2018) [12] introduced a voxel-wise residual network (VoxResNet), a 3D deep network extension of 2D deep residual learning. VoxResNet provides a skip connection to transfer features from one layer to the next. Although 3D U-Net and VoxResNet provide multiple skip connections to facilitate training, their presence generates a short path from the first layer to the last layer, potentially reducing the network to a relatively simple configuration.
Holbrook et al. (2020) [13] proposed using TensorFlow to segregate soft tissue sarcomas in mice and perform radiomics studies on the obtained MRI datasets using a 3D fully convolutional U-Net network. The cross-entropy loss function was used to achieve the best overall segmentation performance. The dice score for T2-weighted images is 0.861, and the dice score for multi-contrast data is 0.863.
Vaidyanathan et al. (2021) [14] proposed a 3D U-Net to build a deep-learning method for automatic segmentation of the inner ear in MRI by using manually segmented inner ear images as a reference standard. A clinical validation set of eight MRI scans in which the labyrinth’s morphology had changed substantially was also used to validate the model. Across images from three different centers, the 3D U-Net model displayed an accurate mean DSC of 0.8790, a high true positive rate (91.5%), and low false discovery rates and false negative rates (14.8% and 8.49%, respectively).
Feng Liu et al. (2022) [15] proposed OSTransnet, a method for segmenting osteosarcoma MRI images. It combines transformer and U-Net models to address challenges related to fuzzy tumor edge segmentation and overfitting. The technique optimizes the dataset by altering the spatial distribution of noise and applying image rotation. By incorporating channel-based transformers, OSTransnet improves upon the limitations of U-Net and achieves a DSC (dice similarity coefficient) of 0.949, resulting in better segmentation results for osteosarcoma MRI images with blurred tumor borders.
Jia Wu et al. (2022) [16] presented a deep CNN system for osteosarcoma MRI image segmentation. It addresses overfitting due to noisy data and improves generalization. The method involves dataset optimization using Mean Teacher and training with noisy data to enhance robustness. Segmentation utilizes a deep separable U-shaped network (SepUNet) and a conditional random field (CRF). SepUNet effectively segments lesions of different sizes at multiple scales, while CRF refines boundaries. This proposed method was evaluated on 80,000 MRI images from three Chinese hospitals, demonstrating an improved DSC of 0.914.

2.3. Summary of Previous Studies

Throughout the research from the previous studies, in the pre-processing process, the contrastivity of MRI images can be enhanced by using histogram equalization or histogram normalization, which are non-linear operations and linear operations, respectively. Other than that, there will be some noise in the MRI images. These noises can be removed by a median filter, which is a non-linear filter, or a Gaussian filter, which is a linear filter. Then, the quality of the processed images can be evaluated by using the peak signal-to-noise ratio (PSNR), mean square error (MSE), and absolute mean brightness error (AMBE). After pre-processing, the segmentation technique by using CNN that was frequently used in previous research is 3D U-Net, which is better at segmenting the 3D model of the region of interest (ROI) as compared to 2D U-Net. Lastly, the accuracy of the automatic segmented ROI can be verified through the dice similarity coefficient (DSC) by comparing it with the manual segmented ROI.

3. Results

This section outlines the process flow for carrying out this study. The Python programming language and software used in this study were Google Colaboratory and PyCharm version 2021.2.3 software for simulation using the K80 graphical processing unit with 12 GB of RAM.
Figure 1 shows the overview flow chart that outlines the events in this study. It started acquiring three types of MRI images of the patients who were diagnosed with osteosarcoma. After that, the MRI images were processed with a series of image enhancement methods to improve the quality of the images, so as to increase the accuracy of the segmentation result later on. Then, a convolutional neural network was built by using 3D U-Net architecture to segment the osteosarcoma cancerous cell. Next, the 3D model of the segmented osteosarcoma was visualized to better observe the location and size of the tumor. Lastly, the comparison between manual segmentation and automatic segmentation was carried out using the dice similarity coefficient (DSC) as evaluation criteria.

3.1. Image Acquisition

Image acquisition is the initial step in every image processing system, which aims to transform an optical image into an array of numerical data that can be modified on a computer. The information was gathered from the records of the Department of Radiology, Hospital Universiti Sains Malaysia (HUSM), picture archiving and communication system (PACS), and radiology information system using a Philips 3 Tesla Achieva Magnetic Resonance Imaging (MRI) scanner (RIS). The MRI scanner was utilized to collect the MRI images of patients who were diagnosed with osteosarcoma, which are T1W, T2W, and T1W + Gd in the Digital Imaging and Communications in Medicine (DICOM) file type. DICOM is a standard for handling, storing, printing, and transferring medical imaging data [17], and it includes the file format description as well as the network communications protocol. The example MRI images collected are shown in Figure 2.

3.2. Image Enhancement

Contrast enhancement techniques are commonly used in medical imaging to improve the visual quality of low-contrast images by emphasizing crucial characteristics or those that are not visible. Contrast enhancement techniques in medical imaging improve the visual quality of low-contrast images by emphasizing important features or invisible details. In this study, contrast-limited adaptive histogram equalization (CLAHE) and contrast stretching (CS) were applied to the MRI images to enhance their contrast, and the best contrast technique was chosen among these two techniques by using the IQA.
CLAHE is a technique used for enhancing local image contrast. CLAHE employed small tiles in the MRI image to compute numerous histograms, each of which compared to a specific area of the image, and then used them to redistribute the image’s brightness or contrast estimation. CLAHE improved the contrast better than normal histogram equalization, which added more detail but amplified noise [18]. A clip limit was then determined for clipping histograms. Each histogram’s height was then redistributed to ensure that it did not exceed the clip limit. The clip limit was calculated as β which can be written in the form as (1) [19], where M × N is the number of pixels in each region, L is the number of grayscales, α is a clip factor (0–100), and s m a x is the maximum allowable slope.
β = M N L 1 + α 100 s m a x 1
Contrast stretching is known as normalization, which is a linear operation, meaning the value of the new pixel changes linearly as the value of the original pixel changes [20]. It is a straightforward image enhancement method that involves stretching the range of intensity values to improve image quality. To stretch the image, the higher and lower pixel value limitations over which the image was normalized must be defined, and the existing lowest and highest pixel values should also be identified. As in (2) [21], it was then scaling each pixel in the MRI image, where Pin is the input pixel and Pout is the resulting pixel, a represents the lower limit and b represents the upper limit, and c and d are the current lowest and maximum pixel values, respectively.
P o u t = P i n c b a d c + a

3.3. Image Denoising

Image quality can be troublesome and poor when acquiring, processing, and storing MRI images. Researchers are still grappling with how to remove noise from original MRI images since noise removal generates artifacts and blurs the images [22]. Different filters can effectively remove different types of noise. In this study, the denoising filters applied include the Gaussian filter and Median filter. The most suitable filter was selected for denoising the MRI images.
The Gaussian filter is a type of linear smoothing filter whose weights are determined by the form of the Gaussian function [23]. The Gaussian filter helps reduce visual noise and minor features drawn from a normal distribution [23]. For image processing, the two-dimensional zero-mean discrete Gaussian function is expressed as (3) [24], where σ is the standard deviation of the distribution, and x and y are the location indices. The value of σ controls the extent of the blurring effect around a pixel by adjusting the variance around the mean value of the Gaussian distribution [24]. It is commonly performed by using a Gaussian kernel to convolve the image. The Gaussian filter is especially useful for filtering images with a lot of noise because the results showed relative independence on the noise features and a significant dependence on the variance value of the Gaussian kernel [25]. The MRI image was denoised using the Gaussian filter by adjusting the standard deviation of the intensity distribution in order to control the blurring effect of the filter, which depends on the effectiveness of the noise to be removed in the MRI image.
G 2 D x , y , σ = 1 2 π σ 2 e x 2 + y 2 2 σ 2
The median filter is a sliding window spatial filter that substitutes the median value of all pixel values in the window for the window’s center value [26]. It is a non-linear filtering technique that helps to remove noise and is capable of removing “impulse” noise from either high or low outliers. The standard median filter was given by (4) [26], where Xi and Yi are the input and output at location i of the filter, and Wi is the r^th order statistic of the samples inside the window. The MRI image was applied with the median filter to replace the particular pixel with the median value of the sample of its particular window throughout the MRI image.
Y i = m e d W i = m e d { X i + r : r W }

3.4. Image Quality Assessment (IQA)

In this study, different techniques were applied for calculating the values of separate parameters such as mean square error (MSE), peak signal-to-noise ratio (PSNR), and absolute mean brightness error (AMBE).
MSE is the most widely used error sensitivity-based image quality assessment, as shown in (5) [23], where M and N are the width and the height of the images, respectively, and xij and yij are the image gray values of reference image x and distorted image y. An average of squared intensity differences in every pixel of a reference image and a distorted image was used to calculate it. The lower value of MSE indicates there is less error in the image [27].
M S E = 1 M N i = 1 M j = 1 N ( x i j y i j ) 2
The PSNR is a quality metric for lossy compressed images. The PSNR is the ratio of the original image’s maximum power to the deformed image’s noisy power. Due to the fact that signals frequently have a wide dynamic range, they are represented in the logarithmic domain. The formula is given as (6), where MAX is the maximum possible pixel value of the MRI image [22] and MSE is the mean square error. A higher PSNR value indicates that the image quality is better.
P S N R = 10   log 10 M A X 2 M S E
AMBE is an assessment for determining brightness preservation. It assesses the effectiveness of contrast enhancement approaches to maintain the original image’s mean brightness [28]. AMBE uses (7) [28] to calculate the absolute mean brightness difference between the acquired image, Iin, and the pre-processed image, Iout. The lower AMBE value indicates a good performance technique with a high-quality image, and the brightness is better preserved.
A M B E = I i n I o u t

3.5. Reconstruct MRI Images into 3D Volumes

Before the images passed to the segmentation process, all the DICOM-format images were converted to NIfTI format. The fundamental distinction between DICOM and NIfTI is that NIfTI saves raw image data as a 3D image, whereas DICOM saves raw image data as 2D image slices ([29], p. 4). In addition, NIfTI is modeled as a three-dimensional image, so it is better than DICOM for several deep learning applications as it is easier to manage a single NIfTI file rather than hundreds of DICOM files. For this study, Pycharm software was used for converting DICOM to NIfTI by using the command “dcm2nii” [30]. All input images were rescaled to the same size before training to maintain optimal image features.

3.6. Segmentation Model for 3D Volumes

The Python programming language was used to create the 3D U-Net model using the open-source deep learning framework Medical Open Network for AI (MONAI), which was combined with the PyTorch Lightning framework and the PyTorch, Numpy, and Matplotlib libraries. The whole segmentation process in this project is shown in Figure 3.
It started with the transformation of the MRI image to enable the image to be better fitted into the convolutional neural network that will be trained for automatic segmentation later. The MONAI’s “compose” function was used to apply several transforms to the same dataset, which allowed for combining any transformations needed. First, use the “LoadImaged” command to load the MRI images and labels from NIfTI format files. Second, the “AddChanneld” command added a channel to the MRI image and label. When it came to tumor segmentation, a channel that played the role of background or tumor was required. Third, the “Orientationd” command unified the data orientation based on the affine matrix. Fourth, based on the affine matrix, the “Spacingd” command was used to adjust the spacing by pixel dimension, “pixdim” = (1.5, 1.5, 2.0). This function assisted in changing the voxel dimensions because the dataset of medical images had different voxel dimensions, which were width, height, and depth. Therefore, it is necessary to generalize all of them to the same dimensions. Moreover, the “CropForegroundd” command helped to remove all zero borders, allowing the focus to be placed on the valid area of the images and labels. Also, the “RandCropByPosNegLabeld” command was included, which helped to randomly crop patch samples from the big image based on the positive-to-negative ratio. The “RandAffined” command followed, which efficiently performed rotation, scaling, shearing, translating, and other operations based on the PyTorch affine transform, which was also applied in this pre-processing. Lastly, the command “EnsureTyped” converted the Numpy array to PyTorch Tensor, which might be used in subsequent phases.
After the transformation, the data loader was applied to speed up the training process and reduce the memory usage of the graphic processing unit since the MRI datasets used to train were in voxel, which required a longer time in the training process of the neural network. For the purpose of doing so, there were two functions employed, “CacheDataset” and “DataLoader”, in MONAI dataset managers. Before the first epoch, “CacheDataset” performed non-random transforms and prepared cache material in the main process, and then all “DataLoader” subprocesses read the same cache content in the main process during training. According to the extent of predicted cache data, preparing cache material may take a long time. In this project, there were two datasets to be built: one to combine the training data with its transforms and the other to combine the validation data with its transforms because there were training and validation sets.
Once the preparation data was performed, the MRI images were ready to train in the 3D U-Net model. A 3D U-Net by MONAI was used in the model design as shown in Figure 4, and the arrows displayed the various operations; the blue boxes showed the feature map at each layer, and the grey boxes described the cropped feature maps from the contracting route. The architecture of the 3D U-Net was roughly divided into two parts, which were the encoder network and decoder network, where each layer had its own encode and decode paths as well as a skip connection between them. Data were down-sampled using strided convolutions in the encoder path, then up-sampled using strided transpose convolutions in the decode path. For the encoder part, it is made up of four blocks, each of which has 3 × 3 × 3 convolutional layers with a PReLU activation function and 2 × 2 × 2 max-pooling layers with strides of two in each dimension. The PReLU activation function used batch normalization to apply a function to the input data in order to boost non-linearity and speed up training. In order to avoid overfitting, the pooling layer downsampled the input values to reduce computing costs and reduce the spatial dimensions of the image. A fully connected layer provided the correlations of the particular class to the high-level features. The number of outputs of the last fully connected layer must be the same as the number of classes [31]. On the other hand, for the decoder path, it was made up of four blocks, each of which contains a 2 × 2 × 2 transposed convolution layer with a stride of 2, followed by two convolutional layers with a size of 3 × 3 × 3, and a PReLU activation function that used batch normalization. The Shortcut connections from equal-resolution layers in the encoder path helped to give the decoder path the necessary high-resolution features. There was also a 1 × 1 × 1 convolution in the final layer with sigmoid output used to reach the feature map with a depth equal to the number of classes, which was 2, where the loss function was determined. Moreover, the high-resolution 3D features in the encoder path were concatenated with up-sampled representations of global low-resolution 3D features in the decoder path to learn and apply local information. The network learned to employ both high-resolution local information and low-resolution global features as a result of this concatenation. During training, the dice loss was used as the loss function, and Adam was used as the optimizer, with a learning rate of 1 × 10−4 using backpropagation to find the gradient of the loss function. Table 1 shows an overview of the hyperparameters and their respective values used in this project.
Moreover, the trained model was evaluated quantitatively and qualitatively. The evaluation metrics used to validate qualitatively were DSC. Throughout the training epoch, the model with the highest validation mean DSC for each MRI image type was saved. The average epoch loss was also recorded, which indicated the error occurred in the validation dataset. On the other hand, the validation mean DSC and average epoch loss through the training epoch were used to plot the curve for statistical purposes. In addition, the overlaid image of the MRI image and the predicted output from the 3D U-Net model (Figure 4) were shown to better observe the true or false prediction from the model. Lastly, for better visualization, the 3D volume of the MRI image with the label of the predicted output was displayed.

3.7. Image Segmentation Performance Validation

In this study, the suggested evaluation metrics used to validate and compare were dice similarity coefficients (DSC). The DSC, also known as the overlap index [32], was a regularly used performance metric in the field of medical image segmentation. It determined the general similarity rate between a given ground truth label and the expected segmentation output of a segmentation technique. DSC can be expressed as (8) and (9). Where Sp is the predicted segmentation output and Sg is the ground truth label. FP, TP, and FN indicate false positives, true positives, and false negatives, respectively. DSC gave a score between 0 and 1, where 1 denotes the best prediction and indicates that the segmentation result was as expected [32].
D S C S p , S g = 2 T P F P + 2 T P + F N
D S C S p , S g = 2 S p     S g S p + S g

4. Discussion

4.1. Image Pre-Processing

Table 2, Table 3 and Table 4 show contrast-enhanced MRI images by CLAHE and contrast stretching before and after implementing the median filter and Gaussian filter for T1W, T2W, and T1W + Gd, respectively. The goal was to improve the contrast and brightness of the MRI images using these enhancement techniques.
The contrast and brightness of the MRI images had improved for both contrast enhancement methods. For CLAHE, the histogram of intensity value was clipped before computing the cumulative distribution function and distributed uniformly to other bins. As a result, a clear, enhanced MRI image without much noise was obtained. On the other hand, contrast stretching increases the difference between the maximum and minimum intensity values in an image. The remaining intensity values were spread out across the range, making the contrast more noticeable. Therefore, in the resulting image from the contrast stretching, one can clearly observe the contrast between the intensities. From the enhanced MRI images in T1W, T2W, and T1W + Gd, it can be observed that the CLAHE technique can give more detail on the MRI. However, in this study, the region of interest (ROI) was not the details of the MRI images, so these details were considered noise, which caused confusion during the segmentation part due to their similar intensity. Consequently, the contrast stretching technique was more suitable to be used in pre-processing the MRI image in this study.
The MRI images that were processed with the Gaussian filter were more blurry than the MRI images processed with the median filter. This was because the Gaussian filter was a linear type of filter that was more effective in smoothing the image and removing noise. On the other hand, the median filter, which is a non-linear type of filter, showed a better result than the Gaussian filter. This was due to the median filter, which removed thin lines or edges and blurred the image but retained useful details. Moreover, both denoising filters did not show good filtering results in the CLAHE-enhanced MRI images.

4.2. Image Quality Assessment

The lower value of MSE indicates there was less error in the image. Overall, from Table 5, the MRI images that were processed with the combination of contrast stretching and median filter scored the lowest MSE value as compared to other combinations. This indicated the combination of contrast stretching and median filter would not affect much on the image quality and would be close to the original MRI image.
The higher the PSNR, the better the image quality after applying filters. In order to calculate the PSNR, we should first have the value of the mean square error (MSE). This was because PSNR was usually expressed in terms of a logarithmic decibel scale to calculate the peak error. From the result in Table 5, as for the PSNR, the combination between the Contrast Stretching and Median filter showed better performance because it got the highest PSNR value.
The lower AMBE indicated better brightness preservation of the image. From Table 5, the AMBE values for the combination of contrast stretching and median filter were very close to zero. So, contrast stretching and median filter showed better brightness preservation than other combinations.
Contrast stretching enhances the contrast between different regions of the cancerous and bone, making its details more visible. By spreading out the intensity values across the range, it increases the separability of different structures and enhances the overall MRI image quality. Thus, this can lead to clearer distinctions between the cancerous region and the surrounding healthy tissue, aiding in their accurate identification and segmentation. This can be crucial for identifying the boundaries, shapes, and other characteristics of bone cancerous cells, which may be critical for accurate segmentation. By applying a Median filter to the MRI images, the filter can effectively remove high-frequency noise components while preserving important details, such as the boundaries and structural characteristics of bone cancerous cells. This preservation of details is essential for accurately delineating and segmenting the cancerous regions. The median filter achieves noise reduction by replacing outlier pixel values, which are likely to be noise, with the median value within a defined neighborhood. This process effectively smoothed the MRI image while retaining the sharpness and integrity of structures, such as the boundaries of the cancerous bone cells. Consequently, the contrast stretching and median filter help improve the visibility and clarity of the cancerous regions within the MRI images.

4.3. Reconstruct MRI Images into 3D Model

The enhanced MRI images with contrast stretching and median filter will then be reconstructed into a 3D volume of MRI images, which involves the conversion of DICOM files to NIfTI files. DICOM image slices are stacked to build a 3D representation of the MRI image. The acquired MRI DICOM images had dimensions of length, width, and height, so the resolution of the reconstructed 3D images may vary from the acquired DICOM images. In this study, DICOM images were obtained from the hospital, so each patient may have a different number of slices. The assembled DICOM into a 3D model for T1W, T2W, and T1W + Gd are shown in Table 6. The reconstruction of the 3D model needed DICOM image slices in different planes, such as axial, coronal, and sagittal.

4.4. Transformation of 3D Volumes

The transformation techniques applied in this project included image loading, adding channels to the 3D volume and label, data orientation based on the affine matrix, spacing adjustment by pixel dimension, scale intensity ranging in the aspect of contrast, cropping the foreground to remove the zero border, randomly cropping patch samples from the big image, and lastly, the random affine that would be able to perform rotation, scaling, shearing, and translation based on the PyTorch affine transform. These transforms are composed with “Compose” to create a fast pipeline. Based on the result obtained as shown in Table 7, the images were resampled to a voxel size of 1.5, 1.5, and 2.0 mm in each dimension to avoid any dimension error that might occur when loading the images into the network. In addition, the 3D sub-volumes were also padded to sizes of 96, 96, and 96 to make sure the input sizes of the images to be loaded into the neural network were consistent. Moreover, the cropping of the foreground and random cropping of patch samples were also effective, as shown in Table 7, where the ROI of the images was amplified to better focus on the ROI when training the neural network. This transformation method needed to be applied to both the input image and the segmentation mask.

4.5. Quantitative and Qualitative Evaluation of 3D U-Net Model

The 3D U-Net was trained and validated with T1W, T2W, and T1W + Gd MRI image datasets through the implementation of Adam as the optimizer algorithm and Sigmoid as the activation function on the GPU in Google Colaboratory. The 46 MRI images for each MRI image type were split into 36 MRI images for training, 5 MRI images for validation, and 5 MRI images for testing. These three types of MRI datasets took about 6 hours to train the 3D U-Net model, as shown in Table 8.
Based on Table 8, the proposed algorithm obtained a good DSC on all the MRI image types. The T2W and T1W + Gd achieved excellent segmentation results with approximately 85% and 87% of the validation mean DSC at epochs 792 and 700, respectively, while the T1W obtained a good result with around 83% of the validation mean DSC at epoch 786. The validation mean DSC of the T1W image dataset was the lowest among these three datasets. This is because T2W optimally shows fluid and abnormalities such as tumors, inflammation, and trauma [33], and T1W + Gd inhibits the fat signal in T1W, which then increases its significance in assessing tumor vascularization [34]. In contrast, T1W optimally shows normal soft tissue anatomy and fat [33]. For this reason, T1W might have a low contrast for the tumors to be identified as compared to T2W and T1W + Gd. The dice loss function is defined as 1 minus the DSC in order to indicate the loss function’s convergence. The epoch average dice loss produced by these three MRI datasets was approximately between 0.15 and 0.17, as shown in Table 8. The smaller the value of the epoch average dice loss, the less error there is in the validation dataset.
Different examples of different MRI image types after implementing the proposed segmentation model are shown in Table 9. Based on the segmentation result, the output test image for each T1W, T2W, and T1W + Gd could be compared with the ground truth. There were 3 samples with their respective 80th slice of predicted segmentation output from the proposed segmentation method’s qualitative results. The output of samples 1 and 2 showed that the majority of the tumor was accurately segmented for T2W and T1W + Gd, with slight border and small hole errors. However, it can be clearly observed that the output of T1W had a false negative as compared to the ground truth, in which some areas of the ROI were not predicted. Other than that, there was another common error shown in sample 3, which was a false positive. It can be observed that there should be no tumor being segmented in sample 3. Yet, a small false-positive tumor was found in the output of the T2 image due to the presence of tissue with a similar look.
In order to have a better observation of the result obtained, the overlaid slices of sample 3 from slices 60th, 70th, 80th, 90th, and 100th were used as representative examples for each T1W, T2W, and T1W + Gd, as shown in Table 10, Table 11 and Table 12. The last column of the table shows the overlaid slices between the ground truth and the predicted output, where the red tint denotes the output predicted by the model, the white tint denotes the true ground truth, and the black color indicates the background.
The predicted output of the 60th slice of each type of MRI image did not show any tumor, which was segmented correctly as compared with ground truth for each T1W, T2W, and T1W + Gd as in Table 12. From the overlaid slice of the 70th slice for these three types of MRI images, it was obviously noticed that there was a bit of false-negative tumor wrongly predicted by the trained model. However, in the 70th slice of these three types of MRI images, T2W performed the best among these MRI image types because it predicted the location of a small detail correctly, while the other two image types did not. In addition, there was also a small false-positive tumor shown in these overlaid slices. Moreover, from the observed results of 80th, 90th, and 100th, even though some of the slices of ground truth were not fully overlapped with the output, most of the slices as shown were predicted almost the same, with a small false-negative tumor as compared with ground truth.

4.6. Visualisation of the Predicted Output

In order to have a clearer visualization of the segmented data, the TensorBoard 3D plugin was used to view the entirety of their three-dimensional deep-learning model output. Thereby, the size and position of the tumor can be clearly observed through this 3D viewer. Figure 5 shows the 3D model output, where the red tint denotes the predicted tumor from the trained model.

4.7. Comparision DSC with Previous Research Works

Based on the comparison of DSC with other researchers’ segmentation models for osteosarcoma MRI images in Figure 6, OSTransnet (by F. Liu, 2022) [15] stands out for its superior segmentation performance, incorporating contextual information and edge enhancement, although its limitations were not outlined. PESNet (Baolong Lv, 2022) enhances tumor localization and segmentation accuracy through a priori generation and feature enrichment networks, but it has a slightly higher computation time than U-Net [35]. MSFCN allows for multi-scale feature integration and can handle objects at various scales. It can be useful for tasks like semantic segmentation and object detection. However, MSFCN may not be explicitly designed for 3D or volumetric data. Its performance can vary depending on the specific task and dataset. SepUNet (by J. Wu, 2022) [16] is a powerful model for osteosarcoma segmentation, providing improved accuracy and efficiency. It excels at handling tumors of various sizes while maintaining a small parameter count, making it accessible and computationally efficient. However, specific limitations were not mentioned. SLIC-S (by E. B. Kayal, 2020) [36] provides efficient superpixel-based segmentation into compact, spatially connected regions. But it may struggle with fine-grained details and be sensitive to the parameter settings of the superpixel generation process. Our 3D U-Net is designed for volumetric segmentation, utilizing 3D convolutions and skip connections to capture spatial information and fuse multi-scale features. However, it requires high memory and computational resources due to volumetric data processing, which provided a slightly lower 0.8762 of DSC. FCM (by E. B. Kayal, 2020) [35] offers flexibility in unsupervised clustering and deals effectively with complex data distributions and partial volume effects. But it may face computational complexity and sensitivity to parameters such as ambiguous data, which caused the lowest DSC of 0.87. From the comparison, SepUNet, PESNet, FCM, and SLIC-S are primarily used for 2D segmentation, whereas 3D U-Net and OSTransnet are tailored for 3D segmentation tasks. They are specifically designed to process 3D volumes or stacks of medical images, allowing for the segmentation of objects or regions in the volumetric space. These models leverage the spatial information and contextual cues present in three-dimensional data to achieve accurate and comprehensive segmentations.

5. Conclusions

The proposed segmentation model was 3D U-Net using MONAI. After a series of training epochs, the model with the highest dice similarity coefficient value was saved for these three different types of MRI images. In the perspective of quantitative analysis, the T1W, T2W, and T1W + Gd achieved good validation mean DSC of 83.75%, 85.45%, and 87.62%, respectively, at epochs 786, 792, and 700. In addition, the epoch average dice loss decreased as the training went over the epoch and got closer to zero, which was about 0.15 to 0.17. From the perspective of qualitative analysis, the overlaid images of the predicted output and ground truth showed high overlapping, with a small false negative and false positive appearing. Although the relative results were good, some of the qualitative data indicate that the proposed method was still constrained by ill-defined borders, missing lesion portions, and other confounding factors. Furthermore, even though all training photos have the determined minimum lesion size maximized, it is possible that some small lesions are still filtered out during testing. In summary, the larger tumor segmentation performed significantly better with the model trained for these three datasets.

Author Contributions

Conceptualization, C.C.L. and Y.F.C.; methodology, A.H.W.L.; software, Y.F.C.; validation, M.E.A. and K.A.; formal analysis, K.A.; investigation, M.Y.M.; resources, K.A.; data curation, M.E.A.; writing—original draft preparation, C.C.L. and A.H.W.L.; writing—review and editing, C.C.L. and M.E.A.; visualization, Y.F.C.; supervision, M.Y.M.; project administration, K.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, approved by the Institutional Review Board of Hospital Universiti Sains Malaysia (JEPeM USM) on 13 December 2020, the approval protocol code USM/JEPeM/20070380.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are unavailable due to privacy or ethical restrictions; a statement is still required.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mandava, R.; Alia, O.M.; Wei, B.C.; Ramachandram, D.; Aziz, M.E.; Shuaib, I.L. Osteosarcoma segmentation in MRI using dynamic Harmony Search based clustering. In Proceedings of the 2010 International Conference of Soft Computing and Pattern Recognition, Cergy-Pontoise, France, 7–10 December 2010; pp. 423–429. [Google Scholar] [CrossRef]
  2. Survival Rates for Osteosarcoma. Available online: https://www.cancer.org/cancer/osteosarcoma/detection-diagnosis-staging/survival-rates.html (accessed on 4 November 2021).
  3. Signs and Symptoms of Osteosarcoma. Available online: https://www.cancer.org/cancer/osteosarcoma/detection-diagnosis-staging/signs-and-symptoms.html (accessed on 4 November 2021).
  4. Nasor, M.; Obaid, W. Segmentation of osteosarcoma in MRI images by K-means clustering, Chan-Vese segmentation, and iterative Gaussian filtering. IET Image Process. 2021, 15, 1310–1318. [Google Scholar] [CrossRef]
  5. Rajeshwari, S.; Sharmila, T.S. Efficient quality analysis of MRI image using preprocessing techniques. In Proceedings of the 2013 IEEE Conference on Information & Communication Technologies, Thuckalay, India, 11–12 April 2013; p. 396. [Google Scholar] [CrossRef]
  6. Suhas, S.; Venugopal, C.R. MRI image preprocessing and noise removal technique using linear and nonlinear filters. In Proceedings of the 2017 International Conference on Electrical, Electronics, Communication, Computer, and Optimization Techniques (ICEECCOT), Mysuru, India, 15–16 December 2017; pp. 1–4. [Google Scholar] [CrossRef]
  7. Mohan, G.; Subashini, M.M. MRI based medical image analysis: Survey on brain tumor grade classification. Biomed. Signal Process. Control 2018, 39, 139–161. [Google Scholar] [CrossRef]
  8. Loizou, C.P.; Pantziaris, M.; Seimenis, I.; Pattichis, C.S. Brain MR image normalization in texture analysis of multiple sclerosis. In Proceedings of the 2009 9th International Conference on Information Technology and Applications in Biomedicine, Larnaka, Cyprus, 4–7 November 2009; pp. 1–5. [Google Scholar] [CrossRef]
  9. Singh, A.; Sengupta, S.; Lakshminarayanan, V. Explainable Deep Learning Models in Medical Image Analysis. J. Imaging 2020, 6, 52. [Google Scholar] [CrossRef] [PubMed]
  10. Debelee, T.G.; Kebede, S.R.; Schwenker, F.; Shewarega, Z.M. Deep Learning in Selected Cancers’ Image Analysis—A Survey. J. Imaging 2020, 6, 121. [Google Scholar] [CrossRef] [PubMed]
  11. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016; Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W., Eds.; Springer International Publishing: Cham, Switzerland, 2016; Volume 9901, pp. 424–432. [Google Scholar] [CrossRef] [Green Version]
  12. Chen, H.; Dou, Q.; Yu, L.; Qin, J.; Heng, P.-A. VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images. NeuroImage 2018, 170, 446–455. [Google Scholar] [CrossRef] [PubMed]
  13. Holbrook, M.D.; Blocker, S.J.; Mowery, Y.M.; Badea, A.; Qi, Y.; Xu, E.S.; Kirsch, D.G.; Johnson, G.A.; Badea, C. MRI-Based Deep Learning Segmentation and Radiomics of Sarcoma in Mice. Tomography 2020, 6, 23–33. [Google Scholar] [CrossRef] [PubMed]
  14. Vaidyanathan, A.; van der Lubbe, M.F.J.A.; Leijenaar, R.T.H.; van Hoof, M.; Zerka, F.; Miraglio, B.; Primakov, S.; Postma, A.A.; Bruintjes, T.D.; Bilderbeek, M.A.L.; et al. Deep learning for the fully automated segmentation of the inner ear on MRI. Sci. Rep. 2021, 11, 2885. [Google Scholar] [CrossRef] [PubMed]
  15. Liu, F.; Zhu, J.; Lv, B.; Yang, L.; Sun, W.; Dai, Z.; Gou, F.; Wu, J. Auxiliary Segmentation Method of Osteosarcoma MRI Image Based on Transformer and U-Net. Comput. Intell. Neurosci. 2022, 2022, 9990092. [Google Scholar] [CrossRef] [PubMed]
  16. Wu, J.; Yang, S.; Gou, F.; Zhou, Z.; Xie, P.; Xu, N.; Dai, Z. Intelligent Segmentation Medical Assistance System for MRI Images of Osteosarcoma in Developing Countries. Comput. Math. Methods Med. 2022, 2022, 1–17. [Google Scholar] [CrossRef] [PubMed]
  17. Labeeb, Y.; Morsy, M.; Abo-Elsoud, M. Preprocessing Technique for Enhancing the DICOM Kidney Images. Int. J. Eng. Res. Technol. 2018, 4, 836–841. [Google Scholar]
  18. Kaur, H.; Rani, J. MRI brain image enhancement using Histogram Equalization techniques. In Proceedings of the 2016 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), Chennai, India, 23–25 March 2016; pp. 770–773. [Google Scholar] [CrossRef]
  19. Koonsanit, K.; Thongvigitmanee, S.; Pongnapang, N.; Thajchayapong, P. Image enhancement on digital X-ray images using N-CLAHE. In Proceedings of the 2017 10th Biomedical Engineering International Conference (BMEiCON), Hokkaido, Japan, 31 August 2017–2 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
  20. Contrast Stretching Using Python and Pillow|Pythontic.com. Available online: https://pythontic.com/image-processing/pillow/contrast%20stretching (accessed on 29 December 2021).
  21. Nuruzzaman, F. Contrast Stretching in Image Processing Using Matlab. 10 February 2019. Available online: https://www.nzfaruqui.com/contrast-stretching-in-image-processing-using-matlab/ (accessed on 29 December 2021).
  22. Ali, H.M. MRI Medical Image Denoising by Fundamental Filters; IntechOpen: Rijeka, Croatia, 2018. [Google Scholar] [CrossRef] [Green Version]
  23. Pawar, M. MRI and CT Image Denoising using Gaussian Filter, Wavelet Transform and Curvelet Transform. Int. J. Eng. Sci. Comput. 2017, 7, 4. [Google Scholar]
  24. Gaussian Blur—An overview|ScienceDirect Topics. Available online: https://www.sciencedirect.com/topics/engineering/gaussian-blur (accessed on 27 December 2021).
  25. Young, I.T.; van Vliet, L.J. Recursive implementation of the Gaussian filter. Signal Process. 1995, 44, 139–151. [Google Scholar] [CrossRef] [Green Version]
  26. 26Evaluating Denoising Performances of Fundamental Filters for T2-Weighted MRI Images|Elsevier Enhanced Reader. Available online: https://reader.elsevier.com/reader/sd/pii/S1877050915023583?token=A8010A91375CC1FBFF21C5C99A2A2F13B2CC26F6A11AB2354554263835A78B6BE1AFA36E1F958D54E39C610CEC98FFB8&originRegion=eu-west-1&originCreation=20211227062644 (accessed on 27 December 2021).
  27. Kaur, R. Comparison of contrast enhancement techniques for medical image. In Proceedings of the 2016 Conference on Emerging Devices and Smart Systems (ICEDSS), Namakkal, India, 4–5 March 2016; pp. 155–159. [Google Scholar] [CrossRef]
  28. Mzoughi, H.; Njeh, I.; Slima, M.B.; Hamida, A.B.; Mhiri, C.; Mahfoudh, K.B. Denoising and contrast-enhancement approach of magnetic resonance imaging glioblastoma brain tumors. J. Med. Imaging 2019, 6, 044002. [Google Scholar] [CrossRef] [PubMed]
  29. Dnuggets, K. Medical Image Analysis with Deep Learning, Part 4. Available online: https://www.kdnuggets.com/medical-image-analysis-with-deep-learning-part-4.html/ (accessed on 3 January 2022).
  30. 30Neuroimaging in Python—Pipelines and Interfaces—Nipy Pipeline and Interfaces Package. Available online: https://nipype.readthedocs.io/en/latest/api/generated/nipype.interfaces.dcm2nii.html (accessed on 3 January 2022).
  31. Haque, I.R.I.; Neubert, J. Deep learning approaches to biomedical image segmentation. Inform. Med. Unlocked 2020, 18, 100297. [Google Scholar] [CrossRef]
  32. Soomro, M.H.; Coppotelli, M.; Conforto, S.; Schmid, M.; Giunta, G.; Del Secco, L.; Neri, E.; Caruso, D.; Rengo, M.; Laghi, A. Automated Segmentation of Colorectal Tumor in 3D MRI Using 3D Multiscale Densely Connected Convolutional Neural Network. J. Heal. Eng. 2019, 2019, 1–11. [Google Scholar] [CrossRef] [PubMed]
  33. 33MSD Manual Professional Edition. Magnetic Resonance Imaging—Special Subjects. Available online: https://www.msdmanuals.com/professional/special-subjects/principles-of-radiologic-imaging/magnetic-resonance-imaging (accessed on 16 June 2022).
  34. Nascimento, D.; Suchard, G.; Hatem, M.; de Abreu, A. The role of magnetic resonance imaging in the evaluation of bone tumours and tumour-like lesions. Insights Imaging 2014, 5, 419–440. [Google Scholar] [CrossRef] [PubMed]
  35. Lv, B.; Liu, F.; Gou, F.; Wu, J. Multi-Scale Tumor Localization Based on Priori Guidance-Based Segmentation Method for Osteosarcoma MRI Images. Mathematics 2022, 10, 2099. [Google Scholar] [CrossRef]
  36. Kayal, E.B.; Kandasamy, D.; Yadav, R.; Bakhshi, S.; Sharma, R.; Mehndiratta, A. Automatic segmentation and RECIST score evaluation in osteosarcoma using diffusion MRI: A computer aided system process. Eur. J. Radiol. 2020, 133, 109359. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overview flow chart.
Figure 1. Overview flow chart.
Diagnostics 13 02377 g001
Figure 2. Types of MRI images. (a) T1-weighted MRI image (T1W); (b) T2-weighted MRI image (T2W); (c) T1-weighted MRI image with an injection of gadolinium (T1W + Gd).
Figure 2. Types of MRI images. (a) T1-weighted MRI image (T1W); (b) T2-weighted MRI image (T2W); (c) T1-weighted MRI image with an injection of gadolinium (T1W + Gd).
Diagnostics 13 02377 g002
Figure 3. Flow chart of the segmentation model for 3D volume.
Figure 3. Flow chart of the segmentation model for 3D volume.
Diagnostics 13 02377 g003
Figure 4. Three-dimensional (3D) U-Net architecture.
Figure 4. Three-dimensional (3D) U-Net architecture.
Diagnostics 13 02377 g004
Figure 5. Three-dimensional model output (red tint denotes the predicted tumor from the trained 3D U-net Model).
Figure 5. Three-dimensional model output (red tint denotes the predicted tumor from the trained 3D U-net Model).
Diagnostics 13 02377 g005
Figure 6. Comparative of DSC between other researchers’ segmentation models for osteosarcoma MRI images [15,16,35,36].
Figure 6. Comparative of DSC between other researchers’ segmentation models for osteosarcoma MRI images [15,16,35,36].
Diagnostics 13 02377 g006
Table 1. Overview of the hyperparameters set for training.
Table 1. Overview of the hyperparameters set for training.
HyperparametersValues
Data split ratio8:1:1
Maximum epochs800
Batch size2
OptimizerAdam
Loss functionDice loss
Activation functionPReLU
Total number of parameters4,808,917
Table 2. Contrast-enhanced T1W MRI image after implementing Gaussian filter and median filter.
Table 2. Contrast-enhanced T1W MRI image after implementing Gaussian filter and median filter.
T1W MRI Image after Contrast EnhancementCombination of Pre-Processing Techniques
CLAHE + Gaussian FilterCLAHE + Median FilterContrast Stretching + Gaussian FilterContrast Stretching + Median Filter
Diagnostics 13 02377 i001Diagnostics 13 02377 i002Diagnostics 13 02377 i003Diagnostics 13 02377 i004Diagnostics 13 02377 i005
Diagnostics 13 02377 i006Diagnostics 13 02377 i007Diagnostics 13 02377 i008Diagnostics 13 02377 i009Diagnostics 13 02377 i010
Table 3. Contrast-enhanced T2W MRI image after implementing Gaussian filter and median filter.
Table 3. Contrast-enhanced T2W MRI image after implementing Gaussian filter and median filter.
T2W MRI Image after Contrast EnhancementAfter Pre-Processing of T2W
CLAHE + Gaussian FilterCLAHE + Median FilterContrast Stretching + Gaussian FilterContrast Stretching + Median Filter
Diagnostics 13 02377 i011Diagnostics 13 02377 i012Diagnostics 13 02377 i013Diagnostics 13 02377 i014Diagnostics 13 02377 i015
Diagnostics 13 02377 i016Diagnostics 13 02377 i017Diagnostics 13 02377 i018Diagnostics 13 02377 i019Diagnostics 13 02377 i020
Table 4. Contrast-enhanced T1W + Gd MRI image after implementing Gaussian filter and median filter.
Table 4. Contrast-enhanced T1W + Gd MRI image after implementing Gaussian filter and median filter.
T1W + Gd MRI Image after Contrast EnhancementAfter Pre-Processing of T1W + Gd
CLAHE +
Gaussian Filter
CLAHE +
Median Filter
Contrast Stretching + Gaussian FilterContrast Stretching + Median Filter
Diagnostics 13 02377 i021Diagnostics 13 02377 i022Diagnostics 13 02377 i023Diagnostics 13 02377 i024Diagnostics 13 02377 i025
Diagnostics 13 02377 i026Diagnostics 13 02377 i027Diagnostics 13 02377 i028Diagnostics 13 02377 i029Diagnostics 13 02377 i030
Table 5. Comparison of MSE, PSNR, and AMBE between T1W, T1W + Gd and T2W implemented with CLAHE or contrast stretching and Gaussian filter or median filter.
Table 5. Comparison of MSE, PSNR, and AMBE between T1W, T1W + Gd and T2W implemented with CLAHE or contrast stretching and Gaussian filter or median filter.
MSEPSNRAMBE
T1WT1W + GdT2WT1WT1W + GdT2WT1WT1W + GdT2W
CLAHE +
Gaussian filter
MRI 199.527098.0830110.434828.151428.214927.69970.060930.070080.08086
MRI 277.634773.5121106.979329.230229.467227.83780.076910.085380.10174
MRI 388.385986.8852108.999528.667028.741327.75660.044680.060800.07764
MRI 493.6140105.3020108.035528.417427.906427.79510.042000.054080.05596
MRI 594.4030100.5126110.521228.380928.108627.69640.043470.050960.05848
AVG90.712992.8589108.994028.569328.487627.75710.053590.064260.07493
CLAHE +
Median
filter
MRI 188.824488.311699.660528.645528.670628.14560.058050.068310.07490
MRI 277.012072.0549100.050629.265229.554228.12860.074750.081460.09748
MRI 381.680482.013897.034529.009628.991928.26150.043610.059400.06829
MRI 476.861691.1990100.102029.273728.530928.12640.383080.050180.05279
MRI 583.494987.9688101.840328.914228.687528.05160.040840.047810.05486
AVG81.574784.309699.737629.021628.887028.14270.12010.06140.0697
Contrast Stretching +
Gaussian filter
MRI 122.508718.850024.159934.607335.377734.15990.012510.015700.02479
MRI 240.481838.774638.274132.058232.245332.30180.046880.044770.04038
MRI 322.082221.936620.200934.690434.719135.07710.011520.016970.01564
MRI 47.368055.70698.903839.457330.671738.63510.012160.034400.01196
MRI 57.455515.07789.904139.406036.347438.17260.010610.020000.01385
AVG19.979230.069220.288636.043833.872235.66930.018740.026470.02132
Contrast Stretching +
Median
filter
MRI 121.713418.588823.622034.763535.438334.39760.011970.015220.02406
MRI 239.523838.635836.755032.162232.260932.47760.046320.044440.03961
MRI 320.948020.848719.348034.919434.940035.26440.010990.016600.01492
MRI 47.667452.15989.198139.284430.957538.49380.011360.033000.01047
MRI 57.592114.869010.350639.327236.408037.98120.009860.019060.01293
AVG19.488929.020419.854736.091334.000935.72290.018100.025660.02040
Note: The best reading of each categories is highlighted and compare between MRI images.
Table 6. Assembling DICOM into 3D model.
Table 6. Assembling DICOM into 3D model.
Type of MRIPlaneDICOM Image3D Model in NIfTI
T1W AxialDiagnostics 13 02377 i031Diagnostics 13 02377 i032
CoronalDiagnostics 13 02377 i033
SagittalDiagnostics 13 02377 i034
T1W + GdAxialDiagnostics 13 02377 i035Diagnostics 13 02377 i036
CoronalDiagnostics 13 02377 i037
SagittalDiagnostics 13 02377 i038
T2WAxialDiagnostics 13 02377 i039Diagnostics 13 02377 i040
CoronalDiagnostics 13 02377 i041
SagittalDiagnostics 13 02377 i042
Table 7. Transformation of 3D volumes.
Table 7. Transformation of 3D volumes.
Slice of Original ImageTransformation
for Ground Truth
After Transformation
T1WT1W + GdT2W
Diagnostics 13 02377 i043Diagnostics 13 02377 i044Diagnostics 13 02377 i045Diagnostics 13 02377 i046Diagnostics 13 02377 i047
Diagnostics 13 02377 i048Diagnostics 13 02377 i049Diagnostics 13 02377 i050Diagnostics 13 02377 i051Diagnostics 13 02377 i052
Diagnostics 13 02377 i053Diagnostics 13 02377 i054Diagnostics 13 02377 i055Diagnostics 13 02377 i056Diagnostics 13 02377 i057
Note: Yellow indicates the ROI of MRI images.
Table 8. The training time, mean dice similarity coefficient (DSC) and the epoch average dice loss of model of different MRI image types.
Table 8. The training time, mean dice similarity coefficient (DSC) and the epoch average dice loss of model of different MRI image types.
Types of MRI ImageEpochTraining Time (Second)Mean DSCEpoch Average Dice Loss
T1W78620,194.5630.83750.1709
T2W79220,429.4270.85450.1563
T1W + Gd70020,020.0690.87620.1534
Table 9. Segmentation output (highlighted in yellow) of the osteosarcoma cancerous cell.
Table 9. Segmentation output (highlighted in yellow) of the osteosarcoma cancerous cell.
Sample No.
(MRI Slice)
InputGround TruthOutput
T1WT2WT1W + Gd
1
(80th slice)
Diagnostics 13 02377 i058Diagnostics 13 02377 i059Diagnostics 13 02377 i060Diagnostics 13 02377 i061Diagnostics 13 02377 i062
2
(80th slice)
Diagnostics 13 02377 i063Diagnostics 13 02377 i064Diagnostics 13 02377 i065Diagnostics 13 02377 i066Diagnostics 13 02377 i067
3
(80th slice)
Diagnostics 13 02377 i068Diagnostics 13 02377 i069Diagnostics 13 02377 i070Diagnostics 13 02377 i071Diagnostics 13 02377 i072
Table 10. Overlaid slices for output of the osteosarcoma cancerous cell in T1W MRI image for sample 3.
Table 10. Overlaid slices for output of the osteosarcoma cancerous cell in T1W MRI image for sample 3.
MRI
Slice
InputGround Truth (G)Output (O)Overlaid Slices
Ground TruthOutputGround Truth
and Output,
(G ∩ O)
60thDiagnostics 13 02377 i073Diagnostics 13 02377 i074Diagnostics 13 02377 i075Diagnostics 13 02377 i076Diagnostics 13 02377 i077Diagnostics 13 02377 i078
70thDiagnostics 13 02377 i079Diagnostics 13 02377 i080Diagnostics 13 02377 i081Diagnostics 13 02377 i082Diagnostics 13 02377 i083Diagnostics 13 02377 i084
80thDiagnostics 13 02377 i085Diagnostics 13 02377 i086Diagnostics 13 02377 i087Diagnostics 13 02377 i088Diagnostics 13 02377 i089Diagnostics 13 02377 i090
90thDiagnostics 13 02377 i091Diagnostics 13 02377 i092Diagnostics 13 02377 i093Diagnostics 13 02377 i094Diagnostics 13 02377 i095Diagnostics 13 02377 i096
100thDiagnostics 13 02377 i097Diagnostics 13 02377 i098Diagnostics 13 02377 i099Diagnostics 13 02377 i100Diagnostics 13 02377 i101Diagnostics 13 02377 i102
Note: Ground truth (G): yellow denotes manually segmented cancerous cell by medical doctors; Output (O): yellow denotes segmented cancerous cell output predicted by 3D U-net model; Ground Truth and Output under overlaid slide (GO): red denotes overlapping area of the ground truth and output, white denotes non-overlapping area of the ground truth and output, black denotes the background.
Table 11. Overlaid slices for output of the osteosarcoma cancerous cell in T1W + Gd MRI image for sample 3.
Table 11. Overlaid slices for output of the osteosarcoma cancerous cell in T1W + Gd MRI image for sample 3.
MRI
Slice
InputGround TruthOutputOverlaid Slices
Ground TruthOutputGround Truth
and Output
60thDiagnostics 13 02377 i103Diagnostics 13 02377 i104Diagnostics 13 02377 i105Diagnostics 13 02377 i106Diagnostics 13 02377 i107Diagnostics 13 02377 i108
70thDiagnostics 13 02377 i109Diagnostics 13 02377 i110Diagnostics 13 02377 i111Diagnostics 13 02377 i112Diagnostics 13 02377 i113Diagnostics 13 02377 i114
80thDiagnostics 13 02377 i115Diagnostics 13 02377 i116Diagnostics 13 02377 i117Diagnostics 13 02377 i118Diagnostics 13 02377 i119Diagnostics 13 02377 i120
90thDiagnostics 13 02377 i121Diagnostics 13 02377 i122Diagnostics 13 02377 i123Diagnostics 13 02377 i124Diagnostics 13 02377 i125Diagnostics 13 02377 i126
100thDiagnostics 13 02377 i127Diagnostics 13 02377 i128Diagnostics 13 02377 i129Diagnostics 13 02377 i130Diagnostics 13 02377 i131Diagnostics 13 02377 i132
Note: Ground truth (G): yellow denotes manually segmented cancerous cell by medical doctors; Output (O): yellow denotes segmented cancerous cell output predicted by 3D U-net model; Ground Truth and Output under overlaid slide (GO): red denotes overlapping area of the ground truth and output, white denotes non-overlapping area of the ground truth and output, black denotes the background.
Table 12. Overlaid slices for output of the osteosarcoma cancerous cell in T2W MRI image sample 3.
Table 12. Overlaid slices for output of the osteosarcoma cancerous cell in T2W MRI image sample 3.
MRI
Slice
InputGround TruthOutputOverlaid Slices
Ground TruthOutputGround Truth
and Output
60th Diagnostics 13 02377 i133Diagnostics 13 02377 i134Diagnostics 13 02377 i135Diagnostics 13 02377 i136Diagnostics 13 02377 i137Diagnostics 13 02377 i138
70th Diagnostics 13 02377 i139Diagnostics 13 02377 i140Diagnostics 13 02377 i141Diagnostics 13 02377 i142Diagnostics 13 02377 i143Diagnostics 13 02377 i144
80th Diagnostics 13 02377 i145Diagnostics 13 02377 i146Diagnostics 13 02377 i147Diagnostics 13 02377 i148Diagnostics 13 02377 i149Diagnostics 13 02377 i150
90th Diagnostics 13 02377 i151Diagnostics 13 02377 i152Diagnostics 13 02377 i153Diagnostics 13 02377 i154Diagnostics 13 02377 i155Diagnostics 13 02377 i156
100th Diagnostics 13 02377 i157Diagnostics 13 02377 i158Diagnostics 13 02377 i159Diagnostics 13 02377 i160Diagnostics 13 02377 i161Diagnostics 13 02377 i162
Note: Ground truth (G): yellow denotes manually segmented cancerous cell by medical doctors; Output (O): yellow denotes segmented cancerous cell output predicted by 3D U-net model; Ground Truth and Output under overlaid slide (GO): red denotes overlapping area of the ground truth and output, white denotes non-overlapping area of the ground truth and output, black denotes the background.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lim, C.C.; Ling, A.H.W.; Chong, Y.F.; Mashor, M.Y.; Alshantti, K.; Aziz, M.E. Comparative Analysis of Image Processing Techniques for Enhanced MRI Image Quality: 3D Reconstruction and Segmentation Using 3D U-Net Architecture. Diagnostics 2023, 13, 2377. https://doi.org/10.3390/diagnostics13142377

AMA Style

Lim CC, Ling AHW, Chong YF, Mashor MY, Alshantti K, Aziz ME. Comparative Analysis of Image Processing Techniques for Enhanced MRI Image Quality: 3D Reconstruction and Segmentation Using 3D U-Net Architecture. Diagnostics. 2023; 13(14):2377. https://doi.org/10.3390/diagnostics13142377

Chicago/Turabian Style

Lim, Chee Chin, Apple Ho Wei Ling, Yen Fook Chong, Mohd Yusoff Mashor, Khalilalrahman Alshantti, and Mohd Ezane Aziz. 2023. "Comparative Analysis of Image Processing Techniques for Enhanced MRI Image Quality: 3D Reconstruction and Segmentation Using 3D U-Net Architecture" Diagnostics 13, no. 14: 2377. https://doi.org/10.3390/diagnostics13142377

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop