Next Article in Journal
Empirical Study of a Room-Level Localization System Based on Bluetooth Low Energy Beacons
Next Article in Special Issue
Diabetic Retinopathy Fundus Image Classification and Lesions Localization System Using Deep Learning
Previous Article in Journal
Diabetic Retinopathy Prediction by Ensemble Learning Based on Biochemical and Physical Data
Previous Article in Special Issue
Quantification of Blood Flow Velocity in the Human Conjunctival Microvessels Using Deep Learning-Based Stabilization Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Precise Identification of Prostate Cancer from DWI Using Transfer Learning

1
Bioengineering Department, University of Louisville, Louisville, KY 40292, USA
2
Faculty of Computers and Information, Mansoura University, Dakahlia 35516, Egypt
3
Radiology Department, Urology and Nephrology Center, University of Mansoura, Dakahlia 35516, Egypt
4
Electrical Engineering Department, Assiut University, Assiut 71515, Egypt
5
College of Computer and Information Science, Princess Nourah Bint Abdulrahman University, Riyadh 11564, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3664; https://doi.org/10.3390/s21113664
Submission received: 27 March 2021 / Revised: 17 May 2021 / Accepted: 18 May 2021 / Published: 25 May 2021
(This article belongs to the Special Issue Computer Aided Diagnosis Sensors)

Abstract

:
Background and Objective: The use of computer-aided detection (CAD) systems can help radiologists make objective decisions and reduce the dependence on invasive techniques. In this study, a CAD system that detects and identifies prostate cancer from diffusion-weighted imaging (DWI) is developed. Methods: The proposed system first uses non-negative matrix factorization (NMF) to integrate three different types of features for the accurate segmentation of prostate regions. Then, discriminatory features in the form of apparent diffusion coefficient (ADC) volumes are estimated from the segmented regions. The ADC maps that constitute these volumes are labeled by a radiologist to identify the ADC maps with malignant or benign tumors. Finally, transfer learning is used to fine-tune two different previously-trained convolutional neural network (CNN) models (AlexNet and VGGNet) for detecting and identifying prostate cancer. Results: Multiple experiments were conducted to evaluate the accuracy of different CNN models using DWI datasets acquired at nine distinct b-values that included both high and low b-values. The average accuracy of AlexNet at the nine b-values was 89.2 ± 1.5 % with average sensitivity and specificity of 87.5 ± 2.3 % and 90.9 ± 1.9 % . These results improved with the use of the deeper CNN model (VGGNet). The average accuracy of VGGNet was 91.2 ± 1.3 % with sensitivity and specificity of 91.7 ± 1.7 % and 90.1 ± 2.8 % . Conclusions: The results of the conducted experiments emphasize the feasibility and accuracy of the developed system and the improvement of this accuracy using the deeper CNN.

1. Introduction

Prostate cancer is a major health problem, especially in western countries. For example, this disease is the second leading cause of mortality among males in the United States [1]. In 2020, the number of new prostate cancer cases and the number of deaths caused by prostate cancer among Americans are expected to be 191,930 and 33,330, respectively [1]. Currently, the definitive technique of diagnosing prostate cancer is transrectal ultrasound (TRUS)-guided biopsy. However, biopsy is an invasive procedure that can miss up to 30 % of cancers [2]. In order to minimize the errors of detecting prostate cancer by TRUS-guided biopsy, many alternatives, such as magnetic resonance imaging (MRI)-guided biopsy, have been investigated [3].
Recently, MRI has evolved in its capabilities of detecting prostate cancer in addition to its use in guiding biopsies for better accuracy. Multiple MRI sequences, which include T2-weighted MRI, diffusion-weighted imaging (DWI), and dynamic contrast-enhanced MRI (DCE-MRI), are used in various clinical tasks, such as active surveillance, localization, and determining the stage of prostate cancer [4,5]. However, detecting and localizing prostate cancer from MRI data is a challenging task. As large volumes of MRI data, from different protocols and sometimes from different scanners, have to be analyzed, these variations can lead to inter-observer differences. Computer-aided detection (CAD) systems can help physicians make objective and fast decisions. These systems can also enhance the quantitative evaluation of prostate cancer.
Developing MRI-based CAD systems for identifying prostate cancer has become a subject of active research [6]. For instance, Viswanath et al. [7] examined the performance of multiple classical classifiers and their bagging and boosting ensembles using a multi-institutional T2-weighed MRI dataset of 85 subjects. These classifiers were fed with radiomic features and their performance was evaluated using the receiver operating characteristic (ROC) curve. The highest average area under the curve (AUC) was obtained by the boosted quadratic discriminant analysis. Riccardo et al. [8] found that the accuracy of targeted biopsy improved by 13.2 % in case of combining physician analysis of multiparametric MRI with CAD results. Rampun et al. [9] compared the accuracy of eleven classifiers using a T2-weighed MRI dataset of 45 subjects. Their system employed feature selection on a feature space of size 215 to select the best discriminating features. The highest resulting accuracy was 85.5 % . More detailed and profound literature review can be found in the recent survey by Gao et al. [10].
Although existing systems have achieved satisfactory results, these systems base their diagnosis on handcrafted features that are validated on small datasets. The good empirical design of these handcrafted features determines their accuracy. An alternative approach for handcrafted features is to learn the discriminating features automatically.
Deep learning structures, especially convolutional neural networks (CNN), are able to automatically learn multiple levels of features from data in a hierarchical manner [11]. These structures have achieved accurate results in multiple computer vision tasks [12,13,14,15,16] as well as lesions detection tasks [17]. Ishioka et al. [18], developed a prostate cancer CAD algorithm that aimed to reduce the variation in the interpretation of T2-weighted MRI. They used histogram smoothing to convert the 12-bit intensity data from the original T2-weighted MR images into 8-bit images. These images were normalized and subjected to data augmentation to increase the employed training data. The detection of prostate cancer was obtained using a CNN architecture that combined both U-net and ResNet50. Their algorithm resulted in an AUC of 0.65. Mehrtash et al. [19] evaluated the performance of CNN with different MRI modalities. The highest performance achieved by their system in terms of AUC was 0.8. This performance was achieved by including zonal information of tumors with DWI and DCE-MRI. Yang et al. [20] proposed a system for detecting and localizing the presence of prostate cancer from T2-weighted MRI and apparent diffusion coefficient (ADC) images. Their system used an individual CNN for each of the used two modalities to produce a response map indicating the malignancy likelihood of each pixel. An average pooling is performed before the last convolutional layer of each CNN to obtain a feature vector. The feature vectors from each modality are concatenated and used as input into a support vector machine (SVM). A sensitivity of 46 % and 97 % was achieved at 0.1 and 10 false positives per normal case. Le et al. [21] integrated a similarity cost function with the CNN cost function to better fuse ADCs with T2-weighted MRI. The authors investigated multiple data augmentation techniques, multiple CNN architectures, and multiple fusion schemes to find out the combination that can lead to the best accuracy. The final differentiation between malignant and benign tissues was based on the integration between the results of CNN and the results of handcrafted features using an SVM classifier. An accuracy of 91.5 % was achieved. Wang et al. [22] found that optimizing the steps of prostate segmentation, multi-modal registration and cancer localization in a joint manner can reduce the computation burdens of optimizing each step individually. Moreover, this joint optimization improves the accuracy by reducing the accumulation of errors over the different steps. Song et al. [23] developed a patch-based CNN system for differentiating between malignant and benign prostate lesions from multiparametric MRI. Their proposed CNN model consisted of three blocks of layers before the final fully-connected (FC) layers. Each block consisted of three convolutional layers followed by a dropout layer and a max-pooling layer. Their model resulted in sensitivity and specificity of 87 % and 90.6 % , respectively. Hosseinzadeh et al. [24] developed 2D-UNet model to produce maps of prostate zones. In their system, the inclusion of zonal information improved the average sensitivity at three different false positives by 5 % .
Schelb et al. [25] used a dataset of two modalities (T2-weighted MRI and DWI) to train a U-Net classifier. Their system showed that the performance of a U-Net classifier is similar to the clinical assessment. Xu et al. [26] developed a system to detect prostate lesions using a residual network. Their system resulted in an accuracy of 93 % . Yuan et al. [27] developed a system for classifying prostate cancer from three MRI modalities. Their systems employed three CNNs. Each CNN was fed with a different modality to learn discriminative features. Their system resulted in an accuracy of 86.9 % . Chen et al. [28] tuned two pre-trained CNNs, namely, InceptionV3 and VGG-16, using PROSTATEx challenge dataset. An AUC of 0.83 was obtained using VGG-16. Abbasi et al. [29] tuned a pre-trained CNN model(GoogLeNet) using a prostate cancer MRI dataset. They compared GoogLeNet with other classifiers, such as, decision tree and SVM. They found that GoogLeNet outperformed these conventional classifiers. A recent survey by Wildeboer et al. [30] listed more than 80 CAD systems for diagnosing prostate cancer. In this survey, the different CAD systems were categorized according to the employed imaging modalities and the employed classifiers.
Training deep architectures with huge numbers of parameters from scratch requires large amounts of data. Typically, the amount of data in the medical domain is small when compared to that of natural images used in conventional computer vision applications. Training deep architectures with small amounts of data can lead to overfitting, which means the network can correctly classify the data used for training but is not able to correctly classify new data (i.e., the network does not generalize well). Moreover, if the amount of medical data is sufficient, annotating this data by multiple experts to prepare it for training can be an impeding factor. In order to overcome these limitations, this work modifies previously-trained CNNs and fine-tunes them for detecting and identifying prostate cancer. The process of fine-tuning such deep networks can be done with small datasets.
The main contribution of this work does not depend on which deep learning network is used; the main contribution is representing the input data in a different form (ADC maps) to be more separable to achieve acceptable accuracy, no matter which deep learning network is used. To prove this point, AlexNet, which is a deep learning network that has only five convolutional layers, is used in the beginning. AlexNet achieved an average accuracy of 89.2 ± 1.5 % at the nine b-values. For further validation, VGGNet, which has more layers that increase its learning ability, is then used. VGGNet achieved an average accuracy of 91.2 ± 1.3 % at the nine b-values. The ADC maps used in the proposed system are calculated at both low and high b-values. This enables the proposed system to capture both blood perfusion and water diffusion for an accurate diagnosis of prostate cancer. The b-value is a measure of the degree of diffusion weighting employed to generate diffusion-weighted images. It is a parameter that relies on the timing and strength of the used gradient pulses. The details of this process are discussed in the following sections.

2. Methods and Materials

The DWI datasets used in this work were collected from 37 subjects (16 benign and 21 malignant) using two distinct scanners (1.5 Tesla (T) scanner and 3T scanner) at nine distinct b-values (100, 200, …, 800, 1000 s/mm2). The acquisition parameters of the 1.5 T scanner in axial plane were: TE = 84.6 ms, TR = 8000 ms, Bandwidth = 142.86 kHz, FOV = 34 cm, slice thickness = 3 mm, inter-slice gap = 0 mm, voxel size = 1.25 × 1.25 × 3.00 mm3. The 3T DW images were acquired in the transverse plane. The acquisition parameters of the second 3T scanner were: TR = 4300–4800 ms, TE = 75–88 ms, acquisition matrix = 128 × 126 , reconstruction matrix = 256 × 256 , FOV = 20 cm. The scanning was done using full-echo and full-k-space. However, a SENSE acceleration factor of 2 was used, thus skipping over every-other line in k-space. Excitation and read-out were spectral fat-suppressed, standard 2D multi-slice 90–180 spin-echo Stejskal-Tanner DWI, with single-shot echo-planar-imaging read-out. A single diffusion encoding direction of [1,1,1] was used (i.e., X,Y,Z gradient channels were on simultaneously) to obtain minimal TE at the maximum b-value. Each b-value was averaged 7 times. Big delta = 47 ms and little delta = 15 ms. All the cases involved in this study performed MRI when there was a clinical suspicion of malignancy. The final diagnosis of the cases was established by biopsy that was carried out after MRI. Therefore, the ground truth of diagnosis of all subjects was based on the results of the biopsy. From the malignant cases, 234 slices were labelled manually by a radiologist as having malignant tumors. The analysis and labeling of all the cases were performed in a slice-wise manner. A similar number, 236, of DW slices from benign cases were chosen to create a balanced dataset of malignant and benign slices. These slices represent all the DW slices of 13 benign cases in addition to 5 slices from the remaining 3 benign cases. The ADC maps of the DWI datasets were calculated, as will be explained in the following subsection. The corresponding 470 ADC maps of the labelled DW slices were used to train and test the performance of the developed model, as explained in the following sections.
Figure 1 shows the general framework of the developed CAD system for detecting and identifying prostate cancer, which incorporates three main steps: prostate segmentation, identifying discriminatory features, and identifying slices with prostate cancer. Prostate segmentation integrates three types of features using non-negative matrix factorization (NMF) to guide the evolution of a level set model. These features are shape-priors of prostates, intensities of voxels and spatial features of neighboring voxels. The effect of incorporating each of these features into the accuracy of the used segmentation approach is explained in detail in [31]. Appearance, shape and spatial information were extracted for each voxel. NMF was used to reduce and make the combined features more separable. Curvature and Euclidean distances between the reduced features (test subject) to the centroids of NMF classes (calculated from training subjects) were used to estimate the voxel-wise guiding force of the level set. If the voxel belongs to the prostate, the level set grows. Otherwise, the level set shrinks. More details about the used segmentation approach can be found in [31]. The second step was identifying discriminatory features, which can discriminate malignant from benign cases. In this study, ADC volumes of the segmented prostates were estimated and used for this purpose. By nature, the prostate is a small organ compared to other organs. To be sure that the proposed model would capture the features that discriminate between malignant tumors and benign tumors, the ADC features were calculated only from the prostate region. This ensured that the system learned the features related only to prostate cancer. The first two processing steps of the proposed framework are illustrated for two different cases (one benign and one malignant) in Figure 2.
The 2D cross sections that constitute these ADC volumes were extracted and used as input to the employed CNN-based model. This process is shown in Figure 3. The final step was identifying slices with tumors using a previously-trained CNN model. In the following subsections, the details of estimating the ADC maps and the use of these maps to fine-tune CNN models for identifying prostate cancer are presented.

2.1. Identifying Discriminatory Features

Currently, DWI is one of the promising modalities utilized for the identification of prostate cancer. DWI is a functional MRI modality similar to DCE-MRI. However, what distinguishes DWI is its fast acquisition time, as there are no contrast agents used. The problem with contrast agents is the potential harm they cause to patients who have kidney disorders. DWI depends on the differences in the motion of water molecules inside the body to create images. DW images visualize and quantify this microscopic motion [32]. The molecules’ motion is random, and there is a positive correlation between the level of randomness and the loss in the signal, which is given by [33]:
S d e b × A D C ,
where b is a parameter that relies on the timing and strength of the used gradient pulses, and ADC is a measure of the magnitude of water molecules’ diffusion within the tissues. The utilization of gradient pulses gives rise to enhanced diffusion sensitivity compared to the steady state gradients [34].
The intensities of pixels in a slice acquired at a specific b-value ( S b ) are equal to the intensities of the congruent pixels of the baseline slice (b = 0 s/mm2) lowered by the signal loss defined in Equation (1). These intensities are given by:
S b = S 0 × e b × A D C .
There is a negative relationship between the cellular density of a tissue region and its ADC values, as regions with dense cells restrict the mobility of water molecules. Since the quality of DWI is low, a large number of researchers choose to utilize the quantitative ADC maps computed from DWI to identify prostate cancer. The reason for the discriminating capabilities of ADCs is that malignant prostate tissues have smaller ADC values than healthy or benign tissues. The following equation shows that two DW images are required to calculate an ADC map:
A D C = l n S b 1 l n S 0 b 1 .
The first image is collected at a specific b-value ( b 1 > 0 s/mm2) while the second is collected at b 0 (b = 0 s/mm2). Another justification for employing ADC maps in this study is that the calculation of ADC maps is not sensitive to the used magnetic field strength [35]. This is suitable for the DWI datasets used in this study, as they were collected by two scanners that have different magnetic field strengths. Moreover, integrating handcrafted features with the automatically-learned features by CNNs can improve accuracy [36]. Since each ADC map represents the difference between two DW images, these maps are themselves images. Therefore, they can be used as input to the CNN model instead of the DW images. As these maps have better discriminating capabilities, their use improves the accuracy of the system.

2.2. Identification of Prostate Cancer

In this study, two different CNN models were used for prostate cancer identification. There are multiple advantages of using CNN over traditional neural networks. First, CNNs typically contain a larger number of layers than traditional neural networks. Augmenting the number of layers allows CNN to learn high levels of abstraction as the first layers learn primitive components while end layers use these learned primitive features to form the high-level features. The process of learning the features is done automatically by CNNs. Second, CNN takes both 2D images and 3D volumes directly as inputs without the need to convert these inputs into vectors, as in the case of neural networks. This preserves the inputs’ spatial information. Third, the network connections and hence the network parameters of CNNs are fewer than the network connections in similar traditional neural networks. This reduction simplifies and expedites the training process of CNNs [37,38].
In this work, DW slices that contained tumors were labeled by a radiologist. The ADC maps that correspond to these labeled DW slices were divided into two groups. The first group contained ADC maps with malignant tumors, and the second group contained ADC maps from benign subjects. The number of ADC maps in the benign group was 236, and the number of ADC maps in the malignant group was 234. These ADC maps were used to train and evaluate two different previously-trained CNN models, which are AlexNet [39] and VGGNet [40].
ALexNet expects an input image of size 227 × 227 × 3 , whereas the size of an input image for VGGNet was 224 × 224 × 3 . The sizes of the calculated ADC maps were the same as the sizes of the corresponding DW images, which were 256 × 256 for the 1.5T images and 144 × 144 for the 3T images. To make these ADC maps suitable as inputs for the employed CNN models, the ADC maps of the larger sizes were center cropped, while the ADC maps of smaller sizes were zero padded. Then, each of these ADC maps was concatenated along the depth dimension to generate a three-channel image, which was the expected input to each of the employed CNN models.
Since the number of ADC maps is considered small for training and evaluating a CNN model from scratch, as training such deep structures from scratch with a small dataset leads to overfitting, a transfer learning model was adopted in this study. The idea of transfer learning is to modify a network that is trained to solve a certain problem and use it to solve a new problem in a different domain. The training of the original network is typically done with millions of images from the original domain. The advantage of transfer learning is that the adoption of this previously-trained network to solve a new problem requires far fewer images from the new domain. This is done by replacing the last few layers, including the output layer of the original network, with new layers appropriate to the new problem. In this work, the original output layer of either AlexNet or VGGNet, which assigns its input image to one of 1000 categories, was replaced with an output layer that classified its input ADC maps into either benign or malignant (Figure 4).
The introduction and success of AlexNet have revolutionized the use of CNNs for multiple classification tasks. AlexNet is a CNN that contains five convolutional layers and three FC layers (Figure 4a). The network depth has a remarkable influence on its accuracy as the accuracy of AlexNet drops by 2 % in the case of removing any of the five convolution layers. Rectified linear units (ReLUs) [41] are employed as activation functions by AlexNet. There are two main advantages of using ReLUs. First, ReLUs are saturation-free even in case of unnormalized inputs. Second, the training time of CNNs with ReLUs is shorter than the training time of CNNs with saturating activations (e.g., sigmoid). AlexNet is trained with a large dataset of natural images (ImageNet). This dataset contains more than one million images that belong to one thousand categories. To reduce the classification error, AlexNet employs overlapping pooling and response normalization [39]. AlexNet employs dropout [42] and data augmentation to overcome the overfitting issue. Two different forms of data augmentation were used: intensity alteration and transformations (translations and horizontal reflections) [39].
VGGNet is another deep CNN that is trained using ImageNet dataset. The main goal of developing VGGNet is to evaluate the effect of the network depth on the accuracies of CNNs. To achieve this goal, the developers examined five different network architectures with different depths, while the other parameters are fixed for a fair comparison. The input image is processed by a sequence of convolution layers, max-pooling layers [43], FC layers, and a softmax layer. VGGNet uses ReLU non-linearity activation. The number of convolutional layers of the different architectures ranges from 8 to 16 layers. The numbers of pooling layers and FC layers are 5 and 3, respectively. These numbers do not differ across the different architectures. The pooling layers follow some, but not all, of the convolutional layers. The convolution layers use kernels of small fixed receptive fields of 3 × 3 to lower the number of parameters. The pooling layers use fixed windows of 2 × 2 and a stride of 2. The number of filters of the first convolutional layer is 64. When a pooling layer is used, the number of filters of the following convolutional layer is doubled until the number of kernels reaches 512 [40].
In this work, the deepest architecture was used. This architecture has 19 layers with weights (16 convolutional layers and 3 FC layers) (Figure 4b). This architecture has 144 million parameters. This large number of parameters increases the training time of VGGNet, especially when compared with other CNNs with smaller numbers of parameters, such as AlexNet (60 million parameters) and GoogLeNet (4 million parameters). However, the performance of VGGNet in many transfer learning tasks is better than GoogLeNet [44]. Both AlexNet and VGGNet were optimized using stochastic gradient descent with momentum and the loss function was cross entropy. The other training parameters were the following: number of epochs = 50, learning rate = 0.0001, momentum = 0.9, mini-batch-size = 10, L2 regularization = 0.0001. The basic architecture and configuration parameters of both the original and the fine-tuned AlexNet and VGGNet are summarized in Table 1.

3. Results

Multiple experiments were conducted to test the performance of the developed system and to compare its performance with other modern machine-learning classifiers. In the first experiment, 70 % of the ADC maps of both the malignant and benign cases at each b-value were used for fine-tuning an AlexNet-based model. The other 30%, which represent 71 ADC maps from benign cases and 70 ADC maps with malignant tumors, were employed to evaluate the accuracy of the tuned model. The ADC slices used for training were from different patients to the ADC slices used for testing to avoid any correlation that could exist between ADC slices of the same patient. The results of this experiment at each b-value are shown in Table 2.
In a similar experiment, 80 % of the ADC maps were used for tuning an AlexNet-based model. The remaining 20 % or 94 ADC maps were used to evaluate the accuracy of the tuned model. The reason behind this 80:20 division of the dataset was to satisfy the Pareto principle. The results of this experiment are reported in Table 3.
Similarly, in another experiment, 70 % of the ADC maps of both the benign and malignant cases at each b-value were randomly chosen for tuning an AlexNet-based model. The other 30% of the ADC maps were used to evaluate the accuracy of the tuned model. Since the ADC maps used for evaluating the accuracy of the system at each b-value were chosen randomly, this experiment was repeated 10 times at each b-value. To ensure the stability of the reported results, the mean accuracy of the 10 experiments conducted at each b-value, in addition to the mean sensitivity and the mean specificity, are listed in Table 4.
In another experiment, 10-fold cross validation was applied using the 470 ADC maps. Each fold contained 47 ADC maps. Nine folds were used for training whereas the remaining fold was used for testing the system. This operation was repeated 10 times with the change of the testing fold each time. The performance of the 10-fold cross validation using AlexNet at the different b-values is reported in Table 5.
In order to evaluate the effect of the depth of the used CNNs on the resulting accuracy, an experiment was conducted. In this experiment, 70 % of the ADC maps were used for tuning a deeper CNN (VGGNet), while the remaining 30 % were used to evaluate the trained model. VGGNet has 16 convolutional layers in comparison to five convolutional layers in AlexNet. The performance using VGGNet is listed in Table 6. As can be noticed, the use of the deeper network improves the accuracy at all b-values except for b-value = 500 s/mm2. The highest improvement in the accuracy is at b-value = 600 s/mm2. The use of VGGNet leads to an average improvement of 1.97 % in the accuracy.
The time required for training VGGNet model is much longer than the time required for training the AlexNet model. For example, the time required for fine-tuning AlexNet in the first experiment was 5 min and 25 s, whereas the time required for fine-tuning the deeper CNN (VGGNet) in the similar experiment was about 97 min. The proposed models were developed using MATLAB 2017b. The training of the proposed model was performed using a workstation with a graphics processing unit (GPU) of type NVIDIA Quadro K1200.
In a similar experiment, 80 % of the ADC maps were used for tuning a VGGNet-based model. The remaining 20 % or 94 ADC maps were used to evaluate the accuracy of the tuned model. The results of this experiment are reported in Table 7.
In another experiment, 10-fold cross validation is applied using VGGNet at the different b-values. The results of this experiment are reported in Table 8.
In order to show the merits of CNNs over classical machine-learning classifiers, the performance of CNN is compared to the performance of SVM with both linear and quadratic kernels. The models of SVMs were trained using 70 % of the ADC maps and evaluated using the remaining 30 % of the ADC maps. The inputs to the SVMs were the same as the inputs to the CNNs. The inputs to the SVMs were the ADC maps (raw data). In order to be used as input to the SVMs, each ADC map has to be transformed into a vector. This vector represents a row in the data matrix used to train the SVM. The results of these SVMs are reported in Table 9. The high performance of the CNN models highlights the importance of the inputs’ spatial information that is preserved in the case of CNNs. However, the inputs’ spatial information is lost in the case of conventional models such as SVMs.
The ROC curves of two CNN models (AlexNet and VGGNet) and two variants of SVMs with linear and quadratic kernels are shown in Figure 5. Since the ROC curves of each of these classifiers at the distinct b-values have similar shapes, the ROC curves at only five b-values are displayed to simplify the figures.
An experiment was conducted to compare the performance of the proposed approach to one of the state-of-the-art CNNs, which is GoogLeNet [45]. This CNN was the winner of the ImageNet challenge in 2014. This deep network consists of 22 layers. However, the number of its parameters is reduced dramatically due to the use of the Inception module and the removal of the FC layers. The resulting performance of GoogLeNet is listed in Table 10. The performance results of GoogLeNet are close to the results of both AlexNet and VGGNet. These results boosts the feasibility of the transfer learning in diagnosing prostate cancer.

4. Discussion

In this study, a transfer learning model is adopted to detect and identify prostate cancer. When the employed CNN models were originally trained using natural images, they used conventional techniques such as, data augmentation and dropout in order to reduce the effect of overfitting. The combination of both conventional overfitting handling techniques and transfer learning can minimize the effect of overfitting.
The proposed system starts with segmentation to limit the region-of-interest (ROI) to the prostate region only. In this system, prostate segmentation is performed using level set due to its capability to provide continuous segmented object. However, any segmentation approach can be integrated with the proposed system, as long as, it provides a continuous segmented object. For example, Comelli et al. [46], presented a fast deep learning network, namely efficient neural network (ENet), for prostate segmentation from T2-weighted MRI. ENet is initially used for image segmentation tasks in self-driving cars where hardware availability is limited and the accuracy is critical for user safety. In this study [46], ENet is trained using a dataset of 85 subjects and results in a dice similarity coefficient of 90.89 % .
Several studies suggested that the use of DWI acquired at higher b-values are preferable for accurate detection and diagnosis of prostate cancer [47,48,49,50,51]. This study shows that the use of ADC maps calculated at lower b-values results in an accuracy close to the accuracy of the ADC maps calculated at higher b-values. There is a slight accuracy increase for the ADC maps calculated at higher b-values. This accuracy increase is more obvious in the case of using a less-deeper CNN (AlexNet). One of the advantages of using ADC maps is that they are insensitive to the magnetic field strengths of the used scanners (1.5T or 3T) [35]. The ADC maps used in this study were calculated from DWI acquired with 1.5T and 3T scanners at nine b-values. The results show that the dependence on ADC maps can also mitigate the differences in the accuracy between higher and lower b-values, especially in the case of using deeper CNN models.
The developed approach performs slice-wise analysis. However, the proposed framework is generic and can perform both slice-wise analysis and prostate-zonal analysis based on how the model is trained. Since the system shows good performance in slice-wise analysis, the authors did not investigate it on zonal analysis. Investigating the performance of the system in prostate-zonal analysis is a good potential for future work.
According to the literature, a sensitivity and a positive predictive value of 80 % and 87 % , have been reported for men with high prostate-specific antigen (PSA) values using positron emission tomography/computed tomography (PET/CT) [52,53].
According to Sun et al. [54], the performance of deep learning networks increases logarithmically based on the size of the training data. One way to improve the performance in the case of limited data is to conduct multiple experiments and choose the best results. In this work, to obtain the best performance, 10-fold cross validation was used to obtain almost the same performance, no matter which fold was used for training and which fold was used for testing.
The use of two different CNN models in this work shows that the depth of the CNN model positively affects the performance of the system. However, much longer processing times are required to train the deeper architectures. The results of this study show that the use of a deeper CNN (VGGNet) increases the accuracy of prostate cancer detection more than the less-deep CNN (AlexNet). However, this accuracy is still far from perfect. Examining the effect of using much deeper CNN models, such as ResNet [55], can be a potential future work. Moreover, in this system, prostate cancer identification from DWI acquired at nine b-values was investigated. This investigation can be extended by performing a statistical analysis of the used b-values to select the best minimal combination of b-values that lead to the best accuracy. Choosing a minimal combination of b-values will reduce both the acquisition time of DWI and the computational complexities. Another area of potential future work could be the use of artificial intelligence optimization techniques on a combination of imaging markers and clinical markers (such as PSA) to optimize prostate cancer management.

5. Conclusions

In conclusion, this paper presents a CAD system for prostate cancer detection and identification from DWI. The identification of prostate cancer is achieved using two previously-trained CNN models (AlexNet and VGGNet) that are fed with the estimated ADC maps of the segmented prostate regions. The conducted experiments show that the use of previously-trained CNN models for detecting prostate cancer is feasible. These previously-trained CNN models learn the discriminatory features automatically. The results section shows that CNN models outperform conventional models. The accuracy of conventional models depends on the good design of the used handcrafted features.

Author Contributions

Conceptualization, I.R.A., A.S., A.M., M.E., A.A., M.A.E.-G., M.E.-M., N.S.A. and A.E.-B.; Formal analysis, I.R.A. and M.E.-M.; Investigation, I.R.A., M.E., A.A., M.A.E.-G., M.E.-M. and A.E.-B.; Methodology, I.R.A., A.S., A.M., M.E., A.A., M.A.E.-G., M.E.-M., N.S.A. and A.E.-B.; Project administration, N.S.A. and A.E.-B.; Software, I.R.A., N.S.A. and A.E.-B.; Supervision, M.E., N.S.A. and A.E.-B.; Validation, M.E.-M.; Visualization, I.R.A. and A.S.; Writing—original draft, I.R.A., A.S., A.M., M.E., A.A., M.A.E.-G., M.E.-M., N.S.A. and A.E.-B.; Writing—review & editing, I.R.A., A.S., A.M., M.E., N.S.A. and A.E.-B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Deanship of Scientific Research at Princess Nourah bint Abdulrahman University through the Fast-track Research Funding Program.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Mansoura University (R.21.04.1289).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. American Cancer Society. Key Statistics for Prostate Cancer. 2019. Available online: https://www.cancer.org/cancer/prostate-cancer/about/key-statistics.html (accessed on 10 February 2020).
  2. Hricak, H. MR imaging and MR spectroscopic imaging in the pre-treatment evaluation of prostate cancer. Br. J. Radiol. 2005, 78, S103–S111. [Google Scholar] [CrossRef]
  3. Brown, A.M.; Elbuluk, O.; Mertan, F.; Sankineni, S.; Margolis, D.J.; Wood, B.J.; Pinto, P.A.; Choyke, P.L.; Turkbey, B. Recent advances in image-guided targeted prostate biopsy. Abdom. Radiol. 2015, 40, 1788–1799. [Google Scholar] [CrossRef] [PubMed]
  4. Hoeks, C.M.; Barentsz, J.O.; Hambrock, T.; Yakar, D.; Somford, D.M.; Heijmink, S.W.; Scheenen, T.W.; Vos, P.C.; Huisman, H.; van Oort, I.M.; et al. Prostate cancer: Multiparametric MR imaging for detection, localization, and staging. Radiology 2011, 261, 46–66. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, L.; Tian, Z.; Zhang, Z.; Fei, B. Computer-aided detection of prostate cancer with MRI: Technology and applications. Acad. Radiol. 2016, 23, 1024–1046. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Fei, B. Computer-aided diagnosis of prostate cancer with MRI. Curr. Opin. Biomed. Eng. 2017, 3, 20–27. [Google Scholar] [CrossRef] [PubMed]
  7. Viswanath, S.E.; Chirra, P.V.; Yim, M.C.; Rofsky, N.M.; Purysko, A.S.; Rosen, M.A.; Bloch, B.N.; Madabhushi, A. Comparing radiomic classifiers and classifier ensembles for detection of peripheral zone prostate tumors on T2-weighted MRI: A multi-site study. BMC Med. Imaging 2019, 19, 22. [Google Scholar] [CrossRef] [PubMed]
  8. Campa, R.; Del Monte, M.; Barchetti, G.; Pecoraro, M.; Salvo, V.; Ceravolo, I.; Indino, E.L.; Ciardi, A.; Catalano, C.; Panebianco, V. Improvement of prostate cancer detection combining a computer-aided diagnostic system with TRUS-MRI targeted biopsy. Abdom. Radiol. 2019, 44, 264–271. [Google Scholar] [CrossRef] [PubMed]
  9. Rampun, A.; Zheng, L.; Malcolm, P.; Tiddeman, B.; Zwiggelaar, R. Computer-aided detection of prostate cancer in T2-weighted MRI within the peripheral zone. Phys. Med. Biol. 2016, 61, 4796. [Google Scholar] [CrossRef] [PubMed]
  10. Gao, J.; Jiang, Q.; Zhou, B.; Chen, D. Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview. Math. Biosci. Eng. 2019, 16, 6536. [Google Scholar] [CrossRef]
  11. Srinivas, S.; Sarvadevabhatla, R.K.; Mopuri, K.R.; Prabhu, N.; Kruthiventi, S.S.; Babu, R.V. A Taxonomy of Deep Convolutional Neural Nets for Computer Vision. Front. Robot AI 2016, 2, 36–48. [Google Scholar] [CrossRef]
  12. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  13. Lu, H.; Li, Y.; Uemura, T.; Kim, H.; Serikawa, S. Low illumination underwater light field images reconstruction using deep convolutional neural networks. Future Gener. Comput. Syst. 2018, 82, 142–148. [Google Scholar] [CrossRef]
  14. Li, Y.; Wang, G.; Nie, L.; Wang, Q.; Tan, W. Distance metric optimization driven convolutional neural network for age invariant face recognition. Pattern Recognit. 2018, 75, 51–62. [Google Scholar] [CrossRef]
  15. Babaee, M.; Dinh, D.T.; Rigoll, G. A deep convolutional neural network for video sequence background subtraction. Pattern Recognit. 2018, 76, 635–649. [Google Scholar] [CrossRef]
  16. Sabokrou, M.; Fayyaz, M.; Fathy, M.; Moayed, Z.; Klette, R. Deep-anomaly: Fully convolutional neural network for fast anomaly detection in crowded scenes. Comput. Vis. Image Underst. 2018, 172, 88–97. [Google Scholar] [CrossRef] [Green Version]
  17. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Ishioka, J.; Matsuoka, Y.; Uehara, S.; Yasuda, Y.; Kijima, T.; Yoshida, S.; Yokoyama, M.; Saito, K.; Kihara, K.; Numao, N.; et al. Computer-aided diagnosis of prostate cancer on magnetic resonance imaging using a convolutional neural network algorithm. BJU Int. 2018, 122, 411–417. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Mehrtash, A.; Sedghi, A.; Ghafoorian, M.; Taghipour, M.; Tempany, C.M.; Wells III, W.M.; Kapur, T.; Mousavi, P.; Abolmaesumi, P.; Fedorov, A. Classification of clinical significance of MRI prostate findings using 3D convolutional neural networks. Medical Imaging 2017: Computer-Aided Diagnosis. Int. Soc. Opt. Photonics 2017, 10134, 101342A. [Google Scholar]
  20. Yang, X.; Liu, C.; Wang, Z.; Yang, J.; Le Min, H.; Wang, L.; Cheng, K.T.T. Co-trained convolutional neural networks for automated detection of prostate cancer in multi-parametric MRI. Med. Image Anal. 2017, 42, 212–227. [Google Scholar] [CrossRef] [PubMed]
  21. Le, M.H.; Chen, J.; Wang, L.; Wang, Z.; Liu, W.; Cheng, K.T.T.; Yang, X. Automated diagnosis of prostate cancer in multi-parametric MRI based on multimodal convolutional neural networks. Phys. Med. Biol. 2017, 62, 6497–6514. [Google Scholar] [CrossRef]
  22. Wang, Z.; Liu, C.; Cheng, D.; Wang, L.; Yang, X.; Cheng, K.T. Automated Detection of Clinically Significant Prostate Cancer in mp-MRI Images Based on an End-to-End Deep Neural Network. IEEE Trans. Med. Imaging 2018, 37, 1127–1139. [Google Scholar] [CrossRef]
  23. Song, Y.; Zhang, Y.D.; Yan, X.; Liu, H.; Zhou, M.; Hu, B.; Yang, G. Computer-aided diagnosis of prostate cancer using a deep convolutional neural network from multiparametric MRI. J. Magn. Reson. Imaging 2018, 48, 1570–1577. [Google Scholar] [CrossRef] [PubMed]
  24. Hosseinzadeh, M.; Brand, P.; Huisman, H. Effect of Adding Probabilistic Zonal Prior in Deep Learning-based Prostate Cancer Detection. arXiv 2019, arXiv:1907.12382. [Google Scholar]
  25. Schelb, P.; Kohl, S.; Radtke, J.P.; Wiesenfarth, M.; Kickingereder, P.; Bickelhaupt, S.; Kuder, T.A.; Stenzinger, A.; Hohenfellner, M.; Schlemmer, H.P.; et al. Classification of cancer at prostate MRI: Deep learning versus clinical PI-RADS assessment. Radiology 2019, 293, 607–617. [Google Scholar] [CrossRef]
  26. Xu, H.; Baxter, J.S.; Akin, O.; Cantor-Rivera, D. Prostate cancer detection using residual networks. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1647–1650. [Google Scholar] [CrossRef] [PubMed]
  27. Yuan, Y.; Qin, W.; Buyyounouski, M.; Ibragimov, B.; Hancock, S.; Han, B.; Xing, L. Prostate cancer classification with multiparametric MRI transfer learning model. Med. Phys. 2019, 46, 756–765. [Google Scholar] [CrossRef] [PubMed]
  28. Chen, Q.; Xu, X.; Hu, S.; Li, X.; Zou, Q.; Li, Y. A transfer learning approach for classification of clinical significant prostate cancers from mpMRI scans. Medical Imaging 2017: Computer-Aided Diagnosis. Int. Soc. Opt. Photonics 2017, 10134, 101344F. [Google Scholar]
  29. Abbasi, A.A.; Hussain, L.; Awan, I.A.; Abbasi, I.; Majid, A.; Nadeem, M.S.A.; Chaudhary, Q.A. Detecting prostate cancer using deep learning convolution neural network with transfer learning approach. Cogn. Neurodyn. 2020, 14, 523–533. [Google Scholar] [CrossRef]
  30. Wildeboer, R.R.; van Sloun, R.J.; Wijkstra, H.; Mischi, M. Artificial intelligence in multiparametric prostate cancer imaging with focus on deep-learning methods. Comput. Methods Programs Biomed. 2020, 189, 105316. [Google Scholar] [CrossRef]
  31. McClure, P.; Khalifa, F.; Soliman, A.; El-Ghar, M.A.; Gimelfarb, G.; Elmagraby, A.; El-Baz, A. A novel NMF guided level-set for DWI prostate segmentation. J. Comput. Sci. Syst. Biol. 2014, 7, 209–216. [Google Scholar] [CrossRef] [Green Version]
  32. Kwee, T.C.; Galbán, C.J.; Tsien, C.; Junck, L.; Sundgren, P.C.; Ivancevic, M.K.; Johnson, T.D.; Meyer, C.R.; Rehemtulla, A.; Ross, B.D.; et al. Comparison of apparent diffusion coefficients and distributed diffusion coefficients in high-grade gliomas. J. Magn. Reson. Imaging 2010, 31, 531–537. [Google Scholar] [CrossRef] [Green Version]
  33. Huisman, T.A.G.M. Diffusion-weighted imaging: Basic concepts and application in cerebral stroke and head trauma. Eur. Radiol. 2003, 13, 2283–2297. [Google Scholar] [CrossRef] [PubMed]
  34. Hrabe, J.; Kaur, G.; Guilfoyle, D.N. Principles and limitations of NMR diffusion measurements. J. Med. Phys. 2007, 32, 34–42. [Google Scholar] [CrossRef] [PubMed]
  35. Jeong, D.; Malalis, C.; Arrington, J.A.; Field, A.S.; Choi, J.W.; Kocak, M. Mean apparent diffusion coefficient values in defining radiotherapy planning target volumes in glioblastoma. Quant. Imaging Med. Surg. 2015, 5, 835–845. [Google Scholar] [PubMed]
  36. Choi, Y.J.; Kim, J.K.; Kim, N.; Kim, K.W.; Choi, E.K.; Cho, K.S. Functional MR imaging of prostate cancer. Radiographics 2007, 27, 63–75. [Google Scholar] [CrossRef]
  37. LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional networks and applications in vision. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems, Paris, France, 30 May–2 June 2010; pp. 253–256. [Google Scholar] [CrossRef] [Green Version]
  38. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
  39. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural. Inf. Process Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  40. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  41. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the International Conference Mach Learning, Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  42. Hinton, G.E.O. Improving neural networks by preventing co-adaptation of feature detectors. arXiv 2012, arXiv:1207.0580. [Google Scholar]
  43. Boureau, Y.L.; Ponce, J.; LeCun, Y. A theoretical analysis of feature pooling in visual recognition. In Proceedings of the International Conference Mach Learning, Haifa, Israel, 21–24 June 2010; pp. 111–118. [Google Scholar]
  44. Xu, B.; Fu, Y.; Jiang, Y.G.; Li, B.; Sigal, L. Heterogeneous Knowledge Transfer in Video Emotion Recognition, Attribution and Summarization. IEEE Trans. Affect. Comput. 2018, 9, 255–270. [Google Scholar] [CrossRef] [Green Version]
  45. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  46. Comelli, A.; Dahiya, N.; Stefano, A.; Vernuccio, F.; Portoghese, M.; Cutaia, G.; Bruno, A.; Salvaggio, G.; Yezzi, A. Deep Learning-Based Methods for Prostate Segmentation in Magnetic Resonance Imaging. Appl. Sci. 2021, 11, 782. [Google Scholar] [CrossRef] [PubMed]
  47. Metens, T.; Miranda, D.; Absil, J.; Matos, C. What is the optimal b value in diffusion-weighted MR imaging to depict prostate cancer at 3T? Eur. Radiol. 2012, 22, 703–709. [Google Scholar] [CrossRef]
  48. Katahira, K.; Takahara, T.; Kwee, T.C.; Oda, S.; Suzuki, Y.; Morishita, S.; Kitani, K.; Hamada, Y.; Kitaoka, M.; Yamashita, Y. Ultra-high-b-value diffusion-weighted MR imaging for the detection of prostate cancer: Evaluation in 201 cases with histopathological correlation. Eur. Radiol. 2011, 21, 188–196. [Google Scholar] [CrossRef] [PubMed]
  49. Grant, K.B.; Agarwal, H.K.; Shih, J.H.; Bernardo, M.; Pang, Y.; Daar, D.; Merino, M.J.; Wood, B.J.; Pinto, P.A.; Choyke, P.L.; et al. Comparison of calculated and acquired high b value diffusion-weighted imaging in prostate cancer. Abdom. Radiol. 2015, 40, 578–586. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Pang, Y.; Turkbey, B.; Bernardo, M.; Kruecker, J.; Kadoury, S.; Merino, M.J.; Wood, B.J.; Pinto, P.A.; Choyke, P.L. Intravoxel incoherent motion MR imaging for prostate cancer: An evaluation of perfusion fraction and diffusion coefficient derived from different b-value combinations. Magn. Reson. Med. 2013, 69, 553–562. [Google Scholar] [CrossRef] [Green Version]
  51. Rosenkrantz, A.B.; Mannelli, L.; Kong, X.; Niver, B.E.; Berkman, D.S.; Babb, J.S.; Melamed, J.; Taneja, S.S. Prostate cancer: Utility of fusion of T2-weighted and high b-value diffusion-weighted images for peripheral zone tumor detection and localization. J. Magn. Reson. Imaging 2011, 34, 95–100. [Google Scholar] [CrossRef] [PubMed]
  52. Wallitt, K.L.; Khan, S.R.; Dubash, S.; Tam, H.H.; Khan, S.; Barwick, T.D. Clinical PET imaging in prostate cancer. Radiographics 2017, 37, 1512–1536. [Google Scholar] [CrossRef] [Green Version]
  53. Tateishi, U.; Morita, S.; Taguri, M.; Shizukuishi, K.; Minamimoto, R.; Kawaguchi, M.; Murano, T.; Terauchi, T.; Inoue, T.; Kim, E.E. A meta-analysis of 18 F-Fluoride positron emission tomography for assessment of metastatic bone tumor. Ann. Nucl. Med. 2010, 24, 523–531. [Google Scholar] [CrossRef] [PubMed]
  54. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 843–852. [Google Scholar]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Figure 1. Overall workflow of the proposed model showing the DWI input data at nine b-values and its three basic steps, which are prostate segmentation, calculation of ADC maps as discriminating features, and the identification of slices with tumor using previously-trained CNN models.
Figure 1. Overall workflow of the proposed model showing the DWI input data at nine b-values and its three basic steps, which are prostate segmentation, calculation of ADC maps as discriminating features, and the identification of slices with tumor using previously-trained CNN models.
Sensors 21 03664 g001
Figure 2. Illustration of the first two processing steps of the proposed framework on two different cases (one benign and one malignant).
Figure 2. Illustration of the first two processing steps of the proposed framework on two different cases (one benign and one malignant).
Sensors 21 03664 g002
Figure 3. Illustration of slice-wise analysis of ADC volumes.
Figure 3. Illustration of slice-wise analysis of ADC volumes.
Sensors 21 03664 g003
Figure 4. Illustration of the two different CNNs used in this study, (a) ALexNet, and (b) VGGNet.
Figure 4. Illustration of the two different CNNs used in this study, (a) ALexNet, and (b) VGGNet.
Sensors 21 03664 g004
Figure 5. ROC curves at five b-values(100, 300, 500, 700, and 1000 s/mm2) of four different classifiers: (a) VGGNet, (b) AlexNet, (c) SVM with linear kernel, and (d) SVM with quadratic kernel. As these figures show, CNN-based models (VGGNet and AlexNet) result in higher AUCs than the two variants of SVMs with linear and quadratic kernels at the distinct b-values.
Figure 5. ROC curves at five b-values(100, 300, 500, 700, and 1000 s/mm2) of four different classifiers: (a) VGGNet, (b) AlexNet, (c) SVM with linear kernel, and (d) SVM with quadratic kernel. As these figures show, CNN-based models (VGGNet and AlexNet) result in higher AUCs than the two variants of SVMs with linear and quadratic kernels at the distinct b-values.
Sensors 21 03664 g005
Table 1. Basic architecture and configuration parameters of both AlexNet and VGGNet, where Conv. means convolutional, FC means fully-connected, SGDM means stochastic gradient descent with momentum, and cross entr. means cross entropy.
Table 1. Basic architecture and configuration parameters of both AlexNet and VGGNet, where Conv. means convolutional, FC means fully-connected, SGDM means stochastic gradient descent with momentum, and cross entr. means cross entropy.
 AlexNetVGGNet
 OriginalFine-TunedOriginalFine-Tuned
No. of training images>1 million329–423>1 million329–423
Size of input images 227 × 227 227 × 227 224 × 224 224 × 224
No. of output categories1000210002
No. of Conv. layers551616
FC layers3333
OptimizerSGDMSGDMSGDMSGDM
Loss functioncross entr.cross entr.cross entr.cross entr.
Table 2. Performance of AlexNet at 9 b-values using 141 ADC maps.
Table 2. Performance of AlexNet at 9 b-values using 141 ADC maps.
b-ValueAccuracy%Sensitivity (Recall)%Specificity%Precision%
100 s/mm286.5284.2988.7388.06
200 s/mm290.0785.7194.3793.75
300 s/mm288.6585.7191.5590.91
400 s/mm288.6588.5788.7391.18
500 s/mm291.4991.4391.5591.43
600 s/mm289.3688.5790.1489.86
700 s/mm287.9485.7190.1489.55
800 s/mm290.0790.0090.1490.00
1000 s/mm290.0787.1492.9692.06
Table 3. Performance of AlexNet at 9 b-values using 94 randomly chosen ADC maps.
Table 3. Performance of AlexNet at 9 b-values using 94 randomly chosen ADC maps.
b-ValueAccuracy%Sensitivity (Recall)%Specificity%Precision%
100 s/mm287.2387.2387.2387.23
200 s/mm290.4389.3691.4991.30
300 s/mm289.3689.3689.3689.36
400 s/mm289.3691.4987.2387.76
500 s/mm290.4391.4989.3689.58
600 s/mm291.4989.3693.6293.33
700 s/mm288.3389.3687.2387.50
800 s/mm291.4993.6289.3689.80
1000 s/mm289.3693.6285.1186.27
Table 4. Average Performance of AlexNet at 9 b-values using 141 randomly chosen ADC maps and repeating the experiment 10 times.
Table 4. Average Performance of AlexNet at 9 b-values using 141 randomly chosen ADC maps and repeating the experiment 10 times.
b-ValueAccuracy%Sensitivity%Specificity%
100 s/mm2 87.52 ± 0.01 89.14 ± 0.04 85.92 ± 0.04
200 s/mm2 89.29 ± 0.01 88.71 ± 0.04 89.86 ± 0.03
300 s/mm2 89.57 ± 0.02 89.00 ± 0.03 90.14 ± 0.03
400 s/mm2 89.57 ± 0.02 89.71 ± 0.03 89.44 ± 0.04
500 s/mm2 89.22 ± 0.03 88.29 ± 0.04 90.14 ± 0.04
600 s/mm2 88.01 ± 0.03 88.00 ± 0.07 88.03 ± 0.05
700 s/mm2 89.72 ± 0.02 89.29 ± 0.04 90.14 ± 0.03
800 s/mm2 89.63 ± 0.03 88.57 ± 0.05 90.70 ± 0.03
1000 s/mm2 90.43 ± 0.02 90.86 ± 0.04 90.00 ± 0.04
Table 5. 10-fold cross validation of AlexNet at 9 b-values.
Table 5. 10-fold cross validation of AlexNet at 9 b-values.
b-Value100 200 300 400 500 600 700 800 1000 
1st fold93.6287.2393.6293.6291.4989.3691.4991.4987.23
2nd fold78.7282.9878.7282.9887.2382.9885.1180.8578.72
3rd fold91.4995.7491.4989.3693.6293.6295.7493.6291.49
4th fold95.7497.8797.8793.6293.6295.7497.8797.8791.49
5th fold87.2389.3691.4991.4989.3693.6287.2391.4991.49
6th fold85.1180.8582.9882.9887.2380.8585.1187.2389.36
7th fold91.4993.6289.3697.8795.7493.6293.6295.7493.62
8th fold89.3687.2387.2389.3689.3689.3687.2389.3689.36
9th fold93.6291.4991.4993.6291.4993.6289.3691.4991.49
10th fold82.9885.1185.1187.2387.2385.1187.2387.2385.11
Average%88.9489.1588.9490.2190.6489.7990.0090.6488.94
Table 6. Performance of VGGNet at 9 b-values using 141 randomly chosen ADC maps.
Table 6. Performance of VGGNet at 9 b-values using 141 randomly chosen ADC maps.
b-ValueAccuracy%Sensitivity (Recall)%Specificity%Precision%
100 s/mm290.0788.5791.5591.18
200 s/mm292.2092.8691.5591.55
300 s/mm290.0790.0090.1490.00
400 s/mm290.0790.0090.1490.00
500 s/mm290.7891.4390.1490.14
600 s/mm293.6292.8694.3794.20
700 s/mm290.0792.8687.3297.84
800 s/mm292.2092.8691.5591.55
1000 s/mm291.4992.8690.1490.28
Table 7. Performance of VGGNet at 9 b-values using 94 randomly chosen ADC maps.
Table 7. Performance of VGGNet at 9 b-values using 94 randomly chosen ADC maps.
b-ValueAccuracy%Sensitivity (Recall)%Specificity%Precision%
100 s/mm291.4987.2395.7495.35
200 s/mm290.4389.3691.4991.30
300 s/mm291.4991.4991.4991.49
400 s/mm290.4387.2393.6293.18
500 s/mm292.5589.3695.7495.45
600 s/mm294.6891.4997.8797.73
700 s/mm290.4387.2393.6293.18
800 s/mm292.5589.3695.7495.45
1000 s/mm292.5593.6291.4991.67
Table 8. 10-fold cross validation of VGGNet at 9 b-values.
Table 8. 10-fold cross validation of VGGNet at 9 b-values.
b-Value100 200 300 400 500 600 700 800 1000 
1st fold85.1191.4989.3691.4995.7495.7495.7495.7495.74
2nd fold82.9889.3689.3687.2389.3689.3689.3691.4987.23
3rd fold89.3691.4985.1193.6291.4991.4995.7497.8791.49
4th fold93.6291.4989.3693.6297.8797.8797.8793.6297.87
5th fold89.3693.6293.6289.3689.3691.4989.3691.4987.23
6th fold89.3680.8587.2387.2382.9885.1189.3689.3682.98
7th fold80.8591.4991.4995.7497.8797.8793.6295.7495.74
8th fold87.2397.8793.6291.4997.8787.2389.3693.6285.11
9th fold93.6295.7497.8797.8797.8795.7493.6295.7493.62
10th fold78.7280.8589.3682.9885.1185.1180.8585.1187.23
Average%87.0290.4390.6491.0692.5591.7091.4992.9890.43
Table 9. Performance of SVMs with linear kernel (lin. ker.) and quadratic kernel (quad. ker.) at 9 b-values using 141 randomly chosen ADC maps, where Acc. means accuracy, Sen. means sensitivity, and Spec. means specificity.
Table 9. Performance of SVMs with linear kernel (lin. ker.) and quadratic kernel (quad. ker.) at 9 b-values using 141 randomly chosen ADC maps, where Acc. means accuracy, Sen. means sensitivity, and Spec. means specificity.
 SVM with Lin. Ker.SVM with Quad. Ker.
b-ValueAcc. %Sen. %Spec. %Acc. %Sen. %Spec. %
100 s/mm275.1865.7184.5178.0171.4384.51
200 s/mm275.8972.8678.8778.7278.5778.87
300 s/mm273.0558.5787.3282.2777.1487.32
400 s/mm268.7975.7161.9780.8588.5773.24
500 s/mm279.4382.8676.0682.2781.4383.10
600 s/mm273.7668.5778.8780.1477.1483.10
700 s/mm268.7971.4366.2080.1480.0080.28
800 s/mm271.6374.2969.0181.5687.1476.06
1000 s/mm273.7670.0077.4681.5682.8680.28
Table 10. Performance of GoogLeNet at 9 b-values using 141 ADC maps.
Table 10. Performance of GoogLeNet at 9 b-values using 141 ADC maps.
b-ValueAccuracy%Sensitivity%Specificity%
100 s/mm285.8285.7185.92
200 s/mm290.7885.7195.77
300 s/mm287.2385.7188.73
400 s/mm287.2382.8691.55
500 s/mm287.9490.0085.92
600 s/mm288.6588.5788.73
700 s/mm290.0791.4388.73
800 s/mm288.6585.7191.55
1000 s/mm289.3690.0088.73
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Abdelmaksoud, I.R.; Shalaby, A.; Mahmoud, A.; Elmogy, M.; Aboelfetouh, A.; Abou El-Ghar, M.; El-Melegy, M.; Alghamdi, N.S.; El-Baz, A. Precise Identification of Prostate Cancer from DWI Using Transfer Learning. Sensors 2021, 21, 3664. https://doi.org/10.3390/s21113664

AMA Style

Abdelmaksoud IR, Shalaby A, Mahmoud A, Elmogy M, Aboelfetouh A, Abou El-Ghar M, El-Melegy M, Alghamdi NS, El-Baz A. Precise Identification of Prostate Cancer from DWI Using Transfer Learning. Sensors. 2021; 21(11):3664. https://doi.org/10.3390/s21113664

Chicago/Turabian Style

Abdelmaksoud, Islam R., Ahmed Shalaby, Ali Mahmoud, Mohammed Elmogy, Ahmed Aboelfetouh, Mohamed Abou El-Ghar, Moumen El-Melegy, Norah Saleh Alghamdi, and Ayman El-Baz. 2021. "Precise Identification of Prostate Cancer from DWI Using Transfer Learning" Sensors 21, no. 11: 3664. https://doi.org/10.3390/s21113664

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop