Next Article in Journal
Maintenance Process Simulation Based Maintainability Evaluation by Using Stochastic Colored Petri Net
Next Article in Special Issue
Quantitative CT Analysis for Predicting the Behavior of Part-Solid Nodules with Solid Components Less than 6 mm: Size, Density and Shape Descriptors
Previous Article in Journal
In Vitro Targeting and Imaging of Neurogenic Differentiation in Mouse Bone-Marrow Derived Mesenchymal Stem Cells with Superparamagnetic Iron Oxide Nanoparticles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Heterogeneous 3D CNN for False-Positive Reduction in Pulmonary Nodule Detection, Based on Chest CT Images

1
School of Electronics and Information Engineering, Tianjin Polytechnic University, Tianjin 300387, China
2
Tianjin Key Laboratory of Optoelectronic Detection Technology and Systems, Tianjin Polytechnic University, Tianjin 300387, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(16), 3261; https://doi.org/10.3390/app9163261
Submission received: 17 June 2019 / Revised: 27 July 2019 / Accepted: 7 August 2019 / Published: 9 August 2019
(This article belongs to the Special Issue Computer-aided Biomedical Imaging 2020: Advances and Prospects)

Abstract

:
Currently, lung cancer has one of the highest mortality rates because it is often caught too late. Therefore, early detection is essential to reduce the risk of death. Pulmonary nodules are considered key indicators of primary lung cancer. Developing an efficient and accurate computer-aided diagnosis system for pulmonary nodule detection is an important goal. Typically, a computer-aided diagnosis system for pulmonary nodule detection consists of two parts: candidate nodule extraction and false-positive reduction of candidate nodules. The reduction of false positives (FPs) of candidate nodules remains an important challenge due to morphological characteristics of nodule height changes and similar characteristics to other organs. In this study, we propose a novel multi-scale heterogeneous three-dimensional (3D) convolutional neural network (MSH-CNN) based on chest computed tomography (CT) images. There are three main strategies of the design: (1) using multi-scale 3D nodule blocks with different levels of contextual information as inputs; (2) using two different branches of 3D CNN to extract the expression features; (3) using a set of weights which are determined by back propagation to fuse the expression features produced by step 2. In order to test the performance of the algorithm, we trained and tested on the Lung Nodule Analysis 2016 (LUNA16) dataset, achieving an average competitive performance metric (CPM) score of 0.874 and a sensitivity of 91.7% at two FPs/scan. Moreover, our framework is universal and can be easily extended to other candidate false-positive reduction tasks in 3D object detection, as well as 3D object classification.

1. Introduction

Lung cancer is one of the deadliest diseases with one of the highest mortality rates in the world [1]. Early asymptomatic manifestations of lung cancer may lead patients to miss the optimal treatment time; thus, early detection and treatments are essential to reduce the risk of death. Pulmonary nodules are considered key indicators of primary lung cancer [2]. The National Lung Screening Trial (NLST) [3] test shows that using low-dose chest computed tomography (CT) to examine pulmonary nodules in high-risk populations can reduce lung cancer mortality by 20%. Therefore, the detection of pulmonary nodules is very important for the early detection of lung cancer. Pulmonary nodules have many clinical manifestations, but one of the most prevalent is when nodules are irregular circles with a diameter of 3 mm to 30 mm [4]. According to pathological knowledge, a computer-aided diagnosis system for pulmonary nodule detection can be developed to reduce the workload of doctors, reduce misdiagnosis rate, and improve the efficiency of diagnosis.
Typically, a computer-aided diagnosis system (CAD) for pulmonary nodule detection consists of two parts: (1) candidate nodule detection, and (2) false-positive (FP) reduction of candidate nodules. In the candidate nodule detection step, the system detects as many candidates as possible to ensure high sensitivity without considering specificity. This is accompanied by a large number of false-positive nodules detected [5]. To reduce false positives, the system classifies the real nodules and non-nodules in the set of candidates detected in the previous step to reduce the number of false positives. False-positive reduction of candidate nodules is the most important part of a computer-aided diagnosis system for pulmonary nodule detection [6]. As far as we know, the current computer-aided diagnosis system (CAD) for pulmonary nodule detection is still in the development stage. Although the CAD system can improve the reading efficiency of radiologists, it is still not well used in clinical applications. Therefore, it is important to develop an advanced algorithm with high performance.
Pathologically, there are many types of nodules (e.g., ground glass nodules, solitary nodules, etc.) [7]. Various nodules have complex morphologies and multiple structures, and they often adhere to the trachea, blood vessels, and other organs. The morphological features of the candidate nodules are often similar to organs in the thoracic cavity (e.g., lymph nodes, blood vessels, etc.). This greatly increases the difficulty of reducing false positives in candidate nodules. Figure 1 shows an example of different types of nodules. This is why many researchers devoted their time to developing stable and effective algorithms to distinguish between nodules and non-nodules, thereby reducing false positives [8,9,10].
Recently, with the rapid development of deep learning techniques for medical image analysis [11,12,13,14,15], more and more researchers successfully applied convolutional neural networks (CNNs) to the detection of pulmonary nodules based on chest CT images [5,8,10,16,17,18], and achieved desirable results. By studying these methods, we can list three findings.
Firstly, the system that uses volumetric information has fewer false positives than methods that use two-dimensional (2D) slice information alone [5,16,17]. For example, Setio et al. [5] proposed a 2D multi-view CNN, which classified nodules based on slices from nine different views. Although the method utilized multiple views of the pulmonary nodules, to some extent, the three-dimensional (3D) information of the nodules was still not fully utilized. Therefore, Ding et al. [17] proposed to use a 3D nodule block as input for 3D-CNN, which was effective in reducing false positives.
Secondly, using multi-scale information of nodules can improve the performance of the algorithm [8,10,18]. For example, Kim et al. [8] proposed a multi-scale gradual integration CNN, which fused three scales of nodules gradually according to two different rules. This method used the 3D nodule information as a channel and used a 2D convolution kernel.
Thirdly, network parallel connections can improve false-positive performance. Dou et al. [10] proposed a multi-scale parallel 3D-CNN, which extracted blocks of three different scales for each candidate nodule. The classification calculations were made for each nodule block separately; then, the three classification results were combined according to a set of weights. The weights were determined manually, not trained from the dataset.
In this paper, we propose a multi-scale heterogeneous 3D CNN (MSH-CNN) for false-positive reduction of pulmonary nodules, which differs from other methods mainly in (1) using the 3D method. By making full use of 3D spatial information, our network can learn representative features with higher discernment than that of 2D CNN. (2) In order to cope with the large variations of nodules and more clearly distinguish them from other similar organizations, we use multi-scale blocks of the nodule as input. According to the characteristics of the nodule and organs surrounding the nodule, a novel heterogeneous network was designed to extract the expressions features. (3) We use a set of weights which are determined by back propagation to fuse the expression features. The experimental performance on the Lung Nodule Analysis 2016 (LUNA16) dataset demonstrates the effectiveness of our method and shows a brighter prospect on large datasets.
Our main contributions are as follows:
(1) We propose a novel 3D CNN for false-positive reduction of pulmonary nodules.
(2) We evaluate the effect of multi-scale heterogeneous 3D CNN with different structures on the reduction of false-positive nodules.
(3) Our method is tested on the Lung Nodule Analysis 2016 (LUNA16) [19] dataset, and the measure of performance surpasses most documented methods.
The rest of this paper is organized as follows: Section 2 introduces the existing methods in the literature. Section 3 describes our method. We introduce the algorithm performance comparison method and evaluate the impact of different structures on the results in Section 4. In Section 5, we discuss the reasons for the effectiveness of our method and the future directions for improvement. The conclusions are drawn in Section 6.

2. Related Work

In early tumor imaging research, researchers mainly focused on using traditional computer vision methods to artificially design expression features according to the appearance of nodules. For example, Messay et al. [20] designed an algorithm for segmentally identifying nodules based on the shape, position, brightness, and gradient characteristics of candidate nodules. Jacobs et al. [21] extracted 128 features based on shape, brightness, texture, and neighboring organ information of the nodule. The method used six classifiers to test and compared the results: support vector machine using radial basis function kernel (SVM-RBF), k-nearest neighbor, GentleBoost, random forest, recent mean, and linear classifiers. The sensitivity of the GentleBoost classifier reached 80% at 1.0 FP/scan. Although this method achieved good results, only pulmonary nodules with certain characteristics (e.g., shape, size, and texture) were found. However, these surface features lacked discrimination because the shape, size, and texture of pulmonary nodules vary widely, making it difficult to distinguish between real nodules and false-positive nodules.
Recently, deep neural networks became more common than manual design features. By studying deep neural network applications to reduce false-positive pulmonary nodules, we can find characteristic patterns. The use of 3D volumetric information, multi-scale nodule block input, and multi-branch network parallelization became more and more popular, and achieved better and better results.
In using 3D volumetric information, researchers gradually shifted away from using multi-view slice information of nodules to volumetric information of nodules. For example, Roth et al. [16] proposed the concept called 2.5D, which extracted three orthogonal slices for each candidate nodule, and used convolutional neural networks to extract expression features for classification. Setio et al. [5] proposed a 2D multi-view convolutional neural network that extracted nine slices of different views for each nodule, inputting them into convolutional neural networks to extract expression features, and finally merging these features by certain methods. In the LIDC-IDRI [22] dataset test, this method achieved good sensitivity, but still had some defects when the false-positive rate was low. Therefore, Ding et al. [17] proposed to extract 3D nodule blocks as the input of 3D-CNN according to the location of candidate nodules.
In the multi-scale nodule block inputting process, Shen et al. [18] proposed a multi-scale CNN, i.e., MCNN, to extract blocks (of different scales) according to candidate nodule coordinates. All the blocks were merged and a convolutional neural network was used to extract expression features for classification. Kim et al. [8] proposed multi-scale gradual integration CNN, using 2D convolution kernels to gradually fuse nodule blocks of different scales, and two branches were input to extract expression features. Finally, the nodules were classified by fusing these expression features. Multi-scale blocks have several advantages over single blocks; the advantages are as follows: (1) multi-scale blocks can completely segment nodules with large diameters; (2) multi-scale blocks can ensure the smaller-diameter nodules maintain a larger proportion of the block; (3) multi-scale blocks can increase the neighboring organ information of the nodule and bring abundant volumetric information while ensuring the integrity of nodule information.
In multi-branch network parallelization, Dou et al. [10] extracted nodule blocks of three scales and input them into three parallel branches for classification. Finally, the classification results were fused by the weights determined manually. When Kim et al. [8] fused the two parallel branches, a simple addition method was used to fuse them. Compared to a single branch network, multi-branch networks can extract more abundant complementary expression features. These complementary expression features can significantly improve the performance of the network after fusion.

3. Method

Our proposed multi-scale heterogeneous 3D CNN framework for false-positive reduction in pulmonary nodule detection shown in Figure 2d consists of three main parts: 3D multi-scale gradual integration, heterogeneous feature extraction, and automatic learning weight fusion. Inspired by Reference [8], for each candidate nodule, we used the coordinates of the center of the nodule to extract nodule blocks of three different scales. The three scales were 40 × 40 × 40 , 30 × 30 × 30 , and 20 × 20 × 20 , and we resized the three scales to transform them into 20 × 20 × 20 blocks, which are represented as I1, I2, and I3 (Figure 2a).

3.1. 3D Multi-Scale Gradual Integration

When humans observe objects, they can retrieve meaningful information by changing the size of the field of view. Inspired by this concept, multi-scale nodule information is effective for false-positive reduction. We propose a 3D multi-scale gradual integration module in Figure 2b that uses 3D convolution kernels to gradually integrate nodule blocks of different scales. There are two fusion sequences, i.e., I1-I2-I3 (“Input1”) and I3-I2-I1 (“Input2”).
For the I1-I2-I3 (“Input1”) fusion order, the nodule block I1 firstly obtains the feature map F1 via two 3D convolution operations; then, the feature maps F1 and nodule blocks I2 are fused according to channel (shown by the symbol “||”) to obtain F1||I2. After each 3D convolution, the non-linear function ReLU is used as the activation function. The output of ReLU is computed via a batch norm. To ensure that the feature map F1 and the nodule block I2 can be fused according to the channel, we use zero padding to ensure F1 and I2 are equivalent sizes. After combining the channels, F1||I2 also uses two 3D convolution operations to obtain the feature map F12; then, F12 and I3 are fused according to channel (shown by the symbol “||”) to obtain F12||I3, i.e., Input 1. For the fusion order of I3-I2-I1, the same calculation is used, except that the input order of the nodule blocks is reversed.
Therefore, nodule block I1 is resized and contains the most information about neighboring organs of the nodule, and I3 contains the least information about neighboring organs. For the I1-I2-I3 case, the input nodules contain less and less information of neighboring organs, mainly to enable the network to gradually focus on the morphological characteristics of the nodules themselves. In contrast, for the I3-I2-I1 case, the input nodule blocks contain more and more information about the neighboring organs. Their main purpose is to let the network gradually pay more and more attention to the morphological characteristics of the neighboring nodule organs. For every two sequences, the network focuses on the morphological features of the nodule and its neighboring organs. Using a 3D convolution kernel, the system is able to more fully extract volume information of the nodules and their neighboring organs. It is worth noting that the multi-scale inputs are computed on the fly during training/testing.

3.2. Heterogeneous Feature Extraction

Considering the important influence of the receptive field and network structure on the classification performance, we input I1-I2-I3 (“Input 1”) and I3-I2-I1 (“Input 2”) into two branches of different structures (as shown in Figure 2d). In theory, the receptive field of the convolutional neural network changes with the network structure. If the receptive field is too small, the network is limited by local information and lacks the ability to discriminate between targets with great variation. If the receptive field is too large, too much redundant information and noise interfere with the training process and reduce the performance of the model. Therefore, different convolutional neural network structures are used for different scale inputs.
I1-I2-I3 (“Input1”) and I3-I2-I1 (“Input2”) gradually integrate the nodule blocks in different order and they mainly focus on different aspects (features). Therefore, by using two branches with different structures, the characteristics of “Input1” and “Input2” can be fully utilized to extract complementary expression features.
The branch structure of input “Input 1” is mainly composed of 3D convolution layers and 3D max-pooling layers. After each 3D convolution, ReLU is used as the activation function and batch norm is used for the output of ReLU.
The branch structure of “Input 2” is mainly composed of 3D convolution layers and 3D inception layers. Similar to the previous branches, after each 3D convolution and 3D inception, ReLU is used as the activation function and batch norm is used for the output of ReLU. A detailed structure of the 3D inception step is shown in Figure 2c.

3.3. Automatic Learning Weight Fusion

In this part, we use a set of weights which are automatically determined by back propagation to fuse the expression features extracted from the two branches, as shown in Figure 2d. The weights γ 1 and γ 2 are determined by the Adam optimizer [23] and back propagation, which is the same way in which the parameters w and b in the convolution kernel are updated. The formula is as follows:
{ m t = β 1 · m t 1 + ( 1 β 1 ) · L ( w , b , γ ) t 1 γ t 1 v t = β 2 · v t 1 + ( 1 β 2 ) · ( L ( w , b , γ ) t 1 γ t 1 ) 2 m t ^ = m t / ( 1 β 1 t ) v t ^ = v t / ( 1 β 2 t ) γ t = γ t 1 α · m t ^ / ( v t ^ + ε ) ,
where β 1 and β 2 are parameters in the Adam optimizer, commonly represented by 0.9 and 0.999, L ( w , b , γ ) is the loss function, and α is the learning rate. The initial values of m t and v t are both 0. The ε is a small value, commonly set to 10 8 .
For each candidate nodule L i , both branch networks can extract the expression features. We fuse the expression features after the flat layer and before the full connection layer. E 1 ( L i ) represents feature vectors of the branch 1 output (similar to branch 2). The feature vectors E f u s i o n ( L i ) are fused by automatic learning weights as follows:
E f u s i o n ( L i ) = γ 1 E 1 ( L i ) + γ 2 E 2 ( L i ) ,
where γ 1 and γ 2 are weights which are automatically determined by back propagation from the training process. The feature vectors E f u s i o n ( L i ) can get the probability that L i is a real nodule through the full connection layer and the softmax. In the fully connected layer, except for the final output layer, the middle layer uses ReLU as the activation function and uses dropout with a threshold of 0.5.

4. Experiments and Results

4.1. Dataset

We used the LUNA16 challenge dataset to evaluate the performance of the network. The LUNA16 dataset was derived from a larger public dataset, the Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) [22]. The LUNA16 dataset is an LIDC-IDRI dataset that removes CT images of slice thicknesses greater than 3 mm, as well as those with space inconsistencies, and missing partial slices, resulting in 888 CT images. It contains 1186 real nodules, where each nodule is an irregular circle with a diameter of more than 3 mm, and most of them are located in the lung parenchyma central, with a few in the lung parenchyma periphery. The transverse plane size of each CT sample is 512 × 512 .
For false-positive reduction of pulmonary nodules, the LUNA16 challenge data provide two versions (V1 and V2) of the datasets. Each dataset contains the central coordinates of candidate nodules, the classification labels, and the corresponding patient identifier (ID). The dataset V1 was detected by three existing candidate detection algorithms [21,24,25] and contains 551,065 candidate nodules. The dataset was separated into nodule and non-nodule groups, where 1351 were labeled as nodules and 549,714 were labeled as non-nodules. Out of the 1186 real nodules previously recognized, 1120 of the 1351 labeled nodules are real nodules, and 231 are non-nodules. The dataset V2 contains 754,975 candidate nodules obtained by References [21,24,25,26,27], where 1557 were labeled as nodules, and 1166 are matched with the 1186 real nodules. Table 1 summarizes the candidate nodule statistics of the dataset in the LUNA16 challenge.

4.2. Evaluations Metrics

We used the competitive performance metric (CPM) [28] score for the evaluation of the algorithm, which is a metric used in the LUNA16 challenge list. The score calculates average sensitivity at seven predefined false positives per scan (FP/scan) indices, i.e., 0.125, 0.25, 0.5, 1, 2, 4, and 8, on a free receiver operation characteristic (FROC) curve. Currently, an FP/scan is widely used in clinical practice in the range of 1 to 4.

4.3. Experimental Settings

In this paper, we used a Pytorch, Intel(R) Xeon(R) E5-2680 v4 central processing unit (CPU) with 2.4 GHz and an NVIDIA GeForce GTX 1080ti graphics processing unit (GPU) as the experimental platform. The learning rate was set to 0.003 and the batch size was set to 128. We applied the batch norm technique at the output of each convolutional layer, and, in order to avoid overfitting, L2 regularization was applied; we also used a dropout technique with a threshold of 0.5 at the fully connected layer. The optimizer chosen was Adam, and β 1 and β 2 were set at respectively 0.9 and 0.999. The loss function used was cross-entropy. The formula for cross entropy is as follows:
L = [ y log y ^ + ( 1 y ) log ( 1 y ^ ) ] ,
where y is labeled with 1 or 0 (i.e., 1 denotes nodule and 0 denotes non-nodule), and y ^ indicates the probability that the model predicts real nodules.
In the data preprocessing stage, we extracted 40 × 40 × 40 , 30 × 30 × 30 , and 20 × 20 × 20 blocks for each sample via the central coordinates of the candidate nodules and used the nearest neighbor interpolation to transform them all into blocks. We used minimum–maximum normalization to ensure the algorithm converged faster by adjusting the data distribution of the sample. The intensities of the blocks were cut to [−1000, 400] Hounsfield units (HU)2 and normalized to [0, 1].
We used a five-fold cross-validation method to evaluate the algorithm, randomly dividing the CT scans into five parts based on patients, such that four parts were used for training and one was used for testing. To deal with the class imbalance between nodules and non-nodules and to avoid overfitting, we used data augmentation for nodule samples. Specifically, we rotated the nodules by 90°, 180°, and 270° on a transverse plane and shifted the center along each axis by 1, 2 pixels; we also added Gaussian noise with a mean of 0 and variance of 0.4 to those added data. The number of nodules and non-nodules in the five-fold cross-validation is shown in Table 2.

4.4. Experimental Results

We used several models from References [5,8,10,29] as baselines. Dou et al. [10] was based on a 3D method, where they used three sizes of nodule blocks as inputs. UACNN [10] selected three slices of the candidate nodule center, 3 mm above and 3 mm below, as inputs. Setio et al. [5] extracted nine different views of the nodules as input based on a 2D method. Sakamoto et al. [29] eliminated prediction inconsistencies by raising the threshold at each iteration. Kim et al. [8] gradually integrated nodule blocks of three sizes as inputs, based on 2D convolution kernels. Table 3 and Table 4 summarize the CPM scores of these different methods on V1 and V2 datasets.
Table 3 summarizes the best performance of our MSH-CNN algorithm on the V1 dataset. The proposed MSH-CNN was superior to most methods in terms of average CPM. Compared to Dou et al. [10], UACNN, and Setio et al. [5] in terms of average CPM, our 3D network had increases by 0.047, 0.050, and 0.036, respectively. Our method was 0.001 lower than Sakamoto’s method [29] at the 0.125 FP/scan, but it had a higher average CPM and sensitivity at the 0.25, 0.5, 1, 2, 4, and 8 FP/scan. In comparison with Kim et al. [8], although the sensitivity of our method was lower when the FP/scan was less than 1, our network still achieved the best performance at the 2, 4, and 8 FP/scan. Our method outperformed most methods in the false-positive rates of range 1–4 which are commonly used in clinical applications [30]. Figure 3 shows some examples of correctly classified candidate nodules. The FROC curves of different methods are shown in Figure 4.
We also tested performance on the large-sized V2 dataset as presented in Table 4. The proposed 3D method was better than other methods at high FP/scan. Specifically, the average CPM scores of our method and the sensitivity at different FP/scan were higher than those of Setio et al. [5]. In comparison with Kim et al. [8], the sensitivity of our network at 0.125, 0.25, and 0.5 FP/scan was lower, but it still achieved the best performance at the 1, 2, 4 and 8 FP/scan. It is worth noting that, comparing the performance of various algorithms on the V1 and V2 datasets, we can see that our sensitivity was higher than that of Kim et al. [8] in the false-positive rates of range 1–4 which are commonly used in clinical applications [30], and, as the dataset increased, our 3D method performed better and better at low FP/scan, even exceeding that of Kim et al. [8] at the 1 FP/scan. In terms of average CPM, Setio et al. [5], Kim et al. [8], and our method increased by 0.016, 0.034, and 0.055, respectively, and our 3D network had the best improvement.

4.5. Effects of the Proposed Methods

In order to quantitatively analyze our proposed MSH-CNN method, we designed the following CNNs for comparison:
(1) MSH-CNN-RI: The 3D multi-scale gradual integration module (I1-I2-I3 and I3-I2-I1) in MSH-CNN was replaced with the radical integration I1||I2||I3 in the same layer. The “||” symbol indicates concatenation by channel, as shown in Figure 5.
(2) MSH-CNN-IC: I1-I2-I3 (“Input 1”) and I3-I2-I1 (“Input 2”) in the MSH-CNN were interchanged, and Input 1 was input into Archi-2, while Input 2 was input into Archi-1, and the structure was similar to that shown in Figure 2d.
(3) MSH-CNN-AF: The fusion method of the MSH-CNN was modified to addition fusion without weights.
(4) MSH-CNN-II: In MSH-CNN, the network in Archi-1 was replaced by that in Archi-2 (including 3D inception), so that both branches contained 3D inception.
(5) MSH-CNN-EI: In MSH-CNN, the network in Archi-2 was replaced by that in Archi-1 (excluding 3D inception), so that neither branch contained 3D inception.
(6) MSH-CNN-LT: The training and test sets were expanded, and the MSH-CNN model was trained and tested on the datasets of LUNA16 + Tianchi. The Tianchi dataset was provided by the Aliabad Cloud of Intelligent Diagnosis of Pulmonary Nodules. The training and validation sets contained a total of 800 CT samples and 1245 nodules.
Table 5 summarizes the best performance of each network. Our method had higher average CPM score than other networks, which can be seen by applying each network to the same dataset. Our algorithm also performed better on the extended dataset, whether it was on the average CPM score or the sensitivity corresponding to different false-positive rates. The following conclusions could be drawn:
(1) MSH-CNN was 0.025 higher than MSH-CNN-RI (average CPM). At the 1 FP/scan, the sensitivity of MSH-CNN exceeded 0.9, but MSH-CNN-RI was a little worse than 0.9. This shows that, at the nodule block integration stage, 3D multi-scale gradual integration can make better use of the structural information of the nodules and its neighboring organs than the radical integration method at the same level.
(2) Comparison between MSH-CNN-IC with MSH-CNN. The sensitivity difference between the two structures was smaller than that of other networks. The sensitivity of MSH-CNN-IC was slightly higher than that of MSH-CNN at 0.125 and 1 FP/scan, but the sensitivity of MSH-CNN was high when the FP/scan was other values. This shows that different networks concern different input characteristics, and that a heterogeneous structure is useful.
(3) This part analyzed the influence of the fusion method on the classification task of the nodule. The average CPM scores of MSH-CNN-AF and the sensitivity of different false-positive rates were lower than those of MSH-CNN. This shows that using a set of weights which are automatically determined by back propagation can better complement the two branches.
(4) To verify the role of the 3D inception layer, we compared MSH-CNN-II and MSH-CNN-EI. As can be seen from Table 5, MSH-CNN-II had high sensitivity at low FP/scan. With the increase in FP/scan, MSH-CNN-EI sensitivity rose rapidly until it exceeded MSH-CNN-II. This shows that the 3D inception layer can improve the sensitivity at low FP/scan, but it has a limited effect on high FP/scan.
(5) Finally, as seen in Table 5, the average CPM and sensitivity corresponding to the different false-positive rates of the MSH-CNN-LT used more training and test data, significantly more than the MSH-CNN. This proves that the proposed algorithm has strong learning ability, and the performance of the algorithm was enhanced with the expansion of the dataset, which provides a bright prospect for further reducing the false-positive rate of pulmonary nodules.

5. Discussion

We quantitatively analyzed the performance of the network via experiments. The success of this method can be summarized by three main aspects. Firstly, the multi-scale nodule blocks were gradually integrated (in two different orders) so that the network could encode rich and complementary volume information according to the morphological features of the nodules and their neighboring organs. Secondly, two different branches of 3D CNN were designed for two inputs with different characteristics. The expression features can be fully extracted according to the characteristics of each input. Thirdly, the weights which were automatically determined by back propagation were used to fuse the expression features. The network can improve its performance by integrating complementary information more reasonably.
Here, we discuss the differences between the 3D and 2D methods. Table 6 shows the cost of Kim et al.’s [8] advanced 2D method and our 3D method. It can be seen that the 3D method takes up more memory and training time than the 2D method in model training. This is due to the fact that, during the training process, the feature map of the 3D method always has one dimension more than the 2D method, which leads to its computational complexity.
Comparing the performance of Kim et al.’s [8] 2D method and our 3D method on the V1 and V2 datasets, as can be seen from Table 3 and Table 4, on the V1 dataset with less data, the 2D method achieved good performance at low FP/scan, and its sensitivities were 0.121, 0.082, 0.048, and 0.011 higher than the 3D method at the 0.125, 0.25, 0.5, and 1 FP/scan, respectively. On the large-sized V2 dataset, the sensitivities of the 2D method were 0.071, 0.064, and 0.011 higher than the 3D method at the 0.125, 0.25, and 0.5 FP/scan, respectively. These differences were reduced on the V2 dataset. As the data increased, the sensitivities at the low FP/scan and the performance (CPM) of the 3D method grew faster than those of the 2D method. The reason was that, due to the complexity of the 3D network, it could not be trained adequately when the dataset was small, resulting in the sensitivity being relatively low at low FP/scan. When the dataset was large, 3D network could be fully trained, and its performance improved rapidly, eventually achieving better performance. It can be expected that, on a large-sized dataset, the 3D method has a brighter future than the 2D method.
Table 7 shows the CPM scores for each round of five-fold cross-validation on the V1 dataset. Fold five with the least training data had the lowest CPM scores; the CPM of fold four with the largest training dataset and positive and negative sample distribution was relatively uniform in the training set, and the test set was the highest.
This paper mainly focused on improving FP reduction rather than developing a whole pulmonary nodule detection system which typically consists of a candidate detection component and an FP reduction component. It is a candidate screening method, and it is essentially a classification problem. This means that our method can combine with existing nodule candidate detection elements to complete the task of pulmonary nodule detection, and also can be combined with other candidate detectors to solve candidate screening problems. Our method can also be used as a classification network alone to solve classification problems of 3D data.
In this paper, our evaluation standard was CPM score, which was extracted from the FRCO curve. Specifically, we measured the sensitivity at seven false-positive rates (1/8, 1/4, 1/2, 1, 2, 4, and 8 FPs per scan) and calculated their average. This performance metric was introduced by van Ginneken et al. [30]. Obviously, a perfect system has a score of 1 and the lowest score is 0. Currently, in clinical applications, the internal thresholds of most CAD systems are set to one, two, or four false positives per scan on average [30]. To make this task more challenging, the false-positive rates we used in the assessment were lower than the false-positive rates used in clinical practice.
Although CNN methods are becoming more prevalent in the field of medical image analysis, most of the work in the field of medical image analysis is still based on 2D CNN. On the other hand, 3D CNN is used less because there is a lack of 3D datasets, which is further complicated by expensive doctor labeling and privacy concerns. Also, 3D CNN training costs much more than 2D CNN. Therefore, we need to increase the computing power to use 3D CNN.
The network achieved good results, but there are still potential limitations, including high training costs, and the finding that the sensitivity at a low false-positive rate was slightly lower when the dataset was relatively small. Our future work will focus on extending the 3D dataset, for example, using generative adversarial networks (GANs) to generate datasets. We will also focus on combining 2D CNN and 3D CNN to reduce computing resources.

6. Conclusions

This paper proposed a novel multi-scale heterogeneous 3D CNN for the reduction of false positives in pulmonary nodules. In this network, the following three methods were used: (1) 3D multi-scale gradual integration; (2) heterogeneous feature extraction; (3) automatic learning weight fusion. We tested the effectiveness of our proposed algorithm by applying the method to the LUNA16 challenge dataset and estimated the effects of multi-scale heterogeneous 3D CNN with different structures. The algorithm of this paper also has certain limitations. In Section 5, we discussed the directions of improvement, as well as the future study directions.

Author Contributions

Conceptualization, Z.X.; validation, Z.X. and N.D.; investigation, L.G.; resources, F.Z.; writing—original draft preparation, N.D.; writing—review and editing, Z.X.; supervision, Y.L.; project administration, J.W.

Funding

This work was supported by Tianjin Science and Technology Major Projects and Engineering under grant No. 17ZXHLSY00040, No. 17ZXSCSY00060, and No. 17ZXSCSY00090, and the Program for Innovative Research Team at the University of Tianjin (No. TD13-5034).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer statistics, 2018. CA Cancer J. Clin. 2018, 68, 277–300. [Google Scholar] [CrossRef] [PubMed]
  2. American Cancer Society. Cancer Facts and Figures 2015. Available online: http://www.cancer.org/acs/groups/content/@editorial/documents/document/acspc-044552.pdf (accessed on 15 June 2019).
  3. Aberle, D.R.; Adams, A.M.; Berg, C.D. Reduced lung-cancer mortality with low-dose computed tomographic screening. N. Engl. J. Med. 2011, 365, 395–409. [Google Scholar] [PubMed]
  4. Gould, M.K.; Fletcher, J.; Iannettoni, M.D.; Lynch, W.R.; Midthun, D.E.; Naidich, D.P.; Ost, D.E. Evaluation of Patients with Pulmonary Nodules: When is it Lung Cancer? ACCP Evidence-Based Clinical Practice Guidelines (2nd edition). Chest 2007, 132, 108S–130S. [Google Scholar] [CrossRef] [PubMed]
  5. Setio, A.A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; Van Riel, S.J.; Wille, M.M.W.; Naqibullah, M.; Sanchez, C.I.; Van Ginneken, B. Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef] [PubMed]
  6. Sluimer, I.; Schilham, A.; Prokop, M. Computer analysis of computed tomography scans of the lung: A survey. IEEE Trans. Med. Imaging 2006, 25, 385–405. [Google Scholar] [CrossRef] [PubMed]
  7. Ciompi, F.; Chung, K.; van Riel, S.J.; Setio, A.A.A.; Gerke, P.K.; Jacobs, C.; Scholten, E.T.; Schaefer-Prokop, C.; Wille, M.M.W.; Marchiano, A.; et al. Towards Automatic Pulmonary Nodule Management in Lung Cancer Screening with Deep Learning. Sci. Rep. 2017, 7, 46479. [Google Scholar] [CrossRef] [PubMed]
  8. Kim, B.C.; Yoon, J.S.; Choi, J.S.; Suk, H.I. Multi-Scale Gradual Integration CNN for False Positive Reduction in Pulmonary Nodule Detection. Neural Netw. 2019, 115, 1–10. [Google Scholar] [CrossRef] [PubMed]
  9. Zhang, Z.C.; Li, X.L.; You, Q.J.; Luo, X.Q. Multicontext 3D residual CNN for false positive reduction of pulmonary nodule detection. Int. J. Imaging Syst. Technol. 2019, 29, 42–49. [Google Scholar] [CrossRef]
  10. Dou, Q.; Chen, H.; Yu, L.; Qin, J.; Heng, P.A. Multilevel Contextual 3-D CNNs for False Positive Reduction in Pulmonary Nodule Detection. IEEE Trans. Biomed. Eng. 2017, 64, 1558–1567. [Google Scholar] [CrossRef]
  11. Abbosh, C.; Birkbak, N.J.; Wilson, G.A.; Jamal-Hanjani, M.; Constantin, T.; Salari, R.; Le Quesne, J.; Moore, D.A.; Veeriah, S.; Rosenthal, R.; et al. Phylogenetic ctDNA analysis depicts early-stage lung cancer evolution. Nature 2018, 554, 264. [Google Scholar] [CrossRef]
  12. Dou, Q.; Chen, H.; Yu, L.; Zhao, L.; Qin, J.; Wang, D.; Mok, V.C.; Shi, L.; Heng, P.A. Automatic Detection of Cerebral Microbleeds from MR Images via 3D Convolutional Neural Networks. IEEE Trans. Med. Imaging 2016, 35, 1182–1195. [Google Scholar] [CrossRef] [PubMed]
  13. Abdelhafiz, D.; Yang, C.; Ammar, R.; Ko, J.; Nabavi, S. Deep convolutional neural networks for mammography: Advances, challenges and applications. BMC Bioinform. 2019, 20, 281. [Google Scholar] [CrossRef] [PubMed]
  14. Yang, J.; Faraji, M.; Basu, A. Robust segmentation of arterial walls in intravascular ultrasound images using Dual Path U-Net. Ultrasonics 2019, 96, 24–33. [Google Scholar] [CrossRef] [PubMed]
  15. Hu, Z.L.; Tang, J.S.; Wang, Z.M.; Zhang, K.; Zhang, L.; Sun, Q.L. Deep learning for image-based cancer detection and diagnosis—A survey. Pattern Recognit. 2018, 83, 134–149. [Google Scholar] [CrossRef]
  16. Roth, H.R.; Lu, L.; Liu, J.; Yao, J.; Seff, A.; Cherry, K.; Kim, L.; Summers, R.M. Improving Computer-Aided Detection using Convolutional Neural Networks and Random View Aggregation. IEEE Trans. Med. Imaging 2016, 35, 1170–1181. [Google Scholar] [CrossRef]
  17. Ding, J.; Li, A.; Hu, Z.Q.; Wang, L.W. Accurate Pulmonary Nodule Detection in Computed Tomography Images Using Deep Convolutional Neural Networks. In Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI 2017), Quebec City, QC, Canada, 11–13 September 2017; pp. 559–567. [Google Scholar]
  18. Shen, W.; Zhou, M.; Yang, F.; Yang, C.; Tian, J. Multi-scale Convolutional Neural Networks for Lung Nodule Classification. Inf. Process. Med. Imaging 2015, 24, 588–599. [Google Scholar] [PubMed]
  19. Setio, A.A.A.; Traverso, A.; de Bel, T.; Berens, M.S.N.; van den Bogaard, C.; Cerello, P.; Chen, H.; Dou, Q.; Fantacci, M.E.; Geurts, B. Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge. Med. Image Anal. 2017, 42, 1–13. [Google Scholar] [CrossRef]
  20. Messay, T.; Hardie, R.C.; Rogers, S.K. A new computationally efficient CAD system for pulmonary nodule detection in CT imagery. Med. Image Anal. 2010, 14, 390–406. [Google Scholar] [CrossRef]
  21. Jacobs, C.; van Rikxoort, E.M.; Twellmann, T.; Scholten, E.T.; de Jong, P.A.; Kuhnigk, J.M.; Oudkerk, M.; de Koning, H.J.; Prokop, M.; Schaefer-Prokop, C.; et al. Automatic Detection of Subsolid Pulmonary Nodules in Thoracic Computed Tomography Images. Med. Image Anal. 2014, 18, 374–384. [Google Scholar] [CrossRef]
  22. Armato, S.G.; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The Lung Image Database Consortium (LIDC) and Image Database Resource Initiative (IDRI): A Completed Reference Database of Lung Nodules on CT Scans. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef]
  23. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv, 2017; arXiv:1412.6980. [Google Scholar]
  24. Murphy, K.; van Ginneken, B.; Schilham, A.M.R.; de Hoop, B.J.; Gietema, H.A.; Prokop, M. A large-scale evaluation of automatic pulmonary nodule detection in chest CT using local image features and k-nearest-neighbour classification. Med. Image Anal. 2009, 13, 757–770. [Google Scholar] [CrossRef] [PubMed]
  25. Tan, M.; Deklerck, R.; Jansen, B.; Bister, M.; Cornelis, J. A Novel Computer-Aided Lung Nodule Detection System for CT Images. Med. Phys. 2011, 38, 5630–5645. [Google Scholar] [CrossRef] [PubMed]
  26. Setio, A.A.A.; Jacobs, C.; Gelderblom, J.; van Ginneken, B. Automatic Detection of Large Pulmonary Solid Nodules in Thoracic CT Images. Med. Phys. 2015, 42, 5642–5653. [Google Scholar] [CrossRef] [PubMed]
  27. Traverso, A.; Torres, E.L.; Fantacci, M.E.; Cerello, P. Computer-Aided Detection Systems to Improve Lung Cancer Early Diagnosis: State-of-the-art and Challenges. In Proceedings of the 7th Young Researcher Meeting, Torino, Italy, 24–26 October 2016. [Google Scholar]
  28. Niemeijer, M.; Loog, M.; Abramoff, M.D.; Viergever, M.A.; Prokop, M.; van Ginneken, B. On Combining Computer-Aided Detection Systems. IEEE Trans. Med. Imaging 2011, 30, 215–223. [Google Scholar] [CrossRef] [PubMed]
  29. Sakamoto, M.; Nakano, H.; Zhao, K.; Sekiyama, T. Multi-stage Neural Networks with Single-Sided Classifiers for False Positive Reduction and Its Evaluation Using Lung X-Ray (CT) Images. In Proceedings of the Image Analysis and Processing (ICIAP 2017), Catania, Italy, 11–15 September 2017; pp. 370–379. [Google Scholar]
  30. van Ginneken, B.; Armato, S.G.; de Hoop, B.; van Amelsvoort-van de Vorst, S.; Duindam, T.; Niemeijer, M.; Murphy, K.; Schilham, A.; Retico, A.; Fantacci, M.E. Comparing and combining algorithms for computer-aided detection of pulmonary nodules in computed tomography scans: The ANODE09 study. Med. Image Anal. 2010, 14, 707–722. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Examples of pulmonary nodules and non-nodules carrying similar structures. Each example is a two-dimensional (2D) transverse plane extracted from the location of the candidate nodule.
Figure 1. Examples of pulmonary nodules and non-nodules carrying similar structures. Each example is a two-dimensional (2D) transverse plane extracted from the location of the candidate nodule.
Applsci 09 03261 g001
Figure 2. Our proposed pulmonary nodule false-positive reduction network framework. The numbers above the thick arrows present a kernel size, e.g., 3 × 3 × 3 . (s: stride, p: padding, Conv: three-dimensional (3D) convolution, MP: 3D max-pooling). The figure after @ denotes the number of channels. (a) Extraction of 3D nodule blocks. According to the coordinate parameters of each candidate nodule, three blocks of different scales are extracted and each block is resized to be equivalent to each other, i.e., I1, I2, and I3. (b) The 3D Multi-scale gradual integration model. Nodule blocks of different scales are gradually integrated and contain two orders, i.e., I1-I2-I3 (“Input1”) and I3-I2-I1 (“Input2”). The “||” symbol represents concatenation by channel. (c) The 3D inception model. The input feature maps are processed by four parallel branches, and the results are integrated into output features map by channels. (d) Multi-scale heterogeneous 3D convolution neural network (CNN). “Input1” and “Input2” are input into two branches of different structures. Then, the expression features of the two branches are fused by automatic learning weights. The entire “Input1” and “Input2” blocks represent gradual integration in Figure 2b. The inception structure is shown in Figure 2c. The flat layer draws feature vectors into a column, and the Fc layer represents the full connection layer. The parameters γ 1 and γ 2 represent weights which are determined by back propagation.
Figure 2. Our proposed pulmonary nodule false-positive reduction network framework. The numbers above the thick arrows present a kernel size, e.g., 3 × 3 × 3 . (s: stride, p: padding, Conv: three-dimensional (3D) convolution, MP: 3D max-pooling). The figure after @ denotes the number of channels. (a) Extraction of 3D nodule blocks. According to the coordinate parameters of each candidate nodule, three blocks of different scales are extracted and each block is resized to be equivalent to each other, i.e., I1, I2, and I3. (b) The 3D Multi-scale gradual integration model. Nodule blocks of different scales are gradually integrated and contain two orders, i.e., I1-I2-I3 (“Input1”) and I3-I2-I1 (“Input2”). The “||” symbol represents concatenation by channel. (c) The 3D inception model. The input feature maps are processed by four parallel branches, and the results are integrated into output features map by channels. (d) Multi-scale heterogeneous 3D convolution neural network (CNN). “Input1” and “Input2” are input into two branches of different structures. Then, the expression features of the two branches are fused by automatic learning weights. The entire “Input1” and “Input2” blocks represent gradual integration in Figure 2b. The inception structure is shown in Figure 2c. The flat layer draws feature vectors into a column, and the Fc layer represents the full connection layer. The parameters γ 1 and γ 2 represent weights which are determined by back propagation.
Applsci 09 03261 g002aApplsci 09 03261 g002b
Figure 3. Examples of candidate nodule classification by our algorithm. Each example is a transverse plane of candidate nodules; P is the probability that the algorithm predicts an example as real nodule.
Figure 3. Examples of candidate nodule classification by our algorithm. Each example is a transverse plane of candidate nodules; P is the probability that the algorithm predicts an example as real nodule.
Applsci 09 03261 g003
Figure 4. The free receiver operation characteristic (FROC) curves of different methods.
Figure 4. The free receiver operation characteristic (FROC) curves of different methods.
Applsci 09 03261 g004
Figure 5. The network structure of multi-scale heterogeneous three-dimensional (3D) convolutional neural network with radical integration (MSH-CNN-RI): radical integration of nodule blocks on the same layer. The two branches of the network have the same structure as the MSH-CNN. The numbers above the thick arrows represent kernel size, e.g., 3 × 3 × 3 (s: stride, p: padding, Conv: 3D convolution). The figure after @ denotes the number of channels. The parameters γ 1 and γ 2 are weights automatically learned from the training dataset.
Figure 5. The network structure of multi-scale heterogeneous three-dimensional (3D) convolutional neural network with radical integration (MSH-CNN-RI): radical integration of nodule blocks on the same layer. The two branches of the network have the same structure as the MSH-CNN. The numbers above the thick arrows represent kernel size, e.g., 3 × 3 × 3 (s: stride, p: padding, Conv: 3D convolution). The figure after @ denotes the number of channels. The parameters γ 1 and γ 2 are weights automatically learned from the training dataset.
Applsci 09 03261 g005
Table 1. Statistics of the dataset in the Lung Nodule Analysis 2016 (LUNA16) challenge. The number in parentheses indicates the number of nodule-labeled matches with 1186 real nodules. V—version.
Table 1. Statistics of the dataset in the Lung Nodule Analysis 2016 (LUNA16) challenge. The number in parentheses indicates the number of nodule-labeled matches with 1186 real nodules. V—version.
DatasetCandidates
NoduleNon-Nodule
V11351 (1120)549,714
V21557 (1166)753,418
Table 2. Statistics of the dataset used for five-fold cross-validation.
Table 2. Statistics of the dataset used for five-fold cross-validation.
DatasetFold 1Fold 2Fold 3Fold 4Fold 5
TrainingTestTrainingTestTrainingTestTrainingTestTrainingTest
V1Patient533133533133533133533133532134
Nodule936184867253910210912208855265
Augmentation101,088-93,636-191,916-98,496-92,340-
Non-nodule439,022110,692440,811108,903438,087111,627444,860104,854436,076113,638
V2Patient711177711177710178710178710178
Nodule970196953213917249930236894272
Augmentation100,440-102,924-99,036-104,760-96,552-
Non-nodule598,891154,527602,299151,119605,820147,598606,027147,391600,635152,783
Table 3. Competitive performance metric (CPM) scores of different methods on V1 dataset of LUNA16. FP—false positive; MSH-CNN—multi-scale heterogeneous three-dimensional (3D) convolutional neural network.
Table 3. Competitive performance metric (CPM) scores of different methods on V1 dataset of LUNA16. FP—false positive; MSH-CNN—multi-scale heterogeneous three-dimensional (3D) convolutional neural network.
FP/Scan0.1250.250.51248Average CPM
Sensitivity
V1Dou et al. [10]0.6780.7380.8160.8480.8790.9070.9220.827
UACNN [10]0.6550.7450.8070.8490.8800.9070.9250.824
Setio et al. [5]0.6920.7710.8090.8630.8950.9140.9230.838
Sakamoto et al. [29]0.7600.7940.8330.8600.8760.8930.9060.846
Kim et al. [8]0.8800.8940.9070.9120.9140.9190.9270.908
MSH-CNN (ours)0.7590.8120.8590.9010.9170.9280.9430.874
Table 4. CPM scores of different methods on V2 dataset of LUNA16.
Table 4. CPM scores of different methods on V2 dataset of LUNA16.
FP/Scan0.1250.250.51248Average CPM
Sensitivity
V2Setio et al. [5]0.6690.7600.8310.8920.9230.9440.9600.854
Kim et al. [8]0.9040.9310.9430.9470.9520.9560.9620.942
MSH-CNN (ours)0.8330.8670.9320.9530.9600.9770.9840.929
Table 5. CPM scores of different structure networks. MSH-CNN-RI radically integrates nodule blocks; MSH-CNN-IC interchange the inputs of two branches; MSH-CNN-AF uses addition fusion without weights; MSH-CNN-II has both branches containing inception; MSH-CNN-EI has neither branch containing Inception; MSH-CNN-LT uses the LUNA16 + Tianchi dataset.
Table 5. CPM scores of different structure networks. MSH-CNN-RI radically integrates nodule blocks; MSH-CNN-IC interchange the inputs of two branches; MSH-CNN-AF uses addition fusion without weights; MSH-CNN-II has both branches containing inception; MSH-CNN-EI has neither branch containing Inception; MSH-CNN-LT uses the LUNA16 + Tianchi dataset.
FP/Scan0.1250.250.51248Average CPM
Sensitivity
MSH-CNN-RI0.7160.7730.8310.8850.9000.9040.9360.849
MSH-CNN-IC0.7630.8110.8400.9050.9090.9260.9420.871
MSH-CNN-AF0.7440.7820.8420.8730.8970.9150.9400.856
MSH-CNN-II0.7710.8180.8390.8850.9050.9170.9380.868
MSH-CNN-EI0.7410.7760.8350.8700.9130.9200.9430.857
MSH-CNN (ours)0.7590.8120.8590.9010.9170.9280.9430.874
MSH-CNN-LT0.9320.9440.9570.9760.9820.9890.9910.967
Table 6. The cost of two-dimensional (2D) and 3D networks. GPU—graphics processing unit.
Table 6. The cost of two-dimensional (2D) and 3D networks. GPU—graphics processing unit.
MethodsConvolution Kernel TypeMemory Usage (GPU)Training TimeTesting Time (One Sample)
Kim et al. [8]2D2089 MBAbout 12 hAbout 0.6 s
MSH-CNN (ours)3D6783 MBAbout 30.5 hAbout 1 s
Table 7. CPM scores for each round of five-fold cross-validation on the V1 dataset. The ratio represents the ratio of nodule to non-nodule in the training/test dataset.
Table 7. CPM scores for each round of five-fold cross-validation on the V1 dataset. The ratio represents the ratio of nodule to non-nodule in the training/test dataset.
IndexFold 1Fold 2Fold 3Fold 4Fold 5
Ratio (training)1:4691:5081:4811:4881:510
Ratio (test)1:6021:4301:5321:5041:429
Average CPM0.8730.8740.8760.8780.869

Share and Cite

MDPI and ACS Style

Xiao, Z.; Du, N.; Geng, L.; Zhang, F.; Wu, J.; Liu, Y. Multi-Scale Heterogeneous 3D CNN for False-Positive Reduction in Pulmonary Nodule Detection, Based on Chest CT Images. Appl. Sci. 2019, 9, 3261. https://doi.org/10.3390/app9163261

AMA Style

Xiao Z, Du N, Geng L, Zhang F, Wu J, Liu Y. Multi-Scale Heterogeneous 3D CNN for False-Positive Reduction in Pulmonary Nodule Detection, Based on Chest CT Images. Applied Sciences. 2019; 9(16):3261. https://doi.org/10.3390/app9163261

Chicago/Turabian Style

Xiao, Zhitao, Naichao Du, Lei Geng, Fang Zhang, Jun Wu, and Yanbei Liu. 2019. "Multi-Scale Heterogeneous 3D CNN for False-Positive Reduction in Pulmonary Nodule Detection, Based on Chest CT Images" Applied Sciences 9, no. 16: 3261. https://doi.org/10.3390/app9163261

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop