Next Article in Journal
Multilevel Task Offloading and Resource Optimization of Edge Computing Networks Considering UAV Relay and Green Energy
Next Article in Special Issue
Visual and Quantitative Evaluation of Amyloid Brain PET Image Synthesis with Generative Adversarial Network
Previous Article in Journal
NURBS-Enhanced Meshfree Method with an Integration Subtraction Technique for Complex Topology
Previous Article in Special Issue
A Novel Computer-Aided-Diagnosis System for Breast Ultrasound Images Based on BI-RADS Categories
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using 2D CNN with Taguchi Parametric Optimization for Lung Cancer Recognition from CT Images

Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung City 411, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(7), 2591; https://doi.org/10.3390/app10072591
Submission received: 13 March 2020 / Revised: 2 April 2020 / Accepted: 7 April 2020 / Published: 9 April 2020
(This article belongs to the Special Issue Machine Learning in Medical Image Processing)

Abstract

:
Lung cancer is one of the common causes of cancer deaths. Early detection and treatment of lung cancer is essential. However, the detection of lung cancer in patients produces many false positives. Therefore, increasing the accuracy of the classification of diagnosis or true detection by computed tomography (CT) is a difficult task. Solving this problem using intelligent and automated methods has become a hot research topic in recent years. Hence, we propose a 2D convolutional neural network (2D CNN) with Taguchi parametric optimization for automatically recognizing lung cancer from CT images. In the Taguchi method, 36 experiments and 8 control factors of mixed levels were selected to determine the optimum parameters of the 2D CNN architecture and improve the classification accuracy of lung cancer. The experimental results show that the average classification accuracy of the 2D CNN with Taguchi parameter optimization and the original 2D CNN in lung cancer recognition are 91.97% and 98.83% on the Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) dataset, and 94.68% and 99.97% on the International Society for Optics and Photonics with the support of the American Association of Physicists in Medicine (SPIE-AAPM) dataset, respectively. The proposed method is 6.86% and 5.29% more accurate than the original 2D CNN on the two datasets, respectively, proving the superiority of proposed model.

1. Introduction

Lung cancer is one of the critical diseases with rapidly rising morbidity and mortality rates, being the primary reason for cancer-related mortality around the world, with 1.8 million deaths annually [1]. Although the early detection rate has increased significantly, distinguishing malignant from benign tumors is one of the most challenging tasks [2,3]. Therefore, early and accurate diagnosis of lung cancer is essential. In recent years, many researchers applied machine learning models to detect and diagnose lung computed tomography (CT) images with the help of various computer assisted detection (CAD) systems, such as convolutional neural networks (CNNs), which demonstrated classification performance on medical images [4,5]. Liu et al. [6] proposed a multi-view multi-scale CNN for lung nodule type classification from CT images. In another study, a CNN was integrated to assess pulmonary nodule malignancy in an automated existing pipeline of lung cancer detection [7]. However, this method is restricted to a limited number of visual parameters and the fixed value range. Thus, increasing the accuracy of lung cancer classification through CNN with parametric optimization will play a vital role in lung cancer recognition from CT images.
CT is widely used in screening procedures to reduce high mortality and to obtain many details about lung nodules and the surrounding structures. To effectively detect lung cancer from CT images, CNNs have demonstrated outstanding performance in applications including vision and images in the medical field in CAD systems [8]. A CNN model was used for classification, including three convolutional layers, three pooling layers, and two fully-connected layers [9]. Two-dimensional (2D) CNN has been applied in various fields with significant results such as image classification, face recognition, and natural language processing [10,11]. The design of the existing CAD requires the training of a large number of parameters, but parameters setting is complicated, so the parameters must be optimized to increase the accuracy of the classification [12]. Yunus and Alsoufi [13] applied the Taguchi method to evaluate the output quality characteristics to predict the optimum parameters. In this study, we used the Taguchi method to find optimal parameter combinations based on the design of experiments and to solve the complicated parameters setting problem in CAD systems.
The Taguchi method is widely used for statistical analysis, and has been effectively used to help design experiments and obtain highly reliable results in studies. To optimize factors that affect the performance characteristics of a process or system, the Taguchi method reduces time and material costs [14]. The 2D CNN contains three convolutional, three max-pooling, and two fully-connected layers, and has the same deep learning architecture, but using a 2D CNN single view to train a large number of parameters might produce unstable predictions [15]. To alleviate this problem, we used a 2D CNN with the Taguchi method to find optimal parameter combinations.
For the purpose of overcoming the challenges associated with improving the classification performance of benign and malignant tumors from CT images, we propose a 2D CNN with Taguchi parametric optimization to address the complicated parameters setting problem and to increase the accuracy of classification. The objective of this study was to use the 2D CNN with Taguchi parametric optimization to determine the optimum parameters of the 2D CNN architecture and improve the lung cancer classification accuracy. The contribution is finding the optimal parameter combinations in the 2D CNN architecture on proven datasets to improve accuracy for lung cancer recognition with the help of various CAD systems.
The remainder of this is study is organized as follows. Section 2 presents the proposed method including the 2D CNN and the experiments in the Taguchi method. The experimental results are reported in Section 3. Finally, the conclusion is provided in Section 4.

2. Methods

The proposed 2D CNN with Taguchi parametric optimization to improve the classification performance and the proposed CAD system framework are shown in Figure 1. For the proposed CAD lung cancer detection system, we used three main steps. First, we use the Taguchi method to evaluate optimal parameter combinations for the 2D CNN model (Section 2.1). Next, we extract the input CT images to the 2D CNN model (Section 2.2). Third, we set the optimization parameters of the 2D CNN model to train and achieve malignant and benign tumor classification (Section 2.3).

2.1. Taguchi Method

The Taguchi method is a statistical approach using an orthogonal array (OA), consisting of factors and levels, to classify the results to optimize process parameters [16]. In the Taguchi method, the experimental design using OA tables enables the optimization process. The results are converted to a signal-to-noise (S/N) ratio to determine the combination of control factor levels to improve system stability and reduce quality loss. In this study, considering the factors that were found to be effective on the 2D CNN parameter design, the parameters with eight factors and three levels are listed in Table 1. To reduce the number of experiments and improve the reliability of experiments, the L36(211,312) OA was selected for the experimental design, and 36 experimental runs generated by Minitab® 19 (Scientific Formosa Inc, Taipei, Taiwan) are given in Table 2.
According to the Taguchi optimization and the S/N ratio to quantify the output quality, the S/N ratio was selected as the optimization criterion. The best experimental results obtained by the calculation of average S/N ratio for each factor and the optimal level of the process parameters were indicated by the level with the largest S/N ratio. We analyzed the S/N ratio using three different performance characteristics: larger-is-better, nominal-is-better, and smaller-is-better [17]. In this study, ‘larger-is-better’ performance characteristics were preferred.
S / N = 10 log ( 1 / n i = 1 n 1 y i 2 )
where n is the total number of replications of per test run and yi is the accuracy of the 2D CNN in the replication experiment for lung cancer recognition i executed under the same experimental conditions per test run.
We used analysis of variance (ANOVA) on the experimental data from the Taguchi method to determine the significance of process parameters statistically affecting extraction and the measurements [18].
SS T = [ i = 1 N y i 2 ] T 2 N
SS A = [ i = 1 K A ( A i 2 n A i ) ] T 2 N
SS e = SS T ( SS A + SS B +   )
DF = K A 1
M S A = S S A D F
M S e = S S E r r o r D F
F A = M S A M S e
where T represents the sum of all observations, N is the total number of experiments, SST is the sum of squares of total change, SSA is the sum of squares due to factor A, KA is number of levels of factor A, Ai is the sum of all observations of level i of factor A, n A i is the number of all observations at level i of factor A, DF is the degrees of freedom, MSA is the variance of the factor, and FA represents the F ratio of the factor.

2.2. Materials

The images used in this study were collected from lung cancer International Society for Optics and Photonics with the support of the American Association of Physicists in Medicine (SPIE-AAPM) Lung CT Challenge and Lung Image Database Consortium and Infectious Disease Research Institute (LIDC-IDRI) datasets, consisting of 22,489 CT and 244,617 CT scans 512 × 512 pixels in size, respectively [19,20]. The sample benign and malignant CT images of lungs from the LIDC-IDRI dataset for 2D CNN model training are shown in Figure 2 and Figure 3, and sample CT images of lungs from the SPIE-AAPM Lung CT Challenge dataset are shown in Figure 4. We used 70% of the CT image cases to train the 2D CNN model, and the other 30% for validation.

2.3. 2D CNN Model

The 2D CNN designs the correct model of single views by reducing the bias of a single network, thereby effectively and accurately reducing the false positive rate of lung nodule [21]. This study proposes a 2D CNN architecture includes of two convolutional, two pooling and two fully-connected layers. The input image sizes 50 × 50 × 3, and the convolutional layer is based on the process of circulating a particular filter over the entire image. Kernel size, stride, padding, and filters can be different sizes, and the proposed 2D CNN architecture is described in Table 3.
We statistically analyzed the classification performance of the 2D CNN in terms of the Taguchi parametric optimization, accuracy, sensitivity, specificity, and false positive rate (FPR) derived from the confusion matrix. Accuracy is the correct prediction from the whole dataset, sensitivity is the ability of the test to correctly identify patients with the disease, and the specificity of a clinical test refers to the ability of the test to correctly identify patients without the disease. The following commonly used validation measurements are shown [3].
Accuracy = ( T P + T N ) ( T P + F N ) + ( F P + T N )
Sensitivity = ( T P ) ( T P + F P )
Specificity = ( T N ) ( T N + F P )
FPR = ( F P ) ( T N + F P )
where true positive (TP) is malignant correctly identified as malignant, false positive (FP) is benign incorrectly identified as malignant, true negative (TN) is benign correctly identified as benign, and false negative (FN) is malignancy incorrectly identified as benign.

3. Experimental Results

For implementing the proposed framework, we trained and validated the 2D CNN model with Taguchi parametric optimization both on LIDC-IDRI [19] and SPIE-AAPM [20] separately. The architecture of the proposed CAD system is shown in Figure 5.

3.1. LIDC-IDRI

LIDC-IDRI [19] consists of 1018 CT scans case, and 244,617 CT images. Pulmonary nodule diameters range from 3 to 30 mm, including only 4702 images of small nodules (<3 mm). For differential diagnostic imaging features, the larger the size (maximum diameter) of the nodule, the higher its malignancy: ≤8 mm is mostly benign and >8 mm is more likely malignant. Other types are very faint with smaller lesions called ground-glass opacity (GGO): ≤10 mm is mostly benign; >10 mm is more malignant. We performed experiments according to the combination of experiments in the OA table to improve the reliability of the experimental results. We obtained observations by combining experiments with three replications. The values of the parameters, the measured accuracies of the 2D CNN for lung cancer recognition, and S/N ratios in each run are listed in Table 4. The S/N ratio values of the accuracy were calculated by implementing of the-higher-the-better concept. A larger S/N ratio means less quality loss, that is, higher quality. Based on the results of the 36 runs, the maximum mean accuracy of 98.48% was achieved with Run #31, whereas the minimum mean accuracy of 78.29% occurred for Run #30.
The mean responses for the S/N ratios and the ANOVA table are shown in Table 5 and Table 6. Table 6 shows that four factors–C1_S, C2_S, C1_KS, and C2_F–have a significant impact. Table 5 and Figure 6 show that the factors affecting the S/N ratios based on yield extraction method are affected for each level of the parameters. The most significant factors can be determined by the larger difference in the S/N ratio. The experimental results indicate that factor C2_F has noticeable delta value of 0.845%, meaning that factor C2_F has the big significant parameter on model response. The importance of parameter was ranked from C2_F > C1_KS > C2_S > C2_P > C1_S > C2_KS > C1_F > C1_P. Hence, the best levels are A1 (level 1 for C1_S), B1 (level 1 for C1_P), C1 (level 1 for C2_S), D2 (level 2 for C2_P), E1 (level 1 for C1_KS), F1 (level 1 for C1_F), G3 (level 3 for C2_KS) and H3 (level 3 for C2_F). Optimum parameter C1_S is 1, C1_P is 0, C2_S is 1, C2_P is 1, C1_KS is 3, C1_F is 4, C2_KS is 7 and C2_F is 32.
To compare our 2D CNN performance with Taguchi parametric optimization with the other neural networks such as deep fully convolutional neural network (DFCNet) [22] and deep features + k-nearest neighbor (kNN) + minimum redundancy maximum relevance (mRMR) [2] based on LIDC-IDRI dataset. The results are shown in Table 7. The best accuracy was 99.51% for the Deep features + kNN + mRMR, but the best sensitivity was produced by 2D CNN with Taguchi parametric optimization with a sensitivity of 99.97%, the best specificity of 99.93%, and FPR of 0.06%.

3.2. SPIE-AAPM

SPIE-AAPM [20] consists of 22,489 CT images for 68 samples, including 31 benign and 37 malignant samples. The mean responses for the S/N ratios and ANOVA are shown in Table 8 and Table 9, respectively. Table 9 shows that four factors—C1_S, C2_P, C1_KS, and C2_F—have a significant impact. From Table 8 and Figure 7, the factors affecting the S/N ratios based on the yield extraction method are affected at each level of the parameters. The most significant factors can be determined by the larger differences in the S/N ratio. The experimental results indicated that factor C2_F has a noticeable delta value of 0.76%, meaning that factor C2_F is the most significant parameter in the model response. In terms of the importance, the parameters were ranked: C2_F > C1_KS > C1_S > C2_P > C1_F > C2_S > C2_KS > C1_P. Hence, the best levels are A1 (level 1 for C1_S), B2 (level 2 for C1_P), C1 (level 1 for C2_S), D2 (level 2 for C2_P), E1 (level 1 for C1_KS), F1 (level 1 for C1_F), G2 (level 2 for C2_KS) and H3 (level 3 for C2_F). Optimum parameter C1_S is 1, C1_P is 1, C2_S is 1, C2_P is 1, C1_KS is 3, C1_F is 4, C2_KS is 5 and C2_F is 32.
To evaluate the classification performance of the proposed 2D CNN with Taguchi parametric optimization, we selected DFCNet [22] and back propagation neural network-variational inertia weight (BPNN-VIW) [23] for comparison based on the SPIE-AAPM dataset. The results are provided in Table 10. The best sensitivity was 100% for BPNN-VIW, but the best accuracy was recorded for 2D CNN with Taguchi parametric optimization with an accuracy of 99.97%, the best specificity of 99.94%, and an FPR of 0.06%.

3.3. Discussion

We compared the performance of the proposed 2D CNN with Taguchi parametric optimization with the original 2D CNN model on the LIDC-IDRI and SPIE-AAPM datasets. We optimized the convolutional neural network model with new data on the LIDC-IDRI and SPIE-AAPM datasets using the best parameter configuration to improve accuracy. The comparison with the original accuracy, and parameters settings are shown in Table 11 and Table 12, respectively. On the LIDC-IDRI dataset, the original 2D CNN model attained an accuracy of 91.97%. After three more experiments with the best combination of parameters, the average accuracy rate was 98.83% (98.04%, 98.91%, and 99.53%). On the SPIE-AAPM dataset for the original 2D CNN model the accuracy was 94.68%. After three more experiments with the best combination of parameters, the average accuracy rate was 99.97% (100%, 99.90%, and 100%). Therefore, compared with the original model, the accuracy significant improved 6.86% and 5.29% for the two datasets, respectively. The results showed that the design model for 2D CNN with parametric optimization is necessary.

4. Conclusions

In this study, we constructed a 2D CNN Taguchi parametric optimization method for recognizing lung cancer from CT images. We successfully applied the Taguchi method to determine optimal 2D CNN architecture to increase the classification accuracy of lung cancer recognition, which was verified by running an eight-factor mixed-level experimental design. Additionally, our results identified the significant factors affecting the extraction yield. In this study, the LIDC-IDRI and SPIE-AAPM datasets were used to verify the proposed method. In the LIDC-IDRI dataset experimental results, the deep features + kNN + mRMR method for lung nodule type classification in a CAD system, the accuracy was 99.51%. Although this is better than the 98.83% result for the proposed method, the sensitivity, specificity, and FPR for the proposed method are superior to their results. In the SPIE-AAPM dataset experimental results, the best accuracy of the BPNN-VIW method in classifying lung CT was 97.2%. The proposed method had an accuracy of 99.97%, so it is better than the previously proposed CAD system. The experimental findings showed the proposed method provides increased classification accuracy, the essential model design with parametric optimization is necessary.
The classification accuracy of lung cancer using the original 2D CNN was 91.97% on the LIDC-IDRI dataset; the lung cancer classification accuracies for the three experiments with the proposed method were 98.04%, 98.91%, and 99.53%. Hence, the average accuracy rate of three observations of the proposed method was 98.83%. In addition, the lung cancer classification accuracy of the original 2D CNN was 94.68% on the SPIE-AAPM dataset; the accuracies were 100%, 99.90%, and 100% for the proposed method. Hence, the average accuracy rate of three observations of the proposed method was 99.97%. The accuracies of the proposed method were 6.86% and 5.29% higher than the original 2D CNN on the LIDC-IDRI and SPIE-AAPM datasets, respectively. The findings show that the proposed 2D CNN with Taguchi parametric optimization is more accurate for this task.
The limitation of this study is the depth of layers. However, the 2D CNN with Taguchi parametric optimization shows potential feature learning and is robust to variable sizes of training and testing datasets. Future studies should focus on the optimal size of the input patch for the deep learning algorithms development of new architectures to improve the computational efficiency and diagnosis of human diseases.

Author Contributions

C.-J.L. and S.-Y.J. conceived and designed the experiments; M.-K.C. performed the experiments; C.-J.L., S.-Y.J. and M.-K.C. analyzed the data; C.-J.L. and S.-Y.J. wrote the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of the Republic of China, grant number MOST 108-2221-E-167 -026.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [Green Version]
  2. Togaçar, M.; Ergen, B.; Cömert, Z. Detection of lung cancer on chest CT images using minimum redundancy maximum relevance feature selection method with convolutional neural networks. Biocybern. Biomed. Eng. 2020, 40, 23–39. [Google Scholar] [CrossRef]
  3. Zhanga, S.; Hanb, F.; Lianga, Z.; Tane, J.; Caoa, W.; Gaoa, Y.; Pomeroyc, M.; Ng, K.; Hou, W. An investigation of CNN models for differentiating malignant from benign lesions using small pathologically proven datasets. Comput. Med Imaging Graph. 2019, 77. [Google Scholar] [CrossRef]
  4. Liu, H.; Zhang, S.; Jiang, X.; Zhang, T.; Huang, H.; Ge, F.; Zhao, L.; Li, X.; Hu, X.; Han, J.; et al. The cerebral cortex is bisectionally segregated into two fundamentally different functional units of gyri and sulci. Cereb. Cortex 2019, 29, 4238–4252. [Google Scholar] [CrossRef]
  5. Zhang, S.; Liu, H.; Huang, H.; Zhao, Y.; Jiang, X.; Bowers, B.; Guo, L.; Hu, X.; Sanchez, M.; Liu, T. Deep learning models unveiled functional difference between cortical gyri and sulci. IEEE Trans. Biomed. Eng. 2019, 66, 1297–1308. [Google Scholar] [CrossRef]
  6. Liu, Y.; Wang, H.; Gu, Y.; Lv, X. Image classification toward lung cancer recognition by learning deep quality model. J. Vis. Commun. Image Represent. 2019, 78. [Google Scholar] [CrossRef]
  7. Bonavita, I.; Rafael-Palou, X.; Ceresa, M.; Piella, G.; Ribas, V.; Ballester, M.A.G. Integration of convolutional neural networks for pulmonary nodule malignancy assessment in a lung cancer classification pipeline. Comput. Methods Programs Biomed. 2020, 185. [Google Scholar] [CrossRef] [Green Version]
  8. Dolz, J.; Desrosiers, C.; Wang, L.; Yuan, J.; Shen, D.; Ayed, I.B. Deep CNN ensembles and suggestive annotations for infant brain MRI segmentation. Comput. Med Imaging Graph. 2020, 79. [Google Scholar] [CrossRef]
  9. Teramoto, A.; Tsukamoto, T.; Kiriyama, Y.; Fujita, H. Automated classification of lung Cancer types from cytological images using deep convolutional neural networks. BioMed Res. Int. 2017, 2017. [Google Scholar] [CrossRef]
  10. Trigueros, D.S.; Meng, L.; Hartnett, M. Enhancing convolutional neural networks for face recognition with occlusion maps and batch triplet loss. Image Vis. Comput. 2018, 79, 99–108. [Google Scholar] [CrossRef] [Green Version]
  11. Giménez, M.; Palanca, J.; Botti, V. Semantic-based padding in convolutional neural networks for improving the performance in natural language processing. A case of study in sentiment analysis. Neurocomputing 2020, 378, 315–323. [Google Scholar] [CrossRef]
  12. Suresh, S.; Mohan, S. NROI based feature learning for automated tumor stage classification of pulmonary lung nodules using deep convolutional neural networks. J. King Saud Univ. Comput. Inf. Sci. 2019. [Google Scholar] [CrossRef]
  13. Yunus, M.; Alsoufi, M.S. A Statistical Analysis of Joint Strength of dissimilar Aluminium Alloys Formed by Friction Stir Welding using Taguchi Design Approach, ANOVA for the Optimization of Process Parameters. IMPACT Int. J. Res. Eng.Technol. 2015, 3, 63–70. [Google Scholar]
  14. Canel, T.; Zeren, M.; Sınmazçelik, T. Laser parameters optimization of surface treating of Al 6082-T6 with Taguchi method. Opt. Laser Technol. 2019, 120. [Google Scholar] [CrossRef]
  15. Özakın, A.N.; Kaya, F. Experimental thermodynamic analysis of air-based PVT system using fins indifferent materials: Optimization of control parameters by Taguchi method and ANOVA. Sol. Energy 2020, 197, 199–211. [Google Scholar] [CrossRef]
  16. Özel, S.; Vural, E.; Binici, M. Optimization of the effect of thermal barrier coating (TBC) on diesel engine performance by Taguchi method. Fuel 2020, 263. [Google Scholar] [CrossRef]
  17. Rezania, A.; Atouei, S.A.; Rosendahl, L. Critical parameters in integration of thermoelectric generators and phase change materials by numerical and Taguchi methods. Mater. Today Energy 2020, 16. [Google Scholar] [CrossRef]
  18. Idris, F.N.; Nadzir, M.M.; Shukor, S.R.A. Optimization of solvent-free microwave extraction of Centella asiatica using Taguchi method. J. Environ. Chem. Eng. 2020. [Google Scholar] [CrossRef]
  19. Armato, S.G., III; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Clarke, L.P. Data from LIDC-IDRI. Cancer Imaging Arch. 2015. [Google Scholar] [CrossRef]
  20. Armato, S.G., III; Hadjiiski, L.; Tourassi, G.D.; Drukker, K.; Giger, M.L.; Li, F.; Redmond, G.; Farahani, K.; Kirby, J.S.; Clarke, L.P. SPIE-AAPM-NCI Lung Nodule Classification Challenge Dataset. Cancer Imaging Arch. 2015. [Google Scholar] [CrossRef]
  21. Eun, H.; Kim, D.; Jung, C.; Kim, C. Single-view 2D CNNs with fully automatic non-nodule categorization for false positive reduction in pulmonary nodule detection. Comput. Methods Programs Biomed. 2018, 165, 215–224. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Masood, A.; Sheng, B.; Li, P.; Hou, X.; Wei, X.; Qin, J.; Feng, D. Computer-Assisted Decision Support System in Pulmonary Cancer detection and stage classification on CT images. J. Biomed. Inform. 2018, 79, 117–128. [Google Scholar] [CrossRef] [PubMed]
  23. Nithila, E.E.; Kumar, S.S. Automatic detection of solitary pulmonary nodules using swarm intelligence optimized neural networks on CT images. Eng. Sci. Technol. Int. J. 2017, 20, 1192–1202. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed computer assisted detection (CAD) system.
Figure 1. The framework of the proposed computer assisted detection (CAD) system.
Applsci 10 02591 g001
Figure 2. Sample benign tumor computed tomography (CT) images from the Lung Image Database Consortium and Infectious Disease Research Institute (LIDC-IDRI) dataset.
Figure 2. Sample benign tumor computed tomography (CT) images from the Lung Image Database Consortium and Infectious Disease Research Institute (LIDC-IDRI) dataset.
Applsci 10 02591 g002
Figure 3. Sample malignant tumor CT images from the LIDC-IDRI dataset.
Figure 3. Sample malignant tumor CT images from the LIDC-IDRI dataset.
Applsci 10 02591 g003
Figure 4. Sample CT images of lungs from the International Society for Optics and Photonics with the support of the American Association of Physicists in Medicine (SPIE-AAPM) dataset.
Figure 4. Sample CT images of lungs from the International Society for Optics and Photonics with the support of the American Association of Physicists in Medicine (SPIE-AAPM) dataset.
Applsci 10 02591 g004
Figure 5. Architecture of the proposed CAD system.
Figure 5. Architecture of the proposed CAD system.
Applsci 10 02591 g005
Figure 6. Response graph of S/N ratio values of factor levels on the LIDC-IDRI dataset.
Figure 6. Response graph of S/N ratio values of factor levels on the LIDC-IDRI dataset.
Applsci 10 02591 g006
Figure 7. Response graph of S/N ratio values of factor levels on the SPIE-AAPM dataset.
Figure 7. Response graph of S/N ratio values of factor levels on the SPIE-AAPM dataset.
Applsci 10 02591 g007
Table 1. Levels of control factors.
Table 1. Levels of control factors.
ColumnsAbbreviationsFactorsLevel 1 Level 2 Level 3
AC1_Sconv1_Stride12
BC1_Pconv1_padding01
CC2_Sconv2_Stride12
DC2_Pconv2_padding01
EC1_KSconv1_Kernel size357
FC1_Fconv1_Filter4612
GC2_KSconv2_Kernel size357
HC2_Fconv2_Filter81632
Table 2. L36 orthogonal array (OA) for experiments the parameters setting.
Table 2. L36 orthogonal array (OA) for experiments the parameters setting.
Exp. NoFactor
C1_S
A
C1_P
B
C2_S
C
C2_P
D
C1_KS
E
C1_F
F
C2_KS
G
C2_F
H
110103438
2101056516
31010712732
410103438
5101056516
61010712732
7102134532
810215678
91021712316
10111134716
11111156332
12111171258
1311203678
141120512316
15112074532
16112136716
171121512332
1811217458
19202136332
20202151258
21202174716
22202036532
23202051278
24202074316
25201131258
26201154716
27201176332
282120312516
29212054732
3021207638
312111312732
3221115438
33211176516
342110312316
35211054532
3621107678
Table 3. The proposed 2D convolutional neural network (CNN) architecture.
Table 3. The proposed 2D convolutional neural network (CNN) architecture.
LayerImage SizeKernel SizeStridePaddingFilter
Input50 × 50 × 3
Convolution Layer 1 5 × 5106
Relu Layer
MaxPooling Layer1 2 × 22 × 2
Convolution Layer 2 5 × 51016
Relu Layer
MaxPooling Layer2 2 × 22 × 2
FullyConnectedLayer 1 × 110120
FullyConnectedLayer
SoftmaxLayer
Classification
Table 4. The signal-to-noise (S/N) ratios of accuracy for different combinations of factors and levels.
Table 4. The signal-to-noise (S/N) ratios of accuracy for different combinations of factors and levels.
Run#FactorResult
A C1_SB C1_PC C2_SD C2_PE C1_KSF C1_FG C2_KSH C2_FY1 (%)Y2 (%)Y3 (%)Yave (%)S/N (Y)
1 1010343897.1990.4593.0693.57−0.498
2 10105651687.8493.9795.1092.30−0.725
3 101071273294.7088.8589.6691.07−0.741
4 1010343894.5896.8095.0095.46−0.405
5 10105651695.1088.5992.1991.96−0.739
6 101071273291.0692.9694.1992.74−0.657
7 10213453297.4797.7796.6697.30−0.238
8 1021567890.2186.6794.7090.53−0.882
9 102171231692.5192.2190.0491.59−0.765
10 11113471696.5892.2596.4695.10−0.443
11 11115633298.1097.8898.6898.22−0.156
12 11117125888.6580.9885.0584.89−1.440
13 1120367886.6991.2491.0089.64−0.957
14 112051231688.0892.3392.0790.83−0.842
15 11207453297.2197.2792.5795.68−0.390
16 11213671691.5093.7798.0094.42−0.509
17 112151233298.6493.7199.4397.26−0.250
18 1121745888.4777.2883.0282.92−1.666
19 20213633295.8396.6696.1496.21−0.336
20 20215125883.7984.4280.1582.79−1.648
21 20217471693.3896.3292.5394.08−0.534
22 20203653295.6794.8696.2695.60−0.392
23 20205127885.1789.6683.6986.17−1.304
24 20207431681.0685.2981.2882.54−1.673
25 20113125892.7691.2492.6592.22−0.705
26 20115471692.5192.0992.0392.21−0.705
27 20117633296.1196.7294.8895.90−0.364
28 212031251690.7987.0585.6987.84−1.134
29 21205473296.4691.6894.8094.31−0.514
30 2120763880.4177.0977.3678.29−2.131
31 211131273297.8698.4499.1398.48−0.134
32 2111543893.1095.0892.5193.56−0.580
33 21117651689.5881.9981.2684.28−1.511
34 211031231694.7092.0391.1292.62−0.670
35 21105453296.8497.8192.7895.81−0.379
36 2110767882.7087.1186.3285.38−1.380
Table 5. Mean responses of the S/N ratios for each level and optimal parameter for the LIDC-IDRI dataset.
Table 5. Mean responses of the S/N ratios for each level and optimal parameter for the LIDC-IDRI dataset.
LevelFactors
A
C1_S
B
C1_P
C
C2_S
D
C2_P
E
C1_KS
F
C1_F
G
C2_KS
H
C2_F
1−0.7002−0.7672−0.6953−0.9153−0.5467−0.6927−0.7514−1.199
2−0.8939−0.8381−0.898−0.7147−0.7258−0.8493−0.9297−0.8646
3 −1.1451−0.8756−0.7365−0.354
Delta0.19380.07080.20280.20060.59850.18290.19330.845
Rank58342761
Best level 11121133
Optimal parameter1 0 1 134 7 32
Table 6. ANOVA for the S/N ratios of the accuracy on the LIDC-IDRI dataset.
Table 6. ANOVA for the S/N ratios of the accuracy on the LIDC-IDRI dataset.
Source Degree of Freedom (DF)Sum of Squares (SS)Mean Squares (MS)Ffactorp-Value
A (C1_S)10.349890.349894.460.047
B (C1_P)10.077220.077220.980.333
C (C2_S)10.37760.37764.810.040
D (C2_P)10.171790.171792.190.155
E (C1_KS)21.604370.8021810.230.001
F (C1_F)20.18530.092651.180.327
G (C2_KS)20.274290.137141.750.20
H (C2_F)23.35451.6772521.380
Error201.568970.07845
Total328.5957
Table 7. Recognition results of different techniques on the LIDC-IDRI dataset.
Table 7. Recognition results of different techniques on the LIDC-IDRI dataset.
Methods Accuracy (%)Sensitivity (%)Specificity (%)False Positive Rate (FPR) (%)
Masood et al. [22]86.0283.9189.3210.68
Togaçar et al. [2]99.5199.7199.710.29
2D CNN with Taguchi parametric optimization98.8399.9799.930.06
Table 8. Mean responses for S/N ratios for each level and optimal parameter on the SPIE-AAPM dataset.
Table 8. Mean responses for S/N ratios for each level and optimal parameter on the SPIE-AAPM dataset.
LevelFactors
A
C1_S
B
C1_P
C
C2_S
D
C2_P
E
C1_KS
F
C1_F
G
C2_KS
H
C2_F
1−0.1761−0.45716−0.33985−0.69967−0.16709−0.30776−0.47113−0.83316
2−0.68205−0.44783−0.5456−0.24574−0.41333−0.60539−0.38385−0.44508
3 −0.77581−0.44307−0.50125−0.07798
Delta0.505950.009330.205750.453930.608720.297630.11740.75518
Rank38642571
Best level 1 2 1 2 1 1 2 3
Optimal parameter1 1 1 1 3 4 5 32
Table 9. ANOVA for the S/N ratios of the accuracy on the SPIE-AAPM dataset.
Table 9. ANOVA for the S/N ratios of the accuracy on the SPIE-AAPM dataset.
SourceDegree of Freedom (DF)Sum of Squares (SS)Mean Squares (MS)Ffactorp-Value
A (C1_S)11.907961.907968.010.01
B (C1_P)10.004450.004450.020.893
C (C2_S)10.378770.378771.590.222
D (C2_P)11.139521.139524.790.041
E (C1_KS)21.688890.844453.550.048
F (C1_F)20.421570.210780.890.428
G (C2_KS)20.058620.029310.120.885
H (C2_F)22.539831.269915.330.014
Error234.761090.23805
Total3512.9007
Table 10. Recognition results of different techniques on the SPIE-AAPM dataset.
Table 10. Recognition results of different techniques on the SPIE-AAPM dataset.
Methods Accuracy (%)Sensitivity (%)Specificity (%)False Positive Rate (FPR) (%)
Masood et al. [22]84.8781.2282.9717.03
Nithila and Kumar [23]97.2010094.45.60
2D CNN with Taguchi parametric optimization99.9799.9499.940.06
Table 11. 2D CNNs architecture parameters setting comparison on the LIDC-IDRI dataset.
Table 11. 2D CNNs architecture parameters setting comparison on the LIDC-IDRI dataset.
Original 2D CNN2D CNN with Taguchi Parametric Optimization
conv1_Kernel size53
conv1_Filter64
conv1_Stride11
conv1_padding00
conv2_Kernel size57
conv2_Filter1632
conv2_Stride11
conv2_padding01
Accuracy91.97%98.83%
Table 12. 2D CNNs architecture parameters setting comparison on the SPIE-AAPM dataset.
Table 12. 2D CNNs architecture parameters setting comparison on the SPIE-AAPM dataset.
Original 2D CNN2D CNN with Taguchi Parametric Optimization
conv1_Kernel size53
conv1_Filter64
conv1_Stride11
conv1_padding01
conv2_Kernel size55
conv2_Filter1632
conv2_Stride11
conv2_padding01
Accuracy94.68%99.97%

Share and Cite

MDPI and ACS Style

Lin, C.-J.; Jeng, S.-Y.; Chen, M.-K. Using 2D CNN with Taguchi Parametric Optimization for Lung Cancer Recognition from CT Images. Appl. Sci. 2020, 10, 2591. https://doi.org/10.3390/app10072591

AMA Style

Lin C-J, Jeng S-Y, Chen M-K. Using 2D CNN with Taguchi Parametric Optimization for Lung Cancer Recognition from CT Images. Applied Sciences. 2020; 10(7):2591. https://doi.org/10.3390/app10072591

Chicago/Turabian Style

Lin, Cheng-Jian, Shiou-Yun Jeng, and Mei-Kuei Chen. 2020. "Using 2D CNN with Taguchi Parametric Optimization for Lung Cancer Recognition from CT Images" Applied Sciences 10, no. 7: 2591. https://doi.org/10.3390/app10072591

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop