Next Article in Journal
Polarization-Charge Inversion at Al2O3/GaN Interfaces through Post-Deposition Annealing
Next Article in Special Issue
Posit Arithmetic Hardware Implementations with The Minimum Cost Divider and SquareRoot
Previous Article in Journal
Three-Dimensional Time-Harmonic Electromagnetic Scattering Problems from Bianisotropic Materials and Metamaterials: Reference Solutions Provided by Converging Finite Element Approximations
Previous Article in Special Issue
A Segmentation Enhancement Method for the Low-Contrast and Narrow-Banded Substances in CBCT Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lung Nodule Classification Using Taguchi-Based Convolutional Neural Networks for Computer Tomography Images

1
Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung 411, Taiwan
2
College of Intelligence, National Taichung University of Science and Technology, Taichung 404, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(7), 1066; https://doi.org/10.3390/electronics9071066
Submission received: 3 June 2020 / Revised: 17 June 2020 / Accepted: 29 June 2020 / Published: 29 June 2020
(This article belongs to the Special Issue Application of Electronic Devices on Intelligent System)

Abstract

:
Lung cancer occurs in the lungs, trachea, or bronchi. This cancer is often caused by malignant nodules. These cancer cells spread uncontrollably to other organs of the body and pose a threat to life. An accurate assessment of disease severity is critical to determining the optimal treatment approach. In this study, a Taguchi-based convolutional neural network (CNN) was proposed for classifying nodules into malignant or benign. For setting parameters in a CNN, most users adopt trial and error to determine structural parameters. This study used the Taguchi method for selecting preliminary factors. The orthogonal table design is used in the Taguchi method. The final optimal parameter combination was determined, as were the most significant parameters. To verify the proposed method, the lung image database consortium data set from the National Cancer Institute was used for analysis. The database contains a total of 16,471 images, including 11,139 malignant nodule images. The experimental results demonstrated that the proposed method with the optimal parameter combination obtained an accuracy of 99.6%.

1. Introduction

Lung cancer is the commonest form of cancer with the highest death rate both in developed and developing countries. In Taiwan, it has ranked first in terms of cancer mortality for more than five years. Female lung cancer is the number one cause of cancer-related death among women in Taiwan. Lung cancer is the leading cause of death in the United States and East Asia [1,2]. The smoking behaviors and environmental factors that cause lung cancer have been extensively studied [3]. Lung cancer is asymptomatic in the initial stage; therefore, approximately 70% to 75% of most newly diagnosed lung cancers are terminal. Traditionally, chest x-rays, computed tomography (CT), and magnetic resonance imaging have been used to physically analyze tissues to diagnose lung cancer [4,5]. Systemic therapy for these patients includes chemotherapy or targeted therapy. Early detection of lung cancer is critical in reducing misdiagnosis by physicians. Because of rapid advancements in computer software, the application of computer aided detection (CAD) software has become common. CAD assists imaging specialists to determine, identify, and evaluate lung lesions and nodules in digital CT images. In particular, the discovery of small nodules between 5 and 15 mm in diameter allows the early detection and treatment of lung diseases.
The current wave of artificial intelligence technology has achieved several breakthroughs, such as recognition, generation, and analysis of images [6,7,8], and speech [9], in the natural user interfaces. Machines have features that enable humans to interact with users. Modern devices not only have more affinity, but also a more reasonable and intelligent judgment and response to the surrounding environment. Such products have large development potential, including driverless cars, drones, smart homes, manufacturing robots, and chat robots. Convolutional neural networks (CNN) are a major development in deep learning. CNNs are capable of classifying images and are widely applied in object detection tasks [10,11,12]. Furthermore, they have been continuously developed and improved. The handwritten font recognition model LeNet-5 [13] was developed in 1994 and is one of the earliest CNNs. LeCun et al. compared the multilayer back propagation network and CNN. The excellent design of LeNet5 helps extract features using operations such as convolution, parameter sharing, and pooling, thus reducing computational costs. Finally, the fully connected neural network is used for classification and recognition. This network is the basis for several neural network architectures. Developed in 2012, AlexNet [14] won the Imagenet competition. ReLU, instead of traditional activation functions, is used in AlexNet. However, the GPU memory limits the network size, Therefore, AlexNet was extended on two GPUs, which were trained in parallel. However, none of the aforementioned architectures has a rapid method of determining optimal parameters. Therefore, in this study, we used the Taguchi method to determine optimal parameters.
Experiments should be conducted to obtain sufficient relevant data to determine the mechanism of the observed phenomena. Most researchers use the trial and error approach to determine optimal parameters. Trial and error is a type of learning in which various responses in a situation are attempted, seemingly at random, until one response achieves the relevant goal. Through successive trials, the successful response is strengthened and occurs earlier. However, this method is time consuming. Therefore, an efficient method should be used to obtain experimental data. Taguchi [15,16] developed a method based on the orthogonal array (OA) experiment, which considerably reduces variance for the experiment by optimal setting control parameters. Therefore, a combination of experimental design and control parameter optimization is used in the Taguchi method to obtain the best results. An OA provides a well-balanced set of experiments. The Taguchi signal-to-noise ratio (S/N) is a logarithmic function of the desired output. It is an optimized objective function, which is helpful for data analysis and the prediction of the optimal results. The Taguchi method is a scientifically standardized mechanism for evaluating and implementing improvements in products, processes, materials, equipment, and facilities. These improvements are designed to optimize the program to produce the best results and simultaneously reducing the number of defects. In recent years, several researchers [17,18] have used the evolutionary computation methods to optimize parameters of CNN. Esteban et al. [17] used evolutionary techniques at unprecedented scales to discover models and DNA-based coding schemes for CIFAR-10 and CIFAR-100 datasets. Ma et al. [18] proposed an autonomous and continuous learning (ACL) algorithm that can automatically generate a deep convolutional neural network (DCNN) architecture for each given visual task. They divided the DCNN into multiple stacked meta-convolutional blocks and fully connected blocks, each of which may contain convolution, pooling, full connection, batch normalization, activation and exit operations to convert the architecture to integer code. According to the above-mentioned methods, these methods can obtain good accuracy. However, they take more time to determine the architecture parameters [17,18]. We used the CNN-based Taguchi method to classify lung nodules in CT images. Our objective was to improve the CNN accuracy and decrease the experimental time.
The remainder of the paper is organized as follows. Section 2 introduces the relevant CNN based Taguchi method. In Section 3, the experimental results of the proposed method are described. Section 4 presents conclusions.

2. CNN-Based Taguchi Method

This section details how a Taguchi-based CNN is used for detecting gender from an image data set. The flow chart of the proposed method is presented in Figure 1.

2.1. CNNs

CNNs consists of one or more convolutional and fully connected layers, and pooling layers. Compared with other deep learning structures, CNNs can provide better results in image and speech recognition. This model can also be trained using back-propagation algorithms. AlexNet was used in this study [14].
The AlexNet network structure has eight layers. The first five layers are convolutional layers (C), the next three layers are fully connected (FC) layers, and the output of the last fully connected layer is passed to a 1000-channel softmax layer, which corresponds to the distribution of 1000 class labels. Lung nodules are classified using two labels, namely benign and malignant. Therefore, we changed the last FC layer to a two-channel softmax layer, which corresponded to the distribution of the two-class label. Table 1 lists the parameters of show the AlexNet architecture.

2.2. Taguchi Method

Statistical methods are used in quality engineering to improve product quality and reduce costs. These methods are also used in biology, marketing, and advertising. Professional statisticians concur with the goals and advancements of quality engineering. The most substantial advantage of the Taguchi method is that useful information can be obtained with fewer experiments. Although it is not as favorable as the full factor method for determining the exact optimization position, it requires fewer trends to determine the optimization trend. The feasibility is considerably greater than the full factor method. The Taguchi method has the following characteristics: (1) quality characteristics based on the quality loss function, (2) definition and selection of experimental factors, (3) S/N ratio, and (4) the Taguchi OA.
In the full factor method, when the control factor and number of levels increase, the number of experiments increases exponentially. Therefore, partial factorial design increases the complexity of the experimental method. If an OA is used, fewer experiments are required to obtain a reliable factor effect. The use of OA to conduct experiments is key to Taguchi quality engineering. Before the OA is constructed, the number of factors that affect the quality characteristics, the level of each factor, and the difficulties that may occur in the experiment should be determined. Then, an appropriate OA should be selected, and the factors should be preconfigured before experiments are conducted. An appropriate OA is configured for the selected control factor for experiments. Finally, the S/N ratio is calculated based on the experimental observations obtained by the experiment, and the best factor is analyzed using the S/N ratio.
The analysis of experimental data is based on the S/N ratio created by Dr. Taguchi. The quality characteristics can be divided into the larger the better (LTB), smaller the better, and nominal the best. This paper presents a static parameter design. A larger quality characteristic and higher experimental is preferred. We used the following LTB formula:
LTB   S / N   ratio   = 10 log 10 1 T i = 1 T ( 1 X i 2 )
where T is the number of experiments, and X i is the i t h experimental result.

3. Experimental Results

3.1. Dataset

The lung image database consortium image collection includes diagnostic and lung cancer screening chest CT scans with annotated lesions. Eight medical imaging companies and seven academic centers collaborated to create this data set which contains 245,931 images (CT scans and X-ray images). For each participant, images from CT scans were obtained with an associated XML file that recorded the results of a two-phase image annotation process performed by four experienced chest radiologists. In the initial blinded-read phase, each radiologist independently reviewed each CT scan and marked lesions belonging to one of three categories “nodules ≥ 3 mm in diameter,” “nodules < 3 mm in diameter,” and “non-nodules ≥ 3 mm in diameter” (see Figure 2).
In this data set nodules are classified as either benign or malignant. We classified 16,471 images as benign and 11,139 images as malignant. Figure 3 shows the malignant and benign lung nodules of computed tomography (CT).

3.2. Taguchi Based CNNs

3.2.1. Defining Control Levels and Factors

First, we defined the three levels and six factors. Factors included kernel size, stride, and padding for first and fifth convolutional layers. Table 2 presents the control levels and factors.

3.2.2. OA

Experiments were performed according to the OA. Here, there were five with three levels and one with two levels. The degree of freedom was 15. We selected the L18 OA. Table 3 presents the L18 OA. The factors C1 Kernel Size, C1 Stride, C1 Padding, C5 Kernel Size, C5 Stride and C5 Padding are shown in Table 3. The initially selected factors and levels required 2 1 × 3 5 = 486 experiments, while only 18 experiments were required after using the Taguchi experiment.
Table 4 presents the test accuracy and S/N ratio. Finally, we analyzed the best levels and rank significance factors. Table 5 presents the best level prediction.

3.2.3. Confirmation Experiment

The best factors that were determined in this experiment were as follows: C1 Kernel Size = 9, C1 Stride = 4, C1 Padding = 2, C5 Kernel Size = 5, C5 Stride = 1, C5 Padding = 2. These best factors were used to perform two confirmation experiments.
To verify that these best factors were the optimal combination, each experiment was performed five times, as depicted in Table 6 and Table 7. Table 6 presents the experimental results of the original AlexNet parameters, and Table 7 presents the experimental results of the best combination. The accuracy of the best combination improved by 2.764%. Experimental results verified the accuracy of the original AlexNet was 96.836%. After using the Taguchi method to determine the architecture parameters of AlexNet, the accuracy was increased to 99.6%. It can be proved that the parameter optimization of Taguchi method improves the accuracy of AlexNet for classifying nodules into malignant or benign. The proposed method assists radiologists in diagnosing lung nodules more accurately and is an excellent CAD.

3.3. Comparison Results of Training Times

In this study, 5 parameters with 3 levels and 1 parameter with 2 levels in AlexNet were chosen, the training times using the Taguchi-based CNN and the trial and error method were 4.5 h, and 5 days and 1.5 h, respectively. Table 8 presents the comparison results of training times using the Taguchi-based CNN and the trial and error method. Recently, Ma et al. [18] proposed an autonomous and continuous learning (ACL) algorithm to determine a DCNN structure for solving classification problems. The MNIST, Fashion-MNIST, EMNIST-Digits, EMNIST-Letters, CIFAR10 and CIFAR100 datasets were used, the average training time of each generation took about 3, 5, 18, 12, 11, and 11 GPU days. Table 9 shows the training times using the ACL algorithm for various databases. It took more training time than the Taguchi-based CNN method. In addition, an ACL algorithm [18] divided the DCNN into multiple stacked meta-convolutional blocks and fully connected blocks, each of which might contain convolution, pooling, full connection, batch normalization, activation and exit operations to convert the architecture to integer code. According to an ACL algorithm, the parameter optimization of the convolution layers was also used in this study.

4. Conclusions

In radiology, it is important to improve the accuracy of lung nodule classification. An excellent CAD can not only assist the radiologist, but also reduce the burden on the radiologist. In this study, we used the Taguchi method to design experiments and improve the classification accuracy of AlexNet. Compared with the trial and error method, the proposed method saves substantial experimental time. The selected five parameters with three levels and one parameter with two levels in AlexNet require 486 experiments using the trial and error method, while only 18 experiments are required using the Taguchi method. The S/N ratio reaction table is used to determine the response of the factor to the system at this level, and then determine the best level. The order of the influence of each factor on the system is C5 Padding, C1 Kernel Size, C5 Stride, C1 Stride, C5 Kernel, and C1 Padding. The accuracy of the best factor using Taguchi-based CNN is improved by 2.764%. The training times using the Taguchi-based CNN and the trial and error method are 4.5 h, and 5 days and 1.5 h, respectively. Experimental results indicate that the proposed method takes less training time than other methods. In future work, the proposed method can also be applied to breast cancer, brain tumors, and liver tumors. In addition, some incremental data methods, such as Generative Adversarial Network (GAN), can be used to increase the amount of training data and improve the accuracy of classification.

Author Contributions

Conceptualization, C.-J.L.; methodology, C.-J.L. and Y.-C.L.; software, C.-J.L. and Y.-C.L.; data curation, C.-H.L.; writing—original draft preparation, C.-J.L. and Y.-C.L.; funding acquisition, C.-J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of the Republic of China, grant number MOST 108-2221-E-167-026.

Acknowledgments

The authors would like to thank the Ministry of Science and Technology of the Republic of China, Taiwan for financially supporting this research under Contract No. MOST 108-2221-E-167-026.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rebecca, L.; Siegel, K.D.; Miller, A.J. Cancer statistics, 2019. CA Cancer J. Clin. 2019, 69, 7–34. [Google Scholar] [CrossRef] [Green Version]
  2. Worrawat, E.; Jonathan, C. An overview of cancer trends in Asia. Neurocomputing 2015, 165. [Google Scholar] [CrossRef]
  3. Cornfield, J.; Haenszel, W.; Hammond, E.C.; Lilienfeld, A.M.; Shimkin, M.B.; Wynder, E.L. Smoking and lung cancer: Recent evidence and a discussion of some questions. Int. J. Epidemiol. 2009, 38, 1175–1191. [Google Scholar] [CrossRef] [PubMed]
  4. Mountain, C.F.; Dresler, C.M. Regional Lymph Node Classification for Lung Cancer Staging. CHEST 1997, 111, 1718–1723. [Google Scholar] [CrossRef] [PubMed]
  5. Mountain, C.F. Revisions in the International System for Staging Lung Cancer. Chest 1997, 111, 1710. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Kumar, A.; Kim, J.; Lyndon, D.; Fulham, M.; Feng, D. An Ensemble of Fine-Tuned Convolutional Neural Networks for Medical Image Classification. IEEE J. Biomed. Health Inform. 2017, 31–40. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Igor, Š.; Aleksej, A. Convolutional Neural Network Based Automatic Object Detection on Aerial Images. IEEE Geosci. Remote. Sens. Lett. 2016, 13, 740–744. [Google Scholar] [CrossRef]
  8. Juan, M.P.; Arturo, A.; José, M.A. Olive-Fruit Variety Classification by Means of Image Processing and Convolutional Neural Networks. IEEE Access 2019, 7, 147629–147641. [Google Scholar] [CrossRef]
  9. Hou, J.C.; Wang, S.S.; Lai, Y.H.; Tsao, Y.; Chang, H.W.; Wang, H.M. Audio-Visual Speech Enhancement Using Multimodal Deep Convolutional Neural Networks. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 2, 117–128. [Google Scholar] [CrossRef] [Green Version]
  10. Changxing, D.; Dacheng, T. Trunk-Branch Ensemble Convolutional Neural Networks for Video-Based Face Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1002–1014. [Google Scholar] [CrossRef] [Green Version]
  11. Pejman, R.; Tõnis, U. Convolutional Neural Network Super Resolution for Face Recognition in Surveillance Monitoring; Springer International Publishing: Cham, Switzerlamd, 2016; pp. 175–184. [Google Scholar] [CrossRef]
  12. Gil, L.; Tal, H. Age and gender classification using convolutional neural networks. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015. [Google Scholar] [CrossRef]
  13. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. IEEE 1998, 88, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  14. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 6. [Google Scholar] [CrossRef]
  15. Yang, W.H.; Tarng, Y.S. Design optimization of cutting parameters for turning operations based on the Taguchi method. J. Mater. Process. Technol. 1998, 84, 122–129. [Google Scholar] [CrossRef]
  16. Ballantyne, K.N.; Oorschot, R.A.; Mitchell, R.J. Reduce optimisation time and effort: Taguchi experimental design methods. Forensic Sci. Int. Genet. Suppl. Ser. 2008, 1, 7–8. [Google Scholar] [CrossRef]
  17. Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.L.; Tan, J.; Le, Q.; Kurakin, A. Large-scale evolution of image classifiers. arXiv 2017, arXiv:1703.01041. [Google Scholar]
  18. Ma, B.; Li, X.; Xia, Y.; Zhang, Y. Autonomous deep learning: A genetic DCNN designer for image classification. Neurocomputing 2020, 379, 152–161. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Flow chart of the proposed method.
Figure 1. Flow chart of the proposed method.
Electronics 09 01066 g001
Figure 2. Lung nodules.
Figure 2. Lung nodules.
Electronics 09 01066 g002
Figure 3. The malignant and benign lung nodules of computed tomography (CT).
Figure 3. The malignant and benign lung nodules of computed tomography (CT).
Electronics 09 01066 g003
Table 1. Parameters of the AlexNet architecture.
Table 1. Parameters of the AlexNet architecture.
LayersKernel SizeStridePadding
C111 × 1140
C25 × 512
C33 × 311
C43 × 311
C53 × 311
Neurons
FC64096
FC74096
FC81000
Table 2. Control levels and factors.
Table 2. Control levels and factors.
First Convolutional Layer (C1) Kernel SizeFirst Convolutional Layer (C1) StrideFirst Convolutional Layer (C1) PaddingFifth Convolutional Layer (C5) Kernel SizeFifth Convolutional Layer (C5) StrideFifth Convolutional Layer (C5) Padding
Level 11342732
Level 21131521
Level 392-310
Table 3. L18 orthogonal array (OA).
Table 3. L18 orthogonal array (OA).
First Convolutional Layer (C1) Kernel SizeFirst Convolutional Layer (C1) StrideFirst Convolutional Layer (C1) PaddingFifth Convolutional Layer (C5) Kernel SizeFifth Convolutional Layer (C5) StrideFifth Convolutional Layer (C5) Padding
11342732
21332521
31322310
41142721
51132510
61122332
7942530
8932322
9922711
101341311
111331730
121321522
131141512
141131331
151121720
16941320
17931712
18921531
Table 4. Accuracy of the test and signal-to-noise (S/N) ratio.
Table 4. Accuracy of the test and signal-to-noise (S/N) ratio.
NO.Test 1Test 2Test 3Test 4Test 5S/N Ratio
196.2296.1195.4396.0995.43−0.36779
291.8797.4390.5990.9191.91−0.68229
397.1195.2789.8893.6993.65−0.55338
499.8299.7499.2999.2597.96−0.06931
597.8892.6193.593.8596.68−0.46014
698.793.393.497.4991−0.4777
793.2896.0793.195.7197.01−0.44597
899.7699.7299.8498.799.31−0.04675
999.2598.2697.1798.1699.7−0.13163
1099.5597.4597.7999.6499.25−0.11164
1193.1291.6489.329391.42−0.75558
1299.3599.6499.997.2198.68−0.09241
1399.9299.5599.7499.3398.64−0.04939
1487.7492.3989.5694.4290.31−0.83863
1587.6692.7689.5290.4783.93−1.03993
1692.8296.1192.1790.9197.65−0.55308
1798.8598.4698.199.3798.6−0.11601
1899.8899.9499.8499.899.88−0.01148
Table 5. Best level prediction.
Table 5. Best level prediction.
Factors LevelsC1 Kernel SizeC1 StrideC1 PaddingC5 Kernel SizeC5 StrideC5 Padding
1−0.427182−0.2662−0.359439−0.41337−0.48286−0.19167
2−0.489182−0.48323−0.396461−0.29028−0.41396−0.3075
3−0.217486−0.38442-−0.4302−0.23703−0.63468
Delta0.2716960.2170340.0370220.139920.2458250.443009
Rank246531
Best levels311231
Best factors942512
Table 6. Experimental results of the original AlexNet parameters.
Table 6. Experimental results of the original AlexNet parameters.
AlexNetTest 1Test 2Test 3Test 4Test 5
Accuracy93.02%97.29%99.62%99.53%94.72%
Sensitivity99.0%99.8%99.3%98.7%98.8%
Specificity89.7%92.0%99.8%99.9%90.5%
Average Accuracy96.836%
Table 7. Experimental results of the best combination.
Table 7. Experimental results of the best combination.
Best FactorsTest 1Test 2Test 3Test 4Test 5
Accuracy99.51%99.78%99.94%99.68%99.09%
Sensitivity99.9%99.4%99.98%99.7%99.5%
Specificity99.3%99.8%99.9%99.7%98.9%
Average Accuracy99.6%
Table 8. The comparison results of training times using various methods.
Table 8. The comparison results of training times using various methods.
DatasetLung image database
MethodTaguchi-based CNNThe trial and error method
Training Time of Each Experiment 15 min 15 min
Number of Experiments18486
Total Training Time4 h 30 min5 days 1 h 30 min
Table 9. The training times using the autonomous and continuous learning (ACL) algorithm for various databases.
Table 9. The training times using the autonomous and continuous learning (ACL) algorithm for various databases.
DatasetMNISTFashion MNISTEMNIST DigitEMNIST LetterCIFAR 10CIFAR 100
MethodACL algorithm [18]
Number of Experiments111111
Training Time of Each Generation3 days5 days18 days12 days12 days11 days

Share and Cite

MDPI and ACS Style

Lin, C.-J.; Li, Y.-C. Lung Nodule Classification Using Taguchi-Based Convolutional Neural Networks for Computer Tomography Images. Electronics 2020, 9, 1066. https://doi.org/10.3390/electronics9071066

AMA Style

Lin C-J, Li Y-C. Lung Nodule Classification Using Taguchi-Based Convolutional Neural Networks for Computer Tomography Images. Electronics. 2020; 9(7):1066. https://doi.org/10.3390/electronics9071066

Chicago/Turabian Style

Lin, Cheng-Jian, and Yu-Chi Li. 2020. "Lung Nodule Classification Using Taguchi-Based Convolutional Neural Networks for Computer Tomography Images" Electronics 9, no. 7: 1066. https://doi.org/10.3390/electronics9071066

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop