You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

11 June 2022

Performance Analysis of State-of-the-Art CNN Architectures for LUNA16

,
,
,
,
and
1
Faculty of Computer Science & Information Technology, The Superior University, Lahore 54600, Pakistan
2
Department of Software, Gachon University, Seongnam 13120, Korea
3
John von Neumann Faculty of Informatics, Obuda University, 1034 Budapest, Hungary
4
Institute of Information Engineering, Automation and Mathematics, Slovak University of Technology in Bratislava, 81107 Bratislava, Slovakia
This article belongs to the Special Issue Sensor Data Fusion Based on Deep Learning for Computer Vision and Medical Applications

Abstract

The convolutional neural network (CNN) has become a powerful tool in machine learning (ML) that is used to solve complex problems such as image recognition, natural language processing, and video analysis. Notably, the idea of exploring convolutional neural network architecture has gained substantial attention as well as popularity. This study focuses on the intrinsic various CNN architectures: LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1, which have been scrutinized and compared with each other for the detection of lung cancer using publicly available LUNA16 datasets. Furthermore, multiple performance optimizers: root mean square propagation (RMSProp), adaptive moment estimation (Adam), and stochastic gradient descent (SGD), were applied for this comparative study. The performances of the three CNN architectures were measured for accuracy, specificity, sensitivity, positive predictive value, false omission rate, negative predictive value, and F1 score. The experimental results showed that the CNN AlexNet architecture with the SGD optimizer achieved the highest validation accuracy for CT lung cancer with an accuracy of 97.42%, misclassification rate of 2.58%, 97.58% sensitivity, 97.25% specificity, 97.58% positive predictive value, 97.25% negative predictive value, false omission rate of 2.75%, and F1 score of 97.58%. AlexNet with the SGD optimizer was the best and outperformed compared to the other state-of-the-art CNN architectures.

1. Introduction

Artificial intelligence (AI) has proven to be a significant success in every field of life [1]. Artificial intelligence is the mimic of human intelligence used by computer programs. It has a subset named the machine learning (ML) technique that helps to train algorithms in making decisions [2]. Currently, deep learning technology has become a promising approach for clinical detection systems [3]. A deep learning computer-aided diagnosis system has been used to analyze medical images, which has proven to be a remarkable advancement in various medical applications [4]. Deep learning algorithms have the potential to solve real-world complex problems, especially in image analysis and computer vision [5]. Convolutional neural network (CNN) is a deep learning technique used in image and text recognition [6]. CNN has proven to have remarkable performance in understanding image segmentation, image classification problems, and detection [7]. Medical imaging is a core standard for the early diagnosis, detection, and treatment of several diseases. Medical images are kept in Digital Imaging and Communications in Medicine (DICOM) format, which can be quickly accessed for quantitative and qualitative analysis [8].
Medical imaging modalities comprise multidisciplinary techniques to attain an accurate diagnosis of various diseases. There are various medical image modalities such as computed tomography, X-ray angiography, mammography, digital radiography, radio fluoroscopy, and computed radiography that are used to analyze image processing [9]. Computed tomography (CT) images are an easier and more accurate method for the correct detection of various diseases [10]. However, it is tedious for medical experts to manually analyze CT scans without any error [11]. In the last few years, a low-dose CT scan has been used to screen for lung cancer diseases that has led to a decrease in the mortality rate of lung cancer [12].
The authors of [13] presented a framework based on deep learning to detect lung cancer and pneumonia abnormalities. The deep learning technique, named Modified AlexNet (MAN), was presented to detect two classes, normal and pneumonia class, by using X-ray images. The MAN technique was applied to classify lung cancer by using a support vector machine (SVM) model. The MAN method performance was evaluated with other deep learning approaches such as ResNet50, VGG19, VGG16, and AlexNet and achieved 97.27% accuracy by using the LIDC-IDRI dataset. Another study [14] presented a network based on a deep convolutional neural network, named AlexNet. The network consisted of an eight-layer two-architecture deep CNN to classify malignant and benign nodules. The network extracted the features automatically from CT scan images. Binary cross-entropy was applied to improve precision in the training phase as well as validation accuracy. The lung nodule classification model achieved 99% training precision and 97% validation accuracy. Similarly, the authors of [15] introduced a lung cancer detection method based on a deep learning algorithm using the segmentation method. Five-fold cross-validation was applied to train and validate the deep learning-based model. In another study, the authors of [16] presented a CAD system for pulmonary pure ground-glass nodules based on the convolutional neural network. In line with the above, various research works have been presented where diverse optimization algorithms have been applied to achieve an improvement in performance for the detection of lung cancer diseases. In this research work, we have investigated and explored an optimization algorithm based on a convolutional neural network to detect lung cancer. Furthermore, this work exploited the optimizers RMSprop, Adam, and SGD as well as their implementation with CNN architectures.
The remaining part of the paper has been organized as follows: Section 2 discusses prior works published in recent years. Section 3 introduces the materials and methods based on CNN, and Section 4 demonstrates the experimental results. Section 5 discusses the present, and finally, limitations and future works are discussed in Section 6.

3. Materials and Methods

Lung cancer has become the main reason for cancer deaths all over the world, as its symptoms appear late. Consequently, a significant detection system is required to detect lung cancer in patients at its early stages. In this study, the existing algorithms LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1, with the Adam, SGD, and RMSprop optimizers were applied to classify lung cancer, and an example is shown in Figure 1.
Figure 1. Convolutional neural network architecture.
The overall procedure adopted to apply convolutional neural network architectures is demonstrated in Figure 1. In the first step, the LUng Nodule Analysis 2016 (LUNA-16) dataset obtains from the publicly available lung cancer nodules [44]. The LIDC-IDRI database is publicly available from The Cancer Imaging Archive (TCIA). This database contains a total number of 1018 CT scans. CT scan images are associated with XML files annotated by four experienced radiologists. Thin-slice CT scans play a significant role in the detection of pulmonary nodules; therefore, the authors of [45] discarded slices that were greater than 3 mm thick, missing slices, and those with inconsistent slice spacing. A total number of 888 scans provided as MetaImage (.mhd) images are accessible from the LUNA-16 website. Two methods are applied on LUNA16 for training and validation purposes. In the first method, a dataset is randomly split into 80% and 20%, and the second method is based on 5-fold cross-validation. After splitting and 5-fold cross-validation, the dataset is forwarded to the CNN architectures to extract the features.
CNNs have various parameters and hyperparameters such as neurons, number of layers, weights, biases, stride, filter size, activation function, learning rate, etc. Convolution operation plays a significant role in image feature extraction [46]. Two types of filters, large size filters and small size filters, are used to extract various information. Large-sized filters are used to extract coarse-grained information, and small-sized filters are used to extract fine-grained information.
LeNet was developed by LeCun in 1998 for zip code recognition. In LeNet, a convolve filter size of 5 × 5 with a stride of 1 was used, and in the subsampling (pooling) layer, a filter size of 2 × 2 was applied with a stride of 2. AlexNet was the first CNN-based method that won the ImageNet Scale Visual Recognition Challenge in 2012. AlexNet comprises 5 convolutions, 3 pooling, and 3 fully connected (FC) layers. The input image size was 227 × 227 × 3, and the rectified linear unit (ReLU) was applied for the first time in AlexNet. The visual geometric group (VGG) has two versions: VGG16 with 16 layers and VGG19 with 19 layers. In VGG16 and VGG19, the number of layers increases, but the size of the filters decreases. In this study, we applied LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 to classify lung cancer on the LUNA16 dataset.
Optimizers are methods or algorithms applied to minimize a loss function and to maximize the efficiency of the model. Optimizers are mathematical functions that are based on the model’s learnable parameters, and they assist to reduce the losses with updated learning rates and weights of the neural network. The learning rate (LR) is known as a tuning parameter that works in an optimization algorithm. LR determines the step size in an optimization algorithm at each iteration when moving toward finding a minimum of the loss function.
Root mean square propagation [47] is also an adaptive learning method to resolve destructive learning rates. RMSprop determines the learning rate after each iteration by using an exponentially weighted average [41].
q t = q t 1   + ( 1 ϒ ) × p t 2
Δ w t = q t q t + ε × p t  
w t + 1 = w t + η × Δ w t  
where
η: initial learning rate;
q t : exponential average of gradients along wj;
p t : gradient at time t along wj;
x t : exponential average of squares of gradients along wj;
ϒ: hyperparameter.
The adaptive moment estimation (Adam) method computes adaptive learning rates for every parameter at each iteration. It is easy to implement with less memory requirements, and it is computationally efficient. Adam uses a combination of RMSprop and gradient descent with momentum to determine the parameter values [41].
q t = ϒ 1 q t 1 ( 1 ϒ 1 ) × p t
s t = ϒ 2 s t 1 ( 1 ϒ 2 ) × p t
Δ w t = p t q t x t + ε × p t
w t + 1 = w t + η × Δ w t
where
Ƞ: initial learning rate;
q t : exponential average of gradients along wj;
p t : gradient at time t along wj;
x t : exponential average of squares of gradients along wj;
ϒ 1 and ϒ 2 : hyperparameters.
A stochastic gradient descent (SGD) focuses on performing updates to the model parameters one at a time; therefore, it is much faster. After each iteration, the cost function minimizes, and SGD performs frequent updates of the model parameter that causes the cost function to fluctuate heavily, which leads the gradient to jump to the global minimum. It requires less memory and permits the use of enormous datasets [41].
f ( x ) = 1 n × i = 0 n f i ( x )  
f ( x ) = 1 n i = 0 n f i ( x )  
x     x η f i   ( x )  
E i f ( x ) = 1 n i = 0 n f i ( x ) = f ( x )  
where
η: initial learning rate;
x: parameters;
n: training dataset;
f i ( x ) : loss function;
f i   ( x ) : stochastic gradient.
In this study, CNN architectures with LeNet, AlexNet, VGG16, ResNet-50, Inception-V1 with the Adam, RMSprop, and SGD optimizers were applied to extract features from the LUNA16 dataset. Subsequently, the flattened layer converts the matrix to vector form and is fed to the fully connected layer for classification purposes. Finally, the classifier softmax classifies lung cancer into benign and malignant.

4. Results and Discussion

In this study, the comparison between state-of-the-art CNN architectures LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 for the detection of lung cancer is explained in detail by using the different optimizers, Adam, RMSprop, and SGD. The performances of the LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 architectures were validated and evaluated in terms of accuracy. These deep learning networks use binary cross-entropy loss. In this section, the results presented were achieved by LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 detection algorithms.
In this study, the performance analysis was implemented in the Keras tool using Python 3.8. The Adam, SGD, and RMSprop optimizers were employed with a learning rate of 0.001, batch size of 20, and 200 epoch values. To measure the performance of LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1, various optimizers such as RMSprop, Adam, and SGD were applied, and various statistical parameters were measured to detect lung cancer.
Various statistical parameters such as accuracy [48], sensitivity [49], specificity [50], positive predictive value [51], negative predictive value [52], false omission rate [53], and F1 score were applied to evaluate the performance of the convolutional neural network architectures with the optimizers.
A c c u r a c y = ( T N + T P ) ( T N + F N + F P + T P ) × 100 %  
S e n s i t i v i t y = T P ( T P + F N ) × 100 %  
S p e c i f i c i t y = T N ( T N + F P ) × 100 %  
P o s i t i v e   p r e d i c t i v e   v a l u e   ( P P V ) = T P ( T P + F P ) × 100 %  
N e g a t i v e   p r e d i c t i v e   v a l u e   ( N P V ) = T P ( T P + F P ) × 100 %  
The validation confusion matrix of the LeNet architecture is shown in Table 2, and the validation performance factors of the LeNet model with different optimizers are shown in Table 3. LeNet with the SGD optimizer achieved a 95.92% accuracy, which is higher than the RMSprop and Adam optimizers. The other statistical parameters of LeNet with the SGD optimizer such as accuracy, sensitivity, specificity, PPV, NPV, FOR, and F1-score were 95.92%, 94.76%, 97.25%, 97.51%, 94.22%, 5.78%, and 96.11%, respectively.
Table 2. LeNet confusion matrix (validation).
Table 3. Validation statistical analysis of the LeNet model.
The validation confusion matrix of AlexNet is shown in Table 4, and the validation performance of the AlexNet model with different optimizers is shown in Table 5. AlexNet with the SGD optimizer achieved 97.42% accuracy, which was higher than the RMSprop and Adam optimizers. The other statistical parameters of AlexNet with the SGD optimizer such as sensitivity, specificity, PPV, NPV, FOR, and F1-score were 97.58%, 97.25%, 97.58%, 97.25%, 2.75%, and 97.58%, respectively.
Table 4. AlexNet confusion matrix (validation).
Table 5. Validation statistical analysis of the AlexNet model.
The VGG16 validation confusion matrix is shown in Table 6, and the validation performance of the VGG16 model with different optimizers is shown in Table 7. VGG16 with the SGD optimizer achieved 93.56% accuracy, which was higher than the RMSprop and Adam optimizers. The other statistical parameters of VGG16 with the SGD optimizer such as sensitivity, specificity, PPV, NPV, FOR, and F1-score were 91.53%, 95.87%, 96.19%, 90.87%, 9.13%, and 93.80%, respectively.
Table 6. VGG16 confusion matrix (validation).
Table 7. Validation Statistical Analysis of the VGG16 model.
The ResNet 50 validation confusion matrix is shown in Table 8, and the validation performance of the ResNet 50 model with different optimizers is shown in Table 9. ResNet 50 with the SGD optimizer achieved 96.35% accuracy, which was higher than the RMSprop and Adam optimizers. The other statistical parameters of ResNet 50 with the SGD optimizer such as sensitivity, specificity, PPV, NPV, FOR, and F1-score were 91.53%, 95.87%, 96.19%, 90.87%, 9.13%, and 93.80%, respectively.
Table 8. ResNet 50 confusion matrix (validation).
Table 9. Validation statistical analysis of the ResNet 50 model.
The Inception-V1 validation confusion matrix is shown in Table 10, and the validation performance of the Inception-V1 model with different optimizers is shown in Table 11. Inception-V1 with the SGD optimizer achieved 93.56% accuracy, which was higher than the RMSprop and Adam optimizers. The other statistical parameters of Inception-V1 with the SGD optimizer such as sensitivity, specificity, PPV, NPV, FOR, and F1-score were 91.53%, 95.87%, 96.19%, 90.87%, 9.13%, and 93.80%, respectively.
Table 10. Inception-V1 confusion matrix (validation).
Table 11. Validation statistical analysis of the Inception-V1 model.
Table 12 demonstrates the results obtained from AlexNet with the SGD optimizer on the Luna16 dataset. The original image was benign, and the CNN architecture AlexNet with the SGD optimizer detected the image as benign. Next, the image was benign and AlexNet with the SGD detected the image as malignant, which was wrongly predict by AlexNet. The next image was malignant, and AlexNet with the SGD optimizer detected it as benign. Finally, the actual image was malignant, and AlexNet with the SGD optimizer detected it as malignant.
Table 12. Detection results of AlexNet with the SGD optimizer on the LUNA16 dataset.
In the training phase, the accuracies of LeNet, AlexNet, VGG16 ResNet-50 and Inception-V1 with the optimizers are illustrated in Figure 2. LeNet with the SGD optimizer achieved 97.47% accuracy, whereas AlexNet with the SGD optimizer obtained 99.09% accuracy. VGG16 with the SGD optimizer achieved 94.07% accuracy. ResNet-50 with the SGD optimizer obtained 98.05% accuracy, while Inception-V1 with the SGD optimizer achieved 97.99% accuracy.
Figure 2. Performance evaluation with statistical parameters for LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 (training accuracy).
The next optimizer, Adam, was performed on LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1. LeNet with the Adam optimizer achieved 97.26% accuracy, whereas AlexNet with Adam obtained 96.13% accuracy, VGG16 with Adam achieved 93.71% accuracy, ResNet-50 with Adam achieved 97.24% accuracy, and Inception-V1 with Adam achieved 96.12% accuracy.
The last optimizer, RMSprop, also measured for training purposes. LeNet with RMSprop achieved 95.27%, AlexNet with RMSprop obtained 95.54% accuracy, VGG16 with RMSprop achieved 93.01% accuracy, ResNet-50 obtained 95.53% accuracy, and Inception-V1 achieved 94.86% accuracy using the RMSprop optimizer.
Figure 3 demonstrates the validation phase, the LeNet algorithm with the SGD optimizer achieved 95.92% validation accuracy. AlexNet with the SGD optimizer obtained 97.42% accuracy, VGG16 with the SGD optimizer obtained 93.56% validated accuracy, ResNet-50 with the SGD optimizer achieved 96.35% validation accuracy, and Inception-V1 with the SGD optimizer obtained 95.06% accuracy.
Figure 3. Performance evaluation with statistical parameters for LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 (validation accuracy).
The next optimizer, Adam, was performed on LeNet, AlexNet, and VGG16. LeNet with the Adam optimizer achieved 95.18% accuracy, whereas AlexNet with Adam obtained 95.71% accuracy, VGG16 achieved 92.70% accuracy, ResNet-50 with the Adam optimizer achieved 95.09% validation accuracy, and Inception-V1 with Adam obtained 93.99% accuracy.
The last optimizer RMSprop also measured for validation purposes. LeNet with RMSprop achieved 95.06%, AlexNet with RMSprop obtained 95.28% accuracy, VGG16 with RMSprop achieved 92.49% accuracy, ResNet-50 with the RMSprop optimizer achieved 93.78% accuracy, and Inception-V1 with RMSprop achieved 91.42% validation accuracy on theLUNA16 dataset.
The second method, five-fold cross-validation, was adopted to train and validate the CNN architectures with the SGD optimizer. The validation statistical analysis is shown in Table 13. When the five-fold cross-validation approach was applied, the AlexNet–SGD optimizer achieved 95.73%, 95.20%, 96.33%, 96.75%, 94.59%, 5.41% and 95.97% accuracy, sensitivity, specificity, PPV, NPV, FOR, and F1-score, respectively, which were the highest scores compared with the other CNN architectures.
Table 13. Validation statistical analysis of five CNN architectures with the SGD optimizer confusion matrix (5-fold cross-validation).
The CNN-based architectures LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1, with different optimizers, were evaluated in this study. It was found that AlexNet with the SGD optimizer achieved the highest accuracy of 97.42%. Table 14 presents the performance analysis of AlexNet with the SGD technique with different methods. Comparatively, the accuracy of AlexNet with the SGD was higher than the other state-of-the-art approaches. Various existing publications used different methods on the different datasets, including LIDC-IDRI, LUNA16, IQ-OTH/NCCD, ELCAP, and private lung datasets for the detection of lung cancer. AlexNet with SGD and with the five-fold cross-validation method obtained 95.73% accuracy and a 4.27% misclassification rate. AlexNet with the SGD achieved the highest accuracy of 97.42% and 2.58% of misclassification rate.
Table 14. Comparison of AlexNet with the SGD model with previously published approaches.

5. Conclusions

In recent years, lung cancer has become a dangerous disease with a low survival rate. Early diagnosis and proper treatment can increase the survival rate. In this study, various deep learning-based architectures were presented to classify lung cancer into benign and malignant. The advanced CNN architectures LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1 were applied for the detection of lung cancer to analyze the performance. Various optimizers, including RMSprop, Adam, and SGD, were used to tune the CNN architectures, which provided different results. The experimental results show that the AlexNet architecture with the SGD optimizer achieved the highest validation accuracy of 97.42%, with a misclassification rate of 2.58% for the detection of lung cancer as benign or malignant while applying the five-fold cross-validation method. AlexNet–SGD achieved 95.73% accuracy, while AlexNet with the SGD optimizer outperformed compared to other state-of-the-art existing CNN architectures and optimizers.

6. Limitations and Future Work

The performance analysis of the state-of-the-art CNN architectures is presented in this study to classify lung cancer into benign and malignant. The study comprised five CNN architectures: LeNet, AlexNet, VGG16, ResNet-50, and Inception-V1, with the Adam, RMSprop, and SGD optimizers. Future work should include a performance analysis that can be increased to improve the classification system using other state-of-the-art CNN architectures, such as Darknet, EfficientNet, VGG19, Xception, Inception-V3, Inception-V4, Inception-ResNet-V2, and ResNeXt50. Various optimizers and cross-validation techniques can be adapted to remove the randomness effect to achieve better accuracy.

Author Contributions

I.N. and S.A. collected data from different resources. I.N., T.M., A.J. and M.A.K. performed the formal analysis and simulations. I.N., S.A. and A.M. contributed to writing and original draft preparation. I.N., A.J. and A.M. drafted the pictures and tables. M.A.K. and A.M. performed revisions and improved the quality of the draft. Writing—review and editing, M.A.K., T.M. and A.M.; supervision, M.A.K. and S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The simulation files/data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gonzalez, T.F. Handbook of Approximation Algorithms and Metaheuristics; Chapman & Hall/CRC: New York, NY, USA, 2007; pp. 1–1432. [Google Scholar] [CrossRef]
  2. Christie, J.R.; Lang, P.; Zelko, L.M.; Palma, D.A.; Abdelrazek, M.; Mattonen, S.A. Artificial Intelligence in Lung Cancer: Bridging the Gap Between Computational Power and Clinical Decision-Making. Can. Assoc. Radiol. J. 2021, 72, 86–97. [Google Scholar] [CrossRef] [PubMed]
  3. Munir, K.; Frezza, F.; Rizzi, A. Deep Learning for Brain Tumor Segmentation. Stud. Comput. Intell. 2021, 908, 189–201. [Google Scholar] [CrossRef]
  4. Rajaraman, S.; Ganesan, P.; Antani, S. Deep learning model calibration for improving performance in class-imbalanced medical image classification tasks. PLoS ONE 2022, 17, e0262838. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, S.; Yang, D.M.; Rong, R.; Zhan, X.; Fujimoto, J.; Liu, H.; Minna, J.; Wistuba, I.I.; Xie, Y.; Xiao, G. Artificial intelligence in lung cancer pathology image analysis. Cancers 2019, 11, 1673. [Google Scholar] [CrossRef]
  6. Melekoodappattu, J.G.; Dhas, A.S.; Kandathil, B.K.; Adarsh, K.S. Breast cancer detection in mammogram: Combining modified CNN and texture feature based approach. J. Ambient Intell. Humaniz. Comput. 2022, 2, 1–10. [Google Scholar] [CrossRef]
  7. Liu, X.; Deng, Z.; Yang, Y. Recent progress in semantic image segmentation. Artif. Intell. Rev. 2019, 52, 1089–1106. [Google Scholar] [CrossRef]
  8. Tunali, I.; Gillies, R.J.; Schabath, M.B. Application of Radiomics and Artificial Intelligence for Lung Cancer Precision Medicine. Cold Spring Harb. Perspect. Med. 2021, 11, a039537. [Google Scholar] [CrossRef]
  9. Santos, M.; Rocha, N.P. Medical Imaging Repository Contributions for Radiation Protection Key Performance Indicators. Procedia Comput. Sci. 2021, 196, 590–597. [Google Scholar] [CrossRef]
  10. Abel, M.F.; Sutherland, D.H.; Wenger, D.R.; Mubarak, S.J. Evaluation of ct scans and 3-D reformatted images for quantitative assessment of the hip. J. Pediatr. Orthop. 1994, 14, 48–53. [Google Scholar] [CrossRef]
  11. Siddiqui, S.Y.; Abbas, S.; Khan, M.A.; Naseer, I.; Masood, T.; Khan, K.M.; Al Ghamdi, M.A.; Almotiri, S.H. Intelligent decision support system for COVID-19 empowered with deep learning. Comput. Mater. Contin. 2020, 66, 1719–1732. [Google Scholar] [CrossRef]
  12. Leleu, O.; Basille, D.; Auquier, M.; Clarot, C.; Hoguet, E.; Baud, M.; Lenel, S.; Milleron, B.; Berna, P.; Jounieaux, V. Results of Second Round Lung Cancer Screening by Low-Dose CT scan—French Cohort Study (DEP-KP80). Clin. Lung Cancer 2022, 23, e54–e59. [Google Scholar] [CrossRef] [PubMed]
  13. Bhandary, A.; Prabhu, G.A.; Rajinikanth, V.; Thanaraj, K.P.; Thanaraj, K.P.; Satapathy, S.C.; Robbins, D.E.; Shasky, C.; Zhang, Y.-D.; Tavares, J.M.R.S.; et al. Deep-learning framework to detect lung abnormality—A study with chest X-ray and lung CT scan images. Pattern Recognit. Lett. 2020, 129, 271–278. [Google Scholar] [CrossRef]
  14. Gupta, P.; Shukla, A.P. Improving Accuracy of Lung Nodule Classification Using AlexNet Model. In Proceedings of the 2021 International Conference on Innovative Computing, Intelligent Communication and Smart Electrical Systems (ICSES), Chennai, India, 24–25 September 2021; pp. 1–6. [Google Scholar] [CrossRef]
  15. Shimazaki, A.; Ueda, D.; Choppin, A.; Yamamoto, A.; Honjo, T.; Shimahara, Y.; Miki, Y. Deep learning-based algorithm for lung cancer detection on chest radiographs using the segmentation method. Sci. Rep. 2022, 12, 1–10. [Google Scholar] [CrossRef] [PubMed]
  16. Qi, L.-L.; Wu, B.-T.; Tang, W.; Zhou, L.-N.; Huang, Y.; Zhao, S.-J.; Li, M.; Zhang, L.; Feng, S.-C.; Hou, D.-H.; et al. Long-term follow-up of persistent pulmonary pure ground-glass nodules with deep learning–assisted nodule segmentation. Eur. Radiol. 2020, 30, 744–755. [Google Scholar] [CrossRef] [PubMed]
  17. Khehrah, N.; Farid, M.S.; Bilal, S.; Khan, M.H. Lung Nodule Detection in CT Images Using Statistical and Shape-Based Features. J. Imaging 2020, 6, 6. [Google Scholar] [CrossRef]
  18. Xie, Y.; Zhang, J.; Xia, Y.; Fulham, M.; Zhang, Y. Fusing texture, shape and deep model-learned information at decision level for automated classification of lung nodules on chest CT. Inf. Fusion 2018, 42, 102–110. [Google Scholar] [CrossRef]
  19. Nasser, I.M.; Abu-Naser, S.S. Lung Cancer Detection Using Artificial Neural Network. Int. J. Eng. Inf. Syst. 2019, 3, 17–23. Available online: www.ijeais.org (accessed on 8 May 2022).
  20. Miah, M.B.A.; Yousuf, M.A. Detection of lung cancer from CT image using image processing and neural network. In Proceedings of the 2015 International Conference on Electrical Engineering and Information Communication Technology (ICEEICT), Savar, Bangladesh, 21–23 May 2015. [Google Scholar] [CrossRef]
  21. LeCun, Y.; Boser, B.; Denker, J.; Henderson, D.; Howard, R.; Hubbard, W.; Jackel, L. Handwritten Digit Recognition with a Back-Propagation Network. Adv. Neural Inf. Process. Syst. 1989, 2, 396–404. [Google Scholar]
  22. Zhang, S.; Sun, F.; Wang, N.; Zhang, C.; Yu, Q.; Zhang, M.; Babyn, P.; Zhong, H. Computer-Aided Diagnosis (CAD) of Pulmonary Nodule of Thoracic CT Image Using Transfer Learning. J. Digit. Imaging 2019, 32, 995–1007. [Google Scholar] [CrossRef]
  23. Zhao, X.; Liu, L.; Qi, S.; Teng, Y.; Li, J.; Qian, W. Agile convolutional neural network for pulmonary nodule classification using CT images. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 585–595. [Google Scholar] [CrossRef]
  24. Krizhevsky, B.A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Processing Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef]
  25. Agarwal, A.; Patni, K.; Rajeswari, D. Lung Cancer Detection and Classification Based on Alexnet CNN. In Proceedings of the 2021 6th International Conference on Communication and Electronics Systems (ICCES), Coimbatre, India, 8–10 July 2021; pp. 1390–1397. [Google Scholar] [CrossRef]
  26. Polat, H.; Mehr, H.D. Classification of pulmonary CT images by using hybrid 3D-deep convolutional neural network architecture. Appl. Sci. 2019, 9, 940. [Google Scholar] [CrossRef]
  27. Rao, P.; Fereira, N.A.; Srinivasan, R. Convolutional neural networks for lung cancer screening in computed tomography (CT) scans. In Proceedings of the 2016 2nd International Conference on Contemporary Computing and Informatics (IC3I), Greater Noida, India, 14–17 December 2016; pp. 489–493. [Google Scholar] [CrossRef]
  28. Lin, C.J.; Li, Y.C. Lung nodule classification using taguchi-based convolutional neural networks for computer tomography images. Electronics 2020, 9, 1066. [Google Scholar] [CrossRef]
  29. Al-Yasriy, H.F.; Al-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Diagnosis of Lung Cancer Based on CT Scans Using CNN. IOP Conf. Ser. Mater. Sci. Eng. 2020, 928, 022035. [Google Scholar] [CrossRef]
  30. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  31. Elnakib, A.; Amer, H.M.; Abou-Chadi, F.E.Z. Early lung cancer detection using deep learning optimization. Int. J. Online Biomed. Eng. 2020, 16, 82–94. [Google Scholar] [CrossRef]
  32. Geng, L.; Zhang, S.; Tong, J.; Xiao, Z. Lung segmentation method with dilated convolution based on VGG-16 network. Comput. Assist. Surg. 2019, 24 (Suppl. 2), 27–33. [Google Scholar] [CrossRef]
  33. Pang, S.; Meng, F.; Wang, X.; Wang, J.; Song, T.; Wang, X.; Cheng, X. VGG16-T: A novel deep convolutional neural network with boosting to identify pathological type of lung cancer in early stage by ct images. Int. J. Comput. Intell. Syst. 2020, 13, 771–780. [Google Scholar] [CrossRef]
  34. Sajja, T.K.; Devarapalli, R.M.; Kalluri, H.K. Lung cancer detection based on CT scan images by using deep transfer learning. Trait. Signal 2019, 36, 339–344. [Google Scholar] [CrossRef]
  35. Nibali, A.; He, Z.; Wollersheim, D. Pulmonary nodule classification with deep residual networks. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1799–1808. [Google Scholar] [CrossRef]
  36. Zheng, G.; Han, G.; Soomro, N.Q. An Inception Module CNN Classifiers Fusion Method on Pulmonary Nodule Diagnosis by Signs. Tsinghua Sci. Technol. 2019, 25, 368–383. [Google Scholar] [CrossRef]
  37. Haibo, L.; Shanli, T.; Shuang, S.; Haoran, L. An improved yolov3 algorithm for pulmonary nodule detection. In Proceedings of the 2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 18–20 June 2021; pp. 1068–1072. [Google Scholar]
  38. Zhang, X.; Lee, V.C.S.; Rong, J.; Liu, F.; Kong, H. Multi-channel convolutional neural network architectures for thyroid cancer detection. PLoS ONE 2022, 17, e0262128. [Google Scholar] [CrossRef] [PubMed]
  39. Leo, M.; Cacagnì, P.; Signore, L.; Benincasa, G.; Laukkanen, M.O.; Distante, C. Improving Colon Carcinoma Grading by Advanced CNN Models. In International Conference on Image Analysis and Processing; Springer: Cham, Switzerland, 2022; pp. 233–244. [Google Scholar] [CrossRef]
  40. Siddiqui, S.Y.; Naseer, I.; Khan, M.A.; Mushtaq, M.F.; Naqvi, R.A.; Hussain, D.; Haider, A. Intelligent breast cancer prediction empowered with fusion and deep learning. Comput. Mater. Contin. 2021, 67, 1033–1049. [Google Scholar] [CrossRef]
  41. Siddiqui, S.Y.; Haider, A.; Ghazal, T.M.; Khan, M.A.; Naseer, I.; Abbas, S.; Rahman, M.; Khan, J.A.; Ahmad, M.; Hasan, M.K.; et al. IoMT Cloud-Based Intelligent Prediction of Breast Cancer Stages Empowered with Deep Learning. IEEE Access 2021, 9, 146478–146491. [Google Scholar] [CrossRef]
  42. Carcagnì, P.; Leo, M.; Celeste, G.; Distante, C.; Cuna, A. A systematic investigation on deep architectures for automatic skin lesions classification. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 8639–8646. [Google Scholar] [CrossRef]
  43. Zhang, N.; Cai, Y.X.; Wang, Y.Y.; Tian, Y.T.; Wang, X.L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med. 2020, 102, 101756. [Google Scholar] [CrossRef] [PubMed]
  44. Setio, A.A.A.; Traverso, A.; de Bel, T.; Berens, M.S.N.; van den Bogaard, C.; Cerello, P.; Chen, H.; Dou, Q.; Fantacci, M.E.; Geurts, B.; et al. Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge. Med. Image Anal. 2017, 42, 1–13. [Google Scholar] [CrossRef]
  45. Kazerooni, E.A.; Austin, J.H.M.; Black, W.C.; Dyer, D.S.; Hazelton, T.R.; Leung, A.N.; McNitt-Gray, M.F.; Munden, R.F.; Pipavath, S. ACR-STR practice parameter for the performance and reporting of lung cancer screening thoracic computed tomography (CT): 2014 (Resolution 4). J. Thorac. Imaging 2014, 29, 310–316. [Google Scholar] [CrossRef]
  46. Li, Y.; Tang, Y. Design on Intelligent Feature Graphics Based on Convolution Operation. Mathematics 2022, 10, 384. [Google Scholar] [CrossRef]
  47. Han, Y.; Li, J.; Lou, X.; Fan, C.; Geng, Z. Energy saving of buildings for reducing carbon dioxide emissions using novel dendrite net integrated adaptive mean square gradient. Appl. Energy 2022, 309, 118409. [Google Scholar] [CrossRef]
  48. Mao, Q.; Zhao, S.; Ren, L.; Li, Z.; Tong, D.; Yuan, X.; Li, H. Intelligent immune clonal optimization algorithm for pulmonary nodule classification. Math. Biosci. Eng. 2021, 18, 4146–4161. [Google Scholar] [CrossRef]
  49. Gao, Y.; Song, F.; Zhang, P.; Liu, J.; Cui, J.; Ma, Y.; Zhang, G.; Luo, J. Improving the Subtype Classification of Non-small Cell Lung Cancer by Elastic Deformation Based Machine Learning. J. Digit. Imaging 2021, 34, 605–617. [Google Scholar] [CrossRef] [PubMed]
  50. Lai, K.D.; Nguyen, T.T.; Le, T.H. Detection of lung nodules on ct images based on the convolutional neural network with attention mechanism. Ann. Emerg. Technol. Comput. 2021, 5, 77–89. [Google Scholar] [CrossRef]
  51. Silva, F.; Pereira, T.; Neves, I.; Morgado, J.; Freitas, C.; Malafaia, M.; Sousa, J.; Fonseca, J.; Negrao, E.; de Lima, B.F.; et al. Towards Machine Learning-Aided Lung Cancer Clinical Routines: Approaches and Open Challenges. J. Pers. Med. 2022, 12(3), 480. [Google Scholar] [CrossRef] [PubMed]
  52. Pradhan, K.; Chawla, P. Medical Internet of things using machine learning algorithms for lung cancer detection. J. Manag. Anal. 2020, 7, 591–623. [Google Scholar] [CrossRef]
  53. Bansal, G.; Chamola, V.; Narang, P.; Kumar, S.; Raman, S. Deep3DScan: Deep residual network and morphological descriptor based framework for lung cancer classification and 3D segmentation. IET Image Process. 2020, 14, 1316–1326. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.