You are currently viewing a new version of our website. To view the old version click .
Healthcare
  • Article
  • Open Access

8 June 2022

A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma

,
,
and
1
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakakah 72312, Saudi Arabia
2
School of Information Technology & Engineering, Vellore Institute of Technology, Vellore 632001, India
3
School of Computer Science (SCS), Taylor’s University, Subang Jaya 47500, Malaysia
*
Author to whom correspondence should be addressed.
This article belongs to the Section Artificial Intelligence in Healthcare

Abstract

Lung cancer is among the most hazardous types of cancer in humans. The correct diagnosis of pathogenic lung disease is critical for medication. Traditionally, determining the pathological form of lung cancer involves an expensive and time-consuming process investigation. Lung cancer is a leading cause of mortality worldwide, with lung tissue nodules being the most prevalent way for doctors to identify it. The proposed model is based on robust deep-learning-based lung cancer detection and recognition. This study uses a deep neural network as an extraction of features approach in a computer-aided diagnosing (CAD) system to assist in detecting lung illnesses at high definition. The proposed model is categorized into three phases: first, data augmentation is performed, classification is then performed using the pretrained CNN model, and lastly, localization is completed. The amount of obtained data in medical image assessment is occasionally inadequate to train the learning network. We train the classifier using a technique known as transfer learning (TL) to solve the issue introduced into the process. The proposed methodology offers a non-invasive diagnostic tool for use in the clinical assessment that is effective. The proposed model has a lower number of parameters that are much smaller compared to the state-of-the-art models. We also examined the desired dataset’s robustness depending on its size. The standard performance metrics are used to assess the effectiveness of the proposed architecture. In this dataset, all TL techniques perform well, and VGG 16, VGG 19, and Xception for 20 epoch structure are compared. Preprocessing functions as a wonderful bridge to build a dependable model and eventually helps to forecast future scenarios by including the interface at a faster phase for any model. At the 20th epoch, the accuracy of VGG 16, VGG 19, and Xception is 98.83 percent, 98.05 percent, and 97.4 percent.

1. Introduction

Cancer is a significant concern across the globe, and it is the second foremost reason for death. Lung cancer stands in third place among cancer deaths. Lung cancer is broadly classified into small and non-small-cell lung cancer. Various subtypes in non-small-cell lung cancer are adenocarcinoma, squamous cell cancer, and large cell carcinoma. Non-small-cell lung cancer is frequently observed, but small cell lung cancer spreads faster and is often fatal. Changes to the affected person’s voice, chest pain, shortness of breath, and wheezing are a few symptoms to list, and more painful symptoms also prevail [1,2]. Chemotherapy, targeted drugs, and immunotherapy are the treatments that have been approved over the decades [3].
Google Trends show the potential research going on in the field of cancer, and this is graphed in Figure 1. Several progressive methods are being devised for earlier detection with the help of artificial intelligence concepts.
Figure 1. Cancer research.
Our main contribution to the work is as follows:
  • We have designed a model which can classify the patients’ level of lung carcinoma by applying the TL process, and it is the first of the type to be carried out with this dataset [4].
  • Using CT images as the system’s input, it can predict the level and helps to take contour action at the earliest time.
  • We have applied three TL approaches here, namely VGG16, VGG19, and Xception, with 20 epochs in the Google Colab platform, and it was shown to be a better model to use for future prediction.
  • Based on the experimental performance for lung carcinoma, VGG16 gives maximum accuracy of 98.83%, whereas Xception shows an accuracy rate of 97.4%.
The work is carried out in a phased manner. Section 2 of the paper discusses the recent work carried out in this domain. Section 3 details the workflow with architecture diagram illustration and fine-tuned hyperparameter values. Section 4 elaborates on the experimental discussion, followed by a comparison with recent results, conclusion, and future work in Section 5.

3. Workflow Architecture

Analysis of the cancer dataset with the TL was very informative and yielded a good model based on the various performance measure perspectives. Capturing the lung cancer image dataset was followed by preprocessing using the image data generator, then the TL model, and finally the prediction of the type of abnormality was achieved, as is portrayed in Figure 2.
Figure 2. TL-based lung carcinoma classification.

3.1. Dataset Description

In the fall of 2019, the Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases (IQ-OTH/NCCD) collected data for three months. The dataset comprised CT scans captured from patients in different stages and scans captured from patients in different stages and healthy people in DICOM format. The following CT protocol was utilized for reading: 120 kV, 1 mm slice thickness, window widths varying from 350 to 1200 HU, and window centers ranging from 50 to 600. An oncologist and radiologist working in the centers marked the slides. From 110 cases, 1190 images were retrieved. All the cases were classified into 3 classes, namely normal, benign, and malignant. In total, 55 cases were classified as normal, 15 cases were classified as benign, and 40 cases were classified as malignant.
Prior to analysis, all photos were de-identified. The oversight review board waived written consent. The institutional review boards of the participating medical centers approved the study. There were multiple slices in each scan. The number of slices varied between 80 and 200, and each one displayed an image of the human chest from various sides and perspectives. Gender, age, educational attainment, residency area, and living status were all different in the 110 cases. Some worked for the Iraqi ministries of transportation and oil, while others were farmers and gainers. Most of them were from Iraq’s central area, specifically the provinces of Baghdad, Wasit, Diyala, Salahuddin, and Babylon [16].

3.2. Image Data Generator

To expand the size of the dataset image, augmentation phase is deployed. In Keras, image data generator is used to achieve the augmentation. Different transformation is applied to the original to make it sizable in applying the deep learning structure. It also integrates variation in the images so that the built model will be reliable. This process requires reduced memory only. In our work, the training data following the augmentation technique have values as follows: rescale = 1.255, shear_range = 0.2, zoom_range = 0.2, and horizontal_flip = True. Upon analysis with various options, we found that this was most suitable.

3.3. TL

To the created image dataset pool, TL algorithms, namely, VGG 16, VGG 19, and Xception were applied with fine-tuning hyperparameters to structure the network for training and testing in turn, in order to help in future prediction for new records. First, work was carried out with deep neural networks, meant for object detection. The alternative to deep learning is the TL method, and it works on four transfer methods, namely, instance-based, feature-based, parameter-based, and relation-based. Applications of TL are substantial and this article provides insight [17,18]. GPU support in the Google Colab platform provided the helping hand to overcome the computation glitches and showed good throughput.

3.4. Fine-Tuned Hyperparameters

Model hyperparameters were used to optimize the model, loss used was categorical_crossentropy, optimizer was adam. Fit_generator was the iterator method integrated to fit the data on the neural network simulation. Parameters were target_size = (224,224), batch_size = 32, and epochs = 20, and class_mode was categorical, based (normal, benign, malignant) on the nature of the dataset. In total, 75% of data were for training purposes. Varying hyperparameters were tried and executed to obtain optimized results.
The following performance measures were used to study the model consistency. The training and testing criteria showed varied performance [19,20].

3.4.1. Accuracy

Accuracy is an ideal classification metric and is easy to understand [21,22]. It is the proportion of true results to total number of results. It is easily suited to binary and multiclass classification scenarios. It is a valid evaluation method for problems where data are not skewed and are well-balanced. Below is the confusion matrix to substantiate the 3-class performance measure.
Accuracy = T P + T N T P + F P + F N + T N
ACTUAL VALUES
NormalBenignMalignant
PREDICTED VALUESNormal+ve
1
−ve
2
−ve
3
Benign−ve
4
+ve
5
−ve
6
Malignant−ve
7
−ve
8
+ve
9
Normal
  • TP = Cell1
  • FP = Cell2 + Cell3
  • TN = Cell5 + Cell6 + Cell8 + Cell9
  • FN = Cell4 + Cell7
Benign
  • TP = Cell5
  • FP = Cell4 + Cell6
  • TN = Cell1 + Cell3 + Cell7 + Cell9
  • FN = Cell2 + Cell8
Malignant
  • TP = Cell9
  • FP = Cell7 + Cell8
  • TN = Cell1 + Cell2 + Cell4 + Cell5
  • FN = Cell3 + Cell6
  • TP—True Positive—predict yes and have lung carcinoma
  • TN—True Negative—correctly predict that they have no lung carcinoma
  • FP—False Positive—incorrectly predict having lung carcinoma but no lung carcinoma (type 1 error)
  • FN—False Negative—predict no lung carcinoma but have lung carcinoma (type 2 error)

3.4.2. Loss

It is a function employed to compute distance between the current output of the algorithm and the expected output. It evaluates the performance of an algorithm’s ability to model data. There are two groups, one for classification and another for regression. The commonly used loss functions are cross-entropy, exponential loss, log loss, and hinge loss.

3.4.3. AUC

It refers to area under the receiver operating characteristic curve. It is a performance indicator for how well the positive class probabilities are separated from negative class probabilities. It is scale-invariant. It indicates how well-ranked the predictions are, rather than the absolute values of predictions. Another benefit of employing it is that it is classification-threshold-invariant such that it measures the prediction’s quality irrespective of the threshold chosen.

3.4.4. Precision

Precision is the ratio of relevant outcomes predicted to actually relevant outcomes [23,24]. It is also called positive prediction value. It takes false positives into account.
Precision = T P T P + F P

3.4.5. Recall

Recall gives the ratio of outcomes that were actually relevant to correctly predicted outcomes. It is also called sensitivity. It takes false negatives into account.
Recall = T P T P + F N

3.4.6. F1 Score

F1 Score is another measure that provides the harmonic mean of precision and recall.
F 1   Score = 2 Precision Recall Precision + Recall

4. Experimental Discussion and Analysis

4.1. VGG 16

VGG 16 is a CNN model for visual recognition proposed by Karen Simonyan and Andrew Zisserman of Visual Geometry Group Lab at Oxford University in 2014. VGG 16 won first place in detecting objects within 200 classes and won second place in classifying images labelled 1 of 1000 categories at the ImageNet Large Scale Visual Recognition Challenge in 2014. Applying this model on the ImageNet dataset containing 1000 classes with 14 million images yielded 92.7% top-5 test accuracy.
It takes image input in dimensions (224,224,3). It has a total of sixteen layers. The first two layers are of 3∗3 filter size and have 64 channels with the same padding. Then there is the (2,2) stride max pool layer followed by two 256 filters and (3,3) filter size convolution layers. This is followed by a (2,2) stride max pool layer. Then there are two (3,3) size and 256 filters convolution layers. It is followed by two sets of three 512 filters and (3,3) size convolution layer and a max pool layer all of the same padding. The image is then passed to the two convolution layers stack. Padding of 1-pixel is used to prevent spatial features of the image after each convolution layer.
The (7,7512) feature map is generated once the image passes through all the layers of VGG 16. This feature map is flattened to obtain the (1,25088) feature vector. There are three fully connected layers. The first obtains input from the last feature vector and gives a (1,4096) vector. The second also produces the same output as the first. While the third outputs 1000 channels for 1000 classes. The classification vector is then normalized, bypassing the output form third layer to softmax layer [25,26,27]. The architecture is pictured in Figure 3.
Figure 3. VGG 16 architecture.
The performance measures for the various epochs over VGG 16 are tabulated and plotted in Table 1 for training and Table 2 for testing data and are shown in Figure 4, Figure 5, Figure 6 and Figure 7. Gradually loss decreases, indicating the building of a good model; other performance measures also indicate the building of a good model and other performance measures show the increasing trend. By the start of the 5th epoch, the accuracy started to show outstanding value of above 95%. Similarly, AUC shows the ideal curve. Without compromising, precision and recall also show the upper trend.
Table 1. Performance measures over VGG 16 for training data.
Table 2. Performance measures over VGG 16 for testing data.
Figure 4. Loss vs. Epoch.
Figure 5. Accuracy vs. Epoch.
Figure 6. AUC vs. Epoch.
Figure 7. Precision vs. Recall.

4.2. VGG 19

A VGG model consisting of 19 layers is VGG 19. It has sixteen convolution layers, three fully connected layers, five maxpool layers, and a softmax layer. VGG 19 is conceptually the same as VGG 16 while the number of layers supported by it is different. Nineteen and sixteen denote the number of layers in the model. This model is trained on millions of images present in the ImageNet database. It can classify pictures into 1000 object categories such as pencil, keyboard, car, and many animals.
The input RGB image is given in the fixed matrix dimension of (224,224,3). It is preprocessed by subtracting RGB mean value from each pixel. Kernels of (3,3) size with 1-pixel stride is used to cover the whole image notion. Spatial resolution of the image is preserved by the usage of spatial padding. The (2,2) max pooling is performed with two-pixel stride. This is followed by rectified linear unit (ReLu) to incorporate non-linearity in the model, which makes the model classify better and reduces computational time. This model proved to be much better than previously used models such as tanh or sigmoid function. It has three completely connected layers, of which the first two are of size 4096. A layer with 1000 channels for 1000 object categories of ILSVRC classification is also present, followed by a final softmax function, as illustrated in Figure 8.
Figure 8. VGG 19 architecture.
Primarily, it was developed to win ILSVRC but later it was used as classification architecture for image datasets. Its models were made available to the public by the author, so it can be used as it is or with any required modifications for similar purposes. It can be used to perform TL tasks such as facial recognition. The VGG 19 model was able to perform facial recognition with higher accuracy even with mask usage [28].
Performance measures for the various epochs over VGG 19 are tabulated and plotted in Table 3 for training and Table 4 for testing data and are shown in Figure 9, Figure 10, Figure 11 and Figure 12. Gradually loss decreases, indicating the building of a good model; other performance measures also indicate the building of a good model and other performance measures show the increasing trend. By the start of the 10th epoch, the accuracy started to show outstanding value of above 95%. Similarly, AUC shows the ideal curve. Without compromising, precision and recall also show the upper trend.
Table 3. Performance measures over VGG 19 for training data.
Table 4. Performance measures over VGG 19 for testing data.
Figure 9. Loss vs. Epoch.
Figure 10. Accuracy vs. Epoch.
Figure 11. AUC vs. Epoch.
Figure 12. Precision vs. Recall.

4.3. Xception

It is a convolutional neural network model, with Google assisting object detection and image analysis. It is an extreme version of inception. It is based on layers of depth-wise separable convolution, which performs better than inception. It is also hypothesized that mapping of spatial correlations and cross-channel correlations in convolutional neural network feature maps can be completely decoupled. This hypothesis is stronger than the one present in inception.
Xception architecture possesses thirty-six convolution layers giving rise to a base where feature extraction can be performed. The thirty-six layers are separated into fourteen modules with linear residual connections surrounding them. Linear residual connections are exceptions in the first and last module. In short, there are depth-wise separable layers of convolution with residual connections in a linear stack. Hence, it is an easily definable and modifiable model. It can be modified by using just a few lines of code and high-level libraries such as Keras or TensorFlow.
The presence of residual connections helps in convergence; it makes it a better performer in terms of classification and speed. The image first goes into entry flow, followed by middle flow, where the same process is repeated eight times. Finally, it goes through exit flow. All depth-wise separable convolution layers and convolution layers are followed by batch normalization. All layers have a depth multiplier of one, such that there is no expansion of depth. Xception models perform well in convolutional neural networks, and Figure 13 depicts the flow [29].
Figure 13. Xception architecture.
The performance measures for the various epochs over Xception are tabulated and plotted in Table 5 for training and Table 6 for testing data and are shown in Figure 14, Figure 15, Figure 16 and Figure 17. Gradually loss decreases, indicating the building of a good model; other performance measures also indicate the building of a good model and other performance measures show the increasing trend. From the beginning itself, accuracy, AUC, precision, and recall are good, and as mentioned, this model performs well in the initial stage itself. Because of the similar values, the Figure 17 graph merges precision vs. recall together.
Table 5. Performance measures over Xception for training data.
Table 6. Performance measures over Xception for training data.
Figure 14. Loss vs. Epoch.
Figure 15. Accuracy vs. Epoch.
Figure 16. AUC vs. Epoch.
Figure 17. Precision vs. Recall.

5. Conclusions and Future Work

On comparison with the recent works carried out using this same dataset, our model yields good results. With the same dataset work carried out using SVM classification, the accuracy achieved is 89.8876% [30]. In work carried out with the AlexNet architecture, accuracy is 93.548% [31].
All TL approaches perform well in this dataset and compare VGG 16, VGG 19, and Xception for the 20-epoch structure. For any model, preprocessing acts as the great bridge to construct a reliable model and eventually helps predict a reliable model and eventually helps in predicting future cases by incorporating the interface at the faster phase. Cumulatively, accuracy of VGG 16, VGG 19, and Xception is 98.83%, 98.05%, and 97.4%, respectively, at the 20th epoch. Loss is somewhat high in the case of Xception over testing data, and accuracy of VGG 16, VGG 19, and Xception is 83.39%, 80.97%, and 89.68%, respectively, at the 20th epoch. In comparison, VGG16 and VGG19 outperform.
Utilization of other clinical data along with this will provide even more clarity. Additional demographic information helps in taking additional steps to create awareness in more prone locations. Normally, lifestyle factors such as eating habits, cultural habits, and social living have a great impact on many diseases and cumulating the same will aid greatly in decision-making.

Author Contributions

Conceptualization, M.H. and R.S.; methodology, R.S.; software, S.N.A. and M.H.; validation, R.S., N.Z.J. and S.N.A.; formal analysis, R.S. and M.H.; investigation, R.S. and M.H.; resources, M.H.; data curation, M.H. and R.S.; writing—original draft preparation, M.H. and R.S.; writing—review and editing, R.S., M.H. and N.Z.J.; visualization, M.H.; supervision, N.Z.J.; project administration, N.Z.J.; funding acquisition, S.N.A. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Deanship of Scientific Research at Jouf University under grant no. (DSR-2021-02-0330).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data will be furnished on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nall, R. What to Know about Lung Cancer. Medical News Today. 2018. Available online: https://www.medicalnewstoday.com/articles/323701 (accessed on 2 April 2022).
  2. Saeed, S.; Jhanjhi, N.Z.; Naqvi, M.; Humyun, M.; Ahmad, M.; Gaur, L. Optimized Breast Cancer Premature Detection Method with Computational Segmentation: A Systematic Review Mapping. In Approaches and Applications of Deep Learning in Virtual Medical Care; IGI Global: Harrisburg, PA, USA, 2022; pp. 24–51. [Google Scholar]
  3. Facts About Lung Cancer. Available online: https://www.lungcancerresearchfoundation.org/lung-cancer-facts/ (accessed on 27 March 2022).
  4. Shukla, D. New Test May Quickly Detect Early-Stage Lung Cancer. Medical News Today. 2022. Available online: https://www.medicalnewstoday.com/articles/new-test-may-quickly-detect-early-stage-lung-cancer (accessed on 28 March 2022).
  5. Zhou, Y.; Lu, Y.; Pei, Z. Accurate diagnosis of early lung cancer based on the convolutional neural network model of the embedded medical system. Microprocess. Microsyst. 2021, 81, 103754. [Google Scholar] [CrossRef]
  6. Nanglia, P.; Kumar, S.; Mahajan, A.N.; Singh, P.; Rathee, D. A hybrid algorithm for lung cancer classification using SVM and Neural Networks. ICT Express 2021, 7, 335–341. [Google Scholar] [CrossRef]
  7. Bai, Y.; Li, D.; Duan, Q.; Chen, X. Analysis of high-resolution reconstruction of medical images based on deep convolutional neural networks in lung cancer diagnostics. Comput. Methods Programs Biomed. 2022, 217, 106592. [Google Scholar] [CrossRef]
  8. Ahmad, A.S.; Mayya, A.M. A new tool to predict lung cancer based on risk factors. Heliyon 2020, 6, e03402. [Google Scholar] [CrossRef]
  9. Ahn, B.-C.; So, J.-W.; Synn, C.-B.; Kim, T.H.; Kim, J.H.; Byeon, Y.; Kim, Y.S.; Heo, S.G.; Yang, S.D.; Yun, M.R.; et al. Clinical decision support algorithm based on machine learning to assess the clinical response to anti–programmed death-1 therapy in patients with non–small-cell lung cancer. Eur. J. Cancer 2021, 153, 179–189. [Google Scholar] [CrossRef]
  10. Agarwal, S.; Yadav, A.S.; Dinesh, V.; Vatsav, K.S.S.; Prakash, K.S.S.; Jaiswal, S. By artificial intelligence algorithms and machine learning models to diagnosis cancer. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  11. Huang, X.; Lei, Q.; Xie, T.; Zhang, Y.; Hu, Z.; Zhou, Q. Deep transfer convolutional neural network and extreme learning machine for lung nodule diagnosis on CT images. Knowl.-Based Syst. 2020, 204, 106230. [Google Scholar] [CrossRef]
  12. Doppalapudi, S.; Qiu, R.G.; Badr, Y. Lung cancer survival period prediction and understanding: Deep learning approaches. Int. J. Med. Inform. 2021, 148, 104371. [Google Scholar] [CrossRef]
  13. Ashraf, S.F.; Yin, K.; Meng, C.X.; Wang, Q.; Wang, Q.; Pu, J.; Dhupar, R. Predicting benign, preinvasive, and invasive lung nodules on computed tomography scans using machine learning. J. Thorac. Cardiovasc. Surg. 2022, 163, 1496–1505. [Google Scholar] [CrossRef]
  14. Shakeel, P.M.; Burhanuddin, M.A.; Desa, M.I. Lung cancer detection from CT image using improved profuse clustering and deep learning instantaneously trained neural networks. Measurement 2019, 145, 702–712. [Google Scholar] [CrossRef]
  15. ALzubi, J.A.; Bharathikannan, B.; Tanwar, S.; Manikandan, R.; Khanna, A.; Thaventhiran, C. Boosted neural network ensemble classification for lung cancer disease diagnosis. Appl. Soft Comput. 2019, 80, 579–591. [Google Scholar] [CrossRef]
  16. Alyasriy, H. The IQ-OTHNCCD Lung Cancer Dataset. Mendeley Data, Version 1. 2020. Available online: https://data.mendeley.com/datasets/bhmdr45bh2/1 (accessed on 28 March 2022).
  17. Szegedy, C.; Toshev, A.; Erhan, D. Deep neural networks for object detection. In Advances in Neural Information Processing Systems, Proceedings of the 26th International Conference on Neural Information Processing Systems, Nevada City, CA, USA, 5–10 December 2013; Morgan Kaufmann Publishers: San Francisco, CA, USA, 2013; Volume 26. [Google Scholar]
  18. Yu, X.; Wang, J.; Hong, Q.Q.; Teku, R.; Wang, S.H.; Zhang, Y.D. Transfer learning for medical images analyses: A survey. Neurocomputing 2022, 489, 230–254. [Google Scholar] [CrossRef]
  19. Hand, D.J.; Till, R.J. A simple generalisation of the area under the ROC curve for multiple class classification problems. Mach. Learn. 2001, 45, 171–186. [Google Scholar] [CrossRef]
  20. Ferri, C.; Hernández-Orallo, J.; Modroiu, R. An experimental comparison of performance measures for classification. Pattern Recognit. Lett. 2009, 30, 27–38. [Google Scholar] [CrossRef]
  21. Humayun, M.; Alsayat, A. Prediction Model for Coronavirus Pandemic Using Deep Learning. Comput. Syst. Sci. Eng. 2022, 40, 947–961. [Google Scholar] [CrossRef]
  22. Khalil, M.I.; Humayun, M.; Jhanjhi, N.; Talib, M.; Tabbakh, T.A. Multi-class Segmentation of Organ at Risk from Abdominal CT Images: A Deep Learning Approach. In Intelligent Computing and Innovation on Data Science; Peng, S.L., Hsieh, S.Y., Gopalakrishnan, S., Duraisamy, B., Eds.; Springer: Berlin, Germany, 2021; pp. 425–434. [Google Scholar]
  23. Gouda, W.; Almurafeh, M.; Humayun, M.; Jhanjhi, N.Z. Detection of COVID-19 Based on Chest X-rays Using Deep Learning. Healthcare 2022, 10, 343. [Google Scholar] [CrossRef]
  24. Khalil, M.I.; Tehsin, S.; Humayun, M.; Jhanjhi, N.; AlZain, M.A. Multi-Scale Network for Thoracic Organs Segmentation. CMC-Comput. Mater. Contin. 2022, 70, 3251–3265. [Google Scholar] [CrossRef]
  25. Cheng, S.; Zhou, G. Facial expression recognition method based on improved VGG convolutional neural network. Int. J. Pattern Recognit. Artif. Intell. 2020, 34, 2056003. [Google Scholar] [CrossRef]
  26. Tammina, S. Transfer learning using VGG-16 with deep convolutional neural network for classifying images. Int. J. Sci. Res. Publ. IJSRP 2019, 9, 143–150. [Google Scholar] [CrossRef]
  27. VGG-16: CNN Model. Available online: https://www.geeksforgeeks.org/vgg-16-cnn-model/ (accessed on 27 March 2022).
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  30. Kareem, H.F.; AL-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Evaluation of SVM Performance in the Detection of Lung Cancer in Marked CT Scan Dataset. Indones. J. Electr. Eng. Comput. Sci. 2021, 21, 1731–1738. [Google Scholar] [CrossRef]
  31. Amiripallia, S.S.; Rao, G.N.; Beharaa, J.; Sanjay, K. Mineral Rock Classification Using Convolutional Neural Network. In Recent Trends in Intensive Computing; IOS Press: Amsterdam, The Netherlands, 2021. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.