Next Article in Journal
Influencing Factors Analysis and Optimization of Land Use Allocation: Combining MAS with MOPSO Procedure
Next Article in Special Issue
A Cost-Effective Fall-Detection Framework for the Elderly Using Sensor-Based Technologies
Previous Article in Journal
Thermal Recycling of Glass Fibre Composites: A Circular Economy Approach
Previous Article in Special Issue
Efficient Scheduling of Home Energy Management Controller (HEMC) Using Heuristic Optimization Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Smart Diagnosis of Adenocarcinoma Using Convolution Neural Networks and Support Vector Machines

by
Balasundaram Ananthakrishnan
1,2,*,
Ayesha Shaik
2,*,
Shubhadip Chakrabarti
2,
Vaishnavi Shukla
2,
Dewanshi Paul
3 and
Muthu Subash Kavitha
4
1
Centre for Cyber Physical Systems, Vellore Institute of Technology, Chennai 600127, India
2
School of Computer Science and Engineering, Vellore Institute of Technology, Chennai 600127, India
3
School of Electronics Engineering, Vellore Institute of Technology, Chennai 600127, India
4
School of Information and Data Sciences, Nagasaki University, Nagasaki 8528521, Japan
*
Authors to whom correspondence should be addressed.
Sustainability 2023, 15(2), 1399; https://doi.org/10.3390/su15021399
Submission received: 22 November 2022 / Revised: 26 December 2022 / Accepted: 27 December 2022 / Published: 11 January 2023

Abstract

:
Adenocarcinoma is a type of cancer that develops in the glands present on the lining of the organs in the human body. It is found that histopathological images, obtained as a result of biopsy, are the most definitive way of diagnosing cancer. The main objective of this work is to use deep learning techniques for the detection and classification of adenocarcinoma using histopathological images of lung and colon tissues with minimal preprocessing. Two approaches have been utilized. The first method entails creating two CNN architectures: CNN with a Softmax classifier (AdenoCanNet) and CNN with an SVM classifier (AdenoCanSVM). The second approach corresponds to training some of the prominent existing architecture such as VGG16, VGG19, LeNet, and ResNet50. The study aims at understanding the performance of various architectures in diagnosing using histopathological images with cases taken separately and taken together, with a full dataset and a subset of the dataset. The LC25000 dataset used consists of 25,000 histopathological images, having both cancerous and normal images from both the lung and colon regions of the human body. The accuracy metric was taken as the defining parameter for determining and comparing the performance of various architectures undertaken during the study. A comparison between the several models used in the study is presented and discussed.

1. Introduction

The world today witnesses cancer as one of the most dreadful diseases impacting precious human life adversely. While cancer happens to be a generic term to refer to a large variety of diseases, it essentially involves enhanced formation of abnormal cells that propagate through the bloodstream and spread across the body, destroying normal cells, leading to the death of the affected individual. As reported by WHO [1] (World Health Organization), this disease is estimated to have affected around 10 million people in the year 2020. The global cancer trend [2] has been so concerning that an estimated 47% increase in the disease’s prevalence worldwide from 2020 to 2040 is predicted.
Adenocarcinoma [3] relates to a common type of cancer found in glandular epithelial cells of human body. Lung, prostate, pancreas, liver, colorectal area, and breast form the primary sites for the adenocarcinoma. Unlike other carcinomas, these cancers do not exhibit any symptoms during their early stages and remain undetected. As data [3] say, adenocarcinoma is responsible for around 40%, 95%, and 96% of non-small-cell lung cancers, pancreatic cancers, and colorectal cancers, respectively. It is also responsible for almost all prostate cancers and most breast cancers.
Histological images continue to be a standard method of cancer diagnosis. The advances made thus far in the area of health informatics have still proven to be unsuccessful in meeting the desired clinical requirements. Most diagnostics to date are still performed manually and rely heavily on the expertise and experience of histopathologists. These diagnostic techniques happen to be very time-consuming and difficult to grade in a reproducible manner.
Computer aided diagnosis using histopathological images has always been a topic of paramount interest in the field of cancer detection. Multiple works have been conducted in this domain using artificial intelligence (AI). Experimentation on AI-based cancer diagnosis using various machine learning and deep learning models has presently evolved as one of the prime areas of interest. Major studies undertaken in this domain include implementation of different AI-based models, improving existing AI-based models, or evaluating existing AI-based models to have an insight into their comparative efficacy. The development of methods to enhance image processing techniques to better extract features is provided as a solution to this problem statement. The development of filters to select the most efficient models among all the multiple machine learning models used as classifiers to improve accuracy is also one of the works conducted in this area. Existing works in this domain have recorded accuracies ranging from 70% to over 90%.
Compared to manual analysis, an AI-based system has the potential to provide rapid and consistent cancer detection and classification results. Therefore, the treatment and analysis of images using advanced machine learning and deep learning techniques need to be introduced to facilitate the increased rate of disease diagnosis in humans.
The primary objective of the present study is to develop an artificial intelligence-based tool that can assist in diagnosing adenocarcinoma from histopathology-based images. The study aims at detecting and classifying adenocarcinoma in the colon and lung regions of the human body using the LC25000 dataset [4] procured from Kaggle.

Literature Study

The use of machine learning (ML), deep learning (DL), and transfer learning (TL) to detect and classify has been the talk of the town for a while and there have been several approaches that performed successfully.
The research work [5] used CNN architectures such as VGG16, VGG19, DenseNet169, and DenseNet201 to extract image characteristics from the LC25000 dataset. The extracted features were put into six widely used ML algorithms including Extreme Gradient Boosting (XGB), Random Forest (RF), Support Vector Machine (SVM), Light Gradient Boosting (LGB), Multi-Layer Perceptron (MLP), and Logistic Regression (LR) to evaluate the performance. Accuracy-based filtering of the findings allowed for the selection of the most effective algorithms. As classifiers, SVM, Logistic Regression, and MLP were chosen because of their superior performance. Using this method, cancers of the lung and colon were found. The authors of [6] used the LC25000 dataset to automate the detection of lung and colon cancer. The pre-processing involved wavelet decomposition and application of 2D Fourier transform on channel-separated images. They used a CNN model with a Softmax classifier for feature extraction and classification tasks, achieving an accuracy of 96.37%. The research in [7] used the LC25000 dataset for classifying histopathology images of lung cancer using CNN. Feature extraction was performed using ResNet50, VGG19, Inception_ResNet_V2, and DenseNet121. The Triplet loss function was used to enhance the performance. CNN having three hidden layers was used to classify the images. In this study, Inception-ResNetv2 performed well, having a test accuracy rate of 99.7%.
The authors in the research work [8] classified adenocarcinoma of the lung region and adenocarcinoma of the colon region using a CNN model. They used the LC25000 dataset for this purpose. The images were first resized to 150 × 150 pixels, then some randomized shear and zoom transformations were applied to the images, followed by the normalization of images. A CNN model was applied separately for the lung dataset and the colon dataset, recording an accuracy of 97% and 96%, respectively. A research work was conducted to perform detection of lung cancer using CNN [9]. The dataset used during the process was LC25000. The images of the dataset were first resized to 180 × 180, and the pixel values were then transformed to a range of (0, 1) to facilitate faster convergence. The CNN model had three hidden layers. The model recorded a training accuracy of 96.11% and validation accuracy of 97.2%. The work [10] aimed at developing a model that classified lung cancer into adenocarcinoma, benign, and squamous cell carcinoma. They used the LC25000 dataset for this purpose. The model consists of a main path responsible for extracting small features and sub-paths which pass the medium- and high-level features to fully connected layers. The model recorded an accuracy of 98.53%.
The work carried out in [11] designed a model for the diagnosis of lung and colon cancer using the LC25000 dataset. A pre-trained AlexNet model, after modifying four of its layers, was used. The model performed well for all classes except the “lung_ssc” class, achieving an accuracy of 89%. To improve the performance, image enhancement techniques such as image contrast enhancement were applied to the underperforming class, improving the accuracy to 98.4%. The research work [12] used the AiCOLON dataset to train a CNN model. transfer learning was applied and results were compared with the built CNN model. The highest accuracy achieved was 96.98% with ResNet. CRC-5000, NCT-CRC-HE-100K, and merged (namely, the CRC-5000, the NCT-CRC-HE-100K) datasets were also used to test the ResNet model, recording an accuracy of 96.77%, 99.76%, and 99.98%, respectively. An accuracy of 98.66%, 99.12%, and 78.39% was achieved by SEGNET for the same datasets. SegNet was concluded to be an efficient model for cancer segmentation.
Both the LC25000 and Kather datasets were used in [13] to develop a super-lightweight plug-and-play module (namely, Pyramidal Deep-Broad Learning (PDBL)), to equip CNN backbones, especially lightweight models, to increase tissue-level classification performance without a re-training burden. Experiments were performed to equip this module on ShuffLeNetV2, EfficientNetb0, and ResNet50, with accuracy of 74.61%, 79.87%, and 85.53% with no re-training. The work [14] used the LC25000 dataset to develop a CNN model to categorize and classify the colon region for adenocarcinoma and benign cells. Lime and DeepLift were used as the optimization techniques to improve the understanding of results predicted via the model. The validation accuracy for diagnosis was found to be higher than 94% for distinguishing adenocarcinoma and benign colonic cells.
Extensive research work was carried out in [15] to utilize the power of machine learning, feature engineering, and image processing for identifying classes of lung and colon cancer. They used the LC25000 dataset in their study. Machine learning models such as XGBoost, SVM, RF, LDA, and MLP were employed. Unsharp masking was used for image preprocessing. The Recursive Feature Elimination (RFE) method was used to eliminate the least important features. XGBoost recorded an accuracy of 99%. The SHAP method was used to show the contributions of each feature in the results predicted by models. The research work [16] used CNN models with max pooling layers and average pooling layers along with MobileNetV2 for analyzing colon cancer using the LC25000 dataset. Using the ImageDataGenerator of the Keras library, images were augmented with flips, shear, zoom, rotation, and width and height range. Images were resized to 224 × 224 pixels and later converted to Numpy arrays for further use. The models were trained with various epochs and reported an accuracy of 97.49%, 95.48%, and 99.67%, respectively, for CNN models with max pooling and average pooling and MobileNetV2.
The research work in [17] developed four different CNN models, varying from two pairs of convolutional layers and max pooling layers to four pairs of convolutional layers and max pooling layers with different number of filters and kernel sizes, using lung images procured from the LC25000 dataset. Three different sizes of input images were taken into consideration. The best result accuracy on the test dataset was 96.6% with input size 768 × 768 pixels with the CNN model having four convolutional layers and max pooling layers. It was observed that, with an increase in input image size and convolutional layers, the accuracy increases. The research carried out in [18] used ResNet18, ResNet30, and ResNet50 architectures for colonic adenocarcinoma diagnosis using histopathological images. Two image datasets were used, LC25000 and CRAG. A total of 10,000 images from the first dataset were used to train and validate the CNN architectures in an 80:20 train and test set ratio. Images from the latter dataset were split into 40% and 60% for training and testing, respectively. This was performed to check how the model behaves using a self-supervised learning step, where networks were trained with the LC25000 dataset. Validation accuracy of 93.91%, 93.04%, and 93.04% was recorded for each of the three CNN architectures, respectively. The authors of [19] proposed the development of homology-based image processing techniques along with conventional texture analysis that gives better results for classification when fed into machine learning models such as Perceptron, Logistic Regression, KNN, SVM with Linear Kernel, SVM with Radial-Basis Function Kernel, Decision Tree, Random Forest, and Gradient Tree Boosting. They primarily used two datasets, private and public. The public dataset was the LC25000. Images were processed by converting them into grayscale, applying binarization, and then finally converting them into betti numbers. The accuracy was 78.33% and 99.43%, respectively, for the private and public dataset.
In the research work [20], a CNN model with three convolutional layers, having 32, 64, and 128 filters, respectively, was designed. The study aimed at the detection of breast cancer using histopathological images from the BreakHis dataset. The images were resized to 350 × 230 pixels and reshaped, maintaining the aspect ratio, achieved using OpenCV library in Python. It achieved an accuracy of 99.86%. The research work [21] used the ACDC@LUNGHP dataset to compare the performance of two prominent CNN architectures, VGG16 and ResNet50. In this case, VGG16 was seen to slightly outperform ResNet50, with an accuracy of 75.41%. The research work [22] used the PatchCamelyon benchmark dataset to compare the performance of 14 existing architectures of CNN. They concluded that DenseNet161 outperformed other architectures, with an AUC score of 0.9924. Three approaches were carried out during the process. The first approach correspondeded to utilizing the weights from the pre-trained network. The second approach related to training only fully connected layers. The last approach corresponded to training the entire model. The work [23] used the BreakHis dataset to detect Breast Cancer using histopathological images using CNN. Image augmentation was applied to improve the performance of the model and it was finally concluded with training accuracy of 96.7% and testing accuracy of 90.4%.

2. Proposed System

2.1. Principles of Diagnosis

This work primarily employs two approaches to achieve the aforementioned objective. The first approach corresponds to designing and developing convolution neural network (CNN) architectures. Two CNN architectures, one with a Softmax classifier (referred to as AdenoCanNet) and the other one with an SVM classifier (AdenoCanSVM), were developed in the first approach. The second approach corresponds to training some of the prominent architectures (such as VGG16, VGG19) using the concept of transfer learning. The study aims at understanding the performance of various architectures in diagnosing lung and colon adenocarcinoma taken together and individually using histopathological images. We also undertook a study on the performance of the architectures in different subsets of the dataset. The LC25000 dataset used in this work consists of 25,000 histopathological images, having both cancerous and normal images from both the lung and colon regions of the human body, evenly distributed over five classes. The accuracy and loss metrics were taken as the defining parameters for determining and comparing the performance of various architectures undertaken during the study. It was found that, when the entire dataset was taken into consideration, the AdenoCanNet model was found to produce the best results, recording a maximum training accuracy and maximum testing accuracy of 99.88% and 99.00%, respectively.

2.2. Dataset Description

The proposed work uses the LC25000 dataset to meet the primary objective of designing an artificial intelligence-based tool for assisting in the diagnosis of adenocarcinoma in the lung and colon regions of the human body. The dataset has five different classes. Histopathological images in two classes are taken from the colon region, while histopathological images in the remaining three classes are taken from the lung region. The dataset has a total of 25,000 histopathological images, with each class having 5000 images in them. Lung adenocarcinoma, lung squamous cell carcinoma, normal lung, colon adenocarcinoma, and normal colon are the five different classes present in the dataset. The LC25000 dataset is derived from an original dataset having 750 lung tissue images and 500 colon tissue images. These images were further augmented and finally presented by the authors of the LC25000 dataset. Figure 1 represents the different classes present in the LC25000 dataset, with one sample image taken from each class.

2.3. Preprocessing

The initial observations on the histopathological images of the dataset concluded that all the images present in the dataset were colored. It was also noted that each class in the dataset had a total of 5000 images. Careful observations of the image dataset led to the observation of their size. It was found that the images had dimensions of (768, 768, 3). The images required certain preprocessing before we could go ahead with the modeling phase. While preprocessing, the histopathological images were initially denoised to remove any kind of noise that might be present in the image. The denoising was performed by applying the Gaussian Blur. During the study, the kernel size, the SigmaX, and the SigmaY parameters of the Gaussian Blur were set to (3, 3), 90, and 90, respectively. This operation helped in eliminating the noises from every histopathological image present in the dataset. After denoising the images, they were then resized to the dimensions of (64, 64, 3). The processed images were then ready for further processing.

2.4. Architectures Used

Our study essentially employs two different approaches to meet the objective of classification. The first approach corresponds to constructing two convolution neural network (CNN) architectures. Two CNN architectures, one with a Softmax classifier (also referred to as AdenoCanNet) and the other one with an SVM classifier (also referred to as AdenoCanSVM) were developed. The second approach corresponds to training some of the prominent architectures such as VGG16, VGG19, LeNet, and ResNet50 on the LC25000. Figure 2 is a diagrammatic representation of the workflow adopted in this study.

2.4.1. Convolution neural network with Softmax classifier (AdenoCanNet)

During the initial phase of the study, a deep CNN architecture was taken into consideration while model building. However, on training and testing the model, the deep neural network was discovered to strongly overfit the dataset. The suggested model is the outcome of several actions that were taken during the process to address the issue of overfitting.
The proposed model has three convolution layers, two max pooling layers and one dropout layer followed by fully connected layers. The proposed model finally used the Softmax classifier for the classification task. Figure 3 represents a graphical form of the proposed AdenoCanNet architecture.
The architecture is designed to take an image input of dimensions (64, 64, 3). The first two layers of the proposed CNN architecture correspond to the convolution layers having 32 and 64 channels. The filter size in both these layers is set to (3, 3). A max pooling layer having a stride of (2, 2) is then included in the architecture. The architecture further includes one more convolution layer having 128 channels and with a filter size defined as (3, 3). The output from this layer is then fed into the max pooling layer, having a stride size of (2, 2). A dropout layer is then finally added before flattening it. After a series of convolution, max pooling, and dropout layers, the feature map obtained is flattened to obtain the equivalent feature vector. The feature vector is passed through a series of dense layers before finally throwing the output class of the input image using the Softmax classifier. It is pertinent to note that the architecture uses Sparse Categorical Cross Entropy as the loss function, Adam as the optimizer, and ReLU as its activation function in the entire process. The primary reason of choosing ReLU over sigmoid and hyperbolic tangent activation functions is that it is computationally less time-intensive [24]. Equation (1) corresponds to the mathematical representation of the ReLU activation function. Figure 4 represents the graphical form of representing the ReLU activation function.
ReLu :   f x = max 0 , x

2.4.2. Convolution neural network with SVM classifier (AdenoCanSVM)

During the initial phase of the study, the CNN–SVM architecture design displayed poor performance. Hence, the architecture was altered to make the CNN deeper, enabling better learning of the features of the image. However, developing a deeper neural network led to poor results. Several measures were taken to overcome the problems, including the addition of the data augmentation concept, a result of which was the proposed model.
The proposed model has three convolution layers, two max pooling layers, and one dropout layer, followed by fully connected layers. The proposed model finally used the SVM classifier for the classification task.
The images of the train and test dataset were augmented using the Keras library with the following properties. Table 1 displays the pre-processing done on the train and the test set of the dataset.
The architecture is designed to take an input image having the dimensions of (64, 64, 3). The first two layers of the proposed CNN architecture correspond to the convolution layers having 32 and 64 channels. The filter size in both the layers is set to (3, 3). A max pooling layer having a stride of (2, 2) is then included, after each layer, in the architecture. Following the max pooling layer, a dropout layer is added. The architecture then includes one more convolution layer having 16 channels and a filter size of (3, 3). The output from this layer is then fed into the max pooling layer, having a stride size of (2, 2), and ends with flattening it. After a series of convolution, max pooling, and dropout layers, the feature map obtained is then flattened to obtain the equivalent feature vector. The feature vector is fed to a series of dense layers before finally throwing the output class of the input image. It is pertinent to note that the architecture uses L2 Regularizers, Squared Hinge as the loss functions, and Adam as the optimizer. It is also noted that ReLU was used as the activation function in the entire process. Figure 5 represents a graphical form of the proposed AdenoCanSVM architecture.

2.4.3. Pre-existing Architectures

The proposed models were compared with pre-existing architectures for analyzing the performance. For this study, models such as LeNet were trained and concepts of transfer learning were applied to VGG16, VGG19, and ResNet50. Figure 6 summarizes all the existing architectures taken up in this study.
Transfer learning is a method by which models trained on one particular task are re-purposed on a different related task [25]. Transfer learning allows us to use pre-trained models as a starting point instead of building an entirely new model from scratch. These models can further be trained for the required datasets of the problem statement, saving a lot of computational power, hence improving the performance of the second model.
VGG16: VGG16 is a prominent CNN architecture, developed in 2014 [26]. The first convolution layer in the VGG architecture introduced small 3 × 3 receptive fields instead of having large 11 × 11 receptive fields found in ALexNet, and a large 7 × 7 receptive field found in ZFNet. A 3 × 3 receptive field was used throughout the network in VGG architecture. Additionally, two non-linear activation layers make the decision functions more discriminative. VGG16 has 16 weight layers and is capable of classifying 1000 objects. For this study, VGG16 was used to gather results on the dataset using a Softmax classifier.
VGG19: Similar to VGG16, VGG19 is a 19-layer deep CNN network consisting of 16 convolution layers and 3 fully connected layers [27]. Average pooling layers and dense layers were added to the architecture. The number of neurons in the final dense layer is equal to the number of classes provided for classification. Softmax was used to perform the classification.
ResNet50: ResNet50 model has 48 convolution layers, 1 max pool layer, and 1 average pool layer. In this study, ResNet50 was applied to the dataset and results were obtained using the Softmax classifier.
LeNet: LeNet-5 was introduced in 1989 and has seven layers, including three convolution layers, two subsampling layers, and two fully connected layers. The convolution kernel has a size of 5 × 5. ReLU is the activation function, and Softmax is the classifier.

3. Results and Discussion

The models were built and run using Google Colab. The Colab environment was used to train the models on a large number of data using GPU acceleration.
For the building of the models, several Python libraries were utilized to make the process more efficient and improve the performance. Tensorflow and Keras libraries were essential in building the models, compiling the models, and training the models as per the used dataset. Numpy was used to handle arrays, and OpenCV was used for reading and pre-processing the images. ImageGenerator was used for data augumentation.
In the pre-processing segment, Gaussian Blur with a kernel size of 3 × 3 was used, and SigmaX and SigmaY values were 90 in both cases. The images were initially set to 128 by 128. However, the image size was large for smoother processing in the Colab environment. Hence, the images were resized to 64 by 64.
The model initially designed severely overfit the training set. Hence, the hyperparameters were altered, and the total number of model parameters was reduced. As a result of the above process, we could finally arrive at the proposed model.
In this section, we discuss the results obtained by the CNN models the classification of lung and colon adenocarcinoma from histopathological images. The models were applied to three cases. In the first case, the models trained only on lung classes, the second case is of only colon classes, and the third case has lung and colon classes combined. In each case, 80% of the images are for training the model, and 20% are for testing.
Accuracy metrics were used to compare the results. Accuracy is a metric generally used when all the classes are equally important. Since the dataset used in this study is balanced, and all the classes have the same significance, accuracy metrics seemed reliable. Accuracy is the ratio of correct predictions to the total number of predictions. Equation (2) corresponds to the mathematical formula for the accuracy metric.
Accuracy = True   Positive   +   True   Negative True   Positive   + False   Positive   +   True   Negative   +   False   Negative
Table 2 displays the training and testing accuracies obtained for all the discussed models on a subset of the data, each class containing 500 images.
For subsets, each model was trained for 20 epochs. For the lung classes of the dataset, best performance was recorded for the ResNet50 model with the training and testing accuracies noted as 99.83% and 93.67%.
For the colon classes of the subset taken alone, LeNet-5 gave us 100% training accuracy. However, it is pertinent to note that the testing accuracy was only 79.50%, which is a case of overfitting. On the other hand, ResNet50 had a training accuracy of 99.87% and a testing accuracy of 94.50%. It is pertinent to note that VGG16 performed better than the rest of the models on the unseen data, recording the highest validation accuracy of 96.00%.
When histopathological images obtained from both lung and colon regions are analyzed, ResNet50 performed the best, with the highest training accuracy and testing accuracy of 100% and 95.40%.
The pre-existing architectures were found to perform better than the newly proposed system when the number of data used for training the model was drastically reduced.
It is pertinent to note that the concept of data augmentation was not implemented in any model run for 20 epochs.
All the models were then trained on the entire dataset i.e., 25,000 images, evenly distributed across five classes. The same train–test split of 80:20 was used. The models were trained for 50 as well as 100 epochs.
Table 3 displays the results obtained when the models were trained for 50 epochs.
For 50 epochs, in the case of histopathological images of only lungs, both ResNet50 and LeNet were found to record the highest training accuracy of 100%, and the validation accuracy of ResNet50 and LeNet was 97.67% and 94.97%. It is pertinent to note that the training and testing accuracies of the AdenoCanNet were 99.87% and 98.80%. The validation accuracy reported by AdenoCanNet architecture was the highest among the other models, indicating its better performance on unseen data. Hence, it can be concluded that AdenoCanNet performs better in the case of histopathological images in the lung region.
For colon images, as visible from Table 3, ResNet50 and LeNet performed the best on the training dataset with 100% training accuracy. The validation accuracy of ResNet and LeNet was 98.90% and 97.40%. It is pertinent to note that the training and validation accuracy achieved by AdenoCanNet were 99.99% and 99.90%. With the training and testing accuracy of AdenoCanNet being almost 100%, it can be concluded that AdenoCanNet performs better in the case of histopathological images obtained from the colon region.
When histopathological images obtained from both lung and colon regions are taken into consideration, ResNet performed well, the training and testing accuracy being 100%. However, the proposed architecture, AdenoCanNet, also recorded a training accuracy of 99.88% and validation accuracy of 98.90%, comparable with the performance of ResNet attained on the complete dataset.
Figure 7 and Figure 8 correspond to the accuracy and loss graphs of AdenoCanNet trained for 50 epochs on different sets of the LC25000.
Table 4 discusses the results obtained when the models were trained for 100 epochs.
The results obtained for the models trained on histopathological images in the lung dataset run for 100 epochs show that ResNet50 and LeNet models achieved the highest training accuracy of 100%. The validation accuracy of ResNet and LeNet models correspond to 97.60% and 96.00%. It is worth noting that, among all the pre-existing architectures, VGG16 displayed a better performance on the validation dataset, achieving an accuracy of 97.20%. It is pertinent to note that the accuracy achieved by AdenoCanNet achieved a training accuracy of 99.96% and a validation accuracy of 98.77%. Among all the models, the suggested AdenoCanNet performs better on the unseen validation dataset. The training accuracy achieved by the CNN model is comparable with the performance of the ResNet and the LeNet models.
For the colon subset taken alone, the suggested AdenoCanNet with LeNet and ResNet recorded the highest accuracy of 100%. The validation accuracy obtained using the mentioned AdenoCanNet was the highest among all the other models, recording a value of 99.80%. It is worth noting that the proposed model of AdenoCanSVM also recorded a good training and validation accuracy of 95.46% and 99.40%. An accuracy of 98.50% and 99.10% reported for training and testing sets for the VGG16 model makes it another good model.
When histopathological images obtained from both lung and colon regions are taken into consideration, ResNet performed well, with both training and testing accuracy being 100%. However, AdenoCanNet attained a training accuracy of 99.88% and validation accuracy of 99.00%, which is comparable with the performance of ResNet on the complete dataset. Figure 9 and Figure 10 correspond to the accuracy and loss graphs of AdenoCanNet trained for 100 epochs on different sets of the LC25000.
With the entire dataset into consideration, ResNet50 performs better in many cases. However, it is pertinent to note that ResNet50 is a 50-layer deep neural network, and training this model requires intense computation and time. The proposed AdenoCanNet consists of three convolution layers, two max pooling layers, and one dropout layer. The accuracies of this model and ResNet turn out to be comparable, and in some cases, AdenoCanNet outperforms the ResNet50 model.
When the models were trained on the entire dataset, the testing accuracy of AdenoCanNet outperformed the ResNet50 model when only lung and colon classes were considered.
LeNet-5 also displayed substantial results, however the AdenoCanNet outperformed the model in terms of testing accuracy for all cases when the entire dataset is considered.
For VGG16 and VGG19, the performance recorded was greatly comparable to the other architectures mentioned; however, AdenoCanNet outperformed both the pre-existing models in terms of training and testing accuracy in the case of the entire dataset.
Table 5 compares the results of the existing work on the detection and classification of lung and colon adenocarcinoma. Most works have opted for the LC25000 dataset and opted for CNN models; hence, these works are directly comparable.
For lung classes taken alone, Sanidhya Mangal et al. achieved an accuracy of 97%, and for colon classes alone Zarrin Tasnim et al. achieved a 96.6% accuracy rate. For CNN models experimented on the LC25000 dataset, the highest recorded accuracy turned out to be 99.8% achieved by Ben Hamida et al. With the obtained results and in comparison to the works discussed in the earlier section of this study, it can be concluded that, when considering the entire dataset of 25,000 histopathological images, AdenoCanNet introduced in this study performs with high accuracy and reliability, outperforming most of the existing models.
Figure 11, Figure 12 and Figure 13 correspond to performance comparison plots for lung, colon, and lung and colon regions of the LC25000 dataset. As observed from the above comparisons and explanation, the AdenoCanNet model exhibited a better performance for the entire dataset for each category considered. Some recent works have used relatively deep neural networks or filtering multiple machine learning models based on accuracy obtained from them. The proposed model, having just three convolution layers, two max pooling layers, and one dropout layer, with minimum image preprocessing, achieved about 99% validation accuracy and 99.88% accuracy for detection and classification on the entire dataset of both lung and colon tissues. A 99.96% accuracy and 98.77% validation accuracy on the entire dataset of lung tissues were achieved, as well as 100% accuracy and 99.80% validation accuracy on the entire dataset of colon tissues with 100 epochs.

4. Conclusions

In this study, a deep learning-based supervised learning model was developed and used for the classification of histopathological images obtained from the lung and colon regions of the human body. We introduced AdenoCanNet and AdenoCanSVM to perform the classification. The experimentation used the LC25000 lung and colon histopathological dataset. Results from pre-existing architectures such as VGG16, VGG19, LeNet-5, and ResNet50 were compared with the results of the devised models. The AdenoCanNet proved to outperform the pre-existing models in most of the cases. The experiment performed for only lung classes gave a testing accuracy of 98.77% for AdenoCanNet. For the colon dataset, the testing accuracy obtained was 99.80%, and for the combined dataset, the testing accuracy was 99.00%.
With only a limited number of layers in the model, high accuracy was achieved, proving to perform better than most of the existing works.
With outstanding results obtained in diagnosing adenocarcinoma in both the lung and colon regions of the human body, the model may now be extended to test the accuracy in diagnosing adenocarcinoma in other regions of the body too. This work could further be deployed as a website as an easy-to-use platform. Utilization of this work would reduce effort, computation, and time.
There is a need for computer-aided diagnosis systems for adenocarcinoma detection from histopathological images. From this study, we concluded that, for small image datasets, existing architectures of deep learning are better, whereas for larger datasets, the proposed models work better.

Author Contributions

Conceptualization, S.C., B.A. and M.S.K.; methodology, S.C., B.A. and A.S.; proposed architecture design and implementation: S.C.; software, S.C., V.S. and D.P.; validation, B.A., M.S.K., A.S. and S.C.; formal analysis, S.C.; investigation, S.C., V.S. and D.P.; resources, S.C.; data curation, S.C.; writing—original draft preparation, S.C., V.S. and D.P.; writing—review and editing, B.A.; visualization, S.C., V.S. and D.P.; supervision, B.A.; project administration, B.A. and S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this work are available in Kaggle.

Acknowledgments

The authors wish to express their thanks to Vellore Institute of Technology (VIT) management and Centre for Cyber Physical Systems, VIT Chennai for their extensive support during this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. Available online: https://www.who.int/news-room/fact-sheets/detail/cancer (accessed on 14 December 2022).
  2. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  3. Cleveland Clinic. Available online: https://my.clevelandclinic.org/health/diseases/21652-adenocarcinoma-cancers#:~:text=Adenocarcinoma%20is%20a%20type%20of,Cancer%20Answer%20Line%20866.223.8100 (accessed on 30 May 2022).
  4. Borkowski, A.A.; Bui, M.M.; Thomas, L.B.; Wilson, C.P.; DeLand, L.A.; Mastorides, S.M. Lung and Colon Cancer Histopathological Image Dataset (LC25000). arXiv 2019. [Google Scholar] [CrossRef]
  5. Talukder, M.A.; Islam, M.M.; Uddin, M.A.; Akhter, A.; Hasan, K.F.; Moni, M.A. Machine Learning-Based Lung and Colon Cancer Detection using Deep Feature Extraction and Ensemble Learning. Expert Syst. Appl. 2022, 205, 117695. [Google Scholar] [CrossRef]
  6. Masud, M.; Sikder, N.; Nahid, A.-A.; Bairagi, A.K.; AlZain, M.A. A Machine Learning Approach to Diagnosing Lung and Colon Cancer using a Deep Learning-Based Classification Framework. Sensors 2021, 21, 748. [Google Scholar] [CrossRef] [PubMed]
  7. Baranwal, N.; Doravari, P.; Kachhoria, R. Classification of Histopathology Images of Lung Cancer Using Convolutional Neural Network (CNN). arXiv 2021. [Google Scholar] [CrossRef]
  8. Mangal, S.; Chaurasia, A.; Khajanchi, A. Convolution Neural Networks for Diagnosing Colon and Lung Cancer Histopathological Images. arXiv 2020. [Google Scholar] [CrossRef]
  9. Hatuwal, B.K.; Thapa, H.C. Lung Cancer Detection using Convolutional Neural Network on Histopathological Images. Int. J. Comput. Trends Technol. 2020, 68, 21–24. [Google Scholar] [CrossRef]
  10. Saif, A.; Qasim, Y.R.H.; Al-Sameai, H.A.M.; Ali, O.A.F.; Hassan, A.A.M. Multi Paths Technique on Convolutional Neural Network for Lung Cancer Detection Based on Histopathological Images. Int. J. Adv. Netw. Appl. 2020, 12, 4549–4554. [Google Scholar] [CrossRef]
  11. Mehmood, S.; Ghazal, T.M.; Khan, M.A.; Zubair, M.; Naseem, M.T.; Faiz, T.; Ahmad, M. Malignancy Detection in Lung and Colon Histopathology Images Using Transfer Learning with Class Selective Image Processing. IEEE Access 2020, 10, 25657–25668. [Google Scholar] [CrossRef]
  12. Hamida, A.B.; Devanne, M.; Weber, J.; Truntzer, C.; Derangère, V.; Ghiringhelli, F.; Forestier, G.; Wemmert, C. Deep Learning for Colon Cancer Histopathological Images Analysis. In Computers in Biology and Medicine; Elsevier: Amsterdam, The Netherlands, 2021. [Google Scholar] [CrossRef]
  13. Lin, J.; Han, G.; Pan, X.; Liu, Z.; Chen, H.; Li, D.; Jia, X.; Shi, Z.; Wang, Z.; Cui, Y.; et al. PDBL: Improving Histopathological Tissue Classification with Plug-and-Play Pyramidal Deep-Broad Learning. arXiv 2021, arXiv:2111.03063. [Google Scholar] [CrossRef] [PubMed]
  14. Hossain, M.; Haque, S.S.; Ahmed, H.; Mahdi, H.A.; Aich, A. Early Stage Detection and Classification of Colon Cancer using Deep Learning and Explainable AI on Histopathological Images. Ph.D. Thesis, Brac University, Dhaka, Bangladesh, 2022. Available online: http://hdl.handle.net/10361/16671 (accessed on 1 June 2022).
  15. Hage Chehade, A.; Abdallah, N.; Marion, J.M.; Oueidat, M.; Chauvet, P. Lung and Colon Cancer Classification using Medical Imaging: A Feature Engineering Approach. Phys. Eng. Sci. Med. 2021, in press. [Google Scholar] [CrossRef] [PubMed]
  16. Tasnim, Z.; Chakraborty, S.; Shamrat, F.M.J.M.; Chowdhury, A.N.; Nuha, H.A.; Karim, A.; Zahir, S.B.; Billah, M.M. Deep Learning Predictive Model for Colon Cancer Patient using CNN-based Classification. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 687–696. [Google Scholar] [CrossRef]
  17. Hlavcheva, D.; Yaloveha, V.; Podorozhniak, A.; Kuchuk, H. Comparison of CNNs for Lung Biopsy Image Classification. In Proceedings of the IEEE Ukraine Conference on Electrical and Computer Engineering (UKRCON), Lviv, Ukraine, 26–28 August 2021. [Google Scholar] [CrossRef]
  18. Bukhari, S.U.K.; Syed, A.; Bokhari, S.K.A.; Hussain, S.S.; Armaghan, S.U.; Shah, S.S.H. The Histological Diagnosis of Colonic Adenocarcinoma by Applying Partial Self Supervised Learning. MedRxiv 2020. [Google Scholar] [CrossRef]
  19. Nishio, M.; Nishio, M.; Jimbo, N.; Nakane, K. Homology-based Image Processing for Automatic Classification of Histopathological Images of Lung Tissue. Cancers 2021, 13, 1192. [Google Scholar] [CrossRef] [PubMed]
  20. Dabeer, S.; Khan, M.M.; Islam, S. Cancer diagnosis in histopathological image: CNN based approach. Inform. Med. Unlocked 2019, 16, 100231. [Google Scholar] [CrossRef]
  21. Šarić, M.; Russo, M.; Stella, M.; Sikora, M. CNN-based Method for Lung Cancer Detection in Whole Slide Histopathology Images. In Proceedings of the 4th International Conference on Smart and Sustainable Technologies (SpliTech), Bol and Split, Croatia, 18–21 June 2019. [Google Scholar] [CrossRef]
  22. Aneja, S.; Aneja, N.; Abas, P.E.; Naim, A.G. Transfer learning for cancer diagnosis in histopathological images. Int. J. Artif. Intell. 2022, 11, 129–136. [Google Scholar] [CrossRef]
  23. Maan, J.; Maan, H. Breast Cancer Detection using Histopathological Images. Int. J. Comput. Sci. Trends Technol. 2022. [Google Scholar] [CrossRef]
  24. DeepAI. Available online: https://deepai.org/machine-learning-glossary-and-terms/relu#:~:text=ReLu%20is%20a%20non%2Dlinear,zero%20and%20the%20input%20value (accessed on 1 June 2022).
  25. A Gentle Introduction to Transfer Learning for Machine Learning. Available online: https://machinelearningmastery.com/transfer-learning-for-deep-learning/ (accessed on 1 June 2022).
  26. Tammina, S. Transfer learning using VGG-16 with Deep Convolutional Neural Network for Classifying Images. Int. J. Sci. Res. Publ. 2019, 9, 9420. [Google Scholar] [CrossRef]
  27. Machine Learning Blog. Available online: https://blog.techcraft.org/vgg-19-convolutional-neural-network/ (accessed on 2 June 2022).
Figure 1. Dataset structure.
Figure 1. Dataset structure.
Sustainability 15 01399 g001
Figure 2. Workflow of the proposed system.
Figure 2. Workflow of the proposed system.
Sustainability 15 01399 g002
Figure 3. AdenoCanNet architecture.
Figure 3. AdenoCanNet architecture.
Sustainability 15 01399 g003
Figure 4. Graph of ReLU activation function.
Figure 4. Graph of ReLU activation function.
Sustainability 15 01399 g004
Figure 5. AdenoCanSVM architecture diagram.
Figure 5. AdenoCanSVM architecture diagram.
Sustainability 15 01399 g005
Figure 6. Existing-architecture implementation diagram.
Figure 6. Existing-architecture implementation diagram.
Sustainability 15 01399 g006
Figure 7. Accuracy graphs of AdenoCanNet trained for 50 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Figure 7. Accuracy graphs of AdenoCanNet trained for 50 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Sustainability 15 01399 g007
Figure 8. Loss graphs of AdenoCanNet trained for 50 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Figure 8. Loss graphs of AdenoCanNet trained for 50 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Sustainability 15 01399 g008
Figure 9. Accuracy graphs of AdenoCanNet trained for 100 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Figure 9. Accuracy graphs of AdenoCanNet trained for 100 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Sustainability 15 01399 g009
Figure 10. Loss graphs of AdenoCanNet trained for 100 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Figure 10. Loss graphs of AdenoCanNet trained for 100 epochs on: (a) lung dataset; (b) colon dataset; (c) both lung and colon dataset taken together.
Sustainability 15 01399 g010aSustainability 15 01399 g010b
Figure 11. Performance-comparison plot for models trained on the lung dataset: (a) accuracy comparison; (b) validation accuracy comparison.
Figure 11. Performance-comparison plot for models trained on the lung dataset: (a) accuracy comparison; (b) validation accuracy comparison.
Sustainability 15 01399 g011
Figure 12. Performance-comparison plot for models trained on the colon dataset: (a) accuracy comparison; (b) validation accuracy comparison.
Figure 12. Performance-comparison plot for models trained on the colon dataset: (a) accuracy comparison; (b) validation accuracy comparison.
Sustainability 15 01399 g012
Figure 13. Performance-comparison plot for models trained on the lung and colon dataset: (a) accuracy comparison; (b) validation accuracy comparison.
Figure 13. Performance-comparison plot for models trained on the lung and colon dataset: (a) accuracy comparison; (b) validation accuracy comparison.
Sustainability 15 01399 g013
Table 1. Pre-processing done on the dataset.
Table 1. Pre-processing done on the dataset.
Train DatasetTest Dataset
Rescale = 1/255
Rescale = 1/255
Zoom Range = 0.2
Zoom Range = 0.2
Horizontal Flip = True
Table 2. Accuracy comparison of models for one subset (20 epochs).
Table 2. Accuracy comparison of models for one subset (20 epochs).
AdenoCanNet
(CNN W/Softmax)
AdenoCanSVM
(CNN W/SVM)
VGG16VGG19LENETRESNET50
Lung85.92
89.00
33.75
34.00
95.25
92.67
94.25
89.33
99.75
90.00
99.83
93.67
Colon51.50
60.00
51.88
43.50
97.25
96.00
98.62
95.00
100
79.50
99.87
94.50
Lung and Colon75.15
78.00
60.25
59.20
93.30
92.00
92.55
92.20
99.56
92.64
100
95.40
Table 3. Accuracy comparison of models for the entire dataset (50 epochs).
Table 3. Accuracy comparison of models for the entire dataset (50 epochs).
AdenoCanNet
(CNN W/Sotmax)
AdenoCanSVM
(CNN W/SVM)
VGG16VGG19LENETRESNET50
Lung99.87
98.80
94.55
95.70
95.57
96.57
93.79
94.67
100
94.97
100
97.67
Colon99.99
99.90
88.87
68.90
97.45
98.60
96.22
96.95
100
97.40
100
98.90
Lung and Colon99.88
98.90
92.02
76.59
94.76
96.36
92.27
94.32
100
96.86
100
100
Table 4. Accuracy comparison of models for the entire dataset (100 epochs).
Table 4. Accuracy comparison of models for the entire dataset (100 epochs).
AdenoCanNet
(CNN W/Softmax)
AdenoCanSVM
(CNN W/SVM)
VGG16VGG19LENETRESNET50
Lung99.96
98.77
95.63
96.26
96.76
97.20
95.96
95.90
100
96.00
100
97.60
Colon100
99.80
95.46
99.40
98.50
99.10
97.46
98.05
100
94.10
100
99.05
Lung and Colon99.88
99.00
94.52
79.14
96.50
97.10
94.36
95.70
100
96.84
100
100
Table 5. Comparison of existing works.
Table 5. Comparison of existing works.
AuthorDatasetModelAccuracy
Md. Alamin Talukder et al.LC25000CNN (Hybrid Model)96.37%
Neha Baranwal et al.LC25000CNN (Inception-ResNetv2)99.7%
Sanidhya Mangal et al.LC25000CNN97% (Lung)
96% (Colon)
Bijaya Hatuwal et al.LC25000CNN97.2%
Amin Saif et al.LC25000CNN98.53%
Mehmood et al.LC25000CNN (AlexNet)98.4%
Ben Hamida et al.AiCOLON CNN (ResNet)99.8%
Jiatai Lin et al. LC25000 and KatherCNN (ResNet)85.53%
Mainul Hossain et al.LC25000CNN94% (Colon)
Aya Hage Chehade et al.LC25000XGBoost99%
Zarrin Tasnim et al.LC25000CNN 99.67% (Colon)
Daria Hlavcheva et al.LC25000CNN96.6% (Lung)
Syed Usama Khalid Bukhari et al.LC25000 and CRAGCNN (ResNet)93.91%
Md. Alamin Talukder et al.LC25000CNN99.30%
Mizuho Nishio et al. LC25000Hybrid ML models99.43%
Sumaiya Dabeer et al.BreakHisCNN99.86%
M. Šarić et al.ACDC@LUNGHPCNN (VGG16)75.41%
AdenoCanNet
(Proposed Model)
LC25000
(Lung)
CNN-Softmax98.77%
AdenoCanNet
(Proposed Model)
LC25000
(Colon)
CNN-Softmax99.8%
Proposed ModelLC25000CNN (ResNet50)100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ananthakrishnan, B.; Shaik, A.; Chakrabarti, S.; Shukla, V.; Paul, D.; Kavitha, M.S. Smart Diagnosis of Adenocarcinoma Using Convolution Neural Networks and Support Vector Machines. Sustainability 2023, 15, 1399. https://doi.org/10.3390/su15021399

AMA Style

Ananthakrishnan B, Shaik A, Chakrabarti S, Shukla V, Paul D, Kavitha MS. Smart Diagnosis of Adenocarcinoma Using Convolution Neural Networks and Support Vector Machines. Sustainability. 2023; 15(2):1399. https://doi.org/10.3390/su15021399

Chicago/Turabian Style

Ananthakrishnan, Balasundaram, Ayesha Shaik, Shubhadip Chakrabarti, Vaishnavi Shukla, Dewanshi Paul, and Muthu Subash Kavitha. 2023. "Smart Diagnosis of Adenocarcinoma Using Convolution Neural Networks and Support Vector Machines" Sustainability 15, no. 2: 1399. https://doi.org/10.3390/su15021399

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop