Next Article in Journal
Efficient Thorax Disease Classification and Localization Using DCNN and Chest X-ray Images
Previous Article in Journal
Antibody Response to SARS-CoV-2 Vaccination in Heart Failure Patients: Retrospective Single-Center Cohort Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Framework with an Intermediate Layer Using the Swarm Intelligence Optimizer for Diagnosing Oral Squamous Cell Carcinoma

by
Bharanidharan Nagarajan
1,
Sannasi Chakravarthy
2,
Vinoth Kumar Venkatesan
1,
Mahesh Thyluru Ramakrishna
3,*,
Surbhi Bhatia Khan
4,5,6,*,
Shakila Basheer
7 and
Eid Albalawi
8
1
School of Computer Science Engineering and Information Systems (SCORE), Vellore Institute of Technology, Vellore 632014, India
2
Department of ECE, Bannari Amman Institute of Technology, Sathyamangalam 638401, India
3
Department of Computer Science and Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-Be University), Bangalore 562112, India
4
Department of Data Science, School of Science Engineering and Environment, University of Salford, Manchester M5 4WT, UK
5
Department of Engineering and Environment, University of Religions and Denominations, Qom 13357, Iran
6
Department of Electrical and Computer Engineering, Lebanese American University, Byblos P.O. Box 13-5053, Lebanon
7
Department of Information Systems, College of Computer and Information Science, Princess Nourah bint Abdulrahman University, Riyadh 11671, Saudi Arabia
8
Department of Computer Science, School of Computer Science and Information Technology, King Faisal University, Al-Ahsa 31982, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(22), 3461; https://doi.org/10.3390/diagnostics13223461
Submission received: 4 October 2023 / Revised: 7 November 2023 / Accepted: 8 November 2023 / Published: 16 November 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
One of the most prevalent cancers is oral squamous cell carcinoma, and preventing mortality from this disease primarily depends on early detection. Clinicians will greatly benefit from automated diagnostic techniques that analyze a patient’s histopathology images to identify abnormal oral lesions. A deep learning framework was designed with an intermediate layer between feature extraction layers and classification layers for classifying the histopathological images into two categories, namely, normal and oral squamous cell carcinoma. The intermediate layer is constructed using the proposed swarm intelligence technique called the Modified Gorilla Troops Optimizer. While there are many optimization algorithms used in the literature for feature selection, weight updating, and optimal parameter identification in deep learning models, this work focuses on using optimization algorithms as an intermediate layer to convert extracted features into features that are better suited for classification. Three datasets comprising 2784 normal and 3632 oral squamous cell carcinoma subjects are considered in this work. Three popular CNN architectures, namely, InceptionV2, MobileNetV3, and EfficientNetB3, are investigated as feature extraction layers. Two fully connected Neural Network layers, batch normalization, and dropout are used as classification layers. With the best accuracy of 0.89 among the examined feature extraction models, MobileNetV3 exhibits good performance. This accuracy is increased to 0.95 when the suggested Modified Gorilla Troops Optimizer is used as an intermediary layer.

1. Introduction

Any neighbor tissue impairment due to uncontrolled cell growth and invasion is called cancer. Oral cancer is ranked as the sixth most prevailing cancer globally, and it falls under the broad category of head and neck cancer. Oral cancer results in malignant cell growth in the lips and various parts of the oral cavity. Worldwide, it is ranked as the fifteenth most common reason for death among various types of cancer. Out of one hundred thousand people, a minimum of four people are affected by this disease across the globe [1,2]. Approximately, seventy-seven new cases and fifty-two thousand deaths are registered every year in India, and one-fourth of the global oral cancer occurs in India [3]. In 2018, around 355,000 oral cancer cases occurred worldwide and resulted in 177,000 deaths. Estimates for the year 2020 include about 53,260 new cases added to the previous year’s cases, and the estimated death toll from this cancer in 2020 was about 10,750 deaths more than previous years [4,5].
The most common types of oral cancer include oral squamous cell carcinoma (OSCC), verrucous carcinoma, minor salivary gland carcinomas, lymphoma, and mucosal melanoma. Among them, OSCC is a predominant type of oral cancer which contributes around 84–97% of oral cancer cases [6]. The major risk factors that lead to the development of OSCC include tobacco usage, frequent chewing of betel quid, alcohol intake, oral infection, and genetic disorders [7]. Detection of OSCC at an early stage is very crucial to avoid deaths since the five-year survival rate of humans with early-stage OSCC is around 85%, while it is only around 40% with advanced stage [8,9]. Hence, early detection is the key to reducing the mortality rate and so there is a huge demand for diagnostic tools that identify OSCC at earlier stages of malignancy.
Apart from physical examination, major diagnostic tools used for the identification of oral cancer include techniques such as endoscope biopsy, liquid biopsy, the vital staining technique, ultrasound imaging, Magnetic Resonance Imaging (MRI), Computed Tomography (CT) imaging, Raman spectroscopy, gene/DNA array-based biomarker detection, enzyme assay-based biomarker detection, and histopathological examination [6]. Among these techniques, histopathologic examination is mainly preferred since it can be used to detect both malignant and benign tumors by identifying the changes in histopathological and molecular levels. Histological assays can be used to reveal the gradual growth of malignant cells in the oral cavity beginning from elementary dysplasia to tumors with a highly invasive nature. It helps analyze cell proliferation, growth of abnormalities, cytoplasmic-level and cellular-level atypia, changes at the surface of the epithelium, and deep tissue-level cytoarchitecture [10]. Usually, abnormalities at the microscopic and clinical levels arise only after abnormalities at the molecular and genetic levels. Histopathological examination is good at capturing these molecular-level changes and so is preferred for early detection [11].
Analysis of histopathological images with visual inspection is usually subjective and prone to errors sometimes. Particularly, the sensitivity and specificity measures will be very poor in the visual inspection of histopathological images when compared with computerized diagnostic tools. Sometimes, human inspection is biased so that many patients are identified as having OSCC even though they do not possess it, which makes the specificity measure low. In some other cases, the clinician may be biased to not detect a diseased person properly, which results in poor sensitivity. Computerized diagnostic tools will be very helpful to assist clinicians in the decision-making process to reduce such errors. Various machine learning (ML) techniques are used nowadays in a variety of fields. Particularly in the healthcare field, the implementation of ML algorithms is increasing day by day. Accuracy and robustness are the key concerns in such healthcare-related decision-making tools. Fortunately, nowadays, deep learning (DL) models are available for solving these issues. Deep learning is a sub-field in machine learning where Artificial Neural Network (ANN) models with many numbers of hidden layers are trained with a large set of training images and labels; labels of new unseen images are predicted using the trained model. The main advantage of deep learning is that it does not require hand-crafted feature engineering, which is required by traditional supervised classifiers such as Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Decision Trees (DT), etc., where domain experts are required to identify the appropriate features and Region of Interest (RoI) [12,13].
Convolutional Neural Networks (CNNs) are a popular deep learning technique where convolution operation is involved in multiple ANN layers. Various CNN architectures have been developed, and they are very efficient in different image classification tasks [14,15]. Popular CNN architectures include ResNet, EffiecientNet, InceptionNet, MobileNet, etc. The main advantage of these architectures is their ability to work well on a classification task even if most of their weights are pre-trained on another classification task. This concept is known as transfer learning and works very well for two similar and unique classification tasks. The two main advantages of transfer learning are a reduction in training time and competence to work well on small datasets [16,17].
To improve the accuracy of such deep learning models, various techniques are used such as fine-tuning, feature selection, regularization, optimal parameter selection, optimization, etc. On the other hand, various population-based swarm intelligence (SI) optimization algorithms are widely used for optimal parameter identification, weight updating, and feature selection in deep learning models for enhancing accuracy. SI algorithms are meta-heuristic iterative algorithms that are usually inspired by the characteristics and nature of the swarm of animals. These algorithms are preferred in many applications, mainly due to their minimalism, derivation-free design, and ability to avoid local optima [18]. Some of the popular SI algorithms include Particle Swarm Optimization (PSO), Ant Colony Optimization, the Grey Wolf Optimizer, Dragonfly Optimization, Elephant Herding Optimization, the Gorilla Troops Optimizer (GTO), etc.
This work primarily focuses on classifying histopathological images into two categories: Normal and OSCC. Histopathological image features are extracted using pre-trained weights of transfer learning-based popular CNN models namely, InceptionV2, MobileNetV3, and EfficientNetB3. Then, the Modified Gorilla Troops Optimizer (MGTO) is used as an intermediate layer in between feature extraction and classification layers. Two fully connected ANN layers, batch normalization, and dropout are used as classification layers.
The key contributions of this research work are listed below:
  • The proposal of a novel deep learning framework that includes a swarm intelligence-based optimization algorithm as an intermediate layer in the deep learning model.
  • The development of MGTO with appropriate modifications that enhance classification accuracy.
  • A comparative analysis of popular deep learning models with and without the proposed intermediate layer in terms of various classification metrics and training times.
The remainder of this paper is organized as follows: Section 2 deals with related works, and Section 3 is related to the background of various transfer-learning models used and the original GTO. Section 4 deals with the methodology used in this research work, and Section 5 presents the implementation procedure for the proposed MGTO as an intermediate layer. Results are presented and discussed in Section 6. The last section summarizes the conclusion and future work.

2. Related Work

Various techniques based on machine learning and deep learning are proposed in the literature to diagnose oral cancer by analyzing medical images. Early publications related to oral cancer diagnosis mainly use feature extraction and traditional supervised classifiers [19,20,21,22]. For example, Krishnan et al. [23] considered features based on texture discrimination using higher order spectra, laws texture energy, and local binary pattern and fed these features to supervised classifiers such as DT, the Gaussian Mixture Model, KNN, the Sugeno Fuzzy Classifier, and the Radial Basis Probabilistic Neural Network. Similarly, Thomas et al. [24] proposed textural change detection using features extracted from digital images of oral lesions using a grey-level cooccurrence matrix and grey-level run length matrix. They used back a propagation-based ANN for classification. Particularly for OSCC diagnosis, Rahman et al. [25,26] proposed texture, shape, and color feature extraction from histopathological images and classification using DT, SVM, and Logistic Regression.
The usage of deep learning models in medical image analysis is increasing rapidly, particularly from the last decade onwards. Various deep learning models are developed and tested for oral cancer diagnosis that involve both binary and multi-class classification. In [27], the authors investigated a customized AlexNet model for detecting OSCC from histopathological images. In [28], the authors applied the DenseNet121 model to oral biopsy images to detect OSCC and found that it performs better than regions with the CNN (R-CNN) model. Other transfer learning models such as Inception-ResNet-V2 [29], Xception [30], and ResNet101 [31] were also investigated for diagnosing oral cancer from medical images. Apart from the above-mentioned works where popular CNN architectures are investigated, some works propose their own CNN model for detecting oral cancer. For example, Lin et al. [32] proposed the HRNet model for diagnosing malignant lesions in oral cavities and compared it with popular ResNet50 and DenseNet169 models. In [33], the authors developed a modified CNN model that performs well when compared with transfer learning-based models such as Resnet-50, VGG-16, VGG-19, and Alexnet. Similarly, Das et al. [31] proposed their own ten-layer CNN model that outperforms pre-trained CNN models in the diagnosis of OSCC from histopathological images. Other than CNN and its variants, capsule networks are also implemented in some works to identify oral malignancy. In [34], the authors tested performance capsule networks to identify OSCC from histopathological images.
Many optimization algorithms are used in many applications to enhance the classification performance and robustness of deep learning models, and some of them are outlined below. A hybrid optimization algorithm was developed that mixes PSO and Al-Biruni Earth Radius Optimization for optimizing the design parameters of CNNs and Deep Belief Networks in malignant oral lesion identification [35]. The segmentation of psoriasis skin images using Adaptive Golden Eagle Optimization was implemented for finding the ideal weight and bias parameters of CNNs [36]. The Artificial Bee Colony optimization algorithm was considered for finding the optimal hyper-parameters of a CNN that worked as a classifier for identifying plant species [37]. The optimal guidance-whale optimization algorithm was used to select features extracted from an AlexNet–ResNet50 model, and the selected features were supplied to bi-directional long short-term memory for Land Use Land Cover classification [38]. Modified Lion Optimization was implemented for selecting the optimal features in a transfer learning-based CNN classification model to build a multimodal biometric system [39]. In this manner, numerous optimization algorithms are incorporated for finding optimal hyper-parameters, training models, and feature selection in deep learning. Comparatively, only a few works are reported regarding the usage of optimization algorithms as a transformation technique. For example, the Crow search optimization algorithm is used was a transformation technique for improving the classification performance of weighted KNN in the severity classification of breast cancer [40].
From the above-related works, the following points can be summarized. Compared with hand-crafted feature extraction and traditional supervised classifiers, deep learning models perform well in the diagnosis of OSCC. But still, they lag in classification accuracy and robustness. To solve these two concerns, optimization algorithms are widely used in various applications for the improvisation of deep learning models in different ways. Hence, this work attempts to use the MGTO optimization algorithm as an intermediate layer between feature extraction and classification layers for enhancing the accuracy of OSCC diagnosis.

3. Background

3.1. CNN

CNN [41]-based deep learning models are widely used to classify images in a variety of applications, mainly due to their capability of recognizing an underlying pattern. Convolution operation at multiple layers acts as the foundation for CNN and generally, a typical CNN contains convolutional layers, pooling layers, and fully connected layers. The goal of convolution layers is to extract the image attributes such as contours, colors, etc. Pooling layers act as a dimensionality reduction layer, i.e., the layer that reduces the number of features. Max and average pooling layers are very popular when compared with others. The last stage is usually built using a fully connected layer called DenseNet, and it is responsible for classification [42].

3.2. InceptionV2

The inception [43] model is an altered version of a CNN in which inception blocks are included. These inception blocks refer to the processing of the same input with different filter sizes before combining them. InceptionV2 is an advanced variant of the original InceptionV1. When compared with Inception V1, two 3 × 3 convolution operations are performed in Inception V2 instead of one 5 × 5 convolution operation. In addition, the filter size n × n is factorized into 1 × n and n × 1 convolutions in Inception V2.

3.3. MobileNetV3

MobileNet [44] is a modified version of a CNN where batch normalization and ReLU activation functions are used instead of a single 3 × 3 convolution layer. In addition, one convolution operation is carried out for each color channel in MobileNet, while the flattening of color channels will happen in typical CNNs. Relatively, MobileNet architectures require minimal computational power and so are mainly preferred in mobile devices and embedded systems. Compared with MobileNetV1, bottlenecks with residuals are implemented in MobileNetV2, while layer removal and swish non-linearity are incorporated in MobileNetV3.

3.4. EfficientNetB3

Unlike typical CNNs, EfficientNet [45] uniformly scales all dimensions with a compound coefficient. A fixed set of scaling coefficients is used to uniformly scale the network depth, width, and resolution. The original EfficientNetB0 version is based on MobileNetV2 combined with squeeze and excitation blocks. EffientNetB3 is developed by scaling up the baseline network of previous versions.

3.5. Gorilla Troops Optimization

GTO is one of the iterative meta-heuristic optimization algorithms that was proposed in the year 2021 [46]. It is based on the social activities and characteristics of a gorilla troop. Usually, each such troop contains one adult male gorilla, called a silverback gorilla, a substantial number of adult female gorillas, and their children. The male gorilla leads the troop, and it is responsible for controlling the troop activities such as identification of sources of food, solving conflicts, and decision-making. GTO is mathematically modeled as a five-stage algorithm where three stages are responsible for exploration, while the remaining two stages are related to exploitation. The positions of gorillas are updated using the following equations:
X ( t + 1 ) = { ( U l L l   ) × r 1 + L l   r a n d < p ( r 2 C ) × X r ( t ) + L × H   r a n d   0.5 X ( t ) L × ( L × ( X ( t ) X r ( t ) ) + r 3 × ( X ( t ) X r ( t ) ) )   r a n d < 0.5  
where X is the position of the current gorilla at iteration t . r 1 ,     r 2 ,   r 3 , and r a n d are random numbers in the range of 0 to 1. p is a parameter whose value will usually lies between 0 and 1. U l and L l are the upper and lower boundaries, respectively. X r is a gorilla randomly chosen at each iteration. The values of C, L, and H are calculated using Equations (2), (4) and (5), respectively.
C = F × ( 1 I t e r M a x i t )
F = cos ( 2 × r 4 ) + 1
L = C × l
H = Z × X ( t )
In Equation (2), I t e r represents the current iteration and M a x i t represents the maximum number of iterations. F in Equation (2) is calculated using Equation (3), and r 4 is a random number in the range of 0 to 1. Here, l is an integer randomly chosen in the range of −1 to 1. In Equation (5), Z is a random number in the range −C to +C. Based on the position of the silverback, the other gorillas change their position while searching for food, and this behavior is represented using Equation (6). The M value mentioned in Equation (6) is computed using Equations (7) and (8).
X ( t + 1 ) = L × M × ( X ( t ) X s i l v e r b a c k ) + X ( t )
M = ( | 1 N | g i = 1 N X i ( t ) ) 1 g
g = 2 L
where X s i l v e r b a c k is the position of the silverback gorilla with the best position when compared with the positions of the other gorillas and N is the total number of gorillas. Gorilla behavior in competing to choose adult females is represented using Equation (9).
X ( t + 1 ) = X s i l v e r b a c k ( X s i l v e r b a c k × Q X ( t ) × Q ) × A
Q = 2 × r 5 1
A = β × E
E = { N 1   r a n d   0.5 N 2   r a n d < 0.5  
In the above equations, r 5 and r a n d are random numbers in the range of 0 to 1, while β is a parameter whose value is crucial in deciding the updated positions of gorillas. N 1 is a random number in the range decided by the problem dimension, while N 2 is a random number that follows a normal distribution in the range [0, 1]. Initially, Equation (1) is used to update all the gorilla positions. Then, the silverback gorilla is found in that iteration. After that, the other gorilla positions will be updated based on the silverback gorilla position. If | C | 1 , then the positions of the gorillas are updated using Equation (6); otherwise, they will be updated using Equation (9).

3.6. Particle Swarm Optimization

PSO [47] is one of the popular and efficient swarm intelligence-based optimization algorithms. PSO is inspired by the characteristics exhibited by flocks of birds while searching for food. Usually, the population will be initialized randomly and updated in each iteration based ona fitness function. The velocity of each particle is mathematically modeled and updated using Equation (13).
v i ( t + 1 ) = w × v i ( t ) + c 1 × r 1 × ( p i ( t ) x i ( t ) ) + c 2 × r 2 × ( g b e s t x i ( t ) )
where v i ( t ) stands for the velocity of the ith particle in iteration t. Three crucial parameters in PSO are w, c1, and c2. The position of the ith particle in iteration t is represented as x i ( t ) ; and p i ( t ) and g b e s t represent the personal best and global best particle positions, respectively. r 1 and r 2 are random numbers in the range of 0 to 1. The position of each particle is updated based on the old position and new velocity, as represented in Equation (14).
x i ( t + 1 ) = x i ( t ) + v i ( t + 1 )
The personal best and global best are computed in each iteration using Equations (15) and (16), respectively.
p i ( t + 1 ) = { p i ( t )   i f   f ( x i ( t + 1 ) ) f ( p i ( t ) ) x i ( t + 1 )   i f   f ( x i ( t + 1 ) ) < f ( p i ( t ) )
g b e s t { p 0 ( t ) , p 1 ( t ) ,   . ,   p m ( t )   } = m i n   { f ( p 0 ( t ) ) , f ( p 1 ( t ) ) ,   , f ( p m ( t ) )   }
where f represents the fitness function, which is crucial in deciding the performance of PSO.

3.7. Elephant Herding Optimization

Elephant Herding Optimization (EHO) [48] is inspired by the behavior of elephants. Like PSO and GTO, EHO also falls under the category of swarm intelligence meta-heuristic algorithms. The position of an elephant is updated using Equation (17).
x i n e w = x i o l d + α   ( x b e s t x i o l d ) r a n  
where x i n e w and x i o l d are the new and old positions of the ith elephant. x b e s t is the best elephant position found using Equation (18). x c e n t e r in Equation (18) is computed using Equation (19). In addition to updating the best elephant position, the worst elephant position x w o r s t is also updated using Equation (20).
x b e s t = β × x c e n t e r
x c e n t e r = 1 n × i = 1 n x i
x w o r s t = x m i n + ( x m a x x m i n + 1 ) r a n d
where α and β are the EHO parameters; r a n d is a random number in the range [0, 1]; and n is the number of elephants. x m a x and x m i n are the maximum and minimum boundaries for elephant positions.

4. Materials and Methods

Three publicly available datasets comprising 2784 Normal and 3632 oral squamous cell carcinoma subjects are considered in this work. The first dataset was obtained from Kaggle [49], and it contains oral histopathological images at both the 100× and 400× zoom levels. The first dataset contains a total of 5192 images, and out of them, 2494 images belong to the Normal class and 2698 belong to OSCC class. The second and third datasets are obtained from the online repository built by Tabassum Yesmin Rahman et al. [50]. Oral histopathological images with zoom levels of 100× and 400× are present in the second and third datasets, respectively. A total of 89 normal images and 439 OSCC images are available in the second dataset, while 201 normal images and 495 OSCC images are available in the third dataset. Some of the sample oral histopathological images belonging to normal and OSCC classes are shown in Figure 1 and Figure 2, respectively.
The typical procedure for implementing oral cancer detection using transfer learning-based feature extraction is shown in Figure 3. Histopathological oral images from the three datasets are fed to the feature extraction layers discretely, and the resultant classification performance metrics are also computed individually. Features are extracted using the transfer learning approach, where the weights are pre-trained for another similar dataset. Three popular CNN architectures, namely, InceptionV2, MobileNetV3, and EfficientNetB3, are investigated in this work for feature extraction. Weights that are pre-trained for the popular ImageNet dataset are considered in all three architectures. The extracted features are then divided into training, validation, and test feature sets using a stratified shuffle split approach in a 70:15:15 ratio, respectively. A validation set is used to avoid overfitting and to improve the performance of the model for the unseen new data. A stratified shuffle split is considered since it randomly selects the samples according to the class ratio in the original dataset. In other words, stratified shuffle split ensures the ratio of each class in all three resultant sets is the same as that shown in Table 1. This approach of data splitting is very crucial in imbalanced datasets. Then, the classification layers are trained using training and validation feature sets, where the ideal weights of Neural Networks for classifying the oral histopathological images are found.
Two fully connected Neural Network layers along with batch normalization and dropout are used as classification layers, as shown in Figure 4. Finally, the trained classification layers with ideal weights are used to classify the test feature set as Normal or OSCC class. In Figure 4, the functional layer depicts the transfer learning-based pre-trained model while the remaining layers are used for classification. The specifications of the classification layer considered in this research work are presented in Table 2. For comparison purposes, the classification layer is unaltered for all the datasets and different feature extraction layers. Specifications related to the number of epochs and batch size during training, optimizer, early stopping, and a reduction in the learning rate on the plateau are also mentioned in Table 2. Based on the transfer model used for the feature extraction layer, the number of trainable parameters of the complete deep learning model will vary, as shown in Table 3. The number of features extracted per input image with the three different feature extraction layers is also shown in Table 3.
The proposed approach for OSCC detection is presented in Figure 5 and Figure 6. An intermediate layer based on MGTO is included in the proposed method when compared with Figure 3 and Figure 4. Similar to the classification layer, the newly introduced intermediate layer also needs to be trained, where it will learn the ideal values for its parameters related to the MGTO algorithm. Hence, it is trained with the original training and validation feature sets. Then, all three original feature sets are supplied as input to the trained MGTO layer, where the feature sets are transformed to produce another three transformed sets, namely, training, testing, and validation. The size of the input and output feature sets remains the same. Then, the transformed sets are considered classification layers for detecting the class of an oral histopathological image.

5. Implementation of the Proposed MGTO

The equations to update the positions of gorillas are modified based on the Sine Cosine Algorithm 1 [51] to increase the exploitation and exploration capabilities of GTO. In MGTO, three equations that update the positions of gorillas are modified. Equations (1), (6) and (9) of GTO are modified, as represented in Equations (21)–(23), respectively, in MGTO. All other equations of GTO remain intact in MGTO.
X ( t + 1 ) = { ( U l L l   ) × r 1 + L l   r a n d < p ( r 2 C )   X r ( t ) + L × H   r a n d   0.5 X ( t ) L × r a d × s i n ( X ( t ) X r ( t ) ) + r a d × c o s ( X ( t ) X r ( t ) ) )   r a n d < 0.5  
X ( t + 1 ) = L × M × r a d × s i n ( X ( t ) X s i l v e r b a c k ) + X ( t )
X ( t + 1 ) = X s i l v e r b a c k r a d × c o s ( X s i l v e r b a c k × Q X ( t ) × Q ) × A  
where r a d in the above equations is computed using Equation (24). const in Equation (24) is a constant, and it is considered to be three, as suggested in [48]. Crnt _   Iter represents the current iteration number, while Max _ Iter represents the maximum number of iterations.
rad   =   const C r n t _ Iter   × ( const Max Iter )  
In the original GTO, the gorilla population is initialized randomly. When using MGTO as an intermediate layer in deep learning models, the gorilla population is initialized with the features extracted from the previous layer. The number of gorillas is equal to the number of features extracted. Then, the gorillas’ positions are updated in each iteration using MGTO equations. The fitness function is very crucial in optimization algorithms, and it is selected based on the problem to be solved. To use MGTO for transforming the features, the fitness function based on the variance metric is considered. The fitness of each gorilla, F ( X i ) depends on its own position and four nearest-neighbor gorillas, as shown in Equation (25).
F ( X i ) =   Variance ( X i 2 , X i 1 , X i , X i + 1 , X i + 2 )  
In MGTO, p and β are the parameters that mainly determine the performance along with Max _ Iter . The ideal values of these parameters are found based on the accuracy attained during training and validation. The validation accuracy for various values of Max _ Iter is plotted in Figure 7 and Figure 8. The ideal value is found where the validation accuracy of 0.77 is reached. While finding the optimal value for Max _ Iter , the other two parameters, namely, p and β are kept at 0.5 (median of range [0, 1]). To find the optimal values for p and β parameters, Max _ Iter is kept at its ideal value of 11. Figure 8 depicts the validation accuracy for various values of the p and β parameters. The highest validation accuracy of 0.95 is attained when p = 0.3 and β = 0.7. Finding the ideal values for the parameters of MGTO is termed training, and for this purpose, training and validation feature sets are used. After training, an MGTO transform is implemented for all three feature sets, namely, the training, validation, and test sets with the ideal parameter values of Max _ Iter = 11 ,   p = 0.3, and β = 0.7. Notably, these are the ideal parameters of MGTO for the first dataset when MobileNetV3 is used as a feature extraction layer. The ideal parametric values may change depending on the input data given to the MGTO layer. The ideal values for other input data will be presented in the next section. The procedure for implementing MGTO as an intermediate feature transform layer for the test feature set is summarized in Algorithm 1.
Algorithm 1: Algorithm to implement the proposed MGTO as an intermediate layer in deep learning models for feature transformation of a test feature set.
Step 1: Extract features using pre-trained transfer learning models for each oral histopathological image.
Step 2: Consider the number of features as the size of the population in MGTO. Initialize the position of gorillas with extracted features.
Step 3: Initialize parameters of MGTO: U l = max ( X 1 , X 2 , . , X N ) , L l = min ( X 1 , X 2 , . , X N ) , Max Iter = 11 ,   p = 0.3, and β = 0.7.
Step 4: Compute the fitness value of each gorilla using Equation (25).
Step 5: Update the position of each gorilla using Equation (21).
Step 6: Identify the silverback gorilla, i.e., the gorilla with the highest fitness.
Step 7: Update the position of each gorilla using Equation (22) if | C | 1 . Otherwise, use Equation (23).
Step 8: Repeat steps 4 to 7 until the maximum number of iterations is reached. If the maximum number of iterations are completed, then go to step 9.
Step 9: Consider the final position of the gorillas as the output of the feature transform and give them as input to the classification layer.

6. Results and Discussion

Initially, the experiment is conducted without any intermediate layers in the deep learning model. Three different transfer learning-based models, namely, InceptionV2, MobileNetV3, and EfficientNetB3, are tested as feature extraction layers. As mentioned in Table 2, the specifications of the classification layer remain the same for all three different feature extraction layers. The confusion matrix attained for these three deep learning models without an intermediate layer on the first dataset in OSCC detection is shown in Figure 9. In Figure 9, the label 0 represents the Normal class and label 1 represents class OSCC class. EfficientNetB3 classifies all the input images as OSCC, and so its True Negative (TN) = 0. This clearly indicates the poor performance of the EfficientNetB3-based deep learning model without an intermediate layer. To detect OSCC, a high TP is required, while to detect the Normal class properly, a high TN is required. Among the remaining two models without an intermediate layer, the highest True Positive (TP) is attained with MobileNetV3, while the highest TN is attained with InceptionV2.
To improve the number of TNs and TPs, an MGTO-based intermediate layer is proposed in this work. Figure 10 shows the confusion matrix of three different feature extraction-based deep learning models with MGTO as the intermediate layer in OSCC detection. When MGTO is not used as an intermediate layer in the EfficientNetB3-based deep learning model, all the oral images are classified as OSCC, while better TN and TP values are attained with the proposed layer. The highest TN and TP values are attained for the proposed MobileNetV3-based feature extraction with MGTO as an intermediate layer.
Based on the confusion matrix, four popular performance metrics, namely, accuracy, precision, recall, and the F1-score, are used in this research work to analyze the performance of the deep learning models. In addition to the deep learning models with and without an MGTO-based intermediate layer, three other swarm intelligence-based optimization algorithms, namely, PSO, EHO, and GTO, are also tested as intermediate layers, and their results are also presented in Table 4. The implementation procedure for PSO, EHO, and GTO as intermediate layers follows Algorithm 1 presented in the previous section. Only the parameters and the way of updating the position of the swarm vary based on the optimization algorithm used. The final ideal parameters of all four tested intermediate layers after training are listed in Table 5.
The main objective of this work is to detect OSCC, so precision, recall, and the F1-score in Table 4 are related to the correct identification of OSCC class, while the accuracy metric is related to the correct identification of both the Normal and OSCC classes. As seen in Table 4, the deep learning models without an intermediate layer provide less accuracy than the proposed deep learning models with MGTO as an intermediate layer.
Among the models without an intermediate layer, MobileNetV3 offers the highest accuracy of 0.89, which is followed by InceptionV2 with an accuracy of 0.88 and EfficientNetB3 with an accuracy of 0.52. The reason for such poor performance of EfficientNetB3 is explained as follows: All three feature extraction models are pre-trained on the ImageNet dataset and features are extracted based on the weights appropriate for the ImageNet dataset. The weights and architecture of EfficientNetB3 fail to capture the significant features from input oral images, while vital features are properly extracted using the remaining two feature extraction models. This statement is further supported by Figure 11, where the training and validation accuracy and loss are presented for all the three investigated deep learning models without an intermediate layer on the first dataset.
Since quality features are extracted with MobileNetV3 and InceptionV2, both training and validation accuracy increase gradually during training. In addition, both training and validation loss also decrease exponentially. But a deep learning model that uses EfficientNetB3 fails to increase in both training and validation accuracy due to the poor features that are extracted from the histopathological oral images. Figure 12 presents the training and validation accuracy and loss when MGTO is used as an intermediate layer on the first dataset. It clearly shows an improved accuracy during both training and validation because of the transformed appropriate features produced with MGTO. To support the findings based on the first dataset, the other two smaller OSCC datasets are tested. The second and third datasets are highly imbalanced since the number of OSCC class samples is much higher than the number of normal class samples. The performance metrics attained on those two datasets are presented in Table 6 and Table 7.
From Table 4, Table 6 and Table 7, it is very clear that MGTO works very well as an intermediate layer when compared with the other tested intermediate layers in all three datasets. The significance of MGTO as an intermediate layer can be clearly seen in Figure 13, where the percentage of accuracy increase attained with the usage of various intermediate layers when compared with a deep learning model without an intermediate layer is depicted.
On all three datasets, the percentage of accuracy increase is much less or even negative when PSO, EHO, and GTO are used as intermediate layers on the features extracted from MobileNetV3 and InceptionNetV2. Notably, already these two feature extraction models without an intermediate layer produce an accuracy of more than 0.8 in all three datasets. Implicitly, these intermediate layers fail to significantly improve the accuracy since they are not able to produce more appropriate transformed features for classification. Only for the features extracted with EfficientNetB3 from the first dataset can these intermediate layers provide a significant accuracy increase. This is because the original features extracted with EfficientNetB3 are very poor on the first dataset, which yields an accuracy of only 0.52. Out of these three intermediate layers, GTO comparatively performs well on all three datasets. Hence evidence for improving GTO further with suitable modifications was found. MGTO is formulated with the modifications stated in the previous section, and it worked well on all three datasets.
In the first dataset, a 73% increase in accuracy is seen using the EfficientNetB3-based DL model due to the usage of MGTO as an intermediate layer. Nearly a 6% accuracy increase is seen due to the usage of MGTO in the MobileNetV3 and InceptionV2-based DL models. Notably, the highest accuracy of 0.95 is produced on the first dataset using the MobileNetV3-MGTO-based DL model. Even on the imbalanced second and third datasets, MGTO is capable of producing significant increases in accuracy. The reason for this better performance is threefold. Firstly, the modification of GTO with the Sine and Cosine algorithm increases its exploration and exploitation capability well. Exploitation is responsible for the local search, i.e., fine-tuning and exploration are responsible for the global search. Secondly, the selection of appropriate fitness functions. Local variance-based fitness functions worked well to transform the features of different classes in different ways. Thirdly, the usage of ideal parameters in MGTO resulted in better accuracy. As shown in Figure 7 and Figure 8, the values of MGTO parameters have a huge impact on accuracy. Due to the above-mentioned reasons, MGTO works soundly as an intermediate layer that transforms the input features into more appropriate features for classification. In other words, the introduction of the proposed intermediate layer helps the classifier to distinguish the features of two different classes. This statement is backed by the scatter plots shown in Figure 14, Figure 15 and Figure 16. In the scatter plots, label 0 represents the Normal class and label 1 represents the OSCC class. To represent the features of the first dataset in a scatter plot, two averages are computed. Average1 is the mean of the first half of the features and Average2 is the mean of the remaining half of the features. For example, 1280 features are extracted with MobileNetV3, and the mean of the first 640 features is considered as Average1, and the mean of the remaining 640 features is considered as Average2.
In the scatter plots, the MobileNetV3 features of two the classes are slightly scattered and overlapped, while the EfficientNetB3 features are heavily overlapped. A comparison of the scatter plots with and without intermediate layers for all three feature extraction models suggests the significance of MGTO. The proposed layer transforms the features in a manner that is more suitable for classification by spreading the two different class features apart to some extent. When these transformed features are used for training and validation, the classification layer is trained well. Finally, better performance is achieved when the transformed test dataset is categorized by the trained classification layer.
Apart from accuracy, other performance metrics are also relevantly important. Precision gives the percentage of correct OSCC predictions among the total number of OSCCs predicted. Recall is related to the percentage of actual OSCC that was identified correctly as OSCC. The F1-score is the harmonic mean of precision and recall. These metrics are depicted for all three datasets in Figure 17. Considering these three metrics, DL models with MGTO as an intermediate layer outperform all other investigated intermediate layers. In the first dataset, the highest performance is offered by the proposed MobileNetV3-MGTO-based DL model, which achieves precision = 0.95, recall = 0.95, and F1-score = 0.95. Even on the second and third datasets, the highest F1-score is archived with the proposed DL model. Though the highest F1-score and accuracy are attained with the proposed DL model on all three datasets, it fails to attain balanced precision and recall in imbalanced datasets. For example, recall is much higher than precision for the proposed DL model on the second dataset, while precision is much higher than recall for the proposed DL model on the third dataset. But it attains almost equal precision and recall on the first dataset, which is a balanced dataset. Hence, wherever higher values of both precision and recall are required on imbalanced datasets, the proposed DL model underperforms the other models. This could be considered as the first limitation of the proposed model.
The training time for all the investigated DL models on the first dataset is presented in Table 8. DL models without an intermediate layer are trained comparatively quickly, while the presence of an intermediate layer may take more training time [52]. MobileNetV3 has a shorter training time since the number of features extracted is 1280, while the number of features extracted with EfficientNetB3 and InceptionNetV2 is 1536. A pie chart is presented in Figure 18 that depicts the percentage of training time required by a DL model when compared with the total training time required by all the DL models. PSO and EHO take relatively less training time than other intermediate layers due to their simple structure. When compared with GTO, the proposed MGTO intermediate layer takes more training time due to the inclusion of Sine and Cosine argument calculations. Only 2% of the total training time is taken by the MobileNetV3 DL model without an intermediate layer, while 7% of the total training time is taken by the proposed DL model. This could be considered as second limitation.
An accuracy comparison of some related works for oral cancer detection is presented in Table 9. Supervised classifiers such as K-Nearest Neighbor, and Support Vector machine attain comparatively lower accuracy than the deep learning models. The proposed deep learning model with MGTO as the intermediate layer offers the highest accuracy of 95%, which shows the importance of the proposed DL model.

7. Conclusions

This research work focuses on developing an enhanced deep-learning model to diagnose OSCC disease. The proposed DL model with MGTO as the intermediate layer and MobileNetV3 as the feature extraction layer could classify 95% of the histopathological oral images correctly. Three oral histopathological image datasets were tested and on all three datasets, and the inclusion of MGTO as an intermediate layer enhanced the accuracy of the DL model. Features were transformed using MGTO to produce more appropriate features for classification. MGTO outperformed other investigated SI algorithms as an intermediate layer when compared with PSO, EHO, and GTO, primarily due to the modifications made to the GTO equations and their fitness function. The limitations of the proposed DL model are a relatively higher training time and loss of either precision or recall scores on an imbalanced dataset. Ouruture work will investigate other SI algorithms as intermediate layers in DL models. In addition, the proposed model needs to be tested for other medical image classification problems.

Author Contributions

Conceptualization, V.K.V., M.T.R. and B.N.; methodology, S.C. and S.B.K.; software, M.T.R.; conceptualization, V.K.V., M.T.R. and S.B.K.; methodology, B.N. and S.B.K.; validation, S.B., E.A. and M.T.R.; formal analysis, M.T.R.; investigation, V.K.V.; resources, M.T.R.; data curation, S.C. and S.B.K.; writing—original draft preparation, V.K.V., S.B. and E.A.; writing—review and editing, M.T.R. and S.B.K.; visualization, S.B. and E.A.; supervision, S.B.K. and M.T.R. All authors have read and agreed to the published version of this manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used for the findings will be shared by the author upon request.

Acknowledgments

This research is supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R195) Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. This work was supported by the Deanship of Scientific Research, Vice President for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia [Grant No. 5094].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gupta, B.; Bray, F.; Kumar, N.; Johnson, N.W. Associations between oral hygiene habits, diet, tobacco and alcohol and risk of oral cancer: A case–control study from India. Cancer Epidemiol. 2017, 51, 7–14. [Google Scholar] [CrossRef]
  2. Ramakrishna, M.T.; Venkatesan, V.K.; Izonin, I.; Havryliuk, M.; Bhat, C.R. Homogeneous Adaboost Ensemble Machine Learning Algorithms with Reduced Entropy on Balanced Data. Entropy 2023, 25, 245. [Google Scholar] [CrossRef] [PubMed]
  3. Laprise, C.; Shahul, H.P.; Madathil, S.A.; Thekkepurakkal, A.S.; Castonguay, G.; Varghese, I.; Shiraz, S.; Allison, P.; Schlecht, N.F.; Rousseau, M.C.; et al. Periodontal diseases and risk of oral cancer in Southern India: Results from the HeNCe Life study. Int. J. Cancer 2016, 139, 1512–1519. [Google Scholar] [CrossRef] [PubMed]
  4. Khayatan, D.; Hussain, A.; Tebyaniyan, H. Exploring animal models in oral cancer research and clinical intervention: A critical review. Vet. Med. Sci. 2023, 9, 1833–1847. [Google Scholar] [CrossRef]
  5. Mosaddad, S.A.; Beigi, K.; Doroodizadeh, T.; Haghnegahdar, M.; Golfeshan, F.; Ranjbar, R.; Tebyanian, H. Therapeutic applications of herbal/synthetic/bio-drug in oral cancer: An update. Eur. J. Pharmacol. 2021, 890, 173657. [Google Scholar] [CrossRef]
  6. Borse, V.; Konwar, A.N.; Buragohain, P. Oral cancer diagnosis and perspectives in India. Sens. Int. 2020, 1, 100046. [Google Scholar] [CrossRef]
  7. Ajay, P.; Ashwinirani, S.; Nayak, A.; Suragimath, G.; Kamala, K.; Sande, A.; Naik, R. Oral cancer prevalence in Western population of Maharashtra, India, for a period of 5 years. J. Oral. Res. Rev. 2018, 10, 11. [Google Scholar] [CrossRef]
  8. Karadaghy, O.A.; Shew, M.; New, J.; Bur, A.M. Development and assessment of a machine learning model to help predict survival among patients with oral squamous cell carcinoma. JAMA Otolaryngol. Head Neck Surg. 2019, 145, 1115–1120. [Google Scholar] [CrossRef]
  9. Seoane-Romero, J.; Vazquez-Mahia, I.; Seoane, J.; Varela-Centelles, P.; Tomas, I.; Lopez-Cedrun, J. Factors related to late stage diagnosis of oral squamous cell carcinoma. Med. Oral Patol. Oral Cir. Bucal 2012, 17, e35–e40. [Google Scholar] [CrossRef]
  10. Dascălu, I.T. Histopathological aspects in oral squamous cell carcinoma. J. Dent. Sci. 2018, 3, 173. [Google Scholar] [CrossRef]
  11. Mangalath, U.; Mikacha, M.K.; Abdul Khadar, A.H.; Aslam, S.; Francis, P.; Kalathingal, J. Recent trends in prevention of oral cancer. J. Int. Soc. Prev. Community Dent. 2014, 4, 131. [Google Scholar] [CrossRef]
  12. O’Mahony, N.; Campbell, S.; Carvalho, A.; Harapanahalli, S.; Hernandez, G.V.; Krpalkova, L.; Riordan, D.; Walsh, J. Deep learning vs. traditional computer vision. In Science and Information Conference; Springer: Berlin/Heidelberg, Germany, 2019; pp. 128–144. [Google Scholar]
  13. Hussein, I.J.; Burhanuddin, M.A.; Mohammed, M.A.; Benameur, N.; Maashi, M.S.; Maashi, M.S. Fully automatic identification of gynaecological abnormality using a new adaptive frequency filter and histogram of oriented gradients (hog). Expert. Syst. 2021, 39, e12789. [Google Scholar] [CrossRef]
  14. Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G. Completely Automated CNN Architecture Design Based on Blocks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 1242–1254. [Google Scholar] [CrossRef]
  15. Johner, F.M.; Wassner, J. Efficient evolutionary architecture search for CNN optimization on GTSRB. In Proceedings of the 18th IEEE International Conference on Machine Learning and Applications, ICMLA, Boca Raton, FL, USA, 16–19 December 2019; pp. 56–61. [Google Scholar]
  16. Mozafari, M.; Farahbakhsh, R.; Crespi, N. A BERT-Based Transfer Learning Approach for Hate Speech Detection in Online Social Media. Stud. Comput. Intell. 2020, 881, 928–940. [Google Scholar] [CrossRef]
  17. Khoh, W.H.; Pang, Y.H.; Teoh, A.B.J.; Ooi, S.Y. In-air hand gesture signature using transfer learning and its forgery attack. Appl. Soft Comput. 2021, 113 Pt A, 108033. [Google Scholar] [CrossRef]
  18. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  19. Krishnan, M.M.R.; Chakraborty, C.; Ray, A.K. Wavelet based texture classification of oral histopathological sections. Int. J. Microsc. Sci. Technol. Appl. Educ. 2010, 2, 897–906. [Google Scholar]
  20. Krishnan, M.M.R.; Shah, P.; Choudhary, A.; Chakraborty, C.; Paul, R.R.; Ray, A.K. Textural characterization of histopathological images for oral sub-mucous fibrosis detection. Tissue Cell 2011, 43, 318–330. [Google Scholar] [CrossRef] [PubMed]
  21. Krishnan, M.; Acharya, U.; Chakraborty, C.; Ray, A. Automated diagnosis of oral cancer using higher order spectra features and local binary pattern: A comparative study. Technol. Cancer Res. Treat. 2011, 10, 443–455. [Google Scholar] [CrossRef] [PubMed]
  22. Patra, R.; Chakraborty, C.; Chatterjee, J. Textural analysis of spinous layer for grading oral submucous fibrosis. Int. J. Comput. Appl. 2012, 47, 975–8887. [Google Scholar] [CrossRef]
  23. Krishnan, M.M.R.; Venkatraghavan, V.; Acharya, U.R.; Pal, M.; Paul, R.R.; Min, L.C.; Ray, A.K.; Chatterjee, J.; Chakraborty, C. Automated oral cancer identification using histopathological images: A hybrid feature extraction paradigm. Micron 2012, 43, 352–364. [Google Scholar] [CrossRef]
  24. Thomas, B.; Kumar, V.; Saini, S. Texture analysis based segmentation and classification of oral cancer lesions in color images using ANN. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Computing and Control (ISPCC), Solan, India, 26–28 September 2013; pp. 1–5. [Google Scholar]
  25. Rahman, T.; Mahanta, L.; Chakraborty, C.; Das, A.; Sarma, J. Textural pattern classification for oral squamous cell carcinoma. J. Microsc. 2018, 269, 85–93. [Google Scholar] [CrossRef] [PubMed]
  26. Rahman, T.Y.; Mahanta, L.B.; Das, A.K.; Sarma, J.D. Automated oral squamous cell carcinoma identification using shape, texture and color features of whole image strips. Tissue Cell 2020, 63, 101322. [Google Scholar] [CrossRef] [PubMed]
  27. Rahman, A.U.; Alqahtani, A.; Aldhaferi, N.; Nasir, M.U.; Khan, M.F.; Khan, M.A.; Mosavi, A. Histopathologic oral cancer prediction using oral squamous cell carcinoma biopsy empowered with transfer learning. Sensors 2022, 22, 3833. [Google Scholar] [CrossRef] [PubMed]
  28. Warin, K.; Limprasert, W.; Suebnukarn, S.; Jinaporntham, S.; Jantana, P. Automatic classifcation and detection of oral cancer in photographic images using deep learning algorithms. J. Oral. Pathol. Med. 2021, 50, 911–918. [Google Scholar] [CrossRef]
  29. Camalan, S.; Mahmood, H.; Binol, H.; Araújo, A.L.D.; Santos-Silva, A.R.; Vargas, P.A.; Lopes, M.A.; Khurram, S.A.; Gurcan, M.N. Convolutional neural network-based clinical predictors of oral dysplasia: Class activation map analysis of deep learning results. Cancers 2021, 13, 1291. [Google Scholar] [CrossRef] [PubMed]
  30. Musulin, J.; Štifanić, D.; Zulijani, A.; Ćabov, T.; Dekanić, A.; Car, Z. An enhanced histopathology analysis: An AI-based system for multiclass grading of oral squamous cell carcinoma and segmenting of epithelial and stromal tissue. Cancers 2021, 13, 1784. [Google Scholar] [CrossRef]
  31. Das, M.; Dash, R.; Mishra, S.K. Automatic detection of oral squamous cell carcinoma from histopathological images of oral mucosa using deep convolutional neural network. Int. J. Environ. Res. Public Health 2023, 20, 2131. [Google Scholar] [CrossRef]
  32. Lin, H.; Chen, H.; Weng, L.; Shao, J.; Lin, J. Automatic detection of oral cancer in smartphone-based images using deep learning for early diagnosis. J. Biomed. Opt. 2021, 26, 086007. [Google Scholar] [CrossRef]
  33. Das, N.; Hussain, E.; Mahanta, L.B. Automated classification of cells into multiple classes in epithelial tissue of oral squamous cell carcinoma using transfer learning and convolutional neural network. Neural Netw. 2020, 128, 47–60. [Google Scholar] [CrossRef] [PubMed]
  34. Panigrahi, S.; Das, J.; Swarnkar, T. Capsule network based analysis of histopathological images of oral squamous cell carcinoma. J. King Saud. Univ. Comput. Inf. Sci. 2022, 34, 4546–4553. [Google Scholar] [CrossRef]
  35. Myriam, H.; Abdelhamid, A.A.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Jamjoom, M.M.; Khafaga, D.S. Advanced meta-heuristic algorithm based on Particle Swarm and Al-biruni Earth Radius optimization methods for oral cancer detection. IEEE Access 2023, 11, 23681–23700. [Google Scholar] [CrossRef]
  36. Panneerselvam, K.; Nayudu, P.P. Improved Golden Eagle Optimization Based CNN for Automatic Segmentation of Psoriasis Skin Images. Wirel. Pers. Commun. 2023, 131, 1817–1831. [Google Scholar] [CrossRef]
  37. Erkan, U.; Toktas, A.; Ustun, D. Hyperparameter optimization of deep CNN classifier for plant species identification using artificial bee colony algorithm. J. Ambient. Intell. Human. Comput. 2023, 14, 8827–8838. [Google Scholar] [CrossRef]
  38. Vinaykumar, V.N.; Babu, J.A.; Frnda, J. Optimal guidance whale optimization algorithm and hybrid deep learning networks for land use land cover classification. Eurasip J. Adv. Signal Process. 2023, 2023, 13. [Google Scholar] [CrossRef]
  39. Anilkumar Gona, M.; Subramoniam, R. Swarnalatha, Transfer learning convolutional neural network with modified Lion optimization for multimodal biometric system. Comput. Electr. Eng. 2023, 108, 108664. [Google Scholar] [CrossRef]
  40. Subashchandrabose, U.; John, R.; Anbazhagu, U.V.; Venkatesan, V.K.; Thyluru Ramakrishna, M. Ensemble Federated Learning Approach for Diagnostics of Multi-Order Lung Cancer. Diagnostics 2023, 13, 3053. [Google Scholar] [CrossRef]
  41. Saab, S., Jr.; Saab, K.; Phoha, S.; Zhu, M.; Ray, A. A multivariate adaptive gradient algorithm with reduced tuning efforts. Neural Netw. 2022, 152, 499–509. [Google Scholar] [CrossRef]
  42. Wang, S.-H.; Phillips, P.; Sui, Y.; Liu, B.; Yang, M.; Cheng, H. Classification of Alzheimer’s disease based on eight-layer convolutional neural network with leaky rectified linear unit and max pooling. J. Med. Syst. 2018, 42, 85. [Google Scholar] [CrossRef] [PubMed]
  43. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  44. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  45. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2020, arXiv:1905.11946v5. [Google Scholar]
  46. Abdollahzadeh, B.; Soleimanian Gharehchopogh, F.; Mirjalili, S. Artificial gorilla troops optimizer: A new nature-inspired metaheuristic algorithm for global optimization problems. Int. J. Intell. Syst. 2021, 36, 5887–5958. [Google Scholar] [CrossRef]
  47. Sayour, M.H.; Kozhaya, S.E.; Saab, S.S. Autonomous robotic manipulation: Real-time, deep-learning approach for grasping of unknown objects. J. Robot. 2022, 2022, 2585656. [Google Scholar] [CrossRef]
  48. Saab, S., Jr.; Fu, Y.; Ray, A.; Hauser, M. A dynamically stabilized recurrent neural network. Neural Process. Lett. 2022, 5, 1195–1209. [Google Scholar] [CrossRef]
  49. Histopathologic Oral Cancer Detection Using CNNs. Available online: https://www.kaggle.com/ashenafifasilkebede/dataset?select=val (accessed on 10 June 2023).
  50. Rahman, T.Y.; Mahanta, L.B.; Das, A.K.; Sarma, J.D. Histopathological imaging database for oral cancer analysis. Data Brief. 2020, 29, 105114. [Google Scholar] [CrossRef]
  51. Lian, Z.; Zeng, Q.; Wang, W.; Gadekallu, T.R.; Su, C. Blockchain-Based Two-Stage Federated Learning with Non-IID Data in IoMT System. IEEE Trans. Comput. Soc. Syst. 2022, 10, 1701–1710. [Google Scholar] [CrossRef]
  52. Aubreville, M.; Knipfer, C.; Oetter, N.; Jaremenko, C.; Rodner, E.; Denzler, J.; Bohr, C.; Neumann, H.; Stelzle, F.; Maier, A. Automatic Classification of Cancerous Tissue in Laserendomicroscopy Images of the Oral Cavity using Deep Learning. Sci. Rep. 2017, 7, 11979. [Google Scholar] [CrossRef] [PubMed]
  53. Alkhadar, H.; Macluskey, M.; White, S.; Ellis, I.; Gardner, A. Comparison of machine learning algorithms for the prediction of five-year survival in oral squamous cell carcinoma. J. Oral. Pathol. Med. 2021, 50, 378–384. [Google Scholar] [CrossRef] [PubMed]
  54. Alhazmi, A.; Alhazmi, Y.; Makrami, A.; Salawi, N.; Masmali, K.; Patil, S. Application of artificial intelligence and machine learning for prediction of oral cancer risk. J. Oral. Pathol. Med. 2021, 50, 444–450. [Google Scholar] [CrossRef] [PubMed]
  55. Arikumar, K.S.; Deepak Kumar, A.; Gadekallu, T.R.; Prathiba, S.B.; Tamilarasi, K. Real-Time 3D Object Detection and Classification in Autonomous Driving Environment Using 3D LiDAR and Camera Sensors. Electronics 2022, 11, 4203. [Google Scholar] [CrossRef]
  56. Welikala, R.A.; Remagnino, P.; Lim, J.H.; Chan, C.S.; Rajendran, S.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R.; et al. Automated Detection and Classification of Oral Lesions Using Deep Learning for Early Detection of Oral Cancer. IEEE Access 2020, 8, 132677–132693. [Google Scholar] [CrossRef]
  57. Shavlokhova, V.; Sandhu, S.; Flechtenmacher, C.; Koveshazi, I.; Neumeier, F.; Padrón-Laso, V.; Jonke, Ž.; Saravi, B.; Vollmer, M.; Vollmer, A.; et al. Deep Learning on Oral Squamous Cell Carcinoma Ex Vivo Fluorescent Confocal Microscopy Data: A Feasibility Study. J. Clin. Med. 2021, 10, 5326. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Sample oral histopathological images belonging to the Normal class.
Figure 1. Sample oral histopathological images belonging to the Normal class.
Diagnostics 13 03461 g001
Figure 2. Sample oral histopathological images belonging to OSCC class.
Figure 2. Sample oral histopathological images belonging to OSCC class.
Diagnostics 13 03461 g002
Figure 3. Typical approach for OSCC detection using transfer learning-based feature extraction.
Figure 3. Typical approach for OSCC detection using transfer learning-based feature extraction.
Diagnostics 13 03461 g003
Figure 4. Typical deep learning architecture with a functional layer depicting the transfer learning model for feature extraction and the remaining layers depicting the classification layers.
Figure 4. Typical deep learning architecture with a functional layer depicting the transfer learning model for feature extraction and the remaining layers depicting the classification layers.
Diagnostics 13 03461 g004
Figure 5. Overview of the proposed approach for OSCC detection.
Figure 5. Overview of the proposed approach for OSCC detection.
Diagnostics 13 03461 g005
Figure 6. Proposed deep learning architecture where MGTO is used as an intermediate layer between the feature extraction (functional) layer and the classification layer.
Figure 6. Proposed deep learning architecture where MGTO is used as an intermediate layer between the feature extraction (functional) layer and the classification layer.
Diagnostics 13 03461 g006
Figure 7. Ideal value computation for Max _ Iter .
Figure 7. Ideal value computation for Max _ Iter .
Diagnostics 13 03461 g007
Figure 8. Ideal value computation of the p and β parameters.
Figure 8. Ideal value computation of the p and β parameters.
Diagnostics 13 03461 g008
Figure 9. Confusion matrix of deep learning models without an intermediate layer.
Figure 9. Confusion matrix of deep learning models without an intermediate layer.
Diagnostics 13 03461 g009
Figure 10. Confusion matrix of deep learning models with MGTO as an intermediate layer.
Figure 10. Confusion matrix of deep learning models with MGTO as an intermediate layer.
Diagnostics 13 03461 g010
Figure 11. Training and validation accuracy and loss of deep learning models without an intermediate layer on the first dataset.
Figure 11. Training and validation accuracy and loss of deep learning models without an intermediate layer on the first dataset.
Diagnostics 13 03461 g011
Figure 12. Training and validation accuracy and loss of deep learning models with MGTO as an intermediate layer on the first dataset.
Figure 12. Training and validation accuracy and loss of deep learning models with MGTO as an intermediate layer on the first dataset.
Diagnostics 13 03461 g012aDiagnostics 13 03461 g012b
Figure 13. Percentage accuracy increase due to the usage of intermediate layers in DL models when compared with the accuracy offered by DL models without intermediate layer.
Figure 13. Percentage accuracy increase due to the usage of intermediate layers in DL models when compared with the accuracy offered by DL models without intermediate layer.
Diagnostics 13 03461 g013
Figure 14. Scatter plot of features extracted with MobileNetV3.
Figure 14. Scatter plot of features extracted with MobileNetV3.
Diagnostics 13 03461 g014
Figure 15. Scatter plot of features extracted with InceptioNetV2.
Figure 15. Scatter plot of features extracted with InceptioNetV2.
Diagnostics 13 03461 g015
Figure 16. Scatter plot of features extracted by EfficientNetB3.
Figure 16. Scatter plot of features extracted by EfficientNetB3.
Diagnostics 13 03461 g016
Figure 17. Precision, recall, and F1-score of various OSCC classification models.
Figure 17. Precision, recall, and F1-score of various OSCC classification models.
Diagnostics 13 03461 g017
Figure 18. Pie-chart representing the percentage of training time taken by each DL model with respect to total training time taken by all DL models.
Figure 18. Pie-chart representing the percentage of training time taken by each DL model with respect to total training time taken by all DL models.
Diagnostics 13 03461 g018
Table 1. Stratified shuffle data split into three datasets.
Table 1. Stratified shuffle data split into three datasets.
DatasetClassTotal Number of SamplesNumber of Training SamplesNumber of Validation SamplesNumber of Test Samples
FirstNormal24941746374374
OSCC26981890404404
SecondNormal89631313
OSCC4393076666
ThirdNormal2011413030
OSCC4953477474
Table 2. Specifications of the classification layers and techniques used.
Table 2. Specifications of the classification layers and techniques used.
Classification Layers and Techniques UsedSpecifications
Batch Normalizationmomentum = 0.99, epsilon = 0.001
Denseunits = 256, kernel regularizer = L2 regularizer with coefficient L = 0.016, activity regularizer = L1 regularizer with coefficient L = 0.006, bias regularizer = L1 regularizer with coefficient L = 0.006, activation = ReLu
Dropoutdrop rate = 0.45
Denseunits = 2, activation = SoftMax
Trainingepochs = 100, batch size = 128, stratified shuffle split: training—70%, testing—15%, validation—15%
OptimizerAdamax with learning rate = 0.001, loss = sparse categorical cross-entropy, metrics = accuracy
Early stopping patience = 5, minimum delta = 0, monitor = validation loss, restore best weights = true, mode = minimum
Reduce learning rate on plateaumonitor = validation loss, factor = 0.2, patience = 4, mode = minimum
Table 3. Specifications of the feature extraction layers.
Table 3. Specifications of the feature extraction layers.
Feature Extraction LayersTotal Number of ParametersNumber of Trainable ParametersThe Number of Features Extracted
Mobilenet V32,591,554331,0101280
Efficientnet B311,183,665397,0581536
InceptionV254,736,866397,0581536
Table 4. Performance metrics computed on the test set of the first dataset.
Table 4. Performance metrics computed on the test set of the first dataset.
Transfer Learning ModelIntermediate LayerAccuracyPrecisionRecallF1-Score
MobilenetV3NO0.890.870.920.89
EfficientnetB3NO0.520.5210.68
InceptionV2NO0.880.890.880.88
MobilenetV3PSO0.790.780.820.8
EfficientnetB3PSO0.750.750.780.77
InceptionV2PSO0.820.850.80.82
MobilenetV3EHO0.770.770.80.79
EfficientnetB3EHO0.80.820.780.8
InceptionV2EHO0.830.850.820.83
MobilenetV3GTO0.870.870.880.88
EfficientnetB3GTO0.810.830.80.81
InceptionV2GTO0.860.860.880.87
MobilenetV3MGTO0.950.950.950.95
EfficientnetB3MGTO0.90.920.90.91
InceptionV2MGTO0.930.930.930.93
Table 5. Ideal parameters of various intermediate layers.
Table 5. Ideal parameters of various intermediate layers.
Feature ExtractionIntermediate LayerIdeal Parameter Values
MobilenetV3PSOMax_Iter = 10, w = 0.6, c1 = 0.7, and c2 = 0.9
EHOMax_Iter = 12, α = 0.9, and β   = 0.8
GTO Max _ Iter = 11 ,   p = 0.2, and β = 0.7
MGTO Max _ Iter = 11 ,   p = 0.3, and β = 0.7
EfficientnetB3PSOMax_Iter = 12, w = 0.4, c1 = 0.7, and c2 = 0.9
EHOMax_Iter = 12, α = 0.7, and β   = 0.8
GTO Max _ Iter = 8 ,   p = 0.5, and β = 0.7
MGTO Max _ Iter = 9 ,   p = 0.3, and β = 0.8
InceptionV2PSOMax_Iter = 12, w = 0.6, c1 = 0.8, and c2 = 0.8
EHOMax_Iter = 11, α = 0.8, and β   = 0.6
GTO Max _ Iter = 7 ,   p = 0.4, and β = 0.7
MGTO Max _ Iter = 9 ,   p = 0.4, and β = 0.6
Table 6. Performance metrics computed on the test set of the second dataset.
Table 6. Performance metrics computed on the test set of the second dataset.
Transfer Learning ModelIntermediate LayerAccuracyPrecisionRecallF1-Score
MobilenetV3NO0.80.80.970.88
EfficientnetB3NO0.740.7410.85
InceptionV2NO0.80.820.930.87
MobilenetV3PSO0.780.810.920.86
EfficientnetB3PSO0.750.790.90.84
InceptionV2PSO0.80.810.950.87
MobilenetV3EHO0.760.80.90.85
EfficientnetB3EHO0.750.810.860.84
InceptionV2EHO0.780.820.90.85
MobilenetV3GTO0.820.820.980.89
EfficientnetB3GTO0.780.810.920.86
InceptionV2GTO0.820.830.970.89
MobilenetV3MGTO0.880.880.970.92
EfficientnetB3MGTO0.810.810.970.88
InceptionV2MGTO0.860.8410.91
Table 7. Performance metrics computed on the test set of the third dataset.
Table 7. Performance metrics computed on the test set of the third dataset.
Transfer Learning ModelIntermediate LayerAccuracyPrecisionRecallF1-Score
MobilenetV3NO0.840.860.920.89
EfficientnetB3NO0.710.7110.83
InceptionV2NO0.820.860.890.88
MobilenetV3PSO0.780.840.850.85
EfficientnetB3PSO0.730.810.810.81
InceptionV2PSO0.780.830.870.85
MobilenetV3EHO0.80.840.890.86
EfficientnetB3EHO0.730.810.830.82
InceptionV2EHO0.810.850.890.87
MobilenetV3GTO0.910.920.960.94
EfficientnetB3GTO0.750.830.830.83
InceptionV2GTO0.860.870.950.9
MobilenetV3MGTO0.940.970.850.96
EfficientnetB3MGTO0.90.930.930.93
InceptionV2MGTO0.930.970.930.95
Table 8. Time taken to train various DL models.
Table 8. Time taken to train various DL models.
DL ModelTraining Time (hh:mm:ss)DL ModelTraining Time (hh:mm:ss)
MN00:05:33IN-EHO00:18:22
EN00:15:42MN-GTO00:15:15
IN00:14:37EN-GTO00:19:17
MN-PSO00:09:23IN-GTO00:21:39
EN-PSO00:17:52MN-MGTO00:15:52
IN-PSO00:17:01EN-MGTO00:20:12
MN-EHO00:12:47IN-MGTO00:22:21
EN-EHO00:18:46
Table 9. Comparison of the accuracy attained in related works.
Table 9. Comparison of the accuracy attained in related works.
Related WorkYearClassification FrameworkAccuracy (%) Attained
Rahman A.U. et al. [27]2022AlexNet90.06%
Aberville M. [52] 2017Convolutional Neural Network88.3%
Alkhadar H. [53]2021KNN, Logistic Regression, Decision Tree, Random Forest76%
Alhazmi A. [54] 2021Artificial Neural Network78.95%
Chu C.S. [55] 2020SVM, KNN70.59%
Welikala R.A. [56] 2020ResNet10178.30%
Shavlokhova V. [57] 2021CNN77.89%
Proposed2023Pre-trained MobileNetV3 for feature extraction and MGTO as an intermediate layer95%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nagarajan, B.; Chakravarthy, S.; Venkatesan, V.K.; Ramakrishna, M.T.; Khan, S.B.; Basheer, S.; Albalawi, E. A Deep Learning Framework with an Intermediate Layer Using the Swarm Intelligence Optimizer for Diagnosing Oral Squamous Cell Carcinoma. Diagnostics 2023, 13, 3461. https://doi.org/10.3390/diagnostics13223461

AMA Style

Nagarajan B, Chakravarthy S, Venkatesan VK, Ramakrishna MT, Khan SB, Basheer S, Albalawi E. A Deep Learning Framework with an Intermediate Layer Using the Swarm Intelligence Optimizer for Diagnosing Oral Squamous Cell Carcinoma. Diagnostics. 2023; 13(22):3461. https://doi.org/10.3390/diagnostics13223461

Chicago/Turabian Style

Nagarajan, Bharanidharan, Sannasi Chakravarthy, Vinoth Kumar Venkatesan, Mahesh Thyluru Ramakrishna, Surbhi Bhatia Khan, Shakila Basheer, and Eid Albalawi. 2023. "A Deep Learning Framework with an Intermediate Layer Using the Swarm Intelligence Optimizer for Diagnosing Oral Squamous Cell Carcinoma" Diagnostics 13, no. 22: 3461. https://doi.org/10.3390/diagnostics13223461

APA Style

Nagarajan, B., Chakravarthy, S., Venkatesan, V. K., Ramakrishna, M. T., Khan, S. B., Basheer, S., & Albalawi, E. (2023). A Deep Learning Framework with an Intermediate Layer Using the Swarm Intelligence Optimizer for Diagnosing Oral Squamous Cell Carcinoma. Diagnostics, 13(22), 3461. https://doi.org/10.3390/diagnostics13223461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop