Next Article in Journal
Computational Modeling Approach to Profile Hemodynamical Behavior in a Healthy Aorta
Next Article in Special Issue
Emotion Recognition Using EEG Signals and Audiovisual Features with Contrastive Learning
Previous Article in Journal
Optical Image Sensors for Smart Analytical Chemiluminescence Biosensors
Previous Article in Special Issue
Discriminant Input Processing Scheme for Self-Assisted Intelligent Healthcare Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Oral Squamous Cell Carcinoma Detection Using Histopathological Images: A Deep Feature Fusion and Improved Haris Hawks Optimization-Based Framework

1
Department of Artificial Intelligence and Robotics, Sejong University, Seoul 05006, Republic of Korea
2
Department of Computer Science and Artificial Intelligence, College of Computing, Umm Al-Qura University, Makkah 24382, Saudi Arabia
3
Department of Computer and Network Engineering, College of Computing, Umm Al-Qura University, Makkah 24382, Saudi Arabia
4
Department of Cybersecurity, College of Computing, Umm Al-Qura University, Makkah 24382, Saudi Arabia
*
Author to whom correspondence should be addressed.
Bioengineering 2024, 11(9), 913; https://doi.org/10.3390/bioengineering11090913
Submission received: 16 August 2024 / Revised: 10 September 2024 / Accepted: 11 September 2024 / Published: 12 September 2024

Abstract

:
Oral cancer, also known as oral squamous cell carcinoma (OSCC), is one of the most prevalent types of cancer and caused 177,757 deaths worldwide in 2020, as reported by the World Health Organization. Early detection and identification of OSCC are highly correlated with survival rates. Therefore, this study presents an automatic image-processing-based machine learning approach for OSCC detection. Histopathological images were used to compute deep features using various pretrained models. Based on the classification performance, the best features (ResNet-101 and EfficientNet-b0) were merged using the canonical correlation feature fusion approach, resulting in an enhanced classification performance. Additionally, the binary-improved Haris Hawks optimization (b-IHHO) algorithm was used to eliminate redundant features and further enhance the classification performance, leading to a high classification rate of 97.78% for OSCC. The b-IHHO trained the k-nearest neighbors model with an average feature vector size of only 899. A comparison with other wrapper-based feature selection approaches showed that the b-IHHO results were statistically more stable, reliable, and significant (p < 0.01). Moreover, comparisons with those other state-of-the-art (SOTA) approaches indicated that the b-IHHO model offered better results, suggesting that the proposed framework may be applicable in clinical settings to aid doctors in OSCC detection.

Graphical Abstract

1. Introduction

Mouth cancer, including oral squamous cell carcinoma (OSCC), is one of the most prevalent and fatal diseases and has long been a significant public health concern worldwide [1]. OSCC is a well-known malignant tumor with a high incidence rate, with an estimated 1,401,931 cases reported globally in 2019 [2,3]. As of 2018, OSCC accounted for approximately 25% of upper aerodigestive tract cancer cases in France [4], ranking sixth in terms of cancer frequency. Although betel quid chewing, tobacco use, and excessive alcohol consumption are the primary risk factors for OSCC [5], individuals that do not participate in these activities, especially those aged <45 years, can still develop this malignancy. According to a previous study [6], young women with tongue cancer are at a high risk of developing OSCC.
In 2020, OSCC had a high annual mortality rate of 170,000, primarily because of late-stage detection [4]. According to a 2008 study [7], OSCC is primarily diagnosed in men and women with median ages of 61.5 and 66.4 years, respectively. Another study published in 2022 reported that the average age of patients with OSCC is 62 years [8]. The cancer stage at diagnosis is highly correlated with survival rates. Early screening and detection are essential for improving patient prognosis, and blood tests and visual examinations are conventionally employed for detecting OSCC.
However, clinical diagnoses are more successful for detecting cancer in the oral cavity, primarily through X-rays [9], computed tomography (CT) scans [10], positron emission tomography (PET) scans [11], magnetic resonance imaging (MRI) [12], and endoscopies [13]. The biopsy procedure involves removing tiny tissue samples from the oral region, including malignant tumors. X-rays are frequently used to track changes in the teeth and jawbone anatomies, which aids in detecting any malignant growths that metastasize to the jawbone tissue. CT scans combine several X-rays obtained from various angles to produce intricate 3D images of the mouth and throat, allowing medical professionals to spot anomalies or indications of malignant growths. Additionally, PET scans involve injecting a small quantity of radioactive material into the body and using a specialized camera to detect the material, whereas MRI scans offer a precise visualization of the oral cavity and adjacent structures. This is beneficial for identifying malignancies located deep within tissues. Furthermore, endoscopy is particularly useful for inspecting the throat and larynx and identifying cancers in difficult-to-reach locations. Fluorescence endoscopy can be used to visualize oral malignancies on the outer surfaces that are difficult to detect using white light. However, physical examinations can be expensive, time consuming, and require specialized knowledge for accurate result interpretations.
The rapidly evolving field of artificial intelligence comprises novel diagnostic techniques that can potentially facilitate early OSCC detection and classification [14], which are essential for providing timely and optimal treatment [15]. Several machine and deep learning models have been developed for this purpose [16]. Amin et al. [17] employed three pretrained models, namely InceptionV3, VGG16, and ResNet50, to extract features from histopathology images and merged them to form a new feature vector, achieving a high accuracy of 96.55% compared to individual models. Subsequently, Das et al. [18] developed a deep learning model that obtained better results (97.5% accuracy) than pretrained models. In 2023, Das et al. [19] designed the simplest 10-layer deep convolutional neural network (CNN) architecture to detect OSCC from histopathological images and obtained promising results. In another study [20], a framework comprising the MobileNet-V2 and Darknet-19 models was used to extract deep features, and traditional machine learning classifiers were used for classification, resulting in an accuracy of 92%. It employed a serial-based feature fusion approach for feature concatenation, and chaotic crow search optimization for optimal feature selection, resulting in a high computational complexity.
In conclusion, the aforementioned studies focused on developing deep learning and pretrained models for OSCC detection. However, they did not modify the layer structure or employ hyperparameter optimization, model selection, information fusion, and optimal feature selection to lower the computational time and increase accuracy.
Therefore, this study presents an automatic machine learning-based approach for OSCC detection using histopathological images, wherein pretrained models were used to obtain the deep features from the acquired images. Based on the classification performance, deep features were merged using the canonical correlation feature fusion approach. Subsequently, various wrapper-based approaches were tested to remove redundant features and enhance the classification accuracy. Finally, based on the results, the binary-improved Haris Hawks optimization (b-IHHO) was used to further enhance the classification performance and reduce the feature vector size. The k-nearest neighbors (KNN) algorithm was used to evaluate the classification performance of the proposed framework. Its results were compared with those of other wrapper-based feature selection approaches, and a t-test was conducted to determine their statistical significance. Additionally, the results were compared with those obtained using other SOTA approaches.

2. Materials and Methods

2.1. Proposed OSCC Detection Framework

Histopathological imaging is a reliable method for detecting and diagnosing OSCC. Typically, histopathological images are obtained from tissue samples and examined under a microscope to detect the presence of abnormal cells, providing additional information to guide subsequent investigations and treatments to achieve better outcomes.
This study presents an automatic detection and identification machine learning framework for OSCC. After acquiring the OSCC images, deep features were extracted using various pretrained deep learning models. Next, the deep features extracted from models with classification accuracies greater than 91% were fused using a canonical correlation approach. Subsequently, wrapper-based optimal feature selection approaches were employed to further enhance the classification performance. A flowchart of the proposed OSCC detection approach is shown in Figure 1.

2.2. Histopathological Images Dataset

This study employed the online biopsy dataset Histopathologic Oral Cancer Detection using CNNs (https://www.kaggle.com/datasets/ashenafifasilkebede/dataset; accessed 13 July 2024). Medical experts collected, prepared, and categorized the slides comprising H&E-stained tissues of 230 patients using a Leica ICC50 HD microscope(Leica Microsystems, Wetzlar, Germany). The details and examples of images in these datasets are presented in Table 1.

2.3. Feature Extraction from Histopathological Images

2.3.1. CNNs

CNNs, also known as ConvNETs, are a subclass of artificial neural networks that handle data in a grid-like layout. They can be used to identify various features in an image, such as corners and edges, and effectively eliminate the need for handcrafted feature extraction approaches by including them in their architecture. They comprise various layers, such as input convolution, rectified linear unit (ReLU), and pooling, for extracting image features and information. Finally, a fully connected layer retrieves features for image classification [21,22]. The other fundamental elements of CNNs are the weights, neurons, bias factors, and activation functions.

2.3.2. Deep Feature Extraction Using CNNs

The performance of a CNN can be improved by using a larger training dataset. Transfer learning is a process that allows transferring knowledge from one domain to another. It involves transferring knowledge from a model trained for solving a particular problem and reusing it to solve another related problem. In this study, we assumed a domain with two components [23,24]:
d m = A + p r o b a ,
where A and p r o b ( a ) denote the feature space and marginal probability, respectively. We assume that a task has the following elements:
t r = B + ω ,
where B and ω are space and objective functions, respectively. Additionally, let d s m and t s r denote the source domain and the task, respectively, and d t m and t t r denote the target domains and tasks, respectively. In transfer learning, the source information is used to learn the conditional probabilities of the target domain. Several pretrained models have been developed for various medical imaging applications [25,26]. Figure 2 shows the basic transfer learning concept employed by AlexNet for deep feature extraction from histopathological images in ImageNet.
In this study, various pretrained deep learning models, such as Xception, SqueezeNet, ShuffleNet, ResNet-18, ResNet-50, ResNet-101, NASNet-Mobile, MobileNet-v2, Inception-v3, Inception-ResNet-v2, GoogLeNet, GoogLeNet365, EfficientNet-b0, DenseNet-201, DarkNet-53, and DarkNet-19, were used for deep feature extraction from histopathological images for OSCC detection.

2.3.3. Feature Fusion Using Canonical Correlation Analysis

This study employed a canonical correlation analysis approach to fuse the deep features acquired from histopathological images. The basic principle of the canonical correlation analysis approach is to maximize the correlation between two features. Assuming that there are two feature sets ( f x R a 1 × b and f y R a 2 × b ) with n features, where a 1 and a 2 denote feature dimensions, they can be defined as
f x = f x 1 , f x 1 , . . . , f x n f y = f y 1 , f y 1 , . . . , f y n .
A linear transfer function for the above equation can be defined as
σ = max W x , W y W x T C x y W y W x T C x x W x W y T C y y W y .
Additionally, within-covariance matrices can be defined as C x x R a 1 × a 1 or C x y R a 1 × a 2 , where C x x R a 1 × a 1 denotes the feature set covariance matrix. Hence, the canonical correlation analysis approach can be defined as
C x x 1 C x y C y y 1 C y x W x = σ W x C y y 1 C y x C x x 1 C x y W y = σ W y .
The following equation can be used to compute the final transformed fused vector:
Z ~ = W x T σ x , i + W y T σ y , i = W x T W y T σ x , i σ y , i .

2.3.4. HHO

HHO is a computationally intelligent approach that replicates predator–prey interaction patterns of Harris hawks [27]. It comprises three primary stages: exploration, transformation, and exploitation. HHO has obtained promising results in mining applications owing to its efficient global search capability and minimal parameter adjustments. It employs the following methods to locate prey in diverse locations:
M t + 1 = M r ( t ) r a M r ( t ) 2 r b M ( t ) , q 0.5 M r a b ( t ) M a v g ( t ) r c L + r d u l , q < 0.5 ,
where M ( t ) , M r a b ( t ) , M r ( t ) , and M a v g ( t ) denote the current, rabbit, random, and average positions of the hawks at t , respectively, whereas r is a random value between 0 and 1. Additionally, u and l represent the lower and upper boundaries, respectively. M a v g ( t ) is calculated as follows:
M a v g t = n = 1 N M n t N ,
where N and M n ( t ) denote the population size and the position of the nth individual, respectively. Depending on the prey’s energy ( E e n e r g y ) (defined in Equation (9)), HHO switches between searching and various developmental actions.
E e n e r g y = 2 × E e n e r g y , o 1 t s ,
where E e n e r g y , o is a random value between −1 and 1, s is the maximum number of iterations, and t is the current iteration. If E e n e r g y > 1, it enters the development phase; otherwise, it remains in the search space. Soft and hard besieges occur depending on the conditions for updating the position, which are obtained as follows:
M ( t + 1 ) = M r a b ( t ) M ( t ) E e n e r g y 2 1 r e M r a b ( t ) M ( t ) , 0.5 E e n e r g y < 1   a n d   r e 0.5 M r a b ( t ) E e n e r g y M r a b ( t ) M ( t ) , E e n e r g y < 0.5   a n d   r e 0.5 ,
where r e and M ( t ) denote a random number and the current prey position, respectively. When 0.5 E e n e r g y < 1   a n d   r e < 0.5 , the algorithm uses the following equations to update the position (soft besiege progressive rapid dives approach):
M ( t + 1 ) = A , f ( A ) < f ( M ( t ) ) B , f ( B ) < f ( M ( t ) ) ,
A = M r a b t E e n e r g y 2 1 r e M r a b t M t ,
B = A + r a n d d i m × l e v y d i m ,
where f ( * ) , r a n d , and l e v y denote the fitness function, random vector size of the problem dimension ( d i m ), and Levi’s flight, respectively. When E e n e r g y < 0.5   a n d   r e < 0.5 , the algorithm uses the following equations to update the position (hard besiege progressive rapid dives approach):
M ( t + 1 ) = A , f ( A ) < f ( M ( t ) ) B , f ( B ) < f ( M ( t ) ) ,
A = M r a b t E e n e r g y 2 1 r e M r a b t M a v g t ,
B = A + r a n d d i m × l e v y d i m .

2.3.5. IHHO

In 2023, Peng et al. [28] presented an improved version of the HHO to enhance the individual linkages between the populations of the HHO such that individuals with better fitness values take the lead and influence the remaining population to adjust their positions. Herein, the individuals are ranked according to their fitness values and denoted as α , β , and γ , respectively. Individual α is the first step, and its position-update formula is obtained as follows: When the ratio of the remaining running times of the algorithm to the total running times with the Cauchy random number is compared, the current position has a certain probability of moving closer to the optimal position, whereas the later-stage replacement probability is more negligible, which effectively ensures that the algorithm does not fall into local optimization.
M i ( t + 1 ) = M r a b i ( t ) , tan π ( r a n d 0.5 ) < 1 t s M r a b i ( t ) + 4 t × 4 s × r a n d M m i ( t ) M n i ( t ) , o t h e r
where M m i ( t ) and M n i ( t ) are randomly selected individuals from the population that do not belong to α and i represents the dimensions. The following equations present the position-update formulas for individuals β and γ , respectively. Both M e and M f were randomly selected.
M i ( t + 1 ) = M i ( t ) , r a n d > 0.5 M α i ( t ) + M β i ( t ) 2 , o t h e r ,
M i t + 1 = M e i t + X f i t 2 , r a n d > 0.5 M α i t + M β i t + M γ i t 3 , o t h e r .
Three excellent individuals ( α , β , and γ ) are involved in local development, whereas the others contribute to the original HHO updates. Individuals’ optimal locations were awarded in decreasing order, which encourages localized growth within the field being investigated. Individuals β and γ are connected to α , which improves the communication between outstanding individuals. Compared to the original HHO, the updating technique for these three individuals is relatively simple. This method, known as IHHO, enables quicker execution and higher accuracy for feature selection tasks, while reducing temporal complexity [28]. Additionally, IHHO is converted into a discrete optimization problem, known as b-IHHO, for feature selection. Further details regarding b-IHHO can be found in [28], and the b-IHHO flowchart for selecting the optimal features is illustrated in Figure 3.
Finally, the KNN is used as a classifier to evaluate the selected features using the following fitness function:
F i t n e s s   f u n c t i o n   J = σ 1 C o r r e c t l y   c l a s s i f i e d   i m a g e s T o t a l   n o . o f   i m a g e s + 1 σ f S L f F L ,
where σ , f S L , and f F L denote the weight factor, number of features selected, and total number of features, respectively. The value of σ is 0.99 [29].

3. Results

In this study, various pretrained deep learning models, such as Xception, SqueezeNet, ShuffleNet, ResNet-18, ResNet-50, ResNet-101, NASNet-Mobile, MobileNet-v2, Inception-v3, Inception-ResNet-v2, GoogLeNet, GoogLeNet365, EfficientNet-b0, DenseNet-201, DarkNet-53, and DarkNet-19, were used for deep feature extraction from histopathological images for OSCC detection. MATLAB 2023b was used for processing, running on a PC with the following specifications: 12th Generation Intel(R) Core (TM) i7 CPU, 1 TB SSD, NVIDIA GeForce RTX 3050 GPU, 32 GB of RAM, and 64-bit Windows 11. Additionally, the 0.2 holdout validation approach is used to train and test the models.
First, all the pretrained models were used to extract the deep features before applying the softmax layer. The KNN models were trained using the deep features extracted using each model; the results are presented in Table 2.
A comprehensive analysis of the results revealed that the deep features extracted using the ResNet-101 and EfficientNet-b0 models, which had feature vector sizes of 2048 and 1280, respectively, yielded the highest accuracies of 91.51% and 91.61%, respectively. The canonical correlation feature fusion approach was applied to enhance the classification performance and reduce the feature vector size, as discussed in Section 2.3.3. The results are presented in Figure 4.
The result in Figure 4 indicates that the canonical correlation feature fusion approach further enhanced the classification accuracy to 92.62%, with a feature vector size of 2560. This is because it removes redundant features and fuses them to form a new feature training vector.
To further enhance the classification performance for OSCC detection, various wrapper-based optimal feature selection approaches, such as the marine predator algorithm, generalized normal distribution optimization, slime mold algorithm, equilibrium optimizer, manta ray foraging optimization, atom search optimization, Henry gas solubility optimization, pathfinder algorithm, poor and rich optimization, HHO, and b-IHHO, were employed. The results are presented in Figure 5 using a box-whisker plot for ten runs.
In wrapper-based approaches, the extracted optimal features tested using a machine learning classifier (k-NN) guarantee high reliability and better classification accuracy. Figure 5 shows the classification performance enhancement for all the wrapper-based optimal feature selection approaches. Each algorithm was run ten times, and the results are presented as box-whisker plots. A careful analysis of the results revealed that HHO exhibited the best classification performance. Therefore, the advanced HHO (b-IHHO) variant was employed to further enhance the classification accuracy to 98.28% (mean = 97.78%), as shown in Figure 5. Specifically, b-IHHO resulted in an average increase of 2.32% in the classification performance compared to the simple HHO.
Subsequently, a two-sample t-test was employed to prove the statistical significance and reliability of the results, which were highly accurate (p < 0.01, 99% confidence interval), as demonstrated by the t-test. Furthermore, Cohen’s d value of −5.24 was obtained for the effect sizes of the HHO and b-IHHO results. This value indicates a large effect size, suggesting a significant difference between the two approaches. Additionally, the negative sign indicates that the mean HHO accuracy was lower than the mean accuracy of b-IHHO. The average numbers of features used to train the models are shown in Figure 6.
The results in Figure 6 indicate that each algorithm removed a considerable number of redundant features compared to the fused feature vector (2560) and b-IHHO employed fewer features on average for training the KNN model to obtain a high classification accuracy.

4. Discussion

Various protocols can help doctors detect and diagnose OSCC. One crucial aspect is the detailed examination of histopathological biopsy images, which helps us to understand the disease progression and stage, enabling appropriate and timely treatment. However, a highly skilled pathologist is required to distinguish between healthy and cancerous cells in oral biopsy images. However, this process is time consuming, leading to delayed detection and treatment. Therefore, an automated OSCC detection approach is required for faster and more accurate OSCC diagnosis.
The automatic OSCC detection approach proposed in this study employs deep learning models, feature fusion, and optimal feature selection. Pretrained deep learning models, such as Xception, SqueezeNet, ShuffleNet, ResNet-18, ResNet-50, ResNet-101, NASNet-Mobile, MobileNet-v2, Inception-v3, Inception-ResNet-v2, GoogLeNet, GoogLeNet365, EfficientNet-b0, DenseNet-201, DarkNet-53, and DarkNet-19, were used for feature extraction. However, the extracted features exhibited low classification performance for the binary class problem (Table 2). Therefore, the deep features of the best pretrained models (ResNet-101 and EfficientNet-b0) were fused using a canonical correlation feature fusion approach, resulting in significantly better classification performance. The use of wrapper-based approaches for optimal feature selection guarantees better classification performance because the features are tested using a machine learning model.
The b-IHHO wrapper-based approach was applied to remove redundant features and enhance the classification performance. The results demonstrated that the proposed framework features a high classification accuracy of 97.78 ± 0.33 (average ± standard deviation). The ability of b-IHHO to select more valuable features for classifying histopathological images owes to its effective search strategy. The conventional HHO only uses the objective function to select the features, which leads to a subpar classification performance. Therefore, the b-IHHO employed in this study uses three advanced search strategies in conjunction with an objective function, as discussed in Section 2.3.5. This enables quicker execution and higher accuracy for feature selection tasks, while reducing temporal complexity. Table 3 compares the classification accuracies of the proposed framework and other SOTA approaches.
Achieving a high OSCC detection accuracy has a tremendous significance and far-reaching implications, particularly for the early diagnosis of the disease. Timely and highly accurate OSCC detection using the proposed framework may significantly improve the prognosis and reduce mortality rates. Moreover, automatic histopathological image analysis may help practitioners maximize their workflow efficiency and enhance the diagnostic precision of OSCC detection.

5. Conclusions

This paper presented an automated OSCC detection framework that uses histopathological images for OSCC classification. First, various pretrained deep learning models were used to extract the deep features. The ResNet-101 and EfficientNet-b0 models yielded the highest accuracies of 91.51 and 91.61%, respectively, with 2048 and 1280 feature vector sizes, respectively. Subsequently, canonical correlation feature fusion analysis was conducted to concatenate the features, and an accuracy of 92.62% was achieved with a feature vector size of 2560. Moreover, the wrapper-based approach b-IHHO was used for feature selection and yielded the highest accuracy of 98.28%, with only 899 features. Additionally, comparisons with other wrapper-based feature selection approaches showed that the results of b-IHHO were statistically more stable, reliable, and significant (p < 0.01). Finally, a comparison with other SOTA methods also demonstrated the superiority and high classification performance of the proposed automated OSCC detection approach.

Author Contributions

Conceptualization, A.Z.; formal analysis, M.K. and M.F.; funding acquisition, S.-H.K.; investigation, M.F. and H.F.M.L.; methodology, A.Z. and S.-H.K.; project administration, supervision S.-H.K.; software, A.Z.; validation, M.K. and S.-H.K.; visualization, T.M.Q.; writing—original draft, A.Z.; writing—review and editing, A.Z., M.K., T.M.Q. and H.F.M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the MSIT (Ministry of Science and ICT (MSIT), Korea, under the ICAN (ICT Challenge and Advanced Network of HRD (ICAN) program (IITP-2024-RS-2022-00156345) supervised by the IITP (Institute of Information & Communications Technology Planning & Evaluation (IITP). This work was also partially supported by the National Research Foundation of Korea (NRF) grant (No. RS-2023-00219051 and RS-2023-00209107), and the Unmanned Vehicles Core Technology Research and Development Program through the NRF and Unmanned Vehicle Advanced Research Center (UVARC), funded by the Ministry of Science and ICT, Republic of Korea (NRF-2023M3C1C1A01098408).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in Kaggle at https://www.kaggle.com/datasets/ashenafifasilkebede/dataset, accessed 13 July 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ghantous, Y.; Abu Elnaaj, I. Global incidence and risk factors of oral cancer. Harefuah 2017, 156, 645–649. [Google Scholar] [PubMed]
  2. Zygogianni, A.G.; Kyrgias, G.; Karakitsos, P.; Psyrri, A.; Kouvaris, J.; Kelekis, N.; Kouloulias, V. Oral squamous cell cancer: Early detection and the role of alcohol and smoking. Head Neck Oncol. 2011, 3, 2. [Google Scholar] [CrossRef] [PubMed]
  3. Khijmatgar, S.; Yong, J.; Rübsamen, N.; Lorusso, F.; Rai, P.; Cenzato, N.; Gaffuri, F.; Del Fabbro, M.; Tartaglia, G.M. Salivary biomarkers for early detection of oral squamous cell carcinoma (OSCC) and head/neck squamous cell carcinoma (HNSCC): A systematic review and network meta-analysis. Jpn. Dent. Sci. Rev. 2024, 60, 32–39. [Google Scholar] [CrossRef] [PubMed]
  4. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef]
  5. Boccia, S.; Hashibe, M.; Gallì, P.; De Feo, E.; Asakage, T.; Hashimoto, T.; Hiraki, A.; Katoh, T.; Nomura, T.; Yokoyama, A.; et al. Aldehyde dehydrogenase 2 and head and neck cancer: A meta-analysis implementing a Mendelian randomization approach. Cancer Epidemiol. Biomark. Prev. 2009, 18, 248–254. [Google Scholar] [CrossRef]
  6. Patel, S.C.; Carpenter, W.R.; Tyree, S.; Couch, M.E.; Weissler, M.; Hackman, T.; Hayes, D.N.; Shores, C.; Chera, B.S. Increasing incidence of oral tongue squamous cell carcinoma in young white women, age 18 to 44 years. J. Clin. Oncol. 2011, 29, 1488–1494. [Google Scholar] [CrossRef]
  7. Müller, S.; Pan, Y.; Li, R.; Chi, A.C. Changing trends in oral squamous cell carcinoma with particular reference to young patients: 1971–2006. The Emory University experience. Head Neck Pathol. 2008, 2, 60–66. [Google Scholar] [CrossRef]
  8. Ferreira E Costa, R.; Leão, M.L.B.; Sant’Ana, M.S.P.; Mesquita, R.A.; Gomez, R.S.; Santos-Silva, A.R.; Khurram, S.A.; Tailor, A.; Schouwstra, C.-M.; Robinson, L.; et al. Oral squamous cell carcinoma frequency in young patients from referral centers around the world. Head Neck Pathol. 2022, 16, 755–762. [Google Scholar] [CrossRef]
  9. Faeli Ghadikolaei, R.; Ghorbani, H.; Seyedmajidi, M.; Ebrahimnejad Gorji, K.; Moudi, E.; Seyedmajidi, S. Genotoxicity and cytotoxicity effects of X-rays on the oral mucosa epithelium at different fields of view: A cone beam computed tomography technique. Casp. J. Intern. Med. 2023, 14, 121–127. [Google Scholar] [CrossRef]
  10. Nien, H.-H.; Wang, L.-Y.; Liao, L.-J.; Lin, P.-Y.; Wu, C.-Y.; Shueng, P.-W.; Chung, C.-S.; Lo, W.-C.; Lin, S.-C.; Hsieh, C.-H. Advances in image-guided radiotherapy in the treatment of oral cavity cancer. Cancers 2022, 14, 4630. [Google Scholar] [CrossRef]
  11. Marcus, C.; Subramaniam, R.M. PET imaging of oral cavity and oropharyngeal cancers. PET Clin. 2022, 17, 223–234. [Google Scholar] [CrossRef]
  12. Maraghelli, D.; Pietragalla, M.; Calistri, L.; Barbato, L.; Locatello, L.G.; Orlandi, M.; Landini, N.; Lo Casto, A.; Nardi, C. Techniques, tricks, and stratagems of oral cavity computed tomography and magnetic resonance imaging. Appl. Sci. 2022, 12, 1473. [Google Scholar] [CrossRef]
  13. Azam, M.A.; Sampieri, C.; Ioppi, A.; Benzi, P.; Giordano, G.G.; De Vecchi, M.; Campagnari, V.; Li, S.; Guastini, L.; Paderno, A.; et al. Videomics of the upper aero-digestive tract cancer: Deep learning applied to white light and narrow band imaging for automatic segmentation of endoscopic images. Front. Oncol. 2022, 12, 900451. [Google Scholar] [CrossRef]
  14. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  15. Tshering Vogel, D.W.T.; Zbaeren, P.; Thoeny, H.C. Cancer of the oral cavity and oropharynx. Cancer Imaging 2010, 10, 62–72. [Google Scholar] [CrossRef]
  16. de Chauveron, J.; Unger, M.; Lescaille, G.; Wendling, L.; Kurtz, C.; Rochefort, J. Artificial intelligence for oral squamous cell carcinoma detection based on oral photographs: A comprehensive literature review. Cancer Med. 2024, 13, e6822. [Google Scholar] [CrossRef]
  17. Amin, I.; Zamir, H.; Khan, F.F. Histopathological image analysis for oral squamous cell carcinoma classification using concatenated deep learning models. medRxiv 2021. [Google Scholar] [CrossRef]
  18. Das, N.; Hussain, E.; Mahanta, L.B. Automated classification of cells into multiple classes in epithelial tissue of oral squamous cell carcinoma using transfer learning and convolutional neural network. Neural Netw. 2020, 128, 47–60. [Google Scholar] [CrossRef]
  19. Das, M.; Dash, R.; Mishra, S.K. Automatic detection of oral squamous cell carcinoma from histopathological images of oral mucosa using deep convolutional neural network. Int. J. Environ. Res. Public Health 2023, 20, 2131. [Google Scholar] [CrossRef]
  20. Khan, M.A.; Mir, M.; Ullah, M.S.; Hamza, A.; Jabeen, K.; Gupta, D. A fusion framework of pre-trained deep learning models for oral squamous cell carcinoma classification. In Proceedings of the Third International Conference on Computing and Communication Networks, Manchester, UK, 17–18 November 2023; Springer: Singapore, 2024; pp. 769–782. [Google Scholar] [CrossRef]
  21. Akram, M.W.; Li, G.; Jin, Y.; Chen, X.; Zhu, C.; Ahmad, A. Automatic detection of photovoltaic module defects in infrared images with isolated and develop-model transfer deep learning. Sol. Energy 2020, 198, 175–186. [Google Scholar] [CrossRef]
  22. Oyetade, I.S.; Ayeni, J.O.; Ogunde, A.O.; Oguntunde, B.O.; Olowookere, T.A. Hybridized deep convolutional neural network and fuzzy support vector machines for breast cancer detection. SN Comput. Sci. 2022, 3, 581. [Google Scholar] [CrossRef]
  23. Fatima, M.; Khan, M.A.; Shaheen, S.; Almujally, N.A.; Wang, S.-H. B2C3NetF2: Breast cancer classification using an end-to-end deep learning feature fusion and satin bowerbird optimization controlled Newton Raphson feature selection. CAAI Trans. Intell. Technol. 2023, 8, 1374–1390. [Google Scholar] [CrossRef]
  24. Zahoor, S.; Shoaib, U.; Lali, I.U. Breast cancer mammograms classification using deep neural network and entropy-controlled whale optimization algorithm. Diagnostics 2022, 12, 557. [Google Scholar] [CrossRef]
  25. Baltruschat, I.M.; Nickisch, H.; Grass, M.; Knopp, T.; Saalbach, A. Comparison of deep learning approaches for multi-label chest X-ray classification. Sci. Rep. 2019, 9, 6381. [Google Scholar] [CrossRef]
  26. Kang, J.; Gwak, J. Ensemble of instance segmentation models for polyp segmentation in colonoscopy images. IEEE Access 2019, 7, 26440–26447. [Google Scholar] [CrossRef]
  27. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  28. Peng, L.; Cai, Z.; Heidari, A.A.; Zhang, L.; Chen, H. Hierarchical Harris hawks optimizer for feature selection. J. Adv. Res. 2023, 53, 261–278. [Google Scholar] [CrossRef]
  29. Agrawal, P.; Abutarboush, H.F.; Ganesh, T.; Mohamed, A.W. Metaheuristic algorithms on feature selection: A survey of one decade of research (2009–2019). IEEE Access 2021, 9, 26766–26791. [Google Scholar] [CrossRef]
  30. Sukegawa, S.; Ono, S.; Tanaka, F.; Inoue, Y.; Hara, T.; Yoshii, K.; Nakano, K.; Takabatake, K.; Kawai, H.; Katsumitsu, S.; et al. Effectiveness of deep learning classifiers in histopathological diagnosis of oral squamous cell carcinoma by pathologists. Sci. Rep. 2023, 13, 11676. [Google Scholar] [CrossRef] [PubMed]
  31. Yu, M.; Ding, J.; Liu, W.; Tang, X.; Xia, J.; Liang, S.; Jing, R.; Zhu, L.; Zhang, T. Deep multi-feature fusion residual network for oral squamous cell carcinoma classification and its intelligent system using Raman spectroscopy. Biomed. Signal Process Control 2023, 86, 105339. [Google Scholar] [CrossRef]
  32. Chang, X.; Yu, M.; Liu, R.; Jing, R.; Ding, J.; Xia, J.; Zhu, Z.; Li, X.; Yao, Q.; Zhu, L.; et al. Deep learning methods for oral cancer detection using Raman spectroscopy. Vib. Spectrosc. 2023, 126, 103522. [Google Scholar] [CrossRef]
  33. Panigrahi, S.; Nanda, B.S.; Bhuyan, R.; Kumar, K.; Ghosh, S.; Swarnkar, T. Classifying histopathological images of oral squamous cell carcinoma using deep transfer learning. Heliyon 2023, 9, e13444. [Google Scholar] [CrossRef] [PubMed]
  34. Yang, Z.; Pan, H.; Shang, J.; Zhang, J.; Liang, Y. Deep-learning-based automated identification and visualization of oral cancer in optical coherence tomography images. Biomedicines 2023, 11, 802. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart of the proposed OSCC-detection framework using histopathological images.
Figure 1. Flowchart of the proposed OSCC-detection framework using histopathological images.
Bioengineering 11 00913 g001
Figure 2. Modified AlexNet for deep feature extraction through transfer learning.
Figure 2. Modified AlexNet for deep feature extraction through transfer learning.
Bioengineering 11 00913 g002
Figure 3. b-IHHO flowchart for selecting optimal deep features.
Figure 3. b-IHHO flowchart for selecting optimal deep features.
Bioengineering 11 00913 g003
Figure 4. Confusion matrix for OSCC detection obtained by applying the canonical correlation feature fusion approach.
Figure 4. Confusion matrix for OSCC detection obtained by applying the canonical correlation feature fusion approach.
Bioengineering 11 00913 g004
Figure 5. Classification performances of various wrapper-based approaches for OSCC detection with b-IHHO over ten runs (p < 0.01). MPA: marine predator algorithm, GNDO: generalized normal distribution optimization, SMA: slime mold algorithm, EO: equilibrium optimizer, MRFO: manta ray foraging optimization, ASO: atom search optimization, HGSO: Henry gas solubility optimization, PFA: pathfinder algorithm, PRO: poor and rich optimization, HHO: Harris hawks optimization, and b-IHHO: binary improved HHO.
Figure 5. Classification performances of various wrapper-based approaches for OSCC detection with b-IHHO over ten runs (p < 0.01). MPA: marine predator algorithm, GNDO: generalized normal distribution optimization, SMA: slime mold algorithm, EO: equilibrium optimizer, MRFO: manta ray foraging optimization, ASO: atom search optimization, HGSO: Henry gas solubility optimization, PFA: pathfinder algorithm, PRO: poor and rich optimization, HHO: Harris hawks optimization, and b-IHHO: binary improved HHO.
Bioengineering 11 00913 g005
Figure 6. Optimal number of features for each wrapper-based method (average ± standard deviation). MPA: marine predator algorithm, GNDO: generalized normal distribution optimization, SMA: slime mold algorithm, EO: equilibrium optimizer, MRFO: manta ray foraging optimization, ASO: atom search optimization, HGSO: Henry gas solubility optimization, PFA: pathfinder algorithm, PRO: poor and rich optimization, HHO: Harris hawks optimization, and b-IHHO: binary improved HHO.
Figure 6. Optimal number of features for each wrapper-based method (average ± standard deviation). MPA: marine predator algorithm, GNDO: generalized normal distribution optimization, SMA: slime mold algorithm, EO: equilibrium optimizer, MRFO: manta ray foraging optimization, ASO: atom search optimization, HGSO: Henry gas solubility optimization, PFA: pathfinder algorithm, PRO: poor and rich optimization, HHO: Harris hawks optimization, and b-IHHO: binary improved HHO.
Bioengineering 11 00913 g006
Table 1. Details and sample images in the OSCC biopsy datasets employed in this study.
Table 1. Details and sample images in the OSCC biopsy datasets employed in this study.
NormalSick (OSCC)
Histopathological imagesBioengineering 11 00913 i001Bioengineering 11 00913 i002
Images per class24352511
Table 2. Classification performances of various pretrained models for OSCC detection.
Table 2. Classification performances of various pretrained models for OSCC detection.
ModelFeature Vector SizeAccuracy (%)
Xception204887.77
SqueezeNet100082.91
ShuffleNet54484.93
ResNet-1851286.05
ResNet-50204889.69
ResNet-101204891.51
NASNet-Mobile105684.73
MobileNet-v2128084.53
Inception-v3204886.86
Inception-ResNet-v2153689.69
GoogLeNet102481.60
GoogLeNet365102485.04
EfficientNet-b0128091.61
DenseNet-201192087.87
DarkNet-53102486.96
DarkNet-19100087.36
Table 3. Classification accuracies of the proposed framework and other SOTA approaches.
Table 3. Classification accuracies of the proposed framework and other SOTA approaches.
StudyAccuracy (%)
Sukegawa et al. [30]86.22
Khan et al. [20]92
Yu et al. [31]92.78
Chang et al. [32]92.81
Panigrahi et al. [33]96.6
Yang et al. [34]92.52
Das et al. [19]97.82
This study98.28 (mean = 97.78)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zafar, A.; Khalid, M.; Farrash, M.; Qadah, T.M.; Lahza, H.F.M.; Kim, S.-H. Enhancing Oral Squamous Cell Carcinoma Detection Using Histopathological Images: A Deep Feature Fusion and Improved Haris Hawks Optimization-Based Framework. Bioengineering 2024, 11, 913. https://doi.org/10.3390/bioengineering11090913

AMA Style

Zafar A, Khalid M, Farrash M, Qadah TM, Lahza HFM, Kim S-H. Enhancing Oral Squamous Cell Carcinoma Detection Using Histopathological Images: A Deep Feature Fusion and Improved Haris Hawks Optimization-Based Framework. Bioengineering. 2024; 11(9):913. https://doi.org/10.3390/bioengineering11090913

Chicago/Turabian Style

Zafar, Amad, Majdi Khalid, Majed Farrash, Thamir M. Qadah, Hassan Fareed M. Lahza, and Seong-Han Kim. 2024. "Enhancing Oral Squamous Cell Carcinoma Detection Using Histopathological Images: A Deep Feature Fusion and Improved Haris Hawks Optimization-Based Framework" Bioengineering 11, no. 9: 913. https://doi.org/10.3390/bioengineering11090913

APA Style

Zafar, A., Khalid, M., Farrash, M., Qadah, T. M., Lahza, H. F. M., & Kim, S. -H. (2024). Enhancing Oral Squamous Cell Carcinoma Detection Using Histopathological Images: A Deep Feature Fusion and Improved Haris Hawks Optimization-Based Framework. Bioengineering, 11(9), 913. https://doi.org/10.3390/bioengineering11090913

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop