Next Article in Journal
Identification of a Twelve-microRNA Signature with Prognostic Value in Stage II Microsatellite Stable Colon Cancer
Next Article in Special Issue
Machine Learning Integrating 99mTc Sestamibi SPECT/CT and Radiomics Data Achieves Optimal Characterization of Renal Oncocytic Tumors
Previous Article in Journal
Optimized Modeling of Metastatic Triple-Negative Invasive Lobular Breast Carcinoma
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Al-Biruni Earth Radius Optimization with Transfer Learning Based Histopathological Image Analysis for Lung and Colon Cancer Detection

1
Information Technology Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
2
Department of Biology, College of Science and Arts at Alkamil, University of Jeddah, Jeddah, Saudi Arabia
3
Department of Software Engineering, College of Computer Science and Engineering, University of Jeddah, Jeddah, Saudi Arabia
4
Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
5
Department of Biochemistry, Faculty of Sciences, King Abdulaziz University, Jeddah 21589, Saudi Arabia
6
Mathematics Department, Faculty of Science, Al-Azhar University, Naser City 11884, Cairo, Egypt
7
King Abdulaziz University-University of Oxford Centre for Artificial Intelligence in Precision Medicines, King Abdulaziz University, Jeddah 21589, Saudi Arabia
*
Author to whom correspondence should be addressed.
Cancers 2023, 15(13), 3300; https://doi.org/10.3390/cancers15133300
Submission received: 1 June 2023 / Revised: 13 June 2023 / Accepted: 14 June 2023 / Published: 23 June 2023
(This article belongs to the Special Issue Image Analysis and Machine Learning in Cancers)

Abstract

:

Simple Summary

Histopathological image analysis can be used for the detection of lung and colon cancer by the investigation of microscopic images of tissue samples. Since manual diagnosis takes a long time and is subjected to differing opinions of doctors, automated lung and colon cancer diagnosis becomes necessary. Therefore, the purpose of the study is to develop a transfer learning approach for lung and colon cancer detection on histopathological image analysis. It involves leveraging pre-trained model to analyze histopathological images. In addition, the proposed model uses improved ShuffleNet with deep convolutional recurrent neural network for feature extraction and classification, respectively. Besides, Al-Biruni Earth Radius Optimization and coati optimization algorithm are employed for hyperparameter tuning process. The experimental result analysis of the proposed model on the LC25000 database shows its promising performance on lung and colon cancer diagnosis.

Abstract

An early diagnosis of lung and colon cancer (LCC) is critical for improved patient outcomes and effective treatment. Histopathological image (HSI) analysis has emerged as a robust tool for cancer diagnosis. HSI analysis for a LCC diagnosis includes the analysis and examination of tissue samples attained from the LCC to recognize lesions or cancerous cells. It has a significant role in the staging and diagnosis of this tumor, which aids in the prognosis and treatment planning, but a manual analysis of the image is subject to human error and is also time-consuming. Therefore, a computer-aided approach is needed for the detection of LCC using HSI. Transfer learning (TL) leverages pretrained deep learning (DL) algorithms that have been trained on a larger dataset for extracting related features from the HIS, which are then used for training a classifier for a tumor diagnosis. This manuscript offers the design of the Al-Biruni Earth Radius Optimization with Transfer Learning-based Histopathological Image Analysis for Lung and Colon Cancer Detection (BERTL-HIALCCD) technique. The purpose of the study is to detect LCC effectually in histopathological images. To execute this, the BERTL-HIALCCD method follows the concepts of computer vision (CV) and transfer learning for accurate LCC detection. When using the BERTL-HIALCCD technique, an improved ShuffleNet model is applied for the feature extraction process, and its hyperparameters are chosen by the BER system. For the effectual recognition of LCC, a deep convolutional recurrent neural network (DCRNN) model is applied. Finally, the coati optimization algorithm (COA) is exploited for the parameter choice of the DCRNN approach. For examining the efficacy of the BERTL-HIALCCD technique, a comprehensive group of experiments was conducted on a large dataset of histopathological images. The experimental outcomes demonstrate that the combination of AER and COA algorithms attain an improved performance in cancer detection over the compared models.

1. Introduction

Cancer can affect any organ of the human body; the foremost areas to be commonly affected are the brain, colon, skin, breasts, stomach, rectum, liver, prostate, and lungs. The common tumors to cause death in females and males are lung and colon cancer (LCC) [1]. When lung cells mutate uncontrollably, malignant cells appear, forming clusters called cancers. Globally, lung and colorectal (rectum and colon) tumors are the general kinds of tumors after breast cancer (BC) [2]. Moreover, colorectal and lung tumors have resulted in death rates of 9.4% and 18% correspondingly among all tumors. Hence, to explore the treatment options in the early phase of diseases, the precise detection of these tumor subtypes are essential. The noninvasive approaches for detection involve computed tomography (CT) imaging and radiography for flexible sigmoidoscopy and lung cancer and CT colonoscopy for colon tumors [3], but dependable classifying of these tumors is not probable utilizing noninvasive means at all times, and invasive processes such as histopathology are essential for accurate disease detection and the enhanced quality of treatments. As well, the manual grading of histopathologic images may be annoying for pathologists [4]. Likewise, the precise grading of the colon and lung tumor subtypes necessitates a trained pathologist, and manual grading can be prone to error. Now, automatic image processing techniques are being applied for lung tumors [5].
Artificial intelligence (AI) methods are utilized in the medical domain, such as the initial detection of biomedical images, health disasters, forecasts of diseases, etc. [6]. Deep learning (DL) methods can examine data in anatomical representations, high-dimensional images, and videos [7]. Likewise, DL methods derive hidden characteristics and features from healthcare images that are invisible to the naked eye for the initial cancer recognition and discrimination between its phases. DL refers to a subdivision of ML that eradicates the necessity for manual feature engineering, and CNN-related DL methods present hierarchical mapping features for superior representations of input images [8]. However, enormous data are needed for training large DL methods; transfer learning (TL) aids in adapting large pretraining methods for downstream tasks. Therefore, TL decreases the necessity for an enormous dataset for training that is rare in particular domains such as medicine [9]. TL and DL execute a crucial role in healthcare in framing automatic diagnostic mechanisms by utilizing healthcare images that include magnetic resonance images, histopathological images, radiographs, retina images, etc. Such automatic mechanisms are mainly utilized for classifier tasks and assist doctors in circumstances of automated quality checking and rapid data acquisition [10].
This manuscript offers the design of the Al-Biruni Earth Radius Optimization with Transfer Learning-based Histopathological Image Analysis for Lung and Colon Cancer Detection (BERTL-HIALCCD) technique. In the BERTL-HIALCCD technique, an improved ShuffleNet model is applied for the feature extraction process, and its hyperparameters are chosen by the BER algorithm. For the effectual detection of LCC, a deep convolutional recurrent neural network (DCRNN) model is applied. At the final stage, the coati optimization algorithm (COA) is exploited for the parameter selection of the DCRNN approach. The design of BER and COA for the hyperparameter tuning of the improved ShuffleNet and DCRNN models demonstrates the novelty of the work. To examine the result of the BERTL-HIALCCD technique, a comprehensive group of experiments was conducted on a large dataset of histopathological images.

2. Related Works

In [11], the authors utilized an AI-supported method and optimization approaches to realize the categorization of histopathologic images of colon and lung tumors. In the presented method, the image classes were trained with the DarkNet-19 technique, one of which was the DL method. With the equilibrium and Manta Ray Foraging optimizer methods, the selection of the ineffective attributes was attained. In the feature set mined from the DarkNet19 method, the potential attributes gained by the two utilized optimization methods were classified and integrated with the SVM approach. In [12], a hybrid classification method that included hog and daisy feature extraction modules and inception_v3 network were built to categorize lung tumors and normal tissues from lung pathological imagery. In [13], the authors presented a brief analysis of two feature extraction approaches for colon and lung tumor classification. In one method presented, six handcrafted extracted features methods dependent on shape, color, structure, and texture. The RF, Gradient Boosting (GB), MLP, and SVM-RBF methods with handcrafted attributes were tested and trained for lung and colon tumor categorization. In [14], the main intention of this case was to utilize digital histopathology images and a multi-input capsule network for framing an enhanced computerized diagnosis mechanism to find adenocarcinomas and squamous cell carcinomas of the lungs, in addition to adenocarcinomas of the colon. During the presented multi-input capsule network, two convolution layer blocks were utilized. The CLB (convolutional layer blocks) considered unprocessed histopathologic images as the input.
The authors of [15] presented a hybrid ensemble extracted feature method to proficiently find the LCC. It integrated ensemble learning and deep feature extraction with high-performance filters for cancer image data. Hamida et al. [16] concentrated on the usage of a DL structure for highlighting and classifying colon tumor regions in a sparsely annotated histopathologic data context. Firstly, the authors compared and reviewed the existing CNN that included the DenseNet, AlexNet, vgg, and ResNet Inception methods. The authors resorted to the use of TL methods to cope with the lack of a rich WSI dataset.
Mangal et al. [17] aimed to devise a computer-aided diagnosis system to find squamous cell carcinomas and adenocarcinomas of the lungs, in addition to adenocarcinomas of the colon, utilizing CNN by assessing the digital pathology images for these cancers. Ding et al. [18] designed FENet for genetic mutation estimation utilizing histopathologic images of colon cancer. Different to traditional methods of analyzing patch-related features alone without concerning their spatial connectivity, FENet incorporated feature enhancements in convolutional graph NN to combine discriminatory attributes with the capture gene mutation status.
In [19], a clinically comparable CNN structure-based approach to carry out an automated classifier of cancer grades and tissue infrastructures in hematoxylin and eosin-stained colon HSI was presented. It contained Enhanced Convolutional Learning Modules (ECLMs), a multi-level Attention Learning Module (ALM), and Transitional Modules (TMs). In [20], the efficiency of an extensive variety of DL-based structures was measured for the automatic tumor segment of colorectal tissue instances. The presented method demonstrated the efficacy of integrating CNN and TL in the encoded part of the segmentation structure for histopathology image diagnosis. In [21], the authors provided a systematic examination of XAI with an initial concentration on the methods that are presently being utilized in the domain of the healthcare system. In [22], the authors established a DL technique to predict disease-specific survival for stage II and III colorectal cancer utilizing 3652 cases. In [23], a novel approach dependent upon GCN for the early detection of COPD was presented that utilized lesser and weakly labeled chest CT image data in the openly accessible Danish Lung Cancer Screening Trial database. Jain et al. [24] demonstrated that lung cancer recognition depends on a histopathological image diagnosis utilizing DL structures. Afterwards, the image features were extracted utilizing Kernel PCA combined with CNN with KPCA (KPCA-CNN) utilized in the feature extracted layer of the CNN.
Although several LCC classification models are available in the literature, it is still required to enhance the detection performance. Most of the existing works did not focus on the hyperparameter optimization process. Generally, hyperparameter optimization helps in the identification of an optimal combination of hyperparameters for a given model architecture and dataset. It proficiently searches the hyperparameter space for the detection of the optimal configuration, which saves time and computational resources by automatically exploring different combinations rather than relying on manual trial-and-error approaches. It helps in quickly finding a good set of hyperparameters without exhaustively evaluating every possible combination. Therefore, in this work, the BER and COA are used for the hyperparameter tuning process.

3. Materials and Methods

In this manuscript, we developed a novel LCC detection model named the BERTL-HIALCCD technique. The study aims to find LCC effectually on histopathological images. To achieve this, the BERTL-HIALCCD approach follows a series of subprocesses, namely improved ShuffleNet feature extraction, BER-based parameter optimization, DCRNN-based classification, and COA-based hyperparameter tuning. Figure 1 signifies the overall flow of the BERTL-HIALCCD system.

3.1. Dataset Used

The LC25000 dataset comprises 25,000 color images with 5 class labels of 5000 images each. The dataset is available at https://www.kaggle.com/datasets/andrewmvd/lung-and-colon-cancer-histopathological-images (accessed on 12 June 2023). Every image is 768 × 768 pixels and in .jpeg file format. The 5 class labels can be lung squamous cell carcinomas, colon adenocarcinomas, lung adenocarcinomas, and benign lung and benign colonic tissues.

3.2. Feature Extraction Using Improved ShuffleNet

In this work, the improved ShuffleNet method extracts features from histopathological images. Depthwise convolution (DW-Conv) is a special case of gathering convolutions in general and then evaluating a typical convolution with  1  ×  1  dimensions for the depthwise convolutional procedure [25]:
G i , j , m = w , h W , H K w , h , m X i + w , j + h , m
In Equation (1),  G   and   K  represent the resultant feature matrix and convolutional kernel weighted matrix, correspondingly.  i ,   j ,   w ,   and   h  represents the coordinates of the related matrix. The depthwise separable convolution replaces the typical convolution, has lesser computations, and is typical for lightweight methods:
Q 1 = D f 2 D k 2 M + D f 2 M
Q 2 = D f 2 D k 2 M N
Q 1 Q 2 = D f 2 D k 2 M + D f 2 M N D f 2 D k 2 M N = 1 N + 1 D k 2
where  D f  and  D k  show the side size of the feature matrix and convolutional kernels, and  Q 1  and  Q 2  denote the computation count of depthwise convolutional and typical convolution correspondingly.  M  and  N  signify the number of channels in the output and input feature maps. Thus, the computation amount of the depthwise convolution is 1/9 of the typical convolution.
The channel shuffle method is a significant innovation point developed by ShuffleNet that realizes the data interchange of channels from the extracted feature method with smaller computation costs. In this work, effective channel attention (ECA) was established to suppress the relevant attributes and accomplish the weighting method of the features for succeeding the classifier models. However, when retaining the lightness, the LSR loss function can be performed when considering the multi-dimensional loss computation and enhances the model noise immunity.
The visual attention modules are drawn from human visual features to emphasize crucial data in images that are advantageous for enhancing the performance of the model. Visual attention mechanism brings accuracy improvements to CNNs by weighting the outcome features but mainly at the cost of enhancing the difficulty, namely convolution block attention mechanism (CBAM) SE. ECA refers to a lightweight attention model that borrowed the idea from SE to construct a channel attention module that is established in CNN to have participated during the end-to-end trained model. ECA exploits 1D convolutions for an extracted feature that prevents feature downscaling and efficiently captures cross-channel data interactions.
Assume the input feature matrix is  F C × H × W , and  W C ,  and  H  characterize the width, channels, and height of the input features, correspondingly. First, the input matrix can be processed via a global average pooling layer that leads to the channel feature description matrix  F C × 1 × 1 . Next, feature extraction can be executed utilizing  a   1 D  convolutional layer, and the output can be processed via a nonlinear activation function:
M c ( F ) = σ ( f 1 d ( F a v g ) )
In Equation (5),  σ  represents the sigmoid activation function, and  f 1 d  designates the  1 D  convolutional process. Lastly, the input feature is multiplied with the attention weight for the channel size:
F = F M c
In Equation (6),   represents elemental multiplication,  M c  is copies within the spatial size to attain  C × H × W  feature matrices that are later point multiplied with another matrix. The ECA belongs to the channel attention that enables the allocation of the weight to the channel of the feature maps, which makes the network focus on the most crucial channels. Thus, after this layer, the ECA module is embedded.

3.3. BER-Based Hyperparameter Optimization

For the effectual identification of the hyperparameter values of the improved ShuffleNet method, the BER algorithm is used. The optimizer technique aims to find the optimum solution for the problem with a set of constraints [26]. In this work, an individual from the population is characterized by the vector  S = { S 1 ,   S 2 ,   ,   S d } R d , where  d  represents the searching space size, and  S j  indicates the feature or parameter from the optimizer problems. The fitness function  f  is used for determining how well an individual performs up to the provided point. The optimization process exploits the subsequent phase to search through the population for the certain optimal vector  S * , which increases the FF. This technique starts with a set of solutions. The following parameters are used for the optimization process: the population size, the FF, the dimension, and the lower and upper boundaries for the solution space.
In the proposed AER algorithm, the population is split into subgroups. The number of individuals in every group can be adapted to enhance the balance between the exploration and exploitation processes. Moreover, to assure the convergence of optimization techniques for the population, the elitism approach can be applied by keeping the leading solution when no best solution is found. If the fitness of the solution does not increase dramatically for 3 iterations, this is the local optima, and subsequently, other exploration individuals are produced by using the mutation process.
Exploration avoids local optimal stagnation by movement towards a better solution and is responsible for finding exciting locations in the search space.
Heading towards the best solution strategy is used to search for prospective regions around its existing location in the search space. This could be obtained by repetitively searching amongst near promising alternatives for a better option with respect to the fitness value:
r = h c o s ( x ) 1 cos ( x )
D = r 1 · ( S ( t ) 1 )
S ( t + 1 ) = S ( t ) + D · ( 2 r 2 1 )
where  D  denotes the diameter of the circle where the searching agent finds the potential area. If  0 < x 180 ,   h  denotes a random number within  [ 0 , 2 ] ,   and   r 1  and  r 2  show the coefficient vector which value was measured by  S ( t ) .
The exploitation team is responsible for improving the present solution. The BER evaluates every individual fitness value at all the cycles and differentiates the optimal individual. The BER applies two dissimilar strategies to accomplish exploitation, as follows.
The following equation is used for moving the searching agent towards the better solution:
S ( t + 1 ) = r 2 ( S ( t ) + D )
D = r 3 ( L ( t ) S ( t ) )
where  S ( t )  indicates the solution vector at  t  iteration,  r 3  Denotes a random vector that controls the movement steps near the better solution,  D  denotes the distance vector, and  L  shows the better solution vector.
The most potential is the area surrounding the lead (better solution). Consequently, some individuals hunt in the surrounding area for a better solution with the possibility of finding the best solution. The BER exploits the subsequent formula to realize these operations.
S ( t + 1 ) = r ( S ( t ) + k )
k = z + 2 × t 2 N 2
From the expression,  S  represents the better solution,  z  denotes the random value within  [ 0 , 1 ] ,   t  indicates the iteration value, and  N  indicates the overall number of iterations.
A mutation is an alternative approach applied by the BER. It can be a genetic operator used for creating and sustaining population diversity. It avoids an earlier convergence by assisting in avoiding the local optima, such as a modification of the searching space as a springboard to other interesting topics. The mutation is crucial for the remarkable exploration of the BER.
S ( t + 1 ) = k z 2 h cos ( x ) 1 cos ( x )
The BER selects the better performance for the following iteration to guarantee the quality of the defined solution. Even though the elitism technique improves the efficacy, it causes an early convergence when using a multi-modal function. Note that the BER gives impressive exploration abilities by applying a mutation method and searching nearby individuals to the exploration group. Due to its strong exploration abilities, the BER could prevent early convergence.

3.4. Detection Using Optimal DCRNN Model

At this stage, the features are passed to the DCRNN model for classification. Recently, a CNN was investigated and proved efficient in high-dimensional and large-scale data learning [27]. Additionally, a RNN is robust in long-term dependency capturing and temporal sequence learning. Here, the CNNs and RNNs are combined to implement feature learning on a MFCC-based representation depending on heart sounds that exploit the long-term dependency captured by the RNN and the encoded local features extracted from the CNN. The learnable kernel size in all the layers is fixed to 3 × 3, and the renowned ReLU function is exploited in all the convolution layers. A max-pooling is employed, followed by each convolutional layer, where 2 × 2 windows are exploited, and the stride is 2 × 2.
The BN layer standardizes the mini-batch via the whole network, which reduces the internal covariate shift caused by progressive transform, and the dropout layer might prevent overfitting and decrease the amount of neurons. Therefore, for all the input samples, a feature map is attained. After the max-pooling and convolutional layers, an LSTM layer is exploited for learning the temporal characteristics amongst the attained mapping features, and an FC layer with sixty-four neurons is effectuated to learn the global features. Eventually, a softmax layer is executed for deriving the probability distribution through two classes respective to normal and abnormal heart sound signals. Figure 2 depicts the infrastructure of the CRNN.
To improve the detection rate, the COA is used for hyperparameter tuning. The updating method of a candidate solution (coatis place) from the COA dependent upon modeling 2 basic performances of coatis are given below [28]:
(i)
Coatis’ escape strategy from the predator.
(ii)
Coatis’ strategy while attacking iguanas.
Consequently, the COA population was upgraded 2 stages.
The primary stage of updating the coatis’ population is modeled by simulating the strategy while attacking the iguanas. Here, a set of coatis climb trees to attain an iguana and scare it. This approach leads the coatis to move toward dissimilar locations in the searching space, which illustrates their exploration capability in a global searching space. Here, the location of the fittest member of the population is considered the location of the iguana. Thus, the rise in the coatis’ location in the tree can be a mathematical formula:
X i P 1 : x i , j P 1 = x i , j + r ( I g u a n a j I x i , j ) ,   f o r   i = 1 ,   2 ,   ,   N 2   and   j = 1 ,   2 ,   ,   m .
Once the iguana falls to the ground, it can be positioned at an arbitrary location from the searching space. As per the arbitrary location, the coatis on the ground moves from the searching space, which can be given as follows:
I g u a n a G : I g u a n a j G = l b j + r · G ( u b j . . I b j ) ,   j = 1 ,   2 ,   ,   m ,
X i P 1 I : x i , j P 1 = { x i , j + r ( I g u o n a j G I · x i , j ) ,   F I g u a n a G < F i , x i , j + r ( x i , j I g u o n o j G ) , e l s e
f o r   i = N 2 + 1 , N 2 + 2 , , N   a n d   j = 1 ,   2 ,   ,   m
The new location evaluated for all the coatis is acceptable for the updating procedure once it enhances the value of the main function or the coatis remain in the prior location.
X i = { X i P 1 , F i P 1 < F i , X i , e l s e .
Now,  r  represents a random real number within  [ 0 , 1 ] X i P 1  denotes the new location evaluated for the  i th  coatis,  x i , j P 1  indicates the  j th  dimension,  F i P 1  shows the objective function value,  I g u a n a  shows the iguana’s location in the search space that represents the location of the fittest member,  I g u ana j  refers to the  j th  dimension,  I g u a n a G  shows the location of the iguana on the ground that is produced at random,  I   I g u a n a j G  stands for the  j th  dimension,  ·  shows the floor function, and  F I g u a n a G  denotes the value of the main function.
The second stage is to update the location of the coatis from the searching space using a mathematical process according to the natural behaviors of the coatis while encountering predators and escaping from them. Once the predator attacks a coati, the animal escapes during the locating. The coatis’ moves during these strategies result in a safer location closer towards the existing location that represents the exploitation capability during a local search.
To simulate these behaviors, a random location is produced nearby the location where every coati is located as follows:
l b j l o c a l = l b j t , u b j l o c a l = u b j t , w h e r e   t = 1 ,   2 ,   ,   T .
X i P 2   :   x i , j P 2 = x i , j + ( 1 2 r ) ( l b j l o c a l + r ( u b j l o c a l l b j l o c a l ) ) ,
i = 1 ,   2 ,   ,   N ,   j = 1 ,   2 ,   ,   m ,
A recently estimated location is suitable once it enhances the value of the main function as follows:
X i = { X i P 2 , F i P 2 < F i , X i , e l s e ,
In Equation (21),  X i P 2  denotes the newest location evaluated for the  i th  coatis,  x i , j P 2  shows the  j th  dimension,  F i P 2  indicates the objective function values,  r  refers to the randomly generated values within  [ 0 , 1 ] ,   t  shows the iteration counter,  l b l o c a l  and  u b l o c a l  denote the local lower, as well as upper, boundaries of the  j th  search space correspondingly, and  l b j  and  u b j  indicate the lower boundaries, as well as upper boundaries, of the  j th  search space, respectively.
The COA iteration is completed after the location of the coatis when the decision variable is updated. The updating process is based on Equations (15)–(21) and reiterated until it attains the maximum iteration.
The COA method uses a fitness function (FF) to obtain a superior efficiency of the classifier. It describes a positive integer to suggest the finest outcome of the candidate solutions. The decline of the classifier rate of errors is assumed in the FF.
f i t n e s s ( x i ) = C l a s s i f i e r E r r o r R a t e ( x i ) = n o .   o f   m i s c l a s s i f i e d   i n s t a n c e s   T o t a l   n o .   o f   i n s t a n c e s 100

4. Performance Validation

In this section, the cancer detection outcomes of the BERTL-HIALCC technique are validated using the LC25000 database [29]. The database contains 25,000 instances with five classes, as provided in Table 1. Figure 3 demonstrates the sample images.
The confusion matrices of the BERTL-HIALCC method in the LCC detection results are depicted in Figure 4. The results identified that the BERTL-HIALCC technique recognized lung and colon cancers effectually.
In Table 2, the LCC recognition outcomes of the BERTL-HIALCC method under 80:20 of TRP/TSP are reported. In Figure 5, the recognition outcomes of the BERTL-HIALCC method are investigated under 80% of TRP. The figure indicates that the BERTL-HIALCC system identified the five classes proficiently. In the Col-Ad class, the BERTL-HIALCC technique provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99.19%, 98.98%, 96.92%, 97.94%, and 98.34%, respectively. Likewise, in the Col-Be class, the BERTL-HIALCC technique provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99.27%, 97.85%, 98.51%, 98.18%, and 98.98%, respectively. Similarly, in the Lun-SC class, the BERTL-HIALCC method provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 98.99%, 97.05%, 97.93%, 97.49%, and 98.59%, correspondingly.
In Figure 6, the recognition results of the BERTL-HIALCC technique are investigated under 20% of TRP. The figure revealed that the BERTL-HIALCC technique identified the five classes proficiently. In the Col-Ad class, the BERTL-HIALCC approach provided  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99.28%, 99.08%, 97.30%, 98.18%, and 98.54%, correspondingly. Similarly, in the Col-Be class, the BERTL-HIALCC method provided  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99.38%, 98.27%, 98.57%, 98.42%, and 99.07%, correspondingly. Similarly, in the Lun-SC class, the BERTL-HIALCC algorithm provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99%, 96.65%, 98.40%, 97.52%, and 98.77%, respectively.
In Table 3, the LCC recognition results of the BERTL-HIALCC method under 80:20 of TRP/TSP are reported. In Figure 7, the recognition outcomes of the BERTL-HIALCC method are investigated under 80% of TRP. The results point out that the BERTL-HIALCC system identified the five classes proficiently. In the Col-Ad class, the BERTL-HIALCC technique provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99.42%, 98.98%, 98.25%, 98.57%, and 98.98%, respectively. Likewise, in the Col-Be class, the BERTL-HIALCC approach offers  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 98.77%, 95.66%, 98.35%, 96.99%, and 98.61%, correspondingly. Additionally, in the Lun-SC class, the BERTL-HIALCC technique provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 98.67%, 96.33%, 97.05%, 96.69%, and 98.06%, correspondingly.
In Figure 8, the recognition results of the BERTL-HIALCC technique are inspected under 20% of TRP. The results demonstrate that the BERTL-HIALCC technique identified the five classes proficiently. In the Col-Ad class, the BERTL-HIALCC technique provides  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 99.40%, 98.56%, 98.36%, 98.46%, and 99.01%, respectively. Additionally, in the Col-Be class, the BERTL-HIALCC method offers  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 98.97%, 96.61%, 98.24%, 97.42%, and 98.70%, respectively. Likewise, in the Lun-SC class, the BERTL-HIALCC approach presents  a c c u y p r e c n r e c a l F s c o r e , and  A U C s c o r e  of 98.89%, 97.22%, 97.29%, 97.25%, and 98.29%, respectively.
Figure 9 inspects the accuracy of the BERTL-HIALCC method in the training and validation on 80:20 of TRP/TSP. The figure specifies that the BERTL-HIALCC approach reaches greater accuracy values with higher epochs. In addition, the greater validation accuracy over training accuracy shows that the BERTL-HIALCC method learns productively on 80:20 of TRP/TSP.
The loss analysis of the BERTL-HIALCC algorithm at the time of training and validation is validated on 80:20 of TRP/TSP in Figure 10. The results point out that the BERTL-HIALCC technique reaches similar values for the training and validation loss. The BERTL-HIALCC method learns productively on 80:20 of TRP/TSP.
A detailed precision–recall (PR) curve of the BERTL-HIALCC method is shown on 80:20 of TRP/TSP in Figure 11. The results show that the BERTL-HIALCC method results in increasing values of PR. In addition, the BERTL-HIALCC approach reached higher PR values in all the classes.
In Figure 12, a ROC study of the BERTL-HIALCC technique is revealed on 80:20 of TRP/TSP. The figure highlights that the BERTL-HIALCC method results in improved ROC values. In addition, the BERTL-HIALCC algorithm extends enhanced ROC values in all the classes.
To illustrate the improved cancer recognition results of the BERTL-HIALCC technique, a brief comparison study is carried out in Table 4 [30]. The results point out that the BERTL-HIALCC technique has improved results. Based on  a c c u y , the BERTL-HIALCC technique gains increasing  a c c u y  of 99.22%, while the MPADL-LC3, mSRC, Faster R-CNN, DAELGNN, ResNet50, CNN, and DL models accomplish decreasing  a c c u y  of 99.09%, 88.21%, 98.79%, 98.73%, 93.64%, and 97.11%, respectively.
Additionally, based on  p r e c n , the BERTL-HIALCC technique has an increasing  p r e c n  of 98.07%, while the MPADL-LC3, mSRC, Faster R-CNN, DAELGNN, ResNet50, CNN, and DL algorithms accomplish decreasing  p r e c n  of 98.01%, 85.21%, 96.53%, 97.95%, 96.12%, and 97.07%, respectively. Last, based on  r e c a l , the BERTL-HIALCC method has an increasing  r e c a l  of 98.06%, while the MPADL-LC3, mSRC, Faster R-CNN, DAELGNN, ResNet50, CNN, and DL approaches accomplish decreasing  r e c a l  of 97.20%, 91.78%, 97.78%, 97.63%, 96.39%, and 96.44%, correspondingly. In addition, the computation time (CT) analysis reported that the BERT-HIALCC technique results in the minimal CT value compared to the other models. Therefore, the proposed model can be employed for accurate LCC detection and classification.

5. Conclusions

In this study, we developed a novel LCC detection model named the BERTL-HIALCCD technique. This study aimed to identify LCC effectually on HSI. To achieve this, the BERTL-HIALCCD method followed a series of subprocesses, namely improved ShuffleNet feature extraction, BER-based parameter optimization, DCRNN-based classification, and COA-based hyperparameter tuning. Finally, a COA-based parameter selection process was carried out for the DCRNN model. To examine the performance of the BERTL-HIALCCD technique, a detailed set of experiments was conducted on a large dataset of HSI. The experimental results demonstrated that the combination of AER and COA algorithms attained an improved performance for cancer detection compared to the other models. In the future, the performance of the BERTL-HIALCDD method can be improved with multimodal fusion approaches. In addition, a histopathological image analysis can be integrated with clinical data, such as patient demographics, medical history, and genetic information, which can offer a better holistic understanding of cancer detection. It can significantly enhance the accuracy and reliability of the detection system and enable personalized medicine approaches. In addition, methods such as attention mechanisms, saliency maps, or feature visualization can assist in the identification of the important regions in histopathological images and provide insights into the decision-making process.

Author Contributions

Conceptualization, R.A. and M.R.; Methodology, M.R., F.Y.A. and R.A.M.; Software, T.O.A. and F.Y.A.; Validation, T.O.A., F.Y.A. and R.A.M.; Formal analysis, M.R., T.O.A. and F.Y.A.; Investigation, R.A., M.R. and R.A.M.; Data curation, T.O.A. and R.A.M.; Writing—original draft, R.A., M.R. and R.A.M.; Writing—review & editing, T.O.A. and F.Y.A.; Supervision, R.A.; Project administration, M.R.; Funding acquisition, R.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by Institutional Fund Projects under grant no. (IFPIP: 193-611-1443). Therefore, the authors gratefully acknowledge technical and financial support provided by the Ministry of Education and Deanship of Scientific Research (DSR), King Abdulaziz University (KAU), Jeddah, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tummala, S.; Kadry, S.; Nadeem, A.; Rauf, H.T.; Gul, N. An Explainable Classification Method Based on Complex Scaling in Histopathology Images for Lung and Colon Cancer. Diagnostics 2023, 13, 1594. [Google Scholar] [CrossRef] [PubMed]
  2. Naga Raju, M.S.; Srinivasa Rao, B. Lung and colon cancer classification using hybrid principle component analysis network-extreme learning machine. Concurr. Comput. Pract. Exp. 2023, 35, e7361. [Google Scholar] [CrossRef]
  3. Mansouri, R.A.; Ragab, M. Equilibrium Optimization Algorithm with Ensemble Learning Based Cervical Precancerous Lesion Classification Model. Healthcare 2023, 11, 55. [Google Scholar] [CrossRef] [PubMed]
  4. Graham, S.; Chen, H.; Gamper, J.; Dou, Q.; Heng, P.A.; Snead, D.; Tsang, Y.W.; Rajpoot, N. MILD-Net: Minimal information loss dilated network for gland instance segmentation in colon histology images. Med. Image Anal. 2019, 52, 199–211. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Ragab, M.; Abdushkour, H.A.; Nahhas, A.F.; Aljedaibi, W.H. Deer hunting optimization with deep learning model for lung cancer classification. CMC Comput. Mater. Continua. 2022, 73, 533–546. [Google Scholar] [CrossRef]
  6. Adu, K.; Yu, Y.; Cai, J.; Owusu-Agyemang, K.; Twumasi, B.A.; Wang, X. DHS-CapsNet: Dual horizontal squash capsule networks for lung and colon cancer classification from whole slide histopathological images. Int. J. Imaging Syst. Technol. 2021, 31, 2075–2092. [Google Scholar] [CrossRef]
  7. Kumar, A.; Vishwakarma, A.; Bajaj, V. Crccn-net: Automated framework for classification of colorectal tissue using histopathological images. Biomed. Signal Process. Control 2023, 79, 104172. [Google Scholar] [CrossRef]
  8. Yu, Y.; Li, J.; Li, J.; Xia, Y.; Ding, Z.; Samali, B. Automated damage diagnosis of concrete jack arch beam using optimized deep stacked autoencoders and multi-sensor fusion. Dev. Built Environ. 2023, 14, 100128. [Google Scholar] [CrossRef]
  9. Yu, Y.; Liang, S.; Samali, B.; Nguyen, T.N.; Zhai, C.; Li, J.; Xie, X. Torsional capacity evaluation of RC beams using an improved bird swarm algorithm optimised 2D convolutional neural network. Eng. Struct. 2022, 273, 115066. [Google Scholar] [CrossRef]
  10. Abdullah, S.; Ragab, M. Tunicate swarm algorithm with deep convolutional neural network-driven colorectal cancer classification from histopathological imaging data. Electron. Res. Arch. 2023, 31, 2793–2812. [Google Scholar]
  11. Toğaçar, M. Disease type detection in lung and colon cancer images using the complement approach of inefficient sets. Comput. Biol. Med. 2021, 137, 104827. [Google Scholar] [CrossRef]
  12. Chen, M.; Huang, S.; Huang, Z.; Zhang, Z. Detection of lung cancer from pathological images using cnn model. In Proceedings of the 2021 IEEE International Conference on Computer Science, Electronic Information Engineering and Intelligent Control Technology (CEI), Fuzhou, China, 24–26 September 2021; IEEE: New York City, NY, USA, 2021; pp. 352–358. [Google Scholar]
  13. Kumar, N.; Sharma, M.; Singh, V.P.; Madan, C.; Mehandia, S. An empirical study of handcrafted and dense feature extraction techniques for lung and colon cancer classification from histopathological images. Biomed. Signal Process. Control 2022, 75, 103596. [Google Scholar] [CrossRef]
  14. Ali, M.; Ali, R. Multi-input dual-stream capsule network for improved lung and colon cancer classification. Diagnostics 2021, 11, 1485. [Google Scholar] [CrossRef] [PubMed]
  15. Talukder, M.A.; Islam, M.M.; Uddin, M.A.; Akhter, A.; Hasan, K.F.; Moni, M.A. Machine learning-based lung and colon cancer detection using deep feature extraction and ensemble learning. Expert Syst. Appl. 2022, 205, 117695. [Google Scholar] [CrossRef]
  16. Hamida, A.B.; Devanne, M.; Weber, J.; Truntzer, C.; Derangère, V.; Ghiringhelli, F.; Forestier, G.; Wemmert, C. Deep learning for colon cancer histopathological images analysis. Comput. Biol. Med. 2021, 136, 104730. [Google Scholar] [CrossRef] [PubMed]
  17. Mangal, S.; Chaurasia, A.; Khajanchi, A. Convolution neural networks for diagnosing colon and lung cancer histopathological images. arXiv 2020, arXiv:2009.03878. [Google Scholar]
  18. Ding, K.; Liu, Q.; Lee, E.; Zhou, M.; Lu, A.; Zhang, S. Feature-enhanced graph networks for genetic mutational prediction using histopathological images in colon cancer. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 294–304. [Google Scholar]
  19. Dabass, M.; Vashisth, S.; Vig, R. A convolution neural network with multi-level convolutional and attention learning for classification of cancer grades and tissue structures in colon histopathological images. Comput. Biol. Med. 2022, 147, 105680. [Google Scholar] [CrossRef]
  20. Kassani, S.H.; Kassani, P.H.; Wesolowski, M.J.; Schneider, K.A.; Deters, R. Deep transfer learning based model for colorectal cancer histopathology segmentation: A comparative study of deep pre-trained models. Int. J. Med. Inform. 2022, 159, 104669. [Google Scholar] [CrossRef]
  21. Bharati, S.; Mondal, M.R.H.; Podder, P. A Review on Explainable Artificial Intelligence for Healthcare: Why, How, and When? In IEEE Transactions on Artificial Intelligence; IEEE: New York City, NY, USA, 2023. [Google Scholar]
  22. Wulczyn, E.; Steiner, D.F.; Moran, M.; Plass, M.; Reihs, R.; Tan, F.; Flament-Auvigne, I.; Brown, T.; Regitnig, P.; Chen, P.H.C.; et al. Interpretable survival prediction for colorectal cancer using deep learning. NPJ Digit. Med. 2021, 4, 71. [Google Scholar] [CrossRef]
  23. Li, Z.; Huang, K.; Liu, L.; Zhang, Z. Early detection of COPD based on graph convolutional network and small and weakly labeled data. Med. Biol. Eng. Comput. 2022, 60, 2321–2333. [Google Scholar] [CrossRef]
  24. Jain, D.K.; Lakshmi, K.M.; Varma, K.P.; Ramachandran, M.; Bharati, S. Lung Cancer Detection Based on Kernel PCA-Convolution Neural Network Feature Extraction and Classification by Fast Deep Belief Neural Network in Disease Management Using Multimedia Data Sources. Comput. Intell. Neurosci. 2022, 2022, 3149406. [Google Scholar] [CrossRef] [PubMed]
  25. Chen, Z.; Yang, J.; Feng, Z.; Chen, L. RSCNet: An Efficient Remote Sensing Scene Classification Model Based on Lightweight Convolution Neural Networks. Electronics 2022, 11, 3727. [Google Scholar] [CrossRef]
  26. El-kenawy, E.S.M.; Abdelhamid, A.A.; Ibrahim, A.; Mirjalili, S.; Khodadad, N.; Al Duailij, M.A.; Alhussan, A.A.; Khafaga, D.S. Al-Biruni Earth Radius (BER) metaheuristic search optimization algorithm. Comput. Syst. Sci. Eng. 2023, 45, 1917–1934. [Google Scholar] [CrossRef]
  27. Deng, M.; Meng, T.; Cao, J.; Wang, S.; Zhang, J.; Fan, H. Heart sound classification is based on improved MFCC features and convolutional recurrent neural networks. Neural Netw. 2020, 130, 22–32. [Google Scholar] [CrossRef]
  28. Dehghani, M.; Montazeri, Z.; Trojovská, E.; Trojovský, P. Coati Optimization Algorithm: A new bio-inspired metaheuristic algorithm for solving optimization problems. Knowl. Based Syst. 2023, 259, 110011. [Google Scholar] [CrossRef]
  29. Borkowski, A.A.; Bui, M.M.; Thomas, L.B.; Wilson, C.P.; DeLand, L.A.; Mastorides, S.M. Lung and Colon Cancer Histopathological Image Dataset (LC25000). Available online: https://www.kaggle.com/datasets/andrewmvd/lung-and-colon-cancer-histopathological-images (accessed on 13 June 2023).
  30. Mengash, H.A.; Alamgeer, M.; Maashi, M.; Othman, M.; Hamza, M.A.; Ibrahim, S.S.; Zamani, A.S.; Yaseen, I. Leveraging Marine Predators Algorithm with Deep Learning for Lung and Colon Cancer Diagnosis. Cancers 2023, 15, 1591. [Google Scholar] [CrossRef]
Figure 1. Overall flow of the BERTL-HIALCCD approach.
Figure 1. Overall flow of the BERTL-HIALCCD approach.
Cancers 15 03300 g001
Figure 2. Framework of the CRNN.
Figure 2. Framework of the CRNN.
Cancers 15 03300 g002
Figure 3. Sample images. (a) Lung cancer (b) Colon cancer.
Figure 3. Sample images. (a) Lung cancer (b) Colon cancer.
Cancers 15 03300 g003
Figure 4. Confusion matrices of the BERTL-HIALCC system: (a,b) 80:20 of TRP/TSP and (c,d) 70:30 of TRP/TSP.
Figure 4. Confusion matrices of the BERTL-HIALCC system: (a,b) 80:20 of TRP/TSP and (c,d) 70:30 of TRP/TSP.
Cancers 15 03300 g004
Figure 5. LCC recognition outcomes of the BERTL-HIALCC system on 70% of TRP.
Figure 5. LCC recognition outcomes of the BERTL-HIALCC system on 70% of TRP.
Cancers 15 03300 g005
Figure 6. LCC recognition outcomes of the BERTL-HIALCC system on 30% of TSP.
Figure 6. LCC recognition outcomes of the BERTL-HIALCC system on 30% of TSP.
Cancers 15 03300 g006
Figure 7. LCC recognition outcomes of the BERTL-HIALCC system on 80% of TRP.
Figure 7. LCC recognition outcomes of the BERTL-HIALCC system on 80% of TRP.
Cancers 15 03300 g007
Figure 8. LCC recognition outcomes of the BERTL-HIALCC system on 30% of TSP.
Figure 8. LCC recognition outcomes of the BERTL-HIALCC system on 30% of TSP.
Cancers 15 03300 g008
Figure 9. Accuracy curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Figure 9. Accuracy curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Cancers 15 03300 g009
Figure 10. Loss curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Figure 10. Loss curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Cancers 15 03300 g010
Figure 11. PR curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Figure 11. PR curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Cancers 15 03300 g011
Figure 12. ROC curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Figure 12. ROC curve of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Cancers 15 03300 g012
Table 1. Details of the database.
Table 1. Details of the database.
Class NameDescriptionNo. of Instances
Col-AdColon Adenocarcinoma5000
Col-BeColon Benign Tissue5000
Lun-AdLung Adenocarcinoma5000
Lun-BeLung Benign Tissue5000
Lun-SCLung Squamous Cell Carcinoma5000
Total No. of Samples25,000
Table 2. LCC recognition outcomes of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Table 2. LCC recognition outcomes of the BERTL-HIALCC system on 80:20 of TRP/TSP.
Class   A c c u y   P r e c n   R e c a l   F s c o r e   A U C s c o r e
Training Phase (80%)
Col-Ad99.1998.9896.9297.9498.34
Col-Be99.2797.8598.5198.1898.98
Lun-Ad99.0597.5497.6697.6098.52
Lun-Be99.2797.9898.3598.1798.92
Lun-SC98.9997.0597.9397.4998.59
Average99.1597.8897.8797.8798.67
Testing Phase (20%)
Col-Ad99.2899.0897.3098.1898.54
Col-Be99.3898.2798.5798.4299.07
Lun-Ad99.1497.9597.8597.9098.66
Lun-Be99.3298.3998.1998.2998.90
Lun-SC99.0096.6598.4097.5298.77
Average99.2298.0798.0698.0698.79
Table 3. LCC recognition outcomes of the BERTL-HIALCC system on 70:30 of TRP/TSP.
Table 3. LCC recognition outcomes of the BERTL-HIALCC system on 70:30 of TRP/TSP.
Class   A c c u y   P r e c n   R e c a l   F s c o r e   A U C s c o r e
Training Phase (70%)
Col-Ad99.4298.8998.2598.5798.98
Col-Be98.7795.6698.3596.9998.61
Lun-Ad98.9598.3896.3197.3497.96
Lun-Be98.9297.6896.8797.2898.15
Lun-SC98.6796.3397.0596.6998.06
Average98.9597.3997.3797.3798.35
Testing Phase (30%)
Col-Ad99.4098.5698.3698.4699.01
Col-Be98.9796.6198.2497.4298.70
Lun-Ad99.2898.6297.8498.2398.75
Lun-Be99.0898.0797.3697.7298.44
Lun-SC98.8997.2297.2997.2598.29
Average99.1397.8297.8297.8298.64
Table 4. Comparative outcomes of the BERTL-HIALCC approach with other methods [30].
Table 4. Comparative outcomes of the BERTL-HIALCC approach with other methods [30].
Methods   A c c u y  (%)   P r e c n  (%)   R e c a l  (%)   F s c o r e  (%) CT (s)
BERTL-HIALCC99.2298.0798.0698.0645
MPADL-LC3 Algorithm99.0998.0197.2097.0752
mSRC Algorithm88.2185.2191.7886.7864
Faster R-CNN Model98.7996.5397.6397.3258
DAELGNN Model98.7397.9596.3996.8055
RESNET-5093.6496.1297.4996.9460
CNN Model97.1197.0797.4497.6154
DL Algorithm96.3296.8696.4497.0857
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

AlGhamdi, R.; Asar, T.O.; Assiri, F.Y.; Mansouri, R.A.; Ragab, M. Al-Biruni Earth Radius Optimization with Transfer Learning Based Histopathological Image Analysis for Lung and Colon Cancer Detection. Cancers 2023, 15, 3300. https://doi.org/10.3390/cancers15133300

AMA Style

AlGhamdi R, Asar TO, Assiri FY, Mansouri RA, Ragab M. Al-Biruni Earth Radius Optimization with Transfer Learning Based Histopathological Image Analysis for Lung and Colon Cancer Detection. Cancers. 2023; 15(13):3300. https://doi.org/10.3390/cancers15133300

Chicago/Turabian Style

AlGhamdi, Rayed, Turky Omar Asar, Fatmah Y. Assiri, Rasha A. Mansouri, and Mahmoud Ragab. 2023. "Al-Biruni Earth Radius Optimization with Transfer Learning Based Histopathological Image Analysis for Lung and Colon Cancer Detection" Cancers 15, no. 13: 3300. https://doi.org/10.3390/cancers15133300

APA Style

AlGhamdi, R., Asar, T. O., Assiri, F. Y., Mansouri, R. A., & Ragab, M. (2023). Al-Biruni Earth Radius Optimization with Transfer Learning Based Histopathological Image Analysis for Lung and Colon Cancer Detection. Cancers, 15(13), 3300. https://doi.org/10.3390/cancers15133300

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop