Next Article in Journal
Sexual Functioning and Depressive Symptoms in Women with Postpartum Thyroiditis
Next Article in Special Issue
ViSwNeXtNet Deep Patch-Wise Ensemble of Vision Transformers and ConvNeXt for Robust Binary Histopathology Classification
Previous Article in Journal
The Clinical and Laboratory Profiles of a Deletional α2-Globin Gene Polyadenylation Signal Sequence (AATAAA > AATA--) [HBA2:c.*93_*94delAA]: The Malaysian Experience
Previous Article in Special Issue
Transformer and Attention-Based Architectures for Segmentation of Coronary Arterial Walls in Intravascular Ultrasound: A Narrative Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiclassification of Colorectal Polyps from Colonoscopy Images Using AI for Early Diagnosis

by
Jothiraj Selvaraj
1,
Kishwar Sadaf
2,*,
Shabnam Mohamed Aslam
3 and
Snekhalatha Umapathy
1,*
1
Department of Biomedical Engineering, College of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu 603203, India
2
Department of Computer Science, College of Computer and Information Sciences, Majmaah University, Al Majmaah 11952, Saudi Arabia
3
Department of Information Technology, College of Computer and Information Sciences, Majmaah University, Al Majmaah 11952, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(10), 1285; https://doi.org/10.3390/diagnostics15101285
Submission received: 18 April 2025 / Revised: 10 May 2025 / Accepted: 12 May 2025 / Published: 20 May 2025

Abstract

Background/Objectives: Colorectal cancer (CRC) remains one of the leading causes of cancer-related mortality worldwide, emphasizing the critical need for the accurate classification of precancerous polyps. This research presents an extensive analysis of the multiclassification framework leveraging various deep learning (DL) architectures for the automated classification of colorectal polyps from colonoscopy images. Methods: The proposed methodology integrates real-time data for training and utilizes a publicly available dataset for testing, ensuring generalizability. The real-time images were cautiously annotated and verified by a panel of experts, including post-graduate medical doctors and gastroenterology specialists. The DL models were designed to categorize the preprocessed colonoscopy images into four clinically significant classes: hyperplastic, serrated, adenoma, and normal. A suite of state-of-the-art models, including VGG16, VGG19, ResNet50, DenseNet121, EfficientNetV2, InceptionNetV3, Vision Transformer (ViT), and the custom-developed CRP-ViT, were trained and rigorously evaluated for this task. Results: Notably, the CRP-ViT model exhibited superior capability in capturing intricate features, achieving an impressive accuracy of 97.28% during training and 96.02% during validation with real-time images. Furthermore, the model demonstrated remarkable performance during testing on the public dataset, attaining an accuracy of 95.69%. To facilitate real-time interaction and clinical applicability, a user-friendly interface was developed using Gradio, allowing healthcare professionals to upload colonoscopy images and receive instant classification results. Conclusions: The CRP-ViT model effectively predicts and categorizes colonoscopy images into clinically relevant classes, aiding gastroenterologists in decision-making. This study highlights the potential of integrating AI-driven models into routine clinical practice to improve colorectal cancer screening outcomes and reduce diagnostic variability.

1. Introduction

The cells within the human body normally undergo a controlled cycle of proliferation and senescence, ensuring tissue homeostasis [1,2]. However, in the development of cancer, this process is disrupted, resulting in the uncontrolled proliferation of abnormal cells [3,4]. These aberrant cells can also invade surrounding tissues and can metastasize to distant organs, compromising normal physiological functions [5,6]. Colorectal cancer (CRC) is a malignant condition that originates in the epithelial tissues of the large intestine (colon) or the final segment of the digestive tract (rectum) [7,8,9], as shown in Figure 1. CRC often originates as a benign growth known as a colorectal polyp (CRP), which can develop into an invasive cancer [10]. While the majority of polyps are benign, certain types have the potential to undergo malignant transformation over an extended period [11]. The probability of a polyp progressing to cancer is determined by its histological classification, encompassing the structural and cellular attributes of the polyp [12,13].
Different types of CRPs include the following:
  • Hyperplastic and inflammatory polyps: These types are more frequently encountered and are generally considered non-precancerous. However, individuals with large hyperplastic polyps (>1 cm) may require more frequent CRC screening through colonoscopy, as a precautionary measure [14,15].
  • Serrated polyps: Sessile serrated polyps (SSPs) and traditional serrated adenomas (TSAs) are two common premalignant lesions that share an increased risk of progression to CRC. SSPs are characterized by their relatively flat or slightly elevated morphology and indistinct borders. Conversely, TSAs often exhibit a more pedunculated or sessile growth pattern and a more pronounced adenomatous behaviour, distinguishing them from SSPs [16,17].
  • Adenomatous polyps (adenomas): These polyps possess the potential to evolve into cancer and thus are regarded as precancerous. Adenomas are further classified into subtypes: tubular, villous, and tubulovillous [18]. Among these, tubular adenomas are the most prevalent, while villous adenomas, though less common, carry a higher risk of malignancy [19].
The key signs and symptoms associated with CRC/CRP [9,20] are persistent alterations in bowel habits and rectal bleeding as represented in Figure 2a. A notable symptom is tenesmus, where patients experience a continuous sensation of incomplete bowel evacuation, unrelieved by defecation. Other associated symptoms include lower abdominal cramps, unexplained fatigue, generalized weakness, and significant unintentional weight loss [21].
The key risk factors associated with CRC may be either changeable or un-changeable. As illustrated in Figure 2b, changeable factors, which can be altered through lifestyle interventions or environmental changes, include obesity, excessive alcohol consumption, type 2 diabetes, smoking, and specific dietary patterns. These factors are manageable and can potentially reduce the risk when addressed. On the other hand, the un-changeable risk factors, which cannot be altered, include age, race/ethnicity, sex, a history of cholecystectomy (gallbladder removal), and a personal or family history of CRC, CRP, or inflammatory bowel disease (IBD) [22,23].
CRC can lead to significant complications as the malignancy progresses, as it can invade adjacent tissue layers, potentially penetrating the muscularis propria and serosa. This invasive growth can facilitate the dissemination of cancer cells via the lymphatic and vascular systems, leading to the formation of metastases in regional lymph nodes and distant organs [24,25]. The stage of CRC, a critical prognostic factor, is determined by the extent of tumour invasion and the presence or absence of distant metastases, reflecting the local and systemic spread of the disease [26]. The progression of colorectal cancer (CRC) from normal colonic tissue to malignant cancer is illustrated in Figure 3a,b.
The statistics of CRC are represented as a bar chart in Figure 4. The CRC incidence (blue bar) and death rates (orange bar) are reported from 2018 to 2022 [27,28,29]. The trend lines highlight the gradual progression from 1.80 to 1.93 million in incidence and 0.86 to 0.90 million death rates over the years. According to statistics from GLOBOCAN, CRC ranks as the third most globally diagnosed cancer and the second leading cause of cancer-related mortality [7].
Figure 5 presents a schematic overview of the diagnostic and therapeutic (D&T) pathways for CRC. Screening procedures serve as an initial assessment to identify potential cases of CRC prior to the onset of overt symptoms [30]. Upon the detection of anomalies in screening tests, confirmatory diagnostic evaluations (biopsy) are conducted to establish a definitive diagnosis of CRC. Subsequent to diagnosis, surgical intervention is employed to excise the primary tumour [31]. Postoperative management may involve adjuvant therapies, including chemotherapy, immunotherapy, or other pharmacological interventions to manage the disease or to prevent recurrence [32].
The three primary techniques for CRC screening, including blood-, stool-, and image-based (visual) examinations [33], are presented in Figure 6. Blood-based assays analyze a patient’s blood for molecular or protein markers indicative of colorectal malignancies. Stool-based non-invasive assays evaluate fecal samples for the potential biomarkers of CRC, such as occult blood or abnormal DNA [34]. While these tests offer a less invasive screening method compared to other options, they generally require more frequent repetition to maintain efficacy [35]. Visual examinations involve the direct visualization of the colon and rectum to identify structural abnormalities, such as polyps or tumours [36]. This can be achieved either via endoscopy (e.g., colonoscopy, sigmoidoscopy, wireless capsule endoscopy (WCE)), equipped with a light source and video camera, or through radiographic techniques like computed tomography (CT) colonography [37].
A CT colonography, where the CT of the colon and rectum is performed, is less sensitive in detecting small polyps (<5 mm in diameter) [38]. Despite the relatively low radiation exposure from CT colonography, it remains a consideration, and since the procedure is non-therapeutic, the detected polyps cannot be removed, potentially necessitating a follow-up colonoscopy [39]. A colonoscopy examination involves a comprehensive visual assessment of the colon and rectum. A colonoscope, a slender, flexible tube equipped with a light source and a miniature camera, is inserted through the anus and gently advanced into the colon. This procedure allows for the direct visualization of the entire colonic lumen, facilitating the identification and evaluation of potential pathological lesions. If necessary, specialized instruments can be passed through the colonoscope to obtain tissue biopsies or perform an endoscopic resection of suspicious abnormalities, such as colonic polyps [40]. The normal and CRC are visually depicted in Figure 7. Sigmoidoscopy, similar to colonoscopy, facilitates examining only the lower third of the colon and the entire rectum, which may increase the likelihood of missing lesions located in the proximal colon [41]. In WCE, a camera-equipped capsule is ingested by the patient, enabling the transmission of images as it traverses the GI tract [42]. The capsule’s battery life is often insufficient to fully capture images of the colon and rectum, especially in patients with slow GI transit time.
In a clinical setting, for visual examination, colonoscopy is preferred rather than other imaging techniques due to its advantages compared to other procedures [43,44]. The visual examination of the colonoscopy images by the physician is tedious and time-consuming [45]. Due to the subjective nature of visual interpretation, clinicians may inadvertently overlook polyps or lesions, particularly those of diminutive size or flat morphology [46].
Traditional colonoscopy, while effective, can be limited by human factors such as fatigue or variability in interpretation. AI algorithms can automatically analyze colonoscopy images with high precision, identifying subtle patterns and abnormalities that might be missed by human eyes [47,48]. This technological integration aids in early detection, improves diagnostic accuracy, and reduces the risk of missed lesions, ultimately leading to better patient outcomes [49]. Machine learning (ML) algorithms, employing manual feature engineering, have exhibited significant proficiency in the detection of CRPs within the colonoscopy images. Contemporary research efforts are increasingly focused on deep learning (DL) architectures, notably convolutional neural networks (CNNs), eliminating the need for manual feature extraction. The neural network’s intrinsic capacity to learn and extract relevant features directly from the input colonic images has demonstrated promising results in both segmenting and classifying the CRP [50].

1.1. Literature Review

Colorectal polyps can be classified based on their characteristics into several categories, including hyperplastic, adenomatous, and serrated polyps. The classification of these polyps is crucial for determining the risk of colorectal cancer, as certain types, particularly neoplastic polyps (adenomatous and serrated), have a higher potential for malignancy. In the endoscopic evaluation of colorectal polyps, classification systems such as Kudo [51], NICE [52], Paris [53], and Sano [54] have a significant role in guiding clinical decisions. The Kudo pit pattern classification focuses on the mucosal surface architecture, using magnifying chromoendoscopy to distinguish between non-neoplastic and neoplastic lesions based on pit patterns I to V, where advanced irregular patterns suggest invasive carcinoma. Complementing this, the narrow-band imaging international colorectal endoscopic (NICE) classification leverages image-enhanced endoscopy to categorize lesions into types 1, 2, and 3 based on colour, vascular pattern, and surface structure. The Paris classification, on the other hand, provides a morphological framework to describe superficial neoplastic lesions, categorizing them into protruding, non-protruding, and excavated types, which is crucial for assessing the risk of submucosal invasion. Meanwhile, the Sano capillary pattern classification emphasizes microvascular architecture under magnification with narrow-band imaging, particularly highlighting irregular or sparse vascular patterns that are indicative of deeper invasion. Integrating these complementary classifications enhances the accuracy of endoscopic assessment, enabling detection and risk stratification for colorectal neoplasia.
This literature review synthesizes findings from multiple studies investigating the effectiveness of AI techniques in enhancing the detection and classification of colorectal polyps. Notably, Komeda et al. [55]. developed a CNN-based computer-aided diagnosis (CAD) system, presenting promising preliminary results that underscore the potential of AI in colon polyp detection. Their pilot study involved a retrospective analysis of colonoscopy video frames to identify polyps that might otherwise be missed by human observers. The researchers specifically focused on classifying colon images into adenomatous and non-adenomatous polyps, providing an early indication of AI’s capability to support diagnostic decision-making in clinical practice.
A systematic review by Sanchez-Peralta et al. [56] analyzed various AI methods for polyp detection, highlighting the advantages and disadvantages of different approaches. The authors insisted on the necessity for a common validation framework that includes a large, annotated database, which is crucial for standardizing results and facilitating comparisons across studies. Many studies have shown that the size and histological type of polyps significantly influence cancer risk, with larger and more complex polyps being associated with a higher likelihood of malignancy [12,57]. Itoh et al. [58] developed an automated binary classification based on the size of the polyp. Saad et al. [59] advanced this approach by implementing a multiclass classification system using the PICCOLO dataset, categorizing polyps into six distinct classes according to the Paris classification. Grosu et al. [60] proposed an ML-based classification of benign and pre-malignant polyps. Sharma et al. [61] developed a two-stage binary classification model for identifying cancerous polyps using an ensemble-based CNN model. Barua et al. [62] identified neoplastic polyps through binary classification methods using CAD.
In contrast to the more common binary classification of polyp presence, the literature contains limited studies on multiclassification. The application of AI for the multiclass classification of CRPs from colonoscopy images has gained substantial attention in recent years, primarily due to its potential to facilitate early diagnosis and improve patient outcomes. Bora et al. [63] proposed a novel approach to quantify shape, texture, and colour features for detecting the stages of dysplasia in polyps using a fuzzy entropy-based feature selection method, achieving an accuracy of 95.24% in the generated dataset and 95.72% in the public dataset. Krenzer et al. [64] focused on the development of two automated classification systems for polyps in the field of gastroenterology: The first system, based on the Paris classification, which categorizes polyps based on their shape, and the second system, based on the NICE classification, which categorizes polyps based on their texture and surface patterns [65].
The study by Wang et al. [66] utilized a deep learning approach for the automatic detection of colonic polyps using CNNs with global average pooling, achieving a high classification accuracy of 98% and reducing model parameters. Carvalho et al. [67] proposed a DL model to classify polyp features according to the NICE classification, achieving an accuracy of over 92% on internal datasets and exceeding 88% on a public dataset, demonstrating its potential to enhance diagnostic decision-making in CRC diagnosis. The present work distinguishes itself from the existing literature by advancing beyond conventional binary and limited multiclass classification systems that describe the severity of CRPs.
A customized four-class classification model that includes hyperplastic, serrated, adenoma, and normal categories is proposed in this research. Our approach also integrates the AI model into an interactive, web-based Gradio interface, enabling real-time, user-friendly deployment for clinical practitioners.

1.2. Contributions

Aim and Objectives: The aim of this investigation is to develop an AI-dependent multiclassification model for the early diagnosis of CRPs from colonoscopy images, enhancing clinical accuracy and aiding in the timely detection and prevention of CRC. The objectives of this study encompass the collection and preprocessing of colonoscopy image datasets for polyp classification, the design and implementation of deep learning algorithms for the multiclass classification of various types of colorectal polyps, and the evaluation of the proposed AI model’s performance using standard metrics. Additionally, the research involves comparing the developed model with existing state-of-the-art techniques in polyp classification and validating the model’s effectiveness to ensure clinical applicability.
Based on the objectives, the author contributions are as follows:
  • Dataset collection and preprocessing: A comprehensive colonoscopy image dataset including real-time data is curated and preprocessed to enhance the quality of input data for effective CRP classification. Additionally, ethical clearance is obtained to ensure the responsible collection and use of real-time images in compliance with regulatory guidelines.
  • AI-based deep learning model development and performance evaluation: A customized DL-based multiclassification model is designed and implemented to accurately classify different types of colorectal polyps, aiding in the early diagnosis of CRC. The proposed model’s performance is assessed using standard evaluation metrics.
  • Comparative analysis: The developed AI model is compared with existing state-of-the-art methods for polyp classification to demonstrate its superiority and effectiveness in clinical diagnosis.
  • Interactive interface deployment using Gradio: To enhance usability and clinical translation, a Gradio-based interface is developed, enabling users to upload colonoscopy images and receive immediate visual feedback on the predicted type of polyp. This interactive feature aids clinicians in real-time decision-making during diagnostic procedures.
This research article is structured to provide a comprehensive exploration of the study. It begins with an introduction, outlining the context and background, followed by an in-depth literature review and a clear statement of the contributions made by the study. The Materials and Methods (Section 2) details the proposed workflow, including data acquisition, preprocessing, and the implementation of deep learning classifiers, along with the integration of a real-time diagnostic interface using Gradio-5.29.0. The Results and Discussion (Section 3) presents and analyzes the findings, providing insights into the model’s performance and potential future directions. Finally, the article concludes (Section 4) with a summary of the research outcomes.

2. Materials and Methods

2.1. Proposed Workflow

Figure 8 illustrates the workflow of the proposed research methodology. The entire experimental framework was implemented using the Python-3.10 programming language. In this study, two datasets were utilized. The real-time dataset, following data augmentation, was employed for training and validation. To evaluate the generalizability of the architecture, non-augmented images from both the publicly available dataset and the real-time dataset were used exclusively for testing. As illustrated in Figure 8, initial experiments were conducted using six conventional CNN architectures, including VGG16 [68], VGG19 [68], ResNet50 [69], DenseNet121 [70], EfficientNetV2 [71], and InceptionNetV3 [72]. VGG16 and VGG19 are known for their uniform layer design and lower computation complexity. ResNet50 introduces residual connections that enable the training of deeper networks without the vanishing gradient problem, leading to improved accuracy and convergence. DenseNet121 enhances features through dense connectivity between layers, resulting in reduced parameters and strengthened gradient flow. InceptionNetV3 incorporates multi-scale processing within each module, enabling the efficient handling of visual information. EfficientNetV2 offers state-of-the-art performance by scaling depth, width, and resolution in a compound manner, achieving high accuracy with optimized computational efficiency. The inclusion of these diverse models allows for a robust comparative evaluation of their capabilities in detecting and classifying colorectal polyps from colonoscopy images.
Additionally, Transformer-based models, specifically the Vision Transformer (ViT) [73], were incorporated. Building upon the performance of these baseline models, a hybrid architecture combining ResNet50 and ViT was developed, herein referred to as CRP-ViT. Furthermore, the Gradio library [74] was utilized to develop the graphical user interface (GUI) for facilitating class prediction.

2.2. Data Acquisition and Data Collection

In this study, two distinct datasets (DSs) were utilized to conduct the investigations: a real-time dataset, denoted as DS-1, and a publicly available/open access (OA) (UAH) dataset [16], referred to as DS-2. DS-1 comprises data collected from the SRM Medical College Hospital and Research Centre, Kattankulathur, India. The research protocol for DS-1 was reviewed and approved by the Institutional Ethics Committee (IEC)—Human Studies of SRM Medical College Hospital and Research Centre under approval number 8677/IEC/2023. The study design adhered to ethical guidelines, ensuring that all procedures, including participant recruitment, informed consent, data collection, and analysis, were conducted, with strict adherence to ethical standards to protect participants’ rights and welfare. Each video within DS-1 is annotated with images obtained from histopathological analysis, complemented by an assessment performed by two expert evaluators and two novice observers referring to NICE and Sano classification. The annotation and labelling methodology, along with additional dataset details, are comprehensively illustrated in Figure 8.
DS-2 [16], on the other hand, was sourced from an open access, publicly available database. Similarly to DS-1, each video in DS-2 is accompanied by ground truth derived from histopathology and expert assessments. However, DS-2 features a more diverse panel of evaluators, consisting of four expert reviewers and three beginners, ensuring a balanced representation of varying levels of expertise. The multi-tiered annotation approach enhances the robustness of the dataset, facilitating a more comprehensive validation of the region of interest in the colonoscopy images.
Table 1 presents an overview of the datasets, including DS-1 and DS-2, utilized for CRP detection through colonoscopy image analysis. DS-1, collected from 71 participants, comprising 100 polyp frames and 100 normal frames, was captured at a high resolution of 1920 × 1080 pixels. DS-2 [16], consisting of 76 polyp frames from 76 participants, with no normal images available, has a resolution of 768 × 576 pixels. The distribution of images under each class of polyp is elaborated in Table 2. It is notable that the distribution is uneven across each class considered in this investigation.

2.3. Image Preprocessing and Data Split

In this study, three essential preprocessing steps, including resizing, data normalization, and data augmentation, were implemented to improve the performance of the DL models [7]. To ensure uniformity across datasets DS-1 and DS-2, all the frames were resized to 512 × 512. Additionally, following the resizing of images, data normalization was performed concurrently to enhance the consistency of the images. DS-1, excluding 40 normal images, was considered for training and validation. DS-2, along with 40 normal images excluded from DS-1, was utilized for testing the DL model. Data augmentation on DS-1 was conducted in two stages: Stage 1 and Stage 2. The primary objective of Stage 1 augmentation was to balance the distribution of polyp classes, while Stage 2 augmentation aimed to increase the dataset size for more effective training and validation.
In Stage 1 augmentation, a balanced dataset was achieved, with each class containing 60 images. As already 60 hyperplastic images were available, augmentation techniques were applied to balance the remaining polyp classes. The serrated class was augmented using one technique, while the adenoma class underwent five augmentation techniques. Among the 100 normal images of DS-1, 60 were considered without any augmentation. In the Stage 2 augmentation, all the images considered for training and validation, totalling 60 images per class across four classes, were augmented using 21 techniques. As a result, the dataset for training and validation following the Stage 2 augmentation was 5280 images, as presented in Table 3. The utilized augmentation techniques are clearly illustrated in Figure 8. Following Stage 2 augmentation, 80% of the data (4224) was used for training, and the rest, 20% of the frames (1056), were used for validation. To assess the generalizability of the proposed system, 116 images, including images from DS-1 (40 normal images) and DS-2 (76 images), were used as detailed in Table 4.

2.4. Deep Learning Classifier

The CRP-ViT architecture, illustrated in Figure 9, originally developed for the binary classification of polyp presence in the previous study [7], was used in the present work as well. However, instead of the sigmoid function, the final layer was replaced with a SoftMax activation to better align with the multiclassification task. The CRP-ViT integrates ResNet50 with ViT encoders to enhance the classification of colonoscopy images. Initially, high-resolution images are passed through a series of convolutional layers, beginning with a 7 × 7 convolution followed by 3 × 3 convolutions, to extract low-level spatial features. These are further processed through a ResNet50 feature extractor consisting of four sequential stages, each containing multiple stages of residual blocks with 1 × 1 and 3 × 3 convolutions that progressively increase the number of channels and capture rich hierarchical representations. The resulting feature maps are then divided into non-overlapping patches and linearly projected into a suitable embedding space to serve as input tokens for the ViT encoder. The ViT encoder, composed of multiple stacked layers of multi-head self-attention, layer normalization, and multi-layer perceptrons (MLPs), enables global contextual modelling across the entire image. The encoded features are then passed through a final MLP head for classification. This hybrid architecture effectively combines the local feature extraction capabilities of CNNs with the global modelling power of Transformers, making it well suited for the accurate detection and classification of CRPs.

2.5. Real-Time Diagnostic Interface Integration Using Gradio

The Gradio, a Python library [74], allows the creation of an interactive web-based interface without any complex skills. In this research, Gradio is utilized to build an intuitive interface through which clinicians can upload colonoscopy images and immediately view the predicted type of colorectal polyp generated by the DL model. This real-time interaction significantly enhances the usability of the system, supporting clinicians in making timely and informed decisions during diagnostic procedures. Moreover, Gradio facilitates the rapid prototyping and sharing of the DL model considered in this study by generating deployable web applications that can be accessed locally or shared via public links. By integrating Gradio, the proposed framework bridges the gap between advanced DL algorithms and practical clinical application, promoting accessibility, transparency, and ease of use in clinical settings.

3. Results and Discussion

The performance metrics of conventional DL models are summarized in Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10. The performance was assessed across four individual classes: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. The evaluation was conducted using three distinct optimizers, with each model trained for 50 epochs. The number of epochs in which the model converged varied depending on the model and the optimizer used. The learning rate of the classifier was set to 0.0001 based on insights from prior work on binary classification.
ResNet50, when optimized with ADAM, exhibited superior performance with the highest overall training accuracy of 89.2% and validation accuracy of 88.07%, indicating its capability to effectively capture complex features and handle multiclassification. VGG19, with ADAM, also demonstrated competitive performance, while VGG16 showed a moderate performance. The ADAM optimizer demonstrated superior convergence and stability across all architectures, while RMSprop optimization showed balanced performance, slightly inferior when compared to ADAM. Overall, the findings emphasize that the choice of optimizer plays a crucial role in enhancing the model’s performance. The ADAM has emerged as the most effective optimizer for improving classification accuracy and stability across different classes in this investigation.
DenseNet121, EfficientNetV2, and InceptionNetV3 demonstrated competitive classification performance, compared with ResNet50, particularly when trained using the ADAM optimizers. Among the evaluated architectures, ResNet50 and EfficientNetV2 yielded closely comparable performance metrics, exhibiting minimal variance across training and validation phases. However, ResNet50 consistently outperformed EfficientNetV2, achieving a training accuracy of 89.2% and a validation accuracy of 88.07%, in contrast to EfficientNetV2’s 88.66% and 87.69%, respectively. A comprehensive breakdown of model-specific performance metrics is detailed in Table 7 and Table 9. Based on this consistent performance advantage, the ResNet50 architecture was selected for further experimentation in subsequent phases of the study.
The Transformer-based model, ViT, was analyzed across the three optimizers and is presented in Table 11. The ViT optimized with ADAM performed well compared to the other optimizers considered in this investigation, with an overall accuracy of 92.38% and 95.36% during training and validation. The RMSprop achieved performance in close association with the ADAM, while the SGD showed the lowest performance, with slower convergence and a limited ability to handle complex data patterns.
The performance achieved by CRP-ViT is tabulated in Table 12. When evaluating the performance of CRP-ViT, the ADAM optimizer achieved the highest metrics, with a training accuracy of 97.28% and a validation accuracy of 96.02%. Across all scenarios, ADAM proved to be the most effective optimizer, owing to its adaptive learning strategy and superior convergence capabilities. Furthermore, when comparing the performance of all models, CRP-ViT exhibited exceptional performance, outperforming its counterparts.
In the comparative evaluation of deep learning architectures for colorectal image classification, the CRP-ViT model demonstrated superior performance across the metrics, outperforming traditional CNN and Transformer models, reflecting its robust feature extraction and class discrimination capabilities. The ViT model also exhibited commendable performance, especially with the ADAM optimizer, surpassing traditional models, thus highlighting the efficacy of Transformer-based architectures in capturing complex patterns within colorectal images. While ResNet50, with its deep residual connections, offered improved accuracy over the VGG architectures, it still lagged behind the ViT and CRP-ViT. VGG16 and VGG19, despite their historical significance and simplicity, showed relatively lower performance, likely due to their limited capacity in modelling intricate spatial relationships compared to ViT and CRP-ViT. Overall, the integration of convolutional mechanisms within the CRP-ViT model substantially enhanced classification efficacy, making it a promising approach for advancing automated colorectal cancer screening.
The results of the K-fold cross-validation, which assessed the fold-wise performance and overall average performance of the CRP-ViT model, are presented in Table 13. A five-fold cross-validation across four distinct classes (Class 0, Class 1, Class 2, and Class 3) was implemented. The performance of the five-fold cross-validation was in close concordance with the 80:20 split. Furthermore, the five-fold cross-validation results demonstrate minimal performance variance across folds, as evidenced by the low standard deviation values across the key metric, accuracy. This suggests that the CRP-ViT maintains stable predictive performance irrespective of data partitioning, and it reinforces the model’s potential applicability in clinical settings.
The difference between the performance of the 80:20 split and five-fold cross validation is provided in Table 14. The marginal difference observed highlights the model’s ability to generalize effectively across different data splits, demonstrating that the model is not biased toward any specific split. The ablation study conducted on CRP-ViT is provided in Table 15. The analysis of ablation studies on different stages of the CRP-ViT architecture reveals significant variations in classification accuracy. Among the individual stages, ablation on Stage 3 of ResNet50 yields an accuracy of 92.78%, and Stage 2 of ResNet50 ablation follows closely with an accuracy of 91.71%. The Stage 4 and Stage 1 of ResNet50 ablations result in accuracies of 90.74% and 90.18%, respectively. Notably, ablation across all stages from Stage 1 to Stage 4 of ResNet50 drastically reduces the accuracy to 67.06%, highlighting the collective importance of these stages for effective feature extraction and learning. Ablation on the fully connected network (FCN) impacts the performance as well, dropping the accuracy to 87.85%. In comparison, the baseline model without any ablation achieves the highest accuracy of 97.28%, emphasizing the significance of maintaining the integrity of the entire architecture for optimal performance.
Figure 10a,b represent the training and validation performance of the CRP-ViT model across the epochs, highlighting both accuracy and loss trends. In the accuracy plot, the model demonstrates a steady increase in accuracy over the epochs, with both training and validation accuracy converging at the epoch of 23, indicating effective learning. In the loss plot, there is a clear downward trend in both training and validation loss values over time. The training loss shows a smooth and continuous decline, reflecting effective optimization. However, the validation loss, while decreasing, exhibits occasional spikes and stabilizes toward the latter epochs. The ROC curves for both the training and validation phases across the four distinct classes considered in this investigation, highlighting the model’s strong classification capability with consistently high AUC values, are presented in Figure 10c. Overall, the close alignment of the ROC curves between the training and validation sets, coupled with the high AUC scores across all classes, confirms the performance of CRP-ViT in multiclass classification, with minimal overfitting and strong predictive accuracy. The confusion matrix of 116 test images is provided in Figure 10d. Out of the 116 test images, the CRP-ViT has classified 111 images correctly, highlighting the model’s strong ability to distinguish between classes, with high true positive rates and minimal misclassifications.
An interactive web-based interface developed using Gradio for real-time colorectal polyp classification is illustrated in Figure 11a–e. The input panel enables healthcare professionals to upload colonoscopy images either by dragging and dropping or selecting files manually from their local system. Upon submission, the system processes the input image through the proposed CRP-ViT model and displays the classification outcomes in the prediction section. The output is systematically categorized into four clinically significant classes: hyperplastic, serrated, adenoma, and normal, with each category distinctly displayed in separate panels to facilitate clear interpretation. Additionally, an inference text box is integrated to present the final decision, indicating the most probable class predicted by the model, thereby enhancing diagnostic clarity. The interface incorporates essential functional controls, such as the Clear button to reset the input selection and the Submit button to initiate the prediction process.
Most of the previous studies have primarily focused on binary classification approaches. Although the UAH DB dataset has been employed in the existing literature, its application has been limited to binary classification tasks. To date, no published studies have demonstrated the use of this dataset for multiclass classification, and therefore, a direct comparison with prior work is not feasible. For a fair performance evaluation, results from different datasets and studies cannot be directly compared. Furthermore, to the best of our knowledge, no publicly available dataset currently supports multiclass classification of this nature. Consequently, we utilized a self-collected dataset for model training and validation, while a publicly available dataset was employed solely for testing purposes.
Due to the absence of multiclass classification studies on CRPs using the same dataset employed in this research, a fair comparison with the existing literature was not feasible. Therefore, the proposed CRP-ViT model was benchmarked against six conventional CNN-based state-of-the-art models within this study. This internal evaluation strategy ensures a consistent and reliable comparison under uniform experimental conditions.
Limitations and future work: In the current study, the dataset comprised 4224 images for training, 1056 images for validation, and 116 images for testing. While the model demonstrated promising performance, the limited size and scope of the dataset constrain its ability to generalize across broader clinical environments. Future work will prioritize the expansion of the dataset by incorporating larger, multi-centre, and heterogeneous image collections to enhance the robustness and external validity of the model. Additionally, the integration of quantum computing techniques will be explored to accelerate and improve the precision of polyp classification tasks. Furthermore, the current model, which operates on static colonoscopy images, will be extended to process continuous colonoscopy video streams to enable real-time polyp detection and classification, thereby enhancing its applicability in live clinical settings and supporting endoscopists during procedures.

4. Conclusions

This study presents a comprehensive DL framework for the multiclass classification of colorectal polyps, aiming to enhance the detection and diagnosis of colorectal cancer (CRC). By leveraging real-time clinical data alongside publicly available datasets, the proposed approach ensures both robustness and generalizability. The custom-developed CRP-ViT model demonstrated exceptional performance, surpassing established architectures, such as VGG16, VGG19, ResNet50, DenseNet121, EfficientNetV2, InceptionNetV3, and ViT, by effectively capturing intricate patterns within colonoscopy images. The integration of the CRP-ViT model with an accessible Gradio-based interface bridges the gap between complex AI solutions and frontline clinical practice, enabling timely, accurate assessments even in resource-constrained settings. By reducing diagnostic variability and supporting medical professionals in real-time decision-making, the proposed framework contributes to more equitable healthcare delivery. Ultimately, this research paves the way for the widespread adoption of AI-assisted screening tools, fostering earlier interventions, improving patient survival rates, and alleviating the broader public health burden associated with colorectal cancer.

Author Contributions

Conceptualization, J.S. and S.U.; Data Curation, J.S.; Formal Analysis, K.S., S.M.A. and S.U.; Funding Acquisition, K.S. and S.M.A.; Investigation, J.S., K.S. and S.U.; Methodology, J.S. and S.U.; Project Administration, K.S. and S.U.; Resources, J.S.; Software, J.S. and S.U.; Supervision, S.U.; Validation, S.M.A.; Visualization, J.S. and K.S.; Writing—Original Draft, J.S. and S.U.; Writing—Review and Editing, S.M.A. and S.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Ethics Committee of SRM Medical College Hospital and Research Centre, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu, Tamil Nadu, India-603203 (Ethical approval number: 8677/IEC/2023 and approval date: 19 July 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data that support the findings of this study are openly available at the following URL: http://www.depeca.uah.es/colonoscopy_dataset/; (accessed on 18 April 2025) and DOI: https://doi.org/10.1109/tmi.2016.2547947.

Acknowledgments

The authors would like to thank the Deanship of Scientific Research at Majmaah University for supporting this work under project number R-2025-1783. We gratefully acknowledge the use of Servier Medical Art’s open access image repository, which has been highly valuable to the development of figures for this project. We also extend our profound appreciation to the distinguished gastroenterologist for their expert input in annotation, which has substantially enriched the quality and accuracy of this work. Additionally, we are thankful to the participants whose involvement has supported the successful completion of this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CADComputer-Aided Diagnosis
CNNConvolutional Neural Network
CRCColorectal Cancer
CRPColorectal Polyp
CTComputed Tomography
D&TDiagnostic and Therapeutic
DADomain Adaptation
DAData Augmentation
DLDeep Learning
DNADeoxyribonucleic Acid
DSDataset
DS-1Dataset 1 (Real-Time (own))
DS-2Dataset 2 (UAH DB)
GUIGraphical User Interface
IBDInflammatory Bowel Disease
IECInstitutional Ethics Committee
MLMachine Learning
MLPMulti-Layer Perceptron
ROCReceiver Operating Characteristic
S1DAStage 1 Data Augmentation
S2DAStage 2 Data Augmentation
SSPSessile Serrated Polyps
TSATraditional Serrated Adenomas
ViTVision Transformer
WCEWireless Capsule Endoscopy

References

  1. Roger, L.; Tomas, F.; Gire, V. Mechanisms and regulation of cellular senescence. Int. J. Mol. Sci. 2021, 22, 13173. [Google Scholar] [CrossRef]
  2. Wan, M.; Gray-Gaillard, E.F.; Elisseeff, J.H. Cellular senescence in musculoskeletal homeostasis, diseases, and regeneration. Bone Res. 2021, 9, 41. [Google Scholar] [CrossRef]
  3. Kashyap, A.K.; Dubey, S.K. Molecular mechanisms in cancer development. In Understanding Cancer; Elsevier: Amsterdam, The Netherlands, 2022; pp. 79–90. [Google Scholar]
  4. Sehgal, P.; Chaturvedi, P. Chromatin and cancer: Implications of disrupted chromatin organization in tumorigenesis and its diversification. Cancers 2023, 15, 466. [Google Scholar] [CrossRef]
  5. Li, Y.; Liu, F.; Cai, Q.; Deng, L.; Ouyang, Q.; Zhang, X.H.-F.; Zheng, J. Invasion and metastasis in cancer: Molecular insights and therapeutic targets. Signal Transduct. Target. Ther. 2025, 10, 57. [Google Scholar] [CrossRef]
  6. Kciuk, M.; Gielecińska, A.; Budzinska, A.; Mojzych, M.; Kontek, R. Metastasis and MAPK pathways. Int. J. Mol. Sci. 2022, 23, 3847. [Google Scholar] [CrossRef]
  7. Selvaraj, J.; Umapathy, S.; Rajesh, N.A. Artificial intelligence based real time colorectal cancer screening study: Polyp segmentation and classification using multi-house database. Biomed. Signal Process. Control 2025, 99, 106928. [Google Scholar] [CrossRef]
  8. Selvaraj, J.; Umapathy, S. CRPU-NET: A deep learning model based semantic segmentation for the detection of colorectal polyp in lower gastrointestinal tract. Biomed. Phys. Eng. Express 2023, 10, 015018. [Google Scholar] [CrossRef]
  9. Sawicki, T.; Ruszkowska, M.; Danielewicz, A.; Niedźwiedzka, E.; Arłukowicz, T.; Przybyłowicz, K.E. A review of colorectal cancer in terms of epidemiology, risk factors, development, symptoms and diagnosis. Cancers 2021, 13, 2025. [Google Scholar] [CrossRef]
  10. Łukaszewicz-Zając, M.; Mroczko, B. Circulating biomarkers of colorectal cancer (CRC)—Their utility in diagnosis and prognosis. J. Clin. Med. 2021, 10, 2391. [Google Scholar] [CrossRef]
  11. Saraiva, S.; Rosa, I.; Fonseca, R.; Pereira, A.D. Colorectal malignant polyps: A modern approach. Ann. Gastroenterol. 2021, 35, 17. [Google Scholar] [CrossRef]
  12. Sullivan, B.A.; Noujaim, M.; Roper, J. Cause, epidemiology, and histology of polyps and pathways to colorectal cancer. Gastrointest. Endosc. Clin. 2022, 32, 177–194. [Google Scholar] [CrossRef]
  13. Wang, J.-D.; Xu, G.-S.; Hu, X.-L.; Li, W.-Q.; Yao, N.; Han, F.-Z.; Zhang, Y.; Qu, J. The histologic features, molecular features, detection and management of serrated polyps: A review. Front. Oncol. 2024, 14, 1356250. [Google Scholar] [CrossRef]
  14. Kővári, B.; Kim, B.H.; Lauwers, G.Y. The pathology of gastric and duodenal polyps: Current concepts. Histopathology 2021, 78, 106–124. [Google Scholar] [CrossRef]
  15. Triadafilopoulos, G. Prevalence of Abnormalities at Tandem Endoscopy in Patients Referred for Colorectal Cancer Screening/Surveillance Colonoscopy. Cancers 2024, 16, 3998. [Google Scholar] [CrossRef]
  16. Mezzapesa, M.; Losurdo, G.; Celiberto, F.; Rizzi, S.; d’Amati, A.; Piscitelli, D.; Ierardi, E.; Di Leo, A. Serrated colorectal lesions: An up-to-date review from histological pattern to molecular pathogenesis. Int. J. Mol. Sci. 2022, 23, 4461. [Google Scholar] [CrossRef]
  17. Nagtegaal, I.D.; Snover, D.C. Head to head: Should we adopt the term ‘sessile serrated lesion’? Histopathology 2022, 80, 1019–1025. [Google Scholar] [CrossRef]
  18. Kahraman, D.S.; Sayhan, S. Colon polyps and their pathologic characteristics. In Colon Polyps and Colorectal Cancer; Springer: Cham, Switzerland, 2021; pp. 167–211. [Google Scholar]
  19. Jungwirth, J.; Urbanova, M.; Boot, A.; Hosek, P.; Bendova, P.; Siskova, A.; Svec, J.; Kment, M.; Tumova, D.; Summerova, S. Mutational analysis of driver genes defines the colorectal adenoma: In situ carcinoma transition. Sci. Rep. 2022, 12, 2570. [Google Scholar] [CrossRef]
  20. Holtedahl, K.; Borgquist, L.; Donker, G.A.; Buntinx, F.; Weller, D.; Campbell, C.; Månsson, J.; Hammersley, V.; Braaten, T.; Parajuli, R. Symptoms and signs of colorectal cancer, with differences between proximal and distal colon cancer: A prospective cohort study of diagnostic accuracy in primary care. BMC Fam. Pract. 2021, 22, 148. [Google Scholar] [CrossRef]
  21. Fritz, C.D.; Otegbeye, E.E.; Zong, X.; Demb, J.; Nickel, K.B.; Olsen, M.A.; Mutch, M.; Davidson, N.O.; Gupta, S.; Cao, Y. Red-flag signs and symptoms for earlier diagnosis of early-onset colorectal cancer. JNCI J. Natl. Cancer Inst. 2023, 115, 909–916. [Google Scholar] [CrossRef]
  22. Lewandowska, A.; Rudzki, G.; Lewandowski, T.; Stryjkowska-Góra, A.; Rudzki, S. Risk factors for the diagnosis of colorectal cancer. Cancer Control 2022, 29, 10732748211056692. [Google Scholar] [CrossRef]
  23. Sninsky, J.A.; Shore, B.M.; Lupu, G.V.; Crockett, S.D. Risk factors for colorectal polyps and cancer. Gastrointest. Endosc. Clin. N. Am. 2022, 32, 195–213. [Google Scholar] [CrossRef]
  24. Leong, S.P.; Naxerova, K.; Keller, L.; Pantel, K.; Witte, M. Molecular mechanisms of cancer metastasis via the lymphatic versus the blood vessels. Clin. Exp. Metastasis 2022, 39, 159–179. [Google Scholar] [CrossRef]
  25. Zhou, H.; Lei, P.-j.; Padera, T.P. Progression of metastasis through lymphatic system. Cells 2021, 10, 627. [Google Scholar] [CrossRef]
  26. Chen, K.; Collins, G.; Wang, H.; Toh, J.W.T. Pathological features and prognostication in colorectal cancer. Curr. Oncol. 2021, 28, 5356–5383. [Google Scholar] [CrossRef]
  27. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef]
  28. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef]
  29. Bray, F.; Laversanne, M.; Sung, H.; Ferlay, J.; Siegel, R.L.; Soerjomataram, I.; Jemal, A. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2024, 74, 229–263. [Google Scholar] [CrossRef]
  30. Burnett-Hartman, A.N.; Lee, J.K.; Demb, J.; Gupta, S. An update on the epidemiology, molecular characterization, diagnosis, and screening strategies for early-onset colorectal cancer. Gastroenterology 2021, 160, 1041–1049. [Google Scholar] [CrossRef]
  31. Naito, Y.; Aburatani, H.; Amano, T.; Baba, E.; Furukawa, T.; Hayashida, T.; Hiyama, E.; Ikeda, S.; Kanai, M.; Kato, M. Clinical practice guidance for next-generation sequencing in cancer diagnosis and treatment (edition 2.1). Int. J. Clin. Oncol. 2021, 26, 233–283. [Google Scholar] [CrossRef]
  32. Ngu, S.-F.; Ngan, H.Y.; Chan, K.K. Role of adjuvant and post-surgical treatment in gynaecological cancer. Best Pract. Res. Clin. Obstet. Gynaecol. 2022, 78, 2–13. [Google Scholar] [CrossRef]
  33. Shaukat, A.; Levin, T.R. Current and future colorectal cancer screening strategies. Nat. Rev. Gastroenterol. Hepatol. 2022, 19, 521–531. [Google Scholar] [CrossRef]
  34. Zygulska, A.L.; Pierzchalski, P. Novel diagnostic biomarkers in colorectal cancer. Int. J. Mol. Sci. 2022, 23, 852. [Google Scholar] [CrossRef]
  35. Ferrari, A.; Neefs, I.; Hoeck, S.; Peeters, M.; Van Hal, G. Towards novel non-invasive colorectal cancer screening methods: A comprehensive review. Cancers 2021, 13, 1820. [Google Scholar] [CrossRef]
  36. Cellina, M.; Cacioppa, L.M.; Cè, M.; Chiarpenello, V.; Costa, M.; Vincenzo, Z.; Pais, D.; Bausano, M.V.; Rossini, N.; Bruno, A. Artificial intelligence in lung cancer screening: The future is now. Cancers 2023, 15, 4344. [Google Scholar] [CrossRef]
  37. Mitsala, A.; Tsalikidis, C.; Pitiakoudis, M.; Simopoulos, C.; Tsaroucha, A.K. Artificial intelligence in colorectal cancer screening, diagnosis and treatment. A new era. Curr. Oncol. 2021, 28, 1581–1607. [Google Scholar] [CrossRef]
  38. Wesp, P.; Grosu, S.; Graser, A.; Maurus, S.; Schulz, C.; Knösel, T.; Fabritius, M.P.; Schachtner, B.; Yeh, B.M.; Cyran, C.C. Deep learning in CT colonography: Differentiating premalignant from benign colorectal polyps. Eur. Radiol. 2022, 32, 4749–4759. [Google Scholar] [CrossRef]
  39. Alsufayan, M.; Sulieman, A.; Moslem, R.; Asiri, A.; Alomary, A.; Alanazi, B.M.; Aldossari, H.; Alonazi, B.; Bradley, D.A. Assessment of imaging protocol and patients radiation exposure in computed tomography colonography. Appl. Sci. 2021, 11, 4761. [Google Scholar] [CrossRef]
  40. Hong, S.M.; Baek, D.H. A review of colonoscopy in intestinal diseases. Diagnostics 2023, 13, 1262. [Google Scholar] [CrossRef]
  41. Selvaraj, J.; Jayanthy, A. Design and development of artificial intelligence-based application programming interface for early detection and diagnosis of colorectal cancer from wireless capsule endoscopy images. Int. J. Imaging Syst. Technol. 2024, 34, e23034. [Google Scholar] [CrossRef]
  42. Selvaraj, J.; Jayanthy, A. Automatic polyp semantic segmentation using wireless capsule endoscopy images with various convolutional neural network and optimization techniques: A comparison and performance evaluation. Biomed. Eng. Appl. Basis Commun. 2023, 35, 2350026. [Google Scholar] [CrossRef]
  43. Shahsavari, D.; Waqar, M.; Chandrasekar, V.T. Image enhanced colonoscopy: Updates and prospects—A review. Transl. Gastroenterol. Hepatol. 2023, 8, 26. [Google Scholar] [CrossRef]
  44. Li, K.; Fathan, M.I.; Patel, K.; Zhang, T.; Zhong, C.; Bansal, A.; Rastogi, A.; Wang, J.S.; Wang, G. Colonoscopy polyp detection and classification: Dataset creation and comparative evaluations. PLoS ONE 2021, 16, e0255809. [Google Scholar] [CrossRef] [PubMed]
  45. Aliyi, S.; Dese, K.; Raj, H. Detection of gastrointestinal tract disorders using deep learning methods from colonoscopy images and videos. Sci. Afr. 2023, 20, e01628. [Google Scholar] [CrossRef]
  46. Jahn, B.; Bundo, M.; Arvandi, M.; Schaffner, M.; Todorovic, J.; Sroczynski, G.; Knudsen, A.; Fischer, T.; Schiller-Fruehwirth, I.; Öfner, D. One in three adenomas could be missed by white-light colonoscopy–findings from a systematic review and meta-analysis. BMC Gastroenterol. 2025, 25, 170. [Google Scholar] [CrossRef] [PubMed]
  47. Ali, S. Where do we stand in AI for endoscopic image analysis? Deciphering gaps and future directions. npj Digit. Med. 2022, 5, 184. [Google Scholar] [CrossRef]
  48. Young, E.; Edwards, L.; Singh, R. The role of artificial intelligence in colorectal cancer screening: Lesion detection and lesion characterization. Cancers 2023, 15, 5126. [Google Scholar] [CrossRef] [PubMed]
  49. Pulumati, A.; Pulumati, A.; Dwarakanath, B.S.; Verma, A.; Papineni, R.V. Technological advancements in cancer diagnostics: Improvements and limitations. Cancer Rep. 2023, 6, e1764. [Google Scholar] [CrossRef]
  50. Khalifa, M.; Albadawy, M. AI in diagnostic imaging: Revolutionising accuracy and efficiency. Comput. Methods Programs Biomed. Update 2024, 5, 100146. [Google Scholar] [CrossRef]
  51. Li, M.; Ali, S.M.; Umm-a-OmarahGilani, S.; Liu, J.; Li, Y.-Q.; Zuo, X.-L. Kudo’s pit pattern classification for colorectal neoplasms: A meta-analysis. World J. Gastroenterol. WJG 2014, 20, 12649. [Google Scholar] [CrossRef]
  52. De Carvalho, T.; Kader, R.; Brandao, P.; Lovat, L.B.; Mountney, P.; Stoyanov, D. NICE polyp feature classification for colonoscopy screening. Int. J. Comput. Assist. Radiol. Surg. 2025, 20, 1015–1024. [Google Scholar] [CrossRef]
  53. Van Doorn, S.C.; Hazewinkel, Y.; East, J.E.; Van Leerdam, M.E.; Rastogi, A.; Pellisé, M.; Sanduleanu-Dascalescu, S.; Bastiaansen, B.A.; Fockens, P.; Dekker, E. Polyp morphology: An interobserver evaluation for the Paris classification among international experts. Off. J. Am. Coll. Gastroenterol. | ACG 2015, 110, 180–187. [Google Scholar] [CrossRef] [PubMed]
  54. Uraoka, T.; Saito, Y.; Ikematsu, H.; Yamamoto, K.; Sano, Y. Sano’s capillary pattern classification for narrow-band imaging of early colorectal lesions. Dig. Endosc. 2011, 23, 112–115. [Google Scholar] [CrossRef]
  55. Komeda, Y.; Handa, H.; Watanabe, T.; Nomura, T.; Kitahashi, M.; Sakurai, T.; Okamoto, A.; Minami, T.; Kono, M.; Arizumi, T. Computer-aided diagnosis based on convolutional neural network system for colorectal polyp classification: Preliminary experience. Oncology 2017, 93, 30–34. [Google Scholar] [CrossRef]
  56. Sánchez-Peralta, L.F.; Bote-Curiel, L.; Picón, A.; Sánchez-Margallo, F.M.; Pagador, J.B. Deep learning to find colorectal polyps in colonoscopy: A systematic literature review. Artif. Intell. Med. 2020, 108, 101923. [Google Scholar] [CrossRef] [PubMed]
  57. Mathews, A.A.; Draganov, P.V.; Yang, D. Endoscopic management of colorectal polyps: From benign to malignant polyps. World J. Gastrointest. Endosc. 2021, 13, 356. [Google Scholar] [CrossRef] [PubMed]
  58. Itoh, H.; Oda, M.; Jiang, K.; Mori, Y.; Misawa, M.; Kudo, S.-E.; Imai, K.; Ito, S.; Hotta, K.; Mori, K. Binary polyp-size classification based on deep-learned spatial information. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 1817–1828. [Google Scholar] [CrossRef]
  59. Saad, A.I.; Maghraby, F.A.; Badawy, O.M. PolyDSS: Computer-aided decision support system for multiclass polyp segmentation and classification using deep learning. Neural Comput. Appl. 2024, 36, 5031–5057. [Google Scholar] [CrossRef]
  60. Grosu, S.; Wesp, P.; Graser, A.; Maurus, S.; Schulz, C.; Knösel, T.; Cyran, C.C.; Ricke, J.; Ingrisch, M.; Kazmierczak, P.M. Machine learning–based differentiation of benign and premalignant colorectal polyps detected with CT colonography in an asymptomatic screening population: A proof-of-concept study. Radiology 2021, 299, 326–335. [Google Scholar] [CrossRef]
  61. Sharma, P.; Balabantaray, B.K.; Bora, K.; Mallik, S.; Kasugai, K.; Zhao, Z. An ensemble-based deep convolutional neural network for computer-aided polyps identification from colonoscopy. Front. Genet. 2022, 13, 844391. [Google Scholar] [CrossRef]
  62. Barua, I.; Vinsard, D.G.; Jodal, H.C.; Løberg, M.; Kalager, M.; Holme, Ø.; Misawa, M.; Bretthauer, M.; Mori, Y. Artificial intelligence for polyp detection during colonoscopy: A systematic review and meta-analysis. Endoscopy 2021, 53, 277–284. [Google Scholar] [CrossRef]
  63. Bora, K.; Bhuyan, M.; Kasugai, K.; Mallik, S.; Zhao, Z. Computational learning of features for automated colonic polyp classification. Sci. Rep. 2021, 11, 4347. [Google Scholar] [CrossRef] [PubMed]
  64. Krenzer, A.; Heil, S.; Fitting, D.; Matti, S.; Zoller, W.G.; Hann, A.; Puppe, F. Automated classification of polyps using deep learning architectures and few-shot learning. BMC Med. Imaging 2023, 23, 59. [Google Scholar] [CrossRef]
  65. Krenzer, A.; Banck, M.; Makowski, K.; Hekalo, A.; Fitting, D.; Troya, J.; Sudarevic, B.; Zoller, W.G.; Hann, A.; Puppe, F. A real-time polyp-detection system with clinical application in colonoscopy using deep convolutional neural networks. J. Imaging 2023, 9, 26. [Google Scholar] [CrossRef] [PubMed]
  66. Wang, W.; Tian, J.; Zhang, C.; Luo, Y.; Wang, X.; Li, J. An improved deep learning approach and its applications on colonic polyp images detection. BMC Med. Imaging 2020, 20, 83. [Google Scholar] [CrossRef] [PubMed]
  67. Buitrago-Tamayo, A.C.; Lombo-Moreno, C.E.; Ursida, V.; Leguizamo-Naranjo, A.M.; Muñoz-Velandia, O.M.; Vargas-Rubio, R.D. Concordance between nice classification and histopathology in colonic polyps: A tertiary center experience. Ther. Adv. Gastrointest. Endosc. 2024, 17, 26317745241231102. [Google Scholar] [CrossRef]
  68. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  69. Targ, S.; Almeida, D.; Lyman, K. Resnet in resnet: Generalizing residual architectures. arXiv 2016, arXiv:1603.08029. [Google Scholar]
  70. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  71. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  72. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  73. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  74. Abid, A.; Abdalla, A.; Abid, A.; Khan, D.; Alfozan, A.; Zou, J. Gradio: Hassle-free sharing and testing of ml models in the wild. arXiv 2019, arXiv:1906.02569. [Google Scholar]
Figure 1. Anatomical representation of the human digestive system.
Figure 1. Anatomical representation of the human digestive system.
Diagnostics 15 01285 g001
Figure 2. CRC: (a) Signs and symptoms and (b) risk factors.
Figure 2. CRC: (a) Signs and symptoms and (b) risk factors.
Diagnostics 15 01285 g002
Figure 3. Progression of CRC. (a) Types of polyps and (b) transition from benign to malignant.
Figure 3. Progression of CRC. (a) Types of polyps and (b) transition from benign to malignant.
Diagnostics 15 01285 g003
Figure 4. CRC statistics: WHO (GLOBOCON).
Figure 4. CRC statistics: WHO (GLOBOCON).
Diagnostics 15 01285 g004
Figure 5. CRC D&T pathways.
Figure 5. CRC D&T pathways.
Diagnostics 15 01285 g005
Figure 6. Types of CRC screening methods.
Figure 6. Types of CRC screening methods.
Diagnostics 15 01285 g006
Figure 7. Colonoscopy images.
Figure 7. Colonoscopy images.
Diagnostics 15 01285 g007
Figure 8. Proposed workflow.
Figure 8. Proposed workflow.
Diagnostics 15 01285 g008
Figure 9. Architecture of CRP-ViT (multi-class).
Figure 9. Architecture of CRP-ViT (multi-class).
Diagnostics 15 01285 g009
Figure 10. CRP-ViT (multi-class) classification results. (a) Accuracy plot, (b) loss plot, (c) ROC plot, and (d) confusion matrix for test results.
Figure 10. CRP-ViT (multi-class) classification results. (a) Accuracy plot, (b) loss plot, (c) ROC plot, and (d) confusion matrix for test results.
Diagnostics 15 01285 g010
Figure 11. Developed user interface using Gradio. (a) Home Page; (b) predicted result: hyperplastic; (c) predicted result: serrated; (d) predicted result: adenoma; (e) predicted result: normal.
Figure 11. Developed user interface using Gradio. (a) Home Page; (b) predicted result: hyperplastic; (c) predicted result: serrated; (d) predicted result: adenoma; (e) predicted result: normal.
Diagnostics 15 01285 g011aDiagnostics 15 01285 g011bDiagnostics 15 01285 g011c
Table 1. Dataset used in the proposed study.
Table 1. Dataset used in the proposed study.
DatabaseParticipantsStill FrameResolutionType
PolypNormal
Real-Time (own) (DS-1)711001001920 × 1080Own
UAH DB (DS-2)767600768 × 576OA
Table 2. Class-wise distribution of the dataset used in this study.
Table 2. Class-wise distribution of the dataset used in this study.
ClassReal-Time (Own) (DS-1)UAH DB (DS-2)
PolypHyperplastic6021
Serrated3015
Adenoma1040
Normal10000
Total20076
Table 3. Dataset details after augmentation for training and validation.
Table 3. Dataset details after augmentation for training and validation.
ClassReal-Time (Own) (DS-1)
Original DatasetStage 1 AugmentationStage 2 Augmentation
Hyperplastic60601320
Serrated30601320
Adenoma10601320
Normal100601320
Total2002405280
Note: Stage 1 data augmentation (S1DA): serrated data were augmented using 1 augmentation techniques’ (30 × 1 = 30) original data with S1DA (30 + 30 = 60); adenoma data were augmented using 5 augmentation techniques (10 × 5 = 50) on original data with S1DA (10 + 50 = 60); Stage 2 data augmentation (S2DA): used 11 augmentation techniques (60 × 21 = 1260) on original data with S1DA (60 + 1260 = 1320).
Table 4. Data split.
Table 4. Data split.
ClassTraining (80%)Validation (20%)Testing (100%)
Real-Time (OWN) (DS-1)Real-Time (Own) (DS-1)UAH DB (DS-2)
Hyperplastic10562640021
Serrated10562640015
Adenoma10562640040
Normal10562644000
Total4224105640 + 76 = 116
Note: After the Stage 2 data augmentation (S2DA): 80% of data (1056 images in each class) and 20% of data (264 images in each class) were used for training and validation, respectively.
Table 5. Classification results achieved using the VGG16 architecture.
Table 5. Classification results achieved using the VGG16 architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD50/50085.1273.4191.5773.290.8268.9971.8671.8671.5971.5971.8667.71
18468.9990.6272.0688.9264.3964.3967.8067.8064.39
283.9270.4890.2771.6989.4264.4764.4766.6766.6764.47
385.9973.2491.1168.9490.8570.6670.6664.7764.7770.66
ADAM48/50090.6183.9394.7885.0494.0379.1979.9379.9382.9582.9579.9377.84
189.9180.2693.8582.7792.9476.0376.0376.8976.8976.03
289.2481.5592.9281.6392.3776.1476.1476.1476.1476.14
391.2984.3395.1680.4994.579.2879.2875.3875.3879.28
RMSprop50/50088.8580.7393.4181.7292.5975.1978.0878.0876.8976.8978.0873.96
188.1278.2992.780.5991.5871.1271.1274.6274.6271.12
289.2378.593.3679.1792.671.0671.0673.4873.4871.06
389.8381.3694.0877.2793.2176.0276.0270.8370.8376.02
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 6. Classification results achieved using the VGG19 architecture.
Table 6. Classification results achieved using the VGG19 architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD50/50085.1273.4191.5773.290.8271.4775.4975.4972.3572.3575.4970.08
18468.9990.6272.0688.9265.3865.3870.8370.8365.38
283.9270.4890.2771.6989.4266.0666.0669.3269.3266.06
385.9973.2491.1168.9490.8574.5874.5867.8067.8074.58
ADAM47/50090.6183.9394.7885.0494.0382.4884.5384.5384.8584.8584.5380.87
189.9180.2693.8582.7792.9477.3477.3481.4481.4477.34
289.2481.5592.9281.6392.3778.2878.2879.1779.1778.28
391.2984.3395.1680.4994.583.7483.7478.0378.0383.74
RMSprop50/50088.8580.7393.4181.7292.5979.6983.4683.4682.2082.2083.4678.13
188.1278.2992.780.5991.5873.5973.5979.1779.1773.59
289.2378.593.3679.1792.674.4474.4476.1476.1474.44
389.8381.3694.0877.2793.2181.8281.8275.0075.0081.82
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 7. Classification results achieved using the ResNet50 architecture.
Table 7. Classification results achieved using the ResNet50 architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD46/50090.1284.3895.1684.9494.4583.7687.8087.8084.4784.4787.8081.91
189.7782.294.5184.3893.6176.8476.8482.9582.9576.84
290.6782.3195.5583.7194.1877.7477.7480.6880.6877.74
391.2586.3496.1282.0195.186.4286.4279.5579.5586.42
ADAM39/50095.1291.0497.5590.4496.6589.290.2690.2691.2991.2990.2688.07
194.6787.296.8889.6895.4585.7185.7188.6488.6485.71
294.9487.5397.289.1196.186.4286.4286.7486.7486.42
395.7891.2297.8887.5997.0290.0490.0485.6185.6190.04
RMSprop43/50094.2288.3696.5589.1195.4987.3887.7387.7389.3989.3987.7386.93
193.6985.8796.1388.0794.6985.2485.2487.5087.5085.24
294.0686.3796.4787.0395.1485.7785.7786.7486.7485.77
395.0289.0397.1285.3296.4389.1689.1684.0984.0989.16
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 8. Classification results achieved using the DenseNet121 architecture.
Table 8. Classification results achieved using the DenseNet121 architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD50/50094.3984.298.5482.7798.5181.8793.5684.4795.9884.4793.5680.21
193.7379.6198.5282.198.2591.4878.1494.9282.5891.48
293.8179.8798.5181.5398.2689.9677.5393.5678.4189.96
394.428498.3181.0698.3390.5380.8992.6775.3890.53
ADAM46/50095.0688.3698.5188.4598.5287.5595.9187.6496.7988.6495.9186.27
194.8286.6698.4887.9798.3294.3685.4595.8386.7494.36
294.8486.7598.4987.4198.3193.8485.1795.7284.8593.84
395.0988.4698.3486.3698.3394.4186.8295.7284.8594.41
RMSprop50/50093.1287.4698.7987.4693.1286.2794.7485.8796.5587.594.7484.66
192.7384.6698.6686.7492.5593.3782.9195.8386.3693.37
292.6485.2298.5386.2792.3792.8483.7795.0884.0992.84
393.1487.8498.6184.8593.1893.6686.2394.6680.6893.66
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 9. Classification results achieved using the EfficientNetV2 architecture.
Table 9. Classification results achieved using the EfficientNetV2 architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD47/50091.3783.3894.8384.5694.3883.0792.885.8294.9584.8595.0181.06
190.7381.4894.6684.1993.6389.277.0394.0582.5891.82
290.8681.2594.1382.4893.6789.277.3793.3780.392.16
392.0286.4693.8181.0695.7792.884.8795.676.5295.6
ADAM43/50094.4290.2496.4489.396.4488.6694.9289.4396.5989.7796.4787.69
193.9787.2796.388.9295.6893.385.0496.0388.2694.81
294.0487.4596.1388.4595.7793.9486.1495.6987.1295.33
394.9789.7696.6587.9796.6595.4590.497.0485.6197.06
RMSprop45/50093.9188.6796.0488.1696.0686.9894.0186.8995.9787.8895.9485.70
192.2685.4595.9287.889592.884.1395.4286.3694.57
292.8485.6795.5286.6595.1793.0884.795.2985.9894.82
394.2788.2496.2685.2396.2693.9487.296.0182.5896.01
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 10. Classification results achieved using the InceptionNetV3 architecture.
Table 10. Classification results achieved using the InceptionNetV3 architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD49/50092.1283.6294.5683.6294.5782.5592.7184.8598.7584.8592.7180.59
191.182.7794.282.7793.2991.4378.798.5682.5891.05
291.382.2994.182.2993.4690.7377.798.1479.1790.36
392.6181.5394.8681.5395.1991.9681.398.0875.7692.33
ADAM45/50096.6189.0798.7488.7598.7488.0496.1587.5598.987.8898.8986.62
196.4786.8898.9588.4598.6495.8485.3398.9287.0398.75
296.586.9698.9787.7898.6595.9285.6198.9386.1798.74
396.889.3398.8587.2298.8796.388.0998.7685.4298.77
RMSprop48/50096.1587.5598.987.8898.8986.6292.986.2598.5687.8892.5184.94
195.8485.3398.9287.0398.7591.8283.1598.1885.9891.34
295.9285.6198.9386.1798.7491.7584.0397.9983.7191.26
396.388.0998.7685.4298.7792.7986.4598.1582.292.45
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 11. Classification results achieved using the ViT architecture.
Table 11. Classification results achieved using the ViT architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD42/50093.4587.0595.8287.8894.9586.3685.7785.7786.7486.7485.7784.85
192.4984.7395.3687.2293.9184.2784.2785.2385.2384.27
29385.5195.686.0894.2883.7783.7784.0984.0983.77
393.8588.2996.1884.2895.2585.6085.6083.3383.3385.60
ADAM35/50096.3693.4597.9293.1897.2192.3896.491.8597.9693.9497.2291.1
195.7191.4797.8591.3896.2995.2789.9397.0891.2996.59
295.891.597.8992.7196.0895.1790.4996.7290.1596.84
396.3593.1298.1792.2396.9295.3692.1696.3889.0297.47
RMSprop39/50094.5391.1196.7491.1995.8990.3495.2788.9397.9691.2996.5989.11
193.8489.8196.6890.1595.4894.2187.8297.0890.1596.84
293.789.5596.5589.395.2594.0787.9296.7288.2697.22
394.190.8996.8690.7295.7494.491.9796.3886.7497.47
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 12. Classification results achieved using CRP-ViT (multi-class) architecture.
Table 12. Classification results achieved using CRP-ViT (multi-class) architecture.
OptimizersEpochClassTrainingValidation
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
AccuracySensitivitySpecificityPrecisionNPVOverall
Accuracy
SGD35/50092.5292.4397.2291.1995.791.7695.5191.6397.9491.2996.8689.77
191.0491.0496.6791.3895.3193.7686.8697.0890.1596.41
291.1191.1196.4491.1995.194.2688.7296.789.3997.12
392.4892.4897.2891.9595.8394.8892.0996.4588.2697.54
ADAM22/50097.9297.5599.3197.4498.997.2898.0197.7598.2998.8698.5196.02
196.4496.4499.1397.2598.595.9393.897.9797.3596.2
297.1697.1699.1596.598.7596.5795.0896.6895.0897.04
397.9897.9899.4397.989996.9797.6196.892.899.02
RMSprop29/50096.4996.4998.9296.2198.695.4597.3595.5498.897.3598.6594.22
194.3994.3998.7595.2798.293.9492.5495.193.9496.15
294.8294.8298.694.7998.392.893.5192.792.896.1
396.1696.1699.0194.7998.592.895.3391.992.899.1
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma. Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Table 13. K-fold cross validation on CRP-ViT (multi-class).
Table 13. K-fold cross validation on CRP-ViT (multi-class).
K-FoldClassAccuracySensitivitySpecificityPrecisionNPVOverall Accuracy
K = 1099.1597.7599.6298.8699.2496.69
198.1194.8599.2397.7398.23
298.0196.298.6195.8398.74
398.1198.0398.1394.3299.37
K = 2099.2397.7799.6299.6299.2497.92
198.4997.3799.0498.1198.49
298.297.3698.8697.7399.24
398.7799.2298.796.2199.79
K = 3097.1196.2797.7397.7397.1295.08
195.993.6897.0495.4596.56
295.5294.6895.8494.3296.83
394.8795.794.7492.898.16
K = 4097.1195.5697.7396.2197.1295.17
195.994.7896.2193.5696.56
295.5294.6495.8493.1896.83
394.8795.7294.7493.1898.16
K = 5098.4897.0198.4896.9798.2496.5
196.9795.5296.9796.5997.25
296.5995.1596.5993.9498.05
395.8398.4195.8393.9499.48
Mean ± SD098.22 ± 0.9496.87 ± 0.8698.64 ± 0.8597.88 ± 1.2498.19 ± 0.9596.27 ± 1.06
197.07 ± 1.0295.24 ± 1.2297.70 ± 1.2196.29 ± 1.6597.42 ± 0.81
296.77 ± 1.0695.61 ± 1.0497.15 ± 1.3395.00 ± 1.6197.94 ± 0.98
396.89 ± 1.5597.42 ± 1.4596.43 ± 1.6894.09 ± 1.1998.99 ± 0.69
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma., and all metrics are presented as percentages.
Table 14. Difference between 80:20 and K-fold.
Table 14. Difference between 80:20 and K-fold.
MethodClassAccuracySensitivitySpecificityPrecisionNPVOverall Accuracy
80:20 Split
(Hold-Out)
(A)
098.0197.7598.2998.8698.5196.02
195.9393.897.9797.3596.2
296.5795.0896.6895.0897.04
396.9797.6196.892.899.02
K-Fold Cross
(Leave-Out)
(B)
098.2296.8798.6497.8898.1996.27
197.0795.2497.796.2997.42
296.7795.6197.159597.94
396.4997.4296.4394.0998.99
Difference Between
(A and B)
00.210.880.350.980.320.25
11.141.440.271.061.22
20.20.530.470.080.9
30.480.190.371.290.03
Note: Class 0—normal, Class 1—hyperplastic, Class 2—serrated, and Class 3—adenoma; all metrics are presented as percentages.
Table 15. Ablation study for CRP-ViT (multi-class) architecture.
Table 15. Ablation study for CRP-ViT (multi-class) architecture.
Removal/Modification Techniques Applied to ResNet50 Architecture TechniquesAccuracy
Stage 1 layer modification 90.18
Stage 2 layer modification91.71
Stage 3 layer modification92.78
Stage 4 layer modification90.74
Simultaneous modifications across Stages 1 to 467.06
Modification at the fully connected (FC) layer87.85
CRP-ViT97.28
Note: Optimal performance metrics are highlighted in bold, and all metrics are presented as percentages.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Selvaraj, J.; Sadaf, K.; Aslam, S.M.; Umapathy, S. Multiclassification of Colorectal Polyps from Colonoscopy Images Using AI for Early Diagnosis. Diagnostics 2025, 15, 1285. https://doi.org/10.3390/diagnostics15101285

AMA Style

Selvaraj J, Sadaf K, Aslam SM, Umapathy S. Multiclassification of Colorectal Polyps from Colonoscopy Images Using AI for Early Diagnosis. Diagnostics. 2025; 15(10):1285. https://doi.org/10.3390/diagnostics15101285

Chicago/Turabian Style

Selvaraj, Jothiraj, Kishwar Sadaf, Shabnam Mohamed Aslam, and Snekhalatha Umapathy. 2025. "Multiclassification of Colorectal Polyps from Colonoscopy Images Using AI for Early Diagnosis" Diagnostics 15, no. 10: 1285. https://doi.org/10.3390/diagnostics15101285

APA Style

Selvaraj, J., Sadaf, K., Aslam, S. M., & Umapathy, S. (2025). Multiclassification of Colorectal Polyps from Colonoscopy Images Using AI for Early Diagnosis. Diagnostics, 15(10), 1285. https://doi.org/10.3390/diagnostics15101285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop