Previous Article in Journal
E-CMCA and LSTM-Enhanced Framework for Cross-Modal MRI-TRUS Registration in Prostate Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review

1
School of Electrical and Electronic Engineering, Engineering Campus, Universiti Sains Malaysia, Pulau Pinang 14300, Malaysia
2
School of Automation and Information Engineering, Sichuan University of Science and Engineering, Yibin 644000, China
3
College of Biomedical Engineering, Chongqing Medical University, Chongqing 400016, China
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(9), 293; https://doi.org/10.3390/jimaging11090293
Submission received: 4 August 2025 / Revised: 25 August 2025 / Accepted: 25 August 2025 / Published: 27 August 2025
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

Colorectal cancer is one of the three most common cancers worldwide. Early detection and assessment of polyps can significantly reduce the risk of developing colorectal cancer. Physicians can obtain information about polyp regions through polyp segmentation techniques, enabling the provision of targeted treatment plans. This study systematically reviews polyp segmentation methods. We investigated 146 papers published between 2018 and 2024 and conducted an in-depth analysis of the methodologies employed. Based on the selected literature, we systematically organized this review. First, we analyzed the development and evolution of the polyp segmentation field. Second, we provided a comprehensive overview of deep learning-based polyp image segmentation methods and the Mamba method, as well as video polyp segmentation methods categorized by network architecture, addressing the challenges faced in polyp segmentation. Subsequently, we evaluated the performance of 44 models, including segmentation performance metrics and real-time analysis capabilities. Additionally, we introduced commonly used datasets for polyp images and videos, along with metrics for assessing segmentation models. Finally, we discussed existing issues and potential future trends in this area.

1. Introduction

Colorectal cancer (CRC) is one of the most common cancers [1]. Most CRC cases develop from small polyps protruding from the surface of the colon and rectum [2]. Colonoscopy can help doctors detect and remove polyps before they turn into cancer. The low contrast of colonoscopy images and the diversity in the size, shape, texture, color, and orientation of polyps pose challenges for precise segmentation [3]. Accurate polyp segmentation is crucial for subsequent treatment. Early examinations [4] relied on manually crafted features such as color and texture, but recent research based on deep learning methods has made polyp segmentation more accurate.

1.1. Research Progress on Polyp Segmentation

Traditional methods for polyp segmentation rely on manually created features, utilizing texture characteristics in images to capture the geometric orientation information of polyps [4,5,6,7,8,9,10,11]. Due to the irregular shapes, varying sizes, and different locations of polyps in colonoscopy images, these traditional methods have limitations, including false detection and poor generalization.
The emergence of Fully Convolutional Networks (FCNs) in 2015 [12] marked the beginning of research based on deep learning models. In 2017, the introduction of Transformer [13] led to the proposal of related Transformer-based models. By the end of 2023, the advent of Mamba [14] initiated research on polyp segmentation based on the Mamba model.
As shown in Figure 1, during the progress of polyp segmentation research, we selected Jerebko et al. [6] as a representative of traditional methods. For Convolutional Neural Network (CNN)-based models from 2015 to 2024, we chose UNet [15], Manjunath et al. [16], Brandao et al. [17], Unet++ [18], ResUNet++ [19], PraNet [20], TransFuse [21], PNS+ [22], ERDUnet [23], and WDFF-Net [24] as representatives for their respective years.
From 2021 to 2024, we selected Polyformer [25], SSFormer [26], PVT-CASCADE [27], and CTNet [28] as representative Transformer-based models for those years. We chose Polyp-Mamba [29] as the representative Mamba-based model for 2024.

1.2. Relevant Review Survey

Relevant research is shown in Table 1. The above review effectively summarizes existing polyp segmentation work but also has some shortcomings: Some reviews lack a systematic method summary table; some lack model performance evaluation analysis; some reviews analyze existing methods from a problem perspective, but with relatively few analyzed literature and papers. Compared to the review papers, this review encompasses 146 papers published between 2018 and 2024, representing both a longer time span and a larger corpus of analyzed literature, and additionally provides statistical analysis of video polyp segmentation methods. Second, the method and model section analyzes the advantages and disadvantages of method models; third, it classifies method models from the perspective of solving specific polyp segmentation problems and network architectures; fourth, the performance evaluation section provides performance analysis of these methods.

1.3. Our Contribution

This review aims to provide a comprehensive review of polyp segmentation methods, including CNN, Transformer, hybrid, and Mamba approaches for image-based segmentation as well as video polyp segmentation methods. It offers a systematic overview and summary, categorizing and analyzing each study based on the specific polyp segmentation problems addressed. Additionally, it evaluates each method in the performance assessment section.
By primarily collecting relevant literature from the past three years, this paper explores the main issues and corresponding solutions in current research in this field, providing a reliable reference for researchers. The main contributions of this review are as follows:
(1)
Classify polyp segmentation methods into five categories based on their network architectures: CNN, Transformer, Hybrid, Mamba, and Other.
(2)
Conduct a systematic study of polyp segmentation methods. Summarize the 87 papers and categorize the specific problems addressed by each polyp segmentation method into five categories.
(3)
Provide a comprehensive compilation of datasets required for image and video polyp segmentation.
(4)
We conducted a performance evaluation of methods for solving different problems, comparing and analyzing 44 models.
(5)
Discuss the existing limitations and future research trends.
We organize the remaining sections of this review as follows: Section 2 introduces the literature search and review process. Section 3 defines polyps and discusses the challenges doctors face when managing them. Section 4 presents methods for polyp segmentation. Section 5 reviews the loss functions used in polyp segmentation. Section 6 describes the datasets used for polyp segmentation. Section 7 lists the metrics for evaluating models. Section 8 assesses the performance of these models. Section 9 explores the ongoing challenges in polyp segmentation. Section 10 investigates future research trends. Section 11 provides the conclusion.

2. Literature Search and Review Procedures

This review systematically assesses the effectiveness of deep learning methods for automated polyp segmentation in colonoscopy images. The primary research questions addressed are as follows: (1) How do various deep learning architectures—including convolutional neural networks (CNNs), Transformer-based models, hybrid approaches, and other emerging architectures—perform in colorectal polyp segmentation? (2) What are the current limitations associated with existing approaches, and what are promising avenues for future research?
Inclusion Criteria: Original research studies developing, validating, or evaluating deep learning models specifically for colorectal polyp segmentation were considered eligible. Studies included were published in English between 1 January 2018 and 10 December 2024; employed deep learning-based pixel-level segmentation techniques; reported at least one standardized segmentation evaluation metric (Dice coefficient or Intersection over Union); and were available as peer-reviewed journal articles or conference proceedings with accessible full texts.
Exclusion Criteria: The following types of articles were excluded: editorials, commentaries, or papers lacking full-text abstracts; studies relying exclusively on traditional image processing without deep learning; research focusing solely on polyp detection or classification without performing segmentation; studies lacking clear methodological descriptions or relevant performance metrics; duplicate publications; and non-English-language publications.
Information Sources and Search Date: We conducted a comprehensive search across Web of Science, IEEE Xplore, Scopus, and SpringerLink databases, employing the query combination: “polyp segmentation AND deep learning OR convolutional neural network”, further enhanced by including specific publicly available datasets (e.g., Kvasir, CVC-ClinicDB). The final literature search was completed on 10 December 2024.
The initial database search identified a total of 2951 records across four electronic databases. After removing 983 duplicates, 1968 articles remained for further screening. Among these, 1689 papers were subsequently excluded based on the following criteria: (1) publications were limited to the past three years (2022–2024); (2) included studies were further assessed for eligibility based on the availability of source code links, citation counts exceeding 5, journal impact factors (IFs) greater than 3, clarity of methodological framework diagrams, and comprehensiveness of experimental validations on relevant polyp datasets. Ultimately, 146 original research articles met the inclusion criteria and were retained for detailed evaluation. The systematic literature review selection process is depicted in the PRISMA flow diagram (Figure 2) [37].
The screening criteria for papers from 2018–2021 include those with over a thousand citations in Google Scholar, as well as those with codes and over one hundred stars. The publication quantity of polyp segmentation research papers from 2018–2024 and their distribution after screening are shown in Figure 3.

3. Review About Polyp

In this section, we will explore three aspects: what polyps are, how doctors detect their presence, and the challenges doctors encounter when treating polyps.
Polyps are abnormal tissue growths that protrude inward or outward from the mucosal surface; they typically appear mushroom-shaped or cauliflower-like and can range in size from a few millimeters to several centimeters [38,39]. The human body commonly finds them in various organs, especially the digestive tract [40]. Colon polyps grow on the inner walls of the colon or the rectum. Figure 4 shows the diversity of polyp appearances across datasets. Although most polyps are benign, tumorous polyps may carry a risk of becoming cancerous [41]. Approximately 70% of colorectal polyps are adenomatous polyps, which have the highest malignant potential, while hyperplastic polyps account for 10–30% and generally have low malignant risk [42,43]. Colon cancer survival rates vary depending on the stage of detection, ranging from over 95% in stage I to less than 35% in stages IV and V [44]. Therefore, timely detection and treatment are very important.
Doctors use ultrasound examinations or colonoscopies to check for the presence of polyps [45,46]. Colonoscopy is considered the gold standard for polyp detection, with a sensitivity of 75–93% for polyps ≥ 6 mm [47]. If doctors find polyps, they may recommend a biopsy and choose appropriate treatments based on the type and size of the polyps [48]. Screening and removal of colonic mucosal polyps are among the effective methods for preventing CRC [49]. Compared to the expected mortality rate in the general population, the CRC mortality rate in patients who had adenomas removed decreased by 53% within a median of 15.8 years [50]. Recent studies have shown that polypectomy can reduce CRC incidence by 76–90% [51].
Due to the varying types of polyps, complex morphologies, and indistinct polyp boundaries, doctors may encounter challenges in identifying polyps, particularly small or flat ones [52,53]. The miss rate for polyps during colonoscopy ranges from 6% to 27%, with higher rates for smaller polyps [54]. This can result in false positives or missed detections. Large polyps, polyps located in hard-to-reach areas such as bends or narrow sections of the intestine, and polyps growing near critical blood vessels are more challenging to segment and carry higher risks [55,56]. Different types of polyps, including adenomatous polyps and inflammatory polyps, have distinct histological characteristics, which can affect the difficulty and strategies for segmentation [57]. Doctors need to guarantee the complete removal of polyps and carry out comprehensive examinations. The quality of endoscopic images impacts physicians’ judgments and procedural accuracy [58]. Factors such as bowel preparation quality, withdrawal time, and endoscopist experience significantly affect polyp detection rates [59,60]. Blurry or unclear images may increase the difficulty and risk of surgery. The issues mentioned above require precise polyp segmentation techniques to assist doctors in handling them. Polyp segmentation is a pixel-level task that can accurately delineate polyps from colonoscopy images [61]. Figure 5 shows samples of polyp images from four datasets along with their corresponding segmentation masks. Continuous innovation and advancement in polyp segmentation models can significantly improve early diagnosis and precise treatment of polyps, thereby enhancing patients’ quality of life [62].

4. Polyp Segmentation Model

In this section, we categorize and describe methods based on network architectures, and within each architecture, we classify different approaches according to the specific problems they address.
Currently, there are six major challenges in polyp segmentation: First, polyps vary in shape and size, exhibiting diversity in morphology, brightness, and color. This variability directly impacts two critical aspects: feature extraction becomes difficult due to the heterogeneous appearance of polyps, and boundary detection is challenged by blurred edges between polyps and surrounding tissue. Second, limited dataset issues arise from insufficient pixel-level annotations, making it difficult to train fully supervised polyp segmentation models. Small datasets can also lead to overfitting problems. Third, generalization performance remains a significant challenge, as trained models often show decreased performance on unseen test datasets. When the distributions of training and test data differ, domain adaptation techniques become necessary to address the decline in model performance. Fourth, real-time performance is crucial for clinical applications, where large parameter counts and computational complexity can hinder the deployment of models in practical endoscopy procedures. Fifth, low-quality frames in endoscopy videos, caused by factors such as motion blur, poor illumination, or artifacts, pose additional challenges for accurate segmentation. Finally, limitations of the architecture itself can constrain performance, as certain network designs may not fully capture the complex features required for precise polyp segmentation.
This section introduces polyp segmentation models based on images and videos from various network architectures. Figure 6 illustrates the classification of network architectures and the challenges faced by polyp segmentation methods.

4.1. CNN-Based Methods

CNN-based polyp segmentation methods center around convolutional neural networks and often utilize architectures like U-Net and ResUNet. These methods perform multi-scale feature extraction and fusion by stacking convolutional and pooling layers. Skip connections are used to retain detailed information, making them well-suited for extracting local features. The advantage of CNN approaches lies in their high computational efficiency, but they may struggle to capture long-range dependencies, which can limit segmentation accuracy.

4.1.1. Polyp Diversity and Boundary Factors

Addressing the diversity of polyps and the issue of blurred boundaries is crucial to the performance of polyp segmentation models. Therefore, this section reviews CNN-based models from two aspects: solving the problem of blurred boundaries and handling the morphological diversity of polyps.
Problems of Difficult Boundary Segmentation
The goal is to solve the difficult problem of precise boundary segmentation caused by unclear polyp boundaries. Bui et al. [63] proposed a multi-scale edge-guided attention network aimed at addressing segmentation challenges, such as the ambiguity of polyp boundaries and their varied morphologies. The method achieves fine edge enhancement by combining an encoder, a decoder, and an edge-guided attention module that uses the Laplacian operator. While this approach effectively highlights weak polyp boundaries and enhances segmentation performance, it requires further optimization for real-time performance. The goal of MISNet [64] is to solve the problems that come up when trying to separate polyps during a colonoscopy, such as changing lighting, polyps’ location and shape, and lines that are not clear. MISNet actively aggregates multi-scale features, generates guidance maps via a selective shared fusion module, and extracts and utilizes boundary information through a Parallel Attention Module and a Balanced Weight Module. This method effectively captures multi-level features, enhances boundary information extraction, and demonstrates excellent generalization ability on new datasets. However, the segmentation performance in extremely low-contrast environments still requires further improvement. Fan et al. [20] conducted a study to resolve the segmentation issue in endoscopic images, which arises due to the variability in polyp appearances and the lack of clear boundaries between them and the mucosa. They proposed the parallel reverse attention network as a solution. The system employs parallel part decoders to integrate high-level features into global maps, while a reverse attention module gradually eliminates estimated regions to unveil boundary details. This method effectively captures the polyp regions and boundary details. However, the model may still overlook precise constraints on polyp boundaries to some extent and has a significant reliance on high-quality annotated data. To solve the problems of appearance diversity and low boundary contrast in polyp segmentation, Liu et al. [65] came up with Focus on Boundary Segmentation (FoBS), a multi-level framework for improving the quality of segmentation boundaries. FoBS generates diverse polyp images through boundary-aware mixing, captures high-frequency boundary information with a variable Laplacian feature refinement module, and corrects degraded pixel regions using a position-sensitive compensation criterion. This framework synergistically enhances segmentation performance at the sample, feature, and optimization levels. It offers plug-and-play advantages, allowing for direct improvement in existing model performance, and is the first to extend the Laplacian operator into a deformable form for polyp segmentation. Its strengths lie in significantly enhancing boundary quality; however, it has limited effectiveness in segmenting atypical polyp morphologies and is sensitive to noise. CFA-Net proposed by Zhou et al. [66] addresses the challenges of polyps of various sizes and shapes, as well as the lack of clear boundaries with the surrounding environment. It integrates a boundary prediction network, a dual-stream segmentation network, and a boundary feature aggregation module to achieve cross-level feature fusion. This method effectively enhances segmentation performance for small polyps and in scenarios with ambiguous boundaries by combining boundary-aware features and multi-scale semantic information. Its strengths are particularly evident in situations with unclear boundaries; however, it does not meet real-time segmentation requirements and lacks the ability to capture fine boundaries for large polyps. Notably, the segmentation accuracy significantly decreases when the appearance of the polyp is similar to the background.
Yue et al. [67] addressed the issue of suboptimal polyp segmentation by proposing a boundary constraint network that integrates cross-layer context and edge information to enhance segmentation performance. This method utilizes a Cross-layer Feature Integration Strategy, which employs an attention mechanism to adaptively fuse high-level features, in conjunction with a Bilateral Boundary Extraction Module, which refines shallow polyp regions and non-polyp regions based on high-level positional information and boundary supervision. Its advantage lies in its excellent performance in situations with ambiguous boundaries and complex backgrounds. Nevertheless, in certain specific complex scenarios, BCNet may still exhibit inaccuracies in segmentation. Lin et al. [68] addressed the challenge of extracting boundary information in polyp segmentation by proposing the Bit-Slice Context Attention Network. This network includes the Bit-Slice Context Attention (BSCA) module, the Segmentation-Squeeze-Bottleneck Union (SSBU) module, the Multi-Path Attention Decoder (MCAD), and the Multi-Path Attention Encoder (MACE), which effectively suppress noise in feature maps and enhance feature representation. The BSCA module focuses on the boundary information between the polyp and surrounding tissue, while SSBU, MCAD, and MACE explore geometric information and detailed features from multiple perspectives. This method can effectively extract complex boundaries and improve segmentation accuracy. However, in scenarios where the background is highly similar to the polyp or where there are complex interferences, it may lose some low-level details, necessitating further robustness improvements.
Issues on Diversity in the Morphology, Size, Brightness, and Color of Polyps
To address the diversity in polyp shape, size, brightness, and color, Chen et al. [69] addressed the issue of color appearance imbalance and diverse morphologies leading to blurred boundaries in colorectal polyps with limited training data by proposing a method that combines Dataset-Level Color Augmentation (DLCA) and a Convolutional Multi-Scale Attention Module (CMAM). DLCA is based on dataset-level color statistics, significantly improving the color diversity of the training data and better addressing color imbalance. CMAM effectively handles the complexities of polyp size, shape, and boundary blurriness through multi-scale feature extraction and attention mechanisms. This method is the first to utilize dataset-level color statistics for data augmentation and deeply explores multi-scale contextual features. However, there is still a need for optimization in computational efficiency, and the issue of data scarcity remains to be resolved. UHA-Net [70] fully utilizes cross-layer and multi-scale features, aiming to address the challenge of scale variation in lesion areas within medical images. This framework includes a Hierarchical Feature Fusion module for coarse localization of targets, an Uncertainty Cross-layer Fusion module to enhance the effective integration of features from adjacent layers, and a Scale Aggregation Module that learns multi-scale features through convolutional kernels of different sizes. This approach excels in segmenting small-sized lesions and accurately locating lesion boundaries, effectively reducing both over-segmentation and under-segmentation issues. Conversely, its multiple-module design incurs a higher computational cost and shows a significant dependence on high-quality pixel-level annotated data. Wang et al. [71] addressed the issue of colon polyps being highly similar to their surrounding environment in terms of color, shape, and texture by proposing a semantic enhanced perception network. This network includes a multi-scale adaptive perception, module to enhance feature extraction, a polyp semantic enhancement module that utilizes coarse segmentation maps to extract semantic information, and a context representation calibration module that calibrates details through a branching network. SEPNet demonstrates excellent generalization ability in segmentation tasks within similar background scenes, achieving an inference speed of up to 62 Frames Per Second (FPS). However, it faces limitations in preserving edge details and has limited utilization of low-level local features. The FEGNet proposed by Jin et al. [72] aims to address the issue of ambiguous boundaries in polyp segmentation caused by variations in shape, size, and texture. The network employs a U-shaped architecture and integrates a feedback mechanism through a Recurrent Gate Module to optimize the attention maps of high-level features, while an edge extraction module is responsible for processing low-level features. The model enhances the learning process through three strategies: multi-scale supervision, recurrent supervision, and edge supervision. It performs exceptionally well in segmentation tasks involving complex shapes and smaller-sized polyps. However, its generalization capability still requires further improvement. CoInNet [73] aims to address the variability and unclear boundaries in polyps in terms of shape, color, size, and texture. This network combines the advantages of convolution and deconvolution operations, capturing the relationships between feature maps through statistical feature attention units and recursively fusing features using the Anomalous Boundary Approximation module to optimize segmentation results. It performs exceptionally well in the detection of small polyps. However, there remains a risk of missed detections and misclassifications when polyps coexist with folds or when there are interfering objects that resemble polyps. Wang et al. [74] addressed the challenge of segmenting small, low-contrast, and morphologically diverse polyps in colonoscopy images by proposing ISCNet, which integrates image-level and surrounding contextual information. This network includes a Global Guided Context Aggregation module and a Diversified Surrounding Context Focusing module, which enhance background–foreground differentiation and multi-scale feature extraction around the polyps, respectively. Through a combined loss strategy of binary cross-entropy and Dice coefficient, the model performs well in handling small polyps and demonstrates strong generalization capabilities. However, it may still encounter segmentation errors and misjudgments in scenarios where the polyp edges and surrounding textures are similar. Zhang et al. [75] proposed LDNet, which is based on a U-shaped structure, to address the diversity in polyps in terms of shape, size, and brightness, as well as the subtle issues related to background contrast. This network enhances the feature contrast between the polyps and the background region through lesion-aware cross-attention and captures long-range contextual relationships using efficient self-attention. The model improves segmentation accuracy by combining a dynamic kernel generation and updating mechanism with these attention-based approaches. However, the study did not delve into the model’s real-time performance and computational efficiency issues. Srivastava et al. [76] proposed a multi-scale residual fusion network to address the challenges of effectively segmenting multi-size targets and the difficulty of training on small-scale and biased datasets. This method introduces a Dual-Scale Dense Fusion block within the network to achieve cross-resolution feature exchange and employs residual dense blocks to reduce reliance on large training datasets. The Shape Stream and triple attention mechanism further enhance spatial accuracy and feature extraction capabilities while maintaining low computational complexity. However, the network’s performance in handling low-contrast images remains relatively weak. Liu et al. [77] addressed the issue that convolutional neural networks can only learn features in the spatial domain, making it difficult to capture more directional, singular, and regular patterns, by proposing a multi-scale contour-guided learning segmentation network. This method extends CNNs from the spatial domain to the spectral domain, obtaining multi-scale and multi-directional features through multi-scale contour sparse representation and enhancing spectral domain feature representation by utilizing contour knowledge to guide learning. Its advantage lies in achieving complementarity between the spatial and spectral domains, using contour kernels to make features more interpretable. However, the real-time performance of this network still needs improvement.
Some researchers propose using dual-branch network models to address challenges such as low polyp edge contrast and diverse polyp morphologies. Cao et al. [24] addressed the significant differences in color, size, shape, appearance, and location of polyps by proposing a weighted dual-branch feature fusion network (WDFF-Net) based on HarDNet68. This network enhances the segmentation performance of small and edge-blurred polyps through two complementary branches: Progressive Feature Fusion and Scale-Aware Feature Fusion, while also incorporating an Object-Aware Attention Mechanism. WDFF-Net demonstrates excellent generalization ability and segmentation performance when handling polyps of various morphologies. However, its drawback is that the inference speed is slightly lower than that of lightweight models such as HarDNet-MSEG. Song et al. [78] addressed the issues of significant variations in the shape and size of polyps, as well as the high similarity of small polyps to their surrounding environment, by proposing a multi-scale parallel network based on an attention mechanism. This network utilizes a multi-scale fusion module to integrate high-level features and comprehensively extract global and local information through parallel attention mechanisms, while optimizing edge details using reverse context fusion. Its advantage lies in balancing features of different scales and significantly improving the segmentation performance of small polyps. However, the model’s performance still needs enhancement when faced with complex real-world scenarios, such as liquid occlusion.
To solve the problem of difficult precise segmentation of small polyps, Liu et al. [79] addressed the issue of insufficient boundary detail and global context information extraction in the segmentation of small polyps by proposing the FTMF-Net model. This model extracts fine boundary features from the frequency domain using a Fourier Transform (FT) module and enhances high-level semantic information on the encoder side through a Multi-Scale Feature Fusion module. This is the first application of the Fourier Transform in small polyp segmentation, significantly strengthening the ability to capture subtle boundaries. However, when faced with complex and difficult samples, the model’s segmentation performance still requires further improvement.

4.1.2. Limited Polyp Dataset

Due to limited annotated datasets, some researchers use semi-supervised models for polyp segmentation. To address the challenges posed by the reliance on large amounts of labeled data in fully supervised methods, as well as the variations in shape, size, and location of lesions, Gu et al. [80] proposed the semi-supervised segmentation framework DEC-Seg. This framework uses cross-layer feature aggregation, a scale-enhanced consistency strategy, and a dual-scale complementary fusion module to get the most out of the small amount of labeled data that is available. The model outperformed existing semi-supervised methods with only 10% and 30% labeled data. However, it still has certain limitations when dealing with ambiguous regions and small lesions in the background. Ren et al. [81] proposed the WS-DefSegNet framework, which combines weak supervision and semi-supervision to reduce the strong reliance on pixel-level annotations for polyp segmentation. This framework utilizes a newly constructed weakly labeled dataset, W-Polyp, and employs a sparse foreground loss function to suppress false positives during weakly supervised training while enhancing features with a deformable Transformer encoder. It can complete training with only 1.9% of pixel annotations, significantly reducing the annotation burden on physicians, and demonstrates good transferability. However, the mean Dice coefficient (mDice) score of approximately 0.768 on datasets like Kvasir still falls short of the fully supervised method PraNet, which achieves a score of 0.898.
Yu et al. [82] addressed the issues of overfitting in small-scale datasets and the high computational complexity of Transformer by proposing a novel non-pretrained deep supervision network called NPD-Net. To avoid overfitting, it uses a non-pretrained deep supervision strategy. The parallel dual attention module combines spatial and channel attention to make the computer work faster while still being able to model features. Compared to most advanced methods, NPD-Net has fewer parameters and can significantly reduce overfitting on small datasets. However, its segmentation performance still falls short when dealing with ambiguous images or those containing bubbles.
Guo et al. [83] proposed a method that combines meta-learning mixed data augmentation and confidence-aware resampling strategies to address the issues of limited training data, significant variations in polyps, and class imbalance. This method utilizes meta-learning to acquire clinical diagnostic knowledge from validation data, generating compatible hard labels for mixed images, and employs a progressive resampling strategy that moves from easy to difficult at both the image and pixel levels to balance the data distribution. It achieved good segmentation results in low-contrast and varying lighting conditions. However, its performance in extremely imbalanced scenarios has not been thoroughly discussed, and real-time performance has not been evaluated in detail.
Wang et al. [84] addressed the issue of high costs associated with dense annotations and the limitations of existing weakly supervised methods that are confined to spatial domain learning by proposing the S2ME segmentation framework based on doodle supervision. This framework performs feature learning in both spatial and spectral domains, generating reliable mixed pseudo-labels by combining an entropy-guided pseudo-label ensemble learning strategy with adaptive pixel-level fusion techniques. High-quality segmentation results can be achieved solely by relying on doodle annotations. However, this method still lacks comparisons with the latest approaches.

4.1.3. Generalization of Training Models and Domain Adaptation

To enhance the model’s generalization performance, Jia et al. [85] focus on the challenges of limited labeled data and class imbalance in polyp segmentation, proposing the PolypMixNet semi-supervised framework. This framework enhances the utilization of unlabeled data through polyp-aware augmentation techniques and dual consistency regularization under the Mean Teacher structure. It achieves near-full-supervised performance with only 15% labeled data but still performs worse than fully supervised methods on smaller datasets such as CVC-ColonDB and ETIS-Larib. Zhang et al. [86] addressed the issue of decreased polyp segmentation performance caused by domain transfer by proposing a domain generalization method based on random global illumination enhancement. This method utilizes an Image Decomposition Module to decompose images into three components: reflectance, local illumination, and global illumination. It then employs an Illumination Transformation Module to generate global illumination in the style of the target domain, thereby increasing data diversity. This method can effectively segment polyps under complex shapes and various lighting conditions. However, there are still shortcomings in handling other domain shift factors, such as color and shape differences.
Lu et al. [87] addressed the issue of decreased polyp segmentation performance caused by distribution differences between the source and target domains by proposing the DCL-PS cross-domain polyp segmentation framework. This framework achieves cross-domain feature alignment through Domain Contrastive Learning (DCL) and Contrastive Learning with Cross-Consistency Training and utilizes a Prototype-guided Self-training strategy to dynamically assign weights to pixels. It demonstrates excellent capability in cross-domain feature alignment. However, there are still shortcomings in capturing edge details of large polyps. Wang and Chen [88] proposed the RPANet unsupervised domain adaptation framework to address the issue of polyp segmentation transfer learning in scenarios where source domain data cannot be obtained. This method employs two modules: foreground-aware contrastive learning and confidence-calibrated pseudo-labeling, to achieve efficient adaptation of the model by self-supervised learning of target domain data features from coarse to fine. It does not require labeled data from the source domain, meeting the privacy and cross-hospital deployment needs of medical scenarios. However, there are still limitations in processing when the differences between foreground and background are not significant, and the issues of noise and errors accumulated in the pseudo-labels have not been fully resolved.

4.1.4. Real-Time Performance of the Model

To improve real-time performance, maintain a smaller model parameter count, and enhance computational efficiency, Li et al. [23] addressed the challenges of extracting global contextual features and the large model parameters that hinder clinical applications by proposing the ERDUnet segmentation network. This network achieves high precision and lightweight segmentation through modules such as the Context Enhanced Encoder and Differential Regional Attention, as well as a dual U-shaped structure. While maintaining a small parameter size and fast inference speed, ERDUnet can accurately handle large-scale targets. However, when the features of the target and background are too similar, the model is still prone to recognition errors.

4.2. Transformer-Based Methods

Transformer-based polyp segmentation methods primarily leverage the self-attention mechanism in Transformer to model global information, thereby better understanding the overall shape and context of polyps. Examples of such architectures include TransUNet and Swin-UNet. Transformers are excellent at capturing long-range dependencies, which can enhance the accuracy and stability of segmentation in complex scenarios.

4.2.1. Polyp Diversity and Boundary Factors

The model needs to have strong feature extraction capabilities to handle the morphological diversity of polyps and sensitivity to boundary information. We review Transformer-based models from these two perspectives.
  • Issues on Diversity in the Morphology, Size, Brightness, and Color of Polyps
To address the challenges of existing methods in capturing high-level semantics and dealing with the camouflage characteristics of polyps, Xiao et al. [28] proposed a novel polyp segmentation framework named CTNet. This framework trains a Transformer backbone network through supervised contrastive learning and incorporates a self-multi-scale interaction module and an information collection module to extract structured features and multi-scale information. Experimental results show that CTNet improves mDic and mean Intersection over Union (mIoU) by 1.5% and 1.6%, respectively, compared to the second-best method, ColonFormer-L. However, it has a high dependency on pixel-level annotated data, and there is still room for improvement in the segmentation performance of small target polyps. Wu et al. [89] addressed the problem of segmenting polyps that are morphologically diverse and difficult to distinguish from the surrounding mucosa by proposing the MSRAformer multi-scale spatial reverse attention network. This method uses a pyramid-structured Swin Transformer as the encoder and combines multi-scale channel attention modules with a spatial reverse attention mechanism to progressively enrich and refine the edge structure and details of the polyp region. The use of circular residual convolution blocks enhances feature learning capabilities and effectively prevents gradient vanishing. The model demonstrates excellent generalization ability in extracting and aggregating multi-scale features, but there is still room for improvement in segmenting tiny lesion areas. Liu et al. [90] proposed the MLFF-Net multi-layer feature fusion network to address the issues of insufficient feature utilization and semantic conflicts arising from feature fusion in existing methods. This network works collaboratively through a multi-scale attention module a high-level feature enhancement module, and a global attention module to integrate shallow details with deep semantic features. It has achieved high accuracy and robustness in segmenting small or irregularly shaped polyps. However, it still has certain limitations when handling the segmentation of multiple polyps. Xu et al. [91] came up with PSTNet, which combines RGB and frequency domain information to improve the accuracy of colon polyp segmentation. This was done because previous methods relied too heavily on RGB information and misaligned features, which made it hard to accurately identify polyps. This network includes a frequency feature attention module, a feature supplement alignment module, and a cross-perception localization module, which extract low-contrast polyp cues from the frequency domain and reduce misalignment interference during multi-scale feature fusion. Compared to existing methods, PSTNet demonstrates improved detection capabilities for small polyps and edge regions while effectively suppressing noise. However, there remains a need for improvements in computational efficiency. Yue et al. [92] addressed the challenges of polyp segmentation in complex backgrounds, varying shapes and sizes, and fuzzy boundaries by proposing the Progressive Feature Enhancement Network. This network employs a pyramid vision Transformer as the encoder and integrates three modules: Cross-Stage Feature Enhancement, Coarse Map Generation, and Foreground Boundary Collaborative Enhancement, enabling the network to progressively refine segmentation results from coarse to fine. By eliminating redundant information and enhancing multi-scale feature interaction, PFENet demonstrates a stronger feature representation capability in recognizing complex backgrounds and deformed polyps. However, when dealing with data from different domains, its generalization performance still requires further improvement. Dong et al. [93] addressed the limitations of CNN backbones in cross-layer feature fusion by proposing the Polyp-PVT framework, which utilizes a Pyramid Vision Transformer to obtain more robust feature representations instead of traditional CNNs. This method includes a Cascade Fusion Module, a Camouflage Identification Module, and a Similarity Aggregation Module, which are used for high-level semantic fusion, low-level camouflage feature identification, and cross-layer feature integration, respectively. It performs exceptionally well in tasks such as video polyp segmentation, particularly surpassing the existing method PNS-Net in feature extraction and generalization capabilities. However, the model’s detection of polyp boundaries still needs improvement in complex areas with overlapping light and shadow. Liu et al. [94] proposed the CAFE-Net segmentation network to address issues such as the loss of small-sized targets, difficulty in recovering fine-grained details, and insufficient multi-scale feature aggregation. This network compensates for missing details and explores potential features through the Feature Supplementation and Exploration Module, retains low-level information and recovers fine-grained features using the Cross-Attention Decoder Module, and maximizes the utilization of learned features with the Multi-Scale Feature Aggregation Module. It demonstrates excellent performance in handling small-sized polyps, complex lighting, and blurred boundaries. However, potential overfitting and computational efficiency require further improvements. Li et al. [95] proposed the Lesion-Aware Context Interaction Network to address the issues of diverse foreground shapes, complex background interference, and high computational costs. This network reduces background noise through a Lesion-Aware Pyramid Mechanism and enhances global feature representation by combining a Non-Local Context Lesion Interaction Module in the Robust Representation Enhanced Decoder and a Triplet-Branch Multi-Scale Perceptual Self-Attention Module. LACINet demonstrates excellent segmentation performance for diverse polyp morphologies while maintaining real-time capability at 28.41 fps. However, the segmentation accuracy for tiny polyps and under extreme lighting conditions still needs improvement.
2.
Problems of Difficult Boundary Segmentation
To solve the problem of difficult precise boundary segmentation caused by unclear polyp boundaries, Cai et al. [96] proposed VANet to address the diversity of polyp appearances and the issue of blurred boundaries caused by changes in the endoscopic viewpoint. This network locates polyp regions through a viewpoint classifier’s class activation map and enhances polyp feature representation and optimizes segmentation boundaries by combining viewpoint-aware Transformer and boundary-aware Transformer. The method guides attention by using viewpoint information to cut down on interference from areas that are not polyps and improves boundary segmentation for polyps even more through multi-scale predictions. VANet adapts well to changes in viewpoint and improves edge recognition accuracy. However, the model has a large number of parameters, and its performance significantly declines when handling very small or very large polyps.

4.2.2. Limited Polyp Dataset

Rahman et al. [97] proposed the PP-SAM perturbation prompt method to address the Segment Anything Model’s (SAM) reliance on a large amount of labeled data when fine-tuning on new data sources. This method enhances the model’s adaptability to imprecise prompts by introducing variable bounding box prompt perturbations during the fine-tuning process, only fine-tuning the image encoder and prompt encoder while freezing the mask decoder. On the Kvasir dataset, using bounding box prompt perturbations of 50 and 100 pixels improved the DICE scores by 20% and 37%, respectively, demonstrating excellent few-shot learning capabilities. However, when the sample size exceeded 50, the performance improvement was no longer significant.

4.2.3. Generalization of Training Models and Domain Adaptation

To address the generalization and detail-handling challenges of deep learning models in polyp segmentation tasks, Wang et al. [26] addressed the challenges of complex polyp morphology and model overfitting by proposing the SSFormer model to enhance generalization ability and detail processing. This model combines a pyramid Transformer encoder with a Progressive Local Decoder and enhances local detail processing capabilities through Local Emphasis and Stepwise Feature Aggregation modules. SSFormer demonstrates strong generalization ability and local feature processing capability across different polyp datasets through multi-stage feature aggregation and progressive feature fusion. However, the model’s performance on extreme or unseen medical images still needs validation, and it requires a long training cycle with high demands on the quality and quantity of training data.

4.2.4. Real-Time Performance of the Model

To improve real-time performance, maintain a smaller model parameter count, and enhance computational efficiency, Jha et al. [98] proposed the Transformer-based Residual Upsampling Network to meet the demand for automatic real-time polyp segmentation. This network effectively avoids over-segmentation and under-segmentation issues while maintaining a real-time processing speed of 47.07 FPS. However, its performance still shows a significant decline on out-of-distribution datasets, indicating certain limitations in generalization. Lin et al. [3] addressed the issues of high computational complexity and large parameter size in existing Vision Transformer (ViT) models for polyp segmentation by proposing a lightweight visual Transformer model called Polyp-LVT. This method significantly reduces the model’s computational complexity and parameter count by replacing the attention layer with a global max pooling layer in the encoder, and it incorporates the IFFM module in the decoder for efficient feature fusion. Polyp-LVT reduces the parameters by approximately 44% compared to the baseline model Polyp-PVT, with a computational requirement of only 13.21 Giga Floating Point Operations Per Second (GFlops). However, there is still room for further improvement in model accuracy.

4.2.5. Limitations of Architecture

Huang et al. [99] addressed the issue of feature information loss caused by patch segmentation in Transformer models by proposing the HST-MRF model, which integrates multi-scale receptive fields. This model combines the Heterogeneous Swin Transformer with a multi-modal bilinear pooling module to effectively integrate features from different receptive fields while also incorporating both low-level and high-level semantic information for more accurate lesion localization. This approach maintains high segmentation performance while reducing computational complexity and effectively minimizing the loss of structural information. However, the precision metric still shows shortcomings, particularly in the recognition of lesion boundaries, which requires further improvement.
Lin et al. [100] addressed the issue of the Transformer block division neglecting pixel-level intrinsic structural features by proposing DS-TransUNet, which incorporates a hierarchical Swin Transformer into the encoder and decoder of a U-shaped architecture. Through a dual-scale encoding mechanism, the model simultaneously captures coarse and fine-grained features and utilizes self-attention to model non-local dependencies and multi-scale contextual information. This strategy effectively enhances the semantic segmentation quality of medical images. However, when both encoding branches use the same patch size, the model’s performance still has certain limitations. Rahman and Marculescu [27] proposed the CASCADE attention decoder to better capture local and long-range dependencies between pixels in medical images based on the Transformer architecture. They suppressed background interference and enhanced detail representation through a cascade design of multi-scale features, attention gates, and convolutional attention modules. A multi-stage feature and loss aggregation framework further improved the accuracy and convergence speed of segmentation. However, the model still requires more experimental validation to demonstrate its robustness in terms of generalization capability.

4.3. Hybrid Architecture Methods

A polyp segmentation method that combines the CNN and Transformer leverages the CNN’s strength in capturing local features and the Transformer’s ability to model global features. Typically, CNNs are used to extract basic features in the early or middle stages of the network, and Transformer modules are integrated later or concurrently for global self-attention learning. This approach can significantly enhance the accuracy and robustness of segmentation.

4.3.1. Polyp Diversity and Boundary Factors

Blurred boundaries increase the difficulty of effective feature extraction, and challenges arising from the natural characteristics of polyps affect polyp segmentation performance. In this subsection, we will review how hybrid polyp segmentation methods based on CNNs and Transformer address these two issues.
  • Issues on Diversity in the Morphology, Size, Brightness, and Color of Polyps
To address the diversity in polyp shape, size, brightness, and color, Zhang et al. [21] proposed the parallel branch architecture TransFuse to simultaneously address the modeling of local spatial details by CNNs and global dependencies by Transformer. This method efficiently fuses multi-level features between parallel branches using an innovative BiFusion module that leverages self-attention and bilinear Hadamard products. While maintaining a low parameter count, TransFuse-S achieves an inference speed of 98.7 FPS. However, the study has not sufficiently validated its capability to handle tiny objects and extremely imbalanced datasets.
Liu et al. [101] proposed MVOA-Net to tackle intra-class confusion stemming from inconsistencies in polyp morphology, size, and location, as well as inter-class indistinguishability resulting from high similarity with surrounding tissues. This model uses a Geometric Orientation Transformer Encoder to pull out features from both horizontal and vertical views. It also uses a Point Affinity Module to improve boundary detection and a Cross-Attention Fusion Module to combine global and local data. In multi-target polyp segmentation, MVOA-Net does a great job in terms of accuracy and parameter efficiency. It has great geometric orientation and point perception skills. However, its computational efficiency still requires further improvement. The Region Self-Attention Enhanced Network was created by Yin et al. [102] to solve the problem of segmentation that arises from the high similarity between the edges of normal tissue and polyps. This network is built on Pyramid Vision Transformer v2 (PVTv2) and makes different feature maps using a dual-decoder structure. It does this by using the Region Self-Attention (RSA) module to obtain more accurate boundary information. The design of the dual decoders combined with RSA accommodates both small and normal-sized polyps, significantly improving segmentation performance. However, the model still faces certain limitations in effectiveness when dealing with complex samples that have reflections and shadows. Yang and Zhang [103] proposed PFD-Net to better capture local and global semantic information in medical images. This network combines the PVTv2 Transformer with a CNN encoder and utilizes fast Fourier convolution and deformable convolution modules. The network becomes better at showing complicated shapes and uneven edges. This leads to clearer target boundary predictions and richer semantic structure extraction. However, the model’s segmentation performance remains unstable in scenarios where the target and background are extremely similar or occluded. To fix the problems of multi-level semantic feature gaps and weak connections between spatial and channel dimensions, Wu et al. [104] suggested a hierarchical mixed architecture they called HD-Former. This model uses a Compression Bottleneck module to improve shallow features and find target areas. It also includes a Dual Cross-Attention Transformer module for flexible fusion of multi-scale features and a Broad Exploration Network for hierarchical dense contextual semantic capture. This approach preserves detailed information while bridging the differences between various semantic levels on a global scale. However, the increased depth of the backbone network significantly raises the computational load. To address the issue of diverse lesion types in medical images that are similar to surrounding tissues, Du et al. [105] proposed the SwinPA-Net. The Dense Multiplicative Feature Fusion module minimizes interference from shallow background noise, while the Local Pyramid Attention module combines global and local attention to guide the network’s focus on the lesion areas. This innovative design provides strong generalization capabilities, but the large number of model parameters still limits its real-time performance. Zhang et al. [106] suggested HSNet, which combines Transformer and a CNN, as a way to fix the problems of polyps changing sizes, shapes, and lighting, as well as the lack of attention to fine textures. It combines long-range and local features through the Cross-Semantic Attention module and the Hybrid Semantic Complement module and utilizes the Multi-Scale Prediction module to adaptively fuse outputs of different scales. This method takes into account both global appearance and detailed textures, significantly enhancing segmentation performance. However, the detection of tiny target polyps still requires further improvements.
Li et al. [107] proposed FMCA-Net to address issues such as inconsistencies within polyps and similarities with surrounding tissues in medical image segmentation. This network uses PVTv2 as the encoder and further explores the initially extracted features through a feature secondary reuse module, supplemented by a group of dilated convolution attention modules to enhance edge segmentation. This approach effectively utilizes early polyp features while improving the problem of blurred segmentation boundaries. However, there is still room for improvement in the model’s performance in scenarios with densely distributed or very small polyps.
Jiang et al. [108] addressed the issue of the overly complex structure of Transformer and the insufficient focus on decoders by proposing the MLP-CNN dual-path complementary network. This network replaces the traditional Transformer with a Multi-Layer Perceptron (MLP), using dual-path decoders to separately reconstruct global and local information. It employs a Dual-Path Complementary (DPC) module to fuse multi-layer features from both MLP and CNNs, ultimately merging the segmentation results through the segmentation mask feature fusion module. The DPC module performs well in multi-scale and fuzzy boundary scenarios while also being parameter-efficient and offering higher training and inference efficiency. However, in complex scenarios, the model still has a significant demand for large-scale data, and its generalization ability requires further validation. Wu et al. [109] addressed the issue of high computational costs and insufficient detail feature learning in traditional Vision Transformer for polyp segmentation by proposing META-Unet. This network introduces a multi-scale efficient Transformer attention mechanism within a U-shaped architecture, generating element-level attention maps through a dual-branch structure that captures features at different scales effectively. This design allows for simultaneous modeling of global context and local details, significantly improving the accuracy and efficiency of segmentation. However, the model’s segmentation performance remains suboptimal for very small, very large, or poorly defined boundary polyps. Sanderson and Matuszewski [110] proposed FCBFormer to address the issues of overfitting in deep learning for segmentation and insufficient resolution in Transformer. This network extracts key features through a transformer branch while utilizing a fully convolutional branch and its dense skip connections to ensure boundary accuracy. Its advantage lies in simultaneously leveraging Transformer for global feature extraction and the CNN for maintaining segmentation details. However, the model still has a strong dependence on training data.
2.
Problems of Difficult Boundary Segmentation
To solve the problem of difficult precise boundary segmentation caused by unclear polyp boundaries. Li et al. [111] addressed the issue of significant internal differences in polyps and the ambiguous boundaries with surrounding tissues by proposing a Cross-layer Information Fusion and Guidance Network. This network extracts multi-scale features through a Transformer encoder. On the CVC-ClinicDB dataset, it improved mDice by 3.9% and mIoU by 4.2% compared to PraNet, demonstrating greater robustness against interferences such as lighting reflections. However, this method did not discuss the limitations related to inference speed and resource consumption in practical applications in detail. Shao et al. [112] proposed the Polyper method to address the segmentation problem of small polyps and ambiguous boundaries. This method draws on clinical experience to extract boundaries and polyp regions using morphological operators. It employs a boundary-sensitive attention mechanism to feedback the features of the internal polyp region to the boundary area, enhancing segmentation accuracy. The method performs well in scenarios involving small polyps and ambiguous boundaries. However, its fixed boundary width design has limitations in handling false positives and false negatives, making it challenging to adequately address the diverse characteristics of polyps. To solve the problems of Transformer not being able to extract enough local features and the CNN having trouble modeling long-range relationships, Cai et al. [113] came up with PPFormer, which combines Transformer’s long-range modeling skills with the CNN’s local feature extraction skills. The PP-guided self-attention mechanism focuses on hard-to-segment areas and utilizes a local-to-global mechanism to obtain more local information while maintaining low memory usage, achieving more accurate polyp localization. This method significantly enhances the perception of polyp boundaries. However, the segmentation performance on small-sized polyps still requires further improvement.

4.3.2. Limited Polyp Dataset

To address the need for reducing the annotation cost of medical image segmentation while maintaining accuracy, Xie et al. [114] proposed SimTxtSeg, a weakly supervised method. This method utilizes a text-to-visual cue converter and a text-visual hybrid attention module to train the segmentation model using simple text prompts and pseudo-labels generated by SAM. Compared to traditional fully supervised methods, SimTxtSeg significantly reduces annotation costs and demonstrates the effectiveness of cross-modal feature fusion. However, the overall performance may suffer if the quality of the pseudo-masks generated by the SAM model is insufficient.

4.3.3. Generalization of Training Models and Domain Adaptation

Li et al. [115] proposed the ASPS method to address the issues of insufficient local detail capture and poor out-of-distribution performance in the SAM model for endoscopic images. This method integrates CNN and ViT encoder features through a cross-branch feature augmentation module, enhancing the capture of local details and high-frequency information. Meanwhile, the uncertainty-guided prediction regularization module reduces prediction bias and improves the generalization ability of out-of-domain segmentation. However, there is still room for further optimization of ASPS in the segmentation of small target polyps. Xiong et al. [116] proposed SAM2-UNet, aiming to create a universal segmentation framework that is applicable to both natural images and medical images. This method is based on the Hiera network of SAM2 as the encoder, incorporating an adapter for parameter-efficient fine-tuning, combined with a classic U-Net decoder. Its advantages include being more friendly to devices with limited memory while demonstrating excellent performance in medical image segmentation. However, its performance on specific datasets, such as ClinicDB, still lags behind some other advanced methods.

4.3.4. Real-Time Performance of the Model

Jha et al. [117] proposed TransNetR to address the shortcomings of existing automated segmentation methods in terms of real-time processing speed and generalization ability. The encoder is a pre-trained ResNet50, and the self-attention mechanism of the Residual Transformer module improves the decoder’s ability to learn features. It achieves high-precision segmentation on polyps of various sizes and shapes, with a processing speed of up to 54.60 FPS. However, the recall metric is slightly lower than that of UACANet on certain specific center data, indicating room for improvement in the identification of very small polyps.

4.4. Mamba-Based Methods

The polyp segmentation method based on Mamba primarily leverages Mamba’s automatic segmentation capabilities and optimization algorithms to achieve fast and accurate segmentation. By employing a selective State Space Model (SSM), it can handle long sequence dependencies with linear complexity, offering greater efficiency than Transformer.

4.4.1. Polyp Diversity and Boundary Factors

Some researchers have employed Mamba-based methods to address the diversity in polyp morphology and the issue of blurred boundaries. We will introduce these methods in this section.
  • Issues on Diversity in the Morphology, Size, Brightness, and Color of Polyps
Xie et al. [118] addressed the problem of segmenting polyps with diverse morphologies and unclear boundaries by proposing the Prompt-Mamba method, which combines Vision-Mamba with box-prompt. This model consists of three components: a lightweight image encoder, a prompt encoder, and a mask decoder. It introduces bidirectional sequence modeling and a reverse SSM branch to better capture image features. The use of box-prompt enhances the model’s stability and generalization, while the combination of Focal Loss and Dice Loss improves segmentation quality. However, its performance on certain datasets is currently slightly lower than that of existing state-of-the-art methods, indicating that there is still room for improvement in generalization capability. Zhang et al. [119] addressed the issues of insufficient long-range dependency modeling in CNNs and the high computational complexity of Transformer by proposing the VM-UNetV2 based on SSM. This model introduces a Visual State Space (VSS) block to capture global contextual information and employs a semantic detail integration module to enhance cross-layer feature fusion. Compared to traditional CNNs and Transformer, VM-UNetV2 makes breakthroughs in both long-range dependencies and computational complexity while also improving segmentation performance through deep supervision. However, its robustness and generalization capabilities still require further empirical validation. Xu et al. [29] addressed the limitations of existing medical image segmentation models in long-range dependencies and cross-scale feature interactions by proposing the Polyp-Mamba framework. By introducing the Scale-Aware Semantic module and the Global Semantic Injection module, this model effectively captures multi-scale semantic information and enhances local features while maintaining linear computational complexity. It demonstrates excellent segmentation capabilities for polyps of various sizes, shapes, and textures. However, the model requires a high number of training iterations, and its interpretability still needs further improvement. Dutta et al. [120] proposed the SAM-Mamba architecture, aimed at addressing issues such as the variability in polyp shapes, ambiguous boundaries, and difficulties in generalization on unseen datasets. The SAM-Mamba method, based on SAM, achieves efficient polyp segmentation by introducing the Mamba-Prior module into the architecture and employing a multi-scale spatial decomposition and two-stage training strategy. Its advantages include excellent zero-shot generalization capability, multi-scale feature capture, and stable performance even under complex boundaries and textures. The main drawbacks are that edge detection accuracy still needs improvement, there may be misjudgments in complex scenes, and there is the computational burden brought by high-resolution inputs of SAM.
2.
Problems of Difficult Boundary Segmentation
Zhu et al. [121] came up with the Hybrid Multi-Frequency Perception Gated Selection Network Polyp-Mamba to deal with problems like unclear polyp boundaries and a significant number of similarities with nearby tissues. This model extracts global and local features through a dual multi-frequency fusion encoder combining Mamba and ResNet and enhances the boundaries of ambiguous regions using multi-frequency spectrum analysis via Discrete Cosine Transform (DCT). In the decoding phase, the gated selection decoder can suppress irrelevant areas and improve segmentation accuracy, while deep supervision further ensures effective alignment between features and labels. Although this method performs excellently across multiple datasets and approaches real-time processing, it may misjudge in certain scenarios with clear boundaries, and DCT still has limitations in capturing details within complex structures.

4.4.2. Limitations of Architecture

Wu et al. [122] addressed the challenges of CNNs in processing long sequence information and the low sensitivity of Vision Transformer to local features, as well as the drawback of the SS2D module introducing redundant information while maintaining a global receptive field. They proposed a high-order visual Mamba UNet based on an SSM. By utilizing a high-order two-dimensional selective scanning module, they reduced redundant information while preserving the advantages of a global receptive field, in conjunction with a Local-SS2D module to enhance local feature representation. This method integrates H-SS2D into a high-order visual state space module, achieving efficient medical image segmentation. However, the model currently lacks a systematic analysis and validation of its robustness.

4.5. Other Architectural Methods

There are also models that do not belong to the above four categories of architecture, such as architectures based on diffusion probabilistic models, models based on the Broad Learning System, and the Segment Anything Model. The emergence of these models has also expanded research in the field of polyp segmentation.

4.5.1. Polyp Diversity and Boundary Factors

Some researchers have used diffusion probability models and the general segmentation model SAM to address the diversity of polyps and the issue of blurred boundaries in polyp segmentation.
  • Issues on Diversity in the Morphology, Size, Brightness, and Color of Polyps
Du et al. [123] addressed the challenges of diverse polyp morphology, data scarcity, and poor contrast with surrounding tissues by proposing the Highlighted Diffusion Model (HDM) and the HDM+ framework. This method uses a two-step training process. first, it trains the HDM to highlight polyp regions while suppressing the background, and then uses the HDM features as prior knowledge for polyp segmentation, achieving feature fusion through a cross-attention mechanism. This design reduces the domain gap between RGB images and segmentation masks while also enhancing segmentation accuracy and efficiency. However, there are still certain challenges in segmenting very small or poorly defined polyps.
2.
Problems of Difficult Boundary Segmentation
Wang et al. [124] focused on improving the efficiency and accuracy of colorectal polyp segmentation, proposing an end-to-end segmentation method based on a Denoising Diffusion Probabilistic Model (DDPM). This method uses U-Net as the backbone, cascading the original image with a noise mask to learn their intrinsic relationship, and generates more precise segmentation results through multiple resampling and majority voting strategies. This is the first application of DDPM in the field of polyp segmentation, effectively handling images of polyps with blurred boundaries. However, the model’s generalization performance still requires further validation and enhancement.

4.5.2. Limited Polyp Dataset

Mansoori et al. [125] aimed to reduce reliance on large-scale labeled data and explored the zero-shot learning capability of the Segment Anything Model 2 (SAM 2) in colorectal polyp segmentation. The model achieved excellent segmentation results on six public datasets without additional training, particularly showing a significant improvement in accuracy with bounding box prompts. In the video polyp segmentation task, SAM 2 demonstrated a notable enhancement compared to the original SAM model, with a 31.4% increase in mIoU on the PolypGen dataset. However, the model’s performance is highly dependent on the type of prompts, with significant differences in effectiveness between different prompting methods.

4.5.3. Real-Time Performance of the Model

Banik et al. [126] addressed the issue of deep learning requiring a large number of parameters and training time for polyp segmentation by proposing the Deep Hierarchical Broad Learning System Network. This model enhances segmentation accuracy and efficiency through two multi-scale feature extraction techniques: 2D-DT-CWT and Sifp-conv, and introduces diligent parameter control to stabilize network training. Its flat network structure eliminates the need for multiple layers to learn complex features, resulting in superior performance in terms of computational cost. However, there are still certain limitations in segmentation effectiveness for morphologically complex and relatively flat polyps.

4.6. Video Polyp Segmentation

Some researchers have also conducted studies on video polyp segmentation. Ji et al. [22] addressed the issue of the lack of fine-grained segmentation annotations in large-scale datasets by constructing the video polyp segmentation dataset SUN-SEG and proposing the baseline model PNS+. This model utilizes global and local encoders to extract long-term and short-term spatiotemporal information and then progressively refines the segmentation results through a normalized self-attention block. This method achieves a real-time inference speed of 170 fps but lacks robustness when dealing with interference factors such as surgical instruments and optical glare. Wu et al. [127] addressed challenges such as the diversity of polyp types, low contrast, and blurred boundaries while also considering real-time requirements by proposing a lightweight, context-aware network called PolypSeg+. This network employs three key components: Adaptive Scale Context for gathering multi-scale information, Efficient Global Context for reducing background noise and improving feature fusion, and a Feature Pyramid Fusion module that effectively utilizes the extracted contextual features. It achieves high segmentation accuracy while ensuring real-time performance. However, the segmentation results still need improvement when dealing with tiny polyps and extremely low contrast scenarios. The FlowICBNet model was created by Wan et al. [128] to solve the problem of polyp segmentation that is hard to perform in endoscopic videos because the camera shakes and the frames become out of focus. This model uses the Iterative Feedback Unit (IFU) to help with video polyp segmentation and adds Reference Frame Selection (RFS) and Flow-Guided Warping (FGW) modules to use past frame prediction results to help segment the current frame better. This approach demonstrates excellent performance in alleviating the difficulties posed by camera shake and frame defocus. However, further optimization of the IFU’s iterative feedback mechanism is still necessary due to its potential for "over-correction". Lu et al. [129] came up with the Diff-VPS video polyp segmentation network, which is based on a multi-task diffusion model and solves the problems of high camouflage and redundant temporal cues in video polyp segmentation. This method uses a multi-task supervision strategy to add classification and detection tasks to the diffusion model. It also creates a Temporal Reasoning Module (TRM) to obtain more dynamic information from video frame sequences. An adversarial self-supervised strategy improves the accuracy of frame reconstruction, which makes it easier to find polyps that are well hidden. However, only the SUN-SEG dataset has validated this method, necessitating further research to explore its generalization to other datasets and more complex scenarios. Chen et al. [130] addressed the challenge of modeling long-range spatiotemporal relationships in colonoscopy videos by proposing the Mixture-Attention Siamese Transformer (MAST) method. This approach effectively captures and models the long-range spatiotemporal relationships between video frames through a Siamese Transformer architecture and a mixed attention mechanism. It achieved excellent segmentation results on the SUN-SEG dataset. However, due to its validation on a single dataset, its generalization capability still requires further evaluation. To fix the problem that traditional polyp segmentation algorithms do not work as well when switching from static images to video scenes, Xu et al. [131] came up with SSTFB, an end-to-end video polyp segmentation scheme that combines self-supervised learning with spatiotemporal self-attention. This method innovatively employs self-supervised learning as an auxiliary task, combining it with a spatiotemporal self-attention mechanism and feature branching techniques to enhance feature representation capabilities through a joint loss optimization strategy. The method achieved significant improvements on the SUN-SEG dataset and realized real-time inference at 126 frames per second. However, the model may still experience some degree of mis-segmentation under extreme lighting conditions, severe artifacts, and complex background environments. Hu et al. [132] proposed the SALI network to address the issue of limited segmentation accuracy in endoscopic videos caused by significant changes between adjacent frames and consecutive low-quality frames. This method utilizes a short-term alignment module to align features of adjacent frames through deformable convolution and designs a long-term interaction module that combines a memory bank and mask attention to reconstruct more reliable polyp features. This design effectively enhances the stability of spatiotemporal features and improves the handling of significant changes and low-quality frames. However, memory cost constraints limit the memory length to 35 frames, leading to inaccurate segmentation in scenarios with extremely low quality or drastic changes. Yang et al. [133] addressed the challenge of modeling long-term spatiotemporal dependencies in medical video segmentation by proposing the Vivim framework based on Video Vision Mamba. This method captures and compresses long-range spatial and temporal dependencies in videos through the Temporal Mamba Block and ST-Mamba module and introduces boundary-aware affine constraints to reduce prediction uncertainty in ambiguous lesion areas. Vivim achieves efficient long-sequence modeling while maintaining linear computational complexity. However, the overall complexity of the model is relatively high, necessitating a balance between operational efficiency and accuracy.

4.7. Remarks

This section summarizes the model methods outlined in this review. Table 2 is a summary of polyp segmentation methods from 2022 to 2024; Table 3 summarizes polyp segmentation methods from 2018 to 2021, and Table 4 summarizes video polyp segmentation methods from 2022 to 2024. In this section, we will summarize the contents of these three tables, discussing their common advantages, common problems, and potential solutions. We will briefly outline the issues and solutions in this section; for a detailed discussion on the ongoing challenges and future trends, please refer to Section 8 and Section 9.
Table 2 summarizes the research progress on polyp segmentation methods from 2022 to 2024, primarily including CNN, Transformer, hybrid architectures, Mamba-based methods, and five other categories. Among these, CNN-based methods mainly focus on feature extraction and optimizing boundary details, but they still have limitations in complex scenarios; Transformer-based methods excel at capturing long-range dependencies but come with high computational costs and perform poorly on small target segmentation; hybrid architectures attempt to combine the local feature extraction capabilities of CNNs with the global modeling capabilities of Transformer to achieve a balance between accuracy and efficiency; meanwhile, emerging Mamba-based methods show potential in long sequence modeling but still require further validation of their robustness. Most of these methods are based on fully supervised strategies and utilize various backbone networks (such as Res2Net, PVTv2, Swin Transformer, etc.). Some works emphasize efficient training in semi-supervised or weakly supervised environments while also exploring scenarios like cross-domain adaptation, real-time segmentation, and few-shot learning. Multi-scale feature fusion and attention mechanisms are widely used to improve the precise localization of morphologically diverse, small-sized, or low-contrast polyp regions. Most methods have been validated on the CVC-ClinicDB and Kvasir-SEG datasets, achieving favorable results. The common advantages of these methods include the following: effective handling of multi-scale features, strong feature extraction and representation capabilities; improved segmentation accuracy through various feature fusion strategies; and some methods achieving good real-time performance. However, these methods also face several common challenges: suboptimal segmentation performance for small polyps; decreased performance in complex scenarios with low contrast and blurred boundaries; inadequate results for simultaneous segmentation of multiple polyps; the need to balance computational efficiency and real-time performance; and limited generalization performance on out-of-domain datasets. Possible solutions to these issues include first developing more effective multi-scale feature extraction and fusion strategies, with a particular focus on feature representation for small targets. Second, introducing stronger boundary-aware mechanisms and contrastive learning strategies. Additionally, designing lightweight yet efficient network structures to balance performance and computational costs. Finally, exploring semi-supervised and weakly supervised methods to reduce annotation dependence and enhancing the model’s domain adaptation capabilities to improve robustness in different scenarios. We anticipate significant advancements in polyp segmentation technology in the future due to these efforts.
Table 3 summarizes various polyp segmentation methods from 2018 to 2021, including TransFuse, PraNet, ResUNet++, and UNet++. These methods are primarily based on convolutional neural network architectures, such as ResNet and Res2Net, and employ fully supervised learning approaches. Each method presents innovative solutions to specific challenges in polyp segmentation: TransFuse integrates multi-layer features through parallel branches and a BiFusion module; PraNet introduces partial decoders and a reverse attention module; ResUNet++ enhances segmentation of polyps of different shapes and sizes using ASPP; and UNet++ redesigns the skip connection paths. A common advantage of these methods is their ability to effectively capture multi-scale features of polyps, improving segmentation accuracy and addressing issues such as diverse polyp shapes and unclear edges to some extent. However, these methods also face shared challenges, including high computational complexity, poor segmentation performance for very tiny targets, and inaccuracies in segmenting complex boundaries. Some ways to solve these problems are to make network architectures that are lighter, come up with better ways to extract and combine features, try out self-supervised or semi-supervised learning methods, use attention mechanisms to improve how features are represented, and make training datasets more diverse and larger. Additionally, integrating the latest deep learning technologies, such as state space models, is expected to further enhance the performance and generalization ability of polyp segmentation. Future research could focus more on model lightweighting, real-time performance, and robustness to extreme conditions, driving the advancement of medical image segmentation technology through interdisciplinary innovative approaches.
Table 4 summarizes various deep learning methods used for video polyp segmentation from 2022 to 2024. These methods include CNN-based PNS+, Mamba-based Vivim, hybrid architecture MAST, FlowICBNet, and SALI, among others. These models primarily utilize backbone networks such as Res2Net and PVTv2, employing either fully supervised or self-supervised learning approaches. By introducing innovative technologies such as spatiotemporal self-attention mechanisms, the Temporal Mamba Block, the Siamese Transformer, and hybrid attention modules, these methods have made significant progress in improving segmentation accuracy and processing efficiency. A common advantage of these methods is their ability to effectively capture the spatiotemporal relationships between video frames, enhancing the accuracy of polyp feature extraction and addressing issues such as the lack of large-scale datasets, low-quality frames, and spatial variations to some extent. However, these methods face common challenges, including high computational complexity, poor segmentation performance for small and highly concealed polyps, and limited model generalization capabilities. Possible solutions to these issues may include introducing more advanced attention mechanisms, such as multi-scale and cross-frame attention; exploring lighter and more efficient network architectures, such as Mamba state space models; developing more robust data augmentation strategies; leveraging self-supervised learning and contrastive learning techniques to enhance the model’s feature representation capabilities; constructing larger and more diverse datasets; and exploring multimodal and cross-modal learning methods. Through continuous innovation and research, it is believed that these challenges can be gradually addressed, leading to more precise video polyp segmentation technology.

5. Polyp Segmentation Loss Function

In polyp segmentation, class imbalance is a pervasive challenge: polyp regions typically occupy only approximately 5–15% of the image, leaving background pixels overwhelmingly dominant. This skew biases models toward the background class and diminishes sensitivity to small lesions. Consequently, careful selection of the loss function is critical to mitigate imbalance-induced bias and improve overall detection and segmentation performance.
The main loss functions are as follows:
  • Weighted Binary Cross-Entropy Loss;
L B C E = α × y × log p ( 1 α ) × ( 1 y ) × l o g ( 1 p )
Here, y denotes the ground truth label, p represents the probability predicted by the model, and α is the weighting factor for the foreground class (polyp), which is typically set between 0.6 and 0.8.
2.
Dice Loss;
Dice loss directly optimizes the overlap between the predicted and ground-truth masks. By emphasizing relative overlap rather than absolute pixel counts, it is inherently more robust to class imbalance.
L D i c e = 1 2 i = 1 N P i G i i = 1 N P i + i = 1 N G i
Pi is the probability that the i-th pixel belongs to the segmentation region, and Gi is the ground truth value of the i-th pixel.
3.
Focal Loss;
L F o c a l = α × 1 p γ × y × log p ( 1 α ) × P γ × ( 1 y ) × l o g ( 1 p )
Here, y represents the true label, P denotes the predicted probability, and α is the class balancing factor, which is typically set to 0.25. By using the modulating factor (1 − p)γ, the weights of easily classified samples are automatically reduced, allowing the model to focus more on hard-to-classify samples. The parameter γ controls the extent of this weighting adjustment between easy and hard samples (usually γ = 2), resulting in excellent performance in challenging regions such as polyp boundaries.
4.
Tversky Loss;
L T v e r s k y = 1 T P T P + α × F P + β × F N
TP denotes true positives, FP denotes false positives, and FN denotes false negatives.
As a generalized form of the Dice loss, the balance between false positives and false negatives can be controlled by adjusting the α and β parameters. When a higher recall is required (for example, to avoid missing small polyps), set β > α.
Single loss functions often struggle to optimize all metrics simultaneously, making combined loss functions a common approach.
L c o m b i n e d = λ 1 × L B C E + λ 2 × L D i c e + λ 3 × L F o c a l
Here, λ1, λ2, and λ3 represent the combination weights for each loss function, with λ1 + λ2 + λ3 = 1. LBCE, LDice, and LFocal denote the respective calculated loss values.

6. Polyp Segmentation Datasets

There are five commonly used image datasets for polyp segmentation, namely CVC-ColonDB [44], ETIS-Larib [139], CVC-ClinicDB [140], CVC-EndoSceneStill [141], and Kvasir-SEG [142]. Some images from CVC-ClinicDB [140] and Kvasir-SEG [142] are commonly used as training data. Figure 7 displays the progress of the polyp segmentation dataset. Table 5 provides detailed information on the related polyp dataset.

6.1. Datasets of Polyp Image

The CVC-Colon DB [44] was created by Bernal et al. in 2012 at the Computer Vision Center and Computer Science Department of Universitat Autònoma de Barcelona situated in Barcelona, Spain. The dataset includes 300 images and their corresponding polyp masks were obtained from 13 polyp video sequences acquired from 13 patients. The ETIS-Larib database [139] was developed in 2014 at the ETIS Laboratory of CY Cergy Paris University. This dataset includes 196 annotated samples extracted from 34 colonoscopy videos. The CVC-ClinicDB database [140] was created by Bernal et al. in 2015 in collaboration with a hospital clinic in Barcelona, Spain. The database includes 612 images, which were derived from 31 polyp video sequences collected from 23 patients. CVC-EndoSceneStill [141] is a dataset formed in 2017 by merging CVC-ColonDB and CVC-ClinicDB. It comprises 912 images derived from 44 video sequences obtained from 36 patients. Kvasir-SEG [142] contains 1000 images, annotated by expert scientists from Oslo University Hospital in 2020. The images and videos in Hyper-Kvasir [143] were collected from actual gastroscopy and colonoscopy tests at Norway’s Baerum Hospital from 2008 to 2016 and were partially annotated by experienced gastroenterologists. It includes 11.62 h of video and 1,059,519 video frames. This dataset was published in 2020. The Piccolo dataset [144] contains 3433 clinical colonoscopy video images, including both white light and narrow-band imaging images, which were obtained from colonoscopy procedures performed on human patients. It includes 76 different lesions from 48 patients. This dataset was published in 2020. The BKAI-IGH [145] dataset was published in 2021 and contains 1200 images. The PolypGen dataset [146] was collected from six different centers across Europe and Africa. It contains 1537 polyp images, 2225 positive video sequences, and 4275 negative frames. A total of 3762 positive frames and 4275 negative frames are provided. This dataset was published in 2023.
Table 5. Comparative analysis of polyp datasets in existing research studies.
Table 5. Comparative analysis of polyp datasets in existing research studies.
Datasets of Polyp Image/VideoDatasetYearNumber
of Images
Number
of Video
Sequences
Resolution
Datasets of polyp imageCVC-ColonDB [44]201230013574 × 500
ETIS-Larib [139]2014196341225 × 966
CVC-ClinicDB [140]201561231384 × 288
CVC-EndoSceneStill [141]201791244384 × 288 to 574 × 500
Kvasir-SEG [142]20201000N/A332 × 487 to 1920 × 1072
PICCOLO [144]2020343340854 × 480 to 1920 × 1080
Hyper-Kvasir [143]2020110,079374332 × 487 to 1920 × 1072
BKAI-IGH [145]20211200N/A1280 × 959
PolypGen [146]20238037N/A384 × 288 to 1920 × 1080
Datasets of polyp videoASU-Mayo [61]201636,45838688 × 550
LDPolypVideo [147]202140,266160560 × 480
Kvasir-Capsule [148]20214,741,504117336×336
SUN-SEG2022158,69010131158 ×1008 to 1240 × 1080

6.2. Datasets of Polyp Video

The ASU-Mayo dataset [61] contains 36,458 consecutive frames from 38 videos and provides 3856 labels for 10 frontal videos. This data was published in 2016. LDPolypVideo [147] includes 160 colonoscopy videos and 40,266 frames. This data was published in 2021. Smedsrud et al. in 2021 introduced the Kvasir-Capsule dataset [148], which includes 47,238 labeled frames and 4,694,266 unlabeled frames. It comprises 117 videos, from which a total of 4,741,504 image frames can be extracted. The colonoscopy videos in SUN-SEG [22] come from the Showa University and Nagoya University database (also known as the SUN database [149]). It contains 158,690 colonoscopy video frames. Various types of annotations are provided, including attributes, object masks, boundaries, scribbles, and polygons. This data was published in 2022.

7. Polyp Segmentation Evaluation Metrics

The metrics for evaluating polyp segmentation performance include Dice Similarity Coefficient (DSC), Intersection over Union (IoU), Recall (R), Average Precision (AP), Precision (P), Pixel Accuracy, and Frames Per Second (FPS) for assessing real-time performance. Among these, the key metrics are mean Dice and mean IoU.
Dice [150,151,152,153] is used for pixel-by-pixel comparison of predicted segmentation results with ground truth data. Its description is as follows:
D i c e A , B = 2 × | A B | A + | B | = 2 × T P 2 × T P + F P + F N
where A represents the predicted segmentation mask and B represents the ground truth segmentation mask. TP, FP, TN, and FN stand for true positives, false positives, true negatives, and false negatives, respectively. The definitions of TP, FP, TN, and FN are shown in Table 6.
IoU [154] is the number of overlapping pixels between the target mask and the predicted mask divided by the total number of pixels present in both masks. It is represented by the mathematical Formula (7).
I o U A , B = A B A B = T P T P + F P + F N
Pixel accuracy [155] is calculated by dividing the number of correct predictions by the total number of predictions. True positives refer to pixels accurately predicted to belong to a specific category, and true negatives refer to pixels identified as not belonging to that category. It is represented by the mathematical Formula (8).
P i x e l   a c c u r a c y = T P + T N T P + T N + F P + F N
Precision is the number of correctly classified positive image samples divided by the number of image samples classified as positive. It is represented by the mathematical Formula (9).
P r e c i s i o n = T P T P + F P
Recall is the number of image samples correctly classified as positive divided by the total number of positive image samples. It is represented by the mathematical Formula (10).
R e c a l l = T P T P + F N

8. Polyp Segmentation Model Performance Evaluation

In this section, we selected 44 models from those mentioned in Section 4 to evaluate their segmentation performance and real-time capabilities. Table 7 displays the performance evaluation results of the above 44 models.
From Table 7, DDPM [124] and ISCNet [74] rank top two on the CVC-ClinicDB dataset, with DDPM being 0.6% and 1.7% higher than ISCNet in mDice and mIoU, respectively; Polyper [112], ISCNet [74], and Polyp-Mamba [29] performed outstandingly on the Kvasir-SEG dataset, with Polyper having the highest mDice and mIoU at 0.948 and 0.904; Polyper [112], TGANet [136], and SAM2 showed excellent results on the CVCColonDB dataset, with SAM2 achieving the highest mDice and mIoU at 0.934 and 0.877; Polyper [112] and SAM2 [125] ranked second and first on the ETIS-Larib dataset, with SAM2 being 7.6% and 10.5% higher than Polyper in mDice and mIoU, respectively; Polyper [112], TGANet [136], and Polyp-Mamba [29] performed exceptionally on the EndoScene dataset, with Polyper having the highest mDice at 92.4% and TGANet having the highest mIoU at 89.9%.
We selected ten models from Table 7 and compared performance evaluation metrics on four datasets: CVC-ClinicDB, Kvasir-SEG, CVC-ColonDB, and ETIS-Larib. In comparison to other models, ISCNet and Polyper demonstrate excellent segmentation performance on the CVC-ClinicDB and Kvasir-SEG datasets. On the CVC-ColonDB and ETIS-Larib datasets, SAM2 achieves the best segmentation performance.
Figure 8 presents a comparative analysis of the performance of ten polyp segmentation models across four benchmark datasets. On the CVC-ClinicDB dataset, the ISCNet model performed exceptionally well, achieving mDice and mIoU scores of 96.1% and 92%, respectively. However, it was still slightly lower than the DDPM model by 0.6% and 1.7%. On the Kvasir-SEG dataset, Polyper demonstrated the best performance, followed by Polyp-Mamba. SAM2 excelled on the CVC-ColonDB dataset, achieving the highest mDice of 93.4% and the highest mIoU of 87.7%. SAM2 also achieved outstanding performance on the ETIS-Larib dataset, with mDice and mIoU scores of 94.1% and 89%, respectively.
Based on Table 7 and Table 8, we selected ten models to analyze the relationship between the segmentation performance metrics and real-time capabilities. As shown in Figure 9, a higher mDice and mIoU segmentation metric indicates better segmentation performance for the model. For models with comparable segmentation performance, a higher FPS real-time metric, along with a moderate number of model parameters, suggests a greater potential for clinical application of that model.
The size of the parameters is an important indicator of model complexity. As shown in Table 8, the number of parameters in the models ranges from 2.54M to 287.75M. Smaller models, such as PolypSeg+ with 2.54M parameters and ERDUNet with 10.21M parameters, have an advantage in terms of parameter count and may perform better in environments with limited memory and computational resources. Larger models, like HST-MRF with 174.73M parameters and DS-TransUNet with 287.75M parameters, may be better at feature extraction and segmentation accuracy, even though they have more parameters. Therefore, when selecting a model, it is necessary to strike a balance between performance and resource consumption.
Flops is another important metric for evaluating the computational complexity of a model. As shown in Table 8, the range of Flops varies from 7.23G (PolypSeg+) to 55.36G (CFA-Net). A higher Flops value indicates that the model requires more computational resources during inference, which may affect real-time performance. For applications that require quick responses, it may be more appropriate to choose a model with lower Flops.
FPS is a direct indicator of a model’s real-time performance. As shown in Table 8, the FPS ranges from 2.06 (I-RA) to 126.00 (SSTFB). A high FPS value indicates that the model can process more frames in less time, making it suitable for real-time monitoring and diagnostic applications.
When selecting a polyp segmentation model, it is important to balance parameter size, computational complexity, and real-time performance based on specific application needs. Future research could further explore how to optimize a model’s real-time performance while maintaining high segmentation accuracy to meet the demands of clinical applications.
The distribution in Figure 9 shows the performance of different models used in polyp segmentation, where the mDice value represents the segmentation performance of polyps and FPS indicates the real-time processing capability of the model. Among the 10 models surveyed, TGANet performed excellently, with moderate parameter count, and achieved the highest mDice and FPS values. In clinical applications, the model’s parameter count cannot be too large, and it must maintain a balance between real-time processing speed and segmentation capability.

9. Open Research Challenges

Polyp segmentation can assist medical professionals in analyzing and diagnosing polyps. Over the past decade, deep learning-based polyp segmentation technology has rapidly advanced. Although current segmentation models perform well, there is still room for improvement in segmenting extremely challenging images that exhibit blurred boundaries, low contrast, or very small polyp sizes, and there are many challenges to successfully applying this technology in clinical settings.

9.1. Lack of Large-Scale Datasets

High-quality polyp segmentation annotation data requires the involvement of specialized doctors. However, the annotation process is often time-consuming, labor-intensive, and costly, which limits the scale and quality of training data for models. Additionally, different doctors may have discrepancies in annotating polyps, leading to inconsistencies in the training data. Currently, existing datasets have insufficient sample sizes and lack diversity, which may not cover various types, shapes, and sizes of polyps. In real clinical environments, doctors may observe multiple types of polyps in the same image. However, many datasets frequently feature images that highlight only a single type of polyp. For example, the current ETIS-LaribPolypDB, CVC-ClinicDB, and Kvasir-SEG datasets primarily consist of images of single polyps. Multiple polyps can appear simultaneously in endoscopic images, exhibiting diverse characteristics in terms of size (e.g., <5 mm vs. 10–15 mm), color (e.g., red, pink, light brown), and shape (e.g., flat, mushroom-like, stalked). Furthermore, an endoscopic image may simultaneously feature adenomatous polyps and hyperplastic polyps. However, the limited number of images depicting multiple polyps in existing datasets hinders models from concurrently learning how to handle different types of polyps during training. Due to the lack of diverse samples, models may struggle to accurately identify and segment different types of polyps in practical applications, thereby affecting the accuracy of clinical decision-making.

9.2. Image Quality and Noise

Various factors, particularly changes in lighting, endoscopic angles, and interference from intestinal contents, affect the image quality in the polyp segmentation dataset. During endoscopic examinations, the intensity and angle of the light source may vary. For instance, when the endoscope is close to a polyp, the light may be too intense, causing excessive reflection from the polyp’s surface and creating highlights; conversely, when the endoscope is farther away, insufficient light may lead to an overall dark image with lost details. These lighting variations can cause the boundaries of the polyp to become blurred, making accurate segmentation difficult. Additionally, changes in the endoscopic angle can also impact image clarity. When the endoscope enters the intestine at different angles, it may distort the shape and position of the polyp, especially in curved sections of the intestine. This change in perspective can make the polyp’s morphology irregular, increasing the difficulty of segmentation. Furthermore, liquids, gases, or other contents within the intestine may interfere with the clarity of the endoscopic images. For example, liquid in the intestine can cause light scattering, resulting in a blurred image. These contents may obscure the polyp, preventing segmentation algorithms from accurately identifying the polyp’s boundaries. In existing datasets, some images in the Kvasir-SEG dataset exhibit uneven lighting, certain images in the CVC-ClinicDB dataset show perspective distortion, and the ETIS-LaribPolypDB dataset contains images with occluded contents. For instance, intestinal contents partially obscure a polyp in one endoscopic image, and due to insufficient lighting, the polyp’s color contrast with the surrounding tissue is low, resulting in unclear boundaries. In such cases, the segmentation model may mistakenly classify the polyp as background or fail to accurately identify its shape and size. In summary, these factors collectively contribute to the decline in image quality within the polyp segmentation dataset, which in turn affects the training and testing performance of models. Therefore, improving image acquisition techniques, enhancing image processing algorithms, and constructing higher-quality datasets are key to increasing the accuracy of polyp segmentation.

9.3. Complex Shapes and Boundaries

The differences in the shape, size, and color of polyps have a significant impact on the polyp segmentation task. Firstly, variations in the shape of polyps can affect the difficulty of edge detection. For example, flat polyps typically present as flat surfaces with blurred edges, nearly level with the intestinal wall, which may confuse them with surrounding tissues, making it challenging for segmentation algorithms to accurately identify them. In contrast, mushroom-shaped polyps have a distinct stalk and umbrella-like top with clear edges, while stalk-type polyps have a slender stalk and a smaller top, giving them a more unique shape. These shape differences may require different segmentation strategies, thereby increasing the complexity of the model. Secondly, differences in the size of polyps can also affect segmentation outcomes. The surrounding tissues may obscure polyps smaller than 5 mm, making them easily overlooked, whereas polyps measuring 10-15 mm are relatively easier to identify. Polyps larger than 20 mm are usually easier to segment but may overlap with other structures, leading to identification difficulties. Noise or other tissues in images may mask small polyps, causing segmentation algorithms to miss them, while larger polyps may obscure smaller surrounding polyps, thereby affecting the segmentation accuracy. Additionally, differences in the color of polyps can similarly influence segmentation results. Red polyps, due to their rich vascularity and bright color, are simple to identify; however, pink polyps may be difficult to distinguish from surrounding normal tissues due to their similar color. Light brown polyps may appear darker due to tissue degeneration or inflammation. These color differences can lead to insufficient contrast and color distortion, making it hard to differentiate polyps of similar colors from surrounding tissues, thus reducing the accuracy of segmentation algorithms. Variations in lighting or endoscopic reflections may also result in the loss of color information, further impacting segmentation performance. In the same image, there may be polyps of different shapes, sizes, and colors simultaneously, which increases the complexity of the segmentation algorithm and its generalization capability. Segmentation algorithms need to possess multi-scale perception abilities to handle various types of polyps, which adds to the difficulty of model training.

9.4. Insufficient Generalization Ability

In the polyp segmentation task, the model’s generalization ability is a key factor. There are significant differences between different datasets; for example, the Kvasir-SEG dataset contains endoscopic images, while the CVC-ClinicDB dataset consists of clinically collected images. The distribution of polyp types and image resolution reflect these differences, resulting in variations in dataset characteristics. For instance, a model trained on the Kvasir-SEG dataset shows a significant drop in segmentation performance when tested on the CVC-ClinicDB dataset, especially in recognizing small or irregularly shaped polyps. Also, models trained on the CVC-ClinicDB dataset tend to focus on certain types of polyps, like mushroom-shaped ones. This means that, when tested on the ETIS-LaribPolypDB dataset, they perform less well, especially when dealing with polyps that are very different in color and shape. Training and testing on different endoscopic devices can also lead to a decrease in segmentation accuracy. Insufficient generalization ability can result in inaccurate edge detection and difficulties in recognizing small polyps. The model may struggle to accurately identify the edges of polyps, leading to blurred segmentation results. In the training dataset, there are fewer samples of small polyps, which further weakens the model’s ability to recognize them. The main reasons for insufficient generalization ability include data distribution shifts and limitations in feature learning. Training a model on a specific dataset may prevent it from learning the features of polyps with different shapes and colors, which could impact its performance in real-world applications.

9.5. High Real-Time Requirements

In polyp segmentation models, scenarios with high real-time requirements pose a severe challenge to computational efficiency. During colonoscopy, doctors need to observe and analyze the condition of polyps in real time to make timely decisions. Therefore, the segmentation model must be capable of processing images and providing segmentation results within a few seconds. Additionally, automated endoscopic systems need to analyze video streams in real-time to quickly identify and label polyps, thereby assisting doctors in diagnosis. To improve processing speed, it may be necessary to simplify the model, which could lead to a decrease in segmentation accuracy and an increased risk of misdiagnosis. During endoscopic examinations, images change rapidly, and insufficient real-time processing capability may prevent the model from accurately capturing the dynamic features of polyps. Thus, polyp segmentation models face an important challenge of maintaining segmentation accuracy while ensuring real-time performance.

9.6. Lack of Interpretability

Doctors often view deep learning models as “black boxes”, lacking interpretability, making it challenging to comprehend the models’ decision-making process. This “black box” issue limits clinical trust, as the internal decision-making process is opaque, and doctors cannot clearly comprehend how the model arrives at a particular segmentation result. For instance, in an endoscopic image, the model may label certain areas as polyps, while these areas do not align with the characteristics of polyps based on the doctor’s experience. The lack of interpretability leads to a decrease in doctors’ trust in the model, creating barriers to clinical application. During critical decision-making, doctors may become skeptical of the model’s judgments and thus be reluctant to rely on its results, which will affect the model’s practical application. If doctors do not trust the model, it may lead to the oversight of potential polyps, ultimately impacting patient health. Therefore, the interpretability of polyp segmentation models is crucial for clinical application. Enhancing the model’s transparency and interpretability can strengthen doctors’ trust, thereby improving the model’s effectiveness in real medical settings. Therefore, the interpretability of polyp segmentation models is crucial for clinical application. Recent research has made significant progress in addressing this challenge. Wickstrøm et al. [156] developed methods combining uncertainty estimation with interpretability techniques for colorectal polyp segmentation, using Monte Carlo Dropout for uncertainty quantification and Guided Backpropagation to visualize which input features the model deems important for predictions. Their findings revealed that models primarily utilize shape and edge information of polyps, and importantly, erroneous predictions exhibit significantly higher uncertainty levels, providing clinicians with valuable confidence indicators. Furthermore, Ramos and Hortúa [157] advanced this field by implementing Bayesian neural networks with multiplicative normalized flows, achieving well-calibrated uncertainty estimates while maintaining high segmentation performance, demonstrating that interpretability and uncertainty quantification can be effectively integrated without compromising accuracy. Enhancing the model’s transparency and interpretability can strengthen doctors’ trust, thereby improving the model’s effectiveness in real medical settings.

9.7. Ethics and Privacy

Due to ethical and privacy concerns, the use of medical data is subject to strict limitations [158,159,160]. Sensitive personal information often resides in medical data, necessitating protection and preventing its use for research or commercial purposes without patient consent [161,162]. Therefore, medical research typically requires review by an ethics committee to ensure that the design and implementation of the study do not harm patients or infringe on their rights [163]. This process can lead to delays and complexities in data acquisition. Ethical and privacy issues directly impact the training and performance of deep learning models for polyp segmentation. Researchers may not be able to obtain sufficient labeled data to train deep learning models [164]; a lack of data can prevent the model from learning enough features, thereby affecting its performance. With limited training data, the model may overfit to the training set, resulting in poor generalization to new data [165]. This means that the model may not effectively identify and segment polyps in practical applications. Additionally, if the available dataset lacks diversity, the model may struggle to adapt to various clinical scenarios, reducing its effectiveness in real-world applications [166].

10. Future Research Directions

Based on the challenges mentioned in the previous chapter, we explore potential research directions for addressing these challenges in this section.

10.1. Develop Larger-Scale and More Diverse Datasets

During the annotation process of the polyp dataset, a multi-expert annotation mechanism was introduced, along with the use of ensemble learning methods and the exploration of automated annotation techniques, aiming to improve the quality of the dataset annotations and address the potential discrepancies that may arise from different doctors annotating polyp images. Variations in doctors’ experiences, knowledge backgrounds, or personal judgments can lead to inconsistent annotation results for the same polyp. By employing multi-expert annotations, collective intelligence can be harnessed to reduce this subjectivity. Additionally, using statistical methods to assess the consistency of annotations among different experts helps identify and analyze the reasons for annotation discrepancies. Experts can combine automated annotation techniques like deep learning with manual review to further enhance annotation quality. Automated annotation systems can continuously learn and optimize based on expert feedback, gradually reducing annotation differences. This comprehensive approach not only improves the accuracy and consistency of annotations but also enhances the quality of the dataset, providing a more reliable foundation for subsequent deep learning model training.
To address the issues of insufficient sample size and lack of diversity in existing datasets, it is particularly important to establish a larger-scale dataset that includes diverse polyp images. On one hand, for the acquired blurry images or blurred video frames, data augmentation methods can be employed to improve image quality. For example, image enhancement techniques can be applied to unclear polyp images to enhance image quality and optimize polyp feature extraction, improving the distinction between polyps and the background, and highlighting the edges and texture features of the polyps. The EndoUIC method proposed by Bai et al. [167] aims to achieve better imaging results; Shang et al. [168] proposed a video deblurring method to recover continuous clear frames from blurry videos; ELKarazle et al. [169] addressed the issue of specular reflection areas in colonoscopy images to improve the readability of medical images; Sharma et al. [170] improved polyp synthesis by controlling the size, shape, and position of polyps. On the other hand, it is also crucial to develop and expand the scale and diversity of the dataset, including images of various polyp categories. For instance, the ColonINST [171] dataset proposed by Ji et al. is a multi-modal instruction-tuning dataset that contains 62 categories and 303,001 colonoscopy images, which can be used for segmentation tasks. These efforts will contribute to enhancing the performance and applicability of polyp segmentation models.

10.2. Exploring and Improving Deep Learning Architectures

Refining deep learning architectures and incorporating attention mechanisms can enhance the performance of polyp segmentation models in addressing issues such as small-sized polyps, low contrast, and blurred boundaries. For example, using frequency feature attention modules [91] and parallel attention mechanisms [80] can help the model focus on important features, especially when it comes to small polyps. Additionally, utilizing multi-scale feature extraction methods through a local and global dual-branch structure can efficiently capture features at different scales, combined with multi-scale fusion modules to capture the details of small-sized polyps, thereby enhancing the model’s ability to recognize polyps of varying sizes.
Integrating contextual information is also an effective strategy for improving model performance. For example, [75] use a global guided context aggregation module that can help the model understand the relationship between polyps and their surrounding background, thus improving segmentation results for low-contrast images. To address color variations, specialized modules can be designed, such as dataset-level color augmentation [70], to enhance the color diversity of training data, thereby improving feature extraction capabilities for low-contrast images. Furthermore, designing shape feature extraction modules, such as [121] using DCT for multi-spectral analysis, can enhance the boundaries of blurred areas.
In terms of boundary handling, a variable Laplacian feature refinement module [66] can capture high-frequency information of boundaries, aiding the model in better recognizing the geometric features of polyps, especially when dealing with blurred boundaries. By optimizing segmentation strategies, the accuracy and robustness of segmentation can be further improved. For instance, considering ensemble learning methods that combine predictions from multiple models through voting or weighted averaging, as demonstrated in a strategy involving multiple resampling and majority voting [124], can generate more precise segmentation results. Additionally, selecting or designing appropriate loss functions, such as joint loss optimization strategies [131], can enhance feature representation, thereby improving segmentation accuracy for small-sized polyps.
Prompt-based architectures, such as fine-tuning SAM [172], may also prove to be an effective strategy. Rahman et al. [97] proposed a robust fine-tuning technique for segmentation tasks involving multiple polyps. Future research could develop a self-prompting SAM to automatically generate high-quality prompts, thereby further enhancing segmentation quality. The comprehensive application of these methods will contribute to improving the overall performance of polyp segmentation models.

10.3. Weak Supervised

Through weakly supervised and unsupervised learning, polyp segmentation models can be effectively trained using only a small amount of labeled data. Traditional, fully supervised learning methods require a large number of labeled images for training, and most current models still rely on comprehensive supervised learning. However, due to the high cost of labeled data, weakly supervised and semi-supervised learning techniques provide an effective alternative. These techniques can utilize a small amount of labeled data along with a large amount of unlabeled data for model training and generate pseudo-labels for the unlabeled data. These pseudo-labels can serve as additional training data, helping the model learn more features and thereby improving segmentation performance. For example, Hu et al. [173] introduced monotonicity constraints into weakly supervised segmentation, proposing a new method for handling imprecise annotations. It is anticipated that there will be more innovative research in the fields of weakly supervised, unsupervised learning, and hybrid supervision strategies. This research will provide new ideas and methods for training polyp segmentation models, further advancing the development of this field.

10.4. Combine with Prior Knowledge

Combining polyp segmentation models with doctors’ prior knowledge—specifically incorporating their expertise and experience during the model’s training and inference processes—can effectively utilize the morphological features and positional information of polyps to guide segmentation. This approach not only enhances the model’s focus on boundary region features but also improves segmentation accuracy and robustness. For instance, doctors often leverage the inherent characteristics of the internal regions of polyps in clinical practice to address boundary ambiguity. To this end, Polyper [112] utilizes features from the internal polyp regions to enhance boundary recognition, resulting in favorable segmentation outcomes. Additionally, bounding box annotations provide relatively accurate semantic and positional information, serving as valuable prior knowledge. W-PolypBox [174] employs the prior information from bounding boxes to enhance the training process for polyp segmentation. This method fully leverages doctors’ expertise, making the model more effective in handling complex medical images, thereby improving polyp segmentation results. By integrating doctors’ experience with deep learning models, the performance of polyp segmentation can be further enhanced, providing more reliable support for clinical applications.

10.5. Domain Adaptation

Domain adaptation is a transfer learning technique aimed at improving model performance in target domains, especially when labeled data in the target domain is scarce.
This method enables the training of models in a source domain, typically rich in labeled data, and then transfers their knowledge to the target domain. The advantages of domain adaptation include reducing the need for labeled data in the target domain, enhancing the model’s generalization ability, and increasing the model’s robustness, all of which can significantly improve the performance of polyp segmentation models. Researchers have proposed various domain adaptation methods [87,175,176,177,178] to address the performance decline caused by domain bias. For example, Pan et al. [179] utilized a frequency-based Federated Domain Generalization framework to tackle the domain generalization problem, while Li et al. [180] compensated for domain gaps through continual learning and multi-scale domain adaptation. By aligning the features of the source and target domains and reducing the distribution differences between them, the model’s performance in the target domain is enhanced, thereby improving the accuracy and reliability of polyp segmentation. These studies indicate that domain adaptation techniques hold significant potential for enhancing applications in medical image analysis, particularly in data-scarce situations.

10.6. Develop Lightweight Models

To improve the real-time performance of polyp segmentation, developing lightweight models is particularly necessary. In clinical applications, doctors need to quickly obtain polyp segmentation results to make timely diagnostic and treatment decisions. Lightweight models can process images in a shorter time, thus meeting the requirements for real-time performance. To achieve this, methods such as model compression, network architecture optimization, feature selection, and the use of lightweight convolutions can be considered to design efficient lightweight models. For example, Polyp-LVT [3] replaces the attention layer with a global max pooling layer in the encoder, significantly reducing the model’s computational complexity and parameter count. This type of lightweight model not only improves processing speed but also maintains high segmentation accuracy, thereby providing a better user experience and decision support in clinical applications. These improvements enhance the application potential of lightweight models in real medical environments.

10.7. Interpretability

The interpretability of polyp segmentation models refers to the ability of the model to clearly explain its decision-making process and output results during polyp segmentation. This interpretability is particularly important in medical image analysis, as doctors and clinical experts need to understand the basis of the model’s judgments in order to trust its results and make corresponding clinical decisions. To enhance the interpretability of the model, various methods can be employed. First, visualization techniques, such as heatmaps, can be used to show the areas of the image that the model focuses on when making segmentation decisions.
Additionally, using model architectures that are inherently more interpretable for preliminary analysis can help in understanding the decision-making process of the model. Recent advances have shown promising directions in this area. Wickstrøm et al. [156] proposed Monte Carlo Guided Backpropagation, a novel method that not only identifies important input features but also quantifies the uncertainty associated with these feature importances, enabling clinicians to understand both what the model is looking at and how confident it is about these features. This dual insight is particularly valuable when models encounter challenging cases such as small or poorly visible polyps. Moreover, Ramos and Hortúa [157] demonstrated that Bayesian approaches using multiplicative normalized flows can provide well-calibrated predictions with interpretable uncertainty maps, showing higher uncertainty at polyp boundaries and in difficult-to-detect regions, which aligns with clinical intuition and enhances trust in model decisions. Incorporating clinical explanations into the model’s outputs, which align with the clinical experience of doctors, further enhances interpretability. At the same time, continuous evaluation and improvement of the model’s interpretability through feedback from doctors is essential to ensure its effectiveness and reliability in practical applications. In [181], it has been summarized that employing uncertainty quantification methods to assess the reliability of machine learning models can improve the interpretability and acceptability of model results. In summary, these methods can effectively enhance the interpretability of polyp segmentation models, thereby promoting their application in clinical practice.

10.8. Federated Learning

Federated learning [182,183,184] is a distributed machine learning method designed to enhance model performance while protecting user privacy. This approach allows multiple devices, such as hospitals and medical institutions, to train models locally without the need to centralize data on a single server. Each device uses local data for training and sends the resulting model updates to a central server, which aggregates these updates to form a global model. In federated learning, the transmission of model updates employs encryption techniques, such as homomorphic encryption or differential privacy, to protect the content of the updates, ensuring that users’ original data remains on local devices, thereby mitigating the risk of data leakage. Furthermore, federated learning can collect data from various medical institutions, where patient populations in different hospitals may exhibit different types and characteristics of polyps. This diversity allows the model to learn more comprehensive features, thereby enhancing its generalization ability. Stelter et al. [185] analyzed the impact of federated learning on polyp segmentation, noting that this method effectively protects patient data privacy. However, the study also mentioned that the controlled noise introduced by differential privacy during training might reduce the model’s accuracy. To address this, the research suggests that future work could utilize Generative Adversarial Networks to generate privacy-preserving data. By increasing the training data for the model while safeguarding data privacy, the segmentation performance and generalization ability of the polyp segmentation model can be further improved.
Polyp segmentation remains a challenging research area that requires the integration of various technical approaches to achieve breakthroughs. Future research trends will place greater emphasis on the robustness, real-time performance, and interpretability of models to promote the widespread application of polyp segmentation technology in clinical settings.

11. Conclusions

This article reviews polyp segmentation models based on deep learning methods and the Mamba method, providing a comprehensive overview of research published between 2018 and 2024. First, the research on polyp segmentation models is categorized from the perspective of network architecture and the problems addressed by the models. Second, a comparative analysis of different polyp segmentation models is conducted, with performance evaluation including segmentation performance metrics, model parameters, and real-time capabilities. Third, we summarize commonly used datasets in polyp segmentation for researchers’ reference. Finally, we discuss the main challenges faced in the field of polyp segmentation and future directions for addressing these challenges.

Author Contributions

Conceptualization, X.L.; Methodology, X.L. and N.A.M.I.; software, X.L.; validation, X.L. and F.L.; formal analysis, X.L. and C.C.; investigation, X.L. and C.C.; resources, X.L. and F.L.; data curation, X.L.; writing—original draft preparation, X.L.; writing—review and editing, X.L., C.C., F.L. and N.A.M.I.; visualization, X.L. and C.C.; supervision, N.A.M.I.; project administration, X.L. and N.A.M.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets referenced and analyzed in this review are publicly available from their original sources: CVC-ColonDB: https://www.kaggle.com/datasets/longvil/cvc-colondb (accessed on 1 December 2024.); ETIS-Larib: https://drive.google.com/drive/folders/10QXjxBJqCf7PAXqbDvoceWmZ-qF07tFi (accessed on 1 December 2024.); CVC-ClinicDB: https://polyp.grand-challenge.org/CVCClinicDB/ (accessed on 1 December 2024.); CVC-EndoSceneStill: https://drive.google.com/file/d/1MuO2SbGgOL_jdBu3ffSf92feBtj8pbnw/view (accessed on 1 December 2024.); Kvasir-SEG: https://datasets.simula.no/kvasir-seg/ (accessed on 1 December 2024.); PICCOLO: https://www.biobancovasco.bioef.eus/en/Sample-and-data-catalog/Databases/PD178-PICCOLO-EN.html (accessed on 1 December 2024.); Hyper-Kvasir: https://datasets.simula.no/hyper-kvasir/ (accessed on 1 December 2024.); BKAI-IGH: https://www.kaggle.com/competitions/bkai-igh-neopolyp/data (accessed on 1 December 2024.); PolypGen: https://www.synapse.org/Synapse:syn26376615/files/ (accessed on 1 December 2024.); LDPolypVideo: https://github.com/dashishi/LDPolypVideo-Benchmark (accessed on 1 December 2024.); Kvasir-Capsule: https://osf.io/dv2ag/ (accessed on 1 December 2024.); SUN-SEG: https://github.com/GewelsJI/VPS/blob/main/docs/DATA_PREPARATION.md (accessed on 1 December 2024.).

Acknowledgments

We thank the authors of the original studies included in this review for making their research publicly available. We also acknowledge the developers and maintainers of the public datasets referenced in this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APAverage Precision
BSCABit-Slice Context Attention
BWMBalanced Weight Module
CMAMConvolutional Multi-Scale Attention Module
CNNConvolutional Neural Network
CRCColorectal Cancer
DCLDomain Contrastive Learning
DCTDiscrete Cosine Transform
DDPMDenoising Diffusion Probabilistic Model
DLCADataset-Level Color Augmentation
DPCDual-Path Complementary
DSCDice Similarity Coefficient
FCNsFully Convolutional Networks
FGWFlow-Guided Warping
FoBSFocus on Boundary Segmentation
FPSFrames Per Second
FTFourier Transform
GFlopsGiga Floating Point Operations Per Second
IFUIterative Feedback Unit
IoUIntersection over Union
MACEMulti-Path Attention Encoder
MASTMixture-Attention Siamese Transformer
MCADMulti-Path Attention Decoder
mDicemean Dice coefficient
mIoUmean Intersection over Union
MLPMulti-Layer Perceptron
PAMParallel Attention Module
PVTv2Pyramid Vision Transformer v2
RFSReference Frame Selection
RSARegion Self-Attention
SAMSegment Anything Model
SAM2Segment Anything Model 2
SSBUSegmentation-Squeeze-Bottleneck Union
SSFMSelective Shared Fusion Module
SSMState Space Model
TRMTemporal Reasoning Module
ViTVision Transformer
VSSVisual State Space
WoSWeb of Science

References

  1. Siegel, R.L.; Miller, K.D.; Wagle, N.S.; Jemal, A. Cancer statistics, 2023. CA Cancer J. Clin. 2023, 73, 17–48. [Google Scholar] [CrossRef]
  2. Haggar, F.A.; Boushey, R.P. Colorectal Cancer Epidemiology: Incidence, Mortality, Survival, and Risk Factors. Clin. Colon Rectal Surg. 2009, 22, 191–197. [Google Scholar] [CrossRef]
  3. Lin, L.; Lv, G.; Wang, B.; Xu, C.; Liu, J. Polyp-LVT: Polyp segmentation with lightweight vision transformers. Knowl.-Based Syst. 2024, 300, 112181. [Google Scholar] [CrossRef]
  4. Mamonov, A.V.; Figueiredo, I.N.; Figueiredo, P.N.; Richard Tsai, Y.-H. Automated Polyp Detection in Colon Capsule Endoscopy. IEEE Trans. Med. Imaging 2014, 33, 1488–1502. [Google Scholar] [CrossRef]
  5. Gross, S.; Kennel, M.; Stehle, T.; Wulff, J.; Tischendorf, J.; Trautwein, C.; Aach, T. Polyp Segmentation in NBI Colonoscopy. In Proceedings of the Bildverarbeitung für die Medizin 2009, Heidelberg, Germany, 21–24 March 2009; Meinzer, H.-P., Deserno, T.M., Handels, H., Tolxdorff, T., Eds.; Springer: Berlin, Heidelberg, 2009; pp. 252–256. [Google Scholar]
  6. Jerebko, A.K.; Teerlink, S.; Franaszek, M.; Summers, R.M. Polyp segmentation method for CT colonography computer-aided detection. In Proceedings of the Medical Imaging 2003, San Diego, CA, USA, 12–20 February 2003; Clough, A.V., Amini, A.A., Eds.; SPIE: Bellingham, WA, USA, 2003; p. 359. [Google Scholar]
  7. Yao, J.; Miller, M.; Franaszek, M.; Summers, R.M. Colonic polyp segmentation in CT colonography-based on fuzzy clustering and deformable models. IEEE Trans. Med. Imaging 2004, 23, 1344–1352. [Google Scholar] [CrossRef]
  8. Lu, L.; Barbu, A.; Wolf, M.; Liang, J.; Salganicoff, M.; Comaniciu, D. Accurate polyp segmentation for 3D CT colongraphy using multi-staged probabilistic binary learning and compositional model. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  9. Bernal, J.; Núñez, J.M.; Sánchez, F.J.; Vilariño, F. Polyp Segmentation Method in Colonoscopy Videos by Means of MSA-DOVA Energy Maps Calculation. In Clinical Image-Based Procedures. Translational Research in Medical Imaging; Linguraru, M.G., Oyarzun Laura, C., Shekhar, R., Wesarg, S., González Ballester, M.Á., Drechsler, K., Sato, Y., Erdt, M., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2014; Volume 8680, pp. 41–49. ISBN 978-3-319-13908-1. [Google Scholar]
  10. Hwang, S.; Celebi, M.E. Polyp detection in Wireless Capsule Endoscopy videos based on image segmentation and geometric feature. In Proceedings of the 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 678–681. [Google Scholar]
  11. Ganz, M.; Yang, X.; Slabaugh, G. Automatic Segmentation of Polyps in Colonoscopic Narrow-Band Imaging Data. IEEE Trans. Biomed. Eng. 2012, 59, 2144–2151. [Google Scholar] [CrossRef]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  13. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2023. [Google Scholar] [CrossRef]
  14. Gu, A.; Dao, T. Mamba: Linear-Time Sequence Modeling with Selective State Spaces. arXiv 2024. [Google Scholar] [CrossRef]
  15. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  16. Manjunath, K.N.; Siddalingaswamy, P.C.; Gopalakrishna Prabhu, K. An Improved Method of Colon Segmentation in Computed Tomography Colonography Images Using Domain Knowledge. J. Med. Imaging Health Inform. 2016, 6, 916–924. [Google Scholar] [CrossRef]
  17. Brandao, P.; Mazomenos, E.; Ciuti, G.; Caliò, R.; Bianchi, F.; Menciassi, A.; Dario, P.; Koulaouzidis, A.; Arezzo, A.; Stoyanov, D. Fully convolutional neural networks for polyp segmentation in colonoscopy. In Proceedings of the Medical Imaging 2017: Computer-Aided Diagnosis, Orlando, FL, USA, 13–16 February 2017; SPIE: Bellingham, WA, USA, 2017; pp. 101–107. [Google Scholar]
  18. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R.S., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  19. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; Lange, T.D.; Halvorsen, P.; Johansen, H.D. ResUNet++: An Advanced Architecture for Medical Image Segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar]
  20. Fan, D.-P.; Ji, G.-P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. PraNet: Parallel Reverse Attention Network for Polyp Segmentation. arXiv 2020, arXiv:2006.11392. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Liu, H.; Hu, Q. TransFuse: Fusing Transformers and CNNs for Medical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Strasbourg, France, 27 September–1 October 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 14–24. [Google Scholar]
  22. Ji, G.-P.; Xiao, G.; Chou, Y.-C.; Fan, D.-P.; Zhao, K.; Chen, G.; Gool, L.V. Video Polyp Segmentation: A Deep Learning Perspective. Mach. Intell. Res. 2022, 19, 531–549. [Google Scholar] [CrossRef]
  23. Li, H.; Zhai, D.-H.; Xia, Y. ERDUnet: An Efficient Residual Double-coding Unet for Medical Image Segmentation. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 2093–2096. [Google Scholar] [CrossRef]
  24. Cao, J.; Wang, X.; Qu, Z.; Zhuo, L.; Li, X.; Zhang, H.; Yang, Y.; Wei, W. WDFF-Net: Weighted Dual-Branch Feature Fusion Network for Polyp Segmentation with Object-Aware Attention Mechanism. IEEE J. Biomed. Health Inform. 2024, 28, 4118–4131. [Google Scholar] [CrossRef] [PubMed]
  25. Li, S.; Sui, X.; Fu, J.; Fu, H.; Luo, X.; Feng, Y.; Xu, X.; Liu, Y.; Ting, D.S.W.; Goh, R.S.M. Few-Shot Domain Adaptation with Polymorphic Transformers. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Strasbourg, France, 27 September–1 October 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 330–340. [Google Scholar]
  26. Wang, J.; Huang, Q.; Tang, F.; Meng, J.; Su, J.; Song, S. Stepwise Feature Fusion: Local Guides Global. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; pp. 110–120. [Google Scholar]
  27. Rahman, M.M.; Marculescu, R. Medical Image Segmentation via Cascaded Attention Decoding. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 6211–6220. [Google Scholar]
  28. Xiao, B.; Hu, J.; Li, W.; Pun, C.-M.; Bi, X. CTNet: Contrastive Transformer Network for Polyp Segmentation. IEEE Trans. Cybern. 2024, 54, 5040–5053. [Google Scholar] [CrossRef]
  29. Xu, Z.; Tang, F.; Chen, Z.; Zhou, Z.; Wu, W.; Yang, Y.; Liang, Y.; Jiang, J.; Cai, X.; Su, J. Polyp-Mamba: Polyp Segmentation with Visual Mamba. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2024, Marrakesh, Morocco, 6–10 October 2024; Linguraru, M.G., Dou, Q., Feragen, A., Giannarou, S., Glocker, B., Lekadir, K., Schnabel, J.A., Eds.; Lecture Notes in Computer Science. Springer Nature: Cham, Switzerland, 2024; Volume 15008, pp. 510–521, ISBN 978-3-031-72110-6. [Google Scholar]
  30. Sánchez-Peralta, L.F.; Bote-Curiel, L.; Picón, A.; Sánchez-Margallo, F.M.; Pagador, J.B. Deep learning to find colorectal polyps in colonoscopy: A systematic literature review. Artif. Intell. Med. 2020, 108, 101923. [Google Scholar] [CrossRef]
  31. Mei, J.; Zhou, T.; Huang, K.; Zhang, Y.; Zhou, Y.; Wu, Y.; Fu, H. A survey on deep learning for polyp segmentation: Techniques, challenges and future trends. Vis. Intell. 2025, 3, 1. [Google Scholar] [CrossRef]
  32. Wu, Z.; Lv, F.; Chen, C.; Hao, A.; Li, S. Colorectal Polyp Segmentation in the Deep Learning Era: A Comprehensive Survey. arXiv 2024. [Google Scholar] [CrossRef]
  33. Gupta, M.; Mishra, A. A systematic review of deep learning based image segmentation to detect polyp. Artif. Intell. Rev. 2024, 57, 7. [Google Scholar] [CrossRef]
  34. Xia, Q.; Zheng, H.; Zou, H.; Luo, D.; Tang, H.; Li, L.; Jiang, B. A comprehensive review of deep learning for medical image segmentation. Neurocomputing 2025, 613, 128740. [Google Scholar] [CrossRef]
  35. Li, S.; Ren, Y.; Yu, Y.; Jiang, Q.; He, X.; Li, H. A survey of deep learning algorithms for colorectal polyp segmentation. Neurocomputing 2025, 614, 128767. [Google Scholar] [CrossRef]
  36. Nemani, P.; Vadali, V.S.S.; Medi, P.R.; Marisetty, A.; Vollala, S.; Kumar, S. Cross-modal hybrid architectures for gastrointestinal tract image analysis: A systematic review and futuristic applications. Image Vis. Comput. 2024, 148, 105068. [Google Scholar] [CrossRef]
  37. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
  38. Pickhardt, P.J.; Kim, D.H.; Pooler, B.D.; Hinshaw, J.L.; Barlow, D.; Jensen, D.; Reichelderfer, M.; Cash, B.D. Assessment of volumetric growth rates of small colorectal polyps with CT colonography: A longitudinal study of natural history. Lancet Oncol. 2013, 14, 711–720. [Google Scholar] [CrossRef]
  39. Kudo, S.; Hirota, S.; Nakajima, T.; Hosobe, S.; Kusaka, H.; Kobayashi, T.; Himori, M.; Yagyuu, A. Colorectal tumours and pit pattern. J. Clin. Pathol. 1994, 47, 880–885. [Google Scholar] [CrossRef] [PubMed]
  40. Putniković, D.; Jevtić, J.; Ristić, N.; Milovanovich, I.D.; Đuknić, M.; Radusinović, M.; Popovac, N.; Đorđić, I.; Leković, Z.; Janković, R.M. Endoscopic and Histopathological Findings in the Upper Gastrointestinal Tract in Pediatric Chron’s Disease. preprints 2024. [Google Scholar] [CrossRef]
  41. Winawer, S.J.; Zauber, A.G.; Ho, M.N.; O’Brien, M.J.; Gottlieb, L.S.; Sternberg, S.S.; Waye, J.D.; Schapiro, M.; Bond, J.H.; Panish, J.F. Prevention of colorectal cancer by colonoscopic polypectomy. The National Polyp Study Workgroup. N. Engl. J. Med. 1993, 329, 1977–1981. [Google Scholar] [CrossRef] [PubMed]
  42. Lieberman, D.; Moravec, M.; Holub, J.; Michaels, L.; Eisen, G. Polyp Size and Advanced Histology in Patients Undergoing Colonoscopy Screening: Implications for CT Colonography. Gastroenterology 2008, 135, 1100–1105. [Google Scholar] [CrossRef]
  43. Snover, D.C. Update on the serrated pathway to colorectal carcinoma. Hum. Pathol. 2011, 42, 1–10. [Google Scholar] [CrossRef]
  44. Bernal, J.; Sánchez, J.; Vilariño, F. Towards automatic polyp detection with a polyp appearance model. Pattern Recognit. 2012, 45, 3166–3182. [Google Scholar] [CrossRef]
  45. Kuipers, E.J.; Rösch, T.; Bretthauer, M. Colorectal cancer screening—Optimizing current strategies and new directions. Nat. Rev. Clin. Oncol. 2013, 10, 130–142. [Google Scholar] [CrossRef]
  46. Rex, D.K.; Johnson, D.A.; Lieberman, D.A.; Burt, R.W.; Sonnenberg, A. Colorectal cancer prevention 2000: Screening recommendations of the American College of Gastroenterology1. Am. J. Gastroenterol. 2000, 95, 868–877. [Google Scholar] [CrossRef]
  47. Van Rijn, J.C.; Reitsma, J.B.; Stoker, J.; Bossuyt, P.M.; Van Deventer, S.J.; Dekker, E. Polyp Miss Rate Determined by Tandem Colonoscopy: A Systematic Review. Am. J. Gastroenterol. 2006, 101, 343–350. [Google Scholar] [CrossRef] [PubMed]
  48. Ferlitsch, M.; Hassan, C.; Bisschops, R.; Bhandari, P.; Dinis-Ribeiro, M.; Risio, M.; Paspatis, G.A.; Moss, A.; Libânio, D.; Lorenzo-Zúñiga, V.; et al. Colorectal polypectomy and endoscopic mucosal resection: European Society of Gastrointestinal Endoscopy (ESGE) Guideline—Update 2024. Endoscopy 2024, 56, 516–545. [Google Scholar] [CrossRef]
  49. Granados-Romero, J.J.; Valderrama-Treviño, A.I.; Contreras-Flores, E.H.; Barrera-Mera, B.; Herrera Enríquez, M.; Uriarte-Ruíz, K.; Ceballos-Villalba, J.C.; Estrada-Mata, A.G.; Alvarado Rodríguez, C.; Arauz-Peña, G. Colorectal cancer: A review. Int. J. Res. Med. Sci. 2017, 5, 4667. [Google Scholar] [CrossRef]
  50. Winawer, S.J.; Zauber, A.G.; O’Brien, M.J.; Geenen, J.; Waye, J.D. The National Polyp Study at 40: Challenges then and now. Gastrointest. Endosc. 2021, 93, 720–726. [Google Scholar] [CrossRef]
  51. Løberg, M.; Kalager, M.; Holme, Ø.; Hoff, G.; Adami, H.-O.; Bretthauer, M. Long-Term Colorectal-Cancer Mortality after Adenoma Removal. N. Engl. J. Med. 2014, 371, 799–807. [Google Scholar] [CrossRef]
  52. Heresbach, D.; Barrioz, T.; Lapalus, M.G.; Coumaros, D.; Bauret, P.; Potier, P.; Sautereau, D.; Boustière, C.; Grimaud, J.C.; Barthélémy, C.; et al. Miss rate for colorectal neoplastic polyps: A prospective multicenter study of back-to-back video colonoscopies. Endoscopy 2008, 40, 284–290. [Google Scholar] [CrossRef] [PubMed]
  53. Pickhardt, P.J.; Nugent, P.A.; Choi, J.R.; Schindler, W.R. Flat Colorectal Lesions in Asymptomatic Adults:</b> Implications for Screening with CT Virtual Colonoscopy. Am. J. Roentgenol. 2004, 183, 1343–1347. [Google Scholar] [CrossRef]
  54. Ahn, S.B.; Han, D.S.; Bae, J.H.; Byun, T.J.; Kim, J.P.; Eun, C.S. The Miss Rate for Colorectal Adenoma Determined by Quality-Adjusted, Back-to-Back Colonoscopies. Gut Liver 2012, 6, 64–70. [Google Scholar] [CrossRef]
  55. Hassan, C.; Repici, A.; Sharma, P.; Correale, L.; Zullo, A.; Bretthauer, M.; Senore, C.; Spada, C.; Bellisario, C.; Bhandari, P.; et al. Efficacy and safety of endoscopic resection of large colorectal polyps: A systematic review and meta-analysis. Gut 2016, 65, 806–820. [Google Scholar] [CrossRef]
  56. Moss, A.; Williams, S.J.; Hourigan, L.F.; Brown, G.; Tam, W.; Singh, R.; Zanati, S.; Burgess, N.G.; Sonson, R.; Byth, K.; et al. Long-term adenoma recurrence following wide-field endoscopic mucosal resection (WF-EMR) for advanced colonic mucosal neoplasia is infrequent: Results and risk factors in 1000 cases from the Australian Colonic EMR (ACE) study. Gut 2015, 64, 57–65. [Google Scholar] [CrossRef]
  57. Jass, J.R. Classification of colorectal cancer based on correlation of clinical, morphological and molecular features. Histopathology 2007, 50, 113–130. [Google Scholar] [CrossRef]
  58. Bressler, B.; Paszat, L.F.; Chen, Z.; Rothwell, D.M.; Vinden, C.; Rabeneck, L. Rates of New or Missed Colorectal Cancers After Colonoscopy and Their Risk Factors: A Population-Based Analysis. Gastroenterology 2007, 132, 96–102. [Google Scholar] [CrossRef]
  59. Kaminski, M.F.; Polkowski, M.; Zwierko, M.; Butruk, E. Quality Indicators for Colonoscopy and the Risk of Interval Cancer. N. Engl. J. Med. 2010, 362, 1795–1803. [Google Scholar] [CrossRef]
  60. Corley, D.A.; Jensen, C.D.; Marks, A.R.; Zhao, W.K.; Lee, J.K.; Doubeni, C.A.; Zauber, A.G.; De Boer, J.; Fireman, B.H.; Schottinger, J.E.; et al. Adenoma Detection Rate and Risk of Colorectal Cancer and Death. N. Engl. J. Med. 2014, 370, 1298–1306. [Google Scholar] [CrossRef]
  61. Tajbakhsh, N.; Gurudu, S.R.; Liang, J. Automated Polyp Detection in Colonoscopy Videos Using Shape and Context Information. IEEE Trans. Med. Imaging 2016, 35, 630–644. [Google Scholar] [CrossRef] [PubMed]
  62. Urban, G.; Tripathi, P.; Alkayali, T.; Mittal, M.; Jalali, F.; Karnes, W.; Baldi, P. Deep Learning Localizes and Identifies Polyps in Real Time With 96% Accuracy in Screening Colonoscopy. Gastroenterology 2018, 155, 1069–1078.e8. [Google Scholar] [CrossRef]
  63. Bui, N.-T.; Hoang, D.-H.; Nguyen, Q.-T.; Tran, M.-T.; Le, N. MEGANet: Multi-Scale Edge-Guided Attention Network for Weak Boundary Polyp Segmentation. In Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–7 January 2024; pp. 7970–7979. [Google Scholar]
  64. Kang, X.; Ma, Z.; Liu, K.; Li, Y.; Miao, Q. Multi-scale information sharing and selection network with boundary attention for polyp segmentation. Eng. Appl. Artif. Intell. 2025, 139, 109467. [Google Scholar] [CrossRef]
  65. Liu, Z.; Zheng, S.; Sun, X.; Zhu, Z.; Zhao, Y.; Yang, X.; Zhao, Y. The Devil Is in the Boundary: Boundary-enhanced Polyp Segmentation. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 5414–5423. [Google Scholar] [CrossRef]
  66. Zhou, T.; Zhou, Y.; He, K.; Gong, C.; Yang, J.; Fu, H.; Shen, D. Cross-level Feature Aggregation Network for Polyp Segmentation. Pattern Recognit. 2023, 140, 109555. [Google Scholar] [CrossRef]
  67. Yue, G.; Han, W.; Jiang, B.; Zhou, T.; Cong, R.; Wang, T. Boundary Constraint Network with Cross Layer Feature Integration for Polyp Segmentation. IEEE J. Biomed. Health Inform. 2022, 26, 4090–4099. [Google Scholar] [CrossRef]
  68. Lin, Y.; Wu, J.; Xiao, G.; Guo, J.; Chen, G.; Ma, J. BSCA-Net: Bit Slicing Context Attention network for polyp segmentation. Pattern Recognit. 2022, 132, 108917. [Google Scholar] [CrossRef]
  69. Chen, H.; Ju, H.; Qin, J.; Song, J.; Lyu, Y.; Liu, X. Dataset-level color augmentation and multi-scale exploration methods for polyp segmentation. Expert. Syst. Appl. 2025, 260, 125395. [Google Scholar] [CrossRef]
  70. Zhou, T.; Zhou, Y.; Li, G.; Chen, G.; Shen, J. Uncertainty-aware Hierarchical Aggregation Network for Medical Image Segmentation. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7440–7453. [Google Scholar] [CrossRef]
  71. Wang, T.; Qi, X.; Yang, G. Polyp Segmentation via Semantic Enhanced Perceptual Network. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 12594–12607. [Google Scholar] [CrossRef]
  72. Jin, Q.; Hou, H.; Zhang, G.; Li, Z. FEGNet: A Feedback Enhancement Gate Network for Automatic Polyp Segmentation. IEEE J. Biomed. Health Inform. 2023, 27, 3420–3430. [Google Scholar] [CrossRef]
  73. Jain, S.; Atale, R.; Gupta, A.; Mishra, U.; Seal, A.; Ojha, A.; Kuncewicz, J.; Krejcar, O. CoInNet: A Convolution-Involution Network with a Novel Statistical Attention for Automatic Polyp Segmentation. IEEE Trans. Med. Imaging 2023, 42, 3987–4000. [Google Scholar] [CrossRef]
  74. Wang, C.; Xu, R.; Xu, S.; Meng, W.; Zhang, X. Automatic polyp segmentation via image-level and surrounding-level context fusion deep neural network. Eng. Appl. Artif. Intell. 2023, 123, 106168. [Google Scholar] [CrossRef]
  75. Zhang, R.; Lai, P.; Wan, X.; Fan, D.-J.; Gao, F.; Wu, X.-J.; Li, G. Lesion-Aware Dynamic Kernel for Polyp Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 99–109. [Google Scholar]
  76. Srivastava, A.; Jha, D.; Chanda, S.; Pal, U.; Johansen, H.D.; Johansen, D.; Riegler, M.A.; Ali, S.; Halvorsen, P. MSRF-Net: A Multi-Scale Residual Fusion Network for Biomedical Image Segmentation. IEEE J. Biomed. Health Inform. 2022, 26, 2252–2263. [Google Scholar] [CrossRef] [PubMed]
  77. Liu, M.; Jiao, L.; Liu, X.; Li, L.; Liu, F.; Yang, S.; Wang, S.; Hou, B. Multi-scale Contourlet Knowledge Guide Learning Segmentation. IEEE Trans. Multimed. 2023, 26, 4831–4845. [Google Scholar] [CrossRef]
  78. Song, P.; Li, J.; Fan, H. Attention based multi-scale parallel network for polyp segmentation. Comput. Biol. Med. 2022, 146, 105476. [Google Scholar] [CrossRef]
  79. Liu, G.; Chen, Z.; Liu, D.; Chang, B.; Dou, Z. FTMF-Net: A Fourier Transform-Multiscale Feature Fusion Network for Segmentation of Small Polyp Objects. IEEE Trans. Instrum. Meas. 2023, 72, 502085. [Google Scholar] [CrossRef]
  80. Gu, Y.; Zhou, T.; Zhang, Y.; Zhou, Y.; He, K.; Gong, C.; Fu, H. Dual-scale enhanced and cross-generative consistency learning for semi-supervised medical image segmentation. Pattern Recognit. 2025, 158, 110962. [Google Scholar] [CrossRef]
  81. Ren, G.; Lazarou, M.; Yuan, J.; Stathaki, T. Towards Automated Polyp Segmentation Using Weakly- and Semi-Supervised Learning and Deformable Transformers. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 17–24 June 2023; pp. 4355–4364. [Google Scholar]
  82. Yu, Z.; Zhao, L.; Liao, T.; Zhang, X.; Chen, G.; Xiao, G. A novel non-pretrained deep supervision network for polyp segmentation. Pattern Recognit. 2024, 154, 110554. [Google Scholar] [CrossRef]
  83. Guo, X.; Chen, Z.; Liu, J.; Yuan, Y. Non-equivalent images and pixels: Confidence-aware resampling with meta-learning mixup for polyp segmentation. Med. Image Anal. 2022, 78, 102394. [Google Scholar] [CrossRef]
  84. Wang, A.; Xu, M.; Zhang, Y.; Islam, M.; Ren, H. S^2ME: Spatial-Spectral Mutual Teaching and Ensemble Learning for Scribble-Supervised Polyp Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; Greenspan, H., Madabhushi, A., Mousavi, P., Salcudean, S., Duncan, J., Syeda-Mahmood, T., Taylor, R., Eds.; Springer Nature: Cham, Switzerland, 2023; pp. 35–45. [Google Scholar]
  85. Jia, X.; Shen, Y.; Yang, J.; Song, R.; Zhang, W.; Meng, M.Q.-H.; Liao, J.C.; Xing, L. PolypMixNet: Enhancing semi-supervised polyp segmentation with polyp-aware augmentation. Comput. Biol. Med. 2024, 170, 108006. [Google Scholar] [CrossRef]
  86. Zhang, Z.; Li, Y.; Shin, B.-S. Generalizable Polyp Segmentation via Randomized Global Illumination Augmentation. IEEE J. Biomed. Health Inform. 2024, 28, 2138–2151. [Google Scholar] [CrossRef]
  87. Lu, Z.; Zhang, Y.; Zhou, Y.; Wu, Y.; Zhou, T. Domain-interactive Contrastive Learning and Prototype-guided Self-training for Cross-domain Polyp Segmentation. IEEE Trans. Med. Imaging 2024. [Google Scholar] [CrossRef] [PubMed]
  88. Wang, J.; Chen, C. Unsupervised Adaptation of Polyp Segmentation Models via Coarse-to-Fine Self-Supervision. In Proceedings of the Information Processing in Medical Imaging, San Carlos de Bariloche, Argentina, 18–23 June 2023; Frangi, A., de Bruijne, M., Wassermann, D., Navab, N., Eds.; Springer Nature: Cham, Switzerland, 2023; pp. 250–262. [Google Scholar]
  89. Wu, C.; Long, C.; Li, S.; Yang, J.; Jiang, F.; Zhou, R. MSRAformer: Multiscale spatial reverse attention network for polyp segmentation. Comput. Biol. Med. 2022, 151, 106274. [Google Scholar] [CrossRef] [PubMed]
  90. Liu, J.; Chen, Q.; Zhang, Y.; Wang, Z.; Deng, X.; Wang, J. Multi-level feature fusion network combining attention mechanisms for polyp segmentation. Inf. Fusion 2024, 104, 102195. [Google Scholar] [CrossRef]
  91. Xu, W.; Xu, R.; Wang, C.; Li, X.; Xu, S.; Guo, L. PSTNet: Enhanced Polyp Segmentation with Multi-Scale Alignment and Frequency Domain Integration. IEEE J. Biomed. Health Inform. 2024, 28, 6042–6053. [Google Scholar] [CrossRef] [PubMed]
  92. Yue, G.; Xiao, H.; Zhou, T.; Tan, S.; Liu, Y.; Yan, W. Progressive Feature Enhancement Network for Automated Colorectal Polyp Segmentation. IEEE Trans. Autom. Sci. Eng. 2024, 22, 5792–5803. [Google Scholar] [CrossRef]
  93. Dong, B.; Wang, W.; Fan, D.-P.; Li, J.; Fu, H.; Shao, L. Polyp-PVT: Polyp Segmentation with Pyramid Vision Transformers. CAAI Artif. Intell. Res. 2023, 2, 9150015. [Google Scholar] [CrossRef]
  94. Liu, G.; Yao, S.; Liu, D.; Chang, B.; Chen, Z.; Wang, J.; Wei, J. CAFE-Net: Cross-Attention and Feature Exploration Network for polyp segmentation. Expert Syst. Appl. 2024, 238, 121754. [Google Scholar] [CrossRef]
  95. Li, W.; Lu, W.; Chu, J.; Fan, F. LACINet: A Lesion-Aware Contextual Interaction Network for Polyp Segmentation. IEEE Trans. Instrum. Meas. 2023, 72, 5029112. [Google Scholar] [CrossRef]
  96. Cai, L.; Chen, L.; Huang, J.; Wang, Y.; Zhang, Y. Know your orientation: A viewpoint-aware framework for polyp segmentation. Med. Image Anal. 2024, 97, 103288. [Google Scholar] [CrossRef]
  97. Rahman, M.M.; Munir, M.; Jha, D.; Bagci, U.; Marculescu, R. PP-SAM: Perturbed Prompts for Robust Adaption of Segment Anything Model for Polyp Segmentation. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024; pp. 4989–4995. [Google Scholar]
  98. Jha, D.; Tomar, N.K.; Bhattacharya, D.; Bagci, U. TransRUPNet for Improved Polyp Segmentation. arXiv 2024. [Google Scholar] [CrossRef]
  99. Huang, X.; Gong, H.; Zhang, J. HST-MRF: Heterogeneous Swin Transformer with Multi-Receptive Field for Medical Image Segmentation. IEEE J. Biomed. Health Inform. 2024, 28, 4048–4061. [Google Scholar] [CrossRef]
  100. Lin, A.; Chen, B.; Xu, J.; Zhang, Z.; Lu, G.; Zhang, D. DS-TransUNet: Dual Swin Transformer U-Net for Medical Image Segmentation. IEEE Trans. Instrum. Meas. 2022, 71, 4005615. [Google Scholar] [CrossRef]
  101. Liu, Y.; Yang, Y.; Jiang, Y.; Xie, Z. Multi-view orientational attention network combining point-based affinity for polyp segmentation. Expert Syst. Appl. 2024, 249, 123663. [Google Scholar] [CrossRef]
  102. Yin, X.; Zeng, J.; Hou, T.; Tang, C.; Gan, C.; Jain, D.K.; García, S. RSAFormer: A method of polyp segmentation with region self-attention transformer. Comput. Biol. Med. 2024, 172, 108268. [Google Scholar] [CrossRef] [PubMed]
  103. Yang, C.; Zhang, Z. PFD-Net: Pyramid Fourier Deformable Network for medical image segmentation. Comput. Biol. Med. 2024, 172, 108302. [Google Scholar] [CrossRef] [PubMed]
  104. Wu, H.; Min, W.; Gai, D.; Huang, Z.; Geng, Y.; Wang, Q.; Chen, R. HD-Former: A hierarchical dependency Transformer for medical image segmentation. Comput. Biol. Med. 2024, 178, 108671. [Google Scholar] [CrossRef]
  105. Du, H.; Wang, J.; Liu, M.; Wang, Y.; Meijering, E. SwinPA-Net: Swin Transformer-Based Multiscale Feature Pyramid Aggregation Network for Medical Image Segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 5355–5366. [Google Scholar] [CrossRef]
  106. Zhang, W.; Fu, C.; Zheng, Y.; Zhang, F.; Zhao, Y.; Sham, C.-W. HSNet: A hybrid semantic network for polyp segmentation. Comput. Biol. Med. 2022, 150, 106173. [Google Scholar] [CrossRef]
  107. Li, W.; Nie, X.; Li, F.; Huang, Z.; Zeng, G. FMCA-Net: A feature secondary multiplexing and dilated convolutional attention polyp segmentation network based on pyramid vision transformer. Expert Syst. Appl. 2025, 260, 125419. [Google Scholar] [CrossRef]
  108. Jiang, X.; Zhu, Y.; Liu, Y.; Wang, N.; Yi, L. MC-DC: An MLP-CNN Based Dual-path Complementary Network for Medical Image Segmentation. Comput. Methods Programs Biomed. 2023, 242, 107846. [Google Scholar] [CrossRef]
  109. Wu, H.; Zhao, Z.; Wang, Z. META-Unet: Multi-Scale Efficient Transformer Attention Unet for Fast and High-Accuracy Polyp Segmentation. IEEE Trans. Autom. Sci. Eng. 2023, 21, 4117–4128. [Google Scholar] [CrossRef]
  110. Sanderson, E.; Matuszewski, B.J. FCN-Transformer Feature Fusion for Polyp Segmentation. In Proceedings of the 26th Annual Conference, MIUA 2022, Cambridge, UK, 27–29 July 2022; Volume 13413, pp. 892–907. [Google Scholar] [CrossRef]
  111. Li, W.; Huang, Z.; Li, F.; Zhao, Y.; Zhang, H. CIFG-Net: Cross-level information fusion and guidance network for Polyp Segmentation. Comput. Biol. Med. 2024, 169, 107931. [Google Scholar] [CrossRef] [PubMed]
  112. Shao, H.; Zhang, Y.; Hou, Q. Polyper: Boundary Sensitive Polyp Segmentation. Proc. AAAI Conf. Artif. Intell. 2024, 38, 4731–4739. [Google Scholar] [CrossRef]
  113. Cai, L.; Wu, M.; Chen, L.; Bai, W.; Yang, M.; Lyu, S.; Zhao, Q. Using Guided Self-Attention with Local Information for Polyp Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 629–638. [Google Scholar]
  114. Xie, Y.; Zhou, T.; Zhou, Y.; Chen, G. SimTxtSeg: Weakly-Supervised Medical Image Segmentation with Simple Text Cues. arXiv 2024. [Google Scholar] [CrossRef]
  115. Li, H.; Zhang, D.; Yao, J.; Han, L.; Li, Z.; Han, J. ASPS: Augmented Segment Anything Model for Polyp Segmentation. arXiv 2024. [Google Scholar] [CrossRef]
  116. Xiong, X.; Wu, Z.; Tan, S.; Li, W.; Tang, F.; Chen, Y.; Li, S.; Ma, J.; Li, G. SAM2-UNet: Segment Anything 2 Makes Strong Encoder for Natural and Medical Image Segmentation. arXiv 2024, arXiv:2408.08870. [Google Scholar] [CrossRef]
  117. Jha, D.; Tomar, N.K.; Sharma, V.; Bagci, U. TransNetR: Transformer-based Residual Network for Polyp Segmentation with Multi-Center Out-of-Distribution Testing. arXiv 2023, arXiv:2303.07428. [Google Scholar]
  118. Xie, J.; Liao, R.; Zhang, Z.; Yi, S.; Zhu, Y.; Luo, G. ProMamba: Prompt-Mamba for polyp segmentation. arXiv 2024. [Google Scholar] [CrossRef]
  119. Zhang, M.; Yu, Y.; Jin, S.; Gu, L.; Ling, T.; Tao, X. VM-UNET-V2: Rethinking Vision Mamba UNet for Medical Image Segmentation. In Proceedings of the Bioinformatics Research and Applications, Milan, Italy, 13–15 September 2024; Peng, W., Cai, Z., Skums, P., Eds.; Springer Nature: Singapore, 2024; pp. 335–346. [Google Scholar]
  120. Dutta, T.K.; Majhi, S.; Nayak, D.R.; Jha, D. SAM-Mamba: Mamba Guided SAM Architecture for Generalized Zero-Shot Polyp Segmentation. arXiv 2024. [Google Scholar] [CrossRef]
  121. Zhu, X.; Wang, W.; Zhang, C.; Wang, H. Polyp-Mamba: A Hybrid Multi-Frequency Perception Gated Selection Network for polyp segmentation. Inf. Fusion 2025, 115, 102759. [Google Scholar] [CrossRef]
  122. Wu, R.; Liu, Y.; Liang, P.; Chang, Q. H-vmunet: High-order Vision Mamba UNet for Medical Image Segmentation. arXiv 2024. [Google Scholar] [CrossRef]
  123. Du, Y.; Jiang, Y.; Tan, S.; Liu, S.-Q.; Li, Z.; Li, G.; Wan, X. Highlighted Diffusion Model as Plug-in Priors for Polyp Segmentation. IEEE J. Biomed. Health Inform. 2024, 29, 1209–1220. [Google Scholar] [CrossRef]
  124. Wang, Z.; Liu, M.; Jiang, J.; Qu, X. Colorectal polyp segmentation with denoising diffusion probabilistic models. Comput. Biol. Med. 2024, 180, 108981. [Google Scholar] [CrossRef]
  125. Mansoori, M.; Shahabodini, S.; Abouei, J.; Plataniotis, K.N.; Mohammadi, A. Polyp SAM 2: Advancing Zero shot Polyp Segmentation in Colorectal Cancer Detection. arXiv 2024. [Google Scholar] [CrossRef]
  126. Banik, D.; Roy, K.; Krejcar, O.; Bhattacharjee, D. dHBLSN: A diligent hierarchical broad learning system network for cogent polyp segmentation. Knowl.-Based Syst. 2024, 300, 112228. [Google Scholar] [CrossRef]
  127. Wu, H.; Zhao, Z.; Zhong, J.; Wang, W.; Wen, Z.; Qin, J. PolypSeg+: A Lightweight Context-Aware Network for Real-Time Polyp Segmentation. IEEE Trans. Cybern. 2023, 53, 2610–2621. [Google Scholar] [CrossRef]
  128. Wan, L.; Chen, Z.; Xiao, Y.; Zhao, J.; Feng, W.; Fu, H. Iterative feedback-based models for image and video polyp segmentation. Comput. Biol. Med. 2024, 177, 108569. [Google Scholar] [CrossRef]
  129. Lu, Y.; Yang, Y.; Xing, Z.; Wang, Q.; Zhu, L. Diff-VPS: Video Polyp Segmentation via a Multi-task Diffusion Network with Adversarial Temporal Reasoning. arXiv 2024, arXiv:2409.07238. [Google Scholar]
  130. Chen, G.; Yang, J.; Pu, X.; Ji, G.-P.; Xiong, H.; Pan, Y.; Cui, H.; Xia, Y. MAST: Video Polyp Segmentation with a Mixture-Attention Siamese Transformer. arXiv 2024, arXiv:2401.12439. [Google Scholar] [CrossRef]
  131. Xu, Z.; Rittscher, J.; Ali, S. SSTFB: Leveraging self-supervised pretext learning and temporal self-attention with feature branching for real-time video polyp segmentation. arXiv 2024, arXiv:2406.10200. [Google Scholar]
  132. Hu, Q.; Yi, Z.; Zhou, Y.; Peng, F.; Liu, M.; Li, Q.; Wang, Z. SALI: Short-term Alignment and Long-term Interaction Network for Colonoscopy Video Polyp Segmentation. arXiv 2024, arXiv:2406.13532. [Google Scholar]
  133. Yang, Y.; Xing, Z.; Zhu, L. Vivim: A Video Vision Mamba for Medical Video Object Segmentation. arXiv 2024, arXiv:2401.14168. [Google Scholar] [CrossRef]
  134. Wang, M.; An, X.; Pei, Z.; Li, N.; Zhang, L.; Liu, G.; Ming, D. An Efficient Multi-Task Synergetic Network for Polyp Segmentation and Classification. IEEE J. Biomed. Health Inform. 2023, 28, 1228–1239. [Google Scholar] [CrossRef]
  135. Mazumdar, H.; Chakraborty, C.; Sathvik, M.; Jayakumar, P.; Kaushik, A. Optimizing Pix2Pix GAN with Attention Mechanisms for AI-Driven Polyp Segmentation in IoMT-Enabled Smart Healthcare. IEEE J. Biomed. Health Inform. 2023, 29, 3825–3832. [Google Scholar] [CrossRef]
  136. Tomar, N.K.; Jha, D.; Bagci, U.; Ali, S. TGANet: Text-Guided Attention for Improved Polyp Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; pp. 151–160. [Google Scholar]
  137. Liu, T.; Ye, X.; Hu, K.; Xiong, D.; Zhang, Y.; Li, X.; Gao, X. Polyp segmentation with distraction separation. Expert Syst. Appl. 2023, 228, 120434. [Google Scholar] [CrossRef]
  138. Ji, Z.; Qian, H.; Ma, X. Progressive Group Convolution Fusion network for colon polyp segmentation. Biomed. Signal Process. Control 2024, 96, 106586. [Google Scholar] [CrossRef]
  139. Silva, J.; Histace, A.; Romain, O.; Dray, X.; Granado, B. Toward embedded detection of polyps in WCE images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 283–293. [Google Scholar] [CrossRef] [PubMed]
  140. Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; Gil, D.; Rodríguez, C.; Vilariño, F. WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Comput. Med. Imaging Graph. 2015, 43, 99–111. [Google Scholar] [CrossRef]
  141. Vázquez, D.; Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; López, A.M.; Romero, A.; Drozdzal, M.; Courville, A. A Benchmark for Endoluminal Scene Segmentation of Colonoscopy Images. J. Healthc. Eng. 2017, 2017, 4037190. [Google Scholar] [CrossRef]
  142. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Halvorsen, P.; de Lange, T.; Johansen, D.; Johansen, H.D. Kvasir-SEG: A Segmented Polyp Dataset. In Proceedings of the MultiMedia Modeling, Daejeon, Republic of Korea, 5–8 January 2020; Ro, Y.M., Cheng, W.-H., Kim, J., Chu, W.-T., Cui, P., Choi, J.-W., Hu, M.-C., De Neve, W., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 451–462. [Google Scholar]
  143. Borgli, H.; Thambawita, V.; Smedsrud, P.H.; Hicks, S.; Jha, D.; Eskeland, S.L.; Randel, K.R.; Pogorelov, K.; Lux, M.; Nguyen, D.T.D.; et al. HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy. Sci. Data 2020, 7, 283. [Google Scholar] [CrossRef]
  144. Sánchez-Peralta, L.F.; Pagador, J.B.; Picón, A.; Calderón, Á.J.; Polo, F.; Andraka, N.; Bilbao, R.; Glover, B.; Saratxaga, C.L.; Sánchez-Margallo, F.M. PICCOLO White-Light and Narrow-Band Imaging Colonoscopic Dataset: A Performance Comparative of Models and Datasets. Appl. Sci. 2020, 10, 8501. [Google Scholar] [CrossRef]
  145. Ngoc Lan, P.; An, N.S.; Hang, D.V.; Long, D.V.; Trung, T.Q.; Thuy, N.T.; Sang, D.V. NeoUNet: Towards Accurate Colon Polyp Segmentation and Neoplasm Detection. In Proceedings of the Advances in Visual Computing, Virtual, 4–6 October 2021; Bebis, G., Athitsos, V., Yan, T., Lau, M., Li, F., Shi, C., Yuan, X., Mousas, C., Bruder, G., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 15–28. [Google Scholar]
  146. Ali, S.; Jha, D.; Ghatwary, N.; Realdon, S.; Cannizzaro, R.; Salem, O.E.; Lamarque, D.; Daul, C.; Riegler, M.A.; Anonsen, K.V.; et al. A multi-centre polyp detection and segmentation dataset for generalisability assessment. Sci. Data 2023, 10, 75. [Google Scholar] [CrossRef] [PubMed]
  147. Ma, Y.; Chen, X.; Cheng, K.; Li, Y.; Sun, B. LDPolypVideo Benchmark: A Large-Scale Colonoscopy Video Dataset of Diverse Polyps. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Strasbourg, France, 27 September–1 October 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 387–396. [Google Scholar]
  148. Smedsrud, P.H.; Thambawita, V.; Hicks, S.A.; Gjestang, H.; Nedrejord, O.O.; Næss, E.; Borgli, H.; Jha, D.; Berstad, T.J.D.; Eskeland, S.L.; et al. Kvasir-Capsule, a video capsule endoscopy dataset. Sci. Data 2021, 8, 142. [Google Scholar] [CrossRef]
  149. Misawa, M.; Kudo, S.; Mori, Y.; Hotta, K.; Ohtsuka, K.; Matsuda, T.; Saito, S.; Kudo, T.; Baba, T.; Ishida, F.; et al. Development of a computer-aided detection system for colonoscopy and a publicly accessible large colonoscopy video database (with video). Gastrointest. Endosc. 2021, 93, 960–967.e3. [Google Scholar] [CrossRef]
  150. Jha, S.; Son, L.H.; Kumar, R.; Priyadarshini, I.; Smarandache, F.; Long, H.V. Neutrosophic image segmentation with Dice Coefficients. Measurement 2019, 134, 762–772. [Google Scholar] [CrossRef]
  151. Shamir, R.R.; Duchin, Y.; Kim, J.; Sapiro, G.; Harel, N. Continuous Dice Coefficient: A Method for Evaluating Probabilistic Segmentations. arXiv 2019. [Google Scholar] [CrossRef]
  152. Zhao, R.; Qian, B.; Zhang, X.; Li, Y.; Wei, R.; Liu, Y.; Pan, Y. Rethinking dice loss for medical image segmentation. In Proceedings of the 2020 IEEE international conference on data mining (ICDM), Sorrento, Italy, 17–20 November 2020; pp. 851–860. [Google Scholar]
  153. Jha, D.; Ali, S.; Tomar, N.; Johansen, H.; Johansen, D.; Rittscher, J.; Riegler, M.; Halvorsen, P. Real-Time Polyp Detection, Localization and Segmentation in Colonoscopy Using Deep Learning. IEEE Access 2021, 9, 40496–40510. [Google Scholar] [CrossRef] [PubMed]
  154. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  155. Coleman, C.; Kang, D.; Narayanan, D.; Nardi, L.; Zhao, T.; Zhang, J.; Bailis, P.; Olukotun, K.; Ré, C.; Zaharia, M. Analysis of DAWNBench, a Time-to-Accuracy Machine Learning Performance Benchmark. SIGOPS Oper. Syst. Rev. 2019, 53, 14–25. [Google Scholar] [CrossRef]
  156. Wickstrøm, K.; Kampffmeyer, M.; Jenssen, R. Uncertainty and interpretability in convolutional neural networks for semantic segmentation of colorectal polyps. Med. Image Anal. 2020, 60, 101619. [Google Scholar] [CrossRef]
  157. Ramos, D.L.; Hortúa, H.J. Deep Bayesian segmentation for colon polyps: Well-calibrated predictions in medical imaging. Biomed. Signal Process. Control 2025, 104, 107383. [Google Scholar] [CrossRef]
  158. Char, D.S.; Shah, N.H.; Magnus, D. Implementing Machine Learning in Health Care—Addressing Ethical Challenges. N. Engl. J. Med. 2018, 378, 981–983. [Google Scholar] [CrossRef]
  159. Price, W.N.; Cohen, I.G. Privacy in the age of medical big data. Nat. Med. 2019, 25, 37–43. [Google Scholar] [CrossRef] [PubMed]
  160. Mittelstadt, B. Principles alone cannot guarantee ethical AI. Nat. Mach. Intell. 2019, 1, 501–507. [Google Scholar] [CrossRef]
  161. Kaissis, G.A.; Makowski, M.R.; Rückert, D.; Braren, R.F. Secure, privacy-preserving and federated machine learning in medical imaging. Nat. Mach. Intell. 2020, 2, 305–311. [Google Scholar] [CrossRef]
  162. Vayena, E.; Blasimme, A.; Cohen, I.G. Machine learning in medicine: Addressing ethical challenges. PLoS Med. 2018, 15, e1002689. [Google Scholar] [CrossRef]
  163. Geis, J.R.; Brady, A.P.; Wu, C.C.; Spencer, J.; Ranschaert, E.; Jaremko, J.L.; Langer, S.G.; Borondy Kitts, A.; Birch, J.; Shields, W.F.; et al. Ethics of Artificial Intelligence in Radiology: Summary of the Joint European and North American Multisociety Statement. Radiology 2019, 293, 436–440. [Google Scholar] [CrossRef] [PubMed]
  164. Willemink, M.J.; Koszek, W.A.; Hardell, C.; Wu, J.; Fleischmann, D.; Harvey, H.; Folio, L.R.; Summers, R.M.; Rubin, D.L.; Lungren, M.P. Preparing Medical Imaging Data for Machine Learning. Radiology 2020, 295, 4–15. [Google Scholar] [CrossRef] [PubMed]
  165. Kelly, C.J.; Karthikesalingam, A.; Suleyman, M.; Corrado, G.; King, D. Key challenges for delivering clinical impact with artificial intelligence. BMC Med. 2019, 17, 195. [Google Scholar] [CrossRef] [PubMed]
  166. Larrazabal, A.J.; Nieto, N.; Peterson, V.; Milone, D.H.; Ferrante, E. Gender imbalance in medical imaging datasets produces biased classifiers for computer-aided diagnosis. Proc. Natl. Acad. Sci. USA 2020, 117, 12592–12594. [Google Scholar] [CrossRef]
  167. Bai, L.; Chen, T.; Tan, Q.; Nah, W.J.; Li, Y.; He, Z.; Yuan, S.; Chen, Z.; Wu, J.; Islam, M.; et al. EndoUIC: Promptable Diffusion Transformer for Unified Illumination Correction in Capsule Endoscopy. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2024, Marrakesh, Morocco, 6–10 October 2024; Linguraru, M.G., Dou, Q., Feragen, A., Giannarou, S., Glocker, B., Lekadir, K., Schnabel, J.A., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 296–306. [Google Scholar]
  168. Shang, W.; Ren, D.; Yang, Y.; Zuo, W. Aggregating nearest sharp features via hybrid transformers for video deblurring. Inf. Sci. 2025, 694, 121689. [Google Scholar] [CrossRef]
  169. ELKarazle, K.; Raman, V.; Chua, C.; Then, P. A Hessian-Based Technique for Specular Reflection Detection and Inpainting in Colonoscopy Images. IEEE J. Biomed. Health Inform. 2024, 28, 4724–4736. [Google Scholar] [CrossRef]
  170. Sharma, V.; Kumar, A.; Jha, D.; Bhuyan, M.K.; Das, P.K.; Bagci, U. ControlPolypNet: Towards Controlled Colon Polyp Synthesis for Improved Polyp Segmentation. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024; pp. 2325–2334. [Google Scholar]
  171. Ji, G.-P.; Liu, J.; Xu, P.; Barnes, N.; Khan, F.S.; Khan, S.; Fan, D.-P. Frontiers in Intelligent Colonoscopy. arXiv 2024. [Google Scholar] [CrossRef]
  172. Schön, R.; Lorenz, J.; Ludwig, K.; Lienhart, R. Adapting the Segment Anything Model During Usage in Novel Situations. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024; pp. 3616–3626. [Google Scholar]
  173. Hu, Q.; Yi, Z.; Zhou, Y.; Huang, F.; Liu, M.; Li, Q.; Wang, Z. MonoBox: Tightness-free Box-supervised Polyp Segmentation using Monotonicity Constraint. arXiv 2024. [Google Scholar] [CrossRef]
  174. Long, J.; Lin, J.; Liu, D. W-PolypBox: Exploring bounding box priors constraints for weakly supervised polyp segmentation. Biomed. Signal Process. Control 2025, 103, 107418. [Google Scholar] [CrossRef]
  175. Zhang, Z.; Jiang, Y.; Wang, Y.; Xie, B.; Zhang, W.; Li, Y.; Chen, Z.; Jin, X.; Zeng, W. Exploring Contrastive Pre-training for Domain Connections in Medical Image Segmentation. IEEE Trans. Med. Imaging 2025, 44, 1686–1698. [Google Scholar] [CrossRef]
  176. Fang, Z.; Liu, Y.; Wu, H.; Qin, J. VP-SAM: Taming Segment Anything Model for Video Polyp Segmentation via Disentanglement and Spatio-Temporal Side Network. In Proceedings of the Computer Vision—ECCV 2024, Milan, Italy, 29 September–4 October 2024; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer Nature: Cham, Switzerland, 2025; pp. 367–383. [Google Scholar]
  177. Xiao, A.; Xuan, W.; Qi, H.; Xing, Y.; Ren, R.; Zhang, X.; Shao, L.; Lu, S. CAT-SAM: Conditional Tuning for Few-Shot Adaptation of Segment Anything Model. In Proceedings of the Computer Vision—ECCV 2024, Milan, Italy, 29 September–4 October 2024; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer Nature: Cham, Switzerland, 2025; pp. 189–206. [Google Scholar]
  178. Wang, L.; Xu, Q.; Chen, C.; Yang, H.; Deng, G. Adaptive cascade decoders for segmenting challenging regions in medical images. Comput. Biol. Med. 2025, 185, 109572. [Google Scholar] [CrossRef]
  179. Pan, H.; Jha, D.; Biswas, K.; Bagci, U. Frequency-Based Federated Domain Generalization for Polyp Segmentation. arXiv 2024. [Google Scholar] [CrossRef]
  180. Li, W.; Zhang, Y.; Zhou, H.; Yang, W.; Xie, Z.; He, Y. CLMS: Bridging domain gaps in medical imaging segmentation with source-free continual learning for robust knowledge transfer and adaptation. Med. Image Anal. 2025, 100, 103404. [Google Scholar] [CrossRef]
  181. Huang, L.; Ruan, S.; Xing, Y.; Feng, M. A review of uncertainty quantification in medical image analysis: Probabilistic and non-probabilistic methods. Med. Image Anal. 2024, 97, 103223. [Google Scholar] [CrossRef]
  182. Wang, J.; Jin, Y.; Stoyanov, D.; Wang, L. FedDP: Dual Personalization in Federated Medical Image Segmentation. IEEE Trans. Med. Imaging 2024, 43, 297–308. [Google Scholar] [CrossRef] [PubMed]
  183. Li, L.; Fan, Y.; Tse, M.; Lin, K.-Y. A review of applications in federated learning. Comput. Ind. Eng. 2020, 149, 106854. [Google Scholar] [CrossRef]
  184. Ma, Y.; Wang, J.; Yang, J.; Wang, L. Model-Heterogeneous Semi-Supervised Federated Learning for Medical Image Segmentation. IEEE Trans. Med. Imaging 2024, 43, 1804–1815. [Google Scholar] [CrossRef] [PubMed]
  185. Stelter, L.; Corbetta, V.; Beets-Tan, R.; Silva, W. Assessing the Impact of Federated Learning and Differential Privacy on Multi-centre Polyp Segmentation. In Proceedings of the 2024 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 15–19 July 2024; pp. 1–4. [Google Scholar]
Figure 1. Progress in polyp segmentation research [6,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29].
Figure 1. Progress in polyp segmentation research [6,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29].
Jimaging 11 00293 g001
Figure 2. Flow diagram of the literature review procedures.
Figure 2. Flow diagram of the literature review procedures.
Jimaging 11 00293 g002
Figure 3. Distribution of published papers and selected papers from 2018–2024.
Figure 3. Distribution of published papers and selected papers from 2018–2024.
Jimaging 11 00293 g003
Figure 4. Diversity of polyp appearances across datasets. (Top row (left to right)): low contrast (ETIS-Larib), blurred boundaries (ETIS-Larib), small polyp (CVC-ClinicDB). (Bottom row (left to right)): contains hemorrhagic regions (Kvasir-SEG), contains multiple polyps (Kvasir-SEG), contains reflective areas (ETIS-Larib).
Figure 4. Diversity of polyp appearances across datasets. (Top row (left to right)): low contrast (ETIS-Larib), blurred boundaries (ETIS-Larib), small polyp (CVC-ClinicDB). (Bottom row (left to right)): contains hemorrhagic regions (Kvasir-SEG), contains multiple polyps (Kvasir-SEG), contains reflective areas (ETIS-Larib).
Jimaging 11 00293 g004
Figure 5. Image samples and corresponding segmentation masks from four benchmark datasets: CVC-ClinicDB, Kvasir-SEG, CVC-ColonDB, and ETIS-Larib.
Figure 5. Image samples and corresponding segmentation masks from four benchmark datasets: CVC-ClinicDB, Kvasir-SEG, CVC-ColonDB, and ETIS-Larib.
Jimaging 11 00293 g005
Figure 6. Classification of network architectures and challenges for polyp segmentation methods.
Figure 6. Classification of network architectures and challenges for polyp segmentation methods.
Jimaging 11 00293 g006
Figure 7. Progress in polyp segmentation datasets.
Figure 7. Progress in polyp segmentation datasets.
Jimaging 11 00293 g007
Figure 8. Comparative performance analysis of 10 polyp segmentation models across four benchmark datasets. Each dataset is represented by two metrics: mDice (solid lines) and mIoU (dashed lines).
Figure 8. Comparative performance analysis of 10 polyp segmentation models across four benchmark datasets. Each dataset is represented by two metrics: mDice (solid lines) and mIoU (dashed lines).
Jimaging 11 00293 g008
Figure 9. Distribution of Params, FPS, and mDice values for 10 models on the CVC-ClinicDB dataset.
Figure 9. Distribution of Params, FPS, and mDice values for 10 models on the CVC-ClinicDB dataset.
Jimaging 11 00293 g009
Table 1. Summary table of previous reviews on polyp segmentation.
Table 1. Summary table of previous reviews on polyp segmentation.
Title of the PaperYearDescriptionLimitations
Deep learning to find colorectal polyps in colonoscopy: A systematic literature review [30]2020This review analyzes 35 papers published between 2015 and 2018 that include polyp detection, localization, and segmentation.The number of analytical studies is relatively small; these are models from 2015–2018.
A survey on deep learning for polyp segmentation: Techniques, challenges, and future trends [31]2024This review examines 58 papers from 2019 to 2023 and analyzes 24 models.There has been no evaluation or analysis of the model’s real-time performance.
Colorectal Polyp Segmentation in the Deep Learning Era: A Comprehensive Survey [32]2024This review examines 115 papers from 2014 to 2023 and analyzes 40 model papers from 2015 to 2023.The models being analyzed do not have real-time performance evaluation.
A systematic review of deep learning based image segmentation to detect polyp [33]2024This review examines 117 papers from 2018 to 2023 on the use of deep learning image segmentation models for polyp segmentation. It analyzes 16 models.The number of models in the statistical table is relatively small.
A Comprehensive Review of Deep Learning for Medical Image Segmentation [34]2024This review analyzes 41 medical image segmentation papers from 2015 to 2024, covering skin lesions, hippocampus, tumors, and polyps.The polyp segmentation model has a small sample size; there is no performance evaluation analysis.
A survey of deep learning algorithms for colorectal polyp segmentation [35]2024This review discusses four challenges encountered by deep learning methods in the task of colorectal polyp segmentation, along with related papers.There is no statistical examination of Mamba-based models, no model statistical table, and no performance evaluation analysis of the model.
Cross-modal hybrid architectures for gastrointestinal tract image analysis: A systematic review and futuristic applications [36]2024This review surveys CNN and Transformer methods for organ and polyp segmentation.There is no statistical examination of Mamba-based models and no assessment of the models’ performance.
Table 2. Summary of polyp segmentation methods from 2022 to 2024.
Table 2. Summary of polyp segmentation methods from 2022 to 2024.
ArchitectureYearMethodsBackboneProblems SolvedAdvantagesLimitations
CNN2024RGIA [86]N/ADomain transfer problem.Accurately segmenting polyps under complex shapes and various lighting conditions.Insufficient handling of domain shift factors such as color and shape differences.
PolypMixNet [85]ResNet34Limited annotated data and class imbalance.Achieving near fully supervised model performance with just 15% labeled data.Performance still lags behind fully supervised methods on some datasets.
DEC-Seg [80]Res2NetChanges in the shape, size, and location of polyps; pixel-by-pixel annotation.Significantly improves segmentation performance with 10% and 30% labeled data.Challenges remain in handling fuzzy areas and unclear boundaries.
MEGANet [63]ResNet-34 and Res2Net-50Complex background, varying sizes and shapes of polyps, unclear boundaries.Edge-guided attention; Laplacian operator to enhance feature extraction of polyp boundaries.Insufficient real-time performance.
NPD-Net [82]Res2Net50Overfitting.Unsupervised deep supervision strategy.The segmentation ability will decrease on abnormal images.
DCL-PS [87]ResNet-101Distribution differences between the source domain and the target domain.Better cross-domain feature alignment.The edge details of large polyps are not captured accurately enough.
DLCA_
CMAM [69]
ConvNeXtThe color appearance in the polyp dataset is imbalanced.Color statistics knowledge for data augmentation.DLCA requires more computation time compared to Color Jittering.
MISNet [64]Res2NetThe location, size, and shape of the polyps vary; the boundaries are unclear.Adaptive fusion of multi-layer features; accurately capturing the details of polyp boundaries.Segmentation performance is poor in extremely low-contrast environments.
WDFF-Net [24]HarDNet68Polyps exhibit significant differences in color, size, shape, appearance, and location.The dual-branch architecture (PFF and SFF) has strong complementarity.The FPS performance is slightly inferior to that of lightweight models like HarDNet-MSEG.
UHA-Net [70]Res2Net-50Scale variation of lesion areas.The segmentation effect of small polyps is good.Complete pixel-level annotated data is needed for training.
SEPNet [71]PVTv2-B2Difficult to learn accurate semantics.Handling polyps with similar backgrounds.Loss of edge details.
FoBS [65]DeepLabV3+The appearance of polyps varies.Multilevel boundary enhancement.Noise-sensitive.
2023S2ME [84]ResNet50Existing weak supervision methods are mainly limited to spatial domain learning.Spatial-frequency dual branch; pixel-level adaptive fusion strategy.Lack of comparison with the latest methods.
EMTSNet [134]Res2Net50How to accurately segment polyps.Effective extraction of multi-scale features.The segmentation results for small-sized polyps are not satisfactory.
GAN-PSNet [135]Res2Net50Polyps with variable appearance.Attention mechanism integrated into the discriminator.Higher time complexity.
FEGNet [72]Res2Net50The shape, size, and texture of polyps vary; the boundaries are unclear.Skilled in handling the segmentation of complex-shaped and small-sized polyps.Insufficient generalization performance.
CoInNet [73]DenseNet121The shape, color, size, and texture of polyps vary, and their boundaries are unclear.It has excellent segmentation capability for small polyps with an area of ≤1000 pixels.Detection difficulties in specific scenarios.
RPANet [88]ResNet101Domain adaptation for polyp segmentation models.No source domain data is required, suitable for cross-hospital deployment.Insufficient handling of situations where the foreground and background have minimal differences.
CFA-Net [66]Res2Net-50Polyps vary in shape and size, with indistinct boundaries.Effective in handling small polyps and unclear boundaries.The segmentation effect of large polyps is not ideal.
PolypSeg+ [127]ResNet50Polyps are diverse, with low contrast against the background and blurred boundaries.High segmentation accuracy; good real-time performance.The segmentation effect of very small polyps is poor.
FTMF-Net [79]Res2NetBoundary details; global context information.Boundary feature extraction from frequency domain perspective.The segmentation of complex samples still needs improvement.
ISCNet [74]ResNet34Low contrast, small size, and a wide variety.Performs well when treating small polyps.Segmentation errors still exist in the edge regions.
I-RA [77]Res2Net50Representation of directionality, singularity, and regularity in the spectral domain.Contour knowledge guidance enhances spectral domain feature representation.Real-time performance still needs improvement.
WS-
DefSegNet [81]
Res2netPixel-wise annotated dataset.Weakly supervised and semi-supervised polyp segmentation framework.The performance on certain datasets is still lower than that of state-of-the-art fully supervised methods.
ERDUnet [23]UnetExtracting global contextual features is difficult; the parameters are too large to be applied clinically.The model has a small parameter size and achieves an average FPS of 40.71, accurately segmenting large-scale targets.Recognition errors may occur when the target and background features are too similar.
2022LDNet [75]Res2Net50The diversity of polyps in shape, size, and brightness, as well as their subtle contrast with the background.Dynamic kernel generation and update mechanism; efficient self-attention and lesion-aware cross-attention.The real-time performance and computational efficiency of the model have not been discussed in detail.
MSRF-Net [76]N/ATargets of different sizes; difficult to train on small-scale and biased datasets.The dual-scale dense fusion; the shape stream.The performance of low-contrast image segmentation is weak.
BSCA-Net [68]Res2Net50Extraction of boundary information.Obtain geometric information from multiple angles.Loss of some low-level details
AMNet [78]Res2Net50Small polyps; similar to the surrounding environment.Multi-scale fusion; parallel attention mechanism; reverse context fusion.Performance still needs to be improved in complex real-world environments.
BCNet [67]Res2Net50Inaccurate segmentation.Excels at handling ambiguous boundaries and complex backgrounds.There are still cases of inaccurate segmentation in specific complex scenarios.
NIP [83]ResNet101Limited training data, significant variation in polyps, and class imbalance.MLMix data augmentation methods and CAR resampling strategies.Real-time performance has not been evaluated in detail.
TGANet [136]ResNet50Changes in polyp size can affect model training.Outstanding performance in the segmentation of small and flat polyps.Real-time performance metrics not discussed in detail.
Transformer2024CTNet [28]MiT-b3Lack of features with advanced semantic details.Supervised contrastive learning strategy.Small polyp segmentation is weak.
HST-MRF [99]Swin TransformerFeature information loss problem caused by patch segmentation.Multi-receptive field patch segmentation strategy.Relatively weak in precision metrics.
MLFF-Net [90]PVTv2Insufficient feature utilization and feature fusion conflicts.Multi-layer feature fusion and multi-scale attention.The segmentation effect of multiple polyps is not ideal.
PSTNet [91]Shunted TransformerFeature misalignment in multi-scale aggregation.Enhance noise suppression capabilities.Need to improve computational efficiency.
Polyp-LVT [3]PVTv2Application limitations in clinical settings.The parameter count is reduced by approximately 44% compared to the baseline model Polyp-PVT.There is still room for improvement in accuracy.
PFENet [92]PVTv2-B2Limitations of global context modeling and cross-layer feature interactions.The CFE module enhances feature representation capabilities; the CMG module fuses features.Decline in out-of-domain generalization performance.
VANet [96]CvT-13Significant differences in appearance; unclear boundaries.VAFormer reduces interference from non-polyp regions; BAFormer optimizes boundary segmentation.Segmentation performance decreases for polyps that are too small (<5%) or too large (>25%).
PP-SAM [97]ViTDependence on large amounts of labeled dataThe PP-SAM framework enhances SAM’s robustness in polyp segmentation.Only supports binary classification and single bounding box prompts.
2023Polyp-PVT [93]PVTv2Differences between features at different levels; feature fusion.Strong feature extraction capability.Boundary detection has its limitations.
PVT-
CASCADE [27]
PVTv2 + TransUNetLimitations of Transformer models.Multi-stage loss and feature aggregation framework.Generalization requires more experiments.
CAFE-Net [94]PVTv2The ability to aggregate multi-scale features is limited.The MFA module maximizes the utilization of previously learned features.The efficiency of calculations needs improvement.
DSNet [137]PVTv2False positive/negative interference.Robust performance.False positives can occur with multiple image noise.
LACINet [95]Shunted Transformer-SComplex backgrounds interfere with pixel-level prediction performance.The LPM mechanism effectively reduces noise interference and redundant information.False detections or missed detections may occur under low light or overexposure conditions.
TransRUPNet [98]pvt_v2_b2Real-time polyp segmentation.Real-time processing speed 47.07 FPS.Performance significantly declines on out-of-distribution datasets.
2022MSRAformer [89]Swin TransformerDistinguish between polyp area and boundary.Enhance the edge details of the segmentation target.There is still room for improvement in the segmentation of small lesion areas.
SSFormer [26]PVTv2The polyps vary in size and have complex shapes; their borders are unclear.Strong generalization ability; progressive local decoder.The findings on previously unseen medical images still require further validation.
DS-TransUNet [100]SwinIgnored the pixel-level intrinsic structural features within each block.Dual-scale encoding mechanism simultaneously captures coarse-grained and fine-grained features.Performance is limited when using the same patch size.
Hybrid Architecture2024CIFG-Net [111]PVTv2The boundaries are unclear.Improved mDice by 3.9% and mIoU by 4.2% compared to PraNet.Inference speed and resource consumption not mentioned.
MVOA-Net [101]PVTv2-b2Intra-class inconsistency; inter-class indistinction.Skilled in multi-target polyp segmentation.The computational efficiency is not competitive.
RSAFormer [102]PVTv2-B4Boundary similarity.RSA module can better extract boundary information.Performs poorly when samples have reflections and shadows.
Polyper [112]Swin-TFuzzy Boundary.The segmentation effect for small polyps (image proportion <6%) is good.Unable to handle false positives or false negatives effectively.
TransNetR [117]ResNet50Inefficient real-time processing speeds.Processing speed reaches 54.60 FPS.Detection of tiny polyps still has room for improvement.
PFD-Net [103]PVTv2Failed to effectively utilize local details and global semantic information.Spatial-frequency joint method enhances the local and global features of feature maps.When the target and background are extremely similar, segmentation performance is poor.
HD-Former [104]ResNet-34Unable to account for multi-level dependencies between spatial and channel dimensions.Dual Cross-Attention Transformer (DCAT) module is used for multi-level feature fusion.Increased computational overhead.
ASPS [115]ViT-B
+MSCAN-L
Ignoring local details.Enhanced OOD performance and domain generalization.There is still room for improvement in the segmentation of small polyps.
PGCF [138]PVTv2Recognition of complex features.PGCF module extracts multi-scale features.Inference time is slow, affecting real-time performance.
FMCA-Net [107]PVT_v2Blurred edges; insufficient feature extraction.The D-BFRM module extracts and enhances polyp features.The segmentation of small polyps is not ideal.
SAM2-Unet [116]HieraDesign a simple and efficient universal segmentation framework.Insert an adapter into the encoder to achieve parameter-efficient fine-tuning.The performance on some datasets is slightly below that of other methods.
SimTxtSeg [114]ConvNeXt-Tiny +BERT-BASEReducing annotation costs while maintaining segmentation performance.Text-to-visual prompt converter and text–visual hybrid attention module.Pseudo-label generation relies on the SAM model and is sensitive to its performance.
2023MC-DC [108]Res2Net-50+Wave-MLPDesign of the feature decoder.Multi-layer features that integrate MLP and CNN.The generalization ability needs improvement.
META-Unet [109]ResNet34Low contrast.The dual-branch structure can simultaneously capture global and local features.The accuracy of polyp segmentation with ambiguous boundaries is low.
2022SwinPA-Net [105]Swin-BThe size and type of lesions vary too much.The dense multiplicative connection module is used for multi-scale feature fusion.Real-time constraints.
HSNet [106]PVTv2+ Res2Net50Ignore the visual details of small polyps.Interactive attention mechanism; MSP module integrates different scales.Detection of small polyps still needs improvement.
PPFormer [113]CvT+VGG-16Transformer has insufficient local feature extraction.PP-guided self-attention mechanism; local-to-global mechanism.Small polyp segmentation is weak.
FCBFormer [110]PVTv2-B3Only able to predict low-resolution segmentation maps.FCN and Transformer parallel branches.Data dependency.
Mamba2024VM-UNetV2 [119]VSSLimitations of CNN and Transformer.VSS blocks capture long-range dependencies.Robustness still needs to be further validated.
Prompt-Mamba [118]vision-mambaPolyps of various shapes and colors.Image feature extraction in Visual Mamba.Performance is slightly lower on some datasets.
Polyp-Mamba [121]ResNet34+
Mamba
The polyp has blurred margins.Mamba and ResNet extract global and local features.Misjudgments may occur in situations with clear boundaries.
Polyp-Mamba [29]VSSCross-scale dependencies and the consistency of feature representations and semantic embeddingsThe SAS module facilitates the interaction of multi-scale semantic information.Limited validation exists for segmentation of small or complex polyp structures
H-vmunet [122]N/ADefects of the SS2D module.The H-SS2D module reduces redundant information.Lack of analysis of model robustness.
Other2024SAM 2 [125]N/ADepends on extensive dataset labeling.Strong zero-shot learning capabilities.The effects vary significantly between different prompts.
DDPM [124]U-Net ϵθEfficient and accurate polyp segmentation.Majority voting strategy enhances performance.Generalization performance needs improvement.
dHBLSN [126]BLS(Broad Learning System)Cost calculation.BLS does not require multiple layers to learn complex features.Difficult to remove complex flat polyps.
HDM [123]Diffusion U-NetDomain disparities; low efficiency.Generation of prior features for polyp segmentation using diffusion models.Unclear boundaries are difficult to accurately segment.
Table 3. Summary of polyp segmentation methods from 2018 to 2021.
Table 3. Summary of polyp segmentation methods from 2018 to 2021.
ArchitectureYearMethodsBackboneProblems SolvedAdvantagesLimitations
Hybrid Architecture2021TransFuse [21]ResNet-34+ DeiT -SEfficiency in modeling the global context; low-level details.Parallel branches; the BiFusion module fuses multi-level features.Performance on extremely small target datasets has not been discussed in detail.
CNN2020PraNet [20]Res2NetPolyps exhibit diverse shapes and have indistinct margins.Parallel partial decoder and reverse attention module.Detail segmentation remains inadequate in complex boundary situations.
CNN2019ResUNet++ [19]ResUNetHow to accurately segment polyps.ASPP improves the ability to segment polyps of various shapes and sizes.Resizing the image may cause some details to be lost.
CNN2018Unet++ [18]UNetThere are semantic differences.Re-designed skip pathways.High computational complexity.
Table 4. Summary of video polyp segmentation methods.
Table 4. Summary of video polyp segmentation methods.
ArchitectureYearMethodsBackboneProblems SolvedAdvantagesLimitations
CNN2022PNS+ [22]Res2Net50Lack of large-scale datasets with fine-grained segmentation annotations.Constructed the SUN-SEG dataset; achieved an inference speed of 170 frames per second.Insufficient robustness.
CNN2024SSTFB [131]Res2Net-50Decline in video polyp segmentation performance.Enhancing representation learning with a spatio-temporal self-attention mechanism.Some mis-segmentation may occur in complex backgrounds.
Mamba2024Vivim [133]Temporal Mamba BlockVanilla SSMs cannot preserve non-causal spatial information.Temporal Mamba Block and ST-Mamba Module.The model is relatively complex.
Hybrid2024MAST [130]PVTv2-B2Modeling long-range spatiotemporal relationships is challenging.The Siamese Transformer and hybrid attention module capture the spatiotemporal relationships between video frames.Generalization ability needs to be verified.
Hybrid2024FlowICBNet [128]PVTPredictive discontinuity problems caused by low-quality frames.RFS and FGW modules select the optimal historical reference frame.The IFU exhibits an over-correction phenomenon.
Hybrid2024SALI [132]PVTv2Low-quality frames limit segmentation accuracy.The SAM module alleviates spatial variations; the LIM module reconstructs polyp features.Segmentation remains inaccurate for frames with drastic changes.
Hybrid2024Diff-VPS [129]N/AHighly camouflaged polyps and redundant temporal clues.Multi-task diffusion model; the TRM module captures dynamic video features.There is still room for improvement in detecting highly concealed polyps.
Table 6. Definitions of TP, FP, TN, and FN.
Table 6. Definitions of TP, FP, TN, and FN.
IndicatorDefinitionMeaning
True Positive (TP)Correctly predicted as positive class.It is a polyp and has been correctly predicted.
False Positive (FP)Incorrectly predicted as positive class.Falsely predicted as a polyp, but not actually a polyp.
True Negative (TN)Correctly predicted as negative class.Not a polyp and correctly predicted.
False Negative (FN)Incorrectly predicted as negative class.It is a polyp but was incorrectly predicted not to be a polyp.
Table 7. Comparative study of performance evaluation of various models for polyp segmentation. The top 2 methods are marked in Green and Red, respectively.
Table 7. Comparative study of performance evaluation of various models for polyp segmentation. The top 2 methods are marked in Green and Red, respectively.
MethodCVC-ClinicDBKvasir-SEGCVC-ColonDBETIS-LaribEndoScene
mDicemIoUmDicemIoUmDicemIoUmDicemIoUmDicemIoU
MEGANet [63]0.9380.8940.9130.8630.7930.7140.7390.6650.8990.834
Polyper [112]0.9450.8990.9480.9040.8370.7460.8650.7850.9240.867
Polyp-Mamba [29]0.9490.9070.9400.8810.8290.7430.8250.7470.9210.875
ISCNet [74]0.9610.9200.9390.8980.8280.7410.8040.7160.8710.778
CTNet [28]0.9360.8870.9170.8630.8130.7340.8100.7340.9080.844
PraNet [20]0.8990.8490.8980.8400.7090.6400.6280.5670.8710.797
CFA-Net [66]0.9330.8830.9150.8610.7430.6650.7320.6550.8930.827
DLCA_CMAM [69]0.9440.9000.9290.8820.8180.7450.8390.7660.9030.837
Polyp-PVT [93]0.9370.8890.9170.8640.8080.7270.7870.7060.9000.833
CAFE-Net [94]0.9430.8990.9330.8890.8200.7400.8220.7380.9010.834
UNet [15]0.8230.7550.8180.7460.5120.4440.3980.3350.7100.627
TGANet [136]0.9460.8990.8980.8330.7550.8240.6360.7820.8850.899
HST-MRF [99]0.9350.8840.9280.8850.8310.7760.7730.7190.9010.846
DS-TransUNet [100]0.9360.8870.9350.8890.7980.7220.7610.6870.9110.846
TransFuse [21]0.9420.8970.9200.8700.7810.7060.7370.6630.8940.826
MVOA-Net [101]0.9470.9020.9350.8910.8240.7450.8200.7430.9040.838
SwinPA-Net [105]0.9410.8940.9250.8760.8070.7260.7620.6840.8930.823
HSNet [106]0.9480.9050.9260.8770.8100.7350.8080.7340.9030.839
SAM 2 [125]0.9300.8700.9390.8850.9340.8770.9410.890--
PFD-Net [103]0.9390.8930.9280.8800.8160.7370.8260.746--
CoInNet [73]0.9300.8870.9260.8720.7970.7290.7590.690--
AMNet [78]0.9360.8880.9120.8650.7620.6900.7560.679--
DEC-Seg [80]0.8590.8040.8930.8300.7210.6400.6340.564--
WS-DefSegNet [81]0.8070.7460.7680.7090.6670.5880.5960.517--
NPD-Net [82]0.9250.8760.9050.8500.7640.6820.7370.659--
MISNet [64]0.9180.8690.9030.8460.7620.6900.7640.686--
UHA-Net [70]0.9270.8810.9080.8570.7690.6950.7460.670--
SEPNet [71]0.9370.8870.9220.8690.8190.7400.7950.718--
FEGNet [72]0.9430.9010.9230.8740.7670.6860.7190.645--
MLFF-Net [90]0.9430.8970.9190.8660.8200.7420.7840.707--
PSTNet [91]0.9450.9010.9350.8950.8270.7480.8000.726--
PFENet [92]0.9400.8970.9310.8860.8210.7450.8090.735--
FMCA-Net [107]0.9440.8980.9190.8660.8040.7230.8410.765--
FCBFormer [110]0.9470.9020.9390.8900.7830.7060.7960.715--
CIFG-Net [111]0.9380.8910.9250.8760.8150.7330.8060.726--
SAM2-UNet [116]0.9070.8560.9280.8790.8080.7300.7960.723--
Prompt-Mamba [118]0.8880.8140.8860.8080.8200.7120.7710.663--
VM-UNetV2 [119]0.9440.8930.9130.8420.7580.6100.8390.723--
Polyp-Mamba [121]0.9410.8960.9190.8670.7910.7130.7560.668--
RPANet [88]0.8000.7190.8580.782--0.6320.552--
TransNetR [117]0.7660.6910.8710.802------
DDPM [124]0.9670.9370.9340.886------
TransRUPNet [98]0.8540.7770.9010.845------
Polyp-LVT [3]0.9350.8820.9090.851----0.9040.835
Table 8. Comparative study of various models in terms of real-time performance.
Table 8. Comparative study of various models in terms of real-time performance.
ModelsParams (M)Flops (G)FPS (Frames/s)
MISNet [64]33.6345.9430.68
CFA-Net [66]25.2455.3623.50
SEPNet [71]25.9612.5262.00
TGANet [136]42.3019.8485.00
MSRF-Net [76]18.3820.2614.38
I-RA [77]35.8515.652.06
PraNet [20]30.5013.0837.31
LACINet [95]22.3611.7428.41
MVOA-Net [101]27.8629.7330.78
TransNetR [117]27.2710.5854.60
Polyp-Mamba [121]49.5027.9023.80
ERDUnet [23]10.2110.3027.03
PolypSeg+ [127]2.547.2331.00
FlowICBNet [128]98.40N/A29.00
SSTFB [131]33.40N/A126.00
UNet [15]34.52N/A55.10
WDFF-Net [24]17.46N/A83.82
META-Unet [109]27.32N/A75.00
Polyp-PVT [93]125.60N/A66.00
MC-DC [108]65.9341.13N/A
FTMF-Net [79]44.778.85N/A
Polyp-LVT [3]25.1113.21N/A
NPD-Net [82]29.2214.51N/A
CTNet [28]44.1915.20N/A
FMCA-Net [107]28.6114.36N/A
PFD-Net [103]28.8916.81N/A
ISCNet [74]23.3415.57N/A
CAFE-Net [94]35.5316.12N/A
DS-TransUNet [100]287.7551.09N/A
MAST [130]25.6921.02N/A
Prompt-Mamba [118]102.00N/AN/A
HSNet [106]29.23N/AN/A
MEGANet [63]44.19N/AN/A
TransFuse [21]26.30N/AN/A
HST-MRF [99]174.73N/AN/A
TransRUPet [98]N/AN/A47.07
Note: N/A indicates data not available.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, X.; Isa, N.A.M.; Chen, C.; Lv, F. Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review. J. Imaging 2025, 11, 293. https://doi.org/10.3390/jimaging11090293

AMA Style

Liu X, Isa NAM, Chen C, Lv F. Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review. Journal of Imaging. 2025; 11(9):293. https://doi.org/10.3390/jimaging11090293

Chicago/Turabian Style

Liu, Xin, Nor Ashidi Mat Isa, Chao Chen, and Fajin Lv. 2025. "Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review" Journal of Imaging 11, no. 9: 293. https://doi.org/10.3390/jimaging11090293

APA Style

Liu, X., Isa, N. A. M., Chen, C., & Lv, F. (2025). Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review. Journal of Imaging, 11(9), 293. https://doi.org/10.3390/jimaging11090293

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop