Next Article in Journal
Clinicians’ Agreement on Extrapulmonary Radiographic Findings in Chest X-Rays Using a Diagnostic Labelling Scheme
Next Article in Special Issue
Advancing Colorectal Cancer Diagnostics from Barium Enema to AI-Assisted Colonoscopy
Previous Article in Journal
A Multicentric Study on Adverse COVID-19 Outcomes Among Pregnant and Nonpregnant Women in Multidisciplinary Hospitals of Kazakhstan
Previous Article in Special Issue
Brain Morphometry and Cognitive Features in the Prediction of Irritable Bowel Syndrome
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor

1
Department of Internal Medicine, Division of Gastroenterology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul 06273, Republic of Korea
2
Waycen Inc., Seoul 06167, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
These authors contributed equally to this work.
Diagnostics 2025, 15(7), 901; https://doi.org/10.3390/diagnostics15070901
Submission received: 10 March 2025 / Revised: 25 March 2025 / Accepted: 31 March 2025 / Published: 1 April 2025
(This article belongs to the Special Issue Application of Artificial Intelligence in Gastrointestinal Disease)

Abstract

:
Background/Objectives: Artificial intelligence (AI) integration in colon polyp detection often exhibits high sensitivity but notably low specificity in real-world settings, primarily due to reliance on publicly available datasets alone. To address this limitation, we proposed a semi-automatic annotation method using real colonoscopy videos to enhance AI model performance and reduce manual labeling labor. Methods: An integrated AI model was trained and validated on 86,258 training images and 17,616 validation images. Model 1 utilized only publicly available datasets, while Model 2 additionally incorporated images obtained from real colonoscopy videos of patients through a semi-automatic annotation process, significantly reducing the labeling burden on expert endoscopists. Results: The integrated AI model (Model 2) significantly outperformed the public-dataset-only model (Model 1). At epoch 35, Model 2 achieved a sensitivity of 90.6%, a specificity of 96.0%, an overall accuracy of 94.5%, and an F1 score of 89.9%. All polyps in the test videos were successfully detected, demonstrating considerable enhancement in detection performance compared to the public-dataset-only model. Conclusions: Integrating real-world colonoscopy video data using semi-automatic annotation markedly improved diagnostic accuracy while potentially reducing the need for extensive manual annotation typically performed by expert endoscopists. However, the findings need validation through multicenter external datasets to ensure generalizability.

1. Introduction

Artificial intelligence (AI) integration into medical imaging, particularly colonoscopy image analysis, has become increasingly relevant in clinical practice due to its potential to enhance polyp detection and diagnosis [1,2,3]. AI models supporting colonoscopy procedures generally fall into two categories: computer-aided detection (CADe), which focuses on detecting abnormalities, such as polyps, and computer-aided diagnosis (CADx), which classifies detected lesions based on their characteristics [4].
Despite advancements in AI models for colonoscopy, the lack of easily accessible, high-quality data for training and validating AI models remains a challenge [5]. Models trained exclusively on publicly available datasets frequently encounter difficulties due to discrepancies in image quality, diversity, and differences in clinical equipment, resulting in high sensitivity but notably low specificity in clinical settings [6,7].
This study aimed to address these limitations through a novel semi-automatic annotation method utilizing real patient colonoscopy videos from Gangnam Severance Hospital. Unlike previous methods relying heavily on manual annotation by expert endoscopists, our approach leveraged preliminary AI inference results for efficient selection of clinically relevant video frames, substantially reducing annotation effort. Although semi-automatic annotation is a common method in general computer vision tasks, our specific approach—AI-driven preliminary selection from colonoscopy video data—is designed explicitly for clinical practice integration.
This study demonstrates the feasibility of developing an accurate and efficient AI polyp detection model by integrating real-world clinical data, significantly reducing manual annotation burdens. Nevertheless, due to the single-center nature of the data collection, additional multicenter validation is necessary to ensure broader applicability.

2. Materials and Methods

2.1. Patients

Colonoscopy video data were retrospectively collected from patients who underwent routine colonoscopies at Gangnam Severance Hospital, a tertiary academic center, between April 2021 and April 2022. Inclusion criteria consisted of adults (≥18 years old) undergoing colonoscopy for indications including colorectal cancer screening, surveillance, or evaluation of gastrointestinal symptoms. Videos with poor bowel preparation were excluded. Personally identifiable information was anonymized prior to analysis.
The dataset included 117 colonoscopy videos totaling 30 h, 57 min, and 46 s, averaging approximately 15 min and 53 s per patient. Overall, the dataset comprised 3,348,994 frames, averaging 28,624 frames per video.

2.2. Public Datasets

Representative public datasets used for colonoscopy research include ETIS-Larib [8], CVC-ClinicDB [9], KVASIR-SEG [10], LDPolypVideo [11], KUMC [12], and PolypGen [13]. Each dataset offered unique features and contributions to polyp detection and segmentation tasks: (a) ETIS-Larib Polyp DB contains 196 image frames with binary masks of polyps. (b) CVC-ClinicDB provides 612 images with binary masks extracted from 29 colonoscopy videos. (c) Kvasir-SEG offers 1000 polyp images with corresponding masks. (d) LDPolypVideo includes 160 labeled and 42 unlabeled videos, containing 33,884 images annotated with bounding boxes. (e) KUMC provides 37,899 frames with bounding box annotations. (f) PolypGen contains 1537 polyp images, each accompanied by binary masks.

2.3. Data Processing

Colonoscopy videos were used to develop two deep-learning-based AI models. The collected colonoscopy videos (n = 117) from Gangnam Severance Hospital were randomly allocated into three groups: training (57 videos), validation (15 videos), and test sets (45 videos), using a computer-generated randomization procedure to minimize selection bias. The extracted frames from these videos, combined with publicly available datasets (ETIS-Larib, CVC-ClinicDB, KVASIR-SEG, LDPolypVideo, KUMC, PolypGen), were utilized to construct comprehensive image datasets for model training, validation, and testing. The detailed distribution of images across the datasets is summarized in Table 1.
Two deep-learning AI models were developed to compare the effects of integrating real-world clinical data. Model 1 was trained exclusively using publicly available datasets, while Model 2 incorporated additional images obtained from colonoscopy videos collected at Gangnam Severance Hospital. Both models were trained using images containing polyps and normal colonoscopic images without polyps, as outlined in Table 2. The final evaluation was performed using the best-performing model, as determined by accuracy on the validation set.

2.4. Semi-Automatic Image Selection

The proposed development process is illustrated in Figure 1. Initially, Model 1 was trained and validated exclusively using public datasets. Subsequently, Model 1 was used to infer frames from 72 colonoscopy videos (57 training, 15 validation) collected from Gangnam Severance Hospital. In this inference step, clinical timestamps indicating the actual removal of polyps during colonoscopy were indirectly used to guide the AI model in identifying frames likely containing polyps. The inference results were automatically categorized into subfolders as frames containing polyps and frames without polyps. Each detected polyp area was highlighted with bounding boxes on corresponding video frames and saved separately, facilitating efficient review by researchers. Frames incorrectly classified by Model 1—such as missed polyps (false negatives) or incorrectly identified polyps (false positives)—were also selectively included to enhance subsequent model training. This combined inference and data-selection approach, termed the semi-automatic image selection process, substantially reduced manual annotation effort, enabling rapid construction of the training dataset for Model 2.
Figure 2 shows representative examples of inference results from Gangnam Severance Hospital using Model 1 trained on public datasets. Each row represents a sequence of consecutive frames. Frames with blue borders indicate instances where polyps were inaccurately detected or completely missed. These misdetected frames were selectively incorporated into the training dataset to enhance the subsequent learning of Model 2.

2.5. Algorithm Training, Validation, and Test Sets

A YOLO-based object detection model, a deep learning approach, was employed to detect polyps in the colonoscopy images. Previous studies have shown that various backbones based on convolutional neural networks (CNNs) exhibit similar performance when applied to endoscopic images [2,3,6,12]. Many existing AI models used for colonoscopy have initially been developed for general image recognition tasks and subsequently adapted for specific object detection tasks, including polyp detection.
In this study, we adopted a CNN-based polyp detection model utilizing the Visual Geometry Group (VGG)16 backbone. The VGG16 backbone has been widely applied in medical image analysis and has demonstrated reliable performance in extracting relevant image features in various endoscopic imaging tasks [14,15].
To train Model 1, we augmented the training data (public datasets only) by a factor of 10 using various data augmentation methods, including vertical and horizontal flips, left–right symmetry, random cropping, and rotation. Input image size was standardized at 224 × 224 pixels, and polyp masks and bounding box annotations were consistently preserved during augmentation. Model 1 was trained for 50 epochs with a VGG16 backbone followed by fully connected layers, optimized using the Adam optimizer with an initial learning rate of 1 × 10−5. Cross-entropy and activation errors served as error functions for backpropagation.
Model 2 integrated both public datasets and additional images derived from colonoscopy videos obtained at Gangnam Severance Hospital. Frames from these videos were efficiently annotated using a semi-automatic selection process based on preliminary inference by Model 1. Model 2 underwent similar training procedures as Model 1, with validation performance assessed at each epoch. The optimal performing model on the validation set was selected for subsequent testing.
The test dataset comprised 4373 previously unseen images, including 1200 polyp-containing and 3173 non-polyp-containing images, extracted from 45 separate test videos. Specifically, 120 polyps were identified in these test videos, and approximately 10 sequential frames per polyp (at 5-frame intervals, ~1.6 s duration per polyp) were extracted for performance evaluation.

2.6. Evaluations

Several metrics were used to evaluate the model performance [16,17]. True positives (TP) represent the number of polyps that were accurately detected by the model. The true negative (TN) indicates the number of normal images correctly classified as normal. A false negative (FN) refers to instances in which a polyp was present but not detected by the model, whereas a false positive (FP) denotes cases in which a polyp was incorrectly identified in a normal image.
Sensitivity (or recall) reflects the proportion of actual polyps that were correctly identified by the model and was computed as TP divided by the sum of TP and FN. Specificity represents the proportion of actual normal images that were correctly classified and was calculated as TN divided by the sum of TN and FP. Accuracy measures the overall proportion of correctly classified images, both polyps and normal, and was determined by dividing the sum of TP and TN by the total number of images (TP + TN + FP + FN). Additionally, the F1 Score, which is the harmonic means of sensitivity and positive predictive value (PPV), provides a balanced measure of precision and recall. The F1 Score was calculated as twice the product of sensitivity and PPV divided by the sum of sensitivity and PPV, where PPV was calculated as TP divided by the sum of TP and FP.
  • S e n s i t i v i t y = T P / T P + F N
  • S p e c i f i c i t y = T N / ( T N + F P )
  • A c c u r a c y = ( T P + T N ) / ( T P + T N + F P + F N )
  • P P V = T P / T P + F P
  • F 1   s c o r e   =   2   ×   ( S e n s i t i v i t y   ×   P P V ) / ( S e n s i t i v i t y   +   P P V )

3. Results

The performance of the proposed method was evaluated using the model from the epoch that demonstrated the highest sensitivity and specificity of the validation data. This evaluation was performed over 50 epochs for Model 1, which was trained solely with public datasets, and Model 2, which was trained with both public and additional collected data. Model 1 achieved optimal performance at epoch 37 based on validation accuracy, whereas Model 2’s best performance occurred at epoch 35. Table 3 presents the polyp detection performance on the test data using the model saved at epoch 37 for Model 1 and the models saved at 5-epoch intervals from epoch 5 to epoch 45 for Model 2. Figure 3 demonstrates stable learning curves for both models without signs of overfitting, indicating effective generalization throughout the training process. Specifically, across 50 epochs, Model 1 showed a mean validation accuracy of 0.943 (standard deviation [SD] = 0.013), and Model 2 showed a mean validation accuracy of 0.939 (SD = 0.012). Since Model 1 exhibited its highest validation accuracy at epoch 37, this epoch was selected to compare performance against Model 2, trained with integrated data. According to Table 3, Model 2 consistently outperformed Model 1 across most evaluated metrics at the selected epochs, confirming the benefits of incorporating additional colonoscopy video data into the training process.
The test results revealed that Model 1, trained only with public datasets, had a sensitivity of 0.778 for detecting polyps, indicating reasonable performance. However, it exhibited a low specificity of 0.059 due to significant false detections, which were attributed to differences in image quality between the collected data and public datasets. In contrast, Model 2, which incorporated additional collected data into the training process, demonstrated an improved performance by addressing these environmental differences. Specifically, Model 2 achieved a sensitivity of 0.906, a specificity of 0.960, and an accuracy of 0.945 at epoch 35, reflecting a notable enhancement in detection capability.
In real-time colonoscopy applications, the polyp detection model must minimize false detections in normal images and ensure that all polyps are accurately identified to prevent them from being overlooked. Frequent false detections can lead to examiner fatigue and reduce the reliability of the detection model. Conversely, accurate detection of polyps occasionally provides opportunities to identify and remove lesions that might otherwise be missed. Table 4 details the number and detection rates of polyps that were never detected out of the 120 polyps, each represented by 10 images, amounting to 1200 test images. Figure 4 shows examples of the 120 undetected polyps. Figure 5 shows the number of polyps accurately detected in each epoch. The results indicated that Model 2, which utilized additional collected data, consistently demonstrated higher sensitivity than Model 1, which relied solely on public datasets. This suggests that Model 2 offers an improved continuity of polyp detection throughout the video sequences.

4. Discussion

Recent advances in AI-based CADe and CADx have significantly improved the detection and characterization of colorectal polyps during colonoscopy, potentially enhancing clinical outcomes through earlier and more accurate diagnosis [18,19]. However, AI models trained exclusively on publicly available datasets frequently demonstrate limited specificity when applied to real-world clinical scenarios, mainly due to differences in image quality, diversity, and endoscopic equipment [20,21].
In this study, we proposed a simple and efficient approach to enhance the performance of AI models by integrating real-world colonoscopy video data into the training process. A distinctive feature of our method was the semi-automatic annotation technique, which utilized preliminary inference results from a publicly trained AI model to rapidly and automatically categorize colonoscopy video frames into polyp-containing and non-polyp-containing categories. Although semi-automatic annotation tools, such as CVAT, are widely used in computer vision, our approach specifically employed AI-guided inference to minimize manual intervention by expert endoscopists, thereby significantly reducing the traditionally labor-intensive annotation workload [22].
By integrating real patient colonoscopy data through this approach, we observed notable improvements in both sensitivity and specificity compared to a model trained solely on public datasets. This highlights that our method not only improves AI model performance but also offers practical advantages by substantially reducing manual annotation efforts.
However, several limitations of our study warrant careful consideration. First, this study used colonoscopy videos from a single institution employing a standardized endoscopic imaging system (Olympus Evia Exera III CV-190, Olympus Corporation, Tokyo, Japan). Therefore, generalizability of our results to other clinical environments and various endoscopic equipment has not yet been confirmed [23]. Future studies should include external validation using multi-center datasets to verify broader clinical applicability and reproducibility of results. Additionally, the study lacked detailed patient characteristics data, including age, gender, clinical indications, polyp morphology, size, location, and histopathological results. Consequently, we could not evaluate the AI model’s performance across different lesion types or diverse patient subpopulations, potentially limiting the clinical interpretation of our results. Further research incorporating comprehensive clinical and histopathological data is required to better understand the AI model’s performance across various lesion types and patient subpopulations, which will be crucial for identifying potential areas of improvement and enhancing clinical relevance. Furthermore, our evaluation focused primarily on per-frame metrics such as sensitivity and specificity, which may not fully reflect clinical relevance. To better evaluate AI systems in clinical settings, future studies should incorporate lesion-level sensitivity, precision metrics per detected polyp, and precision-recall curves or receiver operating characteristic (ROC) analyses.

5. Conclusions

In conclusion, our study introduces a practically meaningful, efficient, and scalable approach to significantly enhance AI model performance for colonoscopy polyp detection using real-world colonoscopy video data. The semi-automatic annotation method employed could substantially reduce the traditionally labor-intensive manual annotation process typically performed by expert endoscopists, making it an attractive strategy for practical clinical implementation. However, further external validation, detailed patient and lesion characteristic analyses, and lesion-level performance assessments are necessary to confirm and maximize the clinical utility and generalizability of our approach.

Author Contributions

Conceptualization, J.-H.K.; Data curation, Y.K., J.-S.K., S.-I.O. and K.-N.K.; Formal analysis, J.-S.K., S.-I.O. and K.-N.K.; Funding acquisition, J.-H.K.; Investigation, J.-H.K.; Methodology, J.-H.K.; Project administration, J.-H.K.; Resources, Y.K., J.-H.K., J.C., Y.-H.Y. and H.P.; Software, J.-S.K., S.-I.O. and K.-N.K.; Supervision, J.-H.K. and J.C.; Validation, J.-S.K.; Visualization, Y.K. and J.-S.K.; Writing—original draft, Y.K. and J.-S.K.; Writing—review & editing, Y.K., J.-S.K. and J.-H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by a grant from the Korean Gastrointestinal Endoscopy Research Foundation (2022 Investigation Grant).

Institutional Review Board Statement

This study was approved by the Institutional Review Board of the Gangnam Severance Hospital (approval no. 3-2022-0309, date: 28 September 2022).

Informed Consent Statement

The requirement for informed consent was waived due to the retrospective study design.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

Ji-Soo Keum, Sang-Il Oh, and Kyung-Nam Kim were employed by the company, Waycen Inc., Seoul 06167, Republic of Korea. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial intelligence
CADeComputer-Aided Detection
CADxComputer-Aided Diagnosis
CNNsConvolutional neural networks
VGGVisual Geometry Group
TPTrue positive
TNTrue negative
FNFalse negative
FPFalse positive
PPVPositive predictive value
ROCReceiver operating characteristics

References

  1. Tavanapong, W.; Oh, J.; Riegler, M.A.; Khaleel, M.; Mittal, B.; de Groen, P.C. Artificial Intelligence for Colonoscopy: Past, Present, and Future. IEEE J. Biomed. Health Inform. 2022, 26, 3950–3965. [Google Scholar] [CrossRef] [PubMed]
  2. Jha, D.; Ali, S.; Tomar, N.K.; Johansen, H.D.; Johansen, D.; Rittscher, J.; Riegler, M.A.; Halvorsen, P. Real-Time Polyp Detection, Localization and Segmentation in Colonoscopy Using Deep Learning. IEEE Access 2021, 9, 40496–40510. [Google Scholar] [CrossRef] [PubMed]
  3. Wan, J.; Chen, B.; Yu, Y. Polyp Detection from Colorectum Images by Using Attentive YOLOv5. Diagnostics 2021, 11, 2264. [Google Scholar] [CrossRef] [PubMed]
  4. Biffi, C.; Salvagnini, P.; Dinh, N.N.; Hassan, C.; Sharma, P.; Antonelli, G.; Awadie, H.; Bernhofer, S.; Carballal, S.; Dinis-Ribeiro, M.; et al. A novel AI device for real-time optical characterization of colorectal polyps. npj Digit. Med. 2022, 5, 84. [Google Scholar] [CrossRef] [PubMed]
  5. Ali, S. Where do we stand in AI for endoscopic image analysis? Deciphering gaps and future directions. npj Digit. Med. 2022, 5, 184. [Google Scholar] [CrossRef] [PubMed]
  6. Nogueira-Rodríguez, A.; Reboiro-Jato, M.; Glez-Peña, D.; López-Fernández, H. Performance of Convolutional Neural Networks for Polyp Localization on Public Colonoscopy Image Datasets. Diagnostics 2022, 12, 898. [Google Scholar] [CrossRef] [PubMed]
  7. Nogueira-Rodriguez, A.; Glez-Pena, D.; Reboiro-Jato, M.; Lopez-Fernandez, H. Negative Samples for Improving Object Detection-A Case Study in AI-Assisted Colonoscopy for Polyp Detection. Diagnostics 2023, 13, 966. [Google Scholar] [CrossRef] [PubMed]
  8. Silva, J.; Histace, A.; Romain, O.; Dray, X.; Granado, B. Toward embedded detection of polyps in WCE images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 283–293. [Google Scholar] [CrossRef] [PubMed]
  9. Bernal, J.; Tajkbaksh, N.; Sanchez, F.J.; Matuszewski, B.J.; Hao, C.; Lequan, Y.; Angermann, Q.; Romain, O.; Rustad, B.; Balasingham, I.; et al. Comparative Validation of Polyp Detection Methods in Video Colonoscopy: Results from the MICCAI 2015 Endoscopic Vision Challenge. IEEE Trans. Med. Imaging 2017, 36, 1231–1249. [Google Scholar] [CrossRef] [PubMed]
  10. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Halvorsen, P.; de Lange, T.; Johansen, D.; Johansen, H.D. Kvasir-SEG: A Segmented Polyp Dataset. In Proceedings of the MultiMedia Modeling: 26th International Conference, MMM 2020, Daejeon, Republic of Korea, 5–8 January 2020; Springer: Cham, Switzerland, 2020; pp. 451–462. [Google Scholar]
  11. Ma, Y.; Chen, X.; Cheng, K.; Li, Y.; Sun, B. LDPolypVideo Benchmark: A Large-Scale Colonoscopy Video Dataset of Diverse Polyps. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Cham, Switzerland, 2021; pp. 387–396. [Google Scholar]
  12. Li, K.; Fathan, M.I.; Patel, K.; Zhang, T.; Zhong, C.; Bansal, A.; Rastogi, A.; Wang, J.S.; Wang, G. Colonoscopy polyp detection and classification: Dataset creation and comparative evaluations. PLoS ONE 2021, 16, e0255809. [Google Scholar] [CrossRef] [PubMed]
  13. Ali, S.; Jha, D.; Ghatwary, N.; Realdon, S.; Cannizzaro, R.; Salem, O.E.; Lamarque, D.; Daul, C.; Riegler, M.A.; Anonsen, K.V.; et al. A multi-centre polyp detection and segmentation dataset for generalisability assessment. Sci. Data 2023, 10, 75. [Google Scholar] [CrossRef] [PubMed]
  14. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  15. Kim, J.H.; Oh, S.I.; Han, S.Y.; Keum, J.S.; Kim, K.N.; Chun, J.Y.; Youn, Y.H.; Park, H. An Optimal Artificial Intelligence System for Real-Time Endoscopic Prediction of Invasion Depth in Early Gastric Cancer. Cancers 2022, 14, 6000. [Google Scholar] [CrossRef] [PubMed]
  16. Canbek, G.; Sagiroglu, S.; Temizel, T.T.; Baykal, N. Binary classification performance measures/metrics: A comprehensive visualized roadmap to gain new insights. In Proceedings of the 2017 International Conference on Computer Science and Engineering (UBMK), Antalya, Turkey, 5–8 October 2017; pp. 821–826. [Google Scholar]
  17. Hicks, S.A.; Strümke, I.; Thambawita, V.; Hammou, M.; Riegler, M.A.; Halvorsen, P.; Parasa, S. On evaluation metrics for medical applications of artificial intelligence. Sci. Rep. 2022, 12, 5979. [Google Scholar] [CrossRef]
  18. Sinagra, E.; Badalamenti, M.; Maida, M.; Spadaccini, M.; Maselli, R.; Rossi, F.; Conoscenti, G.; Raimondo, D.; Pallio, S.; Repici, A.; et al. Use of artificial intelligence in improving adenoma detection rate during colonoscopy: Might both endoscopists and pathologists be further helped. World J. Gastroenterol. 2020, 26, 5911–5918. [Google Scholar] [CrossRef] [PubMed]
  19. Biscaglia, G.; Cocomazzi, F.; Gentile, M.; Loconte, I.; Mileti, A.; Paolillo, R.; Marra, A.; Castellana, S.; Mazza, T.; Di Leo, A.; et al. Real-time, computer-aided, detection-assisted colonoscopy eliminates differences in adenoma detection rate between trainee and experienced endoscopists. Endosc. Int. Open 2022, 10, E616–E621. [Google Scholar] [CrossRef] [PubMed]
  20. Wang, K.W.; Dong, M. Potential applications of artificial intelligence in colorectal polyps and cancer: Recent advances and prospects. World J. Gastroenterol. 2020, 26, 5090–5100. [Google Scholar] [CrossRef] [PubMed]
  21. Li, M.D.; Huang, Z.R.; Shan, Q.Y.; Chen, S.L.; Zhang, N.; Hu, H.T.; Wang, W. Performance and comparison of artificial intelligence and human experts in the detection and classification of colonic polyps. BMC Gastroenterol. 2022, 22, 517. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, W.; Tian, J.; Zhang, C.; Luo, Y.; Wang, X.; Li, J. An improved deep learning approach and its applications on colonic polyp images detection. BMC Med. Imaging. 2020, 20, 83. [Google Scholar] [CrossRef] [PubMed]
  23. Lee, J.Y.; Jeong, J.; Song, E.M.; Ha, C.; Lee, H.J.; Koo, J.E.; Yang, D.-H.; Kim, N.; Byeon, J.-S. Real-time detection of colon polyps during colonoscopy using deep learning: Systematic validation with four independent datasets. Sci. Rep. 2020, 10, 8379. [Google Scholar] [CrossRef]
Figure 1. Process of developing the initial model with public datasets and constructing additional datasets.
Figure 1. Process of developing the initial model with public datasets and constructing additional datasets.
Diagnostics 15 00901 g001
Figure 2. Examples of resulting images from inference in Model 1. The images include detected polyps (a) and normal tissue (b). Blue squares indicate data selected for further training.
Figure 2. Examples of resulting images from inference in Model 1. The images include detected polyps (a) and normal tissue (b). Blue squares indicate data selected for further training.
Diagnostics 15 00901 g002
Figure 3. Training loss and validation accuracy per epoch for Model 1 and Model 2.
Figure 3. Training loss and validation accuracy per epoch for Model 1 and Model 2.
Diagnostics 15 00901 g003
Figure 4. Example of polyps (blue circle) that were not detected by either Model 1 (a) or Model 2 (b) among 120 test polyps.
Figure 4. Example of polyps (blue circle) that were not detected by either Model 1 (a) or Model 2 (b) among 120 test polyps.
Diagnostics 15 00901 g004
Figure 5. The number of TPs for polyps in each epoch. The X-axis represents the number of TPs, while the Y-axis indicates the number of polyps.
Figure 5. The number of TPs for polyps in each epoch. The X-axis represents the number of TPs, while the Y-axis indicates the number of polyps.
Diagnostics 15 00901 g005
Table 1. Summary of the number of images used for training, validation, and test.
Table 1. Summary of the number of images used for training, validation, and test.
DatasetTraining
(polyps)
Validation
(polyps)
Test
(polyps)
Total
(polyps)
A. ETIS-Larib196
(196)
--196
(196)
B. CVC-ClinicDB612
(612)
--612
(612)
C. KVASIR-SEG900
(900)
--900
(900)
D. LDPolypVideo23,723
(1604)
6191
(462)
-29,914
(2066)
E. KUMC27,048
(27,048)
4214
(4214)
-31,262
(31,262)
F. PolypGen980
(980)
200
(200)
-1180
(1180)
G. Gangnam Severance Hospital32,799
(9353)
7011
(623)
4373
(1200)
44,183
(11,176)
Total86,258
(40,693)
17,616
(5499)
4373
(1200)
108,247
(47,392)
Table 2. Dataset configuration used to develop AI models.
Table 2. Dataset configuration used to develop AI models.
ModelDatasetNumber of Images (polyps)
TrainingValidationTestTotal
1A. + B + C + D + E + F
(Public dataset only)
53,459
(31,340)
10,605
(4876)
4373
(1200)
68,437
(37,416)
2A. + B + C + D + E + F + G
(Public + Collected dataset)
86,258
(40,693)
17,616
(5499)
4373
(1200)
108,247
(47,392)
Table 3. Comparison of polyp detection performance between the model trained only with public datasets and the integrated model that includes additional data from colonoscopy videos at Gangnam Severance Hospital.
Table 3. Comparison of polyp detection performance between the model trained only with public datasets and the integrated model that includes additional data from colonoscopy videos at Gangnam Severance Hospital.
Model 1Model 2
Epoch
37
Epoch
5
Epoch 10Epoch 15Epoch 20Epoch 25Epoch 30Epoch 35Epoch 40Epoch 45
TP933948101710569558951018108710511042
FN267252183144245305182113149158
FP3647537365427308160184132180148
TN230310431003100314131553149314231443144
Sensitivity0.7780.7900.8480.8800.7960.7460.8480.9060.8760.868
Specificity0.0590.8530.8950.8790.9110.9520.9450.9600.9460.955
PPV0.2040.6380.7360.7120.7560.8480.8470.8920.8540.876
F1 score0.3230.7060.7880.7870.7750.7940.8480.8990.8650.872
Accuracy0.2290.8370.8830.8790.8810.8970.9190.9450.9270.932
Table 4. Comparison of polyp detection performance between Model 1 and Model 2.
Table 4. Comparison of polyp detection performance between Model 1 and Model 2.
Model 1Model 2
Epoch 37Epoch
5
Epoch 10Epoch 15Epoch 20Epoch 25Epoch 30Epoch 35Epoch 40Epoch 45
Undetected polyp6631444011
Detection rate (%)95.0095.0097.5099.1796.6796.6796.67100.099.1799.17
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, Y.; Keum, J.-S.; Kim, J.-H.; Chun, J.; Oh, S.-I.; Kim, K.-N.; Yoon, Y.-H.; Park, H. Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor. Diagnostics 2025, 15, 901. https://doi.org/10.3390/diagnostics15070901

AMA Style

Kim Y, Keum J-S, Kim J-H, Chun J, Oh S-I, Kim K-N, Yoon Y-H, Park H. Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor. Diagnostics. 2025; 15(7):901. https://doi.org/10.3390/diagnostics15070901

Chicago/Turabian Style

Kim, Yuna, Ji-Soo Keum, Jie-Hyun Kim, Jaeyoung Chun, Sang-Il Oh, Kyung-Nam Kim, Young-Hoon Yoon, and Hyojin Park. 2025. "Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor" Diagnostics 15, no. 7: 901. https://doi.org/10.3390/diagnostics15070901

APA Style

Kim, Y., Keum, J.-S., Kim, J.-H., Chun, J., Oh, S.-I., Kim, K.-N., Yoon, Y.-H., & Park, H. (2025). Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor. Diagnostics, 15(7), 901. https://doi.org/10.3390/diagnostics15070901

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop