Next Article in Journal
The Use of Computer-Driven Technologies in the Treatment of Borderline Personality Disorder: A Systematic Review
Next Article in Special Issue
Safety of Nonagenarians Receiving Therapeutic ERCP, Single Center Experience
Previous Article in Journal
Mental Health and Social Support Are Key Predictors of Resilience in German Women with Endometriosis during the COVID-19 Pandemic
Previous Article in Special Issue
Serum Leucine-Rich Alpha-2 Glycoprotein in Quiescent Crohn’s Disease as a Potential Surrogate Marker for Small-Bowel Ulceration detected by Capsule Endoscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of a Deep-Learning Algorithm for Small Bowel-Lesion Detection and a Study of the Improvement in the False-Positive Rate

1
Center for Diagnostic and Therapeutic Endoscopy, Keio University School of Medicine, 35 Shinanomachi, Shinjuku, Tokyo 160-8582, Japan
2
Division of Gastroenterology and Hepatology, Department of Internal Medicine, Keio University School of Medicine, Tokyo 160-8582, Japan
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2022, 11(13), 3682; https://doi.org/10.3390/jcm11133682
Submission received: 2 May 2022 / Revised: 23 June 2022 / Accepted: 24 June 2022 / Published: 26 June 2022
(This article belongs to the Special Issue Current Status of Endoscopy in Clinical Medicine)

Abstract

:
Deep learning has recently been gaining attention as a promising technology to improve the identification of lesions, and deep-learning algorithms for lesion detection have been actively developed in small-bowel capsule endoscopy (SBCE). We developed a detection algorithm for abnormal findings by deep learning (convolutional neural network) the SBCE imaging data of 30 cases with abnormal findings. To enable the detection of a wide variety of abnormal findings, the training data were balanced to include all major findings identified in SBCE (bleeding, angiodysplasia, ulceration, and neoplastic lesions). To reduce the false-positive rate, “findings that may be responsible for hemorrhage” and “findings that may require therapeutic intervention” were extracted from the images of abnormal findings and added to the training dataset. For the performance evaluation, the sensitivity and the specificity were calculated using 271 detectable findings in 35 cases. The sensitivity was calculated using 68,494 images of non-abnormal findings. The sensitivity and specificity were 93.4% and 97.8%, respectively. The average number of images detected by the algorithm as having abnormal findings was 7514. We developed an image-reading support system using deep learning for SBCE and obtained a good detection performance.

1. Introduction

Small-bowel capsule endoscopy (SBCE) is an easy and non-invasive procedure that allows observation of the whole small bowel, which is difficult to achieve with conventional endoscopes. However, SBCE takes more than 50,000 images, and reading such a large number of endoscopic images imposes a large burden on physicians. To shorten the reading time and reduce the misreading of lesions, several capsule-endoscopy-reading algorithms have been developed that detect red-colored images [1,2] and reduce the number of capsule-endoscopic images by discarding similar images [3,4] or assist in detecting abnormal candidate images. Since various findings (such as vascular, inflammatory, and neoplastic lesions) are observed in the small bowel, there is a need to create an effective lesion-detection algorithm that has a minimal false-positive rate and can detect a variety of abnormalities with high sensitivity. Deep learning has recently been gaining attention as a promising technology to improve the identification of lesions, and lesion-detection algorithms using deep learning have been developed. For example, lesion-specific algorithms have been created for bleeding [5], hookworm [6], angioectasia [7,8], and mucosal breaks [9,10]. More recently, studies have shown that various lesions in the small bowel can be detected with high sensitivity by deep learning; however, there is no report of the false-positive rate [11,12,13]. Moreover, most developed algorithms use the Medtronic platform [7,8,10,11,13,14,15,16].
In the present study, we developed a lesion-detection algorithm for the Olympus platform, using deep learning in collaboration with Olympus, and evaluated its performance by assessing the sensitivity and the false-positive rate, which are important in achieving a reduction in reading time.

2. Materials and Methods

2.1. Study Design

We conducted a retrospective study of 172 cases in which SBCE was performed by the EndoCapsule 10 system (Olympus Corp., Tokyo, Japan). We assumed that the lesion-detection algorithm had a mean sensitivity of 90% and followed the t-distribution, and we set the number of cases for the validation so that the lower limit of the 95% confidence interval was more than 80%. According to this calculation, 35 cases were required for the validation dataset. In addition, we estimated that at least 30 cases would be required for the training dataset, based on our experience in the development of a lesion-detection algorithm using deep learning. Informed consent was obtained from all patients enrolled in the study.
The validation dataset and training dataset comprised images selected from the 172 included cases and were classified so that the findings were equally represented in each dataset. To avoid biases in the validation result through the data selection, each case was randomly allocated to either the training or validation dataset, by lesion category. More specifically, the cases with abnormal findings were selected from the 172 cases and were categorized into four lesion categories usually seen in SBCE (bleeding, angiodysplasia, ulcer, and other). Cases within each lesion category were then equally and randomly allocated to either the validation or training dataset. This resulted in 68 cases with abnormal findings that were classified as bleeding (n = 17), angiodysplasia (n = 19), ulcer (n = 18), and other (n = 14). Thirty-five cases were assigned to the validation dataset, and the remaining thirty-three cases were assigned to the training dataset. An approximately equal number of cases in each lesion category were allocated to the validation and training datasets. If the number of cases in a lesion category was not divisible by two, the extra case was allocated to the validation dataset (Figure 1).

2.2. Training Dataset for the Lesion-Detection Algorithm with Deep Learning

To specify the image data to be used for training, 33 cases assigned to the training dataset were re-read and abnormal findings were detected. If the lesion-detection algorithm is trained using images with a negligible level of minor findings, the sensitivity of the detection of minor abnormal findings will be improved. However, the false-positive rate would likely worsen because slight changes in the endoscopic images would also be detected, even though they were not abnormal. Therefore, to eliminate such drawbacks and minimize the false-positive rate, we defined abnormal findings as a “finding that may be a lesion responsible for bleeding” and a “finding that requires therapeutic intervention”. Subsequently, three cases in the “other” lesion category were excluded from the training dataset after re-reading.
The resultant training dataset consisted of 49,180 images with abnormal findings and 30,012 images without abnormal findings. The training dataset included 18 images of bleeding, 70 of vascular lesions, 133 of ulcerative lesions, and 35 of neoplastic/other lesions. A total of 256 findings were used for the training dataset.

2.3. Lesion-Detection Algorithm with Deep Learning

Recently, deep learning has shown a high level of performance for various image-recognition tasks. In clinical research, deep learning has spread rapidly, and the convolutional neural network (CNN) is most-used. The CNN is a very popular algorithm for image classification and typically comprises convolution layers, activation-function layers, and max-pooling layers. In the first convolution layer, each pixel value of the input images was translated to the feature maps by multiplying the filter weights and sliding the filter over the input images (Figure 2). After the first convolution layer, the subsequent convolution layers were inputted into the feature map of the previous layer instead of the input image.
In the algorithm we developed, ERFNet is used to automatically detect lesion images. The CNN of ERFnet has a novel layer that uses residual connections and factorized convolutions in order to remain highly efficient while still retaining remarkable performance. ERFNet is a deep network with hundreds of layers. The architecture and parameters such as the number of layers, the channels, and the input/output size of each layer are used without any particular modifications [17]. Using ERFNet, segmentation is performed on capsule-endoscopic images to assign a lesion label or a non-lesion label to all pixels in each image. To minimize the influence of misidentified pixels, after segmentation, if there are more than a threshold number of lesion-label pixels in the image, the image is judged as a lesion image.

2.4. Validation

The primary evaluation points were the sensitivity, specificity, and number of images detected by the CNN algorithm as having abnormal findings. We considered that both the specificity and the number of detected images (images detected by the algorithm as having abnormal findings) were significant indicators to evaluate the effect of the decreased false-positive rate in terms of reducing the burden of reading.
Regarding specificity, to confirm the effect of the decreased false-positive rate, normal images were identified from the validation dataset. The criterion for detecting normal images was defined as “images without either an abnormality or minor finding such as small red spot or small lymphoid follicles”. Specificity was calculated based on the number of images detected by the CNN algorithm as a percentage of the total number of normal images that were identified. Although SBCE cases involve many images taken throughout the gastrointestinal tract other than the small bowel, the present study focused on small-bowel evaluation. Hence, the number of images detected by the CNN algorithm was counted within the small-bowel section predefined in the validation dataset, and the detection rate was calculated by dividing the number of detected images by the average number of images taken in the small-bowel section.
In some patients with overt bleeding, blood covers most of the images taken after the origin of bleeding. We assumed that the reader would not usually observe these images in detail when reading such cases and that the number of detected images would not be an appropriate indicator of the burden of the reading. Therefore, we also evaluated the average number of detected images and the detection rate in the small bowel of only the cases without massive bleeding.
Sensitivity was calculated based on the agreement rate between the images with abnormal findings identified in the validation dataset and the detection result of the CNN algorithm. The criteria for identifying abnormal findings were the same as for the training dataset, as the detection targets of the lesion-detection algorithm were the images with abnormal findings that directly contributed to the diagnosis.

3. Results

For the images in the validation dataset, the mean patient age was 69.3 years, and there were 22 men and 13 women (Table 1). The major indication for SCBE was obscure gastrointestinal bleeding, which accounted for 57% (20 cases), followed by anemia (14%). A total of 271 findings were identified, consisting of 30 findings of bleeding, 46 of angiodysplasia, 153 of ulcers/erosions, and 42 of neoplastic lesions, including polyps and lymphoma (Table 1). The total number of images without abnormal findings in the validation dataset was 68,494.
Regarding the performance of the algorithm, the specificity was 97.8%, and the average number of images in the small-bowel section detected as having abnormal findings was 7514 per case. Among the 31 cases without massive bleeding, the average number of detected images was 3576 per case. The average-detection rate (i.e., the number of detected images divided by the average number of images taken in the small-bowel section) was 16.3% for the total 35 cases and 7.8% for the 31 cases without massive bleeding.
The sensitivity for each finding was 100% (30/30) for bleeding, 100% (46/46) for angiodysplasia, 92.7% (102/110) for ulceration, 93.0% (40/43) for erosion, 83.3% (10/12) for polyps, and 57.7% (15/26) for lymphoma (Table 2). Figure 3 shows examples of images correctly detected by the CNN algorithm (true-positive images), while Figure 4 shows images incorrectly detected as having abnormal findings (false-positive images). Representative false-negative images and the training data for the neoplastic lesions are shown in Figure 5. All polyp images missed by the current algorithm were white polyps (upper left, Peutz-Jeghers syndrome; upper middle, Peutz-Jeghers syndrome; upper right, FAP). All of the malignant-lymphoma images missed by the developed algorithm in this study were follicular lymphomas with white-lymph follicle-like bumps. In contrast, the polyp training data were red polyps with pyogenic granuloma. The tumor image in the training data used in this study was covered with bubbles and was very difficult to distinguish.
All polyp images missed by the current algorithm were white polyps (upper left, Peutz-Jeghers syndrome; upper middle; Peutz-Jeghers syndrome; upper right, FAP). All of the malignant lymphoma images missed by the developed algorithm in this study were follicular lymphomas with white-lymph follicle-like bumps (middle left, follicular lymphoma; middle center, follicular lymphoma; middle right, follicular lymphoma). In contrast, the polyp training data were red polyps with pyogenic granuloma. The tumor image in the training data used in this study was covered with bubbles and was very difficult to distinguish.

4. Discussion

We developed a CNN algorithm to detect a variety of clinically significant lesions on images obtained by SBCE. Since the training dataset included images with abnormal findings that contributed to diagnosis, the CNN algorithm developed in the current study realized a high specificity and a relatively small number of detected images.
Previous reports have showed that lesion-specific algorithms have high sensitivity and specificity for angioectasia (approximately 98%–100%) [7,8], erosion (88.2%), and ulceration (90.9%) [10]. A meta-analysis showed that the pooled sensitivity and specificity for the detection of ulceration are 95% (95% confidence interval [CI], 89–98) and 94% (95% CI, 90–96), respectively [18]. Thus, CNN algorithms show high levels of diagnostic performance for angioectasia and erosion and ulceration in SBCE. In contrast, the lesion-detection performance of a previous CNN algorithm for polyps is relatively low (sensitivity 90.7%, specificity 79.8%) [19]. Similarly, our CNN algorithm developed to detect various lesions had a relatively low sensitivity for polyps (83.3%), although the overall sensitivity was relatively high (93.4%). The CNN algorithm developed in the current study also showed a lower sensitivity for lymphoma than for the other findings. After reviewing the images of neoplastic lesions in the training dataset, we assumed that the reason that our CNN algorithm had a decreased sensitivity for polyps and lymphoma was because the training dataset did not include images with similar characteristics to the false-negative images (Figure 5). In SBCE, the prevalences of polyps and lymphoma are relatively low; however, there is a variety of abnormalities such as vascular and neoplastic lesions in the small bowel. Hence, for use in clinical practice, it is important to develop an algorithm that can effectively detect polyps and lymphoma, with increased sensitivity for detecting all small-bowel lesions. To improve the CNN algorithm developed in the present study, there is a need for additional training data including cases with polyps and lymphoma.
The present results show that high specificity can be achieved by appropriately selecting training data; however, we consider that further improvements can be made in the future by examining the false-positive images (examples of which are shown in Figure 3). The main findings incorrectly judged by the CNN algorithm (i.e., the findings seen in the false-positive images) were the assessment of white debris as an ulcer, a normal blood vessel as angiodysplasia, and a dark part of the surface depression as angiodysplasia (Figure 4). If the algorithm results in many false-positive images (e.g., normal vessels identified as abnormal findings), this may greatly increase the reading time because inexperienced physicians may need to spend a lot of time determining whether such findings are normal or abnormal. It is necessary to contrive ways to further improve the specificity of the algorithm, while maintaining the high sensitivity; this may be achieved by training the CNN algorithm with normal images rather than only adjusting the training dataset of abnormal findings, as this could cause a decrease in sensitivity.
The conventional-reading method takes around 30 to 75 min to observe all the images taken in the small bowel [3,20,21]. The CNN algorithm developed in the present study reduced the number of small-bowel images requiring reading to 7.8%, excluding cases with massive bleeding. It is not easy to estimate the reading time directly based on the number of reading images because the reading time varies depending on the experience of the reader and the diseases of the patient being examined. However, if we assume that the number of detected images is substitutable for the number of images requiring reading in the small bowel, the reading could potentially be completed within 10 min with the CNN algorithm. Two comparative studies demonstrated that the reading time of SBCE is around 3 to 5 min using a CNN algorithm in the clinical setting [12,14]. In daily practice, a reduction in reading time will enable a quick diagnosis and notification of the results to the patient, which is especially important for patients requiring urgent treatment. From this perspective, it is very important to prove the effectiveness of the CNN algorithm in the actual clinical setting. Therefore, it is necessary to demonstrate the effect of the reduced reading time by conducting further study to compare the detection rates between the conventional reading method and the reading using the CNN algorithm.

5. Conclusions

In the present study, we developed a lesion-detection algorithm with a high specificity and a reduced number of detected images, which may contribute to a shortened reading time. By conducting appropriate extra training with neoplastic lesions, it is expected that the proposed algorithm can be applied to a SBCE-reading-support system to assist in detecting images of lesions in the future.

Author Contributions

Conceptualization, N.H.; methodology, N.H.; software development, Olympus co.; validation, T.H., A.T., H.S., Y.H., and K.T.; formal analysis, N.H.; investigation, N.H. and K.J.-L.L.K.; data curation, N.H.; writing—original draft preparation, N.H.; writing—review and editing, T.S.; supervision, H.O. and T.K. All authors have read and agreed to the published version of the manuscript.

Funding

This study was conducted in collaboration with Olympus Corporation.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the ethics committee of Keio University Hospital (Approved ID 20170407).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Conflicts of Interest

This study was conducted in collaboration with Olympus Corporation. Olympus Corporation had no role in the design of the study; in the collection, analyses, or interpretation of the data; or in the writing of the manuscript.

References

  1. D’Halluin, P.N.; Delvaux, M.; Lapalus, M.G.; Sacher-Huvelin, S.; Ben Soussan, E.; Heyries, L.; Filoche, B.; Saurin, J.C.; Gay, G.; Heresbach, D. Does the “Suspected Blood Indicator” improve the detection of bleeding lesions by capsule endoscopy? Gastrointest. Endosc. 2005, 61, 243–249. [Google Scholar] [CrossRef]
  2. Ogata, H.; Kumai, K.; Imaeda, H.; Aiura, K.; Hisamatsu, T.; Okamoto, S.; Iwao, Y.; Sugino, Y.; Kitajima, M.; Hibi, T. Clinical impact of a newly developed capsule endoscope: Usefulness of a real-time image viewer for gastric transit abnormality. J. Gastroenterol. 2008, 43, 186–192. [Google Scholar] [CrossRef] [PubMed]
  3. Hosoe, N.; Watanabe, K.; Miyazaki, T.; Shimatani, M.; Wakamatsu, T.; Okazaki, K.; Esaki, M.; Matsumoto, T.; Abe, T.; Kanai, T.; et al. Evaluation of performance of the Omni mode for detecting video capsule endoscopy images: A multicenter randomized controlled trial. Endosc. Int. Open 2016, 4, 878. [Google Scholar] [CrossRef] [PubMed]
  4. Westerhof, J.; Koornstra, J.J.; Weersma, R.K. Can we reduce capsule endoscopy reading times? Gastrointest. Endosc. 2009, 69, 497–502. [Google Scholar] [CrossRef] [PubMed]
  5. Xing, X.; Jia, X.; Meng, M.Q. Bleeding Detection in Wireless Capsule Endoscopy Image Video Using Superpixel-Color Histogram and a Subspace KNN Classifier. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; Volume 2018, pp. 1–4. [Google Scholar] [CrossRef]
  6. He, J.Y.; Wu, X.; Jiang, Y.G.; Peng, Q.; Jain, R. Hookworm Detection in Wireless Capsule Endoscopy Images With Deep Learning. IEEE Trans. Image Process. 2018, 27, 2379–2392. [Google Scholar] [CrossRef] [PubMed]
  7. Leenhardt, R.; Vasseur, P.; Li, C.; Saurin, J.C.; Rahmi, G.; Cholet, F.; Becq, A.; Marteau, P.; Histace, A.; Dray, X. A neural network algorithm for detection of GI angiectasia during small-bowel capsule endoscopy. Gastrointest. Endosc. 2019, 89, 189–194. [Google Scholar] [CrossRef] [PubMed]
  8. Tsuboi, A.; Oka, S.; Aoyama, K.; Saito, H.; Aoki, T.; Yamada, A.; Matsuda, T.; Fujishiro, M.; Ishihara, S.; Nakahori, M.; et al. Artificial intelligence using a convolutional neural network for automatic detection of small-bowel angioectasia in capsule endoscopy images. Dig. Endosc. 2020, 32, 382–390. [Google Scholar] [CrossRef] [PubMed]
  9. Afonso, J.; Saraiva, M.M.; Ferreira, J.P.S.; Cardoso, H.; Ribeiro, T.; Andrade, P.; Parente, M.; Jorge, R.N.; Macedo, G. Automated detection of ulcers and erosions in capsule endoscopy images using a convolutional neural network. Med. Biol. Eng. Comput. 2022, 60, 719–725. [Google Scholar] [CrossRef] [PubMed]
  10. Aoki, T.; Yamada, A.; Aoyama, K.; Saito, H.; Tsuboi, A.; Nakada, A.; Niikura, R.; Fujishiro, M.; Oka, S.; Ishihara, S.; et al. Automatic detection of erosions and ulcerations in wireless capsule endoscopy images based on a deep convolutional neural network. Gastrointest. Endosc. 2019, 89, 357–363.e352. [Google Scholar] [CrossRef] [PubMed]
  11. Aoki, T.; Yamada, A.; Kato, Y.; Saito, H.; Tsuboi, A.; Nakada, A.; Niikura, R.; Fujishiro, M.; Oka, S.; Ishihara, S.; et al. Automatic detection of various abnormalities in capsule endoscopy videos by a deep learning-based system: A multicenter study. Gastrointest. Endosc. 2021, 93, 165–173.e161. [Google Scholar] [CrossRef] [PubMed]
  12. Ding, Z.; Shi, H.; Zhang, H.; Meng, L.; Fan, M.; Han, C.; Zhang, K.; Ming, F.; Xie, X.; Liu, H.; et al. Gastroenterologist-Level Identification of Small-Bowel Diseases and Normal Variants by Capsule Endoscopy Using a Deep-Learning Model. Gastroenterology 2019, 157, 1044–1054.e1045. [Google Scholar] [CrossRef] [PubMed]
  13. Otani, K.; Nakada, A.; Kurose, Y.; Niikura, R.; Yamada, A.; Aoki, T.; Nakanishi, H.; Doyama, H.; Hasatani, K.; Sumiyoshi, T.; et al. Automatic detection of different types of small-bowel lesions on capsule endoscopy images using a newly developed deep convolutional neural network. Endoscopy 2020, 52, 786–791. [Google Scholar] [CrossRef] [PubMed]
  14. Aoki, T.; Yamada, A.; Aoyama, K.; Saito, H.; Fujisawa, G.; Odawara, N.; Kondo, R.; Tsuboi, A.; Ishibashi, R.; Nakada, A.; et al. Clinical usefulness of a deep learning-based system as the first screening on small-bowel capsule endoscopy reading. Dig. Endosc. 2020, 32, 585–591. [Google Scholar] [CrossRef] [PubMed]
  15. Klang, E.; Barash, Y.; Margalit, R.Y.; Soffer, S.; Shimon, O.; Albshesh, A.; Ben-Horin, S.; Amitai, M.M.; Eliakim, R.; Kopylov, U. Deep learning algorithms for automated detection of Crohn’s disease ulcers by video capsule endoscopy. Gastrointest. Endosc. 2020, 91, 606–613.e602. [Google Scholar] [CrossRef] [PubMed]
  16. Park, J.; Hwang, Y.; Nam, J.H.; Oh, D.J.; Kim, K.B.; Song, H.J.; Kim, S.H.; Kang, S.H.; Jung, M.K.; Jeong Lim, Y. Artificial intelligence that determines the clinical significance of capsule endoscopy images can increase the efficiency of reading. PLoS ONE 2020, 15, e0241474. [Google Scholar] [CrossRef] [PubMed]
  17. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. Efficient ConvNet for Real-time Semantic Segmentation. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Redondo Beach, CA, USA, 11–14 June 2017; pp. 1789–1794. [Google Scholar]
  18. Soffer, S.; Klang, E.; Shimon, O.; Nachmias, N.; Eliakim, R.; Ben-Horin, S.; Kopylov, U.; Barash, Y. Deep learning for wireless capsule endoscopy: A systematic review and meta-analysis. Gastrointest. Endosc. 2020, 92, 831–839.e8. [Google Scholar] [CrossRef] [PubMed]
  19. Saito, H.; Aoki, T.; Aoyama, K.; Kato, Y.; Tsuboi, A.; Yamada, A.; Fujishiro, M.; Oka, S.; Ishihara, S.; Matsuda, T.; et al. Automatic detection and classification of protruding lesions in wireless capsule endoscopy images based on a deep convolutional neural network. Gastrointest. Endosc. 2020, 92, 144–151.e1. [Google Scholar] [CrossRef] [PubMed]
  20. Subramanian, V.; Mannath, J.; Telakis, E.; Ragunath, K.; Hawkey, C.J. Efficacy of new playback functions at reducing small-bowel wireless capsule endoscopy reading times. Dig. Dis. Sci. 2012, 57, 1624–1628. [Google Scholar] [CrossRef] [PubMed]
  21. Hosoe, N.; Rey, J.F.; Imaeda, H.; Bessho, R.; Ichikawa, R.; Ida, Y.; Naganuma, M.; Kanai, T.; Hibi, T.; Ogata, H. Evaluations of capsule endoscopy software in reducing the reading time and the rate of false negatives by inexperienced endoscopists. Clin. Res. Hepatol. Gastroenterol. 2012, 36, 66–71. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Data selection for the validation and training datasets.
Figure 1. Data selection for the validation and training datasets.
Jcm 11 03682 g001
Figure 2. Layout of the first-convolution layer. In the first-convolution layer, each pixel value of the input images (X00, X01, X02, etc.) was translated to the feature maps (Y00, Y01, Y02, etc.) by multiplying the filter weights (W00, W01, W02, etc.) and sliding the filter over the input images (e.g., Y = W00X00 + W01X01 + W02X02 + W10X10 + W11X11 + W12X12 + W20X20 + W21X21 + W22X22).
Figure 2. Layout of the first-convolution layer. In the first-convolution layer, each pixel value of the input images (X00, X01, X02, etc.) was translated to the feature maps (Y00, Y01, Y02, etc.) by multiplying the filter weights (W00, W01, W02, etc.) and sliding the filter over the input images (e.g., Y = W00X00 + W01X01 + W02X02 + W10X10 + W11X11 + W12X12 + W20X20 + W21X21 + W22X22).
Jcm 11 03682 g002
Figure 3. Representative true-positive images correctly detected by the CNN algorithm as showing (A) bleeding, (B) angiodysplasia, (C) ulceration, (D) other abnormal findings.
Figure 3. Representative true-positive images correctly detected by the CNN algorithm as showing (A) bleeding, (B) angiodysplasia, (C) ulceration, (D) other abnormal findings.
Jcm 11 03682 g003
Figure 4. Representative false-positive images detected by the CNN algorithm among the normal image group. (A) White debris mistakenly identified as an ulcer, (B) normal blood vessel mistakenly identified as angiodysplasia, (C) dark part of a surface depression mistakenly identified as angiodysplasia.
Figure 4. Representative false-positive images detected by the CNN algorithm among the normal image group. (A) White debris mistakenly identified as an ulcer, (B) normal blood vessel mistakenly identified as angiodysplasia, (C) dark part of a surface depression mistakenly identified as angiodysplasia.
Jcm 11 03682 g004
Figure 5. Representative false-negative images and training data for neoplastic lesions.
Figure 5. Representative false-negative images and training data for neoplastic lesions.
Jcm 11 03682 g005
Table 1. Characteristics of the validation dataset.
Table 1. Characteristics of the validation dataset.
Number of cases35
Number of abnormal findings271
Age, mean ± SD69.3 ± 14.3
Gender male/female22/13
Indication
OGIB20
Anemia5
Abdominal pain2
Lymphoma3
Polyp2
FAP1
Other2
Abnormal findings
Bleeding30
Angiodysplasia46
Ulcer110
Erosion43
Polyp12
Others4
Lymphoma26
OGIB; obscure gastrointestinal bleeding, FAP; familial adenomatous polyposis.
Table 2. Performance of the CNN algorithm.
Table 2. Performance of the CNN algorithm.
Sensitivity, %93.4
Bleeding100.0
Angiodysplasia100.0
Erosion93.0
Ulcer92.7
Polyp83.3
Lymphoma57.7
Others100.0
Specificity, %97.8
Number of detected images7514
Number of detected images excluding massive bleeding cases3576
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hosoe, N.; Horie, T.; Tojo, A.; Sakurai, H.; Hayashi, Y.; Limpias Kamiya, K.J.-L.; Sujino, T.; Takabayashi, K.; Ogata, H.; Kanai, T. Development of a Deep-Learning Algorithm for Small Bowel-Lesion Detection and a Study of the Improvement in the False-Positive Rate. J. Clin. Med. 2022, 11, 3682. https://doi.org/10.3390/jcm11133682

AMA Style

Hosoe N, Horie T, Tojo A, Sakurai H, Hayashi Y, Limpias Kamiya KJ-L, Sujino T, Takabayashi K, Ogata H, Kanai T. Development of a Deep-Learning Algorithm for Small Bowel-Lesion Detection and a Study of the Improvement in the False-Positive Rate. Journal of Clinical Medicine. 2022; 11(13):3682. https://doi.org/10.3390/jcm11133682

Chicago/Turabian Style

Hosoe, Naoki, Tomofumi Horie, Anna Tojo, Hinako Sakurai, Yukie Hayashi, Kenji Jose-Luis Limpias Kamiya, Tomohisa Sujino, Kaoru Takabayashi, Haruhiko Ogata, and Takanori Kanai. 2022. "Development of a Deep-Learning Algorithm for Small Bowel-Lesion Detection and a Study of the Improvement in the False-Positive Rate" Journal of Clinical Medicine 11, no. 13: 3682. https://doi.org/10.3390/jcm11133682

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop